xref: /freebsd/contrib/llvm-project/clang/lib/CodeGen/CGOpenMPRuntime.cpp (revision 0fca6ea1d4eea4c934cfff25ac9ee8ad6fe95583)
1 //===----- CGOpenMPRuntime.cpp - Interface to OpenMP Runtimes -------------===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 //
9 // This provides a class for OpenMP runtime code generation.
10 //
11 //===----------------------------------------------------------------------===//
12 
13 #include "CGOpenMPRuntime.h"
14 #include "ABIInfoImpl.h"
15 #include "CGCXXABI.h"
16 #include "CGCleanup.h"
17 #include "CGRecordLayout.h"
18 #include "CodeGenFunction.h"
19 #include "TargetInfo.h"
20 #include "clang/AST/APValue.h"
21 #include "clang/AST/Attr.h"
22 #include "clang/AST/Decl.h"
23 #include "clang/AST/OpenMPClause.h"
24 #include "clang/AST/StmtOpenMP.h"
25 #include "clang/AST/StmtVisitor.h"
26 #include "clang/Basic/BitmaskEnum.h"
27 #include "clang/Basic/FileManager.h"
28 #include "clang/Basic/OpenMPKinds.h"
29 #include "clang/Basic/SourceManager.h"
30 #include "clang/CodeGen/ConstantInitBuilder.h"
31 #include "llvm/ADT/ArrayRef.h"
32 #include "llvm/ADT/SetOperations.h"
33 #include "llvm/ADT/SmallBitVector.h"
34 #include "llvm/ADT/SmallVector.h"
35 #include "llvm/ADT/StringExtras.h"
36 #include "llvm/Bitcode/BitcodeReader.h"
37 #include "llvm/IR/Constants.h"
38 #include "llvm/IR/DerivedTypes.h"
39 #include "llvm/IR/GlobalValue.h"
40 #include "llvm/IR/InstrTypes.h"
41 #include "llvm/IR/Value.h"
42 #include "llvm/Support/AtomicOrdering.h"
43 #include "llvm/Support/Format.h"
44 #include "llvm/Support/raw_ostream.h"
45 #include <cassert>
46 #include <cstdint>
47 #include <numeric>
48 #include <optional>
49 
50 using namespace clang;
51 using namespace CodeGen;
52 using namespace llvm::omp;
53 
54 namespace {
55 /// Base class for handling code generation inside OpenMP regions.
56 class CGOpenMPRegionInfo : public CodeGenFunction::CGCapturedStmtInfo {
57 public:
58   /// Kinds of OpenMP regions used in codegen.
59   enum CGOpenMPRegionKind {
60     /// Region with outlined function for standalone 'parallel'
61     /// directive.
62     ParallelOutlinedRegion,
63     /// Region with outlined function for standalone 'task' directive.
64     TaskOutlinedRegion,
65     /// Region for constructs that do not require function outlining,
66     /// like 'for', 'sections', 'atomic' etc. directives.
67     InlinedRegion,
68     /// Region with outlined function for standalone 'target' directive.
69     TargetRegion,
70   };
71 
CGOpenMPRegionInfo(const CapturedStmt & CS,const CGOpenMPRegionKind RegionKind,const RegionCodeGenTy & CodeGen,OpenMPDirectiveKind Kind,bool HasCancel)72   CGOpenMPRegionInfo(const CapturedStmt &CS,
73                      const CGOpenMPRegionKind RegionKind,
74                      const RegionCodeGenTy &CodeGen, OpenMPDirectiveKind Kind,
75                      bool HasCancel)
76       : CGCapturedStmtInfo(CS, CR_OpenMP), RegionKind(RegionKind),
77         CodeGen(CodeGen), Kind(Kind), HasCancel(HasCancel) {}
78 
CGOpenMPRegionInfo(const CGOpenMPRegionKind RegionKind,const RegionCodeGenTy & CodeGen,OpenMPDirectiveKind Kind,bool HasCancel)79   CGOpenMPRegionInfo(const CGOpenMPRegionKind RegionKind,
80                      const RegionCodeGenTy &CodeGen, OpenMPDirectiveKind Kind,
81                      bool HasCancel)
82       : CGCapturedStmtInfo(CR_OpenMP), RegionKind(RegionKind), CodeGen(CodeGen),
83         Kind(Kind), HasCancel(HasCancel) {}
84 
85   /// Get a variable or parameter for storing global thread id
86   /// inside OpenMP construct.
87   virtual const VarDecl *getThreadIDVariable() const = 0;
88 
89   /// Emit the captured statement body.
90   void EmitBody(CodeGenFunction &CGF, const Stmt *S) override;
91 
92   /// Get an LValue for the current ThreadID variable.
93   /// \return LValue for thread id variable. This LValue always has type int32*.
94   virtual LValue getThreadIDVariableLValue(CodeGenFunction &CGF);
95 
emitUntiedSwitch(CodeGenFunction &)96   virtual void emitUntiedSwitch(CodeGenFunction & /*CGF*/) {}
97 
getRegionKind() const98   CGOpenMPRegionKind getRegionKind() const { return RegionKind; }
99 
getDirectiveKind() const100   OpenMPDirectiveKind getDirectiveKind() const { return Kind; }
101 
hasCancel() const102   bool hasCancel() const { return HasCancel; }
103 
classof(const CGCapturedStmtInfo * Info)104   static bool classof(const CGCapturedStmtInfo *Info) {
105     return Info->getKind() == CR_OpenMP;
106   }
107 
108   ~CGOpenMPRegionInfo() override = default;
109 
110 protected:
111   CGOpenMPRegionKind RegionKind;
112   RegionCodeGenTy CodeGen;
113   OpenMPDirectiveKind Kind;
114   bool HasCancel;
115 };
116 
117 /// API for captured statement code generation in OpenMP constructs.
118 class CGOpenMPOutlinedRegionInfo final : public CGOpenMPRegionInfo {
119 public:
CGOpenMPOutlinedRegionInfo(const CapturedStmt & CS,const VarDecl * ThreadIDVar,const RegionCodeGenTy & CodeGen,OpenMPDirectiveKind Kind,bool HasCancel,StringRef HelperName)120   CGOpenMPOutlinedRegionInfo(const CapturedStmt &CS, const VarDecl *ThreadIDVar,
121                              const RegionCodeGenTy &CodeGen,
122                              OpenMPDirectiveKind Kind, bool HasCancel,
123                              StringRef HelperName)
124       : CGOpenMPRegionInfo(CS, ParallelOutlinedRegion, CodeGen, Kind,
125                            HasCancel),
126         ThreadIDVar(ThreadIDVar), HelperName(HelperName) {
127     assert(ThreadIDVar != nullptr && "No ThreadID in OpenMP region.");
128   }
129 
130   /// Get a variable or parameter for storing global thread id
131   /// inside OpenMP construct.
getThreadIDVariable() const132   const VarDecl *getThreadIDVariable() const override { return ThreadIDVar; }
133 
134   /// Get the name of the capture helper.
getHelperName() const135   StringRef getHelperName() const override { return HelperName; }
136 
classof(const CGCapturedStmtInfo * Info)137   static bool classof(const CGCapturedStmtInfo *Info) {
138     return CGOpenMPRegionInfo::classof(Info) &&
139            cast<CGOpenMPRegionInfo>(Info)->getRegionKind() ==
140                ParallelOutlinedRegion;
141   }
142 
143 private:
144   /// A variable or parameter storing global thread id for OpenMP
145   /// constructs.
146   const VarDecl *ThreadIDVar;
147   StringRef HelperName;
148 };
149 
150 /// API for captured statement code generation in OpenMP constructs.
151 class CGOpenMPTaskOutlinedRegionInfo final : public CGOpenMPRegionInfo {
152 public:
153   class UntiedTaskActionTy final : public PrePostActionTy {
154     bool Untied;
155     const VarDecl *PartIDVar;
156     const RegionCodeGenTy UntiedCodeGen;
157     llvm::SwitchInst *UntiedSwitch = nullptr;
158 
159   public:
UntiedTaskActionTy(bool Tied,const VarDecl * PartIDVar,const RegionCodeGenTy & UntiedCodeGen)160     UntiedTaskActionTy(bool Tied, const VarDecl *PartIDVar,
161                        const RegionCodeGenTy &UntiedCodeGen)
162         : Untied(!Tied), PartIDVar(PartIDVar), UntiedCodeGen(UntiedCodeGen) {}
Enter(CodeGenFunction & CGF)163     void Enter(CodeGenFunction &CGF) override {
164       if (Untied) {
165         // Emit task switching point.
166         LValue PartIdLVal = CGF.EmitLoadOfPointerLValue(
167             CGF.GetAddrOfLocalVar(PartIDVar),
168             PartIDVar->getType()->castAs<PointerType>());
169         llvm::Value *Res =
170             CGF.EmitLoadOfScalar(PartIdLVal, PartIDVar->getLocation());
171         llvm::BasicBlock *DoneBB = CGF.createBasicBlock(".untied.done.");
172         UntiedSwitch = CGF.Builder.CreateSwitch(Res, DoneBB);
173         CGF.EmitBlock(DoneBB);
174         CGF.EmitBranchThroughCleanup(CGF.ReturnBlock);
175         CGF.EmitBlock(CGF.createBasicBlock(".untied.jmp."));
176         UntiedSwitch->addCase(CGF.Builder.getInt32(0),
177                               CGF.Builder.GetInsertBlock());
178         emitUntiedSwitch(CGF);
179       }
180     }
emitUntiedSwitch(CodeGenFunction & CGF) const181     void emitUntiedSwitch(CodeGenFunction &CGF) const {
182       if (Untied) {
183         LValue PartIdLVal = CGF.EmitLoadOfPointerLValue(
184             CGF.GetAddrOfLocalVar(PartIDVar),
185             PartIDVar->getType()->castAs<PointerType>());
186         CGF.EmitStoreOfScalar(CGF.Builder.getInt32(UntiedSwitch->getNumCases()),
187                               PartIdLVal);
188         UntiedCodeGen(CGF);
189         CodeGenFunction::JumpDest CurPoint =
190             CGF.getJumpDestInCurrentScope(".untied.next.");
191         CGF.EmitBranch(CGF.ReturnBlock.getBlock());
192         CGF.EmitBlock(CGF.createBasicBlock(".untied.jmp."));
193         UntiedSwitch->addCase(CGF.Builder.getInt32(UntiedSwitch->getNumCases()),
194                               CGF.Builder.GetInsertBlock());
195         CGF.EmitBranchThroughCleanup(CurPoint);
196         CGF.EmitBlock(CurPoint.getBlock());
197       }
198     }
getNumberOfParts() const199     unsigned getNumberOfParts() const { return UntiedSwitch->getNumCases(); }
200   };
CGOpenMPTaskOutlinedRegionInfo(const CapturedStmt & CS,const VarDecl * ThreadIDVar,const RegionCodeGenTy & CodeGen,OpenMPDirectiveKind Kind,bool HasCancel,const UntiedTaskActionTy & Action)201   CGOpenMPTaskOutlinedRegionInfo(const CapturedStmt &CS,
202                                  const VarDecl *ThreadIDVar,
203                                  const RegionCodeGenTy &CodeGen,
204                                  OpenMPDirectiveKind Kind, bool HasCancel,
205                                  const UntiedTaskActionTy &Action)
206       : CGOpenMPRegionInfo(CS, TaskOutlinedRegion, CodeGen, Kind, HasCancel),
207         ThreadIDVar(ThreadIDVar), Action(Action) {
208     assert(ThreadIDVar != nullptr && "No ThreadID in OpenMP region.");
209   }
210 
211   /// Get a variable or parameter for storing global thread id
212   /// inside OpenMP construct.
getThreadIDVariable() const213   const VarDecl *getThreadIDVariable() const override { return ThreadIDVar; }
214 
215   /// Get an LValue for the current ThreadID variable.
216   LValue getThreadIDVariableLValue(CodeGenFunction &CGF) override;
217 
218   /// Get the name of the capture helper.
getHelperName() const219   StringRef getHelperName() const override { return ".omp_outlined."; }
220 
emitUntiedSwitch(CodeGenFunction & CGF)221   void emitUntiedSwitch(CodeGenFunction &CGF) override {
222     Action.emitUntiedSwitch(CGF);
223   }
224 
classof(const CGCapturedStmtInfo * Info)225   static bool classof(const CGCapturedStmtInfo *Info) {
226     return CGOpenMPRegionInfo::classof(Info) &&
227            cast<CGOpenMPRegionInfo>(Info)->getRegionKind() ==
228                TaskOutlinedRegion;
229   }
230 
231 private:
232   /// A variable or parameter storing global thread id for OpenMP
233   /// constructs.
234   const VarDecl *ThreadIDVar;
235   /// Action for emitting code for untied tasks.
236   const UntiedTaskActionTy &Action;
237 };
238 
239 /// API for inlined captured statement code generation in OpenMP
240 /// constructs.
241 class CGOpenMPInlinedRegionInfo : public CGOpenMPRegionInfo {
242 public:
CGOpenMPInlinedRegionInfo(CodeGenFunction::CGCapturedStmtInfo * OldCSI,const RegionCodeGenTy & CodeGen,OpenMPDirectiveKind Kind,bool HasCancel)243   CGOpenMPInlinedRegionInfo(CodeGenFunction::CGCapturedStmtInfo *OldCSI,
244                             const RegionCodeGenTy &CodeGen,
245                             OpenMPDirectiveKind Kind, bool HasCancel)
246       : CGOpenMPRegionInfo(InlinedRegion, CodeGen, Kind, HasCancel),
247         OldCSI(OldCSI),
248         OuterRegionInfo(dyn_cast_or_null<CGOpenMPRegionInfo>(OldCSI)) {}
249 
250   // Retrieve the value of the context parameter.
getContextValue() const251   llvm::Value *getContextValue() const override {
252     if (OuterRegionInfo)
253       return OuterRegionInfo->getContextValue();
254     llvm_unreachable("No context value for inlined OpenMP region");
255   }
256 
setContextValue(llvm::Value * V)257   void setContextValue(llvm::Value *V) override {
258     if (OuterRegionInfo) {
259       OuterRegionInfo->setContextValue(V);
260       return;
261     }
262     llvm_unreachable("No context value for inlined OpenMP region");
263   }
264 
265   /// Lookup the captured field decl for a variable.
lookup(const VarDecl * VD) const266   const FieldDecl *lookup(const VarDecl *VD) const override {
267     if (OuterRegionInfo)
268       return OuterRegionInfo->lookup(VD);
269     // If there is no outer outlined region,no need to lookup in a list of
270     // captured variables, we can use the original one.
271     return nullptr;
272   }
273 
getThisFieldDecl() const274   FieldDecl *getThisFieldDecl() const override {
275     if (OuterRegionInfo)
276       return OuterRegionInfo->getThisFieldDecl();
277     return nullptr;
278   }
279 
280   /// Get a variable or parameter for storing global thread id
281   /// inside OpenMP construct.
getThreadIDVariable() const282   const VarDecl *getThreadIDVariable() const override {
283     if (OuterRegionInfo)
284       return OuterRegionInfo->getThreadIDVariable();
285     return nullptr;
286   }
287 
288   /// Get an LValue for the current ThreadID variable.
getThreadIDVariableLValue(CodeGenFunction & CGF)289   LValue getThreadIDVariableLValue(CodeGenFunction &CGF) override {
290     if (OuterRegionInfo)
291       return OuterRegionInfo->getThreadIDVariableLValue(CGF);
292     llvm_unreachable("No LValue for inlined OpenMP construct");
293   }
294 
295   /// Get the name of the capture helper.
getHelperName() const296   StringRef getHelperName() const override {
297     if (auto *OuterRegionInfo = getOldCSI())
298       return OuterRegionInfo->getHelperName();
299     llvm_unreachable("No helper name for inlined OpenMP construct");
300   }
301 
emitUntiedSwitch(CodeGenFunction & CGF)302   void emitUntiedSwitch(CodeGenFunction &CGF) override {
303     if (OuterRegionInfo)
304       OuterRegionInfo->emitUntiedSwitch(CGF);
305   }
306 
getOldCSI() const307   CodeGenFunction::CGCapturedStmtInfo *getOldCSI() const { return OldCSI; }
308 
classof(const CGCapturedStmtInfo * Info)309   static bool classof(const CGCapturedStmtInfo *Info) {
310     return CGOpenMPRegionInfo::classof(Info) &&
311            cast<CGOpenMPRegionInfo>(Info)->getRegionKind() == InlinedRegion;
312   }
313 
314   ~CGOpenMPInlinedRegionInfo() override = default;
315 
316 private:
317   /// CodeGen info about outer OpenMP region.
318   CodeGenFunction::CGCapturedStmtInfo *OldCSI;
319   CGOpenMPRegionInfo *OuterRegionInfo;
320 };
321 
322 /// API for captured statement code generation in OpenMP target
323 /// constructs. For this captures, implicit parameters are used instead of the
324 /// captured fields. The name of the target region has to be unique in a given
325 /// application so it is provided by the client, because only the client has
326 /// the information to generate that.
327 class CGOpenMPTargetRegionInfo final : public CGOpenMPRegionInfo {
328 public:
CGOpenMPTargetRegionInfo(const CapturedStmt & CS,const RegionCodeGenTy & CodeGen,StringRef HelperName)329   CGOpenMPTargetRegionInfo(const CapturedStmt &CS,
330                            const RegionCodeGenTy &CodeGen, StringRef HelperName)
331       : CGOpenMPRegionInfo(CS, TargetRegion, CodeGen, OMPD_target,
332                            /*HasCancel=*/false),
333         HelperName(HelperName) {}
334 
335   /// This is unused for target regions because each starts executing
336   /// with a single thread.
getThreadIDVariable() const337   const VarDecl *getThreadIDVariable() const override { return nullptr; }
338 
339   /// Get the name of the capture helper.
getHelperName() const340   StringRef getHelperName() const override { return HelperName; }
341 
classof(const CGCapturedStmtInfo * Info)342   static bool classof(const CGCapturedStmtInfo *Info) {
343     return CGOpenMPRegionInfo::classof(Info) &&
344            cast<CGOpenMPRegionInfo>(Info)->getRegionKind() == TargetRegion;
345   }
346 
347 private:
348   StringRef HelperName;
349 };
350 
EmptyCodeGen(CodeGenFunction &,PrePostActionTy &)351 static void EmptyCodeGen(CodeGenFunction &, PrePostActionTy &) {
352   llvm_unreachable("No codegen for expressions");
353 }
354 /// API for generation of expressions captured in a innermost OpenMP
355 /// region.
356 class CGOpenMPInnerExprInfo final : public CGOpenMPInlinedRegionInfo {
357 public:
CGOpenMPInnerExprInfo(CodeGenFunction & CGF,const CapturedStmt & CS)358   CGOpenMPInnerExprInfo(CodeGenFunction &CGF, const CapturedStmt &CS)
359       : CGOpenMPInlinedRegionInfo(CGF.CapturedStmtInfo, EmptyCodeGen,
360                                   OMPD_unknown,
361                                   /*HasCancel=*/false),
362         PrivScope(CGF) {
363     // Make sure the globals captured in the provided statement are local by
364     // using the privatization logic. We assume the same variable is not
365     // captured more than once.
366     for (const auto &C : CS.captures()) {
367       if (!C.capturesVariable() && !C.capturesVariableByCopy())
368         continue;
369 
370       const VarDecl *VD = C.getCapturedVar();
371       if (VD->isLocalVarDeclOrParm())
372         continue;
373 
374       DeclRefExpr DRE(CGF.getContext(), const_cast<VarDecl *>(VD),
375                       /*RefersToEnclosingVariableOrCapture=*/false,
376                       VD->getType().getNonReferenceType(), VK_LValue,
377                       C.getLocation());
378       PrivScope.addPrivate(VD, CGF.EmitLValue(&DRE).getAddress());
379     }
380     (void)PrivScope.Privatize();
381   }
382 
383   /// Lookup the captured field decl for a variable.
lookup(const VarDecl * VD) const384   const FieldDecl *lookup(const VarDecl *VD) const override {
385     if (const FieldDecl *FD = CGOpenMPInlinedRegionInfo::lookup(VD))
386       return FD;
387     return nullptr;
388   }
389 
390   /// Emit the captured statement body.
EmitBody(CodeGenFunction & CGF,const Stmt * S)391   void EmitBody(CodeGenFunction &CGF, const Stmt *S) override {
392     llvm_unreachable("No body for expressions");
393   }
394 
395   /// Get a variable or parameter for storing global thread id
396   /// inside OpenMP construct.
getThreadIDVariable() const397   const VarDecl *getThreadIDVariable() const override {
398     llvm_unreachable("No thread id for expressions");
399   }
400 
401   /// Get the name of the capture helper.
getHelperName() const402   StringRef getHelperName() const override {
403     llvm_unreachable("No helper name for expressions");
404   }
405 
classof(const CGCapturedStmtInfo * Info)406   static bool classof(const CGCapturedStmtInfo *Info) { return false; }
407 
408 private:
409   /// Private scope to capture global variables.
410   CodeGenFunction::OMPPrivateScope PrivScope;
411 };
412 
413 /// RAII for emitting code of OpenMP constructs.
414 class InlinedOpenMPRegionRAII {
415   CodeGenFunction &CGF;
416   llvm::DenseMap<const ValueDecl *, FieldDecl *> LambdaCaptureFields;
417   FieldDecl *LambdaThisCaptureField = nullptr;
418   const CodeGen::CGBlockInfo *BlockInfo = nullptr;
419   bool NoInheritance = false;
420 
421 public:
422   /// Constructs region for combined constructs.
423   /// \param CodeGen Code generation sequence for combined directives. Includes
424   /// a list of functions used for code generation of implicitly inlined
425   /// regions.
InlinedOpenMPRegionRAII(CodeGenFunction & CGF,const RegionCodeGenTy & CodeGen,OpenMPDirectiveKind Kind,bool HasCancel,bool NoInheritance=true)426   InlinedOpenMPRegionRAII(CodeGenFunction &CGF, const RegionCodeGenTy &CodeGen,
427                           OpenMPDirectiveKind Kind, bool HasCancel,
428                           bool NoInheritance = true)
429       : CGF(CGF), NoInheritance(NoInheritance) {
430     // Start emission for the construct.
431     CGF.CapturedStmtInfo = new CGOpenMPInlinedRegionInfo(
432         CGF.CapturedStmtInfo, CodeGen, Kind, HasCancel);
433     if (NoInheritance) {
434       std::swap(CGF.LambdaCaptureFields, LambdaCaptureFields);
435       LambdaThisCaptureField = CGF.LambdaThisCaptureField;
436       CGF.LambdaThisCaptureField = nullptr;
437       BlockInfo = CGF.BlockInfo;
438       CGF.BlockInfo = nullptr;
439     }
440   }
441 
~InlinedOpenMPRegionRAII()442   ~InlinedOpenMPRegionRAII() {
443     // Restore original CapturedStmtInfo only if we're done with code emission.
444     auto *OldCSI =
445         cast<CGOpenMPInlinedRegionInfo>(CGF.CapturedStmtInfo)->getOldCSI();
446     delete CGF.CapturedStmtInfo;
447     CGF.CapturedStmtInfo = OldCSI;
448     if (NoInheritance) {
449       std::swap(CGF.LambdaCaptureFields, LambdaCaptureFields);
450       CGF.LambdaThisCaptureField = LambdaThisCaptureField;
451       CGF.BlockInfo = BlockInfo;
452     }
453   }
454 };
455 
456 /// Values for bit flags used in the ident_t to describe the fields.
457 /// All enumeric elements are named and described in accordance with the code
458 /// from https://github.com/llvm/llvm-project/blob/main/openmp/runtime/src/kmp.h
459 enum OpenMPLocationFlags : unsigned {
460   /// Use trampoline for internal microtask.
461   OMP_IDENT_IMD = 0x01,
462   /// Use c-style ident structure.
463   OMP_IDENT_KMPC = 0x02,
464   /// Atomic reduction option for kmpc_reduce.
465   OMP_ATOMIC_REDUCE = 0x10,
466   /// Explicit 'barrier' directive.
467   OMP_IDENT_BARRIER_EXPL = 0x20,
468   /// Implicit barrier in code.
469   OMP_IDENT_BARRIER_IMPL = 0x40,
470   /// Implicit barrier in 'for' directive.
471   OMP_IDENT_BARRIER_IMPL_FOR = 0x40,
472   /// Implicit barrier in 'sections' directive.
473   OMP_IDENT_BARRIER_IMPL_SECTIONS = 0xC0,
474   /// Implicit barrier in 'single' directive.
475   OMP_IDENT_BARRIER_IMPL_SINGLE = 0x140,
476   /// Call of __kmp_for_static_init for static loop.
477   OMP_IDENT_WORK_LOOP = 0x200,
478   /// Call of __kmp_for_static_init for sections.
479   OMP_IDENT_WORK_SECTIONS = 0x400,
480   /// Call of __kmp_for_static_init for distribute.
481   OMP_IDENT_WORK_DISTRIBUTE = 0x800,
482   LLVM_MARK_AS_BITMASK_ENUM(/*LargestValue=*/OMP_IDENT_WORK_DISTRIBUTE)
483 };
484 
485 /// Describes ident structure that describes a source location.
486 /// All descriptions are taken from
487 /// https://github.com/llvm/llvm-project/blob/main/openmp/runtime/src/kmp.h
488 /// Original structure:
489 /// typedef struct ident {
490 ///    kmp_int32 reserved_1;   /**<  might be used in Fortran;
491 ///                                  see above  */
492 ///    kmp_int32 flags;        /**<  also f.flags; KMP_IDENT_xxx flags;
493 ///                                  KMP_IDENT_KMPC identifies this union
494 ///                                  member  */
495 ///    kmp_int32 reserved_2;   /**<  not really used in Fortran any more;
496 ///                                  see above */
497 ///#if USE_ITT_BUILD
498 ///                            /*  but currently used for storing
499 ///                                region-specific ITT */
500 ///                            /*  contextual information. */
501 ///#endif /* USE_ITT_BUILD */
502 ///    kmp_int32 reserved_3;   /**< source[4] in Fortran, do not use for
503 ///                                 C++  */
504 ///    char const *psource;    /**< String describing the source location.
505 ///                            The string is composed of semi-colon separated
506 //                             fields which describe the source file,
507 ///                            the function and a pair of line numbers that
508 ///                            delimit the construct.
509 ///                             */
510 /// } ident_t;
511 enum IdentFieldIndex {
512   /// might be used in Fortran
513   IdentField_Reserved_1,
514   /// OMP_IDENT_xxx flags; OMP_IDENT_KMPC identifies this union member.
515   IdentField_Flags,
516   /// Not really used in Fortran any more
517   IdentField_Reserved_2,
518   /// Source[4] in Fortran, do not use for C++
519   IdentField_Reserved_3,
520   /// String describing the source location. The string is composed of
521   /// semi-colon separated fields which describe the source file, the function
522   /// and a pair of line numbers that delimit the construct.
523   IdentField_PSource
524 };
525 
526 /// Schedule types for 'omp for' loops (these enumerators are taken from
527 /// the enum sched_type in kmp.h).
528 enum OpenMPSchedType {
529   /// Lower bound for default (unordered) versions.
530   OMP_sch_lower = 32,
531   OMP_sch_static_chunked = 33,
532   OMP_sch_static = 34,
533   OMP_sch_dynamic_chunked = 35,
534   OMP_sch_guided_chunked = 36,
535   OMP_sch_runtime = 37,
536   OMP_sch_auto = 38,
537   /// static with chunk adjustment (e.g., simd)
538   OMP_sch_static_balanced_chunked = 45,
539   /// Lower bound for 'ordered' versions.
540   OMP_ord_lower = 64,
541   OMP_ord_static_chunked = 65,
542   OMP_ord_static = 66,
543   OMP_ord_dynamic_chunked = 67,
544   OMP_ord_guided_chunked = 68,
545   OMP_ord_runtime = 69,
546   OMP_ord_auto = 70,
547   OMP_sch_default = OMP_sch_static,
548   /// dist_schedule types
549   OMP_dist_sch_static_chunked = 91,
550   OMP_dist_sch_static = 92,
551   /// Support for OpenMP 4.5 monotonic and nonmonotonic schedule modifiers.
552   /// Set if the monotonic schedule modifier was present.
553   OMP_sch_modifier_monotonic = (1 << 29),
554   /// Set if the nonmonotonic schedule modifier was present.
555   OMP_sch_modifier_nonmonotonic = (1 << 30),
556 };
557 
558 /// A basic class for pre|post-action for advanced codegen sequence for OpenMP
559 /// region.
560 class CleanupTy final : public EHScopeStack::Cleanup {
561   PrePostActionTy *Action;
562 
563 public:
CleanupTy(PrePostActionTy * Action)564   explicit CleanupTy(PrePostActionTy *Action) : Action(Action) {}
Emit(CodeGenFunction & CGF,Flags)565   void Emit(CodeGenFunction &CGF, Flags /*flags*/) override {
566     if (!CGF.HaveInsertPoint())
567       return;
568     Action->Exit(CGF);
569   }
570 };
571 
572 } // anonymous namespace
573 
operator ()(CodeGenFunction & CGF) const574 void RegionCodeGenTy::operator()(CodeGenFunction &CGF) const {
575   CodeGenFunction::RunCleanupsScope Scope(CGF);
576   if (PrePostAction) {
577     CGF.EHStack.pushCleanup<CleanupTy>(NormalAndEHCleanup, PrePostAction);
578     Callback(CodeGen, CGF, *PrePostAction);
579   } else {
580     PrePostActionTy Action;
581     Callback(CodeGen, CGF, Action);
582   }
583 }
584 
585 /// Check if the combiner is a call to UDR combiner and if it is so return the
586 /// UDR decl used for reduction.
587 static const OMPDeclareReductionDecl *
getReductionInit(const Expr * ReductionOp)588 getReductionInit(const Expr *ReductionOp) {
589   if (const auto *CE = dyn_cast<CallExpr>(ReductionOp))
590     if (const auto *OVE = dyn_cast<OpaqueValueExpr>(CE->getCallee()))
591       if (const auto *DRE =
592               dyn_cast<DeclRefExpr>(OVE->getSourceExpr()->IgnoreImpCasts()))
593         if (const auto *DRD = dyn_cast<OMPDeclareReductionDecl>(DRE->getDecl()))
594           return DRD;
595   return nullptr;
596 }
597 
emitInitWithReductionInitializer(CodeGenFunction & CGF,const OMPDeclareReductionDecl * DRD,const Expr * InitOp,Address Private,Address Original,QualType Ty)598 static void emitInitWithReductionInitializer(CodeGenFunction &CGF,
599                                              const OMPDeclareReductionDecl *DRD,
600                                              const Expr *InitOp,
601                                              Address Private, Address Original,
602                                              QualType Ty) {
603   if (DRD->getInitializer()) {
604     std::pair<llvm::Function *, llvm::Function *> Reduction =
605         CGF.CGM.getOpenMPRuntime().getUserDefinedReduction(DRD);
606     const auto *CE = cast<CallExpr>(InitOp);
607     const auto *OVE = cast<OpaqueValueExpr>(CE->getCallee());
608     const Expr *LHS = CE->getArg(/*Arg=*/0)->IgnoreParenImpCasts();
609     const Expr *RHS = CE->getArg(/*Arg=*/1)->IgnoreParenImpCasts();
610     const auto *LHSDRE =
611         cast<DeclRefExpr>(cast<UnaryOperator>(LHS)->getSubExpr());
612     const auto *RHSDRE =
613         cast<DeclRefExpr>(cast<UnaryOperator>(RHS)->getSubExpr());
614     CodeGenFunction::OMPPrivateScope PrivateScope(CGF);
615     PrivateScope.addPrivate(cast<VarDecl>(LHSDRE->getDecl()), Private);
616     PrivateScope.addPrivate(cast<VarDecl>(RHSDRE->getDecl()), Original);
617     (void)PrivateScope.Privatize();
618     RValue Func = RValue::get(Reduction.second);
619     CodeGenFunction::OpaqueValueMapping Map(CGF, OVE, Func);
620     CGF.EmitIgnoredExpr(InitOp);
621   } else {
622     llvm::Constant *Init = CGF.CGM.EmitNullConstant(Ty);
623     std::string Name = CGF.CGM.getOpenMPRuntime().getName({"init"});
624     auto *GV = new llvm::GlobalVariable(
625         CGF.CGM.getModule(), Init->getType(), /*isConstant=*/true,
626         llvm::GlobalValue::PrivateLinkage, Init, Name);
627     LValue LV = CGF.MakeNaturalAlignRawAddrLValue(GV, Ty);
628     RValue InitRVal;
629     switch (CGF.getEvaluationKind(Ty)) {
630     case TEK_Scalar:
631       InitRVal = CGF.EmitLoadOfLValue(LV, DRD->getLocation());
632       break;
633     case TEK_Complex:
634       InitRVal =
635           RValue::getComplex(CGF.EmitLoadOfComplex(LV, DRD->getLocation()));
636       break;
637     case TEK_Aggregate: {
638       OpaqueValueExpr OVE(DRD->getLocation(), Ty, VK_LValue);
639       CodeGenFunction::OpaqueValueMapping OpaqueMap(CGF, &OVE, LV);
640       CGF.EmitAnyExprToMem(&OVE, Private, Ty.getQualifiers(),
641                            /*IsInitializer=*/false);
642       return;
643     }
644     }
645     OpaqueValueExpr OVE(DRD->getLocation(), Ty, VK_PRValue);
646     CodeGenFunction::OpaqueValueMapping OpaqueMap(CGF, &OVE, InitRVal);
647     CGF.EmitAnyExprToMem(&OVE, Private, Ty.getQualifiers(),
648                          /*IsInitializer=*/false);
649   }
650 }
651 
652 /// Emit initialization of arrays of complex types.
653 /// \param DestAddr Address of the array.
654 /// \param Type Type of array.
655 /// \param Init Initial expression of array.
656 /// \param SrcAddr Address of the original array.
EmitOMPAggregateInit(CodeGenFunction & CGF,Address DestAddr,QualType Type,bool EmitDeclareReductionInit,const Expr * Init,const OMPDeclareReductionDecl * DRD,Address SrcAddr=Address::invalid ())657 static void EmitOMPAggregateInit(CodeGenFunction &CGF, Address DestAddr,
658                                  QualType Type, bool EmitDeclareReductionInit,
659                                  const Expr *Init,
660                                  const OMPDeclareReductionDecl *DRD,
661                                  Address SrcAddr = Address::invalid()) {
662   // Perform element-by-element initialization.
663   QualType ElementTy;
664 
665   // Drill down to the base element type on both arrays.
666   const ArrayType *ArrayTy = Type->getAsArrayTypeUnsafe();
667   llvm::Value *NumElements = CGF.emitArrayLength(ArrayTy, ElementTy, DestAddr);
668   if (DRD)
669     SrcAddr = SrcAddr.withElementType(DestAddr.getElementType());
670 
671   llvm::Value *SrcBegin = nullptr;
672   if (DRD)
673     SrcBegin = SrcAddr.emitRawPointer(CGF);
674   llvm::Value *DestBegin = DestAddr.emitRawPointer(CGF);
675   // Cast from pointer to array type to pointer to single element.
676   llvm::Value *DestEnd =
677       CGF.Builder.CreateGEP(DestAddr.getElementType(), DestBegin, NumElements);
678   // The basic structure here is a while-do loop.
679   llvm::BasicBlock *BodyBB = CGF.createBasicBlock("omp.arrayinit.body");
680   llvm::BasicBlock *DoneBB = CGF.createBasicBlock("omp.arrayinit.done");
681   llvm::Value *IsEmpty =
682       CGF.Builder.CreateICmpEQ(DestBegin, DestEnd, "omp.arrayinit.isempty");
683   CGF.Builder.CreateCondBr(IsEmpty, DoneBB, BodyBB);
684 
685   // Enter the loop body, making that address the current address.
686   llvm::BasicBlock *EntryBB = CGF.Builder.GetInsertBlock();
687   CGF.EmitBlock(BodyBB);
688 
689   CharUnits ElementSize = CGF.getContext().getTypeSizeInChars(ElementTy);
690 
691   llvm::PHINode *SrcElementPHI = nullptr;
692   Address SrcElementCurrent = Address::invalid();
693   if (DRD) {
694     SrcElementPHI = CGF.Builder.CreatePHI(SrcBegin->getType(), 2,
695                                           "omp.arraycpy.srcElementPast");
696     SrcElementPHI->addIncoming(SrcBegin, EntryBB);
697     SrcElementCurrent =
698         Address(SrcElementPHI, SrcAddr.getElementType(),
699                 SrcAddr.getAlignment().alignmentOfArrayElement(ElementSize));
700   }
701   llvm::PHINode *DestElementPHI = CGF.Builder.CreatePHI(
702       DestBegin->getType(), 2, "omp.arraycpy.destElementPast");
703   DestElementPHI->addIncoming(DestBegin, EntryBB);
704   Address DestElementCurrent =
705       Address(DestElementPHI, DestAddr.getElementType(),
706               DestAddr.getAlignment().alignmentOfArrayElement(ElementSize));
707 
708   // Emit copy.
709   {
710     CodeGenFunction::RunCleanupsScope InitScope(CGF);
711     if (EmitDeclareReductionInit) {
712       emitInitWithReductionInitializer(CGF, DRD, Init, DestElementCurrent,
713                                        SrcElementCurrent, ElementTy);
714     } else
715       CGF.EmitAnyExprToMem(Init, DestElementCurrent, ElementTy.getQualifiers(),
716                            /*IsInitializer=*/false);
717   }
718 
719   if (DRD) {
720     // Shift the address forward by one element.
721     llvm::Value *SrcElementNext = CGF.Builder.CreateConstGEP1_32(
722         SrcAddr.getElementType(), SrcElementPHI, /*Idx0=*/1,
723         "omp.arraycpy.dest.element");
724     SrcElementPHI->addIncoming(SrcElementNext, CGF.Builder.GetInsertBlock());
725   }
726 
727   // Shift the address forward by one element.
728   llvm::Value *DestElementNext = CGF.Builder.CreateConstGEP1_32(
729       DestAddr.getElementType(), DestElementPHI, /*Idx0=*/1,
730       "omp.arraycpy.dest.element");
731   // Check whether we've reached the end.
732   llvm::Value *Done =
733       CGF.Builder.CreateICmpEQ(DestElementNext, DestEnd, "omp.arraycpy.done");
734   CGF.Builder.CreateCondBr(Done, DoneBB, BodyBB);
735   DestElementPHI->addIncoming(DestElementNext, CGF.Builder.GetInsertBlock());
736 
737   // Done.
738   CGF.EmitBlock(DoneBB, /*IsFinished=*/true);
739 }
740 
emitSharedLValue(CodeGenFunction & CGF,const Expr * E)741 LValue ReductionCodeGen::emitSharedLValue(CodeGenFunction &CGF, const Expr *E) {
742   return CGF.EmitOMPSharedLValue(E);
743 }
744 
emitSharedLValueUB(CodeGenFunction & CGF,const Expr * E)745 LValue ReductionCodeGen::emitSharedLValueUB(CodeGenFunction &CGF,
746                                             const Expr *E) {
747   if (const auto *OASE = dyn_cast<ArraySectionExpr>(E))
748     return CGF.EmitArraySectionExpr(OASE, /*IsLowerBound=*/false);
749   return LValue();
750 }
751 
emitAggregateInitialization(CodeGenFunction & CGF,unsigned N,Address PrivateAddr,Address SharedAddr,const OMPDeclareReductionDecl * DRD)752 void ReductionCodeGen::emitAggregateInitialization(
753     CodeGenFunction &CGF, unsigned N, Address PrivateAddr, Address SharedAddr,
754     const OMPDeclareReductionDecl *DRD) {
755   // Emit VarDecl with copy init for arrays.
756   // Get the address of the original variable captured in current
757   // captured region.
758   const auto *PrivateVD =
759       cast<VarDecl>(cast<DeclRefExpr>(ClausesData[N].Private)->getDecl());
760   bool EmitDeclareReductionInit =
761       DRD && (DRD->getInitializer() || !PrivateVD->hasInit());
762   EmitOMPAggregateInit(CGF, PrivateAddr, PrivateVD->getType(),
763                        EmitDeclareReductionInit,
764                        EmitDeclareReductionInit ? ClausesData[N].ReductionOp
765                                                 : PrivateVD->getInit(),
766                        DRD, SharedAddr);
767 }
768 
ReductionCodeGen(ArrayRef<const Expr * > Shareds,ArrayRef<const Expr * > Origs,ArrayRef<const Expr * > Privates,ArrayRef<const Expr * > ReductionOps)769 ReductionCodeGen::ReductionCodeGen(ArrayRef<const Expr *> Shareds,
770                                    ArrayRef<const Expr *> Origs,
771                                    ArrayRef<const Expr *> Privates,
772                                    ArrayRef<const Expr *> ReductionOps) {
773   ClausesData.reserve(Shareds.size());
774   SharedAddresses.reserve(Shareds.size());
775   Sizes.reserve(Shareds.size());
776   BaseDecls.reserve(Shareds.size());
777   const auto *IOrig = Origs.begin();
778   const auto *IPriv = Privates.begin();
779   const auto *IRed = ReductionOps.begin();
780   for (const Expr *Ref : Shareds) {
781     ClausesData.emplace_back(Ref, *IOrig, *IPriv, *IRed);
782     std::advance(IOrig, 1);
783     std::advance(IPriv, 1);
784     std::advance(IRed, 1);
785   }
786 }
787 
emitSharedOrigLValue(CodeGenFunction & CGF,unsigned N)788 void ReductionCodeGen::emitSharedOrigLValue(CodeGenFunction &CGF, unsigned N) {
789   assert(SharedAddresses.size() == N && OrigAddresses.size() == N &&
790          "Number of generated lvalues must be exactly N.");
791   LValue First = emitSharedLValue(CGF, ClausesData[N].Shared);
792   LValue Second = emitSharedLValueUB(CGF, ClausesData[N].Shared);
793   SharedAddresses.emplace_back(First, Second);
794   if (ClausesData[N].Shared == ClausesData[N].Ref) {
795     OrigAddresses.emplace_back(First, Second);
796   } else {
797     LValue First = emitSharedLValue(CGF, ClausesData[N].Ref);
798     LValue Second = emitSharedLValueUB(CGF, ClausesData[N].Ref);
799     OrigAddresses.emplace_back(First, Second);
800   }
801 }
802 
emitAggregateType(CodeGenFunction & CGF,unsigned N)803 void ReductionCodeGen::emitAggregateType(CodeGenFunction &CGF, unsigned N) {
804   QualType PrivateType = getPrivateType(N);
805   bool AsArraySection = isa<ArraySectionExpr>(ClausesData[N].Ref);
806   if (!PrivateType->isVariablyModifiedType()) {
807     Sizes.emplace_back(
808         CGF.getTypeSize(OrigAddresses[N].first.getType().getNonReferenceType()),
809         nullptr);
810     return;
811   }
812   llvm::Value *Size;
813   llvm::Value *SizeInChars;
814   auto *ElemType = OrigAddresses[N].first.getAddress().getElementType();
815   auto *ElemSizeOf = llvm::ConstantExpr::getSizeOf(ElemType);
816   if (AsArraySection) {
817     Size = CGF.Builder.CreatePtrDiff(ElemType,
818                                      OrigAddresses[N].second.getPointer(CGF),
819                                      OrigAddresses[N].first.getPointer(CGF));
820     Size = CGF.Builder.CreateNUWAdd(
821         Size, llvm::ConstantInt::get(Size->getType(), /*V=*/1));
822     SizeInChars = CGF.Builder.CreateNUWMul(Size, ElemSizeOf);
823   } else {
824     SizeInChars =
825         CGF.getTypeSize(OrigAddresses[N].first.getType().getNonReferenceType());
826     Size = CGF.Builder.CreateExactUDiv(SizeInChars, ElemSizeOf);
827   }
828   Sizes.emplace_back(SizeInChars, Size);
829   CodeGenFunction::OpaqueValueMapping OpaqueMap(
830       CGF,
831       cast<OpaqueValueExpr>(
832           CGF.getContext().getAsVariableArrayType(PrivateType)->getSizeExpr()),
833       RValue::get(Size));
834   CGF.EmitVariablyModifiedType(PrivateType);
835 }
836 
emitAggregateType(CodeGenFunction & CGF,unsigned N,llvm::Value * Size)837 void ReductionCodeGen::emitAggregateType(CodeGenFunction &CGF, unsigned N,
838                                          llvm::Value *Size) {
839   QualType PrivateType = getPrivateType(N);
840   if (!PrivateType->isVariablyModifiedType()) {
841     assert(!Size && !Sizes[N].second &&
842            "Size should be nullptr for non-variably modified reduction "
843            "items.");
844     return;
845   }
846   CodeGenFunction::OpaqueValueMapping OpaqueMap(
847       CGF,
848       cast<OpaqueValueExpr>(
849           CGF.getContext().getAsVariableArrayType(PrivateType)->getSizeExpr()),
850       RValue::get(Size));
851   CGF.EmitVariablyModifiedType(PrivateType);
852 }
853 
emitInitialization(CodeGenFunction & CGF,unsigned N,Address PrivateAddr,Address SharedAddr,llvm::function_ref<bool (CodeGenFunction &)> DefaultInit)854 void ReductionCodeGen::emitInitialization(
855     CodeGenFunction &CGF, unsigned N, Address PrivateAddr, Address SharedAddr,
856     llvm::function_ref<bool(CodeGenFunction &)> DefaultInit) {
857   assert(SharedAddresses.size() > N && "No variable was generated");
858   const auto *PrivateVD =
859       cast<VarDecl>(cast<DeclRefExpr>(ClausesData[N].Private)->getDecl());
860   const OMPDeclareReductionDecl *DRD =
861       getReductionInit(ClausesData[N].ReductionOp);
862   if (CGF.getContext().getAsArrayType(PrivateVD->getType())) {
863     if (DRD && DRD->getInitializer())
864       (void)DefaultInit(CGF);
865     emitAggregateInitialization(CGF, N, PrivateAddr, SharedAddr, DRD);
866   } else if (DRD && (DRD->getInitializer() || !PrivateVD->hasInit())) {
867     (void)DefaultInit(CGF);
868     QualType SharedType = SharedAddresses[N].first.getType();
869     emitInitWithReductionInitializer(CGF, DRD, ClausesData[N].ReductionOp,
870                                      PrivateAddr, SharedAddr, SharedType);
871   } else if (!DefaultInit(CGF) && PrivateVD->hasInit() &&
872              !CGF.isTrivialInitializer(PrivateVD->getInit())) {
873     CGF.EmitAnyExprToMem(PrivateVD->getInit(), PrivateAddr,
874                          PrivateVD->getType().getQualifiers(),
875                          /*IsInitializer=*/false);
876   }
877 }
878 
needCleanups(unsigned N)879 bool ReductionCodeGen::needCleanups(unsigned N) {
880   QualType PrivateType = getPrivateType(N);
881   QualType::DestructionKind DTorKind = PrivateType.isDestructedType();
882   return DTorKind != QualType::DK_none;
883 }
884 
emitCleanups(CodeGenFunction & CGF,unsigned N,Address PrivateAddr)885 void ReductionCodeGen::emitCleanups(CodeGenFunction &CGF, unsigned N,
886                                     Address PrivateAddr) {
887   QualType PrivateType = getPrivateType(N);
888   QualType::DestructionKind DTorKind = PrivateType.isDestructedType();
889   if (needCleanups(N)) {
890     PrivateAddr =
891         PrivateAddr.withElementType(CGF.ConvertTypeForMem(PrivateType));
892     CGF.pushDestroy(DTorKind, PrivateAddr, PrivateType);
893   }
894 }
895 
loadToBegin(CodeGenFunction & CGF,QualType BaseTy,QualType ElTy,LValue BaseLV)896 static LValue loadToBegin(CodeGenFunction &CGF, QualType BaseTy, QualType ElTy,
897                           LValue BaseLV) {
898   BaseTy = BaseTy.getNonReferenceType();
899   while ((BaseTy->isPointerType() || BaseTy->isReferenceType()) &&
900          !CGF.getContext().hasSameType(BaseTy, ElTy)) {
901     if (const auto *PtrTy = BaseTy->getAs<PointerType>()) {
902       BaseLV = CGF.EmitLoadOfPointerLValue(BaseLV.getAddress(), PtrTy);
903     } else {
904       LValue RefLVal = CGF.MakeAddrLValue(BaseLV.getAddress(), BaseTy);
905       BaseLV = CGF.EmitLoadOfReferenceLValue(RefLVal);
906     }
907     BaseTy = BaseTy->getPointeeType();
908   }
909   return CGF.MakeAddrLValue(
910       BaseLV.getAddress().withElementType(CGF.ConvertTypeForMem(ElTy)),
911       BaseLV.getType(), BaseLV.getBaseInfo(),
912       CGF.CGM.getTBAAInfoForSubobject(BaseLV, BaseLV.getType()));
913 }
914 
castToBase(CodeGenFunction & CGF,QualType BaseTy,QualType ElTy,Address OriginalBaseAddress,llvm::Value * Addr)915 static Address castToBase(CodeGenFunction &CGF, QualType BaseTy, QualType ElTy,
916                           Address OriginalBaseAddress, llvm::Value *Addr) {
917   RawAddress Tmp = RawAddress::invalid();
918   Address TopTmp = Address::invalid();
919   Address MostTopTmp = Address::invalid();
920   BaseTy = BaseTy.getNonReferenceType();
921   while ((BaseTy->isPointerType() || BaseTy->isReferenceType()) &&
922          !CGF.getContext().hasSameType(BaseTy, ElTy)) {
923     Tmp = CGF.CreateMemTemp(BaseTy);
924     if (TopTmp.isValid())
925       CGF.Builder.CreateStore(Tmp.getPointer(), TopTmp);
926     else
927       MostTopTmp = Tmp;
928     TopTmp = Tmp;
929     BaseTy = BaseTy->getPointeeType();
930   }
931 
932   if (Tmp.isValid()) {
933     Addr = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
934         Addr, Tmp.getElementType());
935     CGF.Builder.CreateStore(Addr, Tmp);
936     return MostTopTmp;
937   }
938 
939   Addr = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
940       Addr, OriginalBaseAddress.getType());
941   return OriginalBaseAddress.withPointer(Addr, NotKnownNonNull);
942 }
943 
getBaseDecl(const Expr * Ref,const DeclRefExpr * & DE)944 static const VarDecl *getBaseDecl(const Expr *Ref, const DeclRefExpr *&DE) {
945   const VarDecl *OrigVD = nullptr;
946   if (const auto *OASE = dyn_cast<ArraySectionExpr>(Ref)) {
947     const Expr *Base = OASE->getBase()->IgnoreParenImpCasts();
948     while (const auto *TempOASE = dyn_cast<ArraySectionExpr>(Base))
949       Base = TempOASE->getBase()->IgnoreParenImpCasts();
950     while (const auto *TempASE = dyn_cast<ArraySubscriptExpr>(Base))
951       Base = TempASE->getBase()->IgnoreParenImpCasts();
952     DE = cast<DeclRefExpr>(Base);
953     OrigVD = cast<VarDecl>(DE->getDecl());
954   } else if (const auto *ASE = dyn_cast<ArraySubscriptExpr>(Ref)) {
955     const Expr *Base = ASE->getBase()->IgnoreParenImpCasts();
956     while (const auto *TempASE = dyn_cast<ArraySubscriptExpr>(Base))
957       Base = TempASE->getBase()->IgnoreParenImpCasts();
958     DE = cast<DeclRefExpr>(Base);
959     OrigVD = cast<VarDecl>(DE->getDecl());
960   }
961   return OrigVD;
962 }
963 
adjustPrivateAddress(CodeGenFunction & CGF,unsigned N,Address PrivateAddr)964 Address ReductionCodeGen::adjustPrivateAddress(CodeGenFunction &CGF, unsigned N,
965                                                Address PrivateAddr) {
966   const DeclRefExpr *DE;
967   if (const VarDecl *OrigVD = ::getBaseDecl(ClausesData[N].Ref, DE)) {
968     BaseDecls.emplace_back(OrigVD);
969     LValue OriginalBaseLValue = CGF.EmitLValue(DE);
970     LValue BaseLValue =
971         loadToBegin(CGF, OrigVD->getType(), SharedAddresses[N].first.getType(),
972                     OriginalBaseLValue);
973     Address SharedAddr = SharedAddresses[N].first.getAddress();
974     llvm::Value *Adjustment = CGF.Builder.CreatePtrDiff(
975         SharedAddr.getElementType(), BaseLValue.getPointer(CGF),
976         SharedAddr.emitRawPointer(CGF));
977     llvm::Value *PrivatePointer =
978         CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
979             PrivateAddr.emitRawPointer(CGF), SharedAddr.getType());
980     llvm::Value *Ptr = CGF.Builder.CreateGEP(
981         SharedAddr.getElementType(), PrivatePointer, Adjustment);
982     return castToBase(CGF, OrigVD->getType(),
983                       SharedAddresses[N].first.getType(),
984                       OriginalBaseLValue.getAddress(), Ptr);
985   }
986   BaseDecls.emplace_back(
987       cast<VarDecl>(cast<DeclRefExpr>(ClausesData[N].Ref)->getDecl()));
988   return PrivateAddr;
989 }
990 
usesReductionInitializer(unsigned N) const991 bool ReductionCodeGen::usesReductionInitializer(unsigned N) const {
992   const OMPDeclareReductionDecl *DRD =
993       getReductionInit(ClausesData[N].ReductionOp);
994   return DRD && DRD->getInitializer();
995 }
996 
getThreadIDVariableLValue(CodeGenFunction & CGF)997 LValue CGOpenMPRegionInfo::getThreadIDVariableLValue(CodeGenFunction &CGF) {
998   return CGF.EmitLoadOfPointerLValue(
999       CGF.GetAddrOfLocalVar(getThreadIDVariable()),
1000       getThreadIDVariable()->getType()->castAs<PointerType>());
1001 }
1002 
EmitBody(CodeGenFunction & CGF,const Stmt * S)1003 void CGOpenMPRegionInfo::EmitBody(CodeGenFunction &CGF, const Stmt *S) {
1004   if (!CGF.HaveInsertPoint())
1005     return;
1006   // 1.2.2 OpenMP Language Terminology
1007   // Structured block - An executable statement with a single entry at the
1008   // top and a single exit at the bottom.
1009   // The point of exit cannot be a branch out of the structured block.
1010   // longjmp() and throw() must not violate the entry/exit criteria.
1011   CGF.EHStack.pushTerminate();
1012   if (S)
1013     CGF.incrementProfileCounter(S);
1014   CodeGen(CGF);
1015   CGF.EHStack.popTerminate();
1016 }
1017 
getThreadIDVariableLValue(CodeGenFunction & CGF)1018 LValue CGOpenMPTaskOutlinedRegionInfo::getThreadIDVariableLValue(
1019     CodeGenFunction &CGF) {
1020   return CGF.MakeAddrLValue(CGF.GetAddrOfLocalVar(getThreadIDVariable()),
1021                             getThreadIDVariable()->getType(),
1022                             AlignmentSource::Decl);
1023 }
1024 
addFieldToRecordDecl(ASTContext & C,DeclContext * DC,QualType FieldTy)1025 static FieldDecl *addFieldToRecordDecl(ASTContext &C, DeclContext *DC,
1026                                        QualType FieldTy) {
1027   auto *Field = FieldDecl::Create(
1028       C, DC, SourceLocation(), SourceLocation(), /*Id=*/nullptr, FieldTy,
1029       C.getTrivialTypeSourceInfo(FieldTy, SourceLocation()),
1030       /*BW=*/nullptr, /*Mutable=*/false, /*InitStyle=*/ICIS_NoInit);
1031   Field->setAccess(AS_public);
1032   DC->addDecl(Field);
1033   return Field;
1034 }
1035 
CGOpenMPRuntime(CodeGenModule & CGM)1036 CGOpenMPRuntime::CGOpenMPRuntime(CodeGenModule &CGM)
1037     : CGM(CGM), OMPBuilder(CGM.getModule()) {
1038   KmpCriticalNameTy = llvm::ArrayType::get(CGM.Int32Ty, /*NumElements*/ 8);
1039   llvm::OpenMPIRBuilderConfig Config(
1040       CGM.getLangOpts().OpenMPIsTargetDevice, isGPU(),
1041       CGM.getLangOpts().OpenMPOffloadMandatory,
1042       /*HasRequiresReverseOffload*/ false, /*HasRequiresUnifiedAddress*/ false,
1043       hasRequiresUnifiedSharedMemory(), /*HasRequiresDynamicAllocators*/ false);
1044   OMPBuilder.initialize();
1045   OMPBuilder.loadOffloadInfoMetadata(CGM.getLangOpts().OpenMPIsTargetDevice
1046                                          ? CGM.getLangOpts().OMPHostIRFile
1047                                          : StringRef{});
1048   OMPBuilder.setConfig(Config);
1049 
1050   // The user forces the compiler to behave as if omp requires
1051   // unified_shared_memory was given.
1052   if (CGM.getLangOpts().OpenMPForceUSM) {
1053     HasRequiresUnifiedSharedMemory = true;
1054     OMPBuilder.Config.setHasRequiresUnifiedSharedMemory(true);
1055   }
1056 }
1057 
clear()1058 void CGOpenMPRuntime::clear() {
1059   InternalVars.clear();
1060   // Clean non-target variable declarations possibly used only in debug info.
1061   for (const auto &Data : EmittedNonTargetVariables) {
1062     if (!Data.getValue().pointsToAliveValue())
1063       continue;
1064     auto *GV = dyn_cast<llvm::GlobalVariable>(Data.getValue());
1065     if (!GV)
1066       continue;
1067     if (!GV->isDeclaration() || GV->getNumUses() > 0)
1068       continue;
1069     GV->eraseFromParent();
1070   }
1071 }
1072 
getName(ArrayRef<StringRef> Parts) const1073 std::string CGOpenMPRuntime::getName(ArrayRef<StringRef> Parts) const {
1074   return OMPBuilder.createPlatformSpecificName(Parts);
1075 }
1076 
1077 static llvm::Function *
emitCombinerOrInitializer(CodeGenModule & CGM,QualType Ty,const Expr * CombinerInitializer,const VarDecl * In,const VarDecl * Out,bool IsCombiner)1078 emitCombinerOrInitializer(CodeGenModule &CGM, QualType Ty,
1079                           const Expr *CombinerInitializer, const VarDecl *In,
1080                           const VarDecl *Out, bool IsCombiner) {
1081   // void .omp_combiner.(Ty *in, Ty *out);
1082   ASTContext &C = CGM.getContext();
1083   QualType PtrTy = C.getPointerType(Ty).withRestrict();
1084   FunctionArgList Args;
1085   ImplicitParamDecl OmpOutParm(C, /*DC=*/nullptr, Out->getLocation(),
1086                                /*Id=*/nullptr, PtrTy, ImplicitParamKind::Other);
1087   ImplicitParamDecl OmpInParm(C, /*DC=*/nullptr, In->getLocation(),
1088                               /*Id=*/nullptr, PtrTy, ImplicitParamKind::Other);
1089   Args.push_back(&OmpOutParm);
1090   Args.push_back(&OmpInParm);
1091   const CGFunctionInfo &FnInfo =
1092       CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args);
1093   llvm::FunctionType *FnTy = CGM.getTypes().GetFunctionType(FnInfo);
1094   std::string Name = CGM.getOpenMPRuntime().getName(
1095       {IsCombiner ? "omp_combiner" : "omp_initializer", ""});
1096   auto *Fn = llvm::Function::Create(FnTy, llvm::GlobalValue::InternalLinkage,
1097                                     Name, &CGM.getModule());
1098   CGM.SetInternalFunctionAttributes(GlobalDecl(), Fn, FnInfo);
1099   if (CGM.getLangOpts().Optimize) {
1100     Fn->removeFnAttr(llvm::Attribute::NoInline);
1101     Fn->removeFnAttr(llvm::Attribute::OptimizeNone);
1102     Fn->addFnAttr(llvm::Attribute::AlwaysInline);
1103   }
1104   CodeGenFunction CGF(CGM);
1105   // Map "T omp_in;" variable to "*omp_in_parm" value in all expressions.
1106   // Map "T omp_out;" variable to "*omp_out_parm" value in all expressions.
1107   CGF.StartFunction(GlobalDecl(), C.VoidTy, Fn, FnInfo, Args, In->getLocation(),
1108                     Out->getLocation());
1109   CodeGenFunction::OMPPrivateScope Scope(CGF);
1110   Address AddrIn = CGF.GetAddrOfLocalVar(&OmpInParm);
1111   Scope.addPrivate(
1112       In, CGF.EmitLoadOfPointerLValue(AddrIn, PtrTy->castAs<PointerType>())
1113               .getAddress());
1114   Address AddrOut = CGF.GetAddrOfLocalVar(&OmpOutParm);
1115   Scope.addPrivate(
1116       Out, CGF.EmitLoadOfPointerLValue(AddrOut, PtrTy->castAs<PointerType>())
1117                .getAddress());
1118   (void)Scope.Privatize();
1119   if (!IsCombiner && Out->hasInit() &&
1120       !CGF.isTrivialInitializer(Out->getInit())) {
1121     CGF.EmitAnyExprToMem(Out->getInit(), CGF.GetAddrOfLocalVar(Out),
1122                          Out->getType().getQualifiers(),
1123                          /*IsInitializer=*/true);
1124   }
1125   if (CombinerInitializer)
1126     CGF.EmitIgnoredExpr(CombinerInitializer);
1127   Scope.ForceCleanup();
1128   CGF.FinishFunction();
1129   return Fn;
1130 }
1131 
emitUserDefinedReduction(CodeGenFunction * CGF,const OMPDeclareReductionDecl * D)1132 void CGOpenMPRuntime::emitUserDefinedReduction(
1133     CodeGenFunction *CGF, const OMPDeclareReductionDecl *D) {
1134   if (UDRMap.count(D) > 0)
1135     return;
1136   llvm::Function *Combiner = emitCombinerOrInitializer(
1137       CGM, D->getType(), D->getCombiner(),
1138       cast<VarDecl>(cast<DeclRefExpr>(D->getCombinerIn())->getDecl()),
1139       cast<VarDecl>(cast<DeclRefExpr>(D->getCombinerOut())->getDecl()),
1140       /*IsCombiner=*/true);
1141   llvm::Function *Initializer = nullptr;
1142   if (const Expr *Init = D->getInitializer()) {
1143     Initializer = emitCombinerOrInitializer(
1144         CGM, D->getType(),
1145         D->getInitializerKind() == OMPDeclareReductionInitKind::Call ? Init
1146                                                                      : nullptr,
1147         cast<VarDecl>(cast<DeclRefExpr>(D->getInitOrig())->getDecl()),
1148         cast<VarDecl>(cast<DeclRefExpr>(D->getInitPriv())->getDecl()),
1149         /*IsCombiner=*/false);
1150   }
1151   UDRMap.try_emplace(D, Combiner, Initializer);
1152   if (CGF) {
1153     auto &Decls = FunctionUDRMap.FindAndConstruct(CGF->CurFn);
1154     Decls.second.push_back(D);
1155   }
1156 }
1157 
1158 std::pair<llvm::Function *, llvm::Function *>
getUserDefinedReduction(const OMPDeclareReductionDecl * D)1159 CGOpenMPRuntime::getUserDefinedReduction(const OMPDeclareReductionDecl *D) {
1160   auto I = UDRMap.find(D);
1161   if (I != UDRMap.end())
1162     return I->second;
1163   emitUserDefinedReduction(/*CGF=*/nullptr, D);
1164   return UDRMap.lookup(D);
1165 }
1166 
1167 namespace {
1168 // Temporary RAII solution to perform a push/pop stack event on the OpenMP IR
1169 // Builder if one is present.
1170 struct PushAndPopStackRAII {
PushAndPopStackRAII__anon93cce0fb0211::PushAndPopStackRAII1171   PushAndPopStackRAII(llvm::OpenMPIRBuilder *OMPBuilder, CodeGenFunction &CGF,
1172                       bool HasCancel, llvm::omp::Directive Kind)
1173       : OMPBuilder(OMPBuilder) {
1174     if (!OMPBuilder)
1175       return;
1176 
1177     // The following callback is the crucial part of clangs cleanup process.
1178     //
1179     // NOTE:
1180     // Once the OpenMPIRBuilder is used to create parallel regions (and
1181     // similar), the cancellation destination (Dest below) is determined via
1182     // IP. That means if we have variables to finalize we split the block at IP,
1183     // use the new block (=BB) as destination to build a JumpDest (via
1184     // getJumpDestInCurrentScope(BB)) which then is fed to
1185     // EmitBranchThroughCleanup. Furthermore, there will not be the need
1186     // to push & pop an FinalizationInfo object.
1187     // The FiniCB will still be needed but at the point where the
1188     // OpenMPIRBuilder is asked to construct a parallel (or similar) construct.
1189     auto FiniCB = [&CGF](llvm::OpenMPIRBuilder::InsertPointTy IP) {
1190       assert(IP.getBlock()->end() == IP.getPoint() &&
1191              "Clang CG should cause non-terminated block!");
1192       CGBuilderTy::InsertPointGuard IPG(CGF.Builder);
1193       CGF.Builder.restoreIP(IP);
1194       CodeGenFunction::JumpDest Dest =
1195           CGF.getOMPCancelDestination(OMPD_parallel);
1196       CGF.EmitBranchThroughCleanup(Dest);
1197     };
1198 
1199     // TODO: Remove this once we emit parallel regions through the
1200     //       OpenMPIRBuilder as it can do this setup internally.
1201     llvm::OpenMPIRBuilder::FinalizationInfo FI({FiniCB, Kind, HasCancel});
1202     OMPBuilder->pushFinalizationCB(std::move(FI));
1203   }
~PushAndPopStackRAII__anon93cce0fb0211::PushAndPopStackRAII1204   ~PushAndPopStackRAII() {
1205     if (OMPBuilder)
1206       OMPBuilder->popFinalizationCB();
1207   }
1208   llvm::OpenMPIRBuilder *OMPBuilder;
1209 };
1210 } // namespace
1211 
emitParallelOrTeamsOutlinedFunction(CodeGenModule & CGM,const OMPExecutableDirective & D,const CapturedStmt * CS,const VarDecl * ThreadIDVar,OpenMPDirectiveKind InnermostKind,const StringRef OutlinedHelperName,const RegionCodeGenTy & CodeGen)1212 static llvm::Function *emitParallelOrTeamsOutlinedFunction(
1213     CodeGenModule &CGM, const OMPExecutableDirective &D, const CapturedStmt *CS,
1214     const VarDecl *ThreadIDVar, OpenMPDirectiveKind InnermostKind,
1215     const StringRef OutlinedHelperName, const RegionCodeGenTy &CodeGen) {
1216   assert(ThreadIDVar->getType()->isPointerType() &&
1217          "thread id variable must be of type kmp_int32 *");
1218   CodeGenFunction CGF(CGM, true);
1219   bool HasCancel = false;
1220   if (const auto *OPD = dyn_cast<OMPParallelDirective>(&D))
1221     HasCancel = OPD->hasCancel();
1222   else if (const auto *OPD = dyn_cast<OMPTargetParallelDirective>(&D))
1223     HasCancel = OPD->hasCancel();
1224   else if (const auto *OPSD = dyn_cast<OMPParallelSectionsDirective>(&D))
1225     HasCancel = OPSD->hasCancel();
1226   else if (const auto *OPFD = dyn_cast<OMPParallelForDirective>(&D))
1227     HasCancel = OPFD->hasCancel();
1228   else if (const auto *OPFD = dyn_cast<OMPTargetParallelForDirective>(&D))
1229     HasCancel = OPFD->hasCancel();
1230   else if (const auto *OPFD = dyn_cast<OMPDistributeParallelForDirective>(&D))
1231     HasCancel = OPFD->hasCancel();
1232   else if (const auto *OPFD =
1233                dyn_cast<OMPTeamsDistributeParallelForDirective>(&D))
1234     HasCancel = OPFD->hasCancel();
1235   else if (const auto *OPFD =
1236                dyn_cast<OMPTargetTeamsDistributeParallelForDirective>(&D))
1237     HasCancel = OPFD->hasCancel();
1238 
1239   // TODO: Temporarily inform the OpenMPIRBuilder, if any, about the new
1240   //       parallel region to make cancellation barriers work properly.
1241   llvm::OpenMPIRBuilder &OMPBuilder = CGM.getOpenMPRuntime().getOMPBuilder();
1242   PushAndPopStackRAII PSR(&OMPBuilder, CGF, HasCancel, InnermostKind);
1243   CGOpenMPOutlinedRegionInfo CGInfo(*CS, ThreadIDVar, CodeGen, InnermostKind,
1244                                     HasCancel, OutlinedHelperName);
1245   CodeGenFunction::CGCapturedStmtRAII CapInfoRAII(CGF, &CGInfo);
1246   return CGF.GenerateOpenMPCapturedStmtFunction(*CS, D.getBeginLoc());
1247 }
1248 
getOutlinedHelperName(StringRef Name) const1249 std::string CGOpenMPRuntime::getOutlinedHelperName(StringRef Name) const {
1250   std::string Suffix = getName({"omp_outlined"});
1251   return (Name + Suffix).str();
1252 }
1253 
getOutlinedHelperName(CodeGenFunction & CGF) const1254 std::string CGOpenMPRuntime::getOutlinedHelperName(CodeGenFunction &CGF) const {
1255   return getOutlinedHelperName(CGF.CurFn->getName());
1256 }
1257 
getReductionFuncName(StringRef Name) const1258 std::string CGOpenMPRuntime::getReductionFuncName(StringRef Name) const {
1259   std::string Suffix = getName({"omp", "reduction", "reduction_func"});
1260   return (Name + Suffix).str();
1261 }
1262 
emitParallelOutlinedFunction(CodeGenFunction & CGF,const OMPExecutableDirective & D,const VarDecl * ThreadIDVar,OpenMPDirectiveKind InnermostKind,const RegionCodeGenTy & CodeGen)1263 llvm::Function *CGOpenMPRuntime::emitParallelOutlinedFunction(
1264     CodeGenFunction &CGF, const OMPExecutableDirective &D,
1265     const VarDecl *ThreadIDVar, OpenMPDirectiveKind InnermostKind,
1266     const RegionCodeGenTy &CodeGen) {
1267   const CapturedStmt *CS = D.getCapturedStmt(OMPD_parallel);
1268   return emitParallelOrTeamsOutlinedFunction(
1269       CGM, D, CS, ThreadIDVar, InnermostKind, getOutlinedHelperName(CGF),
1270       CodeGen);
1271 }
1272 
emitTeamsOutlinedFunction(CodeGenFunction & CGF,const OMPExecutableDirective & D,const VarDecl * ThreadIDVar,OpenMPDirectiveKind InnermostKind,const RegionCodeGenTy & CodeGen)1273 llvm::Function *CGOpenMPRuntime::emitTeamsOutlinedFunction(
1274     CodeGenFunction &CGF, const OMPExecutableDirective &D,
1275     const VarDecl *ThreadIDVar, OpenMPDirectiveKind InnermostKind,
1276     const RegionCodeGenTy &CodeGen) {
1277   const CapturedStmt *CS = D.getCapturedStmt(OMPD_teams);
1278   return emitParallelOrTeamsOutlinedFunction(
1279       CGM, D, CS, ThreadIDVar, InnermostKind, getOutlinedHelperName(CGF),
1280       CodeGen);
1281 }
1282 
emitTaskOutlinedFunction(const OMPExecutableDirective & D,const VarDecl * ThreadIDVar,const VarDecl * PartIDVar,const VarDecl * TaskTVar,OpenMPDirectiveKind InnermostKind,const RegionCodeGenTy & CodeGen,bool Tied,unsigned & NumberOfParts)1283 llvm::Function *CGOpenMPRuntime::emitTaskOutlinedFunction(
1284     const OMPExecutableDirective &D, const VarDecl *ThreadIDVar,
1285     const VarDecl *PartIDVar, const VarDecl *TaskTVar,
1286     OpenMPDirectiveKind InnermostKind, const RegionCodeGenTy &CodeGen,
1287     bool Tied, unsigned &NumberOfParts) {
1288   auto &&UntiedCodeGen = [this, &D, TaskTVar](CodeGenFunction &CGF,
1289                                               PrePostActionTy &) {
1290     llvm::Value *ThreadID = getThreadID(CGF, D.getBeginLoc());
1291     llvm::Value *UpLoc = emitUpdateLocation(CGF, D.getBeginLoc());
1292     llvm::Value *TaskArgs[] = {
1293         UpLoc, ThreadID,
1294         CGF.EmitLoadOfPointerLValue(CGF.GetAddrOfLocalVar(TaskTVar),
1295                                     TaskTVar->getType()->castAs<PointerType>())
1296             .getPointer(CGF)};
1297     CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
1298                             CGM.getModule(), OMPRTL___kmpc_omp_task),
1299                         TaskArgs);
1300   };
1301   CGOpenMPTaskOutlinedRegionInfo::UntiedTaskActionTy Action(Tied, PartIDVar,
1302                                                             UntiedCodeGen);
1303   CodeGen.setAction(Action);
1304   assert(!ThreadIDVar->getType()->isPointerType() &&
1305          "thread id variable must be of type kmp_int32 for tasks");
1306   const OpenMPDirectiveKind Region =
1307       isOpenMPTaskLoopDirective(D.getDirectiveKind()) ? OMPD_taskloop
1308                                                       : OMPD_task;
1309   const CapturedStmt *CS = D.getCapturedStmt(Region);
1310   bool HasCancel = false;
1311   if (const auto *TD = dyn_cast<OMPTaskDirective>(&D))
1312     HasCancel = TD->hasCancel();
1313   else if (const auto *TD = dyn_cast<OMPTaskLoopDirective>(&D))
1314     HasCancel = TD->hasCancel();
1315   else if (const auto *TD = dyn_cast<OMPMasterTaskLoopDirective>(&D))
1316     HasCancel = TD->hasCancel();
1317   else if (const auto *TD = dyn_cast<OMPParallelMasterTaskLoopDirective>(&D))
1318     HasCancel = TD->hasCancel();
1319 
1320   CodeGenFunction CGF(CGM, true);
1321   CGOpenMPTaskOutlinedRegionInfo CGInfo(*CS, ThreadIDVar, CodeGen,
1322                                         InnermostKind, HasCancel, Action);
1323   CodeGenFunction::CGCapturedStmtRAII CapInfoRAII(CGF, &CGInfo);
1324   llvm::Function *Res = CGF.GenerateCapturedStmtFunction(*CS);
1325   if (!Tied)
1326     NumberOfParts = Action.getNumberOfParts();
1327   return Res;
1328 }
1329 
setLocThreadIdInsertPt(CodeGenFunction & CGF,bool AtCurrentPoint)1330 void CGOpenMPRuntime::setLocThreadIdInsertPt(CodeGenFunction &CGF,
1331                                              bool AtCurrentPoint) {
1332   auto &Elem = OpenMPLocThreadIDMap.FindAndConstruct(CGF.CurFn);
1333   assert(!Elem.second.ServiceInsertPt && "Insert point is set already.");
1334 
1335   llvm::Value *Undef = llvm::UndefValue::get(CGF.Int32Ty);
1336   if (AtCurrentPoint) {
1337     Elem.second.ServiceInsertPt = new llvm::BitCastInst(
1338         Undef, CGF.Int32Ty, "svcpt", CGF.Builder.GetInsertBlock());
1339   } else {
1340     Elem.second.ServiceInsertPt =
1341         new llvm::BitCastInst(Undef, CGF.Int32Ty, "svcpt");
1342     Elem.second.ServiceInsertPt->insertAfter(CGF.AllocaInsertPt);
1343   }
1344 }
1345 
clearLocThreadIdInsertPt(CodeGenFunction & CGF)1346 void CGOpenMPRuntime::clearLocThreadIdInsertPt(CodeGenFunction &CGF) {
1347   auto &Elem = OpenMPLocThreadIDMap.FindAndConstruct(CGF.CurFn);
1348   if (Elem.second.ServiceInsertPt) {
1349     llvm::Instruction *Ptr = Elem.second.ServiceInsertPt;
1350     Elem.second.ServiceInsertPt = nullptr;
1351     Ptr->eraseFromParent();
1352   }
1353 }
1354 
getIdentStringFromSourceLocation(CodeGenFunction & CGF,SourceLocation Loc,SmallString<128> & Buffer)1355 static StringRef getIdentStringFromSourceLocation(CodeGenFunction &CGF,
1356                                                   SourceLocation Loc,
1357                                                   SmallString<128> &Buffer) {
1358   llvm::raw_svector_ostream OS(Buffer);
1359   // Build debug location
1360   PresumedLoc PLoc = CGF.getContext().getSourceManager().getPresumedLoc(Loc);
1361   OS << ";" << PLoc.getFilename() << ";";
1362   if (const auto *FD = dyn_cast_or_null<FunctionDecl>(CGF.CurFuncDecl))
1363     OS << FD->getQualifiedNameAsString();
1364   OS << ";" << PLoc.getLine() << ";" << PLoc.getColumn() << ";;";
1365   return OS.str();
1366 }
1367 
emitUpdateLocation(CodeGenFunction & CGF,SourceLocation Loc,unsigned Flags,bool EmitLoc)1368 llvm::Value *CGOpenMPRuntime::emitUpdateLocation(CodeGenFunction &CGF,
1369                                                  SourceLocation Loc,
1370                                                  unsigned Flags, bool EmitLoc) {
1371   uint32_t SrcLocStrSize;
1372   llvm::Constant *SrcLocStr;
1373   if ((!EmitLoc && CGM.getCodeGenOpts().getDebugInfo() ==
1374                        llvm::codegenoptions::NoDebugInfo) ||
1375       Loc.isInvalid()) {
1376     SrcLocStr = OMPBuilder.getOrCreateDefaultSrcLocStr(SrcLocStrSize);
1377   } else {
1378     std::string FunctionName;
1379     if (const auto *FD = dyn_cast_or_null<FunctionDecl>(CGF.CurFuncDecl))
1380       FunctionName = FD->getQualifiedNameAsString();
1381     PresumedLoc PLoc = CGF.getContext().getSourceManager().getPresumedLoc(Loc);
1382     const char *FileName = PLoc.getFilename();
1383     unsigned Line = PLoc.getLine();
1384     unsigned Column = PLoc.getColumn();
1385     SrcLocStr = OMPBuilder.getOrCreateSrcLocStr(FunctionName, FileName, Line,
1386                                                 Column, SrcLocStrSize);
1387   }
1388   unsigned Reserved2Flags = getDefaultLocationReserved2Flags();
1389   return OMPBuilder.getOrCreateIdent(
1390       SrcLocStr, SrcLocStrSize, llvm::omp::IdentFlag(Flags), Reserved2Flags);
1391 }
1392 
getThreadID(CodeGenFunction & CGF,SourceLocation Loc)1393 llvm::Value *CGOpenMPRuntime::getThreadID(CodeGenFunction &CGF,
1394                                           SourceLocation Loc) {
1395   assert(CGF.CurFn && "No function in current CodeGenFunction.");
1396   // If the OpenMPIRBuilder is used we need to use it for all thread id calls as
1397   // the clang invariants used below might be broken.
1398   if (CGM.getLangOpts().OpenMPIRBuilder) {
1399     SmallString<128> Buffer;
1400     OMPBuilder.updateToLocation(CGF.Builder.saveIP());
1401     uint32_t SrcLocStrSize;
1402     auto *SrcLocStr = OMPBuilder.getOrCreateSrcLocStr(
1403         getIdentStringFromSourceLocation(CGF, Loc, Buffer), SrcLocStrSize);
1404     return OMPBuilder.getOrCreateThreadID(
1405         OMPBuilder.getOrCreateIdent(SrcLocStr, SrcLocStrSize));
1406   }
1407 
1408   llvm::Value *ThreadID = nullptr;
1409   // Check whether we've already cached a load of the thread id in this
1410   // function.
1411   auto I = OpenMPLocThreadIDMap.find(CGF.CurFn);
1412   if (I != OpenMPLocThreadIDMap.end()) {
1413     ThreadID = I->second.ThreadID;
1414     if (ThreadID != nullptr)
1415       return ThreadID;
1416   }
1417   // If exceptions are enabled, do not use parameter to avoid possible crash.
1418   if (auto *OMPRegionInfo =
1419           dyn_cast_or_null<CGOpenMPRegionInfo>(CGF.CapturedStmtInfo)) {
1420     if (OMPRegionInfo->getThreadIDVariable()) {
1421       // Check if this an outlined function with thread id passed as argument.
1422       LValue LVal = OMPRegionInfo->getThreadIDVariableLValue(CGF);
1423       llvm::BasicBlock *TopBlock = CGF.AllocaInsertPt->getParent();
1424       if (!CGF.EHStack.requiresLandingPad() || !CGF.getLangOpts().Exceptions ||
1425           !CGF.getLangOpts().CXXExceptions ||
1426           CGF.Builder.GetInsertBlock() == TopBlock ||
1427           !isa<llvm::Instruction>(LVal.getPointer(CGF)) ||
1428           cast<llvm::Instruction>(LVal.getPointer(CGF))->getParent() ==
1429               TopBlock ||
1430           cast<llvm::Instruction>(LVal.getPointer(CGF))->getParent() ==
1431               CGF.Builder.GetInsertBlock()) {
1432         ThreadID = CGF.EmitLoadOfScalar(LVal, Loc);
1433         // If value loaded in entry block, cache it and use it everywhere in
1434         // function.
1435         if (CGF.Builder.GetInsertBlock() == TopBlock) {
1436           auto &Elem = OpenMPLocThreadIDMap.FindAndConstruct(CGF.CurFn);
1437           Elem.second.ThreadID = ThreadID;
1438         }
1439         return ThreadID;
1440       }
1441     }
1442   }
1443 
1444   // This is not an outlined function region - need to call __kmpc_int32
1445   // kmpc_global_thread_num(ident_t *loc).
1446   // Generate thread id value and cache this value for use across the
1447   // function.
1448   auto &Elem = OpenMPLocThreadIDMap.FindAndConstruct(CGF.CurFn);
1449   if (!Elem.second.ServiceInsertPt)
1450     setLocThreadIdInsertPt(CGF);
1451   CGBuilderTy::InsertPointGuard IPG(CGF.Builder);
1452   CGF.Builder.SetInsertPoint(Elem.second.ServiceInsertPt);
1453   auto DL = ApplyDebugLocation::CreateDefaultArtificial(CGF, Loc);
1454   llvm::CallInst *Call = CGF.Builder.CreateCall(
1455       OMPBuilder.getOrCreateRuntimeFunction(CGM.getModule(),
1456                                             OMPRTL___kmpc_global_thread_num),
1457       emitUpdateLocation(CGF, Loc));
1458   Call->setCallingConv(CGF.getRuntimeCC());
1459   Elem.second.ThreadID = Call;
1460   return Call;
1461 }
1462 
functionFinished(CodeGenFunction & CGF)1463 void CGOpenMPRuntime::functionFinished(CodeGenFunction &CGF) {
1464   assert(CGF.CurFn && "No function in current CodeGenFunction.");
1465   if (OpenMPLocThreadIDMap.count(CGF.CurFn)) {
1466     clearLocThreadIdInsertPt(CGF);
1467     OpenMPLocThreadIDMap.erase(CGF.CurFn);
1468   }
1469   if (FunctionUDRMap.count(CGF.CurFn) > 0) {
1470     for(const auto *D : FunctionUDRMap[CGF.CurFn])
1471       UDRMap.erase(D);
1472     FunctionUDRMap.erase(CGF.CurFn);
1473   }
1474   auto I = FunctionUDMMap.find(CGF.CurFn);
1475   if (I != FunctionUDMMap.end()) {
1476     for(const auto *D : I->second)
1477       UDMMap.erase(D);
1478     FunctionUDMMap.erase(I);
1479   }
1480   LastprivateConditionalToTypes.erase(CGF.CurFn);
1481   FunctionToUntiedTaskStackMap.erase(CGF.CurFn);
1482 }
1483 
getIdentTyPointerTy()1484 llvm::Type *CGOpenMPRuntime::getIdentTyPointerTy() {
1485   return OMPBuilder.IdentPtr;
1486 }
1487 
getKmpc_MicroPointerTy()1488 llvm::Type *CGOpenMPRuntime::getKmpc_MicroPointerTy() {
1489   if (!Kmpc_MicroTy) {
1490     // Build void (*kmpc_micro)(kmp_int32 *global_tid, kmp_int32 *bound_tid,...)
1491     llvm::Type *MicroParams[] = {llvm::PointerType::getUnqual(CGM.Int32Ty),
1492                                  llvm::PointerType::getUnqual(CGM.Int32Ty)};
1493     Kmpc_MicroTy = llvm::FunctionType::get(CGM.VoidTy, MicroParams, true);
1494   }
1495   return llvm::PointerType::getUnqual(Kmpc_MicroTy);
1496 }
1497 
1498 llvm::OffloadEntriesInfoManager::OMPTargetDeviceClauseKind
convertDeviceClause(const VarDecl * VD)1499 convertDeviceClause(const VarDecl *VD) {
1500   std::optional<OMPDeclareTargetDeclAttr::DevTypeTy> DevTy =
1501       OMPDeclareTargetDeclAttr::getDeviceType(VD);
1502   if (!DevTy)
1503     return llvm::OffloadEntriesInfoManager::OMPTargetDeviceClauseNone;
1504 
1505   switch ((int)*DevTy) { // Avoid -Wcovered-switch-default
1506   case OMPDeclareTargetDeclAttr::DT_Host:
1507     return llvm::OffloadEntriesInfoManager::OMPTargetDeviceClauseHost;
1508     break;
1509   case OMPDeclareTargetDeclAttr::DT_NoHost:
1510     return llvm::OffloadEntriesInfoManager::OMPTargetDeviceClauseNoHost;
1511     break;
1512   case OMPDeclareTargetDeclAttr::DT_Any:
1513     return llvm::OffloadEntriesInfoManager::OMPTargetDeviceClauseAny;
1514     break;
1515   default:
1516     return llvm::OffloadEntriesInfoManager::OMPTargetDeviceClauseNone;
1517     break;
1518   }
1519 }
1520 
1521 llvm::OffloadEntriesInfoManager::OMPTargetGlobalVarEntryKind
convertCaptureClause(const VarDecl * VD)1522 convertCaptureClause(const VarDecl *VD) {
1523   std::optional<OMPDeclareTargetDeclAttr::MapTypeTy> MapType =
1524       OMPDeclareTargetDeclAttr::isDeclareTargetDeclaration(VD);
1525   if (!MapType)
1526     return llvm::OffloadEntriesInfoManager::OMPTargetGlobalVarEntryNone;
1527   switch ((int)*MapType) { // Avoid -Wcovered-switch-default
1528   case OMPDeclareTargetDeclAttr::MapTypeTy::MT_To:
1529     return llvm::OffloadEntriesInfoManager::OMPTargetGlobalVarEntryTo;
1530     break;
1531   case OMPDeclareTargetDeclAttr::MapTypeTy::MT_Enter:
1532     return llvm::OffloadEntriesInfoManager::OMPTargetGlobalVarEntryEnter;
1533     break;
1534   case OMPDeclareTargetDeclAttr::MapTypeTy::MT_Link:
1535     return llvm::OffloadEntriesInfoManager::OMPTargetGlobalVarEntryLink;
1536     break;
1537   default:
1538     return llvm::OffloadEntriesInfoManager::OMPTargetGlobalVarEntryNone;
1539     break;
1540   }
1541 }
1542 
getEntryInfoFromPresumedLoc(CodeGenModule & CGM,llvm::OpenMPIRBuilder & OMPBuilder,SourceLocation BeginLoc,llvm::StringRef ParentName="")1543 static llvm::TargetRegionEntryInfo getEntryInfoFromPresumedLoc(
1544     CodeGenModule &CGM, llvm::OpenMPIRBuilder &OMPBuilder,
1545     SourceLocation BeginLoc, llvm::StringRef ParentName = "") {
1546 
1547   auto FileInfoCallBack = [&]() {
1548     SourceManager &SM = CGM.getContext().getSourceManager();
1549     PresumedLoc PLoc = SM.getPresumedLoc(BeginLoc);
1550 
1551     llvm::sys::fs::UniqueID ID;
1552     if (llvm::sys::fs::getUniqueID(PLoc.getFilename(), ID)) {
1553       PLoc = SM.getPresumedLoc(BeginLoc, /*UseLineDirectives=*/false);
1554     }
1555 
1556     return std::pair<std::string, uint64_t>(PLoc.getFilename(), PLoc.getLine());
1557   };
1558 
1559   return OMPBuilder.getTargetEntryUniqueInfo(FileInfoCallBack, ParentName);
1560 }
1561 
getAddrOfDeclareTargetVar(const VarDecl * VD)1562 ConstantAddress CGOpenMPRuntime::getAddrOfDeclareTargetVar(const VarDecl *VD) {
1563   auto AddrOfGlobal = [&VD, this]() { return CGM.GetAddrOfGlobal(VD); };
1564 
1565   auto LinkageForVariable = [&VD, this]() {
1566     return CGM.getLLVMLinkageVarDefinition(VD);
1567   };
1568 
1569   std::vector<llvm::GlobalVariable *> GeneratedRefs;
1570 
1571   llvm::Type *LlvmPtrTy = CGM.getTypes().ConvertTypeForMem(
1572       CGM.getContext().getPointerType(VD->getType()));
1573   llvm::Constant *addr = OMPBuilder.getAddrOfDeclareTargetVar(
1574       convertCaptureClause(VD), convertDeviceClause(VD),
1575       VD->hasDefinition(CGM.getContext()) == VarDecl::DeclarationOnly,
1576       VD->isExternallyVisible(),
1577       getEntryInfoFromPresumedLoc(CGM, OMPBuilder,
1578                                   VD->getCanonicalDecl()->getBeginLoc()),
1579       CGM.getMangledName(VD), GeneratedRefs, CGM.getLangOpts().OpenMPSimd,
1580       CGM.getLangOpts().OMPTargetTriples, LlvmPtrTy, AddrOfGlobal,
1581       LinkageForVariable);
1582 
1583   if (!addr)
1584     return ConstantAddress::invalid();
1585   return ConstantAddress(addr, LlvmPtrTy, CGM.getContext().getDeclAlign(VD));
1586 }
1587 
1588 llvm::Constant *
getOrCreateThreadPrivateCache(const VarDecl * VD)1589 CGOpenMPRuntime::getOrCreateThreadPrivateCache(const VarDecl *VD) {
1590   assert(!CGM.getLangOpts().OpenMPUseTLS ||
1591          !CGM.getContext().getTargetInfo().isTLSSupported());
1592   // Lookup the entry, lazily creating it if necessary.
1593   std::string Suffix = getName({"cache", ""});
1594   return OMPBuilder.getOrCreateInternalVariable(
1595       CGM.Int8PtrPtrTy, Twine(CGM.getMangledName(VD)).concat(Suffix).str());
1596 }
1597 
getAddrOfThreadPrivate(CodeGenFunction & CGF,const VarDecl * VD,Address VDAddr,SourceLocation Loc)1598 Address CGOpenMPRuntime::getAddrOfThreadPrivate(CodeGenFunction &CGF,
1599                                                 const VarDecl *VD,
1600                                                 Address VDAddr,
1601                                                 SourceLocation Loc) {
1602   if (CGM.getLangOpts().OpenMPUseTLS &&
1603       CGM.getContext().getTargetInfo().isTLSSupported())
1604     return VDAddr;
1605 
1606   llvm::Type *VarTy = VDAddr.getElementType();
1607   llvm::Value *Args[] = {
1608       emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc),
1609       CGF.Builder.CreatePointerCast(VDAddr.emitRawPointer(CGF), CGM.Int8PtrTy),
1610       CGM.getSize(CGM.GetTargetTypeStoreSize(VarTy)),
1611       getOrCreateThreadPrivateCache(VD)};
1612   return Address(
1613       CGF.EmitRuntimeCall(
1614           OMPBuilder.getOrCreateRuntimeFunction(
1615               CGM.getModule(), OMPRTL___kmpc_threadprivate_cached),
1616           Args),
1617       CGF.Int8Ty, VDAddr.getAlignment());
1618 }
1619 
emitThreadPrivateVarInit(CodeGenFunction & CGF,Address VDAddr,llvm::Value * Ctor,llvm::Value * CopyCtor,llvm::Value * Dtor,SourceLocation Loc)1620 void CGOpenMPRuntime::emitThreadPrivateVarInit(
1621     CodeGenFunction &CGF, Address VDAddr, llvm::Value *Ctor,
1622     llvm::Value *CopyCtor, llvm::Value *Dtor, SourceLocation Loc) {
1623   // Call kmp_int32 __kmpc_global_thread_num(&loc) to init OpenMP runtime
1624   // library.
1625   llvm::Value *OMPLoc = emitUpdateLocation(CGF, Loc);
1626   CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
1627                           CGM.getModule(), OMPRTL___kmpc_global_thread_num),
1628                       OMPLoc);
1629   // Call __kmpc_threadprivate_register(&loc, &var, ctor, cctor/*NULL*/, dtor)
1630   // to register constructor/destructor for variable.
1631   llvm::Value *Args[] = {
1632       OMPLoc,
1633       CGF.Builder.CreatePointerCast(VDAddr.emitRawPointer(CGF), CGM.VoidPtrTy),
1634       Ctor, CopyCtor, Dtor};
1635   CGF.EmitRuntimeCall(
1636       OMPBuilder.getOrCreateRuntimeFunction(
1637           CGM.getModule(), OMPRTL___kmpc_threadprivate_register),
1638       Args);
1639 }
1640 
emitThreadPrivateVarDefinition(const VarDecl * VD,Address VDAddr,SourceLocation Loc,bool PerformInit,CodeGenFunction * CGF)1641 llvm::Function *CGOpenMPRuntime::emitThreadPrivateVarDefinition(
1642     const VarDecl *VD, Address VDAddr, SourceLocation Loc,
1643     bool PerformInit, CodeGenFunction *CGF) {
1644   if (CGM.getLangOpts().OpenMPUseTLS &&
1645       CGM.getContext().getTargetInfo().isTLSSupported())
1646     return nullptr;
1647 
1648   VD = VD->getDefinition(CGM.getContext());
1649   if (VD && ThreadPrivateWithDefinition.insert(CGM.getMangledName(VD)).second) {
1650     QualType ASTTy = VD->getType();
1651 
1652     llvm::Value *Ctor = nullptr, *CopyCtor = nullptr, *Dtor = nullptr;
1653     const Expr *Init = VD->getAnyInitializer();
1654     if (CGM.getLangOpts().CPlusPlus && PerformInit) {
1655       // Generate function that re-emits the declaration's initializer into the
1656       // threadprivate copy of the variable VD
1657       CodeGenFunction CtorCGF(CGM);
1658       FunctionArgList Args;
1659       ImplicitParamDecl Dst(CGM.getContext(), /*DC=*/nullptr, Loc,
1660                             /*Id=*/nullptr, CGM.getContext().VoidPtrTy,
1661                             ImplicitParamKind::Other);
1662       Args.push_back(&Dst);
1663 
1664       const auto &FI = CGM.getTypes().arrangeBuiltinFunctionDeclaration(
1665           CGM.getContext().VoidPtrTy, Args);
1666       llvm::FunctionType *FTy = CGM.getTypes().GetFunctionType(FI);
1667       std::string Name = getName({"__kmpc_global_ctor_", ""});
1668       llvm::Function *Fn =
1669           CGM.CreateGlobalInitOrCleanUpFunction(FTy, Name, FI, Loc);
1670       CtorCGF.StartFunction(GlobalDecl(), CGM.getContext().VoidPtrTy, Fn, FI,
1671                             Args, Loc, Loc);
1672       llvm::Value *ArgVal = CtorCGF.EmitLoadOfScalar(
1673           CtorCGF.GetAddrOfLocalVar(&Dst), /*Volatile=*/false,
1674           CGM.getContext().VoidPtrTy, Dst.getLocation());
1675       Address Arg(ArgVal, CtorCGF.ConvertTypeForMem(ASTTy),
1676                   VDAddr.getAlignment());
1677       CtorCGF.EmitAnyExprToMem(Init, Arg, Init->getType().getQualifiers(),
1678                                /*IsInitializer=*/true);
1679       ArgVal = CtorCGF.EmitLoadOfScalar(
1680           CtorCGF.GetAddrOfLocalVar(&Dst), /*Volatile=*/false,
1681           CGM.getContext().VoidPtrTy, Dst.getLocation());
1682       CtorCGF.Builder.CreateStore(ArgVal, CtorCGF.ReturnValue);
1683       CtorCGF.FinishFunction();
1684       Ctor = Fn;
1685     }
1686     if (VD->getType().isDestructedType() != QualType::DK_none) {
1687       // Generate function that emits destructor call for the threadprivate copy
1688       // of the variable VD
1689       CodeGenFunction DtorCGF(CGM);
1690       FunctionArgList Args;
1691       ImplicitParamDecl Dst(CGM.getContext(), /*DC=*/nullptr, Loc,
1692                             /*Id=*/nullptr, CGM.getContext().VoidPtrTy,
1693                             ImplicitParamKind::Other);
1694       Args.push_back(&Dst);
1695 
1696       const auto &FI = CGM.getTypes().arrangeBuiltinFunctionDeclaration(
1697           CGM.getContext().VoidTy, Args);
1698       llvm::FunctionType *FTy = CGM.getTypes().GetFunctionType(FI);
1699       std::string Name = getName({"__kmpc_global_dtor_", ""});
1700       llvm::Function *Fn =
1701           CGM.CreateGlobalInitOrCleanUpFunction(FTy, Name, FI, Loc);
1702       auto NL = ApplyDebugLocation::CreateEmpty(DtorCGF);
1703       DtorCGF.StartFunction(GlobalDecl(), CGM.getContext().VoidTy, Fn, FI, Args,
1704                             Loc, Loc);
1705       // Create a scope with an artificial location for the body of this function.
1706       auto AL = ApplyDebugLocation::CreateArtificial(DtorCGF);
1707       llvm::Value *ArgVal = DtorCGF.EmitLoadOfScalar(
1708           DtorCGF.GetAddrOfLocalVar(&Dst),
1709           /*Volatile=*/false, CGM.getContext().VoidPtrTy, Dst.getLocation());
1710       DtorCGF.emitDestroy(
1711           Address(ArgVal, DtorCGF.Int8Ty, VDAddr.getAlignment()), ASTTy,
1712           DtorCGF.getDestroyer(ASTTy.isDestructedType()),
1713           DtorCGF.needsEHCleanup(ASTTy.isDestructedType()));
1714       DtorCGF.FinishFunction();
1715       Dtor = Fn;
1716     }
1717     // Do not emit init function if it is not required.
1718     if (!Ctor && !Dtor)
1719       return nullptr;
1720 
1721     llvm::Type *CopyCtorTyArgs[] = {CGM.VoidPtrTy, CGM.VoidPtrTy};
1722     auto *CopyCtorTy = llvm::FunctionType::get(CGM.VoidPtrTy, CopyCtorTyArgs,
1723                                                /*isVarArg=*/false)
1724                            ->getPointerTo();
1725     // Copying constructor for the threadprivate variable.
1726     // Must be NULL - reserved by runtime, but currently it requires that this
1727     // parameter is always NULL. Otherwise it fires assertion.
1728     CopyCtor = llvm::Constant::getNullValue(CopyCtorTy);
1729     if (Ctor == nullptr) {
1730       auto *CtorTy = llvm::FunctionType::get(CGM.VoidPtrTy, CGM.VoidPtrTy,
1731                                              /*isVarArg=*/false)
1732                          ->getPointerTo();
1733       Ctor = llvm::Constant::getNullValue(CtorTy);
1734     }
1735     if (Dtor == nullptr) {
1736       auto *DtorTy = llvm::FunctionType::get(CGM.VoidTy, CGM.VoidPtrTy,
1737                                              /*isVarArg=*/false)
1738                          ->getPointerTo();
1739       Dtor = llvm::Constant::getNullValue(DtorTy);
1740     }
1741     if (!CGF) {
1742       auto *InitFunctionTy =
1743           llvm::FunctionType::get(CGM.VoidTy, /*isVarArg*/ false);
1744       std::string Name = getName({"__omp_threadprivate_init_", ""});
1745       llvm::Function *InitFunction = CGM.CreateGlobalInitOrCleanUpFunction(
1746           InitFunctionTy, Name, CGM.getTypes().arrangeNullaryFunction());
1747       CodeGenFunction InitCGF(CGM);
1748       FunctionArgList ArgList;
1749       InitCGF.StartFunction(GlobalDecl(), CGM.getContext().VoidTy, InitFunction,
1750                             CGM.getTypes().arrangeNullaryFunction(), ArgList,
1751                             Loc, Loc);
1752       emitThreadPrivateVarInit(InitCGF, VDAddr, Ctor, CopyCtor, Dtor, Loc);
1753       InitCGF.FinishFunction();
1754       return InitFunction;
1755     }
1756     emitThreadPrivateVarInit(*CGF, VDAddr, Ctor, CopyCtor, Dtor, Loc);
1757   }
1758   return nullptr;
1759 }
1760 
emitDeclareTargetFunction(const FunctionDecl * FD,llvm::GlobalValue * GV)1761 void CGOpenMPRuntime::emitDeclareTargetFunction(const FunctionDecl *FD,
1762                                                 llvm::GlobalValue *GV) {
1763   std::optional<OMPDeclareTargetDeclAttr *> ActiveAttr =
1764       OMPDeclareTargetDeclAttr::getActiveAttr(FD);
1765 
1766   // We only need to handle active 'indirect' declare target functions.
1767   if (!ActiveAttr || !(*ActiveAttr)->getIndirect())
1768     return;
1769 
1770   // Get a mangled name to store the new device global in.
1771   llvm::TargetRegionEntryInfo EntryInfo = getEntryInfoFromPresumedLoc(
1772       CGM, OMPBuilder, FD->getCanonicalDecl()->getBeginLoc(), FD->getName());
1773   SmallString<128> Name;
1774   OMPBuilder.OffloadInfoManager.getTargetRegionEntryFnName(Name, EntryInfo);
1775 
1776   // We need to generate a new global to hold the address of the indirectly
1777   // called device function. Doing this allows us to keep the visibility and
1778   // linkage of the associated function unchanged while allowing the runtime to
1779   // access its value.
1780   llvm::GlobalValue *Addr = GV;
1781   if (CGM.getLangOpts().OpenMPIsTargetDevice) {
1782     Addr = new llvm::GlobalVariable(
1783         CGM.getModule(), CGM.VoidPtrTy,
1784         /*isConstant=*/true, llvm::GlobalValue::ExternalLinkage, GV, Name,
1785         nullptr, llvm::GlobalValue::NotThreadLocal,
1786         CGM.getModule().getDataLayout().getDefaultGlobalsAddressSpace());
1787     Addr->setVisibility(llvm::GlobalValue::ProtectedVisibility);
1788   }
1789 
1790   OMPBuilder.OffloadInfoManager.registerDeviceGlobalVarEntryInfo(
1791       Name, Addr, CGM.GetTargetTypeStoreSize(CGM.VoidPtrTy).getQuantity(),
1792       llvm::OffloadEntriesInfoManager::OMPTargetGlobalVarEntryIndirect,
1793       llvm::GlobalValue::WeakODRLinkage);
1794 }
1795 
getAddrOfArtificialThreadPrivate(CodeGenFunction & CGF,QualType VarType,StringRef Name)1796 Address CGOpenMPRuntime::getAddrOfArtificialThreadPrivate(CodeGenFunction &CGF,
1797                                                           QualType VarType,
1798                                                           StringRef Name) {
1799   std::string Suffix = getName({"artificial", ""});
1800   llvm::Type *VarLVType = CGF.ConvertTypeForMem(VarType);
1801   llvm::GlobalVariable *GAddr = OMPBuilder.getOrCreateInternalVariable(
1802       VarLVType, Twine(Name).concat(Suffix).str());
1803   if (CGM.getLangOpts().OpenMP && CGM.getLangOpts().OpenMPUseTLS &&
1804       CGM.getTarget().isTLSSupported()) {
1805     GAddr->setThreadLocal(/*Val=*/true);
1806     return Address(GAddr, GAddr->getValueType(),
1807                    CGM.getContext().getTypeAlignInChars(VarType));
1808   }
1809   std::string CacheSuffix = getName({"cache", ""});
1810   llvm::Value *Args[] = {
1811       emitUpdateLocation(CGF, SourceLocation()),
1812       getThreadID(CGF, SourceLocation()),
1813       CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(GAddr, CGM.VoidPtrTy),
1814       CGF.Builder.CreateIntCast(CGF.getTypeSize(VarType), CGM.SizeTy,
1815                                 /*isSigned=*/false),
1816       OMPBuilder.getOrCreateInternalVariable(
1817           CGM.VoidPtrPtrTy,
1818           Twine(Name).concat(Suffix).concat(CacheSuffix).str())};
1819   return Address(
1820       CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
1821           CGF.EmitRuntimeCall(
1822               OMPBuilder.getOrCreateRuntimeFunction(
1823                   CGM.getModule(), OMPRTL___kmpc_threadprivate_cached),
1824               Args),
1825           VarLVType->getPointerTo(/*AddrSpace=*/0)),
1826       VarLVType, CGM.getContext().getTypeAlignInChars(VarType));
1827 }
1828 
emitIfClause(CodeGenFunction & CGF,const Expr * Cond,const RegionCodeGenTy & ThenGen,const RegionCodeGenTy & ElseGen)1829 void CGOpenMPRuntime::emitIfClause(CodeGenFunction &CGF, const Expr *Cond,
1830                                    const RegionCodeGenTy &ThenGen,
1831                                    const RegionCodeGenTy &ElseGen) {
1832   CodeGenFunction::LexicalScope ConditionScope(CGF, Cond->getSourceRange());
1833 
1834   // If the condition constant folds and can be elided, try to avoid emitting
1835   // the condition and the dead arm of the if/else.
1836   bool CondConstant;
1837   if (CGF.ConstantFoldsToSimpleInteger(Cond, CondConstant)) {
1838     if (CondConstant)
1839       ThenGen(CGF);
1840     else
1841       ElseGen(CGF);
1842     return;
1843   }
1844 
1845   // Otherwise, the condition did not fold, or we couldn't elide it.  Just
1846   // emit the conditional branch.
1847   llvm::BasicBlock *ThenBlock = CGF.createBasicBlock("omp_if.then");
1848   llvm::BasicBlock *ElseBlock = CGF.createBasicBlock("omp_if.else");
1849   llvm::BasicBlock *ContBlock = CGF.createBasicBlock("omp_if.end");
1850   CGF.EmitBranchOnBoolExpr(Cond, ThenBlock, ElseBlock, /*TrueCount=*/0);
1851 
1852   // Emit the 'then' code.
1853   CGF.EmitBlock(ThenBlock);
1854   ThenGen(CGF);
1855   CGF.EmitBranch(ContBlock);
1856   // Emit the 'else' code if present.
1857   // There is no need to emit line number for unconditional branch.
1858   (void)ApplyDebugLocation::CreateEmpty(CGF);
1859   CGF.EmitBlock(ElseBlock);
1860   ElseGen(CGF);
1861   // There is no need to emit line number for unconditional branch.
1862   (void)ApplyDebugLocation::CreateEmpty(CGF);
1863   CGF.EmitBranch(ContBlock);
1864   // Emit the continuation block for code after the if.
1865   CGF.EmitBlock(ContBlock, /*IsFinished=*/true);
1866 }
1867 
emitParallelCall(CodeGenFunction & CGF,SourceLocation Loc,llvm::Function * OutlinedFn,ArrayRef<llvm::Value * > CapturedVars,const Expr * IfCond,llvm::Value * NumThreads)1868 void CGOpenMPRuntime::emitParallelCall(CodeGenFunction &CGF, SourceLocation Loc,
1869                                        llvm::Function *OutlinedFn,
1870                                        ArrayRef<llvm::Value *> CapturedVars,
1871                                        const Expr *IfCond,
1872                                        llvm::Value *NumThreads) {
1873   if (!CGF.HaveInsertPoint())
1874     return;
1875   llvm::Value *RTLoc = emitUpdateLocation(CGF, Loc);
1876   auto &M = CGM.getModule();
1877   auto &&ThenGen = [&M, OutlinedFn, CapturedVars, RTLoc,
1878                     this](CodeGenFunction &CGF, PrePostActionTy &) {
1879     // Build call __kmpc_fork_call(loc, n, microtask, var1, .., varn);
1880     CGOpenMPRuntime &RT = CGF.CGM.getOpenMPRuntime();
1881     llvm::Value *Args[] = {
1882         RTLoc,
1883         CGF.Builder.getInt32(CapturedVars.size()), // Number of captured vars
1884         CGF.Builder.CreateBitCast(OutlinedFn, RT.getKmpc_MicroPointerTy())};
1885     llvm::SmallVector<llvm::Value *, 16> RealArgs;
1886     RealArgs.append(std::begin(Args), std::end(Args));
1887     RealArgs.append(CapturedVars.begin(), CapturedVars.end());
1888 
1889     llvm::FunctionCallee RTLFn =
1890         OMPBuilder.getOrCreateRuntimeFunction(M, OMPRTL___kmpc_fork_call);
1891     CGF.EmitRuntimeCall(RTLFn, RealArgs);
1892   };
1893   auto &&ElseGen = [&M, OutlinedFn, CapturedVars, RTLoc, Loc,
1894                     this](CodeGenFunction &CGF, PrePostActionTy &) {
1895     CGOpenMPRuntime &RT = CGF.CGM.getOpenMPRuntime();
1896     llvm::Value *ThreadID = RT.getThreadID(CGF, Loc);
1897     // Build calls:
1898     // __kmpc_serialized_parallel(&Loc, GTid);
1899     llvm::Value *Args[] = {RTLoc, ThreadID};
1900     CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
1901                             M, OMPRTL___kmpc_serialized_parallel),
1902                         Args);
1903 
1904     // OutlinedFn(&GTid, &zero_bound, CapturedStruct);
1905     Address ThreadIDAddr = RT.emitThreadIDAddress(CGF, Loc);
1906     RawAddress ZeroAddrBound =
1907         CGF.CreateDefaultAlignTempAlloca(CGF.Int32Ty,
1908                                          /*Name=*/".bound.zero.addr");
1909     CGF.Builder.CreateStore(CGF.Builder.getInt32(/*C*/ 0), ZeroAddrBound);
1910     llvm::SmallVector<llvm::Value *, 16> OutlinedFnArgs;
1911     // ThreadId for serialized parallels is 0.
1912     OutlinedFnArgs.push_back(ThreadIDAddr.emitRawPointer(CGF));
1913     OutlinedFnArgs.push_back(ZeroAddrBound.getPointer());
1914     OutlinedFnArgs.append(CapturedVars.begin(), CapturedVars.end());
1915 
1916     // Ensure we do not inline the function. This is trivially true for the ones
1917     // passed to __kmpc_fork_call but the ones called in serialized regions
1918     // could be inlined. This is not a perfect but it is closer to the invariant
1919     // we want, namely, every data environment starts with a new function.
1920     // TODO: We should pass the if condition to the runtime function and do the
1921     //       handling there. Much cleaner code.
1922     OutlinedFn->removeFnAttr(llvm::Attribute::AlwaysInline);
1923     OutlinedFn->addFnAttr(llvm::Attribute::NoInline);
1924     RT.emitOutlinedFunctionCall(CGF, Loc, OutlinedFn, OutlinedFnArgs);
1925 
1926     // __kmpc_end_serialized_parallel(&Loc, GTid);
1927     llvm::Value *EndArgs[] = {RT.emitUpdateLocation(CGF, Loc), ThreadID};
1928     CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
1929                             M, OMPRTL___kmpc_end_serialized_parallel),
1930                         EndArgs);
1931   };
1932   if (IfCond) {
1933     emitIfClause(CGF, IfCond, ThenGen, ElseGen);
1934   } else {
1935     RegionCodeGenTy ThenRCG(ThenGen);
1936     ThenRCG(CGF);
1937   }
1938 }
1939 
1940 // If we're inside an (outlined) parallel region, use the region info's
1941 // thread-ID variable (it is passed in a first argument of the outlined function
1942 // as "kmp_int32 *gtid"). Otherwise, if we're not inside parallel region, but in
1943 // regular serial code region, get thread ID by calling kmp_int32
1944 // kmpc_global_thread_num(ident_t *loc), stash this thread ID in a temporary and
1945 // return the address of that temp.
emitThreadIDAddress(CodeGenFunction & CGF,SourceLocation Loc)1946 Address CGOpenMPRuntime::emitThreadIDAddress(CodeGenFunction &CGF,
1947                                              SourceLocation Loc) {
1948   if (auto *OMPRegionInfo =
1949           dyn_cast_or_null<CGOpenMPRegionInfo>(CGF.CapturedStmtInfo))
1950     if (OMPRegionInfo->getThreadIDVariable())
1951       return OMPRegionInfo->getThreadIDVariableLValue(CGF).getAddress();
1952 
1953   llvm::Value *ThreadID = getThreadID(CGF, Loc);
1954   QualType Int32Ty =
1955       CGF.getContext().getIntTypeForBitwidth(/*DestWidth*/ 32, /*Signed*/ true);
1956   Address ThreadIDTemp = CGF.CreateMemTemp(Int32Ty, /*Name*/ ".threadid_temp.");
1957   CGF.EmitStoreOfScalar(ThreadID,
1958                         CGF.MakeAddrLValue(ThreadIDTemp, Int32Ty));
1959 
1960   return ThreadIDTemp;
1961 }
1962 
getCriticalRegionLock(StringRef CriticalName)1963 llvm::Value *CGOpenMPRuntime::getCriticalRegionLock(StringRef CriticalName) {
1964   std::string Prefix = Twine("gomp_critical_user_", CriticalName).str();
1965   std::string Name = getName({Prefix, "var"});
1966   return OMPBuilder.getOrCreateInternalVariable(KmpCriticalNameTy, Name);
1967 }
1968 
1969 namespace {
1970 /// Common pre(post)-action for different OpenMP constructs.
1971 class CommonActionTy final : public PrePostActionTy {
1972   llvm::FunctionCallee EnterCallee;
1973   ArrayRef<llvm::Value *> EnterArgs;
1974   llvm::FunctionCallee ExitCallee;
1975   ArrayRef<llvm::Value *> ExitArgs;
1976   bool Conditional;
1977   llvm::BasicBlock *ContBlock = nullptr;
1978 
1979 public:
CommonActionTy(llvm::FunctionCallee EnterCallee,ArrayRef<llvm::Value * > EnterArgs,llvm::FunctionCallee ExitCallee,ArrayRef<llvm::Value * > ExitArgs,bool Conditional=false)1980   CommonActionTy(llvm::FunctionCallee EnterCallee,
1981                  ArrayRef<llvm::Value *> EnterArgs,
1982                  llvm::FunctionCallee ExitCallee,
1983                  ArrayRef<llvm::Value *> ExitArgs, bool Conditional = false)
1984       : EnterCallee(EnterCallee), EnterArgs(EnterArgs), ExitCallee(ExitCallee),
1985         ExitArgs(ExitArgs), Conditional(Conditional) {}
Enter(CodeGenFunction & CGF)1986   void Enter(CodeGenFunction &CGF) override {
1987     llvm::Value *EnterRes = CGF.EmitRuntimeCall(EnterCallee, EnterArgs);
1988     if (Conditional) {
1989       llvm::Value *CallBool = CGF.Builder.CreateIsNotNull(EnterRes);
1990       auto *ThenBlock = CGF.createBasicBlock("omp_if.then");
1991       ContBlock = CGF.createBasicBlock("omp_if.end");
1992       // Generate the branch (If-stmt)
1993       CGF.Builder.CreateCondBr(CallBool, ThenBlock, ContBlock);
1994       CGF.EmitBlock(ThenBlock);
1995     }
1996   }
Done(CodeGenFunction & CGF)1997   void Done(CodeGenFunction &CGF) {
1998     // Emit the rest of blocks/branches
1999     CGF.EmitBranch(ContBlock);
2000     CGF.EmitBlock(ContBlock, true);
2001   }
Exit(CodeGenFunction & CGF)2002   void Exit(CodeGenFunction &CGF) override {
2003     CGF.EmitRuntimeCall(ExitCallee, ExitArgs);
2004   }
2005 };
2006 } // anonymous namespace
2007 
emitCriticalRegion(CodeGenFunction & CGF,StringRef CriticalName,const RegionCodeGenTy & CriticalOpGen,SourceLocation Loc,const Expr * Hint)2008 void CGOpenMPRuntime::emitCriticalRegion(CodeGenFunction &CGF,
2009                                          StringRef CriticalName,
2010                                          const RegionCodeGenTy &CriticalOpGen,
2011                                          SourceLocation Loc, const Expr *Hint) {
2012   // __kmpc_critical[_with_hint](ident_t *, gtid, Lock[, hint]);
2013   // CriticalOpGen();
2014   // __kmpc_end_critical(ident_t *, gtid, Lock);
2015   // Prepare arguments and build a call to __kmpc_critical
2016   if (!CGF.HaveInsertPoint())
2017     return;
2018   llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc),
2019                          getCriticalRegionLock(CriticalName)};
2020   llvm::SmallVector<llvm::Value *, 4> EnterArgs(std::begin(Args),
2021                                                 std::end(Args));
2022   if (Hint) {
2023     EnterArgs.push_back(CGF.Builder.CreateIntCast(
2024         CGF.EmitScalarExpr(Hint), CGM.Int32Ty, /*isSigned=*/false));
2025   }
2026   CommonActionTy Action(
2027       OMPBuilder.getOrCreateRuntimeFunction(
2028           CGM.getModule(),
2029           Hint ? OMPRTL___kmpc_critical_with_hint : OMPRTL___kmpc_critical),
2030       EnterArgs,
2031       OMPBuilder.getOrCreateRuntimeFunction(CGM.getModule(),
2032                                             OMPRTL___kmpc_end_critical),
2033       Args);
2034   CriticalOpGen.setAction(Action);
2035   emitInlinedDirective(CGF, OMPD_critical, CriticalOpGen);
2036 }
2037 
emitMasterRegion(CodeGenFunction & CGF,const RegionCodeGenTy & MasterOpGen,SourceLocation Loc)2038 void CGOpenMPRuntime::emitMasterRegion(CodeGenFunction &CGF,
2039                                        const RegionCodeGenTy &MasterOpGen,
2040                                        SourceLocation Loc) {
2041   if (!CGF.HaveInsertPoint())
2042     return;
2043   // if(__kmpc_master(ident_t *, gtid)) {
2044   //   MasterOpGen();
2045   //   __kmpc_end_master(ident_t *, gtid);
2046   // }
2047   // Prepare arguments and build a call to __kmpc_master
2048   llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc)};
2049   CommonActionTy Action(OMPBuilder.getOrCreateRuntimeFunction(
2050                             CGM.getModule(), OMPRTL___kmpc_master),
2051                         Args,
2052                         OMPBuilder.getOrCreateRuntimeFunction(
2053                             CGM.getModule(), OMPRTL___kmpc_end_master),
2054                         Args,
2055                         /*Conditional=*/true);
2056   MasterOpGen.setAction(Action);
2057   emitInlinedDirective(CGF, OMPD_master, MasterOpGen);
2058   Action.Done(CGF);
2059 }
2060 
emitMaskedRegion(CodeGenFunction & CGF,const RegionCodeGenTy & MaskedOpGen,SourceLocation Loc,const Expr * Filter)2061 void CGOpenMPRuntime::emitMaskedRegion(CodeGenFunction &CGF,
2062                                        const RegionCodeGenTy &MaskedOpGen,
2063                                        SourceLocation Loc, const Expr *Filter) {
2064   if (!CGF.HaveInsertPoint())
2065     return;
2066   // if(__kmpc_masked(ident_t *, gtid, filter)) {
2067   //   MaskedOpGen();
2068   //   __kmpc_end_masked(iden_t *, gtid);
2069   // }
2070   // Prepare arguments and build a call to __kmpc_masked
2071   llvm::Value *FilterVal = Filter
2072                                ? CGF.EmitScalarExpr(Filter, CGF.Int32Ty)
2073                                : llvm::ConstantInt::get(CGM.Int32Ty, /*V=*/0);
2074   llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc),
2075                          FilterVal};
2076   llvm::Value *ArgsEnd[] = {emitUpdateLocation(CGF, Loc),
2077                             getThreadID(CGF, Loc)};
2078   CommonActionTy Action(OMPBuilder.getOrCreateRuntimeFunction(
2079                             CGM.getModule(), OMPRTL___kmpc_masked),
2080                         Args,
2081                         OMPBuilder.getOrCreateRuntimeFunction(
2082                             CGM.getModule(), OMPRTL___kmpc_end_masked),
2083                         ArgsEnd,
2084                         /*Conditional=*/true);
2085   MaskedOpGen.setAction(Action);
2086   emitInlinedDirective(CGF, OMPD_masked, MaskedOpGen);
2087   Action.Done(CGF);
2088 }
2089 
emitTaskyieldCall(CodeGenFunction & CGF,SourceLocation Loc)2090 void CGOpenMPRuntime::emitTaskyieldCall(CodeGenFunction &CGF,
2091                                         SourceLocation Loc) {
2092   if (!CGF.HaveInsertPoint())
2093     return;
2094   if (CGF.CGM.getLangOpts().OpenMPIRBuilder) {
2095     OMPBuilder.createTaskyield(CGF.Builder);
2096   } else {
2097     // Build call __kmpc_omp_taskyield(loc, thread_id, 0);
2098     llvm::Value *Args[] = {
2099         emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc),
2100         llvm::ConstantInt::get(CGM.IntTy, /*V=*/0, /*isSigned=*/true)};
2101     CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
2102                             CGM.getModule(), OMPRTL___kmpc_omp_taskyield),
2103                         Args);
2104   }
2105 
2106   if (auto *Region = dyn_cast_or_null<CGOpenMPRegionInfo>(CGF.CapturedStmtInfo))
2107     Region->emitUntiedSwitch(CGF);
2108 }
2109 
emitTaskgroupRegion(CodeGenFunction & CGF,const RegionCodeGenTy & TaskgroupOpGen,SourceLocation Loc)2110 void CGOpenMPRuntime::emitTaskgroupRegion(CodeGenFunction &CGF,
2111                                           const RegionCodeGenTy &TaskgroupOpGen,
2112                                           SourceLocation Loc) {
2113   if (!CGF.HaveInsertPoint())
2114     return;
2115   // __kmpc_taskgroup(ident_t *, gtid);
2116   // TaskgroupOpGen();
2117   // __kmpc_end_taskgroup(ident_t *, gtid);
2118   // Prepare arguments and build a call to __kmpc_taskgroup
2119   llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc)};
2120   CommonActionTy Action(OMPBuilder.getOrCreateRuntimeFunction(
2121                             CGM.getModule(), OMPRTL___kmpc_taskgroup),
2122                         Args,
2123                         OMPBuilder.getOrCreateRuntimeFunction(
2124                             CGM.getModule(), OMPRTL___kmpc_end_taskgroup),
2125                         Args);
2126   TaskgroupOpGen.setAction(Action);
2127   emitInlinedDirective(CGF, OMPD_taskgroup, TaskgroupOpGen);
2128 }
2129 
2130 /// Given an array of pointers to variables, project the address of a
2131 /// given variable.
emitAddrOfVarFromArray(CodeGenFunction & CGF,Address Array,unsigned Index,const VarDecl * Var)2132 static Address emitAddrOfVarFromArray(CodeGenFunction &CGF, Address Array,
2133                                       unsigned Index, const VarDecl *Var) {
2134   // Pull out the pointer to the variable.
2135   Address PtrAddr = CGF.Builder.CreateConstArrayGEP(Array, Index);
2136   llvm::Value *Ptr = CGF.Builder.CreateLoad(PtrAddr);
2137 
2138   llvm::Type *ElemTy = CGF.ConvertTypeForMem(Var->getType());
2139   return Address(
2140       CGF.Builder.CreateBitCast(
2141           Ptr, ElemTy->getPointerTo(Ptr->getType()->getPointerAddressSpace())),
2142       ElemTy, CGF.getContext().getDeclAlign(Var));
2143 }
2144 
emitCopyprivateCopyFunction(CodeGenModule & CGM,llvm::Type * ArgsElemType,ArrayRef<const Expr * > CopyprivateVars,ArrayRef<const Expr * > DestExprs,ArrayRef<const Expr * > SrcExprs,ArrayRef<const Expr * > AssignmentOps,SourceLocation Loc)2145 static llvm::Value *emitCopyprivateCopyFunction(
2146     CodeGenModule &CGM, llvm::Type *ArgsElemType,
2147     ArrayRef<const Expr *> CopyprivateVars, ArrayRef<const Expr *> DestExprs,
2148     ArrayRef<const Expr *> SrcExprs, ArrayRef<const Expr *> AssignmentOps,
2149     SourceLocation Loc) {
2150   ASTContext &C = CGM.getContext();
2151   // void copy_func(void *LHSArg, void *RHSArg);
2152   FunctionArgList Args;
2153   ImplicitParamDecl LHSArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, C.VoidPtrTy,
2154                            ImplicitParamKind::Other);
2155   ImplicitParamDecl RHSArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, C.VoidPtrTy,
2156                            ImplicitParamKind::Other);
2157   Args.push_back(&LHSArg);
2158   Args.push_back(&RHSArg);
2159   const auto &CGFI =
2160       CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args);
2161   std::string Name =
2162       CGM.getOpenMPRuntime().getName({"omp", "copyprivate", "copy_func"});
2163   auto *Fn = llvm::Function::Create(CGM.getTypes().GetFunctionType(CGFI),
2164                                     llvm::GlobalValue::InternalLinkage, Name,
2165                                     &CGM.getModule());
2166   CGM.SetInternalFunctionAttributes(GlobalDecl(), Fn, CGFI);
2167   Fn->setDoesNotRecurse();
2168   CodeGenFunction CGF(CGM);
2169   CGF.StartFunction(GlobalDecl(), C.VoidTy, Fn, CGFI, Args, Loc, Loc);
2170   // Dest = (void*[n])(LHSArg);
2171   // Src = (void*[n])(RHSArg);
2172   Address LHS(CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
2173                   CGF.Builder.CreateLoad(CGF.GetAddrOfLocalVar(&LHSArg)),
2174                   ArgsElemType->getPointerTo()),
2175               ArgsElemType, CGF.getPointerAlign());
2176   Address RHS(CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
2177                   CGF.Builder.CreateLoad(CGF.GetAddrOfLocalVar(&RHSArg)),
2178                   ArgsElemType->getPointerTo()),
2179               ArgsElemType, CGF.getPointerAlign());
2180   // *(Type0*)Dst[0] = *(Type0*)Src[0];
2181   // *(Type1*)Dst[1] = *(Type1*)Src[1];
2182   // ...
2183   // *(Typen*)Dst[n] = *(Typen*)Src[n];
2184   for (unsigned I = 0, E = AssignmentOps.size(); I < E; ++I) {
2185     const auto *DestVar =
2186         cast<VarDecl>(cast<DeclRefExpr>(DestExprs[I])->getDecl());
2187     Address DestAddr = emitAddrOfVarFromArray(CGF, LHS, I, DestVar);
2188 
2189     const auto *SrcVar =
2190         cast<VarDecl>(cast<DeclRefExpr>(SrcExprs[I])->getDecl());
2191     Address SrcAddr = emitAddrOfVarFromArray(CGF, RHS, I, SrcVar);
2192 
2193     const auto *VD = cast<DeclRefExpr>(CopyprivateVars[I])->getDecl();
2194     QualType Type = VD->getType();
2195     CGF.EmitOMPCopy(Type, DestAddr, SrcAddr, DestVar, SrcVar, AssignmentOps[I]);
2196   }
2197   CGF.FinishFunction();
2198   return Fn;
2199 }
2200 
emitSingleRegion(CodeGenFunction & CGF,const RegionCodeGenTy & SingleOpGen,SourceLocation Loc,ArrayRef<const Expr * > CopyprivateVars,ArrayRef<const Expr * > SrcExprs,ArrayRef<const Expr * > DstExprs,ArrayRef<const Expr * > AssignmentOps)2201 void CGOpenMPRuntime::emitSingleRegion(CodeGenFunction &CGF,
2202                                        const RegionCodeGenTy &SingleOpGen,
2203                                        SourceLocation Loc,
2204                                        ArrayRef<const Expr *> CopyprivateVars,
2205                                        ArrayRef<const Expr *> SrcExprs,
2206                                        ArrayRef<const Expr *> DstExprs,
2207                                        ArrayRef<const Expr *> AssignmentOps) {
2208   if (!CGF.HaveInsertPoint())
2209     return;
2210   assert(CopyprivateVars.size() == SrcExprs.size() &&
2211          CopyprivateVars.size() == DstExprs.size() &&
2212          CopyprivateVars.size() == AssignmentOps.size());
2213   ASTContext &C = CGM.getContext();
2214   // int32 did_it = 0;
2215   // if(__kmpc_single(ident_t *, gtid)) {
2216   //   SingleOpGen();
2217   //   __kmpc_end_single(ident_t *, gtid);
2218   //   did_it = 1;
2219   // }
2220   // call __kmpc_copyprivate(ident_t *, gtid, <buf_size>, <copyprivate list>,
2221   // <copy_func>, did_it);
2222 
2223   Address DidIt = Address::invalid();
2224   if (!CopyprivateVars.empty()) {
2225     // int32 did_it = 0;
2226     QualType KmpInt32Ty =
2227         C.getIntTypeForBitwidth(/*DestWidth=*/32, /*Signed=*/1);
2228     DidIt = CGF.CreateMemTemp(KmpInt32Ty, ".omp.copyprivate.did_it");
2229     CGF.Builder.CreateStore(CGF.Builder.getInt32(0), DidIt);
2230   }
2231   // Prepare arguments and build a call to __kmpc_single
2232   llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc)};
2233   CommonActionTy Action(OMPBuilder.getOrCreateRuntimeFunction(
2234                             CGM.getModule(), OMPRTL___kmpc_single),
2235                         Args,
2236                         OMPBuilder.getOrCreateRuntimeFunction(
2237                             CGM.getModule(), OMPRTL___kmpc_end_single),
2238                         Args,
2239                         /*Conditional=*/true);
2240   SingleOpGen.setAction(Action);
2241   emitInlinedDirective(CGF, OMPD_single, SingleOpGen);
2242   if (DidIt.isValid()) {
2243     // did_it = 1;
2244     CGF.Builder.CreateStore(CGF.Builder.getInt32(1), DidIt);
2245   }
2246   Action.Done(CGF);
2247   // call __kmpc_copyprivate(ident_t *, gtid, <buf_size>, <copyprivate list>,
2248   // <copy_func>, did_it);
2249   if (DidIt.isValid()) {
2250     llvm::APInt ArraySize(/*unsigned int numBits=*/32, CopyprivateVars.size());
2251     QualType CopyprivateArrayTy = C.getConstantArrayType(
2252         C.VoidPtrTy, ArraySize, nullptr, ArraySizeModifier::Normal,
2253         /*IndexTypeQuals=*/0);
2254     // Create a list of all private variables for copyprivate.
2255     Address CopyprivateList =
2256         CGF.CreateMemTemp(CopyprivateArrayTy, ".omp.copyprivate.cpr_list");
2257     for (unsigned I = 0, E = CopyprivateVars.size(); I < E; ++I) {
2258       Address Elem = CGF.Builder.CreateConstArrayGEP(CopyprivateList, I);
2259       CGF.Builder.CreateStore(
2260           CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
2261               CGF.EmitLValue(CopyprivateVars[I]).getPointer(CGF),
2262               CGF.VoidPtrTy),
2263           Elem);
2264     }
2265     // Build function that copies private values from single region to all other
2266     // threads in the corresponding parallel region.
2267     llvm::Value *CpyFn = emitCopyprivateCopyFunction(
2268         CGM, CGF.ConvertTypeForMem(CopyprivateArrayTy), CopyprivateVars,
2269         SrcExprs, DstExprs, AssignmentOps, Loc);
2270     llvm::Value *BufSize = CGF.getTypeSize(CopyprivateArrayTy);
2271     Address CL = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
2272         CopyprivateList, CGF.VoidPtrTy, CGF.Int8Ty);
2273     llvm::Value *DidItVal = CGF.Builder.CreateLoad(DidIt);
2274     llvm::Value *Args[] = {
2275         emitUpdateLocation(CGF, Loc), // ident_t *<loc>
2276         getThreadID(CGF, Loc),        // i32 <gtid>
2277         BufSize,                      // size_t <buf_size>
2278         CL.emitRawPointer(CGF),       // void *<copyprivate list>
2279         CpyFn,                        // void (*) (void *, void *) <copy_func>
2280         DidItVal                      // i32 did_it
2281     };
2282     CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
2283                             CGM.getModule(), OMPRTL___kmpc_copyprivate),
2284                         Args);
2285   }
2286 }
2287 
emitOrderedRegion(CodeGenFunction & CGF,const RegionCodeGenTy & OrderedOpGen,SourceLocation Loc,bool IsThreads)2288 void CGOpenMPRuntime::emitOrderedRegion(CodeGenFunction &CGF,
2289                                         const RegionCodeGenTy &OrderedOpGen,
2290                                         SourceLocation Loc, bool IsThreads) {
2291   if (!CGF.HaveInsertPoint())
2292     return;
2293   // __kmpc_ordered(ident_t *, gtid);
2294   // OrderedOpGen();
2295   // __kmpc_end_ordered(ident_t *, gtid);
2296   // Prepare arguments and build a call to __kmpc_ordered
2297   if (IsThreads) {
2298     llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc)};
2299     CommonActionTy Action(OMPBuilder.getOrCreateRuntimeFunction(
2300                               CGM.getModule(), OMPRTL___kmpc_ordered),
2301                           Args,
2302                           OMPBuilder.getOrCreateRuntimeFunction(
2303                               CGM.getModule(), OMPRTL___kmpc_end_ordered),
2304                           Args);
2305     OrderedOpGen.setAction(Action);
2306     emitInlinedDirective(CGF, OMPD_ordered, OrderedOpGen);
2307     return;
2308   }
2309   emitInlinedDirective(CGF, OMPD_ordered, OrderedOpGen);
2310 }
2311 
getDefaultFlagsForBarriers(OpenMPDirectiveKind Kind)2312 unsigned CGOpenMPRuntime::getDefaultFlagsForBarriers(OpenMPDirectiveKind Kind) {
2313   unsigned Flags;
2314   if (Kind == OMPD_for)
2315     Flags = OMP_IDENT_BARRIER_IMPL_FOR;
2316   else if (Kind == OMPD_sections)
2317     Flags = OMP_IDENT_BARRIER_IMPL_SECTIONS;
2318   else if (Kind == OMPD_single)
2319     Flags = OMP_IDENT_BARRIER_IMPL_SINGLE;
2320   else if (Kind == OMPD_barrier)
2321     Flags = OMP_IDENT_BARRIER_EXPL;
2322   else
2323     Flags = OMP_IDENT_BARRIER_IMPL;
2324   return Flags;
2325 }
2326 
getDefaultScheduleAndChunk(CodeGenFunction & CGF,const OMPLoopDirective & S,OpenMPScheduleClauseKind & ScheduleKind,const Expr * & ChunkExpr) const2327 void CGOpenMPRuntime::getDefaultScheduleAndChunk(
2328     CodeGenFunction &CGF, const OMPLoopDirective &S,
2329     OpenMPScheduleClauseKind &ScheduleKind, const Expr *&ChunkExpr) const {
2330   // Check if the loop directive is actually a doacross loop directive. In this
2331   // case choose static, 1 schedule.
2332   if (llvm::any_of(
2333           S.getClausesOfKind<OMPOrderedClause>(),
2334           [](const OMPOrderedClause *C) { return C->getNumForLoops(); })) {
2335     ScheduleKind = OMPC_SCHEDULE_static;
2336     // Chunk size is 1 in this case.
2337     llvm::APInt ChunkSize(32, 1);
2338     ChunkExpr = IntegerLiteral::Create(
2339         CGF.getContext(), ChunkSize,
2340         CGF.getContext().getIntTypeForBitwidth(32, /*Signed=*/0),
2341         SourceLocation());
2342   }
2343 }
2344 
emitBarrierCall(CodeGenFunction & CGF,SourceLocation Loc,OpenMPDirectiveKind Kind,bool EmitChecks,bool ForceSimpleCall)2345 void CGOpenMPRuntime::emitBarrierCall(CodeGenFunction &CGF, SourceLocation Loc,
2346                                       OpenMPDirectiveKind Kind, bool EmitChecks,
2347                                       bool ForceSimpleCall) {
2348   // Check if we should use the OMPBuilder
2349   auto *OMPRegionInfo =
2350       dyn_cast_or_null<CGOpenMPRegionInfo>(CGF.CapturedStmtInfo);
2351   if (CGF.CGM.getLangOpts().OpenMPIRBuilder) {
2352     CGF.Builder.restoreIP(OMPBuilder.createBarrier(
2353         CGF.Builder, Kind, ForceSimpleCall, EmitChecks));
2354     return;
2355   }
2356 
2357   if (!CGF.HaveInsertPoint())
2358     return;
2359   // Build call __kmpc_cancel_barrier(loc, thread_id);
2360   // Build call __kmpc_barrier(loc, thread_id);
2361   unsigned Flags = getDefaultFlagsForBarriers(Kind);
2362   // Build call __kmpc_cancel_barrier(loc, thread_id) or __kmpc_barrier(loc,
2363   // thread_id);
2364   llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc, Flags),
2365                          getThreadID(CGF, Loc)};
2366   if (OMPRegionInfo) {
2367     if (!ForceSimpleCall && OMPRegionInfo->hasCancel()) {
2368       llvm::Value *Result = CGF.EmitRuntimeCall(
2369           OMPBuilder.getOrCreateRuntimeFunction(CGM.getModule(),
2370                                                 OMPRTL___kmpc_cancel_barrier),
2371           Args);
2372       if (EmitChecks) {
2373         // if (__kmpc_cancel_barrier()) {
2374         //   exit from construct;
2375         // }
2376         llvm::BasicBlock *ExitBB = CGF.createBasicBlock(".cancel.exit");
2377         llvm::BasicBlock *ContBB = CGF.createBasicBlock(".cancel.continue");
2378         llvm::Value *Cmp = CGF.Builder.CreateIsNotNull(Result);
2379         CGF.Builder.CreateCondBr(Cmp, ExitBB, ContBB);
2380         CGF.EmitBlock(ExitBB);
2381         //   exit from construct;
2382         CodeGenFunction::JumpDest CancelDestination =
2383             CGF.getOMPCancelDestination(OMPRegionInfo->getDirectiveKind());
2384         CGF.EmitBranchThroughCleanup(CancelDestination);
2385         CGF.EmitBlock(ContBB, /*IsFinished=*/true);
2386       }
2387       return;
2388     }
2389   }
2390   CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
2391                           CGM.getModule(), OMPRTL___kmpc_barrier),
2392                       Args);
2393 }
2394 
emitErrorCall(CodeGenFunction & CGF,SourceLocation Loc,Expr * ME,bool IsFatal)2395 void CGOpenMPRuntime::emitErrorCall(CodeGenFunction &CGF, SourceLocation Loc,
2396                                     Expr *ME, bool IsFatal) {
2397   llvm::Value *MVL =
2398       ME ? CGF.EmitStringLiteralLValue(cast<StringLiteral>(ME)).getPointer(CGF)
2399          : llvm::ConstantPointerNull::get(CGF.VoidPtrTy);
2400   // Build call void __kmpc_error(ident_t *loc, int severity, const char
2401   // *message)
2402   llvm::Value *Args[] = {
2403       emitUpdateLocation(CGF, Loc, /*Flags=*/0, /*GenLoc=*/true),
2404       llvm::ConstantInt::get(CGM.Int32Ty, IsFatal ? 2 : 1),
2405       CGF.Builder.CreatePointerCast(MVL, CGM.Int8PtrTy)};
2406   CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
2407                           CGM.getModule(), OMPRTL___kmpc_error),
2408                       Args);
2409 }
2410 
2411 /// Map the OpenMP loop schedule to the runtime enumeration.
getRuntimeSchedule(OpenMPScheduleClauseKind ScheduleKind,bool Chunked,bool Ordered)2412 static OpenMPSchedType getRuntimeSchedule(OpenMPScheduleClauseKind ScheduleKind,
2413                                           bool Chunked, bool Ordered) {
2414   switch (ScheduleKind) {
2415   case OMPC_SCHEDULE_static:
2416     return Chunked ? (Ordered ? OMP_ord_static_chunked : OMP_sch_static_chunked)
2417                    : (Ordered ? OMP_ord_static : OMP_sch_static);
2418   case OMPC_SCHEDULE_dynamic:
2419     return Ordered ? OMP_ord_dynamic_chunked : OMP_sch_dynamic_chunked;
2420   case OMPC_SCHEDULE_guided:
2421     return Ordered ? OMP_ord_guided_chunked : OMP_sch_guided_chunked;
2422   case OMPC_SCHEDULE_runtime:
2423     return Ordered ? OMP_ord_runtime : OMP_sch_runtime;
2424   case OMPC_SCHEDULE_auto:
2425     return Ordered ? OMP_ord_auto : OMP_sch_auto;
2426   case OMPC_SCHEDULE_unknown:
2427     assert(!Chunked && "chunk was specified but schedule kind not known");
2428     return Ordered ? OMP_ord_static : OMP_sch_static;
2429   }
2430   llvm_unreachable("Unexpected runtime schedule");
2431 }
2432 
2433 /// Map the OpenMP distribute schedule to the runtime enumeration.
2434 static OpenMPSchedType
getRuntimeSchedule(OpenMPDistScheduleClauseKind ScheduleKind,bool Chunked)2435 getRuntimeSchedule(OpenMPDistScheduleClauseKind ScheduleKind, bool Chunked) {
2436   // only static is allowed for dist_schedule
2437   return Chunked ? OMP_dist_sch_static_chunked : OMP_dist_sch_static;
2438 }
2439 
isStaticNonchunked(OpenMPScheduleClauseKind ScheduleKind,bool Chunked) const2440 bool CGOpenMPRuntime::isStaticNonchunked(OpenMPScheduleClauseKind ScheduleKind,
2441                                          bool Chunked) const {
2442   OpenMPSchedType Schedule =
2443       getRuntimeSchedule(ScheduleKind, Chunked, /*Ordered=*/false);
2444   return Schedule == OMP_sch_static;
2445 }
2446 
isStaticNonchunked(OpenMPDistScheduleClauseKind ScheduleKind,bool Chunked) const2447 bool CGOpenMPRuntime::isStaticNonchunked(
2448     OpenMPDistScheduleClauseKind ScheduleKind, bool Chunked) const {
2449   OpenMPSchedType Schedule = getRuntimeSchedule(ScheduleKind, Chunked);
2450   return Schedule == OMP_dist_sch_static;
2451 }
2452 
isStaticChunked(OpenMPScheduleClauseKind ScheduleKind,bool Chunked) const2453 bool CGOpenMPRuntime::isStaticChunked(OpenMPScheduleClauseKind ScheduleKind,
2454                                       bool Chunked) const {
2455   OpenMPSchedType Schedule =
2456       getRuntimeSchedule(ScheduleKind, Chunked, /*Ordered=*/false);
2457   return Schedule == OMP_sch_static_chunked;
2458 }
2459 
isStaticChunked(OpenMPDistScheduleClauseKind ScheduleKind,bool Chunked) const2460 bool CGOpenMPRuntime::isStaticChunked(
2461     OpenMPDistScheduleClauseKind ScheduleKind, bool Chunked) const {
2462   OpenMPSchedType Schedule = getRuntimeSchedule(ScheduleKind, Chunked);
2463   return Schedule == OMP_dist_sch_static_chunked;
2464 }
2465 
isDynamic(OpenMPScheduleClauseKind ScheduleKind) const2466 bool CGOpenMPRuntime::isDynamic(OpenMPScheduleClauseKind ScheduleKind) const {
2467   OpenMPSchedType Schedule =
2468       getRuntimeSchedule(ScheduleKind, /*Chunked=*/false, /*Ordered=*/false);
2469   assert(Schedule != OMP_sch_static_chunked && "cannot be chunked here");
2470   return Schedule != OMP_sch_static;
2471 }
2472 
addMonoNonMonoModifier(CodeGenModule & CGM,OpenMPSchedType Schedule,OpenMPScheduleClauseModifier M1,OpenMPScheduleClauseModifier M2)2473 static int addMonoNonMonoModifier(CodeGenModule &CGM, OpenMPSchedType Schedule,
2474                                   OpenMPScheduleClauseModifier M1,
2475                                   OpenMPScheduleClauseModifier M2) {
2476   int Modifier = 0;
2477   switch (M1) {
2478   case OMPC_SCHEDULE_MODIFIER_monotonic:
2479     Modifier = OMP_sch_modifier_monotonic;
2480     break;
2481   case OMPC_SCHEDULE_MODIFIER_nonmonotonic:
2482     Modifier = OMP_sch_modifier_nonmonotonic;
2483     break;
2484   case OMPC_SCHEDULE_MODIFIER_simd:
2485     if (Schedule == OMP_sch_static_chunked)
2486       Schedule = OMP_sch_static_balanced_chunked;
2487     break;
2488   case OMPC_SCHEDULE_MODIFIER_last:
2489   case OMPC_SCHEDULE_MODIFIER_unknown:
2490     break;
2491   }
2492   switch (M2) {
2493   case OMPC_SCHEDULE_MODIFIER_monotonic:
2494     Modifier = OMP_sch_modifier_monotonic;
2495     break;
2496   case OMPC_SCHEDULE_MODIFIER_nonmonotonic:
2497     Modifier = OMP_sch_modifier_nonmonotonic;
2498     break;
2499   case OMPC_SCHEDULE_MODIFIER_simd:
2500     if (Schedule == OMP_sch_static_chunked)
2501       Schedule = OMP_sch_static_balanced_chunked;
2502     break;
2503   case OMPC_SCHEDULE_MODIFIER_last:
2504   case OMPC_SCHEDULE_MODIFIER_unknown:
2505     break;
2506   }
2507   // OpenMP 5.0, 2.9.2 Worksharing-Loop Construct, Desription.
2508   // If the static schedule kind is specified or if the ordered clause is
2509   // specified, and if the nonmonotonic modifier is not specified, the effect is
2510   // as if the monotonic modifier is specified. Otherwise, unless the monotonic
2511   // modifier is specified, the effect is as if the nonmonotonic modifier is
2512   // specified.
2513   if (CGM.getLangOpts().OpenMP >= 50 && Modifier == 0) {
2514     if (!(Schedule == OMP_sch_static_chunked || Schedule == OMP_sch_static ||
2515           Schedule == OMP_sch_static_balanced_chunked ||
2516           Schedule == OMP_ord_static_chunked || Schedule == OMP_ord_static ||
2517           Schedule == OMP_dist_sch_static_chunked ||
2518           Schedule == OMP_dist_sch_static))
2519       Modifier = OMP_sch_modifier_nonmonotonic;
2520   }
2521   return Schedule | Modifier;
2522 }
2523 
emitForDispatchInit(CodeGenFunction & CGF,SourceLocation Loc,const OpenMPScheduleTy & ScheduleKind,unsigned IVSize,bool IVSigned,bool Ordered,const DispatchRTInput & DispatchValues)2524 void CGOpenMPRuntime::emitForDispatchInit(
2525     CodeGenFunction &CGF, SourceLocation Loc,
2526     const OpenMPScheduleTy &ScheduleKind, unsigned IVSize, bool IVSigned,
2527     bool Ordered, const DispatchRTInput &DispatchValues) {
2528   if (!CGF.HaveInsertPoint())
2529     return;
2530   OpenMPSchedType Schedule = getRuntimeSchedule(
2531       ScheduleKind.Schedule, DispatchValues.Chunk != nullptr, Ordered);
2532   assert(Ordered ||
2533          (Schedule != OMP_sch_static && Schedule != OMP_sch_static_chunked &&
2534           Schedule != OMP_ord_static && Schedule != OMP_ord_static_chunked &&
2535           Schedule != OMP_sch_static_balanced_chunked));
2536   // Call __kmpc_dispatch_init(
2537   //          ident_t *loc, kmp_int32 tid, kmp_int32 schedule,
2538   //          kmp_int[32|64] lower, kmp_int[32|64] upper,
2539   //          kmp_int[32|64] stride, kmp_int[32|64] chunk);
2540 
2541   // If the Chunk was not specified in the clause - use default value 1.
2542   llvm::Value *Chunk = DispatchValues.Chunk ? DispatchValues.Chunk
2543                                             : CGF.Builder.getIntN(IVSize, 1);
2544   llvm::Value *Args[] = {
2545       emitUpdateLocation(CGF, Loc),
2546       getThreadID(CGF, Loc),
2547       CGF.Builder.getInt32(addMonoNonMonoModifier(
2548           CGM, Schedule, ScheduleKind.M1, ScheduleKind.M2)), // Schedule type
2549       DispatchValues.LB,                                     // Lower
2550       DispatchValues.UB,                                     // Upper
2551       CGF.Builder.getIntN(IVSize, 1),                        // Stride
2552       Chunk                                                  // Chunk
2553   };
2554   CGF.EmitRuntimeCall(OMPBuilder.createDispatchInitFunction(IVSize, IVSigned),
2555                       Args);
2556 }
2557 
emitForDispatchDeinit(CodeGenFunction & CGF,SourceLocation Loc)2558 void CGOpenMPRuntime::emitForDispatchDeinit(CodeGenFunction &CGF,
2559                                             SourceLocation Loc) {
2560   if (!CGF.HaveInsertPoint())
2561     return;
2562   // Call __kmpc_dispatch_deinit(ident_t *loc, kmp_int32 tid);
2563   llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc)};
2564   CGF.EmitRuntimeCall(OMPBuilder.createDispatchDeinitFunction(), Args);
2565 }
2566 
emitForStaticInitCall(CodeGenFunction & CGF,llvm::Value * UpdateLocation,llvm::Value * ThreadId,llvm::FunctionCallee ForStaticInitFunction,OpenMPSchedType Schedule,OpenMPScheduleClauseModifier M1,OpenMPScheduleClauseModifier M2,const CGOpenMPRuntime::StaticRTInput & Values)2567 static void emitForStaticInitCall(
2568     CodeGenFunction &CGF, llvm::Value *UpdateLocation, llvm::Value *ThreadId,
2569     llvm::FunctionCallee ForStaticInitFunction, OpenMPSchedType Schedule,
2570     OpenMPScheduleClauseModifier M1, OpenMPScheduleClauseModifier M2,
2571     const CGOpenMPRuntime::StaticRTInput &Values) {
2572   if (!CGF.HaveInsertPoint())
2573     return;
2574 
2575   assert(!Values.Ordered);
2576   assert(Schedule == OMP_sch_static || Schedule == OMP_sch_static_chunked ||
2577          Schedule == OMP_sch_static_balanced_chunked ||
2578          Schedule == OMP_ord_static || Schedule == OMP_ord_static_chunked ||
2579          Schedule == OMP_dist_sch_static ||
2580          Schedule == OMP_dist_sch_static_chunked);
2581 
2582   // Call __kmpc_for_static_init(
2583   //          ident_t *loc, kmp_int32 tid, kmp_int32 schedtype,
2584   //          kmp_int32 *p_lastiter, kmp_int[32|64] *p_lower,
2585   //          kmp_int[32|64] *p_upper, kmp_int[32|64] *p_stride,
2586   //          kmp_int[32|64] incr, kmp_int[32|64] chunk);
2587   llvm::Value *Chunk = Values.Chunk;
2588   if (Chunk == nullptr) {
2589     assert((Schedule == OMP_sch_static || Schedule == OMP_ord_static ||
2590             Schedule == OMP_dist_sch_static) &&
2591            "expected static non-chunked schedule");
2592     // If the Chunk was not specified in the clause - use default value 1.
2593     Chunk = CGF.Builder.getIntN(Values.IVSize, 1);
2594   } else {
2595     assert((Schedule == OMP_sch_static_chunked ||
2596             Schedule == OMP_sch_static_balanced_chunked ||
2597             Schedule == OMP_ord_static_chunked ||
2598             Schedule == OMP_dist_sch_static_chunked) &&
2599            "expected static chunked schedule");
2600   }
2601   llvm::Value *Args[] = {
2602       UpdateLocation,
2603       ThreadId,
2604       CGF.Builder.getInt32(addMonoNonMonoModifier(CGF.CGM, Schedule, M1,
2605                                                   M2)), // Schedule type
2606       Values.IL.emitRawPointer(CGF),                    // &isLastIter
2607       Values.LB.emitRawPointer(CGF),                    // &LB
2608       Values.UB.emitRawPointer(CGF),                    // &UB
2609       Values.ST.emitRawPointer(CGF),                    // &Stride
2610       CGF.Builder.getIntN(Values.IVSize, 1),            // Incr
2611       Chunk                                             // Chunk
2612   };
2613   CGF.EmitRuntimeCall(ForStaticInitFunction, Args);
2614 }
2615 
emitForStaticInit(CodeGenFunction & CGF,SourceLocation Loc,OpenMPDirectiveKind DKind,const OpenMPScheduleTy & ScheduleKind,const StaticRTInput & Values)2616 void CGOpenMPRuntime::emitForStaticInit(CodeGenFunction &CGF,
2617                                         SourceLocation Loc,
2618                                         OpenMPDirectiveKind DKind,
2619                                         const OpenMPScheduleTy &ScheduleKind,
2620                                         const StaticRTInput &Values) {
2621   OpenMPSchedType ScheduleNum = getRuntimeSchedule(
2622       ScheduleKind.Schedule, Values.Chunk != nullptr, Values.Ordered);
2623   assert((isOpenMPWorksharingDirective(DKind) || (DKind == OMPD_loop)) &&
2624          "Expected loop-based or sections-based directive.");
2625   llvm::Value *UpdatedLocation = emitUpdateLocation(CGF, Loc,
2626                                              isOpenMPLoopDirective(DKind)
2627                                                  ? OMP_IDENT_WORK_LOOP
2628                                                  : OMP_IDENT_WORK_SECTIONS);
2629   llvm::Value *ThreadId = getThreadID(CGF, Loc);
2630   llvm::FunctionCallee StaticInitFunction =
2631       OMPBuilder.createForStaticInitFunction(Values.IVSize, Values.IVSigned,
2632                                              false);
2633   auto DL = ApplyDebugLocation::CreateDefaultArtificial(CGF, Loc);
2634   emitForStaticInitCall(CGF, UpdatedLocation, ThreadId, StaticInitFunction,
2635                         ScheduleNum, ScheduleKind.M1, ScheduleKind.M2, Values);
2636 }
2637 
emitDistributeStaticInit(CodeGenFunction & CGF,SourceLocation Loc,OpenMPDistScheduleClauseKind SchedKind,const CGOpenMPRuntime::StaticRTInput & Values)2638 void CGOpenMPRuntime::emitDistributeStaticInit(
2639     CodeGenFunction &CGF, SourceLocation Loc,
2640     OpenMPDistScheduleClauseKind SchedKind,
2641     const CGOpenMPRuntime::StaticRTInput &Values) {
2642   OpenMPSchedType ScheduleNum =
2643       getRuntimeSchedule(SchedKind, Values.Chunk != nullptr);
2644   llvm::Value *UpdatedLocation =
2645       emitUpdateLocation(CGF, Loc, OMP_IDENT_WORK_DISTRIBUTE);
2646   llvm::Value *ThreadId = getThreadID(CGF, Loc);
2647   llvm::FunctionCallee StaticInitFunction;
2648   bool isGPUDistribute =
2649       CGM.getLangOpts().OpenMPIsTargetDevice &&
2650       (CGM.getTriple().isAMDGCN() || CGM.getTriple().isNVPTX());
2651   StaticInitFunction = OMPBuilder.createForStaticInitFunction(
2652       Values.IVSize, Values.IVSigned, isGPUDistribute);
2653 
2654   emitForStaticInitCall(CGF, UpdatedLocation, ThreadId, StaticInitFunction,
2655                         ScheduleNum, OMPC_SCHEDULE_MODIFIER_unknown,
2656                         OMPC_SCHEDULE_MODIFIER_unknown, Values);
2657 }
2658 
emitForStaticFinish(CodeGenFunction & CGF,SourceLocation Loc,OpenMPDirectiveKind DKind)2659 void CGOpenMPRuntime::emitForStaticFinish(CodeGenFunction &CGF,
2660                                           SourceLocation Loc,
2661                                           OpenMPDirectiveKind DKind) {
2662   assert((DKind == OMPD_distribute || DKind == OMPD_for ||
2663           DKind == OMPD_sections) &&
2664          "Expected distribute, for, or sections directive kind");
2665   if (!CGF.HaveInsertPoint())
2666     return;
2667   // Call __kmpc_for_static_fini(ident_t *loc, kmp_int32 tid);
2668   llvm::Value *Args[] = {
2669       emitUpdateLocation(CGF, Loc,
2670                          isOpenMPDistributeDirective(DKind) ||
2671                                  (DKind == OMPD_target_teams_loop)
2672                              ? OMP_IDENT_WORK_DISTRIBUTE
2673                          : isOpenMPLoopDirective(DKind)
2674                              ? OMP_IDENT_WORK_LOOP
2675                              : OMP_IDENT_WORK_SECTIONS),
2676       getThreadID(CGF, Loc)};
2677   auto DL = ApplyDebugLocation::CreateDefaultArtificial(CGF, Loc);
2678   if (isOpenMPDistributeDirective(DKind) &&
2679       CGM.getLangOpts().OpenMPIsTargetDevice &&
2680       (CGM.getTriple().isAMDGCN() || CGM.getTriple().isNVPTX()))
2681     CGF.EmitRuntimeCall(
2682         OMPBuilder.getOrCreateRuntimeFunction(
2683             CGM.getModule(), OMPRTL___kmpc_distribute_static_fini),
2684         Args);
2685   else
2686     CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
2687                             CGM.getModule(), OMPRTL___kmpc_for_static_fini),
2688                         Args);
2689 }
2690 
emitForOrderedIterationEnd(CodeGenFunction & CGF,SourceLocation Loc,unsigned IVSize,bool IVSigned)2691 void CGOpenMPRuntime::emitForOrderedIterationEnd(CodeGenFunction &CGF,
2692                                                  SourceLocation Loc,
2693                                                  unsigned IVSize,
2694                                                  bool IVSigned) {
2695   if (!CGF.HaveInsertPoint())
2696     return;
2697   // Call __kmpc_for_dynamic_fini_(4|8)[u](ident_t *loc, kmp_int32 tid);
2698   llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc)};
2699   CGF.EmitRuntimeCall(OMPBuilder.createDispatchFiniFunction(IVSize, IVSigned),
2700                       Args);
2701 }
2702 
emitForNext(CodeGenFunction & CGF,SourceLocation Loc,unsigned IVSize,bool IVSigned,Address IL,Address LB,Address UB,Address ST)2703 llvm::Value *CGOpenMPRuntime::emitForNext(CodeGenFunction &CGF,
2704                                           SourceLocation Loc, unsigned IVSize,
2705                                           bool IVSigned, Address IL,
2706                                           Address LB, Address UB,
2707                                           Address ST) {
2708   // Call __kmpc_dispatch_next(
2709   //          ident_t *loc, kmp_int32 tid, kmp_int32 *p_lastiter,
2710   //          kmp_int[32|64] *p_lower, kmp_int[32|64] *p_upper,
2711   //          kmp_int[32|64] *p_stride);
2712   llvm::Value *Args[] = {
2713       emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc),
2714       IL.emitRawPointer(CGF), // &isLastIter
2715       LB.emitRawPointer(CGF), // &Lower
2716       UB.emitRawPointer(CGF), // &Upper
2717       ST.emitRawPointer(CGF)  // &Stride
2718   };
2719   llvm::Value *Call = CGF.EmitRuntimeCall(
2720       OMPBuilder.createDispatchNextFunction(IVSize, IVSigned), Args);
2721   return CGF.EmitScalarConversion(
2722       Call, CGF.getContext().getIntTypeForBitwidth(32, /*Signed=*/1),
2723       CGF.getContext().BoolTy, Loc);
2724 }
2725 
emitNumThreadsClause(CodeGenFunction & CGF,llvm::Value * NumThreads,SourceLocation Loc)2726 void CGOpenMPRuntime::emitNumThreadsClause(CodeGenFunction &CGF,
2727                                            llvm::Value *NumThreads,
2728                                            SourceLocation Loc) {
2729   if (!CGF.HaveInsertPoint())
2730     return;
2731   // Build call __kmpc_push_num_threads(&loc, global_tid, num_threads)
2732   llvm::Value *Args[] = {
2733       emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc),
2734       CGF.Builder.CreateIntCast(NumThreads, CGF.Int32Ty, /*isSigned*/ true)};
2735   CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
2736                           CGM.getModule(), OMPRTL___kmpc_push_num_threads),
2737                       Args);
2738 }
2739 
emitProcBindClause(CodeGenFunction & CGF,ProcBindKind ProcBind,SourceLocation Loc)2740 void CGOpenMPRuntime::emitProcBindClause(CodeGenFunction &CGF,
2741                                          ProcBindKind ProcBind,
2742                                          SourceLocation Loc) {
2743   if (!CGF.HaveInsertPoint())
2744     return;
2745   assert(ProcBind != OMP_PROC_BIND_unknown && "Unsupported proc_bind value.");
2746   // Build call __kmpc_push_proc_bind(&loc, global_tid, proc_bind)
2747   llvm::Value *Args[] = {
2748       emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc),
2749       llvm::ConstantInt::get(CGM.IntTy, unsigned(ProcBind), /*isSigned=*/true)};
2750   CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
2751                           CGM.getModule(), OMPRTL___kmpc_push_proc_bind),
2752                       Args);
2753 }
2754 
emitFlush(CodeGenFunction & CGF,ArrayRef<const Expr * >,SourceLocation Loc,llvm::AtomicOrdering AO)2755 void CGOpenMPRuntime::emitFlush(CodeGenFunction &CGF, ArrayRef<const Expr *>,
2756                                 SourceLocation Loc, llvm::AtomicOrdering AO) {
2757   if (CGF.CGM.getLangOpts().OpenMPIRBuilder) {
2758     OMPBuilder.createFlush(CGF.Builder);
2759   } else {
2760     if (!CGF.HaveInsertPoint())
2761       return;
2762     // Build call void __kmpc_flush(ident_t *loc)
2763     CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
2764                             CGM.getModule(), OMPRTL___kmpc_flush),
2765                         emitUpdateLocation(CGF, Loc));
2766   }
2767 }
2768 
2769 namespace {
2770 /// Indexes of fields for type kmp_task_t.
2771 enum KmpTaskTFields {
2772   /// List of shared variables.
2773   KmpTaskTShareds,
2774   /// Task routine.
2775   KmpTaskTRoutine,
2776   /// Partition id for the untied tasks.
2777   KmpTaskTPartId,
2778   /// Function with call of destructors for private variables.
2779   Data1,
2780   /// Task priority.
2781   Data2,
2782   /// (Taskloops only) Lower bound.
2783   KmpTaskTLowerBound,
2784   /// (Taskloops only) Upper bound.
2785   KmpTaskTUpperBound,
2786   /// (Taskloops only) Stride.
2787   KmpTaskTStride,
2788   /// (Taskloops only) Is last iteration flag.
2789   KmpTaskTLastIter,
2790   /// (Taskloops only) Reduction data.
2791   KmpTaskTReductions,
2792 };
2793 } // anonymous namespace
2794 
createOffloadEntriesAndInfoMetadata()2795 void CGOpenMPRuntime::createOffloadEntriesAndInfoMetadata() {
2796   // If we are in simd mode or there are no entries, we don't need to do
2797   // anything.
2798   if (CGM.getLangOpts().OpenMPSimd || OMPBuilder.OffloadInfoManager.empty())
2799     return;
2800 
2801   llvm::OpenMPIRBuilder::EmitMetadataErrorReportFunctionTy &&ErrorReportFn =
2802       [this](llvm::OpenMPIRBuilder::EmitMetadataErrorKind Kind,
2803              const llvm::TargetRegionEntryInfo &EntryInfo) -> void {
2804     SourceLocation Loc;
2805     if (Kind != llvm::OpenMPIRBuilder::EMIT_MD_GLOBAL_VAR_LINK_ERROR) {
2806       for (auto I = CGM.getContext().getSourceManager().fileinfo_begin(),
2807                 E = CGM.getContext().getSourceManager().fileinfo_end();
2808            I != E; ++I) {
2809         if (I->getFirst().getUniqueID().getDevice() == EntryInfo.DeviceID &&
2810             I->getFirst().getUniqueID().getFile() == EntryInfo.FileID) {
2811           Loc = CGM.getContext().getSourceManager().translateFileLineCol(
2812               I->getFirst(), EntryInfo.Line, 1);
2813           break;
2814         }
2815       }
2816     }
2817     switch (Kind) {
2818     case llvm::OpenMPIRBuilder::EMIT_MD_TARGET_REGION_ERROR: {
2819       unsigned DiagID = CGM.getDiags().getCustomDiagID(
2820           DiagnosticsEngine::Error, "Offloading entry for target region in "
2821                                     "%0 is incorrect: either the "
2822                                     "address or the ID is invalid.");
2823       CGM.getDiags().Report(Loc, DiagID) << EntryInfo.ParentName;
2824     } break;
2825     case llvm::OpenMPIRBuilder::EMIT_MD_DECLARE_TARGET_ERROR: {
2826       unsigned DiagID = CGM.getDiags().getCustomDiagID(
2827           DiagnosticsEngine::Error, "Offloading entry for declare target "
2828                                     "variable %0 is incorrect: the "
2829                                     "address is invalid.");
2830       CGM.getDiags().Report(Loc, DiagID) << EntryInfo.ParentName;
2831     } break;
2832     case llvm::OpenMPIRBuilder::EMIT_MD_GLOBAL_VAR_LINK_ERROR: {
2833       unsigned DiagID = CGM.getDiags().getCustomDiagID(
2834           DiagnosticsEngine::Error,
2835           "Offloading entry for declare target variable is incorrect: the "
2836           "address is invalid.");
2837       CGM.getDiags().Report(DiagID);
2838     } break;
2839     }
2840   };
2841 
2842   OMPBuilder.createOffloadEntriesAndInfoMetadata(ErrorReportFn);
2843 }
2844 
emitKmpRoutineEntryT(QualType KmpInt32Ty)2845 void CGOpenMPRuntime::emitKmpRoutineEntryT(QualType KmpInt32Ty) {
2846   if (!KmpRoutineEntryPtrTy) {
2847     // Build typedef kmp_int32 (* kmp_routine_entry_t)(kmp_int32, void *); type.
2848     ASTContext &C = CGM.getContext();
2849     QualType KmpRoutineEntryTyArgs[] = {KmpInt32Ty, C.VoidPtrTy};
2850     FunctionProtoType::ExtProtoInfo EPI;
2851     KmpRoutineEntryPtrQTy = C.getPointerType(
2852         C.getFunctionType(KmpInt32Ty, KmpRoutineEntryTyArgs, EPI));
2853     KmpRoutineEntryPtrTy = CGM.getTypes().ConvertType(KmpRoutineEntryPtrQTy);
2854   }
2855 }
2856 
2857 namespace {
2858 struct PrivateHelpersTy {
PrivateHelpersTy__anon93cce0fb0e11::PrivateHelpersTy2859   PrivateHelpersTy(const Expr *OriginalRef, const VarDecl *Original,
2860                    const VarDecl *PrivateCopy, const VarDecl *PrivateElemInit)
2861       : OriginalRef(OriginalRef), Original(Original), PrivateCopy(PrivateCopy),
2862         PrivateElemInit(PrivateElemInit) {}
PrivateHelpersTy__anon93cce0fb0e11::PrivateHelpersTy2863   PrivateHelpersTy(const VarDecl *Original) : Original(Original) {}
2864   const Expr *OriginalRef = nullptr;
2865   const VarDecl *Original = nullptr;
2866   const VarDecl *PrivateCopy = nullptr;
2867   const VarDecl *PrivateElemInit = nullptr;
isLocalPrivate__anon93cce0fb0e11::PrivateHelpersTy2868   bool isLocalPrivate() const {
2869     return !OriginalRef && !PrivateCopy && !PrivateElemInit;
2870   }
2871 };
2872 typedef std::pair<CharUnits /*Align*/, PrivateHelpersTy> PrivateDataTy;
2873 } // anonymous namespace
2874 
isAllocatableDecl(const VarDecl * VD)2875 static bool isAllocatableDecl(const VarDecl *VD) {
2876   const VarDecl *CVD = VD->getCanonicalDecl();
2877   if (!CVD->hasAttr<OMPAllocateDeclAttr>())
2878     return false;
2879   const auto *AA = CVD->getAttr<OMPAllocateDeclAttr>();
2880   // Use the default allocation.
2881   return !(AA->getAllocatorType() == OMPAllocateDeclAttr::OMPDefaultMemAlloc &&
2882            !AA->getAllocator());
2883 }
2884 
2885 static RecordDecl *
createPrivatesRecordDecl(CodeGenModule & CGM,ArrayRef<PrivateDataTy> Privates)2886 createPrivatesRecordDecl(CodeGenModule &CGM, ArrayRef<PrivateDataTy> Privates) {
2887   if (!Privates.empty()) {
2888     ASTContext &C = CGM.getContext();
2889     // Build struct .kmp_privates_t. {
2890     //         /*  private vars  */
2891     //       };
2892     RecordDecl *RD = C.buildImplicitRecord(".kmp_privates.t");
2893     RD->startDefinition();
2894     for (const auto &Pair : Privates) {
2895       const VarDecl *VD = Pair.second.Original;
2896       QualType Type = VD->getType().getNonReferenceType();
2897       // If the private variable is a local variable with lvalue ref type,
2898       // allocate the pointer instead of the pointee type.
2899       if (Pair.second.isLocalPrivate()) {
2900         if (VD->getType()->isLValueReferenceType())
2901           Type = C.getPointerType(Type);
2902         if (isAllocatableDecl(VD))
2903           Type = C.getPointerType(Type);
2904       }
2905       FieldDecl *FD = addFieldToRecordDecl(C, RD, Type);
2906       if (VD->hasAttrs()) {
2907         for (specific_attr_iterator<AlignedAttr> I(VD->getAttrs().begin()),
2908              E(VD->getAttrs().end());
2909              I != E; ++I)
2910           FD->addAttr(*I);
2911       }
2912     }
2913     RD->completeDefinition();
2914     return RD;
2915   }
2916   return nullptr;
2917 }
2918 
2919 static RecordDecl *
createKmpTaskTRecordDecl(CodeGenModule & CGM,OpenMPDirectiveKind Kind,QualType KmpInt32Ty,QualType KmpRoutineEntryPointerQTy)2920 createKmpTaskTRecordDecl(CodeGenModule &CGM, OpenMPDirectiveKind Kind,
2921                          QualType KmpInt32Ty,
2922                          QualType KmpRoutineEntryPointerQTy) {
2923   ASTContext &C = CGM.getContext();
2924   // Build struct kmp_task_t {
2925   //         void *              shareds;
2926   //         kmp_routine_entry_t routine;
2927   //         kmp_int32           part_id;
2928   //         kmp_cmplrdata_t data1;
2929   //         kmp_cmplrdata_t data2;
2930   // For taskloops additional fields:
2931   //         kmp_uint64          lb;
2932   //         kmp_uint64          ub;
2933   //         kmp_int64           st;
2934   //         kmp_int32           liter;
2935   //         void *              reductions;
2936   //       };
2937   RecordDecl *UD = C.buildImplicitRecord("kmp_cmplrdata_t", TagTypeKind::Union);
2938   UD->startDefinition();
2939   addFieldToRecordDecl(C, UD, KmpInt32Ty);
2940   addFieldToRecordDecl(C, UD, KmpRoutineEntryPointerQTy);
2941   UD->completeDefinition();
2942   QualType KmpCmplrdataTy = C.getRecordType(UD);
2943   RecordDecl *RD = C.buildImplicitRecord("kmp_task_t");
2944   RD->startDefinition();
2945   addFieldToRecordDecl(C, RD, C.VoidPtrTy);
2946   addFieldToRecordDecl(C, RD, KmpRoutineEntryPointerQTy);
2947   addFieldToRecordDecl(C, RD, KmpInt32Ty);
2948   addFieldToRecordDecl(C, RD, KmpCmplrdataTy);
2949   addFieldToRecordDecl(C, RD, KmpCmplrdataTy);
2950   if (isOpenMPTaskLoopDirective(Kind)) {
2951     QualType KmpUInt64Ty =
2952         CGM.getContext().getIntTypeForBitwidth(/*DestWidth=*/64, /*Signed=*/0);
2953     QualType KmpInt64Ty =
2954         CGM.getContext().getIntTypeForBitwidth(/*DestWidth=*/64, /*Signed=*/1);
2955     addFieldToRecordDecl(C, RD, KmpUInt64Ty);
2956     addFieldToRecordDecl(C, RD, KmpUInt64Ty);
2957     addFieldToRecordDecl(C, RD, KmpInt64Ty);
2958     addFieldToRecordDecl(C, RD, KmpInt32Ty);
2959     addFieldToRecordDecl(C, RD, C.VoidPtrTy);
2960   }
2961   RD->completeDefinition();
2962   return RD;
2963 }
2964 
2965 static RecordDecl *
createKmpTaskTWithPrivatesRecordDecl(CodeGenModule & CGM,QualType KmpTaskTQTy,ArrayRef<PrivateDataTy> Privates)2966 createKmpTaskTWithPrivatesRecordDecl(CodeGenModule &CGM, QualType KmpTaskTQTy,
2967                                      ArrayRef<PrivateDataTy> Privates) {
2968   ASTContext &C = CGM.getContext();
2969   // Build struct kmp_task_t_with_privates {
2970   //         kmp_task_t task_data;
2971   //         .kmp_privates_t. privates;
2972   //       };
2973   RecordDecl *RD = C.buildImplicitRecord("kmp_task_t_with_privates");
2974   RD->startDefinition();
2975   addFieldToRecordDecl(C, RD, KmpTaskTQTy);
2976   if (const RecordDecl *PrivateRD = createPrivatesRecordDecl(CGM, Privates))
2977     addFieldToRecordDecl(C, RD, C.getRecordType(PrivateRD));
2978   RD->completeDefinition();
2979   return RD;
2980 }
2981 
2982 /// Emit a proxy function which accepts kmp_task_t as the second
2983 /// argument.
2984 /// \code
2985 /// kmp_int32 .omp_task_entry.(kmp_int32 gtid, kmp_task_t *tt) {
2986 ///   TaskFunction(gtid, tt->part_id, &tt->privates, task_privates_map, tt,
2987 ///   For taskloops:
2988 ///   tt->task_data.lb, tt->task_data.ub, tt->task_data.st, tt->task_data.liter,
2989 ///   tt->reductions, tt->shareds);
2990 ///   return 0;
2991 /// }
2992 /// \endcode
2993 static llvm::Function *
emitProxyTaskFunction(CodeGenModule & CGM,SourceLocation Loc,OpenMPDirectiveKind Kind,QualType KmpInt32Ty,QualType KmpTaskTWithPrivatesPtrQTy,QualType KmpTaskTWithPrivatesQTy,QualType KmpTaskTQTy,QualType SharedsPtrTy,llvm::Function * TaskFunction,llvm::Value * TaskPrivatesMap)2994 emitProxyTaskFunction(CodeGenModule &CGM, SourceLocation Loc,
2995                       OpenMPDirectiveKind Kind, QualType KmpInt32Ty,
2996                       QualType KmpTaskTWithPrivatesPtrQTy,
2997                       QualType KmpTaskTWithPrivatesQTy, QualType KmpTaskTQTy,
2998                       QualType SharedsPtrTy, llvm::Function *TaskFunction,
2999                       llvm::Value *TaskPrivatesMap) {
3000   ASTContext &C = CGM.getContext();
3001   FunctionArgList Args;
3002   ImplicitParamDecl GtidArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, KmpInt32Ty,
3003                             ImplicitParamKind::Other);
3004   ImplicitParamDecl TaskTypeArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr,
3005                                 KmpTaskTWithPrivatesPtrQTy.withRestrict(),
3006                                 ImplicitParamKind::Other);
3007   Args.push_back(&GtidArg);
3008   Args.push_back(&TaskTypeArg);
3009   const auto &TaskEntryFnInfo =
3010       CGM.getTypes().arrangeBuiltinFunctionDeclaration(KmpInt32Ty, Args);
3011   llvm::FunctionType *TaskEntryTy =
3012       CGM.getTypes().GetFunctionType(TaskEntryFnInfo);
3013   std::string Name = CGM.getOpenMPRuntime().getName({"omp_task_entry", ""});
3014   auto *TaskEntry = llvm::Function::Create(
3015       TaskEntryTy, llvm::GlobalValue::InternalLinkage, Name, &CGM.getModule());
3016   CGM.SetInternalFunctionAttributes(GlobalDecl(), TaskEntry, TaskEntryFnInfo);
3017   TaskEntry->setDoesNotRecurse();
3018   CodeGenFunction CGF(CGM);
3019   CGF.StartFunction(GlobalDecl(), KmpInt32Ty, TaskEntry, TaskEntryFnInfo, Args,
3020                     Loc, Loc);
3021 
3022   // TaskFunction(gtid, tt->task_data.part_id, &tt->privates, task_privates_map,
3023   // tt,
3024   // For taskloops:
3025   // tt->task_data.lb, tt->task_data.ub, tt->task_data.st, tt->task_data.liter,
3026   // tt->task_data.shareds);
3027   llvm::Value *GtidParam = CGF.EmitLoadOfScalar(
3028       CGF.GetAddrOfLocalVar(&GtidArg), /*Volatile=*/false, KmpInt32Ty, Loc);
3029   LValue TDBase = CGF.EmitLoadOfPointerLValue(
3030       CGF.GetAddrOfLocalVar(&TaskTypeArg),
3031       KmpTaskTWithPrivatesPtrQTy->castAs<PointerType>());
3032   const auto *KmpTaskTWithPrivatesQTyRD =
3033       cast<RecordDecl>(KmpTaskTWithPrivatesQTy->getAsTagDecl());
3034   LValue Base =
3035       CGF.EmitLValueForField(TDBase, *KmpTaskTWithPrivatesQTyRD->field_begin());
3036   const auto *KmpTaskTQTyRD = cast<RecordDecl>(KmpTaskTQTy->getAsTagDecl());
3037   auto PartIdFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTPartId);
3038   LValue PartIdLVal = CGF.EmitLValueForField(Base, *PartIdFI);
3039   llvm::Value *PartidParam = PartIdLVal.getPointer(CGF);
3040 
3041   auto SharedsFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTShareds);
3042   LValue SharedsLVal = CGF.EmitLValueForField(Base, *SharedsFI);
3043   llvm::Value *SharedsParam = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
3044       CGF.EmitLoadOfScalar(SharedsLVal, Loc),
3045       CGF.ConvertTypeForMem(SharedsPtrTy));
3046 
3047   auto PrivatesFI = std::next(KmpTaskTWithPrivatesQTyRD->field_begin(), 1);
3048   llvm::Value *PrivatesParam;
3049   if (PrivatesFI != KmpTaskTWithPrivatesQTyRD->field_end()) {
3050     LValue PrivatesLVal = CGF.EmitLValueForField(TDBase, *PrivatesFI);
3051     PrivatesParam = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
3052         PrivatesLVal.getPointer(CGF), CGF.VoidPtrTy);
3053   } else {
3054     PrivatesParam = llvm::ConstantPointerNull::get(CGF.VoidPtrTy);
3055   }
3056 
3057   llvm::Value *CommonArgs[] = {
3058       GtidParam, PartidParam, PrivatesParam, TaskPrivatesMap,
3059       CGF.Builder
3060           .CreatePointerBitCastOrAddrSpaceCast(TDBase.getAddress(),
3061                                                CGF.VoidPtrTy, CGF.Int8Ty)
3062           .emitRawPointer(CGF)};
3063   SmallVector<llvm::Value *, 16> CallArgs(std::begin(CommonArgs),
3064                                           std::end(CommonArgs));
3065   if (isOpenMPTaskLoopDirective(Kind)) {
3066     auto LBFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTLowerBound);
3067     LValue LBLVal = CGF.EmitLValueForField(Base, *LBFI);
3068     llvm::Value *LBParam = CGF.EmitLoadOfScalar(LBLVal, Loc);
3069     auto UBFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTUpperBound);
3070     LValue UBLVal = CGF.EmitLValueForField(Base, *UBFI);
3071     llvm::Value *UBParam = CGF.EmitLoadOfScalar(UBLVal, Loc);
3072     auto StFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTStride);
3073     LValue StLVal = CGF.EmitLValueForField(Base, *StFI);
3074     llvm::Value *StParam = CGF.EmitLoadOfScalar(StLVal, Loc);
3075     auto LIFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTLastIter);
3076     LValue LILVal = CGF.EmitLValueForField(Base, *LIFI);
3077     llvm::Value *LIParam = CGF.EmitLoadOfScalar(LILVal, Loc);
3078     auto RFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTReductions);
3079     LValue RLVal = CGF.EmitLValueForField(Base, *RFI);
3080     llvm::Value *RParam = CGF.EmitLoadOfScalar(RLVal, Loc);
3081     CallArgs.push_back(LBParam);
3082     CallArgs.push_back(UBParam);
3083     CallArgs.push_back(StParam);
3084     CallArgs.push_back(LIParam);
3085     CallArgs.push_back(RParam);
3086   }
3087   CallArgs.push_back(SharedsParam);
3088 
3089   CGM.getOpenMPRuntime().emitOutlinedFunctionCall(CGF, Loc, TaskFunction,
3090                                                   CallArgs);
3091   CGF.EmitStoreThroughLValue(RValue::get(CGF.Builder.getInt32(/*C=*/0)),
3092                              CGF.MakeAddrLValue(CGF.ReturnValue, KmpInt32Ty));
3093   CGF.FinishFunction();
3094   return TaskEntry;
3095 }
3096 
emitDestructorsFunction(CodeGenModule & CGM,SourceLocation Loc,QualType KmpInt32Ty,QualType KmpTaskTWithPrivatesPtrQTy,QualType KmpTaskTWithPrivatesQTy)3097 static llvm::Value *emitDestructorsFunction(CodeGenModule &CGM,
3098                                             SourceLocation Loc,
3099                                             QualType KmpInt32Ty,
3100                                             QualType KmpTaskTWithPrivatesPtrQTy,
3101                                             QualType KmpTaskTWithPrivatesQTy) {
3102   ASTContext &C = CGM.getContext();
3103   FunctionArgList Args;
3104   ImplicitParamDecl GtidArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, KmpInt32Ty,
3105                             ImplicitParamKind::Other);
3106   ImplicitParamDecl TaskTypeArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr,
3107                                 KmpTaskTWithPrivatesPtrQTy.withRestrict(),
3108                                 ImplicitParamKind::Other);
3109   Args.push_back(&GtidArg);
3110   Args.push_back(&TaskTypeArg);
3111   const auto &DestructorFnInfo =
3112       CGM.getTypes().arrangeBuiltinFunctionDeclaration(KmpInt32Ty, Args);
3113   llvm::FunctionType *DestructorFnTy =
3114       CGM.getTypes().GetFunctionType(DestructorFnInfo);
3115   std::string Name =
3116       CGM.getOpenMPRuntime().getName({"omp_task_destructor", ""});
3117   auto *DestructorFn =
3118       llvm::Function::Create(DestructorFnTy, llvm::GlobalValue::InternalLinkage,
3119                              Name, &CGM.getModule());
3120   CGM.SetInternalFunctionAttributes(GlobalDecl(), DestructorFn,
3121                                     DestructorFnInfo);
3122   DestructorFn->setDoesNotRecurse();
3123   CodeGenFunction CGF(CGM);
3124   CGF.StartFunction(GlobalDecl(), KmpInt32Ty, DestructorFn, DestructorFnInfo,
3125                     Args, Loc, Loc);
3126 
3127   LValue Base = CGF.EmitLoadOfPointerLValue(
3128       CGF.GetAddrOfLocalVar(&TaskTypeArg),
3129       KmpTaskTWithPrivatesPtrQTy->castAs<PointerType>());
3130   const auto *KmpTaskTWithPrivatesQTyRD =
3131       cast<RecordDecl>(KmpTaskTWithPrivatesQTy->getAsTagDecl());
3132   auto FI = std::next(KmpTaskTWithPrivatesQTyRD->field_begin());
3133   Base = CGF.EmitLValueForField(Base, *FI);
3134   for (const auto *Field :
3135        cast<RecordDecl>(FI->getType()->getAsTagDecl())->fields()) {
3136     if (QualType::DestructionKind DtorKind =
3137             Field->getType().isDestructedType()) {
3138       LValue FieldLValue = CGF.EmitLValueForField(Base, Field);
3139       CGF.pushDestroy(DtorKind, FieldLValue.getAddress(), Field->getType());
3140     }
3141   }
3142   CGF.FinishFunction();
3143   return DestructorFn;
3144 }
3145 
3146 /// Emit a privates mapping function for correct handling of private and
3147 /// firstprivate variables.
3148 /// \code
3149 /// void .omp_task_privates_map.(const .privates. *noalias privs, <ty1>
3150 /// **noalias priv1,...,  <tyn> **noalias privn) {
3151 ///   *priv1 = &.privates.priv1;
3152 ///   ...;
3153 ///   *privn = &.privates.privn;
3154 /// }
3155 /// \endcode
3156 static llvm::Value *
emitTaskPrivateMappingFunction(CodeGenModule & CGM,SourceLocation Loc,const OMPTaskDataTy & Data,QualType PrivatesQTy,ArrayRef<PrivateDataTy> Privates)3157 emitTaskPrivateMappingFunction(CodeGenModule &CGM, SourceLocation Loc,
3158                                const OMPTaskDataTy &Data, QualType PrivatesQTy,
3159                                ArrayRef<PrivateDataTy> Privates) {
3160   ASTContext &C = CGM.getContext();
3161   FunctionArgList Args;
3162   ImplicitParamDecl TaskPrivatesArg(
3163       C, /*DC=*/nullptr, Loc, /*Id=*/nullptr,
3164       C.getPointerType(PrivatesQTy).withConst().withRestrict(),
3165       ImplicitParamKind::Other);
3166   Args.push_back(&TaskPrivatesArg);
3167   llvm::DenseMap<CanonicalDeclPtr<const VarDecl>, unsigned> PrivateVarsPos;
3168   unsigned Counter = 1;
3169   for (const Expr *E : Data.PrivateVars) {
3170     Args.push_back(ImplicitParamDecl::Create(
3171         C, /*DC=*/nullptr, Loc, /*Id=*/nullptr,
3172         C.getPointerType(C.getPointerType(E->getType()))
3173             .withConst()
3174             .withRestrict(),
3175         ImplicitParamKind::Other));
3176     const auto *VD = cast<VarDecl>(cast<DeclRefExpr>(E)->getDecl());
3177     PrivateVarsPos[VD] = Counter;
3178     ++Counter;
3179   }
3180   for (const Expr *E : Data.FirstprivateVars) {
3181     Args.push_back(ImplicitParamDecl::Create(
3182         C, /*DC=*/nullptr, Loc, /*Id=*/nullptr,
3183         C.getPointerType(C.getPointerType(E->getType()))
3184             .withConst()
3185             .withRestrict(),
3186         ImplicitParamKind::Other));
3187     const auto *VD = cast<VarDecl>(cast<DeclRefExpr>(E)->getDecl());
3188     PrivateVarsPos[VD] = Counter;
3189     ++Counter;
3190   }
3191   for (const Expr *E : Data.LastprivateVars) {
3192     Args.push_back(ImplicitParamDecl::Create(
3193         C, /*DC=*/nullptr, Loc, /*Id=*/nullptr,
3194         C.getPointerType(C.getPointerType(E->getType()))
3195             .withConst()
3196             .withRestrict(),
3197         ImplicitParamKind::Other));
3198     const auto *VD = cast<VarDecl>(cast<DeclRefExpr>(E)->getDecl());
3199     PrivateVarsPos[VD] = Counter;
3200     ++Counter;
3201   }
3202   for (const VarDecl *VD : Data.PrivateLocals) {
3203     QualType Ty = VD->getType().getNonReferenceType();
3204     if (VD->getType()->isLValueReferenceType())
3205       Ty = C.getPointerType(Ty);
3206     if (isAllocatableDecl(VD))
3207       Ty = C.getPointerType(Ty);
3208     Args.push_back(ImplicitParamDecl::Create(
3209         C, /*DC=*/nullptr, Loc, /*Id=*/nullptr,
3210         C.getPointerType(C.getPointerType(Ty)).withConst().withRestrict(),
3211         ImplicitParamKind::Other));
3212     PrivateVarsPos[VD] = Counter;
3213     ++Counter;
3214   }
3215   const auto &TaskPrivatesMapFnInfo =
3216       CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args);
3217   llvm::FunctionType *TaskPrivatesMapTy =
3218       CGM.getTypes().GetFunctionType(TaskPrivatesMapFnInfo);
3219   std::string Name =
3220       CGM.getOpenMPRuntime().getName({"omp_task_privates_map", ""});
3221   auto *TaskPrivatesMap = llvm::Function::Create(
3222       TaskPrivatesMapTy, llvm::GlobalValue::InternalLinkage, Name,
3223       &CGM.getModule());
3224   CGM.SetInternalFunctionAttributes(GlobalDecl(), TaskPrivatesMap,
3225                                     TaskPrivatesMapFnInfo);
3226   if (CGM.getLangOpts().Optimize) {
3227     TaskPrivatesMap->removeFnAttr(llvm::Attribute::NoInline);
3228     TaskPrivatesMap->removeFnAttr(llvm::Attribute::OptimizeNone);
3229     TaskPrivatesMap->addFnAttr(llvm::Attribute::AlwaysInline);
3230   }
3231   CodeGenFunction CGF(CGM);
3232   CGF.StartFunction(GlobalDecl(), C.VoidTy, TaskPrivatesMap,
3233                     TaskPrivatesMapFnInfo, Args, Loc, Loc);
3234 
3235   // *privi = &.privates.privi;
3236   LValue Base = CGF.EmitLoadOfPointerLValue(
3237       CGF.GetAddrOfLocalVar(&TaskPrivatesArg),
3238       TaskPrivatesArg.getType()->castAs<PointerType>());
3239   const auto *PrivatesQTyRD = cast<RecordDecl>(PrivatesQTy->getAsTagDecl());
3240   Counter = 0;
3241   for (const FieldDecl *Field : PrivatesQTyRD->fields()) {
3242     LValue FieldLVal = CGF.EmitLValueForField(Base, Field);
3243     const VarDecl *VD = Args[PrivateVarsPos[Privates[Counter].second.Original]];
3244     LValue RefLVal =
3245         CGF.MakeAddrLValue(CGF.GetAddrOfLocalVar(VD), VD->getType());
3246     LValue RefLoadLVal = CGF.EmitLoadOfPointerLValue(
3247         RefLVal.getAddress(), RefLVal.getType()->castAs<PointerType>());
3248     CGF.EmitStoreOfScalar(FieldLVal.getPointer(CGF), RefLoadLVal);
3249     ++Counter;
3250   }
3251   CGF.FinishFunction();
3252   return TaskPrivatesMap;
3253 }
3254 
3255 /// Emit initialization for private variables in task-based directives.
emitPrivatesInit(CodeGenFunction & CGF,const OMPExecutableDirective & D,Address KmpTaskSharedsPtr,LValue TDBase,const RecordDecl * KmpTaskTWithPrivatesQTyRD,QualType SharedsTy,QualType SharedsPtrTy,const OMPTaskDataTy & Data,ArrayRef<PrivateDataTy> Privates,bool ForDup)3256 static void emitPrivatesInit(CodeGenFunction &CGF,
3257                              const OMPExecutableDirective &D,
3258                              Address KmpTaskSharedsPtr, LValue TDBase,
3259                              const RecordDecl *KmpTaskTWithPrivatesQTyRD,
3260                              QualType SharedsTy, QualType SharedsPtrTy,
3261                              const OMPTaskDataTy &Data,
3262                              ArrayRef<PrivateDataTy> Privates, bool ForDup) {
3263   ASTContext &C = CGF.getContext();
3264   auto FI = std::next(KmpTaskTWithPrivatesQTyRD->field_begin());
3265   LValue PrivatesBase = CGF.EmitLValueForField(TDBase, *FI);
3266   OpenMPDirectiveKind Kind = isOpenMPTaskLoopDirective(D.getDirectiveKind())
3267                                  ? OMPD_taskloop
3268                                  : OMPD_task;
3269   const CapturedStmt &CS = *D.getCapturedStmt(Kind);
3270   CodeGenFunction::CGCapturedStmtInfo CapturesInfo(CS);
3271   LValue SrcBase;
3272   bool IsTargetTask =
3273       isOpenMPTargetDataManagementDirective(D.getDirectiveKind()) ||
3274       isOpenMPTargetExecutionDirective(D.getDirectiveKind());
3275   // For target-based directives skip 4 firstprivate arrays BasePointersArray,
3276   // PointersArray, SizesArray, and MappersArray. The original variables for
3277   // these arrays are not captured and we get their addresses explicitly.
3278   if ((!IsTargetTask && !Data.FirstprivateVars.empty() && ForDup) ||
3279       (IsTargetTask && KmpTaskSharedsPtr.isValid())) {
3280     SrcBase = CGF.MakeAddrLValue(
3281         CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
3282             KmpTaskSharedsPtr, CGF.ConvertTypeForMem(SharedsPtrTy),
3283             CGF.ConvertTypeForMem(SharedsTy)),
3284         SharedsTy);
3285   }
3286   FI = cast<RecordDecl>(FI->getType()->getAsTagDecl())->field_begin();
3287   for (const PrivateDataTy &Pair : Privates) {
3288     // Do not initialize private locals.
3289     if (Pair.second.isLocalPrivate()) {
3290       ++FI;
3291       continue;
3292     }
3293     const VarDecl *VD = Pair.second.PrivateCopy;
3294     const Expr *Init = VD->getAnyInitializer();
3295     if (Init && (!ForDup || (isa<CXXConstructExpr>(Init) &&
3296                              !CGF.isTrivialInitializer(Init)))) {
3297       LValue PrivateLValue = CGF.EmitLValueForField(PrivatesBase, *FI);
3298       if (const VarDecl *Elem = Pair.second.PrivateElemInit) {
3299         const VarDecl *OriginalVD = Pair.second.Original;
3300         // Check if the variable is the target-based BasePointersArray,
3301         // PointersArray, SizesArray, or MappersArray.
3302         LValue SharedRefLValue;
3303         QualType Type = PrivateLValue.getType();
3304         const FieldDecl *SharedField = CapturesInfo.lookup(OriginalVD);
3305         if (IsTargetTask && !SharedField) {
3306           assert(isa<ImplicitParamDecl>(OriginalVD) &&
3307                  isa<CapturedDecl>(OriginalVD->getDeclContext()) &&
3308                  cast<CapturedDecl>(OriginalVD->getDeclContext())
3309                          ->getNumParams() == 0 &&
3310                  isa<TranslationUnitDecl>(
3311                      cast<CapturedDecl>(OriginalVD->getDeclContext())
3312                          ->getDeclContext()) &&
3313                  "Expected artificial target data variable.");
3314           SharedRefLValue =
3315               CGF.MakeAddrLValue(CGF.GetAddrOfLocalVar(OriginalVD), Type);
3316         } else if (ForDup) {
3317           SharedRefLValue = CGF.EmitLValueForField(SrcBase, SharedField);
3318           SharedRefLValue = CGF.MakeAddrLValue(
3319               SharedRefLValue.getAddress().withAlignment(
3320                   C.getDeclAlign(OriginalVD)),
3321               SharedRefLValue.getType(), LValueBaseInfo(AlignmentSource::Decl),
3322               SharedRefLValue.getTBAAInfo());
3323         } else if (CGF.LambdaCaptureFields.count(
3324                        Pair.second.Original->getCanonicalDecl()) > 0 ||
3325                    isa_and_nonnull<BlockDecl>(CGF.CurCodeDecl)) {
3326           SharedRefLValue = CGF.EmitLValue(Pair.second.OriginalRef);
3327         } else {
3328           // Processing for implicitly captured variables.
3329           InlinedOpenMPRegionRAII Region(
3330               CGF, [](CodeGenFunction &, PrePostActionTy &) {}, OMPD_unknown,
3331               /*HasCancel=*/false, /*NoInheritance=*/true);
3332           SharedRefLValue = CGF.EmitLValue(Pair.second.OriginalRef);
3333         }
3334         if (Type->isArrayType()) {
3335           // Initialize firstprivate array.
3336           if (!isa<CXXConstructExpr>(Init) || CGF.isTrivialInitializer(Init)) {
3337             // Perform simple memcpy.
3338             CGF.EmitAggregateAssign(PrivateLValue, SharedRefLValue, Type);
3339           } else {
3340             // Initialize firstprivate array using element-by-element
3341             // initialization.
3342             CGF.EmitOMPAggregateAssign(
3343                 PrivateLValue.getAddress(), SharedRefLValue.getAddress(), Type,
3344                 [&CGF, Elem, Init, &CapturesInfo](Address DestElement,
3345                                                   Address SrcElement) {
3346                   // Clean up any temporaries needed by the initialization.
3347                   CodeGenFunction::OMPPrivateScope InitScope(CGF);
3348                   InitScope.addPrivate(Elem, SrcElement);
3349                   (void)InitScope.Privatize();
3350                   // Emit initialization for single element.
3351                   CodeGenFunction::CGCapturedStmtRAII CapInfoRAII(
3352                       CGF, &CapturesInfo);
3353                   CGF.EmitAnyExprToMem(Init, DestElement,
3354                                        Init->getType().getQualifiers(),
3355                                        /*IsInitializer=*/false);
3356                 });
3357           }
3358         } else {
3359           CodeGenFunction::OMPPrivateScope InitScope(CGF);
3360           InitScope.addPrivate(Elem, SharedRefLValue.getAddress());
3361           (void)InitScope.Privatize();
3362           CodeGenFunction::CGCapturedStmtRAII CapInfoRAII(CGF, &CapturesInfo);
3363           CGF.EmitExprAsInit(Init, VD, PrivateLValue,
3364                              /*capturedByInit=*/false);
3365         }
3366       } else {
3367         CGF.EmitExprAsInit(Init, VD, PrivateLValue, /*capturedByInit=*/false);
3368       }
3369     }
3370     ++FI;
3371   }
3372 }
3373 
3374 /// Check if duplication function is required for taskloops.
checkInitIsRequired(CodeGenFunction & CGF,ArrayRef<PrivateDataTy> Privates)3375 static bool checkInitIsRequired(CodeGenFunction &CGF,
3376                                 ArrayRef<PrivateDataTy> Privates) {
3377   bool InitRequired = false;
3378   for (const PrivateDataTy &Pair : Privates) {
3379     if (Pair.second.isLocalPrivate())
3380       continue;
3381     const VarDecl *VD = Pair.second.PrivateCopy;
3382     const Expr *Init = VD->getAnyInitializer();
3383     InitRequired = InitRequired || (isa_and_nonnull<CXXConstructExpr>(Init) &&
3384                                     !CGF.isTrivialInitializer(Init));
3385     if (InitRequired)
3386       break;
3387   }
3388   return InitRequired;
3389 }
3390 
3391 
3392 /// Emit task_dup function (for initialization of
3393 /// private/firstprivate/lastprivate vars and last_iter flag)
3394 /// \code
3395 /// void __task_dup_entry(kmp_task_t *task_dst, const kmp_task_t *task_src, int
3396 /// lastpriv) {
3397 /// // setup lastprivate flag
3398 ///    task_dst->last = lastpriv;
3399 /// // could be constructor calls here...
3400 /// }
3401 /// \endcode
3402 static llvm::Value *
emitTaskDupFunction(CodeGenModule & CGM,SourceLocation Loc,const OMPExecutableDirective & D,QualType KmpTaskTWithPrivatesPtrQTy,const RecordDecl * KmpTaskTWithPrivatesQTyRD,const RecordDecl * KmpTaskTQTyRD,QualType SharedsTy,QualType SharedsPtrTy,const OMPTaskDataTy & Data,ArrayRef<PrivateDataTy> Privates,bool WithLastIter)3403 emitTaskDupFunction(CodeGenModule &CGM, SourceLocation Loc,
3404                     const OMPExecutableDirective &D,
3405                     QualType KmpTaskTWithPrivatesPtrQTy,
3406                     const RecordDecl *KmpTaskTWithPrivatesQTyRD,
3407                     const RecordDecl *KmpTaskTQTyRD, QualType SharedsTy,
3408                     QualType SharedsPtrTy, const OMPTaskDataTy &Data,
3409                     ArrayRef<PrivateDataTy> Privates, bool WithLastIter) {
3410   ASTContext &C = CGM.getContext();
3411   FunctionArgList Args;
3412   ImplicitParamDecl DstArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr,
3413                            KmpTaskTWithPrivatesPtrQTy,
3414                            ImplicitParamKind::Other);
3415   ImplicitParamDecl SrcArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr,
3416                            KmpTaskTWithPrivatesPtrQTy,
3417                            ImplicitParamKind::Other);
3418   ImplicitParamDecl LastprivArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, C.IntTy,
3419                                 ImplicitParamKind::Other);
3420   Args.push_back(&DstArg);
3421   Args.push_back(&SrcArg);
3422   Args.push_back(&LastprivArg);
3423   const auto &TaskDupFnInfo =
3424       CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args);
3425   llvm::FunctionType *TaskDupTy = CGM.getTypes().GetFunctionType(TaskDupFnInfo);
3426   std::string Name = CGM.getOpenMPRuntime().getName({"omp_task_dup", ""});
3427   auto *TaskDup = llvm::Function::Create(
3428       TaskDupTy, llvm::GlobalValue::InternalLinkage, Name, &CGM.getModule());
3429   CGM.SetInternalFunctionAttributes(GlobalDecl(), TaskDup, TaskDupFnInfo);
3430   TaskDup->setDoesNotRecurse();
3431   CodeGenFunction CGF(CGM);
3432   CGF.StartFunction(GlobalDecl(), C.VoidTy, TaskDup, TaskDupFnInfo, Args, Loc,
3433                     Loc);
3434 
3435   LValue TDBase = CGF.EmitLoadOfPointerLValue(
3436       CGF.GetAddrOfLocalVar(&DstArg),
3437       KmpTaskTWithPrivatesPtrQTy->castAs<PointerType>());
3438   // task_dst->liter = lastpriv;
3439   if (WithLastIter) {
3440     auto LIFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTLastIter);
3441     LValue Base = CGF.EmitLValueForField(
3442         TDBase, *KmpTaskTWithPrivatesQTyRD->field_begin());
3443     LValue LILVal = CGF.EmitLValueForField(Base, *LIFI);
3444     llvm::Value *Lastpriv = CGF.EmitLoadOfScalar(
3445         CGF.GetAddrOfLocalVar(&LastprivArg), /*Volatile=*/false, C.IntTy, Loc);
3446     CGF.EmitStoreOfScalar(Lastpriv, LILVal);
3447   }
3448 
3449   // Emit initial values for private copies (if any).
3450   assert(!Privates.empty());
3451   Address KmpTaskSharedsPtr = Address::invalid();
3452   if (!Data.FirstprivateVars.empty()) {
3453     LValue TDBase = CGF.EmitLoadOfPointerLValue(
3454         CGF.GetAddrOfLocalVar(&SrcArg),
3455         KmpTaskTWithPrivatesPtrQTy->castAs<PointerType>());
3456     LValue Base = CGF.EmitLValueForField(
3457         TDBase, *KmpTaskTWithPrivatesQTyRD->field_begin());
3458     KmpTaskSharedsPtr = Address(
3459         CGF.EmitLoadOfScalar(CGF.EmitLValueForField(
3460                                  Base, *std::next(KmpTaskTQTyRD->field_begin(),
3461                                                   KmpTaskTShareds)),
3462                              Loc),
3463         CGF.Int8Ty, CGM.getNaturalTypeAlignment(SharedsTy));
3464   }
3465   emitPrivatesInit(CGF, D, KmpTaskSharedsPtr, TDBase, KmpTaskTWithPrivatesQTyRD,
3466                    SharedsTy, SharedsPtrTy, Data, Privates, /*ForDup=*/true);
3467   CGF.FinishFunction();
3468   return TaskDup;
3469 }
3470 
3471 /// Checks if destructor function is required to be generated.
3472 /// \return true if cleanups are required, false otherwise.
3473 static bool
checkDestructorsRequired(const RecordDecl * KmpTaskTWithPrivatesQTyRD,ArrayRef<PrivateDataTy> Privates)3474 checkDestructorsRequired(const RecordDecl *KmpTaskTWithPrivatesQTyRD,
3475                          ArrayRef<PrivateDataTy> Privates) {
3476   for (const PrivateDataTy &P : Privates) {
3477     if (P.second.isLocalPrivate())
3478       continue;
3479     QualType Ty = P.second.Original->getType().getNonReferenceType();
3480     if (Ty.isDestructedType())
3481       return true;
3482   }
3483   return false;
3484 }
3485 
3486 namespace {
3487 /// Loop generator for OpenMP iterator expression.
3488 class OMPIteratorGeneratorScope final
3489     : public CodeGenFunction::OMPPrivateScope {
3490   CodeGenFunction &CGF;
3491   const OMPIteratorExpr *E = nullptr;
3492   SmallVector<CodeGenFunction::JumpDest, 4> ContDests;
3493   SmallVector<CodeGenFunction::JumpDest, 4> ExitDests;
3494   OMPIteratorGeneratorScope() = delete;
3495   OMPIteratorGeneratorScope(OMPIteratorGeneratorScope &) = delete;
3496 
3497 public:
OMPIteratorGeneratorScope(CodeGenFunction & CGF,const OMPIteratorExpr * E)3498   OMPIteratorGeneratorScope(CodeGenFunction &CGF, const OMPIteratorExpr *E)
3499       : CodeGenFunction::OMPPrivateScope(CGF), CGF(CGF), E(E) {
3500     if (!E)
3501       return;
3502     SmallVector<llvm::Value *, 4> Uppers;
3503     for (unsigned I = 0, End = E->numOfIterators(); I < End; ++I) {
3504       Uppers.push_back(CGF.EmitScalarExpr(E->getHelper(I).Upper));
3505       const auto *VD = cast<VarDecl>(E->getIteratorDecl(I));
3506       addPrivate(VD, CGF.CreateMemTemp(VD->getType(), VD->getName()));
3507       const OMPIteratorHelperData &HelperData = E->getHelper(I);
3508       addPrivate(
3509           HelperData.CounterVD,
3510           CGF.CreateMemTemp(HelperData.CounterVD->getType(), "counter.addr"));
3511     }
3512     Privatize();
3513 
3514     for (unsigned I = 0, End = E->numOfIterators(); I < End; ++I) {
3515       const OMPIteratorHelperData &HelperData = E->getHelper(I);
3516       LValue CLVal =
3517           CGF.MakeAddrLValue(CGF.GetAddrOfLocalVar(HelperData.CounterVD),
3518                              HelperData.CounterVD->getType());
3519       // Counter = 0;
3520       CGF.EmitStoreOfScalar(
3521           llvm::ConstantInt::get(CLVal.getAddress().getElementType(), 0),
3522           CLVal);
3523       CodeGenFunction::JumpDest &ContDest =
3524           ContDests.emplace_back(CGF.getJumpDestInCurrentScope("iter.cont"));
3525       CodeGenFunction::JumpDest &ExitDest =
3526           ExitDests.emplace_back(CGF.getJumpDestInCurrentScope("iter.exit"));
3527       // N = <number-of_iterations>;
3528       llvm::Value *N = Uppers[I];
3529       // cont:
3530       // if (Counter < N) goto body; else goto exit;
3531       CGF.EmitBlock(ContDest.getBlock());
3532       auto *CVal =
3533           CGF.EmitLoadOfScalar(CLVal, HelperData.CounterVD->getLocation());
3534       llvm::Value *Cmp =
3535           HelperData.CounterVD->getType()->isSignedIntegerOrEnumerationType()
3536               ? CGF.Builder.CreateICmpSLT(CVal, N)
3537               : CGF.Builder.CreateICmpULT(CVal, N);
3538       llvm::BasicBlock *BodyBB = CGF.createBasicBlock("iter.body");
3539       CGF.Builder.CreateCondBr(Cmp, BodyBB, ExitDest.getBlock());
3540       // body:
3541       CGF.EmitBlock(BodyBB);
3542       // Iteri = Begini + Counter * Stepi;
3543       CGF.EmitIgnoredExpr(HelperData.Update);
3544     }
3545   }
~OMPIteratorGeneratorScope()3546   ~OMPIteratorGeneratorScope() {
3547     if (!E)
3548       return;
3549     for (unsigned I = E->numOfIterators(); I > 0; --I) {
3550       // Counter = Counter + 1;
3551       const OMPIteratorHelperData &HelperData = E->getHelper(I - 1);
3552       CGF.EmitIgnoredExpr(HelperData.CounterUpdate);
3553       // goto cont;
3554       CGF.EmitBranchThroughCleanup(ContDests[I - 1]);
3555       // exit:
3556       CGF.EmitBlock(ExitDests[I - 1].getBlock(), /*IsFinished=*/I == 1);
3557     }
3558   }
3559 };
3560 } // namespace
3561 
3562 static std::pair<llvm::Value *, llvm::Value *>
getPointerAndSize(CodeGenFunction & CGF,const Expr * E)3563 getPointerAndSize(CodeGenFunction &CGF, const Expr *E) {
3564   const auto *OASE = dyn_cast<OMPArrayShapingExpr>(E);
3565   llvm::Value *Addr;
3566   if (OASE) {
3567     const Expr *Base = OASE->getBase();
3568     Addr = CGF.EmitScalarExpr(Base);
3569   } else {
3570     Addr = CGF.EmitLValue(E).getPointer(CGF);
3571   }
3572   llvm::Value *SizeVal;
3573   QualType Ty = E->getType();
3574   if (OASE) {
3575     SizeVal = CGF.getTypeSize(OASE->getBase()->getType()->getPointeeType());
3576     for (const Expr *SE : OASE->getDimensions()) {
3577       llvm::Value *Sz = CGF.EmitScalarExpr(SE);
3578       Sz = CGF.EmitScalarConversion(
3579           Sz, SE->getType(), CGF.getContext().getSizeType(), SE->getExprLoc());
3580       SizeVal = CGF.Builder.CreateNUWMul(SizeVal, Sz);
3581     }
3582   } else if (const auto *ASE =
3583                  dyn_cast<ArraySectionExpr>(E->IgnoreParenImpCasts())) {
3584     LValue UpAddrLVal = CGF.EmitArraySectionExpr(ASE, /*IsLowerBound=*/false);
3585     Address UpAddrAddress = UpAddrLVal.getAddress();
3586     llvm::Value *UpAddr = CGF.Builder.CreateConstGEP1_32(
3587         UpAddrAddress.getElementType(), UpAddrAddress.emitRawPointer(CGF),
3588         /*Idx0=*/1);
3589     llvm::Value *LowIntPtr = CGF.Builder.CreatePtrToInt(Addr, CGF.SizeTy);
3590     llvm::Value *UpIntPtr = CGF.Builder.CreatePtrToInt(UpAddr, CGF.SizeTy);
3591     SizeVal = CGF.Builder.CreateNUWSub(UpIntPtr, LowIntPtr);
3592   } else {
3593     SizeVal = CGF.getTypeSize(Ty);
3594   }
3595   return std::make_pair(Addr, SizeVal);
3596 }
3597 
3598 /// Builds kmp_depend_info, if it is not built yet, and builds flags type.
getKmpAffinityType(ASTContext & C,QualType & KmpTaskAffinityInfoTy)3599 static void getKmpAffinityType(ASTContext &C, QualType &KmpTaskAffinityInfoTy) {
3600   QualType FlagsTy = C.getIntTypeForBitwidth(32, /*Signed=*/false);
3601   if (KmpTaskAffinityInfoTy.isNull()) {
3602     RecordDecl *KmpAffinityInfoRD =
3603         C.buildImplicitRecord("kmp_task_affinity_info_t");
3604     KmpAffinityInfoRD->startDefinition();
3605     addFieldToRecordDecl(C, KmpAffinityInfoRD, C.getIntPtrType());
3606     addFieldToRecordDecl(C, KmpAffinityInfoRD, C.getSizeType());
3607     addFieldToRecordDecl(C, KmpAffinityInfoRD, FlagsTy);
3608     KmpAffinityInfoRD->completeDefinition();
3609     KmpTaskAffinityInfoTy = C.getRecordType(KmpAffinityInfoRD);
3610   }
3611 }
3612 
3613 CGOpenMPRuntime::TaskResultTy
emitTaskInit(CodeGenFunction & CGF,SourceLocation Loc,const OMPExecutableDirective & D,llvm::Function * TaskFunction,QualType SharedsTy,Address Shareds,const OMPTaskDataTy & Data)3614 CGOpenMPRuntime::emitTaskInit(CodeGenFunction &CGF, SourceLocation Loc,
3615                               const OMPExecutableDirective &D,
3616                               llvm::Function *TaskFunction, QualType SharedsTy,
3617                               Address Shareds, const OMPTaskDataTy &Data) {
3618   ASTContext &C = CGM.getContext();
3619   llvm::SmallVector<PrivateDataTy, 4> Privates;
3620   // Aggregate privates and sort them by the alignment.
3621   const auto *I = Data.PrivateCopies.begin();
3622   for (const Expr *E : Data.PrivateVars) {
3623     const auto *VD = cast<VarDecl>(cast<DeclRefExpr>(E)->getDecl());
3624     Privates.emplace_back(
3625         C.getDeclAlign(VD),
3626         PrivateHelpersTy(E, VD, cast<VarDecl>(cast<DeclRefExpr>(*I)->getDecl()),
3627                          /*PrivateElemInit=*/nullptr));
3628     ++I;
3629   }
3630   I = Data.FirstprivateCopies.begin();
3631   const auto *IElemInitRef = Data.FirstprivateInits.begin();
3632   for (const Expr *E : Data.FirstprivateVars) {
3633     const auto *VD = cast<VarDecl>(cast<DeclRefExpr>(E)->getDecl());
3634     Privates.emplace_back(
3635         C.getDeclAlign(VD),
3636         PrivateHelpersTy(
3637             E, VD, cast<VarDecl>(cast<DeclRefExpr>(*I)->getDecl()),
3638             cast<VarDecl>(cast<DeclRefExpr>(*IElemInitRef)->getDecl())));
3639     ++I;
3640     ++IElemInitRef;
3641   }
3642   I = Data.LastprivateCopies.begin();
3643   for (const Expr *E : Data.LastprivateVars) {
3644     const auto *VD = cast<VarDecl>(cast<DeclRefExpr>(E)->getDecl());
3645     Privates.emplace_back(
3646         C.getDeclAlign(VD),
3647         PrivateHelpersTy(E, VD, cast<VarDecl>(cast<DeclRefExpr>(*I)->getDecl()),
3648                          /*PrivateElemInit=*/nullptr));
3649     ++I;
3650   }
3651   for (const VarDecl *VD : Data.PrivateLocals) {
3652     if (isAllocatableDecl(VD))
3653       Privates.emplace_back(CGM.getPointerAlign(), PrivateHelpersTy(VD));
3654     else
3655       Privates.emplace_back(C.getDeclAlign(VD), PrivateHelpersTy(VD));
3656   }
3657   llvm::stable_sort(Privates,
3658                     [](const PrivateDataTy &L, const PrivateDataTy &R) {
3659                       return L.first > R.first;
3660                     });
3661   QualType KmpInt32Ty = C.getIntTypeForBitwidth(/*DestWidth=*/32, /*Signed=*/1);
3662   // Build type kmp_routine_entry_t (if not built yet).
3663   emitKmpRoutineEntryT(KmpInt32Ty);
3664   // Build type kmp_task_t (if not built yet).
3665   if (isOpenMPTaskLoopDirective(D.getDirectiveKind())) {
3666     if (SavedKmpTaskloopTQTy.isNull()) {
3667       SavedKmpTaskloopTQTy = C.getRecordType(createKmpTaskTRecordDecl(
3668           CGM, D.getDirectiveKind(), KmpInt32Ty, KmpRoutineEntryPtrQTy));
3669     }
3670     KmpTaskTQTy = SavedKmpTaskloopTQTy;
3671   } else {
3672     assert((D.getDirectiveKind() == OMPD_task ||
3673             isOpenMPTargetExecutionDirective(D.getDirectiveKind()) ||
3674             isOpenMPTargetDataManagementDirective(D.getDirectiveKind())) &&
3675            "Expected taskloop, task or target directive");
3676     if (SavedKmpTaskTQTy.isNull()) {
3677       SavedKmpTaskTQTy = C.getRecordType(createKmpTaskTRecordDecl(
3678           CGM, D.getDirectiveKind(), KmpInt32Ty, KmpRoutineEntryPtrQTy));
3679     }
3680     KmpTaskTQTy = SavedKmpTaskTQTy;
3681   }
3682   const auto *KmpTaskTQTyRD = cast<RecordDecl>(KmpTaskTQTy->getAsTagDecl());
3683   // Build particular struct kmp_task_t for the given task.
3684   const RecordDecl *KmpTaskTWithPrivatesQTyRD =
3685       createKmpTaskTWithPrivatesRecordDecl(CGM, KmpTaskTQTy, Privates);
3686   QualType KmpTaskTWithPrivatesQTy = C.getRecordType(KmpTaskTWithPrivatesQTyRD);
3687   QualType KmpTaskTWithPrivatesPtrQTy =
3688       C.getPointerType(KmpTaskTWithPrivatesQTy);
3689   llvm::Type *KmpTaskTWithPrivatesTy = CGF.ConvertType(KmpTaskTWithPrivatesQTy);
3690   llvm::Type *KmpTaskTWithPrivatesPtrTy =
3691       KmpTaskTWithPrivatesTy->getPointerTo();
3692   llvm::Value *KmpTaskTWithPrivatesTySize =
3693       CGF.getTypeSize(KmpTaskTWithPrivatesQTy);
3694   QualType SharedsPtrTy = C.getPointerType(SharedsTy);
3695 
3696   // Emit initial values for private copies (if any).
3697   llvm::Value *TaskPrivatesMap = nullptr;
3698   llvm::Type *TaskPrivatesMapTy =
3699       std::next(TaskFunction->arg_begin(), 3)->getType();
3700   if (!Privates.empty()) {
3701     auto FI = std::next(KmpTaskTWithPrivatesQTyRD->field_begin());
3702     TaskPrivatesMap =
3703         emitTaskPrivateMappingFunction(CGM, Loc, Data, FI->getType(), Privates);
3704     TaskPrivatesMap = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
3705         TaskPrivatesMap, TaskPrivatesMapTy);
3706   } else {
3707     TaskPrivatesMap = llvm::ConstantPointerNull::get(
3708         cast<llvm::PointerType>(TaskPrivatesMapTy));
3709   }
3710   // Build a proxy function kmp_int32 .omp_task_entry.(kmp_int32 gtid,
3711   // kmp_task_t *tt);
3712   llvm::Function *TaskEntry = emitProxyTaskFunction(
3713       CGM, Loc, D.getDirectiveKind(), KmpInt32Ty, KmpTaskTWithPrivatesPtrQTy,
3714       KmpTaskTWithPrivatesQTy, KmpTaskTQTy, SharedsPtrTy, TaskFunction,
3715       TaskPrivatesMap);
3716 
3717   // Build call kmp_task_t * __kmpc_omp_task_alloc(ident_t *, kmp_int32 gtid,
3718   // kmp_int32 flags, size_t sizeof_kmp_task_t, size_t sizeof_shareds,
3719   // kmp_routine_entry_t *task_entry);
3720   // Task flags. Format is taken from
3721   // https://github.com/llvm/llvm-project/blob/main/openmp/runtime/src/kmp.h,
3722   // description of kmp_tasking_flags struct.
3723   enum {
3724     TiedFlag = 0x1,
3725     FinalFlag = 0x2,
3726     DestructorsFlag = 0x8,
3727     PriorityFlag = 0x20,
3728     DetachableFlag = 0x40,
3729   };
3730   unsigned Flags = Data.Tied ? TiedFlag : 0;
3731   bool NeedsCleanup = false;
3732   if (!Privates.empty()) {
3733     NeedsCleanup =
3734         checkDestructorsRequired(KmpTaskTWithPrivatesQTyRD, Privates);
3735     if (NeedsCleanup)
3736       Flags = Flags | DestructorsFlag;
3737   }
3738   if (Data.Priority.getInt())
3739     Flags = Flags | PriorityFlag;
3740   if (D.hasClausesOfKind<OMPDetachClause>())
3741     Flags = Flags | DetachableFlag;
3742   llvm::Value *TaskFlags =
3743       Data.Final.getPointer()
3744           ? CGF.Builder.CreateSelect(Data.Final.getPointer(),
3745                                      CGF.Builder.getInt32(FinalFlag),
3746                                      CGF.Builder.getInt32(/*C=*/0))
3747           : CGF.Builder.getInt32(Data.Final.getInt() ? FinalFlag : 0);
3748   TaskFlags = CGF.Builder.CreateOr(TaskFlags, CGF.Builder.getInt32(Flags));
3749   llvm::Value *SharedsSize = CGM.getSize(C.getTypeSizeInChars(SharedsTy));
3750   SmallVector<llvm::Value *, 8> AllocArgs = {emitUpdateLocation(CGF, Loc),
3751       getThreadID(CGF, Loc), TaskFlags, KmpTaskTWithPrivatesTySize,
3752       SharedsSize, CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
3753           TaskEntry, KmpRoutineEntryPtrTy)};
3754   llvm::Value *NewTask;
3755   if (D.hasClausesOfKind<OMPNowaitClause>()) {
3756     // Check if we have any device clause associated with the directive.
3757     const Expr *Device = nullptr;
3758     if (auto *C = D.getSingleClause<OMPDeviceClause>())
3759       Device = C->getDevice();
3760     // Emit device ID if any otherwise use default value.
3761     llvm::Value *DeviceID;
3762     if (Device)
3763       DeviceID = CGF.Builder.CreateIntCast(CGF.EmitScalarExpr(Device),
3764                                            CGF.Int64Ty, /*isSigned=*/true);
3765     else
3766       DeviceID = CGF.Builder.getInt64(OMP_DEVICEID_UNDEF);
3767     AllocArgs.push_back(DeviceID);
3768     NewTask = CGF.EmitRuntimeCall(
3769         OMPBuilder.getOrCreateRuntimeFunction(
3770             CGM.getModule(), OMPRTL___kmpc_omp_target_task_alloc),
3771         AllocArgs);
3772   } else {
3773     NewTask =
3774         CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
3775                                 CGM.getModule(), OMPRTL___kmpc_omp_task_alloc),
3776                             AllocArgs);
3777   }
3778   // Emit detach clause initialization.
3779   // evt = (typeof(evt))__kmpc_task_allow_completion_event(loc, tid,
3780   // task_descriptor);
3781   if (const auto *DC = D.getSingleClause<OMPDetachClause>()) {
3782     const Expr *Evt = DC->getEventHandler()->IgnoreParenImpCasts();
3783     LValue EvtLVal = CGF.EmitLValue(Evt);
3784 
3785     // Build kmp_event_t *__kmpc_task_allow_completion_event(ident_t *loc_ref,
3786     // int gtid, kmp_task_t *task);
3787     llvm::Value *Loc = emitUpdateLocation(CGF, DC->getBeginLoc());
3788     llvm::Value *Tid = getThreadID(CGF, DC->getBeginLoc());
3789     Tid = CGF.Builder.CreateIntCast(Tid, CGF.IntTy, /*isSigned=*/false);
3790     llvm::Value *EvtVal = CGF.EmitRuntimeCall(
3791         OMPBuilder.getOrCreateRuntimeFunction(
3792             CGM.getModule(), OMPRTL___kmpc_task_allow_completion_event),
3793         {Loc, Tid, NewTask});
3794     EvtVal = CGF.EmitScalarConversion(EvtVal, C.VoidPtrTy, Evt->getType(),
3795                                       Evt->getExprLoc());
3796     CGF.EmitStoreOfScalar(EvtVal, EvtLVal);
3797   }
3798   // Process affinity clauses.
3799   if (D.hasClausesOfKind<OMPAffinityClause>()) {
3800     // Process list of affinity data.
3801     ASTContext &C = CGM.getContext();
3802     Address AffinitiesArray = Address::invalid();
3803     // Calculate number of elements to form the array of affinity data.
3804     llvm::Value *NumOfElements = nullptr;
3805     unsigned NumAffinities = 0;
3806     for (const auto *C : D.getClausesOfKind<OMPAffinityClause>()) {
3807       if (const Expr *Modifier = C->getModifier()) {
3808         const auto *IE = cast<OMPIteratorExpr>(Modifier->IgnoreParenImpCasts());
3809         for (unsigned I = 0, E = IE->numOfIterators(); I < E; ++I) {
3810           llvm::Value *Sz = CGF.EmitScalarExpr(IE->getHelper(I).Upper);
3811           Sz = CGF.Builder.CreateIntCast(Sz, CGF.SizeTy, /*isSigned=*/false);
3812           NumOfElements =
3813               NumOfElements ? CGF.Builder.CreateNUWMul(NumOfElements, Sz) : Sz;
3814         }
3815       } else {
3816         NumAffinities += C->varlist_size();
3817       }
3818     }
3819     getKmpAffinityType(CGM.getContext(), KmpTaskAffinityInfoTy);
3820     // Fields ids in kmp_task_affinity_info record.
3821     enum RTLAffinityInfoFieldsTy { BaseAddr, Len, Flags };
3822 
3823     QualType KmpTaskAffinityInfoArrayTy;
3824     if (NumOfElements) {
3825       NumOfElements = CGF.Builder.CreateNUWAdd(
3826           llvm::ConstantInt::get(CGF.SizeTy, NumAffinities), NumOfElements);
3827       auto *OVE = new (C) OpaqueValueExpr(
3828           Loc,
3829           C.getIntTypeForBitwidth(C.getTypeSize(C.getSizeType()), /*Signed=*/0),
3830           VK_PRValue);
3831       CodeGenFunction::OpaqueValueMapping OpaqueMap(CGF, OVE,
3832                                                     RValue::get(NumOfElements));
3833       KmpTaskAffinityInfoArrayTy = C.getVariableArrayType(
3834           KmpTaskAffinityInfoTy, OVE, ArraySizeModifier::Normal,
3835           /*IndexTypeQuals=*/0, SourceRange(Loc, Loc));
3836       // Properly emit variable-sized array.
3837       auto *PD = ImplicitParamDecl::Create(C, KmpTaskAffinityInfoArrayTy,
3838                                            ImplicitParamKind::Other);
3839       CGF.EmitVarDecl(*PD);
3840       AffinitiesArray = CGF.GetAddrOfLocalVar(PD);
3841       NumOfElements = CGF.Builder.CreateIntCast(NumOfElements, CGF.Int32Ty,
3842                                                 /*isSigned=*/false);
3843     } else {
3844       KmpTaskAffinityInfoArrayTy = C.getConstantArrayType(
3845           KmpTaskAffinityInfoTy,
3846           llvm::APInt(C.getTypeSize(C.getSizeType()), NumAffinities), nullptr,
3847           ArraySizeModifier::Normal, /*IndexTypeQuals=*/0);
3848       AffinitiesArray =
3849           CGF.CreateMemTemp(KmpTaskAffinityInfoArrayTy, ".affs.arr.addr");
3850       AffinitiesArray = CGF.Builder.CreateConstArrayGEP(AffinitiesArray, 0);
3851       NumOfElements = llvm::ConstantInt::get(CGM.Int32Ty, NumAffinities,
3852                                              /*isSigned=*/false);
3853     }
3854 
3855     const auto *KmpAffinityInfoRD = KmpTaskAffinityInfoTy->getAsRecordDecl();
3856     // Fill array by elements without iterators.
3857     unsigned Pos = 0;
3858     bool HasIterator = false;
3859     for (const auto *C : D.getClausesOfKind<OMPAffinityClause>()) {
3860       if (C->getModifier()) {
3861         HasIterator = true;
3862         continue;
3863       }
3864       for (const Expr *E : C->varlists()) {
3865         llvm::Value *Addr;
3866         llvm::Value *Size;
3867         std::tie(Addr, Size) = getPointerAndSize(CGF, E);
3868         LValue Base =
3869             CGF.MakeAddrLValue(CGF.Builder.CreateConstGEP(AffinitiesArray, Pos),
3870                                KmpTaskAffinityInfoTy);
3871         // affs[i].base_addr = &<Affinities[i].second>;
3872         LValue BaseAddrLVal = CGF.EmitLValueForField(
3873             Base, *std::next(KmpAffinityInfoRD->field_begin(), BaseAddr));
3874         CGF.EmitStoreOfScalar(CGF.Builder.CreatePtrToInt(Addr, CGF.IntPtrTy),
3875                               BaseAddrLVal);
3876         // affs[i].len = sizeof(<Affinities[i].second>);
3877         LValue LenLVal = CGF.EmitLValueForField(
3878             Base, *std::next(KmpAffinityInfoRD->field_begin(), Len));
3879         CGF.EmitStoreOfScalar(Size, LenLVal);
3880         ++Pos;
3881       }
3882     }
3883     LValue PosLVal;
3884     if (HasIterator) {
3885       PosLVal = CGF.MakeAddrLValue(
3886           CGF.CreateMemTemp(C.getSizeType(), "affs.counter.addr"),
3887           C.getSizeType());
3888       CGF.EmitStoreOfScalar(llvm::ConstantInt::get(CGF.SizeTy, Pos), PosLVal);
3889     }
3890     // Process elements with iterators.
3891     for (const auto *C : D.getClausesOfKind<OMPAffinityClause>()) {
3892       const Expr *Modifier = C->getModifier();
3893       if (!Modifier)
3894         continue;
3895       OMPIteratorGeneratorScope IteratorScope(
3896           CGF, cast_or_null<OMPIteratorExpr>(Modifier->IgnoreParenImpCasts()));
3897       for (const Expr *E : C->varlists()) {
3898         llvm::Value *Addr;
3899         llvm::Value *Size;
3900         std::tie(Addr, Size) = getPointerAndSize(CGF, E);
3901         llvm::Value *Idx = CGF.EmitLoadOfScalar(PosLVal, E->getExprLoc());
3902         LValue Base =
3903             CGF.MakeAddrLValue(CGF.Builder.CreateGEP(CGF, AffinitiesArray, Idx),
3904                                KmpTaskAffinityInfoTy);
3905         // affs[i].base_addr = &<Affinities[i].second>;
3906         LValue BaseAddrLVal = CGF.EmitLValueForField(
3907             Base, *std::next(KmpAffinityInfoRD->field_begin(), BaseAddr));
3908         CGF.EmitStoreOfScalar(CGF.Builder.CreatePtrToInt(Addr, CGF.IntPtrTy),
3909                               BaseAddrLVal);
3910         // affs[i].len = sizeof(<Affinities[i].second>);
3911         LValue LenLVal = CGF.EmitLValueForField(
3912             Base, *std::next(KmpAffinityInfoRD->field_begin(), Len));
3913         CGF.EmitStoreOfScalar(Size, LenLVal);
3914         Idx = CGF.Builder.CreateNUWAdd(
3915             Idx, llvm::ConstantInt::get(Idx->getType(), 1));
3916         CGF.EmitStoreOfScalar(Idx, PosLVal);
3917       }
3918     }
3919     // Call to kmp_int32 __kmpc_omp_reg_task_with_affinity(ident_t *loc_ref,
3920     // kmp_int32 gtid, kmp_task_t *new_task, kmp_int32
3921     // naffins, kmp_task_affinity_info_t *affin_list);
3922     llvm::Value *LocRef = emitUpdateLocation(CGF, Loc);
3923     llvm::Value *GTid = getThreadID(CGF, Loc);
3924     llvm::Value *AffinListPtr = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
3925         AffinitiesArray.emitRawPointer(CGF), CGM.VoidPtrTy);
3926     // FIXME: Emit the function and ignore its result for now unless the
3927     // runtime function is properly implemented.
3928     (void)CGF.EmitRuntimeCall(
3929         OMPBuilder.getOrCreateRuntimeFunction(
3930             CGM.getModule(), OMPRTL___kmpc_omp_reg_task_with_affinity),
3931         {LocRef, GTid, NewTask, NumOfElements, AffinListPtr});
3932   }
3933   llvm::Value *NewTaskNewTaskTTy =
3934       CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
3935           NewTask, KmpTaskTWithPrivatesPtrTy);
3936   LValue Base = CGF.MakeNaturalAlignRawAddrLValue(NewTaskNewTaskTTy,
3937                                                   KmpTaskTWithPrivatesQTy);
3938   LValue TDBase =
3939       CGF.EmitLValueForField(Base, *KmpTaskTWithPrivatesQTyRD->field_begin());
3940   // Fill the data in the resulting kmp_task_t record.
3941   // Copy shareds if there are any.
3942   Address KmpTaskSharedsPtr = Address::invalid();
3943   if (!SharedsTy->getAsStructureType()->getDecl()->field_empty()) {
3944     KmpTaskSharedsPtr = Address(
3945         CGF.EmitLoadOfScalar(
3946             CGF.EmitLValueForField(
3947                 TDBase,
3948                 *std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTShareds)),
3949             Loc),
3950         CGF.Int8Ty, CGM.getNaturalTypeAlignment(SharedsTy));
3951     LValue Dest = CGF.MakeAddrLValue(KmpTaskSharedsPtr, SharedsTy);
3952     LValue Src = CGF.MakeAddrLValue(Shareds, SharedsTy);
3953     CGF.EmitAggregateCopy(Dest, Src, SharedsTy, AggValueSlot::DoesNotOverlap);
3954   }
3955   // Emit initial values for private copies (if any).
3956   TaskResultTy Result;
3957   if (!Privates.empty()) {
3958     emitPrivatesInit(CGF, D, KmpTaskSharedsPtr, Base, KmpTaskTWithPrivatesQTyRD,
3959                      SharedsTy, SharedsPtrTy, Data, Privates,
3960                      /*ForDup=*/false);
3961     if (isOpenMPTaskLoopDirective(D.getDirectiveKind()) &&
3962         (!Data.LastprivateVars.empty() || checkInitIsRequired(CGF, Privates))) {
3963       Result.TaskDupFn = emitTaskDupFunction(
3964           CGM, Loc, D, KmpTaskTWithPrivatesPtrQTy, KmpTaskTWithPrivatesQTyRD,
3965           KmpTaskTQTyRD, SharedsTy, SharedsPtrTy, Data, Privates,
3966           /*WithLastIter=*/!Data.LastprivateVars.empty());
3967     }
3968   }
3969   // Fields of union "kmp_cmplrdata_t" for destructors and priority.
3970   enum { Priority = 0, Destructors = 1 };
3971   // Provide pointer to function with destructors for privates.
3972   auto FI = std::next(KmpTaskTQTyRD->field_begin(), Data1);
3973   const RecordDecl *KmpCmplrdataUD =
3974       (*FI)->getType()->getAsUnionType()->getDecl();
3975   if (NeedsCleanup) {
3976     llvm::Value *DestructorFn = emitDestructorsFunction(
3977         CGM, Loc, KmpInt32Ty, KmpTaskTWithPrivatesPtrQTy,
3978         KmpTaskTWithPrivatesQTy);
3979     LValue Data1LV = CGF.EmitLValueForField(TDBase, *FI);
3980     LValue DestructorsLV = CGF.EmitLValueForField(
3981         Data1LV, *std::next(KmpCmplrdataUD->field_begin(), Destructors));
3982     CGF.EmitStoreOfScalar(CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
3983                               DestructorFn, KmpRoutineEntryPtrTy),
3984                           DestructorsLV);
3985   }
3986   // Set priority.
3987   if (Data.Priority.getInt()) {
3988     LValue Data2LV = CGF.EmitLValueForField(
3989         TDBase, *std::next(KmpTaskTQTyRD->field_begin(), Data2));
3990     LValue PriorityLV = CGF.EmitLValueForField(
3991         Data2LV, *std::next(KmpCmplrdataUD->field_begin(), Priority));
3992     CGF.EmitStoreOfScalar(Data.Priority.getPointer(), PriorityLV);
3993   }
3994   Result.NewTask = NewTask;
3995   Result.TaskEntry = TaskEntry;
3996   Result.NewTaskNewTaskTTy = NewTaskNewTaskTTy;
3997   Result.TDBase = TDBase;
3998   Result.KmpTaskTQTyRD = KmpTaskTQTyRD;
3999   return Result;
4000 }
4001 
4002 /// Translates internal dependency kind into the runtime kind.
translateDependencyKind(OpenMPDependClauseKind K)4003 static RTLDependenceKindTy translateDependencyKind(OpenMPDependClauseKind K) {
4004   RTLDependenceKindTy DepKind;
4005   switch (K) {
4006   case OMPC_DEPEND_in:
4007     DepKind = RTLDependenceKindTy::DepIn;
4008     break;
4009   // Out and InOut dependencies must use the same code.
4010   case OMPC_DEPEND_out:
4011   case OMPC_DEPEND_inout:
4012     DepKind = RTLDependenceKindTy::DepInOut;
4013     break;
4014   case OMPC_DEPEND_mutexinoutset:
4015     DepKind = RTLDependenceKindTy::DepMutexInOutSet;
4016     break;
4017   case OMPC_DEPEND_inoutset:
4018     DepKind = RTLDependenceKindTy::DepInOutSet;
4019     break;
4020   case OMPC_DEPEND_outallmemory:
4021     DepKind = RTLDependenceKindTy::DepOmpAllMem;
4022     break;
4023   case OMPC_DEPEND_source:
4024   case OMPC_DEPEND_sink:
4025   case OMPC_DEPEND_depobj:
4026   case OMPC_DEPEND_inoutallmemory:
4027   case OMPC_DEPEND_unknown:
4028     llvm_unreachable("Unknown task dependence type");
4029   }
4030   return DepKind;
4031 }
4032 
4033 /// Builds kmp_depend_info, if it is not built yet, and builds flags type.
getDependTypes(ASTContext & C,QualType & KmpDependInfoTy,QualType & FlagsTy)4034 static void getDependTypes(ASTContext &C, QualType &KmpDependInfoTy,
4035                            QualType &FlagsTy) {
4036   FlagsTy = C.getIntTypeForBitwidth(C.getTypeSize(C.BoolTy), /*Signed=*/false);
4037   if (KmpDependInfoTy.isNull()) {
4038     RecordDecl *KmpDependInfoRD = C.buildImplicitRecord("kmp_depend_info");
4039     KmpDependInfoRD->startDefinition();
4040     addFieldToRecordDecl(C, KmpDependInfoRD, C.getIntPtrType());
4041     addFieldToRecordDecl(C, KmpDependInfoRD, C.getSizeType());
4042     addFieldToRecordDecl(C, KmpDependInfoRD, FlagsTy);
4043     KmpDependInfoRD->completeDefinition();
4044     KmpDependInfoTy = C.getRecordType(KmpDependInfoRD);
4045   }
4046 }
4047 
4048 std::pair<llvm::Value *, LValue>
getDepobjElements(CodeGenFunction & CGF,LValue DepobjLVal,SourceLocation Loc)4049 CGOpenMPRuntime::getDepobjElements(CodeGenFunction &CGF, LValue DepobjLVal,
4050                                    SourceLocation Loc) {
4051   ASTContext &C = CGM.getContext();
4052   QualType FlagsTy;
4053   getDependTypes(C, KmpDependInfoTy, FlagsTy);
4054   RecordDecl *KmpDependInfoRD =
4055       cast<RecordDecl>(KmpDependInfoTy->getAsTagDecl());
4056   QualType KmpDependInfoPtrTy = C.getPointerType(KmpDependInfoTy);
4057   LValue Base = CGF.EmitLoadOfPointerLValue(
4058       DepobjLVal.getAddress().withElementType(
4059           CGF.ConvertTypeForMem(KmpDependInfoPtrTy)),
4060       KmpDependInfoPtrTy->castAs<PointerType>());
4061   Address DepObjAddr = CGF.Builder.CreateGEP(
4062       CGF, Base.getAddress(),
4063       llvm::ConstantInt::get(CGF.IntPtrTy, -1, /*isSigned=*/true));
4064   LValue NumDepsBase = CGF.MakeAddrLValue(
4065       DepObjAddr, KmpDependInfoTy, Base.getBaseInfo(), Base.getTBAAInfo());
4066   // NumDeps = deps[i].base_addr;
4067   LValue BaseAddrLVal = CGF.EmitLValueForField(
4068       NumDepsBase,
4069       *std::next(KmpDependInfoRD->field_begin(),
4070                  static_cast<unsigned int>(RTLDependInfoFields::BaseAddr)));
4071   llvm::Value *NumDeps = CGF.EmitLoadOfScalar(BaseAddrLVal, Loc);
4072   return std::make_pair(NumDeps, Base);
4073 }
4074 
emitDependData(CodeGenFunction & CGF,QualType & KmpDependInfoTy,llvm::PointerUnion<unsigned *,LValue * > Pos,const OMPTaskDataTy::DependData & Data,Address DependenciesArray)4075 static void emitDependData(CodeGenFunction &CGF, QualType &KmpDependInfoTy,
4076                            llvm::PointerUnion<unsigned *, LValue *> Pos,
4077                            const OMPTaskDataTy::DependData &Data,
4078                            Address DependenciesArray) {
4079   CodeGenModule &CGM = CGF.CGM;
4080   ASTContext &C = CGM.getContext();
4081   QualType FlagsTy;
4082   getDependTypes(C, KmpDependInfoTy, FlagsTy);
4083   RecordDecl *KmpDependInfoRD =
4084       cast<RecordDecl>(KmpDependInfoTy->getAsTagDecl());
4085   llvm::Type *LLVMFlagsTy = CGF.ConvertTypeForMem(FlagsTy);
4086 
4087   OMPIteratorGeneratorScope IteratorScope(
4088       CGF, cast_or_null<OMPIteratorExpr>(
4089                Data.IteratorExpr ? Data.IteratorExpr->IgnoreParenImpCasts()
4090                                  : nullptr));
4091   for (const Expr *E : Data.DepExprs) {
4092     llvm::Value *Addr;
4093     llvm::Value *Size;
4094 
4095     // The expression will be a nullptr in the 'omp_all_memory' case.
4096     if (E) {
4097       std::tie(Addr, Size) = getPointerAndSize(CGF, E);
4098       Addr = CGF.Builder.CreatePtrToInt(Addr, CGF.IntPtrTy);
4099     } else {
4100       Addr = llvm::ConstantInt::get(CGF.IntPtrTy, 0);
4101       Size = llvm::ConstantInt::get(CGF.SizeTy, 0);
4102     }
4103     LValue Base;
4104     if (unsigned *P = Pos.dyn_cast<unsigned *>()) {
4105       Base = CGF.MakeAddrLValue(
4106           CGF.Builder.CreateConstGEP(DependenciesArray, *P), KmpDependInfoTy);
4107     } else {
4108       assert(E && "Expected a non-null expression");
4109       LValue &PosLVal = *Pos.get<LValue *>();
4110       llvm::Value *Idx = CGF.EmitLoadOfScalar(PosLVal, E->getExprLoc());
4111       Base = CGF.MakeAddrLValue(
4112           CGF.Builder.CreateGEP(CGF, DependenciesArray, Idx), KmpDependInfoTy);
4113     }
4114     // deps[i].base_addr = &<Dependencies[i].second>;
4115     LValue BaseAddrLVal = CGF.EmitLValueForField(
4116         Base,
4117         *std::next(KmpDependInfoRD->field_begin(),
4118                    static_cast<unsigned int>(RTLDependInfoFields::BaseAddr)));
4119     CGF.EmitStoreOfScalar(Addr, BaseAddrLVal);
4120     // deps[i].len = sizeof(<Dependencies[i].second>);
4121     LValue LenLVal = CGF.EmitLValueForField(
4122         Base, *std::next(KmpDependInfoRD->field_begin(),
4123                          static_cast<unsigned int>(RTLDependInfoFields::Len)));
4124     CGF.EmitStoreOfScalar(Size, LenLVal);
4125     // deps[i].flags = <Dependencies[i].first>;
4126     RTLDependenceKindTy DepKind = translateDependencyKind(Data.DepKind);
4127     LValue FlagsLVal = CGF.EmitLValueForField(
4128         Base,
4129         *std::next(KmpDependInfoRD->field_begin(),
4130                    static_cast<unsigned int>(RTLDependInfoFields::Flags)));
4131     CGF.EmitStoreOfScalar(
4132         llvm::ConstantInt::get(LLVMFlagsTy, static_cast<unsigned int>(DepKind)),
4133         FlagsLVal);
4134     if (unsigned *P = Pos.dyn_cast<unsigned *>()) {
4135       ++(*P);
4136     } else {
4137       LValue &PosLVal = *Pos.get<LValue *>();
4138       llvm::Value *Idx = CGF.EmitLoadOfScalar(PosLVal, E->getExprLoc());
4139       Idx = CGF.Builder.CreateNUWAdd(Idx,
4140                                      llvm::ConstantInt::get(Idx->getType(), 1));
4141       CGF.EmitStoreOfScalar(Idx, PosLVal);
4142     }
4143   }
4144 }
4145 
emitDepobjElementsSizes(CodeGenFunction & CGF,QualType & KmpDependInfoTy,const OMPTaskDataTy::DependData & Data)4146 SmallVector<llvm::Value *, 4> CGOpenMPRuntime::emitDepobjElementsSizes(
4147     CodeGenFunction &CGF, QualType &KmpDependInfoTy,
4148     const OMPTaskDataTy::DependData &Data) {
4149   assert(Data.DepKind == OMPC_DEPEND_depobj &&
4150          "Expected depobj dependency kind.");
4151   SmallVector<llvm::Value *, 4> Sizes;
4152   SmallVector<LValue, 4> SizeLVals;
4153   ASTContext &C = CGF.getContext();
4154   {
4155     OMPIteratorGeneratorScope IteratorScope(
4156         CGF, cast_or_null<OMPIteratorExpr>(
4157                  Data.IteratorExpr ? Data.IteratorExpr->IgnoreParenImpCasts()
4158                                    : nullptr));
4159     for (const Expr *E : Data.DepExprs) {
4160       llvm::Value *NumDeps;
4161       LValue Base;
4162       LValue DepobjLVal = CGF.EmitLValue(E->IgnoreParenImpCasts());
4163       std::tie(NumDeps, Base) =
4164           getDepobjElements(CGF, DepobjLVal, E->getExprLoc());
4165       LValue NumLVal = CGF.MakeAddrLValue(
4166           CGF.CreateMemTemp(C.getUIntPtrType(), "depobj.size.addr"),
4167           C.getUIntPtrType());
4168       CGF.Builder.CreateStore(llvm::ConstantInt::get(CGF.IntPtrTy, 0),
4169                               NumLVal.getAddress());
4170       llvm::Value *PrevVal = CGF.EmitLoadOfScalar(NumLVal, E->getExprLoc());
4171       llvm::Value *Add = CGF.Builder.CreateNUWAdd(PrevVal, NumDeps);
4172       CGF.EmitStoreOfScalar(Add, NumLVal);
4173       SizeLVals.push_back(NumLVal);
4174     }
4175   }
4176   for (unsigned I = 0, E = SizeLVals.size(); I < E; ++I) {
4177     llvm::Value *Size =
4178         CGF.EmitLoadOfScalar(SizeLVals[I], Data.DepExprs[I]->getExprLoc());
4179     Sizes.push_back(Size);
4180   }
4181   return Sizes;
4182 }
4183 
emitDepobjElements(CodeGenFunction & CGF,QualType & KmpDependInfoTy,LValue PosLVal,const OMPTaskDataTy::DependData & Data,Address DependenciesArray)4184 void CGOpenMPRuntime::emitDepobjElements(CodeGenFunction &CGF,
4185                                          QualType &KmpDependInfoTy,
4186                                          LValue PosLVal,
4187                                          const OMPTaskDataTy::DependData &Data,
4188                                          Address DependenciesArray) {
4189   assert(Data.DepKind == OMPC_DEPEND_depobj &&
4190          "Expected depobj dependency kind.");
4191   llvm::Value *ElSize = CGF.getTypeSize(KmpDependInfoTy);
4192   {
4193     OMPIteratorGeneratorScope IteratorScope(
4194         CGF, cast_or_null<OMPIteratorExpr>(
4195                  Data.IteratorExpr ? Data.IteratorExpr->IgnoreParenImpCasts()
4196                                    : nullptr));
4197     for (unsigned I = 0, End = Data.DepExprs.size(); I < End; ++I) {
4198       const Expr *E = Data.DepExprs[I];
4199       llvm::Value *NumDeps;
4200       LValue Base;
4201       LValue DepobjLVal = CGF.EmitLValue(E->IgnoreParenImpCasts());
4202       std::tie(NumDeps, Base) =
4203           getDepobjElements(CGF, DepobjLVal, E->getExprLoc());
4204 
4205       // memcopy dependency data.
4206       llvm::Value *Size = CGF.Builder.CreateNUWMul(
4207           ElSize,
4208           CGF.Builder.CreateIntCast(NumDeps, CGF.SizeTy, /*isSigned=*/false));
4209       llvm::Value *Pos = CGF.EmitLoadOfScalar(PosLVal, E->getExprLoc());
4210       Address DepAddr = CGF.Builder.CreateGEP(CGF, DependenciesArray, Pos);
4211       CGF.Builder.CreateMemCpy(DepAddr, Base.getAddress(), Size);
4212 
4213       // Increase pos.
4214       // pos += size;
4215       llvm::Value *Add = CGF.Builder.CreateNUWAdd(Pos, NumDeps);
4216       CGF.EmitStoreOfScalar(Add, PosLVal);
4217     }
4218   }
4219 }
4220 
emitDependClause(CodeGenFunction & CGF,ArrayRef<OMPTaskDataTy::DependData> Dependencies,SourceLocation Loc)4221 std::pair<llvm::Value *, Address> CGOpenMPRuntime::emitDependClause(
4222     CodeGenFunction &CGF, ArrayRef<OMPTaskDataTy::DependData> Dependencies,
4223     SourceLocation Loc) {
4224   if (llvm::all_of(Dependencies, [](const OMPTaskDataTy::DependData &D) {
4225         return D.DepExprs.empty();
4226       }))
4227     return std::make_pair(nullptr, Address::invalid());
4228   // Process list of dependencies.
4229   ASTContext &C = CGM.getContext();
4230   Address DependenciesArray = Address::invalid();
4231   llvm::Value *NumOfElements = nullptr;
4232   unsigned NumDependencies = std::accumulate(
4233       Dependencies.begin(), Dependencies.end(), 0,
4234       [](unsigned V, const OMPTaskDataTy::DependData &D) {
4235         return D.DepKind == OMPC_DEPEND_depobj
4236                    ? V
4237                    : (V + (D.IteratorExpr ? 0 : D.DepExprs.size()));
4238       });
4239   QualType FlagsTy;
4240   getDependTypes(C, KmpDependInfoTy, FlagsTy);
4241   bool HasDepobjDeps = false;
4242   bool HasRegularWithIterators = false;
4243   llvm::Value *NumOfDepobjElements = llvm::ConstantInt::get(CGF.IntPtrTy, 0);
4244   llvm::Value *NumOfRegularWithIterators =
4245       llvm::ConstantInt::get(CGF.IntPtrTy, 0);
4246   // Calculate number of depobj dependencies and regular deps with the
4247   // iterators.
4248   for (const OMPTaskDataTy::DependData &D : Dependencies) {
4249     if (D.DepKind == OMPC_DEPEND_depobj) {
4250       SmallVector<llvm::Value *, 4> Sizes =
4251           emitDepobjElementsSizes(CGF, KmpDependInfoTy, D);
4252       for (llvm::Value *Size : Sizes) {
4253         NumOfDepobjElements =
4254             CGF.Builder.CreateNUWAdd(NumOfDepobjElements, Size);
4255       }
4256       HasDepobjDeps = true;
4257       continue;
4258     }
4259     // Include number of iterations, if any.
4260 
4261     if (const auto *IE = cast_or_null<OMPIteratorExpr>(D.IteratorExpr)) {
4262       llvm::Value *ClauseIteratorSpace =
4263           llvm::ConstantInt::get(CGF.IntPtrTy, 1);
4264       for (unsigned I = 0, E = IE->numOfIterators(); I < E; ++I) {
4265         llvm::Value *Sz = CGF.EmitScalarExpr(IE->getHelper(I).Upper);
4266         Sz = CGF.Builder.CreateIntCast(Sz, CGF.IntPtrTy, /*isSigned=*/false);
4267         ClauseIteratorSpace = CGF.Builder.CreateNUWMul(Sz, ClauseIteratorSpace);
4268       }
4269       llvm::Value *NumClauseDeps = CGF.Builder.CreateNUWMul(
4270           ClauseIteratorSpace,
4271           llvm::ConstantInt::get(CGF.IntPtrTy, D.DepExprs.size()));
4272       NumOfRegularWithIterators =
4273           CGF.Builder.CreateNUWAdd(NumOfRegularWithIterators, NumClauseDeps);
4274       HasRegularWithIterators = true;
4275       continue;
4276     }
4277   }
4278 
4279   QualType KmpDependInfoArrayTy;
4280   if (HasDepobjDeps || HasRegularWithIterators) {
4281     NumOfElements = llvm::ConstantInt::get(CGM.IntPtrTy, NumDependencies,
4282                                            /*isSigned=*/false);
4283     if (HasDepobjDeps) {
4284       NumOfElements =
4285           CGF.Builder.CreateNUWAdd(NumOfDepobjElements, NumOfElements);
4286     }
4287     if (HasRegularWithIterators) {
4288       NumOfElements =
4289           CGF.Builder.CreateNUWAdd(NumOfRegularWithIterators, NumOfElements);
4290     }
4291     auto *OVE = new (C) OpaqueValueExpr(
4292         Loc, C.getIntTypeForBitwidth(/*DestWidth=*/64, /*Signed=*/0),
4293         VK_PRValue);
4294     CodeGenFunction::OpaqueValueMapping OpaqueMap(CGF, OVE,
4295                                                   RValue::get(NumOfElements));
4296     KmpDependInfoArrayTy =
4297         C.getVariableArrayType(KmpDependInfoTy, OVE, ArraySizeModifier::Normal,
4298                                /*IndexTypeQuals=*/0, SourceRange(Loc, Loc));
4299     // CGF.EmitVariablyModifiedType(KmpDependInfoArrayTy);
4300     // Properly emit variable-sized array.
4301     auto *PD = ImplicitParamDecl::Create(C, KmpDependInfoArrayTy,
4302                                          ImplicitParamKind::Other);
4303     CGF.EmitVarDecl(*PD);
4304     DependenciesArray = CGF.GetAddrOfLocalVar(PD);
4305     NumOfElements = CGF.Builder.CreateIntCast(NumOfElements, CGF.Int32Ty,
4306                                               /*isSigned=*/false);
4307   } else {
4308     KmpDependInfoArrayTy = C.getConstantArrayType(
4309         KmpDependInfoTy, llvm::APInt(/*numBits=*/64, NumDependencies), nullptr,
4310         ArraySizeModifier::Normal, /*IndexTypeQuals=*/0);
4311     DependenciesArray =
4312         CGF.CreateMemTemp(KmpDependInfoArrayTy, ".dep.arr.addr");
4313     DependenciesArray = CGF.Builder.CreateConstArrayGEP(DependenciesArray, 0);
4314     NumOfElements = llvm::ConstantInt::get(CGM.Int32Ty, NumDependencies,
4315                                            /*isSigned=*/false);
4316   }
4317   unsigned Pos = 0;
4318   for (unsigned I = 0, End = Dependencies.size(); I < End; ++I) {
4319     if (Dependencies[I].DepKind == OMPC_DEPEND_depobj ||
4320         Dependencies[I].IteratorExpr)
4321       continue;
4322     emitDependData(CGF, KmpDependInfoTy, &Pos, Dependencies[I],
4323                    DependenciesArray);
4324   }
4325   // Copy regular dependencies with iterators.
4326   LValue PosLVal = CGF.MakeAddrLValue(
4327       CGF.CreateMemTemp(C.getSizeType(), "dep.counter.addr"), C.getSizeType());
4328   CGF.EmitStoreOfScalar(llvm::ConstantInt::get(CGF.SizeTy, Pos), PosLVal);
4329   for (unsigned I = 0, End = Dependencies.size(); I < End; ++I) {
4330     if (Dependencies[I].DepKind == OMPC_DEPEND_depobj ||
4331         !Dependencies[I].IteratorExpr)
4332       continue;
4333     emitDependData(CGF, KmpDependInfoTy, &PosLVal, Dependencies[I],
4334                    DependenciesArray);
4335   }
4336   // Copy final depobj arrays without iterators.
4337   if (HasDepobjDeps) {
4338     for (unsigned I = 0, End = Dependencies.size(); I < End; ++I) {
4339       if (Dependencies[I].DepKind != OMPC_DEPEND_depobj)
4340         continue;
4341       emitDepobjElements(CGF, KmpDependInfoTy, PosLVal, Dependencies[I],
4342                          DependenciesArray);
4343     }
4344   }
4345   DependenciesArray = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
4346       DependenciesArray, CGF.VoidPtrTy, CGF.Int8Ty);
4347   return std::make_pair(NumOfElements, DependenciesArray);
4348 }
4349 
emitDepobjDependClause(CodeGenFunction & CGF,const OMPTaskDataTy::DependData & Dependencies,SourceLocation Loc)4350 Address CGOpenMPRuntime::emitDepobjDependClause(
4351     CodeGenFunction &CGF, const OMPTaskDataTy::DependData &Dependencies,
4352     SourceLocation Loc) {
4353   if (Dependencies.DepExprs.empty())
4354     return Address::invalid();
4355   // Process list of dependencies.
4356   ASTContext &C = CGM.getContext();
4357   Address DependenciesArray = Address::invalid();
4358   unsigned NumDependencies = Dependencies.DepExprs.size();
4359   QualType FlagsTy;
4360   getDependTypes(C, KmpDependInfoTy, FlagsTy);
4361   RecordDecl *KmpDependInfoRD =
4362       cast<RecordDecl>(KmpDependInfoTy->getAsTagDecl());
4363 
4364   llvm::Value *Size;
4365   // Define type kmp_depend_info[<Dependencies.size()>];
4366   // For depobj reserve one extra element to store the number of elements.
4367   // It is required to handle depobj(x) update(in) construct.
4368   // kmp_depend_info[<Dependencies.size()>] deps;
4369   llvm::Value *NumDepsVal;
4370   CharUnits Align = C.getTypeAlignInChars(KmpDependInfoTy);
4371   if (const auto *IE =
4372           cast_or_null<OMPIteratorExpr>(Dependencies.IteratorExpr)) {
4373     NumDepsVal = llvm::ConstantInt::get(CGF.SizeTy, 1);
4374     for (unsigned I = 0, E = IE->numOfIterators(); I < E; ++I) {
4375       llvm::Value *Sz = CGF.EmitScalarExpr(IE->getHelper(I).Upper);
4376       Sz = CGF.Builder.CreateIntCast(Sz, CGF.SizeTy, /*isSigned=*/false);
4377       NumDepsVal = CGF.Builder.CreateNUWMul(NumDepsVal, Sz);
4378     }
4379     Size = CGF.Builder.CreateNUWAdd(llvm::ConstantInt::get(CGF.SizeTy, 1),
4380                                     NumDepsVal);
4381     CharUnits SizeInBytes =
4382         C.getTypeSizeInChars(KmpDependInfoTy).alignTo(Align);
4383     llvm::Value *RecSize = CGM.getSize(SizeInBytes);
4384     Size = CGF.Builder.CreateNUWMul(Size, RecSize);
4385     NumDepsVal =
4386         CGF.Builder.CreateIntCast(NumDepsVal, CGF.IntPtrTy, /*isSigned=*/false);
4387   } else {
4388     QualType KmpDependInfoArrayTy = C.getConstantArrayType(
4389         KmpDependInfoTy, llvm::APInt(/*numBits=*/64, NumDependencies + 1),
4390         nullptr, ArraySizeModifier::Normal, /*IndexTypeQuals=*/0);
4391     CharUnits Sz = C.getTypeSizeInChars(KmpDependInfoArrayTy);
4392     Size = CGM.getSize(Sz.alignTo(Align));
4393     NumDepsVal = llvm::ConstantInt::get(CGF.IntPtrTy, NumDependencies);
4394   }
4395   // Need to allocate on the dynamic memory.
4396   llvm::Value *ThreadID = getThreadID(CGF, Loc);
4397   // Use default allocator.
4398   llvm::Value *Allocator = llvm::ConstantPointerNull::get(CGF.VoidPtrTy);
4399   llvm::Value *Args[] = {ThreadID, Size, Allocator};
4400 
4401   llvm::Value *Addr =
4402       CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
4403                               CGM.getModule(), OMPRTL___kmpc_alloc),
4404                           Args, ".dep.arr.addr");
4405   llvm::Type *KmpDependInfoLlvmTy = CGF.ConvertTypeForMem(KmpDependInfoTy);
4406   Addr = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
4407       Addr, KmpDependInfoLlvmTy->getPointerTo());
4408   DependenciesArray = Address(Addr, KmpDependInfoLlvmTy, Align);
4409   // Write number of elements in the first element of array for depobj.
4410   LValue Base = CGF.MakeAddrLValue(DependenciesArray, KmpDependInfoTy);
4411   // deps[i].base_addr = NumDependencies;
4412   LValue BaseAddrLVal = CGF.EmitLValueForField(
4413       Base,
4414       *std::next(KmpDependInfoRD->field_begin(),
4415                  static_cast<unsigned int>(RTLDependInfoFields::BaseAddr)));
4416   CGF.EmitStoreOfScalar(NumDepsVal, BaseAddrLVal);
4417   llvm::PointerUnion<unsigned *, LValue *> Pos;
4418   unsigned Idx = 1;
4419   LValue PosLVal;
4420   if (Dependencies.IteratorExpr) {
4421     PosLVal = CGF.MakeAddrLValue(
4422         CGF.CreateMemTemp(C.getSizeType(), "iterator.counter.addr"),
4423         C.getSizeType());
4424     CGF.EmitStoreOfScalar(llvm::ConstantInt::get(CGF.SizeTy, Idx), PosLVal,
4425                           /*IsInit=*/true);
4426     Pos = &PosLVal;
4427   } else {
4428     Pos = &Idx;
4429   }
4430   emitDependData(CGF, KmpDependInfoTy, Pos, Dependencies, DependenciesArray);
4431   DependenciesArray = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
4432       CGF.Builder.CreateConstGEP(DependenciesArray, 1), CGF.VoidPtrTy,
4433       CGF.Int8Ty);
4434   return DependenciesArray;
4435 }
4436 
emitDestroyClause(CodeGenFunction & CGF,LValue DepobjLVal,SourceLocation Loc)4437 void CGOpenMPRuntime::emitDestroyClause(CodeGenFunction &CGF, LValue DepobjLVal,
4438                                         SourceLocation Loc) {
4439   ASTContext &C = CGM.getContext();
4440   QualType FlagsTy;
4441   getDependTypes(C, KmpDependInfoTy, FlagsTy);
4442   LValue Base = CGF.EmitLoadOfPointerLValue(DepobjLVal.getAddress(),
4443                                             C.VoidPtrTy.castAs<PointerType>());
4444   QualType KmpDependInfoPtrTy = C.getPointerType(KmpDependInfoTy);
4445   Address Addr = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
4446       Base.getAddress(), CGF.ConvertTypeForMem(KmpDependInfoPtrTy),
4447       CGF.ConvertTypeForMem(KmpDependInfoTy));
4448   llvm::Value *DepObjAddr = CGF.Builder.CreateGEP(
4449       Addr.getElementType(), Addr.emitRawPointer(CGF),
4450       llvm::ConstantInt::get(CGF.IntPtrTy, -1, /*isSigned=*/true));
4451   DepObjAddr = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(DepObjAddr,
4452                                                                CGF.VoidPtrTy);
4453   llvm::Value *ThreadID = getThreadID(CGF, Loc);
4454   // Use default allocator.
4455   llvm::Value *Allocator = llvm::ConstantPointerNull::get(CGF.VoidPtrTy);
4456   llvm::Value *Args[] = {ThreadID, DepObjAddr, Allocator};
4457 
4458   // _kmpc_free(gtid, addr, nullptr);
4459   (void)CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
4460                                 CGM.getModule(), OMPRTL___kmpc_free),
4461                             Args);
4462 }
4463 
emitUpdateClause(CodeGenFunction & CGF,LValue DepobjLVal,OpenMPDependClauseKind NewDepKind,SourceLocation Loc)4464 void CGOpenMPRuntime::emitUpdateClause(CodeGenFunction &CGF, LValue DepobjLVal,
4465                                        OpenMPDependClauseKind NewDepKind,
4466                                        SourceLocation Loc) {
4467   ASTContext &C = CGM.getContext();
4468   QualType FlagsTy;
4469   getDependTypes(C, KmpDependInfoTy, FlagsTy);
4470   RecordDecl *KmpDependInfoRD =
4471       cast<RecordDecl>(KmpDependInfoTy->getAsTagDecl());
4472   llvm::Type *LLVMFlagsTy = CGF.ConvertTypeForMem(FlagsTy);
4473   llvm::Value *NumDeps;
4474   LValue Base;
4475   std::tie(NumDeps, Base) = getDepobjElements(CGF, DepobjLVal, Loc);
4476 
4477   Address Begin = Base.getAddress();
4478   // Cast from pointer to array type to pointer to single element.
4479   llvm::Value *End = CGF.Builder.CreateGEP(Begin.getElementType(),
4480                                            Begin.emitRawPointer(CGF), NumDeps);
4481   // The basic structure here is a while-do loop.
4482   llvm::BasicBlock *BodyBB = CGF.createBasicBlock("omp.body");
4483   llvm::BasicBlock *DoneBB = CGF.createBasicBlock("omp.done");
4484   llvm::BasicBlock *EntryBB = CGF.Builder.GetInsertBlock();
4485   CGF.EmitBlock(BodyBB);
4486   llvm::PHINode *ElementPHI =
4487       CGF.Builder.CreatePHI(Begin.getType(), 2, "omp.elementPast");
4488   ElementPHI->addIncoming(Begin.emitRawPointer(CGF), EntryBB);
4489   Begin = Begin.withPointer(ElementPHI, KnownNonNull);
4490   Base = CGF.MakeAddrLValue(Begin, KmpDependInfoTy, Base.getBaseInfo(),
4491                             Base.getTBAAInfo());
4492   // deps[i].flags = NewDepKind;
4493   RTLDependenceKindTy DepKind = translateDependencyKind(NewDepKind);
4494   LValue FlagsLVal = CGF.EmitLValueForField(
4495       Base, *std::next(KmpDependInfoRD->field_begin(),
4496                        static_cast<unsigned int>(RTLDependInfoFields::Flags)));
4497   CGF.EmitStoreOfScalar(
4498       llvm::ConstantInt::get(LLVMFlagsTy, static_cast<unsigned int>(DepKind)),
4499       FlagsLVal);
4500 
4501   // Shift the address forward by one element.
4502   llvm::Value *ElementNext =
4503       CGF.Builder.CreateConstGEP(Begin, /*Index=*/1, "omp.elementNext")
4504           .emitRawPointer(CGF);
4505   ElementPHI->addIncoming(ElementNext, CGF.Builder.GetInsertBlock());
4506   llvm::Value *IsEmpty =
4507       CGF.Builder.CreateICmpEQ(ElementNext, End, "omp.isempty");
4508   CGF.Builder.CreateCondBr(IsEmpty, DoneBB, BodyBB);
4509   // Done.
4510   CGF.EmitBlock(DoneBB, /*IsFinished=*/true);
4511 }
4512 
emitTaskCall(CodeGenFunction & CGF,SourceLocation Loc,const OMPExecutableDirective & D,llvm::Function * TaskFunction,QualType SharedsTy,Address Shareds,const Expr * IfCond,const OMPTaskDataTy & Data)4513 void CGOpenMPRuntime::emitTaskCall(CodeGenFunction &CGF, SourceLocation Loc,
4514                                    const OMPExecutableDirective &D,
4515                                    llvm::Function *TaskFunction,
4516                                    QualType SharedsTy, Address Shareds,
4517                                    const Expr *IfCond,
4518                                    const OMPTaskDataTy &Data) {
4519   if (!CGF.HaveInsertPoint())
4520     return;
4521 
4522   TaskResultTy Result =
4523       emitTaskInit(CGF, Loc, D, TaskFunction, SharedsTy, Shareds, Data);
4524   llvm::Value *NewTask = Result.NewTask;
4525   llvm::Function *TaskEntry = Result.TaskEntry;
4526   llvm::Value *NewTaskNewTaskTTy = Result.NewTaskNewTaskTTy;
4527   LValue TDBase = Result.TDBase;
4528   const RecordDecl *KmpTaskTQTyRD = Result.KmpTaskTQTyRD;
4529   // Process list of dependences.
4530   Address DependenciesArray = Address::invalid();
4531   llvm::Value *NumOfElements;
4532   std::tie(NumOfElements, DependenciesArray) =
4533       emitDependClause(CGF, Data.Dependences, Loc);
4534 
4535   // NOTE: routine and part_id fields are initialized by __kmpc_omp_task_alloc()
4536   // libcall.
4537   // Build kmp_int32 __kmpc_omp_task_with_deps(ident_t *, kmp_int32 gtid,
4538   // kmp_task_t *new_task, kmp_int32 ndeps, kmp_depend_info_t *dep_list,
4539   // kmp_int32 ndeps_noalias, kmp_depend_info_t *noalias_dep_list) if dependence
4540   // list is not empty
4541   llvm::Value *ThreadID = getThreadID(CGF, Loc);
4542   llvm::Value *UpLoc = emitUpdateLocation(CGF, Loc);
4543   llvm::Value *TaskArgs[] = { UpLoc, ThreadID, NewTask };
4544   llvm::Value *DepTaskArgs[7];
4545   if (!Data.Dependences.empty()) {
4546     DepTaskArgs[0] = UpLoc;
4547     DepTaskArgs[1] = ThreadID;
4548     DepTaskArgs[2] = NewTask;
4549     DepTaskArgs[3] = NumOfElements;
4550     DepTaskArgs[4] = DependenciesArray.emitRawPointer(CGF);
4551     DepTaskArgs[5] = CGF.Builder.getInt32(0);
4552     DepTaskArgs[6] = llvm::ConstantPointerNull::get(CGF.VoidPtrTy);
4553   }
4554   auto &&ThenCodeGen = [this, &Data, TDBase, KmpTaskTQTyRD, &TaskArgs,
4555                         &DepTaskArgs](CodeGenFunction &CGF, PrePostActionTy &) {
4556     if (!Data.Tied) {
4557       auto PartIdFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTPartId);
4558       LValue PartIdLVal = CGF.EmitLValueForField(TDBase, *PartIdFI);
4559       CGF.EmitStoreOfScalar(CGF.Builder.getInt32(0), PartIdLVal);
4560     }
4561     if (!Data.Dependences.empty()) {
4562       CGF.EmitRuntimeCall(
4563           OMPBuilder.getOrCreateRuntimeFunction(
4564               CGM.getModule(), OMPRTL___kmpc_omp_task_with_deps),
4565           DepTaskArgs);
4566     } else {
4567       CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
4568                               CGM.getModule(), OMPRTL___kmpc_omp_task),
4569                           TaskArgs);
4570     }
4571     // Check if parent region is untied and build return for untied task;
4572     if (auto *Region =
4573             dyn_cast_or_null<CGOpenMPRegionInfo>(CGF.CapturedStmtInfo))
4574       Region->emitUntiedSwitch(CGF);
4575   };
4576 
4577   llvm::Value *DepWaitTaskArgs[7];
4578   if (!Data.Dependences.empty()) {
4579     DepWaitTaskArgs[0] = UpLoc;
4580     DepWaitTaskArgs[1] = ThreadID;
4581     DepWaitTaskArgs[2] = NumOfElements;
4582     DepWaitTaskArgs[3] = DependenciesArray.emitRawPointer(CGF);
4583     DepWaitTaskArgs[4] = CGF.Builder.getInt32(0);
4584     DepWaitTaskArgs[5] = llvm::ConstantPointerNull::get(CGF.VoidPtrTy);
4585     DepWaitTaskArgs[6] =
4586         llvm::ConstantInt::get(CGF.Int32Ty, Data.HasNowaitClause);
4587   }
4588   auto &M = CGM.getModule();
4589   auto &&ElseCodeGen = [this, &M, &TaskArgs, ThreadID, NewTaskNewTaskTTy,
4590                         TaskEntry, &Data, &DepWaitTaskArgs,
4591                         Loc](CodeGenFunction &CGF, PrePostActionTy &) {
4592     CodeGenFunction::RunCleanupsScope LocalScope(CGF);
4593     // Build void __kmpc_omp_wait_deps(ident_t *, kmp_int32 gtid,
4594     // kmp_int32 ndeps, kmp_depend_info_t *dep_list, kmp_int32
4595     // ndeps_noalias, kmp_depend_info_t *noalias_dep_list); if dependence info
4596     // is specified.
4597     if (!Data.Dependences.empty())
4598       CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
4599                               M, OMPRTL___kmpc_omp_taskwait_deps_51),
4600                           DepWaitTaskArgs);
4601     // Call proxy_task_entry(gtid, new_task);
4602     auto &&CodeGen = [TaskEntry, ThreadID, NewTaskNewTaskTTy,
4603                       Loc](CodeGenFunction &CGF, PrePostActionTy &Action) {
4604       Action.Enter(CGF);
4605       llvm::Value *OutlinedFnArgs[] = {ThreadID, NewTaskNewTaskTTy};
4606       CGF.CGM.getOpenMPRuntime().emitOutlinedFunctionCall(CGF, Loc, TaskEntry,
4607                                                           OutlinedFnArgs);
4608     };
4609 
4610     // Build void __kmpc_omp_task_begin_if0(ident_t *, kmp_int32 gtid,
4611     // kmp_task_t *new_task);
4612     // Build void __kmpc_omp_task_complete_if0(ident_t *, kmp_int32 gtid,
4613     // kmp_task_t *new_task);
4614     RegionCodeGenTy RCG(CodeGen);
4615     CommonActionTy Action(OMPBuilder.getOrCreateRuntimeFunction(
4616                               M, OMPRTL___kmpc_omp_task_begin_if0),
4617                           TaskArgs,
4618                           OMPBuilder.getOrCreateRuntimeFunction(
4619                               M, OMPRTL___kmpc_omp_task_complete_if0),
4620                           TaskArgs);
4621     RCG.setAction(Action);
4622     RCG(CGF);
4623   };
4624 
4625   if (IfCond) {
4626     emitIfClause(CGF, IfCond, ThenCodeGen, ElseCodeGen);
4627   } else {
4628     RegionCodeGenTy ThenRCG(ThenCodeGen);
4629     ThenRCG(CGF);
4630   }
4631 }
4632 
emitTaskLoopCall(CodeGenFunction & CGF,SourceLocation Loc,const OMPLoopDirective & D,llvm::Function * TaskFunction,QualType SharedsTy,Address Shareds,const Expr * IfCond,const OMPTaskDataTy & Data)4633 void CGOpenMPRuntime::emitTaskLoopCall(CodeGenFunction &CGF, SourceLocation Loc,
4634                                        const OMPLoopDirective &D,
4635                                        llvm::Function *TaskFunction,
4636                                        QualType SharedsTy, Address Shareds,
4637                                        const Expr *IfCond,
4638                                        const OMPTaskDataTy &Data) {
4639   if (!CGF.HaveInsertPoint())
4640     return;
4641   TaskResultTy Result =
4642       emitTaskInit(CGF, Loc, D, TaskFunction, SharedsTy, Shareds, Data);
4643   // NOTE: routine and part_id fields are initialized by __kmpc_omp_task_alloc()
4644   // libcall.
4645   // Call to void __kmpc_taskloop(ident_t *loc, int gtid, kmp_task_t *task, int
4646   // if_val, kmp_uint64 *lb, kmp_uint64 *ub, kmp_int64 st, int nogroup, int
4647   // sched, kmp_uint64 grainsize, void *task_dup);
4648   llvm::Value *ThreadID = getThreadID(CGF, Loc);
4649   llvm::Value *UpLoc = emitUpdateLocation(CGF, Loc);
4650   llvm::Value *IfVal;
4651   if (IfCond) {
4652     IfVal = CGF.Builder.CreateIntCast(CGF.EvaluateExprAsBool(IfCond), CGF.IntTy,
4653                                       /*isSigned=*/true);
4654   } else {
4655     IfVal = llvm::ConstantInt::getSigned(CGF.IntTy, /*V=*/1);
4656   }
4657 
4658   LValue LBLVal = CGF.EmitLValueForField(
4659       Result.TDBase,
4660       *std::next(Result.KmpTaskTQTyRD->field_begin(), KmpTaskTLowerBound));
4661   const auto *LBVar =
4662       cast<VarDecl>(cast<DeclRefExpr>(D.getLowerBoundVariable())->getDecl());
4663   CGF.EmitAnyExprToMem(LBVar->getInit(), LBLVal.getAddress(), LBLVal.getQuals(),
4664                        /*IsInitializer=*/true);
4665   LValue UBLVal = CGF.EmitLValueForField(
4666       Result.TDBase,
4667       *std::next(Result.KmpTaskTQTyRD->field_begin(), KmpTaskTUpperBound));
4668   const auto *UBVar =
4669       cast<VarDecl>(cast<DeclRefExpr>(D.getUpperBoundVariable())->getDecl());
4670   CGF.EmitAnyExprToMem(UBVar->getInit(), UBLVal.getAddress(), UBLVal.getQuals(),
4671                        /*IsInitializer=*/true);
4672   LValue StLVal = CGF.EmitLValueForField(
4673       Result.TDBase,
4674       *std::next(Result.KmpTaskTQTyRD->field_begin(), KmpTaskTStride));
4675   const auto *StVar =
4676       cast<VarDecl>(cast<DeclRefExpr>(D.getStrideVariable())->getDecl());
4677   CGF.EmitAnyExprToMem(StVar->getInit(), StLVal.getAddress(), StLVal.getQuals(),
4678                        /*IsInitializer=*/true);
4679   // Store reductions address.
4680   LValue RedLVal = CGF.EmitLValueForField(
4681       Result.TDBase,
4682       *std::next(Result.KmpTaskTQTyRD->field_begin(), KmpTaskTReductions));
4683   if (Data.Reductions) {
4684     CGF.EmitStoreOfScalar(Data.Reductions, RedLVal);
4685   } else {
4686     CGF.EmitNullInitialization(RedLVal.getAddress(),
4687                                CGF.getContext().VoidPtrTy);
4688   }
4689   enum { NoSchedule = 0, Grainsize = 1, NumTasks = 2 };
4690   llvm::Value *TaskArgs[] = {
4691       UpLoc,
4692       ThreadID,
4693       Result.NewTask,
4694       IfVal,
4695       LBLVal.getPointer(CGF),
4696       UBLVal.getPointer(CGF),
4697       CGF.EmitLoadOfScalar(StLVal, Loc),
4698       llvm::ConstantInt::getSigned(
4699           CGF.IntTy, 1), // Always 1 because taskgroup emitted by the compiler
4700       llvm::ConstantInt::getSigned(
4701           CGF.IntTy, Data.Schedule.getPointer()
4702                          ? Data.Schedule.getInt() ? NumTasks : Grainsize
4703                          : NoSchedule),
4704       Data.Schedule.getPointer()
4705           ? CGF.Builder.CreateIntCast(Data.Schedule.getPointer(), CGF.Int64Ty,
4706                                       /*isSigned=*/false)
4707           : llvm::ConstantInt::get(CGF.Int64Ty, /*V=*/0),
4708       Result.TaskDupFn ? CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
4709                              Result.TaskDupFn, CGF.VoidPtrTy)
4710                        : llvm::ConstantPointerNull::get(CGF.VoidPtrTy)};
4711   CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
4712                           CGM.getModule(), OMPRTL___kmpc_taskloop),
4713                       TaskArgs);
4714 }
4715 
4716 /// Emit reduction operation for each element of array (required for
4717 /// array sections) LHS op = RHS.
4718 /// \param Type Type of array.
4719 /// \param LHSVar Variable on the left side of the reduction operation
4720 /// (references element of array in original variable).
4721 /// \param RHSVar Variable on the right side of the reduction operation
4722 /// (references element of array in original variable).
4723 /// \param RedOpGen Generator of reduction operation with use of LHSVar and
4724 /// RHSVar.
EmitOMPAggregateReduction(CodeGenFunction & CGF,QualType Type,const VarDecl * LHSVar,const VarDecl * RHSVar,const llvm::function_ref<void (CodeGenFunction & CGF,const Expr *,const Expr *,const Expr *)> & RedOpGen,const Expr * XExpr=nullptr,const Expr * EExpr=nullptr,const Expr * UpExpr=nullptr)4725 static void EmitOMPAggregateReduction(
4726     CodeGenFunction &CGF, QualType Type, const VarDecl *LHSVar,
4727     const VarDecl *RHSVar,
4728     const llvm::function_ref<void(CodeGenFunction &CGF, const Expr *,
4729                                   const Expr *, const Expr *)> &RedOpGen,
4730     const Expr *XExpr = nullptr, const Expr *EExpr = nullptr,
4731     const Expr *UpExpr = nullptr) {
4732   // Perform element-by-element initialization.
4733   QualType ElementTy;
4734   Address LHSAddr = CGF.GetAddrOfLocalVar(LHSVar);
4735   Address RHSAddr = CGF.GetAddrOfLocalVar(RHSVar);
4736 
4737   // Drill down to the base element type on both arrays.
4738   const ArrayType *ArrayTy = Type->getAsArrayTypeUnsafe();
4739   llvm::Value *NumElements = CGF.emitArrayLength(ArrayTy, ElementTy, LHSAddr);
4740 
4741   llvm::Value *RHSBegin = RHSAddr.emitRawPointer(CGF);
4742   llvm::Value *LHSBegin = LHSAddr.emitRawPointer(CGF);
4743   // Cast from pointer to array type to pointer to single element.
4744   llvm::Value *LHSEnd =
4745       CGF.Builder.CreateGEP(LHSAddr.getElementType(), LHSBegin, NumElements);
4746   // The basic structure here is a while-do loop.
4747   llvm::BasicBlock *BodyBB = CGF.createBasicBlock("omp.arraycpy.body");
4748   llvm::BasicBlock *DoneBB = CGF.createBasicBlock("omp.arraycpy.done");
4749   llvm::Value *IsEmpty =
4750       CGF.Builder.CreateICmpEQ(LHSBegin, LHSEnd, "omp.arraycpy.isempty");
4751   CGF.Builder.CreateCondBr(IsEmpty, DoneBB, BodyBB);
4752 
4753   // Enter the loop body, making that address the current address.
4754   llvm::BasicBlock *EntryBB = CGF.Builder.GetInsertBlock();
4755   CGF.EmitBlock(BodyBB);
4756 
4757   CharUnits ElementSize = CGF.getContext().getTypeSizeInChars(ElementTy);
4758 
4759   llvm::PHINode *RHSElementPHI = CGF.Builder.CreatePHI(
4760       RHSBegin->getType(), 2, "omp.arraycpy.srcElementPast");
4761   RHSElementPHI->addIncoming(RHSBegin, EntryBB);
4762   Address RHSElementCurrent(
4763       RHSElementPHI, RHSAddr.getElementType(),
4764       RHSAddr.getAlignment().alignmentOfArrayElement(ElementSize));
4765 
4766   llvm::PHINode *LHSElementPHI = CGF.Builder.CreatePHI(
4767       LHSBegin->getType(), 2, "omp.arraycpy.destElementPast");
4768   LHSElementPHI->addIncoming(LHSBegin, EntryBB);
4769   Address LHSElementCurrent(
4770       LHSElementPHI, LHSAddr.getElementType(),
4771       LHSAddr.getAlignment().alignmentOfArrayElement(ElementSize));
4772 
4773   // Emit copy.
4774   CodeGenFunction::OMPPrivateScope Scope(CGF);
4775   Scope.addPrivate(LHSVar, LHSElementCurrent);
4776   Scope.addPrivate(RHSVar, RHSElementCurrent);
4777   Scope.Privatize();
4778   RedOpGen(CGF, XExpr, EExpr, UpExpr);
4779   Scope.ForceCleanup();
4780 
4781   // Shift the address forward by one element.
4782   llvm::Value *LHSElementNext = CGF.Builder.CreateConstGEP1_32(
4783       LHSAddr.getElementType(), LHSElementPHI, /*Idx0=*/1,
4784       "omp.arraycpy.dest.element");
4785   llvm::Value *RHSElementNext = CGF.Builder.CreateConstGEP1_32(
4786       RHSAddr.getElementType(), RHSElementPHI, /*Idx0=*/1,
4787       "omp.arraycpy.src.element");
4788   // Check whether we've reached the end.
4789   llvm::Value *Done =
4790       CGF.Builder.CreateICmpEQ(LHSElementNext, LHSEnd, "omp.arraycpy.done");
4791   CGF.Builder.CreateCondBr(Done, DoneBB, BodyBB);
4792   LHSElementPHI->addIncoming(LHSElementNext, CGF.Builder.GetInsertBlock());
4793   RHSElementPHI->addIncoming(RHSElementNext, CGF.Builder.GetInsertBlock());
4794 
4795   // Done.
4796   CGF.EmitBlock(DoneBB, /*IsFinished=*/true);
4797 }
4798 
4799 /// Emit reduction combiner. If the combiner is a simple expression emit it as
4800 /// is, otherwise consider it as combiner of UDR decl and emit it as a call of
4801 /// UDR combiner function.
emitReductionCombiner(CodeGenFunction & CGF,const Expr * ReductionOp)4802 static void emitReductionCombiner(CodeGenFunction &CGF,
4803                                   const Expr *ReductionOp) {
4804   if (const auto *CE = dyn_cast<CallExpr>(ReductionOp))
4805     if (const auto *OVE = dyn_cast<OpaqueValueExpr>(CE->getCallee()))
4806       if (const auto *DRE =
4807               dyn_cast<DeclRefExpr>(OVE->getSourceExpr()->IgnoreImpCasts()))
4808         if (const auto *DRD =
4809                 dyn_cast<OMPDeclareReductionDecl>(DRE->getDecl())) {
4810           std::pair<llvm::Function *, llvm::Function *> Reduction =
4811               CGF.CGM.getOpenMPRuntime().getUserDefinedReduction(DRD);
4812           RValue Func = RValue::get(Reduction.first);
4813           CodeGenFunction::OpaqueValueMapping Map(CGF, OVE, Func);
4814           CGF.EmitIgnoredExpr(ReductionOp);
4815           return;
4816         }
4817   CGF.EmitIgnoredExpr(ReductionOp);
4818 }
4819 
emitReductionFunction(StringRef ReducerName,SourceLocation Loc,llvm::Type * ArgsElemType,ArrayRef<const Expr * > Privates,ArrayRef<const Expr * > LHSExprs,ArrayRef<const Expr * > RHSExprs,ArrayRef<const Expr * > ReductionOps)4820 llvm::Function *CGOpenMPRuntime::emitReductionFunction(
4821     StringRef ReducerName, SourceLocation Loc, llvm::Type *ArgsElemType,
4822     ArrayRef<const Expr *> Privates, ArrayRef<const Expr *> LHSExprs,
4823     ArrayRef<const Expr *> RHSExprs, ArrayRef<const Expr *> ReductionOps) {
4824   ASTContext &C = CGM.getContext();
4825 
4826   // void reduction_func(void *LHSArg, void *RHSArg);
4827   FunctionArgList Args;
4828   ImplicitParamDecl LHSArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, C.VoidPtrTy,
4829                            ImplicitParamKind::Other);
4830   ImplicitParamDecl RHSArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, C.VoidPtrTy,
4831                            ImplicitParamKind::Other);
4832   Args.push_back(&LHSArg);
4833   Args.push_back(&RHSArg);
4834   const auto &CGFI =
4835       CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args);
4836   std::string Name = getReductionFuncName(ReducerName);
4837   auto *Fn = llvm::Function::Create(CGM.getTypes().GetFunctionType(CGFI),
4838                                     llvm::GlobalValue::InternalLinkage, Name,
4839                                     &CGM.getModule());
4840   CGM.SetInternalFunctionAttributes(GlobalDecl(), Fn, CGFI);
4841   Fn->setDoesNotRecurse();
4842   CodeGenFunction CGF(CGM);
4843   CGF.StartFunction(GlobalDecl(), C.VoidTy, Fn, CGFI, Args, Loc, Loc);
4844 
4845   // Dst = (void*[n])(LHSArg);
4846   // Src = (void*[n])(RHSArg);
4847   Address LHS(CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
4848                   CGF.Builder.CreateLoad(CGF.GetAddrOfLocalVar(&LHSArg)),
4849                   ArgsElemType->getPointerTo()),
4850               ArgsElemType, CGF.getPointerAlign());
4851   Address RHS(CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
4852                   CGF.Builder.CreateLoad(CGF.GetAddrOfLocalVar(&RHSArg)),
4853                   ArgsElemType->getPointerTo()),
4854               ArgsElemType, CGF.getPointerAlign());
4855 
4856   //  ...
4857   //  *(Type<i>*)lhs[i] = RedOp<i>(*(Type<i>*)lhs[i], *(Type<i>*)rhs[i]);
4858   //  ...
4859   CodeGenFunction::OMPPrivateScope Scope(CGF);
4860   const auto *IPriv = Privates.begin();
4861   unsigned Idx = 0;
4862   for (unsigned I = 0, E = ReductionOps.size(); I < E; ++I, ++IPriv, ++Idx) {
4863     const auto *RHSVar =
4864         cast<VarDecl>(cast<DeclRefExpr>(RHSExprs[I])->getDecl());
4865     Scope.addPrivate(RHSVar, emitAddrOfVarFromArray(CGF, RHS, Idx, RHSVar));
4866     const auto *LHSVar =
4867         cast<VarDecl>(cast<DeclRefExpr>(LHSExprs[I])->getDecl());
4868     Scope.addPrivate(LHSVar, emitAddrOfVarFromArray(CGF, LHS, Idx, LHSVar));
4869     QualType PrivTy = (*IPriv)->getType();
4870     if (PrivTy->isVariablyModifiedType()) {
4871       // Get array size and emit VLA type.
4872       ++Idx;
4873       Address Elem = CGF.Builder.CreateConstArrayGEP(LHS, Idx);
4874       llvm::Value *Ptr = CGF.Builder.CreateLoad(Elem);
4875       const VariableArrayType *VLA =
4876           CGF.getContext().getAsVariableArrayType(PrivTy);
4877       const auto *OVE = cast<OpaqueValueExpr>(VLA->getSizeExpr());
4878       CodeGenFunction::OpaqueValueMapping OpaqueMap(
4879           CGF, OVE, RValue::get(CGF.Builder.CreatePtrToInt(Ptr, CGF.SizeTy)));
4880       CGF.EmitVariablyModifiedType(PrivTy);
4881     }
4882   }
4883   Scope.Privatize();
4884   IPriv = Privates.begin();
4885   const auto *ILHS = LHSExprs.begin();
4886   const auto *IRHS = RHSExprs.begin();
4887   for (const Expr *E : ReductionOps) {
4888     if ((*IPriv)->getType()->isArrayType()) {
4889       // Emit reduction for array section.
4890       const auto *LHSVar = cast<VarDecl>(cast<DeclRefExpr>(*ILHS)->getDecl());
4891       const auto *RHSVar = cast<VarDecl>(cast<DeclRefExpr>(*IRHS)->getDecl());
4892       EmitOMPAggregateReduction(
4893           CGF, (*IPriv)->getType(), LHSVar, RHSVar,
4894           [=](CodeGenFunction &CGF, const Expr *, const Expr *, const Expr *) {
4895             emitReductionCombiner(CGF, E);
4896           });
4897     } else {
4898       // Emit reduction for array subscript or single variable.
4899       emitReductionCombiner(CGF, E);
4900     }
4901     ++IPriv;
4902     ++ILHS;
4903     ++IRHS;
4904   }
4905   Scope.ForceCleanup();
4906   CGF.FinishFunction();
4907   return Fn;
4908 }
4909 
emitSingleReductionCombiner(CodeGenFunction & CGF,const Expr * ReductionOp,const Expr * PrivateRef,const DeclRefExpr * LHS,const DeclRefExpr * RHS)4910 void CGOpenMPRuntime::emitSingleReductionCombiner(CodeGenFunction &CGF,
4911                                                   const Expr *ReductionOp,
4912                                                   const Expr *PrivateRef,
4913                                                   const DeclRefExpr *LHS,
4914                                                   const DeclRefExpr *RHS) {
4915   if (PrivateRef->getType()->isArrayType()) {
4916     // Emit reduction for array section.
4917     const auto *LHSVar = cast<VarDecl>(LHS->getDecl());
4918     const auto *RHSVar = cast<VarDecl>(RHS->getDecl());
4919     EmitOMPAggregateReduction(
4920         CGF, PrivateRef->getType(), LHSVar, RHSVar,
4921         [=](CodeGenFunction &CGF, const Expr *, const Expr *, const Expr *) {
4922           emitReductionCombiner(CGF, ReductionOp);
4923         });
4924   } else {
4925     // Emit reduction for array subscript or single variable.
4926     emitReductionCombiner(CGF, ReductionOp);
4927   }
4928 }
4929 
emitReduction(CodeGenFunction & CGF,SourceLocation Loc,ArrayRef<const Expr * > Privates,ArrayRef<const Expr * > LHSExprs,ArrayRef<const Expr * > RHSExprs,ArrayRef<const Expr * > ReductionOps,ReductionOptionsTy Options)4930 void CGOpenMPRuntime::emitReduction(CodeGenFunction &CGF, SourceLocation Loc,
4931                                     ArrayRef<const Expr *> Privates,
4932                                     ArrayRef<const Expr *> LHSExprs,
4933                                     ArrayRef<const Expr *> RHSExprs,
4934                                     ArrayRef<const Expr *> ReductionOps,
4935                                     ReductionOptionsTy Options) {
4936   if (!CGF.HaveInsertPoint())
4937     return;
4938 
4939   bool WithNowait = Options.WithNowait;
4940   bool SimpleReduction = Options.SimpleReduction;
4941 
4942   // Next code should be emitted for reduction:
4943   //
4944   // static kmp_critical_name lock = { 0 };
4945   //
4946   // void reduce_func(void *lhs[<n>], void *rhs[<n>]) {
4947   //  *(Type0*)lhs[0] = ReductionOperation0(*(Type0*)lhs[0], *(Type0*)rhs[0]);
4948   //  ...
4949   //  *(Type<n>-1*)lhs[<n>-1] = ReductionOperation<n>-1(*(Type<n>-1*)lhs[<n>-1],
4950   //  *(Type<n>-1*)rhs[<n>-1]);
4951   // }
4952   //
4953   // ...
4954   // void *RedList[<n>] = {&<RHSExprs>[0], ..., &<RHSExprs>[<n>-1]};
4955   // switch (__kmpc_reduce{_nowait}(<loc>, <gtid>, <n>, sizeof(RedList),
4956   // RedList, reduce_func, &<lock>)) {
4957   // case 1:
4958   //  ...
4959   //  <LHSExprs>[i] = RedOp<i>(*<LHSExprs>[i], *<RHSExprs>[i]);
4960   //  ...
4961   // __kmpc_end_reduce{_nowait}(<loc>, <gtid>, &<lock>);
4962   // break;
4963   // case 2:
4964   //  ...
4965   //  Atomic(<LHSExprs>[i] = RedOp<i>(*<LHSExprs>[i], *<RHSExprs>[i]));
4966   //  ...
4967   // [__kmpc_end_reduce(<loc>, <gtid>, &<lock>);]
4968   // break;
4969   // default:;
4970   // }
4971   //
4972   // if SimpleReduction is true, only the next code is generated:
4973   //  ...
4974   //  <LHSExprs>[i] = RedOp<i>(*<LHSExprs>[i], *<RHSExprs>[i]);
4975   //  ...
4976 
4977   ASTContext &C = CGM.getContext();
4978 
4979   if (SimpleReduction) {
4980     CodeGenFunction::RunCleanupsScope Scope(CGF);
4981     const auto *IPriv = Privates.begin();
4982     const auto *ILHS = LHSExprs.begin();
4983     const auto *IRHS = RHSExprs.begin();
4984     for (const Expr *E : ReductionOps) {
4985       emitSingleReductionCombiner(CGF, E, *IPriv, cast<DeclRefExpr>(*ILHS),
4986                                   cast<DeclRefExpr>(*IRHS));
4987       ++IPriv;
4988       ++ILHS;
4989       ++IRHS;
4990     }
4991     return;
4992   }
4993 
4994   // 1. Build a list of reduction variables.
4995   // void *RedList[<n>] = {<ReductionVars>[0], ..., <ReductionVars>[<n>-1]};
4996   auto Size = RHSExprs.size();
4997   for (const Expr *E : Privates) {
4998     if (E->getType()->isVariablyModifiedType())
4999       // Reserve place for array size.
5000       ++Size;
5001   }
5002   llvm::APInt ArraySize(/*unsigned int numBits=*/32, Size);
5003   QualType ReductionArrayTy = C.getConstantArrayType(
5004       C.VoidPtrTy, ArraySize, nullptr, ArraySizeModifier::Normal,
5005       /*IndexTypeQuals=*/0);
5006   RawAddress ReductionList =
5007       CGF.CreateMemTemp(ReductionArrayTy, ".omp.reduction.red_list");
5008   const auto *IPriv = Privates.begin();
5009   unsigned Idx = 0;
5010   for (unsigned I = 0, E = RHSExprs.size(); I < E; ++I, ++IPriv, ++Idx) {
5011     Address Elem = CGF.Builder.CreateConstArrayGEP(ReductionList, Idx);
5012     CGF.Builder.CreateStore(
5013         CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
5014             CGF.EmitLValue(RHSExprs[I]).getPointer(CGF), CGF.VoidPtrTy),
5015         Elem);
5016     if ((*IPriv)->getType()->isVariablyModifiedType()) {
5017       // Store array size.
5018       ++Idx;
5019       Elem = CGF.Builder.CreateConstArrayGEP(ReductionList, Idx);
5020       llvm::Value *Size = CGF.Builder.CreateIntCast(
5021           CGF.getVLASize(
5022                  CGF.getContext().getAsVariableArrayType((*IPriv)->getType()))
5023               .NumElts,
5024           CGF.SizeTy, /*isSigned=*/false);
5025       CGF.Builder.CreateStore(CGF.Builder.CreateIntToPtr(Size, CGF.VoidPtrTy),
5026                               Elem);
5027     }
5028   }
5029 
5030   // 2. Emit reduce_func().
5031   llvm::Function *ReductionFn = emitReductionFunction(
5032       CGF.CurFn->getName(), Loc, CGF.ConvertTypeForMem(ReductionArrayTy),
5033       Privates, LHSExprs, RHSExprs, ReductionOps);
5034 
5035   // 3. Create static kmp_critical_name lock = { 0 };
5036   std::string Name = getName({"reduction"});
5037   llvm::Value *Lock = getCriticalRegionLock(Name);
5038 
5039   // 4. Build res = __kmpc_reduce{_nowait}(<loc>, <gtid>, <n>, sizeof(RedList),
5040   // RedList, reduce_func, &<lock>);
5041   llvm::Value *IdentTLoc = emitUpdateLocation(CGF, Loc, OMP_ATOMIC_REDUCE);
5042   llvm::Value *ThreadId = getThreadID(CGF, Loc);
5043   llvm::Value *ReductionArrayTySize = CGF.getTypeSize(ReductionArrayTy);
5044   llvm::Value *RL = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
5045       ReductionList.getPointer(), CGF.VoidPtrTy);
5046   llvm::Value *Args[] = {
5047       IdentTLoc,                             // ident_t *<loc>
5048       ThreadId,                              // i32 <gtid>
5049       CGF.Builder.getInt32(RHSExprs.size()), // i32 <n>
5050       ReductionArrayTySize,                  // size_type sizeof(RedList)
5051       RL,                                    // void *RedList
5052       ReductionFn, // void (*) (void *, void *) <reduce_func>
5053       Lock         // kmp_critical_name *&<lock>
5054   };
5055   llvm::Value *Res = CGF.EmitRuntimeCall(
5056       OMPBuilder.getOrCreateRuntimeFunction(
5057           CGM.getModule(),
5058           WithNowait ? OMPRTL___kmpc_reduce_nowait : OMPRTL___kmpc_reduce),
5059       Args);
5060 
5061   // 5. Build switch(res)
5062   llvm::BasicBlock *DefaultBB = CGF.createBasicBlock(".omp.reduction.default");
5063   llvm::SwitchInst *SwInst =
5064       CGF.Builder.CreateSwitch(Res, DefaultBB, /*NumCases=*/2);
5065 
5066   // 6. Build case 1:
5067   //  ...
5068   //  <LHSExprs>[i] = RedOp<i>(*<LHSExprs>[i], *<RHSExprs>[i]);
5069   //  ...
5070   // __kmpc_end_reduce{_nowait}(<loc>, <gtid>, &<lock>);
5071   // break;
5072   llvm::BasicBlock *Case1BB = CGF.createBasicBlock(".omp.reduction.case1");
5073   SwInst->addCase(CGF.Builder.getInt32(1), Case1BB);
5074   CGF.EmitBlock(Case1BB);
5075 
5076   // Add emission of __kmpc_end_reduce{_nowait}(<loc>, <gtid>, &<lock>);
5077   llvm::Value *EndArgs[] = {
5078       IdentTLoc, // ident_t *<loc>
5079       ThreadId,  // i32 <gtid>
5080       Lock       // kmp_critical_name *&<lock>
5081   };
5082   auto &&CodeGen = [Privates, LHSExprs, RHSExprs, ReductionOps](
5083                        CodeGenFunction &CGF, PrePostActionTy &Action) {
5084     CGOpenMPRuntime &RT = CGF.CGM.getOpenMPRuntime();
5085     const auto *IPriv = Privates.begin();
5086     const auto *ILHS = LHSExprs.begin();
5087     const auto *IRHS = RHSExprs.begin();
5088     for (const Expr *E : ReductionOps) {
5089       RT.emitSingleReductionCombiner(CGF, E, *IPriv, cast<DeclRefExpr>(*ILHS),
5090                                      cast<DeclRefExpr>(*IRHS));
5091       ++IPriv;
5092       ++ILHS;
5093       ++IRHS;
5094     }
5095   };
5096   RegionCodeGenTy RCG(CodeGen);
5097   CommonActionTy Action(
5098       nullptr, std::nullopt,
5099       OMPBuilder.getOrCreateRuntimeFunction(
5100           CGM.getModule(), WithNowait ? OMPRTL___kmpc_end_reduce_nowait
5101                                       : OMPRTL___kmpc_end_reduce),
5102       EndArgs);
5103   RCG.setAction(Action);
5104   RCG(CGF);
5105 
5106   CGF.EmitBranch(DefaultBB);
5107 
5108   // 7. Build case 2:
5109   //  ...
5110   //  Atomic(<LHSExprs>[i] = RedOp<i>(*<LHSExprs>[i], *<RHSExprs>[i]));
5111   //  ...
5112   // break;
5113   llvm::BasicBlock *Case2BB = CGF.createBasicBlock(".omp.reduction.case2");
5114   SwInst->addCase(CGF.Builder.getInt32(2), Case2BB);
5115   CGF.EmitBlock(Case2BB);
5116 
5117   auto &&AtomicCodeGen = [Loc, Privates, LHSExprs, RHSExprs, ReductionOps](
5118                              CodeGenFunction &CGF, PrePostActionTy &Action) {
5119     const auto *ILHS = LHSExprs.begin();
5120     const auto *IRHS = RHSExprs.begin();
5121     const auto *IPriv = Privates.begin();
5122     for (const Expr *E : ReductionOps) {
5123       const Expr *XExpr = nullptr;
5124       const Expr *EExpr = nullptr;
5125       const Expr *UpExpr = nullptr;
5126       BinaryOperatorKind BO = BO_Comma;
5127       if (const auto *BO = dyn_cast<BinaryOperator>(E)) {
5128         if (BO->getOpcode() == BO_Assign) {
5129           XExpr = BO->getLHS();
5130           UpExpr = BO->getRHS();
5131         }
5132       }
5133       // Try to emit update expression as a simple atomic.
5134       const Expr *RHSExpr = UpExpr;
5135       if (RHSExpr) {
5136         // Analyze RHS part of the whole expression.
5137         if (const auto *ACO = dyn_cast<AbstractConditionalOperator>(
5138                 RHSExpr->IgnoreParenImpCasts())) {
5139           // If this is a conditional operator, analyze its condition for
5140           // min/max reduction operator.
5141           RHSExpr = ACO->getCond();
5142         }
5143         if (const auto *BORHS =
5144                 dyn_cast<BinaryOperator>(RHSExpr->IgnoreParenImpCasts())) {
5145           EExpr = BORHS->getRHS();
5146           BO = BORHS->getOpcode();
5147         }
5148       }
5149       if (XExpr) {
5150         const auto *VD = cast<VarDecl>(cast<DeclRefExpr>(*ILHS)->getDecl());
5151         auto &&AtomicRedGen = [BO, VD,
5152                                Loc](CodeGenFunction &CGF, const Expr *XExpr,
5153                                     const Expr *EExpr, const Expr *UpExpr) {
5154           LValue X = CGF.EmitLValue(XExpr);
5155           RValue E;
5156           if (EExpr)
5157             E = CGF.EmitAnyExpr(EExpr);
5158           CGF.EmitOMPAtomicSimpleUpdateExpr(
5159               X, E, BO, /*IsXLHSInRHSPart=*/true,
5160               llvm::AtomicOrdering::Monotonic, Loc,
5161               [&CGF, UpExpr, VD, Loc](RValue XRValue) {
5162                 CodeGenFunction::OMPPrivateScope PrivateScope(CGF);
5163                 Address LHSTemp = CGF.CreateMemTemp(VD->getType());
5164                 CGF.emitOMPSimpleStore(
5165                     CGF.MakeAddrLValue(LHSTemp, VD->getType()), XRValue,
5166                     VD->getType().getNonReferenceType(), Loc);
5167                 PrivateScope.addPrivate(VD, LHSTemp);
5168                 (void)PrivateScope.Privatize();
5169                 return CGF.EmitAnyExpr(UpExpr);
5170               });
5171         };
5172         if ((*IPriv)->getType()->isArrayType()) {
5173           // Emit atomic reduction for array section.
5174           const auto *RHSVar =
5175               cast<VarDecl>(cast<DeclRefExpr>(*IRHS)->getDecl());
5176           EmitOMPAggregateReduction(CGF, (*IPriv)->getType(), VD, RHSVar,
5177                                     AtomicRedGen, XExpr, EExpr, UpExpr);
5178         } else {
5179           // Emit atomic reduction for array subscript or single variable.
5180           AtomicRedGen(CGF, XExpr, EExpr, UpExpr);
5181         }
5182       } else {
5183         // Emit as a critical region.
5184         auto &&CritRedGen = [E, Loc](CodeGenFunction &CGF, const Expr *,
5185                                            const Expr *, const Expr *) {
5186           CGOpenMPRuntime &RT = CGF.CGM.getOpenMPRuntime();
5187           std::string Name = RT.getName({"atomic_reduction"});
5188           RT.emitCriticalRegion(
5189               CGF, Name,
5190               [=](CodeGenFunction &CGF, PrePostActionTy &Action) {
5191                 Action.Enter(CGF);
5192                 emitReductionCombiner(CGF, E);
5193               },
5194               Loc);
5195         };
5196         if ((*IPriv)->getType()->isArrayType()) {
5197           const auto *LHSVar =
5198               cast<VarDecl>(cast<DeclRefExpr>(*ILHS)->getDecl());
5199           const auto *RHSVar =
5200               cast<VarDecl>(cast<DeclRefExpr>(*IRHS)->getDecl());
5201           EmitOMPAggregateReduction(CGF, (*IPriv)->getType(), LHSVar, RHSVar,
5202                                     CritRedGen);
5203         } else {
5204           CritRedGen(CGF, nullptr, nullptr, nullptr);
5205         }
5206       }
5207       ++ILHS;
5208       ++IRHS;
5209       ++IPriv;
5210     }
5211   };
5212   RegionCodeGenTy AtomicRCG(AtomicCodeGen);
5213   if (!WithNowait) {
5214     // Add emission of __kmpc_end_reduce(<loc>, <gtid>, &<lock>);
5215     llvm::Value *EndArgs[] = {
5216         IdentTLoc, // ident_t *<loc>
5217         ThreadId,  // i32 <gtid>
5218         Lock       // kmp_critical_name *&<lock>
5219     };
5220     CommonActionTy Action(nullptr, std::nullopt,
5221                           OMPBuilder.getOrCreateRuntimeFunction(
5222                               CGM.getModule(), OMPRTL___kmpc_end_reduce),
5223                           EndArgs);
5224     AtomicRCG.setAction(Action);
5225     AtomicRCG(CGF);
5226   } else {
5227     AtomicRCG(CGF);
5228   }
5229 
5230   CGF.EmitBranch(DefaultBB);
5231   CGF.EmitBlock(DefaultBB, /*IsFinished=*/true);
5232 }
5233 
5234 /// Generates unique name for artificial threadprivate variables.
5235 /// Format is: <Prefix> "." <Decl_mangled_name> "_" "<Decl_start_loc_raw_enc>"
generateUniqueName(CodeGenModule & CGM,StringRef Prefix,const Expr * Ref)5236 static std::string generateUniqueName(CodeGenModule &CGM, StringRef Prefix,
5237                                       const Expr *Ref) {
5238   SmallString<256> Buffer;
5239   llvm::raw_svector_ostream Out(Buffer);
5240   const clang::DeclRefExpr *DE;
5241   const VarDecl *D = ::getBaseDecl(Ref, DE);
5242   if (!D)
5243     D = cast<VarDecl>(cast<DeclRefExpr>(Ref)->getDecl());
5244   D = D->getCanonicalDecl();
5245   std::string Name = CGM.getOpenMPRuntime().getName(
5246       {D->isLocalVarDeclOrParm() ? D->getName() : CGM.getMangledName(D)});
5247   Out << Prefix << Name << "_"
5248       << D->getCanonicalDecl()->getBeginLoc().getRawEncoding();
5249   return std::string(Out.str());
5250 }
5251 
5252 /// Emits reduction initializer function:
5253 /// \code
5254 /// void @.red_init(void* %arg, void* %orig) {
5255 /// %0 = bitcast void* %arg to <type>*
5256 /// store <type> <init>, <type>* %0
5257 /// ret void
5258 /// }
5259 /// \endcode
emitReduceInitFunction(CodeGenModule & CGM,SourceLocation Loc,ReductionCodeGen & RCG,unsigned N)5260 static llvm::Value *emitReduceInitFunction(CodeGenModule &CGM,
5261                                            SourceLocation Loc,
5262                                            ReductionCodeGen &RCG, unsigned N) {
5263   ASTContext &C = CGM.getContext();
5264   QualType VoidPtrTy = C.VoidPtrTy;
5265   VoidPtrTy.addRestrict();
5266   FunctionArgList Args;
5267   ImplicitParamDecl Param(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, VoidPtrTy,
5268                           ImplicitParamKind::Other);
5269   ImplicitParamDecl ParamOrig(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, VoidPtrTy,
5270                               ImplicitParamKind::Other);
5271   Args.emplace_back(&Param);
5272   Args.emplace_back(&ParamOrig);
5273   const auto &FnInfo =
5274       CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args);
5275   llvm::FunctionType *FnTy = CGM.getTypes().GetFunctionType(FnInfo);
5276   std::string Name = CGM.getOpenMPRuntime().getName({"red_init", ""});
5277   auto *Fn = llvm::Function::Create(FnTy, llvm::GlobalValue::InternalLinkage,
5278                                     Name, &CGM.getModule());
5279   CGM.SetInternalFunctionAttributes(GlobalDecl(), Fn, FnInfo);
5280   Fn->setDoesNotRecurse();
5281   CodeGenFunction CGF(CGM);
5282   CGF.StartFunction(GlobalDecl(), C.VoidTy, Fn, FnInfo, Args, Loc, Loc);
5283   QualType PrivateType = RCG.getPrivateType(N);
5284   Address PrivateAddr = CGF.EmitLoadOfPointer(
5285       CGF.GetAddrOfLocalVar(&Param).withElementType(
5286           CGF.ConvertTypeForMem(PrivateType)->getPointerTo()),
5287       C.getPointerType(PrivateType)->castAs<PointerType>());
5288   llvm::Value *Size = nullptr;
5289   // If the size of the reduction item is non-constant, load it from global
5290   // threadprivate variable.
5291   if (RCG.getSizes(N).second) {
5292     Address SizeAddr = CGM.getOpenMPRuntime().getAddrOfArtificialThreadPrivate(
5293         CGF, CGM.getContext().getSizeType(),
5294         generateUniqueName(CGM, "reduction_size", RCG.getRefExpr(N)));
5295     Size = CGF.EmitLoadOfScalar(SizeAddr, /*Volatile=*/false,
5296                                 CGM.getContext().getSizeType(), Loc);
5297   }
5298   RCG.emitAggregateType(CGF, N, Size);
5299   Address OrigAddr = Address::invalid();
5300   // If initializer uses initializer from declare reduction construct, emit a
5301   // pointer to the address of the original reduction item (reuired by reduction
5302   // initializer)
5303   if (RCG.usesReductionInitializer(N)) {
5304     Address SharedAddr = CGF.GetAddrOfLocalVar(&ParamOrig);
5305     OrigAddr = CGF.EmitLoadOfPointer(
5306         SharedAddr,
5307         CGM.getContext().VoidPtrTy.castAs<PointerType>()->getTypePtr());
5308   }
5309   // Emit the initializer:
5310   // %0 = bitcast void* %arg to <type>*
5311   // store <type> <init>, <type>* %0
5312   RCG.emitInitialization(CGF, N, PrivateAddr, OrigAddr,
5313                          [](CodeGenFunction &) { return false; });
5314   CGF.FinishFunction();
5315   return Fn;
5316 }
5317 
5318 /// Emits reduction combiner function:
5319 /// \code
5320 /// void @.red_comb(void* %arg0, void* %arg1) {
5321 /// %lhs = bitcast void* %arg0 to <type>*
5322 /// %rhs = bitcast void* %arg1 to <type>*
5323 /// %2 = <ReductionOp>(<type>* %lhs, <type>* %rhs)
5324 /// store <type> %2, <type>* %lhs
5325 /// ret void
5326 /// }
5327 /// \endcode
emitReduceCombFunction(CodeGenModule & CGM,SourceLocation Loc,ReductionCodeGen & RCG,unsigned N,const Expr * ReductionOp,const Expr * LHS,const Expr * RHS,const Expr * PrivateRef)5328 static llvm::Value *emitReduceCombFunction(CodeGenModule &CGM,
5329                                            SourceLocation Loc,
5330                                            ReductionCodeGen &RCG, unsigned N,
5331                                            const Expr *ReductionOp,
5332                                            const Expr *LHS, const Expr *RHS,
5333                                            const Expr *PrivateRef) {
5334   ASTContext &C = CGM.getContext();
5335   const auto *LHSVD = cast<VarDecl>(cast<DeclRefExpr>(LHS)->getDecl());
5336   const auto *RHSVD = cast<VarDecl>(cast<DeclRefExpr>(RHS)->getDecl());
5337   FunctionArgList Args;
5338   ImplicitParamDecl ParamInOut(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr,
5339                                C.VoidPtrTy, ImplicitParamKind::Other);
5340   ImplicitParamDecl ParamIn(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, C.VoidPtrTy,
5341                             ImplicitParamKind::Other);
5342   Args.emplace_back(&ParamInOut);
5343   Args.emplace_back(&ParamIn);
5344   const auto &FnInfo =
5345       CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args);
5346   llvm::FunctionType *FnTy = CGM.getTypes().GetFunctionType(FnInfo);
5347   std::string Name = CGM.getOpenMPRuntime().getName({"red_comb", ""});
5348   auto *Fn = llvm::Function::Create(FnTy, llvm::GlobalValue::InternalLinkage,
5349                                     Name, &CGM.getModule());
5350   CGM.SetInternalFunctionAttributes(GlobalDecl(), Fn, FnInfo);
5351   Fn->setDoesNotRecurse();
5352   CodeGenFunction CGF(CGM);
5353   CGF.StartFunction(GlobalDecl(), C.VoidTy, Fn, FnInfo, Args, Loc, Loc);
5354   llvm::Value *Size = nullptr;
5355   // If the size of the reduction item is non-constant, load it from global
5356   // threadprivate variable.
5357   if (RCG.getSizes(N).second) {
5358     Address SizeAddr = CGM.getOpenMPRuntime().getAddrOfArtificialThreadPrivate(
5359         CGF, CGM.getContext().getSizeType(),
5360         generateUniqueName(CGM, "reduction_size", RCG.getRefExpr(N)));
5361     Size = CGF.EmitLoadOfScalar(SizeAddr, /*Volatile=*/false,
5362                                 CGM.getContext().getSizeType(), Loc);
5363   }
5364   RCG.emitAggregateType(CGF, N, Size);
5365   // Remap lhs and rhs variables to the addresses of the function arguments.
5366   // %lhs = bitcast void* %arg0 to <type>*
5367   // %rhs = bitcast void* %arg1 to <type>*
5368   CodeGenFunction::OMPPrivateScope PrivateScope(CGF);
5369   PrivateScope.addPrivate(
5370       LHSVD,
5371       // Pull out the pointer to the variable.
5372       CGF.EmitLoadOfPointer(
5373           CGF.GetAddrOfLocalVar(&ParamInOut)
5374               .withElementType(
5375                   CGF.ConvertTypeForMem(LHSVD->getType())->getPointerTo()),
5376           C.getPointerType(LHSVD->getType())->castAs<PointerType>()));
5377   PrivateScope.addPrivate(
5378       RHSVD,
5379       // Pull out the pointer to the variable.
5380       CGF.EmitLoadOfPointer(
5381           CGF.GetAddrOfLocalVar(&ParamIn).withElementType(
5382               CGF.ConvertTypeForMem(RHSVD->getType())->getPointerTo()),
5383           C.getPointerType(RHSVD->getType())->castAs<PointerType>()));
5384   PrivateScope.Privatize();
5385   // Emit the combiner body:
5386   // %2 = <ReductionOp>(<type> *%lhs, <type> *%rhs)
5387   // store <type> %2, <type>* %lhs
5388   CGM.getOpenMPRuntime().emitSingleReductionCombiner(
5389       CGF, ReductionOp, PrivateRef, cast<DeclRefExpr>(LHS),
5390       cast<DeclRefExpr>(RHS));
5391   CGF.FinishFunction();
5392   return Fn;
5393 }
5394 
5395 /// Emits reduction finalizer function:
5396 /// \code
5397 /// void @.red_fini(void* %arg) {
5398 /// %0 = bitcast void* %arg to <type>*
5399 /// <destroy>(<type>* %0)
5400 /// ret void
5401 /// }
5402 /// \endcode
emitReduceFiniFunction(CodeGenModule & CGM,SourceLocation Loc,ReductionCodeGen & RCG,unsigned N)5403 static llvm::Value *emitReduceFiniFunction(CodeGenModule &CGM,
5404                                            SourceLocation Loc,
5405                                            ReductionCodeGen &RCG, unsigned N) {
5406   if (!RCG.needCleanups(N))
5407     return nullptr;
5408   ASTContext &C = CGM.getContext();
5409   FunctionArgList Args;
5410   ImplicitParamDecl Param(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, C.VoidPtrTy,
5411                           ImplicitParamKind::Other);
5412   Args.emplace_back(&Param);
5413   const auto &FnInfo =
5414       CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args);
5415   llvm::FunctionType *FnTy = CGM.getTypes().GetFunctionType(FnInfo);
5416   std::string Name = CGM.getOpenMPRuntime().getName({"red_fini", ""});
5417   auto *Fn = llvm::Function::Create(FnTy, llvm::GlobalValue::InternalLinkage,
5418                                     Name, &CGM.getModule());
5419   CGM.SetInternalFunctionAttributes(GlobalDecl(), Fn, FnInfo);
5420   Fn->setDoesNotRecurse();
5421   CodeGenFunction CGF(CGM);
5422   CGF.StartFunction(GlobalDecl(), C.VoidTy, Fn, FnInfo, Args, Loc, Loc);
5423   Address PrivateAddr = CGF.EmitLoadOfPointer(
5424       CGF.GetAddrOfLocalVar(&Param), C.VoidPtrTy.castAs<PointerType>());
5425   llvm::Value *Size = nullptr;
5426   // If the size of the reduction item is non-constant, load it from global
5427   // threadprivate variable.
5428   if (RCG.getSizes(N).second) {
5429     Address SizeAddr = CGM.getOpenMPRuntime().getAddrOfArtificialThreadPrivate(
5430         CGF, CGM.getContext().getSizeType(),
5431         generateUniqueName(CGM, "reduction_size", RCG.getRefExpr(N)));
5432     Size = CGF.EmitLoadOfScalar(SizeAddr, /*Volatile=*/false,
5433                                 CGM.getContext().getSizeType(), Loc);
5434   }
5435   RCG.emitAggregateType(CGF, N, Size);
5436   // Emit the finalizer body:
5437   // <destroy>(<type>* %0)
5438   RCG.emitCleanups(CGF, N, PrivateAddr);
5439   CGF.FinishFunction(Loc);
5440   return Fn;
5441 }
5442 
emitTaskReductionInit(CodeGenFunction & CGF,SourceLocation Loc,ArrayRef<const Expr * > LHSExprs,ArrayRef<const Expr * > RHSExprs,const OMPTaskDataTy & Data)5443 llvm::Value *CGOpenMPRuntime::emitTaskReductionInit(
5444     CodeGenFunction &CGF, SourceLocation Loc, ArrayRef<const Expr *> LHSExprs,
5445     ArrayRef<const Expr *> RHSExprs, const OMPTaskDataTy &Data) {
5446   if (!CGF.HaveInsertPoint() || Data.ReductionVars.empty())
5447     return nullptr;
5448 
5449   // Build typedef struct:
5450   // kmp_taskred_input {
5451   //   void *reduce_shar; // shared reduction item
5452   //   void *reduce_orig; // original reduction item used for initialization
5453   //   size_t reduce_size; // size of data item
5454   //   void *reduce_init; // data initialization routine
5455   //   void *reduce_fini; // data finalization routine
5456   //   void *reduce_comb; // data combiner routine
5457   //   kmp_task_red_flags_t flags; // flags for additional info from compiler
5458   // } kmp_taskred_input_t;
5459   ASTContext &C = CGM.getContext();
5460   RecordDecl *RD = C.buildImplicitRecord("kmp_taskred_input_t");
5461   RD->startDefinition();
5462   const FieldDecl *SharedFD = addFieldToRecordDecl(C, RD, C.VoidPtrTy);
5463   const FieldDecl *OrigFD = addFieldToRecordDecl(C, RD, C.VoidPtrTy);
5464   const FieldDecl *SizeFD = addFieldToRecordDecl(C, RD, C.getSizeType());
5465   const FieldDecl *InitFD  = addFieldToRecordDecl(C, RD, C.VoidPtrTy);
5466   const FieldDecl *FiniFD = addFieldToRecordDecl(C, RD, C.VoidPtrTy);
5467   const FieldDecl *CombFD = addFieldToRecordDecl(C, RD, C.VoidPtrTy);
5468   const FieldDecl *FlagsFD = addFieldToRecordDecl(
5469       C, RD, C.getIntTypeForBitwidth(/*DestWidth=*/32, /*Signed=*/false));
5470   RD->completeDefinition();
5471   QualType RDType = C.getRecordType(RD);
5472   unsigned Size = Data.ReductionVars.size();
5473   llvm::APInt ArraySize(/*numBits=*/64, Size);
5474   QualType ArrayRDType =
5475       C.getConstantArrayType(RDType, ArraySize, nullptr,
5476                              ArraySizeModifier::Normal, /*IndexTypeQuals=*/0);
5477   // kmp_task_red_input_t .rd_input.[Size];
5478   RawAddress TaskRedInput = CGF.CreateMemTemp(ArrayRDType, ".rd_input.");
5479   ReductionCodeGen RCG(Data.ReductionVars, Data.ReductionOrigs,
5480                        Data.ReductionCopies, Data.ReductionOps);
5481   for (unsigned Cnt = 0; Cnt < Size; ++Cnt) {
5482     // kmp_task_red_input_t &ElemLVal = .rd_input.[Cnt];
5483     llvm::Value *Idxs[] = {llvm::ConstantInt::get(CGM.SizeTy, /*V=*/0),
5484                            llvm::ConstantInt::get(CGM.SizeTy, Cnt)};
5485     llvm::Value *GEP = CGF.EmitCheckedInBoundsGEP(
5486         TaskRedInput.getElementType(), TaskRedInput.getPointer(), Idxs,
5487         /*SignedIndices=*/false, /*IsSubtraction=*/false, Loc,
5488         ".rd_input.gep.");
5489     LValue ElemLVal = CGF.MakeNaturalAlignRawAddrLValue(GEP, RDType);
5490     // ElemLVal.reduce_shar = &Shareds[Cnt];
5491     LValue SharedLVal = CGF.EmitLValueForField(ElemLVal, SharedFD);
5492     RCG.emitSharedOrigLValue(CGF, Cnt);
5493     llvm::Value *Shared = RCG.getSharedLValue(Cnt).getPointer(CGF);
5494     CGF.EmitStoreOfScalar(Shared, SharedLVal);
5495     // ElemLVal.reduce_orig = &Origs[Cnt];
5496     LValue OrigLVal = CGF.EmitLValueForField(ElemLVal, OrigFD);
5497     llvm::Value *Orig = RCG.getOrigLValue(Cnt).getPointer(CGF);
5498     CGF.EmitStoreOfScalar(Orig, OrigLVal);
5499     RCG.emitAggregateType(CGF, Cnt);
5500     llvm::Value *SizeValInChars;
5501     llvm::Value *SizeVal;
5502     std::tie(SizeValInChars, SizeVal) = RCG.getSizes(Cnt);
5503     // We use delayed creation/initialization for VLAs and array sections. It is
5504     // required because runtime does not provide the way to pass the sizes of
5505     // VLAs/array sections to initializer/combiner/finalizer functions. Instead
5506     // threadprivate global variables are used to store these values and use
5507     // them in the functions.
5508     bool DelayedCreation = !!SizeVal;
5509     SizeValInChars = CGF.Builder.CreateIntCast(SizeValInChars, CGM.SizeTy,
5510                                                /*isSigned=*/false);
5511     LValue SizeLVal = CGF.EmitLValueForField(ElemLVal, SizeFD);
5512     CGF.EmitStoreOfScalar(SizeValInChars, SizeLVal);
5513     // ElemLVal.reduce_init = init;
5514     LValue InitLVal = CGF.EmitLValueForField(ElemLVal, InitFD);
5515     llvm::Value *InitAddr = emitReduceInitFunction(CGM, Loc, RCG, Cnt);
5516     CGF.EmitStoreOfScalar(InitAddr, InitLVal);
5517     // ElemLVal.reduce_fini = fini;
5518     LValue FiniLVal = CGF.EmitLValueForField(ElemLVal, FiniFD);
5519     llvm::Value *Fini = emitReduceFiniFunction(CGM, Loc, RCG, Cnt);
5520     llvm::Value *FiniAddr =
5521         Fini ? Fini : llvm::ConstantPointerNull::get(CGM.VoidPtrTy);
5522     CGF.EmitStoreOfScalar(FiniAddr, FiniLVal);
5523     // ElemLVal.reduce_comb = comb;
5524     LValue CombLVal = CGF.EmitLValueForField(ElemLVal, CombFD);
5525     llvm::Value *CombAddr = emitReduceCombFunction(
5526         CGM, Loc, RCG, Cnt, Data.ReductionOps[Cnt], LHSExprs[Cnt],
5527         RHSExprs[Cnt], Data.ReductionCopies[Cnt]);
5528     CGF.EmitStoreOfScalar(CombAddr, CombLVal);
5529     // ElemLVal.flags = 0;
5530     LValue FlagsLVal = CGF.EmitLValueForField(ElemLVal, FlagsFD);
5531     if (DelayedCreation) {
5532       CGF.EmitStoreOfScalar(
5533           llvm::ConstantInt::get(CGM.Int32Ty, /*V=*/1, /*isSigned=*/true),
5534           FlagsLVal);
5535     } else
5536       CGF.EmitNullInitialization(FlagsLVal.getAddress(), FlagsLVal.getType());
5537   }
5538   if (Data.IsReductionWithTaskMod) {
5539     // Build call void *__kmpc_taskred_modifier_init(ident_t *loc, int gtid, int
5540     // is_ws, int num, void *data);
5541     llvm::Value *IdentTLoc = emitUpdateLocation(CGF, Loc);
5542     llvm::Value *GTid = CGF.Builder.CreateIntCast(getThreadID(CGF, Loc),
5543                                                   CGM.IntTy, /*isSigned=*/true);
5544     llvm::Value *Args[] = {
5545         IdentTLoc, GTid,
5546         llvm::ConstantInt::get(CGM.IntTy, Data.IsWorksharingReduction ? 1 : 0,
5547                                /*isSigned=*/true),
5548         llvm::ConstantInt::get(CGM.IntTy, Size, /*isSigned=*/true),
5549         CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
5550             TaskRedInput.getPointer(), CGM.VoidPtrTy)};
5551     return CGF.EmitRuntimeCall(
5552         OMPBuilder.getOrCreateRuntimeFunction(
5553             CGM.getModule(), OMPRTL___kmpc_taskred_modifier_init),
5554         Args);
5555   }
5556   // Build call void *__kmpc_taskred_init(int gtid, int num_data, void *data);
5557   llvm::Value *Args[] = {
5558       CGF.Builder.CreateIntCast(getThreadID(CGF, Loc), CGM.IntTy,
5559                                 /*isSigned=*/true),
5560       llvm::ConstantInt::get(CGM.IntTy, Size, /*isSigned=*/true),
5561       CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(TaskRedInput.getPointer(),
5562                                                       CGM.VoidPtrTy)};
5563   return CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
5564                                  CGM.getModule(), OMPRTL___kmpc_taskred_init),
5565                              Args);
5566 }
5567 
emitTaskReductionFini(CodeGenFunction & CGF,SourceLocation Loc,bool IsWorksharingReduction)5568 void CGOpenMPRuntime::emitTaskReductionFini(CodeGenFunction &CGF,
5569                                             SourceLocation Loc,
5570                                             bool IsWorksharingReduction) {
5571   // Build call void *__kmpc_taskred_modifier_init(ident_t *loc, int gtid, int
5572   // is_ws, int num, void *data);
5573   llvm::Value *IdentTLoc = emitUpdateLocation(CGF, Loc);
5574   llvm::Value *GTid = CGF.Builder.CreateIntCast(getThreadID(CGF, Loc),
5575                                                 CGM.IntTy, /*isSigned=*/true);
5576   llvm::Value *Args[] = {IdentTLoc, GTid,
5577                          llvm::ConstantInt::get(CGM.IntTy,
5578                                                 IsWorksharingReduction ? 1 : 0,
5579                                                 /*isSigned=*/true)};
5580   (void)CGF.EmitRuntimeCall(
5581       OMPBuilder.getOrCreateRuntimeFunction(
5582           CGM.getModule(), OMPRTL___kmpc_task_reduction_modifier_fini),
5583       Args);
5584 }
5585 
emitTaskReductionFixups(CodeGenFunction & CGF,SourceLocation Loc,ReductionCodeGen & RCG,unsigned N)5586 void CGOpenMPRuntime::emitTaskReductionFixups(CodeGenFunction &CGF,
5587                                               SourceLocation Loc,
5588                                               ReductionCodeGen &RCG,
5589                                               unsigned N) {
5590   auto Sizes = RCG.getSizes(N);
5591   // Emit threadprivate global variable if the type is non-constant
5592   // (Sizes.second = nullptr).
5593   if (Sizes.second) {
5594     llvm::Value *SizeVal = CGF.Builder.CreateIntCast(Sizes.second, CGM.SizeTy,
5595                                                      /*isSigned=*/false);
5596     Address SizeAddr = getAddrOfArtificialThreadPrivate(
5597         CGF, CGM.getContext().getSizeType(),
5598         generateUniqueName(CGM, "reduction_size", RCG.getRefExpr(N)));
5599     CGF.Builder.CreateStore(SizeVal, SizeAddr, /*IsVolatile=*/false);
5600   }
5601 }
5602 
getTaskReductionItem(CodeGenFunction & CGF,SourceLocation Loc,llvm::Value * ReductionsPtr,LValue SharedLVal)5603 Address CGOpenMPRuntime::getTaskReductionItem(CodeGenFunction &CGF,
5604                                               SourceLocation Loc,
5605                                               llvm::Value *ReductionsPtr,
5606                                               LValue SharedLVal) {
5607   // Build call void *__kmpc_task_reduction_get_th_data(int gtid, void *tg, void
5608   // *d);
5609   llvm::Value *Args[] = {CGF.Builder.CreateIntCast(getThreadID(CGF, Loc),
5610                                                    CGM.IntTy,
5611                                                    /*isSigned=*/true),
5612                          ReductionsPtr,
5613                          CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
5614                              SharedLVal.getPointer(CGF), CGM.VoidPtrTy)};
5615   return Address(
5616       CGF.EmitRuntimeCall(
5617           OMPBuilder.getOrCreateRuntimeFunction(
5618               CGM.getModule(), OMPRTL___kmpc_task_reduction_get_th_data),
5619           Args),
5620       CGF.Int8Ty, SharedLVal.getAlignment());
5621 }
5622 
emitTaskwaitCall(CodeGenFunction & CGF,SourceLocation Loc,const OMPTaskDataTy & Data)5623 void CGOpenMPRuntime::emitTaskwaitCall(CodeGenFunction &CGF, SourceLocation Loc,
5624                                        const OMPTaskDataTy &Data) {
5625   if (!CGF.HaveInsertPoint())
5626     return;
5627 
5628   if (CGF.CGM.getLangOpts().OpenMPIRBuilder && Data.Dependences.empty()) {
5629     // TODO: Need to support taskwait with dependences in the OpenMPIRBuilder.
5630     OMPBuilder.createTaskwait(CGF.Builder);
5631   } else {
5632     llvm::Value *ThreadID = getThreadID(CGF, Loc);
5633     llvm::Value *UpLoc = emitUpdateLocation(CGF, Loc);
5634     auto &M = CGM.getModule();
5635     Address DependenciesArray = Address::invalid();
5636     llvm::Value *NumOfElements;
5637     std::tie(NumOfElements, DependenciesArray) =
5638         emitDependClause(CGF, Data.Dependences, Loc);
5639     if (!Data.Dependences.empty()) {
5640       llvm::Value *DepWaitTaskArgs[7];
5641       DepWaitTaskArgs[0] = UpLoc;
5642       DepWaitTaskArgs[1] = ThreadID;
5643       DepWaitTaskArgs[2] = NumOfElements;
5644       DepWaitTaskArgs[3] = DependenciesArray.emitRawPointer(CGF);
5645       DepWaitTaskArgs[4] = CGF.Builder.getInt32(0);
5646       DepWaitTaskArgs[5] = llvm::ConstantPointerNull::get(CGF.VoidPtrTy);
5647       DepWaitTaskArgs[6] =
5648           llvm::ConstantInt::get(CGF.Int32Ty, Data.HasNowaitClause);
5649 
5650       CodeGenFunction::RunCleanupsScope LocalScope(CGF);
5651 
5652       // Build void __kmpc_omp_taskwait_deps_51(ident_t *, kmp_int32 gtid,
5653       // kmp_int32 ndeps, kmp_depend_info_t *dep_list, kmp_int32
5654       // ndeps_noalias, kmp_depend_info_t *noalias_dep_list,
5655       // kmp_int32 has_no_wait); if dependence info is specified.
5656       CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
5657                               M, OMPRTL___kmpc_omp_taskwait_deps_51),
5658                           DepWaitTaskArgs);
5659 
5660     } else {
5661 
5662       // Build call kmp_int32 __kmpc_omp_taskwait(ident_t *loc, kmp_int32
5663       // global_tid);
5664       llvm::Value *Args[] = {UpLoc, ThreadID};
5665       // Ignore return result until untied tasks are supported.
5666       CGF.EmitRuntimeCall(
5667           OMPBuilder.getOrCreateRuntimeFunction(M, OMPRTL___kmpc_omp_taskwait),
5668           Args);
5669     }
5670   }
5671 
5672   if (auto *Region = dyn_cast_or_null<CGOpenMPRegionInfo>(CGF.CapturedStmtInfo))
5673     Region->emitUntiedSwitch(CGF);
5674 }
5675 
emitInlinedDirective(CodeGenFunction & CGF,OpenMPDirectiveKind InnerKind,const RegionCodeGenTy & CodeGen,bool HasCancel)5676 void CGOpenMPRuntime::emitInlinedDirective(CodeGenFunction &CGF,
5677                                            OpenMPDirectiveKind InnerKind,
5678                                            const RegionCodeGenTy &CodeGen,
5679                                            bool HasCancel) {
5680   if (!CGF.HaveInsertPoint())
5681     return;
5682   InlinedOpenMPRegionRAII Region(CGF, CodeGen, InnerKind, HasCancel,
5683                                  InnerKind != OMPD_critical &&
5684                                      InnerKind != OMPD_master &&
5685                                      InnerKind != OMPD_masked);
5686   CGF.CapturedStmtInfo->EmitBody(CGF, /*S=*/nullptr);
5687 }
5688 
5689 namespace {
5690 enum RTCancelKind {
5691   CancelNoreq = 0,
5692   CancelParallel = 1,
5693   CancelLoop = 2,
5694   CancelSections = 3,
5695   CancelTaskgroup = 4
5696 };
5697 } // anonymous namespace
5698 
getCancellationKind(OpenMPDirectiveKind CancelRegion)5699 static RTCancelKind getCancellationKind(OpenMPDirectiveKind CancelRegion) {
5700   RTCancelKind CancelKind = CancelNoreq;
5701   if (CancelRegion == OMPD_parallel)
5702     CancelKind = CancelParallel;
5703   else if (CancelRegion == OMPD_for)
5704     CancelKind = CancelLoop;
5705   else if (CancelRegion == OMPD_sections)
5706     CancelKind = CancelSections;
5707   else {
5708     assert(CancelRegion == OMPD_taskgroup);
5709     CancelKind = CancelTaskgroup;
5710   }
5711   return CancelKind;
5712 }
5713 
emitCancellationPointCall(CodeGenFunction & CGF,SourceLocation Loc,OpenMPDirectiveKind CancelRegion)5714 void CGOpenMPRuntime::emitCancellationPointCall(
5715     CodeGenFunction &CGF, SourceLocation Loc,
5716     OpenMPDirectiveKind CancelRegion) {
5717   if (!CGF.HaveInsertPoint())
5718     return;
5719   // Build call kmp_int32 __kmpc_cancellationpoint(ident_t *loc, kmp_int32
5720   // global_tid, kmp_int32 cncl_kind);
5721   if (auto *OMPRegionInfo =
5722           dyn_cast_or_null<CGOpenMPRegionInfo>(CGF.CapturedStmtInfo)) {
5723     // For 'cancellation point taskgroup', the task region info may not have a
5724     // cancel. This may instead happen in another adjacent task.
5725     if (CancelRegion == OMPD_taskgroup || OMPRegionInfo->hasCancel()) {
5726       llvm::Value *Args[] = {
5727           emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc),
5728           CGF.Builder.getInt32(getCancellationKind(CancelRegion))};
5729       // Ignore return result until untied tasks are supported.
5730       llvm::Value *Result = CGF.EmitRuntimeCall(
5731           OMPBuilder.getOrCreateRuntimeFunction(
5732               CGM.getModule(), OMPRTL___kmpc_cancellationpoint),
5733           Args);
5734       // if (__kmpc_cancellationpoint()) {
5735       //   call i32 @__kmpc_cancel_barrier( // for parallel cancellation only
5736       //   exit from construct;
5737       // }
5738       llvm::BasicBlock *ExitBB = CGF.createBasicBlock(".cancel.exit");
5739       llvm::BasicBlock *ContBB = CGF.createBasicBlock(".cancel.continue");
5740       llvm::Value *Cmp = CGF.Builder.CreateIsNotNull(Result);
5741       CGF.Builder.CreateCondBr(Cmp, ExitBB, ContBB);
5742       CGF.EmitBlock(ExitBB);
5743       if (CancelRegion == OMPD_parallel)
5744         emitBarrierCall(CGF, Loc, OMPD_unknown, /*EmitChecks=*/false);
5745       // exit from construct;
5746       CodeGenFunction::JumpDest CancelDest =
5747           CGF.getOMPCancelDestination(OMPRegionInfo->getDirectiveKind());
5748       CGF.EmitBranchThroughCleanup(CancelDest);
5749       CGF.EmitBlock(ContBB, /*IsFinished=*/true);
5750     }
5751   }
5752 }
5753 
emitCancelCall(CodeGenFunction & CGF,SourceLocation Loc,const Expr * IfCond,OpenMPDirectiveKind CancelRegion)5754 void CGOpenMPRuntime::emitCancelCall(CodeGenFunction &CGF, SourceLocation Loc,
5755                                      const Expr *IfCond,
5756                                      OpenMPDirectiveKind CancelRegion) {
5757   if (!CGF.HaveInsertPoint())
5758     return;
5759   // Build call kmp_int32 __kmpc_cancel(ident_t *loc, kmp_int32 global_tid,
5760   // kmp_int32 cncl_kind);
5761   auto &M = CGM.getModule();
5762   if (auto *OMPRegionInfo =
5763           dyn_cast_or_null<CGOpenMPRegionInfo>(CGF.CapturedStmtInfo)) {
5764     auto &&ThenGen = [this, &M, Loc, CancelRegion,
5765                       OMPRegionInfo](CodeGenFunction &CGF, PrePostActionTy &) {
5766       CGOpenMPRuntime &RT = CGF.CGM.getOpenMPRuntime();
5767       llvm::Value *Args[] = {
5768           RT.emitUpdateLocation(CGF, Loc), RT.getThreadID(CGF, Loc),
5769           CGF.Builder.getInt32(getCancellationKind(CancelRegion))};
5770       // Ignore return result until untied tasks are supported.
5771       llvm::Value *Result = CGF.EmitRuntimeCall(
5772           OMPBuilder.getOrCreateRuntimeFunction(M, OMPRTL___kmpc_cancel), Args);
5773       // if (__kmpc_cancel()) {
5774       //   call i32 @__kmpc_cancel_barrier( // for parallel cancellation only
5775       //   exit from construct;
5776       // }
5777       llvm::BasicBlock *ExitBB = CGF.createBasicBlock(".cancel.exit");
5778       llvm::BasicBlock *ContBB = CGF.createBasicBlock(".cancel.continue");
5779       llvm::Value *Cmp = CGF.Builder.CreateIsNotNull(Result);
5780       CGF.Builder.CreateCondBr(Cmp, ExitBB, ContBB);
5781       CGF.EmitBlock(ExitBB);
5782       if (CancelRegion == OMPD_parallel)
5783         RT.emitBarrierCall(CGF, Loc, OMPD_unknown, /*EmitChecks=*/false);
5784       // exit from construct;
5785       CodeGenFunction::JumpDest CancelDest =
5786           CGF.getOMPCancelDestination(OMPRegionInfo->getDirectiveKind());
5787       CGF.EmitBranchThroughCleanup(CancelDest);
5788       CGF.EmitBlock(ContBB, /*IsFinished=*/true);
5789     };
5790     if (IfCond) {
5791       emitIfClause(CGF, IfCond, ThenGen,
5792                    [](CodeGenFunction &, PrePostActionTy &) {});
5793     } else {
5794       RegionCodeGenTy ThenRCG(ThenGen);
5795       ThenRCG(CGF);
5796     }
5797   }
5798 }
5799 
5800 namespace {
5801 /// Cleanup action for uses_allocators support.
5802 class OMPUsesAllocatorsActionTy final : public PrePostActionTy {
5803   ArrayRef<std::pair<const Expr *, const Expr *>> Allocators;
5804 
5805 public:
OMPUsesAllocatorsActionTy(ArrayRef<std::pair<const Expr *,const Expr * >> Allocators)5806   OMPUsesAllocatorsActionTy(
5807       ArrayRef<std::pair<const Expr *, const Expr *>> Allocators)
5808       : Allocators(Allocators) {}
Enter(CodeGenFunction & CGF)5809   void Enter(CodeGenFunction &CGF) override {
5810     if (!CGF.HaveInsertPoint())
5811       return;
5812     for (const auto &AllocatorData : Allocators) {
5813       CGF.CGM.getOpenMPRuntime().emitUsesAllocatorsInit(
5814           CGF, AllocatorData.first, AllocatorData.second);
5815     }
5816   }
Exit(CodeGenFunction & CGF)5817   void Exit(CodeGenFunction &CGF) override {
5818     if (!CGF.HaveInsertPoint())
5819       return;
5820     for (const auto &AllocatorData : Allocators) {
5821       CGF.CGM.getOpenMPRuntime().emitUsesAllocatorsFini(CGF,
5822                                                         AllocatorData.first);
5823     }
5824   }
5825 };
5826 } // namespace
5827 
emitTargetOutlinedFunction(const OMPExecutableDirective & D,StringRef ParentName,llvm::Function * & OutlinedFn,llvm::Constant * & OutlinedFnID,bool IsOffloadEntry,const RegionCodeGenTy & CodeGen)5828 void CGOpenMPRuntime::emitTargetOutlinedFunction(
5829     const OMPExecutableDirective &D, StringRef ParentName,
5830     llvm::Function *&OutlinedFn, llvm::Constant *&OutlinedFnID,
5831     bool IsOffloadEntry, const RegionCodeGenTy &CodeGen) {
5832   assert(!ParentName.empty() && "Invalid target entry parent name!");
5833   HasEmittedTargetRegion = true;
5834   SmallVector<std::pair<const Expr *, const Expr *>, 4> Allocators;
5835   for (const auto *C : D.getClausesOfKind<OMPUsesAllocatorsClause>()) {
5836     for (unsigned I = 0, E = C->getNumberOfAllocators(); I < E; ++I) {
5837       const OMPUsesAllocatorsClause::Data D = C->getAllocatorData(I);
5838       if (!D.AllocatorTraits)
5839         continue;
5840       Allocators.emplace_back(D.Allocator, D.AllocatorTraits);
5841     }
5842   }
5843   OMPUsesAllocatorsActionTy UsesAllocatorAction(Allocators);
5844   CodeGen.setAction(UsesAllocatorAction);
5845   emitTargetOutlinedFunctionHelper(D, ParentName, OutlinedFn, OutlinedFnID,
5846                                    IsOffloadEntry, CodeGen);
5847 }
5848 
emitUsesAllocatorsInit(CodeGenFunction & CGF,const Expr * Allocator,const Expr * AllocatorTraits)5849 void CGOpenMPRuntime::emitUsesAllocatorsInit(CodeGenFunction &CGF,
5850                                              const Expr *Allocator,
5851                                              const Expr *AllocatorTraits) {
5852   llvm::Value *ThreadId = getThreadID(CGF, Allocator->getExprLoc());
5853   ThreadId = CGF.Builder.CreateIntCast(ThreadId, CGF.IntTy, /*isSigned=*/true);
5854   // Use default memspace handle.
5855   llvm::Value *MemSpaceHandle = llvm::ConstantPointerNull::get(CGF.VoidPtrTy);
5856   llvm::Value *NumTraits = llvm::ConstantInt::get(
5857       CGF.IntTy, cast<ConstantArrayType>(
5858                      AllocatorTraits->getType()->getAsArrayTypeUnsafe())
5859                      ->getSize()
5860                      .getLimitedValue());
5861   LValue AllocatorTraitsLVal = CGF.EmitLValue(AllocatorTraits);
5862   Address Addr = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
5863       AllocatorTraitsLVal.getAddress(), CGF.VoidPtrPtrTy, CGF.VoidPtrTy);
5864   AllocatorTraitsLVal = CGF.MakeAddrLValue(Addr, CGF.getContext().VoidPtrTy,
5865                                            AllocatorTraitsLVal.getBaseInfo(),
5866                                            AllocatorTraitsLVal.getTBAAInfo());
5867   llvm::Value *Traits = Addr.emitRawPointer(CGF);
5868 
5869   llvm::Value *AllocatorVal =
5870       CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
5871                               CGM.getModule(), OMPRTL___kmpc_init_allocator),
5872                           {ThreadId, MemSpaceHandle, NumTraits, Traits});
5873   // Store to allocator.
5874   CGF.EmitAutoVarAlloca(*cast<VarDecl>(
5875       cast<DeclRefExpr>(Allocator->IgnoreParenImpCasts())->getDecl()));
5876   LValue AllocatorLVal = CGF.EmitLValue(Allocator->IgnoreParenImpCasts());
5877   AllocatorVal =
5878       CGF.EmitScalarConversion(AllocatorVal, CGF.getContext().VoidPtrTy,
5879                                Allocator->getType(), Allocator->getExprLoc());
5880   CGF.EmitStoreOfScalar(AllocatorVal, AllocatorLVal);
5881 }
5882 
emitUsesAllocatorsFini(CodeGenFunction & CGF,const Expr * Allocator)5883 void CGOpenMPRuntime::emitUsesAllocatorsFini(CodeGenFunction &CGF,
5884                                              const Expr *Allocator) {
5885   llvm::Value *ThreadId = getThreadID(CGF, Allocator->getExprLoc());
5886   ThreadId = CGF.Builder.CreateIntCast(ThreadId, CGF.IntTy, /*isSigned=*/true);
5887   LValue AllocatorLVal = CGF.EmitLValue(Allocator->IgnoreParenImpCasts());
5888   llvm::Value *AllocatorVal =
5889       CGF.EmitLoadOfScalar(AllocatorLVal, Allocator->getExprLoc());
5890   AllocatorVal = CGF.EmitScalarConversion(AllocatorVal, Allocator->getType(),
5891                                           CGF.getContext().VoidPtrTy,
5892                                           Allocator->getExprLoc());
5893   (void)CGF.EmitRuntimeCall(
5894       OMPBuilder.getOrCreateRuntimeFunction(CGM.getModule(),
5895                                             OMPRTL___kmpc_destroy_allocator),
5896       {ThreadId, AllocatorVal});
5897 }
5898 
computeMinAndMaxThreadsAndTeams(const OMPExecutableDirective & D,CodeGenFunction & CGF,int32_t & MinThreadsVal,int32_t & MaxThreadsVal,int32_t & MinTeamsVal,int32_t & MaxTeamsVal)5899 void CGOpenMPRuntime::computeMinAndMaxThreadsAndTeams(
5900     const OMPExecutableDirective &D, CodeGenFunction &CGF,
5901     int32_t &MinThreadsVal, int32_t &MaxThreadsVal, int32_t &MinTeamsVal,
5902     int32_t &MaxTeamsVal) {
5903 
5904   getNumTeamsExprForTargetDirective(CGF, D, MinTeamsVal, MaxTeamsVal);
5905   getNumThreadsExprForTargetDirective(CGF, D, MaxThreadsVal,
5906                                       /*UpperBoundOnly=*/true);
5907 
5908   for (auto *C : D.getClausesOfKind<OMPXAttributeClause>()) {
5909     for (auto *A : C->getAttrs()) {
5910       int32_t AttrMinThreadsVal = 1, AttrMaxThreadsVal = -1;
5911       int32_t AttrMinBlocksVal = 1, AttrMaxBlocksVal = -1;
5912       if (auto *Attr = dyn_cast<CUDALaunchBoundsAttr>(A))
5913         CGM.handleCUDALaunchBoundsAttr(nullptr, Attr, &AttrMaxThreadsVal,
5914                                        &AttrMinBlocksVal, &AttrMaxBlocksVal);
5915       else if (auto *Attr = dyn_cast<AMDGPUFlatWorkGroupSizeAttr>(A))
5916         CGM.handleAMDGPUFlatWorkGroupSizeAttr(
5917             nullptr, Attr, /*ReqdWGS=*/nullptr, &AttrMinThreadsVal,
5918             &AttrMaxThreadsVal);
5919       else
5920         continue;
5921 
5922       MinThreadsVal = std::max(MinThreadsVal, AttrMinThreadsVal);
5923       if (AttrMaxThreadsVal > 0)
5924         MaxThreadsVal = MaxThreadsVal > 0
5925                             ? std::min(MaxThreadsVal, AttrMaxThreadsVal)
5926                             : AttrMaxThreadsVal;
5927       MinTeamsVal = std::max(MinTeamsVal, AttrMinBlocksVal);
5928       if (AttrMaxBlocksVal > 0)
5929         MaxTeamsVal = MaxTeamsVal > 0 ? std::min(MaxTeamsVal, AttrMaxBlocksVal)
5930                                       : AttrMaxBlocksVal;
5931     }
5932   }
5933 }
5934 
emitTargetOutlinedFunctionHelper(const OMPExecutableDirective & D,StringRef ParentName,llvm::Function * & OutlinedFn,llvm::Constant * & OutlinedFnID,bool IsOffloadEntry,const RegionCodeGenTy & CodeGen)5935 void CGOpenMPRuntime::emitTargetOutlinedFunctionHelper(
5936     const OMPExecutableDirective &D, StringRef ParentName,
5937     llvm::Function *&OutlinedFn, llvm::Constant *&OutlinedFnID,
5938     bool IsOffloadEntry, const RegionCodeGenTy &CodeGen) {
5939 
5940   llvm::TargetRegionEntryInfo EntryInfo =
5941       getEntryInfoFromPresumedLoc(CGM, OMPBuilder, D.getBeginLoc(), ParentName);
5942 
5943   CodeGenFunction CGF(CGM, true);
5944   llvm::OpenMPIRBuilder::FunctionGenCallback &&GenerateOutlinedFunction =
5945       [&CGF, &D, &CodeGen](StringRef EntryFnName) {
5946         const CapturedStmt &CS = *D.getCapturedStmt(OMPD_target);
5947 
5948         CGOpenMPTargetRegionInfo CGInfo(CS, CodeGen, EntryFnName);
5949         CodeGenFunction::CGCapturedStmtRAII CapInfoRAII(CGF, &CGInfo);
5950         return CGF.GenerateOpenMPCapturedStmtFunction(CS, D.getBeginLoc());
5951       };
5952 
5953   OMPBuilder.emitTargetRegionFunction(EntryInfo, GenerateOutlinedFunction,
5954                                       IsOffloadEntry, OutlinedFn, OutlinedFnID);
5955 
5956   if (!OutlinedFn)
5957     return;
5958 
5959   CGM.getTargetCodeGenInfo().setTargetAttributes(nullptr, OutlinedFn, CGM);
5960 
5961   for (auto *C : D.getClausesOfKind<OMPXAttributeClause>()) {
5962     for (auto *A : C->getAttrs()) {
5963       if (auto *Attr = dyn_cast<AMDGPUWavesPerEUAttr>(A))
5964         CGM.handleAMDGPUWavesPerEUAttr(OutlinedFn, Attr);
5965     }
5966   }
5967 }
5968 
5969 /// Checks if the expression is constant or does not have non-trivial function
5970 /// calls.
isTrivial(ASTContext & Ctx,const Expr * E)5971 static bool isTrivial(ASTContext &Ctx, const Expr * E) {
5972   // We can skip constant expressions.
5973   // We can skip expressions with trivial calls or simple expressions.
5974   return (E->isEvaluatable(Ctx, Expr::SE_AllowUndefinedBehavior) ||
5975           !E->hasNonTrivialCall(Ctx)) &&
5976          !E->HasSideEffects(Ctx, /*IncludePossibleEffects=*/true);
5977 }
5978 
getSingleCompoundChild(ASTContext & Ctx,const Stmt * Body)5979 const Stmt *CGOpenMPRuntime::getSingleCompoundChild(ASTContext &Ctx,
5980                                                     const Stmt *Body) {
5981   const Stmt *Child = Body->IgnoreContainers();
5982   while (const auto *C = dyn_cast_or_null<CompoundStmt>(Child)) {
5983     Child = nullptr;
5984     for (const Stmt *S : C->body()) {
5985       if (const auto *E = dyn_cast<Expr>(S)) {
5986         if (isTrivial(Ctx, E))
5987           continue;
5988       }
5989       // Some of the statements can be ignored.
5990       if (isa<AsmStmt>(S) || isa<NullStmt>(S) || isa<OMPFlushDirective>(S) ||
5991           isa<OMPBarrierDirective>(S) || isa<OMPTaskyieldDirective>(S))
5992         continue;
5993       // Analyze declarations.
5994       if (const auto *DS = dyn_cast<DeclStmt>(S)) {
5995         if (llvm::all_of(DS->decls(), [](const Decl *D) {
5996               if (isa<EmptyDecl>(D) || isa<DeclContext>(D) ||
5997                   isa<TypeDecl>(D) || isa<PragmaCommentDecl>(D) ||
5998                   isa<PragmaDetectMismatchDecl>(D) || isa<UsingDecl>(D) ||
5999                   isa<UsingDirectiveDecl>(D) ||
6000                   isa<OMPDeclareReductionDecl>(D) ||
6001                   isa<OMPThreadPrivateDecl>(D) || isa<OMPAllocateDecl>(D))
6002                 return true;
6003               const auto *VD = dyn_cast<VarDecl>(D);
6004               if (!VD)
6005                 return false;
6006               return VD->hasGlobalStorage() || !VD->isUsed();
6007             }))
6008           continue;
6009       }
6010       // Found multiple children - cannot get the one child only.
6011       if (Child)
6012         return nullptr;
6013       Child = S;
6014     }
6015     if (Child)
6016       Child = Child->IgnoreContainers();
6017   }
6018   return Child;
6019 }
6020 
getNumTeamsExprForTargetDirective(CodeGenFunction & CGF,const OMPExecutableDirective & D,int32_t & MinTeamsVal,int32_t & MaxTeamsVal)6021 const Expr *CGOpenMPRuntime::getNumTeamsExprForTargetDirective(
6022     CodeGenFunction &CGF, const OMPExecutableDirective &D, int32_t &MinTeamsVal,
6023     int32_t &MaxTeamsVal) {
6024 
6025   OpenMPDirectiveKind DirectiveKind = D.getDirectiveKind();
6026   assert(isOpenMPTargetExecutionDirective(DirectiveKind) &&
6027          "Expected target-based executable directive.");
6028   switch (DirectiveKind) {
6029   case OMPD_target: {
6030     const auto *CS = D.getInnermostCapturedStmt();
6031     const auto *Body =
6032         CS->getCapturedStmt()->IgnoreContainers(/*IgnoreCaptured=*/true);
6033     const Stmt *ChildStmt =
6034         CGOpenMPRuntime::getSingleCompoundChild(CGF.getContext(), Body);
6035     if (const auto *NestedDir =
6036             dyn_cast_or_null<OMPExecutableDirective>(ChildStmt)) {
6037       if (isOpenMPTeamsDirective(NestedDir->getDirectiveKind())) {
6038         if (NestedDir->hasClausesOfKind<OMPNumTeamsClause>()) {
6039           const Expr *NumTeams =
6040               NestedDir->getSingleClause<OMPNumTeamsClause>()->getNumTeams();
6041           if (NumTeams->isIntegerConstantExpr(CGF.getContext()))
6042             if (auto Constant =
6043                     NumTeams->getIntegerConstantExpr(CGF.getContext()))
6044               MinTeamsVal = MaxTeamsVal = Constant->getExtValue();
6045           return NumTeams;
6046         }
6047         MinTeamsVal = MaxTeamsVal = 0;
6048         return nullptr;
6049       }
6050       if (isOpenMPParallelDirective(NestedDir->getDirectiveKind()) ||
6051           isOpenMPSimdDirective(NestedDir->getDirectiveKind())) {
6052         MinTeamsVal = MaxTeamsVal = 1;
6053         return nullptr;
6054       }
6055       MinTeamsVal = MaxTeamsVal = 1;
6056       return nullptr;
6057     }
6058     // A value of -1 is used to check if we need to emit no teams region
6059     MinTeamsVal = MaxTeamsVal = -1;
6060     return nullptr;
6061   }
6062   case OMPD_target_teams_loop:
6063   case OMPD_target_teams:
6064   case OMPD_target_teams_distribute:
6065   case OMPD_target_teams_distribute_simd:
6066   case OMPD_target_teams_distribute_parallel_for:
6067   case OMPD_target_teams_distribute_parallel_for_simd: {
6068     if (D.hasClausesOfKind<OMPNumTeamsClause>()) {
6069       const Expr *NumTeams =
6070           D.getSingleClause<OMPNumTeamsClause>()->getNumTeams();
6071       if (NumTeams->isIntegerConstantExpr(CGF.getContext()))
6072         if (auto Constant = NumTeams->getIntegerConstantExpr(CGF.getContext()))
6073           MinTeamsVal = MaxTeamsVal = Constant->getExtValue();
6074       return NumTeams;
6075     }
6076     MinTeamsVal = MaxTeamsVal = 0;
6077     return nullptr;
6078   }
6079   case OMPD_target_parallel:
6080   case OMPD_target_parallel_for:
6081   case OMPD_target_parallel_for_simd:
6082   case OMPD_target_parallel_loop:
6083   case OMPD_target_simd:
6084     MinTeamsVal = MaxTeamsVal = 1;
6085     return nullptr;
6086   case OMPD_parallel:
6087   case OMPD_for:
6088   case OMPD_parallel_for:
6089   case OMPD_parallel_loop:
6090   case OMPD_parallel_master:
6091   case OMPD_parallel_sections:
6092   case OMPD_for_simd:
6093   case OMPD_parallel_for_simd:
6094   case OMPD_cancel:
6095   case OMPD_cancellation_point:
6096   case OMPD_ordered:
6097   case OMPD_threadprivate:
6098   case OMPD_allocate:
6099   case OMPD_task:
6100   case OMPD_simd:
6101   case OMPD_tile:
6102   case OMPD_unroll:
6103   case OMPD_sections:
6104   case OMPD_section:
6105   case OMPD_single:
6106   case OMPD_master:
6107   case OMPD_critical:
6108   case OMPD_taskyield:
6109   case OMPD_barrier:
6110   case OMPD_taskwait:
6111   case OMPD_taskgroup:
6112   case OMPD_atomic:
6113   case OMPD_flush:
6114   case OMPD_depobj:
6115   case OMPD_scan:
6116   case OMPD_teams:
6117   case OMPD_target_data:
6118   case OMPD_target_exit_data:
6119   case OMPD_target_enter_data:
6120   case OMPD_distribute:
6121   case OMPD_distribute_simd:
6122   case OMPD_distribute_parallel_for:
6123   case OMPD_distribute_parallel_for_simd:
6124   case OMPD_teams_distribute:
6125   case OMPD_teams_distribute_simd:
6126   case OMPD_teams_distribute_parallel_for:
6127   case OMPD_teams_distribute_parallel_for_simd:
6128   case OMPD_target_update:
6129   case OMPD_declare_simd:
6130   case OMPD_declare_variant:
6131   case OMPD_begin_declare_variant:
6132   case OMPD_end_declare_variant:
6133   case OMPD_declare_target:
6134   case OMPD_end_declare_target:
6135   case OMPD_declare_reduction:
6136   case OMPD_declare_mapper:
6137   case OMPD_taskloop:
6138   case OMPD_taskloop_simd:
6139   case OMPD_master_taskloop:
6140   case OMPD_master_taskloop_simd:
6141   case OMPD_parallel_master_taskloop:
6142   case OMPD_parallel_master_taskloop_simd:
6143   case OMPD_requires:
6144   case OMPD_metadirective:
6145   case OMPD_unknown:
6146     break;
6147   default:
6148     break;
6149   }
6150   llvm_unreachable("Unexpected directive kind.");
6151 }
6152 
emitNumTeamsForTargetDirective(CodeGenFunction & CGF,const OMPExecutableDirective & D)6153 llvm::Value *CGOpenMPRuntime::emitNumTeamsForTargetDirective(
6154     CodeGenFunction &CGF, const OMPExecutableDirective &D) {
6155   assert(!CGF.getLangOpts().OpenMPIsTargetDevice &&
6156          "Clauses associated with the teams directive expected to be emitted "
6157          "only for the host!");
6158   CGBuilderTy &Bld = CGF.Builder;
6159   int32_t MinNT = -1, MaxNT = -1;
6160   const Expr *NumTeams =
6161       getNumTeamsExprForTargetDirective(CGF, D, MinNT, MaxNT);
6162   if (NumTeams != nullptr) {
6163     OpenMPDirectiveKind DirectiveKind = D.getDirectiveKind();
6164 
6165     switch (DirectiveKind) {
6166     case OMPD_target: {
6167       const auto *CS = D.getInnermostCapturedStmt();
6168       CGOpenMPInnerExprInfo CGInfo(CGF, *CS);
6169       CodeGenFunction::CGCapturedStmtRAII CapInfoRAII(CGF, &CGInfo);
6170       llvm::Value *NumTeamsVal = CGF.EmitScalarExpr(NumTeams,
6171                                                   /*IgnoreResultAssign*/ true);
6172       return Bld.CreateIntCast(NumTeamsVal, CGF.Int32Ty,
6173                              /*isSigned=*/true);
6174     }
6175     case OMPD_target_teams:
6176     case OMPD_target_teams_distribute:
6177     case OMPD_target_teams_distribute_simd:
6178     case OMPD_target_teams_distribute_parallel_for:
6179     case OMPD_target_teams_distribute_parallel_for_simd: {
6180       CodeGenFunction::RunCleanupsScope NumTeamsScope(CGF);
6181       llvm::Value *NumTeamsVal = CGF.EmitScalarExpr(NumTeams,
6182                                                   /*IgnoreResultAssign*/ true);
6183       return Bld.CreateIntCast(NumTeamsVal, CGF.Int32Ty,
6184                              /*isSigned=*/true);
6185     }
6186     default:
6187       break;
6188     }
6189   }
6190 
6191   assert(MinNT == MaxNT && "Num threads ranges require handling here.");
6192   return llvm::ConstantInt::get(CGF.Int32Ty, MinNT);
6193 }
6194 
6195 /// Check for a num threads constant value (stored in \p DefaultVal), or
6196 /// expression (stored in \p E). If the value is conditional (via an if-clause),
6197 /// store the condition in \p CondVal. If \p E, and \p CondVal respectively, are
6198 /// nullptr, no expression evaluation is perfomed.
getNumThreads(CodeGenFunction & CGF,const CapturedStmt * CS,const Expr ** E,int32_t & UpperBound,bool UpperBoundOnly,llvm::Value ** CondVal)6199 static void getNumThreads(CodeGenFunction &CGF, const CapturedStmt *CS,
6200                           const Expr **E, int32_t &UpperBound,
6201                           bool UpperBoundOnly, llvm::Value **CondVal) {
6202   const Stmt *Child = CGOpenMPRuntime::getSingleCompoundChild(
6203       CGF.getContext(), CS->getCapturedStmt());
6204   const auto *Dir = dyn_cast_or_null<OMPExecutableDirective>(Child);
6205   if (!Dir)
6206     return;
6207 
6208   if (isOpenMPParallelDirective(Dir->getDirectiveKind())) {
6209     // Handle if clause. If if clause present, the number of threads is
6210     // calculated as <cond> ? (<numthreads> ? <numthreads> : 0 ) : 1.
6211     if (CondVal && Dir->hasClausesOfKind<OMPIfClause>()) {
6212       CGOpenMPInnerExprInfo CGInfo(CGF, *CS);
6213       CodeGenFunction::CGCapturedStmtRAII CapInfoRAII(CGF, &CGInfo);
6214       const OMPIfClause *IfClause = nullptr;
6215       for (const auto *C : Dir->getClausesOfKind<OMPIfClause>()) {
6216         if (C->getNameModifier() == OMPD_unknown ||
6217             C->getNameModifier() == OMPD_parallel) {
6218           IfClause = C;
6219           break;
6220         }
6221       }
6222       if (IfClause) {
6223         const Expr *CondExpr = IfClause->getCondition();
6224         bool Result;
6225         if (CondExpr->EvaluateAsBooleanCondition(Result, CGF.getContext())) {
6226           if (!Result) {
6227             UpperBound = 1;
6228             return;
6229           }
6230         } else {
6231           CodeGenFunction::LexicalScope Scope(CGF, CondExpr->getSourceRange());
6232           if (const auto *PreInit =
6233                   cast_or_null<DeclStmt>(IfClause->getPreInitStmt())) {
6234             for (const auto *I : PreInit->decls()) {
6235               if (!I->hasAttr<OMPCaptureNoInitAttr>()) {
6236                 CGF.EmitVarDecl(cast<VarDecl>(*I));
6237               } else {
6238                 CodeGenFunction::AutoVarEmission Emission =
6239                     CGF.EmitAutoVarAlloca(cast<VarDecl>(*I));
6240                 CGF.EmitAutoVarCleanups(Emission);
6241               }
6242             }
6243             *CondVal = CGF.EvaluateExprAsBool(CondExpr);
6244           }
6245         }
6246       }
6247     }
6248     // Check the value of num_threads clause iff if clause was not specified
6249     // or is not evaluated to false.
6250     if (Dir->hasClausesOfKind<OMPNumThreadsClause>()) {
6251       CGOpenMPInnerExprInfo CGInfo(CGF, *CS);
6252       CodeGenFunction::CGCapturedStmtRAII CapInfoRAII(CGF, &CGInfo);
6253       const auto *NumThreadsClause =
6254           Dir->getSingleClause<OMPNumThreadsClause>();
6255       const Expr *NTExpr = NumThreadsClause->getNumThreads();
6256       if (NTExpr->isIntegerConstantExpr(CGF.getContext()))
6257         if (auto Constant = NTExpr->getIntegerConstantExpr(CGF.getContext()))
6258           UpperBound =
6259               UpperBound
6260                   ? Constant->getZExtValue()
6261                   : std::min(UpperBound,
6262                              static_cast<int32_t>(Constant->getZExtValue()));
6263       // If we haven't found a upper bound, remember we saw a thread limiting
6264       // clause.
6265       if (UpperBound == -1)
6266         UpperBound = 0;
6267       if (!E)
6268         return;
6269       CodeGenFunction::LexicalScope Scope(CGF, NTExpr->getSourceRange());
6270       if (const auto *PreInit =
6271               cast_or_null<DeclStmt>(NumThreadsClause->getPreInitStmt())) {
6272         for (const auto *I : PreInit->decls()) {
6273           if (!I->hasAttr<OMPCaptureNoInitAttr>()) {
6274             CGF.EmitVarDecl(cast<VarDecl>(*I));
6275           } else {
6276             CodeGenFunction::AutoVarEmission Emission =
6277                 CGF.EmitAutoVarAlloca(cast<VarDecl>(*I));
6278             CGF.EmitAutoVarCleanups(Emission);
6279           }
6280         }
6281       }
6282       *E = NTExpr;
6283     }
6284     return;
6285   }
6286   if (isOpenMPSimdDirective(Dir->getDirectiveKind()))
6287     UpperBound = 1;
6288 }
6289 
getNumThreadsExprForTargetDirective(CodeGenFunction & CGF,const OMPExecutableDirective & D,int32_t & UpperBound,bool UpperBoundOnly,llvm::Value ** CondVal,const Expr ** ThreadLimitExpr)6290 const Expr *CGOpenMPRuntime::getNumThreadsExprForTargetDirective(
6291     CodeGenFunction &CGF, const OMPExecutableDirective &D, int32_t &UpperBound,
6292     bool UpperBoundOnly, llvm::Value **CondVal, const Expr **ThreadLimitExpr) {
6293   assert((!CGF.getLangOpts().OpenMPIsTargetDevice || UpperBoundOnly) &&
6294          "Clauses associated with the teams directive expected to be emitted "
6295          "only for the host!");
6296   OpenMPDirectiveKind DirectiveKind = D.getDirectiveKind();
6297   assert(isOpenMPTargetExecutionDirective(DirectiveKind) &&
6298          "Expected target-based executable directive.");
6299 
6300   const Expr *NT = nullptr;
6301   const Expr **NTPtr = UpperBoundOnly ? nullptr : &NT;
6302 
6303   auto CheckForConstExpr = [&](const Expr *E, const Expr **EPtr) {
6304     if (E->isIntegerConstantExpr(CGF.getContext())) {
6305       if (auto Constant = E->getIntegerConstantExpr(CGF.getContext()))
6306         UpperBound = UpperBound ? Constant->getZExtValue()
6307                                 : std::min(UpperBound,
6308                                            int32_t(Constant->getZExtValue()));
6309     }
6310     // If we haven't found a upper bound, remember we saw a thread limiting
6311     // clause.
6312     if (UpperBound == -1)
6313       UpperBound = 0;
6314     if (EPtr)
6315       *EPtr = E;
6316   };
6317 
6318   auto ReturnSequential = [&]() {
6319     UpperBound = 1;
6320     return NT;
6321   };
6322 
6323   switch (DirectiveKind) {
6324   case OMPD_target: {
6325     const CapturedStmt *CS = D.getInnermostCapturedStmt();
6326     getNumThreads(CGF, CS, NTPtr, UpperBound, UpperBoundOnly, CondVal);
6327     const Stmt *Child = CGOpenMPRuntime::getSingleCompoundChild(
6328         CGF.getContext(), CS->getCapturedStmt());
6329     // TODO: The standard is not clear how to resolve two thread limit clauses,
6330     //       let's pick the teams one if it's present, otherwise the target one.
6331     const auto *ThreadLimitClause = D.getSingleClause<OMPThreadLimitClause>();
6332     if (const auto *Dir = dyn_cast_or_null<OMPExecutableDirective>(Child)) {
6333       if (const auto *TLC = Dir->getSingleClause<OMPThreadLimitClause>()) {
6334         ThreadLimitClause = TLC;
6335         if (ThreadLimitExpr) {
6336           CGOpenMPInnerExprInfo CGInfo(CGF, *CS);
6337           CodeGenFunction::CGCapturedStmtRAII CapInfoRAII(CGF, &CGInfo);
6338           CodeGenFunction::LexicalScope Scope(
6339               CGF, ThreadLimitClause->getThreadLimit()->getSourceRange());
6340           if (const auto *PreInit =
6341                   cast_or_null<DeclStmt>(ThreadLimitClause->getPreInitStmt())) {
6342             for (const auto *I : PreInit->decls()) {
6343               if (!I->hasAttr<OMPCaptureNoInitAttr>()) {
6344                 CGF.EmitVarDecl(cast<VarDecl>(*I));
6345               } else {
6346                 CodeGenFunction::AutoVarEmission Emission =
6347                     CGF.EmitAutoVarAlloca(cast<VarDecl>(*I));
6348                 CGF.EmitAutoVarCleanups(Emission);
6349               }
6350             }
6351           }
6352         }
6353       }
6354     }
6355     if (ThreadLimitClause)
6356       CheckForConstExpr(ThreadLimitClause->getThreadLimit(), ThreadLimitExpr);
6357     if (const auto *Dir = dyn_cast_or_null<OMPExecutableDirective>(Child)) {
6358       if (isOpenMPTeamsDirective(Dir->getDirectiveKind()) &&
6359           !isOpenMPDistributeDirective(Dir->getDirectiveKind())) {
6360         CS = Dir->getInnermostCapturedStmt();
6361         const Stmt *Child = CGOpenMPRuntime::getSingleCompoundChild(
6362             CGF.getContext(), CS->getCapturedStmt());
6363         Dir = dyn_cast_or_null<OMPExecutableDirective>(Child);
6364       }
6365       if (Dir && isOpenMPParallelDirective(Dir->getDirectiveKind())) {
6366         CS = Dir->getInnermostCapturedStmt();
6367         getNumThreads(CGF, CS, NTPtr, UpperBound, UpperBoundOnly, CondVal);
6368       } else if (Dir && isOpenMPSimdDirective(Dir->getDirectiveKind()))
6369         return ReturnSequential();
6370     }
6371     return NT;
6372   }
6373   case OMPD_target_teams: {
6374     if (D.hasClausesOfKind<OMPThreadLimitClause>()) {
6375       CodeGenFunction::RunCleanupsScope ThreadLimitScope(CGF);
6376       const auto *ThreadLimitClause = D.getSingleClause<OMPThreadLimitClause>();
6377       CheckForConstExpr(ThreadLimitClause->getThreadLimit(), ThreadLimitExpr);
6378     }
6379     const CapturedStmt *CS = D.getInnermostCapturedStmt();
6380     getNumThreads(CGF, CS, NTPtr, UpperBound, UpperBoundOnly, CondVal);
6381     const Stmt *Child = CGOpenMPRuntime::getSingleCompoundChild(
6382         CGF.getContext(), CS->getCapturedStmt());
6383     if (const auto *Dir = dyn_cast_or_null<OMPExecutableDirective>(Child)) {
6384       if (Dir->getDirectiveKind() == OMPD_distribute) {
6385         CS = Dir->getInnermostCapturedStmt();
6386         getNumThreads(CGF, CS, NTPtr, UpperBound, UpperBoundOnly, CondVal);
6387       }
6388     }
6389     return NT;
6390   }
6391   case OMPD_target_teams_distribute:
6392     if (D.hasClausesOfKind<OMPThreadLimitClause>()) {
6393       CodeGenFunction::RunCleanupsScope ThreadLimitScope(CGF);
6394       const auto *ThreadLimitClause = D.getSingleClause<OMPThreadLimitClause>();
6395       CheckForConstExpr(ThreadLimitClause->getThreadLimit(), ThreadLimitExpr);
6396     }
6397     getNumThreads(CGF, D.getInnermostCapturedStmt(), NTPtr, UpperBound,
6398                   UpperBoundOnly, CondVal);
6399     return NT;
6400   case OMPD_target_teams_loop:
6401   case OMPD_target_parallel_loop:
6402   case OMPD_target_parallel:
6403   case OMPD_target_parallel_for:
6404   case OMPD_target_parallel_for_simd:
6405   case OMPD_target_teams_distribute_parallel_for:
6406   case OMPD_target_teams_distribute_parallel_for_simd: {
6407     if (CondVal && D.hasClausesOfKind<OMPIfClause>()) {
6408       const OMPIfClause *IfClause = nullptr;
6409       for (const auto *C : D.getClausesOfKind<OMPIfClause>()) {
6410         if (C->getNameModifier() == OMPD_unknown ||
6411             C->getNameModifier() == OMPD_parallel) {
6412           IfClause = C;
6413           break;
6414         }
6415       }
6416       if (IfClause) {
6417         const Expr *Cond = IfClause->getCondition();
6418         bool Result;
6419         if (Cond->EvaluateAsBooleanCondition(Result, CGF.getContext())) {
6420           if (!Result)
6421             return ReturnSequential();
6422         } else {
6423           CodeGenFunction::RunCleanupsScope Scope(CGF);
6424           *CondVal = CGF.EvaluateExprAsBool(Cond);
6425         }
6426       }
6427     }
6428     if (D.hasClausesOfKind<OMPThreadLimitClause>()) {
6429       CodeGenFunction::RunCleanupsScope ThreadLimitScope(CGF);
6430       const auto *ThreadLimitClause = D.getSingleClause<OMPThreadLimitClause>();
6431       CheckForConstExpr(ThreadLimitClause->getThreadLimit(), ThreadLimitExpr);
6432     }
6433     if (D.hasClausesOfKind<OMPNumThreadsClause>()) {
6434       CodeGenFunction::RunCleanupsScope NumThreadsScope(CGF);
6435       const auto *NumThreadsClause = D.getSingleClause<OMPNumThreadsClause>();
6436       CheckForConstExpr(NumThreadsClause->getNumThreads(), nullptr);
6437       return NumThreadsClause->getNumThreads();
6438     }
6439     return NT;
6440   }
6441   case OMPD_target_teams_distribute_simd:
6442   case OMPD_target_simd:
6443     return ReturnSequential();
6444   default:
6445     break;
6446   }
6447   llvm_unreachable("Unsupported directive kind.");
6448 }
6449 
emitNumThreadsForTargetDirective(CodeGenFunction & CGF,const OMPExecutableDirective & D)6450 llvm::Value *CGOpenMPRuntime::emitNumThreadsForTargetDirective(
6451     CodeGenFunction &CGF, const OMPExecutableDirective &D) {
6452   llvm::Value *NumThreadsVal = nullptr;
6453   llvm::Value *CondVal = nullptr;
6454   llvm::Value *ThreadLimitVal = nullptr;
6455   const Expr *ThreadLimitExpr = nullptr;
6456   int32_t UpperBound = -1;
6457 
6458   const Expr *NT = getNumThreadsExprForTargetDirective(
6459       CGF, D, UpperBound, /* UpperBoundOnly */ false, &CondVal,
6460       &ThreadLimitExpr);
6461 
6462   // Thread limit expressions are used below, emit them.
6463   if (ThreadLimitExpr) {
6464     ThreadLimitVal =
6465         CGF.EmitScalarExpr(ThreadLimitExpr, /*IgnoreResultAssign=*/true);
6466     ThreadLimitVal = CGF.Builder.CreateIntCast(ThreadLimitVal, CGF.Int32Ty,
6467                                                /*isSigned=*/false);
6468   }
6469 
6470   // Generate the num teams expression.
6471   if (UpperBound == 1) {
6472     NumThreadsVal = CGF.Builder.getInt32(UpperBound);
6473   } else if (NT) {
6474     NumThreadsVal = CGF.EmitScalarExpr(NT, /*IgnoreResultAssign=*/true);
6475     NumThreadsVal = CGF.Builder.CreateIntCast(NumThreadsVal, CGF.Int32Ty,
6476                                               /*isSigned=*/false);
6477   } else if (ThreadLimitVal) {
6478     // If we do not have a num threads value but a thread limit, replace the
6479     // former with the latter. We know handled the thread limit expression.
6480     NumThreadsVal = ThreadLimitVal;
6481     ThreadLimitVal = nullptr;
6482   } else {
6483     // Default to "0" which means runtime choice.
6484     assert(!ThreadLimitVal && "Default not applicable with thread limit value");
6485     NumThreadsVal = CGF.Builder.getInt32(0);
6486   }
6487 
6488   // Handle if clause. If if clause present, the number of threads is
6489   // calculated as <cond> ? (<numthreads> ? <numthreads> : 0 ) : 1.
6490   if (CondVal) {
6491     CodeGenFunction::RunCleanupsScope Scope(CGF);
6492     NumThreadsVal = CGF.Builder.CreateSelect(CondVal, NumThreadsVal,
6493                                              CGF.Builder.getInt32(1));
6494   }
6495 
6496   // If the thread limit and num teams expression were present, take the
6497   // minimum.
6498   if (ThreadLimitVal) {
6499     NumThreadsVal = CGF.Builder.CreateSelect(
6500         CGF.Builder.CreateICmpULT(ThreadLimitVal, NumThreadsVal),
6501         ThreadLimitVal, NumThreadsVal);
6502   }
6503 
6504   return NumThreadsVal;
6505 }
6506 
6507 namespace {
6508 LLVM_ENABLE_BITMASK_ENUMS_IN_NAMESPACE();
6509 
6510 // Utility to handle information from clauses associated with a given
6511 // construct that use mappable expressions (e.g. 'map' clause, 'to' clause).
6512 // It provides a convenient interface to obtain the information and generate
6513 // code for that information.
6514 class MappableExprsHandler {
6515 public:
6516   /// Get the offset of the OMP_MAP_MEMBER_OF field.
getFlagMemberOffset()6517   static unsigned getFlagMemberOffset() {
6518     unsigned Offset = 0;
6519     for (uint64_t Remain =
6520              static_cast<std::underlying_type_t<OpenMPOffloadMappingFlags>>(
6521                  OpenMPOffloadMappingFlags::OMP_MAP_MEMBER_OF);
6522          !(Remain & 1); Remain = Remain >> 1)
6523       Offset++;
6524     return Offset;
6525   }
6526 
6527   /// Class that holds debugging information for a data mapping to be passed to
6528   /// the runtime library.
6529   class MappingExprInfo {
6530     /// The variable declaration used for the data mapping.
6531     const ValueDecl *MapDecl = nullptr;
6532     /// The original expression used in the map clause, or null if there is
6533     /// none.
6534     const Expr *MapExpr = nullptr;
6535 
6536   public:
MappingExprInfo(const ValueDecl * MapDecl,const Expr * MapExpr=nullptr)6537     MappingExprInfo(const ValueDecl *MapDecl, const Expr *MapExpr = nullptr)
6538         : MapDecl(MapDecl), MapExpr(MapExpr) {}
6539 
getMapDecl() const6540     const ValueDecl *getMapDecl() const { return MapDecl; }
getMapExpr() const6541     const Expr *getMapExpr() const { return MapExpr; }
6542   };
6543 
6544   using DeviceInfoTy = llvm::OpenMPIRBuilder::DeviceInfoTy;
6545   using MapBaseValuesArrayTy = llvm::OpenMPIRBuilder::MapValuesArrayTy;
6546   using MapValuesArrayTy = llvm::OpenMPIRBuilder::MapValuesArrayTy;
6547   using MapFlagsArrayTy = llvm::OpenMPIRBuilder::MapFlagsArrayTy;
6548   using MapDimArrayTy = llvm::OpenMPIRBuilder::MapDimArrayTy;
6549   using MapNonContiguousArrayTy =
6550       llvm::OpenMPIRBuilder::MapNonContiguousArrayTy;
6551   using MapExprsArrayTy = SmallVector<MappingExprInfo, 4>;
6552   using MapValueDeclsArrayTy = SmallVector<const ValueDecl *, 4>;
6553 
6554   /// This structure contains combined information generated for mappable
6555   /// clauses, including base pointers, pointers, sizes, map types, user-defined
6556   /// mappers, and non-contiguous information.
6557   struct MapCombinedInfoTy : llvm::OpenMPIRBuilder::MapInfosTy {
6558     MapExprsArrayTy Exprs;
6559     MapValueDeclsArrayTy Mappers;
6560     MapValueDeclsArrayTy DevicePtrDecls;
6561 
6562     /// Append arrays in \a CurInfo.
append__anon93cce0fb2c11::MappableExprsHandler::MapCombinedInfoTy6563     void append(MapCombinedInfoTy &CurInfo) {
6564       Exprs.append(CurInfo.Exprs.begin(), CurInfo.Exprs.end());
6565       DevicePtrDecls.append(CurInfo.DevicePtrDecls.begin(),
6566                             CurInfo.DevicePtrDecls.end());
6567       Mappers.append(CurInfo.Mappers.begin(), CurInfo.Mappers.end());
6568       llvm::OpenMPIRBuilder::MapInfosTy::append(CurInfo);
6569     }
6570   };
6571 
6572   /// Map between a struct and the its lowest & highest elements which have been
6573   /// mapped.
6574   /// [ValueDecl *] --> {LE(FieldIndex, Pointer),
6575   ///                    HE(FieldIndex, Pointer)}
6576   struct StructRangeInfoTy {
6577     MapCombinedInfoTy PreliminaryMapData;
6578     std::pair<unsigned /*FieldIndex*/, Address /*Pointer*/> LowestElem = {
6579         0, Address::invalid()};
6580     std::pair<unsigned /*FieldIndex*/, Address /*Pointer*/> HighestElem = {
6581         0, Address::invalid()};
6582     Address Base = Address::invalid();
6583     Address LB = Address::invalid();
6584     bool IsArraySection = false;
6585     bool HasCompleteRecord = false;
6586   };
6587 
6588 private:
6589   /// Kind that defines how a device pointer has to be returned.
6590   struct MapInfo {
6591     OMPClauseMappableExprCommon::MappableExprComponentListRef Components;
6592     OpenMPMapClauseKind MapType = OMPC_MAP_unknown;
6593     ArrayRef<OpenMPMapModifierKind> MapModifiers;
6594     ArrayRef<OpenMPMotionModifierKind> MotionModifiers;
6595     bool ReturnDevicePointer = false;
6596     bool IsImplicit = false;
6597     const ValueDecl *Mapper = nullptr;
6598     const Expr *VarRef = nullptr;
6599     bool ForDeviceAddr = false;
6600 
6601     MapInfo() = default;
MapInfo__anon93cce0fb2c11::MappableExprsHandler::MapInfo6602     MapInfo(
6603         OMPClauseMappableExprCommon::MappableExprComponentListRef Components,
6604         OpenMPMapClauseKind MapType,
6605         ArrayRef<OpenMPMapModifierKind> MapModifiers,
6606         ArrayRef<OpenMPMotionModifierKind> MotionModifiers,
6607         bool ReturnDevicePointer, bool IsImplicit,
6608         const ValueDecl *Mapper = nullptr, const Expr *VarRef = nullptr,
6609         bool ForDeviceAddr = false)
6610         : Components(Components), MapType(MapType), MapModifiers(MapModifiers),
6611           MotionModifiers(MotionModifiers),
6612           ReturnDevicePointer(ReturnDevicePointer), IsImplicit(IsImplicit),
6613           Mapper(Mapper), VarRef(VarRef), ForDeviceAddr(ForDeviceAddr) {}
6614   };
6615 
6616   /// If use_device_ptr or use_device_addr is used on a decl which is a struct
6617   /// member and there is no map information about it, then emission of that
6618   /// entry is deferred until the whole struct has been processed.
6619   struct DeferredDevicePtrEntryTy {
6620     const Expr *IE = nullptr;
6621     const ValueDecl *VD = nullptr;
6622     bool ForDeviceAddr = false;
6623 
DeferredDevicePtrEntryTy__anon93cce0fb2c11::MappableExprsHandler::DeferredDevicePtrEntryTy6624     DeferredDevicePtrEntryTy(const Expr *IE, const ValueDecl *VD,
6625                              bool ForDeviceAddr)
6626         : IE(IE), VD(VD), ForDeviceAddr(ForDeviceAddr) {}
6627   };
6628 
6629   /// The target directive from where the mappable clauses were extracted. It
6630   /// is either a executable directive or a user-defined mapper directive.
6631   llvm::PointerUnion<const OMPExecutableDirective *,
6632                      const OMPDeclareMapperDecl *>
6633       CurDir;
6634 
6635   /// Function the directive is being generated for.
6636   CodeGenFunction &CGF;
6637 
6638   /// Set of all first private variables in the current directive.
6639   /// bool data is set to true if the variable is implicitly marked as
6640   /// firstprivate, false otherwise.
6641   llvm::DenseMap<CanonicalDeclPtr<const VarDecl>, bool> FirstPrivateDecls;
6642 
6643   /// Map between device pointer declarations and their expression components.
6644   /// The key value for declarations in 'this' is null.
6645   llvm::DenseMap<
6646       const ValueDecl *,
6647       SmallVector<OMPClauseMappableExprCommon::MappableExprComponentListRef, 4>>
6648       DevPointersMap;
6649 
6650   /// Map between device addr declarations and their expression components.
6651   /// The key value for declarations in 'this' is null.
6652   llvm::DenseMap<
6653       const ValueDecl *,
6654       SmallVector<OMPClauseMappableExprCommon::MappableExprComponentListRef, 4>>
6655       HasDevAddrsMap;
6656 
6657   /// Map between lambda declarations and their map type.
6658   llvm::DenseMap<const ValueDecl *, const OMPMapClause *> LambdasMap;
6659 
getExprTypeSize(const Expr * E) const6660   llvm::Value *getExprTypeSize(const Expr *E) const {
6661     QualType ExprTy = E->getType().getCanonicalType();
6662 
6663     // Calculate the size for array shaping expression.
6664     if (const auto *OAE = dyn_cast<OMPArrayShapingExpr>(E)) {
6665       llvm::Value *Size =
6666           CGF.getTypeSize(OAE->getBase()->getType()->getPointeeType());
6667       for (const Expr *SE : OAE->getDimensions()) {
6668         llvm::Value *Sz = CGF.EmitScalarExpr(SE);
6669         Sz = CGF.EmitScalarConversion(Sz, SE->getType(),
6670                                       CGF.getContext().getSizeType(),
6671                                       SE->getExprLoc());
6672         Size = CGF.Builder.CreateNUWMul(Size, Sz);
6673       }
6674       return Size;
6675     }
6676 
6677     // Reference types are ignored for mapping purposes.
6678     if (const auto *RefTy = ExprTy->getAs<ReferenceType>())
6679       ExprTy = RefTy->getPointeeType().getCanonicalType();
6680 
6681     // Given that an array section is considered a built-in type, we need to
6682     // do the calculation based on the length of the section instead of relying
6683     // on CGF.getTypeSize(E->getType()).
6684     if (const auto *OAE = dyn_cast<ArraySectionExpr>(E)) {
6685       QualType BaseTy = ArraySectionExpr::getBaseOriginalType(
6686                             OAE->getBase()->IgnoreParenImpCasts())
6687                             .getCanonicalType();
6688 
6689       // If there is no length associated with the expression and lower bound is
6690       // not specified too, that means we are using the whole length of the
6691       // base.
6692       if (!OAE->getLength() && OAE->getColonLocFirst().isValid() &&
6693           !OAE->getLowerBound())
6694         return CGF.getTypeSize(BaseTy);
6695 
6696       llvm::Value *ElemSize;
6697       if (const auto *PTy = BaseTy->getAs<PointerType>()) {
6698         ElemSize = CGF.getTypeSize(PTy->getPointeeType().getCanonicalType());
6699       } else {
6700         const auto *ATy = cast<ArrayType>(BaseTy.getTypePtr());
6701         assert(ATy && "Expecting array type if not a pointer type.");
6702         ElemSize = CGF.getTypeSize(ATy->getElementType().getCanonicalType());
6703       }
6704 
6705       // If we don't have a length at this point, that is because we have an
6706       // array section with a single element.
6707       if (!OAE->getLength() && OAE->getColonLocFirst().isInvalid())
6708         return ElemSize;
6709 
6710       if (const Expr *LenExpr = OAE->getLength()) {
6711         llvm::Value *LengthVal = CGF.EmitScalarExpr(LenExpr);
6712         LengthVal = CGF.EmitScalarConversion(LengthVal, LenExpr->getType(),
6713                                              CGF.getContext().getSizeType(),
6714                                              LenExpr->getExprLoc());
6715         return CGF.Builder.CreateNUWMul(LengthVal, ElemSize);
6716       }
6717       assert(!OAE->getLength() && OAE->getColonLocFirst().isValid() &&
6718              OAE->getLowerBound() && "expected array_section[lb:].");
6719       // Size = sizetype - lb * elemtype;
6720       llvm::Value *LengthVal = CGF.getTypeSize(BaseTy);
6721       llvm::Value *LBVal = CGF.EmitScalarExpr(OAE->getLowerBound());
6722       LBVal = CGF.EmitScalarConversion(LBVal, OAE->getLowerBound()->getType(),
6723                                        CGF.getContext().getSizeType(),
6724                                        OAE->getLowerBound()->getExprLoc());
6725       LBVal = CGF.Builder.CreateNUWMul(LBVal, ElemSize);
6726       llvm::Value *Cmp = CGF.Builder.CreateICmpUGT(LengthVal, LBVal);
6727       llvm::Value *TrueVal = CGF.Builder.CreateNUWSub(LengthVal, LBVal);
6728       LengthVal = CGF.Builder.CreateSelect(
6729           Cmp, TrueVal, llvm::ConstantInt::get(CGF.SizeTy, 0));
6730       return LengthVal;
6731     }
6732     return CGF.getTypeSize(ExprTy);
6733   }
6734 
6735   /// Return the corresponding bits for a given map clause modifier. Add
6736   /// a flag marking the map as a pointer if requested. Add a flag marking the
6737   /// map as the first one of a series of maps that relate to the same map
6738   /// expression.
getMapTypeBits(OpenMPMapClauseKind MapType,ArrayRef<OpenMPMapModifierKind> MapModifiers,ArrayRef<OpenMPMotionModifierKind> MotionModifiers,bool IsImplicit,bool AddPtrFlag,bool AddIsTargetParamFlag,bool IsNonContiguous) const6739   OpenMPOffloadMappingFlags getMapTypeBits(
6740       OpenMPMapClauseKind MapType, ArrayRef<OpenMPMapModifierKind> MapModifiers,
6741       ArrayRef<OpenMPMotionModifierKind> MotionModifiers, bool IsImplicit,
6742       bool AddPtrFlag, bool AddIsTargetParamFlag, bool IsNonContiguous) const {
6743     OpenMPOffloadMappingFlags Bits =
6744         IsImplicit ? OpenMPOffloadMappingFlags::OMP_MAP_IMPLICIT
6745                    : OpenMPOffloadMappingFlags::OMP_MAP_NONE;
6746     switch (MapType) {
6747     case OMPC_MAP_alloc:
6748     case OMPC_MAP_release:
6749       // alloc and release is the default behavior in the runtime library,  i.e.
6750       // if we don't pass any bits alloc/release that is what the runtime is
6751       // going to do. Therefore, we don't need to signal anything for these two
6752       // type modifiers.
6753       break;
6754     case OMPC_MAP_to:
6755       Bits |= OpenMPOffloadMappingFlags::OMP_MAP_TO;
6756       break;
6757     case OMPC_MAP_from:
6758       Bits |= OpenMPOffloadMappingFlags::OMP_MAP_FROM;
6759       break;
6760     case OMPC_MAP_tofrom:
6761       Bits |= OpenMPOffloadMappingFlags::OMP_MAP_TO |
6762               OpenMPOffloadMappingFlags::OMP_MAP_FROM;
6763       break;
6764     case OMPC_MAP_delete:
6765       Bits |= OpenMPOffloadMappingFlags::OMP_MAP_DELETE;
6766       break;
6767     case OMPC_MAP_unknown:
6768       llvm_unreachable("Unexpected map type!");
6769     }
6770     if (AddPtrFlag)
6771       Bits |= OpenMPOffloadMappingFlags::OMP_MAP_PTR_AND_OBJ;
6772     if (AddIsTargetParamFlag)
6773       Bits |= OpenMPOffloadMappingFlags::OMP_MAP_TARGET_PARAM;
6774     if (llvm::is_contained(MapModifiers, OMPC_MAP_MODIFIER_always))
6775       Bits |= OpenMPOffloadMappingFlags::OMP_MAP_ALWAYS;
6776     if (llvm::is_contained(MapModifiers, OMPC_MAP_MODIFIER_close))
6777       Bits |= OpenMPOffloadMappingFlags::OMP_MAP_CLOSE;
6778     if (llvm::is_contained(MapModifiers, OMPC_MAP_MODIFIER_present) ||
6779         llvm::is_contained(MotionModifiers, OMPC_MOTION_MODIFIER_present))
6780       Bits |= OpenMPOffloadMappingFlags::OMP_MAP_PRESENT;
6781     if (llvm::is_contained(MapModifiers, OMPC_MAP_MODIFIER_ompx_hold))
6782       Bits |= OpenMPOffloadMappingFlags::OMP_MAP_OMPX_HOLD;
6783     if (IsNonContiguous)
6784       Bits |= OpenMPOffloadMappingFlags::OMP_MAP_NON_CONTIG;
6785     return Bits;
6786   }
6787 
6788   /// Return true if the provided expression is a final array section. A
6789   /// final array section, is one whose length can't be proved to be one.
isFinalArraySectionExpression(const Expr * E) const6790   bool isFinalArraySectionExpression(const Expr *E) const {
6791     const auto *OASE = dyn_cast<ArraySectionExpr>(E);
6792 
6793     // It is not an array section and therefore not a unity-size one.
6794     if (!OASE)
6795       return false;
6796 
6797     // An array section with no colon always refer to a single element.
6798     if (OASE->getColonLocFirst().isInvalid())
6799       return false;
6800 
6801     const Expr *Length = OASE->getLength();
6802 
6803     // If we don't have a length we have to check if the array has size 1
6804     // for this dimension. Also, we should always expect a length if the
6805     // base type is pointer.
6806     if (!Length) {
6807       QualType BaseQTy = ArraySectionExpr::getBaseOriginalType(
6808                              OASE->getBase()->IgnoreParenImpCasts())
6809                              .getCanonicalType();
6810       if (const auto *ATy = dyn_cast<ConstantArrayType>(BaseQTy.getTypePtr()))
6811         return ATy->getSExtSize() != 1;
6812       // If we don't have a constant dimension length, we have to consider
6813       // the current section as having any size, so it is not necessarily
6814       // unitary. If it happen to be unity size, that's user fault.
6815       return true;
6816     }
6817 
6818     // Check if the length evaluates to 1.
6819     Expr::EvalResult Result;
6820     if (!Length->EvaluateAsInt(Result, CGF.getContext()))
6821       return true; // Can have more that size 1.
6822 
6823     llvm::APSInt ConstLength = Result.Val.getInt();
6824     return ConstLength.getSExtValue() != 1;
6825   }
6826 
6827   /// Generate the base pointers, section pointers, sizes, map type bits, and
6828   /// user-defined mappers (all included in \a CombinedInfo) for the provided
6829   /// map type, map or motion modifiers, and expression components.
6830   /// \a IsFirstComponent should be set to true if the provided set of
6831   /// components is the first associated with a capture.
generateInfoForComponentList(OpenMPMapClauseKind MapType,ArrayRef<OpenMPMapModifierKind> MapModifiers,ArrayRef<OpenMPMotionModifierKind> MotionModifiers,OMPClauseMappableExprCommon::MappableExprComponentListRef Components,MapCombinedInfoTy & CombinedInfo,MapCombinedInfoTy & StructBaseCombinedInfo,StructRangeInfoTy & PartialStruct,bool IsFirstComponentList,bool IsImplicit,bool GenerateAllInfoForClauses,const ValueDecl * Mapper=nullptr,bool ForDeviceAddr=false,const ValueDecl * BaseDecl=nullptr,const Expr * MapExpr=nullptr,ArrayRef<OMPClauseMappableExprCommon::MappableExprComponentListRef> OverlappedElements=std::nullopt,bool AreBothBasePtrAndPteeMapped=false) const6832   void generateInfoForComponentList(
6833       OpenMPMapClauseKind MapType, ArrayRef<OpenMPMapModifierKind> MapModifiers,
6834       ArrayRef<OpenMPMotionModifierKind> MotionModifiers,
6835       OMPClauseMappableExprCommon::MappableExprComponentListRef Components,
6836       MapCombinedInfoTy &CombinedInfo,
6837       MapCombinedInfoTy &StructBaseCombinedInfo,
6838       StructRangeInfoTy &PartialStruct, bool IsFirstComponentList,
6839       bool IsImplicit, bool GenerateAllInfoForClauses,
6840       const ValueDecl *Mapper = nullptr, bool ForDeviceAddr = false,
6841       const ValueDecl *BaseDecl = nullptr, const Expr *MapExpr = nullptr,
6842       ArrayRef<OMPClauseMappableExprCommon::MappableExprComponentListRef>
6843           OverlappedElements = std::nullopt,
6844       bool AreBothBasePtrAndPteeMapped = false) const {
6845     // The following summarizes what has to be generated for each map and the
6846     // types below. The generated information is expressed in this order:
6847     // base pointer, section pointer, size, flags
6848     // (to add to the ones that come from the map type and modifier).
6849     //
6850     // double d;
6851     // int i[100];
6852     // float *p;
6853     // int **a = &i;
6854     //
6855     // struct S1 {
6856     //   int i;
6857     //   float f[50];
6858     // }
6859     // struct S2 {
6860     //   int i;
6861     //   float f[50];
6862     //   S1 s;
6863     //   double *p;
6864     //   struct S2 *ps;
6865     //   int &ref;
6866     // }
6867     // S2 s;
6868     // S2 *ps;
6869     //
6870     // map(d)
6871     // &d, &d, sizeof(double), TARGET_PARAM | TO | FROM
6872     //
6873     // map(i)
6874     // &i, &i, 100*sizeof(int), TARGET_PARAM | TO | FROM
6875     //
6876     // map(i[1:23])
6877     // &i(=&i[0]), &i[1], 23*sizeof(int), TARGET_PARAM | TO | FROM
6878     //
6879     // map(p)
6880     // &p, &p, sizeof(float*), TARGET_PARAM | TO | FROM
6881     //
6882     // map(p[1:24])
6883     // &p, &p[1], 24*sizeof(float), TARGET_PARAM | TO | FROM | PTR_AND_OBJ
6884     // in unified shared memory mode or for local pointers
6885     // p, &p[1], 24*sizeof(float), TARGET_PARAM | TO | FROM
6886     //
6887     // map((*a)[0:3])
6888     // &(*a), &(*a), sizeof(pointer), TARGET_PARAM | TO | FROM
6889     // &(*a), &(*a)[0], 3*sizeof(int), PTR_AND_OBJ | TO | FROM
6890     //
6891     // map(**a)
6892     // &(*a), &(*a), sizeof(pointer), TARGET_PARAM | TO | FROM
6893     // &(*a), &(**a), sizeof(int), PTR_AND_OBJ | TO | FROM
6894     //
6895     // map(s)
6896     // &s, &s, sizeof(S2), TARGET_PARAM | TO | FROM
6897     //
6898     // map(s.i)
6899     // &s, &(s.i), sizeof(int), TARGET_PARAM | TO | FROM
6900     //
6901     // map(s.s.f)
6902     // &s, &(s.s.f[0]), 50*sizeof(float), TARGET_PARAM | TO | FROM
6903     //
6904     // map(s.p)
6905     // &s, &(s.p), sizeof(double*), TARGET_PARAM | TO | FROM
6906     //
6907     // map(to: s.p[:22])
6908     // &s, &(s.p), sizeof(double*), TARGET_PARAM (*)
6909     // &s, &(s.p), sizeof(double*), MEMBER_OF(1) (**)
6910     // &(s.p), &(s.p[0]), 22*sizeof(double),
6911     //   MEMBER_OF(1) | PTR_AND_OBJ | TO (***)
6912     // (*) alloc space for struct members, only this is a target parameter
6913     // (**) map the pointer (nothing to be mapped in this example) (the compiler
6914     //      optimizes this entry out, same in the examples below)
6915     // (***) map the pointee (map: to)
6916     //
6917     // map(to: s.ref)
6918     // &s, &(s.ref), sizeof(int*), TARGET_PARAM (*)
6919     // &s, &(s.ref), sizeof(int), MEMBER_OF(1) | PTR_AND_OBJ | TO (***)
6920     // (*) alloc space for struct members, only this is a target parameter
6921     // (**) map the pointer (nothing to be mapped in this example) (the compiler
6922     //      optimizes this entry out, same in the examples below)
6923     // (***) map the pointee (map: to)
6924     //
6925     // map(s.ps)
6926     // &s, &(s.ps), sizeof(S2*), TARGET_PARAM | TO | FROM
6927     //
6928     // map(from: s.ps->s.i)
6929     // &s, &(s.ps), sizeof(S2*), TARGET_PARAM
6930     // &s, &(s.ps), sizeof(S2*), MEMBER_OF(1)
6931     // &(s.ps), &(s.ps->s.i), sizeof(int), MEMBER_OF(1) | PTR_AND_OBJ  | FROM
6932     //
6933     // map(to: s.ps->ps)
6934     // &s, &(s.ps), sizeof(S2*), TARGET_PARAM
6935     // &s, &(s.ps), sizeof(S2*), MEMBER_OF(1)
6936     // &(s.ps), &(s.ps->ps), sizeof(S2*), MEMBER_OF(1) | PTR_AND_OBJ  | TO
6937     //
6938     // map(s.ps->ps->ps)
6939     // &s, &(s.ps), sizeof(S2*), TARGET_PARAM
6940     // &s, &(s.ps), sizeof(S2*), MEMBER_OF(1)
6941     // &(s.ps), &(s.ps->ps), sizeof(S2*), MEMBER_OF(1) | PTR_AND_OBJ
6942     // &(s.ps->ps), &(s.ps->ps->ps), sizeof(S2*), PTR_AND_OBJ | TO | FROM
6943     //
6944     // map(to: s.ps->ps->s.f[:22])
6945     // &s, &(s.ps), sizeof(S2*), TARGET_PARAM
6946     // &s, &(s.ps), sizeof(S2*), MEMBER_OF(1)
6947     // &(s.ps), &(s.ps->ps), sizeof(S2*), MEMBER_OF(1) | PTR_AND_OBJ
6948     // &(s.ps->ps), &(s.ps->ps->s.f[0]), 22*sizeof(float), PTR_AND_OBJ | TO
6949     //
6950     // map(ps)
6951     // &ps, &ps, sizeof(S2*), TARGET_PARAM | TO | FROM
6952     //
6953     // map(ps->i)
6954     // ps, &(ps->i), sizeof(int), TARGET_PARAM | TO | FROM
6955     //
6956     // map(ps->s.f)
6957     // ps, &(ps->s.f[0]), 50*sizeof(float), TARGET_PARAM | TO | FROM
6958     //
6959     // map(from: ps->p)
6960     // ps, &(ps->p), sizeof(double*), TARGET_PARAM | FROM
6961     //
6962     // map(to: ps->p[:22])
6963     // ps, &(ps->p), sizeof(double*), TARGET_PARAM
6964     // ps, &(ps->p), sizeof(double*), MEMBER_OF(1)
6965     // &(ps->p), &(ps->p[0]), 22*sizeof(double), MEMBER_OF(1) | PTR_AND_OBJ | TO
6966     //
6967     // map(ps->ps)
6968     // ps, &(ps->ps), sizeof(S2*), TARGET_PARAM | TO | FROM
6969     //
6970     // map(from: ps->ps->s.i)
6971     // ps, &(ps->ps), sizeof(S2*), TARGET_PARAM
6972     // ps, &(ps->ps), sizeof(S2*), MEMBER_OF(1)
6973     // &(ps->ps), &(ps->ps->s.i), sizeof(int), MEMBER_OF(1) | PTR_AND_OBJ | FROM
6974     //
6975     // map(from: ps->ps->ps)
6976     // ps, &(ps->ps), sizeof(S2*), TARGET_PARAM
6977     // ps, &(ps->ps), sizeof(S2*), MEMBER_OF(1)
6978     // &(ps->ps), &(ps->ps->ps), sizeof(S2*), MEMBER_OF(1) | PTR_AND_OBJ | FROM
6979     //
6980     // map(ps->ps->ps->ps)
6981     // ps, &(ps->ps), sizeof(S2*), TARGET_PARAM
6982     // ps, &(ps->ps), sizeof(S2*), MEMBER_OF(1)
6983     // &(ps->ps), &(ps->ps->ps), sizeof(S2*), MEMBER_OF(1) | PTR_AND_OBJ
6984     // &(ps->ps->ps), &(ps->ps->ps->ps), sizeof(S2*), PTR_AND_OBJ | TO | FROM
6985     //
6986     // map(to: ps->ps->ps->s.f[:22])
6987     // ps, &(ps->ps), sizeof(S2*), TARGET_PARAM
6988     // ps, &(ps->ps), sizeof(S2*), MEMBER_OF(1)
6989     // &(ps->ps), &(ps->ps->ps), sizeof(S2*), MEMBER_OF(1) | PTR_AND_OBJ
6990     // &(ps->ps->ps), &(ps->ps->ps->s.f[0]), 22*sizeof(float), PTR_AND_OBJ | TO
6991     //
6992     // map(to: s.f[:22]) map(from: s.p[:33])
6993     // &s, &(s.f[0]), 50*sizeof(float) + sizeof(struct S1) +
6994     //     sizeof(double*) (**), TARGET_PARAM
6995     // &s, &(s.f[0]), 22*sizeof(float), MEMBER_OF(1) | TO
6996     // &s, &(s.p), sizeof(double*), MEMBER_OF(1)
6997     // &(s.p), &(s.p[0]), 33*sizeof(double), MEMBER_OF(1) | PTR_AND_OBJ | FROM
6998     // (*) allocate contiguous space needed to fit all mapped members even if
6999     //     we allocate space for members not mapped (in this example,
7000     //     s.f[22..49] and s.s are not mapped, yet we must allocate space for
7001     //     them as well because they fall between &s.f[0] and &s.p)
7002     //
7003     // map(from: s.f[:22]) map(to: ps->p[:33])
7004     // &s, &(s.f[0]), 22*sizeof(float), TARGET_PARAM | FROM
7005     // ps, &(ps->p), sizeof(S2*), TARGET_PARAM
7006     // ps, &(ps->p), sizeof(double*), MEMBER_OF(2) (*)
7007     // &(ps->p), &(ps->p[0]), 33*sizeof(double), MEMBER_OF(2) | PTR_AND_OBJ | TO
7008     // (*) the struct this entry pertains to is the 2nd element in the list of
7009     //     arguments, hence MEMBER_OF(2)
7010     //
7011     // map(from: s.f[:22], s.s) map(to: ps->p[:33])
7012     // &s, &(s.f[0]), 50*sizeof(float) + sizeof(struct S1), TARGET_PARAM
7013     // &s, &(s.f[0]), 22*sizeof(float), MEMBER_OF(1) | FROM
7014     // &s, &(s.s), sizeof(struct S1), MEMBER_OF(1) | FROM
7015     // ps, &(ps->p), sizeof(S2*), TARGET_PARAM
7016     // ps, &(ps->p), sizeof(double*), MEMBER_OF(4) (*)
7017     // &(ps->p), &(ps->p[0]), 33*sizeof(double), MEMBER_OF(4) | PTR_AND_OBJ | TO
7018     // (*) the struct this entry pertains to is the 4th element in the list
7019     //     of arguments, hence MEMBER_OF(4)
7020     //
7021     // map(p, p[:100])
7022     // ===> map(p[:100])
7023     // &p, &p[0], 100*sizeof(float), TARGET_PARAM | PTR_AND_OBJ | TO | FROM
7024 
7025     // Track if the map information being generated is the first for a capture.
7026     bool IsCaptureFirstInfo = IsFirstComponentList;
7027     // When the variable is on a declare target link or in a to clause with
7028     // unified memory, a reference is needed to hold the host/device address
7029     // of the variable.
7030     bool RequiresReference = false;
7031 
7032     // Scan the components from the base to the complete expression.
7033     auto CI = Components.rbegin();
7034     auto CE = Components.rend();
7035     auto I = CI;
7036 
7037     // Track if the map information being generated is the first for a list of
7038     // components.
7039     bool IsExpressionFirstInfo = true;
7040     bool FirstPointerInComplexData = false;
7041     Address BP = Address::invalid();
7042     const Expr *AssocExpr = I->getAssociatedExpression();
7043     const auto *AE = dyn_cast<ArraySubscriptExpr>(AssocExpr);
7044     const auto *OASE = dyn_cast<ArraySectionExpr>(AssocExpr);
7045     const auto *OAShE = dyn_cast<OMPArrayShapingExpr>(AssocExpr);
7046 
7047     if (AreBothBasePtrAndPteeMapped && std::next(I) == CE)
7048       return;
7049     if (isa<MemberExpr>(AssocExpr)) {
7050       // The base is the 'this' pointer. The content of the pointer is going
7051       // to be the base of the field being mapped.
7052       BP = CGF.LoadCXXThisAddress();
7053     } else if ((AE && isa<CXXThisExpr>(AE->getBase()->IgnoreParenImpCasts())) ||
7054                (OASE &&
7055                 isa<CXXThisExpr>(OASE->getBase()->IgnoreParenImpCasts()))) {
7056       BP = CGF.EmitOMPSharedLValue(AssocExpr).getAddress();
7057     } else if (OAShE &&
7058                isa<CXXThisExpr>(OAShE->getBase()->IgnoreParenCasts())) {
7059       BP = Address(
7060           CGF.EmitScalarExpr(OAShE->getBase()),
7061           CGF.ConvertTypeForMem(OAShE->getBase()->getType()->getPointeeType()),
7062           CGF.getContext().getTypeAlignInChars(OAShE->getBase()->getType()));
7063     } else {
7064       // The base is the reference to the variable.
7065       // BP = &Var.
7066       BP = CGF.EmitOMPSharedLValue(AssocExpr).getAddress();
7067       if (const auto *VD =
7068               dyn_cast_or_null<VarDecl>(I->getAssociatedDeclaration())) {
7069         if (std::optional<OMPDeclareTargetDeclAttr::MapTypeTy> Res =
7070                 OMPDeclareTargetDeclAttr::isDeclareTargetDeclaration(VD)) {
7071           if ((*Res == OMPDeclareTargetDeclAttr::MT_Link) ||
7072               ((*Res == OMPDeclareTargetDeclAttr::MT_To ||
7073                 *Res == OMPDeclareTargetDeclAttr::MT_Enter) &&
7074                CGF.CGM.getOpenMPRuntime().hasRequiresUnifiedSharedMemory())) {
7075             RequiresReference = true;
7076             BP = CGF.CGM.getOpenMPRuntime().getAddrOfDeclareTargetVar(VD);
7077           }
7078         }
7079       }
7080 
7081       // If the variable is a pointer and is being dereferenced (i.e. is not
7082       // the last component), the base has to be the pointer itself, not its
7083       // reference. References are ignored for mapping purposes.
7084       QualType Ty =
7085           I->getAssociatedDeclaration()->getType().getNonReferenceType();
7086       if (Ty->isAnyPointerType() && std::next(I) != CE) {
7087         // No need to generate individual map information for the pointer, it
7088         // can be associated with the combined storage if shared memory mode is
7089         // active or the base declaration is not global variable.
7090         const auto *VD = dyn_cast<VarDecl>(I->getAssociatedDeclaration());
7091         if (!AreBothBasePtrAndPteeMapped &&
7092             (CGF.CGM.getOpenMPRuntime().hasRequiresUnifiedSharedMemory() ||
7093              !VD || VD->hasLocalStorage()))
7094           BP = CGF.EmitLoadOfPointer(BP, Ty->castAs<PointerType>());
7095         else
7096           FirstPointerInComplexData = true;
7097         ++I;
7098       }
7099     }
7100 
7101     // Track whether a component of the list should be marked as MEMBER_OF some
7102     // combined entry (for partial structs). Only the first PTR_AND_OBJ entry
7103     // in a component list should be marked as MEMBER_OF, all subsequent entries
7104     // do not belong to the base struct. E.g.
7105     // struct S2 s;
7106     // s.ps->ps->ps->f[:]
7107     //   (1) (2) (3) (4)
7108     // ps(1) is a member pointer, ps(2) is a pointee of ps(1), so it is a
7109     // PTR_AND_OBJ entry; the PTR is ps(1), so MEMBER_OF the base struct. ps(3)
7110     // is the pointee of ps(2) which is not member of struct s, so it should not
7111     // be marked as such (it is still PTR_AND_OBJ).
7112     // The variable is initialized to false so that PTR_AND_OBJ entries which
7113     // are not struct members are not considered (e.g. array of pointers to
7114     // data).
7115     bool ShouldBeMemberOf = false;
7116 
7117     // Variable keeping track of whether or not we have encountered a component
7118     // in the component list which is a member expression. Useful when we have a
7119     // pointer or a final array section, in which case it is the previous
7120     // component in the list which tells us whether we have a member expression.
7121     // E.g. X.f[:]
7122     // While processing the final array section "[:]" it is "f" which tells us
7123     // whether we are dealing with a member of a declared struct.
7124     const MemberExpr *EncounteredME = nullptr;
7125 
7126     // Track for the total number of dimension. Start from one for the dummy
7127     // dimension.
7128     uint64_t DimSize = 1;
7129 
7130     bool IsNonContiguous = CombinedInfo.NonContigInfo.IsNonContiguous;
7131     bool IsPrevMemberReference = false;
7132 
7133     // We need to check if we will be encountering any MEs. If we do not
7134     // encounter any ME expression it means we will be mapping the whole struct.
7135     // In that case we need to skip adding an entry for the struct to the
7136     // CombinedInfo list and instead add an entry to the StructBaseCombinedInfo
7137     // list only when generating all info for clauses.
7138     bool IsMappingWholeStruct = true;
7139     if (!GenerateAllInfoForClauses) {
7140       IsMappingWholeStruct = false;
7141     } else {
7142       for (auto TempI = I; TempI != CE; ++TempI) {
7143         const MemberExpr *PossibleME =
7144             dyn_cast<MemberExpr>(TempI->getAssociatedExpression());
7145         if (PossibleME) {
7146           IsMappingWholeStruct = false;
7147           break;
7148         }
7149       }
7150     }
7151 
7152     for (; I != CE; ++I) {
7153       // If the current component is member of a struct (parent struct) mark it.
7154       if (!EncounteredME) {
7155         EncounteredME = dyn_cast<MemberExpr>(I->getAssociatedExpression());
7156         // If we encounter a PTR_AND_OBJ entry from now on it should be marked
7157         // as MEMBER_OF the parent struct.
7158         if (EncounteredME) {
7159           ShouldBeMemberOf = true;
7160           // Do not emit as complex pointer if this is actually not array-like
7161           // expression.
7162           if (FirstPointerInComplexData) {
7163             QualType Ty = std::prev(I)
7164                               ->getAssociatedDeclaration()
7165                               ->getType()
7166                               .getNonReferenceType();
7167             BP = CGF.EmitLoadOfPointer(BP, Ty->castAs<PointerType>());
7168             FirstPointerInComplexData = false;
7169           }
7170         }
7171       }
7172 
7173       auto Next = std::next(I);
7174 
7175       // We need to generate the addresses and sizes if this is the last
7176       // component, if the component is a pointer or if it is an array section
7177       // whose length can't be proved to be one. If this is a pointer, it
7178       // becomes the base address for the following components.
7179 
7180       // A final array section, is one whose length can't be proved to be one.
7181       // If the map item is non-contiguous then we don't treat any array section
7182       // as final array section.
7183       bool IsFinalArraySection =
7184           !IsNonContiguous &&
7185           isFinalArraySectionExpression(I->getAssociatedExpression());
7186 
7187       // If we have a declaration for the mapping use that, otherwise use
7188       // the base declaration of the map clause.
7189       const ValueDecl *MapDecl = (I->getAssociatedDeclaration())
7190                                      ? I->getAssociatedDeclaration()
7191                                      : BaseDecl;
7192       MapExpr = (I->getAssociatedExpression()) ? I->getAssociatedExpression()
7193                                                : MapExpr;
7194 
7195       // Get information on whether the element is a pointer. Have to do a
7196       // special treatment for array sections given that they are built-in
7197       // types.
7198       const auto *OASE =
7199           dyn_cast<ArraySectionExpr>(I->getAssociatedExpression());
7200       const auto *OAShE =
7201           dyn_cast<OMPArrayShapingExpr>(I->getAssociatedExpression());
7202       const auto *UO = dyn_cast<UnaryOperator>(I->getAssociatedExpression());
7203       const auto *BO = dyn_cast<BinaryOperator>(I->getAssociatedExpression());
7204       bool IsPointer =
7205           OAShE ||
7206           (OASE && ArraySectionExpr::getBaseOriginalType(OASE)
7207                        .getCanonicalType()
7208                        ->isAnyPointerType()) ||
7209           I->getAssociatedExpression()->getType()->isAnyPointerType();
7210       bool IsMemberReference = isa<MemberExpr>(I->getAssociatedExpression()) &&
7211                                MapDecl &&
7212                                MapDecl->getType()->isLValueReferenceType();
7213       bool IsNonDerefPointer = IsPointer &&
7214                                !(UO && UO->getOpcode() != UO_Deref) && !BO &&
7215                                !IsNonContiguous;
7216 
7217       if (OASE)
7218         ++DimSize;
7219 
7220       if (Next == CE || IsMemberReference || IsNonDerefPointer ||
7221           IsFinalArraySection) {
7222         // If this is not the last component, we expect the pointer to be
7223         // associated with an array expression or member expression.
7224         assert((Next == CE ||
7225                 isa<MemberExpr>(Next->getAssociatedExpression()) ||
7226                 isa<ArraySubscriptExpr>(Next->getAssociatedExpression()) ||
7227                 isa<ArraySectionExpr>(Next->getAssociatedExpression()) ||
7228                 isa<OMPArrayShapingExpr>(Next->getAssociatedExpression()) ||
7229                 isa<UnaryOperator>(Next->getAssociatedExpression()) ||
7230                 isa<BinaryOperator>(Next->getAssociatedExpression())) &&
7231                "Unexpected expression");
7232 
7233         Address LB = Address::invalid();
7234         Address LowestElem = Address::invalid();
7235         auto &&EmitMemberExprBase = [](CodeGenFunction &CGF,
7236                                        const MemberExpr *E) {
7237           const Expr *BaseExpr = E->getBase();
7238           // If this is s.x, emit s as an lvalue.  If it is s->x, emit s as a
7239           // scalar.
7240           LValue BaseLV;
7241           if (E->isArrow()) {
7242             LValueBaseInfo BaseInfo;
7243             TBAAAccessInfo TBAAInfo;
7244             Address Addr =
7245                 CGF.EmitPointerWithAlignment(BaseExpr, &BaseInfo, &TBAAInfo);
7246             QualType PtrTy = BaseExpr->getType()->getPointeeType();
7247             BaseLV = CGF.MakeAddrLValue(Addr, PtrTy, BaseInfo, TBAAInfo);
7248           } else {
7249             BaseLV = CGF.EmitOMPSharedLValue(BaseExpr);
7250           }
7251           return BaseLV;
7252         };
7253         if (OAShE) {
7254           LowestElem = LB =
7255               Address(CGF.EmitScalarExpr(OAShE->getBase()),
7256                       CGF.ConvertTypeForMem(
7257                           OAShE->getBase()->getType()->getPointeeType()),
7258                       CGF.getContext().getTypeAlignInChars(
7259                           OAShE->getBase()->getType()));
7260         } else if (IsMemberReference) {
7261           const auto *ME = cast<MemberExpr>(I->getAssociatedExpression());
7262           LValue BaseLVal = EmitMemberExprBase(CGF, ME);
7263           LowestElem = CGF.EmitLValueForFieldInitialization(
7264                               BaseLVal, cast<FieldDecl>(MapDecl))
7265                            .getAddress();
7266           LB = CGF.EmitLoadOfReferenceLValue(LowestElem, MapDecl->getType())
7267                    .getAddress();
7268         } else {
7269           LowestElem = LB =
7270               CGF.EmitOMPSharedLValue(I->getAssociatedExpression())
7271                   .getAddress();
7272         }
7273 
7274         // If this component is a pointer inside the base struct then we don't
7275         // need to create any entry for it - it will be combined with the object
7276         // it is pointing to into a single PTR_AND_OBJ entry.
7277         bool IsMemberPointerOrAddr =
7278             EncounteredME &&
7279             (((IsPointer || ForDeviceAddr) &&
7280               I->getAssociatedExpression() == EncounteredME) ||
7281              (IsPrevMemberReference && !IsPointer) ||
7282              (IsMemberReference && Next != CE &&
7283               !Next->getAssociatedExpression()->getType()->isPointerType()));
7284         if (!OverlappedElements.empty() && Next == CE) {
7285           // Handle base element with the info for overlapped elements.
7286           assert(!PartialStruct.Base.isValid() && "The base element is set.");
7287           assert(!IsPointer &&
7288                  "Unexpected base element with the pointer type.");
7289           // Mark the whole struct as the struct that requires allocation on the
7290           // device.
7291           PartialStruct.LowestElem = {0, LowestElem};
7292           CharUnits TypeSize = CGF.getContext().getTypeSizeInChars(
7293               I->getAssociatedExpression()->getType());
7294           Address HB = CGF.Builder.CreateConstGEP(
7295               CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
7296                   LowestElem, CGF.VoidPtrTy, CGF.Int8Ty),
7297               TypeSize.getQuantity() - 1);
7298           PartialStruct.HighestElem = {
7299               std::numeric_limits<decltype(
7300                   PartialStruct.HighestElem.first)>::max(),
7301               HB};
7302           PartialStruct.Base = BP;
7303           PartialStruct.LB = LB;
7304           assert(
7305               PartialStruct.PreliminaryMapData.BasePointers.empty() &&
7306               "Overlapped elements must be used only once for the variable.");
7307           std::swap(PartialStruct.PreliminaryMapData, CombinedInfo);
7308           // Emit data for non-overlapped data.
7309           OpenMPOffloadMappingFlags Flags =
7310               OpenMPOffloadMappingFlags::OMP_MAP_MEMBER_OF |
7311               getMapTypeBits(MapType, MapModifiers, MotionModifiers, IsImplicit,
7312                              /*AddPtrFlag=*/false,
7313                              /*AddIsTargetParamFlag=*/false, IsNonContiguous);
7314           llvm::Value *Size = nullptr;
7315           // Do bitcopy of all non-overlapped structure elements.
7316           for (OMPClauseMappableExprCommon::MappableExprComponentListRef
7317                    Component : OverlappedElements) {
7318             Address ComponentLB = Address::invalid();
7319             for (const OMPClauseMappableExprCommon::MappableComponent &MC :
7320                  Component) {
7321               if (const ValueDecl *VD = MC.getAssociatedDeclaration()) {
7322                 const auto *FD = dyn_cast<FieldDecl>(VD);
7323                 if (FD && FD->getType()->isLValueReferenceType()) {
7324                   const auto *ME =
7325                       cast<MemberExpr>(MC.getAssociatedExpression());
7326                   LValue BaseLVal = EmitMemberExprBase(CGF, ME);
7327                   ComponentLB =
7328                       CGF.EmitLValueForFieldInitialization(BaseLVal, FD)
7329                           .getAddress();
7330                 } else {
7331                   ComponentLB =
7332                       CGF.EmitOMPSharedLValue(MC.getAssociatedExpression())
7333                           .getAddress();
7334                 }
7335                 llvm::Value *ComponentLBPtr = ComponentLB.emitRawPointer(CGF);
7336                 llvm::Value *LBPtr = LB.emitRawPointer(CGF);
7337                 Size = CGF.Builder.CreatePtrDiff(CGF.Int8Ty, ComponentLBPtr,
7338                                                  LBPtr);
7339                 break;
7340               }
7341             }
7342             assert(Size && "Failed to determine structure size");
7343             CombinedInfo.Exprs.emplace_back(MapDecl, MapExpr);
7344             CombinedInfo.BasePointers.push_back(BP.emitRawPointer(CGF));
7345             CombinedInfo.DevicePtrDecls.push_back(nullptr);
7346             CombinedInfo.DevicePointers.push_back(DeviceInfoTy::None);
7347             CombinedInfo.Pointers.push_back(LB.emitRawPointer(CGF));
7348             CombinedInfo.Sizes.push_back(CGF.Builder.CreateIntCast(
7349                 Size, CGF.Int64Ty, /*isSigned=*/true));
7350             CombinedInfo.Types.push_back(Flags);
7351             CombinedInfo.Mappers.push_back(nullptr);
7352             CombinedInfo.NonContigInfo.Dims.push_back(IsNonContiguous ? DimSize
7353                                                                       : 1);
7354             LB = CGF.Builder.CreateConstGEP(ComponentLB, 1);
7355           }
7356           CombinedInfo.Exprs.emplace_back(MapDecl, MapExpr);
7357           CombinedInfo.BasePointers.push_back(BP.emitRawPointer(CGF));
7358           CombinedInfo.DevicePtrDecls.push_back(nullptr);
7359           CombinedInfo.DevicePointers.push_back(DeviceInfoTy::None);
7360           CombinedInfo.Pointers.push_back(LB.emitRawPointer(CGF));
7361           llvm::Value *LBPtr = LB.emitRawPointer(CGF);
7362           Size = CGF.Builder.CreatePtrDiff(
7363               CGF.Int8Ty, CGF.Builder.CreateConstGEP(HB, 1).emitRawPointer(CGF),
7364               LBPtr);
7365           CombinedInfo.Sizes.push_back(
7366               CGF.Builder.CreateIntCast(Size, CGF.Int64Ty, /*isSigned=*/true));
7367           CombinedInfo.Types.push_back(Flags);
7368           CombinedInfo.Mappers.push_back(nullptr);
7369           CombinedInfo.NonContigInfo.Dims.push_back(IsNonContiguous ? DimSize
7370                                                                     : 1);
7371           break;
7372         }
7373         llvm::Value *Size = getExprTypeSize(I->getAssociatedExpression());
7374         // Skip adding an entry in the CurInfo of this combined entry if the
7375         // whole struct is currently being mapped. The struct needs to be added
7376         // in the first position before any data internal to the struct is being
7377         // mapped.
7378         if (!IsMemberPointerOrAddr ||
7379             (Next == CE && MapType != OMPC_MAP_unknown)) {
7380           if (!IsMappingWholeStruct) {
7381             CombinedInfo.Exprs.emplace_back(MapDecl, MapExpr);
7382             CombinedInfo.BasePointers.push_back(BP.emitRawPointer(CGF));
7383             CombinedInfo.DevicePtrDecls.push_back(nullptr);
7384             CombinedInfo.DevicePointers.push_back(DeviceInfoTy::None);
7385             CombinedInfo.Pointers.push_back(LB.emitRawPointer(CGF));
7386             CombinedInfo.Sizes.push_back(CGF.Builder.CreateIntCast(
7387                 Size, CGF.Int64Ty, /*isSigned=*/true));
7388             CombinedInfo.NonContigInfo.Dims.push_back(IsNonContiguous ? DimSize
7389                                                                       : 1);
7390           } else {
7391             StructBaseCombinedInfo.Exprs.emplace_back(MapDecl, MapExpr);
7392             StructBaseCombinedInfo.BasePointers.push_back(
7393                 BP.emitRawPointer(CGF));
7394             StructBaseCombinedInfo.DevicePtrDecls.push_back(nullptr);
7395             StructBaseCombinedInfo.DevicePointers.push_back(DeviceInfoTy::None);
7396             StructBaseCombinedInfo.Pointers.push_back(LB.emitRawPointer(CGF));
7397             StructBaseCombinedInfo.Sizes.push_back(CGF.Builder.CreateIntCast(
7398                 Size, CGF.Int64Ty, /*isSigned=*/true));
7399             StructBaseCombinedInfo.NonContigInfo.Dims.push_back(
7400                 IsNonContiguous ? DimSize : 1);
7401           }
7402 
7403           // If Mapper is valid, the last component inherits the mapper.
7404           bool HasMapper = Mapper && Next == CE;
7405           if (!IsMappingWholeStruct)
7406             CombinedInfo.Mappers.push_back(HasMapper ? Mapper : nullptr);
7407           else
7408             StructBaseCombinedInfo.Mappers.push_back(HasMapper ? Mapper
7409                                                                : nullptr);
7410 
7411           // We need to add a pointer flag for each map that comes from the
7412           // same expression except for the first one. We also need to signal
7413           // this map is the first one that relates with the current capture
7414           // (there is a set of entries for each capture).
7415           OpenMPOffloadMappingFlags Flags =
7416               getMapTypeBits(MapType, MapModifiers, MotionModifiers, IsImplicit,
7417                              !IsExpressionFirstInfo || RequiresReference ||
7418                                  FirstPointerInComplexData || IsMemberReference,
7419                              AreBothBasePtrAndPteeMapped ||
7420                                  (IsCaptureFirstInfo && !RequiresReference),
7421                              IsNonContiguous);
7422 
7423           if (!IsExpressionFirstInfo || IsMemberReference) {
7424             // If we have a PTR_AND_OBJ pair where the OBJ is a pointer as well,
7425             // then we reset the TO/FROM/ALWAYS/DELETE/CLOSE flags.
7426             if (IsPointer || (IsMemberReference && Next != CE))
7427               Flags &= ~(OpenMPOffloadMappingFlags::OMP_MAP_TO |
7428                          OpenMPOffloadMappingFlags::OMP_MAP_FROM |
7429                          OpenMPOffloadMappingFlags::OMP_MAP_ALWAYS |
7430                          OpenMPOffloadMappingFlags::OMP_MAP_DELETE |
7431                          OpenMPOffloadMappingFlags::OMP_MAP_CLOSE);
7432 
7433             if (ShouldBeMemberOf) {
7434               // Set placeholder value MEMBER_OF=FFFF to indicate that the flag
7435               // should be later updated with the correct value of MEMBER_OF.
7436               Flags |= OpenMPOffloadMappingFlags::OMP_MAP_MEMBER_OF;
7437               // From now on, all subsequent PTR_AND_OBJ entries should not be
7438               // marked as MEMBER_OF.
7439               ShouldBeMemberOf = false;
7440             }
7441           }
7442 
7443           if (!IsMappingWholeStruct)
7444             CombinedInfo.Types.push_back(Flags);
7445           else
7446             StructBaseCombinedInfo.Types.push_back(Flags);
7447         }
7448 
7449         // If we have encountered a member expression so far, keep track of the
7450         // mapped member. If the parent is "*this", then the value declaration
7451         // is nullptr.
7452         if (EncounteredME) {
7453           const auto *FD = cast<FieldDecl>(EncounteredME->getMemberDecl());
7454           unsigned FieldIndex = FD->getFieldIndex();
7455 
7456           // Update info about the lowest and highest elements for this struct
7457           if (!PartialStruct.Base.isValid()) {
7458             PartialStruct.LowestElem = {FieldIndex, LowestElem};
7459             if (IsFinalArraySection) {
7460               Address HB =
7461                   CGF.EmitArraySectionExpr(OASE, /*IsLowerBound=*/false)
7462                       .getAddress();
7463               PartialStruct.HighestElem = {FieldIndex, HB};
7464             } else {
7465               PartialStruct.HighestElem = {FieldIndex, LowestElem};
7466             }
7467             PartialStruct.Base = BP;
7468             PartialStruct.LB = BP;
7469           } else if (FieldIndex < PartialStruct.LowestElem.first) {
7470             PartialStruct.LowestElem = {FieldIndex, LowestElem};
7471           } else if (FieldIndex > PartialStruct.HighestElem.first) {
7472             if (IsFinalArraySection) {
7473               Address HB =
7474                   CGF.EmitArraySectionExpr(OASE, /*IsLowerBound=*/false)
7475                       .getAddress();
7476               PartialStruct.HighestElem = {FieldIndex, HB};
7477             } else {
7478               PartialStruct.HighestElem = {FieldIndex, LowestElem};
7479             }
7480           }
7481         }
7482 
7483         // Need to emit combined struct for array sections.
7484         if (IsFinalArraySection || IsNonContiguous)
7485           PartialStruct.IsArraySection = true;
7486 
7487         // If we have a final array section, we are done with this expression.
7488         if (IsFinalArraySection)
7489           break;
7490 
7491         // The pointer becomes the base for the next element.
7492         if (Next != CE)
7493           BP = IsMemberReference ? LowestElem : LB;
7494 
7495         IsExpressionFirstInfo = false;
7496         IsCaptureFirstInfo = false;
7497         FirstPointerInComplexData = false;
7498         IsPrevMemberReference = IsMemberReference;
7499       } else if (FirstPointerInComplexData) {
7500         QualType Ty = Components.rbegin()
7501                           ->getAssociatedDeclaration()
7502                           ->getType()
7503                           .getNonReferenceType();
7504         BP = CGF.EmitLoadOfPointer(BP, Ty->castAs<PointerType>());
7505         FirstPointerInComplexData = false;
7506       }
7507     }
7508     // If ran into the whole component - allocate the space for the whole
7509     // record.
7510     if (!EncounteredME)
7511       PartialStruct.HasCompleteRecord = true;
7512 
7513     if (!IsNonContiguous)
7514       return;
7515 
7516     const ASTContext &Context = CGF.getContext();
7517 
7518     // For supporting stride in array section, we need to initialize the first
7519     // dimension size as 1, first offset as 0, and first count as 1
7520     MapValuesArrayTy CurOffsets = {llvm::ConstantInt::get(CGF.CGM.Int64Ty, 0)};
7521     MapValuesArrayTy CurCounts = {llvm::ConstantInt::get(CGF.CGM.Int64Ty, 1)};
7522     MapValuesArrayTy CurStrides;
7523     MapValuesArrayTy DimSizes{llvm::ConstantInt::get(CGF.CGM.Int64Ty, 1)};
7524     uint64_t ElementTypeSize;
7525 
7526     // Collect Size information for each dimension and get the element size as
7527     // the first Stride. For example, for `int arr[10][10]`, the DimSizes
7528     // should be [10, 10] and the first stride is 4 btyes.
7529     for (const OMPClauseMappableExprCommon::MappableComponent &Component :
7530          Components) {
7531       const Expr *AssocExpr = Component.getAssociatedExpression();
7532       const auto *OASE = dyn_cast<ArraySectionExpr>(AssocExpr);
7533 
7534       if (!OASE)
7535         continue;
7536 
7537       QualType Ty = ArraySectionExpr::getBaseOriginalType(OASE->getBase());
7538       auto *CAT = Context.getAsConstantArrayType(Ty);
7539       auto *VAT = Context.getAsVariableArrayType(Ty);
7540 
7541       // We need all the dimension size except for the last dimension.
7542       assert((VAT || CAT || &Component == &*Components.begin()) &&
7543              "Should be either ConstantArray or VariableArray if not the "
7544              "first Component");
7545 
7546       // Get element size if CurStrides is empty.
7547       if (CurStrides.empty()) {
7548         const Type *ElementType = nullptr;
7549         if (CAT)
7550           ElementType = CAT->getElementType().getTypePtr();
7551         else if (VAT)
7552           ElementType = VAT->getElementType().getTypePtr();
7553         else
7554           assert(&Component == &*Components.begin() &&
7555                  "Only expect pointer (non CAT or VAT) when this is the "
7556                  "first Component");
7557         // If ElementType is null, then it means the base is a pointer
7558         // (neither CAT nor VAT) and we'll attempt to get ElementType again
7559         // for next iteration.
7560         if (ElementType) {
7561           // For the case that having pointer as base, we need to remove one
7562           // level of indirection.
7563           if (&Component != &*Components.begin())
7564             ElementType = ElementType->getPointeeOrArrayElementType();
7565           ElementTypeSize =
7566               Context.getTypeSizeInChars(ElementType).getQuantity();
7567           CurStrides.push_back(
7568               llvm::ConstantInt::get(CGF.Int64Ty, ElementTypeSize));
7569         }
7570       }
7571       // Get dimension value except for the last dimension since we don't need
7572       // it.
7573       if (DimSizes.size() < Components.size() - 1) {
7574         if (CAT)
7575           DimSizes.push_back(
7576               llvm::ConstantInt::get(CGF.Int64Ty, CAT->getZExtSize()));
7577         else if (VAT)
7578           DimSizes.push_back(CGF.Builder.CreateIntCast(
7579               CGF.EmitScalarExpr(VAT->getSizeExpr()), CGF.Int64Ty,
7580               /*IsSigned=*/false));
7581       }
7582     }
7583 
7584     // Skip the dummy dimension since we have already have its information.
7585     auto *DI = DimSizes.begin() + 1;
7586     // Product of dimension.
7587     llvm::Value *DimProd =
7588         llvm::ConstantInt::get(CGF.CGM.Int64Ty, ElementTypeSize);
7589 
7590     // Collect info for non-contiguous. Notice that offset, count, and stride
7591     // are only meaningful for array-section, so we insert a null for anything
7592     // other than array-section.
7593     // Also, the size of offset, count, and stride are not the same as
7594     // pointers, base_pointers, sizes, or dims. Instead, the size of offset,
7595     // count, and stride are the same as the number of non-contiguous
7596     // declaration in target update to/from clause.
7597     for (const OMPClauseMappableExprCommon::MappableComponent &Component :
7598          Components) {
7599       const Expr *AssocExpr = Component.getAssociatedExpression();
7600 
7601       if (const auto *AE = dyn_cast<ArraySubscriptExpr>(AssocExpr)) {
7602         llvm::Value *Offset = CGF.Builder.CreateIntCast(
7603             CGF.EmitScalarExpr(AE->getIdx()), CGF.Int64Ty,
7604             /*isSigned=*/false);
7605         CurOffsets.push_back(Offset);
7606         CurCounts.push_back(llvm::ConstantInt::get(CGF.Int64Ty, /*V=*/1));
7607         CurStrides.push_back(CurStrides.back());
7608         continue;
7609       }
7610 
7611       const auto *OASE = dyn_cast<ArraySectionExpr>(AssocExpr);
7612 
7613       if (!OASE)
7614         continue;
7615 
7616       // Offset
7617       const Expr *OffsetExpr = OASE->getLowerBound();
7618       llvm::Value *Offset = nullptr;
7619       if (!OffsetExpr) {
7620         // If offset is absent, then we just set it to zero.
7621         Offset = llvm::ConstantInt::get(CGF.Int64Ty, 0);
7622       } else {
7623         Offset = CGF.Builder.CreateIntCast(CGF.EmitScalarExpr(OffsetExpr),
7624                                            CGF.Int64Ty,
7625                                            /*isSigned=*/false);
7626       }
7627       CurOffsets.push_back(Offset);
7628 
7629       // Count
7630       const Expr *CountExpr = OASE->getLength();
7631       llvm::Value *Count = nullptr;
7632       if (!CountExpr) {
7633         // In Clang, once a high dimension is an array section, we construct all
7634         // the lower dimension as array section, however, for case like
7635         // arr[0:2][2], Clang construct the inner dimension as an array section
7636         // but it actually is not in an array section form according to spec.
7637         if (!OASE->getColonLocFirst().isValid() &&
7638             !OASE->getColonLocSecond().isValid()) {
7639           Count = llvm::ConstantInt::get(CGF.Int64Ty, 1);
7640         } else {
7641           // OpenMP 5.0, 2.1.5 Array Sections, Description.
7642           // When the length is absent it defaults to ⌈(size −
7643           // lower-bound)/stride⌉, where size is the size of the array
7644           // dimension.
7645           const Expr *StrideExpr = OASE->getStride();
7646           llvm::Value *Stride =
7647               StrideExpr
7648                   ? CGF.Builder.CreateIntCast(CGF.EmitScalarExpr(StrideExpr),
7649                                               CGF.Int64Ty, /*isSigned=*/false)
7650                   : nullptr;
7651           if (Stride)
7652             Count = CGF.Builder.CreateUDiv(
7653                 CGF.Builder.CreateNUWSub(*DI, Offset), Stride);
7654           else
7655             Count = CGF.Builder.CreateNUWSub(*DI, Offset);
7656         }
7657       } else {
7658         Count = CGF.EmitScalarExpr(CountExpr);
7659       }
7660       Count = CGF.Builder.CreateIntCast(Count, CGF.Int64Ty, /*isSigned=*/false);
7661       CurCounts.push_back(Count);
7662 
7663       // Stride_n' = Stride_n * (D_0 * D_1 ... * D_n-1) * Unit size
7664       // Take `int arr[5][5][5]` and `arr[0:2:2][1:2:1][0:2:2]` as an example:
7665       //              Offset      Count     Stride
7666       //    D0          0           1         4    (int)    <- dummy dimension
7667       //    D1          0           2         8    (2 * (1) * 4)
7668       //    D2          1           2         20   (1 * (1 * 5) * 4)
7669       //    D3          0           2         200  (2 * (1 * 5 * 4) * 4)
7670       const Expr *StrideExpr = OASE->getStride();
7671       llvm::Value *Stride =
7672           StrideExpr
7673               ? CGF.Builder.CreateIntCast(CGF.EmitScalarExpr(StrideExpr),
7674                                           CGF.Int64Ty, /*isSigned=*/false)
7675               : nullptr;
7676       DimProd = CGF.Builder.CreateNUWMul(DimProd, *(DI - 1));
7677       if (Stride)
7678         CurStrides.push_back(CGF.Builder.CreateNUWMul(DimProd, Stride));
7679       else
7680         CurStrides.push_back(DimProd);
7681       if (DI != DimSizes.end())
7682         ++DI;
7683     }
7684 
7685     CombinedInfo.NonContigInfo.Offsets.push_back(CurOffsets);
7686     CombinedInfo.NonContigInfo.Counts.push_back(CurCounts);
7687     CombinedInfo.NonContigInfo.Strides.push_back(CurStrides);
7688   }
7689 
7690   /// Return the adjusted map modifiers if the declaration a capture refers to
7691   /// appears in a first-private clause. This is expected to be used only with
7692   /// directives that start with 'target'.
7693   OpenMPOffloadMappingFlags
getMapModifiersForPrivateClauses(const CapturedStmt::Capture & Cap) const7694   getMapModifiersForPrivateClauses(const CapturedStmt::Capture &Cap) const {
7695     assert(Cap.capturesVariable() && "Expected capture by reference only!");
7696 
7697     // A first private variable captured by reference will use only the
7698     // 'private ptr' and 'map to' flag. Return the right flags if the captured
7699     // declaration is known as first-private in this handler.
7700     if (FirstPrivateDecls.count(Cap.getCapturedVar())) {
7701       if (Cap.getCapturedVar()->getType()->isAnyPointerType())
7702         return OpenMPOffloadMappingFlags::OMP_MAP_TO |
7703                OpenMPOffloadMappingFlags::OMP_MAP_PTR_AND_OBJ;
7704       return OpenMPOffloadMappingFlags::OMP_MAP_PRIVATE |
7705              OpenMPOffloadMappingFlags::OMP_MAP_TO;
7706     }
7707     auto I = LambdasMap.find(Cap.getCapturedVar()->getCanonicalDecl());
7708     if (I != LambdasMap.end())
7709       // for map(to: lambda): using user specified map type.
7710       return getMapTypeBits(
7711           I->getSecond()->getMapType(), I->getSecond()->getMapTypeModifiers(),
7712           /*MotionModifiers=*/std::nullopt, I->getSecond()->isImplicit(),
7713           /*AddPtrFlag=*/false,
7714           /*AddIsTargetParamFlag=*/false,
7715           /*isNonContiguous=*/false);
7716     return OpenMPOffloadMappingFlags::OMP_MAP_TO |
7717            OpenMPOffloadMappingFlags::OMP_MAP_FROM;
7718   }
7719 
getPlainLayout(const CXXRecordDecl * RD,llvm::SmallVectorImpl<const FieldDecl * > & Layout,bool AsBase) const7720   void getPlainLayout(const CXXRecordDecl *RD,
7721                       llvm::SmallVectorImpl<const FieldDecl *> &Layout,
7722                       bool AsBase) const {
7723     const CGRecordLayout &RL = CGF.getTypes().getCGRecordLayout(RD);
7724 
7725     llvm::StructType *St =
7726         AsBase ? RL.getBaseSubobjectLLVMType() : RL.getLLVMType();
7727 
7728     unsigned NumElements = St->getNumElements();
7729     llvm::SmallVector<
7730         llvm::PointerUnion<const CXXRecordDecl *, const FieldDecl *>, 4>
7731         RecordLayout(NumElements);
7732 
7733     // Fill bases.
7734     for (const auto &I : RD->bases()) {
7735       if (I.isVirtual())
7736         continue;
7737 
7738       QualType BaseTy = I.getType();
7739       const auto *Base = BaseTy->getAsCXXRecordDecl();
7740       // Ignore empty bases.
7741       if (isEmptyRecordForLayout(CGF.getContext(), BaseTy) ||
7742           CGF.getContext()
7743               .getASTRecordLayout(Base)
7744               .getNonVirtualSize()
7745               .isZero())
7746         continue;
7747 
7748       unsigned FieldIndex = RL.getNonVirtualBaseLLVMFieldNo(Base);
7749       RecordLayout[FieldIndex] = Base;
7750     }
7751     // Fill in virtual bases.
7752     for (const auto &I : RD->vbases()) {
7753       QualType BaseTy = I.getType();
7754       // Ignore empty bases.
7755       if (isEmptyRecordForLayout(CGF.getContext(), BaseTy))
7756         continue;
7757 
7758       const auto *Base = BaseTy->getAsCXXRecordDecl();
7759       unsigned FieldIndex = RL.getVirtualBaseIndex(Base);
7760       if (RecordLayout[FieldIndex])
7761         continue;
7762       RecordLayout[FieldIndex] = Base;
7763     }
7764     // Fill in all the fields.
7765     assert(!RD->isUnion() && "Unexpected union.");
7766     for (const auto *Field : RD->fields()) {
7767       // Fill in non-bitfields. (Bitfields always use a zero pattern, which we
7768       // will fill in later.)
7769       if (!Field->isBitField() &&
7770           !isEmptyFieldForLayout(CGF.getContext(), Field)) {
7771         unsigned FieldIndex = RL.getLLVMFieldNo(Field);
7772         RecordLayout[FieldIndex] = Field;
7773       }
7774     }
7775     for (const llvm::PointerUnion<const CXXRecordDecl *, const FieldDecl *>
7776              &Data : RecordLayout) {
7777       if (Data.isNull())
7778         continue;
7779       if (const auto *Base = Data.dyn_cast<const CXXRecordDecl *>())
7780         getPlainLayout(Base, Layout, /*AsBase=*/true);
7781       else
7782         Layout.push_back(Data.get<const FieldDecl *>());
7783     }
7784   }
7785 
7786   /// Generate all the base pointers, section pointers, sizes, map types, and
7787   /// mappers for the extracted mappable expressions (all included in \a
7788   /// CombinedInfo). Also, for each item that relates with a device pointer, a
7789   /// pair of the relevant declaration and index where it occurs is appended to
7790   /// the device pointers info array.
generateAllInfoForClauses(ArrayRef<const OMPClause * > Clauses,MapCombinedInfoTy & CombinedInfo,llvm::OpenMPIRBuilder & OMPBuilder,const llvm::DenseSet<CanonicalDeclPtr<const Decl>> & SkipVarSet=llvm::DenseSet<CanonicalDeclPtr<const Decl>> ()) const7791   void generateAllInfoForClauses(
7792       ArrayRef<const OMPClause *> Clauses, MapCombinedInfoTy &CombinedInfo,
7793       llvm::OpenMPIRBuilder &OMPBuilder,
7794       const llvm::DenseSet<CanonicalDeclPtr<const Decl>> &SkipVarSet =
7795           llvm::DenseSet<CanonicalDeclPtr<const Decl>>()) const {
7796     // We have to process the component lists that relate with the same
7797     // declaration in a single chunk so that we can generate the map flags
7798     // correctly. Therefore, we organize all lists in a map.
7799     enum MapKind { Present, Allocs, Other, Total };
7800     llvm::MapVector<CanonicalDeclPtr<const Decl>,
7801                     SmallVector<SmallVector<MapInfo, 8>, 4>>
7802         Info;
7803 
7804     // Helper function to fill the information map for the different supported
7805     // clauses.
7806     auto &&InfoGen =
7807         [&Info, &SkipVarSet](
7808             const ValueDecl *D, MapKind Kind,
7809             OMPClauseMappableExprCommon::MappableExprComponentListRef L,
7810             OpenMPMapClauseKind MapType,
7811             ArrayRef<OpenMPMapModifierKind> MapModifiers,
7812             ArrayRef<OpenMPMotionModifierKind> MotionModifiers,
7813             bool ReturnDevicePointer, bool IsImplicit, const ValueDecl *Mapper,
7814             const Expr *VarRef = nullptr, bool ForDeviceAddr = false) {
7815           if (SkipVarSet.contains(D))
7816             return;
7817           auto It = Info.find(D);
7818           if (It == Info.end())
7819             It = Info
7820                      .insert(std::make_pair(
7821                          D, SmallVector<SmallVector<MapInfo, 8>, 4>(Total)))
7822                      .first;
7823           It->second[Kind].emplace_back(
7824               L, MapType, MapModifiers, MotionModifiers, ReturnDevicePointer,
7825               IsImplicit, Mapper, VarRef, ForDeviceAddr);
7826         };
7827 
7828     for (const auto *Cl : Clauses) {
7829       const auto *C = dyn_cast<OMPMapClause>(Cl);
7830       if (!C)
7831         continue;
7832       MapKind Kind = Other;
7833       if (llvm::is_contained(C->getMapTypeModifiers(),
7834                              OMPC_MAP_MODIFIER_present))
7835         Kind = Present;
7836       else if (C->getMapType() == OMPC_MAP_alloc)
7837         Kind = Allocs;
7838       const auto *EI = C->getVarRefs().begin();
7839       for (const auto L : C->component_lists()) {
7840         const Expr *E = (C->getMapLoc().isValid()) ? *EI : nullptr;
7841         InfoGen(std::get<0>(L), Kind, std::get<1>(L), C->getMapType(),
7842                 C->getMapTypeModifiers(), std::nullopt,
7843                 /*ReturnDevicePointer=*/false, C->isImplicit(), std::get<2>(L),
7844                 E);
7845         ++EI;
7846       }
7847     }
7848     for (const auto *Cl : Clauses) {
7849       const auto *C = dyn_cast<OMPToClause>(Cl);
7850       if (!C)
7851         continue;
7852       MapKind Kind = Other;
7853       if (llvm::is_contained(C->getMotionModifiers(),
7854                              OMPC_MOTION_MODIFIER_present))
7855         Kind = Present;
7856       const auto *EI = C->getVarRefs().begin();
7857       for (const auto L : C->component_lists()) {
7858         InfoGen(std::get<0>(L), Kind, std::get<1>(L), OMPC_MAP_to, std::nullopt,
7859                 C->getMotionModifiers(), /*ReturnDevicePointer=*/false,
7860                 C->isImplicit(), std::get<2>(L), *EI);
7861         ++EI;
7862       }
7863     }
7864     for (const auto *Cl : Clauses) {
7865       const auto *C = dyn_cast<OMPFromClause>(Cl);
7866       if (!C)
7867         continue;
7868       MapKind Kind = Other;
7869       if (llvm::is_contained(C->getMotionModifiers(),
7870                              OMPC_MOTION_MODIFIER_present))
7871         Kind = Present;
7872       const auto *EI = C->getVarRefs().begin();
7873       for (const auto L : C->component_lists()) {
7874         InfoGen(std::get<0>(L), Kind, std::get<1>(L), OMPC_MAP_from,
7875                 std::nullopt, C->getMotionModifiers(),
7876                 /*ReturnDevicePointer=*/false, C->isImplicit(), std::get<2>(L),
7877                 *EI);
7878         ++EI;
7879       }
7880     }
7881 
7882     // Look at the use_device_ptr and use_device_addr clauses information and
7883     // mark the existing map entries as such. If there is no map information for
7884     // an entry in the use_device_ptr and use_device_addr list, we create one
7885     // with map type 'alloc' and zero size section. It is the user fault if that
7886     // was not mapped before. If there is no map information and the pointer is
7887     // a struct member, then we defer the emission of that entry until the whole
7888     // struct has been processed.
7889     llvm::MapVector<CanonicalDeclPtr<const Decl>,
7890                     SmallVector<DeferredDevicePtrEntryTy, 4>>
7891         DeferredInfo;
7892     MapCombinedInfoTy UseDeviceDataCombinedInfo;
7893 
7894     auto &&UseDeviceDataCombinedInfoGen =
7895         [&UseDeviceDataCombinedInfo](const ValueDecl *VD, llvm::Value *Ptr,
7896                                      CodeGenFunction &CGF, bool IsDevAddr) {
7897           UseDeviceDataCombinedInfo.Exprs.push_back(VD);
7898           UseDeviceDataCombinedInfo.BasePointers.emplace_back(Ptr);
7899           UseDeviceDataCombinedInfo.DevicePtrDecls.emplace_back(VD);
7900           UseDeviceDataCombinedInfo.DevicePointers.emplace_back(
7901               IsDevAddr ? DeviceInfoTy::Address : DeviceInfoTy::Pointer);
7902           UseDeviceDataCombinedInfo.Pointers.push_back(Ptr);
7903           UseDeviceDataCombinedInfo.Sizes.push_back(
7904               llvm::Constant::getNullValue(CGF.Int64Ty));
7905           UseDeviceDataCombinedInfo.Types.push_back(
7906               OpenMPOffloadMappingFlags::OMP_MAP_RETURN_PARAM);
7907           UseDeviceDataCombinedInfo.Mappers.push_back(nullptr);
7908         };
7909 
7910     auto &&MapInfoGen =
7911         [&DeferredInfo, &UseDeviceDataCombinedInfoGen,
7912          &InfoGen](CodeGenFunction &CGF, const Expr *IE, const ValueDecl *VD,
7913                    OMPClauseMappableExprCommon::MappableExprComponentListRef
7914                        Components,
7915                    bool IsImplicit, bool IsDevAddr) {
7916           // We didn't find any match in our map information - generate a zero
7917           // size array section - if the pointer is a struct member we defer
7918           // this action until the whole struct has been processed.
7919           if (isa<MemberExpr>(IE)) {
7920             // Insert the pointer into Info to be processed by
7921             // generateInfoForComponentList. Because it is a member pointer
7922             // without a pointee, no entry will be generated for it, therefore
7923             // we need to generate one after the whole struct has been
7924             // processed. Nonetheless, generateInfoForComponentList must be
7925             // called to take the pointer into account for the calculation of
7926             // the range of the partial struct.
7927             InfoGen(nullptr, Other, Components, OMPC_MAP_unknown, std::nullopt,
7928                     std::nullopt, /*ReturnDevicePointer=*/false, IsImplicit,
7929                     nullptr, nullptr, IsDevAddr);
7930             DeferredInfo[nullptr].emplace_back(IE, VD, IsDevAddr);
7931           } else {
7932             llvm::Value *Ptr;
7933             if (IsDevAddr) {
7934               if (IE->isGLValue())
7935                 Ptr = CGF.EmitLValue(IE).getPointer(CGF);
7936               else
7937                 Ptr = CGF.EmitScalarExpr(IE);
7938             } else {
7939               Ptr = CGF.EmitLoadOfScalar(CGF.EmitLValue(IE), IE->getExprLoc());
7940             }
7941             UseDeviceDataCombinedInfoGen(VD, Ptr, CGF, IsDevAddr);
7942           }
7943         };
7944 
7945     auto &&IsMapInfoExist = [&Info](CodeGenFunction &CGF, const ValueDecl *VD,
7946                                     const Expr *IE, bool IsDevAddr) -> bool {
7947       // We potentially have map information for this declaration already.
7948       // Look for the first set of components that refer to it. If found,
7949       // return true.
7950       // If the first component is a member expression, we have to look into
7951       // 'this', which maps to null in the map of map information. Otherwise
7952       // look directly for the information.
7953       auto It = Info.find(isa<MemberExpr>(IE) ? nullptr : VD);
7954       if (It != Info.end()) {
7955         bool Found = false;
7956         for (auto &Data : It->second) {
7957           auto *CI = llvm::find_if(Data, [VD](const MapInfo &MI) {
7958             return MI.Components.back().getAssociatedDeclaration() == VD;
7959           });
7960           // If we found a map entry, signal that the pointer has to be
7961           // returned and move on to the next declaration. Exclude cases where
7962           // the base pointer is mapped as array subscript, array section or
7963           // array shaping. The base address is passed as a pointer to base in
7964           // this case and cannot be used as a base for use_device_ptr list
7965           // item.
7966           if (CI != Data.end()) {
7967             if (IsDevAddr) {
7968               CI->ForDeviceAddr = IsDevAddr;
7969               CI->ReturnDevicePointer = true;
7970               Found = true;
7971               break;
7972             } else {
7973               auto PrevCI = std::next(CI->Components.rbegin());
7974               const auto *VarD = dyn_cast<VarDecl>(VD);
7975               if (CGF.CGM.getOpenMPRuntime().hasRequiresUnifiedSharedMemory() ||
7976                   isa<MemberExpr>(IE) ||
7977                   !VD->getType().getNonReferenceType()->isPointerType() ||
7978                   PrevCI == CI->Components.rend() ||
7979                   isa<MemberExpr>(PrevCI->getAssociatedExpression()) || !VarD ||
7980                   VarD->hasLocalStorage()) {
7981                 CI->ForDeviceAddr = IsDevAddr;
7982                 CI->ReturnDevicePointer = true;
7983                 Found = true;
7984                 break;
7985               }
7986             }
7987           }
7988         }
7989         return Found;
7990       }
7991       return false;
7992     };
7993 
7994     // Look at the use_device_ptr clause information and mark the existing map
7995     // entries as such. If there is no map information for an entry in the
7996     // use_device_ptr list, we create one with map type 'alloc' and zero size
7997     // section. It is the user fault if that was not mapped before. If there is
7998     // no map information and the pointer is a struct member, then we defer the
7999     // emission of that entry until the whole struct has been processed.
8000     for (const auto *Cl : Clauses) {
8001       const auto *C = dyn_cast<OMPUseDevicePtrClause>(Cl);
8002       if (!C)
8003         continue;
8004       for (const auto L : C->component_lists()) {
8005         OMPClauseMappableExprCommon::MappableExprComponentListRef Components =
8006             std::get<1>(L);
8007         assert(!Components.empty() &&
8008                "Not expecting empty list of components!");
8009         const ValueDecl *VD = Components.back().getAssociatedDeclaration();
8010         VD = cast<ValueDecl>(VD->getCanonicalDecl());
8011         const Expr *IE = Components.back().getAssociatedExpression();
8012         if (IsMapInfoExist(CGF, VD, IE, /*IsDevAddr=*/false))
8013           continue;
8014         MapInfoGen(CGF, IE, VD, Components, C->isImplicit(),
8015                    /*IsDevAddr=*/false);
8016       }
8017     }
8018 
8019     llvm::SmallDenseSet<CanonicalDeclPtr<const Decl>, 4> Processed;
8020     for (const auto *Cl : Clauses) {
8021       const auto *C = dyn_cast<OMPUseDeviceAddrClause>(Cl);
8022       if (!C)
8023         continue;
8024       for (const auto L : C->component_lists()) {
8025         OMPClauseMappableExprCommon::MappableExprComponentListRef Components =
8026             std::get<1>(L);
8027         assert(!std::get<1>(L).empty() &&
8028                "Not expecting empty list of components!");
8029         const ValueDecl *VD = std::get<1>(L).back().getAssociatedDeclaration();
8030         if (!Processed.insert(VD).second)
8031           continue;
8032         VD = cast<ValueDecl>(VD->getCanonicalDecl());
8033         const Expr *IE = std::get<1>(L).back().getAssociatedExpression();
8034         if (IsMapInfoExist(CGF, VD, IE, /*IsDevAddr=*/true))
8035           continue;
8036         MapInfoGen(CGF, IE, VD, Components, C->isImplicit(),
8037                    /*IsDevAddr=*/true);
8038       }
8039     }
8040 
8041     for (const auto &Data : Info) {
8042       StructRangeInfoTy PartialStruct;
8043       // Current struct information:
8044       MapCombinedInfoTy CurInfo;
8045       // Current struct base information:
8046       MapCombinedInfoTy StructBaseCurInfo;
8047       const Decl *D = Data.first;
8048       const ValueDecl *VD = cast_or_null<ValueDecl>(D);
8049       bool HasMapBasePtr = false;
8050       bool HasMapArraySec = false;
8051       if (VD && VD->getType()->isAnyPointerType()) {
8052         for (const auto &M : Data.second) {
8053           HasMapBasePtr = any_of(M, [](const MapInfo &L) {
8054             return isa_and_present<DeclRefExpr>(L.VarRef);
8055           });
8056           HasMapArraySec = any_of(M, [](const MapInfo &L) {
8057             return isa_and_present<ArraySectionExpr, ArraySubscriptExpr>(
8058                 L.VarRef);
8059           });
8060           if (HasMapBasePtr && HasMapArraySec)
8061             break;
8062         }
8063       }
8064       for (const auto &M : Data.second) {
8065         for (const MapInfo &L : M) {
8066           assert(!L.Components.empty() &&
8067                  "Not expecting declaration with no component lists.");
8068 
8069           // Remember the current base pointer index.
8070           unsigned CurrentBasePointersIdx = CurInfo.BasePointers.size();
8071           unsigned StructBasePointersIdx =
8072               StructBaseCurInfo.BasePointers.size();
8073           CurInfo.NonContigInfo.IsNonContiguous =
8074               L.Components.back().isNonContiguous();
8075           generateInfoForComponentList(
8076               L.MapType, L.MapModifiers, L.MotionModifiers, L.Components,
8077               CurInfo, StructBaseCurInfo, PartialStruct,
8078               /*IsFirstComponentList=*/false, L.IsImplicit,
8079               /*GenerateAllInfoForClauses*/ true, L.Mapper, L.ForDeviceAddr, VD,
8080               L.VarRef, /*OverlappedElements*/ std::nullopt,
8081               HasMapBasePtr && HasMapArraySec);
8082 
8083           // If this entry relates to a device pointer, set the relevant
8084           // declaration and add the 'return pointer' flag.
8085           if (L.ReturnDevicePointer) {
8086             // Check whether a value was added to either CurInfo or
8087             // StructBaseCurInfo and error if no value was added to either of
8088             // them:
8089             assert((CurrentBasePointersIdx < CurInfo.BasePointers.size() ||
8090                     StructBasePointersIdx <
8091                         StructBaseCurInfo.BasePointers.size()) &&
8092                    "Unexpected number of mapped base pointers.");
8093 
8094             // Choose a base pointer index which is always valid:
8095             const ValueDecl *RelevantVD =
8096                 L.Components.back().getAssociatedDeclaration();
8097             assert(RelevantVD &&
8098                    "No relevant declaration related with device pointer??");
8099 
8100             // If StructBaseCurInfo has been updated this iteration then work on
8101             // the first new entry added to it i.e. make sure that when multiple
8102             // values are added to any of the lists, the first value added is
8103             // being modified by the assignments below (not the last value
8104             // added).
8105             if (StructBasePointersIdx < StructBaseCurInfo.BasePointers.size()) {
8106               StructBaseCurInfo.DevicePtrDecls[StructBasePointersIdx] =
8107                   RelevantVD;
8108               StructBaseCurInfo.DevicePointers[StructBasePointersIdx] =
8109                   L.ForDeviceAddr ? DeviceInfoTy::Address
8110                                   : DeviceInfoTy::Pointer;
8111               StructBaseCurInfo.Types[StructBasePointersIdx] |=
8112                   OpenMPOffloadMappingFlags::OMP_MAP_RETURN_PARAM;
8113             } else {
8114               CurInfo.DevicePtrDecls[CurrentBasePointersIdx] = RelevantVD;
8115               CurInfo.DevicePointers[CurrentBasePointersIdx] =
8116                   L.ForDeviceAddr ? DeviceInfoTy::Address
8117                                   : DeviceInfoTy::Pointer;
8118               CurInfo.Types[CurrentBasePointersIdx] |=
8119                   OpenMPOffloadMappingFlags::OMP_MAP_RETURN_PARAM;
8120             }
8121           }
8122         }
8123       }
8124 
8125       // Append any pending zero-length pointers which are struct members and
8126       // used with use_device_ptr or use_device_addr.
8127       auto CI = DeferredInfo.find(Data.first);
8128       if (CI != DeferredInfo.end()) {
8129         for (const DeferredDevicePtrEntryTy &L : CI->second) {
8130           llvm::Value *BasePtr;
8131           llvm::Value *Ptr;
8132           if (L.ForDeviceAddr) {
8133             if (L.IE->isGLValue())
8134               Ptr = this->CGF.EmitLValue(L.IE).getPointer(CGF);
8135             else
8136               Ptr = this->CGF.EmitScalarExpr(L.IE);
8137             BasePtr = Ptr;
8138             // Entry is RETURN_PARAM. Also, set the placeholder value
8139             // MEMBER_OF=FFFF so that the entry is later updated with the
8140             // correct value of MEMBER_OF.
8141             CurInfo.Types.push_back(
8142                 OpenMPOffloadMappingFlags::OMP_MAP_RETURN_PARAM |
8143                 OpenMPOffloadMappingFlags::OMP_MAP_MEMBER_OF);
8144           } else {
8145             BasePtr = this->CGF.EmitLValue(L.IE).getPointer(CGF);
8146             Ptr = this->CGF.EmitLoadOfScalar(this->CGF.EmitLValue(L.IE),
8147                                              L.IE->getExprLoc());
8148             // Entry is PTR_AND_OBJ and RETURN_PARAM. Also, set the
8149             // placeholder value MEMBER_OF=FFFF so that the entry is later
8150             // updated with the correct value of MEMBER_OF.
8151             CurInfo.Types.push_back(
8152                 OpenMPOffloadMappingFlags::OMP_MAP_PTR_AND_OBJ |
8153                 OpenMPOffloadMappingFlags::OMP_MAP_RETURN_PARAM |
8154                 OpenMPOffloadMappingFlags::OMP_MAP_MEMBER_OF);
8155           }
8156           CurInfo.Exprs.push_back(L.VD);
8157           CurInfo.BasePointers.emplace_back(BasePtr);
8158           CurInfo.DevicePtrDecls.emplace_back(L.VD);
8159           CurInfo.DevicePointers.emplace_back(
8160               L.ForDeviceAddr ? DeviceInfoTy::Address : DeviceInfoTy::Pointer);
8161           CurInfo.Pointers.push_back(Ptr);
8162           CurInfo.Sizes.push_back(
8163               llvm::Constant::getNullValue(this->CGF.Int64Ty));
8164           CurInfo.Mappers.push_back(nullptr);
8165         }
8166       }
8167 
8168       // Unify entries in one list making sure the struct mapping precedes the
8169       // individual fields:
8170       MapCombinedInfoTy UnionCurInfo;
8171       UnionCurInfo.append(StructBaseCurInfo);
8172       UnionCurInfo.append(CurInfo);
8173 
8174       // If there is an entry in PartialStruct it means we have a struct with
8175       // individual members mapped. Emit an extra combined entry.
8176       if (PartialStruct.Base.isValid()) {
8177         UnionCurInfo.NonContigInfo.Dims.push_back(0);
8178         // Emit a combined entry:
8179         emitCombinedEntry(CombinedInfo, UnionCurInfo.Types, PartialStruct,
8180                           /*IsMapThis*/ !VD, OMPBuilder, VD);
8181       }
8182 
8183       // We need to append the results of this capture to what we already have.
8184       CombinedInfo.append(UnionCurInfo);
8185     }
8186     // Append data for use_device_ptr clauses.
8187     CombinedInfo.append(UseDeviceDataCombinedInfo);
8188   }
8189 
8190 public:
MappableExprsHandler(const OMPExecutableDirective & Dir,CodeGenFunction & CGF)8191   MappableExprsHandler(const OMPExecutableDirective &Dir, CodeGenFunction &CGF)
8192       : CurDir(&Dir), CGF(CGF) {
8193     // Extract firstprivate clause information.
8194     for (const auto *C : Dir.getClausesOfKind<OMPFirstprivateClause>())
8195       for (const auto *D : C->varlists())
8196         FirstPrivateDecls.try_emplace(
8197             cast<VarDecl>(cast<DeclRefExpr>(D)->getDecl()), C->isImplicit());
8198     // Extract implicit firstprivates from uses_allocators clauses.
8199     for (const auto *C : Dir.getClausesOfKind<OMPUsesAllocatorsClause>()) {
8200       for (unsigned I = 0, E = C->getNumberOfAllocators(); I < E; ++I) {
8201         OMPUsesAllocatorsClause::Data D = C->getAllocatorData(I);
8202         if (const auto *DRE = dyn_cast_or_null<DeclRefExpr>(D.AllocatorTraits))
8203           FirstPrivateDecls.try_emplace(cast<VarDecl>(DRE->getDecl()),
8204                                         /*Implicit=*/true);
8205         else if (const auto *VD = dyn_cast<VarDecl>(
8206                      cast<DeclRefExpr>(D.Allocator->IgnoreParenImpCasts())
8207                          ->getDecl()))
8208           FirstPrivateDecls.try_emplace(VD, /*Implicit=*/true);
8209       }
8210     }
8211     // Extract device pointer clause information.
8212     for (const auto *C : Dir.getClausesOfKind<OMPIsDevicePtrClause>())
8213       for (auto L : C->component_lists())
8214         DevPointersMap[std::get<0>(L)].push_back(std::get<1>(L));
8215     // Extract device addr clause information.
8216     for (const auto *C : Dir.getClausesOfKind<OMPHasDeviceAddrClause>())
8217       for (auto L : C->component_lists())
8218         HasDevAddrsMap[std::get<0>(L)].push_back(std::get<1>(L));
8219     // Extract map information.
8220     for (const auto *C : Dir.getClausesOfKind<OMPMapClause>()) {
8221       if (C->getMapType() != OMPC_MAP_to)
8222         continue;
8223       for (auto L : C->component_lists()) {
8224         const ValueDecl *VD = std::get<0>(L);
8225         const auto *RD = VD ? VD->getType()
8226                                   .getCanonicalType()
8227                                   .getNonReferenceType()
8228                                   ->getAsCXXRecordDecl()
8229                             : nullptr;
8230         if (RD && RD->isLambda())
8231           LambdasMap.try_emplace(std::get<0>(L), C);
8232       }
8233     }
8234   }
8235 
8236   /// Constructor for the declare mapper directive.
MappableExprsHandler(const OMPDeclareMapperDecl & Dir,CodeGenFunction & CGF)8237   MappableExprsHandler(const OMPDeclareMapperDecl &Dir, CodeGenFunction &CGF)
8238       : CurDir(&Dir), CGF(CGF) {}
8239 
8240   /// Generate code for the combined entry if we have a partially mapped struct
8241   /// and take care of the mapping flags of the arguments corresponding to
8242   /// individual struct members.
emitCombinedEntry(MapCombinedInfoTy & CombinedInfo,MapFlagsArrayTy & CurTypes,const StructRangeInfoTy & PartialStruct,bool IsMapThis,llvm::OpenMPIRBuilder & OMPBuilder,const ValueDecl * VD=nullptr,bool NotTargetParams=true) const8243   void emitCombinedEntry(MapCombinedInfoTy &CombinedInfo,
8244                          MapFlagsArrayTy &CurTypes,
8245                          const StructRangeInfoTy &PartialStruct, bool IsMapThis,
8246                          llvm::OpenMPIRBuilder &OMPBuilder,
8247                          const ValueDecl *VD = nullptr,
8248                          bool NotTargetParams = true) const {
8249     if (CurTypes.size() == 1 &&
8250         ((CurTypes.back() & OpenMPOffloadMappingFlags::OMP_MAP_MEMBER_OF) !=
8251          OpenMPOffloadMappingFlags::OMP_MAP_MEMBER_OF) &&
8252         !PartialStruct.IsArraySection)
8253       return;
8254     Address LBAddr = PartialStruct.LowestElem.second;
8255     Address HBAddr = PartialStruct.HighestElem.second;
8256     if (PartialStruct.HasCompleteRecord) {
8257       LBAddr = PartialStruct.LB;
8258       HBAddr = PartialStruct.LB;
8259     }
8260     CombinedInfo.Exprs.push_back(VD);
8261     // Base is the base of the struct
8262     CombinedInfo.BasePointers.push_back(PartialStruct.Base.emitRawPointer(CGF));
8263     CombinedInfo.DevicePtrDecls.push_back(nullptr);
8264     CombinedInfo.DevicePointers.push_back(DeviceInfoTy::None);
8265     // Pointer is the address of the lowest element
8266     llvm::Value *LB = LBAddr.emitRawPointer(CGF);
8267     const CXXMethodDecl *MD =
8268         CGF.CurFuncDecl ? dyn_cast<CXXMethodDecl>(CGF.CurFuncDecl) : nullptr;
8269     const CXXRecordDecl *RD = MD ? MD->getParent() : nullptr;
8270     bool HasBaseClass = RD && IsMapThis ? RD->getNumBases() > 0 : false;
8271     // There should not be a mapper for a combined entry.
8272     if (HasBaseClass) {
8273       // OpenMP 5.2 148:21:
8274       // If the target construct is within a class non-static member function,
8275       // and a variable is an accessible data member of the object for which the
8276       // non-static data member function is invoked, the variable is treated as
8277       // if the this[:1] expression had appeared in a map clause with a map-type
8278       // of tofrom.
8279       // Emit this[:1]
8280       CombinedInfo.Pointers.push_back(PartialStruct.Base.emitRawPointer(CGF));
8281       QualType Ty = MD->getFunctionObjectParameterType();
8282       llvm::Value *Size =
8283           CGF.Builder.CreateIntCast(CGF.getTypeSize(Ty), CGF.Int64Ty,
8284                                     /*isSigned=*/true);
8285       CombinedInfo.Sizes.push_back(Size);
8286     } else {
8287       CombinedInfo.Pointers.push_back(LB);
8288       // Size is (addr of {highest+1} element) - (addr of lowest element)
8289       llvm::Value *HB = HBAddr.emitRawPointer(CGF);
8290       llvm::Value *HAddr = CGF.Builder.CreateConstGEP1_32(
8291           HBAddr.getElementType(), HB, /*Idx0=*/1);
8292       llvm::Value *CLAddr = CGF.Builder.CreatePointerCast(LB, CGF.VoidPtrTy);
8293       llvm::Value *CHAddr = CGF.Builder.CreatePointerCast(HAddr, CGF.VoidPtrTy);
8294       llvm::Value *Diff = CGF.Builder.CreatePtrDiff(CGF.Int8Ty, CHAddr, CLAddr);
8295       llvm::Value *Size = CGF.Builder.CreateIntCast(Diff, CGF.Int64Ty,
8296                                                     /*isSigned=*/false);
8297       CombinedInfo.Sizes.push_back(Size);
8298     }
8299     CombinedInfo.Mappers.push_back(nullptr);
8300     // Map type is always TARGET_PARAM, if generate info for captures.
8301     CombinedInfo.Types.push_back(
8302         NotTargetParams ? OpenMPOffloadMappingFlags::OMP_MAP_NONE
8303                         : OpenMPOffloadMappingFlags::OMP_MAP_TARGET_PARAM);
8304     // If any element has the present modifier, then make sure the runtime
8305     // doesn't attempt to allocate the struct.
8306     if (CurTypes.end() !=
8307         llvm::find_if(CurTypes, [](OpenMPOffloadMappingFlags Type) {
8308           return static_cast<std::underlying_type_t<OpenMPOffloadMappingFlags>>(
8309               Type & OpenMPOffloadMappingFlags::OMP_MAP_PRESENT);
8310         }))
8311       CombinedInfo.Types.back() |= OpenMPOffloadMappingFlags::OMP_MAP_PRESENT;
8312     // Remove TARGET_PARAM flag from the first element
8313     (*CurTypes.begin()) &= ~OpenMPOffloadMappingFlags::OMP_MAP_TARGET_PARAM;
8314     // If any element has the ompx_hold modifier, then make sure the runtime
8315     // uses the hold reference count for the struct as a whole so that it won't
8316     // be unmapped by an extra dynamic reference count decrement.  Add it to all
8317     // elements as well so the runtime knows which reference count to check
8318     // when determining whether it's time for device-to-host transfers of
8319     // individual elements.
8320     if (CurTypes.end() !=
8321         llvm::find_if(CurTypes, [](OpenMPOffloadMappingFlags Type) {
8322           return static_cast<std::underlying_type_t<OpenMPOffloadMappingFlags>>(
8323               Type & OpenMPOffloadMappingFlags::OMP_MAP_OMPX_HOLD);
8324         })) {
8325       CombinedInfo.Types.back() |= OpenMPOffloadMappingFlags::OMP_MAP_OMPX_HOLD;
8326       for (auto &M : CurTypes)
8327         M |= OpenMPOffloadMappingFlags::OMP_MAP_OMPX_HOLD;
8328     }
8329 
8330     // All other current entries will be MEMBER_OF the combined entry
8331     // (except for PTR_AND_OBJ entries which do not have a placeholder value
8332     // 0xFFFF in the MEMBER_OF field).
8333     OpenMPOffloadMappingFlags MemberOfFlag =
8334         OMPBuilder.getMemberOfFlag(CombinedInfo.BasePointers.size() - 1);
8335     for (auto &M : CurTypes)
8336       OMPBuilder.setCorrectMemberOfFlag(M, MemberOfFlag);
8337   }
8338 
8339   /// Generate all the base pointers, section pointers, sizes, map types, and
8340   /// mappers for the extracted mappable expressions (all included in \a
8341   /// CombinedInfo). Also, for each item that relates with a device pointer, a
8342   /// pair of the relevant declaration and index where it occurs is appended to
8343   /// the device pointers info array.
generateAllInfo(MapCombinedInfoTy & CombinedInfo,llvm::OpenMPIRBuilder & OMPBuilder,const llvm::DenseSet<CanonicalDeclPtr<const Decl>> & SkipVarSet=llvm::DenseSet<CanonicalDeclPtr<const Decl>> ()) const8344   void generateAllInfo(
8345       MapCombinedInfoTy &CombinedInfo, llvm::OpenMPIRBuilder &OMPBuilder,
8346       const llvm::DenseSet<CanonicalDeclPtr<const Decl>> &SkipVarSet =
8347           llvm::DenseSet<CanonicalDeclPtr<const Decl>>()) const {
8348     assert(CurDir.is<const OMPExecutableDirective *>() &&
8349            "Expect a executable directive");
8350     const auto *CurExecDir = CurDir.get<const OMPExecutableDirective *>();
8351     generateAllInfoForClauses(CurExecDir->clauses(), CombinedInfo, OMPBuilder,
8352                               SkipVarSet);
8353   }
8354 
8355   /// Generate all the base pointers, section pointers, sizes, map types, and
8356   /// mappers for the extracted map clauses of user-defined mapper (all included
8357   /// in \a CombinedInfo).
generateAllInfoForMapper(MapCombinedInfoTy & CombinedInfo,llvm::OpenMPIRBuilder & OMPBuilder) const8358   void generateAllInfoForMapper(MapCombinedInfoTy &CombinedInfo,
8359                                 llvm::OpenMPIRBuilder &OMPBuilder) const {
8360     assert(CurDir.is<const OMPDeclareMapperDecl *>() &&
8361            "Expect a declare mapper directive");
8362     const auto *CurMapperDir = CurDir.get<const OMPDeclareMapperDecl *>();
8363     generateAllInfoForClauses(CurMapperDir->clauses(), CombinedInfo,
8364                               OMPBuilder);
8365   }
8366 
8367   /// Emit capture info for lambdas for variables captured by reference.
generateInfoForLambdaCaptures(const ValueDecl * VD,llvm::Value * Arg,MapCombinedInfoTy & CombinedInfo,llvm::DenseMap<llvm::Value *,llvm::Value * > & LambdaPointers) const8368   void generateInfoForLambdaCaptures(
8369       const ValueDecl *VD, llvm::Value *Arg, MapCombinedInfoTy &CombinedInfo,
8370       llvm::DenseMap<llvm::Value *, llvm::Value *> &LambdaPointers) const {
8371     QualType VDType = VD->getType().getCanonicalType().getNonReferenceType();
8372     const auto *RD = VDType->getAsCXXRecordDecl();
8373     if (!RD || !RD->isLambda())
8374       return;
8375     Address VDAddr(Arg, CGF.ConvertTypeForMem(VDType),
8376                    CGF.getContext().getDeclAlign(VD));
8377     LValue VDLVal = CGF.MakeAddrLValue(VDAddr, VDType);
8378     llvm::DenseMap<const ValueDecl *, FieldDecl *> Captures;
8379     FieldDecl *ThisCapture = nullptr;
8380     RD->getCaptureFields(Captures, ThisCapture);
8381     if (ThisCapture) {
8382       LValue ThisLVal =
8383           CGF.EmitLValueForFieldInitialization(VDLVal, ThisCapture);
8384       LValue ThisLValVal = CGF.EmitLValueForField(VDLVal, ThisCapture);
8385       LambdaPointers.try_emplace(ThisLVal.getPointer(CGF),
8386                                  VDLVal.getPointer(CGF));
8387       CombinedInfo.Exprs.push_back(VD);
8388       CombinedInfo.BasePointers.push_back(ThisLVal.getPointer(CGF));
8389       CombinedInfo.DevicePtrDecls.push_back(nullptr);
8390       CombinedInfo.DevicePointers.push_back(DeviceInfoTy::None);
8391       CombinedInfo.Pointers.push_back(ThisLValVal.getPointer(CGF));
8392       CombinedInfo.Sizes.push_back(
8393           CGF.Builder.CreateIntCast(CGF.getTypeSize(CGF.getContext().VoidPtrTy),
8394                                     CGF.Int64Ty, /*isSigned=*/true));
8395       CombinedInfo.Types.push_back(
8396           OpenMPOffloadMappingFlags::OMP_MAP_PTR_AND_OBJ |
8397           OpenMPOffloadMappingFlags::OMP_MAP_LITERAL |
8398           OpenMPOffloadMappingFlags::OMP_MAP_MEMBER_OF |
8399           OpenMPOffloadMappingFlags::OMP_MAP_IMPLICIT);
8400       CombinedInfo.Mappers.push_back(nullptr);
8401     }
8402     for (const LambdaCapture &LC : RD->captures()) {
8403       if (!LC.capturesVariable())
8404         continue;
8405       const VarDecl *VD = cast<VarDecl>(LC.getCapturedVar());
8406       if (LC.getCaptureKind() != LCK_ByRef && !VD->getType()->isPointerType())
8407         continue;
8408       auto It = Captures.find(VD);
8409       assert(It != Captures.end() && "Found lambda capture without field.");
8410       LValue VarLVal = CGF.EmitLValueForFieldInitialization(VDLVal, It->second);
8411       if (LC.getCaptureKind() == LCK_ByRef) {
8412         LValue VarLValVal = CGF.EmitLValueForField(VDLVal, It->second);
8413         LambdaPointers.try_emplace(VarLVal.getPointer(CGF),
8414                                    VDLVal.getPointer(CGF));
8415         CombinedInfo.Exprs.push_back(VD);
8416         CombinedInfo.BasePointers.push_back(VarLVal.getPointer(CGF));
8417         CombinedInfo.DevicePtrDecls.push_back(nullptr);
8418         CombinedInfo.DevicePointers.push_back(DeviceInfoTy::None);
8419         CombinedInfo.Pointers.push_back(VarLValVal.getPointer(CGF));
8420         CombinedInfo.Sizes.push_back(CGF.Builder.CreateIntCast(
8421             CGF.getTypeSize(
8422                 VD->getType().getCanonicalType().getNonReferenceType()),
8423             CGF.Int64Ty, /*isSigned=*/true));
8424       } else {
8425         RValue VarRVal = CGF.EmitLoadOfLValue(VarLVal, RD->getLocation());
8426         LambdaPointers.try_emplace(VarLVal.getPointer(CGF),
8427                                    VDLVal.getPointer(CGF));
8428         CombinedInfo.Exprs.push_back(VD);
8429         CombinedInfo.BasePointers.push_back(VarLVal.getPointer(CGF));
8430         CombinedInfo.DevicePtrDecls.push_back(nullptr);
8431         CombinedInfo.DevicePointers.push_back(DeviceInfoTy::None);
8432         CombinedInfo.Pointers.push_back(VarRVal.getScalarVal());
8433         CombinedInfo.Sizes.push_back(llvm::ConstantInt::get(CGF.Int64Ty, 0));
8434       }
8435       CombinedInfo.Types.push_back(
8436           OpenMPOffloadMappingFlags::OMP_MAP_PTR_AND_OBJ |
8437           OpenMPOffloadMappingFlags::OMP_MAP_LITERAL |
8438           OpenMPOffloadMappingFlags::OMP_MAP_MEMBER_OF |
8439           OpenMPOffloadMappingFlags::OMP_MAP_IMPLICIT);
8440       CombinedInfo.Mappers.push_back(nullptr);
8441     }
8442   }
8443 
8444   /// Set correct indices for lambdas captures.
adjustMemberOfForLambdaCaptures(llvm::OpenMPIRBuilder & OMPBuilder,const llvm::DenseMap<llvm::Value *,llvm::Value * > & LambdaPointers,MapBaseValuesArrayTy & BasePointers,MapValuesArrayTy & Pointers,MapFlagsArrayTy & Types) const8445   void adjustMemberOfForLambdaCaptures(
8446       llvm::OpenMPIRBuilder &OMPBuilder,
8447       const llvm::DenseMap<llvm::Value *, llvm::Value *> &LambdaPointers,
8448       MapBaseValuesArrayTy &BasePointers, MapValuesArrayTy &Pointers,
8449       MapFlagsArrayTy &Types) const {
8450     for (unsigned I = 0, E = Types.size(); I < E; ++I) {
8451       // Set correct member_of idx for all implicit lambda captures.
8452       if (Types[I] != (OpenMPOffloadMappingFlags::OMP_MAP_PTR_AND_OBJ |
8453                        OpenMPOffloadMappingFlags::OMP_MAP_LITERAL |
8454                        OpenMPOffloadMappingFlags::OMP_MAP_MEMBER_OF |
8455                        OpenMPOffloadMappingFlags::OMP_MAP_IMPLICIT))
8456         continue;
8457       llvm::Value *BasePtr = LambdaPointers.lookup(BasePointers[I]);
8458       assert(BasePtr && "Unable to find base lambda address.");
8459       int TgtIdx = -1;
8460       for (unsigned J = I; J > 0; --J) {
8461         unsigned Idx = J - 1;
8462         if (Pointers[Idx] != BasePtr)
8463           continue;
8464         TgtIdx = Idx;
8465         break;
8466       }
8467       assert(TgtIdx != -1 && "Unable to find parent lambda.");
8468       // All other current entries will be MEMBER_OF the combined entry
8469       // (except for PTR_AND_OBJ entries which do not have a placeholder value
8470       // 0xFFFF in the MEMBER_OF field).
8471       OpenMPOffloadMappingFlags MemberOfFlag =
8472           OMPBuilder.getMemberOfFlag(TgtIdx);
8473       OMPBuilder.setCorrectMemberOfFlag(Types[I], MemberOfFlag);
8474     }
8475   }
8476 
8477   /// Generate the base pointers, section pointers, sizes, map types, and
8478   /// mappers associated to a given capture (all included in \a CombinedInfo).
generateInfoForCapture(const CapturedStmt::Capture * Cap,llvm::Value * Arg,MapCombinedInfoTy & CombinedInfo,StructRangeInfoTy & PartialStruct) const8479   void generateInfoForCapture(const CapturedStmt::Capture *Cap,
8480                               llvm::Value *Arg, MapCombinedInfoTy &CombinedInfo,
8481                               StructRangeInfoTy &PartialStruct) const {
8482     assert(!Cap->capturesVariableArrayType() &&
8483            "Not expecting to generate map info for a variable array type!");
8484 
8485     // We need to know when we generating information for the first component
8486     const ValueDecl *VD = Cap->capturesThis()
8487                               ? nullptr
8488                               : Cap->getCapturedVar()->getCanonicalDecl();
8489 
8490     // for map(to: lambda): skip here, processing it in
8491     // generateDefaultMapInfo
8492     if (LambdasMap.count(VD))
8493       return;
8494 
8495     // If this declaration appears in a is_device_ptr clause we just have to
8496     // pass the pointer by value. If it is a reference to a declaration, we just
8497     // pass its value.
8498     if (VD && (DevPointersMap.count(VD) || HasDevAddrsMap.count(VD))) {
8499       CombinedInfo.Exprs.push_back(VD);
8500       CombinedInfo.BasePointers.emplace_back(Arg);
8501       CombinedInfo.DevicePtrDecls.emplace_back(VD);
8502       CombinedInfo.DevicePointers.emplace_back(DeviceInfoTy::Pointer);
8503       CombinedInfo.Pointers.push_back(Arg);
8504       CombinedInfo.Sizes.push_back(CGF.Builder.CreateIntCast(
8505           CGF.getTypeSize(CGF.getContext().VoidPtrTy), CGF.Int64Ty,
8506           /*isSigned=*/true));
8507       CombinedInfo.Types.push_back(
8508           OpenMPOffloadMappingFlags::OMP_MAP_LITERAL |
8509           OpenMPOffloadMappingFlags::OMP_MAP_TARGET_PARAM);
8510       CombinedInfo.Mappers.push_back(nullptr);
8511       return;
8512     }
8513 
8514     using MapData =
8515         std::tuple<OMPClauseMappableExprCommon::MappableExprComponentListRef,
8516                    OpenMPMapClauseKind, ArrayRef<OpenMPMapModifierKind>, bool,
8517                    const ValueDecl *, const Expr *>;
8518     SmallVector<MapData, 4> DeclComponentLists;
8519     // For member fields list in is_device_ptr, store it in
8520     // DeclComponentLists for generating components info.
8521     static const OpenMPMapModifierKind Unknown = OMPC_MAP_MODIFIER_unknown;
8522     auto It = DevPointersMap.find(VD);
8523     if (It != DevPointersMap.end())
8524       for (const auto &MCL : It->second)
8525         DeclComponentLists.emplace_back(MCL, OMPC_MAP_to, Unknown,
8526                                         /*IsImpicit = */ true, nullptr,
8527                                         nullptr);
8528     auto I = HasDevAddrsMap.find(VD);
8529     if (I != HasDevAddrsMap.end())
8530       for (const auto &MCL : I->second)
8531         DeclComponentLists.emplace_back(MCL, OMPC_MAP_tofrom, Unknown,
8532                                         /*IsImpicit = */ true, nullptr,
8533                                         nullptr);
8534     assert(CurDir.is<const OMPExecutableDirective *>() &&
8535            "Expect a executable directive");
8536     const auto *CurExecDir = CurDir.get<const OMPExecutableDirective *>();
8537     bool HasMapBasePtr = false;
8538     bool HasMapArraySec = false;
8539     for (const auto *C : CurExecDir->getClausesOfKind<OMPMapClause>()) {
8540       const auto *EI = C->getVarRefs().begin();
8541       for (const auto L : C->decl_component_lists(VD)) {
8542         const ValueDecl *VDecl, *Mapper;
8543         // The Expression is not correct if the mapping is implicit
8544         const Expr *E = (C->getMapLoc().isValid()) ? *EI : nullptr;
8545         OMPClauseMappableExprCommon::MappableExprComponentListRef Components;
8546         std::tie(VDecl, Components, Mapper) = L;
8547         assert(VDecl == VD && "We got information for the wrong declaration??");
8548         assert(!Components.empty() &&
8549                "Not expecting declaration with no component lists.");
8550         if (VD && E && VD->getType()->isAnyPointerType() && isa<DeclRefExpr>(E))
8551           HasMapBasePtr = true;
8552         if (VD && E && VD->getType()->isAnyPointerType() &&
8553             (isa<ArraySectionExpr>(E) || isa<ArraySubscriptExpr>(E)))
8554           HasMapArraySec = true;
8555         DeclComponentLists.emplace_back(Components, C->getMapType(),
8556                                         C->getMapTypeModifiers(),
8557                                         C->isImplicit(), Mapper, E);
8558         ++EI;
8559       }
8560     }
8561     llvm::stable_sort(DeclComponentLists, [](const MapData &LHS,
8562                                              const MapData &RHS) {
8563       ArrayRef<OpenMPMapModifierKind> MapModifiers = std::get<2>(LHS);
8564       OpenMPMapClauseKind MapType = std::get<1>(RHS);
8565       bool HasPresent =
8566           llvm::is_contained(MapModifiers, clang::OMPC_MAP_MODIFIER_present);
8567       bool HasAllocs = MapType == OMPC_MAP_alloc;
8568       MapModifiers = std::get<2>(RHS);
8569       MapType = std::get<1>(LHS);
8570       bool HasPresentR =
8571           llvm::is_contained(MapModifiers, clang::OMPC_MAP_MODIFIER_present);
8572       bool HasAllocsR = MapType == OMPC_MAP_alloc;
8573       return (HasPresent && !HasPresentR) || (HasAllocs && !HasAllocsR);
8574     });
8575 
8576     // Find overlapping elements (including the offset from the base element).
8577     llvm::SmallDenseMap<
8578         const MapData *,
8579         llvm::SmallVector<
8580             OMPClauseMappableExprCommon::MappableExprComponentListRef, 4>,
8581         4>
8582         OverlappedData;
8583     size_t Count = 0;
8584     for (const MapData &L : DeclComponentLists) {
8585       OMPClauseMappableExprCommon::MappableExprComponentListRef Components;
8586       OpenMPMapClauseKind MapType;
8587       ArrayRef<OpenMPMapModifierKind> MapModifiers;
8588       bool IsImplicit;
8589       const ValueDecl *Mapper;
8590       const Expr *VarRef;
8591       std::tie(Components, MapType, MapModifiers, IsImplicit, Mapper, VarRef) =
8592           L;
8593       ++Count;
8594       for (const MapData &L1 : ArrayRef(DeclComponentLists).slice(Count)) {
8595         OMPClauseMappableExprCommon::MappableExprComponentListRef Components1;
8596         std::tie(Components1, MapType, MapModifiers, IsImplicit, Mapper,
8597                  VarRef) = L1;
8598         auto CI = Components.rbegin();
8599         auto CE = Components.rend();
8600         auto SI = Components1.rbegin();
8601         auto SE = Components1.rend();
8602         for (; CI != CE && SI != SE; ++CI, ++SI) {
8603           if (CI->getAssociatedExpression()->getStmtClass() !=
8604               SI->getAssociatedExpression()->getStmtClass())
8605             break;
8606           // Are we dealing with different variables/fields?
8607           if (CI->getAssociatedDeclaration() != SI->getAssociatedDeclaration())
8608             break;
8609         }
8610         // Found overlapping if, at least for one component, reached the head
8611         // of the components list.
8612         if (CI == CE || SI == SE) {
8613           // Ignore it if it is the same component.
8614           if (CI == CE && SI == SE)
8615             continue;
8616           const auto It = (SI == SE) ? CI : SI;
8617           // If one component is a pointer and another one is a kind of
8618           // dereference of this pointer (array subscript, section, dereference,
8619           // etc.), it is not an overlapping.
8620           // Same, if one component is a base and another component is a
8621           // dereferenced pointer memberexpr with the same base.
8622           if (!isa<MemberExpr>(It->getAssociatedExpression()) ||
8623               (std::prev(It)->getAssociatedDeclaration() &&
8624                std::prev(It)
8625                    ->getAssociatedDeclaration()
8626                    ->getType()
8627                    ->isPointerType()) ||
8628               (It->getAssociatedDeclaration() &&
8629                It->getAssociatedDeclaration()->getType()->isPointerType() &&
8630                std::next(It) != CE && std::next(It) != SE))
8631             continue;
8632           const MapData &BaseData = CI == CE ? L : L1;
8633           OMPClauseMappableExprCommon::MappableExprComponentListRef SubData =
8634               SI == SE ? Components : Components1;
8635           auto &OverlappedElements = OverlappedData.FindAndConstruct(&BaseData);
8636           OverlappedElements.getSecond().push_back(SubData);
8637         }
8638       }
8639     }
8640     // Sort the overlapped elements for each item.
8641     llvm::SmallVector<const FieldDecl *, 4> Layout;
8642     if (!OverlappedData.empty()) {
8643       const Type *BaseType = VD->getType().getCanonicalType().getTypePtr();
8644       const Type *OrigType = BaseType->getPointeeOrArrayElementType();
8645       while (BaseType != OrigType) {
8646         BaseType = OrigType->getCanonicalTypeInternal().getTypePtr();
8647         OrigType = BaseType->getPointeeOrArrayElementType();
8648       }
8649 
8650       if (const auto *CRD = BaseType->getAsCXXRecordDecl())
8651         getPlainLayout(CRD, Layout, /*AsBase=*/false);
8652       else {
8653         const auto *RD = BaseType->getAsRecordDecl();
8654         Layout.append(RD->field_begin(), RD->field_end());
8655       }
8656     }
8657     for (auto &Pair : OverlappedData) {
8658       llvm::stable_sort(
8659           Pair.getSecond(),
8660           [&Layout](
8661               OMPClauseMappableExprCommon::MappableExprComponentListRef First,
8662               OMPClauseMappableExprCommon::MappableExprComponentListRef
8663                   Second) {
8664             auto CI = First.rbegin();
8665             auto CE = First.rend();
8666             auto SI = Second.rbegin();
8667             auto SE = Second.rend();
8668             for (; CI != CE && SI != SE; ++CI, ++SI) {
8669               if (CI->getAssociatedExpression()->getStmtClass() !=
8670                   SI->getAssociatedExpression()->getStmtClass())
8671                 break;
8672               // Are we dealing with different variables/fields?
8673               if (CI->getAssociatedDeclaration() !=
8674                   SI->getAssociatedDeclaration())
8675                 break;
8676             }
8677 
8678             // Lists contain the same elements.
8679             if (CI == CE && SI == SE)
8680               return false;
8681 
8682             // List with less elements is less than list with more elements.
8683             if (CI == CE || SI == SE)
8684               return CI == CE;
8685 
8686             const auto *FD1 = cast<FieldDecl>(CI->getAssociatedDeclaration());
8687             const auto *FD2 = cast<FieldDecl>(SI->getAssociatedDeclaration());
8688             if (FD1->getParent() == FD2->getParent())
8689               return FD1->getFieldIndex() < FD2->getFieldIndex();
8690             const auto *It =
8691                 llvm::find_if(Layout, [FD1, FD2](const FieldDecl *FD) {
8692                   return FD == FD1 || FD == FD2;
8693                 });
8694             return *It == FD1;
8695           });
8696     }
8697 
8698     // Associated with a capture, because the mapping flags depend on it.
8699     // Go through all of the elements with the overlapped elements.
8700     bool IsFirstComponentList = true;
8701     MapCombinedInfoTy StructBaseCombinedInfo;
8702     for (const auto &Pair : OverlappedData) {
8703       const MapData &L = *Pair.getFirst();
8704       OMPClauseMappableExprCommon::MappableExprComponentListRef Components;
8705       OpenMPMapClauseKind MapType;
8706       ArrayRef<OpenMPMapModifierKind> MapModifiers;
8707       bool IsImplicit;
8708       const ValueDecl *Mapper;
8709       const Expr *VarRef;
8710       std::tie(Components, MapType, MapModifiers, IsImplicit, Mapper, VarRef) =
8711           L;
8712       ArrayRef<OMPClauseMappableExprCommon::MappableExprComponentListRef>
8713           OverlappedComponents = Pair.getSecond();
8714       generateInfoForComponentList(
8715           MapType, MapModifiers, std::nullopt, Components, CombinedInfo,
8716           StructBaseCombinedInfo, PartialStruct, IsFirstComponentList,
8717           IsImplicit, /*GenerateAllInfoForClauses*/ false, Mapper,
8718           /*ForDeviceAddr=*/false, VD, VarRef, OverlappedComponents);
8719       IsFirstComponentList = false;
8720     }
8721     // Go through other elements without overlapped elements.
8722     for (const MapData &L : DeclComponentLists) {
8723       OMPClauseMappableExprCommon::MappableExprComponentListRef Components;
8724       OpenMPMapClauseKind MapType;
8725       ArrayRef<OpenMPMapModifierKind> MapModifiers;
8726       bool IsImplicit;
8727       const ValueDecl *Mapper;
8728       const Expr *VarRef;
8729       std::tie(Components, MapType, MapModifiers, IsImplicit, Mapper, VarRef) =
8730           L;
8731       auto It = OverlappedData.find(&L);
8732       if (It == OverlappedData.end())
8733         generateInfoForComponentList(
8734             MapType, MapModifiers, std::nullopt, Components, CombinedInfo,
8735             StructBaseCombinedInfo, PartialStruct, IsFirstComponentList,
8736             IsImplicit, /*GenerateAllInfoForClauses*/ false, Mapper,
8737             /*ForDeviceAddr=*/false, VD, VarRef,
8738             /*OverlappedElements*/ std::nullopt,
8739             HasMapBasePtr && HasMapArraySec);
8740       IsFirstComponentList = false;
8741     }
8742   }
8743 
8744   /// Generate the default map information for a given capture \a CI,
8745   /// record field declaration \a RI and captured value \a CV.
generateDefaultMapInfo(const CapturedStmt::Capture & CI,const FieldDecl & RI,llvm::Value * CV,MapCombinedInfoTy & CombinedInfo) const8746   void generateDefaultMapInfo(const CapturedStmt::Capture &CI,
8747                               const FieldDecl &RI, llvm::Value *CV,
8748                               MapCombinedInfoTy &CombinedInfo) const {
8749     bool IsImplicit = true;
8750     // Do the default mapping.
8751     if (CI.capturesThis()) {
8752       CombinedInfo.Exprs.push_back(nullptr);
8753       CombinedInfo.BasePointers.push_back(CV);
8754       CombinedInfo.DevicePtrDecls.push_back(nullptr);
8755       CombinedInfo.DevicePointers.push_back(DeviceInfoTy::None);
8756       CombinedInfo.Pointers.push_back(CV);
8757       const auto *PtrTy = cast<PointerType>(RI.getType().getTypePtr());
8758       CombinedInfo.Sizes.push_back(
8759           CGF.Builder.CreateIntCast(CGF.getTypeSize(PtrTy->getPointeeType()),
8760                                     CGF.Int64Ty, /*isSigned=*/true));
8761       // Default map type.
8762       CombinedInfo.Types.push_back(OpenMPOffloadMappingFlags::OMP_MAP_TO |
8763                                    OpenMPOffloadMappingFlags::OMP_MAP_FROM);
8764     } else if (CI.capturesVariableByCopy()) {
8765       const VarDecl *VD = CI.getCapturedVar();
8766       CombinedInfo.Exprs.push_back(VD->getCanonicalDecl());
8767       CombinedInfo.BasePointers.push_back(CV);
8768       CombinedInfo.DevicePtrDecls.push_back(nullptr);
8769       CombinedInfo.DevicePointers.push_back(DeviceInfoTy::None);
8770       CombinedInfo.Pointers.push_back(CV);
8771       if (!RI.getType()->isAnyPointerType()) {
8772         // We have to signal to the runtime captures passed by value that are
8773         // not pointers.
8774         CombinedInfo.Types.push_back(
8775             OpenMPOffloadMappingFlags::OMP_MAP_LITERAL);
8776         CombinedInfo.Sizes.push_back(CGF.Builder.CreateIntCast(
8777             CGF.getTypeSize(RI.getType()), CGF.Int64Ty, /*isSigned=*/true));
8778       } else {
8779         // Pointers are implicitly mapped with a zero size and no flags
8780         // (other than first map that is added for all implicit maps).
8781         CombinedInfo.Types.push_back(OpenMPOffloadMappingFlags::OMP_MAP_NONE);
8782         CombinedInfo.Sizes.push_back(llvm::Constant::getNullValue(CGF.Int64Ty));
8783       }
8784       auto I = FirstPrivateDecls.find(VD);
8785       if (I != FirstPrivateDecls.end())
8786         IsImplicit = I->getSecond();
8787     } else {
8788       assert(CI.capturesVariable() && "Expected captured reference.");
8789       const auto *PtrTy = cast<ReferenceType>(RI.getType().getTypePtr());
8790       QualType ElementType = PtrTy->getPointeeType();
8791       CombinedInfo.Sizes.push_back(CGF.Builder.CreateIntCast(
8792           CGF.getTypeSize(ElementType), CGF.Int64Ty, /*isSigned=*/true));
8793       // The default map type for a scalar/complex type is 'to' because by
8794       // default the value doesn't have to be retrieved. For an aggregate
8795       // type, the default is 'tofrom'.
8796       CombinedInfo.Types.push_back(getMapModifiersForPrivateClauses(CI));
8797       const VarDecl *VD = CI.getCapturedVar();
8798       auto I = FirstPrivateDecls.find(VD);
8799       CombinedInfo.Exprs.push_back(VD->getCanonicalDecl());
8800       CombinedInfo.BasePointers.push_back(CV);
8801       CombinedInfo.DevicePtrDecls.push_back(nullptr);
8802       CombinedInfo.DevicePointers.push_back(DeviceInfoTy::None);
8803       if (I != FirstPrivateDecls.end() && ElementType->isAnyPointerType()) {
8804         Address PtrAddr = CGF.EmitLoadOfReference(CGF.MakeAddrLValue(
8805             CV, ElementType, CGF.getContext().getDeclAlign(VD),
8806             AlignmentSource::Decl));
8807         CombinedInfo.Pointers.push_back(PtrAddr.emitRawPointer(CGF));
8808       } else {
8809         CombinedInfo.Pointers.push_back(CV);
8810       }
8811       if (I != FirstPrivateDecls.end())
8812         IsImplicit = I->getSecond();
8813     }
8814     // Every default map produces a single argument which is a target parameter.
8815     CombinedInfo.Types.back() |=
8816         OpenMPOffloadMappingFlags::OMP_MAP_TARGET_PARAM;
8817 
8818     // Add flag stating this is an implicit map.
8819     if (IsImplicit)
8820       CombinedInfo.Types.back() |= OpenMPOffloadMappingFlags::OMP_MAP_IMPLICIT;
8821 
8822     // No user-defined mapper for default mapping.
8823     CombinedInfo.Mappers.push_back(nullptr);
8824   }
8825 };
8826 } // anonymous namespace
8827 
8828 // Try to extract the base declaration from a `this->x` expression if possible.
getDeclFromThisExpr(const Expr * E)8829 static ValueDecl *getDeclFromThisExpr(const Expr *E) {
8830   if (!E)
8831     return nullptr;
8832 
8833   if (const auto *OASE = dyn_cast<ArraySectionExpr>(E->IgnoreParenCasts()))
8834     if (const MemberExpr *ME =
8835             dyn_cast<MemberExpr>(OASE->getBase()->IgnoreParenImpCasts()))
8836       return ME->getMemberDecl();
8837   return nullptr;
8838 }
8839 
8840 /// Emit a string constant containing the names of the values mapped to the
8841 /// offloading runtime library.
8842 llvm::Constant *
emitMappingInformation(CodeGenFunction & CGF,llvm::OpenMPIRBuilder & OMPBuilder,MappableExprsHandler::MappingExprInfo & MapExprs)8843 emitMappingInformation(CodeGenFunction &CGF, llvm::OpenMPIRBuilder &OMPBuilder,
8844                        MappableExprsHandler::MappingExprInfo &MapExprs) {
8845 
8846   uint32_t SrcLocStrSize;
8847   if (!MapExprs.getMapDecl() && !MapExprs.getMapExpr())
8848     return OMPBuilder.getOrCreateDefaultSrcLocStr(SrcLocStrSize);
8849 
8850   SourceLocation Loc;
8851   if (!MapExprs.getMapDecl() && MapExprs.getMapExpr()) {
8852     if (const ValueDecl *VD = getDeclFromThisExpr(MapExprs.getMapExpr()))
8853       Loc = VD->getLocation();
8854     else
8855       Loc = MapExprs.getMapExpr()->getExprLoc();
8856   } else {
8857     Loc = MapExprs.getMapDecl()->getLocation();
8858   }
8859 
8860   std::string ExprName;
8861   if (MapExprs.getMapExpr()) {
8862     PrintingPolicy P(CGF.getContext().getLangOpts());
8863     llvm::raw_string_ostream OS(ExprName);
8864     MapExprs.getMapExpr()->printPretty(OS, nullptr, P);
8865     OS.flush();
8866   } else {
8867     ExprName = MapExprs.getMapDecl()->getNameAsString();
8868   }
8869 
8870   PresumedLoc PLoc = CGF.getContext().getSourceManager().getPresumedLoc(Loc);
8871   return OMPBuilder.getOrCreateSrcLocStr(PLoc.getFilename(), ExprName,
8872                                          PLoc.getLine(), PLoc.getColumn(),
8873                                          SrcLocStrSize);
8874 }
8875 
8876 /// Emit the arrays used to pass the captures and map information to the
8877 /// offloading runtime library. If there is no map or capture information,
8878 /// return nullptr by reference.
emitOffloadingArrays(CodeGenFunction & CGF,MappableExprsHandler::MapCombinedInfoTy & CombinedInfo,CGOpenMPRuntime::TargetDataInfo & Info,llvm::OpenMPIRBuilder & OMPBuilder,bool IsNonContiguous=false)8879 static void emitOffloadingArrays(
8880     CodeGenFunction &CGF, MappableExprsHandler::MapCombinedInfoTy &CombinedInfo,
8881     CGOpenMPRuntime::TargetDataInfo &Info, llvm::OpenMPIRBuilder &OMPBuilder,
8882     bool IsNonContiguous = false) {
8883   CodeGenModule &CGM = CGF.CGM;
8884 
8885   // Reset the array information.
8886   Info.clearArrayInfo();
8887   Info.NumberOfPtrs = CombinedInfo.BasePointers.size();
8888 
8889   using InsertPointTy = llvm::OpenMPIRBuilder::InsertPointTy;
8890   InsertPointTy AllocaIP(CGF.AllocaInsertPt->getParent(),
8891                          CGF.AllocaInsertPt->getIterator());
8892   InsertPointTy CodeGenIP(CGF.Builder.GetInsertBlock(),
8893                           CGF.Builder.GetInsertPoint());
8894 
8895   auto FillInfoMap = [&](MappableExprsHandler::MappingExprInfo &MapExpr) {
8896     return emitMappingInformation(CGF, OMPBuilder, MapExpr);
8897   };
8898   if (CGM.getCodeGenOpts().getDebugInfo() !=
8899       llvm::codegenoptions::NoDebugInfo) {
8900     CombinedInfo.Names.resize(CombinedInfo.Exprs.size());
8901     llvm::transform(CombinedInfo.Exprs, CombinedInfo.Names.begin(),
8902                     FillInfoMap);
8903   }
8904 
8905   auto DeviceAddrCB = [&](unsigned int I, llvm::Value *NewDecl) {
8906     if (const ValueDecl *DevVD = CombinedInfo.DevicePtrDecls[I]) {
8907       Info.CaptureDeviceAddrMap.try_emplace(DevVD, NewDecl);
8908     }
8909   };
8910 
8911   auto CustomMapperCB = [&](unsigned int I) {
8912     llvm::Value *MFunc = nullptr;
8913     if (CombinedInfo.Mappers[I]) {
8914       Info.HasMapper = true;
8915       MFunc = CGF.CGM.getOpenMPRuntime().getOrCreateUserDefinedMapperFunc(
8916           cast<OMPDeclareMapperDecl>(CombinedInfo.Mappers[I]));
8917     }
8918     return MFunc;
8919   };
8920   OMPBuilder.emitOffloadingArrays(AllocaIP, CodeGenIP, CombinedInfo, Info,
8921                                   /*IsNonContiguous=*/true, DeviceAddrCB,
8922                                   CustomMapperCB);
8923 }
8924 
8925 /// Check for inner distribute directive.
8926 static const OMPExecutableDirective *
getNestedDistributeDirective(ASTContext & Ctx,const OMPExecutableDirective & D)8927 getNestedDistributeDirective(ASTContext &Ctx, const OMPExecutableDirective &D) {
8928   const auto *CS = D.getInnermostCapturedStmt();
8929   const auto *Body =
8930       CS->getCapturedStmt()->IgnoreContainers(/*IgnoreCaptured=*/true);
8931   const Stmt *ChildStmt =
8932       CGOpenMPSIMDRuntime::getSingleCompoundChild(Ctx, Body);
8933 
8934   if (const auto *NestedDir =
8935           dyn_cast_or_null<OMPExecutableDirective>(ChildStmt)) {
8936     OpenMPDirectiveKind DKind = NestedDir->getDirectiveKind();
8937     switch (D.getDirectiveKind()) {
8938     case OMPD_target:
8939       // For now, treat 'target' with nested 'teams loop' as if it's
8940       // distributed (target teams distribute).
8941       if (isOpenMPDistributeDirective(DKind) || DKind == OMPD_teams_loop)
8942         return NestedDir;
8943       if (DKind == OMPD_teams) {
8944         Body = NestedDir->getInnermostCapturedStmt()->IgnoreContainers(
8945             /*IgnoreCaptured=*/true);
8946         if (!Body)
8947           return nullptr;
8948         ChildStmt = CGOpenMPSIMDRuntime::getSingleCompoundChild(Ctx, Body);
8949         if (const auto *NND =
8950                 dyn_cast_or_null<OMPExecutableDirective>(ChildStmt)) {
8951           DKind = NND->getDirectiveKind();
8952           if (isOpenMPDistributeDirective(DKind))
8953             return NND;
8954         }
8955       }
8956       return nullptr;
8957     case OMPD_target_teams:
8958       if (isOpenMPDistributeDirective(DKind))
8959         return NestedDir;
8960       return nullptr;
8961     case OMPD_target_parallel:
8962     case OMPD_target_simd:
8963     case OMPD_target_parallel_for:
8964     case OMPD_target_parallel_for_simd:
8965       return nullptr;
8966     case OMPD_target_teams_distribute:
8967     case OMPD_target_teams_distribute_simd:
8968     case OMPD_target_teams_distribute_parallel_for:
8969     case OMPD_target_teams_distribute_parallel_for_simd:
8970     case OMPD_parallel:
8971     case OMPD_for:
8972     case OMPD_parallel_for:
8973     case OMPD_parallel_master:
8974     case OMPD_parallel_sections:
8975     case OMPD_for_simd:
8976     case OMPD_parallel_for_simd:
8977     case OMPD_cancel:
8978     case OMPD_cancellation_point:
8979     case OMPD_ordered:
8980     case OMPD_threadprivate:
8981     case OMPD_allocate:
8982     case OMPD_task:
8983     case OMPD_simd:
8984     case OMPD_tile:
8985     case OMPD_unroll:
8986     case OMPD_sections:
8987     case OMPD_section:
8988     case OMPD_single:
8989     case OMPD_master:
8990     case OMPD_critical:
8991     case OMPD_taskyield:
8992     case OMPD_barrier:
8993     case OMPD_taskwait:
8994     case OMPD_taskgroup:
8995     case OMPD_atomic:
8996     case OMPD_flush:
8997     case OMPD_depobj:
8998     case OMPD_scan:
8999     case OMPD_teams:
9000     case OMPD_target_data:
9001     case OMPD_target_exit_data:
9002     case OMPD_target_enter_data:
9003     case OMPD_distribute:
9004     case OMPD_distribute_simd:
9005     case OMPD_distribute_parallel_for:
9006     case OMPD_distribute_parallel_for_simd:
9007     case OMPD_teams_distribute:
9008     case OMPD_teams_distribute_simd:
9009     case OMPD_teams_distribute_parallel_for:
9010     case OMPD_teams_distribute_parallel_for_simd:
9011     case OMPD_target_update:
9012     case OMPD_declare_simd:
9013     case OMPD_declare_variant:
9014     case OMPD_begin_declare_variant:
9015     case OMPD_end_declare_variant:
9016     case OMPD_declare_target:
9017     case OMPD_end_declare_target:
9018     case OMPD_declare_reduction:
9019     case OMPD_declare_mapper:
9020     case OMPD_taskloop:
9021     case OMPD_taskloop_simd:
9022     case OMPD_master_taskloop:
9023     case OMPD_master_taskloop_simd:
9024     case OMPD_parallel_master_taskloop:
9025     case OMPD_parallel_master_taskloop_simd:
9026     case OMPD_requires:
9027     case OMPD_metadirective:
9028     case OMPD_unknown:
9029     default:
9030       llvm_unreachable("Unexpected directive.");
9031     }
9032   }
9033 
9034   return nullptr;
9035 }
9036 
9037 /// Emit the user-defined mapper function. The code generation follows the
9038 /// pattern in the example below.
9039 /// \code
9040 /// void .omp_mapper.<type_name>.<mapper_id>.(void *rt_mapper_handle,
9041 ///                                           void *base, void *begin,
9042 ///                                           int64_t size, int64_t type,
9043 ///                                           void *name = nullptr) {
9044 ///   // Allocate space for an array section first or add a base/begin for
9045 ///   // pointer dereference.
9046 ///   if ((size > 1 || (base != begin && maptype.IsPtrAndObj)) &&
9047 ///       !maptype.IsDelete)
9048 ///     __tgt_push_mapper_component(rt_mapper_handle, base, begin,
9049 ///                                 size*sizeof(Ty), clearToFromMember(type));
9050 ///   // Map members.
9051 ///   for (unsigned i = 0; i < size; i++) {
9052 ///     // For each component specified by this mapper:
9053 ///     for (auto c : begin[i]->all_components) {
9054 ///       if (c.hasMapper())
9055 ///         (*c.Mapper())(rt_mapper_handle, c.arg_base, c.arg_begin, c.arg_size,
9056 ///                       c.arg_type, c.arg_name);
9057 ///       else
9058 ///         __tgt_push_mapper_component(rt_mapper_handle, c.arg_base,
9059 ///                                     c.arg_begin, c.arg_size, c.arg_type,
9060 ///                                     c.arg_name);
9061 ///     }
9062 ///   }
9063 ///   // Delete the array section.
9064 ///   if (size > 1 && maptype.IsDelete)
9065 ///     __tgt_push_mapper_component(rt_mapper_handle, base, begin,
9066 ///                                 size*sizeof(Ty), clearToFromMember(type));
9067 /// }
9068 /// \endcode
emitUserDefinedMapper(const OMPDeclareMapperDecl * D,CodeGenFunction * CGF)9069 void CGOpenMPRuntime::emitUserDefinedMapper(const OMPDeclareMapperDecl *D,
9070                                             CodeGenFunction *CGF) {
9071   if (UDMMap.count(D) > 0)
9072     return;
9073   ASTContext &C = CGM.getContext();
9074   QualType Ty = D->getType();
9075   QualType PtrTy = C.getPointerType(Ty).withRestrict();
9076   QualType Int64Ty = C.getIntTypeForBitwidth(/*DestWidth=*/64, /*Signed=*/true);
9077   auto *MapperVarDecl =
9078       cast<VarDecl>(cast<DeclRefExpr>(D->getMapperVarRef())->getDecl());
9079   SourceLocation Loc = D->getLocation();
9080   CharUnits ElementSize = C.getTypeSizeInChars(Ty);
9081   llvm::Type *ElemTy = CGM.getTypes().ConvertTypeForMem(Ty);
9082 
9083   // Prepare mapper function arguments and attributes.
9084   ImplicitParamDecl HandleArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr,
9085                               C.VoidPtrTy, ImplicitParamKind::Other);
9086   ImplicitParamDecl BaseArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, C.VoidPtrTy,
9087                             ImplicitParamKind::Other);
9088   ImplicitParamDecl BeginArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr,
9089                              C.VoidPtrTy, ImplicitParamKind::Other);
9090   ImplicitParamDecl SizeArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, Int64Ty,
9091                             ImplicitParamKind::Other);
9092   ImplicitParamDecl TypeArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, Int64Ty,
9093                             ImplicitParamKind::Other);
9094   ImplicitParamDecl NameArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, C.VoidPtrTy,
9095                             ImplicitParamKind::Other);
9096   FunctionArgList Args;
9097   Args.push_back(&HandleArg);
9098   Args.push_back(&BaseArg);
9099   Args.push_back(&BeginArg);
9100   Args.push_back(&SizeArg);
9101   Args.push_back(&TypeArg);
9102   Args.push_back(&NameArg);
9103   const CGFunctionInfo &FnInfo =
9104       CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args);
9105   llvm::FunctionType *FnTy = CGM.getTypes().GetFunctionType(FnInfo);
9106   SmallString<64> TyStr;
9107   llvm::raw_svector_ostream Out(TyStr);
9108   CGM.getCXXABI().getMangleContext().mangleCanonicalTypeName(Ty, Out);
9109   std::string Name = getName({"omp_mapper", TyStr, D->getName()});
9110   auto *Fn = llvm::Function::Create(FnTy, llvm::GlobalValue::InternalLinkage,
9111                                     Name, &CGM.getModule());
9112   CGM.SetInternalFunctionAttributes(GlobalDecl(), Fn, FnInfo);
9113   Fn->removeFnAttr(llvm::Attribute::OptimizeNone);
9114   // Start the mapper function code generation.
9115   CodeGenFunction MapperCGF(CGM);
9116   MapperCGF.StartFunction(GlobalDecl(), C.VoidTy, Fn, FnInfo, Args, Loc, Loc);
9117   // Compute the starting and end addresses of array elements.
9118   llvm::Value *Size = MapperCGF.EmitLoadOfScalar(
9119       MapperCGF.GetAddrOfLocalVar(&SizeArg), /*Volatile=*/false,
9120       C.getPointerType(Int64Ty), Loc);
9121   // Prepare common arguments for array initiation and deletion.
9122   llvm::Value *Handle = MapperCGF.EmitLoadOfScalar(
9123       MapperCGF.GetAddrOfLocalVar(&HandleArg),
9124       /*Volatile=*/false, C.getPointerType(C.VoidPtrTy), Loc);
9125   llvm::Value *BaseIn = MapperCGF.EmitLoadOfScalar(
9126       MapperCGF.GetAddrOfLocalVar(&BaseArg),
9127       /*Volatile=*/false, C.getPointerType(C.VoidPtrTy), Loc);
9128   llvm::Value *BeginIn = MapperCGF.EmitLoadOfScalar(
9129       MapperCGF.GetAddrOfLocalVar(&BeginArg),
9130       /*Volatile=*/false, C.getPointerType(C.VoidPtrTy), Loc);
9131   // Convert the size in bytes into the number of array elements.
9132   Size = MapperCGF.Builder.CreateExactUDiv(
9133       Size, MapperCGF.Builder.getInt64(ElementSize.getQuantity()));
9134   llvm::Value *PtrBegin = MapperCGF.Builder.CreateBitCast(
9135       BeginIn, CGM.getTypes().ConvertTypeForMem(PtrTy));
9136   llvm::Value *PtrEnd = MapperCGF.Builder.CreateGEP(ElemTy, PtrBegin, Size);
9137   llvm::Value *MapType = MapperCGF.EmitLoadOfScalar(
9138       MapperCGF.GetAddrOfLocalVar(&TypeArg), /*Volatile=*/false,
9139       C.getPointerType(Int64Ty), Loc);
9140   llvm::Value *MapName = MapperCGF.EmitLoadOfScalar(
9141       MapperCGF.GetAddrOfLocalVar(&NameArg),
9142       /*Volatile=*/false, C.getPointerType(C.VoidPtrTy), Loc);
9143 
9144   // Emit array initiation if this is an array section and \p MapType indicates
9145   // that memory allocation is required.
9146   llvm::BasicBlock *HeadBB = MapperCGF.createBasicBlock("omp.arraymap.head");
9147   emitUDMapperArrayInitOrDel(MapperCGF, Handle, BaseIn, BeginIn, Size, MapType,
9148                              MapName, ElementSize, HeadBB, /*IsInit=*/true);
9149 
9150   // Emit a for loop to iterate through SizeArg of elements and map all of them.
9151 
9152   // Emit the loop header block.
9153   MapperCGF.EmitBlock(HeadBB);
9154   llvm::BasicBlock *BodyBB = MapperCGF.createBasicBlock("omp.arraymap.body");
9155   llvm::BasicBlock *DoneBB = MapperCGF.createBasicBlock("omp.done");
9156   // Evaluate whether the initial condition is satisfied.
9157   llvm::Value *IsEmpty =
9158       MapperCGF.Builder.CreateICmpEQ(PtrBegin, PtrEnd, "omp.arraymap.isempty");
9159   MapperCGF.Builder.CreateCondBr(IsEmpty, DoneBB, BodyBB);
9160   llvm::BasicBlock *EntryBB = MapperCGF.Builder.GetInsertBlock();
9161 
9162   // Emit the loop body block.
9163   MapperCGF.EmitBlock(BodyBB);
9164   llvm::BasicBlock *LastBB = BodyBB;
9165   llvm::PHINode *PtrPHI = MapperCGF.Builder.CreatePHI(
9166       PtrBegin->getType(), 2, "omp.arraymap.ptrcurrent");
9167   PtrPHI->addIncoming(PtrBegin, EntryBB);
9168   Address PtrCurrent(PtrPHI, ElemTy,
9169                      MapperCGF.GetAddrOfLocalVar(&BeginArg)
9170                          .getAlignment()
9171                          .alignmentOfArrayElement(ElementSize));
9172   // Privatize the declared variable of mapper to be the current array element.
9173   CodeGenFunction::OMPPrivateScope Scope(MapperCGF);
9174   Scope.addPrivate(MapperVarDecl, PtrCurrent);
9175   (void)Scope.Privatize();
9176 
9177   // Get map clause information. Fill up the arrays with all mapped variables.
9178   MappableExprsHandler::MapCombinedInfoTy Info;
9179   MappableExprsHandler MEHandler(*D, MapperCGF);
9180   MEHandler.generateAllInfoForMapper(Info, OMPBuilder);
9181 
9182   // Call the runtime API __tgt_mapper_num_components to get the number of
9183   // pre-existing components.
9184   llvm::Value *OffloadingArgs[] = {Handle};
9185   llvm::Value *PreviousSize = MapperCGF.EmitRuntimeCall(
9186       OMPBuilder.getOrCreateRuntimeFunction(CGM.getModule(),
9187                                             OMPRTL___tgt_mapper_num_components),
9188       OffloadingArgs);
9189   llvm::Value *ShiftedPreviousSize = MapperCGF.Builder.CreateShl(
9190       PreviousSize,
9191       MapperCGF.Builder.getInt64(MappableExprsHandler::getFlagMemberOffset()));
9192 
9193   // Fill up the runtime mapper handle for all components.
9194   for (unsigned I = 0; I < Info.BasePointers.size(); ++I) {
9195     llvm::Value *CurBaseArg = MapperCGF.Builder.CreateBitCast(
9196         Info.BasePointers[I], CGM.getTypes().ConvertTypeForMem(C.VoidPtrTy));
9197     llvm::Value *CurBeginArg = MapperCGF.Builder.CreateBitCast(
9198         Info.Pointers[I], CGM.getTypes().ConvertTypeForMem(C.VoidPtrTy));
9199     llvm::Value *CurSizeArg = Info.Sizes[I];
9200     llvm::Value *CurNameArg =
9201         (CGM.getCodeGenOpts().getDebugInfo() ==
9202          llvm::codegenoptions::NoDebugInfo)
9203             ? llvm::ConstantPointerNull::get(CGM.VoidPtrTy)
9204             : emitMappingInformation(MapperCGF, OMPBuilder, Info.Exprs[I]);
9205 
9206     // Extract the MEMBER_OF field from the map type.
9207     llvm::Value *OriMapType = MapperCGF.Builder.getInt64(
9208         static_cast<std::underlying_type_t<OpenMPOffloadMappingFlags>>(
9209             Info.Types[I]));
9210     llvm::Value *MemberMapType =
9211         MapperCGF.Builder.CreateNUWAdd(OriMapType, ShiftedPreviousSize);
9212 
9213     // Combine the map type inherited from user-defined mapper with that
9214     // specified in the program. According to the OMP_MAP_TO and OMP_MAP_FROM
9215     // bits of the \a MapType, which is the input argument of the mapper
9216     // function, the following code will set the OMP_MAP_TO and OMP_MAP_FROM
9217     // bits of MemberMapType.
9218     // [OpenMP 5.0], 1.2.6. map-type decay.
9219     //        | alloc |  to   | from  | tofrom | release | delete
9220     // ----------------------------------------------------------
9221     // alloc  | alloc | alloc | alloc | alloc  | release | delete
9222     // to     | alloc |  to   | alloc |   to   | release | delete
9223     // from   | alloc | alloc | from  |  from  | release | delete
9224     // tofrom | alloc |  to   | from  | tofrom | release | delete
9225     llvm::Value *LeftToFrom = MapperCGF.Builder.CreateAnd(
9226         MapType,
9227         MapperCGF.Builder.getInt64(
9228             static_cast<std::underlying_type_t<OpenMPOffloadMappingFlags>>(
9229                 OpenMPOffloadMappingFlags::OMP_MAP_TO |
9230                 OpenMPOffloadMappingFlags::OMP_MAP_FROM)));
9231     llvm::BasicBlock *AllocBB = MapperCGF.createBasicBlock("omp.type.alloc");
9232     llvm::BasicBlock *AllocElseBB =
9233         MapperCGF.createBasicBlock("omp.type.alloc.else");
9234     llvm::BasicBlock *ToBB = MapperCGF.createBasicBlock("omp.type.to");
9235     llvm::BasicBlock *ToElseBB = MapperCGF.createBasicBlock("omp.type.to.else");
9236     llvm::BasicBlock *FromBB = MapperCGF.createBasicBlock("omp.type.from");
9237     llvm::BasicBlock *EndBB = MapperCGF.createBasicBlock("omp.type.end");
9238     llvm::Value *IsAlloc = MapperCGF.Builder.CreateIsNull(LeftToFrom);
9239     MapperCGF.Builder.CreateCondBr(IsAlloc, AllocBB, AllocElseBB);
9240     // In case of alloc, clear OMP_MAP_TO and OMP_MAP_FROM.
9241     MapperCGF.EmitBlock(AllocBB);
9242     llvm::Value *AllocMapType = MapperCGF.Builder.CreateAnd(
9243         MemberMapType,
9244         MapperCGF.Builder.getInt64(
9245             ~static_cast<std::underlying_type_t<OpenMPOffloadMappingFlags>>(
9246                 OpenMPOffloadMappingFlags::OMP_MAP_TO |
9247                 OpenMPOffloadMappingFlags::OMP_MAP_FROM)));
9248     MapperCGF.Builder.CreateBr(EndBB);
9249     MapperCGF.EmitBlock(AllocElseBB);
9250     llvm::Value *IsTo = MapperCGF.Builder.CreateICmpEQ(
9251         LeftToFrom,
9252         MapperCGF.Builder.getInt64(
9253             static_cast<std::underlying_type_t<OpenMPOffloadMappingFlags>>(
9254                 OpenMPOffloadMappingFlags::OMP_MAP_TO)));
9255     MapperCGF.Builder.CreateCondBr(IsTo, ToBB, ToElseBB);
9256     // In case of to, clear OMP_MAP_FROM.
9257     MapperCGF.EmitBlock(ToBB);
9258     llvm::Value *ToMapType = MapperCGF.Builder.CreateAnd(
9259         MemberMapType,
9260         MapperCGF.Builder.getInt64(
9261             ~static_cast<std::underlying_type_t<OpenMPOffloadMappingFlags>>(
9262                 OpenMPOffloadMappingFlags::OMP_MAP_FROM)));
9263     MapperCGF.Builder.CreateBr(EndBB);
9264     MapperCGF.EmitBlock(ToElseBB);
9265     llvm::Value *IsFrom = MapperCGF.Builder.CreateICmpEQ(
9266         LeftToFrom,
9267         MapperCGF.Builder.getInt64(
9268             static_cast<std::underlying_type_t<OpenMPOffloadMappingFlags>>(
9269                 OpenMPOffloadMappingFlags::OMP_MAP_FROM)));
9270     MapperCGF.Builder.CreateCondBr(IsFrom, FromBB, EndBB);
9271     // In case of from, clear OMP_MAP_TO.
9272     MapperCGF.EmitBlock(FromBB);
9273     llvm::Value *FromMapType = MapperCGF.Builder.CreateAnd(
9274         MemberMapType,
9275         MapperCGF.Builder.getInt64(
9276             ~static_cast<std::underlying_type_t<OpenMPOffloadMappingFlags>>(
9277                 OpenMPOffloadMappingFlags::OMP_MAP_TO)));
9278     // In case of tofrom, do nothing.
9279     MapperCGF.EmitBlock(EndBB);
9280     LastBB = EndBB;
9281     llvm::PHINode *CurMapType =
9282         MapperCGF.Builder.CreatePHI(CGM.Int64Ty, 4, "omp.maptype");
9283     CurMapType->addIncoming(AllocMapType, AllocBB);
9284     CurMapType->addIncoming(ToMapType, ToBB);
9285     CurMapType->addIncoming(FromMapType, FromBB);
9286     CurMapType->addIncoming(MemberMapType, ToElseBB);
9287 
9288     llvm::Value *OffloadingArgs[] = {Handle,     CurBaseArg, CurBeginArg,
9289                                      CurSizeArg, CurMapType, CurNameArg};
9290     if (Info.Mappers[I]) {
9291       // Call the corresponding mapper function.
9292       llvm::Function *MapperFunc = getOrCreateUserDefinedMapperFunc(
9293           cast<OMPDeclareMapperDecl>(Info.Mappers[I]));
9294       assert(MapperFunc && "Expect a valid mapper function is available.");
9295       MapperCGF.EmitNounwindRuntimeCall(MapperFunc, OffloadingArgs);
9296     } else {
9297       // Call the runtime API __tgt_push_mapper_component to fill up the runtime
9298       // data structure.
9299       MapperCGF.EmitRuntimeCall(
9300           OMPBuilder.getOrCreateRuntimeFunction(
9301               CGM.getModule(), OMPRTL___tgt_push_mapper_component),
9302           OffloadingArgs);
9303     }
9304   }
9305 
9306   // Update the pointer to point to the next element that needs to be mapped,
9307   // and check whether we have mapped all elements.
9308   llvm::Value *PtrNext = MapperCGF.Builder.CreateConstGEP1_32(
9309       ElemTy, PtrPHI, /*Idx0=*/1, "omp.arraymap.next");
9310   PtrPHI->addIncoming(PtrNext, LastBB);
9311   llvm::Value *IsDone =
9312       MapperCGF.Builder.CreateICmpEQ(PtrNext, PtrEnd, "omp.arraymap.isdone");
9313   llvm::BasicBlock *ExitBB = MapperCGF.createBasicBlock("omp.arraymap.exit");
9314   MapperCGF.Builder.CreateCondBr(IsDone, ExitBB, BodyBB);
9315 
9316   MapperCGF.EmitBlock(ExitBB);
9317   // Emit array deletion if this is an array section and \p MapType indicates
9318   // that deletion is required.
9319   emitUDMapperArrayInitOrDel(MapperCGF, Handle, BaseIn, BeginIn, Size, MapType,
9320                              MapName, ElementSize, DoneBB, /*IsInit=*/false);
9321 
9322   // Emit the function exit block.
9323   MapperCGF.EmitBlock(DoneBB, /*IsFinished=*/true);
9324   MapperCGF.FinishFunction();
9325   UDMMap.try_emplace(D, Fn);
9326   if (CGF) {
9327     auto &Decls = FunctionUDMMap.FindAndConstruct(CGF->CurFn);
9328     Decls.second.push_back(D);
9329   }
9330 }
9331 
9332 /// Emit the array initialization or deletion portion for user-defined mapper
9333 /// code generation. First, it evaluates whether an array section is mapped and
9334 /// whether the \a MapType instructs to delete this section. If \a IsInit is
9335 /// true, and \a MapType indicates to not delete this array, array
9336 /// initialization code is generated. If \a IsInit is false, and \a MapType
9337 /// indicates to not this array, array deletion code is generated.
emitUDMapperArrayInitOrDel(CodeGenFunction & MapperCGF,llvm::Value * Handle,llvm::Value * Base,llvm::Value * Begin,llvm::Value * Size,llvm::Value * MapType,llvm::Value * MapName,CharUnits ElementSize,llvm::BasicBlock * ExitBB,bool IsInit)9338 void CGOpenMPRuntime::emitUDMapperArrayInitOrDel(
9339     CodeGenFunction &MapperCGF, llvm::Value *Handle, llvm::Value *Base,
9340     llvm::Value *Begin, llvm::Value *Size, llvm::Value *MapType,
9341     llvm::Value *MapName, CharUnits ElementSize, llvm::BasicBlock *ExitBB,
9342     bool IsInit) {
9343   StringRef Prefix = IsInit ? ".init" : ".del";
9344 
9345   // Evaluate if this is an array section.
9346   llvm::BasicBlock *BodyBB =
9347       MapperCGF.createBasicBlock(getName({"omp.array", Prefix}));
9348   llvm::Value *IsArray = MapperCGF.Builder.CreateICmpSGT(
9349       Size, MapperCGF.Builder.getInt64(1), "omp.arrayinit.isarray");
9350   llvm::Value *DeleteBit = MapperCGF.Builder.CreateAnd(
9351       MapType,
9352       MapperCGF.Builder.getInt64(
9353           static_cast<std::underlying_type_t<OpenMPOffloadMappingFlags>>(
9354               OpenMPOffloadMappingFlags::OMP_MAP_DELETE)));
9355   llvm::Value *DeleteCond;
9356   llvm::Value *Cond;
9357   if (IsInit) {
9358     // base != begin?
9359     llvm::Value *BaseIsBegin = MapperCGF.Builder.CreateICmpNE(Base, Begin);
9360     // IsPtrAndObj?
9361     llvm::Value *PtrAndObjBit = MapperCGF.Builder.CreateAnd(
9362         MapType,
9363         MapperCGF.Builder.getInt64(
9364             static_cast<std::underlying_type_t<OpenMPOffloadMappingFlags>>(
9365                 OpenMPOffloadMappingFlags::OMP_MAP_PTR_AND_OBJ)));
9366     PtrAndObjBit = MapperCGF.Builder.CreateIsNotNull(PtrAndObjBit);
9367     BaseIsBegin = MapperCGF.Builder.CreateAnd(BaseIsBegin, PtrAndObjBit);
9368     Cond = MapperCGF.Builder.CreateOr(IsArray, BaseIsBegin);
9369     DeleteCond = MapperCGF.Builder.CreateIsNull(
9370         DeleteBit, getName({"omp.array", Prefix, ".delete"}));
9371   } else {
9372     Cond = IsArray;
9373     DeleteCond = MapperCGF.Builder.CreateIsNotNull(
9374         DeleteBit, getName({"omp.array", Prefix, ".delete"}));
9375   }
9376   Cond = MapperCGF.Builder.CreateAnd(Cond, DeleteCond);
9377   MapperCGF.Builder.CreateCondBr(Cond, BodyBB, ExitBB);
9378 
9379   MapperCGF.EmitBlock(BodyBB);
9380   // Get the array size by multiplying element size and element number (i.e., \p
9381   // Size).
9382   llvm::Value *ArraySize = MapperCGF.Builder.CreateNUWMul(
9383       Size, MapperCGF.Builder.getInt64(ElementSize.getQuantity()));
9384   // Remove OMP_MAP_TO and OMP_MAP_FROM from the map type, so that it achieves
9385   // memory allocation/deletion purpose only.
9386   llvm::Value *MapTypeArg = MapperCGF.Builder.CreateAnd(
9387       MapType,
9388       MapperCGF.Builder.getInt64(
9389           ~static_cast<std::underlying_type_t<OpenMPOffloadMappingFlags>>(
9390               OpenMPOffloadMappingFlags::OMP_MAP_TO |
9391               OpenMPOffloadMappingFlags::OMP_MAP_FROM)));
9392   MapTypeArg = MapperCGF.Builder.CreateOr(
9393       MapTypeArg,
9394       MapperCGF.Builder.getInt64(
9395           static_cast<std::underlying_type_t<OpenMPOffloadMappingFlags>>(
9396               OpenMPOffloadMappingFlags::OMP_MAP_IMPLICIT)));
9397 
9398   // Call the runtime API __tgt_push_mapper_component to fill up the runtime
9399   // data structure.
9400   llvm::Value *OffloadingArgs[] = {Handle,    Base,       Begin,
9401                                    ArraySize, MapTypeArg, MapName};
9402   MapperCGF.EmitRuntimeCall(
9403       OMPBuilder.getOrCreateRuntimeFunction(CGM.getModule(),
9404                                             OMPRTL___tgt_push_mapper_component),
9405       OffloadingArgs);
9406 }
9407 
getOrCreateUserDefinedMapperFunc(const OMPDeclareMapperDecl * D)9408 llvm::Function *CGOpenMPRuntime::getOrCreateUserDefinedMapperFunc(
9409     const OMPDeclareMapperDecl *D) {
9410   auto I = UDMMap.find(D);
9411   if (I != UDMMap.end())
9412     return I->second;
9413   emitUserDefinedMapper(D);
9414   return UDMMap.lookup(D);
9415 }
9416 
emitTargetNumIterationsCall(CodeGenFunction & CGF,const OMPExecutableDirective & D,llvm::function_ref<llvm::Value * (CodeGenFunction & CGF,const OMPLoopDirective & D)> SizeEmitter)9417 llvm::Value *CGOpenMPRuntime::emitTargetNumIterationsCall(
9418     CodeGenFunction &CGF, const OMPExecutableDirective &D,
9419     llvm::function_ref<llvm::Value *(CodeGenFunction &CGF,
9420                                      const OMPLoopDirective &D)>
9421         SizeEmitter) {
9422   OpenMPDirectiveKind Kind = D.getDirectiveKind();
9423   const OMPExecutableDirective *TD = &D;
9424   // Get nested teams distribute kind directive, if any. For now, treat
9425   // 'target_teams_loop' as if it's really a target_teams_distribute.
9426   if ((!isOpenMPDistributeDirective(Kind) || !isOpenMPTeamsDirective(Kind)) &&
9427       Kind != OMPD_target_teams_loop)
9428     TD = getNestedDistributeDirective(CGM.getContext(), D);
9429   if (!TD)
9430     return llvm::ConstantInt::get(CGF.Int64Ty, 0);
9431 
9432   const auto *LD = cast<OMPLoopDirective>(TD);
9433   if (llvm::Value *NumIterations = SizeEmitter(CGF, *LD))
9434     return NumIterations;
9435   return llvm::ConstantInt::get(CGF.Int64Ty, 0);
9436 }
9437 
9438 static void
emitTargetCallFallback(CGOpenMPRuntime * OMPRuntime,llvm::Function * OutlinedFn,const OMPExecutableDirective & D,llvm::SmallVectorImpl<llvm::Value * > & CapturedVars,bool RequiresOuterTask,const CapturedStmt & CS,bool OffloadingMandatory,CodeGenFunction & CGF)9439 emitTargetCallFallback(CGOpenMPRuntime *OMPRuntime, llvm::Function *OutlinedFn,
9440                        const OMPExecutableDirective &D,
9441                        llvm::SmallVectorImpl<llvm::Value *> &CapturedVars,
9442                        bool RequiresOuterTask, const CapturedStmt &CS,
9443                        bool OffloadingMandatory, CodeGenFunction &CGF) {
9444   if (OffloadingMandatory) {
9445     CGF.Builder.CreateUnreachable();
9446   } else {
9447     if (RequiresOuterTask) {
9448       CapturedVars.clear();
9449       CGF.GenerateOpenMPCapturedVars(CS, CapturedVars);
9450     }
9451     OMPRuntime->emitOutlinedFunctionCall(CGF, D.getBeginLoc(), OutlinedFn,
9452                                          CapturedVars);
9453   }
9454 }
9455 
emitDeviceID(llvm::PointerIntPair<const Expr *,2,OpenMPDeviceClauseModifier> Device,CodeGenFunction & CGF)9456 static llvm::Value *emitDeviceID(
9457     llvm::PointerIntPair<const Expr *, 2, OpenMPDeviceClauseModifier> Device,
9458     CodeGenFunction &CGF) {
9459   // Emit device ID if any.
9460   llvm::Value *DeviceID;
9461   if (Device.getPointer()) {
9462     assert((Device.getInt() == OMPC_DEVICE_unknown ||
9463             Device.getInt() == OMPC_DEVICE_device_num) &&
9464            "Expected device_num modifier.");
9465     llvm::Value *DevVal = CGF.EmitScalarExpr(Device.getPointer());
9466     DeviceID =
9467         CGF.Builder.CreateIntCast(DevVal, CGF.Int64Ty, /*isSigned=*/true);
9468   } else {
9469     DeviceID = CGF.Builder.getInt64(OMP_DEVICEID_UNDEF);
9470   }
9471   return DeviceID;
9472 }
9473 
emitDynCGGroupMem(const OMPExecutableDirective & D,CodeGenFunction & CGF)9474 llvm::Value *emitDynCGGroupMem(const OMPExecutableDirective &D,
9475                                CodeGenFunction &CGF) {
9476   llvm::Value *DynCGroupMem = CGF.Builder.getInt32(0);
9477 
9478   if (auto *DynMemClause = D.getSingleClause<OMPXDynCGroupMemClause>()) {
9479     CodeGenFunction::RunCleanupsScope DynCGroupMemScope(CGF);
9480     llvm::Value *DynCGroupMemVal = CGF.EmitScalarExpr(
9481         DynMemClause->getSize(), /*IgnoreResultAssign=*/true);
9482     DynCGroupMem = CGF.Builder.CreateIntCast(DynCGroupMemVal, CGF.Int32Ty,
9483                                              /*isSigned=*/false);
9484   }
9485   return DynCGroupMem;
9486 }
9487 
emitTargetCallKernelLaunch(CGOpenMPRuntime * OMPRuntime,llvm::Function * OutlinedFn,const OMPExecutableDirective & D,llvm::SmallVectorImpl<llvm::Value * > & CapturedVars,bool RequiresOuterTask,const CapturedStmt & CS,bool OffloadingMandatory,llvm::PointerIntPair<const Expr *,2,OpenMPDeviceClauseModifier> Device,llvm::Value * OutlinedFnID,CodeGenFunction::OMPTargetDataInfo & InputInfo,llvm::Value * & MapTypesArray,llvm::Value * & MapNamesArray,llvm::function_ref<llvm::Value * (CodeGenFunction & CGF,const OMPLoopDirective & D)> SizeEmitter,CodeGenFunction & CGF,CodeGenModule & CGM)9488 static void emitTargetCallKernelLaunch(
9489     CGOpenMPRuntime *OMPRuntime, llvm::Function *OutlinedFn,
9490     const OMPExecutableDirective &D,
9491     llvm::SmallVectorImpl<llvm::Value *> &CapturedVars, bool RequiresOuterTask,
9492     const CapturedStmt &CS, bool OffloadingMandatory,
9493     llvm::PointerIntPair<const Expr *, 2, OpenMPDeviceClauseModifier> Device,
9494     llvm::Value *OutlinedFnID, CodeGenFunction::OMPTargetDataInfo &InputInfo,
9495     llvm::Value *&MapTypesArray, llvm::Value *&MapNamesArray,
9496     llvm::function_ref<llvm::Value *(CodeGenFunction &CGF,
9497                                      const OMPLoopDirective &D)>
9498         SizeEmitter,
9499     CodeGenFunction &CGF, CodeGenModule &CGM) {
9500   llvm::OpenMPIRBuilder &OMPBuilder = OMPRuntime->getOMPBuilder();
9501 
9502   // Fill up the arrays with all the captured variables.
9503   MappableExprsHandler::MapCombinedInfoTy CombinedInfo;
9504 
9505   // Get mappable expression information.
9506   MappableExprsHandler MEHandler(D, CGF);
9507   llvm::DenseMap<llvm::Value *, llvm::Value *> LambdaPointers;
9508   llvm::DenseSet<CanonicalDeclPtr<const Decl>> MappedVarSet;
9509 
9510   auto RI = CS.getCapturedRecordDecl()->field_begin();
9511   auto *CV = CapturedVars.begin();
9512   for (CapturedStmt::const_capture_iterator CI = CS.capture_begin(),
9513                                             CE = CS.capture_end();
9514        CI != CE; ++CI, ++RI, ++CV) {
9515     MappableExprsHandler::MapCombinedInfoTy CurInfo;
9516     MappableExprsHandler::StructRangeInfoTy PartialStruct;
9517 
9518     // VLA sizes are passed to the outlined region by copy and do not have map
9519     // information associated.
9520     if (CI->capturesVariableArrayType()) {
9521       CurInfo.Exprs.push_back(nullptr);
9522       CurInfo.BasePointers.push_back(*CV);
9523       CurInfo.DevicePtrDecls.push_back(nullptr);
9524       CurInfo.DevicePointers.push_back(
9525           MappableExprsHandler::DeviceInfoTy::None);
9526       CurInfo.Pointers.push_back(*CV);
9527       CurInfo.Sizes.push_back(CGF.Builder.CreateIntCast(
9528           CGF.getTypeSize(RI->getType()), CGF.Int64Ty, /*isSigned=*/true));
9529       // Copy to the device as an argument. No need to retrieve it.
9530       CurInfo.Types.push_back(OpenMPOffloadMappingFlags::OMP_MAP_LITERAL |
9531                               OpenMPOffloadMappingFlags::OMP_MAP_TARGET_PARAM |
9532                               OpenMPOffloadMappingFlags::OMP_MAP_IMPLICIT);
9533       CurInfo.Mappers.push_back(nullptr);
9534     } else {
9535       // If we have any information in the map clause, we use it, otherwise we
9536       // just do a default mapping.
9537       MEHandler.generateInfoForCapture(CI, *CV, CurInfo, PartialStruct);
9538       if (!CI->capturesThis())
9539         MappedVarSet.insert(CI->getCapturedVar());
9540       else
9541         MappedVarSet.insert(nullptr);
9542       if (CurInfo.BasePointers.empty() && !PartialStruct.Base.isValid())
9543         MEHandler.generateDefaultMapInfo(*CI, **RI, *CV, CurInfo);
9544       // Generate correct mapping for variables captured by reference in
9545       // lambdas.
9546       if (CI->capturesVariable())
9547         MEHandler.generateInfoForLambdaCaptures(CI->getCapturedVar(), *CV,
9548                                                 CurInfo, LambdaPointers);
9549     }
9550     // We expect to have at least an element of information for this capture.
9551     assert((!CurInfo.BasePointers.empty() || PartialStruct.Base.isValid()) &&
9552            "Non-existing map pointer for capture!");
9553     assert(CurInfo.BasePointers.size() == CurInfo.Pointers.size() &&
9554            CurInfo.BasePointers.size() == CurInfo.Sizes.size() &&
9555            CurInfo.BasePointers.size() == CurInfo.Types.size() &&
9556            CurInfo.BasePointers.size() == CurInfo.Mappers.size() &&
9557            "Inconsistent map information sizes!");
9558 
9559     // If there is an entry in PartialStruct it means we have a struct with
9560     // individual members mapped. Emit an extra combined entry.
9561     if (PartialStruct.Base.isValid()) {
9562       CombinedInfo.append(PartialStruct.PreliminaryMapData);
9563       MEHandler.emitCombinedEntry(
9564           CombinedInfo, CurInfo.Types, PartialStruct, CI->capturesThis(),
9565           OMPBuilder, nullptr,
9566           !PartialStruct.PreliminaryMapData.BasePointers.empty());
9567     }
9568 
9569     // We need to append the results of this capture to what we already have.
9570     CombinedInfo.append(CurInfo);
9571   }
9572   // Adjust MEMBER_OF flags for the lambdas captures.
9573   MEHandler.adjustMemberOfForLambdaCaptures(
9574       OMPBuilder, LambdaPointers, CombinedInfo.BasePointers,
9575       CombinedInfo.Pointers, CombinedInfo.Types);
9576   // Map any list items in a map clause that were not captures because they
9577   // weren't referenced within the construct.
9578   MEHandler.generateAllInfo(CombinedInfo, OMPBuilder, MappedVarSet);
9579 
9580   CGOpenMPRuntime::TargetDataInfo Info;
9581   // Fill up the arrays and create the arguments.
9582   emitOffloadingArrays(CGF, CombinedInfo, Info, OMPBuilder);
9583   bool EmitDebug = CGF.CGM.getCodeGenOpts().getDebugInfo() !=
9584                    llvm::codegenoptions::NoDebugInfo;
9585   OMPBuilder.emitOffloadingArraysArgument(CGF.Builder, Info.RTArgs, Info,
9586                                           EmitDebug,
9587                                           /*ForEndCall=*/false);
9588 
9589   InputInfo.NumberOfTargetItems = Info.NumberOfPtrs;
9590   InputInfo.BasePointersArray = Address(Info.RTArgs.BasePointersArray,
9591                                         CGF.VoidPtrTy, CGM.getPointerAlign());
9592   InputInfo.PointersArray =
9593       Address(Info.RTArgs.PointersArray, CGF.VoidPtrTy, CGM.getPointerAlign());
9594   InputInfo.SizesArray =
9595       Address(Info.RTArgs.SizesArray, CGF.Int64Ty, CGM.getPointerAlign());
9596   InputInfo.MappersArray =
9597       Address(Info.RTArgs.MappersArray, CGF.VoidPtrTy, CGM.getPointerAlign());
9598   MapTypesArray = Info.RTArgs.MapTypesArray;
9599   MapNamesArray = Info.RTArgs.MapNamesArray;
9600 
9601   auto &&ThenGen = [&OMPRuntime, OutlinedFn, &D, &CapturedVars,
9602                     RequiresOuterTask, &CS, OffloadingMandatory, Device,
9603                     OutlinedFnID, &InputInfo, &MapTypesArray, &MapNamesArray,
9604                     SizeEmitter](CodeGenFunction &CGF, PrePostActionTy &) {
9605     bool IsReverseOffloading = Device.getInt() == OMPC_DEVICE_ancestor;
9606 
9607     if (IsReverseOffloading) {
9608       // Reverse offloading is not supported, so just execute on the host.
9609       // FIXME: This fallback solution is incorrect since it ignores the
9610       // OMP_TARGET_OFFLOAD environment variable. Instead it would be better to
9611       // assert here and ensure SEMA emits an error.
9612       emitTargetCallFallback(OMPRuntime, OutlinedFn, D, CapturedVars,
9613                              RequiresOuterTask, CS, OffloadingMandatory, CGF);
9614       return;
9615     }
9616 
9617     bool HasNoWait = D.hasClausesOfKind<OMPNowaitClause>();
9618     unsigned NumTargetItems = InputInfo.NumberOfTargetItems;
9619 
9620     llvm::Value *BasePointersArray =
9621         InputInfo.BasePointersArray.emitRawPointer(CGF);
9622     llvm::Value *PointersArray = InputInfo.PointersArray.emitRawPointer(CGF);
9623     llvm::Value *SizesArray = InputInfo.SizesArray.emitRawPointer(CGF);
9624     llvm::Value *MappersArray = InputInfo.MappersArray.emitRawPointer(CGF);
9625 
9626     auto &&EmitTargetCallFallbackCB =
9627         [&OMPRuntime, OutlinedFn, &D, &CapturedVars, RequiresOuterTask, &CS,
9628          OffloadingMandatory, &CGF](llvm::OpenMPIRBuilder::InsertPointTy IP)
9629         -> llvm::OpenMPIRBuilder::InsertPointTy {
9630       CGF.Builder.restoreIP(IP);
9631       emitTargetCallFallback(OMPRuntime, OutlinedFn, D, CapturedVars,
9632                              RequiresOuterTask, CS, OffloadingMandatory, CGF);
9633       return CGF.Builder.saveIP();
9634     };
9635 
9636     llvm::Value *DeviceID = emitDeviceID(Device, CGF);
9637     llvm::Value *NumTeams = OMPRuntime->emitNumTeamsForTargetDirective(CGF, D);
9638     llvm::Value *NumThreads =
9639         OMPRuntime->emitNumThreadsForTargetDirective(CGF, D);
9640     llvm::Value *RTLoc = OMPRuntime->emitUpdateLocation(CGF, D.getBeginLoc());
9641     llvm::Value *NumIterations =
9642         OMPRuntime->emitTargetNumIterationsCall(CGF, D, SizeEmitter);
9643     llvm::Value *DynCGGroupMem = emitDynCGGroupMem(D, CGF);
9644     llvm::OpenMPIRBuilder::InsertPointTy AllocaIP(
9645         CGF.AllocaInsertPt->getParent(), CGF.AllocaInsertPt->getIterator());
9646 
9647     llvm::OpenMPIRBuilder::TargetDataRTArgs RTArgs(
9648         BasePointersArray, PointersArray, SizesArray, MapTypesArray,
9649         nullptr /* MapTypesArrayEnd */, MappersArray, MapNamesArray);
9650 
9651     llvm::OpenMPIRBuilder::TargetKernelArgs Args(
9652         NumTargetItems, RTArgs, NumIterations, NumTeams, NumThreads,
9653         DynCGGroupMem, HasNoWait);
9654 
9655     CGF.Builder.restoreIP(OMPRuntime->getOMPBuilder().emitKernelLaunch(
9656         CGF.Builder, OutlinedFn, OutlinedFnID, EmitTargetCallFallbackCB, Args,
9657         DeviceID, RTLoc, AllocaIP));
9658   };
9659 
9660   if (RequiresOuterTask)
9661     CGF.EmitOMPTargetTaskBasedDirective(D, ThenGen, InputInfo);
9662   else
9663     OMPRuntime->emitInlinedDirective(CGF, D.getDirectiveKind(), ThenGen);
9664 }
9665 
9666 static void
emitTargetCallElse(CGOpenMPRuntime * OMPRuntime,llvm::Function * OutlinedFn,const OMPExecutableDirective & D,llvm::SmallVectorImpl<llvm::Value * > & CapturedVars,bool RequiresOuterTask,const CapturedStmt & CS,bool OffloadingMandatory,CodeGenFunction & CGF)9667 emitTargetCallElse(CGOpenMPRuntime *OMPRuntime, llvm::Function *OutlinedFn,
9668                    const OMPExecutableDirective &D,
9669                    llvm::SmallVectorImpl<llvm::Value *> &CapturedVars,
9670                    bool RequiresOuterTask, const CapturedStmt &CS,
9671                    bool OffloadingMandatory, CodeGenFunction &CGF) {
9672 
9673   // Notify that the host version must be executed.
9674   auto &&ElseGen =
9675       [&OMPRuntime, OutlinedFn, &D, &CapturedVars, RequiresOuterTask, &CS,
9676        OffloadingMandatory](CodeGenFunction &CGF, PrePostActionTy &) {
9677         emitTargetCallFallback(OMPRuntime, OutlinedFn, D, CapturedVars,
9678                                RequiresOuterTask, CS, OffloadingMandatory, CGF);
9679       };
9680 
9681   if (RequiresOuterTask) {
9682     CodeGenFunction::OMPTargetDataInfo InputInfo;
9683     CGF.EmitOMPTargetTaskBasedDirective(D, ElseGen, InputInfo);
9684   } else {
9685     OMPRuntime->emitInlinedDirective(CGF, D.getDirectiveKind(), ElseGen);
9686   }
9687 }
9688 
emitTargetCall(CodeGenFunction & CGF,const OMPExecutableDirective & D,llvm::Function * OutlinedFn,llvm::Value * OutlinedFnID,const Expr * IfCond,llvm::PointerIntPair<const Expr *,2,OpenMPDeviceClauseModifier> Device,llvm::function_ref<llvm::Value * (CodeGenFunction & CGF,const OMPLoopDirective & D)> SizeEmitter)9689 void CGOpenMPRuntime::emitTargetCall(
9690     CodeGenFunction &CGF, const OMPExecutableDirective &D,
9691     llvm::Function *OutlinedFn, llvm::Value *OutlinedFnID, const Expr *IfCond,
9692     llvm::PointerIntPair<const Expr *, 2, OpenMPDeviceClauseModifier> Device,
9693     llvm::function_ref<llvm::Value *(CodeGenFunction &CGF,
9694                                      const OMPLoopDirective &D)>
9695         SizeEmitter) {
9696   if (!CGF.HaveInsertPoint())
9697     return;
9698 
9699   const bool OffloadingMandatory = !CGM.getLangOpts().OpenMPIsTargetDevice &&
9700                                    CGM.getLangOpts().OpenMPOffloadMandatory;
9701 
9702   assert((OffloadingMandatory || OutlinedFn) && "Invalid outlined function!");
9703 
9704   const bool RequiresOuterTask =
9705       D.hasClausesOfKind<OMPDependClause>() ||
9706       D.hasClausesOfKind<OMPNowaitClause>() ||
9707       D.hasClausesOfKind<OMPInReductionClause>() ||
9708       (CGM.getLangOpts().OpenMP >= 51 &&
9709        needsTaskBasedThreadLimit(D.getDirectiveKind()) &&
9710        D.hasClausesOfKind<OMPThreadLimitClause>());
9711   llvm::SmallVector<llvm::Value *, 16> CapturedVars;
9712   const CapturedStmt &CS = *D.getCapturedStmt(OMPD_target);
9713   auto &&ArgsCodegen = [&CS, &CapturedVars](CodeGenFunction &CGF,
9714                                             PrePostActionTy &) {
9715     CGF.GenerateOpenMPCapturedVars(CS, CapturedVars);
9716   };
9717   emitInlinedDirective(CGF, OMPD_unknown, ArgsCodegen);
9718 
9719   CodeGenFunction::OMPTargetDataInfo InputInfo;
9720   llvm::Value *MapTypesArray = nullptr;
9721   llvm::Value *MapNamesArray = nullptr;
9722 
9723   auto &&TargetThenGen = [this, OutlinedFn, &D, &CapturedVars,
9724                           RequiresOuterTask, &CS, OffloadingMandatory, Device,
9725                           OutlinedFnID, &InputInfo, &MapTypesArray,
9726                           &MapNamesArray, SizeEmitter](CodeGenFunction &CGF,
9727                                                        PrePostActionTy &) {
9728     emitTargetCallKernelLaunch(this, OutlinedFn, D, CapturedVars,
9729                                RequiresOuterTask, CS, OffloadingMandatory,
9730                                Device, OutlinedFnID, InputInfo, MapTypesArray,
9731                                MapNamesArray, SizeEmitter, CGF, CGM);
9732   };
9733 
9734   auto &&TargetElseGen =
9735       [this, OutlinedFn, &D, &CapturedVars, RequiresOuterTask, &CS,
9736        OffloadingMandatory](CodeGenFunction &CGF, PrePostActionTy &) {
9737         emitTargetCallElse(this, OutlinedFn, D, CapturedVars, RequiresOuterTask,
9738                            CS, OffloadingMandatory, CGF);
9739       };
9740 
9741   // If we have a target function ID it means that we need to support
9742   // offloading, otherwise, just execute on the host. We need to execute on host
9743   // regardless of the conditional in the if clause if, e.g., the user do not
9744   // specify target triples.
9745   if (OutlinedFnID) {
9746     if (IfCond) {
9747       emitIfClause(CGF, IfCond, TargetThenGen, TargetElseGen);
9748     } else {
9749       RegionCodeGenTy ThenRCG(TargetThenGen);
9750       ThenRCG(CGF);
9751     }
9752   } else {
9753     RegionCodeGenTy ElseRCG(TargetElseGen);
9754     ElseRCG(CGF);
9755   }
9756 }
9757 
scanForTargetRegionsFunctions(const Stmt * S,StringRef ParentName)9758 void CGOpenMPRuntime::scanForTargetRegionsFunctions(const Stmt *S,
9759                                                     StringRef ParentName) {
9760   if (!S)
9761     return;
9762 
9763   // Codegen OMP target directives that offload compute to the device.
9764   bool RequiresDeviceCodegen =
9765       isa<OMPExecutableDirective>(S) &&
9766       isOpenMPTargetExecutionDirective(
9767           cast<OMPExecutableDirective>(S)->getDirectiveKind());
9768 
9769   if (RequiresDeviceCodegen) {
9770     const auto &E = *cast<OMPExecutableDirective>(S);
9771 
9772     llvm::TargetRegionEntryInfo EntryInfo = getEntryInfoFromPresumedLoc(
9773         CGM, OMPBuilder, E.getBeginLoc(), ParentName);
9774 
9775     // Is this a target region that should not be emitted as an entry point? If
9776     // so just signal we are done with this target region.
9777     if (!OMPBuilder.OffloadInfoManager.hasTargetRegionEntryInfo(EntryInfo))
9778       return;
9779 
9780     switch (E.getDirectiveKind()) {
9781     case OMPD_target:
9782       CodeGenFunction::EmitOMPTargetDeviceFunction(CGM, ParentName,
9783                                                    cast<OMPTargetDirective>(E));
9784       break;
9785     case OMPD_target_parallel:
9786       CodeGenFunction::EmitOMPTargetParallelDeviceFunction(
9787           CGM, ParentName, cast<OMPTargetParallelDirective>(E));
9788       break;
9789     case OMPD_target_teams:
9790       CodeGenFunction::EmitOMPTargetTeamsDeviceFunction(
9791           CGM, ParentName, cast<OMPTargetTeamsDirective>(E));
9792       break;
9793     case OMPD_target_teams_distribute:
9794       CodeGenFunction::EmitOMPTargetTeamsDistributeDeviceFunction(
9795           CGM, ParentName, cast<OMPTargetTeamsDistributeDirective>(E));
9796       break;
9797     case OMPD_target_teams_distribute_simd:
9798       CodeGenFunction::EmitOMPTargetTeamsDistributeSimdDeviceFunction(
9799           CGM, ParentName, cast<OMPTargetTeamsDistributeSimdDirective>(E));
9800       break;
9801     case OMPD_target_parallel_for:
9802       CodeGenFunction::EmitOMPTargetParallelForDeviceFunction(
9803           CGM, ParentName, cast<OMPTargetParallelForDirective>(E));
9804       break;
9805     case OMPD_target_parallel_for_simd:
9806       CodeGenFunction::EmitOMPTargetParallelForSimdDeviceFunction(
9807           CGM, ParentName, cast<OMPTargetParallelForSimdDirective>(E));
9808       break;
9809     case OMPD_target_simd:
9810       CodeGenFunction::EmitOMPTargetSimdDeviceFunction(
9811           CGM, ParentName, cast<OMPTargetSimdDirective>(E));
9812       break;
9813     case OMPD_target_teams_distribute_parallel_for:
9814       CodeGenFunction::EmitOMPTargetTeamsDistributeParallelForDeviceFunction(
9815           CGM, ParentName,
9816           cast<OMPTargetTeamsDistributeParallelForDirective>(E));
9817       break;
9818     case OMPD_target_teams_distribute_parallel_for_simd:
9819       CodeGenFunction::
9820           EmitOMPTargetTeamsDistributeParallelForSimdDeviceFunction(
9821               CGM, ParentName,
9822               cast<OMPTargetTeamsDistributeParallelForSimdDirective>(E));
9823       break;
9824     case OMPD_target_teams_loop:
9825       CodeGenFunction::EmitOMPTargetTeamsGenericLoopDeviceFunction(
9826           CGM, ParentName, cast<OMPTargetTeamsGenericLoopDirective>(E));
9827       break;
9828     case OMPD_target_parallel_loop:
9829       CodeGenFunction::EmitOMPTargetParallelGenericLoopDeviceFunction(
9830           CGM, ParentName, cast<OMPTargetParallelGenericLoopDirective>(E));
9831       break;
9832     case OMPD_parallel:
9833     case OMPD_for:
9834     case OMPD_parallel_for:
9835     case OMPD_parallel_master:
9836     case OMPD_parallel_sections:
9837     case OMPD_for_simd:
9838     case OMPD_parallel_for_simd:
9839     case OMPD_cancel:
9840     case OMPD_cancellation_point:
9841     case OMPD_ordered:
9842     case OMPD_threadprivate:
9843     case OMPD_allocate:
9844     case OMPD_task:
9845     case OMPD_simd:
9846     case OMPD_tile:
9847     case OMPD_unroll:
9848     case OMPD_sections:
9849     case OMPD_section:
9850     case OMPD_single:
9851     case OMPD_master:
9852     case OMPD_critical:
9853     case OMPD_taskyield:
9854     case OMPD_barrier:
9855     case OMPD_taskwait:
9856     case OMPD_taskgroup:
9857     case OMPD_atomic:
9858     case OMPD_flush:
9859     case OMPD_depobj:
9860     case OMPD_scan:
9861     case OMPD_teams:
9862     case OMPD_target_data:
9863     case OMPD_target_exit_data:
9864     case OMPD_target_enter_data:
9865     case OMPD_distribute:
9866     case OMPD_distribute_simd:
9867     case OMPD_distribute_parallel_for:
9868     case OMPD_distribute_parallel_for_simd:
9869     case OMPD_teams_distribute:
9870     case OMPD_teams_distribute_simd:
9871     case OMPD_teams_distribute_parallel_for:
9872     case OMPD_teams_distribute_parallel_for_simd:
9873     case OMPD_target_update:
9874     case OMPD_declare_simd:
9875     case OMPD_declare_variant:
9876     case OMPD_begin_declare_variant:
9877     case OMPD_end_declare_variant:
9878     case OMPD_declare_target:
9879     case OMPD_end_declare_target:
9880     case OMPD_declare_reduction:
9881     case OMPD_declare_mapper:
9882     case OMPD_taskloop:
9883     case OMPD_taskloop_simd:
9884     case OMPD_master_taskloop:
9885     case OMPD_master_taskloop_simd:
9886     case OMPD_parallel_master_taskloop:
9887     case OMPD_parallel_master_taskloop_simd:
9888     case OMPD_requires:
9889     case OMPD_metadirective:
9890     case OMPD_unknown:
9891     default:
9892       llvm_unreachable("Unknown target directive for OpenMP device codegen.");
9893     }
9894     return;
9895   }
9896 
9897   if (const auto *E = dyn_cast<OMPExecutableDirective>(S)) {
9898     if (!E->hasAssociatedStmt() || !E->getAssociatedStmt())
9899       return;
9900 
9901     scanForTargetRegionsFunctions(E->getRawStmt(), ParentName);
9902     return;
9903   }
9904 
9905   // If this is a lambda function, look into its body.
9906   if (const auto *L = dyn_cast<LambdaExpr>(S))
9907     S = L->getBody();
9908 
9909   // Keep looking for target regions recursively.
9910   for (const Stmt *II : S->children())
9911     scanForTargetRegionsFunctions(II, ParentName);
9912 }
9913 
isAssumedToBeNotEmitted(const ValueDecl * VD,bool IsDevice)9914 static bool isAssumedToBeNotEmitted(const ValueDecl *VD, bool IsDevice) {
9915   std::optional<OMPDeclareTargetDeclAttr::DevTypeTy> DevTy =
9916       OMPDeclareTargetDeclAttr::getDeviceType(VD);
9917   if (!DevTy)
9918     return false;
9919   // Do not emit device_type(nohost) functions for the host.
9920   if (!IsDevice && DevTy == OMPDeclareTargetDeclAttr::DT_NoHost)
9921     return true;
9922   // Do not emit device_type(host) functions for the device.
9923   if (IsDevice && DevTy == OMPDeclareTargetDeclAttr::DT_Host)
9924     return true;
9925   return false;
9926 }
9927 
emitTargetFunctions(GlobalDecl GD)9928 bool CGOpenMPRuntime::emitTargetFunctions(GlobalDecl GD) {
9929   // If emitting code for the host, we do not process FD here. Instead we do
9930   // the normal code generation.
9931   if (!CGM.getLangOpts().OpenMPIsTargetDevice) {
9932     if (const auto *FD = dyn_cast<FunctionDecl>(GD.getDecl()))
9933       if (isAssumedToBeNotEmitted(cast<ValueDecl>(FD),
9934                                   CGM.getLangOpts().OpenMPIsTargetDevice))
9935         return true;
9936     return false;
9937   }
9938 
9939   const ValueDecl *VD = cast<ValueDecl>(GD.getDecl());
9940   // Try to detect target regions in the function.
9941   if (const auto *FD = dyn_cast<FunctionDecl>(VD)) {
9942     StringRef Name = CGM.getMangledName(GD);
9943     scanForTargetRegionsFunctions(FD->getBody(), Name);
9944     if (isAssumedToBeNotEmitted(cast<ValueDecl>(FD),
9945                                 CGM.getLangOpts().OpenMPIsTargetDevice))
9946       return true;
9947   }
9948 
9949   // Do not to emit function if it is not marked as declare target.
9950   return !OMPDeclareTargetDeclAttr::isDeclareTargetDeclaration(VD) &&
9951          AlreadyEmittedTargetDecls.count(VD) == 0;
9952 }
9953 
emitTargetGlobalVariable(GlobalDecl GD)9954 bool CGOpenMPRuntime::emitTargetGlobalVariable(GlobalDecl GD) {
9955   if (isAssumedToBeNotEmitted(cast<ValueDecl>(GD.getDecl()),
9956                               CGM.getLangOpts().OpenMPIsTargetDevice))
9957     return true;
9958 
9959   if (!CGM.getLangOpts().OpenMPIsTargetDevice)
9960     return false;
9961 
9962   // Check if there are Ctors/Dtors in this declaration and look for target
9963   // regions in it. We use the complete variant to produce the kernel name
9964   // mangling.
9965   QualType RDTy = cast<VarDecl>(GD.getDecl())->getType();
9966   if (const auto *RD = RDTy->getBaseElementTypeUnsafe()->getAsCXXRecordDecl()) {
9967     for (const CXXConstructorDecl *Ctor : RD->ctors()) {
9968       StringRef ParentName =
9969           CGM.getMangledName(GlobalDecl(Ctor, Ctor_Complete));
9970       scanForTargetRegionsFunctions(Ctor->getBody(), ParentName);
9971     }
9972     if (const CXXDestructorDecl *Dtor = RD->getDestructor()) {
9973       StringRef ParentName =
9974           CGM.getMangledName(GlobalDecl(Dtor, Dtor_Complete));
9975       scanForTargetRegionsFunctions(Dtor->getBody(), ParentName);
9976     }
9977   }
9978 
9979   // Do not to emit variable if it is not marked as declare target.
9980   std::optional<OMPDeclareTargetDeclAttr::MapTypeTy> Res =
9981       OMPDeclareTargetDeclAttr::isDeclareTargetDeclaration(
9982           cast<VarDecl>(GD.getDecl()));
9983   if (!Res || *Res == OMPDeclareTargetDeclAttr::MT_Link ||
9984       ((*Res == OMPDeclareTargetDeclAttr::MT_To ||
9985         *Res == OMPDeclareTargetDeclAttr::MT_Enter) &&
9986        HasRequiresUnifiedSharedMemory)) {
9987     DeferredGlobalVariables.insert(cast<VarDecl>(GD.getDecl()));
9988     return true;
9989   }
9990   return false;
9991 }
9992 
registerTargetGlobalVariable(const VarDecl * VD,llvm::Constant * Addr)9993 void CGOpenMPRuntime::registerTargetGlobalVariable(const VarDecl *VD,
9994                                                    llvm::Constant *Addr) {
9995   if (CGM.getLangOpts().OMPTargetTriples.empty() &&
9996       !CGM.getLangOpts().OpenMPIsTargetDevice)
9997     return;
9998 
9999   std::optional<OMPDeclareTargetDeclAttr::MapTypeTy> Res =
10000       OMPDeclareTargetDeclAttr::isDeclareTargetDeclaration(VD);
10001 
10002   // If this is an 'extern' declaration we defer to the canonical definition and
10003   // do not emit an offloading entry.
10004   if (Res && *Res != OMPDeclareTargetDeclAttr::MT_Link &&
10005       VD->hasExternalStorage())
10006     return;
10007 
10008   if (!Res) {
10009     if (CGM.getLangOpts().OpenMPIsTargetDevice) {
10010       // Register non-target variables being emitted in device code (debug info
10011       // may cause this).
10012       StringRef VarName = CGM.getMangledName(VD);
10013       EmittedNonTargetVariables.try_emplace(VarName, Addr);
10014     }
10015     return;
10016   }
10017 
10018   auto AddrOfGlobal = [&VD, this]() { return CGM.GetAddrOfGlobal(VD); };
10019   auto LinkageForVariable = [&VD, this]() {
10020     return CGM.getLLVMLinkageVarDefinition(VD);
10021   };
10022 
10023   std::vector<llvm::GlobalVariable *> GeneratedRefs;
10024   OMPBuilder.registerTargetGlobalVariable(
10025       convertCaptureClause(VD), convertDeviceClause(VD),
10026       VD->hasDefinition(CGM.getContext()) == VarDecl::DeclarationOnly,
10027       VD->isExternallyVisible(),
10028       getEntryInfoFromPresumedLoc(CGM, OMPBuilder,
10029                                   VD->getCanonicalDecl()->getBeginLoc()),
10030       CGM.getMangledName(VD), GeneratedRefs, CGM.getLangOpts().OpenMPSimd,
10031       CGM.getLangOpts().OMPTargetTriples, AddrOfGlobal, LinkageForVariable,
10032       CGM.getTypes().ConvertTypeForMem(
10033           CGM.getContext().getPointerType(VD->getType())),
10034       Addr);
10035 
10036   for (auto *ref : GeneratedRefs)
10037     CGM.addCompilerUsedGlobal(ref);
10038 }
10039 
emitTargetGlobal(GlobalDecl GD)10040 bool CGOpenMPRuntime::emitTargetGlobal(GlobalDecl GD) {
10041   if (isa<FunctionDecl>(GD.getDecl()) ||
10042       isa<OMPDeclareReductionDecl>(GD.getDecl()))
10043     return emitTargetFunctions(GD);
10044 
10045   return emitTargetGlobalVariable(GD);
10046 }
10047 
emitDeferredTargetDecls() const10048 void CGOpenMPRuntime::emitDeferredTargetDecls() const {
10049   for (const VarDecl *VD : DeferredGlobalVariables) {
10050     std::optional<OMPDeclareTargetDeclAttr::MapTypeTy> Res =
10051         OMPDeclareTargetDeclAttr::isDeclareTargetDeclaration(VD);
10052     if (!Res)
10053       continue;
10054     if ((*Res == OMPDeclareTargetDeclAttr::MT_To ||
10055          *Res == OMPDeclareTargetDeclAttr::MT_Enter) &&
10056         !HasRequiresUnifiedSharedMemory) {
10057       CGM.EmitGlobal(VD);
10058     } else {
10059       assert((*Res == OMPDeclareTargetDeclAttr::MT_Link ||
10060               ((*Res == OMPDeclareTargetDeclAttr::MT_To ||
10061                 *Res == OMPDeclareTargetDeclAttr::MT_Enter) &&
10062                HasRequiresUnifiedSharedMemory)) &&
10063              "Expected link clause or to clause with unified memory.");
10064       (void)CGM.getOpenMPRuntime().getAddrOfDeclareTargetVar(VD);
10065     }
10066   }
10067 }
10068 
adjustTargetSpecificDataForLambdas(CodeGenFunction & CGF,const OMPExecutableDirective & D) const10069 void CGOpenMPRuntime::adjustTargetSpecificDataForLambdas(
10070     CodeGenFunction &CGF, const OMPExecutableDirective &D) const {
10071   assert(isOpenMPTargetExecutionDirective(D.getDirectiveKind()) &&
10072          " Expected target-based directive.");
10073 }
10074 
processRequiresDirective(const OMPRequiresDecl * D)10075 void CGOpenMPRuntime::processRequiresDirective(const OMPRequiresDecl *D) {
10076   for (const OMPClause *Clause : D->clauselists()) {
10077     if (Clause->getClauseKind() == OMPC_unified_shared_memory) {
10078       HasRequiresUnifiedSharedMemory = true;
10079       OMPBuilder.Config.setHasRequiresUnifiedSharedMemory(true);
10080     } else if (const auto *AC =
10081                    dyn_cast<OMPAtomicDefaultMemOrderClause>(Clause)) {
10082       switch (AC->getAtomicDefaultMemOrderKind()) {
10083       case OMPC_ATOMIC_DEFAULT_MEM_ORDER_acq_rel:
10084         RequiresAtomicOrdering = llvm::AtomicOrdering::AcquireRelease;
10085         break;
10086       case OMPC_ATOMIC_DEFAULT_MEM_ORDER_seq_cst:
10087         RequiresAtomicOrdering = llvm::AtomicOrdering::SequentiallyConsistent;
10088         break;
10089       case OMPC_ATOMIC_DEFAULT_MEM_ORDER_relaxed:
10090         RequiresAtomicOrdering = llvm::AtomicOrdering::Monotonic;
10091         break;
10092       case OMPC_ATOMIC_DEFAULT_MEM_ORDER_unknown:
10093         break;
10094       }
10095     }
10096   }
10097 }
10098 
getDefaultMemoryOrdering() const10099 llvm::AtomicOrdering CGOpenMPRuntime::getDefaultMemoryOrdering() const {
10100   return RequiresAtomicOrdering;
10101 }
10102 
hasAllocateAttributeForGlobalVar(const VarDecl * VD,LangAS & AS)10103 bool CGOpenMPRuntime::hasAllocateAttributeForGlobalVar(const VarDecl *VD,
10104                                                        LangAS &AS) {
10105   if (!VD || !VD->hasAttr<OMPAllocateDeclAttr>())
10106     return false;
10107   const auto *A = VD->getAttr<OMPAllocateDeclAttr>();
10108   switch(A->getAllocatorType()) {
10109   case OMPAllocateDeclAttr::OMPNullMemAlloc:
10110   case OMPAllocateDeclAttr::OMPDefaultMemAlloc:
10111   // Not supported, fallback to the default mem space.
10112   case OMPAllocateDeclAttr::OMPLargeCapMemAlloc:
10113   case OMPAllocateDeclAttr::OMPCGroupMemAlloc:
10114   case OMPAllocateDeclAttr::OMPHighBWMemAlloc:
10115   case OMPAllocateDeclAttr::OMPLowLatMemAlloc:
10116   case OMPAllocateDeclAttr::OMPThreadMemAlloc:
10117   case OMPAllocateDeclAttr::OMPConstMemAlloc:
10118   case OMPAllocateDeclAttr::OMPPTeamMemAlloc:
10119     AS = LangAS::Default;
10120     return true;
10121   case OMPAllocateDeclAttr::OMPUserDefinedMemAlloc:
10122     llvm_unreachable("Expected predefined allocator for the variables with the "
10123                      "static storage.");
10124   }
10125   return false;
10126 }
10127 
hasRequiresUnifiedSharedMemory() const10128 bool CGOpenMPRuntime::hasRequiresUnifiedSharedMemory() const {
10129   return HasRequiresUnifiedSharedMemory;
10130 }
10131 
DisableAutoDeclareTargetRAII(CodeGenModule & CGM)10132 CGOpenMPRuntime::DisableAutoDeclareTargetRAII::DisableAutoDeclareTargetRAII(
10133     CodeGenModule &CGM)
10134     : CGM(CGM) {
10135   if (CGM.getLangOpts().OpenMPIsTargetDevice) {
10136     SavedShouldMarkAsGlobal = CGM.getOpenMPRuntime().ShouldMarkAsGlobal;
10137     CGM.getOpenMPRuntime().ShouldMarkAsGlobal = false;
10138   }
10139 }
10140 
~DisableAutoDeclareTargetRAII()10141 CGOpenMPRuntime::DisableAutoDeclareTargetRAII::~DisableAutoDeclareTargetRAII() {
10142   if (CGM.getLangOpts().OpenMPIsTargetDevice)
10143     CGM.getOpenMPRuntime().ShouldMarkAsGlobal = SavedShouldMarkAsGlobal;
10144 }
10145 
markAsGlobalTarget(GlobalDecl GD)10146 bool CGOpenMPRuntime::markAsGlobalTarget(GlobalDecl GD) {
10147   if (!CGM.getLangOpts().OpenMPIsTargetDevice || !ShouldMarkAsGlobal)
10148     return true;
10149 
10150   const auto *D = cast<FunctionDecl>(GD.getDecl());
10151   // Do not to emit function if it is marked as declare target as it was already
10152   // emitted.
10153   if (OMPDeclareTargetDeclAttr::isDeclareTargetDeclaration(D)) {
10154     if (D->hasBody() && AlreadyEmittedTargetDecls.count(D) == 0) {
10155       if (auto *F = dyn_cast_or_null<llvm::Function>(
10156               CGM.GetGlobalValue(CGM.getMangledName(GD))))
10157         return !F->isDeclaration();
10158       return false;
10159     }
10160     return true;
10161   }
10162 
10163   return !AlreadyEmittedTargetDecls.insert(D).second;
10164 }
10165 
emitTeamsCall(CodeGenFunction & CGF,const OMPExecutableDirective & D,SourceLocation Loc,llvm::Function * OutlinedFn,ArrayRef<llvm::Value * > CapturedVars)10166 void CGOpenMPRuntime::emitTeamsCall(CodeGenFunction &CGF,
10167                                     const OMPExecutableDirective &D,
10168                                     SourceLocation Loc,
10169                                     llvm::Function *OutlinedFn,
10170                                     ArrayRef<llvm::Value *> CapturedVars) {
10171   if (!CGF.HaveInsertPoint())
10172     return;
10173 
10174   llvm::Value *RTLoc = emitUpdateLocation(CGF, Loc);
10175   CodeGenFunction::RunCleanupsScope Scope(CGF);
10176 
10177   // Build call __kmpc_fork_teams(loc, n, microtask, var1, .., varn);
10178   llvm::Value *Args[] = {
10179       RTLoc,
10180       CGF.Builder.getInt32(CapturedVars.size()), // Number of captured vars
10181       CGF.Builder.CreateBitCast(OutlinedFn, getKmpc_MicroPointerTy())};
10182   llvm::SmallVector<llvm::Value *, 16> RealArgs;
10183   RealArgs.append(std::begin(Args), std::end(Args));
10184   RealArgs.append(CapturedVars.begin(), CapturedVars.end());
10185 
10186   llvm::FunctionCallee RTLFn = OMPBuilder.getOrCreateRuntimeFunction(
10187       CGM.getModule(), OMPRTL___kmpc_fork_teams);
10188   CGF.EmitRuntimeCall(RTLFn, RealArgs);
10189 }
10190 
emitNumTeamsClause(CodeGenFunction & CGF,const Expr * NumTeams,const Expr * ThreadLimit,SourceLocation Loc)10191 void CGOpenMPRuntime::emitNumTeamsClause(CodeGenFunction &CGF,
10192                                          const Expr *NumTeams,
10193                                          const Expr *ThreadLimit,
10194                                          SourceLocation Loc) {
10195   if (!CGF.HaveInsertPoint())
10196     return;
10197 
10198   llvm::Value *RTLoc = emitUpdateLocation(CGF, Loc);
10199 
10200   llvm::Value *NumTeamsVal =
10201       NumTeams
10202           ? CGF.Builder.CreateIntCast(CGF.EmitScalarExpr(NumTeams),
10203                                       CGF.CGM.Int32Ty, /* isSigned = */ true)
10204           : CGF.Builder.getInt32(0);
10205 
10206   llvm::Value *ThreadLimitVal =
10207       ThreadLimit
10208           ? CGF.Builder.CreateIntCast(CGF.EmitScalarExpr(ThreadLimit),
10209                                       CGF.CGM.Int32Ty, /* isSigned = */ true)
10210           : CGF.Builder.getInt32(0);
10211 
10212   // Build call __kmpc_push_num_teamss(&loc, global_tid, num_teams, thread_limit)
10213   llvm::Value *PushNumTeamsArgs[] = {RTLoc, getThreadID(CGF, Loc), NumTeamsVal,
10214                                      ThreadLimitVal};
10215   CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
10216                           CGM.getModule(), OMPRTL___kmpc_push_num_teams),
10217                       PushNumTeamsArgs);
10218 }
10219 
emitThreadLimitClause(CodeGenFunction & CGF,const Expr * ThreadLimit,SourceLocation Loc)10220 void CGOpenMPRuntime::emitThreadLimitClause(CodeGenFunction &CGF,
10221                                             const Expr *ThreadLimit,
10222                                             SourceLocation Loc) {
10223   llvm::Value *RTLoc = emitUpdateLocation(CGF, Loc);
10224   llvm::Value *ThreadLimitVal =
10225       ThreadLimit
10226           ? CGF.Builder.CreateIntCast(CGF.EmitScalarExpr(ThreadLimit),
10227                                       CGF.CGM.Int32Ty, /* isSigned = */ true)
10228           : CGF.Builder.getInt32(0);
10229 
10230   // Build call __kmpc_set_thread_limit(&loc, global_tid, thread_limit)
10231   llvm::Value *ThreadLimitArgs[] = {RTLoc, getThreadID(CGF, Loc),
10232                                     ThreadLimitVal};
10233   CGF.EmitRuntimeCall(OMPBuilder.getOrCreateRuntimeFunction(
10234                           CGM.getModule(), OMPRTL___kmpc_set_thread_limit),
10235                       ThreadLimitArgs);
10236 }
10237 
emitTargetDataCalls(CodeGenFunction & CGF,const OMPExecutableDirective & D,const Expr * IfCond,const Expr * Device,const RegionCodeGenTy & CodeGen,CGOpenMPRuntime::TargetDataInfo & Info)10238 void CGOpenMPRuntime::emitTargetDataCalls(
10239     CodeGenFunction &CGF, const OMPExecutableDirective &D, const Expr *IfCond,
10240     const Expr *Device, const RegionCodeGenTy &CodeGen,
10241     CGOpenMPRuntime::TargetDataInfo &Info) {
10242   if (!CGF.HaveInsertPoint())
10243     return;
10244 
10245   // Action used to replace the default codegen action and turn privatization
10246   // off.
10247   PrePostActionTy NoPrivAction;
10248 
10249   using InsertPointTy = llvm::OpenMPIRBuilder::InsertPointTy;
10250 
10251   llvm::Value *IfCondVal = nullptr;
10252   if (IfCond)
10253     IfCondVal = CGF.EvaluateExprAsBool(IfCond);
10254 
10255   // Emit device ID if any.
10256   llvm::Value *DeviceID = nullptr;
10257   if (Device) {
10258     DeviceID = CGF.Builder.CreateIntCast(CGF.EmitScalarExpr(Device),
10259                                          CGF.Int64Ty, /*isSigned=*/true);
10260   } else {
10261     DeviceID = CGF.Builder.getInt64(OMP_DEVICEID_UNDEF);
10262   }
10263 
10264   // Fill up the arrays with all the mapped variables.
10265   MappableExprsHandler::MapCombinedInfoTy CombinedInfo;
10266   auto GenMapInfoCB =
10267       [&](InsertPointTy CodeGenIP) -> llvm::OpenMPIRBuilder::MapInfosTy & {
10268     CGF.Builder.restoreIP(CodeGenIP);
10269     // Get map clause information.
10270     MappableExprsHandler MEHandler(D, CGF);
10271     MEHandler.generateAllInfo(CombinedInfo, OMPBuilder);
10272 
10273     auto FillInfoMap = [&](MappableExprsHandler::MappingExprInfo &MapExpr) {
10274       return emitMappingInformation(CGF, OMPBuilder, MapExpr);
10275     };
10276     if (CGM.getCodeGenOpts().getDebugInfo() !=
10277         llvm::codegenoptions::NoDebugInfo) {
10278       CombinedInfo.Names.resize(CombinedInfo.Exprs.size());
10279       llvm::transform(CombinedInfo.Exprs, CombinedInfo.Names.begin(),
10280                       FillInfoMap);
10281     }
10282 
10283     return CombinedInfo;
10284   };
10285   using BodyGenTy = llvm::OpenMPIRBuilder::BodyGenTy;
10286   auto BodyCB = [&](InsertPointTy CodeGenIP, BodyGenTy BodyGenType) {
10287     CGF.Builder.restoreIP(CodeGenIP);
10288     switch (BodyGenType) {
10289     case BodyGenTy::Priv:
10290       if (!Info.CaptureDeviceAddrMap.empty())
10291         CodeGen(CGF);
10292       break;
10293     case BodyGenTy::DupNoPriv:
10294       if (!Info.CaptureDeviceAddrMap.empty()) {
10295         CodeGen.setAction(NoPrivAction);
10296         CodeGen(CGF);
10297       }
10298       break;
10299     case BodyGenTy::NoPriv:
10300       if (Info.CaptureDeviceAddrMap.empty()) {
10301         CodeGen.setAction(NoPrivAction);
10302         CodeGen(CGF);
10303       }
10304       break;
10305     }
10306     return InsertPointTy(CGF.Builder.GetInsertBlock(),
10307                          CGF.Builder.GetInsertPoint());
10308   };
10309 
10310   auto DeviceAddrCB = [&](unsigned int I, llvm::Value *NewDecl) {
10311     if (const ValueDecl *DevVD = CombinedInfo.DevicePtrDecls[I]) {
10312       Info.CaptureDeviceAddrMap.try_emplace(DevVD, NewDecl);
10313     }
10314   };
10315 
10316   auto CustomMapperCB = [&](unsigned int I) {
10317     llvm::Value *MFunc = nullptr;
10318     if (CombinedInfo.Mappers[I]) {
10319       Info.HasMapper = true;
10320       MFunc = CGF.CGM.getOpenMPRuntime().getOrCreateUserDefinedMapperFunc(
10321           cast<OMPDeclareMapperDecl>(CombinedInfo.Mappers[I]));
10322     }
10323     return MFunc;
10324   };
10325 
10326   // Source location for the ident struct
10327   llvm::Value *RTLoc = emitUpdateLocation(CGF, D.getBeginLoc());
10328 
10329   InsertPointTy AllocaIP(CGF.AllocaInsertPt->getParent(),
10330                          CGF.AllocaInsertPt->getIterator());
10331   InsertPointTy CodeGenIP(CGF.Builder.GetInsertBlock(),
10332                           CGF.Builder.GetInsertPoint());
10333   llvm::OpenMPIRBuilder::LocationDescription OmpLoc(CodeGenIP);
10334   CGF.Builder.restoreIP(OMPBuilder.createTargetData(
10335       OmpLoc, AllocaIP, CodeGenIP, DeviceID, IfCondVal, Info, GenMapInfoCB,
10336       /*MapperFunc=*/nullptr, BodyCB, DeviceAddrCB, CustomMapperCB, RTLoc));
10337 }
10338 
emitTargetDataStandAloneCall(CodeGenFunction & CGF,const OMPExecutableDirective & D,const Expr * IfCond,const Expr * Device)10339 void CGOpenMPRuntime::emitTargetDataStandAloneCall(
10340     CodeGenFunction &CGF, const OMPExecutableDirective &D, const Expr *IfCond,
10341     const Expr *Device) {
10342   if (!CGF.HaveInsertPoint())
10343     return;
10344 
10345   assert((isa<OMPTargetEnterDataDirective>(D) ||
10346           isa<OMPTargetExitDataDirective>(D) ||
10347           isa<OMPTargetUpdateDirective>(D)) &&
10348          "Expecting either target enter, exit data, or update directives.");
10349 
10350   CodeGenFunction::OMPTargetDataInfo InputInfo;
10351   llvm::Value *MapTypesArray = nullptr;
10352   llvm::Value *MapNamesArray = nullptr;
10353   // Generate the code for the opening of the data environment.
10354   auto &&ThenGen = [this, &D, Device, &InputInfo, &MapTypesArray,
10355                     &MapNamesArray](CodeGenFunction &CGF, PrePostActionTy &) {
10356     // Emit device ID if any.
10357     llvm::Value *DeviceID = nullptr;
10358     if (Device) {
10359       DeviceID = CGF.Builder.CreateIntCast(CGF.EmitScalarExpr(Device),
10360                                            CGF.Int64Ty, /*isSigned=*/true);
10361     } else {
10362       DeviceID = CGF.Builder.getInt64(OMP_DEVICEID_UNDEF);
10363     }
10364 
10365     // Emit the number of elements in the offloading arrays.
10366     llvm::Constant *PointerNum =
10367         CGF.Builder.getInt32(InputInfo.NumberOfTargetItems);
10368 
10369     // Source location for the ident struct
10370     llvm::Value *RTLoc = emitUpdateLocation(CGF, D.getBeginLoc());
10371 
10372     SmallVector<llvm::Value *, 13> OffloadingArgs(
10373         {RTLoc, DeviceID, PointerNum,
10374          InputInfo.BasePointersArray.emitRawPointer(CGF),
10375          InputInfo.PointersArray.emitRawPointer(CGF),
10376          InputInfo.SizesArray.emitRawPointer(CGF), MapTypesArray, MapNamesArray,
10377          InputInfo.MappersArray.emitRawPointer(CGF)});
10378 
10379     // Select the right runtime function call for each standalone
10380     // directive.
10381     const bool HasNowait = D.hasClausesOfKind<OMPNowaitClause>();
10382     RuntimeFunction RTLFn;
10383     switch (D.getDirectiveKind()) {
10384     case OMPD_target_enter_data:
10385       RTLFn = HasNowait ? OMPRTL___tgt_target_data_begin_nowait_mapper
10386                         : OMPRTL___tgt_target_data_begin_mapper;
10387       break;
10388     case OMPD_target_exit_data:
10389       RTLFn = HasNowait ? OMPRTL___tgt_target_data_end_nowait_mapper
10390                         : OMPRTL___tgt_target_data_end_mapper;
10391       break;
10392     case OMPD_target_update:
10393       RTLFn = HasNowait ? OMPRTL___tgt_target_data_update_nowait_mapper
10394                         : OMPRTL___tgt_target_data_update_mapper;
10395       break;
10396     case OMPD_parallel:
10397     case OMPD_for:
10398     case OMPD_parallel_for:
10399     case OMPD_parallel_master:
10400     case OMPD_parallel_sections:
10401     case OMPD_for_simd:
10402     case OMPD_parallel_for_simd:
10403     case OMPD_cancel:
10404     case OMPD_cancellation_point:
10405     case OMPD_ordered:
10406     case OMPD_threadprivate:
10407     case OMPD_allocate:
10408     case OMPD_task:
10409     case OMPD_simd:
10410     case OMPD_tile:
10411     case OMPD_unroll:
10412     case OMPD_sections:
10413     case OMPD_section:
10414     case OMPD_single:
10415     case OMPD_master:
10416     case OMPD_critical:
10417     case OMPD_taskyield:
10418     case OMPD_barrier:
10419     case OMPD_taskwait:
10420     case OMPD_taskgroup:
10421     case OMPD_atomic:
10422     case OMPD_flush:
10423     case OMPD_depobj:
10424     case OMPD_scan:
10425     case OMPD_teams:
10426     case OMPD_target_data:
10427     case OMPD_distribute:
10428     case OMPD_distribute_simd:
10429     case OMPD_distribute_parallel_for:
10430     case OMPD_distribute_parallel_for_simd:
10431     case OMPD_teams_distribute:
10432     case OMPD_teams_distribute_simd:
10433     case OMPD_teams_distribute_parallel_for:
10434     case OMPD_teams_distribute_parallel_for_simd:
10435     case OMPD_declare_simd:
10436     case OMPD_declare_variant:
10437     case OMPD_begin_declare_variant:
10438     case OMPD_end_declare_variant:
10439     case OMPD_declare_target:
10440     case OMPD_end_declare_target:
10441     case OMPD_declare_reduction:
10442     case OMPD_declare_mapper:
10443     case OMPD_taskloop:
10444     case OMPD_taskloop_simd:
10445     case OMPD_master_taskloop:
10446     case OMPD_master_taskloop_simd:
10447     case OMPD_parallel_master_taskloop:
10448     case OMPD_parallel_master_taskloop_simd:
10449     case OMPD_target:
10450     case OMPD_target_simd:
10451     case OMPD_target_teams_distribute:
10452     case OMPD_target_teams_distribute_simd:
10453     case OMPD_target_teams_distribute_parallel_for:
10454     case OMPD_target_teams_distribute_parallel_for_simd:
10455     case OMPD_target_teams:
10456     case OMPD_target_parallel:
10457     case OMPD_target_parallel_for:
10458     case OMPD_target_parallel_for_simd:
10459     case OMPD_requires:
10460     case OMPD_metadirective:
10461     case OMPD_unknown:
10462     default:
10463       llvm_unreachable("Unexpected standalone target data directive.");
10464       break;
10465     }
10466     if (HasNowait) {
10467       OffloadingArgs.push_back(llvm::Constant::getNullValue(CGF.Int32Ty));
10468       OffloadingArgs.push_back(llvm::Constant::getNullValue(CGF.VoidPtrTy));
10469       OffloadingArgs.push_back(llvm::Constant::getNullValue(CGF.Int32Ty));
10470       OffloadingArgs.push_back(llvm::Constant::getNullValue(CGF.VoidPtrTy));
10471     }
10472     CGF.EmitRuntimeCall(
10473         OMPBuilder.getOrCreateRuntimeFunction(CGM.getModule(), RTLFn),
10474         OffloadingArgs);
10475   };
10476 
10477   auto &&TargetThenGen = [this, &ThenGen, &D, &InputInfo, &MapTypesArray,
10478                           &MapNamesArray](CodeGenFunction &CGF,
10479                                           PrePostActionTy &) {
10480     // Fill up the arrays with all the mapped variables.
10481     MappableExprsHandler::MapCombinedInfoTy CombinedInfo;
10482 
10483     // Get map clause information.
10484     MappableExprsHandler MEHandler(D, CGF);
10485     MEHandler.generateAllInfo(CombinedInfo, OMPBuilder);
10486 
10487     CGOpenMPRuntime::TargetDataInfo Info;
10488     // Fill up the arrays and create the arguments.
10489     emitOffloadingArrays(CGF, CombinedInfo, Info, OMPBuilder,
10490                          /*IsNonContiguous=*/true);
10491     bool RequiresOuterTask = D.hasClausesOfKind<OMPDependClause>() ||
10492                              D.hasClausesOfKind<OMPNowaitClause>();
10493     bool EmitDebug = CGF.CGM.getCodeGenOpts().getDebugInfo() !=
10494                      llvm::codegenoptions::NoDebugInfo;
10495     OMPBuilder.emitOffloadingArraysArgument(CGF.Builder, Info.RTArgs, Info,
10496                                             EmitDebug,
10497                                             /*ForEndCall=*/false);
10498     InputInfo.NumberOfTargetItems = Info.NumberOfPtrs;
10499     InputInfo.BasePointersArray = Address(Info.RTArgs.BasePointersArray,
10500                                           CGF.VoidPtrTy, CGM.getPointerAlign());
10501     InputInfo.PointersArray = Address(Info.RTArgs.PointersArray, CGF.VoidPtrTy,
10502                                       CGM.getPointerAlign());
10503     InputInfo.SizesArray =
10504         Address(Info.RTArgs.SizesArray, CGF.Int64Ty, CGM.getPointerAlign());
10505     InputInfo.MappersArray =
10506         Address(Info.RTArgs.MappersArray, CGF.VoidPtrTy, CGM.getPointerAlign());
10507     MapTypesArray = Info.RTArgs.MapTypesArray;
10508     MapNamesArray = Info.RTArgs.MapNamesArray;
10509     if (RequiresOuterTask)
10510       CGF.EmitOMPTargetTaskBasedDirective(D, ThenGen, InputInfo);
10511     else
10512       emitInlinedDirective(CGF, D.getDirectiveKind(), ThenGen);
10513   };
10514 
10515   if (IfCond) {
10516     emitIfClause(CGF, IfCond, TargetThenGen,
10517                  [](CodeGenFunction &CGF, PrePostActionTy &) {});
10518   } else {
10519     RegionCodeGenTy ThenRCG(TargetThenGen);
10520     ThenRCG(CGF);
10521   }
10522 }
10523 
10524 namespace {
10525   /// Kind of parameter in a function with 'declare simd' directive.
10526 enum ParamKindTy {
10527   Linear,
10528   LinearRef,
10529   LinearUVal,
10530   LinearVal,
10531   Uniform,
10532   Vector,
10533 };
10534 /// Attribute set of the parameter.
10535 struct ParamAttrTy {
10536   ParamKindTy Kind = Vector;
10537   llvm::APSInt StrideOrArg;
10538   llvm::APSInt Alignment;
10539   bool HasVarStride = false;
10540 };
10541 } // namespace
10542 
evaluateCDTSize(const FunctionDecl * FD,ArrayRef<ParamAttrTy> ParamAttrs)10543 static unsigned evaluateCDTSize(const FunctionDecl *FD,
10544                                 ArrayRef<ParamAttrTy> ParamAttrs) {
10545   // Every vector variant of a SIMD-enabled function has a vector length (VLEN).
10546   // If OpenMP clause "simdlen" is used, the VLEN is the value of the argument
10547   // of that clause. The VLEN value must be power of 2.
10548   // In other case the notion of the function`s "characteristic data type" (CDT)
10549   // is used to compute the vector length.
10550   // CDT is defined in the following order:
10551   //   a) For non-void function, the CDT is the return type.
10552   //   b) If the function has any non-uniform, non-linear parameters, then the
10553   //   CDT is the type of the first such parameter.
10554   //   c) If the CDT determined by a) or b) above is struct, union, or class
10555   //   type which is pass-by-value (except for the type that maps to the
10556   //   built-in complex data type), the characteristic data type is int.
10557   //   d) If none of the above three cases is applicable, the CDT is int.
10558   // The VLEN is then determined based on the CDT and the size of vector
10559   // register of that ISA for which current vector version is generated. The
10560   // VLEN is computed using the formula below:
10561   //   VLEN  = sizeof(vector_register) / sizeof(CDT),
10562   // where vector register size specified in section 3.2.1 Registers and the
10563   // Stack Frame of original AMD64 ABI document.
10564   QualType RetType = FD->getReturnType();
10565   if (RetType.isNull())
10566     return 0;
10567   ASTContext &C = FD->getASTContext();
10568   QualType CDT;
10569   if (!RetType.isNull() && !RetType->isVoidType()) {
10570     CDT = RetType;
10571   } else {
10572     unsigned Offset = 0;
10573     if (const auto *MD = dyn_cast<CXXMethodDecl>(FD)) {
10574       if (ParamAttrs[Offset].Kind == Vector)
10575         CDT = C.getPointerType(C.getRecordType(MD->getParent()));
10576       ++Offset;
10577     }
10578     if (CDT.isNull()) {
10579       for (unsigned I = 0, E = FD->getNumParams(); I < E; ++I) {
10580         if (ParamAttrs[I + Offset].Kind == Vector) {
10581           CDT = FD->getParamDecl(I)->getType();
10582           break;
10583         }
10584       }
10585     }
10586   }
10587   if (CDT.isNull())
10588     CDT = C.IntTy;
10589   CDT = CDT->getCanonicalTypeUnqualified();
10590   if (CDT->isRecordType() || CDT->isUnionType())
10591     CDT = C.IntTy;
10592   return C.getTypeSize(CDT);
10593 }
10594 
10595 /// Mangle the parameter part of the vector function name according to
10596 /// their OpenMP classification. The mangling function is defined in
10597 /// section 4.5 of the AAVFABI(2021Q1).
mangleVectorParameters(ArrayRef<ParamAttrTy> ParamAttrs)10598 static std::string mangleVectorParameters(ArrayRef<ParamAttrTy> ParamAttrs) {
10599   SmallString<256> Buffer;
10600   llvm::raw_svector_ostream Out(Buffer);
10601   for (const auto &ParamAttr : ParamAttrs) {
10602     switch (ParamAttr.Kind) {
10603     case Linear:
10604       Out << 'l';
10605       break;
10606     case LinearRef:
10607       Out << 'R';
10608       break;
10609     case LinearUVal:
10610       Out << 'U';
10611       break;
10612     case LinearVal:
10613       Out << 'L';
10614       break;
10615     case Uniform:
10616       Out << 'u';
10617       break;
10618     case Vector:
10619       Out << 'v';
10620       break;
10621     }
10622     if (ParamAttr.HasVarStride)
10623       Out << "s" << ParamAttr.StrideOrArg;
10624     else if (ParamAttr.Kind == Linear || ParamAttr.Kind == LinearRef ||
10625              ParamAttr.Kind == LinearUVal || ParamAttr.Kind == LinearVal) {
10626       // Don't print the step value if it is not present or if it is
10627       // equal to 1.
10628       if (ParamAttr.StrideOrArg < 0)
10629         Out << 'n' << -ParamAttr.StrideOrArg;
10630       else if (ParamAttr.StrideOrArg != 1)
10631         Out << ParamAttr.StrideOrArg;
10632     }
10633 
10634     if (!!ParamAttr.Alignment)
10635       Out << 'a' << ParamAttr.Alignment;
10636   }
10637 
10638   return std::string(Out.str());
10639 }
10640 
10641 static void
emitX86DeclareSimdFunction(const FunctionDecl * FD,llvm::Function * Fn,const llvm::APSInt & VLENVal,ArrayRef<ParamAttrTy> ParamAttrs,OMPDeclareSimdDeclAttr::BranchStateTy State)10642 emitX86DeclareSimdFunction(const FunctionDecl *FD, llvm::Function *Fn,
10643                            const llvm::APSInt &VLENVal,
10644                            ArrayRef<ParamAttrTy> ParamAttrs,
10645                            OMPDeclareSimdDeclAttr::BranchStateTy State) {
10646   struct ISADataTy {
10647     char ISA;
10648     unsigned VecRegSize;
10649   };
10650   ISADataTy ISAData[] = {
10651       {
10652           'b', 128
10653       }, // SSE
10654       {
10655           'c', 256
10656       }, // AVX
10657       {
10658           'd', 256
10659       }, // AVX2
10660       {
10661           'e', 512
10662       }, // AVX512
10663   };
10664   llvm::SmallVector<char, 2> Masked;
10665   switch (State) {
10666   case OMPDeclareSimdDeclAttr::BS_Undefined:
10667     Masked.push_back('N');
10668     Masked.push_back('M');
10669     break;
10670   case OMPDeclareSimdDeclAttr::BS_Notinbranch:
10671     Masked.push_back('N');
10672     break;
10673   case OMPDeclareSimdDeclAttr::BS_Inbranch:
10674     Masked.push_back('M');
10675     break;
10676   }
10677   for (char Mask : Masked) {
10678     for (const ISADataTy &Data : ISAData) {
10679       SmallString<256> Buffer;
10680       llvm::raw_svector_ostream Out(Buffer);
10681       Out << "_ZGV" << Data.ISA << Mask;
10682       if (!VLENVal) {
10683         unsigned NumElts = evaluateCDTSize(FD, ParamAttrs);
10684         assert(NumElts && "Non-zero simdlen/cdtsize expected");
10685         Out << llvm::APSInt::getUnsigned(Data.VecRegSize / NumElts);
10686       } else {
10687         Out << VLENVal;
10688       }
10689       Out << mangleVectorParameters(ParamAttrs);
10690       Out << '_' << Fn->getName();
10691       Fn->addFnAttr(Out.str());
10692     }
10693   }
10694 }
10695 
10696 // This are the Functions that are needed to mangle the name of the
10697 // vector functions generated by the compiler, according to the rules
10698 // defined in the "Vector Function ABI specifications for AArch64",
10699 // available at
10700 // https://developer.arm.com/products/software-development-tools/hpc/arm-compiler-for-hpc/vector-function-abi.
10701 
10702 /// Maps To Vector (MTV), as defined in 4.1.1 of the AAVFABI (2021Q1).
getAArch64MTV(QualType QT,ParamKindTy Kind)10703 static bool getAArch64MTV(QualType QT, ParamKindTy Kind) {
10704   QT = QT.getCanonicalType();
10705 
10706   if (QT->isVoidType())
10707     return false;
10708 
10709   if (Kind == ParamKindTy::Uniform)
10710     return false;
10711 
10712   if (Kind == ParamKindTy::LinearUVal || Kind == ParamKindTy::LinearRef)
10713     return false;
10714 
10715   if ((Kind == ParamKindTy::Linear || Kind == ParamKindTy::LinearVal) &&
10716       !QT->isReferenceType())
10717     return false;
10718 
10719   return true;
10720 }
10721 
10722 /// Pass By Value (PBV), as defined in 3.1.2 of the AAVFABI.
getAArch64PBV(QualType QT,ASTContext & C)10723 static bool getAArch64PBV(QualType QT, ASTContext &C) {
10724   QT = QT.getCanonicalType();
10725   unsigned Size = C.getTypeSize(QT);
10726 
10727   // Only scalars and complex within 16 bytes wide set PVB to true.
10728   if (Size != 8 && Size != 16 && Size != 32 && Size != 64 && Size != 128)
10729     return false;
10730 
10731   if (QT->isFloatingType())
10732     return true;
10733 
10734   if (QT->isIntegerType())
10735     return true;
10736 
10737   if (QT->isPointerType())
10738     return true;
10739 
10740   // TODO: Add support for complex types (section 3.1.2, item 2).
10741 
10742   return false;
10743 }
10744 
10745 /// Computes the lane size (LS) of a return type or of an input parameter,
10746 /// as defined by `LS(P)` in 3.2.1 of the AAVFABI.
10747 /// TODO: Add support for references, section 3.2.1, item 1.
getAArch64LS(QualType QT,ParamKindTy Kind,ASTContext & C)10748 static unsigned getAArch64LS(QualType QT, ParamKindTy Kind, ASTContext &C) {
10749   if (!getAArch64MTV(QT, Kind) && QT.getCanonicalType()->isPointerType()) {
10750     QualType PTy = QT.getCanonicalType()->getPointeeType();
10751     if (getAArch64PBV(PTy, C))
10752       return C.getTypeSize(PTy);
10753   }
10754   if (getAArch64PBV(QT, C))
10755     return C.getTypeSize(QT);
10756 
10757   return C.getTypeSize(C.getUIntPtrType());
10758 }
10759 
10760 // Get Narrowest Data Size (NDS) and Widest Data Size (WDS) from the
10761 // signature of the scalar function, as defined in 3.2.2 of the
10762 // AAVFABI.
10763 static std::tuple<unsigned, unsigned, bool>
getNDSWDS(const FunctionDecl * FD,ArrayRef<ParamAttrTy> ParamAttrs)10764 getNDSWDS(const FunctionDecl *FD, ArrayRef<ParamAttrTy> ParamAttrs) {
10765   QualType RetType = FD->getReturnType().getCanonicalType();
10766 
10767   ASTContext &C = FD->getASTContext();
10768 
10769   bool OutputBecomesInput = false;
10770 
10771   llvm::SmallVector<unsigned, 8> Sizes;
10772   if (!RetType->isVoidType()) {
10773     Sizes.push_back(getAArch64LS(RetType, ParamKindTy::Vector, C));
10774     if (!getAArch64PBV(RetType, C) && getAArch64MTV(RetType, {}))
10775       OutputBecomesInput = true;
10776   }
10777   for (unsigned I = 0, E = FD->getNumParams(); I < E; ++I) {
10778     QualType QT = FD->getParamDecl(I)->getType().getCanonicalType();
10779     Sizes.push_back(getAArch64LS(QT, ParamAttrs[I].Kind, C));
10780   }
10781 
10782   assert(!Sizes.empty() && "Unable to determine NDS and WDS.");
10783   // The LS of a function parameter / return value can only be a power
10784   // of 2, starting from 8 bits, up to 128.
10785   assert(llvm::all_of(Sizes,
10786                       [](unsigned Size) {
10787                         return Size == 8 || Size == 16 || Size == 32 ||
10788                                Size == 64 || Size == 128;
10789                       }) &&
10790          "Invalid size");
10791 
10792   return std::make_tuple(*std::min_element(std::begin(Sizes), std::end(Sizes)),
10793                          *std::max_element(std::begin(Sizes), std::end(Sizes)),
10794                          OutputBecomesInput);
10795 }
10796 
10797 // Function used to add the attribute. The parameter `VLEN` is
10798 // templated to allow the use of "x" when targeting scalable functions
10799 // for SVE.
10800 template <typename T>
addAArch64VectorName(T VLEN,StringRef LMask,StringRef Prefix,char ISA,StringRef ParSeq,StringRef MangledName,bool OutputBecomesInput,llvm::Function * Fn)10801 static void addAArch64VectorName(T VLEN, StringRef LMask, StringRef Prefix,
10802                                  char ISA, StringRef ParSeq,
10803                                  StringRef MangledName, bool OutputBecomesInput,
10804                                  llvm::Function *Fn) {
10805   SmallString<256> Buffer;
10806   llvm::raw_svector_ostream Out(Buffer);
10807   Out << Prefix << ISA << LMask << VLEN;
10808   if (OutputBecomesInput)
10809     Out << "v";
10810   Out << ParSeq << "_" << MangledName;
10811   Fn->addFnAttr(Out.str());
10812 }
10813 
10814 // Helper function to generate the Advanced SIMD names depending on
10815 // the value of the NDS when simdlen is not present.
addAArch64AdvSIMDNDSNames(unsigned NDS,StringRef Mask,StringRef Prefix,char ISA,StringRef ParSeq,StringRef MangledName,bool OutputBecomesInput,llvm::Function * Fn)10816 static void addAArch64AdvSIMDNDSNames(unsigned NDS, StringRef Mask,
10817                                       StringRef Prefix, char ISA,
10818                                       StringRef ParSeq, StringRef MangledName,
10819                                       bool OutputBecomesInput,
10820                                       llvm::Function *Fn) {
10821   switch (NDS) {
10822   case 8:
10823     addAArch64VectorName(8, Mask, Prefix, ISA, ParSeq, MangledName,
10824                          OutputBecomesInput, Fn);
10825     addAArch64VectorName(16, Mask, Prefix, ISA, ParSeq, MangledName,
10826                          OutputBecomesInput, Fn);
10827     break;
10828   case 16:
10829     addAArch64VectorName(4, Mask, Prefix, ISA, ParSeq, MangledName,
10830                          OutputBecomesInput, Fn);
10831     addAArch64VectorName(8, Mask, Prefix, ISA, ParSeq, MangledName,
10832                          OutputBecomesInput, Fn);
10833     break;
10834   case 32:
10835     addAArch64VectorName(2, Mask, Prefix, ISA, ParSeq, MangledName,
10836                          OutputBecomesInput, Fn);
10837     addAArch64VectorName(4, Mask, Prefix, ISA, ParSeq, MangledName,
10838                          OutputBecomesInput, Fn);
10839     break;
10840   case 64:
10841   case 128:
10842     addAArch64VectorName(2, Mask, Prefix, ISA, ParSeq, MangledName,
10843                          OutputBecomesInput, Fn);
10844     break;
10845   default:
10846     llvm_unreachable("Scalar type is too wide.");
10847   }
10848 }
10849 
10850 /// Emit vector function attributes for AArch64, as defined in the AAVFABI.
emitAArch64DeclareSimdFunction(CodeGenModule & CGM,const FunctionDecl * FD,unsigned UserVLEN,ArrayRef<ParamAttrTy> ParamAttrs,OMPDeclareSimdDeclAttr::BranchStateTy State,StringRef MangledName,char ISA,unsigned VecRegSize,llvm::Function * Fn,SourceLocation SLoc)10851 static void emitAArch64DeclareSimdFunction(
10852     CodeGenModule &CGM, const FunctionDecl *FD, unsigned UserVLEN,
10853     ArrayRef<ParamAttrTy> ParamAttrs,
10854     OMPDeclareSimdDeclAttr::BranchStateTy State, StringRef MangledName,
10855     char ISA, unsigned VecRegSize, llvm::Function *Fn, SourceLocation SLoc) {
10856 
10857   // Get basic data for building the vector signature.
10858   const auto Data = getNDSWDS(FD, ParamAttrs);
10859   const unsigned NDS = std::get<0>(Data);
10860   const unsigned WDS = std::get<1>(Data);
10861   const bool OutputBecomesInput = std::get<2>(Data);
10862 
10863   // Check the values provided via `simdlen` by the user.
10864   // 1. A `simdlen(1)` doesn't produce vector signatures,
10865   if (UserVLEN == 1) {
10866     unsigned DiagID = CGM.getDiags().getCustomDiagID(
10867         DiagnosticsEngine::Warning,
10868         "The clause simdlen(1) has no effect when targeting aarch64.");
10869     CGM.getDiags().Report(SLoc, DiagID);
10870     return;
10871   }
10872 
10873   // 2. Section 3.3.1, item 1: user input must be a power of 2 for
10874   // Advanced SIMD output.
10875   if (ISA == 'n' && UserVLEN && !llvm::isPowerOf2_32(UserVLEN)) {
10876     unsigned DiagID = CGM.getDiags().getCustomDiagID(
10877         DiagnosticsEngine::Warning, "The value specified in simdlen must be a "
10878                                     "power of 2 when targeting Advanced SIMD.");
10879     CGM.getDiags().Report(SLoc, DiagID);
10880     return;
10881   }
10882 
10883   // 3. Section 3.4.1. SVE fixed lengh must obey the architectural
10884   // limits.
10885   if (ISA == 's' && UserVLEN != 0) {
10886     if ((UserVLEN * WDS > 2048) || (UserVLEN * WDS % 128 != 0)) {
10887       unsigned DiagID = CGM.getDiags().getCustomDiagID(
10888           DiagnosticsEngine::Warning, "The clause simdlen must fit the %0-bit "
10889                                       "lanes in the architectural constraints "
10890                                       "for SVE (min is 128-bit, max is "
10891                                       "2048-bit, by steps of 128-bit)");
10892       CGM.getDiags().Report(SLoc, DiagID) << WDS;
10893       return;
10894     }
10895   }
10896 
10897   // Sort out parameter sequence.
10898   const std::string ParSeq = mangleVectorParameters(ParamAttrs);
10899   StringRef Prefix = "_ZGV";
10900   // Generate simdlen from user input (if any).
10901   if (UserVLEN) {
10902     if (ISA == 's') {
10903       // SVE generates only a masked function.
10904       addAArch64VectorName(UserVLEN, "M", Prefix, ISA, ParSeq, MangledName,
10905                            OutputBecomesInput, Fn);
10906     } else {
10907       assert(ISA == 'n' && "Expected ISA either 's' or 'n'.");
10908       // Advanced SIMD generates one or two functions, depending on
10909       // the `[not]inbranch` clause.
10910       switch (State) {
10911       case OMPDeclareSimdDeclAttr::BS_Undefined:
10912         addAArch64VectorName(UserVLEN, "N", Prefix, ISA, ParSeq, MangledName,
10913                              OutputBecomesInput, Fn);
10914         addAArch64VectorName(UserVLEN, "M", Prefix, ISA, ParSeq, MangledName,
10915                              OutputBecomesInput, Fn);
10916         break;
10917       case OMPDeclareSimdDeclAttr::BS_Notinbranch:
10918         addAArch64VectorName(UserVLEN, "N", Prefix, ISA, ParSeq, MangledName,
10919                              OutputBecomesInput, Fn);
10920         break;
10921       case OMPDeclareSimdDeclAttr::BS_Inbranch:
10922         addAArch64VectorName(UserVLEN, "M", Prefix, ISA, ParSeq, MangledName,
10923                              OutputBecomesInput, Fn);
10924         break;
10925       }
10926     }
10927   } else {
10928     // If no user simdlen is provided, follow the AAVFABI rules for
10929     // generating the vector length.
10930     if (ISA == 's') {
10931       // SVE, section 3.4.1, item 1.
10932       addAArch64VectorName("x", "M", Prefix, ISA, ParSeq, MangledName,
10933                            OutputBecomesInput, Fn);
10934     } else {
10935       assert(ISA == 'n' && "Expected ISA either 's' or 'n'.");
10936       // Advanced SIMD, Section 3.3.1 of the AAVFABI, generates one or
10937       // two vector names depending on the use of the clause
10938       // `[not]inbranch`.
10939       switch (State) {
10940       case OMPDeclareSimdDeclAttr::BS_Undefined:
10941         addAArch64AdvSIMDNDSNames(NDS, "N", Prefix, ISA, ParSeq, MangledName,
10942                                   OutputBecomesInput, Fn);
10943         addAArch64AdvSIMDNDSNames(NDS, "M", Prefix, ISA, ParSeq, MangledName,
10944                                   OutputBecomesInput, Fn);
10945         break;
10946       case OMPDeclareSimdDeclAttr::BS_Notinbranch:
10947         addAArch64AdvSIMDNDSNames(NDS, "N", Prefix, ISA, ParSeq, MangledName,
10948                                   OutputBecomesInput, Fn);
10949         break;
10950       case OMPDeclareSimdDeclAttr::BS_Inbranch:
10951         addAArch64AdvSIMDNDSNames(NDS, "M", Prefix, ISA, ParSeq, MangledName,
10952                                   OutputBecomesInput, Fn);
10953         break;
10954       }
10955     }
10956   }
10957 }
10958 
emitDeclareSimdFunction(const FunctionDecl * FD,llvm::Function * Fn)10959 void CGOpenMPRuntime::emitDeclareSimdFunction(const FunctionDecl *FD,
10960                                               llvm::Function *Fn) {
10961   ASTContext &C = CGM.getContext();
10962   FD = FD->getMostRecentDecl();
10963   while (FD) {
10964     // Map params to their positions in function decl.
10965     llvm::DenseMap<const Decl *, unsigned> ParamPositions;
10966     if (isa<CXXMethodDecl>(FD))
10967       ParamPositions.try_emplace(FD, 0);
10968     unsigned ParamPos = ParamPositions.size();
10969     for (const ParmVarDecl *P : FD->parameters()) {
10970       ParamPositions.try_emplace(P->getCanonicalDecl(), ParamPos);
10971       ++ParamPos;
10972     }
10973     for (const auto *Attr : FD->specific_attrs<OMPDeclareSimdDeclAttr>()) {
10974       llvm::SmallVector<ParamAttrTy, 8> ParamAttrs(ParamPositions.size());
10975       // Mark uniform parameters.
10976       for (const Expr *E : Attr->uniforms()) {
10977         E = E->IgnoreParenImpCasts();
10978         unsigned Pos;
10979         if (isa<CXXThisExpr>(E)) {
10980           Pos = ParamPositions[FD];
10981         } else {
10982           const auto *PVD = cast<ParmVarDecl>(cast<DeclRefExpr>(E)->getDecl())
10983                                 ->getCanonicalDecl();
10984           auto It = ParamPositions.find(PVD);
10985           assert(It != ParamPositions.end() && "Function parameter not found");
10986           Pos = It->second;
10987         }
10988         ParamAttrs[Pos].Kind = Uniform;
10989       }
10990       // Get alignment info.
10991       auto *NI = Attr->alignments_begin();
10992       for (const Expr *E : Attr->aligneds()) {
10993         E = E->IgnoreParenImpCasts();
10994         unsigned Pos;
10995         QualType ParmTy;
10996         if (isa<CXXThisExpr>(E)) {
10997           Pos = ParamPositions[FD];
10998           ParmTy = E->getType();
10999         } else {
11000           const auto *PVD = cast<ParmVarDecl>(cast<DeclRefExpr>(E)->getDecl())
11001                                 ->getCanonicalDecl();
11002           auto It = ParamPositions.find(PVD);
11003           assert(It != ParamPositions.end() && "Function parameter not found");
11004           Pos = It->second;
11005           ParmTy = PVD->getType();
11006         }
11007         ParamAttrs[Pos].Alignment =
11008             (*NI)
11009                 ? (*NI)->EvaluateKnownConstInt(C)
11010                 : llvm::APSInt::getUnsigned(
11011                       C.toCharUnitsFromBits(C.getOpenMPDefaultSimdAlign(ParmTy))
11012                           .getQuantity());
11013         ++NI;
11014       }
11015       // Mark linear parameters.
11016       auto *SI = Attr->steps_begin();
11017       auto *MI = Attr->modifiers_begin();
11018       for (const Expr *E : Attr->linears()) {
11019         E = E->IgnoreParenImpCasts();
11020         unsigned Pos;
11021         bool IsReferenceType = false;
11022         // Rescaling factor needed to compute the linear parameter
11023         // value in the mangled name.
11024         unsigned PtrRescalingFactor = 1;
11025         if (isa<CXXThisExpr>(E)) {
11026           Pos = ParamPositions[FD];
11027           auto *P = cast<PointerType>(E->getType());
11028           PtrRescalingFactor = CGM.getContext()
11029                                    .getTypeSizeInChars(P->getPointeeType())
11030                                    .getQuantity();
11031         } else {
11032           const auto *PVD = cast<ParmVarDecl>(cast<DeclRefExpr>(E)->getDecl())
11033                                 ->getCanonicalDecl();
11034           auto It = ParamPositions.find(PVD);
11035           assert(It != ParamPositions.end() && "Function parameter not found");
11036           Pos = It->second;
11037           if (auto *P = dyn_cast<PointerType>(PVD->getType()))
11038             PtrRescalingFactor = CGM.getContext()
11039                                      .getTypeSizeInChars(P->getPointeeType())
11040                                      .getQuantity();
11041           else if (PVD->getType()->isReferenceType()) {
11042             IsReferenceType = true;
11043             PtrRescalingFactor =
11044                 CGM.getContext()
11045                     .getTypeSizeInChars(PVD->getType().getNonReferenceType())
11046                     .getQuantity();
11047           }
11048         }
11049         ParamAttrTy &ParamAttr = ParamAttrs[Pos];
11050         if (*MI == OMPC_LINEAR_ref)
11051           ParamAttr.Kind = LinearRef;
11052         else if (*MI == OMPC_LINEAR_uval)
11053           ParamAttr.Kind = LinearUVal;
11054         else if (IsReferenceType)
11055           ParamAttr.Kind = LinearVal;
11056         else
11057           ParamAttr.Kind = Linear;
11058         // Assuming a stride of 1, for `linear` without modifiers.
11059         ParamAttr.StrideOrArg = llvm::APSInt::getUnsigned(1);
11060         if (*SI) {
11061           Expr::EvalResult Result;
11062           if (!(*SI)->EvaluateAsInt(Result, C, Expr::SE_AllowSideEffects)) {
11063             if (const auto *DRE =
11064                     cast<DeclRefExpr>((*SI)->IgnoreParenImpCasts())) {
11065               if (const auto *StridePVD =
11066                       dyn_cast<ParmVarDecl>(DRE->getDecl())) {
11067                 ParamAttr.HasVarStride = true;
11068                 auto It = ParamPositions.find(StridePVD->getCanonicalDecl());
11069                 assert(It != ParamPositions.end() &&
11070                        "Function parameter not found");
11071                 ParamAttr.StrideOrArg = llvm::APSInt::getUnsigned(It->second);
11072               }
11073             }
11074           } else {
11075             ParamAttr.StrideOrArg = Result.Val.getInt();
11076           }
11077         }
11078         // If we are using a linear clause on a pointer, we need to
11079         // rescale the value of linear_step with the byte size of the
11080         // pointee type.
11081         if (!ParamAttr.HasVarStride &&
11082             (ParamAttr.Kind == Linear || ParamAttr.Kind == LinearRef))
11083           ParamAttr.StrideOrArg = ParamAttr.StrideOrArg * PtrRescalingFactor;
11084         ++SI;
11085         ++MI;
11086       }
11087       llvm::APSInt VLENVal;
11088       SourceLocation ExprLoc;
11089       const Expr *VLENExpr = Attr->getSimdlen();
11090       if (VLENExpr) {
11091         VLENVal = VLENExpr->EvaluateKnownConstInt(C);
11092         ExprLoc = VLENExpr->getExprLoc();
11093       }
11094       OMPDeclareSimdDeclAttr::BranchStateTy State = Attr->getBranchState();
11095       if (CGM.getTriple().isX86()) {
11096         emitX86DeclareSimdFunction(FD, Fn, VLENVal, ParamAttrs, State);
11097       } else if (CGM.getTriple().getArch() == llvm::Triple::aarch64) {
11098         unsigned VLEN = VLENVal.getExtValue();
11099         StringRef MangledName = Fn->getName();
11100         if (CGM.getTarget().hasFeature("sve"))
11101           emitAArch64DeclareSimdFunction(CGM, FD, VLEN, ParamAttrs, State,
11102                                          MangledName, 's', 128, Fn, ExprLoc);
11103         else if (CGM.getTarget().hasFeature("neon"))
11104           emitAArch64DeclareSimdFunction(CGM, FD, VLEN, ParamAttrs, State,
11105                                          MangledName, 'n', 128, Fn, ExprLoc);
11106       }
11107     }
11108     FD = FD->getPreviousDecl();
11109   }
11110 }
11111 
11112 namespace {
11113 /// Cleanup action for doacross support.
11114 class DoacrossCleanupTy final : public EHScopeStack::Cleanup {
11115 public:
11116   static const int DoacrossFinArgs = 2;
11117 
11118 private:
11119   llvm::FunctionCallee RTLFn;
11120   llvm::Value *Args[DoacrossFinArgs];
11121 
11122 public:
DoacrossCleanupTy(llvm::FunctionCallee RTLFn,ArrayRef<llvm::Value * > CallArgs)11123   DoacrossCleanupTy(llvm::FunctionCallee RTLFn,
11124                     ArrayRef<llvm::Value *> CallArgs)
11125       : RTLFn(RTLFn) {
11126     assert(CallArgs.size() == DoacrossFinArgs);
11127     std::copy(CallArgs.begin(), CallArgs.end(), std::begin(Args));
11128   }
Emit(CodeGenFunction & CGF,Flags)11129   void Emit(CodeGenFunction &CGF, Flags /*flags*/) override {
11130     if (!CGF.HaveInsertPoint())
11131       return;
11132     CGF.EmitRuntimeCall(RTLFn, Args);
11133   }
11134 };
11135 } // namespace
11136 
emitDoacrossInit(CodeGenFunction & CGF,const OMPLoopDirective & D,ArrayRef<Expr * > NumIterations)11137 void CGOpenMPRuntime::emitDoacrossInit(CodeGenFunction &CGF,
11138                                        const OMPLoopDirective &D,
11139                                        ArrayRef<Expr *> NumIterations) {
11140   if (!CGF.HaveInsertPoint())
11141     return;
11142 
11143   ASTContext &C = CGM.getContext();
11144   QualType Int64Ty = C.getIntTypeForBitwidth(/*DestWidth=*/64, /*Signed=*/true);
11145   RecordDecl *RD;
11146   if (KmpDimTy.isNull()) {
11147     // Build struct kmp_dim {  // loop bounds info casted to kmp_int64
11148     //  kmp_int64 lo; // lower
11149     //  kmp_int64 up; // upper
11150     //  kmp_int64 st; // stride
11151     // };
11152     RD = C.buildImplicitRecord("kmp_dim");
11153     RD->startDefinition();
11154     addFieldToRecordDecl(C, RD, Int64Ty);
11155     addFieldToRecordDecl(C, RD, Int64Ty);
11156     addFieldToRecordDecl(C, RD, Int64Ty);
11157     RD->completeDefinition();
11158     KmpDimTy = C.getRecordType(RD);
11159   } else {
11160     RD = cast<RecordDecl>(KmpDimTy->getAsTagDecl());
11161   }
11162   llvm::APInt Size(/*numBits=*/32, NumIterations.size());
11163   QualType ArrayTy = C.getConstantArrayType(KmpDimTy, Size, nullptr,
11164                                             ArraySizeModifier::Normal, 0);
11165 
11166   Address DimsAddr = CGF.CreateMemTemp(ArrayTy, "dims");
11167   CGF.EmitNullInitialization(DimsAddr, ArrayTy);
11168   enum { LowerFD = 0, UpperFD, StrideFD };
11169   // Fill dims with data.
11170   for (unsigned I = 0, E = NumIterations.size(); I < E; ++I) {
11171     LValue DimsLVal = CGF.MakeAddrLValue(
11172         CGF.Builder.CreateConstArrayGEP(DimsAddr, I), KmpDimTy);
11173     // dims.upper = num_iterations;
11174     LValue UpperLVal = CGF.EmitLValueForField(
11175         DimsLVal, *std::next(RD->field_begin(), UpperFD));
11176     llvm::Value *NumIterVal = CGF.EmitScalarConversion(
11177         CGF.EmitScalarExpr(NumIterations[I]), NumIterations[I]->getType(),
11178         Int64Ty, NumIterations[I]->getExprLoc());
11179     CGF.EmitStoreOfScalar(NumIterVal, UpperLVal);
11180     // dims.stride = 1;
11181     LValue StrideLVal = CGF.EmitLValueForField(
11182         DimsLVal, *std::next(RD->field_begin(), StrideFD));
11183     CGF.EmitStoreOfScalar(llvm::ConstantInt::getSigned(CGM.Int64Ty, /*V=*/1),
11184                           StrideLVal);
11185   }
11186 
11187   // Build call void __kmpc_doacross_init(ident_t *loc, kmp_int32 gtid,
11188   // kmp_int32 num_dims, struct kmp_dim * dims);
11189   llvm::Value *Args[] = {
11190       emitUpdateLocation(CGF, D.getBeginLoc()),
11191       getThreadID(CGF, D.getBeginLoc()),
11192       llvm::ConstantInt::getSigned(CGM.Int32Ty, NumIterations.size()),
11193       CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
11194           CGF.Builder.CreateConstArrayGEP(DimsAddr, 0).emitRawPointer(CGF),
11195           CGM.VoidPtrTy)};
11196 
11197   llvm::FunctionCallee RTLFn = OMPBuilder.getOrCreateRuntimeFunction(
11198       CGM.getModule(), OMPRTL___kmpc_doacross_init);
11199   CGF.EmitRuntimeCall(RTLFn, Args);
11200   llvm::Value *FiniArgs[DoacrossCleanupTy::DoacrossFinArgs] = {
11201       emitUpdateLocation(CGF, D.getEndLoc()), getThreadID(CGF, D.getEndLoc())};
11202   llvm::FunctionCallee FiniRTLFn = OMPBuilder.getOrCreateRuntimeFunction(
11203       CGM.getModule(), OMPRTL___kmpc_doacross_fini);
11204   CGF.EHStack.pushCleanup<DoacrossCleanupTy>(NormalAndEHCleanup, FiniRTLFn,
11205                                              llvm::ArrayRef(FiniArgs));
11206 }
11207 
11208 template <typename T>
EmitDoacrossOrdered(CodeGenFunction & CGF,CodeGenModule & CGM,const T * C,llvm::Value * ULoc,llvm::Value * ThreadID)11209 static void EmitDoacrossOrdered(CodeGenFunction &CGF, CodeGenModule &CGM,
11210                                 const T *C, llvm::Value *ULoc,
11211                                 llvm::Value *ThreadID) {
11212   QualType Int64Ty =
11213       CGM.getContext().getIntTypeForBitwidth(/*DestWidth=*/64, /*Signed=*/1);
11214   llvm::APInt Size(/*numBits=*/32, C->getNumLoops());
11215   QualType ArrayTy = CGM.getContext().getConstantArrayType(
11216       Int64Ty, Size, nullptr, ArraySizeModifier::Normal, 0);
11217   Address CntAddr = CGF.CreateMemTemp(ArrayTy, ".cnt.addr");
11218   for (unsigned I = 0, E = C->getNumLoops(); I < E; ++I) {
11219     const Expr *CounterVal = C->getLoopData(I);
11220     assert(CounterVal);
11221     llvm::Value *CntVal = CGF.EmitScalarConversion(
11222         CGF.EmitScalarExpr(CounterVal), CounterVal->getType(), Int64Ty,
11223         CounterVal->getExprLoc());
11224     CGF.EmitStoreOfScalar(CntVal, CGF.Builder.CreateConstArrayGEP(CntAddr, I),
11225                           /*Volatile=*/false, Int64Ty);
11226   }
11227   llvm::Value *Args[] = {
11228       ULoc, ThreadID,
11229       CGF.Builder.CreateConstArrayGEP(CntAddr, 0).emitRawPointer(CGF)};
11230   llvm::FunctionCallee RTLFn;
11231   llvm::OpenMPIRBuilder &OMPBuilder = CGM.getOpenMPRuntime().getOMPBuilder();
11232   OMPDoacrossKind<T> ODK;
11233   if (ODK.isSource(C)) {
11234     RTLFn = OMPBuilder.getOrCreateRuntimeFunction(CGM.getModule(),
11235                                                   OMPRTL___kmpc_doacross_post);
11236   } else {
11237     assert(ODK.isSink(C) && "Expect sink modifier.");
11238     RTLFn = OMPBuilder.getOrCreateRuntimeFunction(CGM.getModule(),
11239                                                   OMPRTL___kmpc_doacross_wait);
11240   }
11241   CGF.EmitRuntimeCall(RTLFn, Args);
11242 }
11243 
emitDoacrossOrdered(CodeGenFunction & CGF,const OMPDependClause * C)11244 void CGOpenMPRuntime::emitDoacrossOrdered(CodeGenFunction &CGF,
11245                                           const OMPDependClause *C) {
11246   return EmitDoacrossOrdered<OMPDependClause>(
11247       CGF, CGM, C, emitUpdateLocation(CGF, C->getBeginLoc()),
11248       getThreadID(CGF, C->getBeginLoc()));
11249 }
11250 
emitDoacrossOrdered(CodeGenFunction & CGF,const OMPDoacrossClause * C)11251 void CGOpenMPRuntime::emitDoacrossOrdered(CodeGenFunction &CGF,
11252                                           const OMPDoacrossClause *C) {
11253   return EmitDoacrossOrdered<OMPDoacrossClause>(
11254       CGF, CGM, C, emitUpdateLocation(CGF, C->getBeginLoc()),
11255       getThreadID(CGF, C->getBeginLoc()));
11256 }
11257 
emitCall(CodeGenFunction & CGF,SourceLocation Loc,llvm::FunctionCallee Callee,ArrayRef<llvm::Value * > Args) const11258 void CGOpenMPRuntime::emitCall(CodeGenFunction &CGF, SourceLocation Loc,
11259                                llvm::FunctionCallee Callee,
11260                                ArrayRef<llvm::Value *> Args) const {
11261   assert(Loc.isValid() && "Outlined function call location must be valid.");
11262   auto DL = ApplyDebugLocation::CreateDefaultArtificial(CGF, Loc);
11263 
11264   if (auto *Fn = dyn_cast<llvm::Function>(Callee.getCallee())) {
11265     if (Fn->doesNotThrow()) {
11266       CGF.EmitNounwindRuntimeCall(Fn, Args);
11267       return;
11268     }
11269   }
11270   CGF.EmitRuntimeCall(Callee, Args);
11271 }
11272 
emitOutlinedFunctionCall(CodeGenFunction & CGF,SourceLocation Loc,llvm::FunctionCallee OutlinedFn,ArrayRef<llvm::Value * > Args) const11273 void CGOpenMPRuntime::emitOutlinedFunctionCall(
11274     CodeGenFunction &CGF, SourceLocation Loc, llvm::FunctionCallee OutlinedFn,
11275     ArrayRef<llvm::Value *> Args) const {
11276   emitCall(CGF, Loc, OutlinedFn, Args);
11277 }
11278 
emitFunctionProlog(CodeGenFunction & CGF,const Decl * D)11279 void CGOpenMPRuntime::emitFunctionProlog(CodeGenFunction &CGF, const Decl *D) {
11280   if (const auto *FD = dyn_cast<FunctionDecl>(D))
11281     if (OMPDeclareTargetDeclAttr::isDeclareTargetDeclaration(FD))
11282       HasEmittedDeclareTargetRegion = true;
11283 }
11284 
getParameterAddress(CodeGenFunction & CGF,const VarDecl * NativeParam,const VarDecl * TargetParam) const11285 Address CGOpenMPRuntime::getParameterAddress(CodeGenFunction &CGF,
11286                                              const VarDecl *NativeParam,
11287                                              const VarDecl *TargetParam) const {
11288   return CGF.GetAddrOfLocalVar(NativeParam);
11289 }
11290 
11291 /// Return allocator value from expression, or return a null allocator (default
11292 /// when no allocator specified).
getAllocatorVal(CodeGenFunction & CGF,const Expr * Allocator)11293 static llvm::Value *getAllocatorVal(CodeGenFunction &CGF,
11294                                     const Expr *Allocator) {
11295   llvm::Value *AllocVal;
11296   if (Allocator) {
11297     AllocVal = CGF.EmitScalarExpr(Allocator);
11298     // According to the standard, the original allocator type is a enum
11299     // (integer). Convert to pointer type, if required.
11300     AllocVal = CGF.EmitScalarConversion(AllocVal, Allocator->getType(),
11301                                         CGF.getContext().VoidPtrTy,
11302                                         Allocator->getExprLoc());
11303   } else {
11304     // If no allocator specified, it defaults to the null allocator.
11305     AllocVal = llvm::Constant::getNullValue(
11306         CGF.CGM.getTypes().ConvertType(CGF.getContext().VoidPtrTy));
11307   }
11308   return AllocVal;
11309 }
11310 
11311 /// Return the alignment from an allocate directive if present.
getAlignmentValue(CodeGenModule & CGM,const VarDecl * VD)11312 static llvm::Value *getAlignmentValue(CodeGenModule &CGM, const VarDecl *VD) {
11313   std::optional<CharUnits> AllocateAlignment = CGM.getOMPAllocateAlignment(VD);
11314 
11315   if (!AllocateAlignment)
11316     return nullptr;
11317 
11318   return llvm::ConstantInt::get(CGM.SizeTy, AllocateAlignment->getQuantity());
11319 }
11320 
getAddressOfLocalVariable(CodeGenFunction & CGF,const VarDecl * VD)11321 Address CGOpenMPRuntime::getAddressOfLocalVariable(CodeGenFunction &CGF,
11322                                                    const VarDecl *VD) {
11323   if (!VD)
11324     return Address::invalid();
11325   Address UntiedAddr = Address::invalid();
11326   Address UntiedRealAddr = Address::invalid();
11327   auto It = FunctionToUntiedTaskStackMap.find(CGF.CurFn);
11328   if (It != FunctionToUntiedTaskStackMap.end()) {
11329     const UntiedLocalVarsAddressesMap &UntiedData =
11330         UntiedLocalVarsStack[It->second];
11331     auto I = UntiedData.find(VD);
11332     if (I != UntiedData.end()) {
11333       UntiedAddr = I->second.first;
11334       UntiedRealAddr = I->second.second;
11335     }
11336   }
11337   const VarDecl *CVD = VD->getCanonicalDecl();
11338   if (CVD->hasAttr<OMPAllocateDeclAttr>()) {
11339     // Use the default allocation.
11340     if (!isAllocatableDecl(VD))
11341       return UntiedAddr;
11342     llvm::Value *Size;
11343     CharUnits Align = CGM.getContext().getDeclAlign(CVD);
11344     if (CVD->getType()->isVariablyModifiedType()) {
11345       Size = CGF.getTypeSize(CVD->getType());
11346       // Align the size: ((size + align - 1) / align) * align
11347       Size = CGF.Builder.CreateNUWAdd(
11348           Size, CGM.getSize(Align - CharUnits::fromQuantity(1)));
11349       Size = CGF.Builder.CreateUDiv(Size, CGM.getSize(Align));
11350       Size = CGF.Builder.CreateNUWMul(Size, CGM.getSize(Align));
11351     } else {
11352       CharUnits Sz = CGM.getContext().getTypeSizeInChars(CVD->getType());
11353       Size = CGM.getSize(Sz.alignTo(Align));
11354     }
11355     llvm::Value *ThreadID = getThreadID(CGF, CVD->getBeginLoc());
11356     const auto *AA = CVD->getAttr<OMPAllocateDeclAttr>();
11357     const Expr *Allocator = AA->getAllocator();
11358     llvm::Value *AllocVal = getAllocatorVal(CGF, Allocator);
11359     llvm::Value *Alignment = getAlignmentValue(CGM, CVD);
11360     SmallVector<llvm::Value *, 4> Args;
11361     Args.push_back(ThreadID);
11362     if (Alignment)
11363       Args.push_back(Alignment);
11364     Args.push_back(Size);
11365     Args.push_back(AllocVal);
11366     llvm::omp::RuntimeFunction FnID =
11367         Alignment ? OMPRTL___kmpc_aligned_alloc : OMPRTL___kmpc_alloc;
11368     llvm::Value *Addr = CGF.EmitRuntimeCall(
11369         OMPBuilder.getOrCreateRuntimeFunction(CGM.getModule(), FnID), Args,
11370         getName({CVD->getName(), ".void.addr"}));
11371     llvm::FunctionCallee FiniRTLFn = OMPBuilder.getOrCreateRuntimeFunction(
11372         CGM.getModule(), OMPRTL___kmpc_free);
11373     QualType Ty = CGM.getContext().getPointerType(CVD->getType());
11374     Addr = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
11375         Addr, CGF.ConvertTypeForMem(Ty), getName({CVD->getName(), ".addr"}));
11376     if (UntiedAddr.isValid())
11377       CGF.EmitStoreOfScalar(Addr, UntiedAddr, /*Volatile=*/false, Ty);
11378 
11379     // Cleanup action for allocate support.
11380     class OMPAllocateCleanupTy final : public EHScopeStack::Cleanup {
11381       llvm::FunctionCallee RTLFn;
11382       SourceLocation::UIntTy LocEncoding;
11383       Address Addr;
11384       const Expr *AllocExpr;
11385 
11386     public:
11387       OMPAllocateCleanupTy(llvm::FunctionCallee RTLFn,
11388                            SourceLocation::UIntTy LocEncoding, Address Addr,
11389                            const Expr *AllocExpr)
11390           : RTLFn(RTLFn), LocEncoding(LocEncoding), Addr(Addr),
11391             AllocExpr(AllocExpr) {}
11392       void Emit(CodeGenFunction &CGF, Flags /*flags*/) override {
11393         if (!CGF.HaveInsertPoint())
11394           return;
11395         llvm::Value *Args[3];
11396         Args[0] = CGF.CGM.getOpenMPRuntime().getThreadID(
11397             CGF, SourceLocation::getFromRawEncoding(LocEncoding));
11398         Args[1] = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
11399             Addr.emitRawPointer(CGF), CGF.VoidPtrTy);
11400         llvm::Value *AllocVal = getAllocatorVal(CGF, AllocExpr);
11401         Args[2] = AllocVal;
11402         CGF.EmitRuntimeCall(RTLFn, Args);
11403       }
11404     };
11405     Address VDAddr =
11406         UntiedRealAddr.isValid()
11407             ? UntiedRealAddr
11408             : Address(Addr, CGF.ConvertTypeForMem(CVD->getType()), Align);
11409     CGF.EHStack.pushCleanup<OMPAllocateCleanupTy>(
11410         NormalAndEHCleanup, FiniRTLFn, CVD->getLocation().getRawEncoding(),
11411         VDAddr, Allocator);
11412     if (UntiedRealAddr.isValid())
11413       if (auto *Region =
11414               dyn_cast_or_null<CGOpenMPRegionInfo>(CGF.CapturedStmtInfo))
11415         Region->emitUntiedSwitch(CGF);
11416     return VDAddr;
11417   }
11418   return UntiedAddr;
11419 }
11420 
isLocalVarInUntiedTask(CodeGenFunction & CGF,const VarDecl * VD) const11421 bool CGOpenMPRuntime::isLocalVarInUntiedTask(CodeGenFunction &CGF,
11422                                              const VarDecl *VD) const {
11423   auto It = FunctionToUntiedTaskStackMap.find(CGF.CurFn);
11424   if (It == FunctionToUntiedTaskStackMap.end())
11425     return false;
11426   return UntiedLocalVarsStack[It->second].count(VD) > 0;
11427 }
11428 
NontemporalDeclsRAII(CodeGenModule & CGM,const OMPLoopDirective & S)11429 CGOpenMPRuntime::NontemporalDeclsRAII::NontemporalDeclsRAII(
11430     CodeGenModule &CGM, const OMPLoopDirective &S)
11431     : CGM(CGM), NeedToPush(S.hasClausesOfKind<OMPNontemporalClause>()) {
11432   assert(CGM.getLangOpts().OpenMP && "Not in OpenMP mode.");
11433   if (!NeedToPush)
11434     return;
11435   NontemporalDeclsSet &DS =
11436       CGM.getOpenMPRuntime().NontemporalDeclsStack.emplace_back();
11437   for (const auto *C : S.getClausesOfKind<OMPNontemporalClause>()) {
11438     for (const Stmt *Ref : C->private_refs()) {
11439       const auto *SimpleRefExpr = cast<Expr>(Ref)->IgnoreParenImpCasts();
11440       const ValueDecl *VD;
11441       if (const auto *DRE = dyn_cast<DeclRefExpr>(SimpleRefExpr)) {
11442         VD = DRE->getDecl();
11443       } else {
11444         const auto *ME = cast<MemberExpr>(SimpleRefExpr);
11445         assert((ME->isImplicitCXXThis() ||
11446                 isa<CXXThisExpr>(ME->getBase()->IgnoreParenImpCasts())) &&
11447                "Expected member of current class.");
11448         VD = ME->getMemberDecl();
11449       }
11450       DS.insert(VD);
11451     }
11452   }
11453 }
11454 
~NontemporalDeclsRAII()11455 CGOpenMPRuntime::NontemporalDeclsRAII::~NontemporalDeclsRAII() {
11456   if (!NeedToPush)
11457     return;
11458   CGM.getOpenMPRuntime().NontemporalDeclsStack.pop_back();
11459 }
11460 
UntiedTaskLocalDeclsRAII(CodeGenFunction & CGF,const llvm::MapVector<CanonicalDeclPtr<const VarDecl>,std::pair<Address,Address>> & LocalVars)11461 CGOpenMPRuntime::UntiedTaskLocalDeclsRAII::UntiedTaskLocalDeclsRAII(
11462     CodeGenFunction &CGF,
11463     const llvm::MapVector<CanonicalDeclPtr<const VarDecl>,
11464                           std::pair<Address, Address>> &LocalVars)
11465     : CGM(CGF.CGM), NeedToPush(!LocalVars.empty()) {
11466   if (!NeedToPush)
11467     return;
11468   CGM.getOpenMPRuntime().FunctionToUntiedTaskStackMap.try_emplace(
11469       CGF.CurFn, CGM.getOpenMPRuntime().UntiedLocalVarsStack.size());
11470   CGM.getOpenMPRuntime().UntiedLocalVarsStack.push_back(LocalVars);
11471 }
11472 
~UntiedTaskLocalDeclsRAII()11473 CGOpenMPRuntime::UntiedTaskLocalDeclsRAII::~UntiedTaskLocalDeclsRAII() {
11474   if (!NeedToPush)
11475     return;
11476   CGM.getOpenMPRuntime().UntiedLocalVarsStack.pop_back();
11477 }
11478 
isNontemporalDecl(const ValueDecl * VD) const11479 bool CGOpenMPRuntime::isNontemporalDecl(const ValueDecl *VD) const {
11480   assert(CGM.getLangOpts().OpenMP && "Not in OpenMP mode.");
11481 
11482   return llvm::any_of(
11483       CGM.getOpenMPRuntime().NontemporalDeclsStack,
11484       [VD](const NontemporalDeclsSet &Set) { return Set.contains(VD); });
11485 }
11486 
tryToDisableInnerAnalysis(const OMPExecutableDirective & S,llvm::DenseSet<CanonicalDeclPtr<const Decl>> & NeedToAddForLPCsAsDisabled) const11487 void CGOpenMPRuntime::LastprivateConditionalRAII::tryToDisableInnerAnalysis(
11488     const OMPExecutableDirective &S,
11489     llvm::DenseSet<CanonicalDeclPtr<const Decl>> &NeedToAddForLPCsAsDisabled)
11490     const {
11491   llvm::DenseSet<CanonicalDeclPtr<const Decl>> NeedToCheckForLPCs;
11492   // Vars in target/task regions must be excluded completely.
11493   if (isOpenMPTargetExecutionDirective(S.getDirectiveKind()) ||
11494       isOpenMPTaskingDirective(S.getDirectiveKind())) {
11495     SmallVector<OpenMPDirectiveKind, 4> CaptureRegions;
11496     getOpenMPCaptureRegions(CaptureRegions, S.getDirectiveKind());
11497     const CapturedStmt *CS = S.getCapturedStmt(CaptureRegions.front());
11498     for (const CapturedStmt::Capture &Cap : CS->captures()) {
11499       if (Cap.capturesVariable() || Cap.capturesVariableByCopy())
11500         NeedToCheckForLPCs.insert(Cap.getCapturedVar());
11501     }
11502   }
11503   // Exclude vars in private clauses.
11504   for (const auto *C : S.getClausesOfKind<OMPPrivateClause>()) {
11505     for (const Expr *Ref : C->varlists()) {
11506       if (!Ref->getType()->isScalarType())
11507         continue;
11508       const auto *DRE = dyn_cast<DeclRefExpr>(Ref->IgnoreParenImpCasts());
11509       if (!DRE)
11510         continue;
11511       NeedToCheckForLPCs.insert(DRE->getDecl());
11512     }
11513   }
11514   for (const auto *C : S.getClausesOfKind<OMPFirstprivateClause>()) {
11515     for (const Expr *Ref : C->varlists()) {
11516       if (!Ref->getType()->isScalarType())
11517         continue;
11518       const auto *DRE = dyn_cast<DeclRefExpr>(Ref->IgnoreParenImpCasts());
11519       if (!DRE)
11520         continue;
11521       NeedToCheckForLPCs.insert(DRE->getDecl());
11522     }
11523   }
11524   for (const auto *C : S.getClausesOfKind<OMPLastprivateClause>()) {
11525     for (const Expr *Ref : C->varlists()) {
11526       if (!Ref->getType()->isScalarType())
11527         continue;
11528       const auto *DRE = dyn_cast<DeclRefExpr>(Ref->IgnoreParenImpCasts());
11529       if (!DRE)
11530         continue;
11531       NeedToCheckForLPCs.insert(DRE->getDecl());
11532     }
11533   }
11534   for (const auto *C : S.getClausesOfKind<OMPReductionClause>()) {
11535     for (const Expr *Ref : C->varlists()) {
11536       if (!Ref->getType()->isScalarType())
11537         continue;
11538       const auto *DRE = dyn_cast<DeclRefExpr>(Ref->IgnoreParenImpCasts());
11539       if (!DRE)
11540         continue;
11541       NeedToCheckForLPCs.insert(DRE->getDecl());
11542     }
11543   }
11544   for (const auto *C : S.getClausesOfKind<OMPLinearClause>()) {
11545     for (const Expr *Ref : C->varlists()) {
11546       if (!Ref->getType()->isScalarType())
11547         continue;
11548       const auto *DRE = dyn_cast<DeclRefExpr>(Ref->IgnoreParenImpCasts());
11549       if (!DRE)
11550         continue;
11551       NeedToCheckForLPCs.insert(DRE->getDecl());
11552     }
11553   }
11554   for (const Decl *VD : NeedToCheckForLPCs) {
11555     for (const LastprivateConditionalData &Data :
11556          llvm::reverse(CGM.getOpenMPRuntime().LastprivateConditionalStack)) {
11557       if (Data.DeclToUniqueName.count(VD) > 0) {
11558         if (!Data.Disabled)
11559           NeedToAddForLPCsAsDisabled.insert(VD);
11560         break;
11561       }
11562     }
11563   }
11564 }
11565 
LastprivateConditionalRAII(CodeGenFunction & CGF,const OMPExecutableDirective & S,LValue IVLVal)11566 CGOpenMPRuntime::LastprivateConditionalRAII::LastprivateConditionalRAII(
11567     CodeGenFunction &CGF, const OMPExecutableDirective &S, LValue IVLVal)
11568     : CGM(CGF.CGM),
11569       Action((CGM.getLangOpts().OpenMP >= 50 &&
11570               llvm::any_of(S.getClausesOfKind<OMPLastprivateClause>(),
11571                            [](const OMPLastprivateClause *C) {
11572                              return C->getKind() ==
11573                                     OMPC_LASTPRIVATE_conditional;
11574                            }))
11575                  ? ActionToDo::PushAsLastprivateConditional
11576                  : ActionToDo::DoNotPush) {
11577   assert(CGM.getLangOpts().OpenMP && "Not in OpenMP mode.");
11578   if (CGM.getLangOpts().OpenMP < 50 || Action == ActionToDo::DoNotPush)
11579     return;
11580   assert(Action == ActionToDo::PushAsLastprivateConditional &&
11581          "Expected a push action.");
11582   LastprivateConditionalData &Data =
11583       CGM.getOpenMPRuntime().LastprivateConditionalStack.emplace_back();
11584   for (const auto *C : S.getClausesOfKind<OMPLastprivateClause>()) {
11585     if (C->getKind() != OMPC_LASTPRIVATE_conditional)
11586       continue;
11587 
11588     for (const Expr *Ref : C->varlists()) {
11589       Data.DeclToUniqueName.insert(std::make_pair(
11590           cast<DeclRefExpr>(Ref->IgnoreParenImpCasts())->getDecl(),
11591           SmallString<16>(generateUniqueName(CGM, "pl_cond", Ref))));
11592     }
11593   }
11594   Data.IVLVal = IVLVal;
11595   Data.Fn = CGF.CurFn;
11596 }
11597 
LastprivateConditionalRAII(CodeGenFunction & CGF,const OMPExecutableDirective & S)11598 CGOpenMPRuntime::LastprivateConditionalRAII::LastprivateConditionalRAII(
11599     CodeGenFunction &CGF, const OMPExecutableDirective &S)
11600     : CGM(CGF.CGM), Action(ActionToDo::DoNotPush) {
11601   assert(CGM.getLangOpts().OpenMP && "Not in OpenMP mode.");
11602   if (CGM.getLangOpts().OpenMP < 50)
11603     return;
11604   llvm::DenseSet<CanonicalDeclPtr<const Decl>> NeedToAddForLPCsAsDisabled;
11605   tryToDisableInnerAnalysis(S, NeedToAddForLPCsAsDisabled);
11606   if (!NeedToAddForLPCsAsDisabled.empty()) {
11607     Action = ActionToDo::DisableLastprivateConditional;
11608     LastprivateConditionalData &Data =
11609         CGM.getOpenMPRuntime().LastprivateConditionalStack.emplace_back();
11610     for (const Decl *VD : NeedToAddForLPCsAsDisabled)
11611       Data.DeclToUniqueName.insert(std::make_pair(VD, SmallString<16>()));
11612     Data.Fn = CGF.CurFn;
11613     Data.Disabled = true;
11614   }
11615 }
11616 
11617 CGOpenMPRuntime::LastprivateConditionalRAII
disable(CodeGenFunction & CGF,const OMPExecutableDirective & S)11618 CGOpenMPRuntime::LastprivateConditionalRAII::disable(
11619     CodeGenFunction &CGF, const OMPExecutableDirective &S) {
11620   return LastprivateConditionalRAII(CGF, S);
11621 }
11622 
~LastprivateConditionalRAII()11623 CGOpenMPRuntime::LastprivateConditionalRAII::~LastprivateConditionalRAII() {
11624   if (CGM.getLangOpts().OpenMP < 50)
11625     return;
11626   if (Action == ActionToDo::DisableLastprivateConditional) {
11627     assert(CGM.getOpenMPRuntime().LastprivateConditionalStack.back().Disabled &&
11628            "Expected list of disabled private vars.");
11629     CGM.getOpenMPRuntime().LastprivateConditionalStack.pop_back();
11630   }
11631   if (Action == ActionToDo::PushAsLastprivateConditional) {
11632     assert(
11633         !CGM.getOpenMPRuntime().LastprivateConditionalStack.back().Disabled &&
11634         "Expected list of lastprivate conditional vars.");
11635     CGM.getOpenMPRuntime().LastprivateConditionalStack.pop_back();
11636   }
11637 }
11638 
emitLastprivateConditionalInit(CodeGenFunction & CGF,const VarDecl * VD)11639 Address CGOpenMPRuntime::emitLastprivateConditionalInit(CodeGenFunction &CGF,
11640                                                         const VarDecl *VD) {
11641   ASTContext &C = CGM.getContext();
11642   auto I = LastprivateConditionalToTypes.find(CGF.CurFn);
11643   if (I == LastprivateConditionalToTypes.end())
11644     I = LastprivateConditionalToTypes.try_emplace(CGF.CurFn).first;
11645   QualType NewType;
11646   const FieldDecl *VDField;
11647   const FieldDecl *FiredField;
11648   LValue BaseLVal;
11649   auto VI = I->getSecond().find(VD);
11650   if (VI == I->getSecond().end()) {
11651     RecordDecl *RD = C.buildImplicitRecord("lasprivate.conditional");
11652     RD->startDefinition();
11653     VDField = addFieldToRecordDecl(C, RD, VD->getType().getNonReferenceType());
11654     FiredField = addFieldToRecordDecl(C, RD, C.CharTy);
11655     RD->completeDefinition();
11656     NewType = C.getRecordType(RD);
11657     Address Addr = CGF.CreateMemTemp(NewType, C.getDeclAlign(VD), VD->getName());
11658     BaseLVal = CGF.MakeAddrLValue(Addr, NewType, AlignmentSource::Decl);
11659     I->getSecond().try_emplace(VD, NewType, VDField, FiredField, BaseLVal);
11660   } else {
11661     NewType = std::get<0>(VI->getSecond());
11662     VDField = std::get<1>(VI->getSecond());
11663     FiredField = std::get<2>(VI->getSecond());
11664     BaseLVal = std::get<3>(VI->getSecond());
11665   }
11666   LValue FiredLVal =
11667       CGF.EmitLValueForField(BaseLVal, FiredField);
11668   CGF.EmitStoreOfScalar(
11669       llvm::ConstantInt::getNullValue(CGF.ConvertTypeForMem(C.CharTy)),
11670       FiredLVal);
11671   return CGF.EmitLValueForField(BaseLVal, VDField).getAddress();
11672 }
11673 
11674 namespace {
11675 /// Checks if the lastprivate conditional variable is referenced in LHS.
11676 class LastprivateConditionalRefChecker final
11677     : public ConstStmtVisitor<LastprivateConditionalRefChecker, bool> {
11678   ArrayRef<CGOpenMPRuntime::LastprivateConditionalData> LPM;
11679   const Expr *FoundE = nullptr;
11680   const Decl *FoundD = nullptr;
11681   StringRef UniqueDeclName;
11682   LValue IVLVal;
11683   llvm::Function *FoundFn = nullptr;
11684   SourceLocation Loc;
11685 
11686 public:
VisitDeclRefExpr(const DeclRefExpr * E)11687   bool VisitDeclRefExpr(const DeclRefExpr *E) {
11688     for (const CGOpenMPRuntime::LastprivateConditionalData &D :
11689          llvm::reverse(LPM)) {
11690       auto It = D.DeclToUniqueName.find(E->getDecl());
11691       if (It == D.DeclToUniqueName.end())
11692         continue;
11693       if (D.Disabled)
11694         return false;
11695       FoundE = E;
11696       FoundD = E->getDecl()->getCanonicalDecl();
11697       UniqueDeclName = It->second;
11698       IVLVal = D.IVLVal;
11699       FoundFn = D.Fn;
11700       break;
11701     }
11702     return FoundE == E;
11703   }
VisitMemberExpr(const MemberExpr * E)11704   bool VisitMemberExpr(const MemberExpr *E) {
11705     if (!CodeGenFunction::IsWrappedCXXThis(E->getBase()))
11706       return false;
11707     for (const CGOpenMPRuntime::LastprivateConditionalData &D :
11708          llvm::reverse(LPM)) {
11709       auto It = D.DeclToUniqueName.find(E->getMemberDecl());
11710       if (It == D.DeclToUniqueName.end())
11711         continue;
11712       if (D.Disabled)
11713         return false;
11714       FoundE = E;
11715       FoundD = E->getMemberDecl()->getCanonicalDecl();
11716       UniqueDeclName = It->second;
11717       IVLVal = D.IVLVal;
11718       FoundFn = D.Fn;
11719       break;
11720     }
11721     return FoundE == E;
11722   }
VisitStmt(const Stmt * S)11723   bool VisitStmt(const Stmt *S) {
11724     for (const Stmt *Child : S->children()) {
11725       if (!Child)
11726         continue;
11727       if (const auto *E = dyn_cast<Expr>(Child))
11728         if (!E->isGLValue())
11729           continue;
11730       if (Visit(Child))
11731         return true;
11732     }
11733     return false;
11734   }
LastprivateConditionalRefChecker(ArrayRef<CGOpenMPRuntime::LastprivateConditionalData> LPM)11735   explicit LastprivateConditionalRefChecker(
11736       ArrayRef<CGOpenMPRuntime::LastprivateConditionalData> LPM)
11737       : LPM(LPM) {}
11738   std::tuple<const Expr *, const Decl *, StringRef, LValue, llvm::Function *>
getFoundData() const11739   getFoundData() const {
11740     return std::make_tuple(FoundE, FoundD, UniqueDeclName, IVLVal, FoundFn);
11741   }
11742 };
11743 } // namespace
11744 
emitLastprivateConditionalUpdate(CodeGenFunction & CGF,LValue IVLVal,StringRef UniqueDeclName,LValue LVal,SourceLocation Loc)11745 void CGOpenMPRuntime::emitLastprivateConditionalUpdate(CodeGenFunction &CGF,
11746                                                        LValue IVLVal,
11747                                                        StringRef UniqueDeclName,
11748                                                        LValue LVal,
11749                                                        SourceLocation Loc) {
11750   // Last updated loop counter for the lastprivate conditional var.
11751   // int<xx> last_iv = 0;
11752   llvm::Type *LLIVTy = CGF.ConvertTypeForMem(IVLVal.getType());
11753   llvm::Constant *LastIV = OMPBuilder.getOrCreateInternalVariable(
11754       LLIVTy, getName({UniqueDeclName, "iv"}));
11755   cast<llvm::GlobalVariable>(LastIV)->setAlignment(
11756       IVLVal.getAlignment().getAsAlign());
11757   LValue LastIVLVal =
11758       CGF.MakeNaturalAlignRawAddrLValue(LastIV, IVLVal.getType());
11759 
11760   // Last value of the lastprivate conditional.
11761   // decltype(priv_a) last_a;
11762   llvm::GlobalVariable *Last = OMPBuilder.getOrCreateInternalVariable(
11763       CGF.ConvertTypeForMem(LVal.getType()), UniqueDeclName);
11764   cast<llvm::GlobalVariable>(Last)->setAlignment(
11765       LVal.getAlignment().getAsAlign());
11766   LValue LastLVal =
11767       CGF.MakeRawAddrLValue(Last, LVal.getType(), LVal.getAlignment());
11768 
11769   // Global loop counter. Required to handle inner parallel-for regions.
11770   // iv
11771   llvm::Value *IVVal = CGF.EmitLoadOfScalar(IVLVal, Loc);
11772 
11773   // #pragma omp critical(a)
11774   // if (last_iv <= iv) {
11775   //   last_iv = iv;
11776   //   last_a = priv_a;
11777   // }
11778   auto &&CodeGen = [&LastIVLVal, &IVLVal, IVVal, &LVal, &LastLVal,
11779                     Loc](CodeGenFunction &CGF, PrePostActionTy &Action) {
11780     Action.Enter(CGF);
11781     llvm::Value *LastIVVal = CGF.EmitLoadOfScalar(LastIVLVal, Loc);
11782     // (last_iv <= iv) ? Check if the variable is updated and store new
11783     // value in global var.
11784     llvm::Value *CmpRes;
11785     if (IVLVal.getType()->isSignedIntegerType()) {
11786       CmpRes = CGF.Builder.CreateICmpSLE(LastIVVal, IVVal);
11787     } else {
11788       assert(IVLVal.getType()->isUnsignedIntegerType() &&
11789              "Loop iteration variable must be integer.");
11790       CmpRes = CGF.Builder.CreateICmpULE(LastIVVal, IVVal);
11791     }
11792     llvm::BasicBlock *ThenBB = CGF.createBasicBlock("lp_cond_then");
11793     llvm::BasicBlock *ExitBB = CGF.createBasicBlock("lp_cond_exit");
11794     CGF.Builder.CreateCondBr(CmpRes, ThenBB, ExitBB);
11795     // {
11796     CGF.EmitBlock(ThenBB);
11797 
11798     //   last_iv = iv;
11799     CGF.EmitStoreOfScalar(IVVal, LastIVLVal);
11800 
11801     //   last_a = priv_a;
11802     switch (CGF.getEvaluationKind(LVal.getType())) {
11803     case TEK_Scalar: {
11804       llvm::Value *PrivVal = CGF.EmitLoadOfScalar(LVal, Loc);
11805       CGF.EmitStoreOfScalar(PrivVal, LastLVal);
11806       break;
11807     }
11808     case TEK_Complex: {
11809       CodeGenFunction::ComplexPairTy PrivVal = CGF.EmitLoadOfComplex(LVal, Loc);
11810       CGF.EmitStoreOfComplex(PrivVal, LastLVal, /*isInit=*/false);
11811       break;
11812     }
11813     case TEK_Aggregate:
11814       llvm_unreachable(
11815           "Aggregates are not supported in lastprivate conditional.");
11816     }
11817     // }
11818     CGF.EmitBranch(ExitBB);
11819     // There is no need to emit line number for unconditional branch.
11820     (void)ApplyDebugLocation::CreateEmpty(CGF);
11821     CGF.EmitBlock(ExitBB, /*IsFinished=*/true);
11822   };
11823 
11824   if (CGM.getLangOpts().OpenMPSimd) {
11825     // Do not emit as a critical region as no parallel region could be emitted.
11826     RegionCodeGenTy ThenRCG(CodeGen);
11827     ThenRCG(CGF);
11828   } else {
11829     emitCriticalRegion(CGF, UniqueDeclName, CodeGen, Loc);
11830   }
11831 }
11832 
checkAndEmitLastprivateConditional(CodeGenFunction & CGF,const Expr * LHS)11833 void CGOpenMPRuntime::checkAndEmitLastprivateConditional(CodeGenFunction &CGF,
11834                                                          const Expr *LHS) {
11835   if (CGF.getLangOpts().OpenMP < 50 || LastprivateConditionalStack.empty())
11836     return;
11837   LastprivateConditionalRefChecker Checker(LastprivateConditionalStack);
11838   if (!Checker.Visit(LHS))
11839     return;
11840   const Expr *FoundE;
11841   const Decl *FoundD;
11842   StringRef UniqueDeclName;
11843   LValue IVLVal;
11844   llvm::Function *FoundFn;
11845   std::tie(FoundE, FoundD, UniqueDeclName, IVLVal, FoundFn) =
11846       Checker.getFoundData();
11847   if (FoundFn != CGF.CurFn) {
11848     // Special codegen for inner parallel regions.
11849     // ((struct.lastprivate.conditional*)&priv_a)->Fired = 1;
11850     auto It = LastprivateConditionalToTypes[FoundFn].find(FoundD);
11851     assert(It != LastprivateConditionalToTypes[FoundFn].end() &&
11852            "Lastprivate conditional is not found in outer region.");
11853     QualType StructTy = std::get<0>(It->getSecond());
11854     const FieldDecl* FiredDecl = std::get<2>(It->getSecond());
11855     LValue PrivLVal = CGF.EmitLValue(FoundE);
11856     Address StructAddr = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(
11857         PrivLVal.getAddress(),
11858         CGF.ConvertTypeForMem(CGF.getContext().getPointerType(StructTy)),
11859         CGF.ConvertTypeForMem(StructTy));
11860     LValue BaseLVal =
11861         CGF.MakeAddrLValue(StructAddr, StructTy, AlignmentSource::Decl);
11862     LValue FiredLVal = CGF.EmitLValueForField(BaseLVal, FiredDecl);
11863     CGF.EmitAtomicStore(RValue::get(llvm::ConstantInt::get(
11864                             CGF.ConvertTypeForMem(FiredDecl->getType()), 1)),
11865                         FiredLVal, llvm::AtomicOrdering::Unordered,
11866                         /*IsVolatile=*/true, /*isInit=*/false);
11867     return;
11868   }
11869 
11870   // Private address of the lastprivate conditional in the current context.
11871   // priv_a
11872   LValue LVal = CGF.EmitLValue(FoundE);
11873   emitLastprivateConditionalUpdate(CGF, IVLVal, UniqueDeclName, LVal,
11874                                    FoundE->getExprLoc());
11875 }
11876 
checkAndEmitSharedLastprivateConditional(CodeGenFunction & CGF,const OMPExecutableDirective & D,const llvm::DenseSet<CanonicalDeclPtr<const VarDecl>> & IgnoredDecls)11877 void CGOpenMPRuntime::checkAndEmitSharedLastprivateConditional(
11878     CodeGenFunction &CGF, const OMPExecutableDirective &D,
11879     const llvm::DenseSet<CanonicalDeclPtr<const VarDecl>> &IgnoredDecls) {
11880   if (CGF.getLangOpts().OpenMP < 50 || LastprivateConditionalStack.empty())
11881     return;
11882   auto Range = llvm::reverse(LastprivateConditionalStack);
11883   auto It = llvm::find_if(
11884       Range, [](const LastprivateConditionalData &D) { return !D.Disabled; });
11885   if (It == Range.end() || It->Fn != CGF.CurFn)
11886     return;
11887   auto LPCI = LastprivateConditionalToTypes.find(It->Fn);
11888   assert(LPCI != LastprivateConditionalToTypes.end() &&
11889          "Lastprivates must be registered already.");
11890   SmallVector<OpenMPDirectiveKind, 4> CaptureRegions;
11891   getOpenMPCaptureRegions(CaptureRegions, D.getDirectiveKind());
11892   const CapturedStmt *CS = D.getCapturedStmt(CaptureRegions.back());
11893   for (const auto &Pair : It->DeclToUniqueName) {
11894     const auto *VD = cast<VarDecl>(Pair.first->getCanonicalDecl());
11895     if (!CS->capturesVariable(VD) || IgnoredDecls.contains(VD))
11896       continue;
11897     auto I = LPCI->getSecond().find(Pair.first);
11898     assert(I != LPCI->getSecond().end() &&
11899            "Lastprivate must be rehistered already.");
11900     // bool Cmp = priv_a.Fired != 0;
11901     LValue BaseLVal = std::get<3>(I->getSecond());
11902     LValue FiredLVal =
11903         CGF.EmitLValueForField(BaseLVal, std::get<2>(I->getSecond()));
11904     llvm::Value *Res = CGF.EmitLoadOfScalar(FiredLVal, D.getBeginLoc());
11905     llvm::Value *Cmp = CGF.Builder.CreateIsNotNull(Res);
11906     llvm::BasicBlock *ThenBB = CGF.createBasicBlock("lpc.then");
11907     llvm::BasicBlock *DoneBB = CGF.createBasicBlock("lpc.done");
11908     // if (Cmp) {
11909     CGF.Builder.CreateCondBr(Cmp, ThenBB, DoneBB);
11910     CGF.EmitBlock(ThenBB);
11911     Address Addr = CGF.GetAddrOfLocalVar(VD);
11912     LValue LVal;
11913     if (VD->getType()->isReferenceType())
11914       LVal = CGF.EmitLoadOfReferenceLValue(Addr, VD->getType(),
11915                                            AlignmentSource::Decl);
11916     else
11917       LVal = CGF.MakeAddrLValue(Addr, VD->getType().getNonReferenceType(),
11918                                 AlignmentSource::Decl);
11919     emitLastprivateConditionalUpdate(CGF, It->IVLVal, Pair.second, LVal,
11920                                      D.getBeginLoc());
11921     auto AL = ApplyDebugLocation::CreateArtificial(CGF);
11922     CGF.EmitBlock(DoneBB, /*IsFinal=*/true);
11923     // }
11924   }
11925 }
11926 
emitLastprivateConditionalFinalUpdate(CodeGenFunction & CGF,LValue PrivLVal,const VarDecl * VD,SourceLocation Loc)11927 void CGOpenMPRuntime::emitLastprivateConditionalFinalUpdate(
11928     CodeGenFunction &CGF, LValue PrivLVal, const VarDecl *VD,
11929     SourceLocation Loc) {
11930   if (CGF.getLangOpts().OpenMP < 50)
11931     return;
11932   auto It = LastprivateConditionalStack.back().DeclToUniqueName.find(VD);
11933   assert(It != LastprivateConditionalStack.back().DeclToUniqueName.end() &&
11934          "Unknown lastprivate conditional variable.");
11935   StringRef UniqueName = It->second;
11936   llvm::GlobalVariable *GV = CGM.getModule().getNamedGlobal(UniqueName);
11937   // The variable was not updated in the region - exit.
11938   if (!GV)
11939     return;
11940   LValue LPLVal = CGF.MakeRawAddrLValue(
11941       GV, PrivLVal.getType().getNonReferenceType(), PrivLVal.getAlignment());
11942   llvm::Value *Res = CGF.EmitLoadOfScalar(LPLVal, Loc);
11943   CGF.EmitStoreOfScalar(Res, PrivLVal);
11944 }
11945 
emitParallelOutlinedFunction(CodeGenFunction & CGF,const OMPExecutableDirective & D,const VarDecl * ThreadIDVar,OpenMPDirectiveKind InnermostKind,const RegionCodeGenTy & CodeGen)11946 llvm::Function *CGOpenMPSIMDRuntime::emitParallelOutlinedFunction(
11947     CodeGenFunction &CGF, const OMPExecutableDirective &D,
11948     const VarDecl *ThreadIDVar, OpenMPDirectiveKind InnermostKind,
11949     const RegionCodeGenTy &CodeGen) {
11950   llvm_unreachable("Not supported in SIMD-only mode");
11951 }
11952 
emitTeamsOutlinedFunction(CodeGenFunction & CGF,const OMPExecutableDirective & D,const VarDecl * ThreadIDVar,OpenMPDirectiveKind InnermostKind,const RegionCodeGenTy & CodeGen)11953 llvm::Function *CGOpenMPSIMDRuntime::emitTeamsOutlinedFunction(
11954     CodeGenFunction &CGF, const OMPExecutableDirective &D,
11955     const VarDecl *ThreadIDVar, OpenMPDirectiveKind InnermostKind,
11956     const RegionCodeGenTy &CodeGen) {
11957   llvm_unreachable("Not supported in SIMD-only mode");
11958 }
11959 
emitTaskOutlinedFunction(const OMPExecutableDirective & D,const VarDecl * ThreadIDVar,const VarDecl * PartIDVar,const VarDecl * TaskTVar,OpenMPDirectiveKind InnermostKind,const RegionCodeGenTy & CodeGen,bool Tied,unsigned & NumberOfParts)11960 llvm::Function *CGOpenMPSIMDRuntime::emitTaskOutlinedFunction(
11961     const OMPExecutableDirective &D, const VarDecl *ThreadIDVar,
11962     const VarDecl *PartIDVar, const VarDecl *TaskTVar,
11963     OpenMPDirectiveKind InnermostKind, const RegionCodeGenTy &CodeGen,
11964     bool Tied, unsigned &NumberOfParts) {
11965   llvm_unreachable("Not supported in SIMD-only mode");
11966 }
11967 
emitParallelCall(CodeGenFunction & CGF,SourceLocation Loc,llvm::Function * OutlinedFn,ArrayRef<llvm::Value * > CapturedVars,const Expr * IfCond,llvm::Value * NumThreads)11968 void CGOpenMPSIMDRuntime::emitParallelCall(CodeGenFunction &CGF,
11969                                            SourceLocation Loc,
11970                                            llvm::Function *OutlinedFn,
11971                                            ArrayRef<llvm::Value *> CapturedVars,
11972                                            const Expr *IfCond,
11973                                            llvm::Value *NumThreads) {
11974   llvm_unreachable("Not supported in SIMD-only mode");
11975 }
11976 
emitCriticalRegion(CodeGenFunction & CGF,StringRef CriticalName,const RegionCodeGenTy & CriticalOpGen,SourceLocation Loc,const Expr * Hint)11977 void CGOpenMPSIMDRuntime::emitCriticalRegion(
11978     CodeGenFunction &CGF, StringRef CriticalName,
11979     const RegionCodeGenTy &CriticalOpGen, SourceLocation Loc,
11980     const Expr *Hint) {
11981   llvm_unreachable("Not supported in SIMD-only mode");
11982 }
11983 
emitMasterRegion(CodeGenFunction & CGF,const RegionCodeGenTy & MasterOpGen,SourceLocation Loc)11984 void CGOpenMPSIMDRuntime::emitMasterRegion(CodeGenFunction &CGF,
11985                                            const RegionCodeGenTy &MasterOpGen,
11986                                            SourceLocation Loc) {
11987   llvm_unreachable("Not supported in SIMD-only mode");
11988 }
11989 
emitMaskedRegion(CodeGenFunction & CGF,const RegionCodeGenTy & MasterOpGen,SourceLocation Loc,const Expr * Filter)11990 void CGOpenMPSIMDRuntime::emitMaskedRegion(CodeGenFunction &CGF,
11991                                            const RegionCodeGenTy &MasterOpGen,
11992                                            SourceLocation Loc,
11993                                            const Expr *Filter) {
11994   llvm_unreachable("Not supported in SIMD-only mode");
11995 }
11996 
emitTaskyieldCall(CodeGenFunction & CGF,SourceLocation Loc)11997 void CGOpenMPSIMDRuntime::emitTaskyieldCall(CodeGenFunction &CGF,
11998                                             SourceLocation Loc) {
11999   llvm_unreachable("Not supported in SIMD-only mode");
12000 }
12001 
emitTaskgroupRegion(CodeGenFunction & CGF,const RegionCodeGenTy & TaskgroupOpGen,SourceLocation Loc)12002 void CGOpenMPSIMDRuntime::emitTaskgroupRegion(
12003     CodeGenFunction &CGF, const RegionCodeGenTy &TaskgroupOpGen,
12004     SourceLocation Loc) {
12005   llvm_unreachable("Not supported in SIMD-only mode");
12006 }
12007 
emitSingleRegion(CodeGenFunction & CGF,const RegionCodeGenTy & SingleOpGen,SourceLocation Loc,ArrayRef<const Expr * > CopyprivateVars,ArrayRef<const Expr * > DestExprs,ArrayRef<const Expr * > SrcExprs,ArrayRef<const Expr * > AssignmentOps)12008 void CGOpenMPSIMDRuntime::emitSingleRegion(
12009     CodeGenFunction &CGF, const RegionCodeGenTy &SingleOpGen,
12010     SourceLocation Loc, ArrayRef<const Expr *> CopyprivateVars,
12011     ArrayRef<const Expr *> DestExprs, ArrayRef<const Expr *> SrcExprs,
12012     ArrayRef<const Expr *> AssignmentOps) {
12013   llvm_unreachable("Not supported in SIMD-only mode");
12014 }
12015 
emitOrderedRegion(CodeGenFunction & CGF,const RegionCodeGenTy & OrderedOpGen,SourceLocation Loc,bool IsThreads)12016 void CGOpenMPSIMDRuntime::emitOrderedRegion(CodeGenFunction &CGF,
12017                                             const RegionCodeGenTy &OrderedOpGen,
12018                                             SourceLocation Loc,
12019                                             bool IsThreads) {
12020   llvm_unreachable("Not supported in SIMD-only mode");
12021 }
12022 
emitBarrierCall(CodeGenFunction & CGF,SourceLocation Loc,OpenMPDirectiveKind Kind,bool EmitChecks,bool ForceSimpleCall)12023 void CGOpenMPSIMDRuntime::emitBarrierCall(CodeGenFunction &CGF,
12024                                           SourceLocation Loc,
12025                                           OpenMPDirectiveKind Kind,
12026                                           bool EmitChecks,
12027                                           bool ForceSimpleCall) {
12028   llvm_unreachable("Not supported in SIMD-only mode");
12029 }
12030 
emitForDispatchInit(CodeGenFunction & CGF,SourceLocation Loc,const OpenMPScheduleTy & ScheduleKind,unsigned IVSize,bool IVSigned,bool Ordered,const DispatchRTInput & DispatchValues)12031 void CGOpenMPSIMDRuntime::emitForDispatchInit(
12032     CodeGenFunction &CGF, SourceLocation Loc,
12033     const OpenMPScheduleTy &ScheduleKind, unsigned IVSize, bool IVSigned,
12034     bool Ordered, const DispatchRTInput &DispatchValues) {
12035   llvm_unreachable("Not supported in SIMD-only mode");
12036 }
12037 
emitForDispatchDeinit(CodeGenFunction & CGF,SourceLocation Loc)12038 void CGOpenMPSIMDRuntime::emitForDispatchDeinit(CodeGenFunction &CGF,
12039                                                 SourceLocation Loc) {
12040   llvm_unreachable("Not supported in SIMD-only mode");
12041 }
12042 
emitForStaticInit(CodeGenFunction & CGF,SourceLocation Loc,OpenMPDirectiveKind DKind,const OpenMPScheduleTy & ScheduleKind,const StaticRTInput & Values)12043 void CGOpenMPSIMDRuntime::emitForStaticInit(
12044     CodeGenFunction &CGF, SourceLocation Loc, OpenMPDirectiveKind DKind,
12045     const OpenMPScheduleTy &ScheduleKind, const StaticRTInput &Values) {
12046   llvm_unreachable("Not supported in SIMD-only mode");
12047 }
12048 
emitDistributeStaticInit(CodeGenFunction & CGF,SourceLocation Loc,OpenMPDistScheduleClauseKind SchedKind,const StaticRTInput & Values)12049 void CGOpenMPSIMDRuntime::emitDistributeStaticInit(
12050     CodeGenFunction &CGF, SourceLocation Loc,
12051     OpenMPDistScheduleClauseKind SchedKind, const StaticRTInput &Values) {
12052   llvm_unreachable("Not supported in SIMD-only mode");
12053 }
12054 
emitForOrderedIterationEnd(CodeGenFunction & CGF,SourceLocation Loc,unsigned IVSize,bool IVSigned)12055 void CGOpenMPSIMDRuntime::emitForOrderedIterationEnd(CodeGenFunction &CGF,
12056                                                      SourceLocation Loc,
12057                                                      unsigned IVSize,
12058                                                      bool IVSigned) {
12059   llvm_unreachable("Not supported in SIMD-only mode");
12060 }
12061 
emitForStaticFinish(CodeGenFunction & CGF,SourceLocation Loc,OpenMPDirectiveKind DKind)12062 void CGOpenMPSIMDRuntime::emitForStaticFinish(CodeGenFunction &CGF,
12063                                               SourceLocation Loc,
12064                                               OpenMPDirectiveKind DKind) {
12065   llvm_unreachable("Not supported in SIMD-only mode");
12066 }
12067 
emitForNext(CodeGenFunction & CGF,SourceLocation Loc,unsigned IVSize,bool IVSigned,Address IL,Address LB,Address UB,Address ST)12068 llvm::Value *CGOpenMPSIMDRuntime::emitForNext(CodeGenFunction &CGF,
12069                                               SourceLocation Loc,
12070                                               unsigned IVSize, bool IVSigned,
12071                                               Address IL, Address LB,
12072                                               Address UB, Address ST) {
12073   llvm_unreachable("Not supported in SIMD-only mode");
12074 }
12075 
emitNumThreadsClause(CodeGenFunction & CGF,llvm::Value * NumThreads,SourceLocation Loc)12076 void CGOpenMPSIMDRuntime::emitNumThreadsClause(CodeGenFunction &CGF,
12077                                                llvm::Value *NumThreads,
12078                                                SourceLocation Loc) {
12079   llvm_unreachable("Not supported in SIMD-only mode");
12080 }
12081 
emitProcBindClause(CodeGenFunction & CGF,ProcBindKind ProcBind,SourceLocation Loc)12082 void CGOpenMPSIMDRuntime::emitProcBindClause(CodeGenFunction &CGF,
12083                                              ProcBindKind ProcBind,
12084                                              SourceLocation Loc) {
12085   llvm_unreachable("Not supported in SIMD-only mode");
12086 }
12087 
getAddrOfThreadPrivate(CodeGenFunction & CGF,const VarDecl * VD,Address VDAddr,SourceLocation Loc)12088 Address CGOpenMPSIMDRuntime::getAddrOfThreadPrivate(CodeGenFunction &CGF,
12089                                                     const VarDecl *VD,
12090                                                     Address VDAddr,
12091                                                     SourceLocation Loc) {
12092   llvm_unreachable("Not supported in SIMD-only mode");
12093 }
12094 
emitThreadPrivateVarDefinition(const VarDecl * VD,Address VDAddr,SourceLocation Loc,bool PerformInit,CodeGenFunction * CGF)12095 llvm::Function *CGOpenMPSIMDRuntime::emitThreadPrivateVarDefinition(
12096     const VarDecl *VD, Address VDAddr, SourceLocation Loc, bool PerformInit,
12097     CodeGenFunction *CGF) {
12098   llvm_unreachable("Not supported in SIMD-only mode");
12099 }
12100 
getAddrOfArtificialThreadPrivate(CodeGenFunction & CGF,QualType VarType,StringRef Name)12101 Address CGOpenMPSIMDRuntime::getAddrOfArtificialThreadPrivate(
12102     CodeGenFunction &CGF, QualType VarType, StringRef Name) {
12103   llvm_unreachable("Not supported in SIMD-only mode");
12104 }
12105 
emitFlush(CodeGenFunction & CGF,ArrayRef<const Expr * > Vars,SourceLocation Loc,llvm::AtomicOrdering AO)12106 void CGOpenMPSIMDRuntime::emitFlush(CodeGenFunction &CGF,
12107                                     ArrayRef<const Expr *> Vars,
12108                                     SourceLocation Loc,
12109                                     llvm::AtomicOrdering AO) {
12110   llvm_unreachable("Not supported in SIMD-only mode");
12111 }
12112 
emitTaskCall(CodeGenFunction & CGF,SourceLocation Loc,const OMPExecutableDirective & D,llvm::Function * TaskFunction,QualType SharedsTy,Address Shareds,const Expr * IfCond,const OMPTaskDataTy & Data)12113 void CGOpenMPSIMDRuntime::emitTaskCall(CodeGenFunction &CGF, SourceLocation Loc,
12114                                        const OMPExecutableDirective &D,
12115                                        llvm::Function *TaskFunction,
12116                                        QualType SharedsTy, Address Shareds,
12117                                        const Expr *IfCond,
12118                                        const OMPTaskDataTy &Data) {
12119   llvm_unreachable("Not supported in SIMD-only mode");
12120 }
12121 
emitTaskLoopCall(CodeGenFunction & CGF,SourceLocation Loc,const OMPLoopDirective & D,llvm::Function * TaskFunction,QualType SharedsTy,Address Shareds,const Expr * IfCond,const OMPTaskDataTy & Data)12122 void CGOpenMPSIMDRuntime::emitTaskLoopCall(
12123     CodeGenFunction &CGF, SourceLocation Loc, const OMPLoopDirective &D,
12124     llvm::Function *TaskFunction, QualType SharedsTy, Address Shareds,
12125     const Expr *IfCond, const OMPTaskDataTy &Data) {
12126   llvm_unreachable("Not supported in SIMD-only mode");
12127 }
12128 
emitReduction(CodeGenFunction & CGF,SourceLocation Loc,ArrayRef<const Expr * > Privates,ArrayRef<const Expr * > LHSExprs,ArrayRef<const Expr * > RHSExprs,ArrayRef<const Expr * > ReductionOps,ReductionOptionsTy Options)12129 void CGOpenMPSIMDRuntime::emitReduction(
12130     CodeGenFunction &CGF, SourceLocation Loc, ArrayRef<const Expr *> Privates,
12131     ArrayRef<const Expr *> LHSExprs, ArrayRef<const Expr *> RHSExprs,
12132     ArrayRef<const Expr *> ReductionOps, ReductionOptionsTy Options) {
12133   assert(Options.SimpleReduction && "Only simple reduction is expected.");
12134   CGOpenMPRuntime::emitReduction(CGF, Loc, Privates, LHSExprs, RHSExprs,
12135                                  ReductionOps, Options);
12136 }
12137 
emitTaskReductionInit(CodeGenFunction & CGF,SourceLocation Loc,ArrayRef<const Expr * > LHSExprs,ArrayRef<const Expr * > RHSExprs,const OMPTaskDataTy & Data)12138 llvm::Value *CGOpenMPSIMDRuntime::emitTaskReductionInit(
12139     CodeGenFunction &CGF, SourceLocation Loc, ArrayRef<const Expr *> LHSExprs,
12140     ArrayRef<const Expr *> RHSExprs, const OMPTaskDataTy &Data) {
12141   llvm_unreachable("Not supported in SIMD-only mode");
12142 }
12143 
emitTaskReductionFini(CodeGenFunction & CGF,SourceLocation Loc,bool IsWorksharingReduction)12144 void CGOpenMPSIMDRuntime::emitTaskReductionFini(CodeGenFunction &CGF,
12145                                                 SourceLocation Loc,
12146                                                 bool IsWorksharingReduction) {
12147   llvm_unreachable("Not supported in SIMD-only mode");
12148 }
12149 
emitTaskReductionFixups(CodeGenFunction & CGF,SourceLocation Loc,ReductionCodeGen & RCG,unsigned N)12150 void CGOpenMPSIMDRuntime::emitTaskReductionFixups(CodeGenFunction &CGF,
12151                                                   SourceLocation Loc,
12152                                                   ReductionCodeGen &RCG,
12153                                                   unsigned N) {
12154   llvm_unreachable("Not supported in SIMD-only mode");
12155 }
12156 
getTaskReductionItem(CodeGenFunction & CGF,SourceLocation Loc,llvm::Value * ReductionsPtr,LValue SharedLVal)12157 Address CGOpenMPSIMDRuntime::getTaskReductionItem(CodeGenFunction &CGF,
12158                                                   SourceLocation Loc,
12159                                                   llvm::Value *ReductionsPtr,
12160                                                   LValue SharedLVal) {
12161   llvm_unreachable("Not supported in SIMD-only mode");
12162 }
12163 
emitTaskwaitCall(CodeGenFunction & CGF,SourceLocation Loc,const OMPTaskDataTy & Data)12164 void CGOpenMPSIMDRuntime::emitTaskwaitCall(CodeGenFunction &CGF,
12165                                            SourceLocation Loc,
12166                                            const OMPTaskDataTy &Data) {
12167   llvm_unreachable("Not supported in SIMD-only mode");
12168 }
12169 
emitCancellationPointCall(CodeGenFunction & CGF,SourceLocation Loc,OpenMPDirectiveKind CancelRegion)12170 void CGOpenMPSIMDRuntime::emitCancellationPointCall(
12171     CodeGenFunction &CGF, SourceLocation Loc,
12172     OpenMPDirectiveKind CancelRegion) {
12173   llvm_unreachable("Not supported in SIMD-only mode");
12174 }
12175 
emitCancelCall(CodeGenFunction & CGF,SourceLocation Loc,const Expr * IfCond,OpenMPDirectiveKind CancelRegion)12176 void CGOpenMPSIMDRuntime::emitCancelCall(CodeGenFunction &CGF,
12177                                          SourceLocation Loc, const Expr *IfCond,
12178                                          OpenMPDirectiveKind CancelRegion) {
12179   llvm_unreachable("Not supported in SIMD-only mode");
12180 }
12181 
emitTargetOutlinedFunction(const OMPExecutableDirective & D,StringRef ParentName,llvm::Function * & OutlinedFn,llvm::Constant * & OutlinedFnID,bool IsOffloadEntry,const RegionCodeGenTy & CodeGen)12182 void CGOpenMPSIMDRuntime::emitTargetOutlinedFunction(
12183     const OMPExecutableDirective &D, StringRef ParentName,
12184     llvm::Function *&OutlinedFn, llvm::Constant *&OutlinedFnID,
12185     bool IsOffloadEntry, const RegionCodeGenTy &CodeGen) {
12186   llvm_unreachable("Not supported in SIMD-only mode");
12187 }
12188 
emitTargetCall(CodeGenFunction & CGF,const OMPExecutableDirective & D,llvm::Function * OutlinedFn,llvm::Value * OutlinedFnID,const Expr * IfCond,llvm::PointerIntPair<const Expr *,2,OpenMPDeviceClauseModifier> Device,llvm::function_ref<llvm::Value * (CodeGenFunction & CGF,const OMPLoopDirective & D)> SizeEmitter)12189 void CGOpenMPSIMDRuntime::emitTargetCall(
12190     CodeGenFunction &CGF, const OMPExecutableDirective &D,
12191     llvm::Function *OutlinedFn, llvm::Value *OutlinedFnID, const Expr *IfCond,
12192     llvm::PointerIntPair<const Expr *, 2, OpenMPDeviceClauseModifier> Device,
12193     llvm::function_ref<llvm::Value *(CodeGenFunction &CGF,
12194                                      const OMPLoopDirective &D)>
12195         SizeEmitter) {
12196   llvm_unreachable("Not supported in SIMD-only mode");
12197 }
12198 
emitTargetFunctions(GlobalDecl GD)12199 bool CGOpenMPSIMDRuntime::emitTargetFunctions(GlobalDecl GD) {
12200   llvm_unreachable("Not supported in SIMD-only mode");
12201 }
12202 
emitTargetGlobalVariable(GlobalDecl GD)12203 bool CGOpenMPSIMDRuntime::emitTargetGlobalVariable(GlobalDecl GD) {
12204   llvm_unreachable("Not supported in SIMD-only mode");
12205 }
12206 
emitTargetGlobal(GlobalDecl GD)12207 bool CGOpenMPSIMDRuntime::emitTargetGlobal(GlobalDecl GD) {
12208   return false;
12209 }
12210 
emitTeamsCall(CodeGenFunction & CGF,const OMPExecutableDirective & D,SourceLocation Loc,llvm::Function * OutlinedFn,ArrayRef<llvm::Value * > CapturedVars)12211 void CGOpenMPSIMDRuntime::emitTeamsCall(CodeGenFunction &CGF,
12212                                         const OMPExecutableDirective &D,
12213                                         SourceLocation Loc,
12214                                         llvm::Function *OutlinedFn,
12215                                         ArrayRef<llvm::Value *> CapturedVars) {
12216   llvm_unreachable("Not supported in SIMD-only mode");
12217 }
12218 
emitNumTeamsClause(CodeGenFunction & CGF,const Expr * NumTeams,const Expr * ThreadLimit,SourceLocation Loc)12219 void CGOpenMPSIMDRuntime::emitNumTeamsClause(CodeGenFunction &CGF,
12220                                              const Expr *NumTeams,
12221                                              const Expr *ThreadLimit,
12222                                              SourceLocation Loc) {
12223   llvm_unreachable("Not supported in SIMD-only mode");
12224 }
12225 
emitTargetDataCalls(CodeGenFunction & CGF,const OMPExecutableDirective & D,const Expr * IfCond,const Expr * Device,const RegionCodeGenTy & CodeGen,CGOpenMPRuntime::TargetDataInfo & Info)12226 void CGOpenMPSIMDRuntime::emitTargetDataCalls(
12227     CodeGenFunction &CGF, const OMPExecutableDirective &D, const Expr *IfCond,
12228     const Expr *Device, const RegionCodeGenTy &CodeGen,
12229     CGOpenMPRuntime::TargetDataInfo &Info) {
12230   llvm_unreachable("Not supported in SIMD-only mode");
12231 }
12232 
emitTargetDataStandAloneCall(CodeGenFunction & CGF,const OMPExecutableDirective & D,const Expr * IfCond,const Expr * Device)12233 void CGOpenMPSIMDRuntime::emitTargetDataStandAloneCall(
12234     CodeGenFunction &CGF, const OMPExecutableDirective &D, const Expr *IfCond,
12235     const Expr *Device) {
12236   llvm_unreachable("Not supported in SIMD-only mode");
12237 }
12238 
emitDoacrossInit(CodeGenFunction & CGF,const OMPLoopDirective & D,ArrayRef<Expr * > NumIterations)12239 void CGOpenMPSIMDRuntime::emitDoacrossInit(CodeGenFunction &CGF,
12240                                            const OMPLoopDirective &D,
12241                                            ArrayRef<Expr *> NumIterations) {
12242   llvm_unreachable("Not supported in SIMD-only mode");
12243 }
12244 
emitDoacrossOrdered(CodeGenFunction & CGF,const OMPDependClause * C)12245 void CGOpenMPSIMDRuntime::emitDoacrossOrdered(CodeGenFunction &CGF,
12246                                               const OMPDependClause *C) {
12247   llvm_unreachable("Not supported in SIMD-only mode");
12248 }
12249 
emitDoacrossOrdered(CodeGenFunction & CGF,const OMPDoacrossClause * C)12250 void CGOpenMPSIMDRuntime::emitDoacrossOrdered(CodeGenFunction &CGF,
12251                                               const OMPDoacrossClause *C) {
12252   llvm_unreachable("Not supported in SIMD-only mode");
12253 }
12254 
12255 const VarDecl *
translateParameter(const FieldDecl * FD,const VarDecl * NativeParam) const12256 CGOpenMPSIMDRuntime::translateParameter(const FieldDecl *FD,
12257                                         const VarDecl *NativeParam) const {
12258   llvm_unreachable("Not supported in SIMD-only mode");
12259 }
12260 
12261 Address
getParameterAddress(CodeGenFunction & CGF,const VarDecl * NativeParam,const VarDecl * TargetParam) const12262 CGOpenMPSIMDRuntime::getParameterAddress(CodeGenFunction &CGF,
12263                                          const VarDecl *NativeParam,
12264                                          const VarDecl *TargetParam) const {
12265   llvm_unreachable("Not supported in SIMD-only mode");
12266 }
12267