1 //===----- CGCoroutine.cpp - Emit LLVM Code for C++ coroutines ------------===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 //
9 // This contains code dealing with C++ code generation of coroutines.
10 //
11 //===----------------------------------------------------------------------===//
12
13 #include "CGCleanup.h"
14 #include "CGDebugInfo.h"
15 #include "CodeGenFunction.h"
16 #include "clang/AST/StmtCXX.h"
17 #include "clang/AST/StmtVisitor.h"
18 #include "llvm/ADT/ScopeExit.h"
19
20 using namespace clang;
21 using namespace CodeGen;
22
23 using llvm::Value;
24 using llvm::BasicBlock;
25
26 namespace {
27 enum class AwaitKind { Init, Normal, Yield, Final };
28 static constexpr llvm::StringLiteral AwaitKindStr[] = {"init", "await", "yield",
29 "final"};
30 }
31
32 struct clang::CodeGen::CGCoroData {
33 // What is the current await expression kind and how many
34 // await/yield expressions were encountered so far.
35 // These are used to generate pretty labels for await expressions in LLVM IR.
36 AwaitKind CurrentAwaitKind = AwaitKind::Init;
37 unsigned AwaitNum = 0;
38 unsigned YieldNum = 0;
39
40 // How many co_return statements are in the coroutine. Used to decide whether
41 // we need to add co_return; equivalent at the end of the user authored body.
42 unsigned CoreturnCount = 0;
43
44 // A branch to this block is emitted when coroutine needs to suspend.
45 llvm::BasicBlock *SuspendBB = nullptr;
46
47 // The promise type's 'unhandled_exception' handler, if it defines one.
48 Stmt *ExceptionHandler = nullptr;
49
50 // A temporary i1 alloca that stores whether 'await_resume' threw an
51 // exception. If it did, 'true' is stored in this variable, and the coroutine
52 // body must be skipped. If the promise type does not define an exception
53 // handler, this is null.
54 llvm::Value *ResumeEHVar = nullptr;
55
56 // Stores the jump destination just before the coroutine memory is freed.
57 // This is the destination that every suspend point jumps to for the cleanup
58 // branch.
59 CodeGenFunction::JumpDest CleanupJD;
60
61 // Stores the jump destination just before the final suspend. The co_return
62 // statements jumps to this point after calling return_xxx promise member.
63 CodeGenFunction::JumpDest FinalJD;
64
65 // Stores the llvm.coro.id emitted in the function so that we can supply it
66 // as the first argument to coro.begin, coro.alloc and coro.free intrinsics.
67 // Note: llvm.coro.id returns a token that cannot be directly expressed in a
68 // builtin.
69 llvm::CallInst *CoroId = nullptr;
70
71 // Stores the llvm.coro.begin emitted in the function so that we can replace
72 // all coro.frame intrinsics with direct SSA value of coro.begin that returns
73 // the address of the coroutine frame of the current coroutine.
74 llvm::CallInst *CoroBegin = nullptr;
75
76 // Stores the last emitted coro.free for the deallocate expressions, we use it
77 // to wrap dealloc code with if(auto mem = coro.free) dealloc(mem).
78 llvm::CallInst *LastCoroFree = nullptr;
79
80 // If coro.id came from the builtin, remember the expression to give better
81 // diagnostic. If CoroIdExpr is nullptr, the coro.id was created by
82 // EmitCoroutineBody.
83 CallExpr const *CoroIdExpr = nullptr;
84 };
85
86 // Defining these here allows to keep CGCoroData private to this file.
CGCoroInfo()87 clang::CodeGen::CodeGenFunction::CGCoroInfo::CGCoroInfo() {}
~CGCoroInfo()88 CodeGenFunction::CGCoroInfo::~CGCoroInfo() {}
89
createCoroData(CodeGenFunction & CGF,CodeGenFunction::CGCoroInfo & CurCoro,llvm::CallInst * CoroId,CallExpr const * CoroIdExpr=nullptr)90 static void createCoroData(CodeGenFunction &CGF,
91 CodeGenFunction::CGCoroInfo &CurCoro,
92 llvm::CallInst *CoroId,
93 CallExpr const *CoroIdExpr = nullptr) {
94 if (CurCoro.Data) {
95 if (CurCoro.Data->CoroIdExpr)
96 CGF.CGM.Error(CoroIdExpr->getBeginLoc(),
97 "only one __builtin_coro_id can be used in a function");
98 else if (CoroIdExpr)
99 CGF.CGM.Error(CoroIdExpr->getBeginLoc(),
100 "__builtin_coro_id shall not be used in a C++ coroutine");
101 else
102 llvm_unreachable("EmitCoroutineBodyStatement called twice?");
103
104 return;
105 }
106
107 CurCoro.Data = std::make_unique<CGCoroData>();
108 CurCoro.Data->CoroId = CoroId;
109 CurCoro.Data->CoroIdExpr = CoroIdExpr;
110 }
111
112 // Synthesize a pretty name for a suspend point.
buildSuspendPrefixStr(CGCoroData & Coro,AwaitKind Kind)113 static SmallString<32> buildSuspendPrefixStr(CGCoroData &Coro, AwaitKind Kind) {
114 unsigned No = 0;
115 switch (Kind) {
116 case AwaitKind::Init:
117 case AwaitKind::Final:
118 break;
119 case AwaitKind::Normal:
120 No = ++Coro.AwaitNum;
121 break;
122 case AwaitKind::Yield:
123 No = ++Coro.YieldNum;
124 break;
125 }
126 SmallString<32> Prefix(AwaitKindStr[static_cast<unsigned>(Kind)]);
127 if (No > 1) {
128 Twine(No).toVector(Prefix);
129 }
130 return Prefix;
131 }
132
133 // Check if function can throw based on prototype noexcept, also works for
134 // destructors which are implicitly noexcept but can be marked noexcept(false).
FunctionCanThrow(const FunctionDecl * D)135 static bool FunctionCanThrow(const FunctionDecl *D) {
136 const auto *Proto = D->getType()->getAs<FunctionProtoType>();
137 if (!Proto) {
138 // Function proto is not found, we conservatively assume throwing.
139 return true;
140 }
141 return !isNoexceptExceptionSpec(Proto->getExceptionSpecType()) ||
142 Proto->canThrow() != CT_Cannot;
143 }
144
StmtCanThrow(const Stmt * S)145 static bool StmtCanThrow(const Stmt *S) {
146 if (const auto *CE = dyn_cast<CallExpr>(S)) {
147 const auto *Callee = CE->getDirectCallee();
148 if (!Callee)
149 // We don't have direct callee. Conservatively assume throwing.
150 return true;
151
152 if (FunctionCanThrow(Callee))
153 return true;
154
155 // Fall through to visit the children.
156 }
157
158 if (const auto *TE = dyn_cast<CXXBindTemporaryExpr>(S)) {
159 // Special handling of CXXBindTemporaryExpr here as calling of Dtor of the
160 // temporary is not part of `children()` as covered in the fall through.
161 // We need to mark entire statement as throwing if the destructor of the
162 // temporary throws.
163 const auto *Dtor = TE->getTemporary()->getDestructor();
164 if (FunctionCanThrow(Dtor))
165 return true;
166
167 // Fall through to visit the children.
168 }
169
170 for (const auto *child : S->children())
171 if (StmtCanThrow(child))
172 return true;
173
174 return false;
175 }
176
177 // Emit suspend expression which roughly looks like:
178 //
179 // auto && x = CommonExpr();
180 // if (!x.await_ready()) {
181 // llvm_coro_save();
182 // llvm_coro_await_suspend(&x, frame, wrapper) (*) (**)
183 // llvm_coro_suspend(); (***)
184 // }
185 // x.await_resume();
186 //
187 // where the result of the entire expression is the result of x.await_resume()
188 //
189 // (*) llvm_coro_await_suspend_{void, bool, handle} is lowered to
190 // wrapper(&x, frame) when it's certain not to interfere with
191 // coroutine transform. await_suspend expression is
192 // asynchronous to the coroutine body and not all analyses
193 // and transformations can handle it correctly at the moment.
194 //
195 // Wrapper function encapsulates x.await_suspend(...) call and looks like:
196 //
197 // auto __await_suspend_wrapper(auto& awaiter, void* frame) {
198 // std::coroutine_handle<> handle(frame);
199 // return awaiter.await_suspend(handle);
200 // }
201 //
202 // (**) If x.await_suspend return type is bool, it allows to veto a suspend:
203 // if (x.await_suspend(...))
204 // llvm_coro_suspend();
205 //
206 // (***) llvm_coro_suspend() encodes three possible continuations as
207 // a switch instruction:
208 //
209 // %where-to = call i8 @llvm.coro.suspend(...)
210 // switch i8 %where-to, label %coro.ret [ ; jump to epilogue to suspend
211 // i8 0, label %yield.ready ; go here when resumed
212 // i8 1, label %yield.cleanup ; go here when destroyed
213 // ]
214 //
215 // See llvm's docs/Coroutines.rst for more details.
216 //
217 namespace {
218 struct LValueOrRValue {
219 LValue LV;
220 RValue RV;
221 };
222 }
emitSuspendExpression(CodeGenFunction & CGF,CGCoroData & Coro,CoroutineSuspendExpr const & S,AwaitKind Kind,AggValueSlot aggSlot,bool ignoreResult,bool forLValue)223 static LValueOrRValue emitSuspendExpression(CodeGenFunction &CGF, CGCoroData &Coro,
224 CoroutineSuspendExpr const &S,
225 AwaitKind Kind, AggValueSlot aggSlot,
226 bool ignoreResult, bool forLValue) {
227 auto *E = S.getCommonExpr();
228
229 auto CommonBinder =
230 CodeGenFunction::OpaqueValueMappingData::bind(CGF, S.getOpaqueValue(), E);
231 auto UnbindCommonOnExit =
232 llvm::make_scope_exit([&] { CommonBinder.unbind(CGF); });
233
234 auto Prefix = buildSuspendPrefixStr(Coro, Kind);
235 BasicBlock *ReadyBlock = CGF.createBasicBlock(Prefix + Twine(".ready"));
236 BasicBlock *SuspendBlock = CGF.createBasicBlock(Prefix + Twine(".suspend"));
237 BasicBlock *CleanupBlock = CGF.createBasicBlock(Prefix + Twine(".cleanup"));
238
239 // If expression is ready, no need to suspend.
240 CGF.EmitBranchOnBoolExpr(S.getReadyExpr(), ReadyBlock, SuspendBlock, 0);
241
242 // Otherwise, emit suspend logic.
243 CGF.EmitBlock(SuspendBlock);
244
245 auto &Builder = CGF.Builder;
246 llvm::Function *CoroSave = CGF.CGM.getIntrinsic(llvm::Intrinsic::coro_save);
247 auto *NullPtr = llvm::ConstantPointerNull::get(CGF.CGM.Int8PtrTy);
248 auto *SaveCall = Builder.CreateCall(CoroSave, {NullPtr});
249
250 auto SuspendWrapper = CodeGenFunction(CGF.CGM).generateAwaitSuspendWrapper(
251 CGF.CurFn->getName(), Prefix, S);
252
253 CGF.CurCoro.InSuspendBlock = true;
254
255 assert(CGF.CurCoro.Data && CGF.CurCoro.Data->CoroBegin &&
256 "expected to be called in coroutine context");
257
258 SmallVector<llvm::Value *, 3> SuspendIntrinsicCallArgs;
259 SuspendIntrinsicCallArgs.push_back(
260 CGF.getOrCreateOpaqueLValueMapping(S.getOpaqueValue()).getPointer(CGF));
261
262 SuspendIntrinsicCallArgs.push_back(CGF.CurCoro.Data->CoroBegin);
263 SuspendIntrinsicCallArgs.push_back(SuspendWrapper);
264
265 const auto SuspendReturnType = S.getSuspendReturnType();
266 llvm::Intrinsic::ID AwaitSuspendIID;
267
268 switch (SuspendReturnType) {
269 case CoroutineSuspendExpr::SuspendReturnType::SuspendVoid:
270 AwaitSuspendIID = llvm::Intrinsic::coro_await_suspend_void;
271 break;
272 case CoroutineSuspendExpr::SuspendReturnType::SuspendBool:
273 AwaitSuspendIID = llvm::Intrinsic::coro_await_suspend_bool;
274 break;
275 case CoroutineSuspendExpr::SuspendReturnType::SuspendHandle:
276 AwaitSuspendIID = llvm::Intrinsic::coro_await_suspend_handle;
277 break;
278 }
279
280 llvm::Function *AwaitSuspendIntrinsic = CGF.CGM.getIntrinsic(AwaitSuspendIID);
281
282 // SuspendHandle might throw since it also resumes the returned handle.
283 const bool AwaitSuspendCanThrow =
284 SuspendReturnType ==
285 CoroutineSuspendExpr::SuspendReturnType::SuspendHandle ||
286 StmtCanThrow(S.getSuspendExpr());
287
288 llvm::CallBase *SuspendRet = nullptr;
289 // FIXME: add call attributes?
290 if (AwaitSuspendCanThrow)
291 SuspendRet =
292 CGF.EmitCallOrInvoke(AwaitSuspendIntrinsic, SuspendIntrinsicCallArgs);
293 else
294 SuspendRet = CGF.EmitNounwindRuntimeCall(AwaitSuspendIntrinsic,
295 SuspendIntrinsicCallArgs);
296
297 assert(SuspendRet);
298 CGF.CurCoro.InSuspendBlock = false;
299
300 switch (SuspendReturnType) {
301 case CoroutineSuspendExpr::SuspendReturnType::SuspendVoid:
302 assert(SuspendRet->getType()->isVoidTy());
303 break;
304 case CoroutineSuspendExpr::SuspendReturnType::SuspendBool: {
305 assert(SuspendRet->getType()->isIntegerTy());
306
307 // Veto suspension if requested by bool returning await_suspend.
308 BasicBlock *RealSuspendBlock =
309 CGF.createBasicBlock(Prefix + Twine(".suspend.bool"));
310 CGF.Builder.CreateCondBr(SuspendRet, RealSuspendBlock, ReadyBlock);
311 CGF.EmitBlock(RealSuspendBlock);
312 break;
313 }
314 case CoroutineSuspendExpr::SuspendReturnType::SuspendHandle: {
315 assert(SuspendRet->getType()->isVoidTy());
316 break;
317 }
318 }
319
320 // Emit the suspend point.
321 const bool IsFinalSuspend = (Kind == AwaitKind::Final);
322 llvm::Function *CoroSuspend =
323 CGF.CGM.getIntrinsic(llvm::Intrinsic::coro_suspend);
324 auto *SuspendResult = Builder.CreateCall(
325 CoroSuspend, {SaveCall, Builder.getInt1(IsFinalSuspend)});
326
327 // Create a switch capturing three possible continuations.
328 auto *Switch = Builder.CreateSwitch(SuspendResult, Coro.SuspendBB, 2);
329 Switch->addCase(Builder.getInt8(0), ReadyBlock);
330 Switch->addCase(Builder.getInt8(1), CleanupBlock);
331
332 // Emit cleanup for this suspend point.
333 CGF.EmitBlock(CleanupBlock);
334 CGF.EmitBranchThroughCleanup(Coro.CleanupJD);
335
336 // Emit await_resume expression.
337 CGF.EmitBlock(ReadyBlock);
338
339 // Exception handling requires additional IR. If the 'await_resume' function
340 // is marked as 'noexcept', we avoid generating this additional IR.
341 CXXTryStmt *TryStmt = nullptr;
342 if (Coro.ExceptionHandler && Kind == AwaitKind::Init &&
343 StmtCanThrow(S.getResumeExpr())) {
344 Coro.ResumeEHVar =
345 CGF.CreateTempAlloca(Builder.getInt1Ty(), Prefix + Twine("resume.eh"));
346 Builder.CreateFlagStore(true, Coro.ResumeEHVar);
347
348 auto Loc = S.getResumeExpr()->getExprLoc();
349 auto *Catch = new (CGF.getContext())
350 CXXCatchStmt(Loc, /*exDecl=*/nullptr, Coro.ExceptionHandler);
351 auto *TryBody = CompoundStmt::Create(CGF.getContext(), S.getResumeExpr(),
352 FPOptionsOverride(), Loc, Loc);
353 TryStmt = CXXTryStmt::Create(CGF.getContext(), Loc, TryBody, Catch);
354 CGF.EnterCXXTryStmt(*TryStmt);
355 CGF.EmitStmt(TryBody);
356 // We don't use EmitCXXTryStmt here. We need to store to ResumeEHVar that
357 // doesn't exist in the body.
358 Builder.CreateFlagStore(false, Coro.ResumeEHVar);
359 CGF.ExitCXXTryStmt(*TryStmt);
360 LValueOrRValue Res;
361 // We are not supposed to obtain the value from init suspend await_resume().
362 Res.RV = RValue::getIgnored();
363 return Res;
364 }
365
366 LValueOrRValue Res;
367 if (forLValue)
368 Res.LV = CGF.EmitLValue(S.getResumeExpr());
369 else
370 Res.RV = CGF.EmitAnyExpr(S.getResumeExpr(), aggSlot, ignoreResult);
371
372 return Res;
373 }
374
EmitCoawaitExpr(const CoawaitExpr & E,AggValueSlot aggSlot,bool ignoreResult)375 RValue CodeGenFunction::EmitCoawaitExpr(const CoawaitExpr &E,
376 AggValueSlot aggSlot,
377 bool ignoreResult) {
378 return emitSuspendExpression(*this, *CurCoro.Data, E,
379 CurCoro.Data->CurrentAwaitKind, aggSlot,
380 ignoreResult, /*forLValue*/false).RV;
381 }
EmitCoyieldExpr(const CoyieldExpr & E,AggValueSlot aggSlot,bool ignoreResult)382 RValue CodeGenFunction::EmitCoyieldExpr(const CoyieldExpr &E,
383 AggValueSlot aggSlot,
384 bool ignoreResult) {
385 return emitSuspendExpression(*this, *CurCoro.Data, E, AwaitKind::Yield,
386 aggSlot, ignoreResult, /*forLValue*/false).RV;
387 }
388
EmitCoreturnStmt(CoreturnStmt const & S)389 void CodeGenFunction::EmitCoreturnStmt(CoreturnStmt const &S) {
390 ++CurCoro.Data->CoreturnCount;
391 const Expr *RV = S.getOperand();
392 if (RV && RV->getType()->isVoidType() && !isa<InitListExpr>(RV)) {
393 // Make sure to evaluate the non initlist expression of a co_return
394 // with a void expression for side effects.
395 RunCleanupsScope cleanupScope(*this);
396 EmitIgnoredExpr(RV);
397 }
398 EmitStmt(S.getPromiseCall());
399 EmitBranchThroughCleanup(CurCoro.Data->FinalJD);
400 }
401
402
403 #ifndef NDEBUG
getCoroutineSuspendExprReturnType(const ASTContext & Ctx,const CoroutineSuspendExpr * E)404 static QualType getCoroutineSuspendExprReturnType(const ASTContext &Ctx,
405 const CoroutineSuspendExpr *E) {
406 const auto *RE = E->getResumeExpr();
407 // Is it possible for RE to be a CXXBindTemporaryExpr wrapping
408 // a MemberCallExpr?
409 assert(isa<CallExpr>(RE) && "unexpected suspend expression type");
410 return cast<CallExpr>(RE)->getCallReturnType(Ctx);
411 }
412 #endif
413
414 llvm::Function *
generateAwaitSuspendWrapper(Twine const & CoroName,Twine const & SuspendPointName,CoroutineSuspendExpr const & S)415 CodeGenFunction::generateAwaitSuspendWrapper(Twine const &CoroName,
416 Twine const &SuspendPointName,
417 CoroutineSuspendExpr const &S) {
418 std::string FuncName =
419 (CoroName + ".__await_suspend_wrapper__" + SuspendPointName).str();
420
421 ASTContext &C = getContext();
422
423 FunctionArgList args;
424
425 ImplicitParamDecl AwaiterDecl(C, C.VoidPtrTy, ImplicitParamKind::Other);
426 ImplicitParamDecl FrameDecl(C, C.VoidPtrTy, ImplicitParamKind::Other);
427 QualType ReturnTy = S.getSuspendExpr()->getType();
428
429 args.push_back(&AwaiterDecl);
430 args.push_back(&FrameDecl);
431
432 const CGFunctionInfo &FI =
433 CGM.getTypes().arrangeBuiltinFunctionDeclaration(ReturnTy, args);
434
435 llvm::FunctionType *LTy = CGM.getTypes().GetFunctionType(FI);
436
437 llvm::Function *Fn = llvm::Function::Create(
438 LTy, llvm::GlobalValue::PrivateLinkage, FuncName, &CGM.getModule());
439
440 Fn->addParamAttr(0, llvm::Attribute::AttrKind::NonNull);
441 Fn->addParamAttr(0, llvm::Attribute::AttrKind::NoUndef);
442
443 Fn->addParamAttr(1, llvm::Attribute::AttrKind::NoUndef);
444
445 Fn->setMustProgress();
446 Fn->addFnAttr(llvm::Attribute::AttrKind::AlwaysInline);
447
448 StartFunction(GlobalDecl(), ReturnTy, Fn, FI, args);
449
450 // FIXME: add TBAA metadata to the loads
451 llvm::Value *AwaiterPtr = Builder.CreateLoad(GetAddrOfLocalVar(&AwaiterDecl));
452 auto AwaiterLValue =
453 MakeNaturalAlignAddrLValue(AwaiterPtr, AwaiterDecl.getType());
454
455 CurAwaitSuspendWrapper.FramePtr =
456 Builder.CreateLoad(GetAddrOfLocalVar(&FrameDecl));
457
458 auto AwaiterBinder = CodeGenFunction::OpaqueValueMappingData::bind(
459 *this, S.getOpaqueValue(), AwaiterLValue);
460
461 auto *SuspendRet = EmitScalarExpr(S.getSuspendExpr());
462
463 auto UnbindCommonOnExit =
464 llvm::make_scope_exit([&] { AwaiterBinder.unbind(*this); });
465 if (SuspendRet != nullptr) {
466 Fn->addRetAttr(llvm::Attribute::AttrKind::NoUndef);
467 Builder.CreateStore(SuspendRet, ReturnValue);
468 }
469
470 CurAwaitSuspendWrapper.FramePtr = nullptr;
471 FinishFunction();
472 return Fn;
473 }
474
475 LValue
EmitCoawaitLValue(const CoawaitExpr * E)476 CodeGenFunction::EmitCoawaitLValue(const CoawaitExpr *E) {
477 assert(getCoroutineSuspendExprReturnType(getContext(), E)->isReferenceType() &&
478 "Can't have a scalar return unless the return type is a "
479 "reference type!");
480 return emitSuspendExpression(*this, *CurCoro.Data, *E,
481 CurCoro.Data->CurrentAwaitKind, AggValueSlot::ignored(),
482 /*ignoreResult*/false, /*forLValue*/true).LV;
483 }
484
485 LValue
EmitCoyieldLValue(const CoyieldExpr * E)486 CodeGenFunction::EmitCoyieldLValue(const CoyieldExpr *E) {
487 assert(getCoroutineSuspendExprReturnType(getContext(), E)->isReferenceType() &&
488 "Can't have a scalar return unless the return type is a "
489 "reference type!");
490 return emitSuspendExpression(*this, *CurCoro.Data, *E,
491 AwaitKind::Yield, AggValueSlot::ignored(),
492 /*ignoreResult*/false, /*forLValue*/true).LV;
493 }
494
495 // Hunts for the parameter reference in the parameter copy/move declaration.
496 namespace {
497 struct GetParamRef : public StmtVisitor<GetParamRef> {
498 public:
499 DeclRefExpr *Expr = nullptr;
GetParamRef__anon89ab4d200511::GetParamRef500 GetParamRef() {}
VisitDeclRefExpr__anon89ab4d200511::GetParamRef501 void VisitDeclRefExpr(DeclRefExpr *E) {
502 assert(Expr == nullptr && "multilple declref in param move");
503 Expr = E;
504 }
VisitStmt__anon89ab4d200511::GetParamRef505 void VisitStmt(Stmt *S) {
506 for (auto *C : S->children()) {
507 if (C)
508 Visit(C);
509 }
510 }
511 };
512 }
513
514 // This class replaces references to parameters to their copies by changing
515 // the addresses in CGF.LocalDeclMap and restoring back the original values in
516 // its destructor.
517
518 namespace {
519 struct ParamReferenceReplacerRAII {
520 CodeGenFunction::DeclMapTy SavedLocals;
521 CodeGenFunction::DeclMapTy& LocalDeclMap;
522
ParamReferenceReplacerRAII__anon89ab4d200611::ParamReferenceReplacerRAII523 ParamReferenceReplacerRAII(CodeGenFunction::DeclMapTy &LocalDeclMap)
524 : LocalDeclMap(LocalDeclMap) {}
525
addCopy__anon89ab4d200611::ParamReferenceReplacerRAII526 void addCopy(DeclStmt const *PM) {
527 // Figure out what param it refers to.
528
529 assert(PM->isSingleDecl());
530 VarDecl const*VD = static_cast<VarDecl const*>(PM->getSingleDecl());
531 Expr const *InitExpr = VD->getInit();
532 GetParamRef Visitor;
533 Visitor.Visit(const_cast<Expr*>(InitExpr));
534 assert(Visitor.Expr);
535 DeclRefExpr *DREOrig = Visitor.Expr;
536 auto *PD = DREOrig->getDecl();
537
538 auto it = LocalDeclMap.find(PD);
539 assert(it != LocalDeclMap.end() && "parameter is not found");
540 SavedLocals.insert({ PD, it->second });
541
542 auto copyIt = LocalDeclMap.find(VD);
543 assert(copyIt != LocalDeclMap.end() && "parameter copy is not found");
544 it->second = copyIt->getSecond();
545 }
546
~ParamReferenceReplacerRAII__anon89ab4d200611::ParamReferenceReplacerRAII547 ~ParamReferenceReplacerRAII() {
548 for (auto&& SavedLocal : SavedLocals) {
549 LocalDeclMap.insert({SavedLocal.first, SavedLocal.second});
550 }
551 }
552 };
553 }
554
555 // For WinEH exception representation backend needs to know what funclet coro.end
556 // belongs to. That information is passed in a funclet bundle.
557 static SmallVector<llvm::OperandBundleDef, 1>
getBundlesForCoroEnd(CodeGenFunction & CGF)558 getBundlesForCoroEnd(CodeGenFunction &CGF) {
559 SmallVector<llvm::OperandBundleDef, 1> BundleList;
560
561 if (llvm::Instruction *EHPad = CGF.CurrentFuncletPad)
562 BundleList.emplace_back("funclet", EHPad);
563
564 return BundleList;
565 }
566
567 namespace {
568 // We will insert coro.end to cut any of the destructors for objects that
569 // do not need to be destroyed once the coroutine is resumed.
570 // See llvm/docs/Coroutines.rst for more details about coro.end.
571 struct CallCoroEnd final : public EHScopeStack::Cleanup {
Emit__anon89ab4d200711::CallCoroEnd572 void Emit(CodeGenFunction &CGF, Flags flags) override {
573 auto &CGM = CGF.CGM;
574 auto *NullPtr = llvm::ConstantPointerNull::get(CGF.Int8PtrTy);
575 llvm::Function *CoroEndFn = CGM.getIntrinsic(llvm::Intrinsic::coro_end);
576 // See if we have a funclet bundle to associate coro.end with. (WinEH)
577 auto Bundles = getBundlesForCoroEnd(CGF);
578 auto *CoroEnd =
579 CGF.Builder.CreateCall(CoroEndFn,
580 {NullPtr, CGF.Builder.getTrue(),
581 llvm::ConstantTokenNone::get(CoroEndFn->getContext())},
582 Bundles);
583 if (Bundles.empty()) {
584 // Otherwise, (landingpad model), create a conditional branch that leads
585 // either to a cleanup block or a block with EH resume instruction.
586 auto *ResumeBB = CGF.getEHResumeBlock(/*isCleanup=*/true);
587 auto *CleanupContBB = CGF.createBasicBlock("cleanup.cont");
588 CGF.Builder.CreateCondBr(CoroEnd, ResumeBB, CleanupContBB);
589 CGF.EmitBlock(CleanupContBB);
590 }
591 }
592 };
593 }
594
595 namespace {
596 // Make sure to call coro.delete on scope exit.
597 struct CallCoroDelete final : public EHScopeStack::Cleanup {
598 Stmt *Deallocate;
599
600 // Emit "if (coro.free(CoroId, CoroBegin)) Deallocate;"
601
602 // Note: That deallocation will be emitted twice: once for a normal exit and
603 // once for exceptional exit. This usage is safe because Deallocate does not
604 // contain any declarations. The SubStmtBuilder::makeNewAndDeleteExpr()
605 // builds a single call to a deallocation function which is safe to emit
606 // multiple times.
Emit__anon89ab4d200811::CallCoroDelete607 void Emit(CodeGenFunction &CGF, Flags) override {
608 // Remember the current point, as we are going to emit deallocation code
609 // first to get to coro.free instruction that is an argument to a delete
610 // call.
611 BasicBlock *SaveInsertBlock = CGF.Builder.GetInsertBlock();
612
613 auto *FreeBB = CGF.createBasicBlock("coro.free");
614 CGF.EmitBlock(FreeBB);
615 CGF.EmitStmt(Deallocate);
616
617 auto *AfterFreeBB = CGF.createBasicBlock("after.coro.free");
618 CGF.EmitBlock(AfterFreeBB);
619
620 // We should have captured coro.free from the emission of deallocate.
621 auto *CoroFree = CGF.CurCoro.Data->LastCoroFree;
622 if (!CoroFree) {
623 CGF.CGM.Error(Deallocate->getBeginLoc(),
624 "Deallocation expressoin does not refer to coro.free");
625 return;
626 }
627
628 // Get back to the block we were originally and move coro.free there.
629 auto *InsertPt = SaveInsertBlock->getTerminator();
630 CoroFree->moveBefore(InsertPt->getIterator());
631 CGF.Builder.SetInsertPoint(InsertPt);
632
633 // Add if (auto *mem = coro.free) Deallocate;
634 auto *NullPtr = llvm::ConstantPointerNull::get(CGF.Int8PtrTy);
635 auto *Cond = CGF.Builder.CreateICmpNE(CoroFree, NullPtr);
636 CGF.Builder.CreateCondBr(Cond, FreeBB, AfterFreeBB);
637
638 // No longer need old terminator.
639 InsertPt->eraseFromParent();
640 CGF.Builder.SetInsertPoint(AfterFreeBB);
641 }
CallCoroDelete__anon89ab4d200811::CallCoroDelete642 explicit CallCoroDelete(Stmt *DeallocStmt) : Deallocate(DeallocStmt) {}
643 };
644 }
645
646 namespace {
647 struct GetReturnObjectManager {
648 CodeGenFunction &CGF;
649 CGBuilderTy &Builder;
650 const CoroutineBodyStmt &S;
651 // When true, performs RVO for the return object.
652 bool DirectEmit = false;
653
654 Address GroActiveFlag;
655 CodeGenFunction::AutoVarEmission GroEmission;
656
GetReturnObjectManager__anon89ab4d200911::GetReturnObjectManager657 GetReturnObjectManager(CodeGenFunction &CGF, const CoroutineBodyStmt &S)
658 : CGF(CGF), Builder(CGF.Builder), S(S), GroActiveFlag(Address::invalid()),
659 GroEmission(CodeGenFunction::AutoVarEmission::invalid()) {
660 // The call to get_return_object is sequenced before the call to
661 // initial_suspend and is invoked at most once, but there are caveats
662 // regarding on whether the prvalue result object may be initialized
663 // directly/eager or delayed, depending on the types involved.
664 //
665 // More info at https://github.com/cplusplus/papers/issues/1414
666 //
667 // The general cases:
668 // 1. Same type of get_return_object and coroutine return type (direct
669 // emission):
670 // - Constructed in the return slot.
671 // 2. Different types (delayed emission):
672 // - Constructed temporary object prior to initial suspend initialized with
673 // a call to get_return_object()
674 // - When coroutine needs to to return to the caller and needs to construct
675 // return value for the coroutine it is initialized with expiring value of
676 // the temporary obtained above.
677 //
678 // Direct emission for void returning coroutines or GROs.
679 DirectEmit = [&]() {
680 auto *RVI = S.getReturnValueInit();
681 assert(RVI && "expected RVI");
682 auto GroType = RVI->getType();
683 return CGF.getContext().hasSameType(GroType, CGF.FnRetTy);
684 }();
685 }
686
687 // The gro variable has to outlive coroutine frame and coroutine promise, but,
688 // it can only be initialized after coroutine promise was created, thus, we
689 // split its emission in two parts. EmitGroAlloca emits an alloca and sets up
690 // cleanups. Later when coroutine promise is available we initialize the gro
691 // and sets the flag that the cleanup is now active.
EmitGroAlloca__anon89ab4d200911::GetReturnObjectManager692 void EmitGroAlloca() {
693 if (DirectEmit)
694 return;
695
696 auto *GroDeclStmt = dyn_cast_or_null<DeclStmt>(S.getResultDecl());
697 if (!GroDeclStmt) {
698 // If get_return_object returns void, no need to do an alloca.
699 return;
700 }
701
702 auto *GroVarDecl = cast<VarDecl>(GroDeclStmt->getSingleDecl());
703
704 // Set GRO flag that it is not initialized yet
705 GroActiveFlag = CGF.CreateTempAlloca(Builder.getInt1Ty(), CharUnits::One(),
706 "gro.active");
707 Builder.CreateStore(Builder.getFalse(), GroActiveFlag);
708
709 GroEmission = CGF.EmitAutoVarAlloca(*GroVarDecl);
710
711 if (!GroVarDecl->isNRVOVariable()) {
712 // NRVO variables don't have allocas and won't have the same issue.
713 auto *GroAlloca = dyn_cast_or_null<llvm::AllocaInst>(
714 GroEmission.getOriginalAllocatedAddress().getPointer());
715 assert(GroAlloca && "expected alloca to be emitted");
716 GroAlloca->setMetadata(llvm::LLVMContext::MD_coro_outside_frame,
717 llvm::MDNode::get(CGF.CGM.getLLVMContext(), {}));
718 }
719
720 // Remember the top of EHStack before emitting the cleanup.
721 auto old_top = CGF.EHStack.stable_begin();
722 CGF.EmitAutoVarCleanups(GroEmission);
723 auto top = CGF.EHStack.stable_begin();
724
725 // Make the cleanup conditional on gro.active
726 for (auto b = CGF.EHStack.find(top), e = CGF.EHStack.find(old_top); b != e;
727 b++) {
728 if (auto *Cleanup = dyn_cast<EHCleanupScope>(&*b)) {
729 assert(!Cleanup->hasActiveFlag() && "cleanup already has active flag?");
730 Cleanup->setActiveFlag(GroActiveFlag);
731 Cleanup->setTestFlagInEHCleanup();
732 Cleanup->setTestFlagInNormalCleanup();
733 }
734 }
735 }
736
EmitGroInit__anon89ab4d200911::GetReturnObjectManager737 void EmitGroInit() {
738 if (DirectEmit) {
739 // ReturnValue should be valid as long as the coroutine's return type
740 // is not void. The assertion could help us to reduce the check later.
741 assert(CGF.ReturnValue.isValid() == (bool)S.getReturnStmt());
742 // Now we have the promise, initialize the GRO.
743 // We need to emit `get_return_object` first. According to:
744 // [dcl.fct.def.coroutine]p7
745 // The call to get_return_object is sequenced before the call to
746 // initial_suspend and is invoked at most once.
747 //
748 // So we couldn't emit return value when we emit return statment,
749 // otherwise the call to get_return_object wouldn't be in front
750 // of initial_suspend.
751 if (CGF.ReturnValue.isValid()) {
752 CGF.EmitAnyExprToMem(S.getReturnValue(), CGF.ReturnValue,
753 S.getReturnValue()->getType().getQualifiers(),
754 /*IsInit*/ true);
755 }
756 return;
757 }
758
759 if (!GroActiveFlag.isValid()) {
760 // No Gro variable was allocated. Simply emit the call to
761 // get_return_object.
762 CGF.EmitStmt(S.getResultDecl());
763 return;
764 }
765
766 CGF.EmitAutoVarInit(GroEmission);
767 Builder.CreateStore(Builder.getTrue(), GroActiveFlag);
768 }
769 };
770 } // namespace
771
emitBodyAndFallthrough(CodeGenFunction & CGF,const CoroutineBodyStmt & S,Stmt * Body)772 static void emitBodyAndFallthrough(CodeGenFunction &CGF,
773 const CoroutineBodyStmt &S, Stmt *Body) {
774 CGF.EmitStmt(Body);
775 const bool CanFallthrough = CGF.Builder.GetInsertBlock();
776 if (CanFallthrough)
777 if (Stmt *OnFallthrough = S.getFallthroughHandler())
778 CGF.EmitStmt(OnFallthrough);
779 }
780
EmitCoroutineBody(const CoroutineBodyStmt & S)781 void CodeGenFunction::EmitCoroutineBody(const CoroutineBodyStmt &S) {
782 auto *NullPtr = llvm::ConstantPointerNull::get(Builder.getPtrTy());
783 auto &TI = CGM.getContext().getTargetInfo();
784 unsigned NewAlign = TI.getNewAlign() / TI.getCharWidth();
785
786 auto *EntryBB = Builder.GetInsertBlock();
787 auto *AllocBB = createBasicBlock("coro.alloc");
788 auto *InitBB = createBasicBlock("coro.init");
789 auto *FinalBB = createBasicBlock("coro.final");
790 auto *RetBB = createBasicBlock("coro.ret");
791
792 auto *CoroId = Builder.CreateCall(
793 CGM.getIntrinsic(llvm::Intrinsic::coro_id),
794 {Builder.getInt32(NewAlign), NullPtr, NullPtr, NullPtr});
795 createCoroData(*this, CurCoro, CoroId);
796 CurCoro.Data->SuspendBB = RetBB;
797 assert(ShouldEmitLifetimeMarkers &&
798 "Must emit lifetime intrinsics for coroutines");
799
800 // Backend is allowed to elide memory allocations, to help it, emit
801 // auto mem = coro.alloc() ? 0 : ... allocation code ...;
802 auto *CoroAlloc = Builder.CreateCall(
803 CGM.getIntrinsic(llvm::Intrinsic::coro_alloc), {CoroId});
804
805 Builder.CreateCondBr(CoroAlloc, AllocBB, InitBB);
806
807 EmitBlock(AllocBB);
808 auto *AllocateCall = EmitScalarExpr(S.getAllocate());
809 auto *AllocOrInvokeContBB = Builder.GetInsertBlock();
810
811 // Handle allocation failure if 'ReturnStmtOnAllocFailure' was provided.
812 if (auto *RetOnAllocFailure = S.getReturnStmtOnAllocFailure()) {
813 auto *RetOnFailureBB = createBasicBlock("coro.ret.on.failure");
814
815 // See if allocation was successful.
816 auto *NullPtr = llvm::ConstantPointerNull::get(Int8PtrTy);
817 auto *Cond = Builder.CreateICmpNE(AllocateCall, NullPtr);
818 // Expect the allocation to be successful.
819 emitCondLikelihoodViaExpectIntrinsic(Cond, Stmt::LH_Likely);
820 Builder.CreateCondBr(Cond, InitBB, RetOnFailureBB);
821
822 // If not, return OnAllocFailure object.
823 EmitBlock(RetOnFailureBB);
824 EmitStmt(RetOnAllocFailure);
825 }
826 else {
827 Builder.CreateBr(InitBB);
828 }
829
830 EmitBlock(InitBB);
831
832 // Pass the result of the allocation to coro.begin.
833 auto *Phi = Builder.CreatePHI(VoidPtrTy, 2);
834 Phi->addIncoming(NullPtr, EntryBB);
835 Phi->addIncoming(AllocateCall, AllocOrInvokeContBB);
836 auto *CoroBegin = Builder.CreateCall(
837 CGM.getIntrinsic(llvm::Intrinsic::coro_begin), {CoroId, Phi});
838 CurCoro.Data->CoroBegin = CoroBegin;
839
840 GetReturnObjectManager GroManager(*this, S);
841 GroManager.EmitGroAlloca();
842
843 CurCoro.Data->CleanupJD = getJumpDestInCurrentScope(RetBB);
844 {
845 CGDebugInfo *DI = getDebugInfo();
846 ParamReferenceReplacerRAII ParamReplacer(LocalDeclMap);
847 CodeGenFunction::RunCleanupsScope ResumeScope(*this);
848 EHStack.pushCleanup<CallCoroDelete>(NormalAndEHCleanup, S.getDeallocate());
849
850 // Create mapping between parameters and copy-params for coroutine function.
851 llvm::ArrayRef<const Stmt *> ParamMoves = S.getParamMoves();
852 assert(
853 (ParamMoves.size() == 0 || (ParamMoves.size() == FnArgs.size())) &&
854 "ParamMoves and FnArgs should be the same size for coroutine function");
855 if (ParamMoves.size() == FnArgs.size() && DI)
856 for (const auto Pair : llvm::zip(FnArgs, ParamMoves))
857 DI->getCoroutineParameterMappings().insert(
858 {std::get<0>(Pair), std::get<1>(Pair)});
859
860 // Create parameter copies. We do it before creating a promise, since an
861 // evolution of coroutine TS may allow promise constructor to observe
862 // parameter copies.
863 for (const ParmVarDecl *Parm : FnArgs) {
864 // If the original param is in an alloca, exclude it from the coroutine
865 // frame. The parameter copy will be part of the frame, but the original
866 // parameter memory should remain on the stack. This is necessary to
867 // ensure that parameters destroyed in callees, as with `trivial_abi` or
868 // in the MSVC C++ ABI, are appropriately destroyed after setting up the
869 // coroutine.
870 Address ParmAddr = GetAddrOfLocalVar(Parm);
871 if (auto *ParmAlloca =
872 dyn_cast<llvm::AllocaInst>(ParmAddr.getBasePointer())) {
873 ParmAlloca->setMetadata(llvm::LLVMContext::MD_coro_outside_frame,
874 llvm::MDNode::get(CGM.getLLVMContext(), {}));
875 }
876 }
877 for (auto *PM : S.getParamMoves()) {
878 EmitStmt(PM);
879 ParamReplacer.addCopy(cast<DeclStmt>(PM));
880 // TODO: if(CoroParam(...)) need to surround ctor and dtor
881 // for the copy, so that llvm can elide it if the copy is
882 // not needed.
883 }
884
885 EmitStmt(S.getPromiseDeclStmt());
886
887 Address PromiseAddr = GetAddrOfLocalVar(S.getPromiseDecl());
888 auto *PromiseAddrVoidPtr =
889 new llvm::BitCastInst(PromiseAddr.emitRawPointer(*this), VoidPtrTy, "",
890 CoroId->getIterator());
891 // Update CoroId to refer to the promise. We could not do it earlier because
892 // promise local variable was not emitted yet.
893 CoroId->setArgOperand(1, PromiseAddrVoidPtr);
894
895 // Now we have the promise, initialize the GRO
896 GroManager.EmitGroInit();
897
898 EHStack.pushCleanup<CallCoroEnd>(EHCleanup);
899
900 CurCoro.Data->CurrentAwaitKind = AwaitKind::Init;
901 CurCoro.Data->ExceptionHandler = S.getExceptionHandler();
902 EmitStmt(S.getInitSuspendStmt());
903 CurCoro.Data->FinalJD = getJumpDestInCurrentScope(FinalBB);
904
905 CurCoro.Data->CurrentAwaitKind = AwaitKind::Normal;
906
907 if (CurCoro.Data->ExceptionHandler) {
908 // If we generated IR to record whether an exception was thrown from
909 // 'await_resume', then use that IR to determine whether the coroutine
910 // body should be skipped.
911 // If we didn't generate the IR (perhaps because 'await_resume' was marked
912 // as 'noexcept'), then we skip this check.
913 BasicBlock *ContBB = nullptr;
914 if (CurCoro.Data->ResumeEHVar) {
915 BasicBlock *BodyBB = createBasicBlock("coro.resumed.body");
916 ContBB = createBasicBlock("coro.resumed.cont");
917 Value *SkipBody = Builder.CreateFlagLoad(CurCoro.Data->ResumeEHVar,
918 "coro.resumed.eh");
919 Builder.CreateCondBr(SkipBody, ContBB, BodyBB);
920 EmitBlock(BodyBB);
921 }
922
923 auto Loc = S.getBeginLoc();
924 CXXCatchStmt Catch(Loc, /*exDecl=*/nullptr,
925 CurCoro.Data->ExceptionHandler);
926 auto *TryStmt =
927 CXXTryStmt::Create(getContext(), Loc, S.getBody(), &Catch);
928
929 EnterCXXTryStmt(*TryStmt);
930 emitBodyAndFallthrough(*this, S, TryStmt->getTryBlock());
931 ExitCXXTryStmt(*TryStmt);
932
933 if (ContBB)
934 EmitBlock(ContBB);
935 }
936 else {
937 emitBodyAndFallthrough(*this, S, S.getBody());
938 }
939
940 // See if we need to generate final suspend.
941 const bool CanFallthrough = Builder.GetInsertBlock();
942 const bool HasCoreturns = CurCoro.Data->CoreturnCount > 0;
943 if (CanFallthrough || HasCoreturns) {
944 EmitBlock(FinalBB);
945 CurCoro.Data->CurrentAwaitKind = AwaitKind::Final;
946 EmitStmt(S.getFinalSuspendStmt());
947 } else {
948 // We don't need FinalBB. Emit it to make sure the block is deleted.
949 EmitBlock(FinalBB, /*IsFinished=*/true);
950 }
951 }
952
953 EmitBlock(RetBB);
954 // Emit coro.end before getReturnStmt (and parameter destructors), since
955 // resume and destroy parts of the coroutine should not include them.
956 llvm::Function *CoroEnd = CGM.getIntrinsic(llvm::Intrinsic::coro_end);
957 Builder.CreateCall(CoroEnd,
958 {NullPtr, Builder.getFalse(),
959 llvm::ConstantTokenNone::get(CoroEnd->getContext())});
960
961 if (Stmt *Ret = S.getReturnStmt()) {
962 // Since we already emitted the return value above, so we shouldn't
963 // emit it again here.
964 Expr *PreviousRetValue = nullptr;
965 if (GroManager.DirectEmit) {
966 PreviousRetValue = cast<ReturnStmt>(Ret)->getRetValue();
967 cast<ReturnStmt>(Ret)->setRetValue(nullptr);
968 }
969 EmitStmt(Ret);
970 // Set the return value back. The code generator, as the AST **Consumer**,
971 // shouldn't change the AST.
972 if (PreviousRetValue)
973 cast<ReturnStmt>(Ret)->setRetValue(PreviousRetValue);
974 }
975
976 // LLVM require the frontend to mark the coroutine.
977 CurFn->setPresplitCoroutine();
978
979 if (CXXRecordDecl *RD = FnRetTy->getAsCXXRecordDecl();
980 RD && RD->hasAttr<CoroOnlyDestroyWhenCompleteAttr>())
981 CurFn->setCoroDestroyOnlyWhenComplete();
982 }
983
984 // Emit coroutine intrinsic and patch up arguments of the token type.
EmitCoroutineIntrinsic(const CallExpr * E,unsigned int IID)985 RValue CodeGenFunction::EmitCoroutineIntrinsic(const CallExpr *E,
986 unsigned int IID) {
987 SmallVector<llvm::Value *, 8> Args;
988 switch (IID) {
989 default:
990 break;
991 // The coro.frame builtin is replaced with an SSA value of the coro.begin
992 // intrinsic.
993 case llvm::Intrinsic::coro_frame: {
994 if (CurCoro.Data && CurCoro.Data->CoroBegin) {
995 return RValue::get(CurCoro.Data->CoroBegin);
996 }
997
998 if (CurAwaitSuspendWrapper.FramePtr) {
999 return RValue::get(CurAwaitSuspendWrapper.FramePtr);
1000 }
1001
1002 CGM.Error(E->getBeginLoc(), "this builtin expect that __builtin_coro_begin "
1003 "has been used earlier in this function");
1004 auto *NullPtr = llvm::ConstantPointerNull::get(Builder.getPtrTy());
1005 return RValue::get(NullPtr);
1006 }
1007 case llvm::Intrinsic::coro_size: {
1008 auto &Context = getContext();
1009 CanQualType SizeTy = Context.getSizeType();
1010 llvm::IntegerType *T = Builder.getIntNTy(Context.getTypeSize(SizeTy));
1011 llvm::Function *F = CGM.getIntrinsic(llvm::Intrinsic::coro_size, T);
1012 return RValue::get(Builder.CreateCall(F));
1013 }
1014 case llvm::Intrinsic::coro_align: {
1015 auto &Context = getContext();
1016 CanQualType SizeTy = Context.getSizeType();
1017 llvm::IntegerType *T = Builder.getIntNTy(Context.getTypeSize(SizeTy));
1018 llvm::Function *F = CGM.getIntrinsic(llvm::Intrinsic::coro_align, T);
1019 return RValue::get(Builder.CreateCall(F));
1020 }
1021 // The following three intrinsics take a token parameter referring to a token
1022 // returned by earlier call to @llvm.coro.id. Since we cannot represent it in
1023 // builtins, we patch it up here.
1024 case llvm::Intrinsic::coro_alloc:
1025 case llvm::Intrinsic::coro_begin:
1026 case llvm::Intrinsic::coro_free: {
1027 if (CurCoro.Data && CurCoro.Data->CoroId) {
1028 Args.push_back(CurCoro.Data->CoroId);
1029 break;
1030 }
1031 CGM.Error(E->getBeginLoc(), "this builtin expect that __builtin_coro_id has"
1032 " been used earlier in this function");
1033 // Fallthrough to the next case to add TokenNone as the first argument.
1034 [[fallthrough]];
1035 }
1036 // @llvm.coro.suspend takes a token parameter. Add token 'none' as the first
1037 // argument.
1038 case llvm::Intrinsic::coro_suspend:
1039 Args.push_back(llvm::ConstantTokenNone::get(getLLVMContext()));
1040 break;
1041 }
1042 for (const Expr *Arg : E->arguments())
1043 Args.push_back(EmitScalarExpr(Arg));
1044 // @llvm.coro.end takes a token parameter. Add token 'none' as the last
1045 // argument.
1046 if (IID == llvm::Intrinsic::coro_end)
1047 Args.push_back(llvm::ConstantTokenNone::get(getLLVMContext()));
1048
1049 llvm::Function *F = CGM.getIntrinsic(IID);
1050 llvm::CallInst *Call = Builder.CreateCall(F, Args);
1051
1052 // Note: The following code is to enable to emit coro.id and coro.begin by
1053 // hand to experiment with coroutines in C.
1054 // If we see @llvm.coro.id remember it in the CoroData. We will update
1055 // coro.alloc, coro.begin and coro.free intrinsics to refer to it.
1056 if (IID == llvm::Intrinsic::coro_id) {
1057 createCoroData(*this, CurCoro, Call, E);
1058 }
1059 else if (IID == llvm::Intrinsic::coro_begin) {
1060 if (CurCoro.Data)
1061 CurCoro.Data->CoroBegin = Call;
1062 }
1063 else if (IID == llvm::Intrinsic::coro_free) {
1064 // Remember the last coro_free as we need it to build the conditional
1065 // deletion of the coroutine frame.
1066 if (CurCoro.Data)
1067 CurCoro.Data->LastCoroFree = Call;
1068 }
1069 return RValue::get(Call);
1070 }
1071