1 //===- AArch64AsmPrinter.cpp - AArch64 LLVM assembly writer ---------------===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 //
9 // This file contains a printer that converts from our internal representation
10 // of machine-dependent LLVM code to the AArch64 assembly language.
11 //
12 //===----------------------------------------------------------------------===//
13
14 #include "AArch64.h"
15 #include "AArch64MCInstLower.h"
16 #include "AArch64MachineFunctionInfo.h"
17 #include "AArch64RegisterInfo.h"
18 #include "AArch64Subtarget.h"
19 #include "AArch64TargetObjectFile.h"
20 #include "MCTargetDesc/AArch64AddressingModes.h"
21 #include "MCTargetDesc/AArch64InstPrinter.h"
22 #include "MCTargetDesc/AArch64MCAsmInfo.h"
23 #include "MCTargetDesc/AArch64MCTargetDesc.h"
24 #include "MCTargetDesc/AArch64TargetStreamer.h"
25 #include "TargetInfo/AArch64TargetInfo.h"
26 #include "Utils/AArch64BaseInfo.h"
27 #include "llvm/ADT/DenseMap.h"
28 #include "llvm/ADT/ScopeExit.h"
29 #include "llvm/ADT/SmallString.h"
30 #include "llvm/ADT/SmallVector.h"
31 #include "llvm/ADT/StringRef.h"
32 #include "llvm/ADT/Twine.h"
33 #include "llvm/BinaryFormat/COFF.h"
34 #include "llvm/BinaryFormat/ELF.h"
35 #include "llvm/BinaryFormat/MachO.h"
36 #include "llvm/CodeGen/AsmPrinter.h"
37 #include "llvm/CodeGen/FaultMaps.h"
38 #include "llvm/CodeGen/MachineBasicBlock.h"
39 #include "llvm/CodeGen/MachineFunction.h"
40 #include "llvm/CodeGen/MachineInstr.h"
41 #include "llvm/CodeGen/MachineJumpTableInfo.h"
42 #include "llvm/CodeGen/MachineModuleInfoImpls.h"
43 #include "llvm/CodeGen/MachineOperand.h"
44 #include "llvm/CodeGen/StackMaps.h"
45 #include "llvm/CodeGen/TargetRegisterInfo.h"
46 #include "llvm/IR/DataLayout.h"
47 #include "llvm/IR/DebugInfoMetadata.h"
48 #include "llvm/IR/Mangler.h"
49 #include "llvm/IR/Module.h"
50 #include "llvm/MC/MCAsmInfo.h"
51 #include "llvm/MC/MCContext.h"
52 #include "llvm/MC/MCInst.h"
53 #include "llvm/MC/MCInstBuilder.h"
54 #include "llvm/MC/MCSectionELF.h"
55 #include "llvm/MC/MCSectionMachO.h"
56 #include "llvm/MC/MCStreamer.h"
57 #include "llvm/MC/MCSymbol.h"
58 #include "llvm/MC/TargetRegistry.h"
59 #include "llvm/Support/Casting.h"
60 #include "llvm/Support/CommandLine.h"
61 #include "llvm/Support/Compiler.h"
62 #include "llvm/Support/ErrorHandling.h"
63 #include "llvm/Support/raw_ostream.h"
64 #include "llvm/Target/TargetMachine.h"
65 #include "llvm/TargetParser/Triple.h"
66 #include "llvm/Transforms/Instrumentation/HWAddressSanitizer.h"
67 #include <cassert>
68 #include <cstdint>
69 #include <map>
70 #include <memory>
71
72 using namespace llvm;
73
74 enum PtrauthCheckMode { Default, Unchecked, Poison, Trap };
75 static cl::opt<PtrauthCheckMode> PtrauthAuthChecks(
76 "aarch64-ptrauth-auth-checks", cl::Hidden,
77 cl::values(clEnumValN(Unchecked, "none", "don't test for failure"),
78 clEnumValN(Poison, "poison", "poison on failure"),
79 clEnumValN(Trap, "trap", "trap on failure")),
80 cl::desc("Check pointer authentication auth/resign failures"),
81 cl::init(Default));
82
83 #define DEBUG_TYPE "asm-printer"
84
85 namespace {
86
87 class AArch64AsmPrinter : public AsmPrinter {
88 AArch64MCInstLower MCInstLowering;
89 FaultMaps FM;
90 const AArch64Subtarget *STI;
91 bool ShouldEmitWeakSwiftAsyncExtendedFramePointerFlags = false;
92 #ifndef NDEBUG
93 unsigned InstsEmitted;
94 #endif
95 bool EnableImportCallOptimization = false;
96 DenseMap<MCSection *, std::vector<std::pair<MCSymbol *, MCSymbol *>>>
97 SectionToImportedFunctionCalls;
98
99 public:
100 static char ID;
101
AArch64AsmPrinter(TargetMachine & TM,std::unique_ptr<MCStreamer> Streamer)102 AArch64AsmPrinter(TargetMachine &TM, std::unique_ptr<MCStreamer> Streamer)
103 : AsmPrinter(TM, std::move(Streamer), ID),
104 MCInstLowering(OutContext, *this), FM(*this) {}
105
getPassName() const106 StringRef getPassName() const override { return "AArch64 Assembly Printer"; }
107
108 /// Wrapper for MCInstLowering.lowerOperand() for the
109 /// tblgen'erated pseudo lowering.
lowerOperand(const MachineOperand & MO,MCOperand & MCOp) const110 bool lowerOperand(const MachineOperand &MO, MCOperand &MCOp) const {
111 return MCInstLowering.lowerOperand(MO, MCOp);
112 }
113
114 const MCExpr *lowerConstantPtrAuth(const ConstantPtrAuth &CPA) override;
115
116 const MCExpr *lowerBlockAddressConstant(const BlockAddress &BA) override;
117
118 void emitStartOfAsmFile(Module &M) override;
119 void emitJumpTableImpl(const MachineJumpTableInfo &MJTI,
120 ArrayRef<unsigned> JumpTableIndices) override;
121 std::tuple<const MCSymbol *, uint64_t, const MCSymbol *,
122 codeview::JumpTableEntrySize>
123 getCodeViewJumpTableInfo(int JTI, const MachineInstr *BranchInstr,
124 const MCSymbol *BranchLabel) const override;
125
126 void emitFunctionEntryLabel() override;
127
128 void emitXXStructor(const DataLayout &DL, const Constant *CV) override;
129
130 void LowerJumpTableDest(MCStreamer &OutStreamer, const MachineInstr &MI);
131
132 void LowerHardenedBRJumpTable(const MachineInstr &MI);
133
134 void LowerMOPS(MCStreamer &OutStreamer, const MachineInstr &MI);
135
136 void LowerSTACKMAP(MCStreamer &OutStreamer, StackMaps &SM,
137 const MachineInstr &MI);
138 void LowerPATCHPOINT(MCStreamer &OutStreamer, StackMaps &SM,
139 const MachineInstr &MI);
140 void LowerSTATEPOINT(MCStreamer &OutStreamer, StackMaps &SM,
141 const MachineInstr &MI);
142 void LowerFAULTING_OP(const MachineInstr &MI);
143
144 void LowerPATCHABLE_FUNCTION_ENTER(const MachineInstr &MI);
145 void LowerPATCHABLE_FUNCTION_EXIT(const MachineInstr &MI);
146 void LowerPATCHABLE_TAIL_CALL(const MachineInstr &MI);
147 void LowerPATCHABLE_EVENT_CALL(const MachineInstr &MI, bool Typed);
148
149 typedef std::tuple<unsigned, bool, uint32_t, bool, uint64_t>
150 HwasanMemaccessTuple;
151 std::map<HwasanMemaccessTuple, MCSymbol *> HwasanMemaccessSymbols;
152 void LowerKCFI_CHECK(const MachineInstr &MI);
153 void LowerHWASAN_CHECK_MEMACCESS(const MachineInstr &MI);
154 void emitHwasanMemaccessSymbols(Module &M);
155
156 void emitSled(const MachineInstr &MI, SledKind Kind);
157
158 // Emit the sequence for BRA/BLRA (authenticate + branch/call).
159 void emitPtrauthBranch(const MachineInstr *MI);
160
161 void emitPtrauthCheckAuthenticatedValue(Register TestedReg,
162 Register ScratchReg,
163 AArch64PACKey::ID Key,
164 AArch64PAuth::AuthCheckMethod Method,
165 bool ShouldTrap,
166 const MCSymbol *OnFailure);
167
168 // Check authenticated LR before tail calling.
169 void emitPtrauthTailCallHardening(const MachineInstr *TC);
170
171 // Emit the sequence for AUT or AUTPAC.
172 void emitPtrauthAuthResign(Register AUTVal, AArch64PACKey::ID AUTKey,
173 uint64_t AUTDisc,
174 const MachineOperand *AUTAddrDisc,
175 Register Scratch,
176 std::optional<AArch64PACKey::ID> PACKey,
177 uint64_t PACDisc, Register PACAddrDisc);
178
179 // Emit the sequence to compute the discriminator.
180 //
181 // The returned register is either unmodified AddrDisc or ScratchReg.
182 //
183 // If the expanded pseudo is allowed to clobber AddrDisc register, setting
184 // MayUseAddrAsScratch may save one MOV instruction, provided the address
185 // is already in x16/x17 (i.e. return x16/x17 which is the *modified* AddrDisc
186 // register at the same time) or the OS doesn't make it safer to use x16/x17
187 // (see AArch64Subtarget::isX16X17Safer()):
188 //
189 // mov x17, x16
190 // movk x17, #1234, lsl #48
191 // ; x16 is not used anymore
192 //
193 // can be replaced by
194 //
195 // movk x16, #1234, lsl #48
196 Register emitPtrauthDiscriminator(uint16_t Disc, Register AddrDisc,
197 Register ScratchReg,
198 bool MayUseAddrAsScratch = false);
199
200 // Emit the sequence for LOADauthptrstatic
201 void LowerLOADauthptrstatic(const MachineInstr &MI);
202
203 // Emit the sequence for LOADgotPAC/MOVaddrPAC (either GOT adrp-ldr or
204 // adrp-add followed by PAC sign)
205 void LowerMOVaddrPAC(const MachineInstr &MI);
206
207 // Emit the sequence for LOADgotAUTH (load signed pointer from signed ELF GOT
208 // and authenticate it with, if FPAC bit is not set, check+trap sequence after
209 // authenticating)
210 void LowerLOADgotAUTH(const MachineInstr &MI);
211
212 /// tblgen'erated driver function for lowering simple MI->MC
213 /// pseudo instructions.
214 bool lowerPseudoInstExpansion(const MachineInstr *MI, MCInst &Inst);
215
216 // Emit Build Attributes
217 void emitAttributes(unsigned Flags, uint64_t PAuthABIPlatform,
218 uint64_t PAuthABIVersion, AArch64TargetStreamer *TS);
219
220 // Emit expansion of Compare-and-branch pseudo instructions
221 void emitCBPseudoExpansion(const MachineInstr *MI);
222
223 void EmitToStreamer(MCStreamer &S, const MCInst &Inst);
EmitToStreamer(const MCInst & Inst)224 void EmitToStreamer(const MCInst &Inst) {
225 EmitToStreamer(*OutStreamer, Inst);
226 }
227
228 void emitInstruction(const MachineInstr *MI) override;
229
230 void emitFunctionHeaderComment() override;
231
getAnalysisUsage(AnalysisUsage & AU) const232 void getAnalysisUsage(AnalysisUsage &AU) const override {
233 AsmPrinter::getAnalysisUsage(AU);
234 AU.setPreservesAll();
235 }
236
runOnMachineFunction(MachineFunction & MF)237 bool runOnMachineFunction(MachineFunction &MF) override {
238 if (auto *PSIW = getAnalysisIfAvailable<ProfileSummaryInfoWrapperPass>())
239 PSI = &PSIW->getPSI();
240 if (auto *SDPIW =
241 getAnalysisIfAvailable<StaticDataProfileInfoWrapperPass>())
242 SDPI = &SDPIW->getStaticDataProfileInfo();
243
244 AArch64FI = MF.getInfo<AArch64FunctionInfo>();
245 STI = &MF.getSubtarget<AArch64Subtarget>();
246
247 SetupMachineFunction(MF);
248
249 if (STI->isTargetCOFF()) {
250 bool Local = MF.getFunction().hasLocalLinkage();
251 COFF::SymbolStorageClass Scl =
252 Local ? COFF::IMAGE_SYM_CLASS_STATIC : COFF::IMAGE_SYM_CLASS_EXTERNAL;
253 int Type =
254 COFF::IMAGE_SYM_DTYPE_FUNCTION << COFF::SCT_COMPLEX_TYPE_SHIFT;
255
256 OutStreamer->beginCOFFSymbolDef(CurrentFnSym);
257 OutStreamer->emitCOFFSymbolStorageClass(Scl);
258 OutStreamer->emitCOFFSymbolType(Type);
259 OutStreamer->endCOFFSymbolDef();
260 }
261
262 // Emit the rest of the function body.
263 emitFunctionBody();
264
265 // Emit the XRay table for this function.
266 emitXRayTable();
267
268 // We didn't modify anything.
269 return false;
270 }
271
272 const MCExpr *lowerConstant(const Constant *CV,
273 const Constant *BaseCV = nullptr,
274 uint64_t Offset = 0) override;
275
276 private:
277 void printOperand(const MachineInstr *MI, unsigned OpNum, raw_ostream &O);
278 bool printAsmMRegister(const MachineOperand &MO, char Mode, raw_ostream &O);
279 bool printAsmRegInClass(const MachineOperand &MO,
280 const TargetRegisterClass *RC, unsigned AltName,
281 raw_ostream &O);
282
283 bool PrintAsmOperand(const MachineInstr *MI, unsigned OpNum,
284 const char *ExtraCode, raw_ostream &O) override;
285 bool PrintAsmMemoryOperand(const MachineInstr *MI, unsigned OpNum,
286 const char *ExtraCode, raw_ostream &O) override;
287
288 void PrintDebugValueComment(const MachineInstr *MI, raw_ostream &OS);
289
290 void emitFunctionBodyEnd() override;
291 void emitGlobalAlias(const Module &M, const GlobalAlias &GA) override;
292
293 MCSymbol *GetCPISymbol(unsigned CPID) const override;
294 void emitEndOfAsmFile(Module &M) override;
295
296 AArch64FunctionInfo *AArch64FI = nullptr;
297
298 /// Emit the LOHs contained in AArch64FI.
299 void emitLOHs();
300
301 void emitMovXReg(Register Dest, Register Src);
302 void emitMOVZ(Register Dest, uint64_t Imm, unsigned Shift);
303 void emitMOVK(Register Dest, uint64_t Imm, unsigned Shift);
304
305 /// Emit instruction to set float register to zero.
306 void emitFMov0(const MachineInstr &MI);
307
308 using MInstToMCSymbol = std::map<const MachineInstr *, MCSymbol *>;
309
310 MInstToMCSymbol LOHInstToLabel;
311
shouldEmitWeakSwiftAsyncExtendedFramePointerFlags() const312 bool shouldEmitWeakSwiftAsyncExtendedFramePointerFlags() const override {
313 return ShouldEmitWeakSwiftAsyncExtendedFramePointerFlags;
314 }
315
getIFuncMCSubtargetInfo() const316 const MCSubtargetInfo *getIFuncMCSubtargetInfo() const override {
317 assert(STI);
318 return STI;
319 }
320 void emitMachOIFuncStubBody(Module &M, const GlobalIFunc &GI,
321 MCSymbol *LazyPointer) override;
322 void emitMachOIFuncStubHelperBody(Module &M, const GlobalIFunc &GI,
323 MCSymbol *LazyPointer) override;
324
325 /// Checks if this instruction is part of a sequence that is eligle for import
326 /// call optimization and, if so, records it to be emitted in the import call
327 /// section.
328 void recordIfImportCall(const MachineInstr *BranchInst);
329 };
330
331 } // end anonymous namespace
332
emitStartOfAsmFile(Module & M)333 void AArch64AsmPrinter::emitStartOfAsmFile(Module &M) {
334 const Triple &TT = TM.getTargetTriple();
335
336 if (TT.isOSBinFormatCOFF()) {
337 emitCOFFFeatureSymbol(M);
338 emitCOFFReplaceableFunctionData(M);
339
340 if (M.getModuleFlag("import-call-optimization"))
341 EnableImportCallOptimization = true;
342 }
343
344 if (!TT.isOSBinFormatELF())
345 return;
346
347 // For emitting build attributes and .note.gnu.property section
348 auto *TS =
349 static_cast<AArch64TargetStreamer *>(OutStreamer->getTargetStreamer());
350 // Assemble feature flags that may require creation of build attributes and a
351 // note section.
352 unsigned BAFlags = 0;
353 unsigned GNUFlags = 0;
354 if (const auto *BTE = mdconst::extract_or_null<ConstantInt>(
355 M.getModuleFlag("branch-target-enforcement"))) {
356 if (!BTE->isZero()) {
357 BAFlags |= AArch64BuildAttributes::FeatureAndBitsFlag::Feature_BTI_Flag;
358 GNUFlags |= ELF::GNU_PROPERTY_AARCH64_FEATURE_1_BTI;
359 }
360 }
361
362 if (const auto *GCS = mdconst::extract_or_null<ConstantInt>(
363 M.getModuleFlag("guarded-control-stack"))) {
364 if (!GCS->isZero()) {
365 BAFlags |= AArch64BuildAttributes::FeatureAndBitsFlag::Feature_GCS_Flag;
366 GNUFlags |= ELF::GNU_PROPERTY_AARCH64_FEATURE_1_GCS;
367 }
368 }
369
370 if (const auto *Sign = mdconst::extract_or_null<ConstantInt>(
371 M.getModuleFlag("sign-return-address"))) {
372 if (!Sign->isZero()) {
373 BAFlags |= AArch64BuildAttributes::FeatureAndBitsFlag::Feature_PAC_Flag;
374 GNUFlags |= ELF::GNU_PROPERTY_AARCH64_FEATURE_1_PAC;
375 }
376 }
377
378 uint64_t PAuthABIPlatform = -1;
379 if (const auto *PAP = mdconst::extract_or_null<ConstantInt>(
380 M.getModuleFlag("aarch64-elf-pauthabi-platform"))) {
381 PAuthABIPlatform = PAP->getZExtValue();
382 }
383
384 uint64_t PAuthABIVersion = -1;
385 if (const auto *PAV = mdconst::extract_or_null<ConstantInt>(
386 M.getModuleFlag("aarch64-elf-pauthabi-version"))) {
387 PAuthABIVersion = PAV->getZExtValue();
388 }
389
390 // Emit AArch64 Build Attributes
391 emitAttributes(BAFlags, PAuthABIPlatform, PAuthABIVersion, TS);
392 // Emit a .note.gnu.property section with the flags.
393 TS->emitNoteSection(GNUFlags, PAuthABIPlatform, PAuthABIVersion);
394 }
395
emitFunctionHeaderComment()396 void AArch64AsmPrinter::emitFunctionHeaderComment() {
397 const AArch64FunctionInfo *FI = MF->getInfo<AArch64FunctionInfo>();
398 std::optional<std::string> OutlinerString = FI->getOutliningStyle();
399 if (OutlinerString != std::nullopt)
400 OutStreamer->getCommentOS() << ' ' << OutlinerString;
401 }
402
LowerPATCHABLE_FUNCTION_ENTER(const MachineInstr & MI)403 void AArch64AsmPrinter::LowerPATCHABLE_FUNCTION_ENTER(const MachineInstr &MI)
404 {
405 const Function &F = MF->getFunction();
406 if (F.hasFnAttribute("patchable-function-entry")) {
407 unsigned Num;
408 if (F.getFnAttribute("patchable-function-entry")
409 .getValueAsString()
410 .getAsInteger(10, Num))
411 return;
412 emitNops(Num);
413 return;
414 }
415
416 emitSled(MI, SledKind::FUNCTION_ENTER);
417 }
418
LowerPATCHABLE_FUNCTION_EXIT(const MachineInstr & MI)419 void AArch64AsmPrinter::LowerPATCHABLE_FUNCTION_EXIT(const MachineInstr &MI) {
420 emitSled(MI, SledKind::FUNCTION_EXIT);
421 }
422
LowerPATCHABLE_TAIL_CALL(const MachineInstr & MI)423 void AArch64AsmPrinter::LowerPATCHABLE_TAIL_CALL(const MachineInstr &MI) {
424 emitSled(MI, SledKind::TAIL_CALL);
425 }
426
emitSled(const MachineInstr & MI,SledKind Kind)427 void AArch64AsmPrinter::emitSled(const MachineInstr &MI, SledKind Kind) {
428 static const int8_t NoopsInSledCount = 7;
429 // We want to emit the following pattern:
430 //
431 // .Lxray_sled_N:
432 // ALIGN
433 // B #32
434 // ; 7 NOP instructions (28 bytes)
435 // .tmpN
436 //
437 // We need the 28 bytes (7 instructions) because at runtime, we'd be patching
438 // over the full 32 bytes (8 instructions) with the following pattern:
439 //
440 // STP X0, X30, [SP, #-16]! ; push X0 and the link register to the stack
441 // LDR W17, #12 ; W17 := function ID
442 // LDR X16,#12 ; X16 := addr of __xray_FunctionEntry or __xray_FunctionExit
443 // BLR X16 ; call the tracing trampoline
444 // ;DATA: 32 bits of function ID
445 // ;DATA: lower 32 bits of the address of the trampoline
446 // ;DATA: higher 32 bits of the address of the trampoline
447 // LDP X0, X30, [SP], #16 ; pop X0 and the link register from the stack
448 //
449 OutStreamer->emitCodeAlignment(Align(4), &getSubtargetInfo());
450 auto CurSled = OutContext.createTempSymbol("xray_sled_", true);
451 OutStreamer->emitLabel(CurSled);
452 auto Target = OutContext.createTempSymbol();
453
454 // Emit "B #32" instruction, which jumps over the next 28 bytes.
455 // The operand has to be the number of 4-byte instructions to jump over,
456 // including the current instruction.
457 EmitToStreamer(*OutStreamer, MCInstBuilder(AArch64::B).addImm(8));
458
459 for (int8_t I = 0; I < NoopsInSledCount; I++)
460 EmitToStreamer(*OutStreamer, MCInstBuilder(AArch64::HINT).addImm(0));
461
462 OutStreamer->emitLabel(Target);
463 recordSled(CurSled, MI, Kind, 2);
464 }
465
emitAttributes(unsigned Flags,uint64_t PAuthABIPlatform,uint64_t PAuthABIVersion,AArch64TargetStreamer * TS)466 void AArch64AsmPrinter::emitAttributes(unsigned Flags,
467 uint64_t PAuthABIPlatform,
468 uint64_t PAuthABIVersion,
469 AArch64TargetStreamer *TS) {
470
471 PAuthABIPlatform = (uint64_t(-1) == PAuthABIPlatform) ? 0 : PAuthABIPlatform;
472 PAuthABIVersion = (uint64_t(-1) == PAuthABIVersion) ? 0 : PAuthABIVersion;
473
474 if (PAuthABIPlatform || PAuthABIVersion) {
475 TS->emitAttributesSubsection(
476 AArch64BuildAttributes::getVendorName(
477 AArch64BuildAttributes::AEABI_PAUTHABI),
478 AArch64BuildAttributes::SubsectionOptional::REQUIRED,
479 AArch64BuildAttributes::SubsectionType::ULEB128);
480 TS->emitAttribute(AArch64BuildAttributes::getVendorName(
481 AArch64BuildAttributes::AEABI_PAUTHABI),
482 AArch64BuildAttributes::TAG_PAUTH_PLATFORM,
483 PAuthABIPlatform, "");
484 TS->emitAttribute(AArch64BuildAttributes::getVendorName(
485 AArch64BuildAttributes::AEABI_PAUTHABI),
486 AArch64BuildAttributes::TAG_PAUTH_SCHEMA, PAuthABIVersion,
487 "");
488 }
489
490 unsigned BTIValue =
491 (Flags & AArch64BuildAttributes::Feature_BTI_Flag) ? 1 : 0;
492 unsigned PACValue =
493 (Flags & AArch64BuildAttributes::Feature_PAC_Flag) ? 1 : 0;
494 unsigned GCSValue =
495 (Flags & AArch64BuildAttributes::Feature_GCS_Flag) ? 1 : 0;
496
497 if (BTIValue || PACValue || GCSValue) {
498 TS->emitAttributesSubsection(
499 AArch64BuildAttributes::getVendorName(
500 AArch64BuildAttributes::AEABI_FEATURE_AND_BITS),
501 AArch64BuildAttributes::SubsectionOptional::OPTIONAL,
502 AArch64BuildAttributes::SubsectionType::ULEB128);
503 TS->emitAttribute(AArch64BuildAttributes::getVendorName(
504 AArch64BuildAttributes::AEABI_FEATURE_AND_BITS),
505 AArch64BuildAttributes::TAG_FEATURE_BTI, BTIValue, "");
506 TS->emitAttribute(AArch64BuildAttributes::getVendorName(
507 AArch64BuildAttributes::AEABI_FEATURE_AND_BITS),
508 AArch64BuildAttributes::TAG_FEATURE_PAC, PACValue, "");
509 TS->emitAttribute(AArch64BuildAttributes::getVendorName(
510 AArch64BuildAttributes::AEABI_FEATURE_AND_BITS),
511 AArch64BuildAttributes::TAG_FEATURE_GCS, GCSValue, "");
512 }
513 }
514
515 // Emit the following code for Intrinsic::{xray_customevent,xray_typedevent}
516 // (built-in functions __xray_customevent/__xray_typedevent).
517 //
518 // .Lxray_event_sled_N:
519 // b 1f
520 // save x0 and x1 (and also x2 for TYPED_EVENT_CALL)
521 // set up x0 and x1 (and also x2 for TYPED_EVENT_CALL)
522 // bl __xray_CustomEvent or __xray_TypedEvent
523 // restore x0 and x1 (and also x2 for TYPED_EVENT_CALL)
524 // 1:
525 //
526 // There are 6 instructions for EVENT_CALL and 9 for TYPED_EVENT_CALL.
527 //
528 // Then record a sled of kind CUSTOM_EVENT or TYPED_EVENT.
529 // After patching, b .+N will become a nop.
LowerPATCHABLE_EVENT_CALL(const MachineInstr & MI,bool Typed)530 void AArch64AsmPrinter::LowerPATCHABLE_EVENT_CALL(const MachineInstr &MI,
531 bool Typed) {
532 auto &O = *OutStreamer;
533 MCSymbol *CurSled = OutContext.createTempSymbol("xray_sled_", true);
534 O.emitLabel(CurSled);
535 bool MachO = TM.getTargetTriple().isOSBinFormatMachO();
536 auto *Sym = MCSymbolRefExpr::create(
537 OutContext.getOrCreateSymbol(
538 Twine(MachO ? "_" : "") +
539 (Typed ? "__xray_TypedEvent" : "__xray_CustomEvent")),
540 OutContext);
541 if (Typed) {
542 O.AddComment("Begin XRay typed event");
543 EmitToStreamer(O, MCInstBuilder(AArch64::B).addImm(9));
544 EmitToStreamer(O, MCInstBuilder(AArch64::STPXpre)
545 .addReg(AArch64::SP)
546 .addReg(AArch64::X0)
547 .addReg(AArch64::X1)
548 .addReg(AArch64::SP)
549 .addImm(-4));
550 EmitToStreamer(O, MCInstBuilder(AArch64::STRXui)
551 .addReg(AArch64::X2)
552 .addReg(AArch64::SP)
553 .addImm(2));
554 emitMovXReg(AArch64::X0, MI.getOperand(0).getReg());
555 emitMovXReg(AArch64::X1, MI.getOperand(1).getReg());
556 emitMovXReg(AArch64::X2, MI.getOperand(2).getReg());
557 EmitToStreamer(O, MCInstBuilder(AArch64::BL).addExpr(Sym));
558 EmitToStreamer(O, MCInstBuilder(AArch64::LDRXui)
559 .addReg(AArch64::X2)
560 .addReg(AArch64::SP)
561 .addImm(2));
562 O.AddComment("End XRay typed event");
563 EmitToStreamer(O, MCInstBuilder(AArch64::LDPXpost)
564 .addReg(AArch64::SP)
565 .addReg(AArch64::X0)
566 .addReg(AArch64::X1)
567 .addReg(AArch64::SP)
568 .addImm(4));
569
570 recordSled(CurSled, MI, SledKind::TYPED_EVENT, 2);
571 } else {
572 O.AddComment("Begin XRay custom event");
573 EmitToStreamer(O, MCInstBuilder(AArch64::B).addImm(6));
574 EmitToStreamer(O, MCInstBuilder(AArch64::STPXpre)
575 .addReg(AArch64::SP)
576 .addReg(AArch64::X0)
577 .addReg(AArch64::X1)
578 .addReg(AArch64::SP)
579 .addImm(-2));
580 emitMovXReg(AArch64::X0, MI.getOperand(0).getReg());
581 emitMovXReg(AArch64::X1, MI.getOperand(1).getReg());
582 EmitToStreamer(O, MCInstBuilder(AArch64::BL).addExpr(Sym));
583 O.AddComment("End XRay custom event");
584 EmitToStreamer(O, MCInstBuilder(AArch64::LDPXpost)
585 .addReg(AArch64::SP)
586 .addReg(AArch64::X0)
587 .addReg(AArch64::X1)
588 .addReg(AArch64::SP)
589 .addImm(2));
590
591 recordSled(CurSled, MI, SledKind::CUSTOM_EVENT, 2);
592 }
593 }
594
LowerKCFI_CHECK(const MachineInstr & MI)595 void AArch64AsmPrinter::LowerKCFI_CHECK(const MachineInstr &MI) {
596 Register AddrReg = MI.getOperand(0).getReg();
597 assert(std::next(MI.getIterator())->isCall() &&
598 "KCFI_CHECK not followed by a call instruction");
599 assert(std::next(MI.getIterator())->getOperand(0).getReg() == AddrReg &&
600 "KCFI_CHECK call target doesn't match call operand");
601
602 // Default to using the intra-procedure-call temporary registers for
603 // comparing the hashes.
604 unsigned ScratchRegs[] = {AArch64::W16, AArch64::W17};
605 if (AddrReg == AArch64::XZR) {
606 // Checking XZR makes no sense. Instead of emitting a load, zero
607 // ScratchRegs[0] and use it for the ESR AddrIndex below.
608 AddrReg = getXRegFromWReg(ScratchRegs[0]);
609 emitMovXReg(AddrReg, AArch64::XZR);
610 } else {
611 // If one of the scratch registers is used for the call target (e.g.
612 // with AArch64::TCRETURNriBTI), we can clobber another caller-saved
613 // temporary register instead (in this case, AArch64::W9) as the check
614 // is immediately followed by the call instruction.
615 for (auto &Reg : ScratchRegs) {
616 if (Reg == getWRegFromXReg(AddrReg)) {
617 Reg = AArch64::W9;
618 break;
619 }
620 }
621 assert(ScratchRegs[0] != AddrReg && ScratchRegs[1] != AddrReg &&
622 "Invalid scratch registers for KCFI_CHECK");
623
624 // Adjust the offset for patchable-function-prefix. This assumes that
625 // patchable-function-prefix is the same for all functions.
626 int64_t PrefixNops = 0;
627 (void)MI.getMF()
628 ->getFunction()
629 .getFnAttribute("patchable-function-prefix")
630 .getValueAsString()
631 .getAsInteger(10, PrefixNops);
632
633 // Load the target function type hash.
634 EmitToStreamer(*OutStreamer, MCInstBuilder(AArch64::LDURWi)
635 .addReg(ScratchRegs[0])
636 .addReg(AddrReg)
637 .addImm(-(PrefixNops * 4 + 4)));
638 }
639
640 // Load the expected type hash.
641 const int64_t Type = MI.getOperand(1).getImm();
642 emitMOVK(ScratchRegs[1], Type & 0xFFFF, 0);
643 emitMOVK(ScratchRegs[1], (Type >> 16) & 0xFFFF, 16);
644
645 // Compare the hashes and trap if there's a mismatch.
646 EmitToStreamer(*OutStreamer, MCInstBuilder(AArch64::SUBSWrs)
647 .addReg(AArch64::WZR)
648 .addReg(ScratchRegs[0])
649 .addReg(ScratchRegs[1])
650 .addImm(0));
651
652 MCSymbol *Pass = OutContext.createTempSymbol();
653 EmitToStreamer(*OutStreamer,
654 MCInstBuilder(AArch64::Bcc)
655 .addImm(AArch64CC::EQ)
656 .addExpr(MCSymbolRefExpr::create(Pass, OutContext)));
657
658 // The base ESR is 0x8000 and the register information is encoded in bits
659 // 0-9 as follows:
660 // - 0-4: n, where the register Xn contains the target address
661 // - 5-9: m, where the register Wm contains the expected type hash
662 // Where n, m are in [0, 30].
663 unsigned TypeIndex = ScratchRegs[1] - AArch64::W0;
664 unsigned AddrIndex;
665 switch (AddrReg) {
666 default:
667 AddrIndex = AddrReg - AArch64::X0;
668 break;
669 case AArch64::FP:
670 AddrIndex = 29;
671 break;
672 case AArch64::LR:
673 AddrIndex = 30;
674 break;
675 }
676
677 assert(AddrIndex < 31 && TypeIndex < 31);
678
679 unsigned ESR = 0x8000 | ((TypeIndex & 31) << 5) | (AddrIndex & 31);
680 EmitToStreamer(*OutStreamer, MCInstBuilder(AArch64::BRK).addImm(ESR));
681 OutStreamer->emitLabel(Pass);
682 }
683
LowerHWASAN_CHECK_MEMACCESS(const MachineInstr & MI)684 void AArch64AsmPrinter::LowerHWASAN_CHECK_MEMACCESS(const MachineInstr &MI) {
685 Register Reg = MI.getOperand(0).getReg();
686
687 // The HWASan pass won't emit a CHECK_MEMACCESS intrinsic with a pointer
688 // statically known to be zero. However, conceivably, the HWASan pass may
689 // encounter a "cannot currently statically prove to be null" pointer (and is
690 // therefore unable to omit the intrinsic) that later optimization passes
691 // convert into a statically known-null pointer.
692 if (Reg == AArch64::XZR)
693 return;
694
695 bool IsShort =
696 ((MI.getOpcode() == AArch64::HWASAN_CHECK_MEMACCESS_SHORTGRANULES) ||
697 (MI.getOpcode() ==
698 AArch64::HWASAN_CHECK_MEMACCESS_SHORTGRANULES_FIXEDSHADOW));
699 uint32_t AccessInfo = MI.getOperand(1).getImm();
700 bool IsFixedShadow =
701 ((MI.getOpcode() == AArch64::HWASAN_CHECK_MEMACCESS_FIXEDSHADOW) ||
702 (MI.getOpcode() ==
703 AArch64::HWASAN_CHECK_MEMACCESS_SHORTGRANULES_FIXEDSHADOW));
704 uint64_t FixedShadowOffset = IsFixedShadow ? MI.getOperand(2).getImm() : 0;
705
706 MCSymbol *&Sym = HwasanMemaccessSymbols[HwasanMemaccessTuple(
707 Reg, IsShort, AccessInfo, IsFixedShadow, FixedShadowOffset)];
708 if (!Sym) {
709 // FIXME: Make this work on non-ELF.
710 if (!TM.getTargetTriple().isOSBinFormatELF())
711 report_fatal_error("llvm.hwasan.check.memaccess only supported on ELF");
712
713 std::string SymName = "__hwasan_check_x" + utostr(Reg - AArch64::X0) + "_" +
714 utostr(AccessInfo);
715 if (IsFixedShadow)
716 SymName += "_fixed_" + utostr(FixedShadowOffset);
717 if (IsShort)
718 SymName += "_short_v2";
719 Sym = OutContext.getOrCreateSymbol(SymName);
720 }
721
722 EmitToStreamer(*OutStreamer,
723 MCInstBuilder(AArch64::BL)
724 .addExpr(MCSymbolRefExpr::create(Sym, OutContext)));
725 }
726
emitHwasanMemaccessSymbols(Module & M)727 void AArch64AsmPrinter::emitHwasanMemaccessSymbols(Module &M) {
728 if (HwasanMemaccessSymbols.empty())
729 return;
730
731 const Triple &TT = TM.getTargetTriple();
732 assert(TT.isOSBinFormatELF());
733 std::unique_ptr<MCSubtargetInfo> STI(
734 TM.getTarget().createMCSubtargetInfo(TT.str(), "", ""));
735 assert(STI && "Unable to create subtarget info");
736 this->STI = static_cast<const AArch64Subtarget *>(&*STI);
737
738 MCSymbol *HwasanTagMismatchV1Sym =
739 OutContext.getOrCreateSymbol("__hwasan_tag_mismatch");
740 MCSymbol *HwasanTagMismatchV2Sym =
741 OutContext.getOrCreateSymbol("__hwasan_tag_mismatch_v2");
742
743 const MCSymbolRefExpr *HwasanTagMismatchV1Ref =
744 MCSymbolRefExpr::create(HwasanTagMismatchV1Sym, OutContext);
745 const MCSymbolRefExpr *HwasanTagMismatchV2Ref =
746 MCSymbolRefExpr::create(HwasanTagMismatchV2Sym, OutContext);
747
748 for (auto &P : HwasanMemaccessSymbols) {
749 unsigned Reg = std::get<0>(P.first);
750 bool IsShort = std::get<1>(P.first);
751 uint32_t AccessInfo = std::get<2>(P.first);
752 bool IsFixedShadow = std::get<3>(P.first);
753 uint64_t FixedShadowOffset = std::get<4>(P.first);
754 const MCSymbolRefExpr *HwasanTagMismatchRef =
755 IsShort ? HwasanTagMismatchV2Ref : HwasanTagMismatchV1Ref;
756 MCSymbol *Sym = P.second;
757
758 bool HasMatchAllTag =
759 (AccessInfo >> HWASanAccessInfo::HasMatchAllShift) & 1;
760 uint8_t MatchAllTag =
761 (AccessInfo >> HWASanAccessInfo::MatchAllShift) & 0xff;
762 unsigned Size =
763 1 << ((AccessInfo >> HWASanAccessInfo::AccessSizeShift) & 0xf);
764 bool CompileKernel =
765 (AccessInfo >> HWASanAccessInfo::CompileKernelShift) & 1;
766
767 OutStreamer->switchSection(OutContext.getELFSection(
768 ".text.hot", ELF::SHT_PROGBITS,
769 ELF::SHF_EXECINSTR | ELF::SHF_ALLOC | ELF::SHF_GROUP, 0, Sym->getName(),
770 /*IsComdat=*/true));
771
772 OutStreamer->emitSymbolAttribute(Sym, MCSA_ELF_TypeFunction);
773 OutStreamer->emitSymbolAttribute(Sym, MCSA_Weak);
774 OutStreamer->emitSymbolAttribute(Sym, MCSA_Hidden);
775 OutStreamer->emitLabel(Sym);
776
777 EmitToStreamer(MCInstBuilder(AArch64::SBFMXri)
778 .addReg(AArch64::X16)
779 .addReg(Reg)
780 .addImm(4)
781 .addImm(55));
782
783 if (IsFixedShadow) {
784 // Aarch64 makes it difficult to embed large constants in the code.
785 // Fortuitously, kShadowBaseAlignment == 32, so we use the 32-bit
786 // left-shift option in the MOV instruction. Combined with the 16-bit
787 // immediate, this is enough to represent any offset up to 2**48.
788 emitMOVZ(AArch64::X17, FixedShadowOffset >> 32, 32);
789 EmitToStreamer(MCInstBuilder(AArch64::LDRBBroX)
790 .addReg(AArch64::W16)
791 .addReg(AArch64::X17)
792 .addReg(AArch64::X16)
793 .addImm(0)
794 .addImm(0));
795 } else {
796 EmitToStreamer(MCInstBuilder(AArch64::LDRBBroX)
797 .addReg(AArch64::W16)
798 .addReg(IsShort ? AArch64::X20 : AArch64::X9)
799 .addReg(AArch64::X16)
800 .addImm(0)
801 .addImm(0));
802 }
803
804 EmitToStreamer(MCInstBuilder(AArch64::SUBSXrs)
805 .addReg(AArch64::XZR)
806 .addReg(AArch64::X16)
807 .addReg(Reg)
808 .addImm(AArch64_AM::getShifterImm(AArch64_AM::LSR, 56)));
809 MCSymbol *HandleMismatchOrPartialSym = OutContext.createTempSymbol();
810 EmitToStreamer(MCInstBuilder(AArch64::Bcc)
811 .addImm(AArch64CC::NE)
812 .addExpr(MCSymbolRefExpr::create(
813 HandleMismatchOrPartialSym, OutContext)));
814 MCSymbol *ReturnSym = OutContext.createTempSymbol();
815 OutStreamer->emitLabel(ReturnSym);
816 EmitToStreamer(MCInstBuilder(AArch64::RET).addReg(AArch64::LR));
817 OutStreamer->emitLabel(HandleMismatchOrPartialSym);
818
819 if (HasMatchAllTag) {
820 EmitToStreamer(MCInstBuilder(AArch64::UBFMXri)
821 .addReg(AArch64::X17)
822 .addReg(Reg)
823 .addImm(56)
824 .addImm(63));
825 EmitToStreamer(MCInstBuilder(AArch64::SUBSXri)
826 .addReg(AArch64::XZR)
827 .addReg(AArch64::X17)
828 .addImm(MatchAllTag)
829 .addImm(0));
830 EmitToStreamer(
831 MCInstBuilder(AArch64::Bcc)
832 .addImm(AArch64CC::EQ)
833 .addExpr(MCSymbolRefExpr::create(ReturnSym, OutContext)));
834 }
835
836 if (IsShort) {
837 EmitToStreamer(MCInstBuilder(AArch64::SUBSWri)
838 .addReg(AArch64::WZR)
839 .addReg(AArch64::W16)
840 .addImm(15)
841 .addImm(0));
842 MCSymbol *HandleMismatchSym = OutContext.createTempSymbol();
843 EmitToStreamer(
844 MCInstBuilder(AArch64::Bcc)
845 .addImm(AArch64CC::HI)
846 .addExpr(MCSymbolRefExpr::create(HandleMismatchSym, OutContext)));
847
848 EmitToStreamer(MCInstBuilder(AArch64::ANDXri)
849 .addReg(AArch64::X17)
850 .addReg(Reg)
851 .addImm(AArch64_AM::encodeLogicalImmediate(0xf, 64)));
852 if (Size != 1)
853 EmitToStreamer(MCInstBuilder(AArch64::ADDXri)
854 .addReg(AArch64::X17)
855 .addReg(AArch64::X17)
856 .addImm(Size - 1)
857 .addImm(0));
858 EmitToStreamer(MCInstBuilder(AArch64::SUBSWrs)
859 .addReg(AArch64::WZR)
860 .addReg(AArch64::W16)
861 .addReg(AArch64::W17)
862 .addImm(0));
863 EmitToStreamer(
864 MCInstBuilder(AArch64::Bcc)
865 .addImm(AArch64CC::LS)
866 .addExpr(MCSymbolRefExpr::create(HandleMismatchSym, OutContext)));
867
868 EmitToStreamer(MCInstBuilder(AArch64::ORRXri)
869 .addReg(AArch64::X16)
870 .addReg(Reg)
871 .addImm(AArch64_AM::encodeLogicalImmediate(0xf, 64)));
872 EmitToStreamer(MCInstBuilder(AArch64::LDRBBui)
873 .addReg(AArch64::W16)
874 .addReg(AArch64::X16)
875 .addImm(0));
876 EmitToStreamer(
877 MCInstBuilder(AArch64::SUBSXrs)
878 .addReg(AArch64::XZR)
879 .addReg(AArch64::X16)
880 .addReg(Reg)
881 .addImm(AArch64_AM::getShifterImm(AArch64_AM::LSR, 56)));
882 EmitToStreamer(
883 MCInstBuilder(AArch64::Bcc)
884 .addImm(AArch64CC::EQ)
885 .addExpr(MCSymbolRefExpr::create(ReturnSym, OutContext)));
886
887 OutStreamer->emitLabel(HandleMismatchSym);
888 }
889
890 EmitToStreamer(MCInstBuilder(AArch64::STPXpre)
891 .addReg(AArch64::SP)
892 .addReg(AArch64::X0)
893 .addReg(AArch64::X1)
894 .addReg(AArch64::SP)
895 .addImm(-32));
896 EmitToStreamer(MCInstBuilder(AArch64::STPXi)
897 .addReg(AArch64::FP)
898 .addReg(AArch64::LR)
899 .addReg(AArch64::SP)
900 .addImm(29));
901
902 if (Reg != AArch64::X0)
903 emitMovXReg(AArch64::X0, Reg);
904 emitMOVZ(AArch64::X1, AccessInfo & HWASanAccessInfo::RuntimeMask, 0);
905
906 if (CompileKernel) {
907 // The Linux kernel's dynamic loader doesn't support GOT relative
908 // relocations, but it doesn't support late binding either, so just call
909 // the function directly.
910 EmitToStreamer(MCInstBuilder(AArch64::B).addExpr(HwasanTagMismatchRef));
911 } else {
912 // Intentionally load the GOT entry and branch to it, rather than possibly
913 // late binding the function, which may clobber the registers before we
914 // have a chance to save them.
915 EmitToStreamer(MCInstBuilder(AArch64::ADRP)
916 .addReg(AArch64::X16)
917 .addExpr(MCSpecifierExpr::create(HwasanTagMismatchRef,
918 AArch64::S_GOT_PAGE,
919 OutContext)));
920 EmitToStreamer(MCInstBuilder(AArch64::LDRXui)
921 .addReg(AArch64::X16)
922 .addReg(AArch64::X16)
923 .addExpr(MCSpecifierExpr::create(HwasanTagMismatchRef,
924 AArch64::S_GOT_LO12,
925 OutContext)));
926 EmitToStreamer(MCInstBuilder(AArch64::BR).addReg(AArch64::X16));
927 }
928 }
929 this->STI = nullptr;
930 }
931
emitAuthenticatedPointer(MCStreamer & OutStreamer,MCSymbol * StubLabel,const MCExpr * StubAuthPtrRef)932 static void emitAuthenticatedPointer(MCStreamer &OutStreamer,
933 MCSymbol *StubLabel,
934 const MCExpr *StubAuthPtrRef) {
935 // sym$auth_ptr$key$disc:
936 OutStreamer.emitLabel(StubLabel);
937 OutStreamer.emitValue(StubAuthPtrRef, /*size=*/8);
938 }
939
emitEndOfAsmFile(Module & M)940 void AArch64AsmPrinter::emitEndOfAsmFile(Module &M) {
941 emitHwasanMemaccessSymbols(M);
942
943 const Triple &TT = TM.getTargetTriple();
944 if (TT.isOSBinFormatMachO()) {
945 // Output authenticated pointers as indirect symbols, if we have any.
946 MachineModuleInfoMachO &MMIMacho =
947 MMI->getObjFileInfo<MachineModuleInfoMachO>();
948
949 auto Stubs = MMIMacho.getAuthGVStubList();
950
951 if (!Stubs.empty()) {
952 // Switch to the "__auth_ptr" section.
953 OutStreamer->switchSection(
954 OutContext.getMachOSection("__DATA", "__auth_ptr", MachO::S_REGULAR,
955 SectionKind::getMetadata()));
956 emitAlignment(Align(8));
957
958 for (const auto &Stub : Stubs)
959 emitAuthenticatedPointer(*OutStreamer, Stub.first, Stub.second);
960
961 OutStreamer->addBlankLine();
962 }
963
964 // Funny Darwin hack: This flag tells the linker that no global symbols
965 // contain code that falls through to other global symbols (e.g. the obvious
966 // implementation of multiple entry points). If this doesn't occur, the
967 // linker can safely perform dead code stripping. Since LLVM never
968 // generates code that does this, it is always safe to set.
969 OutStreamer->emitSubsectionsViaSymbols();
970 }
971
972 if (TT.isOSBinFormatELF()) {
973 // Output authenticated pointers as indirect symbols, if we have any.
974 MachineModuleInfoELF &MMIELF = MMI->getObjFileInfo<MachineModuleInfoELF>();
975
976 auto Stubs = MMIELF.getAuthGVStubList();
977
978 if (!Stubs.empty()) {
979 const TargetLoweringObjectFile &TLOF = getObjFileLowering();
980 OutStreamer->switchSection(TLOF.getDataSection());
981 emitAlignment(Align(8));
982
983 for (const auto &Stub : Stubs)
984 emitAuthenticatedPointer(*OutStreamer, Stub.first, Stub.second);
985
986 OutStreamer->addBlankLine();
987 }
988
989 // With signed ELF GOT enabled, the linker looks at the symbol type to
990 // choose between keys IA (for STT_FUNC) and DA (for other types). Symbols
991 // for functions not defined in the module have STT_NOTYPE type by default.
992 // This makes linker to emit signing schema with DA key (instead of IA) for
993 // corresponding R_AARCH64_AUTH_GLOB_DAT dynamic reloc. To avoid that, force
994 // all function symbols used in the module to have STT_FUNC type. See
995 // https://github.com/ARM-software/abi-aa/blob/main/pauthabielf64/pauthabielf64.rst#default-signing-schema
996 const auto *PtrAuthELFGOTFlag = mdconst::extract_or_null<ConstantInt>(
997 M.getModuleFlag("ptrauth-elf-got"));
998 if (PtrAuthELFGOTFlag && PtrAuthELFGOTFlag->getZExtValue() == 1)
999 for (const GlobalValue &GV : M.global_values())
1000 if (!GV.use_empty() && isa<Function>(GV) &&
1001 !GV.getName().starts_with("llvm."))
1002 OutStreamer->emitSymbolAttribute(getSymbol(&GV),
1003 MCSA_ELF_TypeFunction);
1004 }
1005
1006 // Emit stack and fault map information.
1007 FM.serializeToFaultMapSection();
1008
1009 // If import call optimization is enabled, emit the appropriate section.
1010 // We do this whether or not we recorded any import calls.
1011 if (EnableImportCallOptimization && TT.isOSBinFormatCOFF()) {
1012 OutStreamer->switchSection(getObjFileLowering().getImportCallSection());
1013
1014 // Section always starts with some magic.
1015 constexpr char ImpCallMagic[12] = "Imp_Call_V1";
1016 OutStreamer->emitBytes(StringRef{ImpCallMagic, sizeof(ImpCallMagic)});
1017
1018 // Layout of this section is:
1019 // Per section that contains calls to imported functions:
1020 // uint32_t SectionSize: Size in bytes for information in this section.
1021 // uint32_t Section Number
1022 // Per call to imported function in section:
1023 // uint32_t Kind: the kind of imported function.
1024 // uint32_t BranchOffset: the offset of the branch instruction in its
1025 // parent section.
1026 // uint32_t TargetSymbolId: the symbol id of the called function.
1027 for (auto &[Section, CallsToImportedFuncs] :
1028 SectionToImportedFunctionCalls) {
1029 unsigned SectionSize =
1030 sizeof(uint32_t) * (2 + 3 * CallsToImportedFuncs.size());
1031 OutStreamer->emitInt32(SectionSize);
1032 OutStreamer->emitCOFFSecNumber(Section->getBeginSymbol());
1033 for (auto &[CallsiteSymbol, CalledSymbol] : CallsToImportedFuncs) {
1034 // Kind is always IMAGE_REL_ARM64_DYNAMIC_IMPORT_CALL (0x13).
1035 OutStreamer->emitInt32(0x13);
1036 OutStreamer->emitCOFFSecOffset(CallsiteSymbol);
1037 OutStreamer->emitCOFFSymbolIndex(CalledSymbol);
1038 }
1039 }
1040 }
1041 }
1042
emitLOHs()1043 void AArch64AsmPrinter::emitLOHs() {
1044 SmallVector<MCSymbol *, 3> MCArgs;
1045
1046 for (const auto &D : AArch64FI->getLOHContainer()) {
1047 for (const MachineInstr *MI : D.getArgs()) {
1048 MInstToMCSymbol::iterator LabelIt = LOHInstToLabel.find(MI);
1049 assert(LabelIt != LOHInstToLabel.end() &&
1050 "Label hasn't been inserted for LOH related instruction");
1051 MCArgs.push_back(LabelIt->second);
1052 }
1053 OutStreamer->emitLOHDirective(D.getKind(), MCArgs);
1054 MCArgs.clear();
1055 }
1056 }
1057
emitFunctionBodyEnd()1058 void AArch64AsmPrinter::emitFunctionBodyEnd() {
1059 if (!AArch64FI->getLOHRelated().empty())
1060 emitLOHs();
1061 }
1062
1063 /// GetCPISymbol - Return the symbol for the specified constant pool entry.
GetCPISymbol(unsigned CPID) const1064 MCSymbol *AArch64AsmPrinter::GetCPISymbol(unsigned CPID) const {
1065 // Darwin uses a linker-private symbol name for constant-pools (to
1066 // avoid addends on the relocation?), ELF has no such concept and
1067 // uses a normal private symbol.
1068 if (!getDataLayout().getLinkerPrivateGlobalPrefix().empty())
1069 return OutContext.getOrCreateSymbol(
1070 Twine(getDataLayout().getLinkerPrivateGlobalPrefix()) + "CPI" +
1071 Twine(getFunctionNumber()) + "_" + Twine(CPID));
1072
1073 return AsmPrinter::GetCPISymbol(CPID);
1074 }
1075
printOperand(const MachineInstr * MI,unsigned OpNum,raw_ostream & O)1076 void AArch64AsmPrinter::printOperand(const MachineInstr *MI, unsigned OpNum,
1077 raw_ostream &O) {
1078 const MachineOperand &MO = MI->getOperand(OpNum);
1079 switch (MO.getType()) {
1080 default:
1081 llvm_unreachable("<unknown operand type>");
1082 case MachineOperand::MO_Register: {
1083 Register Reg = MO.getReg();
1084 assert(Reg.isPhysical());
1085 assert(!MO.getSubReg() && "Subregs should be eliminated!");
1086 O << AArch64InstPrinter::getRegisterName(Reg);
1087 break;
1088 }
1089 case MachineOperand::MO_Immediate: {
1090 O << MO.getImm();
1091 break;
1092 }
1093 case MachineOperand::MO_GlobalAddress: {
1094 PrintSymbolOperand(MO, O);
1095 break;
1096 }
1097 case MachineOperand::MO_BlockAddress: {
1098 MCSymbol *Sym = GetBlockAddressSymbol(MO.getBlockAddress());
1099 Sym->print(O, MAI);
1100 break;
1101 }
1102 }
1103 }
1104
printAsmMRegister(const MachineOperand & MO,char Mode,raw_ostream & O)1105 bool AArch64AsmPrinter::printAsmMRegister(const MachineOperand &MO, char Mode,
1106 raw_ostream &O) {
1107 Register Reg = MO.getReg();
1108 switch (Mode) {
1109 default:
1110 return true; // Unknown mode.
1111 case 'w':
1112 Reg = getWRegFromXReg(Reg);
1113 break;
1114 case 'x':
1115 Reg = getXRegFromWReg(Reg);
1116 break;
1117 case 't':
1118 Reg = getXRegFromXRegTuple(Reg);
1119 break;
1120 }
1121
1122 O << AArch64InstPrinter::getRegisterName(Reg);
1123 return false;
1124 }
1125
1126 // Prints the register in MO using class RC using the offset in the
1127 // new register class. This should not be used for cross class
1128 // printing.
printAsmRegInClass(const MachineOperand & MO,const TargetRegisterClass * RC,unsigned AltName,raw_ostream & O)1129 bool AArch64AsmPrinter::printAsmRegInClass(const MachineOperand &MO,
1130 const TargetRegisterClass *RC,
1131 unsigned AltName, raw_ostream &O) {
1132 assert(MO.isReg() && "Should only get here with a register!");
1133 const TargetRegisterInfo *RI = STI->getRegisterInfo();
1134 Register Reg = MO.getReg();
1135 MCRegister RegToPrint = RC->getRegister(RI->getEncodingValue(Reg));
1136 if (!RI->regsOverlap(RegToPrint, Reg))
1137 return true;
1138 O << AArch64InstPrinter::getRegisterName(RegToPrint, AltName);
1139 return false;
1140 }
1141
PrintAsmOperand(const MachineInstr * MI,unsigned OpNum,const char * ExtraCode,raw_ostream & O)1142 bool AArch64AsmPrinter::PrintAsmOperand(const MachineInstr *MI, unsigned OpNum,
1143 const char *ExtraCode, raw_ostream &O) {
1144 const MachineOperand &MO = MI->getOperand(OpNum);
1145
1146 // First try the generic code, which knows about modifiers like 'c' and 'n'.
1147 if (!AsmPrinter::PrintAsmOperand(MI, OpNum, ExtraCode, O))
1148 return false;
1149
1150 // Does this asm operand have a single letter operand modifier?
1151 if (ExtraCode && ExtraCode[0]) {
1152 if (ExtraCode[1] != 0)
1153 return true; // Unknown modifier.
1154
1155 switch (ExtraCode[0]) {
1156 default:
1157 return true; // Unknown modifier.
1158 case 'w': // Print W register
1159 case 'x': // Print X register
1160 if (MO.isReg())
1161 return printAsmMRegister(MO, ExtraCode[0], O);
1162 if (MO.isImm() && MO.getImm() == 0) {
1163 unsigned Reg = ExtraCode[0] == 'w' ? AArch64::WZR : AArch64::XZR;
1164 O << AArch64InstPrinter::getRegisterName(Reg);
1165 return false;
1166 }
1167 printOperand(MI, OpNum, O);
1168 return false;
1169 case 'b': // Print B register.
1170 case 'h': // Print H register.
1171 case 's': // Print S register.
1172 case 'd': // Print D register.
1173 case 'q': // Print Q register.
1174 case 'z': // Print Z register.
1175 if (MO.isReg()) {
1176 const TargetRegisterClass *RC;
1177 switch (ExtraCode[0]) {
1178 case 'b':
1179 RC = &AArch64::FPR8RegClass;
1180 break;
1181 case 'h':
1182 RC = &AArch64::FPR16RegClass;
1183 break;
1184 case 's':
1185 RC = &AArch64::FPR32RegClass;
1186 break;
1187 case 'd':
1188 RC = &AArch64::FPR64RegClass;
1189 break;
1190 case 'q':
1191 RC = &AArch64::FPR128RegClass;
1192 break;
1193 case 'z':
1194 RC = &AArch64::ZPRRegClass;
1195 break;
1196 default:
1197 return true;
1198 }
1199 return printAsmRegInClass(MO, RC, AArch64::NoRegAltName, O);
1200 }
1201 printOperand(MI, OpNum, O);
1202 return false;
1203 }
1204 }
1205
1206 // According to ARM, we should emit x and v registers unless we have a
1207 // modifier.
1208 if (MO.isReg()) {
1209 Register Reg = MO.getReg();
1210
1211 // If this is a w or x register, print an x register.
1212 if (AArch64::GPR32allRegClass.contains(Reg) ||
1213 AArch64::GPR64allRegClass.contains(Reg))
1214 return printAsmMRegister(MO, 'x', O);
1215
1216 // If this is an x register tuple, print an x register.
1217 if (AArch64::GPR64x8ClassRegClass.contains(Reg))
1218 return printAsmMRegister(MO, 't', O);
1219
1220 unsigned AltName = AArch64::NoRegAltName;
1221 const TargetRegisterClass *RegClass;
1222 if (AArch64::ZPRRegClass.contains(Reg)) {
1223 RegClass = &AArch64::ZPRRegClass;
1224 } else if (AArch64::PPRRegClass.contains(Reg)) {
1225 RegClass = &AArch64::PPRRegClass;
1226 } else if (AArch64::PNRRegClass.contains(Reg)) {
1227 RegClass = &AArch64::PNRRegClass;
1228 } else {
1229 RegClass = &AArch64::FPR128RegClass;
1230 AltName = AArch64::vreg;
1231 }
1232
1233 // If this is a b, h, s, d, or q register, print it as a v register.
1234 return printAsmRegInClass(MO, RegClass, AltName, O);
1235 }
1236
1237 printOperand(MI, OpNum, O);
1238 return false;
1239 }
1240
PrintAsmMemoryOperand(const MachineInstr * MI,unsigned OpNum,const char * ExtraCode,raw_ostream & O)1241 bool AArch64AsmPrinter::PrintAsmMemoryOperand(const MachineInstr *MI,
1242 unsigned OpNum,
1243 const char *ExtraCode,
1244 raw_ostream &O) {
1245 if (ExtraCode && ExtraCode[0] && ExtraCode[0] != 'a')
1246 return true; // Unknown modifier.
1247
1248 const MachineOperand &MO = MI->getOperand(OpNum);
1249 assert(MO.isReg() && "unexpected inline asm memory operand");
1250 O << "[" << AArch64InstPrinter::getRegisterName(MO.getReg()) << "]";
1251 return false;
1252 }
1253
PrintDebugValueComment(const MachineInstr * MI,raw_ostream & OS)1254 void AArch64AsmPrinter::PrintDebugValueComment(const MachineInstr *MI,
1255 raw_ostream &OS) {
1256 unsigned NOps = MI->getNumOperands();
1257 assert(NOps == 4);
1258 OS << '\t' << MAI->getCommentString() << "DEBUG_VALUE: ";
1259 // cast away const; DIetc do not take const operands for some reason.
1260 OS << MI->getDebugVariable()->getName();
1261 OS << " <- ";
1262 // Frame address. Currently handles register +- offset only.
1263 assert(MI->isIndirectDebugValue());
1264 OS << '[';
1265 for (unsigned I = 0, E = std::distance(MI->debug_operands().begin(),
1266 MI->debug_operands().end());
1267 I < E; ++I) {
1268 if (I != 0)
1269 OS << ", ";
1270 printOperand(MI, I, OS);
1271 }
1272 OS << ']';
1273 OS << "+";
1274 printOperand(MI, NOps - 2, OS);
1275 }
1276
emitJumpTableImpl(const MachineJumpTableInfo & MJTI,ArrayRef<unsigned> JumpTableIndices)1277 void AArch64AsmPrinter::emitJumpTableImpl(const MachineJumpTableInfo &MJTI,
1278 ArrayRef<unsigned> JumpTableIndices) {
1279 // Fast return if there is nothing to emit to avoid creating empty sections.
1280 if (JumpTableIndices.empty())
1281 return;
1282 const TargetLoweringObjectFile &TLOF = getObjFileLowering();
1283 const auto &F = MF->getFunction();
1284 ArrayRef<MachineJumpTableEntry> JT = MJTI.getJumpTables();
1285
1286 MCSection *ReadOnlySec = nullptr;
1287 if (TM.Options.EnableStaticDataPartitioning) {
1288 ReadOnlySec =
1289 TLOF.getSectionForJumpTable(F, TM, &JT[JumpTableIndices.front()]);
1290 } else {
1291 ReadOnlySec = TLOF.getSectionForJumpTable(F, TM);
1292 }
1293 OutStreamer->switchSection(ReadOnlySec);
1294
1295 auto AFI = MF->getInfo<AArch64FunctionInfo>();
1296 for (unsigned JTI : JumpTableIndices) {
1297 const std::vector<MachineBasicBlock*> &JTBBs = JT[JTI].MBBs;
1298
1299 // If this jump table was deleted, ignore it.
1300 if (JTBBs.empty()) continue;
1301
1302 unsigned Size = AFI->getJumpTableEntrySize(JTI);
1303 emitAlignment(Align(Size));
1304 OutStreamer->emitLabel(GetJTISymbol(JTI));
1305
1306 const MCSymbol *BaseSym = AArch64FI->getJumpTableEntryPCRelSymbol(JTI);
1307 const MCExpr *Base = MCSymbolRefExpr::create(BaseSym, OutContext);
1308
1309 for (auto *JTBB : JTBBs) {
1310 const MCExpr *Value =
1311 MCSymbolRefExpr::create(JTBB->getSymbol(), OutContext);
1312
1313 // Each entry is:
1314 // .byte/.hword (LBB - Lbase)>>2
1315 // or plain:
1316 // .word LBB - Lbase
1317 Value = MCBinaryExpr::createSub(Value, Base, OutContext);
1318 if (Size != 4)
1319 Value = MCBinaryExpr::createLShr(
1320 Value, MCConstantExpr::create(2, OutContext), OutContext);
1321
1322 OutStreamer->emitValue(Value, Size);
1323 }
1324 }
1325 }
1326
1327 std::tuple<const MCSymbol *, uint64_t, const MCSymbol *,
1328 codeview::JumpTableEntrySize>
getCodeViewJumpTableInfo(int JTI,const MachineInstr * BranchInstr,const MCSymbol * BranchLabel) const1329 AArch64AsmPrinter::getCodeViewJumpTableInfo(int JTI,
1330 const MachineInstr *BranchInstr,
1331 const MCSymbol *BranchLabel) const {
1332 const auto AFI = MF->getInfo<AArch64FunctionInfo>();
1333 const auto Base = AArch64FI->getJumpTableEntryPCRelSymbol(JTI);
1334 codeview::JumpTableEntrySize EntrySize;
1335 switch (AFI->getJumpTableEntrySize(JTI)) {
1336 case 1:
1337 EntrySize = codeview::JumpTableEntrySize::UInt8ShiftLeft;
1338 break;
1339 case 2:
1340 EntrySize = codeview::JumpTableEntrySize::UInt16ShiftLeft;
1341 break;
1342 case 4:
1343 EntrySize = codeview::JumpTableEntrySize::Int32;
1344 break;
1345 default:
1346 llvm_unreachable("Unexpected jump table entry size");
1347 }
1348 return std::make_tuple(Base, 0, BranchLabel, EntrySize);
1349 }
1350
emitFunctionEntryLabel()1351 void AArch64AsmPrinter::emitFunctionEntryLabel() {
1352 const Triple &TT = TM.getTargetTriple();
1353 if (TT.isOSBinFormatELF() &&
1354 (MF->getFunction().getCallingConv() == CallingConv::AArch64_VectorCall ||
1355 MF->getFunction().getCallingConv() ==
1356 CallingConv::AArch64_SVE_VectorCall ||
1357 MF->getInfo<AArch64FunctionInfo>()->isSVECC())) {
1358 auto *TS =
1359 static_cast<AArch64TargetStreamer *>(OutStreamer->getTargetStreamer());
1360 TS->emitDirectiveVariantPCS(CurrentFnSym);
1361 }
1362
1363 AsmPrinter::emitFunctionEntryLabel();
1364
1365 if (TT.isWindowsArm64EC() && !MF->getFunction().hasLocalLinkage()) {
1366 // For ARM64EC targets, a function definition's name is mangled differently
1367 // from the normal symbol, emit required aliases here.
1368 auto emitFunctionAlias = [&](MCSymbol *Src, MCSymbol *Dst) {
1369 OutStreamer->emitSymbolAttribute(Src, MCSA_WeakAntiDep);
1370 OutStreamer->emitAssignment(
1371 Src, MCSymbolRefExpr::create(Dst, MMI->getContext()));
1372 };
1373
1374 auto getSymbolFromMetadata = [&](StringRef Name) {
1375 MCSymbol *Sym = nullptr;
1376 if (MDNode *Node = MF->getFunction().getMetadata(Name)) {
1377 StringRef NameStr = cast<MDString>(Node->getOperand(0))->getString();
1378 Sym = MMI->getContext().getOrCreateSymbol(NameStr);
1379 }
1380 return Sym;
1381 };
1382
1383 SmallVector<MDNode *> UnmangledNames;
1384 MF->getFunction().getMetadata("arm64ec_unmangled_name", UnmangledNames);
1385 for (MDNode *Node : UnmangledNames) {
1386 StringRef NameStr = cast<MDString>(Node->getOperand(0))->getString();
1387 MCSymbol *UnmangledSym = MMI->getContext().getOrCreateSymbol(NameStr);
1388 if (std::optional<std::string> MangledName =
1389 getArm64ECMangledFunctionName(UnmangledSym->getName())) {
1390 MCSymbol *ECMangledSym =
1391 MMI->getContext().getOrCreateSymbol(*MangledName);
1392 emitFunctionAlias(UnmangledSym, ECMangledSym);
1393 }
1394 }
1395 if (MCSymbol *ECMangledSym =
1396 getSymbolFromMetadata("arm64ec_ecmangled_name"))
1397 emitFunctionAlias(ECMangledSym, CurrentFnSym);
1398 }
1399 }
1400
emitXXStructor(const DataLayout & DL,const Constant * CV)1401 void AArch64AsmPrinter::emitXXStructor(const DataLayout &DL,
1402 const Constant *CV) {
1403 if (const auto *CPA = dyn_cast<ConstantPtrAuth>(CV))
1404 if (CPA->hasAddressDiscriminator() &&
1405 !CPA->hasSpecialAddressDiscriminator(
1406 ConstantPtrAuth::AddrDiscriminator_CtorsDtors))
1407 report_fatal_error(
1408 "unexpected address discrimination value for ctors/dtors entry, only "
1409 "'ptr inttoptr (i64 1 to ptr)' is allowed");
1410 // If we have signed pointers in xxstructors list, they'll be lowered to @AUTH
1411 // MCExpr's via AArch64AsmPrinter::lowerConstantPtrAuth. It does not look at
1412 // actual address discrimination value and only checks
1413 // hasAddressDiscriminator(), so it's OK to leave special address
1414 // discrimination value here.
1415 AsmPrinter::emitXXStructor(DL, CV);
1416 }
1417
emitGlobalAlias(const Module & M,const GlobalAlias & GA)1418 void AArch64AsmPrinter::emitGlobalAlias(const Module &M,
1419 const GlobalAlias &GA) {
1420 if (auto F = dyn_cast_or_null<Function>(GA.getAliasee())) {
1421 // Global aliases must point to a definition, but unmangled patchable
1422 // symbols are special and need to point to an undefined symbol with "EXP+"
1423 // prefix. Such undefined symbol is resolved by the linker by creating
1424 // x86 thunk that jumps back to the actual EC target.
1425 if (MDNode *Node = F->getMetadata("arm64ec_exp_name")) {
1426 StringRef ExpStr = cast<MDString>(Node->getOperand(0))->getString();
1427 MCSymbol *ExpSym = MMI->getContext().getOrCreateSymbol(ExpStr);
1428 MCSymbol *Sym = MMI->getContext().getOrCreateSymbol(GA.getName());
1429
1430 OutStreamer->beginCOFFSymbolDef(ExpSym);
1431 OutStreamer->emitCOFFSymbolStorageClass(COFF::IMAGE_SYM_CLASS_EXTERNAL);
1432 OutStreamer->emitCOFFSymbolType(COFF::IMAGE_SYM_DTYPE_FUNCTION
1433 << COFF::SCT_COMPLEX_TYPE_SHIFT);
1434 OutStreamer->endCOFFSymbolDef();
1435
1436 OutStreamer->beginCOFFSymbolDef(Sym);
1437 OutStreamer->emitCOFFSymbolStorageClass(COFF::IMAGE_SYM_CLASS_EXTERNAL);
1438 OutStreamer->emitCOFFSymbolType(COFF::IMAGE_SYM_DTYPE_FUNCTION
1439 << COFF::SCT_COMPLEX_TYPE_SHIFT);
1440 OutStreamer->endCOFFSymbolDef();
1441 OutStreamer->emitSymbolAttribute(Sym, MCSA_Weak);
1442 OutStreamer->emitAssignment(
1443 Sym, MCSymbolRefExpr::create(ExpSym, MMI->getContext()));
1444 return;
1445 }
1446 }
1447 AsmPrinter::emitGlobalAlias(M, GA);
1448 }
1449
1450 /// Small jump tables contain an unsigned byte or half, representing the offset
1451 /// from the lowest-addressed possible destination to the desired basic
1452 /// block. Since all instructions are 4-byte aligned, this is further compressed
1453 /// by counting in instructions rather than bytes (i.e. divided by 4). So, to
1454 /// materialize the correct destination we need:
1455 ///
1456 /// adr xDest, .LBB0_0
1457 /// ldrb wScratch, [xTable, xEntry] (with "lsl #1" for ldrh).
1458 /// add xDest, xDest, xScratch (with "lsl #2" for smaller entries)
LowerJumpTableDest(llvm::MCStreamer & OutStreamer,const llvm::MachineInstr & MI)1459 void AArch64AsmPrinter::LowerJumpTableDest(llvm::MCStreamer &OutStreamer,
1460 const llvm::MachineInstr &MI) {
1461 Register DestReg = MI.getOperand(0).getReg();
1462 Register ScratchReg = MI.getOperand(1).getReg();
1463 Register ScratchRegW =
1464 STI->getRegisterInfo()->getSubReg(ScratchReg, AArch64::sub_32);
1465 Register TableReg = MI.getOperand(2).getReg();
1466 Register EntryReg = MI.getOperand(3).getReg();
1467 int JTIdx = MI.getOperand(4).getIndex();
1468 int Size = AArch64FI->getJumpTableEntrySize(JTIdx);
1469
1470 // This has to be first because the compression pass based its reachability
1471 // calculations on the start of the JumpTableDest instruction.
1472 auto Label =
1473 MF->getInfo<AArch64FunctionInfo>()->getJumpTableEntryPCRelSymbol(JTIdx);
1474
1475 // If we don't already have a symbol to use as the base, use the ADR
1476 // instruction itself.
1477 if (!Label) {
1478 Label = MF->getContext().createTempSymbol();
1479 AArch64FI->setJumpTableEntryInfo(JTIdx, Size, Label);
1480 OutStreamer.emitLabel(Label);
1481 }
1482
1483 auto LabelExpr = MCSymbolRefExpr::create(Label, MF->getContext());
1484 EmitToStreamer(OutStreamer, MCInstBuilder(AArch64::ADR)
1485 .addReg(DestReg)
1486 .addExpr(LabelExpr));
1487
1488 // Load the number of instruction-steps to offset from the label.
1489 unsigned LdrOpcode;
1490 switch (Size) {
1491 case 1: LdrOpcode = AArch64::LDRBBroX; break;
1492 case 2: LdrOpcode = AArch64::LDRHHroX; break;
1493 case 4: LdrOpcode = AArch64::LDRSWroX; break;
1494 default:
1495 llvm_unreachable("Unknown jump table size");
1496 }
1497
1498 EmitToStreamer(OutStreamer, MCInstBuilder(LdrOpcode)
1499 .addReg(Size == 4 ? ScratchReg : ScratchRegW)
1500 .addReg(TableReg)
1501 .addReg(EntryReg)
1502 .addImm(0)
1503 .addImm(Size == 1 ? 0 : 1));
1504
1505 // Add to the already materialized base label address, multiplying by 4 if
1506 // compressed.
1507 EmitToStreamer(OutStreamer, MCInstBuilder(AArch64::ADDXrs)
1508 .addReg(DestReg)
1509 .addReg(DestReg)
1510 .addReg(ScratchReg)
1511 .addImm(Size == 4 ? 0 : 2));
1512 }
1513
LowerHardenedBRJumpTable(const MachineInstr & MI)1514 void AArch64AsmPrinter::LowerHardenedBRJumpTable(const MachineInstr &MI) {
1515 const MachineJumpTableInfo *MJTI = MF->getJumpTableInfo();
1516 assert(MJTI && "Can't lower jump-table dispatch without JTI");
1517
1518 const std::vector<MachineJumpTableEntry> &JTs = MJTI->getJumpTables();
1519 assert(!JTs.empty() && "Invalid JT index for jump-table dispatch");
1520
1521 // Emit:
1522 // mov x17, #<size of table> ; depending on table size, with MOVKs
1523 // cmp x16, x17 ; or #imm if table size fits in 12-bit
1524 // csel x16, x16, xzr, ls ; check for index overflow
1525 //
1526 // adrp x17, Ltable@PAGE ; materialize table address
1527 // add x17, Ltable@PAGEOFF
1528 // ldrsw x16, [x17, x16, lsl #2] ; load table entry
1529 //
1530 // Lanchor:
1531 // adr x17, Lanchor ; compute target address
1532 // add x16, x17, x16
1533 // br x16 ; branch to target
1534
1535 MachineOperand JTOp = MI.getOperand(0);
1536
1537 unsigned JTI = JTOp.getIndex();
1538 assert(!AArch64FI->getJumpTableEntryPCRelSymbol(JTI) &&
1539 "unsupported compressed jump table");
1540
1541 const uint64_t NumTableEntries = JTs[JTI].MBBs.size();
1542
1543 // cmp only supports a 12-bit immediate. If we need more, materialize the
1544 // immediate, using x17 as a scratch register.
1545 uint64_t MaxTableEntry = NumTableEntries - 1;
1546 if (isUInt<12>(MaxTableEntry)) {
1547 EmitToStreamer(*OutStreamer, MCInstBuilder(AArch64::SUBSXri)
1548 .addReg(AArch64::XZR)
1549 .addReg(AArch64::X16)
1550 .addImm(MaxTableEntry)
1551 .addImm(0));
1552 } else {
1553 emitMOVZ(AArch64::X17, static_cast<uint16_t>(MaxTableEntry), 0);
1554 // It's sad that we have to manually materialize instructions, but we can't
1555 // trivially reuse the main pseudo expansion logic.
1556 // A MOVK sequence is easy enough to generate and handles the general case.
1557 for (int Offset = 16; Offset < 64; Offset += 16) {
1558 if ((MaxTableEntry >> Offset) == 0)
1559 break;
1560 emitMOVK(AArch64::X17, static_cast<uint16_t>(MaxTableEntry >> Offset),
1561 Offset);
1562 }
1563 EmitToStreamer(*OutStreamer, MCInstBuilder(AArch64::SUBSXrs)
1564 .addReg(AArch64::XZR)
1565 .addReg(AArch64::X16)
1566 .addReg(AArch64::X17)
1567 .addImm(0));
1568 }
1569
1570 // This picks entry #0 on failure.
1571 // We might want to trap instead.
1572 EmitToStreamer(*OutStreamer, MCInstBuilder(AArch64::CSELXr)
1573 .addReg(AArch64::X16)
1574 .addReg(AArch64::X16)
1575 .addReg(AArch64::XZR)
1576 .addImm(AArch64CC::LS));
1577
1578 // Prepare the @PAGE/@PAGEOFF low/high operands.
1579 MachineOperand JTMOHi(JTOp), JTMOLo(JTOp);
1580 MCOperand JTMCHi, JTMCLo;
1581
1582 JTMOHi.setTargetFlags(AArch64II::MO_PAGE);
1583 JTMOLo.setTargetFlags(AArch64II::MO_PAGEOFF | AArch64II::MO_NC);
1584
1585 MCInstLowering.lowerOperand(JTMOHi, JTMCHi);
1586 MCInstLowering.lowerOperand(JTMOLo, JTMCLo);
1587
1588 EmitToStreamer(
1589 *OutStreamer,
1590 MCInstBuilder(AArch64::ADRP).addReg(AArch64::X17).addOperand(JTMCHi));
1591
1592 EmitToStreamer(*OutStreamer, MCInstBuilder(AArch64::ADDXri)
1593 .addReg(AArch64::X17)
1594 .addReg(AArch64::X17)
1595 .addOperand(JTMCLo)
1596 .addImm(0));
1597
1598 EmitToStreamer(*OutStreamer, MCInstBuilder(AArch64::LDRSWroX)
1599 .addReg(AArch64::X16)
1600 .addReg(AArch64::X17)
1601 .addReg(AArch64::X16)
1602 .addImm(0)
1603 .addImm(1));
1604
1605 MCSymbol *AdrLabel = MF->getContext().createTempSymbol();
1606 const auto *AdrLabelE = MCSymbolRefExpr::create(AdrLabel, MF->getContext());
1607 AArch64FI->setJumpTableEntryInfo(JTI, 4, AdrLabel);
1608
1609 OutStreamer->emitLabel(AdrLabel);
1610 EmitToStreamer(
1611 *OutStreamer,
1612 MCInstBuilder(AArch64::ADR).addReg(AArch64::X17).addExpr(AdrLabelE));
1613
1614 EmitToStreamer(*OutStreamer, MCInstBuilder(AArch64::ADDXrs)
1615 .addReg(AArch64::X16)
1616 .addReg(AArch64::X17)
1617 .addReg(AArch64::X16)
1618 .addImm(0));
1619
1620 EmitToStreamer(*OutStreamer, MCInstBuilder(AArch64::BR).addReg(AArch64::X16));
1621 }
1622
LowerMOPS(llvm::MCStreamer & OutStreamer,const llvm::MachineInstr & MI)1623 void AArch64AsmPrinter::LowerMOPS(llvm::MCStreamer &OutStreamer,
1624 const llvm::MachineInstr &MI) {
1625 unsigned Opcode = MI.getOpcode();
1626 assert(STI->hasMOPS());
1627 assert(STI->hasMTE() || Opcode != AArch64::MOPSMemorySetTaggingPseudo);
1628
1629 const auto Ops = [Opcode]() -> std::array<unsigned, 3> {
1630 if (Opcode == AArch64::MOPSMemoryCopyPseudo)
1631 return {AArch64::CPYFP, AArch64::CPYFM, AArch64::CPYFE};
1632 if (Opcode == AArch64::MOPSMemoryMovePseudo)
1633 return {AArch64::CPYP, AArch64::CPYM, AArch64::CPYE};
1634 if (Opcode == AArch64::MOPSMemorySetPseudo)
1635 return {AArch64::SETP, AArch64::SETM, AArch64::SETE};
1636 if (Opcode == AArch64::MOPSMemorySetTaggingPseudo)
1637 return {AArch64::SETGP, AArch64::SETGM, AArch64::MOPSSETGE};
1638 llvm_unreachable("Unhandled memory operation pseudo");
1639 }();
1640 const bool IsSet = Opcode == AArch64::MOPSMemorySetPseudo ||
1641 Opcode == AArch64::MOPSMemorySetTaggingPseudo;
1642
1643 for (auto Op : Ops) {
1644 int i = 0;
1645 auto MCIB = MCInstBuilder(Op);
1646 // Destination registers
1647 MCIB.addReg(MI.getOperand(i++).getReg());
1648 MCIB.addReg(MI.getOperand(i++).getReg());
1649 if (!IsSet)
1650 MCIB.addReg(MI.getOperand(i++).getReg());
1651 // Input registers
1652 MCIB.addReg(MI.getOperand(i++).getReg());
1653 MCIB.addReg(MI.getOperand(i++).getReg());
1654 MCIB.addReg(MI.getOperand(i++).getReg());
1655
1656 EmitToStreamer(OutStreamer, MCIB);
1657 }
1658 }
1659
LowerSTACKMAP(MCStreamer & OutStreamer,StackMaps & SM,const MachineInstr & MI)1660 void AArch64AsmPrinter::LowerSTACKMAP(MCStreamer &OutStreamer, StackMaps &SM,
1661 const MachineInstr &MI) {
1662 unsigned NumNOPBytes = StackMapOpers(&MI).getNumPatchBytes();
1663
1664 auto &Ctx = OutStreamer.getContext();
1665 MCSymbol *MILabel = Ctx.createTempSymbol();
1666 OutStreamer.emitLabel(MILabel);
1667
1668 SM.recordStackMap(*MILabel, MI);
1669 assert(NumNOPBytes % 4 == 0 && "Invalid number of NOP bytes requested!");
1670
1671 // Scan ahead to trim the shadow.
1672 const MachineBasicBlock &MBB = *MI.getParent();
1673 MachineBasicBlock::const_iterator MII(MI);
1674 ++MII;
1675 while (NumNOPBytes > 0) {
1676 if (MII == MBB.end() || MII->isCall() ||
1677 MII->getOpcode() == AArch64::DBG_VALUE ||
1678 MII->getOpcode() == TargetOpcode::PATCHPOINT ||
1679 MII->getOpcode() == TargetOpcode::STACKMAP)
1680 break;
1681 ++MII;
1682 NumNOPBytes -= 4;
1683 }
1684
1685 // Emit nops.
1686 for (unsigned i = 0; i < NumNOPBytes; i += 4)
1687 EmitToStreamer(OutStreamer, MCInstBuilder(AArch64::HINT).addImm(0));
1688 }
1689
1690 // Lower a patchpoint of the form:
1691 // [<def>], <id>, <numBytes>, <target>, <numArgs>
LowerPATCHPOINT(MCStreamer & OutStreamer,StackMaps & SM,const MachineInstr & MI)1692 void AArch64AsmPrinter::LowerPATCHPOINT(MCStreamer &OutStreamer, StackMaps &SM,
1693 const MachineInstr &MI) {
1694 auto &Ctx = OutStreamer.getContext();
1695 MCSymbol *MILabel = Ctx.createTempSymbol();
1696 OutStreamer.emitLabel(MILabel);
1697 SM.recordPatchPoint(*MILabel, MI);
1698
1699 PatchPointOpers Opers(&MI);
1700
1701 int64_t CallTarget = Opers.getCallTarget().getImm();
1702 unsigned EncodedBytes = 0;
1703 if (CallTarget) {
1704 assert((CallTarget & 0xFFFFFFFFFFFF) == CallTarget &&
1705 "High 16 bits of call target should be zero.");
1706 Register ScratchReg = MI.getOperand(Opers.getNextScratchIdx()).getReg();
1707 EncodedBytes = 16;
1708 // Materialize the jump address:
1709 emitMOVZ(ScratchReg, (CallTarget >> 32) & 0xFFFF, 32);
1710 emitMOVK(ScratchReg, (CallTarget >> 16) & 0xFFFF, 16);
1711 emitMOVK(ScratchReg, CallTarget & 0xFFFF, 0);
1712 EmitToStreamer(OutStreamer, MCInstBuilder(AArch64::BLR).addReg(ScratchReg));
1713 }
1714 // Emit padding.
1715 unsigned NumBytes = Opers.getNumPatchBytes();
1716 assert(NumBytes >= EncodedBytes &&
1717 "Patchpoint can't request size less than the length of a call.");
1718 assert((NumBytes - EncodedBytes) % 4 == 0 &&
1719 "Invalid number of NOP bytes requested!");
1720 for (unsigned i = EncodedBytes; i < NumBytes; i += 4)
1721 EmitToStreamer(OutStreamer, MCInstBuilder(AArch64::HINT).addImm(0));
1722 }
1723
LowerSTATEPOINT(MCStreamer & OutStreamer,StackMaps & SM,const MachineInstr & MI)1724 void AArch64AsmPrinter::LowerSTATEPOINT(MCStreamer &OutStreamer, StackMaps &SM,
1725 const MachineInstr &MI) {
1726 StatepointOpers SOpers(&MI);
1727 if (unsigned PatchBytes = SOpers.getNumPatchBytes()) {
1728 assert(PatchBytes % 4 == 0 && "Invalid number of NOP bytes requested!");
1729 for (unsigned i = 0; i < PatchBytes; i += 4)
1730 EmitToStreamer(OutStreamer, MCInstBuilder(AArch64::HINT).addImm(0));
1731 } else {
1732 // Lower call target and choose correct opcode
1733 const MachineOperand &CallTarget = SOpers.getCallTarget();
1734 MCOperand CallTargetMCOp;
1735 unsigned CallOpcode;
1736 switch (CallTarget.getType()) {
1737 case MachineOperand::MO_GlobalAddress:
1738 case MachineOperand::MO_ExternalSymbol:
1739 MCInstLowering.lowerOperand(CallTarget, CallTargetMCOp);
1740 CallOpcode = AArch64::BL;
1741 break;
1742 case MachineOperand::MO_Immediate:
1743 CallTargetMCOp = MCOperand::createImm(CallTarget.getImm());
1744 CallOpcode = AArch64::BL;
1745 break;
1746 case MachineOperand::MO_Register:
1747 CallTargetMCOp = MCOperand::createReg(CallTarget.getReg());
1748 CallOpcode = AArch64::BLR;
1749 break;
1750 default:
1751 llvm_unreachable("Unsupported operand type in statepoint call target");
1752 break;
1753 }
1754
1755 EmitToStreamer(OutStreamer,
1756 MCInstBuilder(CallOpcode).addOperand(CallTargetMCOp));
1757 }
1758
1759 auto &Ctx = OutStreamer.getContext();
1760 MCSymbol *MILabel = Ctx.createTempSymbol();
1761 OutStreamer.emitLabel(MILabel);
1762 SM.recordStatepoint(*MILabel, MI);
1763 }
1764
LowerFAULTING_OP(const MachineInstr & FaultingMI)1765 void AArch64AsmPrinter::LowerFAULTING_OP(const MachineInstr &FaultingMI) {
1766 // FAULTING_LOAD_OP <def>, <faltinf type>, <MBB handler>,
1767 // <opcode>, <operands>
1768
1769 Register DefRegister = FaultingMI.getOperand(0).getReg();
1770 FaultMaps::FaultKind FK =
1771 static_cast<FaultMaps::FaultKind>(FaultingMI.getOperand(1).getImm());
1772 MCSymbol *HandlerLabel = FaultingMI.getOperand(2).getMBB()->getSymbol();
1773 unsigned Opcode = FaultingMI.getOperand(3).getImm();
1774 unsigned OperandsBeginIdx = 4;
1775
1776 auto &Ctx = OutStreamer->getContext();
1777 MCSymbol *FaultingLabel = Ctx.createTempSymbol();
1778 OutStreamer->emitLabel(FaultingLabel);
1779
1780 assert(FK < FaultMaps::FaultKindMax && "Invalid Faulting Kind!");
1781 FM.recordFaultingOp(FK, FaultingLabel, HandlerLabel);
1782
1783 MCInst MI;
1784 MI.setOpcode(Opcode);
1785
1786 if (DefRegister != (Register)0)
1787 MI.addOperand(MCOperand::createReg(DefRegister));
1788
1789 for (const MachineOperand &MO :
1790 llvm::drop_begin(FaultingMI.operands(), OperandsBeginIdx)) {
1791 MCOperand Dest;
1792 lowerOperand(MO, Dest);
1793 MI.addOperand(Dest);
1794 }
1795
1796 OutStreamer->AddComment("on-fault: " + HandlerLabel->getName());
1797 EmitToStreamer(MI);
1798 }
1799
emitMovXReg(Register Dest,Register Src)1800 void AArch64AsmPrinter::emitMovXReg(Register Dest, Register Src) {
1801 EmitToStreamer(*OutStreamer, MCInstBuilder(AArch64::ORRXrs)
1802 .addReg(Dest)
1803 .addReg(AArch64::XZR)
1804 .addReg(Src)
1805 .addImm(0));
1806 }
1807
emitMOVZ(Register Dest,uint64_t Imm,unsigned Shift)1808 void AArch64AsmPrinter::emitMOVZ(Register Dest, uint64_t Imm, unsigned Shift) {
1809 bool Is64Bit = AArch64::GPR64RegClass.contains(Dest);
1810 EmitToStreamer(*OutStreamer,
1811 MCInstBuilder(Is64Bit ? AArch64::MOVZXi : AArch64::MOVZWi)
1812 .addReg(Dest)
1813 .addImm(Imm)
1814 .addImm(Shift));
1815 }
1816
emitMOVK(Register Dest,uint64_t Imm,unsigned Shift)1817 void AArch64AsmPrinter::emitMOVK(Register Dest, uint64_t Imm, unsigned Shift) {
1818 bool Is64Bit = AArch64::GPR64RegClass.contains(Dest);
1819 EmitToStreamer(*OutStreamer,
1820 MCInstBuilder(Is64Bit ? AArch64::MOVKXi : AArch64::MOVKWi)
1821 .addReg(Dest)
1822 .addReg(Dest)
1823 .addImm(Imm)
1824 .addImm(Shift));
1825 }
1826
emitFMov0(const MachineInstr & MI)1827 void AArch64AsmPrinter::emitFMov0(const MachineInstr &MI) {
1828 Register DestReg = MI.getOperand(0).getReg();
1829 if (STI->hasZeroCycleZeroingFP() && !STI->hasZeroCycleZeroingFPWorkaround() &&
1830 STI->isNeonAvailable()) {
1831 // Convert H/S register to corresponding D register
1832 if (AArch64::H0 <= DestReg && DestReg <= AArch64::H31)
1833 DestReg = AArch64::D0 + (DestReg - AArch64::H0);
1834 else if (AArch64::S0 <= DestReg && DestReg <= AArch64::S31)
1835 DestReg = AArch64::D0 + (DestReg - AArch64::S0);
1836 else
1837 assert(AArch64::D0 <= DestReg && DestReg <= AArch64::D31);
1838
1839 MCInst MOVI;
1840 MOVI.setOpcode(AArch64::MOVID);
1841 MOVI.addOperand(MCOperand::createReg(DestReg));
1842 MOVI.addOperand(MCOperand::createImm(0));
1843 EmitToStreamer(*OutStreamer, MOVI);
1844 } else {
1845 MCInst FMov;
1846 switch (MI.getOpcode()) {
1847 default: llvm_unreachable("Unexpected opcode");
1848 case AArch64::FMOVH0:
1849 FMov.setOpcode(STI->hasFullFP16() ? AArch64::FMOVWHr : AArch64::FMOVWSr);
1850 if (!STI->hasFullFP16())
1851 DestReg = (AArch64::S0 + (DestReg - AArch64::H0));
1852 FMov.addOperand(MCOperand::createReg(DestReg));
1853 FMov.addOperand(MCOperand::createReg(AArch64::WZR));
1854 break;
1855 case AArch64::FMOVS0:
1856 FMov.setOpcode(AArch64::FMOVWSr);
1857 FMov.addOperand(MCOperand::createReg(DestReg));
1858 FMov.addOperand(MCOperand::createReg(AArch64::WZR));
1859 break;
1860 case AArch64::FMOVD0:
1861 FMov.setOpcode(AArch64::FMOVXDr);
1862 FMov.addOperand(MCOperand::createReg(DestReg));
1863 FMov.addOperand(MCOperand::createReg(AArch64::XZR));
1864 break;
1865 }
1866 EmitToStreamer(*OutStreamer, FMov);
1867 }
1868 }
1869
emitPtrauthDiscriminator(uint16_t Disc,Register AddrDisc,Register ScratchReg,bool MayUseAddrAsScratch)1870 Register AArch64AsmPrinter::emitPtrauthDiscriminator(uint16_t Disc,
1871 Register AddrDisc,
1872 Register ScratchReg,
1873 bool MayUseAddrAsScratch) {
1874 assert(ScratchReg == AArch64::X16 || ScratchReg == AArch64::X17 ||
1875 !STI->isX16X17Safer());
1876 // So far we've used NoRegister in pseudos. Now we need real encodings.
1877 if (AddrDisc == AArch64::NoRegister)
1878 AddrDisc = AArch64::XZR;
1879
1880 // If there is no constant discriminator, there's no blend involved:
1881 // just use the address discriminator register as-is (XZR or not).
1882 if (!Disc)
1883 return AddrDisc;
1884
1885 // If there's only a constant discriminator, MOV it into the scratch register.
1886 if (AddrDisc == AArch64::XZR) {
1887 emitMOVZ(ScratchReg, Disc, 0);
1888 return ScratchReg;
1889 }
1890
1891 // If there are both, emit a blend into the scratch register.
1892
1893 // Check if we can save one MOV instruction.
1894 assert(MayUseAddrAsScratch || ScratchReg != AddrDisc);
1895 bool AddrDiscIsSafe = AddrDisc == AArch64::X16 || AddrDisc == AArch64::X17 ||
1896 !STI->isX16X17Safer();
1897 if (MayUseAddrAsScratch && AddrDiscIsSafe)
1898 ScratchReg = AddrDisc;
1899 else
1900 emitMovXReg(ScratchReg, AddrDisc);
1901
1902 emitMOVK(ScratchReg, Disc, 48);
1903 return ScratchReg;
1904 }
1905
1906 /// Emits a code sequence to check an authenticated pointer value.
1907 ///
1908 /// If OnFailure argument is passed, jump there on check failure instead
1909 /// of proceeding to the next instruction (only if ShouldTrap is false).
emitPtrauthCheckAuthenticatedValue(Register TestedReg,Register ScratchReg,AArch64PACKey::ID Key,AArch64PAuth::AuthCheckMethod Method,bool ShouldTrap,const MCSymbol * OnFailure)1910 void AArch64AsmPrinter::emitPtrauthCheckAuthenticatedValue(
1911 Register TestedReg, Register ScratchReg, AArch64PACKey::ID Key,
1912 AArch64PAuth::AuthCheckMethod Method, bool ShouldTrap,
1913 const MCSymbol *OnFailure) {
1914 // Insert a sequence to check if authentication of TestedReg succeeded,
1915 // such as:
1916 //
1917 // - checked and clearing:
1918 // ; x16 is TestedReg, x17 is ScratchReg
1919 // mov x17, x16
1920 // xpaci x17
1921 // cmp x16, x17
1922 // b.eq Lsuccess
1923 // mov x16, x17
1924 // b Lend
1925 // Lsuccess:
1926 // ; skipped if authentication failed
1927 // Lend:
1928 // ...
1929 //
1930 // - checked and trapping:
1931 // mov x17, x16
1932 // xpaci x17
1933 // cmp x16, x17
1934 // b.eq Lsuccess
1935 // brk #<0xc470 + aut key>
1936 // Lsuccess:
1937 // ...
1938 //
1939 // See the documentation on AuthCheckMethod enumeration constants for
1940 // the specific code sequences that can be used to perform the check.
1941 using AArch64PAuth::AuthCheckMethod;
1942
1943 if (Method == AuthCheckMethod::None)
1944 return;
1945 if (Method == AuthCheckMethod::DummyLoad) {
1946 EmitToStreamer(MCInstBuilder(AArch64::LDRWui)
1947 .addReg(getWRegFromXReg(ScratchReg))
1948 .addReg(TestedReg)
1949 .addImm(0));
1950 assert(ShouldTrap && !OnFailure && "DummyLoad always traps on error");
1951 return;
1952 }
1953
1954 MCSymbol *SuccessSym = createTempSymbol("auth_success_");
1955 if (Method == AuthCheckMethod::XPAC || Method == AuthCheckMethod::XPACHint) {
1956 // mov Xscratch, Xtested
1957 emitMovXReg(ScratchReg, TestedReg);
1958
1959 if (Method == AuthCheckMethod::XPAC) {
1960 // xpac(i|d) Xscratch
1961 unsigned XPACOpc = getXPACOpcodeForKey(Key);
1962 EmitToStreamer(
1963 MCInstBuilder(XPACOpc).addReg(ScratchReg).addReg(ScratchReg));
1964 } else {
1965 // xpaclri
1966
1967 // Note that this method applies XPAC to TestedReg instead of ScratchReg.
1968 assert(TestedReg == AArch64::LR &&
1969 "XPACHint mode is only compatible with checking the LR register");
1970 assert((Key == AArch64PACKey::IA || Key == AArch64PACKey::IB) &&
1971 "XPACHint mode is only compatible with I-keys");
1972 EmitToStreamer(MCInstBuilder(AArch64::XPACLRI));
1973 }
1974
1975 // cmp Xtested, Xscratch
1976 EmitToStreamer(MCInstBuilder(AArch64::SUBSXrs)
1977 .addReg(AArch64::XZR)
1978 .addReg(TestedReg)
1979 .addReg(ScratchReg)
1980 .addImm(0));
1981
1982 // b.eq Lsuccess
1983 EmitToStreamer(
1984 MCInstBuilder(AArch64::Bcc)
1985 .addImm(AArch64CC::EQ)
1986 .addExpr(MCSymbolRefExpr::create(SuccessSym, OutContext)));
1987 } else if (Method == AuthCheckMethod::HighBitsNoTBI) {
1988 // eor Xscratch, Xtested, Xtested, lsl #1
1989 EmitToStreamer(MCInstBuilder(AArch64::EORXrs)
1990 .addReg(ScratchReg)
1991 .addReg(TestedReg)
1992 .addReg(TestedReg)
1993 .addImm(1));
1994 // tbz Xscratch, #62, Lsuccess
1995 EmitToStreamer(
1996 MCInstBuilder(AArch64::TBZX)
1997 .addReg(ScratchReg)
1998 .addImm(62)
1999 .addExpr(MCSymbolRefExpr::create(SuccessSym, OutContext)));
2000 } else {
2001 llvm_unreachable("Unsupported check method");
2002 }
2003
2004 if (ShouldTrap) {
2005 assert(!OnFailure && "Cannot specify OnFailure with ShouldTrap");
2006 // Trapping sequences do a 'brk'.
2007 // brk #<0xc470 + aut key>
2008 EmitToStreamer(MCInstBuilder(AArch64::BRK).addImm(0xc470 | Key));
2009 } else {
2010 // Non-trapping checked sequences return the stripped result in TestedReg,
2011 // skipping over success-only code (such as re-signing the pointer) if
2012 // there is one.
2013 // Note that this can introduce an authentication oracle (such as based on
2014 // the high bits of the re-signed value).
2015
2016 // FIXME: The XPAC method can be optimized by applying XPAC to TestedReg
2017 // instead of ScratchReg, thus eliminating one `mov` instruction.
2018 // Both XPAC and XPACHint can be further optimized by not using a
2019 // conditional branch jumping over an unconditional one.
2020
2021 switch (Method) {
2022 case AuthCheckMethod::XPACHint:
2023 // LR is already XPAC-ed at this point.
2024 break;
2025 case AuthCheckMethod::XPAC:
2026 // mov Xtested, Xscratch
2027 emitMovXReg(TestedReg, ScratchReg);
2028 break;
2029 default:
2030 // If Xtested was not XPAC-ed so far, emit XPAC here.
2031 // xpac(i|d) Xtested
2032 unsigned XPACOpc = getXPACOpcodeForKey(Key);
2033 EmitToStreamer(
2034 MCInstBuilder(XPACOpc).addReg(TestedReg).addReg(TestedReg));
2035 }
2036
2037 if (OnFailure) {
2038 // b Lend
2039 EmitToStreamer(
2040 MCInstBuilder(AArch64::B)
2041 .addExpr(MCSymbolRefExpr::create(OnFailure, OutContext)));
2042 }
2043 }
2044
2045 // If the auth check succeeds, we can continue.
2046 // Lsuccess:
2047 OutStreamer->emitLabel(SuccessSym);
2048 }
2049
2050 // With Pointer Authentication, it may be needed to explicitly check the
2051 // authenticated value in LR before performing a tail call.
2052 // Otherwise, the callee may re-sign the invalid return address,
2053 // introducing a signing oracle.
emitPtrauthTailCallHardening(const MachineInstr * TC)2054 void AArch64AsmPrinter::emitPtrauthTailCallHardening(const MachineInstr *TC) {
2055 if (!AArch64FI->shouldSignReturnAddress(*MF))
2056 return;
2057
2058 auto LRCheckMethod = STI->getAuthenticatedLRCheckMethod(*MF);
2059 if (LRCheckMethod == AArch64PAuth::AuthCheckMethod::None)
2060 return;
2061
2062 const AArch64RegisterInfo *TRI = STI->getRegisterInfo();
2063 Register ScratchReg =
2064 TC->readsRegister(AArch64::X16, TRI) ? AArch64::X17 : AArch64::X16;
2065 assert(!TC->readsRegister(ScratchReg, TRI) &&
2066 "Neither x16 nor x17 is available as a scratch register");
2067 AArch64PACKey::ID Key =
2068 AArch64FI->shouldSignWithBKey() ? AArch64PACKey::IB : AArch64PACKey::IA;
2069 emitPtrauthCheckAuthenticatedValue(
2070 AArch64::LR, ScratchReg, Key, LRCheckMethod,
2071 /*ShouldTrap=*/true, /*OnFailure=*/nullptr);
2072 }
2073
emitPtrauthAuthResign(Register AUTVal,AArch64PACKey::ID AUTKey,uint64_t AUTDisc,const MachineOperand * AUTAddrDisc,Register Scratch,std::optional<AArch64PACKey::ID> PACKey,uint64_t PACDisc,Register PACAddrDisc)2074 void AArch64AsmPrinter::emitPtrauthAuthResign(
2075 Register AUTVal, AArch64PACKey::ID AUTKey, uint64_t AUTDisc,
2076 const MachineOperand *AUTAddrDisc, Register Scratch,
2077 std::optional<AArch64PACKey::ID> PACKey, uint64_t PACDisc,
2078 Register PACAddrDisc) {
2079 const bool IsAUTPAC = PACKey.has_value();
2080
2081 // We expand AUT/AUTPAC into a sequence of the form
2082 //
2083 // ; authenticate x16
2084 // ; check pointer in x16
2085 // Lsuccess:
2086 // ; sign x16 (if AUTPAC)
2087 // Lend: ; if not trapping on failure
2088 //
2089 // with the checking sequence chosen depending on whether/how we should check
2090 // the pointer and whether we should trap on failure.
2091
2092 // By default, auth/resign sequences check for auth failures.
2093 bool ShouldCheck = true;
2094 // In the checked sequence, we only trap if explicitly requested.
2095 bool ShouldTrap = MF->getFunction().hasFnAttribute("ptrauth-auth-traps");
2096
2097 // On an FPAC CPU, you get traps whether you want them or not: there's
2098 // no point in emitting checks or traps.
2099 if (STI->hasFPAC())
2100 ShouldCheck = ShouldTrap = false;
2101
2102 // However, command-line flags can override this, for experimentation.
2103 switch (PtrauthAuthChecks) {
2104 case PtrauthCheckMode::Default:
2105 break;
2106 case PtrauthCheckMode::Unchecked:
2107 ShouldCheck = ShouldTrap = false;
2108 break;
2109 case PtrauthCheckMode::Poison:
2110 ShouldCheck = true;
2111 ShouldTrap = false;
2112 break;
2113 case PtrauthCheckMode::Trap:
2114 ShouldCheck = ShouldTrap = true;
2115 break;
2116 }
2117
2118 // Compute aut discriminator
2119 assert(isUInt<16>(AUTDisc));
2120 Register AUTDiscReg = emitPtrauthDiscriminator(
2121 AUTDisc, AUTAddrDisc->getReg(), Scratch, AUTAddrDisc->isKill());
2122 bool AUTZero = AUTDiscReg == AArch64::XZR;
2123 unsigned AUTOpc = getAUTOpcodeForKey(AUTKey, AUTZero);
2124
2125 // autiza x16 ; if AUTZero
2126 // autia x16, x17 ; if !AUTZero
2127 MCInst AUTInst;
2128 AUTInst.setOpcode(AUTOpc);
2129 AUTInst.addOperand(MCOperand::createReg(AUTVal));
2130 AUTInst.addOperand(MCOperand::createReg(AUTVal));
2131 if (!AUTZero)
2132 AUTInst.addOperand(MCOperand::createReg(AUTDiscReg));
2133 EmitToStreamer(*OutStreamer, AUTInst);
2134
2135 // Unchecked or checked-but-non-trapping AUT is just an "AUT": we're done.
2136 if (!IsAUTPAC && (!ShouldCheck || !ShouldTrap))
2137 return;
2138
2139 MCSymbol *EndSym = nullptr;
2140
2141 if (ShouldCheck) {
2142 if (IsAUTPAC && !ShouldTrap)
2143 EndSym = createTempSymbol("resign_end_");
2144
2145 emitPtrauthCheckAuthenticatedValue(AUTVal, Scratch, AUTKey,
2146 AArch64PAuth::AuthCheckMethod::XPAC,
2147 ShouldTrap, EndSym);
2148 }
2149
2150 // We already emitted unchecked and checked-but-non-trapping AUTs.
2151 // That left us with trapping AUTs, and AUTPACs.
2152 // Trapping AUTs don't need PAC: we're done.
2153 if (!IsAUTPAC)
2154 return;
2155
2156 // Compute pac discriminator
2157 assert(isUInt<16>(PACDisc));
2158 Register PACDiscReg =
2159 emitPtrauthDiscriminator(PACDisc, PACAddrDisc, Scratch);
2160 bool PACZero = PACDiscReg == AArch64::XZR;
2161 unsigned PACOpc = getPACOpcodeForKey(*PACKey, PACZero);
2162
2163 // pacizb x16 ; if PACZero
2164 // pacib x16, x17 ; if !PACZero
2165 MCInst PACInst;
2166 PACInst.setOpcode(PACOpc);
2167 PACInst.addOperand(MCOperand::createReg(AUTVal));
2168 PACInst.addOperand(MCOperand::createReg(AUTVal));
2169 if (!PACZero)
2170 PACInst.addOperand(MCOperand::createReg(PACDiscReg));
2171 EmitToStreamer(*OutStreamer, PACInst);
2172
2173 // Lend:
2174 if (EndSym)
2175 OutStreamer->emitLabel(EndSym);
2176 }
2177
emitPtrauthBranch(const MachineInstr * MI)2178 void AArch64AsmPrinter::emitPtrauthBranch(const MachineInstr *MI) {
2179 bool IsCall = MI->getOpcode() == AArch64::BLRA;
2180 unsigned BrTarget = MI->getOperand(0).getReg();
2181
2182 auto Key = (AArch64PACKey::ID)MI->getOperand(1).getImm();
2183 assert((Key == AArch64PACKey::IA || Key == AArch64PACKey::IB) &&
2184 "Invalid auth call key");
2185
2186 uint64_t Disc = MI->getOperand(2).getImm();
2187 assert(isUInt<16>(Disc));
2188
2189 unsigned AddrDisc = MI->getOperand(3).getReg();
2190
2191 // Make sure AddrDisc is solely used to compute the discriminator.
2192 // While hardly meaningful, it is still possible to describe an authentication
2193 // of a pointer against its own value (instead of storage address) with
2194 // intrinsics, so use report_fatal_error instead of assert.
2195 if (BrTarget == AddrDisc)
2196 report_fatal_error("Branch target is signed with its own value");
2197
2198 // If we are printing BLRA pseudo, try to save one MOV by making use of the
2199 // fact that x16 and x17 are described as clobbered by the MI instruction and
2200 // AddrDisc is not used as any other input.
2201 //
2202 // Back in the day, emitPtrauthDiscriminator was restricted to only returning
2203 // either x16 or x17, meaning the returned register is always among the
2204 // implicit-def'ed registers of BLRA pseudo. Now this property can be violated
2205 // if isX16X17Safer predicate is false, thus manually check if AddrDisc is
2206 // among x16 and x17 to prevent clobbering unexpected registers.
2207 //
2208 // Unlike BLRA, BRA pseudo is used to perform computed goto, and thus not
2209 // declared as clobbering x16/x17.
2210 //
2211 // FIXME: Make use of `killed` flags and register masks instead.
2212 bool AddrDiscIsImplicitDef =
2213 IsCall && (AddrDisc == AArch64::X16 || AddrDisc == AArch64::X17);
2214 Register DiscReg = emitPtrauthDiscriminator(Disc, AddrDisc, AArch64::X17,
2215 AddrDiscIsImplicitDef);
2216 bool IsZeroDisc = DiscReg == AArch64::XZR;
2217
2218 unsigned Opc;
2219 if (IsCall) {
2220 if (Key == AArch64PACKey::IA)
2221 Opc = IsZeroDisc ? AArch64::BLRAAZ : AArch64::BLRAA;
2222 else
2223 Opc = IsZeroDisc ? AArch64::BLRABZ : AArch64::BLRAB;
2224 } else {
2225 if (Key == AArch64PACKey::IA)
2226 Opc = IsZeroDisc ? AArch64::BRAAZ : AArch64::BRAA;
2227 else
2228 Opc = IsZeroDisc ? AArch64::BRABZ : AArch64::BRAB;
2229 }
2230
2231 MCInst BRInst;
2232 BRInst.setOpcode(Opc);
2233 BRInst.addOperand(MCOperand::createReg(BrTarget));
2234 if (!IsZeroDisc)
2235 BRInst.addOperand(MCOperand::createReg(DiscReg));
2236 EmitToStreamer(*OutStreamer, BRInst);
2237 }
2238
2239 const MCExpr *
lowerConstantPtrAuth(const ConstantPtrAuth & CPA)2240 AArch64AsmPrinter::lowerConstantPtrAuth(const ConstantPtrAuth &CPA) {
2241 MCContext &Ctx = OutContext;
2242
2243 // Figure out the base symbol and the addend, if any.
2244 APInt Offset(64, 0);
2245 const Value *BaseGV = CPA.getPointer()->stripAndAccumulateConstantOffsets(
2246 getDataLayout(), Offset, /*AllowNonInbounds=*/true);
2247
2248 auto *BaseGVB = dyn_cast<GlobalValue>(BaseGV);
2249
2250 // If we can't understand the referenced ConstantExpr, there's nothing
2251 // else we can do: emit an error.
2252 if (!BaseGVB) {
2253 BaseGV->getContext().emitError(
2254 "cannot resolve target base/addend of ptrauth constant");
2255 return nullptr;
2256 }
2257
2258 // If there is an addend, turn that into the appropriate MCExpr.
2259 const MCExpr *Sym = MCSymbolRefExpr::create(getSymbol(BaseGVB), Ctx);
2260 if (Offset.sgt(0))
2261 Sym = MCBinaryExpr::createAdd(
2262 Sym, MCConstantExpr::create(Offset.getSExtValue(), Ctx), Ctx);
2263 else if (Offset.slt(0))
2264 Sym = MCBinaryExpr::createSub(
2265 Sym, MCConstantExpr::create((-Offset).getSExtValue(), Ctx), Ctx);
2266
2267 uint64_t KeyID = CPA.getKey()->getZExtValue();
2268 // We later rely on valid KeyID value in AArch64PACKeyIDToString call from
2269 // AArch64AuthMCExpr::printImpl, so fail fast.
2270 if (KeyID > AArch64PACKey::LAST) {
2271 CPA.getContext().emitError("AArch64 PAC Key ID '" + Twine(KeyID) +
2272 "' out of range [0, " +
2273 Twine((unsigned)AArch64PACKey::LAST) + "]");
2274 KeyID = 0;
2275 }
2276
2277 uint64_t Disc = CPA.getDiscriminator()->getZExtValue();
2278 if (!isUInt<16>(Disc)) {
2279 CPA.getContext().emitError("AArch64 PAC Discriminator '" + Twine(Disc) +
2280 "' out of range [0, 0xFFFF]");
2281 Disc = 0;
2282 }
2283
2284 // Finally build the complete @AUTH expr.
2285 return AArch64AuthMCExpr::create(Sym, Disc, AArch64PACKey::ID(KeyID),
2286 CPA.hasAddressDiscriminator(), Ctx);
2287 }
2288
LowerLOADauthptrstatic(const MachineInstr & MI)2289 void AArch64AsmPrinter::LowerLOADauthptrstatic(const MachineInstr &MI) {
2290 unsigned DstReg = MI.getOperand(0).getReg();
2291 const MachineOperand &GAOp = MI.getOperand(1);
2292 const uint64_t KeyC = MI.getOperand(2).getImm();
2293 assert(KeyC <= AArch64PACKey::LAST &&
2294 "key is out of range [0, AArch64PACKey::LAST]");
2295 const auto Key = (AArch64PACKey::ID)KeyC;
2296 const uint64_t Disc = MI.getOperand(3).getImm();
2297 assert(isUInt<16>(Disc) &&
2298 "constant discriminator is out of range [0, 0xffff]");
2299
2300 // Emit instruction sequence like the following:
2301 // ADRP x16, symbol$auth_ptr$key$disc
2302 // LDR x16, [x16, :lo12:symbol$auth_ptr$key$disc]
2303 //
2304 // Where the $auth_ptr$ symbol is the stub slot containing the signed pointer
2305 // to symbol.
2306 MCSymbol *AuthPtrStubSym;
2307 if (TM.getTargetTriple().isOSBinFormatELF()) {
2308 const auto &TLOF =
2309 static_cast<const AArch64_ELFTargetObjectFile &>(getObjFileLowering());
2310
2311 assert(GAOp.getOffset() == 0 &&
2312 "non-zero offset for $auth_ptr$ stub slots is not supported");
2313 const MCSymbol *GASym = TM.getSymbol(GAOp.getGlobal());
2314 AuthPtrStubSym = TLOF.getAuthPtrSlotSymbol(TM, MMI, GASym, Key, Disc);
2315 } else {
2316 assert(TM.getTargetTriple().isOSBinFormatMachO() &&
2317 "LOADauthptrstatic is implemented only for MachO/ELF");
2318
2319 const auto &TLOF = static_cast<const AArch64_MachoTargetObjectFile &>(
2320 getObjFileLowering());
2321
2322 assert(GAOp.getOffset() == 0 &&
2323 "non-zero offset for $auth_ptr$ stub slots is not supported");
2324 const MCSymbol *GASym = TM.getSymbol(GAOp.getGlobal());
2325 AuthPtrStubSym = TLOF.getAuthPtrSlotSymbol(TM, MMI, GASym, Key, Disc);
2326 }
2327
2328 MachineOperand StubMOHi =
2329 MachineOperand::CreateMCSymbol(AuthPtrStubSym, AArch64II::MO_PAGE);
2330 MachineOperand StubMOLo = MachineOperand::CreateMCSymbol(
2331 AuthPtrStubSym, AArch64II::MO_PAGEOFF | AArch64II::MO_NC);
2332 MCOperand StubMCHi, StubMCLo;
2333
2334 MCInstLowering.lowerOperand(StubMOHi, StubMCHi);
2335 MCInstLowering.lowerOperand(StubMOLo, StubMCLo);
2336
2337 EmitToStreamer(
2338 *OutStreamer,
2339 MCInstBuilder(AArch64::ADRP).addReg(DstReg).addOperand(StubMCHi));
2340
2341 EmitToStreamer(*OutStreamer, MCInstBuilder(AArch64::LDRXui)
2342 .addReg(DstReg)
2343 .addReg(DstReg)
2344 .addOperand(StubMCLo));
2345 }
2346
LowerMOVaddrPAC(const MachineInstr & MI)2347 void AArch64AsmPrinter::LowerMOVaddrPAC(const MachineInstr &MI) {
2348 const bool IsGOTLoad = MI.getOpcode() == AArch64::LOADgotPAC;
2349 const bool IsELFSignedGOT = MI.getParent()
2350 ->getParent()
2351 ->getInfo<AArch64FunctionInfo>()
2352 ->hasELFSignedGOT();
2353 MachineOperand GAOp = MI.getOperand(0);
2354 const uint64_t KeyC = MI.getOperand(1).getImm();
2355 assert(KeyC <= AArch64PACKey::LAST &&
2356 "key is out of range [0, AArch64PACKey::LAST]");
2357 const auto Key = (AArch64PACKey::ID)KeyC;
2358 const unsigned AddrDisc = MI.getOperand(2).getReg();
2359 const uint64_t Disc = MI.getOperand(3).getImm();
2360 assert(isUInt<16>(Disc) &&
2361 "constant discriminator is out of range [0, 0xffff]");
2362
2363 const int64_t Offset = GAOp.getOffset();
2364 GAOp.setOffset(0);
2365
2366 // Emit:
2367 // target materialization:
2368 // - via GOT:
2369 // - unsigned GOT:
2370 // adrp x16, :got:target
2371 // ldr x16, [x16, :got_lo12:target]
2372 // add offset to x16 if offset != 0
2373 // - ELF signed GOT:
2374 // adrp x17, :got:target
2375 // add x17, x17, :got_auth_lo12:target
2376 // ldr x16, [x17]
2377 // aut{i|d}a x16, x17
2378 // check+trap sequence (if no FPAC)
2379 // add offset to x16 if offset != 0
2380 //
2381 // - direct:
2382 // adrp x16, target
2383 // add x16, x16, :lo12:target
2384 // add offset to x16 if offset != 0
2385 //
2386 // add offset to x16:
2387 // - abs(offset) fits 24 bits:
2388 // add/sub x16, x16, #<offset>[, #lsl 12] (up to 2 instructions)
2389 // - abs(offset) does not fit 24 bits:
2390 // - offset < 0:
2391 // movn+movk sequence filling x17 register with the offset (up to 4
2392 // instructions)
2393 // add x16, x16, x17
2394 // - offset > 0:
2395 // movz+movk sequence filling x17 register with the offset (up to 4
2396 // instructions)
2397 // add x16, x16, x17
2398 //
2399 // signing:
2400 // - 0 discriminator:
2401 // paciza x16
2402 // - Non-0 discriminator, no address discriminator:
2403 // mov x17, #Disc
2404 // pacia x16, x17
2405 // - address discriminator (with potentially folded immediate discriminator):
2406 // pacia x16, xAddrDisc
2407
2408 MachineOperand GAMOHi(GAOp), GAMOLo(GAOp);
2409 MCOperand GAMCHi, GAMCLo;
2410
2411 GAMOHi.setTargetFlags(AArch64II::MO_PAGE);
2412 GAMOLo.setTargetFlags(AArch64II::MO_PAGEOFF | AArch64II::MO_NC);
2413 if (IsGOTLoad) {
2414 GAMOHi.addTargetFlag(AArch64II::MO_GOT);
2415 GAMOLo.addTargetFlag(AArch64II::MO_GOT);
2416 }
2417
2418 MCInstLowering.lowerOperand(GAMOHi, GAMCHi);
2419 MCInstLowering.lowerOperand(GAMOLo, GAMCLo);
2420
2421 EmitToStreamer(
2422 MCInstBuilder(AArch64::ADRP)
2423 .addReg(IsGOTLoad && IsELFSignedGOT ? AArch64::X17 : AArch64::X16)
2424 .addOperand(GAMCHi));
2425
2426 if (IsGOTLoad) {
2427 if (IsELFSignedGOT) {
2428 EmitToStreamer(MCInstBuilder(AArch64::ADDXri)
2429 .addReg(AArch64::X17)
2430 .addReg(AArch64::X17)
2431 .addOperand(GAMCLo)
2432 .addImm(0));
2433
2434 EmitToStreamer(MCInstBuilder(AArch64::LDRXui)
2435 .addReg(AArch64::X16)
2436 .addReg(AArch64::X17)
2437 .addImm(0));
2438
2439 assert(GAOp.isGlobal());
2440 assert(GAOp.getGlobal()->getValueType() != nullptr);
2441 unsigned AuthOpcode = GAOp.getGlobal()->getValueType()->isFunctionTy()
2442 ? AArch64::AUTIA
2443 : AArch64::AUTDA;
2444
2445 EmitToStreamer(MCInstBuilder(AuthOpcode)
2446 .addReg(AArch64::X16)
2447 .addReg(AArch64::X16)
2448 .addReg(AArch64::X17));
2449
2450 if (!STI->hasFPAC()) {
2451 auto AuthKey = (AuthOpcode == AArch64::AUTIA ? AArch64PACKey::IA
2452 : AArch64PACKey::DA);
2453
2454 emitPtrauthCheckAuthenticatedValue(AArch64::X16, AArch64::X17, AuthKey,
2455 AArch64PAuth::AuthCheckMethod::XPAC,
2456 /*ShouldTrap=*/true,
2457 /*OnFailure=*/nullptr);
2458 }
2459 } else {
2460 EmitToStreamer(MCInstBuilder(AArch64::LDRXui)
2461 .addReg(AArch64::X16)
2462 .addReg(AArch64::X16)
2463 .addOperand(GAMCLo));
2464 }
2465 } else {
2466 EmitToStreamer(MCInstBuilder(AArch64::ADDXri)
2467 .addReg(AArch64::X16)
2468 .addReg(AArch64::X16)
2469 .addOperand(GAMCLo)
2470 .addImm(0));
2471 }
2472
2473 if (Offset != 0) {
2474 const uint64_t AbsOffset = (Offset > 0 ? Offset : -((uint64_t)Offset));
2475 const bool IsNeg = Offset < 0;
2476 if (isUInt<24>(AbsOffset)) {
2477 for (int BitPos = 0; BitPos != 24 && (AbsOffset >> BitPos);
2478 BitPos += 12) {
2479 EmitToStreamer(
2480 MCInstBuilder(IsNeg ? AArch64::SUBXri : AArch64::ADDXri)
2481 .addReg(AArch64::X16)
2482 .addReg(AArch64::X16)
2483 .addImm((AbsOffset >> BitPos) & 0xfff)
2484 .addImm(AArch64_AM::getShifterImm(AArch64_AM::LSL, BitPos)));
2485 }
2486 } else {
2487 const uint64_t UOffset = Offset;
2488 EmitToStreamer(MCInstBuilder(IsNeg ? AArch64::MOVNXi : AArch64::MOVZXi)
2489 .addReg(AArch64::X17)
2490 .addImm((IsNeg ? ~UOffset : UOffset) & 0xffff)
2491 .addImm(/*shift=*/0));
2492 auto NeedMovk = [IsNeg, UOffset](int BitPos) -> bool {
2493 assert(BitPos == 16 || BitPos == 32 || BitPos == 48);
2494 uint64_t Shifted = UOffset >> BitPos;
2495 if (!IsNeg)
2496 return Shifted != 0;
2497 for (int I = 0; I != 64 - BitPos; I += 16)
2498 if (((Shifted >> I) & 0xffff) != 0xffff)
2499 return true;
2500 return false;
2501 };
2502 for (int BitPos = 16; BitPos != 64 && NeedMovk(BitPos); BitPos += 16)
2503 emitMOVK(AArch64::X17, (UOffset >> BitPos) & 0xffff, BitPos);
2504
2505 EmitToStreamer(MCInstBuilder(AArch64::ADDXrs)
2506 .addReg(AArch64::X16)
2507 .addReg(AArch64::X16)
2508 .addReg(AArch64::X17)
2509 .addImm(/*shift=*/0));
2510 }
2511 }
2512
2513 Register DiscReg = emitPtrauthDiscriminator(Disc, AddrDisc, AArch64::X17);
2514
2515 auto MIB = MCInstBuilder(getPACOpcodeForKey(Key, DiscReg == AArch64::XZR))
2516 .addReg(AArch64::X16)
2517 .addReg(AArch64::X16);
2518 if (DiscReg != AArch64::XZR)
2519 MIB.addReg(DiscReg);
2520 EmitToStreamer(MIB);
2521 }
2522
LowerLOADgotAUTH(const MachineInstr & MI)2523 void AArch64AsmPrinter::LowerLOADgotAUTH(const MachineInstr &MI) {
2524 Register DstReg = MI.getOperand(0).getReg();
2525 Register AuthResultReg = STI->hasFPAC() ? DstReg : AArch64::X16;
2526 const MachineOperand &GAMO = MI.getOperand(1);
2527 assert(GAMO.getOffset() == 0);
2528
2529 if (MI.getMF()->getTarget().getCodeModel() == CodeModel::Tiny) {
2530 MCOperand GAMC;
2531 MCInstLowering.lowerOperand(GAMO, GAMC);
2532 EmitToStreamer(
2533 MCInstBuilder(AArch64::ADR).addReg(AArch64::X17).addOperand(GAMC));
2534 EmitToStreamer(MCInstBuilder(AArch64::LDRXui)
2535 .addReg(AuthResultReg)
2536 .addReg(AArch64::X17)
2537 .addImm(0));
2538 } else {
2539 MachineOperand GAHiOp(GAMO);
2540 MachineOperand GALoOp(GAMO);
2541 GAHiOp.addTargetFlag(AArch64II::MO_PAGE);
2542 GALoOp.addTargetFlag(AArch64II::MO_PAGEOFF | AArch64II::MO_NC);
2543
2544 MCOperand GAMCHi, GAMCLo;
2545 MCInstLowering.lowerOperand(GAHiOp, GAMCHi);
2546 MCInstLowering.lowerOperand(GALoOp, GAMCLo);
2547
2548 EmitToStreamer(
2549 MCInstBuilder(AArch64::ADRP).addReg(AArch64::X17).addOperand(GAMCHi));
2550
2551 EmitToStreamer(MCInstBuilder(AArch64::ADDXri)
2552 .addReg(AArch64::X17)
2553 .addReg(AArch64::X17)
2554 .addOperand(GAMCLo)
2555 .addImm(0));
2556
2557 EmitToStreamer(MCInstBuilder(AArch64::LDRXui)
2558 .addReg(AuthResultReg)
2559 .addReg(AArch64::X17)
2560 .addImm(0));
2561 }
2562
2563 assert(GAMO.isGlobal());
2564 MCSymbol *UndefWeakSym;
2565 if (GAMO.getGlobal()->hasExternalWeakLinkage()) {
2566 UndefWeakSym = createTempSymbol("undef_weak");
2567 EmitToStreamer(
2568 MCInstBuilder(AArch64::CBZX)
2569 .addReg(AuthResultReg)
2570 .addExpr(MCSymbolRefExpr::create(UndefWeakSym, OutContext)));
2571 }
2572
2573 assert(GAMO.getGlobal()->getValueType() != nullptr);
2574 unsigned AuthOpcode = GAMO.getGlobal()->getValueType()->isFunctionTy()
2575 ? AArch64::AUTIA
2576 : AArch64::AUTDA;
2577 EmitToStreamer(MCInstBuilder(AuthOpcode)
2578 .addReg(AuthResultReg)
2579 .addReg(AuthResultReg)
2580 .addReg(AArch64::X17));
2581
2582 if (GAMO.getGlobal()->hasExternalWeakLinkage())
2583 OutStreamer->emitLabel(UndefWeakSym);
2584
2585 if (!STI->hasFPAC()) {
2586 auto AuthKey =
2587 (AuthOpcode == AArch64::AUTIA ? AArch64PACKey::IA : AArch64PACKey::DA);
2588
2589 emitPtrauthCheckAuthenticatedValue(AuthResultReg, AArch64::X17, AuthKey,
2590 AArch64PAuth::AuthCheckMethod::XPAC,
2591 /*ShouldTrap=*/true,
2592 /*OnFailure=*/nullptr);
2593
2594 emitMovXReg(DstReg, AuthResultReg);
2595 }
2596 }
2597
2598 const MCExpr *
lowerBlockAddressConstant(const BlockAddress & BA)2599 AArch64AsmPrinter::lowerBlockAddressConstant(const BlockAddress &BA) {
2600 const MCExpr *BAE = AsmPrinter::lowerBlockAddressConstant(BA);
2601 const Function &Fn = *BA.getFunction();
2602
2603 if (std::optional<uint16_t> BADisc =
2604 STI->getPtrAuthBlockAddressDiscriminatorIfEnabled(Fn))
2605 return AArch64AuthMCExpr::create(BAE, *BADisc, AArch64PACKey::IA,
2606 /*HasAddressDiversity=*/false, OutContext);
2607
2608 return BAE;
2609 }
2610
emitCBPseudoExpansion(const MachineInstr * MI)2611 void AArch64AsmPrinter::emitCBPseudoExpansion(const MachineInstr *MI) {
2612 bool IsImm = false;
2613 bool Is32Bit = false;
2614
2615 switch (MI->getOpcode()) {
2616 default:
2617 llvm_unreachable("This is not a CB pseudo instruction");
2618 case AArch64::CBWPrr:
2619 Is32Bit = true;
2620 break;
2621 case AArch64::CBXPrr:
2622 Is32Bit = false;
2623 break;
2624 case AArch64::CBWPri:
2625 IsImm = true;
2626 Is32Bit = true;
2627 break;
2628 case AArch64::CBXPri:
2629 IsImm = true;
2630 break;
2631 }
2632
2633 AArch64CC::CondCode CC =
2634 static_cast<AArch64CC::CondCode>(MI->getOperand(0).getImm());
2635 bool NeedsRegSwap = false;
2636 bool NeedsImmDec = false;
2637 bool NeedsImmInc = false;
2638
2639 // Decide if we need to either swap register operands or increment/decrement
2640 // immediate operands
2641 unsigned MCOpC;
2642 switch (CC) {
2643 default:
2644 llvm_unreachable("Invalid CB condition code");
2645 case AArch64CC::EQ:
2646 MCOpC = IsImm ? (Is32Bit ? AArch64::CBEQWri : AArch64::CBEQXri)
2647 : (Is32Bit ? AArch64::CBEQWrr : AArch64::CBEQXrr);
2648 break;
2649 case AArch64CC::NE:
2650 MCOpC = IsImm ? (Is32Bit ? AArch64::CBNEWri : AArch64::CBNEXri)
2651 : (Is32Bit ? AArch64::CBNEWrr : AArch64::CBNEXrr);
2652 break;
2653 case AArch64CC::HS:
2654 MCOpC = IsImm ? (Is32Bit ? AArch64::CBHIWri : AArch64::CBHIXri)
2655 : (Is32Bit ? AArch64::CBHSWrr : AArch64::CBHSXrr);
2656 NeedsImmDec = IsImm;
2657 break;
2658 case AArch64CC::LO:
2659 MCOpC = IsImm ? (Is32Bit ? AArch64::CBLOWri : AArch64::CBLOXri)
2660 : (Is32Bit ? AArch64::CBHIWrr : AArch64::CBHIXrr);
2661 NeedsRegSwap = !IsImm;
2662 break;
2663 case AArch64CC::HI:
2664 MCOpC = IsImm ? (Is32Bit ? AArch64::CBHIWri : AArch64::CBHIXri)
2665 : (Is32Bit ? AArch64::CBHIWrr : AArch64::CBHIXrr);
2666 break;
2667 case AArch64CC::LS:
2668 MCOpC = IsImm ? (Is32Bit ? AArch64::CBLOWri : AArch64::CBLOXri)
2669 : (Is32Bit ? AArch64::CBHSWrr : AArch64::CBHSXrr);
2670 NeedsRegSwap = !IsImm;
2671 NeedsImmInc = IsImm;
2672 break;
2673 case AArch64CC::GE:
2674 MCOpC = IsImm ? (Is32Bit ? AArch64::CBGTWri : AArch64::CBGTXri)
2675 : (Is32Bit ? AArch64::CBGEWrr : AArch64::CBGEXrr);
2676 NeedsImmDec = IsImm;
2677 break;
2678 case AArch64CC::LT:
2679 MCOpC = IsImm ? (Is32Bit ? AArch64::CBLTWri : AArch64::CBLTXri)
2680 : (Is32Bit ? AArch64::CBGTWrr : AArch64::CBGTXrr);
2681 NeedsRegSwap = !IsImm;
2682 break;
2683 case AArch64CC::GT:
2684 MCOpC = IsImm ? (Is32Bit ? AArch64::CBGTWri : AArch64::CBGTXri)
2685 : (Is32Bit ? AArch64::CBGTWrr : AArch64::CBGTXrr);
2686 break;
2687 case AArch64CC::LE:
2688 MCOpC = IsImm ? (Is32Bit ? AArch64::CBLTWri : AArch64::CBLTXri)
2689 : (Is32Bit ? AArch64::CBGEWrr : AArch64::CBGEXrr);
2690 NeedsRegSwap = !IsImm;
2691 NeedsImmInc = IsImm;
2692 break;
2693 }
2694
2695 MCInst Inst;
2696 Inst.setOpcode(MCOpC);
2697
2698 MCOperand Lhs, Rhs, Trgt;
2699 lowerOperand(MI->getOperand(1), Lhs);
2700 lowerOperand(MI->getOperand(2), Rhs);
2701 lowerOperand(MI->getOperand(3), Trgt);
2702
2703 // Now swap, increment or decrement
2704 if (NeedsRegSwap) {
2705 assert(Lhs.isReg() && "Expected register operand for CB");
2706 assert(Rhs.isReg() && "Expected register operand for CB");
2707 Inst.addOperand(Rhs);
2708 Inst.addOperand(Lhs);
2709 } else if (NeedsImmDec) {
2710 Rhs.setImm(Rhs.getImm() - 1);
2711 Inst.addOperand(Lhs);
2712 Inst.addOperand(Rhs);
2713 } else if (NeedsImmInc) {
2714 Rhs.setImm(Rhs.getImm() + 1);
2715 Inst.addOperand(Lhs);
2716 Inst.addOperand(Rhs);
2717 } else {
2718 Inst.addOperand(Lhs);
2719 Inst.addOperand(Rhs);
2720 }
2721
2722 assert((!IsImm || (Rhs.getImm() >= 0 && Rhs.getImm() < 64)) &&
2723 "CB immediate operand out-of-bounds");
2724
2725 Inst.addOperand(Trgt);
2726 EmitToStreamer(*OutStreamer, Inst);
2727 }
2728
2729 // Simple pseudo-instructions have their lowering (with expansion to real
2730 // instructions) auto-generated.
2731 #include "AArch64GenMCPseudoLowering.inc"
2732
EmitToStreamer(MCStreamer & S,const MCInst & Inst)2733 void AArch64AsmPrinter::EmitToStreamer(MCStreamer &S, const MCInst &Inst) {
2734 S.emitInstruction(Inst, *STI);
2735 #ifndef NDEBUG
2736 ++InstsEmitted;
2737 #endif
2738 }
2739
emitInstruction(const MachineInstr * MI)2740 void AArch64AsmPrinter::emitInstruction(const MachineInstr *MI) {
2741 AArch64_MC::verifyInstructionPredicates(MI->getOpcode(), STI->getFeatureBits());
2742
2743 #ifndef NDEBUG
2744 InstsEmitted = 0;
2745 auto CheckMISize = make_scope_exit([&]() {
2746 assert(STI->getInstrInfo()->getInstSizeInBytes(*MI) >= InstsEmitted * 4);
2747 });
2748 #endif
2749
2750 // Do any auto-generated pseudo lowerings.
2751 if (MCInst OutInst; lowerPseudoInstExpansion(MI, OutInst)) {
2752 EmitToStreamer(*OutStreamer, OutInst);
2753 return;
2754 }
2755
2756 if (MI->getOpcode() == AArch64::ADRP) {
2757 for (auto &Opd : MI->operands()) {
2758 if (Opd.isSymbol() && StringRef(Opd.getSymbolName()) ==
2759 "swift_async_extendedFramePointerFlags") {
2760 ShouldEmitWeakSwiftAsyncExtendedFramePointerFlags = true;
2761 }
2762 }
2763 }
2764
2765 if (AArch64FI->getLOHRelated().count(MI)) {
2766 // Generate a label for LOH related instruction
2767 MCSymbol *LOHLabel = createTempSymbol("loh");
2768 // Associate the instruction with the label
2769 LOHInstToLabel[MI] = LOHLabel;
2770 OutStreamer->emitLabel(LOHLabel);
2771 }
2772
2773 AArch64TargetStreamer *TS =
2774 static_cast<AArch64TargetStreamer *>(OutStreamer->getTargetStreamer());
2775 // Do any manual lowerings.
2776 switch (MI->getOpcode()) {
2777 default:
2778 assert(!AArch64InstrInfo::isTailCallReturnInst(*MI) &&
2779 "Unhandled tail call instruction");
2780 break;
2781 case AArch64::HINT: {
2782 // CurrentPatchableFunctionEntrySym can be CurrentFnBegin only for
2783 // -fpatchable-function-entry=N,0. The entry MBB is guaranteed to be
2784 // non-empty. If MI is the initial BTI, place the
2785 // __patchable_function_entries label after BTI.
2786 if (CurrentPatchableFunctionEntrySym &&
2787 CurrentPatchableFunctionEntrySym == CurrentFnBegin &&
2788 MI == &MF->front().front()) {
2789 int64_t Imm = MI->getOperand(0).getImm();
2790 if ((Imm & 32) && (Imm & 6)) {
2791 MCInst Inst;
2792 MCInstLowering.Lower(MI, Inst);
2793 EmitToStreamer(*OutStreamer, Inst);
2794 CurrentPatchableFunctionEntrySym = createTempSymbol("patch");
2795 OutStreamer->emitLabel(CurrentPatchableFunctionEntrySym);
2796 return;
2797 }
2798 }
2799 break;
2800 }
2801 case AArch64::MOVMCSym: {
2802 Register DestReg = MI->getOperand(0).getReg();
2803 const MachineOperand &MO_Sym = MI->getOperand(1);
2804 MachineOperand Hi_MOSym(MO_Sym), Lo_MOSym(MO_Sym);
2805 MCOperand Hi_MCSym, Lo_MCSym;
2806
2807 Hi_MOSym.setTargetFlags(AArch64II::MO_G1 | AArch64II::MO_S);
2808 Lo_MOSym.setTargetFlags(AArch64II::MO_G0 | AArch64II::MO_NC);
2809
2810 MCInstLowering.lowerOperand(Hi_MOSym, Hi_MCSym);
2811 MCInstLowering.lowerOperand(Lo_MOSym, Lo_MCSym);
2812
2813 MCInst MovZ;
2814 MovZ.setOpcode(AArch64::MOVZXi);
2815 MovZ.addOperand(MCOperand::createReg(DestReg));
2816 MovZ.addOperand(Hi_MCSym);
2817 MovZ.addOperand(MCOperand::createImm(16));
2818 EmitToStreamer(*OutStreamer, MovZ);
2819
2820 MCInst MovK;
2821 MovK.setOpcode(AArch64::MOVKXi);
2822 MovK.addOperand(MCOperand::createReg(DestReg));
2823 MovK.addOperand(MCOperand::createReg(DestReg));
2824 MovK.addOperand(Lo_MCSym);
2825 MovK.addOperand(MCOperand::createImm(0));
2826 EmitToStreamer(*OutStreamer, MovK);
2827 return;
2828 }
2829 case AArch64::MOVIv2d_ns:
2830 // It is generally beneficial to rewrite "fmov s0, wzr" to "movi d0, #0".
2831 // as movi is more efficient across all cores. Newer cores can eliminate
2832 // fmovs early and there is no difference with movi, but this not true for
2833 // all implementations.
2834 //
2835 // The floating-point version doesn't quite work in rare cases on older
2836 // CPUs, so on those targets we lower this instruction to movi.16b instead.
2837 if (STI->hasZeroCycleZeroingFPWorkaround() &&
2838 MI->getOperand(1).getImm() == 0) {
2839 MCInst TmpInst;
2840 TmpInst.setOpcode(AArch64::MOVIv16b_ns);
2841 TmpInst.addOperand(MCOperand::createReg(MI->getOperand(0).getReg()));
2842 TmpInst.addOperand(MCOperand::createImm(MI->getOperand(1).getImm()));
2843 EmitToStreamer(*OutStreamer, TmpInst);
2844 return;
2845 }
2846 break;
2847
2848 case AArch64::DBG_VALUE:
2849 case AArch64::DBG_VALUE_LIST:
2850 if (isVerbose() && OutStreamer->hasRawTextSupport()) {
2851 SmallString<128> TmpStr;
2852 raw_svector_ostream OS(TmpStr);
2853 PrintDebugValueComment(MI, OS);
2854 OutStreamer->emitRawText(StringRef(OS.str()));
2855 }
2856 return;
2857
2858 case AArch64::EMITBKEY: {
2859 ExceptionHandling ExceptionHandlingType = MAI->getExceptionHandlingType();
2860 if (ExceptionHandlingType != ExceptionHandling::DwarfCFI &&
2861 ExceptionHandlingType != ExceptionHandling::ARM)
2862 return;
2863
2864 if (getFunctionCFISectionType(*MF) == CFISection::None)
2865 return;
2866
2867 OutStreamer->emitCFIBKeyFrame();
2868 return;
2869 }
2870
2871 case AArch64::EMITMTETAGGED: {
2872 ExceptionHandling ExceptionHandlingType = MAI->getExceptionHandlingType();
2873 if (ExceptionHandlingType != ExceptionHandling::DwarfCFI &&
2874 ExceptionHandlingType != ExceptionHandling::ARM)
2875 return;
2876
2877 if (getFunctionCFISectionType(*MF) != CFISection::None)
2878 OutStreamer->emitCFIMTETaggedFrame();
2879 return;
2880 }
2881
2882 case AArch64::AUTx16x17:
2883 emitPtrauthAuthResign(AArch64::X16,
2884 (AArch64PACKey::ID)MI->getOperand(0).getImm(),
2885 MI->getOperand(1).getImm(), &MI->getOperand(2),
2886 AArch64::X17, std::nullopt, 0, 0);
2887 return;
2888
2889 case AArch64::AUTxMxN:
2890 emitPtrauthAuthResign(MI->getOperand(0).getReg(),
2891 (AArch64PACKey::ID)MI->getOperand(3).getImm(),
2892 MI->getOperand(4).getImm(), &MI->getOperand(5),
2893 MI->getOperand(1).getReg(), std::nullopt, 0, 0);
2894 return;
2895
2896 case AArch64::AUTPAC:
2897 emitPtrauthAuthResign(
2898 AArch64::X16, (AArch64PACKey::ID)MI->getOperand(0).getImm(),
2899 MI->getOperand(1).getImm(), &MI->getOperand(2), AArch64::X17,
2900 (AArch64PACKey::ID)MI->getOperand(3).getImm(),
2901 MI->getOperand(4).getImm(), MI->getOperand(5).getReg());
2902 return;
2903
2904 case AArch64::LOADauthptrstatic:
2905 LowerLOADauthptrstatic(*MI);
2906 return;
2907
2908 case AArch64::LOADgotPAC:
2909 case AArch64::MOVaddrPAC:
2910 LowerMOVaddrPAC(*MI);
2911 return;
2912
2913 case AArch64::LOADgotAUTH:
2914 LowerLOADgotAUTH(*MI);
2915 return;
2916
2917 case AArch64::BRA:
2918 case AArch64::BLRA:
2919 emitPtrauthBranch(MI);
2920 return;
2921
2922 // Tail calls use pseudo instructions so they have the proper code-gen
2923 // attributes (isCall, isReturn, etc.). We lower them to the real
2924 // instruction here.
2925 case AArch64::AUTH_TCRETURN:
2926 case AArch64::AUTH_TCRETURN_BTI: {
2927 Register Callee = MI->getOperand(0).getReg();
2928 const uint64_t Key = MI->getOperand(2).getImm();
2929 assert((Key == AArch64PACKey::IA || Key == AArch64PACKey::IB) &&
2930 "Invalid auth key for tail-call return");
2931
2932 const uint64_t Disc = MI->getOperand(3).getImm();
2933 assert(isUInt<16>(Disc) && "Integer discriminator is too wide");
2934
2935 Register AddrDisc = MI->getOperand(4).getReg();
2936
2937 Register ScratchReg = Callee == AArch64::X16 ? AArch64::X17 : AArch64::X16;
2938
2939 emitPtrauthTailCallHardening(MI);
2940
2941 // See the comments in emitPtrauthBranch.
2942 if (Callee == AddrDisc)
2943 report_fatal_error("Call target is signed with its own value");
2944
2945 // After isX16X17Safer predicate was introduced, emitPtrauthDiscriminator is
2946 // no longer restricted to only reusing AddrDisc when it is X16 or X17
2947 // (which are implicit-def'ed by AUTH_TCRETURN pseudos), thus impose this
2948 // restriction manually not to clobber an unexpected register.
2949 bool AddrDiscIsImplicitDef =
2950 AddrDisc == AArch64::X16 || AddrDisc == AArch64::X17;
2951 Register DiscReg = emitPtrauthDiscriminator(Disc, AddrDisc, ScratchReg,
2952 AddrDiscIsImplicitDef);
2953
2954 const bool IsZero = DiscReg == AArch64::XZR;
2955 const unsigned Opcodes[2][2] = {{AArch64::BRAA, AArch64::BRAAZ},
2956 {AArch64::BRAB, AArch64::BRABZ}};
2957
2958 MCInst TmpInst;
2959 TmpInst.setOpcode(Opcodes[Key][IsZero]);
2960 TmpInst.addOperand(MCOperand::createReg(Callee));
2961 if (!IsZero)
2962 TmpInst.addOperand(MCOperand::createReg(DiscReg));
2963 EmitToStreamer(*OutStreamer, TmpInst);
2964 return;
2965 }
2966
2967 case AArch64::TCRETURNri:
2968 case AArch64::TCRETURNrix16x17:
2969 case AArch64::TCRETURNrix17:
2970 case AArch64::TCRETURNrinotx16:
2971 case AArch64::TCRETURNriALL: {
2972 emitPtrauthTailCallHardening(MI);
2973
2974 recordIfImportCall(MI);
2975 MCInst TmpInst;
2976 TmpInst.setOpcode(AArch64::BR);
2977 TmpInst.addOperand(MCOperand::createReg(MI->getOperand(0).getReg()));
2978 EmitToStreamer(*OutStreamer, TmpInst);
2979 return;
2980 }
2981 case AArch64::TCRETURNdi: {
2982 emitPtrauthTailCallHardening(MI);
2983
2984 MCOperand Dest;
2985 MCInstLowering.lowerOperand(MI->getOperand(0), Dest);
2986 recordIfImportCall(MI);
2987 MCInst TmpInst;
2988 TmpInst.setOpcode(AArch64::B);
2989 TmpInst.addOperand(Dest);
2990 EmitToStreamer(*OutStreamer, TmpInst);
2991 return;
2992 }
2993 case AArch64::SpeculationBarrierISBDSBEndBB: {
2994 // Print DSB SYS + ISB
2995 MCInst TmpInstDSB;
2996 TmpInstDSB.setOpcode(AArch64::DSB);
2997 TmpInstDSB.addOperand(MCOperand::createImm(0xf));
2998 EmitToStreamer(*OutStreamer, TmpInstDSB);
2999 MCInst TmpInstISB;
3000 TmpInstISB.setOpcode(AArch64::ISB);
3001 TmpInstISB.addOperand(MCOperand::createImm(0xf));
3002 EmitToStreamer(*OutStreamer, TmpInstISB);
3003 return;
3004 }
3005 case AArch64::SpeculationBarrierSBEndBB: {
3006 // Print SB
3007 MCInst TmpInstSB;
3008 TmpInstSB.setOpcode(AArch64::SB);
3009 EmitToStreamer(*OutStreamer, TmpInstSB);
3010 return;
3011 }
3012 case AArch64::TLSDESC_AUTH_CALLSEQ: {
3013 /// lower this to:
3014 /// adrp x0, :tlsdesc_auth:var
3015 /// ldr x16, [x0, #:tlsdesc_auth_lo12:var]
3016 /// add x0, x0, #:tlsdesc_auth_lo12:var
3017 /// blraa x16, x0
3018 /// (TPIDR_EL0 offset now in x0)
3019 const MachineOperand &MO_Sym = MI->getOperand(0);
3020 MachineOperand MO_TLSDESC_LO12(MO_Sym), MO_TLSDESC(MO_Sym);
3021 MCOperand SymTLSDescLo12, SymTLSDesc;
3022 MO_TLSDESC_LO12.setTargetFlags(AArch64II::MO_TLS | AArch64II::MO_PAGEOFF);
3023 MO_TLSDESC.setTargetFlags(AArch64II::MO_TLS | AArch64II::MO_PAGE);
3024 MCInstLowering.lowerOperand(MO_TLSDESC_LO12, SymTLSDescLo12);
3025 MCInstLowering.lowerOperand(MO_TLSDESC, SymTLSDesc);
3026
3027 MCInst Adrp;
3028 Adrp.setOpcode(AArch64::ADRP);
3029 Adrp.addOperand(MCOperand::createReg(AArch64::X0));
3030 Adrp.addOperand(SymTLSDesc);
3031 EmitToStreamer(*OutStreamer, Adrp);
3032
3033 MCInst Ldr;
3034 Ldr.setOpcode(AArch64::LDRXui);
3035 Ldr.addOperand(MCOperand::createReg(AArch64::X16));
3036 Ldr.addOperand(MCOperand::createReg(AArch64::X0));
3037 Ldr.addOperand(SymTLSDescLo12);
3038 Ldr.addOperand(MCOperand::createImm(0));
3039 EmitToStreamer(*OutStreamer, Ldr);
3040
3041 MCInst Add;
3042 Add.setOpcode(AArch64::ADDXri);
3043 Add.addOperand(MCOperand::createReg(AArch64::X0));
3044 Add.addOperand(MCOperand::createReg(AArch64::X0));
3045 Add.addOperand(SymTLSDescLo12);
3046 Add.addOperand(MCOperand::createImm(AArch64_AM::getShiftValue(0)));
3047 EmitToStreamer(*OutStreamer, Add);
3048
3049 // Authenticated TLSDESC accesses are not relaxed.
3050 // Thus, do not emit .tlsdesccall for AUTH TLSDESC.
3051
3052 MCInst Blraa;
3053 Blraa.setOpcode(AArch64::BLRAA);
3054 Blraa.addOperand(MCOperand::createReg(AArch64::X16));
3055 Blraa.addOperand(MCOperand::createReg(AArch64::X0));
3056 EmitToStreamer(*OutStreamer, Blraa);
3057
3058 return;
3059 }
3060 case AArch64::TLSDESC_CALLSEQ: {
3061 /// lower this to:
3062 /// adrp x0, :tlsdesc:var
3063 /// ldr x1, [x0, #:tlsdesc_lo12:var]
3064 /// add x0, x0, #:tlsdesc_lo12:var
3065 /// .tlsdesccall var
3066 /// blr x1
3067 /// (TPIDR_EL0 offset now in x0)
3068 const MachineOperand &MO_Sym = MI->getOperand(0);
3069 MachineOperand MO_TLSDESC_LO12(MO_Sym), MO_TLSDESC(MO_Sym);
3070 MCOperand Sym, SymTLSDescLo12, SymTLSDesc;
3071 MO_TLSDESC_LO12.setTargetFlags(AArch64II::MO_TLS | AArch64II::MO_PAGEOFF);
3072 MO_TLSDESC.setTargetFlags(AArch64II::MO_TLS | AArch64II::MO_PAGE);
3073 MCInstLowering.lowerOperand(MO_Sym, Sym);
3074 MCInstLowering.lowerOperand(MO_TLSDESC_LO12, SymTLSDescLo12);
3075 MCInstLowering.lowerOperand(MO_TLSDESC, SymTLSDesc);
3076
3077 MCInst Adrp;
3078 Adrp.setOpcode(AArch64::ADRP);
3079 Adrp.addOperand(MCOperand::createReg(AArch64::X0));
3080 Adrp.addOperand(SymTLSDesc);
3081 EmitToStreamer(*OutStreamer, Adrp);
3082
3083 MCInst Ldr;
3084 if (STI->isTargetILP32()) {
3085 Ldr.setOpcode(AArch64::LDRWui);
3086 Ldr.addOperand(MCOperand::createReg(AArch64::W1));
3087 } else {
3088 Ldr.setOpcode(AArch64::LDRXui);
3089 Ldr.addOperand(MCOperand::createReg(AArch64::X1));
3090 }
3091 Ldr.addOperand(MCOperand::createReg(AArch64::X0));
3092 Ldr.addOperand(SymTLSDescLo12);
3093 Ldr.addOperand(MCOperand::createImm(0));
3094 EmitToStreamer(*OutStreamer, Ldr);
3095
3096 MCInst Add;
3097 if (STI->isTargetILP32()) {
3098 Add.setOpcode(AArch64::ADDWri);
3099 Add.addOperand(MCOperand::createReg(AArch64::W0));
3100 Add.addOperand(MCOperand::createReg(AArch64::W0));
3101 } else {
3102 Add.setOpcode(AArch64::ADDXri);
3103 Add.addOperand(MCOperand::createReg(AArch64::X0));
3104 Add.addOperand(MCOperand::createReg(AArch64::X0));
3105 }
3106 Add.addOperand(SymTLSDescLo12);
3107 Add.addOperand(MCOperand::createImm(AArch64_AM::getShiftValue(0)));
3108 EmitToStreamer(*OutStreamer, Add);
3109
3110 // Emit a relocation-annotation. This expands to no code, but requests
3111 // the following instruction gets an R_AARCH64_TLSDESC_CALL.
3112 MCInst TLSDescCall;
3113 TLSDescCall.setOpcode(AArch64::TLSDESCCALL);
3114 TLSDescCall.addOperand(Sym);
3115 EmitToStreamer(*OutStreamer, TLSDescCall);
3116 #ifndef NDEBUG
3117 --InstsEmitted; // no code emitted
3118 #endif
3119
3120 MCInst Blr;
3121 Blr.setOpcode(AArch64::BLR);
3122 Blr.addOperand(MCOperand::createReg(AArch64::X1));
3123 EmitToStreamer(*OutStreamer, Blr);
3124
3125 return;
3126 }
3127
3128 case AArch64::JumpTableDest32:
3129 case AArch64::JumpTableDest16:
3130 case AArch64::JumpTableDest8:
3131 LowerJumpTableDest(*OutStreamer, *MI);
3132 return;
3133
3134 case AArch64::BR_JumpTable:
3135 LowerHardenedBRJumpTable(*MI);
3136 return;
3137
3138 case AArch64::FMOVH0:
3139 case AArch64::FMOVS0:
3140 case AArch64::FMOVD0:
3141 emitFMov0(*MI);
3142 return;
3143
3144 case AArch64::MOPSMemoryCopyPseudo:
3145 case AArch64::MOPSMemoryMovePseudo:
3146 case AArch64::MOPSMemorySetPseudo:
3147 case AArch64::MOPSMemorySetTaggingPseudo:
3148 LowerMOPS(*OutStreamer, *MI);
3149 return;
3150
3151 case TargetOpcode::STACKMAP:
3152 return LowerSTACKMAP(*OutStreamer, SM, *MI);
3153
3154 case TargetOpcode::PATCHPOINT:
3155 return LowerPATCHPOINT(*OutStreamer, SM, *MI);
3156
3157 case TargetOpcode::STATEPOINT:
3158 return LowerSTATEPOINT(*OutStreamer, SM, *MI);
3159
3160 case TargetOpcode::FAULTING_OP:
3161 return LowerFAULTING_OP(*MI);
3162
3163 case TargetOpcode::PATCHABLE_FUNCTION_ENTER:
3164 LowerPATCHABLE_FUNCTION_ENTER(*MI);
3165 return;
3166
3167 case TargetOpcode::PATCHABLE_FUNCTION_EXIT:
3168 LowerPATCHABLE_FUNCTION_EXIT(*MI);
3169 return;
3170
3171 case TargetOpcode::PATCHABLE_TAIL_CALL:
3172 LowerPATCHABLE_TAIL_CALL(*MI);
3173 return;
3174 case TargetOpcode::PATCHABLE_EVENT_CALL:
3175 return LowerPATCHABLE_EVENT_CALL(*MI, false);
3176 case TargetOpcode::PATCHABLE_TYPED_EVENT_CALL:
3177 return LowerPATCHABLE_EVENT_CALL(*MI, true);
3178
3179 case AArch64::KCFI_CHECK:
3180 LowerKCFI_CHECK(*MI);
3181 return;
3182
3183 case AArch64::HWASAN_CHECK_MEMACCESS:
3184 case AArch64::HWASAN_CHECK_MEMACCESS_SHORTGRANULES:
3185 case AArch64::HWASAN_CHECK_MEMACCESS_FIXEDSHADOW:
3186 case AArch64::HWASAN_CHECK_MEMACCESS_SHORTGRANULES_FIXEDSHADOW:
3187 LowerHWASAN_CHECK_MEMACCESS(*MI);
3188 return;
3189
3190 case AArch64::SEH_StackAlloc:
3191 TS->emitARM64WinCFIAllocStack(MI->getOperand(0).getImm());
3192 return;
3193
3194 case AArch64::SEH_SaveFPLR:
3195 TS->emitARM64WinCFISaveFPLR(MI->getOperand(0).getImm());
3196 return;
3197
3198 case AArch64::SEH_SaveFPLR_X:
3199 assert(MI->getOperand(0).getImm() < 0 &&
3200 "Pre increment SEH opcode must have a negative offset");
3201 TS->emitARM64WinCFISaveFPLRX(-MI->getOperand(0).getImm());
3202 return;
3203
3204 case AArch64::SEH_SaveReg:
3205 TS->emitARM64WinCFISaveReg(MI->getOperand(0).getImm(),
3206 MI->getOperand(1).getImm());
3207 return;
3208
3209 case AArch64::SEH_SaveReg_X:
3210 assert(MI->getOperand(1).getImm() < 0 &&
3211 "Pre increment SEH opcode must have a negative offset");
3212 TS->emitARM64WinCFISaveRegX(MI->getOperand(0).getImm(),
3213 -MI->getOperand(1).getImm());
3214 return;
3215
3216 case AArch64::SEH_SaveRegP:
3217 if (MI->getOperand(1).getImm() == 30 && MI->getOperand(0).getImm() >= 19 &&
3218 MI->getOperand(0).getImm() <= 28) {
3219 assert((MI->getOperand(0).getImm() - 19) % 2 == 0 &&
3220 "Register paired with LR must be odd");
3221 TS->emitARM64WinCFISaveLRPair(MI->getOperand(0).getImm(),
3222 MI->getOperand(2).getImm());
3223 return;
3224 }
3225 assert((MI->getOperand(1).getImm() - MI->getOperand(0).getImm() == 1) &&
3226 "Non-consecutive registers not allowed for save_regp");
3227 TS->emitARM64WinCFISaveRegP(MI->getOperand(0).getImm(),
3228 MI->getOperand(2).getImm());
3229 return;
3230
3231 case AArch64::SEH_SaveRegP_X:
3232 assert((MI->getOperand(1).getImm() - MI->getOperand(0).getImm() == 1) &&
3233 "Non-consecutive registers not allowed for save_regp_x");
3234 assert(MI->getOperand(2).getImm() < 0 &&
3235 "Pre increment SEH opcode must have a negative offset");
3236 TS->emitARM64WinCFISaveRegPX(MI->getOperand(0).getImm(),
3237 -MI->getOperand(2).getImm());
3238 return;
3239
3240 case AArch64::SEH_SaveFReg:
3241 TS->emitARM64WinCFISaveFReg(MI->getOperand(0).getImm(),
3242 MI->getOperand(1).getImm());
3243 return;
3244
3245 case AArch64::SEH_SaveFReg_X:
3246 assert(MI->getOperand(1).getImm() < 0 &&
3247 "Pre increment SEH opcode must have a negative offset");
3248 TS->emitARM64WinCFISaveFRegX(MI->getOperand(0).getImm(),
3249 -MI->getOperand(1).getImm());
3250 return;
3251
3252 case AArch64::SEH_SaveFRegP:
3253 assert((MI->getOperand(1).getImm() - MI->getOperand(0).getImm() == 1) &&
3254 "Non-consecutive registers not allowed for save_regp");
3255 TS->emitARM64WinCFISaveFRegP(MI->getOperand(0).getImm(),
3256 MI->getOperand(2).getImm());
3257 return;
3258
3259 case AArch64::SEH_SaveFRegP_X:
3260 assert((MI->getOperand(1).getImm() - MI->getOperand(0).getImm() == 1) &&
3261 "Non-consecutive registers not allowed for save_regp_x");
3262 assert(MI->getOperand(2).getImm() < 0 &&
3263 "Pre increment SEH opcode must have a negative offset");
3264 TS->emitARM64WinCFISaveFRegPX(MI->getOperand(0).getImm(),
3265 -MI->getOperand(2).getImm());
3266 return;
3267
3268 case AArch64::SEH_SetFP:
3269 TS->emitARM64WinCFISetFP();
3270 return;
3271
3272 case AArch64::SEH_AddFP:
3273 TS->emitARM64WinCFIAddFP(MI->getOperand(0).getImm());
3274 return;
3275
3276 case AArch64::SEH_Nop:
3277 TS->emitARM64WinCFINop();
3278 return;
3279
3280 case AArch64::SEH_PrologEnd:
3281 TS->emitARM64WinCFIPrologEnd();
3282 return;
3283
3284 case AArch64::SEH_EpilogStart:
3285 TS->emitARM64WinCFIEpilogStart();
3286 return;
3287
3288 case AArch64::SEH_EpilogEnd:
3289 TS->emitARM64WinCFIEpilogEnd();
3290 return;
3291
3292 case AArch64::SEH_PACSignLR:
3293 TS->emitARM64WinCFIPACSignLR();
3294 return;
3295
3296 case AArch64::SEH_SaveAnyRegQP:
3297 assert(MI->getOperand(1).getImm() - MI->getOperand(0).getImm() == 1 &&
3298 "Non-consecutive registers not allowed for save_any_reg");
3299 assert(MI->getOperand(2).getImm() >= 0 &&
3300 "SaveAnyRegQP SEH opcode offset must be non-negative");
3301 assert(MI->getOperand(2).getImm() <= 1008 &&
3302 "SaveAnyRegQP SEH opcode offset must fit into 6 bits");
3303 TS->emitARM64WinCFISaveAnyRegQP(MI->getOperand(0).getImm(),
3304 MI->getOperand(2).getImm());
3305 return;
3306
3307 case AArch64::SEH_SaveAnyRegQPX:
3308 assert(MI->getOperand(1).getImm() - MI->getOperand(0).getImm() == 1 &&
3309 "Non-consecutive registers not allowed for save_any_reg");
3310 assert(MI->getOperand(2).getImm() < 0 &&
3311 "SaveAnyRegQPX SEH opcode offset must be negative");
3312 assert(MI->getOperand(2).getImm() >= -1008 &&
3313 "SaveAnyRegQPX SEH opcode offset must fit into 6 bits");
3314 TS->emitARM64WinCFISaveAnyRegQPX(MI->getOperand(0).getImm(),
3315 -MI->getOperand(2).getImm());
3316 return;
3317
3318 case AArch64::SEH_AllocZ:
3319 assert(MI->getOperand(0).getImm() >= 0 &&
3320 "AllocZ SEH opcode offset must be non-negative");
3321 assert(MI->getOperand(0).getImm() <= 255 &&
3322 "AllocZ SEH opcode offset must fit into 8 bits");
3323 TS->emitARM64WinCFIAllocZ(MI->getOperand(0).getImm());
3324 return;
3325
3326 case AArch64::SEH_SaveZReg:
3327 assert(MI->getOperand(1).getImm() >= 0 &&
3328 "SaveZReg SEH opcode offset must be non-negative");
3329 assert(MI->getOperand(1).getImm() <= 255 &&
3330 "SaveZReg SEH opcode offset must fit into 8 bits");
3331 TS->emitARM64WinCFISaveZReg(MI->getOperand(0).getImm(),
3332 MI->getOperand(1).getImm());
3333 return;
3334
3335 case AArch64::SEH_SavePReg:
3336 assert(MI->getOperand(1).getImm() >= 0 &&
3337 "SavePReg SEH opcode offset must be non-negative");
3338 assert(MI->getOperand(1).getImm() <= 255 &&
3339 "SavePReg SEH opcode offset must fit into 8 bits");
3340 TS->emitARM64WinCFISavePReg(MI->getOperand(0).getImm(),
3341 MI->getOperand(1).getImm());
3342 return;
3343
3344 case AArch64::BLR:
3345 case AArch64::BR: {
3346 recordIfImportCall(MI);
3347 MCInst TmpInst;
3348 MCInstLowering.Lower(MI, TmpInst);
3349 EmitToStreamer(*OutStreamer, TmpInst);
3350 return;
3351 }
3352 case AArch64::CBWPri:
3353 case AArch64::CBXPri:
3354 case AArch64::CBWPrr:
3355 case AArch64::CBXPrr:
3356 emitCBPseudoExpansion(MI);
3357 return;
3358 }
3359
3360 // Finally, do the automated lowerings for everything else.
3361 MCInst TmpInst;
3362 MCInstLowering.Lower(MI, TmpInst);
3363 EmitToStreamer(*OutStreamer, TmpInst);
3364 }
3365
recordIfImportCall(const llvm::MachineInstr * BranchInst)3366 void AArch64AsmPrinter::recordIfImportCall(
3367 const llvm::MachineInstr *BranchInst) {
3368 if (!EnableImportCallOptimization)
3369 return;
3370
3371 auto [GV, OpFlags] = BranchInst->getMF()->tryGetCalledGlobal(BranchInst);
3372 if (GV && GV->hasDLLImportStorageClass()) {
3373 auto *CallSiteSymbol = MMI->getContext().createNamedTempSymbol("impcall");
3374 OutStreamer->emitLabel(CallSiteSymbol);
3375
3376 auto *CalledSymbol = MCInstLowering.GetGlobalValueSymbol(GV, OpFlags);
3377 SectionToImportedFunctionCalls[OutStreamer->getCurrentSectionOnly()]
3378 .push_back({CallSiteSymbol, CalledSymbol});
3379 }
3380 }
3381
emitMachOIFuncStubBody(Module & M,const GlobalIFunc & GI,MCSymbol * LazyPointer)3382 void AArch64AsmPrinter::emitMachOIFuncStubBody(Module &M, const GlobalIFunc &GI,
3383 MCSymbol *LazyPointer) {
3384 // _ifunc:
3385 // adrp x16, lazy_pointer@GOTPAGE
3386 // ldr x16, [x16, lazy_pointer@GOTPAGEOFF]
3387 // ldr x16, [x16]
3388 // br x16
3389
3390 {
3391 MCInst Adrp;
3392 Adrp.setOpcode(AArch64::ADRP);
3393 Adrp.addOperand(MCOperand::createReg(AArch64::X16));
3394 MCOperand SymPage;
3395 MCInstLowering.lowerOperand(
3396 MachineOperand::CreateMCSymbol(LazyPointer,
3397 AArch64II::MO_GOT | AArch64II::MO_PAGE),
3398 SymPage);
3399 Adrp.addOperand(SymPage);
3400 EmitToStreamer(Adrp);
3401 }
3402
3403 {
3404 MCInst Ldr;
3405 Ldr.setOpcode(AArch64::LDRXui);
3406 Ldr.addOperand(MCOperand::createReg(AArch64::X16));
3407 Ldr.addOperand(MCOperand::createReg(AArch64::X16));
3408 MCOperand SymPageOff;
3409 MCInstLowering.lowerOperand(
3410 MachineOperand::CreateMCSymbol(LazyPointer, AArch64II::MO_GOT |
3411 AArch64II::MO_PAGEOFF),
3412 SymPageOff);
3413 Ldr.addOperand(SymPageOff);
3414 Ldr.addOperand(MCOperand::createImm(0));
3415 EmitToStreamer(Ldr);
3416 }
3417
3418 EmitToStreamer(MCInstBuilder(AArch64::LDRXui)
3419 .addReg(AArch64::X16)
3420 .addReg(AArch64::X16)
3421 .addImm(0));
3422
3423 EmitToStreamer(MCInstBuilder(TM.getTargetTriple().isArm64e() ? AArch64::BRAAZ
3424 : AArch64::BR)
3425 .addReg(AArch64::X16));
3426 }
3427
emitMachOIFuncStubHelperBody(Module & M,const GlobalIFunc & GI,MCSymbol * LazyPointer)3428 void AArch64AsmPrinter::emitMachOIFuncStubHelperBody(Module &M,
3429 const GlobalIFunc &GI,
3430 MCSymbol *LazyPointer) {
3431 // These stub helpers are only ever called once, so here we're optimizing for
3432 // minimum size by using the pre-indexed store variants, which saves a few
3433 // bytes of instructions to bump & restore sp.
3434
3435 // _ifunc.stub_helper:
3436 // stp fp, lr, [sp, #-16]!
3437 // mov fp, sp
3438 // stp x1, x0, [sp, #-16]!
3439 // stp x3, x2, [sp, #-16]!
3440 // stp x5, x4, [sp, #-16]!
3441 // stp x7, x6, [sp, #-16]!
3442 // stp d1, d0, [sp, #-16]!
3443 // stp d3, d2, [sp, #-16]!
3444 // stp d5, d4, [sp, #-16]!
3445 // stp d7, d6, [sp, #-16]!
3446 // bl _resolver
3447 // adrp x16, lazy_pointer@GOTPAGE
3448 // ldr x16, [x16, lazy_pointer@GOTPAGEOFF]
3449 // str x0, [x16]
3450 // mov x16, x0
3451 // ldp d7, d6, [sp], #16
3452 // ldp d5, d4, [sp], #16
3453 // ldp d3, d2, [sp], #16
3454 // ldp d1, d0, [sp], #16
3455 // ldp x7, x6, [sp], #16
3456 // ldp x5, x4, [sp], #16
3457 // ldp x3, x2, [sp], #16
3458 // ldp x1, x0, [sp], #16
3459 // ldp fp, lr, [sp], #16
3460 // br x16
3461
3462 EmitToStreamer(MCInstBuilder(AArch64::STPXpre)
3463 .addReg(AArch64::SP)
3464 .addReg(AArch64::FP)
3465 .addReg(AArch64::LR)
3466 .addReg(AArch64::SP)
3467 .addImm(-2));
3468
3469 EmitToStreamer(MCInstBuilder(AArch64::ADDXri)
3470 .addReg(AArch64::FP)
3471 .addReg(AArch64::SP)
3472 .addImm(0)
3473 .addImm(0));
3474
3475 for (int I = 0; I != 4; ++I)
3476 EmitToStreamer(MCInstBuilder(AArch64::STPXpre)
3477 .addReg(AArch64::SP)
3478 .addReg(AArch64::X1 + 2 * I)
3479 .addReg(AArch64::X0 + 2 * I)
3480 .addReg(AArch64::SP)
3481 .addImm(-2));
3482
3483 for (int I = 0; I != 4; ++I)
3484 EmitToStreamer(MCInstBuilder(AArch64::STPDpre)
3485 .addReg(AArch64::SP)
3486 .addReg(AArch64::D1 + 2 * I)
3487 .addReg(AArch64::D0 + 2 * I)
3488 .addReg(AArch64::SP)
3489 .addImm(-2));
3490
3491 EmitToStreamer(
3492 MCInstBuilder(AArch64::BL)
3493 .addOperand(MCOperand::createExpr(lowerConstant(GI.getResolver()))));
3494
3495 {
3496 MCInst Adrp;
3497 Adrp.setOpcode(AArch64::ADRP);
3498 Adrp.addOperand(MCOperand::createReg(AArch64::X16));
3499 MCOperand SymPage;
3500 MCInstLowering.lowerOperand(
3501 MachineOperand::CreateES(LazyPointer->getName().data() + 1,
3502 AArch64II::MO_GOT | AArch64II::MO_PAGE),
3503 SymPage);
3504 Adrp.addOperand(SymPage);
3505 EmitToStreamer(Adrp);
3506 }
3507
3508 {
3509 MCInst Ldr;
3510 Ldr.setOpcode(AArch64::LDRXui);
3511 Ldr.addOperand(MCOperand::createReg(AArch64::X16));
3512 Ldr.addOperand(MCOperand::createReg(AArch64::X16));
3513 MCOperand SymPageOff;
3514 MCInstLowering.lowerOperand(
3515 MachineOperand::CreateES(LazyPointer->getName().data() + 1,
3516 AArch64II::MO_GOT | AArch64II::MO_PAGEOFF),
3517 SymPageOff);
3518 Ldr.addOperand(SymPageOff);
3519 Ldr.addOperand(MCOperand::createImm(0));
3520 EmitToStreamer(Ldr);
3521 }
3522
3523 EmitToStreamer(MCInstBuilder(AArch64::STRXui)
3524 .addReg(AArch64::X0)
3525 .addReg(AArch64::X16)
3526 .addImm(0));
3527
3528 EmitToStreamer(MCInstBuilder(AArch64::ADDXri)
3529 .addReg(AArch64::X16)
3530 .addReg(AArch64::X0)
3531 .addImm(0)
3532 .addImm(0));
3533
3534 for (int I = 3; I != -1; --I)
3535 EmitToStreamer(MCInstBuilder(AArch64::LDPDpost)
3536 .addReg(AArch64::SP)
3537 .addReg(AArch64::D1 + 2 * I)
3538 .addReg(AArch64::D0 + 2 * I)
3539 .addReg(AArch64::SP)
3540 .addImm(2));
3541
3542 for (int I = 3; I != -1; --I)
3543 EmitToStreamer(MCInstBuilder(AArch64::LDPXpost)
3544 .addReg(AArch64::SP)
3545 .addReg(AArch64::X1 + 2 * I)
3546 .addReg(AArch64::X0 + 2 * I)
3547 .addReg(AArch64::SP)
3548 .addImm(2));
3549
3550 EmitToStreamer(MCInstBuilder(AArch64::LDPXpost)
3551 .addReg(AArch64::SP)
3552 .addReg(AArch64::FP)
3553 .addReg(AArch64::LR)
3554 .addReg(AArch64::SP)
3555 .addImm(2));
3556
3557 EmitToStreamer(MCInstBuilder(TM.getTargetTriple().isArm64e() ? AArch64::BRAAZ
3558 : AArch64::BR)
3559 .addReg(AArch64::X16));
3560 }
3561
lowerConstant(const Constant * CV,const Constant * BaseCV,uint64_t Offset)3562 const MCExpr *AArch64AsmPrinter::lowerConstant(const Constant *CV,
3563 const Constant *BaseCV,
3564 uint64_t Offset) {
3565 if (const GlobalValue *GV = dyn_cast<GlobalValue>(CV)) {
3566 return MCSymbolRefExpr::create(MCInstLowering.GetGlobalValueSymbol(GV, 0),
3567 OutContext);
3568 }
3569
3570 return AsmPrinter::lowerConstant(CV, BaseCV, Offset);
3571 }
3572
3573 char AArch64AsmPrinter::ID = 0;
3574
3575 INITIALIZE_PASS(AArch64AsmPrinter, "aarch64-asm-printer",
3576 "AArch64 Assembly Printer", false, false)
3577
3578 // Force static initialization.
3579 extern "C" LLVM_ABI LLVM_EXTERNAL_VISIBILITY void
LLVMInitializeAArch64AsmPrinter()3580 LLVMInitializeAArch64AsmPrinter() {
3581 RegisterAsmPrinter<AArch64AsmPrinter> X(getTheAArch64leTarget());
3582 RegisterAsmPrinter<AArch64AsmPrinter> Y(getTheAArch64beTarget());
3583 RegisterAsmPrinter<AArch64AsmPrinter> Z(getTheARM64Target());
3584 RegisterAsmPrinter<AArch64AsmPrinter> W(getTheARM64_32Target());
3585 RegisterAsmPrinter<AArch64AsmPrinter> V(getTheAArch64_32Target());
3586 }
3587