xref: /freebsd/contrib/llvm-project/lld/ELF/Arch/ARM.cpp (revision 2c2ec6bbc9cc7762a250ffe903bda6c2e44d25ff)
1 //===- ARM.cpp ------------------------------------------------------------===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 
9 #include "InputFiles.h"
10 #include "OutputSections.h"
11 #include "SymbolTable.h"
12 #include "Symbols.h"
13 #include "SyntheticSections.h"
14 #include "Target.h"
15 #include "lld/Common/Filesystem.h"
16 #include "llvm/BinaryFormat/ELF.h"
17 #include "llvm/Support/Endian.h"
18 
19 using namespace llvm;
20 using namespace llvm::support::endian;
21 using namespace llvm::support;
22 using namespace llvm::ELF;
23 using namespace lld;
24 using namespace lld::elf;
25 using namespace llvm::object;
26 
27 namespace {
28 class ARM final : public TargetInfo {
29 public:
30   ARM(Ctx &);
31   uint32_t calcEFlags() const override;
32   RelExpr getRelExpr(RelType type, const Symbol &s,
33                      const uint8_t *loc) const override;
34   RelType getDynRel(RelType type) const override;
35   int64_t getImplicitAddend(const uint8_t *buf, RelType type) const override;
36   void writeGotPlt(uint8_t *buf, const Symbol &s) const override;
37   void writeIgotPlt(uint8_t *buf, const Symbol &s) const override;
38   void writePltHeader(uint8_t *buf) const override;
39   void writePlt(uint8_t *buf, const Symbol &sym,
40                 uint64_t pltEntryAddr) const override;
41   void addPltSymbols(InputSection &isec, uint64_t off) const override;
42   void addPltHeaderSymbols(InputSection &isd) const override;
43   bool needsThunk(RelExpr expr, RelType type, const InputFile *file,
44                   uint64_t branchAddr, const Symbol &s,
45                   int64_t a) const override;
46   uint32_t getThunkSectionSpacing() const override;
47   bool inBranchRange(RelType type, uint64_t src, uint64_t dst) const override;
48   void relocate(uint8_t *loc, const Relocation &rel,
49                 uint64_t val) const override;
50 
51   DenseMap<InputSection *, SmallVector<const Defined *, 0>> sectionMap;
52 
53 private:
54   void encodeAluGroup(uint8_t *loc, const Relocation &rel, uint64_t val,
55                       int group, bool check) const;
56 };
57 enum class CodeState { Data = 0, Thumb = 2, Arm = 4 };
58 } // namespace
59 
60 ARM::ARM(Ctx &ctx) : TargetInfo(ctx) {
61   copyRel = R_ARM_COPY;
62   relativeRel = R_ARM_RELATIVE;
63   iRelativeRel = R_ARM_IRELATIVE;
64   gotRel = R_ARM_GLOB_DAT;
65   pltRel = R_ARM_JUMP_SLOT;
66   symbolicRel = R_ARM_ABS32;
67   tlsGotRel = R_ARM_TLS_TPOFF32;
68   tlsModuleIndexRel = R_ARM_TLS_DTPMOD32;
69   tlsOffsetRel = R_ARM_TLS_DTPOFF32;
70   pltHeaderSize = 32;
71   pltEntrySize = 16;
72   ipltEntrySize = 16;
73   trapInstr = {0xd4, 0xd4, 0xd4, 0xd4};
74   needsThunks = true;
75   defaultMaxPageSize = 65536;
76 }
77 
78 uint32_t ARM::calcEFlags() const {
79   // The ABIFloatType is used by loaders to detect the floating point calling
80   // convention.
81   uint32_t abiFloatType = 0;
82 
83   // Set the EF_ARM_BE8 flag in the ELF header, if ELF file is big-endian
84   // with BE-8 code.
85   uint32_t armBE8 = 0;
86 
87   if (ctx.arg.armVFPArgs == ARMVFPArgKind::Base ||
88       ctx.arg.armVFPArgs == ARMVFPArgKind::Default)
89     abiFloatType = EF_ARM_ABI_FLOAT_SOFT;
90   else if (ctx.arg.armVFPArgs == ARMVFPArgKind::VFP)
91     abiFloatType = EF_ARM_ABI_FLOAT_HARD;
92 
93   if (!ctx.arg.isLE && ctx.arg.armBe8)
94     armBE8 = EF_ARM_BE8;
95 
96   // We don't currently use any features incompatible with EF_ARM_EABI_VER5,
97   // but we don't have any firm guarantees of conformance. Linux AArch64
98   // kernels (as of 2016) require an EABI version to be set.
99   return EF_ARM_EABI_VER5 | abiFloatType | armBE8;
100 }
101 
102 RelExpr ARM::getRelExpr(RelType type, const Symbol &s,
103                         const uint8_t *loc) const {
104   switch (type) {
105   case R_ARM_ABS32:
106   case R_ARM_MOVW_ABS_NC:
107   case R_ARM_MOVT_ABS:
108   case R_ARM_THM_MOVW_ABS_NC:
109   case R_ARM_THM_MOVT_ABS:
110   case R_ARM_THM_ALU_ABS_G0_NC:
111   case R_ARM_THM_ALU_ABS_G1_NC:
112   case R_ARM_THM_ALU_ABS_G2_NC:
113   case R_ARM_THM_ALU_ABS_G3:
114     return R_ABS;
115   case R_ARM_THM_JUMP8:
116   case R_ARM_THM_JUMP11:
117     return R_PC;
118   case R_ARM_CALL:
119   case R_ARM_JUMP24:
120   case R_ARM_PC24:
121   case R_ARM_PLT32:
122   case R_ARM_PREL31:
123   case R_ARM_THM_JUMP19:
124   case R_ARM_THM_JUMP24:
125   case R_ARM_THM_CALL:
126     return R_PLT_PC;
127   case R_ARM_GOTOFF32:
128     // (S + A) - GOT_ORG
129     return R_GOTREL;
130   case R_ARM_GOT_BREL:
131     // GOT(S) + A - GOT_ORG
132     return R_GOT_OFF;
133   case R_ARM_GOT_PREL:
134   case R_ARM_TLS_IE32:
135     // GOT(S) + A - P
136     return R_GOT_PC;
137   case R_ARM_SBREL32:
138     return RE_ARM_SBREL;
139   case R_ARM_TARGET1:
140     return ctx.arg.target1Rel ? R_PC : R_ABS;
141   case R_ARM_TARGET2:
142     if (ctx.arg.target2 == Target2Policy::Rel)
143       return R_PC;
144     if (ctx.arg.target2 == Target2Policy::Abs)
145       return R_ABS;
146     return R_GOT_PC;
147   case R_ARM_TLS_GD32:
148     return R_TLSGD_PC;
149   case R_ARM_TLS_LDM32:
150     return R_TLSLD_PC;
151   case R_ARM_TLS_LDO32:
152     return R_DTPREL;
153   case R_ARM_BASE_PREL:
154     // B(S) + A - P
155     // FIXME: currently B(S) assumed to be .got, this may not hold for all
156     // platforms.
157     return R_GOTONLY_PC;
158   case R_ARM_MOVW_PREL_NC:
159   case R_ARM_MOVT_PREL:
160   case R_ARM_REL32:
161   case R_ARM_THM_MOVW_PREL_NC:
162   case R_ARM_THM_MOVT_PREL:
163     return R_PC;
164   case R_ARM_ALU_PC_G0:
165   case R_ARM_ALU_PC_G0_NC:
166   case R_ARM_ALU_PC_G1:
167   case R_ARM_ALU_PC_G1_NC:
168   case R_ARM_ALU_PC_G2:
169   case R_ARM_LDR_PC_G0:
170   case R_ARM_LDR_PC_G1:
171   case R_ARM_LDR_PC_G2:
172   case R_ARM_LDRS_PC_G0:
173   case R_ARM_LDRS_PC_G1:
174   case R_ARM_LDRS_PC_G2:
175   case R_ARM_THM_ALU_PREL_11_0:
176   case R_ARM_THM_PC8:
177   case R_ARM_THM_PC12:
178     return RE_ARM_PCA;
179   case R_ARM_MOVW_BREL_NC:
180   case R_ARM_MOVW_BREL:
181   case R_ARM_MOVT_BREL:
182   case R_ARM_THM_MOVW_BREL_NC:
183   case R_ARM_THM_MOVW_BREL:
184   case R_ARM_THM_MOVT_BREL:
185     return RE_ARM_SBREL;
186   case R_ARM_NONE:
187     return R_NONE;
188   case R_ARM_TLS_LE32:
189     return R_TPREL;
190   case R_ARM_V4BX:
191     // V4BX is just a marker to indicate there's a "bx rN" instruction at the
192     // given address. It can be used to implement a special linker mode which
193     // rewrites ARMv4T inputs to ARMv4. Since we support only ARMv4 input and
194     // not ARMv4 output, we can just ignore it.
195     return R_NONE;
196   default:
197     Err(ctx) << getErrorLoc(ctx, loc) << "unknown relocation (" << type.v
198              << ") against symbol " << &s;
199     return R_NONE;
200   }
201 }
202 
203 RelType ARM::getDynRel(RelType type) const {
204   if ((type == R_ARM_ABS32) || (type == R_ARM_TARGET1 && !ctx.arg.target1Rel))
205     return R_ARM_ABS32;
206   return R_ARM_NONE;
207 }
208 
209 void ARM::writeGotPlt(uint8_t *buf, const Symbol &) const {
210   write32(ctx, buf, ctx.in.plt->getVA());
211 }
212 
213 void ARM::writeIgotPlt(uint8_t *buf, const Symbol &s) const {
214   // An ARM entry is the address of the ifunc resolver function.
215   write32(ctx, buf, s.getVA(ctx));
216 }
217 
218 // Long form PLT Header that does not have any restrictions on the displacement
219 // of the .plt from the .got.plt.
220 static void writePltHeaderLong(Ctx &ctx, uint8_t *buf) {
221   write32(ctx, buf + 0, 0xe52de004);  //     str lr, [sp,#-4]!
222   write32(ctx, buf + 4, 0xe59fe004);  //     ldr lr, L2
223   write32(ctx, buf + 8, 0xe08fe00e);  // L1: add lr, pc, lr
224   write32(ctx, buf + 12, 0xe5bef008); //     ldr pc, [lr, #8]
225   write32(ctx, buf + 16, 0x00000000); // L2: .word   &(.got.plt) - L1 - 8
226   write32(ctx, buf + 20, 0xd4d4d4d4); //     Pad to 32-byte boundary
227   write32(ctx, buf + 24, 0xd4d4d4d4); //     Pad to 32-byte boundary
228   write32(ctx, buf + 28, 0xd4d4d4d4);
229   uint64_t gotPlt = ctx.in.gotPlt->getVA();
230   uint64_t l1 = ctx.in.plt->getVA() + 8;
231   write32(ctx, buf + 16, gotPlt - l1 - 8);
232 }
233 
234 // True if we should use Thumb PLTs, which currently require Thumb2, and are
235 // only used if the target does not have the ARM ISA.
236 static bool useThumbPLTs(Ctx &ctx) {
237   return ctx.arg.armHasThumb2ISA && !ctx.arg.armHasArmISA;
238 }
239 
240 // The default PLT header requires the .got.plt to be within 128 Mb of the
241 // .plt in the positive direction.
242 void ARM::writePltHeader(uint8_t *buf) const {
243   if (useThumbPLTs(ctx)) {
244     // The instruction sequence for thumb:
245     //
246     // 0: b500          push    {lr}
247     // 2: f8df e008     ldr.w   lr, [pc, #0x8]          @ 0xe <func+0xe>
248     // 6: 44fe          add     lr, pc
249     // 8: f85e ff08     ldr     pc, [lr, #8]!
250     // e:               .word   .got.plt - .plt - 16
251     //
252     // At 0x8, we want to jump to .got.plt, the -16 accounts for 8 bytes from
253     // `pc` in the add instruction and 8 bytes for the `lr` adjustment.
254     //
255     uint64_t offset = ctx.in.gotPlt->getVA() - ctx.in.plt->getVA() - 16;
256     assert(llvm::isUInt<32>(offset) && "This should always fit into a 32-bit offset");
257     write16(ctx, buf + 0, 0xb500);
258     // Split into two halves to support endianness correctly.
259     write16(ctx, buf + 2, 0xf8df);
260     write16(ctx, buf + 4, 0xe008);
261     write16(ctx, buf + 6, 0x44fe);
262     // Split into two halves to support endianness correctly.
263     write16(ctx, buf + 8, 0xf85e);
264     write16(ctx, buf + 10, 0xff08);
265     write32(ctx, buf + 12, offset);
266 
267     memcpy(buf + 16, trapInstr.data(), 4);  // Pad to 32-byte boundary
268     memcpy(buf + 20, trapInstr.data(), 4);
269     memcpy(buf + 24, trapInstr.data(), 4);
270     memcpy(buf + 28, trapInstr.data(), 4);
271   } else {
272     // Use a similar sequence to that in writePlt(), the difference is the
273     // calling conventions mean we use lr instead of ip. The PLT entry is
274     // responsible for saving lr on the stack, the dynamic loader is responsible
275     // for reloading it.
276     const uint32_t pltData[] = {
277         0xe52de004, // L1: str lr, [sp,#-4]!
278         0xe28fe600, //     add lr, pc,  #0x0NN00000 &(.got.plt - L1 - 4)
279         0xe28eea00, //     add lr, lr,  #0x000NN000 &(.got.plt - L1 - 4)
280         0xe5bef000, //     ldr pc, [lr, #0x00000NNN] &(.got.plt -L1 - 4)
281     };
282 
283     uint64_t offset = ctx.in.gotPlt->getVA() - ctx.in.plt->getVA() - 4;
284     if (!llvm::isUInt<27>(offset)) {
285       // We cannot encode the Offset, use the long form.
286       writePltHeaderLong(ctx, buf);
287       return;
288     }
289     write32(ctx, buf + 0, pltData[0]);
290     write32(ctx, buf + 4, pltData[1] | ((offset >> 20) & 0xff));
291     write32(ctx, buf + 8, pltData[2] | ((offset >> 12) & 0xff));
292     write32(ctx, buf + 12, pltData[3] | (offset & 0xfff));
293     memcpy(buf + 16, trapInstr.data(), 4); // Pad to 32-byte boundary
294     memcpy(buf + 20, trapInstr.data(), 4);
295     memcpy(buf + 24, trapInstr.data(), 4);
296     memcpy(buf + 28, trapInstr.data(), 4);
297   }
298 }
299 
300 void ARM::addPltHeaderSymbols(InputSection &isec) const {
301   if (useThumbPLTs(ctx)) {
302     addSyntheticLocal(ctx, "$t", STT_NOTYPE, 0, 0, isec);
303     addSyntheticLocal(ctx, "$d", STT_NOTYPE, 12, 0, isec);
304   } else {
305     addSyntheticLocal(ctx, "$a", STT_NOTYPE, 0, 0, isec);
306     addSyntheticLocal(ctx, "$d", STT_NOTYPE, 16, 0, isec);
307   }
308 }
309 
310 // Long form PLT entries that do not have any restrictions on the displacement
311 // of the .plt from the .got.plt.
312 static void writePltLong(Ctx &ctx, uint8_t *buf, uint64_t gotPltEntryAddr,
313                          uint64_t pltEntryAddr) {
314   write32(ctx, buf + 0, 0xe59fc004);  //     ldr ip, L2
315   write32(ctx, buf + 4, 0xe08cc00f);  // L1: add ip, ip, pc
316   write32(ctx, buf + 8, 0xe59cf000);  //     ldr pc, [ip]
317   write32(ctx, buf + 12, 0x00000000); // L2: .word   Offset(&(.got.plt) - L1 - 8
318   uint64_t l1 = pltEntryAddr + 4;
319   write32(ctx, buf + 12, gotPltEntryAddr - l1 - 8);
320 }
321 
322 // The default PLT entries require the .got.plt to be within 128 Mb of the
323 // .plt in the positive direction.
324 void ARM::writePlt(uint8_t *buf, const Symbol &sym,
325                    uint64_t pltEntryAddr) const {
326   if (!useThumbPLTs(ctx)) {
327     uint64_t offset = sym.getGotPltVA(ctx) - pltEntryAddr - 8;
328 
329     // The PLT entry is similar to the example given in Appendix A of ELF for
330     // the Arm Architecture. Instead of using the Group Relocations to find the
331     // optimal rotation for the 8-bit immediate used in the add instructions we
332     // hard code the most compact rotations for simplicity. This saves a load
333     // instruction over the long plt sequences.
334     const uint32_t pltData[] = {
335         0xe28fc600, // L1: add ip, pc,  #0x0NN00000  Offset(&(.got.plt) - L1 - 8
336         0xe28cca00, //     add ip, ip,  #0x000NN000  Offset(&(.got.plt) - L1 - 8
337         0xe5bcf000, //     ldr pc, [ip, #0x00000NNN] Offset(&(.got.plt) - L1 - 8
338     };
339     if (!llvm::isUInt<27>(offset)) {
340       // We cannot encode the Offset, use the long form.
341       writePltLong(ctx, buf, sym.getGotPltVA(ctx), pltEntryAddr);
342       return;
343     }
344     write32(ctx, buf + 0, pltData[0] | ((offset >> 20) & 0xff));
345     write32(ctx, buf + 4, pltData[1] | ((offset >> 12) & 0xff));
346     write32(ctx, buf + 8, pltData[2] | (offset & 0xfff));
347     memcpy(buf + 12, trapInstr.data(), 4); // Pad to 16-byte boundary
348   } else {
349     uint64_t offset = sym.getGotPltVA(ctx) - pltEntryAddr - 12;
350     assert(llvm::isUInt<32>(offset) && "This should always fit into a 32-bit offset");
351 
352     // A PLT entry will be:
353     //
354     //       movw ip, #<lower 16 bits>
355     //       movt ip, #<upper 16 bits>
356     //       add ip, pc
357     //   L1: ldr.w pc, [ip]
358     //       b L1
359     //
360     // where ip = r12 = 0xc
361 
362     // movw ip, #<lower 16 bits>
363     write16(ctx, buf + 2, 0x0c00); // use `ip`
364     relocateNoSym(buf, R_ARM_THM_MOVW_ABS_NC, offset);
365 
366     // movt ip, #<upper 16 bits>
367     write16(ctx, buf + 6, 0x0c00); // use `ip`
368     relocateNoSym(buf + 4, R_ARM_THM_MOVT_ABS, offset);
369 
370     write16(ctx, buf + 8, 0x44fc);  // add ip, pc
371     write16(ctx, buf + 10, 0xf8dc); // ldr.w   pc, [ip] (bottom half)
372     write16(ctx, buf + 12, 0xf000); // ldr.w   pc, [ip] (upper half)
373     write16(ctx, buf + 14, 0xe7fc); // Branch to previous instruction
374   }
375 }
376 
377 void ARM::addPltSymbols(InputSection &isec, uint64_t off) const {
378   if (useThumbPLTs(ctx)) {
379     addSyntheticLocal(ctx, "$t", STT_NOTYPE, off, 0, isec);
380   } else {
381     addSyntheticLocal(ctx, "$a", STT_NOTYPE, off, 0, isec);
382     addSyntheticLocal(ctx, "$d", STT_NOTYPE, off + 12, 0, isec);
383   }
384 }
385 
386 bool ARM::needsThunk(RelExpr expr, RelType type, const InputFile *file,
387                      uint64_t branchAddr, const Symbol &s,
388                      int64_t a) const {
389   // If s is an undefined weak symbol and does not have a PLT entry then it will
390   // be resolved as a branch to the next instruction. If it is hidden, its
391   // binding has been converted to local, so we just check isUndefined() here. A
392   // undefined non-weak symbol will have been errored.
393   if (s.isUndefined() && !s.isInPlt(ctx))
394     return false;
395   // A state change from ARM to Thumb and vice versa must go through an
396   // interworking thunk if the relocation type is not R_ARM_CALL or
397   // R_ARM_THM_CALL.
398   switch (type) {
399   case R_ARM_PC24:
400   case R_ARM_PLT32:
401   case R_ARM_JUMP24:
402     // Source is ARM, all PLT entries are ARM so no interworking required.
403     // Otherwise we need to interwork if STT_FUNC Symbol has bit 0 set (Thumb).
404     assert(!useThumbPLTs(ctx) &&
405            "If the source is ARM, we should not need Thumb PLTs");
406     if (s.isFunc() && expr == R_PC && (s.getVA(ctx) & 1))
407       return true;
408     [[fallthrough]];
409   case R_ARM_CALL: {
410     uint64_t dst = (expr == R_PLT_PC) ? s.getPltVA(ctx) : s.getVA(ctx);
411     return !inBranchRange(type, branchAddr, dst + a) ||
412            (!ctx.arg.armHasBlx && (s.getVA(ctx) & 1));
413   }
414   case R_ARM_THM_JUMP19:
415   case R_ARM_THM_JUMP24:
416     // Source is Thumb, when all PLT entries are ARM interworking is required.
417     // Otherwise we need to interwork if STT_FUNC Symbol has bit 0 clear (ARM).
418     if ((expr == R_PLT_PC && !useThumbPLTs(ctx)) ||
419         (s.isFunc() && (s.getVA(ctx) & 1) == 0))
420       return true;
421     [[fallthrough]];
422   case R_ARM_THM_CALL: {
423     uint64_t dst = (expr == R_PLT_PC) ? s.getPltVA(ctx) : s.getVA(ctx);
424     return !inBranchRange(type, branchAddr, dst + a) ||
425            (!ctx.arg.armHasBlx && (s.getVA(ctx) & 1) == 0);
426   }
427   }
428   return false;
429 }
430 
431 uint32_t ARM::getThunkSectionSpacing() const {
432   // The placing of pre-created ThunkSections is controlled by the value
433   // thunkSectionSpacing returned by getThunkSectionSpacing(). The aim is to
434   // place the ThunkSection such that all branches from the InputSections
435   // prior to the ThunkSection can reach a Thunk placed at the end of the
436   // ThunkSection. Graphically:
437   // | up to thunkSectionSpacing .text input sections |
438   // | ThunkSection                                   |
439   // | up to thunkSectionSpacing .text input sections |
440   // | ThunkSection                                   |
441 
442   // Pre-created ThunkSections are spaced roughly 16MiB apart on ARMv7. This
443   // is to match the most common expected case of a Thumb 2 encoded BL, BLX or
444   // B.W:
445   // ARM B, BL, BLX range +/- 32MiB
446   // Thumb B.W, BL, BLX range +/- 16MiB
447   // Thumb B<cc>.W range +/- 1MiB
448   // If a branch cannot reach a pre-created ThunkSection a new one will be
449   // created so we can handle the rare cases of a Thumb 2 conditional branch.
450   // We intentionally use a lower size for thunkSectionSpacing than the maximum
451   // branch range so the end of the ThunkSection is more likely to be within
452   // range of the branch instruction that is furthest away. The value we shorten
453   // thunkSectionSpacing by is set conservatively to allow us to create 16,384
454   // 12 byte Thunks at any offset in a ThunkSection without risk of a branch to
455   // one of the Thunks going out of range.
456 
457   // On Arm the thunkSectionSpacing depends on the range of the Thumb Branch
458   // range. On earlier Architectures such as ARMv4, ARMv5 and ARMv6 (except
459   // ARMv6T2) the range is +/- 4MiB.
460 
461   return (ctx.arg.armJ1J2BranchEncoding) ? 0x1000000 - 0x30000
462                                          : 0x400000 - 0x7500;
463 }
464 
465 bool ARM::inBranchRange(RelType type, uint64_t src, uint64_t dst) const {
466   if ((dst & 0x1) == 0)
467     // Destination is ARM, if ARM caller then Src is already 4-byte aligned.
468     // If Thumb Caller (BLX) the Src address has bottom 2 bits cleared to ensure
469     // destination will be 4 byte aligned.
470     src &= ~0x3;
471   else
472     // Bit 0 == 1 denotes Thumb state, it is not part of the range.
473     dst &= ~0x1;
474 
475   int64_t offset = dst - src;
476   switch (type) {
477   case R_ARM_PC24:
478   case R_ARM_PLT32:
479   case R_ARM_JUMP24:
480   case R_ARM_CALL:
481     return llvm::isInt<26>(offset);
482   case R_ARM_THM_JUMP19:
483     return llvm::isInt<21>(offset);
484   case R_ARM_THM_JUMP24:
485   case R_ARM_THM_CALL:
486     return ctx.arg.armJ1J2BranchEncoding ? llvm::isInt<25>(offset)
487                                          : llvm::isInt<23>(offset);
488   default:
489     return true;
490   }
491 }
492 
493 // Helper to produce message text when LLD detects that a CALL relocation to
494 // a non STT_FUNC symbol that may result in incorrect interworking between ARM
495 // or Thumb.
496 static void stateChangeWarning(Ctx &ctx, uint8_t *loc, RelType relt,
497                                const Symbol &s) {
498   assert(!s.isFunc());
499   const ErrorPlace place = getErrorPlace(ctx, loc);
500   std::string hint;
501   if (!place.srcLoc.empty())
502     hint = "; " + place.srcLoc;
503   if (s.isSection()) {
504     // Section symbols must be defined and in a section. Users cannot change
505     // the type. Use the section name as getName() returns an empty string.
506     Warn(ctx) << place.loc << "branch and link relocation: " << relt
507               << " to STT_SECTION symbol " << cast<Defined>(s).section->name
508               << " ; interworking not performed" << hint;
509   } else {
510     // Warn with hint on how to alter the symbol type.
511     Warn(ctx)
512         << getErrorLoc(ctx, loc) << "branch and link relocation: " << relt
513         << " to non STT_FUNC symbol: " << s.getName()
514         << " interworking not performed; consider using directive '.type "
515         << s.getName()
516         << ", %function' to give symbol type STT_FUNC if interworking between "
517            "ARM and Thumb is required"
518         << hint;
519   }
520 }
521 
522 // Rotate a 32-bit unsigned value right by a specified amt of bits.
523 static uint32_t rotr32(uint32_t val, uint32_t amt) {
524   assert(amt < 32 && "Invalid rotate amount");
525   return (val >> amt) | (val << ((32 - amt) & 31));
526 }
527 
528 static std::pair<uint32_t, uint32_t> getRemAndLZForGroup(unsigned group,
529                                                          uint32_t val) {
530   uint32_t rem, lz;
531   do {
532     lz = llvm::countl_zero(val) & ~1;
533     rem = val;
534     if (lz == 32) // implies rem == 0
535       break;
536     val &= 0xffffff >> lz;
537   } while (group--);
538   return {rem, lz};
539 }
540 
541 void ARM::encodeAluGroup(uint8_t *loc, const Relocation &rel, uint64_t val,
542                            int group, bool check) const {
543   // ADD/SUB (immediate) add = bit23, sub = bit22
544   // immediate field carries is a 12-bit modified immediate, made up of a 4-bit
545   // even rotate right and an 8-bit immediate.
546   uint32_t opcode = 0x00800000;
547   if (val >> 63) {
548     opcode = 0x00400000;
549     val = -val;
550   }
551   uint32_t imm, lz;
552   std::tie(imm, lz) = getRemAndLZForGroup(group, val);
553   uint32_t rot = 0;
554   if (lz < 24) {
555     imm = rotr32(imm, 24 - lz);
556     rot = (lz + 8) << 7;
557   }
558   if (check && imm > 0xff)
559     Err(ctx) << getErrorLoc(ctx, loc) << "unencodeable immediate " << val
560              << " for relocation " << rel.type;
561   write32(ctx, loc,
562           (read32(ctx, loc) & 0xff3ff000) | opcode | rot | (imm & 0xff));
563 }
564 
565 static void encodeLdrGroup(Ctx &ctx, uint8_t *loc, const Relocation &rel,
566                            uint64_t val, int group) {
567   // R_ARM_LDR_PC_Gn is S + A - P, we have ((S + A) | T) - P, if S is a
568   // function then addr is 0 (modulo 2) and Pa is 0 (modulo 4) so we can clear
569   // bottom bit to recover S + A - P.
570   if (rel.sym->isFunc())
571     val &= ~0x1;
572   // LDR (literal) u = bit23
573   uint32_t opcode = 0x00800000;
574   if (val >> 63) {
575     opcode = 0x0;
576     val = -val;
577   }
578   uint32_t imm = getRemAndLZForGroup(group, val).first;
579   checkUInt(ctx, loc, imm, 12, rel);
580   write32(ctx, loc, (read32(ctx, loc) & 0xff7ff000) | opcode | imm);
581 }
582 
583 static void encodeLdrsGroup(Ctx &ctx, uint8_t *loc, const Relocation &rel,
584                             uint64_t val, int group) {
585   // R_ARM_LDRS_PC_Gn is S + A - P, we have ((S + A) | T) - P, if S is a
586   // function then addr is 0 (modulo 2) and Pa is 0 (modulo 4) so we can clear
587   // bottom bit to recover S + A - P.
588   if (rel.sym->isFunc())
589     val &= ~0x1;
590   // LDRD/LDRH/LDRSB/LDRSH (literal) u = bit23
591   uint32_t opcode = 0x00800000;
592   if (val >> 63) {
593     opcode = 0x0;
594     val = -val;
595   }
596   uint32_t imm = getRemAndLZForGroup(group, val).first;
597   checkUInt(ctx, loc, imm, 8, rel);
598   write32(ctx, loc,
599           (read32(ctx, loc) & 0xff7ff0f0) | opcode | ((imm & 0xf0) << 4) |
600               (imm & 0xf));
601 }
602 
603 void ARM::relocate(uint8_t *loc, const Relocation &rel, uint64_t val) const {
604   switch (rel.type) {
605   case R_ARM_ABS32:
606   case R_ARM_BASE_PREL:
607   case R_ARM_GOTOFF32:
608   case R_ARM_GOT_BREL:
609   case R_ARM_GOT_PREL:
610   case R_ARM_REL32:
611   case R_ARM_RELATIVE:
612   case R_ARM_SBREL32:
613   case R_ARM_TARGET1:
614   case R_ARM_TARGET2:
615   case R_ARM_TLS_GD32:
616   case R_ARM_TLS_IE32:
617   case R_ARM_TLS_LDM32:
618   case R_ARM_TLS_LDO32:
619   case R_ARM_TLS_LE32:
620   case R_ARM_TLS_TPOFF32:
621   case R_ARM_TLS_DTPOFF32:
622     write32(ctx, loc, val);
623     break;
624   case R_ARM_PREL31:
625     checkInt(ctx, loc, val, 31, rel);
626     write32(ctx, loc, (read32(ctx, loc) & 0x80000000) | (val & ~0x80000000));
627     break;
628   case R_ARM_CALL: {
629     // R_ARM_CALL is used for BL and BLX instructions, for symbols of type
630     // STT_FUNC we choose whether to write a BL or BLX depending on the
631     // value of bit 0 of Val. With bit 0 == 1 denoting Thumb. If the symbol is
632     // not of type STT_FUNC then we must preserve the original instruction.
633     assert(rel.sym); // R_ARM_CALL is always reached via relocate().
634     bool bit0Thumb = val & 1;
635     bool isBlx = (read32(ctx, loc) & 0xfe000000) == 0xfa000000;
636     // lld 10.0 and before always used bit0Thumb when deciding to write a BLX
637     // even when type not STT_FUNC.
638     if (!rel.sym->isFunc() && isBlx != bit0Thumb)
639       stateChangeWarning(ctx, loc, rel.type, *rel.sym);
640     if (rel.sym->isFunc() ? bit0Thumb : isBlx) {
641       // The BLX encoding is 0xfa:H:imm24 where Val = imm24:H:'1'
642       checkInt(ctx, loc, val, 26, rel);
643       write32(ctx, loc,
644               0xfa000000 |                    // opcode
645                   ((val & 2) << 23) |         // H
646                   ((val >> 2) & 0x00ffffff)); // imm24
647       break;
648     }
649     // BLX (always unconditional) instruction to an ARM Target, select an
650     // unconditional BL.
651     write32(ctx, loc, 0xeb000000 | (read32(ctx, loc) & 0x00ffffff));
652     // fall through as BL encoding is shared with B
653   }
654     [[fallthrough]];
655   case R_ARM_JUMP24:
656   case R_ARM_PC24:
657   case R_ARM_PLT32:
658     checkInt(ctx, loc, val, 26, rel);
659     write32(ctx, loc,
660             (read32(ctx, loc) & ~0x00ffffff) | ((val >> 2) & 0x00ffffff));
661     break;
662   case R_ARM_THM_JUMP8:
663     // We do a 9 bit check because val is right-shifted by 1 bit.
664     checkInt(ctx, loc, val, 9, rel);
665     write16(ctx, loc, (read16(ctx, loc) & 0xff00) | ((val >> 1) & 0x00ff));
666     break;
667   case R_ARM_THM_JUMP11:
668     // We do a 12 bit check because val is right-shifted by 1 bit.
669     checkInt(ctx, loc, val, 12, rel);
670     write16(ctx, loc, (read16(ctx, loc) & 0xf800) | ((val >> 1) & 0x07ff));
671     break;
672   case R_ARM_THM_JUMP19:
673     // Encoding T3: Val = S:J2:J1:imm6:imm11:0
674     checkInt(ctx, loc, val, 21, rel);
675     write16(ctx, loc,
676             (read16(ctx, loc) & 0xfbc0) | // opcode cond
677                 ((val >> 10) & 0x0400) |  // S
678                 ((val >> 12) & 0x003f));  // imm6
679     write16(ctx, loc + 2,
680             0x8000 |                    // opcode
681                 ((val >> 8) & 0x0800) | // J2
682                 ((val >> 5) & 0x2000) | // J1
683                 ((val >> 1) & 0x07ff)); // imm11
684     break;
685   case R_ARM_THM_CALL: {
686     // R_ARM_THM_CALL is used for BL and BLX instructions, for symbols of type
687     // STT_FUNC we choose whether to write a BL or BLX depending on the
688     // value of bit 0 of Val. With bit 0 == 0 denoting ARM, if the symbol is
689     // not of type STT_FUNC then we must preserve the original instruction.
690     // PLT entries are always ARM state so we know we need to interwork.
691     assert(rel.sym); // R_ARM_THM_CALL is always reached via relocate().
692     bool bit0Thumb = val & 1;
693     bool useThumb = bit0Thumb || useThumbPLTs(ctx);
694     bool isBlx = (read16(ctx, loc + 2) & 0x1000) == 0;
695     // lld 10.0 and before always used bit0Thumb when deciding to write a BLX
696     // even when type not STT_FUNC.
697     if (!rel.sym->isFunc() && !rel.sym->isInPlt(ctx) && isBlx == useThumb)
698       stateChangeWarning(ctx, loc, rel.type, *rel.sym);
699     if ((rel.sym->isFunc() || rel.sym->isInPlt(ctx)) ? !useThumb : isBlx) {
700       // We are writing a BLX. Ensure BLX destination is 4-byte aligned. As
701       // the BLX instruction may only be two byte aligned. This must be done
702       // before overflow check.
703       val = alignTo(val, 4);
704       write16(ctx, loc + 2, read16(ctx, loc + 2) & ~0x1000);
705     } else {
706       write16(ctx, loc + 2, (read16(ctx, loc + 2) & ~0x1000) | 1 << 12);
707     }
708     if (!ctx.arg.armJ1J2BranchEncoding) {
709       // Older Arm architectures do not support R_ARM_THM_JUMP24 and have
710       // different encoding rules and range due to J1 and J2 always being 1.
711       checkInt(ctx, loc, val, 23, rel);
712       write16(ctx, loc,
713               0xf000 |                     // opcode
714                   ((val >> 12) & 0x07ff)); // imm11
715       write16(ctx, loc + 2,
716               (read16(ctx, loc + 2) & 0xd000) | // opcode
717                   0x2800 |                      // J1 == J2 == 1
718                   ((val >> 1) & 0x07ff));       // imm11
719       break;
720     }
721   }
722     // Fall through as rest of encoding is the same as B.W
723     [[fallthrough]];
724   case R_ARM_THM_JUMP24:
725     // Encoding B  T4, BL T1, BLX T2: Val = S:I1:I2:imm10:imm11:0
726     checkInt(ctx, loc, val, 25, rel);
727     write16(ctx, loc,
728             0xf000 |                     // opcode
729                 ((val >> 14) & 0x0400) | // S
730                 ((val >> 12) & 0x03ff)); // imm10
731     write16(ctx, loc + 2,
732             (read16(ctx, loc + 2) & 0xd000) |               // opcode
733                 (((~(val >> 10)) ^ (val >> 11)) & 0x2000) | // J1
734                 (((~(val >> 11)) ^ (val >> 13)) & 0x0800) | // J2
735                 ((val >> 1) & 0x07ff));                     // imm11
736     break;
737   case R_ARM_MOVW_ABS_NC:
738   case R_ARM_MOVW_PREL_NC:
739   case R_ARM_MOVW_BREL_NC:
740     write32(ctx, loc,
741             (read32(ctx, loc) & ~0x000f0fff) | ((val & 0xf000) << 4) |
742                 (val & 0x0fff));
743     break;
744   case R_ARM_MOVT_ABS:
745   case R_ARM_MOVT_PREL:
746   case R_ARM_MOVT_BREL:
747     write32(ctx, loc,
748             (read32(ctx, loc) & ~0x000f0fff) | (((val >> 16) & 0xf000) << 4) |
749                 ((val >> 16) & 0xfff));
750     break;
751   case R_ARM_THM_MOVT_ABS:
752   case R_ARM_THM_MOVT_PREL:
753   case R_ARM_THM_MOVT_BREL:
754     // Encoding T1: A = imm4:i:imm3:imm8
755 
756     write16(ctx, loc,
757             0xf2c0 |                     // opcode
758                 ((val >> 17) & 0x0400) | // i
759                 ((val >> 28) & 0x000f)); // imm4
760 
761     write16(ctx, loc + 2,
762             (read16(ctx, loc + 2) & 0x8f00) | // opcode
763                 ((val >> 12) & 0x7000) |      // imm3
764                 ((val >> 16) & 0x00ff));      // imm8
765     break;
766   case R_ARM_THM_MOVW_ABS_NC:
767   case R_ARM_THM_MOVW_PREL_NC:
768   case R_ARM_THM_MOVW_BREL_NC:
769     // Encoding T3: A = imm4:i:imm3:imm8
770     write16(ctx, loc,
771             0xf240 |                     // opcode
772                 ((val >> 1) & 0x0400) |  // i
773                 ((val >> 12) & 0x000f)); // imm4
774     write16(ctx, loc + 2,
775             (read16(ctx, loc + 2) & 0x8f00) | // opcode
776                 ((val << 4) & 0x7000) |       // imm3
777                 (val & 0x00ff));              // imm8
778     break;
779   case R_ARM_THM_ALU_ABS_G3:
780     write16(ctx, loc, (read16(ctx, loc) & ~0x00ff) | ((val >> 24) & 0x00ff));
781     break;
782   case R_ARM_THM_ALU_ABS_G2_NC:
783     write16(ctx, loc, (read16(ctx, loc) & ~0x00ff) | ((val >> 16) & 0x00ff));
784     break;
785   case R_ARM_THM_ALU_ABS_G1_NC:
786     write16(ctx, loc, (read16(ctx, loc) & ~0x00ff) | ((val >> 8) & 0x00ff));
787     break;
788   case R_ARM_THM_ALU_ABS_G0_NC:
789     write16(ctx, loc, (read16(ctx, loc) & ~0x00ff) | (val & 0x00ff));
790     break;
791   case R_ARM_ALU_PC_G0:
792     encodeAluGroup(loc, rel, val, 0, true);
793     break;
794   case R_ARM_ALU_PC_G0_NC:
795     encodeAluGroup(loc, rel, val, 0, false);
796     break;
797   case R_ARM_ALU_PC_G1:
798     encodeAluGroup(loc, rel, val, 1, true);
799     break;
800   case R_ARM_ALU_PC_G1_NC:
801     encodeAluGroup(loc, rel, val, 1, false);
802     break;
803   case R_ARM_ALU_PC_G2:
804     encodeAluGroup(loc, rel, val, 2, true);
805     break;
806   case R_ARM_LDR_PC_G0:
807     encodeLdrGroup(ctx, loc, rel, val, 0);
808     break;
809   case R_ARM_LDR_PC_G1:
810     encodeLdrGroup(ctx, loc, rel, val, 1);
811     break;
812   case R_ARM_LDR_PC_G2:
813     encodeLdrGroup(ctx, loc, rel, val, 2);
814     break;
815   case R_ARM_LDRS_PC_G0:
816     encodeLdrsGroup(ctx, loc, rel, val, 0);
817     break;
818   case R_ARM_LDRS_PC_G1:
819     encodeLdrsGroup(ctx, loc, rel, val, 1);
820     break;
821   case R_ARM_LDRS_PC_G2:
822     encodeLdrsGroup(ctx, loc, rel, val, 2);
823     break;
824   case R_ARM_THM_ALU_PREL_11_0: {
825     // ADR encoding T2 (sub), T3 (add) i:imm3:imm8
826     int64_t imm = val;
827     uint16_t sub = 0;
828     if (imm < 0) {
829       imm = -imm;
830       sub = 0x00a0;
831     }
832     checkUInt(ctx, loc, imm, 12, rel);
833     write16(ctx, loc, (read16(ctx, loc) & 0xfb0f) | sub | (imm & 0x800) >> 1);
834     write16(ctx, loc + 2,
835             (read16(ctx, loc + 2) & 0x8f00) | (imm & 0x700) << 4 |
836                 (imm & 0xff));
837     break;
838   }
839   case R_ARM_THM_PC8:
840     // ADR and LDR literal encoding T1 positive offset only imm8:00
841     // R_ARM_THM_PC8 is S + A - Pa, we have ((S + A) | T) - Pa, if S is a
842     // function then addr is 0 (modulo 2) and Pa is 0 (modulo 4) so we can clear
843     // bottom bit to recover S + A - Pa.
844     if (rel.sym->isFunc())
845       val &= ~0x1;
846     checkUInt(ctx, loc, val, 10, rel);
847     checkAlignment(ctx, loc, val, 4, rel);
848     write16(ctx, loc, (read16(ctx, loc) & 0xff00) | (val & 0x3fc) >> 2);
849     break;
850   case R_ARM_THM_PC12: {
851     // LDR (literal) encoding T2, add = (U == '1') imm12
852     // imm12 is unsigned
853     // R_ARM_THM_PC12 is S + A - Pa, we have ((S + A) | T) - Pa, if S is a
854     // function then addr is 0 (modulo 2) and Pa is 0 (modulo 4) so we can clear
855     // bottom bit to recover S + A - Pa.
856     if (rel.sym->isFunc())
857       val &= ~0x1;
858     int64_t imm12 = val;
859     uint16_t u = 0x0080;
860     if (imm12 < 0) {
861       imm12 = -imm12;
862       u = 0;
863     }
864     checkUInt(ctx, loc, imm12, 12, rel);
865     write16(ctx, loc, read16(ctx, loc) | u);
866     write16(ctx, loc + 2, (read16(ctx, loc + 2) & 0xf000) | imm12);
867     break;
868   }
869   default:
870     llvm_unreachable("unknown relocation");
871   }
872 }
873 
874 int64_t ARM::getImplicitAddend(const uint8_t *buf, RelType type) const {
875   switch (type) {
876   default:
877     InternalErr(ctx, buf) << "cannot read addend for relocation " << type;
878     return 0;
879   case R_ARM_ABS32:
880   case R_ARM_BASE_PREL:
881   case R_ARM_GLOB_DAT:
882   case R_ARM_GOTOFF32:
883   case R_ARM_GOT_BREL:
884   case R_ARM_GOT_PREL:
885   case R_ARM_IRELATIVE:
886   case R_ARM_REL32:
887   case R_ARM_RELATIVE:
888   case R_ARM_SBREL32:
889   case R_ARM_TARGET1:
890   case R_ARM_TARGET2:
891   case R_ARM_TLS_DTPMOD32:
892   case R_ARM_TLS_DTPOFF32:
893   case R_ARM_TLS_GD32:
894   case R_ARM_TLS_IE32:
895   case R_ARM_TLS_LDM32:
896   case R_ARM_TLS_LE32:
897   case R_ARM_TLS_LDO32:
898   case R_ARM_TLS_TPOFF32:
899     return SignExtend64<32>(read32(ctx, buf));
900   case R_ARM_PREL31:
901     return SignExtend64<31>(read32(ctx, buf));
902   case R_ARM_CALL:
903   case R_ARM_JUMP24:
904   case R_ARM_PC24:
905   case R_ARM_PLT32:
906     return SignExtend64<26>(read32(ctx, buf) << 2);
907   case R_ARM_THM_JUMP8:
908     return SignExtend64<9>(read16(ctx, buf) << 1);
909   case R_ARM_THM_JUMP11:
910     return SignExtend64<12>(read16(ctx, buf) << 1);
911   case R_ARM_THM_JUMP19: {
912     // Encoding T3: A = S:J2:J1:imm10:imm6:0
913     uint16_t hi = read16(ctx, buf);
914     uint16_t lo = read16(ctx, buf + 2);
915     return SignExtend64<20>(((hi & 0x0400) << 10) | // S
916                             ((lo & 0x0800) << 8) |  // J2
917                             ((lo & 0x2000) << 5) |  // J1
918                             ((hi & 0x003f) << 12) | // imm6
919                             ((lo & 0x07ff) << 1));  // imm11:0
920   }
921   case R_ARM_THM_CALL:
922     if (!ctx.arg.armJ1J2BranchEncoding) {
923       // Older Arm architectures do not support R_ARM_THM_JUMP24 and have
924       // different encoding rules and range due to J1 and J2 always being 1.
925       uint16_t hi = read16(ctx, buf);
926       uint16_t lo = read16(ctx, buf + 2);
927       return SignExtend64<22>(((hi & 0x7ff) << 12) | // imm11
928                               ((lo & 0x7ff) << 1));  // imm11:0
929       break;
930     }
931     [[fallthrough]];
932   case R_ARM_THM_JUMP24: {
933     // Encoding B T4, BL T1, BLX T2: A = S:I1:I2:imm10:imm11:0
934     // I1 = NOT(J1 EOR S), I2 = NOT(J2 EOR S)
935     uint16_t hi = read16(ctx, buf);
936     uint16_t lo = read16(ctx, buf + 2);
937     return SignExtend64<24>(((hi & 0x0400) << 14) |                    // S
938                             (~((lo ^ (hi << 3)) << 10) & 0x00800000) | // I1
939                             (~((lo ^ (hi << 1)) << 11) & 0x00400000) | // I2
940                             ((hi & 0x003ff) << 12) |                   // imm0
941                             ((lo & 0x007ff) << 1)); // imm11:0
942   }
943   // ELF for the ARM Architecture 4.6.1.1 the implicit addend for MOVW and
944   // MOVT is in the range -32768 <= A < 32768
945   case R_ARM_MOVW_ABS_NC:
946   case R_ARM_MOVT_ABS:
947   case R_ARM_MOVW_PREL_NC:
948   case R_ARM_MOVT_PREL:
949   case R_ARM_MOVW_BREL_NC:
950   case R_ARM_MOVT_BREL: {
951     uint64_t val = read32(ctx, buf) & 0x000f0fff;
952     return SignExtend64<16>(((val & 0x000f0000) >> 4) | (val & 0x00fff));
953   }
954   case R_ARM_THM_MOVW_ABS_NC:
955   case R_ARM_THM_MOVT_ABS:
956   case R_ARM_THM_MOVW_PREL_NC:
957   case R_ARM_THM_MOVT_PREL:
958   case R_ARM_THM_MOVW_BREL_NC:
959   case R_ARM_THM_MOVT_BREL: {
960     // Encoding T3: A = imm4:i:imm3:imm8
961     uint16_t hi = read16(ctx, buf);
962     uint16_t lo = read16(ctx, buf + 2);
963     return SignExtend64<16>(((hi & 0x000f) << 12) | // imm4
964                             ((hi & 0x0400) << 1) |  // i
965                             ((lo & 0x7000) >> 4) |  // imm3
966                             (lo & 0x00ff));         // imm8
967   }
968   case R_ARM_THM_ALU_ABS_G0_NC:
969   case R_ARM_THM_ALU_ABS_G1_NC:
970   case R_ARM_THM_ALU_ABS_G2_NC:
971   case R_ARM_THM_ALU_ABS_G3:
972     return read16(ctx, buf) & 0xff;
973   case R_ARM_ALU_PC_G0:
974   case R_ARM_ALU_PC_G0_NC:
975   case R_ARM_ALU_PC_G1:
976   case R_ARM_ALU_PC_G1_NC:
977   case R_ARM_ALU_PC_G2: {
978     // 12-bit immediate is a modified immediate made up of a 4-bit even
979     // right rotation and 8-bit constant. After the rotation the value
980     // is zero-extended. When bit 23 is set the instruction is an add, when
981     // bit 22 is set it is a sub.
982     uint32_t instr = read32(ctx, buf);
983     uint32_t val = rotr32(instr & 0xff, ((instr & 0xf00) >> 8) * 2);
984     return (instr & 0x00400000) ? -val : val;
985   }
986   case R_ARM_LDR_PC_G0:
987   case R_ARM_LDR_PC_G1:
988   case R_ARM_LDR_PC_G2: {
989     // ADR (literal) add = bit23, sub = bit22
990     // LDR (literal) u = bit23 unsigned imm12
991     bool u = read32(ctx, buf) & 0x00800000;
992     uint32_t imm12 = read32(ctx, buf) & 0xfff;
993     return u ? imm12 : -imm12;
994   }
995   case R_ARM_LDRS_PC_G0:
996   case R_ARM_LDRS_PC_G1:
997   case R_ARM_LDRS_PC_G2: {
998     // LDRD/LDRH/LDRSB/LDRSH (literal) u = bit23 unsigned imm8
999     uint32_t opcode = read32(ctx, buf);
1000     bool u = opcode & 0x00800000;
1001     uint32_t imm4l = opcode & 0xf;
1002     uint32_t imm4h = (opcode & 0xf00) >> 4;
1003     return u ? (imm4h | imm4l) : -(imm4h | imm4l);
1004   }
1005   case R_ARM_THM_ALU_PREL_11_0: {
1006     // Thumb2 ADR, which is an alias for a sub or add instruction with an
1007     // unsigned immediate.
1008     // ADR encoding T2 (sub), T3 (add) i:imm3:imm8
1009     uint16_t hi = read16(ctx, buf);
1010     uint16_t lo = read16(ctx, buf + 2);
1011     uint64_t imm = (hi & 0x0400) << 1 | // i
1012                    (lo & 0x7000) >> 4 | // imm3
1013                    (lo & 0x00ff);       // imm8
1014     // For sub, addend is negative, add is positive.
1015     return (hi & 0x00f0) ? -imm : imm;
1016   }
1017   case R_ARM_THM_PC8:
1018     // ADR and LDR (literal) encoding T1
1019     // From ELF for the ARM Architecture the initial signed addend is formed
1020     // from an unsigned field using expression (((imm8:00 + 4) & 0x3ff) – 4)
1021     // this trick permits the PC bias of -4 to be encoded using imm8 = 0xff
1022     return ((((read16(ctx, buf) & 0xff) << 2) + 4) & 0x3ff) - 4;
1023   case R_ARM_THM_PC12: {
1024     // LDR (literal) encoding T2, add = (U == '1') imm12
1025     bool u = read16(ctx, buf) & 0x0080;
1026     uint64_t imm12 = read16(ctx, buf + 2) & 0x0fff;
1027     return u ? imm12 : -imm12;
1028   }
1029   case R_ARM_NONE:
1030   case R_ARM_V4BX:
1031   case R_ARM_JUMP_SLOT:
1032     // These relocations are defined as not having an implicit addend.
1033     return 0;
1034   }
1035 }
1036 
1037 static bool isArmMapSymbol(const Symbol *b) {
1038   return b->getName() == "$a" || b->getName().starts_with("$a.");
1039 }
1040 
1041 static bool isThumbMapSymbol(const Symbol *s) {
1042   return s->getName() == "$t" || s->getName().starts_with("$t.");
1043 }
1044 
1045 static bool isDataMapSymbol(const Symbol *b) {
1046   return b->getName() == "$d" || b->getName().starts_with("$d.");
1047 }
1048 
1049 void elf::sortArmMappingSymbols(Ctx &ctx) {
1050   // For each input section make sure the mapping symbols are sorted in
1051   // ascending order.
1052   for (auto &kv : static_cast<ARM &>(*ctx.target).sectionMap) {
1053     SmallVector<const Defined *, 0> &mapSyms = kv.second;
1054     llvm::stable_sort(mapSyms, [](const Defined *a, const Defined *b) {
1055       return a->value < b->value;
1056     });
1057   }
1058 }
1059 
1060 void elf::addArmInputSectionMappingSymbols(Ctx &ctx) {
1061   // Collect mapping symbols for every executable input sections.
1062   // The linker generated mapping symbols for all the synthetic
1063   // sections are adding into the sectionmap through the function
1064   // addArmSyntheitcSectionMappingSymbol.
1065   auto &sectionMap = static_cast<ARM &>(*ctx.target).sectionMap;
1066   for (ELFFileBase *file : ctx.objectFiles) {
1067     for (Symbol *sym : file->getLocalSymbols()) {
1068       auto *def = dyn_cast<Defined>(sym);
1069       if (!def)
1070         continue;
1071       if (!isArmMapSymbol(def) && !isDataMapSymbol(def) &&
1072           !isThumbMapSymbol(def))
1073         continue;
1074       if (auto *sec = cast_if_present<InputSection>(def->section))
1075         if (sec->flags & SHF_EXECINSTR)
1076           sectionMap[sec].push_back(def);
1077     }
1078   }
1079 }
1080 
1081 // Synthetic sections are not backed by an ELF file where we can access the
1082 // symbol table, instead mapping symbols added to synthetic sections are stored
1083 // in the synthetic symbol table. Due to the presence of strip (--strip-all),
1084 // we can not rely on the synthetic symbol table retaining the mapping symbols.
1085 // Instead we record the mapping symbols locally.
1086 void elf::addArmSyntheticSectionMappingSymbol(Defined *sym) {
1087   if (!isArmMapSymbol(sym) && !isDataMapSymbol(sym) && !isThumbMapSymbol(sym))
1088     return;
1089   if (auto *sec = cast_if_present<InputSection>(sym->section))
1090     if (sec->flags & SHF_EXECINSTR)
1091       static_cast<ARM &>(*sec->file->ctx.target).sectionMap[sec].push_back(sym);
1092 }
1093 
1094 static void toLittleEndianInstructions(uint8_t *buf, uint64_t start,
1095                                        uint64_t end, uint64_t width) {
1096   CodeState curState = static_cast<CodeState>(width);
1097   if (curState == CodeState::Arm)
1098     for (uint64_t i = start; i < end; i += width)
1099       write32le(buf + i, read32be(buf + i));
1100 
1101   if (curState == CodeState::Thumb)
1102     for (uint64_t i = start; i < end; i += width)
1103       write16le(buf + i, read16be(buf + i));
1104 }
1105 
1106 // Arm BE8 big endian format requires instructions to be little endian, with
1107 // the initial contents big-endian. Convert the big-endian instructions to
1108 // little endian leaving literal data untouched. We use mapping symbols to
1109 // identify half open intervals of Arm code [$a, non $a) and Thumb code
1110 // [$t, non $t) and convert these to little endian a word or half word at a
1111 // time respectively.
1112 void elf::convertArmInstructionstoBE8(Ctx &ctx, InputSection *sec,
1113                                       uint8_t *buf) {
1114   auto &sectionMap = static_cast<ARM &>(*ctx.target).sectionMap;
1115   auto it = sectionMap.find(sec);
1116   if (it == sectionMap.end())
1117     return;
1118 
1119   SmallVector<const Defined *, 0> &mapSyms = it->second;
1120 
1121   if (mapSyms.empty())
1122     return;
1123 
1124   CodeState curState = CodeState::Data;
1125   uint64_t start = 0, width = 0, size = sec->getSize();
1126   for (auto &msym : mapSyms) {
1127     CodeState newState = CodeState::Data;
1128     if (isThumbMapSymbol(msym))
1129       newState = CodeState::Thumb;
1130     else if (isArmMapSymbol(msym))
1131       newState = CodeState::Arm;
1132 
1133     if (newState == curState)
1134       continue;
1135 
1136     if (curState != CodeState::Data) {
1137       width = static_cast<uint64_t>(curState);
1138       toLittleEndianInstructions(buf, start, msym->value, width);
1139     }
1140     start = msym->value;
1141     curState = newState;
1142   }
1143 
1144   // Passed last mapping symbol, may need to reverse
1145   // up to end of section.
1146   if (curState != CodeState::Data) {
1147     width = static_cast<uint64_t>(curState);
1148     toLittleEndianInstructions(buf, start, size, width);
1149   }
1150 }
1151 
1152 // The Arm Cortex-M Security Extensions (CMSE) splits a system into two parts;
1153 // the non-secure and secure states with the secure state inaccessible from the
1154 // non-secure state, apart from an area of memory in secure state called the
1155 // secure gateway which is accessible from non-secure state. The secure gateway
1156 // contains one or more entry points which must start with a landing pad
1157 // instruction SG. Arm recommends that the secure gateway consists only of
1158 // secure gateway veneers, which are made up of a SG instruction followed by a
1159 // branch to the destination in secure state. Full details can be found in Arm
1160 // v8-M Security Extensions Requirements on Development Tools.
1161 //
1162 // The CMSE model of software development requires the non-secure and secure
1163 // states to be developed as two separate programs. The non-secure developer is
1164 // provided with an import library defining symbols describing the entry points
1165 // in the secure gateway. No additional linker support is required for the
1166 // non-secure state.
1167 //
1168 // Development of the secure state requires linker support to manage the secure
1169 // gateway veneers. The management consists of:
1170 // - Creation of new secure gateway veneers based on symbol conventions.
1171 // - Checking the address of existing secure gateway veneers.
1172 // - Warning when existing secure gateway veneers removed.
1173 //
1174 // The secure gateway veneers are created in an import library, which is just an
1175 // ELF object with a symbol table. The import library is controlled by two
1176 // command line options:
1177 // --in-implib (specify an input import library from a previous revision of the
1178 // program).
1179 // --out-implib (specify an output import library to be created by the linker).
1180 //
1181 // The input import library is used to manage consistency of the secure entry
1182 // points. The output import library is for new and updated secure entry points.
1183 //
1184 // The symbol convention that identifies secure entry functions is the prefix
1185 // __acle_se_ for a symbol called name the linker is expected to create a secure
1186 // gateway veneer if symbols __acle_se_name and name have the same address.
1187 // After creating a secure gateway veneer the symbol name labels the secure
1188 // gateway veneer and the __acle_se_name labels the function definition.
1189 //
1190 // The LLD implementation:
1191 // - Reads an existing import library with importCmseSymbols().
1192 // - Determines which new secure gateway veneers to create and redirects calls
1193 //   within the secure state to the __acle_se_ prefixed symbol with
1194 //   processArmCmseSymbols().
1195 // - Models the SG veneers as a synthetic section.
1196 
1197 // Initialize symbols. symbols is a parallel array to the corresponding ELF
1198 // symbol table.
1199 template <class ELFT> void ObjFile<ELFT>::importCmseSymbols() {
1200   ArrayRef<Elf_Sym> eSyms = getELFSyms<ELFT>();
1201   // Error for local symbols. The symbol at index 0 is LOCAL. So skip it.
1202   for (size_t i = 1, end = firstGlobal; i != end; ++i) {
1203     Err(ctx) << "CMSE symbol '" << CHECK2(eSyms[i].getName(stringTable), this)
1204              << "' in import library '" << this << "' is not global";
1205   }
1206 
1207   for (size_t i = firstGlobal, end = eSyms.size(); i != end; ++i) {
1208     const Elf_Sym &eSym = eSyms[i];
1209     Defined *sym = reinterpret_cast<Defined *>(make<SymbolUnion>());
1210 
1211     // Initialize symbol fields.
1212     memset(static_cast<void *>(sym), 0, sizeof(Symbol));
1213     sym->setName(CHECK2(eSyms[i].getName(stringTable), this));
1214     sym->value = eSym.st_value;
1215     sym->size = eSym.st_size;
1216     sym->type = eSym.getType();
1217     sym->binding = eSym.getBinding();
1218     sym->stOther = eSym.st_other;
1219 
1220     if (eSym.st_shndx != SHN_ABS) {
1221       Err(ctx) << "CMSE symbol '" << sym->getName() << "' in import library '"
1222                << this << "' is not absolute";
1223       continue;
1224     }
1225 
1226     if (!(eSym.st_value & 1) || (eSym.getType() != STT_FUNC)) {
1227       Err(ctx) << "CMSE symbol '" << sym->getName() << "' in import library '"
1228                << this << "' is not a Thumb function definition";
1229       continue;
1230     }
1231 
1232     if (ctx.symtab->cmseImportLib.count(sym->getName())) {
1233       Err(ctx) << "CMSE symbol '" << sym->getName()
1234                << "' is multiply defined in import library '" << this << "'";
1235       continue;
1236     }
1237 
1238     if (eSym.st_size != ACLESESYM_SIZE) {
1239       Warn(ctx) << "CMSE symbol '" << sym->getName() << "' in import library '"
1240                 << this << "' does not have correct size of " << ACLESESYM_SIZE
1241                 << " bytes";
1242     }
1243 
1244     ctx.symtab->cmseImportLib[sym->getName()] = sym;
1245   }
1246 }
1247 
1248 // Check symbol attributes of the acleSeSym, sym pair.
1249 // Both symbols should be global/weak Thumb code symbol definitions.
1250 static std::string checkCmseSymAttributes(Ctx &ctx, Symbol *acleSeSym,
1251                                           Symbol *sym) {
1252   auto check = [&](Symbol *s, StringRef type) -> std::optional<std::string> {
1253     auto d = dyn_cast_or_null<Defined>(s);
1254     if (!(d && d->isFunc() && (d->value & 1)))
1255       return (Twine(toStr(ctx, s->file)) + ": cmse " + type + " symbol '" +
1256               s->getName() + "' is not a Thumb function definition")
1257           .str();
1258     if (!d->section)
1259       return (Twine(toStr(ctx, s->file)) + ": cmse " + type + " symbol '" +
1260               s->getName() + "' cannot be an absolute symbol")
1261           .str();
1262     return std::nullopt;
1263   };
1264   for (auto [sym, type] :
1265        {std::make_pair(acleSeSym, "special"), std::make_pair(sym, "entry")})
1266     if (auto err = check(sym, type))
1267       return *err;
1268   return "";
1269 }
1270 
1271 // Look for [__acle_se_<sym>, <sym>] pairs, as specified in the Cortex-M
1272 // Security Extensions specification.
1273 // 1) <sym> : A standard function name.
1274 // 2) __acle_se_<sym> : A special symbol that prefixes the standard function
1275 // name with __acle_se_.
1276 // Both these symbols are Thumb function symbols with external linkage.
1277 // <sym> may be redefined in .gnu.sgstubs.
1278 void elf::processArmCmseSymbols(Ctx &ctx) {
1279   if (!ctx.arg.cmseImplib)
1280     return;
1281   // Only symbols with external linkage end up in ctx.symtab, so no need to do
1282   // linkage checks. Only check symbol type.
1283   for (Symbol *acleSeSym : ctx.symtab->getSymbols()) {
1284     if (!acleSeSym->getName().starts_with(ACLESESYM_PREFIX))
1285       continue;
1286     // If input object build attributes do not support CMSE, error and disable
1287     // further scanning for <sym>, __acle_se_<sym> pairs.
1288     if (!ctx.arg.armCMSESupport) {
1289       Err(ctx) << "CMSE is only supported by ARMv8-M architecture or later";
1290       ctx.arg.cmseImplib = false;
1291       break;
1292     }
1293 
1294     // Try to find the associated symbol definition.
1295     // Symbol must have external linkage.
1296     StringRef name = acleSeSym->getName().substr(std::strlen(ACLESESYM_PREFIX));
1297     Symbol *sym = ctx.symtab->find(name);
1298     if (!sym) {
1299       Err(ctx) << acleSeSym->file << ": cmse special symbol '"
1300                << acleSeSym->getName()
1301                << "' detected, but no associated entry function definition '"
1302                << name << "' with external linkage found";
1303       continue;
1304     }
1305 
1306     std::string errMsg = checkCmseSymAttributes(ctx, acleSeSym, sym);
1307     if (!errMsg.empty()) {
1308       Err(ctx) << errMsg;
1309       continue;
1310     }
1311 
1312     // <sym> may be redefined later in the link in .gnu.sgstubs
1313     ctx.symtab->cmseSymMap[name] = {acleSeSym, sym};
1314   }
1315 
1316   // If this is an Arm CMSE secure app, replace references to entry symbol <sym>
1317   // with its corresponding special symbol __acle_se_<sym>.
1318   parallelForEach(ctx.objectFiles, [&](InputFile *file) {
1319     MutableArrayRef<Symbol *> syms = file->getMutableSymbols();
1320     for (Symbol *&sym : syms) {
1321       StringRef symName = sym->getName();
1322       auto it = ctx.symtab->cmseSymMap.find(symName);
1323       if (it != ctx.symtab->cmseSymMap.end())
1324         sym = it->second.acleSeSym;
1325     }
1326   });
1327 }
1328 
1329 ArmCmseSGSection::ArmCmseSGSection(Ctx &ctx)
1330     : SyntheticSection(ctx, ".gnu.sgstubs", SHT_PROGBITS,
1331                        SHF_ALLOC | SHF_EXECINSTR,
1332                        /*addralign=*/32) {
1333   entsize = ACLESESYM_SIZE;
1334   // The range of addresses used in the CMSE import library should be fixed.
1335   for (auto &[_, sym] : ctx.symtab->cmseImportLib) {
1336     if (impLibMaxAddr <= sym->value)
1337       impLibMaxAddr = sym->value + sym->size;
1338   }
1339   if (ctx.symtab->cmseSymMap.empty())
1340     return;
1341   addMappingSymbol();
1342   for (auto &[_, entryFunc] : ctx.symtab->cmseSymMap)
1343     addSGVeneer(cast<Defined>(entryFunc.acleSeSym),
1344                 cast<Defined>(entryFunc.sym));
1345   for (auto &[_, sym] : ctx.symtab->cmseImportLib) {
1346     if (!ctx.symtab->inCMSEOutImpLib.count(sym->getName()))
1347       Warn(ctx)
1348           << "entry function '" << sym->getName()
1349           << "' from CMSE import library is not present in secure application";
1350   }
1351 
1352   if (!ctx.symtab->cmseImportLib.empty() && ctx.arg.cmseOutputLib.empty()) {
1353     for (auto &[_, entryFunc] : ctx.symtab->cmseSymMap) {
1354       Symbol *sym = entryFunc.sym;
1355       if (!ctx.symtab->inCMSEOutImpLib.count(sym->getName()))
1356         Warn(ctx) << "new entry function '" << sym->getName()
1357                   << "' introduced but no output import library specified";
1358     }
1359   }
1360 }
1361 
1362 void ArmCmseSGSection::addSGVeneer(Symbol *acleSeSym, Symbol *sym) {
1363   entries.emplace_back(acleSeSym, sym);
1364   if (ctx.symtab->cmseImportLib.count(sym->getName()))
1365     ctx.symtab->inCMSEOutImpLib[sym->getName()] = true;
1366   // Symbol addresses different, nothing to do.
1367   if (acleSeSym->file != sym->file ||
1368       cast<Defined>(*acleSeSym).value != cast<Defined>(*sym).value)
1369     return;
1370   // Only secure symbols with values equal to that of it's non-secure
1371   // counterpart needs to be in the .gnu.sgstubs section.
1372   std::unique_ptr<ArmCmseSGVeneer> ss;
1373   auto it = ctx.symtab->cmseImportLib.find(sym->getName());
1374   if (it != ctx.symtab->cmseImportLib.end()) {
1375     Defined *impSym = it->second;
1376     ss = std::make_unique<ArmCmseSGVeneer>(sym, acleSeSym, impSym->value);
1377   } else {
1378     ss = std::make_unique<ArmCmseSGVeneer>(sym, acleSeSym);
1379     ++newEntries;
1380   }
1381   sgVeneers.emplace_back(std::move(ss));
1382 }
1383 
1384 void ArmCmseSGSection::writeTo(uint8_t *buf) {
1385   for (std::unique_ptr<ArmCmseSGVeneer> &s : sgVeneers) {
1386     uint8_t *p = buf + s->offset;
1387     write16(ctx, p + 0, 0xe97f); // SG
1388     write16(ctx, p + 2, 0xe97f);
1389     write16(ctx, p + 4, 0xf000); // B.W S
1390     write16(ctx, p + 6, 0xb000);
1391     ctx.target->relocateNoSym(p + 4, R_ARM_THM_JUMP24,
1392                               s->acleSeSym->getVA(ctx) -
1393                                   (getVA() + s->offset + s->size));
1394   }
1395 }
1396 
1397 void ArmCmseSGSection::addMappingSymbol() {
1398   addSyntheticLocal(ctx, "$t", STT_NOTYPE, /*off=*/0, /*size=*/0, *this);
1399 }
1400 
1401 size_t ArmCmseSGSection::getSize() const {
1402   if (sgVeneers.empty())
1403     return (impLibMaxAddr ? impLibMaxAddr - getVA() : 0) + newEntries * entsize;
1404 
1405   return entries.size() * entsize;
1406 }
1407 
1408 void ArmCmseSGSection::finalizeContents() {
1409   if (sgVeneers.empty())
1410     return;
1411 
1412   auto it =
1413       std::stable_partition(sgVeneers.begin(), sgVeneers.end(),
1414                             [](auto &i) { return i->getAddr().has_value(); });
1415   std::sort(sgVeneers.begin(), it, [](auto &a, auto &b) {
1416     return a->getAddr().value() < b->getAddr().value();
1417   });
1418   // This is the partition of the veneers with fixed addresses.
1419   uint64_t addr = (*sgVeneers.begin())->getAddr().has_value()
1420                       ? (*sgVeneers.begin())->getAddr().value()
1421                       : getVA();
1422   // Check if the start address of '.gnu.sgstubs' correspond to the
1423   // linker-synthesized veneer with the lowest address.
1424   if ((getVA() & ~1) != (addr & ~1)) {
1425     Err(ctx)
1426         << "start address of '.gnu.sgstubs' is different from previous link";
1427     return;
1428   }
1429 
1430   for (auto [i, s] : enumerate(sgVeneers)) {
1431     s->offset = i * s->size;
1432     Defined(ctx, file, StringRef(), s->sym->binding, s->sym->stOther,
1433             s->sym->type, s->offset | 1, s->size, this)
1434         .overwrite(*s->sym);
1435   }
1436 }
1437 
1438 // Write the CMSE import library to disk.
1439 // The CMSE import library is a relocatable object with only a symbol table.
1440 // The symbols are copies of the (absolute) symbols of the secure gateways
1441 // in the executable output by this link.
1442 // See Arm® v8-M Security Extensions: Requirements on Development Tools
1443 // https://developer.arm.com/documentation/ecm0359818/latest
1444 template <typename ELFT> void elf::writeARMCmseImportLib(Ctx &ctx) {
1445   auto shstrtab =
1446       std::make_unique<StringTableSection>(ctx, ".shstrtab", /*dynamic=*/false);
1447   auto strtab =
1448       std::make_unique<StringTableSection>(ctx, ".strtab", /*dynamic=*/false);
1449   auto impSymTab = std::make_unique<SymbolTableSection<ELFT>>(ctx, *strtab);
1450 
1451   SmallVector<std::pair<std::unique_ptr<OutputSection>, SyntheticSection *>, 0>
1452       osIsPairs;
1453   osIsPairs.emplace_back(
1454       std::make_unique<OutputSection>(ctx, strtab->name, 0, 0), strtab.get());
1455   osIsPairs.emplace_back(
1456       std::make_unique<OutputSection>(ctx, impSymTab->name, 0, 0),
1457       impSymTab.get());
1458   osIsPairs.emplace_back(
1459       std::make_unique<OutputSection>(ctx, shstrtab->name, 0, 0),
1460       shstrtab.get());
1461 
1462   llvm::sort(ctx.symtab->cmseSymMap, [&](const auto &a, const auto &b) {
1463     return a.second.sym->getVA(ctx) < b.second.sym->getVA(ctx);
1464   });
1465   // Copy the secure gateway entry symbols to the import library symbol table.
1466   for (auto &p : ctx.symtab->cmseSymMap) {
1467     Defined *d = cast<Defined>(p.second.sym);
1468     impSymTab->addSymbol(makeDefined(
1469         ctx, ctx.internalFile, d->getName(), d->computeBinding(ctx),
1470         /*stOther=*/0, STT_FUNC, d->getVA(ctx), d->getSize(), nullptr));
1471   }
1472 
1473   size_t idx = 0;
1474   uint64_t off = sizeof(typename ELFT::Ehdr);
1475   for (auto &[osec, isec] : osIsPairs) {
1476     osec->sectionIndex = ++idx;
1477     osec->recordSection(isec);
1478     osec->finalizeInputSections();
1479     osec->shName = shstrtab->addString(osec->name);
1480     osec->size = isec->getSize();
1481     isec->finalizeContents();
1482     osec->offset = alignToPowerOf2(off, osec->addralign);
1483     off = osec->offset + osec->size;
1484   }
1485 
1486   const uint64_t sectionHeaderOff = alignToPowerOf2(off, ctx.arg.wordsize);
1487   const auto shnum = osIsPairs.size() + 1;
1488   const uint64_t fileSize =
1489       sectionHeaderOff + shnum * sizeof(typename ELFT::Shdr);
1490   const unsigned flags =
1491       ctx.arg.mmapOutputFile ? (unsigned)FileOutputBuffer::F_mmap : 0;
1492   unlinkAsync(ctx.arg.cmseOutputLib);
1493   Expected<std::unique_ptr<FileOutputBuffer>> bufferOrErr =
1494       FileOutputBuffer::create(ctx.arg.cmseOutputLib, fileSize, flags);
1495   if (!bufferOrErr) {
1496     Err(ctx) << "failed to open " << ctx.arg.cmseOutputLib << ": "
1497              << bufferOrErr.takeError();
1498     return;
1499   }
1500 
1501   // Write the ELF Header
1502   std::unique_ptr<FileOutputBuffer> &buffer = *bufferOrErr;
1503   uint8_t *const buf = buffer->getBufferStart();
1504   memcpy(buf, "\177ELF", 4);
1505   auto *eHdr = reinterpret_cast<typename ELFT::Ehdr *>(buf);
1506   eHdr->e_type = ET_REL;
1507   eHdr->e_entry = 0;
1508   eHdr->e_shoff = sectionHeaderOff;
1509   eHdr->e_ident[EI_CLASS] = ELFCLASS32;
1510   eHdr->e_ident[EI_DATA] = ctx.arg.isLE ? ELFDATA2LSB : ELFDATA2MSB;
1511   eHdr->e_ident[EI_VERSION] = EV_CURRENT;
1512   eHdr->e_ident[EI_OSABI] = ctx.arg.osabi;
1513   eHdr->e_ident[EI_ABIVERSION] = 0;
1514   eHdr->e_machine = EM_ARM;
1515   eHdr->e_version = EV_CURRENT;
1516   eHdr->e_flags = ctx.arg.eflags;
1517   eHdr->e_ehsize = sizeof(typename ELFT::Ehdr);
1518   eHdr->e_phnum = 0;
1519   eHdr->e_shentsize = sizeof(typename ELFT::Shdr);
1520   eHdr->e_phoff = 0;
1521   eHdr->e_phentsize = 0;
1522   eHdr->e_shnum = shnum;
1523   eHdr->e_shstrndx = shstrtab->getParent()->sectionIndex;
1524 
1525   // Write the section header table.
1526   auto *sHdrs = reinterpret_cast<typename ELFT::Shdr *>(buf + eHdr->e_shoff);
1527   for (auto &[osec, _] : osIsPairs)
1528     osec->template writeHeaderTo<ELFT>(++sHdrs);
1529 
1530   // Write section contents to a mmap'ed file.
1531   {
1532     parallel::TaskGroup tg;
1533     for (auto &[osec, _] : osIsPairs)
1534       osec->template writeTo<ELFT>(ctx, buf + osec->offset, tg);
1535   }
1536 
1537   if (auto e = buffer->commit())
1538     Err(ctx) << "failed to write output '" << buffer->getPath()
1539              << "': " << std::move(e);
1540 }
1541 
1542 void elf::setARMTargetInfo(Ctx &ctx) { ctx.target.reset(new ARM(ctx)); }
1543 
1544 template void elf::writeARMCmseImportLib<ELF32LE>(Ctx &);
1545 template void elf::writeARMCmseImportLib<ELF32BE>(Ctx &);
1546 template void elf::writeARMCmseImportLib<ELF64LE>(Ctx &);
1547 template void elf::writeARMCmseImportLib<ELF64BE>(Ctx &);
1548 
1549 template void ObjFile<ELF32LE>::importCmseSymbols();
1550 template void ObjFile<ELF32BE>::importCmseSymbols();
1551 template void ObjFile<ELF64LE>::importCmseSymbols();
1552 template void ObjFile<ELF64BE>::importCmseSymbols();
1553