xref: /freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/MCTargetDesc/AMDGPUAsmBackend.cpp (revision f126890ac5386406dadf7c4cfa9566cbb56537c5)
1 //===-- AMDGPUAsmBackend.cpp - AMDGPU Assembler Backend -------------------===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 /// \file
8 //===----------------------------------------------------------------------===//
9 
10 #include "MCTargetDesc/AMDGPUFixupKinds.h"
11 #include "MCTargetDesc/AMDGPUMCTargetDesc.h"
12 #include "Utils/AMDGPUBaseInfo.h"
13 #include "llvm/BinaryFormat/ELF.h"
14 #include "llvm/MC/MCAsmBackend.h"
15 #include "llvm/MC/MCAssembler.h"
16 #include "llvm/MC/MCContext.h"
17 #include "llvm/MC/MCFixupKindInfo.h"
18 #include "llvm/MC/MCObjectWriter.h"
19 #include "llvm/MC/MCSubtargetInfo.h"
20 #include "llvm/MC/TargetRegistry.h"
21 #include "llvm/Support/EndianStream.h"
22 #include "llvm/TargetParser/TargetParser.h"
23 
24 using namespace llvm;
25 using namespace llvm::AMDGPU;
26 
27 namespace {
28 
29 class AMDGPUAsmBackend : public MCAsmBackend {
30 public:
31   AMDGPUAsmBackend(const Target &T) : MCAsmBackend(support::little) {}
32 
33   unsigned getNumFixupKinds() const override { return AMDGPU::NumTargetFixupKinds; };
34 
35   void applyFixup(const MCAssembler &Asm, const MCFixup &Fixup,
36                   const MCValue &Target, MutableArrayRef<char> Data,
37                   uint64_t Value, bool IsResolved,
38                   const MCSubtargetInfo *STI) const override;
39   bool fixupNeedsRelaxation(const MCFixup &Fixup, uint64_t Value,
40                             const MCRelaxableFragment *DF,
41                             const MCAsmLayout &Layout) const override;
42 
43   void relaxInstruction(MCInst &Inst,
44                         const MCSubtargetInfo &STI) const override;
45 
46   bool mayNeedRelaxation(const MCInst &Inst,
47                          const MCSubtargetInfo &STI) const override;
48 
49   unsigned getMinimumNopSize() const override;
50   bool writeNopData(raw_ostream &OS, uint64_t Count,
51                     const MCSubtargetInfo *STI) const override;
52 
53   std::optional<MCFixupKind> getFixupKind(StringRef Name) const override;
54   const MCFixupKindInfo &getFixupKindInfo(MCFixupKind Kind) const override;
55   bool shouldForceRelocation(const MCAssembler &Asm, const MCFixup &Fixup,
56                              const MCValue &Target) override;
57 };
58 
59 } //End anonymous namespace
60 
61 void AMDGPUAsmBackend::relaxInstruction(MCInst &Inst,
62                                         const MCSubtargetInfo &STI) const {
63   MCInst Res;
64   unsigned RelaxedOpcode = AMDGPU::getSOPPWithRelaxation(Inst.getOpcode());
65   Res.setOpcode(RelaxedOpcode);
66   Res.addOperand(Inst.getOperand(0));
67   Inst = std::move(Res);
68 }
69 
70 bool AMDGPUAsmBackend::fixupNeedsRelaxation(const MCFixup &Fixup,
71                                             uint64_t Value,
72                                             const MCRelaxableFragment *DF,
73                                             const MCAsmLayout &Layout) const {
74   // if the branch target has an offset of x3f this needs to be relaxed to
75   // add a s_nop 0 immediately after branch to effectively increment offset
76   // for hardware workaround in gfx1010
77   return (((int64_t(Value)/4)-1) == 0x3f);
78 }
79 
80 bool AMDGPUAsmBackend::mayNeedRelaxation(const MCInst &Inst,
81                        const MCSubtargetInfo &STI) const {
82   if (!STI.hasFeature(AMDGPU::FeatureOffset3fBug))
83     return false;
84 
85   if (AMDGPU::getSOPPWithRelaxation(Inst.getOpcode()) >= 0)
86     return true;
87 
88   return false;
89 }
90 
91 static unsigned getFixupKindNumBytes(unsigned Kind) {
92   switch (Kind) {
93   case AMDGPU::fixup_si_sopp_br:
94     return 2;
95   case FK_SecRel_1:
96   case FK_Data_1:
97     return 1;
98   case FK_SecRel_2:
99   case FK_Data_2:
100     return 2;
101   case FK_SecRel_4:
102   case FK_Data_4:
103   case FK_PCRel_4:
104     return 4;
105   case FK_SecRel_8:
106   case FK_Data_8:
107     return 8;
108   default:
109     llvm_unreachable("Unknown fixup kind!");
110   }
111 }
112 
113 static uint64_t adjustFixupValue(const MCFixup &Fixup, uint64_t Value,
114                                  MCContext *Ctx) {
115   int64_t SignedValue = static_cast<int64_t>(Value);
116 
117   switch (Fixup.getTargetKind()) {
118   case AMDGPU::fixup_si_sopp_br: {
119     int64_t BrImm = (SignedValue - 4) / 4;
120 
121     if (Ctx && !isInt<16>(BrImm))
122       Ctx->reportError(Fixup.getLoc(), "branch size exceeds simm16");
123 
124     return BrImm;
125   }
126   case FK_Data_1:
127   case FK_Data_2:
128   case FK_Data_4:
129   case FK_Data_8:
130   case FK_PCRel_4:
131   case FK_SecRel_4:
132     return Value;
133   default:
134     llvm_unreachable("unhandled fixup kind");
135   }
136 }
137 
138 void AMDGPUAsmBackend::applyFixup(const MCAssembler &Asm, const MCFixup &Fixup,
139                                   const MCValue &Target,
140                                   MutableArrayRef<char> Data, uint64_t Value,
141                                   bool IsResolved,
142                                   const MCSubtargetInfo *STI) const {
143   if (Fixup.getKind() >= FirstLiteralRelocationKind)
144     return;
145 
146   Value = adjustFixupValue(Fixup, Value, &Asm.getContext());
147   if (!Value)
148     return; // Doesn't change encoding.
149 
150   MCFixupKindInfo Info = getFixupKindInfo(Fixup.getKind());
151 
152   // Shift the value into position.
153   Value <<= Info.TargetOffset;
154 
155   unsigned NumBytes = getFixupKindNumBytes(Fixup.getKind());
156   uint32_t Offset = Fixup.getOffset();
157   assert(Offset + NumBytes <= Data.size() && "Invalid fixup offset!");
158 
159   // For each byte of the fragment that the fixup touches, mask in the bits from
160   // the fixup value.
161   for (unsigned i = 0; i != NumBytes; ++i)
162     Data[Offset + i] |= static_cast<uint8_t>((Value >> (i * 8)) & 0xff);
163 }
164 
165 std::optional<MCFixupKind>
166 AMDGPUAsmBackend::getFixupKind(StringRef Name) const {
167   return StringSwitch<std::optional<MCFixupKind>>(Name)
168 #define ELF_RELOC(Name, Value)                                                 \
169   .Case(#Name, MCFixupKind(FirstLiteralRelocationKind + Value))
170 #include "llvm/BinaryFormat/ELFRelocs/AMDGPU.def"
171 #undef ELF_RELOC
172       .Default(std::nullopt);
173 }
174 
175 const MCFixupKindInfo &AMDGPUAsmBackend::getFixupKindInfo(
176                                                        MCFixupKind Kind) const {
177   const static MCFixupKindInfo Infos[AMDGPU::NumTargetFixupKinds] = {
178     // name                   offset bits  flags
179     { "fixup_si_sopp_br",     0,     16,   MCFixupKindInfo::FKF_IsPCRel },
180   };
181 
182   if (Kind >= FirstLiteralRelocationKind)
183     return MCAsmBackend::getFixupKindInfo(FK_NONE);
184 
185   if (Kind < FirstTargetFixupKind)
186     return MCAsmBackend::getFixupKindInfo(Kind);
187 
188   return Infos[Kind - FirstTargetFixupKind];
189 }
190 
191 bool AMDGPUAsmBackend::shouldForceRelocation(const MCAssembler &,
192                                              const MCFixup &Fixup,
193                                              const MCValue &) {
194   return Fixup.getKind() >= FirstLiteralRelocationKind;
195 }
196 
197 unsigned AMDGPUAsmBackend::getMinimumNopSize() const {
198   return 4;
199 }
200 
201 bool AMDGPUAsmBackend::writeNopData(raw_ostream &OS, uint64_t Count,
202                                     const MCSubtargetInfo *STI) const {
203   // If the count is not 4-byte aligned, we must be writing data into the text
204   // section (otherwise we have unaligned instructions, and thus have far
205   // bigger problems), so just write zeros instead.
206   OS.write_zeros(Count % 4);
207 
208   // We are properly aligned, so write NOPs as requested.
209   Count /= 4;
210 
211   // FIXME: R600 support.
212   // s_nop 0
213   const uint32_t Encoded_S_NOP_0 = 0xbf800000;
214 
215   for (uint64_t I = 0; I != Count; ++I)
216     support::endian::write<uint32_t>(OS, Encoded_S_NOP_0, Endian);
217 
218   return true;
219 }
220 
221 //===----------------------------------------------------------------------===//
222 // ELFAMDGPUAsmBackend class
223 //===----------------------------------------------------------------------===//
224 
225 namespace {
226 
227 class ELFAMDGPUAsmBackend : public AMDGPUAsmBackend {
228   bool Is64Bit;
229   bool HasRelocationAddend;
230   uint8_t OSABI = ELF::ELFOSABI_NONE;
231   uint8_t ABIVersion = 0;
232 
233 public:
234   ELFAMDGPUAsmBackend(const Target &T, const Triple &TT, uint8_t ABIVersion) :
235       AMDGPUAsmBackend(T), Is64Bit(TT.getArch() == Triple::amdgcn),
236       HasRelocationAddend(TT.getOS() == Triple::AMDHSA),
237       ABIVersion(ABIVersion) {
238     switch (TT.getOS()) {
239     case Triple::AMDHSA:
240       OSABI = ELF::ELFOSABI_AMDGPU_HSA;
241       break;
242     case Triple::AMDPAL:
243       OSABI = ELF::ELFOSABI_AMDGPU_PAL;
244       break;
245     case Triple::Mesa3D:
246       OSABI = ELF::ELFOSABI_AMDGPU_MESA3D;
247       break;
248     default:
249       break;
250     }
251   }
252 
253   std::unique_ptr<MCObjectTargetWriter>
254   createObjectTargetWriter() const override {
255     return createAMDGPUELFObjectWriter(Is64Bit, OSABI, HasRelocationAddend,
256                                        ABIVersion);
257   }
258 };
259 
260 } // end anonymous namespace
261 
262 MCAsmBackend *llvm::createAMDGPUAsmBackend(const Target &T,
263                                            const MCSubtargetInfo &STI,
264                                            const MCRegisterInfo &MRI,
265                                            const MCTargetOptions &Options) {
266   return new ELFAMDGPUAsmBackend(T, STI.getTargetTriple(),
267                                  getHsaAbiVersion(&STI).value_or(0));
268 }
269