xref: /freebsd/contrib/llvm-project/llvm/lib/Target/Sparc/SparcInstrInfo.cpp (revision 84823cc70824c8d842f503d8c2e6d7b0c2d95b61)
1 //===-- SparcInstrInfo.cpp - Sparc Instruction Information ----------------===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 //
9 // This file contains the Sparc implementation of the TargetInstrInfo class.
10 //
11 //===----------------------------------------------------------------------===//
12 
13 #include "SparcInstrInfo.h"
14 #include "Sparc.h"
15 #include "SparcMachineFunctionInfo.h"
16 #include "SparcSubtarget.h"
17 #include "llvm/ADT/STLExtras.h"
18 #include "llvm/ADT/SmallVector.h"
19 #include "llvm/CodeGen/MachineFrameInfo.h"
20 #include "llvm/CodeGen/MachineInstrBuilder.h"
21 #include "llvm/CodeGen/MachineMemOperand.h"
22 #include "llvm/CodeGen/MachineRegisterInfo.h"
23 #include "llvm/MC/TargetRegistry.h"
24 #include "llvm/Support/ErrorHandling.h"
25 
26 using namespace llvm;
27 
28 #define GET_INSTRINFO_CTOR_DTOR
29 #include "SparcGenInstrInfo.inc"
30 
31 // Pin the vtable to this file.
32 void SparcInstrInfo::anchor() {}
33 
34 SparcInstrInfo::SparcInstrInfo(SparcSubtarget &ST)
35     : SparcGenInstrInfo(SP::ADJCALLSTACKDOWN, SP::ADJCALLSTACKUP), RI(),
36       Subtarget(ST) {}
37 
38 /// isLoadFromStackSlot - If the specified machine instruction is a direct
39 /// load from a stack slot, return the virtual or physical register number of
40 /// the destination along with the FrameIndex of the loaded stack slot.  If
41 /// not, return 0.  This predicate must return 0 if the instruction has
42 /// any side effects other than loading from the stack slot.
43 unsigned SparcInstrInfo::isLoadFromStackSlot(const MachineInstr &MI,
44                                              int &FrameIndex) const {
45   if (MI.getOpcode() == SP::LDri || MI.getOpcode() == SP::LDXri ||
46       MI.getOpcode() == SP::LDFri || MI.getOpcode() == SP::LDDFri ||
47       MI.getOpcode() == SP::LDQFri) {
48     if (MI.getOperand(1).isFI() && MI.getOperand(2).isImm() &&
49         MI.getOperand(2).getImm() == 0) {
50       FrameIndex = MI.getOperand(1).getIndex();
51       return MI.getOperand(0).getReg();
52     }
53   }
54   return 0;
55 }
56 
57 /// isStoreToStackSlot - If the specified machine instruction is a direct
58 /// store to a stack slot, return the virtual or physical register number of
59 /// the source reg along with the FrameIndex of the loaded stack slot.  If
60 /// not, return 0.  This predicate must return 0 if the instruction has
61 /// any side effects other than storing to the stack slot.
62 unsigned SparcInstrInfo::isStoreToStackSlot(const MachineInstr &MI,
63                                             int &FrameIndex) const {
64   if (MI.getOpcode() == SP::STri || MI.getOpcode() == SP::STXri ||
65       MI.getOpcode() == SP::STFri || MI.getOpcode() == SP::STDFri ||
66       MI.getOpcode() == SP::STQFri) {
67     if (MI.getOperand(0).isFI() && MI.getOperand(1).isImm() &&
68         MI.getOperand(1).getImm() == 0) {
69       FrameIndex = MI.getOperand(0).getIndex();
70       return MI.getOperand(2).getReg();
71     }
72   }
73   return 0;
74 }
75 
76 static bool IsIntegerCC(unsigned CC)
77 {
78   return  (CC <= SPCC::ICC_VC);
79 }
80 
81 static SPCC::CondCodes GetOppositeBranchCondition(SPCC::CondCodes CC)
82 {
83   switch(CC) {
84   case SPCC::ICC_A:    return SPCC::ICC_N;
85   case SPCC::ICC_N:    return SPCC::ICC_A;
86   case SPCC::ICC_NE:   return SPCC::ICC_E;
87   case SPCC::ICC_E:    return SPCC::ICC_NE;
88   case SPCC::ICC_G:    return SPCC::ICC_LE;
89   case SPCC::ICC_LE:   return SPCC::ICC_G;
90   case SPCC::ICC_GE:   return SPCC::ICC_L;
91   case SPCC::ICC_L:    return SPCC::ICC_GE;
92   case SPCC::ICC_GU:   return SPCC::ICC_LEU;
93   case SPCC::ICC_LEU:  return SPCC::ICC_GU;
94   case SPCC::ICC_CC:   return SPCC::ICC_CS;
95   case SPCC::ICC_CS:   return SPCC::ICC_CC;
96   case SPCC::ICC_POS:  return SPCC::ICC_NEG;
97   case SPCC::ICC_NEG:  return SPCC::ICC_POS;
98   case SPCC::ICC_VC:   return SPCC::ICC_VS;
99   case SPCC::ICC_VS:   return SPCC::ICC_VC;
100 
101   case SPCC::FCC_A:    return SPCC::FCC_N;
102   case SPCC::FCC_N:    return SPCC::FCC_A;
103   case SPCC::FCC_U:    return SPCC::FCC_O;
104   case SPCC::FCC_O:    return SPCC::FCC_U;
105   case SPCC::FCC_G:    return SPCC::FCC_ULE;
106   case SPCC::FCC_LE:   return SPCC::FCC_UG;
107   case SPCC::FCC_UG:   return SPCC::FCC_LE;
108   case SPCC::FCC_ULE:  return SPCC::FCC_G;
109   case SPCC::FCC_L:    return SPCC::FCC_UGE;
110   case SPCC::FCC_GE:   return SPCC::FCC_UL;
111   case SPCC::FCC_UL:   return SPCC::FCC_GE;
112   case SPCC::FCC_UGE:  return SPCC::FCC_L;
113   case SPCC::FCC_LG:   return SPCC::FCC_UE;
114   case SPCC::FCC_UE:   return SPCC::FCC_LG;
115   case SPCC::FCC_NE:   return SPCC::FCC_E;
116   case SPCC::FCC_E:    return SPCC::FCC_NE;
117 
118   case SPCC::CPCC_A:   return SPCC::CPCC_N;
119   case SPCC::CPCC_N:   return SPCC::CPCC_A;
120   case SPCC::CPCC_3:   LLVM_FALLTHROUGH;
121   case SPCC::CPCC_2:   LLVM_FALLTHROUGH;
122   case SPCC::CPCC_23:  LLVM_FALLTHROUGH;
123   case SPCC::CPCC_1:   LLVM_FALLTHROUGH;
124   case SPCC::CPCC_13:  LLVM_FALLTHROUGH;
125   case SPCC::CPCC_12:  LLVM_FALLTHROUGH;
126   case SPCC::CPCC_123: LLVM_FALLTHROUGH;
127   case SPCC::CPCC_0:   LLVM_FALLTHROUGH;
128   case SPCC::CPCC_03:  LLVM_FALLTHROUGH;
129   case SPCC::CPCC_02:  LLVM_FALLTHROUGH;
130   case SPCC::CPCC_023: LLVM_FALLTHROUGH;
131   case SPCC::CPCC_01:  LLVM_FALLTHROUGH;
132   case SPCC::CPCC_013: LLVM_FALLTHROUGH;
133   case SPCC::CPCC_012:
134       // "Opposite" code is not meaningful, as we don't know
135       // what the CoProc condition means here. The cond-code will
136       // only be used in inline assembler, so this code should
137       // not be reached in a normal compilation pass.
138       llvm_unreachable("Meaningless inversion of co-processor cond code");
139   }
140   llvm_unreachable("Invalid cond code");
141 }
142 
143 static bool isUncondBranchOpcode(int Opc) { return Opc == SP::BA; }
144 
145 static bool isCondBranchOpcode(int Opc) {
146   return Opc == SP::FBCOND || Opc == SP::BCOND;
147 }
148 
149 static bool isIndirectBranchOpcode(int Opc) {
150   return Opc == SP::BINDrr || Opc == SP::BINDri;
151 }
152 
153 static void parseCondBranch(MachineInstr *LastInst, MachineBasicBlock *&Target,
154                             SmallVectorImpl<MachineOperand> &Cond) {
155   Cond.push_back(MachineOperand::CreateImm(LastInst->getOperand(1).getImm()));
156   Target = LastInst->getOperand(0).getMBB();
157 }
158 
159 bool SparcInstrInfo::analyzeBranch(MachineBasicBlock &MBB,
160                                    MachineBasicBlock *&TBB,
161                                    MachineBasicBlock *&FBB,
162                                    SmallVectorImpl<MachineOperand> &Cond,
163                                    bool AllowModify) const {
164   MachineBasicBlock::iterator I = MBB.getLastNonDebugInstr();
165   if (I == MBB.end())
166     return false;
167 
168   if (!isUnpredicatedTerminator(*I))
169     return false;
170 
171   // Get the last instruction in the block.
172   MachineInstr *LastInst = &*I;
173   unsigned LastOpc = LastInst->getOpcode();
174 
175   // If there is only one terminator instruction, process it.
176   if (I == MBB.begin() || !isUnpredicatedTerminator(*--I)) {
177     if (isUncondBranchOpcode(LastOpc)) {
178       TBB = LastInst->getOperand(0).getMBB();
179       return false;
180     }
181     if (isCondBranchOpcode(LastOpc)) {
182       // Block ends with fall-through condbranch.
183       parseCondBranch(LastInst, TBB, Cond);
184       return false;
185     }
186     return true; // Can't handle indirect branch.
187   }
188 
189   // Get the instruction before it if it is a terminator.
190   MachineInstr *SecondLastInst = &*I;
191   unsigned SecondLastOpc = SecondLastInst->getOpcode();
192 
193   // If AllowModify is true and the block ends with two or more unconditional
194   // branches, delete all but the first unconditional branch.
195   if (AllowModify && isUncondBranchOpcode(LastOpc)) {
196     while (isUncondBranchOpcode(SecondLastOpc)) {
197       LastInst->eraseFromParent();
198       LastInst = SecondLastInst;
199       LastOpc = LastInst->getOpcode();
200       if (I == MBB.begin() || !isUnpredicatedTerminator(*--I)) {
201         // Return now the only terminator is an unconditional branch.
202         TBB = LastInst->getOperand(0).getMBB();
203         return false;
204       } else {
205         SecondLastInst = &*I;
206         SecondLastOpc = SecondLastInst->getOpcode();
207       }
208     }
209   }
210 
211   // If there are three terminators, we don't know what sort of block this is.
212   if (SecondLastInst && I != MBB.begin() && isUnpredicatedTerminator(*--I))
213     return true;
214 
215   // If the block ends with a B and a Bcc, handle it.
216   if (isCondBranchOpcode(SecondLastOpc) && isUncondBranchOpcode(LastOpc)) {
217     parseCondBranch(SecondLastInst, TBB, Cond);
218     FBB = LastInst->getOperand(0).getMBB();
219     return false;
220   }
221 
222   // If the block ends with two unconditional branches, handle it.  The second
223   // one is not executed.
224   if (isUncondBranchOpcode(SecondLastOpc) && isUncondBranchOpcode(LastOpc)) {
225     TBB = SecondLastInst->getOperand(0).getMBB();
226     return false;
227   }
228 
229   // ...likewise if it ends with an indirect branch followed by an unconditional
230   // branch.
231   if (isIndirectBranchOpcode(SecondLastOpc) && isUncondBranchOpcode(LastOpc)) {
232     I = LastInst;
233     if (AllowModify)
234       I->eraseFromParent();
235     return true;
236   }
237 
238   // Otherwise, can't handle this.
239   return true;
240 }
241 
242 unsigned SparcInstrInfo::insertBranch(MachineBasicBlock &MBB,
243                                       MachineBasicBlock *TBB,
244                                       MachineBasicBlock *FBB,
245                                       ArrayRef<MachineOperand> Cond,
246                                       const DebugLoc &DL,
247                                       int *BytesAdded) const {
248   assert(TBB && "insertBranch must not be told to insert a fallthrough");
249   assert((Cond.size() == 1 || Cond.size() == 0) &&
250          "Sparc branch conditions should have one component!");
251   assert(!BytesAdded && "code size not handled");
252 
253   if (Cond.empty()) {
254     assert(!FBB && "Unconditional branch with multiple successors!");
255     BuildMI(&MBB, DL, get(SP::BA)).addMBB(TBB);
256     return 1;
257   }
258 
259   // Conditional branch
260   unsigned CC = Cond[0].getImm();
261 
262   if (IsIntegerCC(CC))
263     BuildMI(&MBB, DL, get(SP::BCOND)).addMBB(TBB).addImm(CC);
264   else
265     BuildMI(&MBB, DL, get(SP::FBCOND)).addMBB(TBB).addImm(CC);
266   if (!FBB)
267     return 1;
268 
269   BuildMI(&MBB, DL, get(SP::BA)).addMBB(FBB);
270   return 2;
271 }
272 
273 unsigned SparcInstrInfo::removeBranch(MachineBasicBlock &MBB,
274                                       int *BytesRemoved) const {
275   assert(!BytesRemoved && "code size not handled");
276 
277   MachineBasicBlock::iterator I = MBB.end();
278   unsigned Count = 0;
279   while (I != MBB.begin()) {
280     --I;
281 
282     if (I->isDebugInstr())
283       continue;
284 
285     if (I->getOpcode() != SP::BA
286         && I->getOpcode() != SP::BCOND
287         && I->getOpcode() != SP::FBCOND)
288       break; // Not a branch
289 
290     I->eraseFromParent();
291     I = MBB.end();
292     ++Count;
293   }
294   return Count;
295 }
296 
297 bool SparcInstrInfo::reverseBranchCondition(
298     SmallVectorImpl<MachineOperand> &Cond) const {
299   assert(Cond.size() == 1);
300   SPCC::CondCodes CC = static_cast<SPCC::CondCodes>(Cond[0].getImm());
301   Cond[0].setImm(GetOppositeBranchCondition(CC));
302   return false;
303 }
304 
305 void SparcInstrInfo::copyPhysReg(MachineBasicBlock &MBB,
306                                  MachineBasicBlock::iterator I,
307                                  const DebugLoc &DL, MCRegister DestReg,
308                                  MCRegister SrcReg, bool KillSrc) const {
309   unsigned numSubRegs = 0;
310   unsigned movOpc     = 0;
311   const unsigned *subRegIdx = nullptr;
312   bool ExtraG0 = false;
313 
314   const unsigned DW_SubRegsIdx[]  = { SP::sub_even, SP::sub_odd };
315   const unsigned DFP_FP_SubRegsIdx[]  = { SP::sub_even, SP::sub_odd };
316   const unsigned QFP_DFP_SubRegsIdx[] = { SP::sub_even64, SP::sub_odd64 };
317   const unsigned QFP_FP_SubRegsIdx[]  = { SP::sub_even, SP::sub_odd,
318                                           SP::sub_odd64_then_sub_even,
319                                           SP::sub_odd64_then_sub_odd };
320 
321   if (SP::IntRegsRegClass.contains(DestReg, SrcReg))
322     BuildMI(MBB, I, DL, get(SP::ORrr), DestReg).addReg(SP::G0)
323       .addReg(SrcReg, getKillRegState(KillSrc));
324   else if (SP::IntPairRegClass.contains(DestReg, SrcReg)) {
325     subRegIdx  = DW_SubRegsIdx;
326     numSubRegs = 2;
327     movOpc     = SP::ORrr;
328     ExtraG0 = true;
329   } else if (SP::FPRegsRegClass.contains(DestReg, SrcReg))
330     BuildMI(MBB, I, DL, get(SP::FMOVS), DestReg)
331       .addReg(SrcReg, getKillRegState(KillSrc));
332   else if (SP::DFPRegsRegClass.contains(DestReg, SrcReg)) {
333     if (Subtarget.isV9()) {
334       BuildMI(MBB, I, DL, get(SP::FMOVD), DestReg)
335         .addReg(SrcReg, getKillRegState(KillSrc));
336     } else {
337       // Use two FMOVS instructions.
338       subRegIdx  = DFP_FP_SubRegsIdx;
339       numSubRegs = 2;
340       movOpc     = SP::FMOVS;
341     }
342   } else if (SP::QFPRegsRegClass.contains(DestReg, SrcReg)) {
343     if (Subtarget.isV9()) {
344       if (Subtarget.hasHardQuad()) {
345         BuildMI(MBB, I, DL, get(SP::FMOVQ), DestReg)
346           .addReg(SrcReg, getKillRegState(KillSrc));
347       } else {
348         // Use two FMOVD instructions.
349         subRegIdx  = QFP_DFP_SubRegsIdx;
350         numSubRegs = 2;
351         movOpc     = SP::FMOVD;
352       }
353     } else {
354       // Use four FMOVS instructions.
355       subRegIdx  = QFP_FP_SubRegsIdx;
356       numSubRegs = 4;
357       movOpc     = SP::FMOVS;
358     }
359   } else if (SP::ASRRegsRegClass.contains(DestReg) &&
360              SP::IntRegsRegClass.contains(SrcReg)) {
361     BuildMI(MBB, I, DL, get(SP::WRASRrr), DestReg)
362         .addReg(SP::G0)
363         .addReg(SrcReg, getKillRegState(KillSrc));
364   } else if (SP::IntRegsRegClass.contains(DestReg) &&
365              SP::ASRRegsRegClass.contains(SrcReg)) {
366     BuildMI(MBB, I, DL, get(SP::RDASR), DestReg)
367         .addReg(SrcReg, getKillRegState(KillSrc));
368   } else
369     llvm_unreachable("Impossible reg-to-reg copy");
370 
371   if (numSubRegs == 0 || subRegIdx == nullptr || movOpc == 0)
372     return;
373 
374   const TargetRegisterInfo *TRI = &getRegisterInfo();
375   MachineInstr *MovMI = nullptr;
376 
377   for (unsigned i = 0; i != numSubRegs; ++i) {
378     Register Dst = TRI->getSubReg(DestReg, subRegIdx[i]);
379     Register Src = TRI->getSubReg(SrcReg, subRegIdx[i]);
380     assert(Dst && Src && "Bad sub-register");
381 
382     MachineInstrBuilder MIB = BuildMI(MBB, I, DL, get(movOpc), Dst);
383     if (ExtraG0)
384       MIB.addReg(SP::G0);
385     MIB.addReg(Src);
386     MovMI = MIB.getInstr();
387   }
388   // Add implicit super-register defs and kills to the last MovMI.
389   MovMI->addRegisterDefined(DestReg, TRI);
390   if (KillSrc)
391     MovMI->addRegisterKilled(SrcReg, TRI);
392 }
393 
394 void SparcInstrInfo::
395 storeRegToStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator I,
396                     Register SrcReg, bool isKill, int FI,
397                     const TargetRegisterClass *RC,
398                     const TargetRegisterInfo *TRI) const {
399   DebugLoc DL;
400   if (I != MBB.end()) DL = I->getDebugLoc();
401 
402   MachineFunction *MF = MBB.getParent();
403   const MachineFrameInfo &MFI = MF->getFrameInfo();
404   MachineMemOperand *MMO = MF->getMachineMemOperand(
405       MachinePointerInfo::getFixedStack(*MF, FI), MachineMemOperand::MOStore,
406       MFI.getObjectSize(FI), MFI.getObjectAlign(FI));
407 
408   // On the order of operands here: think "[FrameIdx + 0] = SrcReg".
409   if (RC == &SP::I64RegsRegClass)
410     BuildMI(MBB, I, DL, get(SP::STXri)).addFrameIndex(FI).addImm(0)
411       .addReg(SrcReg, getKillRegState(isKill)).addMemOperand(MMO);
412   else if (RC == &SP::IntRegsRegClass)
413     BuildMI(MBB, I, DL, get(SP::STri)).addFrameIndex(FI).addImm(0)
414       .addReg(SrcReg, getKillRegState(isKill)).addMemOperand(MMO);
415   else if (RC == &SP::IntPairRegClass)
416     BuildMI(MBB, I, DL, get(SP::STDri)).addFrameIndex(FI).addImm(0)
417       .addReg(SrcReg, getKillRegState(isKill)).addMemOperand(MMO);
418   else if (RC == &SP::FPRegsRegClass)
419     BuildMI(MBB, I, DL, get(SP::STFri)).addFrameIndex(FI).addImm(0)
420       .addReg(SrcReg,  getKillRegState(isKill)).addMemOperand(MMO);
421   else if (SP::DFPRegsRegClass.hasSubClassEq(RC))
422     BuildMI(MBB, I, DL, get(SP::STDFri)).addFrameIndex(FI).addImm(0)
423       .addReg(SrcReg,  getKillRegState(isKill)).addMemOperand(MMO);
424   else if (SP::QFPRegsRegClass.hasSubClassEq(RC))
425     // Use STQFri irrespective of its legality. If STQ is not legal, it will be
426     // lowered into two STDs in eliminateFrameIndex.
427     BuildMI(MBB, I, DL, get(SP::STQFri)).addFrameIndex(FI).addImm(0)
428       .addReg(SrcReg,  getKillRegState(isKill)).addMemOperand(MMO);
429   else
430     llvm_unreachable("Can't store this register to stack slot");
431 }
432 
433 void SparcInstrInfo::
434 loadRegFromStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator I,
435                      Register DestReg, int FI,
436                      const TargetRegisterClass *RC,
437                      const TargetRegisterInfo *TRI) const {
438   DebugLoc DL;
439   if (I != MBB.end()) DL = I->getDebugLoc();
440 
441   MachineFunction *MF = MBB.getParent();
442   const MachineFrameInfo &MFI = MF->getFrameInfo();
443   MachineMemOperand *MMO = MF->getMachineMemOperand(
444       MachinePointerInfo::getFixedStack(*MF, FI), MachineMemOperand::MOLoad,
445       MFI.getObjectSize(FI), MFI.getObjectAlign(FI));
446 
447   if (RC == &SP::I64RegsRegClass)
448     BuildMI(MBB, I, DL, get(SP::LDXri), DestReg).addFrameIndex(FI).addImm(0)
449       .addMemOperand(MMO);
450   else if (RC == &SP::IntRegsRegClass)
451     BuildMI(MBB, I, DL, get(SP::LDri), DestReg).addFrameIndex(FI).addImm(0)
452       .addMemOperand(MMO);
453   else if (RC == &SP::IntPairRegClass)
454     BuildMI(MBB, I, DL, get(SP::LDDri), DestReg).addFrameIndex(FI).addImm(0)
455       .addMemOperand(MMO);
456   else if (RC == &SP::FPRegsRegClass)
457     BuildMI(MBB, I, DL, get(SP::LDFri), DestReg).addFrameIndex(FI).addImm(0)
458       .addMemOperand(MMO);
459   else if (SP::DFPRegsRegClass.hasSubClassEq(RC))
460     BuildMI(MBB, I, DL, get(SP::LDDFri), DestReg).addFrameIndex(FI).addImm(0)
461       .addMemOperand(MMO);
462   else if (SP::QFPRegsRegClass.hasSubClassEq(RC))
463     // Use LDQFri irrespective of its legality. If LDQ is not legal, it will be
464     // lowered into two LDDs in eliminateFrameIndex.
465     BuildMI(MBB, I, DL, get(SP::LDQFri), DestReg).addFrameIndex(FI).addImm(0)
466       .addMemOperand(MMO);
467   else
468     llvm_unreachable("Can't load this register from stack slot");
469 }
470 
471 Register SparcInstrInfo::getGlobalBaseReg(MachineFunction *MF) const {
472   SparcMachineFunctionInfo *SparcFI = MF->getInfo<SparcMachineFunctionInfo>();
473   Register GlobalBaseReg = SparcFI->getGlobalBaseReg();
474   if (GlobalBaseReg)
475     return GlobalBaseReg;
476 
477   // Insert the set of GlobalBaseReg into the first MBB of the function
478   MachineBasicBlock &FirstMBB = MF->front();
479   MachineBasicBlock::iterator MBBI = FirstMBB.begin();
480   MachineRegisterInfo &RegInfo = MF->getRegInfo();
481 
482   const TargetRegisterClass *PtrRC =
483     Subtarget.is64Bit() ? &SP::I64RegsRegClass : &SP::IntRegsRegClass;
484   GlobalBaseReg = RegInfo.createVirtualRegister(PtrRC);
485 
486   DebugLoc dl;
487 
488   BuildMI(FirstMBB, MBBI, dl, get(SP::GETPCX), GlobalBaseReg);
489   SparcFI->setGlobalBaseReg(GlobalBaseReg);
490   return GlobalBaseReg;
491 }
492 
493 bool SparcInstrInfo::expandPostRAPseudo(MachineInstr &MI) const {
494   switch (MI.getOpcode()) {
495   case TargetOpcode::LOAD_STACK_GUARD: {
496     assert(Subtarget.isTargetLinux() &&
497            "Only Linux target is expected to contain LOAD_STACK_GUARD");
498     // offsetof(tcbhead_t, stack_guard) from sysdeps/sparc/nptl/tls.h in glibc.
499     const int64_t Offset = Subtarget.is64Bit() ? 0x28 : 0x14;
500     MI.setDesc(get(Subtarget.is64Bit() ? SP::LDXri : SP::LDri));
501     MachineInstrBuilder(*MI.getParent()->getParent(), MI)
502         .addReg(SP::G7)
503         .addImm(Offset);
504     return true;
505   }
506   }
507   return false;
508 }
509