xref: /freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/SILowerSGPRSpills.cpp (revision 0fca6ea1d4eea4c934cfff25ac9ee8ad6fe95583)
1 //===-- SILowerSGPRSPills.cpp ---------------------------------------------===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 //
9 // Handle SGPR spills. This pass takes the place of PrologEpilogInserter for all
10 // SGPR spills, so must insert CSR SGPR spills as well as expand them.
11 //
12 // This pass must never create new SGPR virtual registers.
13 //
14 // FIXME: Must stop RegScavenger spills in later passes.
15 //
16 //===----------------------------------------------------------------------===//
17 
18 #include "AMDGPU.h"
19 #include "GCNSubtarget.h"
20 #include "MCTargetDesc/AMDGPUMCTargetDesc.h"
21 #include "SIMachineFunctionInfo.h"
22 #include "llvm/CodeGen/LiveIntervals.h"
23 #include "llvm/CodeGen/MachineFrameInfo.h"
24 #include "llvm/CodeGen/RegisterScavenging.h"
25 #include "llvm/InitializePasses.h"
26 
27 using namespace llvm;
28 
29 #define DEBUG_TYPE "si-lower-sgpr-spills"
30 
31 using MBBVector = SmallVector<MachineBasicBlock *, 4>;
32 
33 namespace {
34 
35 class SILowerSGPRSpills : public MachineFunctionPass {
36 private:
37   const SIRegisterInfo *TRI = nullptr;
38   const SIInstrInfo *TII = nullptr;
39   LiveIntervals *LIS = nullptr;
40   SlotIndexes *Indexes = nullptr;
41 
42   // Save and Restore blocks of the current function. Typically there is a
43   // single save block, unless Windows EH funclets are involved.
44   MBBVector SaveBlocks;
45   MBBVector RestoreBlocks;
46 
47 public:
48   static char ID;
49 
SILowerSGPRSpills()50   SILowerSGPRSpills() : MachineFunctionPass(ID) {}
51 
52   void calculateSaveRestoreBlocks(MachineFunction &MF);
53   bool spillCalleeSavedRegs(MachineFunction &MF,
54                             SmallVectorImpl<int> &CalleeSavedFIs);
55   void extendWWMVirtRegLiveness(MachineFunction &MF, LiveIntervals *LIS);
56 
57   bool runOnMachineFunction(MachineFunction &MF) override;
58 
getAnalysisUsage(AnalysisUsage & AU) const59   void getAnalysisUsage(AnalysisUsage &AU) const override {
60     AU.setPreservesAll();
61     MachineFunctionPass::getAnalysisUsage(AU);
62   }
63 
getClearedProperties() const64   MachineFunctionProperties getClearedProperties() const override {
65     // SILowerSGPRSpills introduces new Virtual VGPRs for spilling SGPRs.
66     return MachineFunctionProperties()
67         .set(MachineFunctionProperties::Property::IsSSA)
68         .set(MachineFunctionProperties::Property::NoVRegs);
69   }
70 };
71 
72 } // end anonymous namespace
73 
74 char SILowerSGPRSpills::ID = 0;
75 
76 INITIALIZE_PASS_BEGIN(SILowerSGPRSpills, DEBUG_TYPE,
77                       "SI lower SGPR spill instructions", false, false)
78 INITIALIZE_PASS_DEPENDENCY(LiveIntervalsWrapperPass)
79 INITIALIZE_PASS_DEPENDENCY(VirtRegMap)
80 INITIALIZE_PASS_END(SILowerSGPRSpills, DEBUG_TYPE,
81                     "SI lower SGPR spill instructions", false, false)
82 
83 char &llvm::SILowerSGPRSpillsID = SILowerSGPRSpills::ID;
84 
85 /// Insert spill code for the callee-saved registers used in the function.
insertCSRSaves(MachineBasicBlock & SaveBlock,ArrayRef<CalleeSavedInfo> CSI,SlotIndexes * Indexes,LiveIntervals * LIS)86 static void insertCSRSaves(MachineBasicBlock &SaveBlock,
87                            ArrayRef<CalleeSavedInfo> CSI, SlotIndexes *Indexes,
88                            LiveIntervals *LIS) {
89   MachineFunction &MF = *SaveBlock.getParent();
90   const TargetInstrInfo &TII = *MF.getSubtarget().getInstrInfo();
91   const TargetFrameLowering *TFI = MF.getSubtarget().getFrameLowering();
92   const TargetRegisterInfo *TRI = MF.getSubtarget().getRegisterInfo();
93   const GCNSubtarget &ST = MF.getSubtarget<GCNSubtarget>();
94   const SIRegisterInfo *RI = ST.getRegisterInfo();
95 
96   MachineBasicBlock::iterator I = SaveBlock.begin();
97   if (!TFI->spillCalleeSavedRegisters(SaveBlock, I, CSI, TRI)) {
98     const MachineRegisterInfo &MRI = MF.getRegInfo();
99 
100     for (const CalleeSavedInfo &CS : CSI) {
101       // Insert the spill to the stack frame.
102       MCRegister Reg = CS.getReg();
103 
104       MachineInstrSpan MIS(I, &SaveBlock);
105       const TargetRegisterClass *RC = TRI->getMinimalPhysRegClass(
106           Reg, Reg == RI->getReturnAddressReg(MF) ? MVT::i64 : MVT::i32);
107 
108       // If this value was already livein, we probably have a direct use of the
109       // incoming register value, so don't kill at the spill point. This happens
110       // since we pass some special inputs (workgroup IDs) in the callee saved
111       // range.
112       const bool IsLiveIn = MRI.isLiveIn(Reg);
113       TII.storeRegToStackSlot(SaveBlock, I, Reg, !IsLiveIn, CS.getFrameIdx(),
114                               RC, TRI, Register());
115 
116       if (Indexes) {
117         assert(std::distance(MIS.begin(), I) == 1);
118         MachineInstr &Inst = *std::prev(I);
119         Indexes->insertMachineInstrInMaps(Inst);
120       }
121 
122       if (LIS)
123         LIS->removeAllRegUnitsForPhysReg(Reg);
124     }
125   }
126 }
127 
128 /// Insert restore code for the callee-saved registers used in the function.
insertCSRRestores(MachineBasicBlock & RestoreBlock,MutableArrayRef<CalleeSavedInfo> CSI,SlotIndexes * Indexes,LiveIntervals * LIS)129 static void insertCSRRestores(MachineBasicBlock &RestoreBlock,
130                               MutableArrayRef<CalleeSavedInfo> CSI,
131                               SlotIndexes *Indexes, LiveIntervals *LIS) {
132   MachineFunction &MF = *RestoreBlock.getParent();
133   const TargetInstrInfo &TII = *MF.getSubtarget().getInstrInfo();
134   const TargetFrameLowering *TFI = MF.getSubtarget().getFrameLowering();
135   const TargetRegisterInfo *TRI = MF.getSubtarget().getRegisterInfo();
136   const GCNSubtarget &ST = MF.getSubtarget<GCNSubtarget>();
137   const SIRegisterInfo *RI = ST.getRegisterInfo();
138   // Restore all registers immediately before the return and any
139   // terminators that precede it.
140   MachineBasicBlock::iterator I = RestoreBlock.getFirstTerminator();
141 
142   // FIXME: Just emit the readlane/writelane directly
143   if (!TFI->restoreCalleeSavedRegisters(RestoreBlock, I, CSI, TRI)) {
144     for (const CalleeSavedInfo &CI : reverse(CSI)) {
145       Register Reg = CI.getReg();
146       const TargetRegisterClass *RC = TRI->getMinimalPhysRegClass(
147           Reg, Reg == RI->getReturnAddressReg(MF) ? MVT::i64 : MVT::i32);
148 
149       TII.loadRegFromStackSlot(RestoreBlock, I, Reg, CI.getFrameIdx(), RC, TRI,
150                                Register());
151       assert(I != RestoreBlock.begin() &&
152              "loadRegFromStackSlot didn't insert any code!");
153       // Insert in reverse order.  loadRegFromStackSlot can insert
154       // multiple instructions.
155 
156       if (Indexes) {
157         MachineInstr &Inst = *std::prev(I);
158         Indexes->insertMachineInstrInMaps(Inst);
159       }
160 
161       if (LIS)
162         LIS->removeAllRegUnitsForPhysReg(Reg);
163     }
164   }
165 }
166 
167 /// Compute the sets of entry and return blocks for saving and restoring
168 /// callee-saved registers, and placing prolog and epilog code.
calculateSaveRestoreBlocks(MachineFunction & MF)169 void SILowerSGPRSpills::calculateSaveRestoreBlocks(MachineFunction &MF) {
170   const MachineFrameInfo &MFI = MF.getFrameInfo();
171 
172   // Even when we do not change any CSR, we still want to insert the
173   // prologue and epilogue of the function.
174   // So set the save points for those.
175 
176   // Use the points found by shrink-wrapping, if any.
177   if (MFI.getSavePoint()) {
178     SaveBlocks.push_back(MFI.getSavePoint());
179     assert(MFI.getRestorePoint() && "Both restore and save must be set");
180     MachineBasicBlock *RestoreBlock = MFI.getRestorePoint();
181     // If RestoreBlock does not have any successor and is not a return block
182     // then the end point is unreachable and we do not need to insert any
183     // epilogue.
184     if (!RestoreBlock->succ_empty() || RestoreBlock->isReturnBlock())
185       RestoreBlocks.push_back(RestoreBlock);
186     return;
187   }
188 
189   // Save refs to entry and return blocks.
190   SaveBlocks.push_back(&MF.front());
191   for (MachineBasicBlock &MBB : MF) {
192     if (MBB.isEHFuncletEntry())
193       SaveBlocks.push_back(&MBB);
194     if (MBB.isReturnBlock())
195       RestoreBlocks.push_back(&MBB);
196   }
197 }
198 
199 // TODO: To support shrink wrapping, this would need to copy
200 // PrologEpilogInserter's updateLiveness.
updateLiveness(MachineFunction & MF,ArrayRef<CalleeSavedInfo> CSI)201 static void updateLiveness(MachineFunction &MF, ArrayRef<CalleeSavedInfo> CSI) {
202   MachineBasicBlock &EntryBB = MF.front();
203 
204   for (const CalleeSavedInfo &CSIReg : CSI)
205     EntryBB.addLiveIn(CSIReg.getReg());
206   EntryBB.sortUniqueLiveIns();
207 }
208 
spillCalleeSavedRegs(MachineFunction & MF,SmallVectorImpl<int> & CalleeSavedFIs)209 bool SILowerSGPRSpills::spillCalleeSavedRegs(
210     MachineFunction &MF, SmallVectorImpl<int> &CalleeSavedFIs) {
211   MachineRegisterInfo &MRI = MF.getRegInfo();
212   const Function &F = MF.getFunction();
213   const GCNSubtarget &ST = MF.getSubtarget<GCNSubtarget>();
214   const SIFrameLowering *TFI = ST.getFrameLowering();
215   MachineFrameInfo &MFI = MF.getFrameInfo();
216   RegScavenger *RS = nullptr;
217 
218   // Determine which of the registers in the callee save list should be saved.
219   BitVector SavedRegs;
220   TFI->determineCalleeSavesSGPR(MF, SavedRegs, RS);
221 
222   // Add the code to save and restore the callee saved registers.
223   if (!F.hasFnAttribute(Attribute::Naked)) {
224     // FIXME: This is a lie. The CalleeSavedInfo is incomplete, but this is
225     // necessary for verifier liveness checks.
226     MFI.setCalleeSavedInfoValid(true);
227 
228     std::vector<CalleeSavedInfo> CSI;
229     const MCPhysReg *CSRegs = MRI.getCalleeSavedRegs();
230 
231     for (unsigned I = 0; CSRegs[I]; ++I) {
232       MCRegister Reg = CSRegs[I];
233 
234       if (SavedRegs.test(Reg)) {
235         const TargetRegisterClass *RC =
236           TRI->getMinimalPhysRegClass(Reg, MVT::i32);
237         int JunkFI = MFI.CreateStackObject(TRI->getSpillSize(*RC),
238                                            TRI->getSpillAlign(*RC), true);
239 
240         CSI.emplace_back(Reg, JunkFI);
241         CalleeSavedFIs.push_back(JunkFI);
242       }
243     }
244 
245     if (!CSI.empty()) {
246       for (MachineBasicBlock *SaveBlock : SaveBlocks)
247         insertCSRSaves(*SaveBlock, CSI, Indexes, LIS);
248 
249       // Add live ins to save blocks.
250       assert(SaveBlocks.size() == 1 && "shrink wrapping not fully implemented");
251       updateLiveness(MF, CSI);
252 
253       for (MachineBasicBlock *RestoreBlock : RestoreBlocks)
254         insertCSRRestores(*RestoreBlock, CSI, Indexes, LIS);
255       return true;
256     }
257   }
258 
259   return false;
260 }
261 
extendWWMVirtRegLiveness(MachineFunction & MF,LiveIntervals * LIS)262 void SILowerSGPRSpills::extendWWMVirtRegLiveness(MachineFunction &MF,
263                                                  LiveIntervals *LIS) {
264   // TODO: This is a workaround to avoid the unmodelled liveness computed with
265   // whole-wave virtual registers when allocated together with the regular VGPR
266   // virtual registers. Presently, the liveness computed during the regalloc is
267   // only uniform (or single lane aware) and it doesn't take account of the
268   // divergent control flow that exists for our GPUs. Since the WWM registers
269   // can modify inactive lanes, the wave-aware liveness should be computed for
270   // the virtual registers to accurately plot their interferences. Without
271   // having the divergent CFG for the function, it is difficult to implement the
272   // wave-aware liveness info. Until then, we conservatively extend the liveness
273   // of the wwm registers into the entire function so that they won't be reused
274   // without first spilling/splitting their liveranges.
275   SIMachineFunctionInfo *MFI = MF.getInfo<SIMachineFunctionInfo>();
276 
277   // Insert the IMPLICIT_DEF for the wwm-registers in the entry blocks.
278   for (auto Reg : MFI->getSGPRSpillVGPRs()) {
279     for (MachineBasicBlock *SaveBlock : SaveBlocks) {
280       MachineBasicBlock::iterator InsertBefore = SaveBlock->begin();
281       DebugLoc DL = SaveBlock->findDebugLoc(InsertBefore);
282       auto MIB = BuildMI(*SaveBlock, InsertBefore, DL,
283                          TII->get(AMDGPU::IMPLICIT_DEF), Reg);
284       MFI->setFlag(Reg, AMDGPU::VirtRegFlag::WWM_REG);
285       // Set SGPR_SPILL asm printer flag
286       MIB->setAsmPrinterFlag(AMDGPU::SGPR_SPILL);
287       if (LIS) {
288         LIS->InsertMachineInstrInMaps(*MIB);
289       }
290     }
291   }
292 
293   // Insert the KILL in the return blocks to extend their liveness untill the
294   // end of function. Insert a separate KILL for each VGPR.
295   for (MachineBasicBlock *RestoreBlock : RestoreBlocks) {
296     MachineBasicBlock::iterator InsertBefore =
297         RestoreBlock->getFirstTerminator();
298     DebugLoc DL = RestoreBlock->findDebugLoc(InsertBefore);
299     for (auto Reg : MFI->getSGPRSpillVGPRs()) {
300       auto MIB = BuildMI(*RestoreBlock, InsertBefore, DL,
301                          TII->get(TargetOpcode::KILL));
302       MIB.addReg(Reg);
303       if (LIS)
304         LIS->InsertMachineInstrInMaps(*MIB);
305     }
306   }
307 }
308 
runOnMachineFunction(MachineFunction & MF)309 bool SILowerSGPRSpills::runOnMachineFunction(MachineFunction &MF) {
310   const GCNSubtarget &ST = MF.getSubtarget<GCNSubtarget>();
311   TII = ST.getInstrInfo();
312   TRI = &TII->getRegisterInfo();
313 
314   auto *LISWrapper = getAnalysisIfAvailable<LiveIntervalsWrapperPass>();
315   LIS = LISWrapper ? &LISWrapper->getLIS() : nullptr;
316   auto *SIWrapper = getAnalysisIfAvailable<SlotIndexesWrapperPass>();
317   Indexes = SIWrapper ? &SIWrapper->getSI() : nullptr;
318 
319   assert(SaveBlocks.empty() && RestoreBlocks.empty());
320 
321   // First, expose any CSR SGPR spills. This is mostly the same as what PEI
322   // does, but somewhat simpler.
323   calculateSaveRestoreBlocks(MF);
324   SmallVector<int> CalleeSavedFIs;
325   bool HasCSRs = spillCalleeSavedRegs(MF, CalleeSavedFIs);
326 
327   MachineFrameInfo &MFI = MF.getFrameInfo();
328   MachineRegisterInfo &MRI = MF.getRegInfo();
329   SIMachineFunctionInfo *FuncInfo = MF.getInfo<SIMachineFunctionInfo>();
330 
331   if (!MFI.hasStackObjects() && !HasCSRs) {
332     SaveBlocks.clear();
333     RestoreBlocks.clear();
334     return false;
335   }
336 
337   bool MadeChange = false;
338   bool SpilledToVirtVGPRLanes = false;
339 
340   // TODO: CSR VGPRs will never be spilled to AGPRs. These can probably be
341   // handled as SpilledToReg in regular PrologEpilogInserter.
342   const bool HasSGPRSpillToVGPR = TRI->spillSGPRToVGPR() &&
343                                   (HasCSRs || FuncInfo->hasSpilledSGPRs());
344   if (HasSGPRSpillToVGPR) {
345     // Process all SGPR spills before frame offsets are finalized. Ideally SGPRs
346     // are spilled to VGPRs, in which case we can eliminate the stack usage.
347     //
348     // This operates under the assumption that only other SGPR spills are users
349     // of the frame index.
350 
351     // To track the spill frame indices handled in this pass.
352     BitVector SpillFIs(MFI.getObjectIndexEnd(), false);
353 
354     for (MachineBasicBlock &MBB : MF) {
355       for (MachineInstr &MI : llvm::make_early_inc_range(MBB)) {
356         if (!TII->isSGPRSpill(MI))
357           continue;
358 
359         int FI = TII->getNamedOperand(MI, AMDGPU::OpName::addr)->getIndex();
360         assert(MFI.getStackID(FI) == TargetStackID::SGPRSpill);
361 
362         bool IsCalleeSaveSGPRSpill = llvm::is_contained(CalleeSavedFIs, FI);
363         if (IsCalleeSaveSGPRSpill) {
364           // Spill callee-saved SGPRs into physical VGPR lanes.
365 
366           // TODO: This is to ensure the CFIs are static for efficient frame
367           // unwinding in the debugger. Spilling them into virtual VGPR lanes
368           // involve regalloc to allocate the physical VGPRs and that might
369           // cause intermediate spill/split of such liveranges for successful
370           // allocation. This would result in broken CFI encoding unless the
371           // regalloc aware CFI generation to insert new CFIs along with the
372           // intermediate spills is implemented. There is no such support
373           // currently exist in the LLVM compiler.
374           if (FuncInfo->allocateSGPRSpillToVGPRLane(
375                   MF, FI, /*SpillToPhysVGPRLane=*/true)) {
376             bool Spilled = TRI->eliminateSGPRToVGPRSpillFrameIndex(
377                 MI, FI, nullptr, Indexes, LIS, true);
378             if (!Spilled)
379               llvm_unreachable(
380                   "failed to spill SGPR to physical VGPR lane when allocated");
381           }
382         } else {
383           if (FuncInfo->allocateSGPRSpillToVGPRLane(MF, FI)) {
384             bool Spilled = TRI->eliminateSGPRToVGPRSpillFrameIndex(
385                 MI, FI, nullptr, Indexes, LIS);
386             if (!Spilled)
387               llvm_unreachable(
388                   "failed to spill SGPR to virtual VGPR lane when allocated");
389             SpillFIs.set(FI);
390             SpilledToVirtVGPRLanes = true;
391           }
392         }
393       }
394     }
395 
396     if (SpilledToVirtVGPRLanes) {
397       extendWWMVirtRegLiveness(MF, LIS);
398       if (LIS) {
399         // Compute the LiveInterval for the newly created virtual registers.
400         for (auto Reg : FuncInfo->getSGPRSpillVGPRs())
401           LIS->createAndComputeVirtRegInterval(Reg);
402       }
403     }
404 
405     for (MachineBasicBlock &MBB : MF) {
406       // FIXME: The dead frame indices are replaced with a null register from
407       // the debug value instructions. We should instead, update it with the
408       // correct register value. But not sure the register value alone is
409       // adequate to lower the DIExpression. It should be worked out later.
410       for (MachineInstr &MI : MBB) {
411         if (MI.isDebugValue() && MI.getOperand(0).isFI() &&
412             !MFI.isFixedObjectIndex(MI.getOperand(0).getIndex()) &&
413             SpillFIs[MI.getOperand(0).getIndex()]) {
414           MI.getOperand(0).ChangeToRegister(Register(), false /*isDef*/);
415         }
416       }
417     }
418 
419     // All those frame indices which are dead by now should be removed from the
420     // function frame. Otherwise, there is a side effect such as re-mapping of
421     // free frame index ids by the later pass(es) like "stack slot coloring"
422     // which in turn could mess-up with the book keeping of "frame index to VGPR
423     // lane".
424     FuncInfo->removeDeadFrameIndices(MFI, /*ResetSGPRSpillStackIDs*/ false);
425 
426     MadeChange = true;
427   }
428 
429   if (SpilledToVirtVGPRLanes) {
430     const TargetRegisterClass *RC = TRI->getWaveMaskRegClass();
431     // Shift back the reserved SGPR for EXEC copy into the lowest range.
432     // This SGPR is reserved to handle the whole-wave spill/copy operations
433     // that might get inserted during vgpr regalloc.
434     Register UnusedLowSGPR = TRI->findUnusedRegister(MRI, RC, MF);
435     if (UnusedLowSGPR && TRI->getHWRegIndex(UnusedLowSGPR) <
436                              TRI->getHWRegIndex(FuncInfo->getSGPRForEXECCopy()))
437       FuncInfo->setSGPRForEXECCopy(UnusedLowSGPR);
438   } else {
439     // No SGPR spills to virtual VGPR lanes and hence there won't be any WWM
440     // spills/copies. Reset the SGPR reserved for EXEC copy.
441     FuncInfo->setSGPRForEXECCopy(AMDGPU::NoRegister);
442   }
443 
444   SaveBlocks.clear();
445   RestoreBlocks.clear();
446 
447   return MadeChange;
448 }
449