xref: /freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/SIInstrInfo.h (revision d9a42747950146bf03cda7f6e25d219253f8a57a)
1 //===- SIInstrInfo.h - SI Instruction Info Interface ------------*- C++ -*-===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 //
9 /// \file
10 /// Interface definition for SIInstrInfo.
11 //
12 //===----------------------------------------------------------------------===//
13 
14 #ifndef LLVM_LIB_TARGET_AMDGPU_SIINSTRINFO_H
15 #define LLVM_LIB_TARGET_AMDGPU_SIINSTRINFO_H
16 
17 #include "AMDGPUMIRFormatter.h"
18 #include "MCTargetDesc/AMDGPUMCTargetDesc.h"
19 #include "SIRegisterInfo.h"
20 #include "Utils/AMDGPUBaseInfo.h"
21 #include "llvm/ADT/SetVector.h"
22 #include "llvm/CodeGen/TargetInstrInfo.h"
23 #include "llvm/CodeGen/TargetSchedule.h"
24 
25 #define GET_INSTRINFO_HEADER
26 #include "AMDGPUGenInstrInfo.inc"
27 
28 namespace llvm {
29 
30 class APInt;
31 class GCNSubtarget;
32 class LiveVariables;
33 class MachineDominatorTree;
34 class MachineRegisterInfo;
35 class RegScavenger;
36 class TargetRegisterClass;
37 class ScheduleHazardRecognizer;
38 
39 /// Mark the MMO of a uniform load if there are no potentially clobbering stores
40 /// on any path from the start of an entry function to this load.
41 static const MachineMemOperand::Flags MONoClobber =
42     MachineMemOperand::MOTargetFlag1;
43 
44 class SIInstrInfo final : public AMDGPUGenInstrInfo {
45 private:
46   const SIRegisterInfo RI;
47   const GCNSubtarget &ST;
48   TargetSchedModel SchedModel;
49   mutable std::unique_ptr<AMDGPUMIRFormatter> Formatter;
50 
51   // The inverse predicate should have the negative value.
52   enum BranchPredicate {
53     INVALID_BR = 0,
54     SCC_TRUE = 1,
55     SCC_FALSE = -1,
56     VCCNZ = 2,
57     VCCZ = -2,
58     EXECNZ = -3,
59     EXECZ = 3
60   };
61 
62   using SetVectorType = SmallSetVector<MachineInstr *, 32>;
63 
64   static unsigned getBranchOpcode(BranchPredicate Cond);
65   static BranchPredicate getBranchPredicate(unsigned Opcode);
66 
67 public:
68   unsigned buildExtractSubReg(MachineBasicBlock::iterator MI,
69                               MachineRegisterInfo &MRI,
70                               MachineOperand &SuperReg,
71                               const TargetRegisterClass *SuperRC,
72                               unsigned SubIdx,
73                               const TargetRegisterClass *SubRC) const;
74   MachineOperand buildExtractSubRegOrImm(MachineBasicBlock::iterator MI,
75                                          MachineRegisterInfo &MRI,
76                                          MachineOperand &SuperReg,
77                                          const TargetRegisterClass *SuperRC,
78                                          unsigned SubIdx,
79                                          const TargetRegisterClass *SubRC) const;
80 private:
81   void swapOperands(MachineInstr &Inst) const;
82 
83   std::pair<bool, MachineBasicBlock *>
84   moveScalarAddSub(SetVectorType &Worklist, MachineInstr &Inst,
85                    MachineDominatorTree *MDT = nullptr) const;
86 
87   void lowerSelect(SetVectorType &Worklist, MachineInstr &Inst,
88                    MachineDominatorTree *MDT = nullptr) const;
89 
90   void lowerScalarAbs(SetVectorType &Worklist,
91                       MachineInstr &Inst) const;
92 
93   void lowerScalarXnor(SetVectorType &Worklist,
94                        MachineInstr &Inst) const;
95 
96   void splitScalarNotBinop(SetVectorType &Worklist,
97                            MachineInstr &Inst,
98                            unsigned Opcode) const;
99 
100   void splitScalarBinOpN2(SetVectorType &Worklist,
101                           MachineInstr &Inst,
102                           unsigned Opcode) const;
103 
104   void splitScalar64BitUnaryOp(SetVectorType &Worklist,
105                                MachineInstr &Inst, unsigned Opcode,
106                                bool Swap = false) const;
107 
108   void splitScalar64BitAddSub(SetVectorType &Worklist, MachineInstr &Inst,
109                               MachineDominatorTree *MDT = nullptr) const;
110 
111   void splitScalar64BitBinaryOp(SetVectorType &Worklist, MachineInstr &Inst,
112                                 unsigned Opcode,
113                                 MachineDominatorTree *MDT = nullptr) const;
114 
115   void splitScalar64BitXnor(SetVectorType &Worklist, MachineInstr &Inst,
116                                 MachineDominatorTree *MDT = nullptr) const;
117 
118   void splitScalar64BitBCNT(SetVectorType &Worklist,
119                             MachineInstr &Inst) const;
120   void splitScalar64BitBFE(SetVectorType &Worklist,
121                            MachineInstr &Inst) const;
122   void movePackToVALU(SetVectorType &Worklist,
123                       MachineRegisterInfo &MRI,
124                       MachineInstr &Inst) const;
125 
126   void addUsersToMoveToVALUWorklist(Register Reg, MachineRegisterInfo &MRI,
127                                     SetVectorType &Worklist) const;
128 
129   void addSCCDefUsersToVALUWorklist(MachineOperand &Op,
130                                     MachineInstr &SCCDefInst,
131                                     SetVectorType &Worklist,
132                                     Register NewCond = Register()) const;
133   void addSCCDefsToVALUWorklist(MachineOperand &Op,
134                                 SetVectorType &Worklist) const;
135 
136   const TargetRegisterClass *
137   getDestEquivalentVGPRClass(const MachineInstr &Inst) const;
138 
139   bool checkInstOffsetsDoNotOverlap(const MachineInstr &MIa,
140                                     const MachineInstr &MIb) const;
141 
142   Register findUsedSGPR(const MachineInstr &MI, int OpIndices[3]) const;
143 
144 protected:
145   bool swapSourceModifiers(MachineInstr &MI,
146                            MachineOperand &Src0, unsigned Src0OpName,
147                            MachineOperand &Src1, unsigned Src1OpName) const;
148 
149   MachineInstr *commuteInstructionImpl(MachineInstr &MI, bool NewMI,
150                                        unsigned OpIdx0,
151                                        unsigned OpIdx1) const override;
152 
153 public:
154   enum TargetOperandFlags {
155     MO_MASK = 0xf,
156 
157     MO_NONE = 0,
158     // MO_GOTPCREL -> symbol@GOTPCREL -> R_AMDGPU_GOTPCREL.
159     MO_GOTPCREL = 1,
160     // MO_GOTPCREL32_LO -> symbol@gotpcrel32@lo -> R_AMDGPU_GOTPCREL32_LO.
161     MO_GOTPCREL32 = 2,
162     MO_GOTPCREL32_LO = 2,
163     // MO_GOTPCREL32_HI -> symbol@gotpcrel32@hi -> R_AMDGPU_GOTPCREL32_HI.
164     MO_GOTPCREL32_HI = 3,
165     // MO_REL32_LO -> symbol@rel32@lo -> R_AMDGPU_REL32_LO.
166     MO_REL32 = 4,
167     MO_REL32_LO = 4,
168     // MO_REL32_HI -> symbol@rel32@hi -> R_AMDGPU_REL32_HI.
169     MO_REL32_HI = 5,
170 
171     MO_FAR_BRANCH_OFFSET = 6,
172 
173     MO_ABS32_LO = 8,
174     MO_ABS32_HI = 9,
175   };
176 
177   explicit SIInstrInfo(const GCNSubtarget &ST);
178 
179   const SIRegisterInfo &getRegisterInfo() const {
180     return RI;
181   }
182 
183   const GCNSubtarget &getSubtarget() const {
184     return ST;
185   }
186 
187   bool isReallyTriviallyReMaterializable(const MachineInstr &MI) const override;
188 
189   bool isIgnorableUse(const MachineOperand &MO) const override;
190 
191   bool areLoadsFromSameBasePtr(SDNode *Load1, SDNode *Load2,
192                                int64_t &Offset1,
193                                int64_t &Offset2) const override;
194 
195   bool getMemOperandsWithOffsetWidth(
196       const MachineInstr &LdSt,
197       SmallVectorImpl<const MachineOperand *> &BaseOps, int64_t &Offset,
198       bool &OffsetIsScalable, unsigned &Width,
199       const TargetRegisterInfo *TRI) const final;
200 
201   bool shouldClusterMemOps(ArrayRef<const MachineOperand *> BaseOps1,
202                            ArrayRef<const MachineOperand *> BaseOps2,
203                            unsigned NumLoads, unsigned NumBytes) const override;
204 
205   bool shouldScheduleLoadsNear(SDNode *Load0, SDNode *Load1, int64_t Offset0,
206                                int64_t Offset1, unsigned NumLoads) const override;
207 
208   void copyPhysReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI,
209                    const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg,
210                    bool KillSrc) const override;
211 
212   void materializeImmediate(MachineBasicBlock &MBB,
213                             MachineBasicBlock::iterator MI,
214                             const DebugLoc &DL,
215                             unsigned DestReg,
216                             int64_t Value) const;
217 
218   const TargetRegisterClass *getPreferredSelectRegClass(
219                                unsigned Size) const;
220 
221   Register insertNE(MachineBasicBlock *MBB,
222                     MachineBasicBlock::iterator I, const DebugLoc &DL,
223                     Register SrcReg, int Value) const;
224 
225   Register insertEQ(MachineBasicBlock *MBB,
226                     MachineBasicBlock::iterator I, const DebugLoc &DL,
227                     Register SrcReg, int Value)  const;
228 
229   void storeRegToStackSlot(MachineBasicBlock &MBB,
230                            MachineBasicBlock::iterator MI, Register SrcReg,
231                            bool isKill, int FrameIndex,
232                            const TargetRegisterClass *RC,
233                            const TargetRegisterInfo *TRI) const override;
234 
235   void loadRegFromStackSlot(MachineBasicBlock &MBB,
236                             MachineBasicBlock::iterator MI, Register DestReg,
237                             int FrameIndex, const TargetRegisterClass *RC,
238                             const TargetRegisterInfo *TRI) const override;
239 
240   bool expandPostRAPseudo(MachineInstr &MI) const override;
241 
242   // Splits a V_MOV_B64_DPP_PSEUDO opcode into a pair of v_mov_b32_dpp
243   // instructions. Returns a pair of generated instructions.
244   // Can split either post-RA with physical registers or pre-RA with
245   // virtual registers. In latter case IR needs to be in SSA form and
246   // and a REG_SEQUENCE is produced to define original register.
247   std::pair<MachineInstr*, MachineInstr*>
248   expandMovDPP64(MachineInstr &MI) const;
249 
250   // Returns an opcode that can be used to move a value to a \p DstRC
251   // register.  If there is no hardware instruction that can store to \p
252   // DstRC, then AMDGPU::COPY is returned.
253   unsigned getMovOpcode(const TargetRegisterClass *DstRC) const;
254 
255   const MCInstrDesc &getIndirectRegWriteMovRelPseudo(unsigned VecSize,
256                                                      unsigned EltSize,
257                                                      bool IsSGPR) const;
258 
259   const MCInstrDesc &getIndirectGPRIDXPseudo(unsigned VecSize,
260                                              bool IsIndirectSrc) const;
261   LLVM_READONLY
262   int commuteOpcode(unsigned Opc) const;
263 
264   LLVM_READONLY
265   inline int commuteOpcode(const MachineInstr &MI) const {
266     return commuteOpcode(MI.getOpcode());
267   }
268 
269   bool findCommutedOpIndices(const MachineInstr &MI, unsigned &SrcOpIdx1,
270                              unsigned &SrcOpIdx2) const override;
271 
272   bool findCommutedOpIndices(MCInstrDesc Desc, unsigned & SrcOpIdx0,
273    unsigned & SrcOpIdx1) const;
274 
275   bool isBranchOffsetInRange(unsigned BranchOpc,
276                              int64_t BrOffset) const override;
277 
278   MachineBasicBlock *getBranchDestBlock(const MachineInstr &MI) const override;
279 
280   void insertIndirectBranch(MachineBasicBlock &MBB,
281                             MachineBasicBlock &NewDestBB,
282                             MachineBasicBlock &RestoreBB, const DebugLoc &DL,
283                             int64_t BrOffset, RegScavenger *RS) const override;
284 
285   bool analyzeBranchImpl(MachineBasicBlock &MBB,
286                          MachineBasicBlock::iterator I,
287                          MachineBasicBlock *&TBB,
288                          MachineBasicBlock *&FBB,
289                          SmallVectorImpl<MachineOperand> &Cond,
290                          bool AllowModify) const;
291 
292   bool analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB,
293                      MachineBasicBlock *&FBB,
294                      SmallVectorImpl<MachineOperand> &Cond,
295                      bool AllowModify = false) const override;
296 
297   unsigned removeBranch(MachineBasicBlock &MBB,
298                         int *BytesRemoved = nullptr) const override;
299 
300   unsigned insertBranch(MachineBasicBlock &MBB, MachineBasicBlock *TBB,
301                         MachineBasicBlock *FBB, ArrayRef<MachineOperand> Cond,
302                         const DebugLoc &DL,
303                         int *BytesAdded = nullptr) const override;
304 
305   bool reverseBranchCondition(
306     SmallVectorImpl<MachineOperand> &Cond) const override;
307 
308   bool canInsertSelect(const MachineBasicBlock &MBB,
309                        ArrayRef<MachineOperand> Cond, Register DstReg,
310                        Register TrueReg, Register FalseReg, int &CondCycles,
311                        int &TrueCycles, int &FalseCycles) const override;
312 
313   void insertSelect(MachineBasicBlock &MBB,
314                     MachineBasicBlock::iterator I, const DebugLoc &DL,
315                     Register DstReg, ArrayRef<MachineOperand> Cond,
316                     Register TrueReg, Register FalseReg) const override;
317 
318   void insertVectorSelect(MachineBasicBlock &MBB,
319                           MachineBasicBlock::iterator I, const DebugLoc &DL,
320                           Register DstReg, ArrayRef<MachineOperand> Cond,
321                           Register TrueReg, Register FalseReg) const;
322 
323   bool analyzeCompare(const MachineInstr &MI, Register &SrcReg,
324                       Register &SrcReg2, int64_t &CmpMask,
325                       int64_t &CmpValue) const override;
326 
327   bool optimizeCompareInstr(MachineInstr &CmpInstr, Register SrcReg,
328                             Register SrcReg2, int64_t CmpMask, int64_t CmpValue,
329                             const MachineRegisterInfo *MRI) const override;
330 
331   bool
332   areMemAccessesTriviallyDisjoint(const MachineInstr &MIa,
333                                   const MachineInstr &MIb) const override;
334 
335   static bool isFoldableCopy(const MachineInstr &MI);
336 
337   void removeModOperands(MachineInstr &MI) const;
338 
339   bool FoldImmediate(MachineInstr &UseMI, MachineInstr &DefMI, Register Reg,
340                      MachineRegisterInfo *MRI) const final;
341 
342   unsigned getMachineCSELookAheadLimit() const override { return 500; }
343 
344   MachineInstr *convertToThreeAddress(MachineInstr &MI, LiveVariables *LV,
345                                       LiveIntervals *LIS) const override;
346 
347   bool isSchedulingBoundary(const MachineInstr &MI,
348                             const MachineBasicBlock *MBB,
349                             const MachineFunction &MF) const override;
350 
351   static bool isSALU(const MachineInstr &MI) {
352     return MI.getDesc().TSFlags & SIInstrFlags::SALU;
353   }
354 
355   bool isSALU(uint16_t Opcode) const {
356     return get(Opcode).TSFlags & SIInstrFlags::SALU;
357   }
358 
359   static bool isVALU(const MachineInstr &MI) {
360     return MI.getDesc().TSFlags & SIInstrFlags::VALU;
361   }
362 
363   bool isVALU(uint16_t Opcode) const {
364     return get(Opcode).TSFlags & SIInstrFlags::VALU;
365   }
366 
367   static bool isVMEM(const MachineInstr &MI) {
368     return isMUBUF(MI) || isMTBUF(MI) || isMIMG(MI);
369   }
370 
371   bool isVMEM(uint16_t Opcode) const {
372     return isMUBUF(Opcode) || isMTBUF(Opcode) || isMIMG(Opcode);
373   }
374 
375   static bool isSOP1(const MachineInstr &MI) {
376     return MI.getDesc().TSFlags & SIInstrFlags::SOP1;
377   }
378 
379   bool isSOP1(uint16_t Opcode) const {
380     return get(Opcode).TSFlags & SIInstrFlags::SOP1;
381   }
382 
383   static bool isSOP2(const MachineInstr &MI) {
384     return MI.getDesc().TSFlags & SIInstrFlags::SOP2;
385   }
386 
387   bool isSOP2(uint16_t Opcode) const {
388     return get(Opcode).TSFlags & SIInstrFlags::SOP2;
389   }
390 
391   static bool isSOPC(const MachineInstr &MI) {
392     return MI.getDesc().TSFlags & SIInstrFlags::SOPC;
393   }
394 
395   bool isSOPC(uint16_t Opcode) const {
396     return get(Opcode).TSFlags & SIInstrFlags::SOPC;
397   }
398 
399   static bool isSOPK(const MachineInstr &MI) {
400     return MI.getDesc().TSFlags & SIInstrFlags::SOPK;
401   }
402 
403   bool isSOPK(uint16_t Opcode) const {
404     return get(Opcode).TSFlags & SIInstrFlags::SOPK;
405   }
406 
407   static bool isSOPP(const MachineInstr &MI) {
408     return MI.getDesc().TSFlags & SIInstrFlags::SOPP;
409   }
410 
411   bool isSOPP(uint16_t Opcode) const {
412     return get(Opcode).TSFlags & SIInstrFlags::SOPP;
413   }
414 
415   static bool isPacked(const MachineInstr &MI) {
416     return MI.getDesc().TSFlags & SIInstrFlags::IsPacked;
417   }
418 
419   bool isPacked(uint16_t Opcode) const {
420     return get(Opcode).TSFlags & SIInstrFlags::IsPacked;
421   }
422 
423   static bool isVOP1(const MachineInstr &MI) {
424     return MI.getDesc().TSFlags & SIInstrFlags::VOP1;
425   }
426 
427   bool isVOP1(uint16_t Opcode) const {
428     return get(Opcode).TSFlags & SIInstrFlags::VOP1;
429   }
430 
431   static bool isVOP2(const MachineInstr &MI) {
432     return MI.getDesc().TSFlags & SIInstrFlags::VOP2;
433   }
434 
435   bool isVOP2(uint16_t Opcode) const {
436     return get(Opcode).TSFlags & SIInstrFlags::VOP2;
437   }
438 
439   static bool isVOP3(const MachineInstr &MI) {
440     return MI.getDesc().TSFlags & SIInstrFlags::VOP3;
441   }
442 
443   bool isVOP3(uint16_t Opcode) const {
444     return get(Opcode).TSFlags & SIInstrFlags::VOP3;
445   }
446 
447   static bool isSDWA(const MachineInstr &MI) {
448     return MI.getDesc().TSFlags & SIInstrFlags::SDWA;
449   }
450 
451   bool isSDWA(uint16_t Opcode) const {
452     return get(Opcode).TSFlags & SIInstrFlags::SDWA;
453   }
454 
455   static bool isVOPC(const MachineInstr &MI) {
456     return MI.getDesc().TSFlags & SIInstrFlags::VOPC;
457   }
458 
459   bool isVOPC(uint16_t Opcode) const {
460     return get(Opcode).TSFlags & SIInstrFlags::VOPC;
461   }
462 
463   static bool isMUBUF(const MachineInstr &MI) {
464     return MI.getDesc().TSFlags & SIInstrFlags::MUBUF;
465   }
466 
467   bool isMUBUF(uint16_t Opcode) const {
468     return get(Opcode).TSFlags & SIInstrFlags::MUBUF;
469   }
470 
471   static bool isMTBUF(const MachineInstr &MI) {
472     return MI.getDesc().TSFlags & SIInstrFlags::MTBUF;
473   }
474 
475   bool isMTBUF(uint16_t Opcode) const {
476     return get(Opcode).TSFlags & SIInstrFlags::MTBUF;
477   }
478 
479   static bool isSMRD(const MachineInstr &MI) {
480     return MI.getDesc().TSFlags & SIInstrFlags::SMRD;
481   }
482 
483   bool isSMRD(uint16_t Opcode) const {
484     return get(Opcode).TSFlags & SIInstrFlags::SMRD;
485   }
486 
487   bool isBufferSMRD(const MachineInstr &MI) const;
488 
489   static bool isDS(const MachineInstr &MI) {
490     return MI.getDesc().TSFlags & SIInstrFlags::DS;
491   }
492 
493   bool isDS(uint16_t Opcode) const {
494     return get(Opcode).TSFlags & SIInstrFlags::DS;
495   }
496 
497   bool isAlwaysGDS(uint16_t Opcode) const;
498 
499   static bool isMIMG(const MachineInstr &MI) {
500     return MI.getDesc().TSFlags & SIInstrFlags::MIMG;
501   }
502 
503   bool isMIMG(uint16_t Opcode) const {
504     return get(Opcode).TSFlags & SIInstrFlags::MIMG;
505   }
506 
507   static bool isGather4(const MachineInstr &MI) {
508     return MI.getDesc().TSFlags & SIInstrFlags::Gather4;
509   }
510 
511   bool isGather4(uint16_t Opcode) const {
512     return get(Opcode).TSFlags & SIInstrFlags::Gather4;
513   }
514 
515   static bool isFLAT(const MachineInstr &MI) {
516     return MI.getDesc().TSFlags & SIInstrFlags::FLAT;
517   }
518 
519   // Is a FLAT encoded instruction which accesses a specific segment,
520   // i.e. global_* or scratch_*.
521   static bool isSegmentSpecificFLAT(const MachineInstr &MI) {
522     auto Flags = MI.getDesc().TSFlags;
523     return Flags & (SIInstrFlags::FlatGlobal | SIInstrFlags::FlatScratch);
524   }
525 
526   bool isSegmentSpecificFLAT(uint16_t Opcode) const {
527     auto Flags = get(Opcode).TSFlags;
528     return Flags & (SIInstrFlags::FlatGlobal | SIInstrFlags::FlatScratch);
529   }
530 
531   static bool isFLATGlobal(const MachineInstr &MI) {
532     return MI.getDesc().TSFlags & SIInstrFlags::FlatGlobal;
533   }
534 
535   bool isFLATGlobal(uint16_t Opcode) const {
536     return get(Opcode).TSFlags & SIInstrFlags::FlatGlobal;
537   }
538 
539   static bool isFLATScratch(const MachineInstr &MI) {
540     return MI.getDesc().TSFlags & SIInstrFlags::FlatScratch;
541   }
542 
543   bool isFLATScratch(uint16_t Opcode) const {
544     return get(Opcode).TSFlags & SIInstrFlags::FlatScratch;
545   }
546 
547   // Any FLAT encoded instruction, including global_* and scratch_*.
548   bool isFLAT(uint16_t Opcode) const {
549     return get(Opcode).TSFlags & SIInstrFlags::FLAT;
550   }
551 
552   static bool isEXP(const MachineInstr &MI) {
553     return MI.getDesc().TSFlags & SIInstrFlags::EXP;
554   }
555 
556   static bool isDualSourceBlendEXP(const MachineInstr &MI) {
557     if (!isEXP(MI))
558       return false;
559     unsigned Target = MI.getOperand(0).getImm();
560     return Target == AMDGPU::Exp::ET_DUAL_SRC_BLEND0 ||
561            Target == AMDGPU::Exp::ET_DUAL_SRC_BLEND1;
562   }
563 
564   bool isEXP(uint16_t Opcode) const {
565     return get(Opcode).TSFlags & SIInstrFlags::EXP;
566   }
567 
568   static bool isAtomicNoRet(const MachineInstr &MI) {
569     return MI.getDesc().TSFlags & SIInstrFlags::IsAtomicNoRet;
570   }
571 
572   bool isAtomicNoRet(uint16_t Opcode) const {
573     return get(Opcode).TSFlags & SIInstrFlags::IsAtomicNoRet;
574   }
575 
576   static bool isAtomicRet(const MachineInstr &MI) {
577     return MI.getDesc().TSFlags & SIInstrFlags::IsAtomicRet;
578   }
579 
580   bool isAtomicRet(uint16_t Opcode) const {
581     return get(Opcode).TSFlags & SIInstrFlags::IsAtomicRet;
582   }
583 
584   static bool isAtomic(const MachineInstr &MI) {
585     return MI.getDesc().TSFlags & (SIInstrFlags::IsAtomicRet |
586                                    SIInstrFlags::IsAtomicNoRet);
587   }
588 
589   bool isAtomic(uint16_t Opcode) const {
590     return get(Opcode).TSFlags & (SIInstrFlags::IsAtomicRet |
591                                   SIInstrFlags::IsAtomicNoRet);
592   }
593 
594   static bool isWQM(const MachineInstr &MI) {
595     return MI.getDesc().TSFlags & SIInstrFlags::WQM;
596   }
597 
598   bool isWQM(uint16_t Opcode) const {
599     return get(Opcode).TSFlags & SIInstrFlags::WQM;
600   }
601 
602   static bool isDisableWQM(const MachineInstr &MI) {
603     return MI.getDesc().TSFlags & SIInstrFlags::DisableWQM;
604   }
605 
606   bool isDisableWQM(uint16_t Opcode) const {
607     return get(Opcode).TSFlags & SIInstrFlags::DisableWQM;
608   }
609 
610   static bool isVGPRSpill(const MachineInstr &MI) {
611     return MI.getDesc().TSFlags & SIInstrFlags::VGPRSpill;
612   }
613 
614   bool isVGPRSpill(uint16_t Opcode) const {
615     return get(Opcode).TSFlags & SIInstrFlags::VGPRSpill;
616   }
617 
618   static bool isSGPRSpill(const MachineInstr &MI) {
619     return MI.getDesc().TSFlags & SIInstrFlags::SGPRSpill;
620   }
621 
622   bool isSGPRSpill(uint16_t Opcode) const {
623     return get(Opcode).TSFlags & SIInstrFlags::SGPRSpill;
624   }
625 
626   static bool isDPP(const MachineInstr &MI) {
627     return MI.getDesc().TSFlags & SIInstrFlags::DPP;
628   }
629 
630   bool isDPP(uint16_t Opcode) const {
631     return get(Opcode).TSFlags & SIInstrFlags::DPP;
632   }
633 
634   static bool isTRANS(const MachineInstr &MI) {
635     return MI.getDesc().TSFlags & SIInstrFlags::TRANS;
636   }
637 
638   bool isTRANS(uint16_t Opcode) const {
639     return get(Opcode).TSFlags & SIInstrFlags::TRANS;
640   }
641 
642   static bool isVOP3P(const MachineInstr &MI) {
643     return MI.getDesc().TSFlags & SIInstrFlags::VOP3P;
644   }
645 
646   bool isVOP3P(uint16_t Opcode) const {
647     return get(Opcode).TSFlags & SIInstrFlags::VOP3P;
648   }
649 
650   static bool isVINTRP(const MachineInstr &MI) {
651     return MI.getDesc().TSFlags & SIInstrFlags::VINTRP;
652   }
653 
654   bool isVINTRP(uint16_t Opcode) const {
655     return get(Opcode).TSFlags & SIInstrFlags::VINTRP;
656   }
657 
658   static bool isMAI(const MachineInstr &MI) {
659     return MI.getDesc().TSFlags & SIInstrFlags::IsMAI;
660   }
661 
662   bool isMAI(uint16_t Opcode) const {
663     return get(Opcode).TSFlags & SIInstrFlags::IsMAI;
664   }
665 
666   static bool isMFMA(const MachineInstr &MI) {
667     return isMAI(MI) && MI.getOpcode() != AMDGPU::V_ACCVGPR_WRITE_B32_e64 &&
668            MI.getOpcode() != AMDGPU::V_ACCVGPR_READ_B32_e64;
669   }
670 
671   static bool isDOT(const MachineInstr &MI) {
672     return MI.getDesc().TSFlags & SIInstrFlags::IsDOT;
673   }
674 
675   static bool isWMMA(const MachineInstr &MI) {
676     return MI.getDesc().TSFlags & SIInstrFlags::IsWMMA;
677   }
678 
679   bool isWMMA(uint16_t Opcode) const {
680     return get(Opcode).TSFlags & SIInstrFlags::IsWMMA;
681   }
682 
683   bool isDOT(uint16_t Opcode) const {
684     return get(Opcode).TSFlags & SIInstrFlags::IsDOT;
685   }
686 
687   static bool isLDSDIR(const MachineInstr &MI) {
688     return MI.getDesc().TSFlags & SIInstrFlags::LDSDIR;
689   }
690 
691   bool isLDSDIR(uint16_t Opcode) const {
692     return get(Opcode).TSFlags & SIInstrFlags::LDSDIR;
693   }
694 
695   static bool isVINTERP(const MachineInstr &MI) {
696     return MI.getDesc().TSFlags & SIInstrFlags::VINTERP;
697   }
698 
699   bool isVINTERP(uint16_t Opcode) const {
700     return get(Opcode).TSFlags & SIInstrFlags::VINTERP;
701   }
702 
703   static bool isScalarUnit(const MachineInstr &MI) {
704     return MI.getDesc().TSFlags & (SIInstrFlags::SALU | SIInstrFlags::SMRD);
705   }
706 
707   static bool usesVM_CNT(const MachineInstr &MI) {
708     return MI.getDesc().TSFlags & SIInstrFlags::VM_CNT;
709   }
710 
711   static bool usesLGKM_CNT(const MachineInstr &MI) {
712     return MI.getDesc().TSFlags & SIInstrFlags::LGKM_CNT;
713   }
714 
715   static bool sopkIsZext(const MachineInstr &MI) {
716     return MI.getDesc().TSFlags & SIInstrFlags::SOPK_ZEXT;
717   }
718 
719   bool sopkIsZext(uint16_t Opcode) const {
720     return get(Opcode).TSFlags & SIInstrFlags::SOPK_ZEXT;
721   }
722 
723   /// \returns true if this is an s_store_dword* instruction. This is more
724   /// specific than than isSMEM && mayStore.
725   static bool isScalarStore(const MachineInstr &MI) {
726     return MI.getDesc().TSFlags & SIInstrFlags::SCALAR_STORE;
727   }
728 
729   bool isScalarStore(uint16_t Opcode) const {
730     return get(Opcode).TSFlags & SIInstrFlags::SCALAR_STORE;
731   }
732 
733   static bool isFixedSize(const MachineInstr &MI) {
734     return MI.getDesc().TSFlags & SIInstrFlags::FIXED_SIZE;
735   }
736 
737   bool isFixedSize(uint16_t Opcode) const {
738     return get(Opcode).TSFlags & SIInstrFlags::FIXED_SIZE;
739   }
740 
741   static bool hasFPClamp(const MachineInstr &MI) {
742     return MI.getDesc().TSFlags & SIInstrFlags::FPClamp;
743   }
744 
745   bool hasFPClamp(uint16_t Opcode) const {
746     return get(Opcode).TSFlags & SIInstrFlags::FPClamp;
747   }
748 
749   static bool hasIntClamp(const MachineInstr &MI) {
750     return MI.getDesc().TSFlags & SIInstrFlags::IntClamp;
751   }
752 
753   uint64_t getClampMask(const MachineInstr &MI) const {
754     const uint64_t ClampFlags = SIInstrFlags::FPClamp |
755                                 SIInstrFlags::IntClamp |
756                                 SIInstrFlags::ClampLo |
757                                 SIInstrFlags::ClampHi;
758       return MI.getDesc().TSFlags & ClampFlags;
759   }
760 
761   static bool usesFPDPRounding(const MachineInstr &MI) {
762     return MI.getDesc().TSFlags & SIInstrFlags::FPDPRounding;
763   }
764 
765   bool usesFPDPRounding(uint16_t Opcode) const {
766     return get(Opcode).TSFlags & SIInstrFlags::FPDPRounding;
767   }
768 
769   static bool isFPAtomic(const MachineInstr &MI) {
770     return MI.getDesc().TSFlags & SIInstrFlags::FPAtomic;
771   }
772 
773   bool isFPAtomic(uint16_t Opcode) const {
774     return get(Opcode).TSFlags & SIInstrFlags::FPAtomic;
775   }
776 
777   bool isVGPRCopy(const MachineInstr &MI) const {
778     assert(MI.isCopy());
779     Register Dest = MI.getOperand(0).getReg();
780     const MachineFunction &MF = *MI.getParent()->getParent();
781     const MachineRegisterInfo &MRI = MF.getRegInfo();
782     return !RI.isSGPRReg(MRI, Dest);
783   }
784 
785   bool hasVGPRUses(const MachineInstr &MI) const {
786     const MachineFunction &MF = *MI.getParent()->getParent();
787     const MachineRegisterInfo &MRI = MF.getRegInfo();
788     return llvm::any_of(MI.explicit_uses(),
789                         [&MRI, this](const MachineOperand &MO) {
790       return MO.isReg() && RI.isVGPR(MRI, MO.getReg());});
791   }
792 
793   /// Return true if the instruction modifies the mode register.q
794   static bool modifiesModeRegister(const MachineInstr &MI);
795 
796   /// Whether we must prevent this instruction from executing with EXEC = 0.
797   bool hasUnwantedEffectsWhenEXECEmpty(const MachineInstr &MI) const;
798 
799   /// Returns true if the instruction could potentially depend on the value of
800   /// exec. If false, exec dependencies may safely be ignored.
801   bool mayReadEXEC(const MachineRegisterInfo &MRI, const MachineInstr &MI) const;
802 
803   bool isInlineConstant(const APInt &Imm) const;
804 
805   bool isInlineConstant(const APFloat &Imm) const {
806     return isInlineConstant(Imm.bitcastToAPInt());
807   }
808 
809   bool isInlineConstant(const MachineOperand &MO, uint8_t OperandType) const;
810 
811   bool isInlineConstant(const MachineOperand &MO,
812                         const MCOperandInfo &OpInfo) const {
813     return isInlineConstant(MO, OpInfo.OperandType);
814   }
815 
816   /// \p returns true if \p UseMO is substituted with \p DefMO in \p MI it would
817   /// be an inline immediate.
818   bool isInlineConstant(const MachineInstr &MI,
819                         const MachineOperand &UseMO,
820                         const MachineOperand &DefMO) const {
821     assert(UseMO.getParent() == &MI);
822     int OpIdx = MI.getOperandNo(&UseMO);
823     if (!MI.getDesc().OpInfo || OpIdx >= MI.getDesc().NumOperands) {
824       return false;
825     }
826 
827     return isInlineConstant(DefMO, MI.getDesc().OpInfo[OpIdx]);
828   }
829 
830   /// \p returns true if the operand \p OpIdx in \p MI is a valid inline
831   /// immediate.
832   bool isInlineConstant(const MachineInstr &MI, unsigned OpIdx) const {
833     const MachineOperand &MO = MI.getOperand(OpIdx);
834     return isInlineConstant(MO, MI.getDesc().OpInfo[OpIdx].OperandType);
835   }
836 
837   bool isInlineConstant(const MachineInstr &MI, unsigned OpIdx,
838                         const MachineOperand &MO) const {
839     if (!MI.getDesc().OpInfo || OpIdx >= MI.getDesc().NumOperands)
840       return false;
841 
842     if (MI.isCopy()) {
843       unsigned Size = getOpSize(MI, OpIdx);
844       assert(Size == 8 || Size == 4);
845 
846       uint8_t OpType = (Size == 8) ?
847         AMDGPU::OPERAND_REG_IMM_INT64 : AMDGPU::OPERAND_REG_IMM_INT32;
848       return isInlineConstant(MO, OpType);
849     }
850 
851     return isInlineConstant(MO, MI.getDesc().OpInfo[OpIdx].OperandType);
852   }
853 
854   bool isInlineConstant(const MachineOperand &MO) const {
855     const MachineInstr *Parent = MO.getParent();
856     return isInlineConstant(*Parent, Parent->getOperandNo(&MO));
857   }
858 
859   bool isLiteralConstant(const MachineOperand &MO,
860                          const MCOperandInfo &OpInfo) const {
861     return MO.isImm() && !isInlineConstant(MO, OpInfo.OperandType);
862   }
863 
864   bool isLiteralConstant(const MachineInstr &MI, int OpIdx) const {
865     const MachineOperand &MO = MI.getOperand(OpIdx);
866     return MO.isImm() && !isInlineConstant(MI, OpIdx);
867   }
868 
869   // Returns true if this operand could potentially require a 32-bit literal
870   // operand, but not necessarily. A FrameIndex for example could resolve to an
871   // inline immediate value that will not require an additional 4-bytes; this
872   // assumes that it will.
873   bool isLiteralConstantLike(const MachineOperand &MO,
874                              const MCOperandInfo &OpInfo) const;
875 
876   bool isImmOperandLegal(const MachineInstr &MI, unsigned OpNo,
877                          const MachineOperand &MO) const;
878 
879   /// Return true if this 64-bit VALU instruction has a 32-bit encoding.
880   /// This function will return false if you pass it a 32-bit instruction.
881   bool hasVALU32BitEncoding(unsigned Opcode) const;
882 
883   /// Returns true if this operand uses the constant bus.
884   bool usesConstantBus(const MachineRegisterInfo &MRI,
885                        const MachineOperand &MO,
886                        const MCOperandInfo &OpInfo) const;
887 
888   /// Return true if this instruction has any modifiers.
889   ///  e.g. src[012]_mod, omod, clamp.
890   bool hasModifiers(unsigned Opcode) const;
891 
892   bool hasModifiersSet(const MachineInstr &MI,
893                        unsigned OpName) const;
894   bool hasAnyModifiersSet(const MachineInstr &MI) const;
895 
896   bool canShrink(const MachineInstr &MI,
897                  const MachineRegisterInfo &MRI) const;
898 
899   MachineInstr *buildShrunkInst(MachineInstr &MI,
900                                 unsigned NewOpcode) const;
901 
902   bool verifyInstruction(const MachineInstr &MI,
903                          StringRef &ErrInfo) const override;
904 
905   unsigned getVALUOp(const MachineInstr &MI) const;
906 
907   /// Return the correct register class for \p OpNo.  For target-specific
908   /// instructions, this will return the register class that has been defined
909   /// in tablegen.  For generic instructions, like REG_SEQUENCE it will return
910   /// the register class of its machine operand.
911   /// to infer the correct register class base on the other operands.
912   const TargetRegisterClass *getOpRegClass(const MachineInstr &MI,
913                                            unsigned OpNo) const;
914 
915   /// Return the size in bytes of the operand OpNo on the given
916   // instruction opcode.
917   unsigned getOpSize(uint16_t Opcode, unsigned OpNo) const {
918     const MCOperandInfo &OpInfo = get(Opcode).OpInfo[OpNo];
919 
920     if (OpInfo.RegClass == -1) {
921       // If this is an immediate operand, this must be a 32-bit literal.
922       assert(OpInfo.OperandType == MCOI::OPERAND_IMMEDIATE);
923       return 4;
924     }
925 
926     return RI.getRegSizeInBits(*RI.getRegClass(OpInfo.RegClass)) / 8;
927   }
928 
929   /// This form should usually be preferred since it handles operands
930   /// with unknown register classes.
931   unsigned getOpSize(const MachineInstr &MI, unsigned OpNo) const {
932     const MachineOperand &MO = MI.getOperand(OpNo);
933     if (MO.isReg()) {
934       if (unsigned SubReg = MO.getSubReg()) {
935         return RI.getSubRegIdxSize(SubReg) / 8;
936       }
937     }
938     return RI.getRegSizeInBits(*getOpRegClass(MI, OpNo)) / 8;
939   }
940 
941   /// Legalize the \p OpIndex operand of this instruction by inserting
942   /// a MOV.  For example:
943   /// ADD_I32_e32 VGPR0, 15
944   /// to
945   /// MOV VGPR1, 15
946   /// ADD_I32_e32 VGPR0, VGPR1
947   ///
948   /// If the operand being legalized is a register, then a COPY will be used
949   /// instead of MOV.
950   void legalizeOpWithMove(MachineInstr &MI, unsigned OpIdx) const;
951 
952   /// Check if \p MO is a legal operand if it was the \p OpIdx Operand
953   /// for \p MI.
954   bool isOperandLegal(const MachineInstr &MI, unsigned OpIdx,
955                       const MachineOperand *MO = nullptr) const;
956 
957   /// Check if \p MO would be a valid operand for the given operand
958   /// definition \p OpInfo. Note this does not attempt to validate constant bus
959   /// restrictions (e.g. literal constant usage).
960   bool isLegalVSrcOperand(const MachineRegisterInfo &MRI,
961                           const MCOperandInfo &OpInfo,
962                           const MachineOperand &MO) const;
963 
964   /// Check if \p MO (a register operand) is a legal register for the
965   /// given operand description.
966   bool isLegalRegOperand(const MachineRegisterInfo &MRI,
967                          const MCOperandInfo &OpInfo,
968                          const MachineOperand &MO) const;
969 
970   /// Legalize operands in \p MI by either commuting it or inserting a
971   /// copy of src1.
972   void legalizeOperandsVOP2(MachineRegisterInfo &MRI, MachineInstr &MI) const;
973 
974   /// Fix operands in \p MI to satisfy constant bus requirements.
975   void legalizeOperandsVOP3(MachineRegisterInfo &MRI, MachineInstr &MI) const;
976 
977   /// Copy a value from a VGPR (\p SrcReg) to SGPR.  This function can only
978   /// be used when it is know that the value in SrcReg is same across all
979   /// threads in the wave.
980   /// \returns The SGPR register that \p SrcReg was copied to.
981   Register readlaneVGPRToSGPR(Register SrcReg, MachineInstr &UseMI,
982                               MachineRegisterInfo &MRI) const;
983 
984   void legalizeOperandsSMRD(MachineRegisterInfo &MRI, MachineInstr &MI) const;
985   void legalizeOperandsFLAT(MachineRegisterInfo &MRI, MachineInstr &MI) const;
986 
987   void legalizeGenericOperand(MachineBasicBlock &InsertMBB,
988                               MachineBasicBlock::iterator I,
989                               const TargetRegisterClass *DstRC,
990                               MachineOperand &Op, MachineRegisterInfo &MRI,
991                               const DebugLoc &DL) const;
992 
993   /// Legalize all operands in this instruction.  This function may create new
994   /// instructions and control-flow around \p MI.  If present, \p MDT is
995   /// updated.
996   /// \returns A new basic block that contains \p MI if new blocks were created.
997   MachineBasicBlock *
998   legalizeOperands(MachineInstr &MI, MachineDominatorTree *MDT = nullptr) const;
999 
1000   /// Change SADDR form of a FLAT \p Inst to its VADDR form if saddr operand
1001   /// was moved to VGPR. \returns true if succeeded.
1002   bool moveFlatAddrToVGPR(MachineInstr &Inst) const;
1003 
1004   /// Replace this instruction's opcode with the equivalent VALU
1005   /// opcode.  This function will also move the users of \p MI to the
1006   /// VALU if necessary. If present, \p MDT is updated.
1007   MachineBasicBlock *moveToVALU(MachineInstr &MI,
1008                                 MachineDominatorTree *MDT = nullptr) const;
1009 
1010   void insertNoop(MachineBasicBlock &MBB,
1011                   MachineBasicBlock::iterator MI) const override;
1012 
1013   void insertNoops(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI,
1014                    unsigned Quantity) const override;
1015 
1016   void insertReturn(MachineBasicBlock &MBB) const;
1017   /// Return the number of wait states that result from executing this
1018   /// instruction.
1019   static unsigned getNumWaitStates(const MachineInstr &MI);
1020 
1021   /// Returns the operand named \p Op.  If \p MI does not have an
1022   /// operand named \c Op, this function returns nullptr.
1023   LLVM_READONLY
1024   MachineOperand *getNamedOperand(MachineInstr &MI, unsigned OperandName) const;
1025 
1026   LLVM_READONLY
1027   const MachineOperand *getNamedOperand(const MachineInstr &MI,
1028                                         unsigned OpName) const {
1029     return getNamedOperand(const_cast<MachineInstr &>(MI), OpName);
1030   }
1031 
1032   /// Get required immediate operand
1033   int64_t getNamedImmOperand(const MachineInstr &MI, unsigned OpName) const {
1034     int Idx = AMDGPU::getNamedOperandIdx(MI.getOpcode(), OpName);
1035     return MI.getOperand(Idx).getImm();
1036   }
1037 
1038   uint64_t getDefaultRsrcDataFormat() const;
1039   uint64_t getScratchRsrcWords23() const;
1040 
1041   bool isLowLatencyInstruction(const MachineInstr &MI) const;
1042   bool isHighLatencyDef(int Opc) const override;
1043 
1044   /// Return the descriptor of the target-specific machine instruction
1045   /// that corresponds to the specified pseudo or native opcode.
1046   const MCInstrDesc &getMCOpcodeFromPseudo(unsigned Opcode) const {
1047     return get(pseudoToMCOpcode(Opcode));
1048   }
1049 
1050   unsigned isStackAccess(const MachineInstr &MI, int &FrameIndex) const;
1051   unsigned isSGPRStackAccess(const MachineInstr &MI, int &FrameIndex) const;
1052 
1053   unsigned isLoadFromStackSlot(const MachineInstr &MI,
1054                                int &FrameIndex) const override;
1055   unsigned isStoreToStackSlot(const MachineInstr &MI,
1056                               int &FrameIndex) const override;
1057 
1058   unsigned getInstBundleSize(const MachineInstr &MI) const;
1059   unsigned getInstSizeInBytes(const MachineInstr &MI) const override;
1060 
1061   bool mayAccessFlatAddressSpace(const MachineInstr &MI) const;
1062 
1063   bool isNonUniformBranchInstr(MachineInstr &Instr) const;
1064 
1065   void convertNonUniformIfRegion(MachineBasicBlock *IfEntry,
1066                                  MachineBasicBlock *IfEnd) const;
1067 
1068   void convertNonUniformLoopRegion(MachineBasicBlock *LoopEntry,
1069                                    MachineBasicBlock *LoopEnd) const;
1070 
1071   std::pair<unsigned, unsigned>
1072   decomposeMachineOperandsTargetFlags(unsigned TF) const override;
1073 
1074   ArrayRef<std::pair<int, const char *>>
1075   getSerializableTargetIndices() const override;
1076 
1077   ArrayRef<std::pair<unsigned, const char *>>
1078   getSerializableDirectMachineOperandTargetFlags() const override;
1079 
1080   ArrayRef<std::pair<MachineMemOperand::Flags, const char *>>
1081   getSerializableMachineMemOperandTargetFlags() const override;
1082 
1083   ScheduleHazardRecognizer *
1084   CreateTargetPostRAHazardRecognizer(const InstrItineraryData *II,
1085                                  const ScheduleDAG *DAG) const override;
1086 
1087   ScheduleHazardRecognizer *
1088   CreateTargetPostRAHazardRecognizer(const MachineFunction &MF) const override;
1089 
1090   ScheduleHazardRecognizer *
1091   CreateTargetMIHazardRecognizer(const InstrItineraryData *II,
1092                                  const ScheduleDAGMI *DAG) const override;
1093 
1094   bool isBasicBlockPrologue(const MachineInstr &MI) const override;
1095 
1096   MachineInstr *createPHIDestinationCopy(MachineBasicBlock &MBB,
1097                                          MachineBasicBlock::iterator InsPt,
1098                                          const DebugLoc &DL, Register Src,
1099                                          Register Dst) const override;
1100 
1101   MachineInstr *createPHISourceCopy(MachineBasicBlock &MBB,
1102                                     MachineBasicBlock::iterator InsPt,
1103                                     const DebugLoc &DL, Register Src,
1104                                     unsigned SrcSubReg,
1105                                     Register Dst) const override;
1106 
1107   bool isWave32() const;
1108 
1109   /// Return a partially built integer add instruction without carry.
1110   /// Caller must add source operands.
1111   /// For pre-GFX9 it will generate unused carry destination operand.
1112   /// TODO: After GFX9 it should return a no-carry operation.
1113   MachineInstrBuilder getAddNoCarry(MachineBasicBlock &MBB,
1114                                     MachineBasicBlock::iterator I,
1115                                     const DebugLoc &DL,
1116                                     Register DestReg) const;
1117 
1118   MachineInstrBuilder getAddNoCarry(MachineBasicBlock &MBB,
1119                                     MachineBasicBlock::iterator I,
1120                                     const DebugLoc &DL,
1121                                     Register DestReg,
1122                                     RegScavenger &RS) const;
1123 
1124   static bool isKillTerminator(unsigned Opcode);
1125   const MCInstrDesc &getKillTerminatorFromPseudo(unsigned Opcode) const;
1126 
1127   static bool isLegalMUBUFImmOffset(unsigned Imm) {
1128     return isUInt<12>(Imm);
1129   }
1130 
1131   /// Returns if \p Offset is legal for the subtarget as the offset to a FLAT
1132   /// encoded instruction. If \p Signed, this is for an instruction that
1133   /// interprets the offset as signed.
1134   bool isLegalFLATOffset(int64_t Offset, unsigned AddrSpace,
1135                          uint64_t FlatVariant) const;
1136 
1137   /// Split \p COffsetVal into {immediate offset field, remainder offset}
1138   /// values.
1139   std::pair<int64_t, int64_t> splitFlatOffset(int64_t COffsetVal,
1140                                               unsigned AddrSpace,
1141                                               uint64_t FlatVariant) const;
1142 
1143   /// \brief Return a target-specific opcode if Opcode is a pseudo instruction.
1144   /// Return -1 if the target-specific opcode for the pseudo instruction does
1145   /// not exist. If Opcode is not a pseudo instruction, this is identity.
1146   int pseudoToMCOpcode(int Opcode) const;
1147 
1148   /// \brief Check if this instruction should only be used by assembler.
1149   /// Return true if this opcode should not be used by codegen.
1150   bool isAsmOnlyOpcode(int MCOp) const;
1151 
1152   const TargetRegisterClass *getRegClass(const MCInstrDesc &TID, unsigned OpNum,
1153                                          const TargetRegisterInfo *TRI,
1154                                          const MachineFunction &MF)
1155     const override;
1156 
1157   void fixImplicitOperands(MachineInstr &MI) const;
1158 
1159   MachineInstr *foldMemoryOperandImpl(MachineFunction &MF, MachineInstr &MI,
1160                                       ArrayRef<unsigned> Ops,
1161                                       MachineBasicBlock::iterator InsertPt,
1162                                       int FrameIndex,
1163                                       LiveIntervals *LIS = nullptr,
1164                                       VirtRegMap *VRM = nullptr) const override;
1165 
1166   unsigned getInstrLatency(const InstrItineraryData *ItinData,
1167                            const MachineInstr &MI,
1168                            unsigned *PredCost = nullptr) const override;
1169 
1170   const MIRFormatter *getMIRFormatter() const override {
1171     if (!Formatter.get())
1172       Formatter = std::make_unique<AMDGPUMIRFormatter>();
1173     return Formatter.get();
1174   }
1175 
1176   static unsigned getDSShaderTypeValue(const MachineFunction &MF);
1177 
1178   const TargetSchedModel &getSchedModel() const { return SchedModel; }
1179 
1180   // Enforce operand's \p OpName even alignment if required by target.
1181   // This is used if an operand is a 32 bit register but needs to be aligned
1182   // regardless.
1183   void enforceOperandRCAlignment(MachineInstr &MI, unsigned OpName) const;
1184 };
1185 
1186 /// \brief Returns true if a reg:subreg pair P has a TRC class
1187 inline bool isOfRegClass(const TargetInstrInfo::RegSubRegPair &P,
1188                          const TargetRegisterClass &TRC,
1189                          MachineRegisterInfo &MRI) {
1190   auto *RC = MRI.getRegClass(P.Reg);
1191   if (!P.SubReg)
1192     return RC == &TRC;
1193   auto *TRI = MRI.getTargetRegisterInfo();
1194   return RC == TRI->getMatchingSuperRegClass(RC, &TRC, P.SubReg);
1195 }
1196 
1197 /// \brief Create RegSubRegPair from a register MachineOperand
1198 inline
1199 TargetInstrInfo::RegSubRegPair getRegSubRegPair(const MachineOperand &O) {
1200   assert(O.isReg());
1201   return TargetInstrInfo::RegSubRegPair(O.getReg(), O.getSubReg());
1202 }
1203 
1204 /// \brief Return the SubReg component from REG_SEQUENCE
1205 TargetInstrInfo::RegSubRegPair getRegSequenceSubReg(MachineInstr &MI,
1206                                                     unsigned SubReg);
1207 
1208 /// \brief Return the defining instruction for a given reg:subreg pair
1209 /// skipping copy like instructions and subreg-manipulation pseudos.
1210 /// Following another subreg of a reg:subreg isn't supported.
1211 MachineInstr *getVRegSubRegDef(const TargetInstrInfo::RegSubRegPair &P,
1212                                MachineRegisterInfo &MRI);
1213 
1214 /// \brief Return false if EXEC is not changed between the def of \p VReg at \p
1215 /// DefMI and the use at \p UseMI. Should be run on SSA. Currently does not
1216 /// attempt to track between blocks.
1217 bool execMayBeModifiedBeforeUse(const MachineRegisterInfo &MRI,
1218                                 Register VReg,
1219                                 const MachineInstr &DefMI,
1220                                 const MachineInstr &UseMI);
1221 
1222 /// \brief Return false if EXEC is not changed between the def of \p VReg at \p
1223 /// DefMI and all its uses. Should be run on SSA. Currently does not attempt to
1224 /// track between blocks.
1225 bool execMayBeModifiedBeforeAnyUse(const MachineRegisterInfo &MRI,
1226                                    Register VReg,
1227                                    const MachineInstr &DefMI);
1228 
1229 namespace AMDGPU {
1230 
1231   LLVM_READONLY
1232   int getVOPe64(uint16_t Opcode);
1233 
1234   LLVM_READONLY
1235   int getVOPe32(uint16_t Opcode);
1236 
1237   LLVM_READONLY
1238   int getSDWAOp(uint16_t Opcode);
1239 
1240   LLVM_READONLY
1241   int getDPPOp32(uint16_t Opcode);
1242 
1243   LLVM_READONLY
1244   int getDPPOp64(uint16_t Opcode);
1245 
1246   LLVM_READONLY
1247   int getBasicFromSDWAOp(uint16_t Opcode);
1248 
1249   LLVM_READONLY
1250   int getCommuteRev(uint16_t Opcode);
1251 
1252   LLVM_READONLY
1253   int getCommuteOrig(uint16_t Opcode);
1254 
1255   LLVM_READONLY
1256   int getAddr64Inst(uint16_t Opcode);
1257 
1258   /// Check if \p Opcode is an Addr64 opcode.
1259   ///
1260   /// \returns \p Opcode if it is an Addr64 opcode, otherwise -1.
1261   LLVM_READONLY
1262   int getIfAddr64Inst(uint16_t Opcode);
1263 
1264   LLVM_READONLY
1265   int getAtomicNoRetOp(uint16_t Opcode);
1266 
1267   LLVM_READONLY
1268   int getSOPKOp(uint16_t Opcode);
1269 
1270   /// \returns SADDR form of a FLAT Global instruction given an \p Opcode
1271   /// of a VADDR form.
1272   LLVM_READONLY
1273   int getGlobalSaddrOp(uint16_t Opcode);
1274 
1275   /// \returns VADDR form of a FLAT Global instruction given an \p Opcode
1276   /// of a SADDR form.
1277   LLVM_READONLY
1278   int getGlobalVaddrOp(uint16_t Opcode);
1279 
1280   LLVM_READONLY
1281   int getVCMPXNoSDstOp(uint16_t Opcode);
1282 
1283   /// \returns ST form with only immediate offset of a FLAT Scratch instruction
1284   /// given an \p Opcode of an SS (SADDR) form.
1285   LLVM_READONLY
1286   int getFlatScratchInstSTfromSS(uint16_t Opcode);
1287 
1288   /// \returns SV (VADDR) form of a FLAT Scratch instruction given an \p Opcode
1289   /// of an SVS (SADDR + VADDR) form.
1290   LLVM_READONLY
1291   int getFlatScratchInstSVfromSVS(uint16_t Opcode);
1292 
1293   /// \returns SS (SADDR) form of a FLAT Scratch instruction given an \p Opcode
1294   /// of an SV (VADDR) form.
1295   LLVM_READONLY
1296   int getFlatScratchInstSSfromSV(uint16_t Opcode);
1297 
1298   /// \returns SV (VADDR) form of a FLAT Scratch instruction given an \p Opcode
1299   /// of an SS (SADDR) form.
1300   LLVM_READONLY
1301   int getFlatScratchInstSVfromSS(uint16_t Opcode);
1302 
1303   /// \returns earlyclobber version of a MAC MFMA is exists.
1304   LLVM_READONLY
1305   int getMFMAEarlyClobberOp(uint16_t Opcode);
1306 
1307   /// \returns v_cmpx version of a v_cmp instruction.
1308   LLVM_READONLY
1309   int getVCMPXOpFromVCMP(uint16_t Opcode);
1310 
1311   const uint64_t RSRC_DATA_FORMAT = 0xf00000000000LL;
1312   const uint64_t RSRC_ELEMENT_SIZE_SHIFT = (32 + 19);
1313   const uint64_t RSRC_INDEX_STRIDE_SHIFT = (32 + 21);
1314   const uint64_t RSRC_TID_ENABLE = UINT64_C(1) << (32 + 23);
1315 
1316 } // end namespace AMDGPU
1317 
1318 namespace SI {
1319 namespace KernelInputOffsets {
1320 
1321 /// Offsets in bytes from the start of the input buffer
1322 enum Offsets {
1323   NGROUPS_X = 0,
1324   NGROUPS_Y = 4,
1325   NGROUPS_Z = 8,
1326   GLOBAL_SIZE_X = 12,
1327   GLOBAL_SIZE_Y = 16,
1328   GLOBAL_SIZE_Z = 20,
1329   LOCAL_SIZE_X = 24,
1330   LOCAL_SIZE_Y = 28,
1331   LOCAL_SIZE_Z = 32
1332 };
1333 
1334 } // end namespace KernelInputOffsets
1335 } // end namespace SI
1336 
1337 } // end namespace llvm
1338 
1339 #endif // LLVM_LIB_TARGET_AMDGPU_SIINSTRINFO_H
1340