1 //===- SIInstrInfo.h - SI Instruction Info Interface ------------*- C++ -*-===// 2 // 3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. 4 // See https://llvm.org/LICENSE.txt for license information. 5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception 6 // 7 //===----------------------------------------------------------------------===// 8 // 9 /// \file 10 /// Interface definition for SIInstrInfo. 11 // 12 //===----------------------------------------------------------------------===// 13 14 #ifndef LLVM_LIB_TARGET_AMDGPU_SIINSTRINFO_H 15 #define LLVM_LIB_TARGET_AMDGPU_SIINSTRINFO_H 16 17 #include "AMDGPUMIRFormatter.h" 18 #include "SIRegisterInfo.h" 19 #include "Utils/AMDGPUBaseInfo.h" 20 #include "llvm/ADT/SetVector.h" 21 #include "llvm/CodeGen/TargetInstrInfo.h" 22 #include "llvm/CodeGen/TargetSchedule.h" 23 24 #define GET_INSTRINFO_HEADER 25 #include "AMDGPUGenInstrInfo.inc" 26 27 namespace llvm { 28 29 class APInt; 30 class GCNSubtarget; 31 class LiveVariables; 32 class MachineDominatorTree; 33 class MachineRegisterInfo; 34 class RegScavenger; 35 class TargetRegisterClass; 36 class ScheduleHazardRecognizer; 37 38 class SIInstrInfo final : public AMDGPUGenInstrInfo { 39 private: 40 const SIRegisterInfo RI; 41 const GCNSubtarget &ST; 42 TargetSchedModel SchedModel; 43 mutable std::unique_ptr<AMDGPUMIRFormatter> Formatter; 44 45 // The inverse predicate should have the negative value. 46 enum BranchPredicate { 47 INVALID_BR = 0, 48 SCC_TRUE = 1, 49 SCC_FALSE = -1, 50 VCCNZ = 2, 51 VCCZ = -2, 52 EXECNZ = -3, 53 EXECZ = 3 54 }; 55 56 using SetVectorType = SmallSetVector<MachineInstr *, 32>; 57 58 static unsigned getBranchOpcode(BranchPredicate Cond); 59 static BranchPredicate getBranchPredicate(unsigned Opcode); 60 61 public: 62 unsigned buildExtractSubReg(MachineBasicBlock::iterator MI, 63 MachineRegisterInfo &MRI, 64 MachineOperand &SuperReg, 65 const TargetRegisterClass *SuperRC, 66 unsigned SubIdx, 67 const TargetRegisterClass *SubRC) const; 68 MachineOperand buildExtractSubRegOrImm(MachineBasicBlock::iterator MI, 69 MachineRegisterInfo &MRI, 70 MachineOperand &SuperReg, 71 const TargetRegisterClass *SuperRC, 72 unsigned SubIdx, 73 const TargetRegisterClass *SubRC) const; 74 private: 75 void swapOperands(MachineInstr &Inst) const; 76 77 std::pair<bool, MachineBasicBlock *> 78 moveScalarAddSub(SetVectorType &Worklist, MachineInstr &Inst, 79 MachineDominatorTree *MDT = nullptr) const; 80 81 void lowerSelect(SetVectorType &Worklist, MachineInstr &Inst, 82 MachineDominatorTree *MDT = nullptr) const; 83 84 void lowerScalarAbs(SetVectorType &Worklist, 85 MachineInstr &Inst) const; 86 87 void lowerScalarXnor(SetVectorType &Worklist, 88 MachineInstr &Inst) const; 89 90 void splitScalarNotBinop(SetVectorType &Worklist, 91 MachineInstr &Inst, 92 unsigned Opcode) const; 93 94 void splitScalarBinOpN2(SetVectorType &Worklist, 95 MachineInstr &Inst, 96 unsigned Opcode) const; 97 98 void splitScalar64BitUnaryOp(SetVectorType &Worklist, 99 MachineInstr &Inst, unsigned Opcode, 100 bool Swap = false) const; 101 102 void splitScalar64BitAddSub(SetVectorType &Worklist, MachineInstr &Inst, 103 MachineDominatorTree *MDT = nullptr) const; 104 105 void splitScalar64BitBinaryOp(SetVectorType &Worklist, MachineInstr &Inst, 106 unsigned Opcode, 107 MachineDominatorTree *MDT = nullptr) const; 108 109 void splitScalar64BitXnor(SetVectorType &Worklist, MachineInstr &Inst, 110 MachineDominatorTree *MDT = nullptr) const; 111 112 void splitScalar64BitBCNT(SetVectorType &Worklist, 113 MachineInstr &Inst) const; 114 void splitScalar64BitBFE(SetVectorType &Worklist, 115 MachineInstr &Inst) const; 116 void movePackToVALU(SetVectorType &Worklist, 117 MachineRegisterInfo &MRI, 118 MachineInstr &Inst) const; 119 120 void addUsersToMoveToVALUWorklist(Register Reg, MachineRegisterInfo &MRI, 121 SetVectorType &Worklist) const; 122 123 void addSCCDefUsersToVALUWorklist(MachineOperand &Op, 124 MachineInstr &SCCDefInst, 125 SetVectorType &Worklist) const; 126 void addSCCDefsToVALUWorklist(MachineOperand &Op, 127 SetVectorType &Worklist) const; 128 129 const TargetRegisterClass * 130 getDestEquivalentVGPRClass(const MachineInstr &Inst) const; 131 132 bool checkInstOffsetsDoNotOverlap(const MachineInstr &MIa, 133 const MachineInstr &MIb) const; 134 135 Register findUsedSGPR(const MachineInstr &MI, int OpIndices[3]) const; 136 137 protected: 138 bool swapSourceModifiers(MachineInstr &MI, 139 MachineOperand &Src0, unsigned Src0OpName, 140 MachineOperand &Src1, unsigned Src1OpName) const; 141 142 MachineInstr *commuteInstructionImpl(MachineInstr &MI, bool NewMI, 143 unsigned OpIdx0, 144 unsigned OpIdx1) const override; 145 146 public: 147 enum TargetOperandFlags { 148 MO_MASK = 0xf, 149 150 MO_NONE = 0, 151 // MO_GOTPCREL -> symbol@GOTPCREL -> R_AMDGPU_GOTPCREL. 152 MO_GOTPCREL = 1, 153 // MO_GOTPCREL32_LO -> symbol@gotpcrel32@lo -> R_AMDGPU_GOTPCREL32_LO. 154 MO_GOTPCREL32 = 2, 155 MO_GOTPCREL32_LO = 2, 156 // MO_GOTPCREL32_HI -> symbol@gotpcrel32@hi -> R_AMDGPU_GOTPCREL32_HI. 157 MO_GOTPCREL32_HI = 3, 158 // MO_REL32_LO -> symbol@rel32@lo -> R_AMDGPU_REL32_LO. 159 MO_REL32 = 4, 160 MO_REL32_LO = 4, 161 // MO_REL32_HI -> symbol@rel32@hi -> R_AMDGPU_REL32_HI. 162 MO_REL32_HI = 5, 163 164 MO_FAR_BRANCH_OFFSET = 6, 165 166 MO_ABS32_LO = 8, 167 MO_ABS32_HI = 9, 168 }; 169 170 explicit SIInstrInfo(const GCNSubtarget &ST); 171 172 const SIRegisterInfo &getRegisterInfo() const { 173 return RI; 174 } 175 176 const GCNSubtarget &getSubtarget() const { 177 return ST; 178 } 179 180 bool isReallyTriviallyReMaterializable(const MachineInstr &MI, 181 AAResults *AA) const override; 182 183 bool isIgnorableUse(const MachineOperand &MO) const override; 184 185 bool areLoadsFromSameBasePtr(SDNode *Load1, SDNode *Load2, 186 int64_t &Offset1, 187 int64_t &Offset2) const override; 188 189 bool getMemOperandsWithOffsetWidth( 190 const MachineInstr &LdSt, 191 SmallVectorImpl<const MachineOperand *> &BaseOps, int64_t &Offset, 192 bool &OffsetIsScalable, unsigned &Width, 193 const TargetRegisterInfo *TRI) const final; 194 195 bool shouldClusterMemOps(ArrayRef<const MachineOperand *> BaseOps1, 196 ArrayRef<const MachineOperand *> BaseOps2, 197 unsigned NumLoads, unsigned NumBytes) const override; 198 199 bool shouldScheduleLoadsNear(SDNode *Load0, SDNode *Load1, int64_t Offset0, 200 int64_t Offset1, unsigned NumLoads) const override; 201 202 void copyPhysReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI, 203 const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg, 204 bool KillSrc) const override; 205 206 void materializeImmediate(MachineBasicBlock &MBB, 207 MachineBasicBlock::iterator MI, 208 const DebugLoc &DL, 209 unsigned DestReg, 210 int64_t Value) const; 211 212 const TargetRegisterClass *getPreferredSelectRegClass( 213 unsigned Size) const; 214 215 Register insertNE(MachineBasicBlock *MBB, 216 MachineBasicBlock::iterator I, const DebugLoc &DL, 217 Register SrcReg, int Value) const; 218 219 Register insertEQ(MachineBasicBlock *MBB, 220 MachineBasicBlock::iterator I, const DebugLoc &DL, 221 Register SrcReg, int Value) const; 222 223 void storeRegToStackSlot(MachineBasicBlock &MBB, 224 MachineBasicBlock::iterator MI, Register SrcReg, 225 bool isKill, int FrameIndex, 226 const TargetRegisterClass *RC, 227 const TargetRegisterInfo *TRI) const override; 228 229 void loadRegFromStackSlot(MachineBasicBlock &MBB, 230 MachineBasicBlock::iterator MI, Register DestReg, 231 int FrameIndex, const TargetRegisterClass *RC, 232 const TargetRegisterInfo *TRI) const override; 233 234 bool expandPostRAPseudo(MachineInstr &MI) const override; 235 236 // Splits a V_MOV_B64_DPP_PSEUDO opcode into a pair of v_mov_b32_dpp 237 // instructions. Returns a pair of generated instructions. 238 // Can split either post-RA with physical registers or pre-RA with 239 // virtual registers. In latter case IR needs to be in SSA form and 240 // and a REG_SEQUENCE is produced to define original register. 241 std::pair<MachineInstr*, MachineInstr*> 242 expandMovDPP64(MachineInstr &MI) const; 243 244 // Returns an opcode that can be used to move a value to a \p DstRC 245 // register. If there is no hardware instruction that can store to \p 246 // DstRC, then AMDGPU::COPY is returned. 247 unsigned getMovOpcode(const TargetRegisterClass *DstRC) const; 248 249 const MCInstrDesc &getIndirectRegWriteMovRelPseudo(unsigned VecSize, 250 unsigned EltSize, 251 bool IsSGPR) const; 252 253 const MCInstrDesc &getIndirectGPRIDXPseudo(unsigned VecSize, 254 bool IsIndirectSrc) const; 255 LLVM_READONLY 256 int commuteOpcode(unsigned Opc) const; 257 258 LLVM_READONLY 259 inline int commuteOpcode(const MachineInstr &MI) const { 260 return commuteOpcode(MI.getOpcode()); 261 } 262 263 bool findCommutedOpIndices(const MachineInstr &MI, unsigned &SrcOpIdx1, 264 unsigned &SrcOpIdx2) const override; 265 266 bool findCommutedOpIndices(MCInstrDesc Desc, unsigned & SrcOpIdx0, 267 unsigned & SrcOpIdx1) const; 268 269 bool isBranchOffsetInRange(unsigned BranchOpc, 270 int64_t BrOffset) const override; 271 272 MachineBasicBlock *getBranchDestBlock(const MachineInstr &MI) const override; 273 274 unsigned insertIndirectBranch(MachineBasicBlock &MBB, 275 MachineBasicBlock &NewDestBB, 276 const DebugLoc &DL, 277 int64_t BrOffset, 278 RegScavenger *RS = nullptr) const override; 279 280 bool analyzeBranchImpl(MachineBasicBlock &MBB, 281 MachineBasicBlock::iterator I, 282 MachineBasicBlock *&TBB, 283 MachineBasicBlock *&FBB, 284 SmallVectorImpl<MachineOperand> &Cond, 285 bool AllowModify) const; 286 287 bool analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB, 288 MachineBasicBlock *&FBB, 289 SmallVectorImpl<MachineOperand> &Cond, 290 bool AllowModify = false) const override; 291 292 unsigned removeBranch(MachineBasicBlock &MBB, 293 int *BytesRemoved = nullptr) const override; 294 295 unsigned insertBranch(MachineBasicBlock &MBB, MachineBasicBlock *TBB, 296 MachineBasicBlock *FBB, ArrayRef<MachineOperand> Cond, 297 const DebugLoc &DL, 298 int *BytesAdded = nullptr) const override; 299 300 bool reverseBranchCondition( 301 SmallVectorImpl<MachineOperand> &Cond) const override; 302 303 bool canInsertSelect(const MachineBasicBlock &MBB, 304 ArrayRef<MachineOperand> Cond, Register DstReg, 305 Register TrueReg, Register FalseReg, int &CondCycles, 306 int &TrueCycles, int &FalseCycles) const override; 307 308 void insertSelect(MachineBasicBlock &MBB, 309 MachineBasicBlock::iterator I, const DebugLoc &DL, 310 Register DstReg, ArrayRef<MachineOperand> Cond, 311 Register TrueReg, Register FalseReg) const override; 312 313 void insertVectorSelect(MachineBasicBlock &MBB, 314 MachineBasicBlock::iterator I, const DebugLoc &DL, 315 Register DstReg, ArrayRef<MachineOperand> Cond, 316 Register TrueReg, Register FalseReg) const; 317 318 unsigned getAddressSpaceForPseudoSourceKind( 319 unsigned Kind) const override; 320 321 bool 322 areMemAccessesTriviallyDisjoint(const MachineInstr &MIa, 323 const MachineInstr &MIb) const override; 324 325 bool isFoldableCopy(const MachineInstr &MI) const; 326 327 bool FoldImmediate(MachineInstr &UseMI, MachineInstr &DefMI, Register Reg, 328 MachineRegisterInfo *MRI) const final; 329 330 unsigned getMachineCSELookAheadLimit() const override { return 500; } 331 332 MachineInstr *convertToThreeAddress(MachineFunction::iterator &MBB, 333 MachineInstr &MI, 334 LiveVariables *LV) const override; 335 336 bool isSchedulingBoundary(const MachineInstr &MI, 337 const MachineBasicBlock *MBB, 338 const MachineFunction &MF) const override; 339 340 static bool isSALU(const MachineInstr &MI) { 341 return MI.getDesc().TSFlags & SIInstrFlags::SALU; 342 } 343 344 bool isSALU(uint16_t Opcode) const { 345 return get(Opcode).TSFlags & SIInstrFlags::SALU; 346 } 347 348 static bool isVALU(const MachineInstr &MI) { 349 return MI.getDesc().TSFlags & SIInstrFlags::VALU; 350 } 351 352 bool isVALU(uint16_t Opcode) const { 353 return get(Opcode).TSFlags & SIInstrFlags::VALU; 354 } 355 356 static bool isVMEM(const MachineInstr &MI) { 357 return isMUBUF(MI) || isMTBUF(MI) || isMIMG(MI); 358 } 359 360 bool isVMEM(uint16_t Opcode) const { 361 return isMUBUF(Opcode) || isMTBUF(Opcode) || isMIMG(Opcode); 362 } 363 364 static bool isSOP1(const MachineInstr &MI) { 365 return MI.getDesc().TSFlags & SIInstrFlags::SOP1; 366 } 367 368 bool isSOP1(uint16_t Opcode) const { 369 return get(Opcode).TSFlags & SIInstrFlags::SOP1; 370 } 371 372 static bool isSOP2(const MachineInstr &MI) { 373 return MI.getDesc().TSFlags & SIInstrFlags::SOP2; 374 } 375 376 bool isSOP2(uint16_t Opcode) const { 377 return get(Opcode).TSFlags & SIInstrFlags::SOP2; 378 } 379 380 static bool isSOPC(const MachineInstr &MI) { 381 return MI.getDesc().TSFlags & SIInstrFlags::SOPC; 382 } 383 384 bool isSOPC(uint16_t Opcode) const { 385 return get(Opcode).TSFlags & SIInstrFlags::SOPC; 386 } 387 388 static bool isSOPK(const MachineInstr &MI) { 389 return MI.getDesc().TSFlags & SIInstrFlags::SOPK; 390 } 391 392 bool isSOPK(uint16_t Opcode) const { 393 return get(Opcode).TSFlags & SIInstrFlags::SOPK; 394 } 395 396 static bool isSOPP(const MachineInstr &MI) { 397 return MI.getDesc().TSFlags & SIInstrFlags::SOPP; 398 } 399 400 bool isSOPP(uint16_t Opcode) const { 401 return get(Opcode).TSFlags & SIInstrFlags::SOPP; 402 } 403 404 static bool isPacked(const MachineInstr &MI) { 405 return MI.getDesc().TSFlags & SIInstrFlags::IsPacked; 406 } 407 408 bool isPacked(uint16_t Opcode) const { 409 return get(Opcode).TSFlags & SIInstrFlags::IsPacked; 410 } 411 412 static bool isVOP1(const MachineInstr &MI) { 413 return MI.getDesc().TSFlags & SIInstrFlags::VOP1; 414 } 415 416 bool isVOP1(uint16_t Opcode) const { 417 return get(Opcode).TSFlags & SIInstrFlags::VOP1; 418 } 419 420 static bool isVOP2(const MachineInstr &MI) { 421 return MI.getDesc().TSFlags & SIInstrFlags::VOP2; 422 } 423 424 bool isVOP2(uint16_t Opcode) const { 425 return get(Opcode).TSFlags & SIInstrFlags::VOP2; 426 } 427 428 static bool isVOP3(const MachineInstr &MI) { 429 return MI.getDesc().TSFlags & SIInstrFlags::VOP3; 430 } 431 432 bool isVOP3(uint16_t Opcode) const { 433 return get(Opcode).TSFlags & SIInstrFlags::VOP3; 434 } 435 436 static bool isSDWA(const MachineInstr &MI) { 437 return MI.getDesc().TSFlags & SIInstrFlags::SDWA; 438 } 439 440 bool isSDWA(uint16_t Opcode) const { 441 return get(Opcode).TSFlags & SIInstrFlags::SDWA; 442 } 443 444 static bool isVOPC(const MachineInstr &MI) { 445 return MI.getDesc().TSFlags & SIInstrFlags::VOPC; 446 } 447 448 bool isVOPC(uint16_t Opcode) const { 449 return get(Opcode).TSFlags & SIInstrFlags::VOPC; 450 } 451 452 static bool isMUBUF(const MachineInstr &MI) { 453 return MI.getDesc().TSFlags & SIInstrFlags::MUBUF; 454 } 455 456 bool isMUBUF(uint16_t Opcode) const { 457 return get(Opcode).TSFlags & SIInstrFlags::MUBUF; 458 } 459 460 static bool isMTBUF(const MachineInstr &MI) { 461 return MI.getDesc().TSFlags & SIInstrFlags::MTBUF; 462 } 463 464 bool isMTBUF(uint16_t Opcode) const { 465 return get(Opcode).TSFlags & SIInstrFlags::MTBUF; 466 } 467 468 static bool isSMRD(const MachineInstr &MI) { 469 return MI.getDesc().TSFlags & SIInstrFlags::SMRD; 470 } 471 472 bool isSMRD(uint16_t Opcode) const { 473 return get(Opcode).TSFlags & SIInstrFlags::SMRD; 474 } 475 476 bool isBufferSMRD(const MachineInstr &MI) const; 477 478 static bool isDS(const MachineInstr &MI) { 479 return MI.getDesc().TSFlags & SIInstrFlags::DS; 480 } 481 482 bool isDS(uint16_t Opcode) const { 483 return get(Opcode).TSFlags & SIInstrFlags::DS; 484 } 485 486 bool isAlwaysGDS(uint16_t Opcode) const; 487 488 static bool isMIMG(const MachineInstr &MI) { 489 return MI.getDesc().TSFlags & SIInstrFlags::MIMG; 490 } 491 492 bool isMIMG(uint16_t Opcode) const { 493 return get(Opcode).TSFlags & SIInstrFlags::MIMG; 494 } 495 496 static bool isGather4(const MachineInstr &MI) { 497 return MI.getDesc().TSFlags & SIInstrFlags::Gather4; 498 } 499 500 bool isGather4(uint16_t Opcode) const { 501 return get(Opcode).TSFlags & SIInstrFlags::Gather4; 502 } 503 504 static bool isFLAT(const MachineInstr &MI) { 505 return MI.getDesc().TSFlags & SIInstrFlags::FLAT; 506 } 507 508 // Is a FLAT encoded instruction which accesses a specific segment, 509 // i.e. global_* or scratch_*. 510 static bool isSegmentSpecificFLAT(const MachineInstr &MI) { 511 auto Flags = MI.getDesc().TSFlags; 512 return Flags & (SIInstrFlags::FlatGlobal | SIInstrFlags::FlatScratch); 513 } 514 515 bool isSegmentSpecificFLAT(uint16_t Opcode) const { 516 auto Flags = get(Opcode).TSFlags; 517 return Flags & (SIInstrFlags::FlatGlobal | SIInstrFlags::FlatScratch); 518 } 519 520 static bool isFLATGlobal(const MachineInstr &MI) { 521 return MI.getDesc().TSFlags & SIInstrFlags::FlatGlobal; 522 } 523 524 bool isFLATGlobal(uint16_t Opcode) const { 525 return get(Opcode).TSFlags & SIInstrFlags::FlatGlobal; 526 } 527 528 static bool isFLATScratch(const MachineInstr &MI) { 529 return MI.getDesc().TSFlags & SIInstrFlags::FlatScratch; 530 } 531 532 bool isFLATScratch(uint16_t Opcode) const { 533 return get(Opcode).TSFlags & SIInstrFlags::FlatScratch; 534 } 535 536 // Any FLAT encoded instruction, including global_* and scratch_*. 537 bool isFLAT(uint16_t Opcode) const { 538 return get(Opcode).TSFlags & SIInstrFlags::FLAT; 539 } 540 541 static bool isEXP(const MachineInstr &MI) { 542 return MI.getDesc().TSFlags & SIInstrFlags::EXP; 543 } 544 545 bool isEXP(uint16_t Opcode) const { 546 return get(Opcode).TSFlags & SIInstrFlags::EXP; 547 } 548 549 static bool isAtomicNoRet(const MachineInstr &MI) { 550 return MI.getDesc().TSFlags & SIInstrFlags::IsAtomicNoRet; 551 } 552 553 bool isAtomicNoRet(uint16_t Opcode) const { 554 return get(Opcode).TSFlags & SIInstrFlags::IsAtomicNoRet; 555 } 556 557 static bool isAtomicRet(const MachineInstr &MI) { 558 return MI.getDesc().TSFlags & SIInstrFlags::IsAtomicRet; 559 } 560 561 bool isAtomicRet(uint16_t Opcode) const { 562 return get(Opcode).TSFlags & SIInstrFlags::IsAtomicRet; 563 } 564 565 static bool isAtomic(const MachineInstr &MI) { 566 return MI.getDesc().TSFlags & (SIInstrFlags::IsAtomicRet | 567 SIInstrFlags::IsAtomicNoRet); 568 } 569 570 bool isAtomic(uint16_t Opcode) const { 571 return get(Opcode).TSFlags & (SIInstrFlags::IsAtomicRet | 572 SIInstrFlags::IsAtomicNoRet); 573 } 574 575 static bool isWQM(const MachineInstr &MI) { 576 return MI.getDesc().TSFlags & SIInstrFlags::WQM; 577 } 578 579 bool isWQM(uint16_t Opcode) const { 580 return get(Opcode).TSFlags & SIInstrFlags::WQM; 581 } 582 583 static bool isDisableWQM(const MachineInstr &MI) { 584 return MI.getDesc().TSFlags & SIInstrFlags::DisableWQM; 585 } 586 587 bool isDisableWQM(uint16_t Opcode) const { 588 return get(Opcode).TSFlags & SIInstrFlags::DisableWQM; 589 } 590 591 static bool isVGPRSpill(const MachineInstr &MI) { 592 return MI.getDesc().TSFlags & SIInstrFlags::VGPRSpill; 593 } 594 595 bool isVGPRSpill(uint16_t Opcode) const { 596 return get(Opcode).TSFlags & SIInstrFlags::VGPRSpill; 597 } 598 599 static bool isSGPRSpill(const MachineInstr &MI) { 600 return MI.getDesc().TSFlags & SIInstrFlags::SGPRSpill; 601 } 602 603 bool isSGPRSpill(uint16_t Opcode) const { 604 return get(Opcode).TSFlags & SIInstrFlags::SGPRSpill; 605 } 606 607 static bool isDPP(const MachineInstr &MI) { 608 return MI.getDesc().TSFlags & SIInstrFlags::DPP; 609 } 610 611 bool isDPP(uint16_t Opcode) const { 612 return get(Opcode).TSFlags & SIInstrFlags::DPP; 613 } 614 615 static bool isTRANS(const MachineInstr &MI) { 616 return MI.getDesc().TSFlags & SIInstrFlags::TRANS; 617 } 618 619 bool isTRANS(uint16_t Opcode) const { 620 return get(Opcode).TSFlags & SIInstrFlags::TRANS; 621 } 622 623 static bool isVOP3P(const MachineInstr &MI) { 624 return MI.getDesc().TSFlags & SIInstrFlags::VOP3P; 625 } 626 627 bool isVOP3P(uint16_t Opcode) const { 628 return get(Opcode).TSFlags & SIInstrFlags::VOP3P; 629 } 630 631 static bool isVINTRP(const MachineInstr &MI) { 632 return MI.getDesc().TSFlags & SIInstrFlags::VINTRP; 633 } 634 635 bool isVINTRP(uint16_t Opcode) const { 636 return get(Opcode).TSFlags & SIInstrFlags::VINTRP; 637 } 638 639 static bool isMAI(const MachineInstr &MI) { 640 return MI.getDesc().TSFlags & SIInstrFlags::IsMAI; 641 } 642 643 bool isMAI(uint16_t Opcode) const { 644 return get(Opcode).TSFlags & SIInstrFlags::IsMAI; 645 } 646 647 static bool isDOT(const MachineInstr &MI) { 648 return MI.getDesc().TSFlags & SIInstrFlags::IsDOT; 649 } 650 651 bool isDOT(uint16_t Opcode) const { 652 return get(Opcode).TSFlags & SIInstrFlags::IsDOT; 653 } 654 655 static bool isScalarUnit(const MachineInstr &MI) { 656 return MI.getDesc().TSFlags & (SIInstrFlags::SALU | SIInstrFlags::SMRD); 657 } 658 659 static bool usesVM_CNT(const MachineInstr &MI) { 660 return MI.getDesc().TSFlags & SIInstrFlags::VM_CNT; 661 } 662 663 static bool usesLGKM_CNT(const MachineInstr &MI) { 664 return MI.getDesc().TSFlags & SIInstrFlags::LGKM_CNT; 665 } 666 667 static bool sopkIsZext(const MachineInstr &MI) { 668 return MI.getDesc().TSFlags & SIInstrFlags::SOPK_ZEXT; 669 } 670 671 bool sopkIsZext(uint16_t Opcode) const { 672 return get(Opcode).TSFlags & SIInstrFlags::SOPK_ZEXT; 673 } 674 675 /// \returns true if this is an s_store_dword* instruction. This is more 676 /// specific than than isSMEM && mayStore. 677 static bool isScalarStore(const MachineInstr &MI) { 678 return MI.getDesc().TSFlags & SIInstrFlags::SCALAR_STORE; 679 } 680 681 bool isScalarStore(uint16_t Opcode) const { 682 return get(Opcode).TSFlags & SIInstrFlags::SCALAR_STORE; 683 } 684 685 static bool isFixedSize(const MachineInstr &MI) { 686 return MI.getDesc().TSFlags & SIInstrFlags::FIXED_SIZE; 687 } 688 689 bool isFixedSize(uint16_t Opcode) const { 690 return get(Opcode).TSFlags & SIInstrFlags::FIXED_SIZE; 691 } 692 693 static bool hasFPClamp(const MachineInstr &MI) { 694 return MI.getDesc().TSFlags & SIInstrFlags::FPClamp; 695 } 696 697 bool hasFPClamp(uint16_t Opcode) const { 698 return get(Opcode).TSFlags & SIInstrFlags::FPClamp; 699 } 700 701 static bool hasIntClamp(const MachineInstr &MI) { 702 return MI.getDesc().TSFlags & SIInstrFlags::IntClamp; 703 } 704 705 uint64_t getClampMask(const MachineInstr &MI) const { 706 const uint64_t ClampFlags = SIInstrFlags::FPClamp | 707 SIInstrFlags::IntClamp | 708 SIInstrFlags::ClampLo | 709 SIInstrFlags::ClampHi; 710 return MI.getDesc().TSFlags & ClampFlags; 711 } 712 713 static bool usesFPDPRounding(const MachineInstr &MI) { 714 return MI.getDesc().TSFlags & SIInstrFlags::FPDPRounding; 715 } 716 717 bool usesFPDPRounding(uint16_t Opcode) const { 718 return get(Opcode).TSFlags & SIInstrFlags::FPDPRounding; 719 } 720 721 static bool isFPAtomic(const MachineInstr &MI) { 722 return MI.getDesc().TSFlags & SIInstrFlags::FPAtomic; 723 } 724 725 bool isFPAtomic(uint16_t Opcode) const { 726 return get(Opcode).TSFlags & SIInstrFlags::FPAtomic; 727 } 728 729 bool isVGPRCopy(const MachineInstr &MI) const { 730 assert(MI.isCopy()); 731 Register Dest = MI.getOperand(0).getReg(); 732 const MachineFunction &MF = *MI.getParent()->getParent(); 733 const MachineRegisterInfo &MRI = MF.getRegInfo(); 734 return !RI.isSGPRReg(MRI, Dest); 735 } 736 737 bool hasVGPRUses(const MachineInstr &MI) const { 738 const MachineFunction &MF = *MI.getParent()->getParent(); 739 const MachineRegisterInfo &MRI = MF.getRegInfo(); 740 return llvm::any_of(MI.explicit_uses(), 741 [&MRI, this](const MachineOperand &MO) { 742 return MO.isReg() && RI.isVGPR(MRI, MO.getReg());}); 743 } 744 745 /// Return true if the instruction modifies the mode register.q 746 static bool modifiesModeRegister(const MachineInstr &MI); 747 748 /// Whether we must prevent this instruction from executing with EXEC = 0. 749 bool hasUnwantedEffectsWhenEXECEmpty(const MachineInstr &MI) const; 750 751 /// Returns true if the instruction could potentially depend on the value of 752 /// exec. If false, exec dependencies may safely be ignored. 753 bool mayReadEXEC(const MachineRegisterInfo &MRI, const MachineInstr &MI) const; 754 755 bool isInlineConstant(const APInt &Imm) const; 756 757 bool isInlineConstant(const APFloat &Imm) const { 758 return isInlineConstant(Imm.bitcastToAPInt()); 759 } 760 761 bool isInlineConstant(const MachineOperand &MO, uint8_t OperandType) const; 762 763 bool isInlineConstant(const MachineOperand &MO, 764 const MCOperandInfo &OpInfo) const { 765 return isInlineConstant(MO, OpInfo.OperandType); 766 } 767 768 /// \p returns true if \p UseMO is substituted with \p DefMO in \p MI it would 769 /// be an inline immediate. 770 bool isInlineConstant(const MachineInstr &MI, 771 const MachineOperand &UseMO, 772 const MachineOperand &DefMO) const { 773 assert(UseMO.getParent() == &MI); 774 int OpIdx = MI.getOperandNo(&UseMO); 775 if (!MI.getDesc().OpInfo || OpIdx >= MI.getDesc().NumOperands) { 776 return false; 777 } 778 779 return isInlineConstant(DefMO, MI.getDesc().OpInfo[OpIdx]); 780 } 781 782 /// \p returns true if the operand \p OpIdx in \p MI is a valid inline 783 /// immediate. 784 bool isInlineConstant(const MachineInstr &MI, unsigned OpIdx) const { 785 const MachineOperand &MO = MI.getOperand(OpIdx); 786 return isInlineConstant(MO, MI.getDesc().OpInfo[OpIdx].OperandType); 787 } 788 789 bool isInlineConstant(const MachineInstr &MI, unsigned OpIdx, 790 const MachineOperand &MO) const { 791 if (!MI.getDesc().OpInfo || OpIdx >= MI.getDesc().NumOperands) 792 return false; 793 794 if (MI.isCopy()) { 795 unsigned Size = getOpSize(MI, OpIdx); 796 assert(Size == 8 || Size == 4); 797 798 uint8_t OpType = (Size == 8) ? 799 AMDGPU::OPERAND_REG_IMM_INT64 : AMDGPU::OPERAND_REG_IMM_INT32; 800 return isInlineConstant(MO, OpType); 801 } 802 803 return isInlineConstant(MO, MI.getDesc().OpInfo[OpIdx].OperandType); 804 } 805 806 bool isInlineConstant(const MachineOperand &MO) const { 807 const MachineInstr *Parent = MO.getParent(); 808 return isInlineConstant(*Parent, Parent->getOperandNo(&MO)); 809 } 810 811 bool isLiteralConstant(const MachineOperand &MO, 812 const MCOperandInfo &OpInfo) const { 813 return MO.isImm() && !isInlineConstant(MO, OpInfo.OperandType); 814 } 815 816 bool isLiteralConstant(const MachineInstr &MI, int OpIdx) const { 817 const MachineOperand &MO = MI.getOperand(OpIdx); 818 return MO.isImm() && !isInlineConstant(MI, OpIdx); 819 } 820 821 // Returns true if this operand could potentially require a 32-bit literal 822 // operand, but not necessarily. A FrameIndex for example could resolve to an 823 // inline immediate value that will not require an additional 4-bytes; this 824 // assumes that it will. 825 bool isLiteralConstantLike(const MachineOperand &MO, 826 const MCOperandInfo &OpInfo) const; 827 828 bool isImmOperandLegal(const MachineInstr &MI, unsigned OpNo, 829 const MachineOperand &MO) const; 830 831 /// Return true if this 64-bit VALU instruction has a 32-bit encoding. 832 /// This function will return false if you pass it a 32-bit instruction. 833 bool hasVALU32BitEncoding(unsigned Opcode) const; 834 835 /// Returns true if this operand uses the constant bus. 836 bool usesConstantBus(const MachineRegisterInfo &MRI, 837 const MachineOperand &MO, 838 const MCOperandInfo &OpInfo) const; 839 840 /// Return true if this instruction has any modifiers. 841 /// e.g. src[012]_mod, omod, clamp. 842 bool hasModifiers(unsigned Opcode) const; 843 844 bool hasModifiersSet(const MachineInstr &MI, 845 unsigned OpName) const; 846 bool hasAnyModifiersSet(const MachineInstr &MI) const; 847 848 bool canShrink(const MachineInstr &MI, 849 const MachineRegisterInfo &MRI) const; 850 851 MachineInstr *buildShrunkInst(MachineInstr &MI, 852 unsigned NewOpcode) const; 853 854 bool verifyInstruction(const MachineInstr &MI, 855 StringRef &ErrInfo) const override; 856 857 unsigned getVALUOp(const MachineInstr &MI) const; 858 859 /// Return the correct register class for \p OpNo. For target-specific 860 /// instructions, this will return the register class that has been defined 861 /// in tablegen. For generic instructions, like REG_SEQUENCE it will return 862 /// the register class of its machine operand. 863 /// to infer the correct register class base on the other operands. 864 const TargetRegisterClass *getOpRegClass(const MachineInstr &MI, 865 unsigned OpNo) const; 866 867 /// Return the size in bytes of the operand OpNo on the given 868 // instruction opcode. 869 unsigned getOpSize(uint16_t Opcode, unsigned OpNo) const { 870 const MCOperandInfo &OpInfo = get(Opcode).OpInfo[OpNo]; 871 872 if (OpInfo.RegClass == -1) { 873 // If this is an immediate operand, this must be a 32-bit literal. 874 assert(OpInfo.OperandType == MCOI::OPERAND_IMMEDIATE); 875 return 4; 876 } 877 878 return RI.getRegSizeInBits(*RI.getRegClass(OpInfo.RegClass)) / 8; 879 } 880 881 /// This form should usually be preferred since it handles operands 882 /// with unknown register classes. 883 unsigned getOpSize(const MachineInstr &MI, unsigned OpNo) const { 884 const MachineOperand &MO = MI.getOperand(OpNo); 885 if (MO.isReg()) { 886 if (unsigned SubReg = MO.getSubReg()) { 887 return RI.getSubRegIdxSize(SubReg) / 8; 888 } 889 } 890 return RI.getRegSizeInBits(*getOpRegClass(MI, OpNo)) / 8; 891 } 892 893 /// Legalize the \p OpIndex operand of this instruction by inserting 894 /// a MOV. For example: 895 /// ADD_I32_e32 VGPR0, 15 896 /// to 897 /// MOV VGPR1, 15 898 /// ADD_I32_e32 VGPR0, VGPR1 899 /// 900 /// If the operand being legalized is a register, then a COPY will be used 901 /// instead of MOV. 902 void legalizeOpWithMove(MachineInstr &MI, unsigned OpIdx) const; 903 904 /// Check if \p MO is a legal operand if it was the \p OpIdx Operand 905 /// for \p MI. 906 bool isOperandLegal(const MachineInstr &MI, unsigned OpIdx, 907 const MachineOperand *MO = nullptr) const; 908 909 /// Check if \p MO would be a valid operand for the given operand 910 /// definition \p OpInfo. Note this does not attempt to validate constant bus 911 /// restrictions (e.g. literal constant usage). 912 bool isLegalVSrcOperand(const MachineRegisterInfo &MRI, 913 const MCOperandInfo &OpInfo, 914 const MachineOperand &MO) const; 915 916 /// Check if \p MO (a register operand) is a legal register for the 917 /// given operand description. 918 bool isLegalRegOperand(const MachineRegisterInfo &MRI, 919 const MCOperandInfo &OpInfo, 920 const MachineOperand &MO) const; 921 922 /// Legalize operands in \p MI by either commuting it or inserting a 923 /// copy of src1. 924 void legalizeOperandsVOP2(MachineRegisterInfo &MRI, MachineInstr &MI) const; 925 926 /// Fix operands in \p MI to satisfy constant bus requirements. 927 void legalizeOperandsVOP3(MachineRegisterInfo &MRI, MachineInstr &MI) const; 928 929 /// Copy a value from a VGPR (\p SrcReg) to SGPR. This function can only 930 /// be used when it is know that the value in SrcReg is same across all 931 /// threads in the wave. 932 /// \returns The SGPR register that \p SrcReg was copied to. 933 Register readlaneVGPRToSGPR(Register SrcReg, MachineInstr &UseMI, 934 MachineRegisterInfo &MRI) const; 935 936 void legalizeOperandsSMRD(MachineRegisterInfo &MRI, MachineInstr &MI) const; 937 void legalizeOperandsFLAT(MachineRegisterInfo &MRI, MachineInstr &MI) const; 938 939 void legalizeGenericOperand(MachineBasicBlock &InsertMBB, 940 MachineBasicBlock::iterator I, 941 const TargetRegisterClass *DstRC, 942 MachineOperand &Op, MachineRegisterInfo &MRI, 943 const DebugLoc &DL) const; 944 945 /// Legalize all operands in this instruction. This function may create new 946 /// instructions and control-flow around \p MI. If present, \p MDT is 947 /// updated. 948 /// \returns A new basic block that contains \p MI if new blocks were created. 949 MachineBasicBlock * 950 legalizeOperands(MachineInstr &MI, MachineDominatorTree *MDT = nullptr) const; 951 952 /// Change SADDR form of a FLAT \p Inst to its VADDR form if saddr operand 953 /// was moved to VGPR. \returns true if succeeded. 954 bool moveFlatAddrToVGPR(MachineInstr &Inst) const; 955 956 /// Replace this instruction's opcode with the equivalent VALU 957 /// opcode. This function will also move the users of \p MI to the 958 /// VALU if necessary. If present, \p MDT is updated. 959 MachineBasicBlock *moveToVALU(MachineInstr &MI, 960 MachineDominatorTree *MDT = nullptr) const; 961 962 void insertNoop(MachineBasicBlock &MBB, 963 MachineBasicBlock::iterator MI) const override; 964 965 void insertNoops(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI, 966 unsigned Quantity) const override; 967 968 void insertReturn(MachineBasicBlock &MBB) const; 969 /// Return the number of wait states that result from executing this 970 /// instruction. 971 static unsigned getNumWaitStates(const MachineInstr &MI); 972 973 /// Returns the operand named \p Op. If \p MI does not have an 974 /// operand named \c Op, this function returns nullptr. 975 LLVM_READONLY 976 MachineOperand *getNamedOperand(MachineInstr &MI, unsigned OperandName) const; 977 978 LLVM_READONLY 979 const MachineOperand *getNamedOperand(const MachineInstr &MI, 980 unsigned OpName) const { 981 return getNamedOperand(const_cast<MachineInstr &>(MI), OpName); 982 } 983 984 /// Get required immediate operand 985 int64_t getNamedImmOperand(const MachineInstr &MI, unsigned OpName) const { 986 int Idx = AMDGPU::getNamedOperandIdx(MI.getOpcode(), OpName); 987 return MI.getOperand(Idx).getImm(); 988 } 989 990 uint64_t getDefaultRsrcDataFormat() const; 991 uint64_t getScratchRsrcWords23() const; 992 993 bool isLowLatencyInstruction(const MachineInstr &MI) const; 994 bool isHighLatencyDef(int Opc) const override; 995 996 /// Return the descriptor of the target-specific machine instruction 997 /// that corresponds to the specified pseudo or native opcode. 998 const MCInstrDesc &getMCOpcodeFromPseudo(unsigned Opcode) const { 999 return get(pseudoToMCOpcode(Opcode)); 1000 } 1001 1002 unsigned isStackAccess(const MachineInstr &MI, int &FrameIndex) const; 1003 unsigned isSGPRStackAccess(const MachineInstr &MI, int &FrameIndex) const; 1004 1005 unsigned isLoadFromStackSlot(const MachineInstr &MI, 1006 int &FrameIndex) const override; 1007 unsigned isStoreToStackSlot(const MachineInstr &MI, 1008 int &FrameIndex) const override; 1009 1010 unsigned getInstBundleSize(const MachineInstr &MI) const; 1011 unsigned getInstSizeInBytes(const MachineInstr &MI) const override; 1012 1013 bool mayAccessFlatAddressSpace(const MachineInstr &MI) const; 1014 1015 bool isNonUniformBranchInstr(MachineInstr &Instr) const; 1016 1017 void convertNonUniformIfRegion(MachineBasicBlock *IfEntry, 1018 MachineBasicBlock *IfEnd) const; 1019 1020 void convertNonUniformLoopRegion(MachineBasicBlock *LoopEntry, 1021 MachineBasicBlock *LoopEnd) const; 1022 1023 std::pair<unsigned, unsigned> 1024 decomposeMachineOperandsTargetFlags(unsigned TF) const override; 1025 1026 ArrayRef<std::pair<int, const char *>> 1027 getSerializableTargetIndices() const override; 1028 1029 ArrayRef<std::pair<unsigned, const char *>> 1030 getSerializableDirectMachineOperandTargetFlags() const override; 1031 1032 ScheduleHazardRecognizer * 1033 CreateTargetPostRAHazardRecognizer(const InstrItineraryData *II, 1034 const ScheduleDAG *DAG) const override; 1035 1036 ScheduleHazardRecognizer * 1037 CreateTargetPostRAHazardRecognizer(const MachineFunction &MF) const override; 1038 1039 bool isBasicBlockPrologue(const MachineInstr &MI) const override; 1040 1041 MachineInstr *createPHIDestinationCopy(MachineBasicBlock &MBB, 1042 MachineBasicBlock::iterator InsPt, 1043 const DebugLoc &DL, Register Src, 1044 Register Dst) const override; 1045 1046 MachineInstr *createPHISourceCopy(MachineBasicBlock &MBB, 1047 MachineBasicBlock::iterator InsPt, 1048 const DebugLoc &DL, Register Src, 1049 unsigned SrcSubReg, 1050 Register Dst) const override; 1051 1052 bool isWave32() const; 1053 1054 /// Return a partially built integer add instruction without carry. 1055 /// Caller must add source operands. 1056 /// For pre-GFX9 it will generate unused carry destination operand. 1057 /// TODO: After GFX9 it should return a no-carry operation. 1058 MachineInstrBuilder getAddNoCarry(MachineBasicBlock &MBB, 1059 MachineBasicBlock::iterator I, 1060 const DebugLoc &DL, 1061 Register DestReg) const; 1062 1063 MachineInstrBuilder getAddNoCarry(MachineBasicBlock &MBB, 1064 MachineBasicBlock::iterator I, 1065 const DebugLoc &DL, 1066 Register DestReg, 1067 RegScavenger &RS) const; 1068 1069 static bool isKillTerminator(unsigned Opcode); 1070 const MCInstrDesc &getKillTerminatorFromPseudo(unsigned Opcode) const; 1071 1072 static bool isLegalMUBUFImmOffset(unsigned Imm) { 1073 return isUInt<12>(Imm); 1074 } 1075 1076 /// Returns if \p Offset is legal for the subtarget as the offset to a FLAT 1077 /// encoded instruction. If \p Signed, this is for an instruction that 1078 /// interprets the offset as signed. 1079 bool isLegalFLATOffset(int64_t Offset, unsigned AddrSpace, 1080 uint64_t FlatVariant) const; 1081 1082 /// Split \p COffsetVal into {immediate offset field, remainder offset} 1083 /// values. 1084 std::pair<int64_t, int64_t> splitFlatOffset(int64_t COffsetVal, 1085 unsigned AddrSpace, 1086 uint64_t FlatVariant) const; 1087 1088 /// \brief Return a target-specific opcode if Opcode is a pseudo instruction. 1089 /// Return -1 if the target-specific opcode for the pseudo instruction does 1090 /// not exist. If Opcode is not a pseudo instruction, this is identity. 1091 int pseudoToMCOpcode(int Opcode) const; 1092 1093 /// \brief Check if this instruction should only be used by assembler. 1094 /// Return true if this opcode should not be used by codegen. 1095 bool isAsmOnlyOpcode(int MCOp) const; 1096 1097 const TargetRegisterClass *getRegClass(const MCInstrDesc &TID, unsigned OpNum, 1098 const TargetRegisterInfo *TRI, 1099 const MachineFunction &MF) 1100 const override; 1101 1102 void fixImplicitOperands(MachineInstr &MI) const; 1103 1104 MachineInstr *foldMemoryOperandImpl(MachineFunction &MF, MachineInstr &MI, 1105 ArrayRef<unsigned> Ops, 1106 MachineBasicBlock::iterator InsertPt, 1107 int FrameIndex, 1108 LiveIntervals *LIS = nullptr, 1109 VirtRegMap *VRM = nullptr) const override; 1110 1111 unsigned getInstrLatency(const InstrItineraryData *ItinData, 1112 const MachineInstr &MI, 1113 unsigned *PredCost = nullptr) const override; 1114 1115 const MIRFormatter *getMIRFormatter() const override { 1116 if (!Formatter.get()) 1117 Formatter = std::make_unique<AMDGPUMIRFormatter>(); 1118 return Formatter.get(); 1119 } 1120 1121 static unsigned getDSShaderTypeValue(const MachineFunction &MF); 1122 }; 1123 1124 /// \brief Returns true if a reg:subreg pair P has a TRC class 1125 inline bool isOfRegClass(const TargetInstrInfo::RegSubRegPair &P, 1126 const TargetRegisterClass &TRC, 1127 MachineRegisterInfo &MRI) { 1128 auto *RC = MRI.getRegClass(P.Reg); 1129 if (!P.SubReg) 1130 return RC == &TRC; 1131 auto *TRI = MRI.getTargetRegisterInfo(); 1132 return RC == TRI->getMatchingSuperRegClass(RC, &TRC, P.SubReg); 1133 } 1134 1135 /// \brief Create RegSubRegPair from a register MachineOperand 1136 inline 1137 TargetInstrInfo::RegSubRegPair getRegSubRegPair(const MachineOperand &O) { 1138 assert(O.isReg()); 1139 return TargetInstrInfo::RegSubRegPair(O.getReg(), O.getSubReg()); 1140 } 1141 1142 /// \brief Return the SubReg component from REG_SEQUENCE 1143 TargetInstrInfo::RegSubRegPair getRegSequenceSubReg(MachineInstr &MI, 1144 unsigned SubReg); 1145 1146 /// \brief Return the defining instruction for a given reg:subreg pair 1147 /// skipping copy like instructions and subreg-manipulation pseudos. 1148 /// Following another subreg of a reg:subreg isn't supported. 1149 MachineInstr *getVRegSubRegDef(const TargetInstrInfo::RegSubRegPair &P, 1150 MachineRegisterInfo &MRI); 1151 1152 /// \brief Return false if EXEC is not changed between the def of \p VReg at \p 1153 /// DefMI and the use at \p UseMI. Should be run on SSA. Currently does not 1154 /// attempt to track between blocks. 1155 bool execMayBeModifiedBeforeUse(const MachineRegisterInfo &MRI, 1156 Register VReg, 1157 const MachineInstr &DefMI, 1158 const MachineInstr &UseMI); 1159 1160 /// \brief Return false if EXEC is not changed between the def of \p VReg at \p 1161 /// DefMI and all its uses. Should be run on SSA. Currently does not attempt to 1162 /// track between blocks. 1163 bool execMayBeModifiedBeforeAnyUse(const MachineRegisterInfo &MRI, 1164 Register VReg, 1165 const MachineInstr &DefMI); 1166 1167 namespace AMDGPU { 1168 1169 LLVM_READONLY 1170 int getVOPe64(uint16_t Opcode); 1171 1172 LLVM_READONLY 1173 int getVOPe32(uint16_t Opcode); 1174 1175 LLVM_READONLY 1176 int getSDWAOp(uint16_t Opcode); 1177 1178 LLVM_READONLY 1179 int getDPPOp32(uint16_t Opcode); 1180 1181 LLVM_READONLY 1182 int getBasicFromSDWAOp(uint16_t Opcode); 1183 1184 LLVM_READONLY 1185 int getCommuteRev(uint16_t Opcode); 1186 1187 LLVM_READONLY 1188 int getCommuteOrig(uint16_t Opcode); 1189 1190 LLVM_READONLY 1191 int getAddr64Inst(uint16_t Opcode); 1192 1193 /// Check if \p Opcode is an Addr64 opcode. 1194 /// 1195 /// \returns \p Opcode if it is an Addr64 opcode, otherwise -1. 1196 LLVM_READONLY 1197 int getIfAddr64Inst(uint16_t Opcode); 1198 1199 LLVM_READONLY 1200 int getMUBUFNoLdsInst(uint16_t Opcode); 1201 1202 LLVM_READONLY 1203 int getAtomicNoRetOp(uint16_t Opcode); 1204 1205 LLVM_READONLY 1206 int getSOPKOp(uint16_t Opcode); 1207 1208 /// \returns SADDR form of a FLAT Global instruction given an \p Opcode 1209 /// of a VADDR form. 1210 LLVM_READONLY 1211 int getGlobalSaddrOp(uint16_t Opcode); 1212 1213 /// \returns VADDR form of a FLAT Global instruction given an \p Opcode 1214 /// of a SADDR form. 1215 LLVM_READONLY 1216 int getGlobalVaddrOp(uint16_t Opcode); 1217 1218 LLVM_READONLY 1219 int getVCMPXNoSDstOp(uint16_t Opcode); 1220 1221 /// \returns ST form with only immediate offset of a FLAT Scratch instruction 1222 /// given an \p Opcode of an SS (SADDR) form. 1223 LLVM_READONLY 1224 int getFlatScratchInstSTfromSS(uint16_t Opcode); 1225 1226 /// \returns SS (SADDR) form of a FLAT Scratch instruction given an \p Opcode 1227 /// of an SV (VADDR) form. 1228 LLVM_READONLY 1229 int getFlatScratchInstSSfromSV(uint16_t Opcode); 1230 1231 /// \returns SV (VADDR) form of a FLAT Scratch instruction given an \p Opcode 1232 /// of an SS (SADDR) form. 1233 LLVM_READONLY 1234 int getFlatScratchInstSVfromSS(uint16_t Opcode); 1235 1236 const uint64_t RSRC_DATA_FORMAT = 0xf00000000000LL; 1237 const uint64_t RSRC_ELEMENT_SIZE_SHIFT = (32 + 19); 1238 const uint64_t RSRC_INDEX_STRIDE_SHIFT = (32 + 21); 1239 const uint64_t RSRC_TID_ENABLE = UINT64_C(1) << (32 + 23); 1240 1241 } // end namespace AMDGPU 1242 1243 namespace SI { 1244 namespace KernelInputOffsets { 1245 1246 /// Offsets in bytes from the start of the input buffer 1247 enum Offsets { 1248 NGROUPS_X = 0, 1249 NGROUPS_Y = 4, 1250 NGROUPS_Z = 8, 1251 GLOBAL_SIZE_X = 12, 1252 GLOBAL_SIZE_Y = 16, 1253 GLOBAL_SIZE_Z = 20, 1254 LOCAL_SIZE_X = 24, 1255 LOCAL_SIZE_Y = 28, 1256 LOCAL_SIZE_Z = 32 1257 }; 1258 1259 } // end namespace KernelInputOffsets 1260 } // end namespace SI 1261 1262 } // end namespace llvm 1263 1264 #endif // LLVM_LIB_TARGET_AMDGPU_SIINSTRINFO_H 1265