1 //===- SIMemoryLegalizer.cpp ----------------------------------------------===// 2 // 3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. 4 // See https://llvm.org/LICENSE.txt for license information. 5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception 6 // 7 //===----------------------------------------------------------------------===// 8 // 9 /// \file 10 /// Memory legalizer - implements memory model. More information can be 11 /// found here: 12 /// http://llvm.org/docs/AMDGPUUsage.html#memory-model 13 // 14 //===----------------------------------------------------------------------===// 15 16 #include "AMDGPU.h" 17 #include "AMDGPUMachineModuleInfo.h" 18 #include "AMDGPUSubtarget.h" 19 #include "SIDefines.h" 20 #include "SIInstrInfo.h" 21 #include "MCTargetDesc/AMDGPUMCTargetDesc.h" 22 #include "Utils/AMDGPUBaseInfo.h" 23 #include "llvm/ADT/BitmaskEnum.h" 24 #include "llvm/ADT/None.h" 25 #include "llvm/ADT/Optional.h" 26 #include "llvm/CodeGen/MachineBasicBlock.h" 27 #include "llvm/CodeGen/MachineFunction.h" 28 #include "llvm/CodeGen/MachineFunctionPass.h" 29 #include "llvm/CodeGen/MachineInstrBuilder.h" 30 #include "llvm/CodeGen/MachineMemOperand.h" 31 #include "llvm/CodeGen/MachineModuleInfo.h" 32 #include "llvm/CodeGen/MachineOperand.h" 33 #include "llvm/IR/DebugLoc.h" 34 #include "llvm/IR/DiagnosticInfo.h" 35 #include "llvm/IR/Function.h" 36 #include "llvm/IR/LLVMContext.h" 37 #include "llvm/MC/MCInstrDesc.h" 38 #include "llvm/Pass.h" 39 #include "llvm/Support/AtomicOrdering.h" 40 #include "llvm/Support/MathExtras.h" 41 #include <cassert> 42 #include <list> 43 44 using namespace llvm; 45 using namespace llvm::AMDGPU; 46 47 #define DEBUG_TYPE "si-memory-legalizer" 48 #define PASS_NAME "SI Memory Legalizer" 49 50 namespace { 51 52 LLVM_ENABLE_BITMASK_ENUMS_IN_NAMESPACE(); 53 54 /// Memory operation flags. Can be ORed together. 55 enum class SIMemOp { 56 NONE = 0u, 57 LOAD = 1u << 0, 58 STORE = 1u << 1, 59 LLVM_MARK_AS_BITMASK_ENUM(/* LargestFlag = */ STORE) 60 }; 61 62 /// Position to insert a new instruction relative to an existing 63 /// instruction. 64 enum class Position { 65 BEFORE, 66 AFTER 67 }; 68 69 /// The atomic synchronization scopes supported by the AMDGPU target. 70 enum class SIAtomicScope { 71 NONE, 72 SINGLETHREAD, 73 WAVEFRONT, 74 WORKGROUP, 75 AGENT, 76 SYSTEM 77 }; 78 79 /// The distinct address spaces supported by the AMDGPU target for 80 /// atomic memory operation. Can be ORed toether. 81 enum class SIAtomicAddrSpace { 82 NONE = 0u, 83 GLOBAL = 1u << 0, 84 LDS = 1u << 1, 85 SCRATCH = 1u << 2, 86 GDS = 1u << 3, 87 OTHER = 1u << 4, 88 89 /// The address spaces that can be accessed by a FLAT instruction. 90 FLAT = GLOBAL | LDS | SCRATCH, 91 92 /// The address spaces that support atomic instructions. 93 ATOMIC = GLOBAL | LDS | SCRATCH | GDS, 94 95 /// All address spaces. 96 ALL = GLOBAL | LDS | SCRATCH | GDS | OTHER, 97 98 LLVM_MARK_AS_BITMASK_ENUM(/* LargestFlag = */ ALL) 99 }; 100 101 /// Sets named bit \p BitName to "true" if present in instruction \p MI. 102 /// \returns Returns true if \p MI is modified, false otherwise. 103 template <uint16_t BitName> 104 bool enableNamedBit(const MachineBasicBlock::iterator &MI) { 105 int BitIdx = AMDGPU::getNamedOperandIdx(MI->getOpcode(), BitName); 106 if (BitIdx == -1) 107 return false; 108 109 MachineOperand &Bit = MI->getOperand(BitIdx); 110 if (Bit.getImm() != 0) 111 return false; 112 113 Bit.setImm(1); 114 return true; 115 } 116 117 class SIMemOpInfo final { 118 private: 119 120 friend class SIMemOpAccess; 121 122 AtomicOrdering Ordering = AtomicOrdering::NotAtomic; 123 AtomicOrdering FailureOrdering = AtomicOrdering::NotAtomic; 124 SIAtomicScope Scope = SIAtomicScope::SYSTEM; 125 SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE; 126 SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::NONE; 127 bool IsCrossAddressSpaceOrdering = false; 128 bool IsNonTemporal = false; 129 130 SIMemOpInfo(AtomicOrdering Ordering = AtomicOrdering::SequentiallyConsistent, 131 SIAtomicScope Scope = SIAtomicScope::SYSTEM, 132 SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::ATOMIC, 133 SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::ALL, 134 bool IsCrossAddressSpaceOrdering = true, 135 AtomicOrdering FailureOrdering = 136 AtomicOrdering::SequentiallyConsistent, 137 bool IsNonTemporal = false) 138 : Ordering(Ordering), FailureOrdering(FailureOrdering), 139 Scope(Scope), OrderingAddrSpace(OrderingAddrSpace), 140 InstrAddrSpace(InstrAddrSpace), 141 IsCrossAddressSpaceOrdering(IsCrossAddressSpaceOrdering), 142 IsNonTemporal(IsNonTemporal) { 143 // There is also no cross address space ordering if the ordering 144 // address space is the same as the instruction address space and 145 // only contains a single address space. 146 if ((OrderingAddrSpace == InstrAddrSpace) && 147 isPowerOf2_32(uint32_t(InstrAddrSpace))) 148 this->IsCrossAddressSpaceOrdering = false; 149 } 150 151 public: 152 /// \returns Atomic synchronization scope of the machine instruction used to 153 /// create this SIMemOpInfo. 154 SIAtomicScope getScope() const { 155 return Scope; 156 } 157 158 /// \returns Ordering constraint of the machine instruction used to 159 /// create this SIMemOpInfo. 160 AtomicOrdering getOrdering() const { 161 return Ordering; 162 } 163 164 /// \returns Failure ordering constraint of the machine instruction used to 165 /// create this SIMemOpInfo. 166 AtomicOrdering getFailureOrdering() const { 167 return FailureOrdering; 168 } 169 170 /// \returns The address spaces be accessed by the machine 171 /// instruction used to create this SiMemOpInfo. 172 SIAtomicAddrSpace getInstrAddrSpace() const { 173 return InstrAddrSpace; 174 } 175 176 /// \returns The address spaces that must be ordered by the machine 177 /// instruction used to create this SiMemOpInfo. 178 SIAtomicAddrSpace getOrderingAddrSpace() const { 179 return OrderingAddrSpace; 180 } 181 182 /// \returns Return true iff memory ordering of operations on 183 /// different address spaces is required. 184 bool getIsCrossAddressSpaceOrdering() const { 185 return IsCrossAddressSpaceOrdering; 186 } 187 188 /// \returns True if memory access of the machine instruction used to 189 /// create this SIMemOpInfo is non-temporal, false otherwise. 190 bool isNonTemporal() const { 191 return IsNonTemporal; 192 } 193 194 /// \returns True if ordering constraint of the machine instruction used to 195 /// create this SIMemOpInfo is unordered or higher, false otherwise. 196 bool isAtomic() const { 197 return Ordering != AtomicOrdering::NotAtomic; 198 } 199 200 }; 201 202 class SIMemOpAccess final { 203 private: 204 AMDGPUMachineModuleInfo *MMI = nullptr; 205 206 /// Reports unsupported message \p Msg for \p MI to LLVM context. 207 void reportUnsupported(const MachineBasicBlock::iterator &MI, 208 const char *Msg) const; 209 210 /// Inspects the target synchonization scope \p SSID and determines 211 /// the SI atomic scope it corresponds to, the address spaces it 212 /// covers, and whether the memory ordering applies between address 213 /// spaces. 214 Optional<std::tuple<SIAtomicScope, SIAtomicAddrSpace, bool>> 215 toSIAtomicScope(SyncScope::ID SSID, SIAtomicAddrSpace InstrScope) const; 216 217 /// \return Return a bit set of the address spaces accessed by \p AS. 218 SIAtomicAddrSpace toSIAtomicAddrSpace(unsigned AS) const; 219 220 /// \returns Info constructed from \p MI, which has at least machine memory 221 /// operand. 222 Optional<SIMemOpInfo> constructFromMIWithMMO( 223 const MachineBasicBlock::iterator &MI) const; 224 225 public: 226 /// Construct class to support accessing the machine memory operands 227 /// of instructions in the machine function \p MF. 228 SIMemOpAccess(MachineFunction &MF); 229 230 /// \returns Load info if \p MI is a load operation, "None" otherwise. 231 Optional<SIMemOpInfo> getLoadInfo( 232 const MachineBasicBlock::iterator &MI) const; 233 234 /// \returns Store info if \p MI is a store operation, "None" otherwise. 235 Optional<SIMemOpInfo> getStoreInfo( 236 const MachineBasicBlock::iterator &MI) const; 237 238 /// \returns Atomic fence info if \p MI is an atomic fence operation, 239 /// "None" otherwise. 240 Optional<SIMemOpInfo> getAtomicFenceInfo( 241 const MachineBasicBlock::iterator &MI) const; 242 243 /// \returns Atomic cmpxchg/rmw info if \p MI is an atomic cmpxchg or 244 /// rmw operation, "None" otherwise. 245 Optional<SIMemOpInfo> getAtomicCmpxchgOrRmwInfo( 246 const MachineBasicBlock::iterator &MI) const; 247 }; 248 249 class SICacheControl { 250 protected: 251 252 /// Instruction info. 253 const SIInstrInfo *TII = nullptr; 254 255 IsaVersion IV; 256 257 /// Whether to insert cache invalidation instructions. 258 bool InsertCacheInv; 259 260 SICacheControl(const GCNSubtarget &ST); 261 262 public: 263 264 /// Create a cache control for the subtarget \p ST. 265 static std::unique_ptr<SICacheControl> create(const GCNSubtarget &ST); 266 267 /// Update \p MI memory load instruction to bypass any caches up to 268 /// the \p Scope memory scope for address spaces \p 269 /// AddrSpace. Return true iff the instruction was modified. 270 virtual bool enableLoadCacheBypass(const MachineBasicBlock::iterator &MI, 271 SIAtomicScope Scope, 272 SIAtomicAddrSpace AddrSpace) const = 0; 273 274 /// Update \p MI memory instruction to indicate it is 275 /// nontemporal. Return true iff the instruction was modified. 276 virtual bool enableNonTemporal(const MachineBasicBlock::iterator &MI) 277 const = 0; 278 279 /// Inserts any necessary instructions at position \p Pos relative 280 /// to instruction \p MI to ensure any caches associated with 281 /// address spaces \p AddrSpace for memory scopes up to memory scope 282 /// \p Scope are invalidated. Returns true iff any instructions 283 /// inserted. 284 virtual bool insertCacheInvalidate(MachineBasicBlock::iterator &MI, 285 SIAtomicScope Scope, 286 SIAtomicAddrSpace AddrSpace, 287 Position Pos) const = 0; 288 289 /// Inserts any necessary instructions at position \p Pos relative 290 /// to instruction \p MI to ensure memory instructions of kind \p Op 291 /// associated with address spaces \p AddrSpace have completed as 292 /// observed by other memory instructions executing in memory scope 293 /// \p Scope. \p IsCrossAddrSpaceOrdering indicates if the memory 294 /// ordering is between address spaces. Returns true iff any 295 /// instructions inserted. 296 virtual bool insertWait(MachineBasicBlock::iterator &MI, 297 SIAtomicScope Scope, 298 SIAtomicAddrSpace AddrSpace, 299 SIMemOp Op, 300 bool IsCrossAddrSpaceOrdering, 301 Position Pos) const = 0; 302 303 /// Virtual destructor to allow derivations to be deleted. 304 virtual ~SICacheControl() = default; 305 306 }; 307 308 class SIGfx6CacheControl : public SICacheControl { 309 protected: 310 311 /// Sets GLC bit to "true" if present in \p MI. Returns true if \p MI 312 /// is modified, false otherwise. 313 bool enableGLCBit(const MachineBasicBlock::iterator &MI) const { 314 return enableNamedBit<AMDGPU::OpName::glc>(MI); 315 } 316 317 /// Sets SLC bit to "true" if present in \p MI. Returns true if \p MI 318 /// is modified, false otherwise. 319 bool enableSLCBit(const MachineBasicBlock::iterator &MI) const { 320 return enableNamedBit<AMDGPU::OpName::slc>(MI); 321 } 322 323 public: 324 325 SIGfx6CacheControl(const GCNSubtarget &ST) : SICacheControl(ST) {}; 326 327 bool enableLoadCacheBypass(const MachineBasicBlock::iterator &MI, 328 SIAtomicScope Scope, 329 SIAtomicAddrSpace AddrSpace) const override; 330 331 bool enableNonTemporal(const MachineBasicBlock::iterator &MI) const override; 332 333 bool insertCacheInvalidate(MachineBasicBlock::iterator &MI, 334 SIAtomicScope Scope, 335 SIAtomicAddrSpace AddrSpace, 336 Position Pos) const override; 337 338 bool insertWait(MachineBasicBlock::iterator &MI, 339 SIAtomicScope Scope, 340 SIAtomicAddrSpace AddrSpace, 341 SIMemOp Op, 342 bool IsCrossAddrSpaceOrdering, 343 Position Pos) const override; 344 }; 345 346 class SIGfx7CacheControl : public SIGfx6CacheControl { 347 public: 348 349 SIGfx7CacheControl(const GCNSubtarget &ST) : SIGfx6CacheControl(ST) {}; 350 351 bool insertCacheInvalidate(MachineBasicBlock::iterator &MI, 352 SIAtomicScope Scope, 353 SIAtomicAddrSpace AddrSpace, 354 Position Pos) const override; 355 356 }; 357 358 class SIGfx10CacheControl : public SIGfx7CacheControl { 359 protected: 360 bool CuMode = false; 361 362 /// Sets DLC bit to "true" if present in \p MI. Returns true if \p MI 363 /// is modified, false otherwise. 364 bool enableDLCBit(const MachineBasicBlock::iterator &MI) const { 365 return enableNamedBit<AMDGPU::OpName::dlc>(MI); 366 } 367 368 public: 369 370 SIGfx10CacheControl(const GCNSubtarget &ST, bool CuMode) : 371 SIGfx7CacheControl(ST), CuMode(CuMode) {}; 372 373 bool enableLoadCacheBypass(const MachineBasicBlock::iterator &MI, 374 SIAtomicScope Scope, 375 SIAtomicAddrSpace AddrSpace) const override; 376 377 bool enableNonTemporal(const MachineBasicBlock::iterator &MI) const override; 378 379 bool insertCacheInvalidate(MachineBasicBlock::iterator &MI, 380 SIAtomicScope Scope, 381 SIAtomicAddrSpace AddrSpace, 382 Position Pos) const override; 383 384 bool insertWait(MachineBasicBlock::iterator &MI, 385 SIAtomicScope Scope, 386 SIAtomicAddrSpace AddrSpace, 387 SIMemOp Op, 388 bool IsCrossAddrSpaceOrdering, 389 Position Pos) const override; 390 }; 391 392 class SIMemoryLegalizer final : public MachineFunctionPass { 393 private: 394 395 /// Cache Control. 396 std::unique_ptr<SICacheControl> CC = nullptr; 397 398 /// List of atomic pseudo instructions. 399 std::list<MachineBasicBlock::iterator> AtomicPseudoMIs; 400 401 /// Return true iff instruction \p MI is a atomic instruction that 402 /// returns a result. 403 bool isAtomicRet(const MachineInstr &MI) const { 404 return AMDGPU::getAtomicNoRetOp(MI.getOpcode()) != -1; 405 } 406 407 /// Removes all processed atomic pseudo instructions from the current 408 /// function. Returns true if current function is modified, false otherwise. 409 bool removeAtomicPseudoMIs(); 410 411 /// Expands load operation \p MI. Returns true if instructions are 412 /// added/deleted or \p MI is modified, false otherwise. 413 bool expandLoad(const SIMemOpInfo &MOI, 414 MachineBasicBlock::iterator &MI); 415 /// Expands store operation \p MI. Returns true if instructions are 416 /// added/deleted or \p MI is modified, false otherwise. 417 bool expandStore(const SIMemOpInfo &MOI, 418 MachineBasicBlock::iterator &MI); 419 /// Expands atomic fence operation \p MI. Returns true if 420 /// instructions are added/deleted or \p MI is modified, false otherwise. 421 bool expandAtomicFence(const SIMemOpInfo &MOI, 422 MachineBasicBlock::iterator &MI); 423 /// Expands atomic cmpxchg or rmw operation \p MI. Returns true if 424 /// instructions are added/deleted or \p MI is modified, false otherwise. 425 bool expandAtomicCmpxchgOrRmw(const SIMemOpInfo &MOI, 426 MachineBasicBlock::iterator &MI); 427 428 public: 429 static char ID; 430 431 SIMemoryLegalizer() : MachineFunctionPass(ID) {} 432 433 void getAnalysisUsage(AnalysisUsage &AU) const override { 434 AU.setPreservesCFG(); 435 MachineFunctionPass::getAnalysisUsage(AU); 436 } 437 438 StringRef getPassName() const override { 439 return PASS_NAME; 440 } 441 442 bool runOnMachineFunction(MachineFunction &MF) override; 443 }; 444 445 } // end namespace anonymous 446 447 void SIMemOpAccess::reportUnsupported(const MachineBasicBlock::iterator &MI, 448 const char *Msg) const { 449 const Function &Func = MI->getParent()->getParent()->getFunction(); 450 DiagnosticInfoUnsupported Diag(Func, Msg, MI->getDebugLoc()); 451 Func.getContext().diagnose(Diag); 452 } 453 454 Optional<std::tuple<SIAtomicScope, SIAtomicAddrSpace, bool>> 455 SIMemOpAccess::toSIAtomicScope(SyncScope::ID SSID, 456 SIAtomicAddrSpace InstrScope) const { 457 if (SSID == SyncScope::System) 458 return std::make_tuple(SIAtomicScope::SYSTEM, 459 SIAtomicAddrSpace::ATOMIC, 460 true); 461 if (SSID == MMI->getAgentSSID()) 462 return std::make_tuple(SIAtomicScope::AGENT, 463 SIAtomicAddrSpace::ATOMIC, 464 true); 465 if (SSID == MMI->getWorkgroupSSID()) 466 return std::make_tuple(SIAtomicScope::WORKGROUP, 467 SIAtomicAddrSpace::ATOMIC, 468 true); 469 if (SSID == MMI->getWavefrontSSID()) 470 return std::make_tuple(SIAtomicScope::WAVEFRONT, 471 SIAtomicAddrSpace::ATOMIC, 472 true); 473 if (SSID == SyncScope::SingleThread) 474 return std::make_tuple(SIAtomicScope::SINGLETHREAD, 475 SIAtomicAddrSpace::ATOMIC, 476 true); 477 if (SSID == MMI->getSystemOneAddressSpaceSSID()) 478 return std::make_tuple(SIAtomicScope::SYSTEM, 479 SIAtomicAddrSpace::ATOMIC & InstrScope, 480 false); 481 if (SSID == MMI->getAgentOneAddressSpaceSSID()) 482 return std::make_tuple(SIAtomicScope::AGENT, 483 SIAtomicAddrSpace::ATOMIC & InstrScope, 484 false); 485 if (SSID == MMI->getWorkgroupOneAddressSpaceSSID()) 486 return std::make_tuple(SIAtomicScope::WORKGROUP, 487 SIAtomicAddrSpace::ATOMIC & InstrScope, 488 false); 489 if (SSID == MMI->getWavefrontOneAddressSpaceSSID()) 490 return std::make_tuple(SIAtomicScope::WAVEFRONT, 491 SIAtomicAddrSpace::ATOMIC & InstrScope, 492 false); 493 if (SSID == MMI->getSingleThreadOneAddressSpaceSSID()) 494 return std::make_tuple(SIAtomicScope::SINGLETHREAD, 495 SIAtomicAddrSpace::ATOMIC & InstrScope, 496 false); 497 return None; 498 } 499 500 SIAtomicAddrSpace SIMemOpAccess::toSIAtomicAddrSpace(unsigned AS) const { 501 if (AS == AMDGPUAS::FLAT_ADDRESS) 502 return SIAtomicAddrSpace::FLAT; 503 if (AS == AMDGPUAS::GLOBAL_ADDRESS) 504 return SIAtomicAddrSpace::GLOBAL; 505 if (AS == AMDGPUAS::LOCAL_ADDRESS) 506 return SIAtomicAddrSpace::LDS; 507 if (AS == AMDGPUAS::PRIVATE_ADDRESS) 508 return SIAtomicAddrSpace::SCRATCH; 509 if (AS == AMDGPUAS::REGION_ADDRESS) 510 return SIAtomicAddrSpace::GDS; 511 512 return SIAtomicAddrSpace::OTHER; 513 } 514 515 SIMemOpAccess::SIMemOpAccess(MachineFunction &MF) { 516 MMI = &MF.getMMI().getObjFileInfo<AMDGPUMachineModuleInfo>(); 517 } 518 519 Optional<SIMemOpInfo> SIMemOpAccess::constructFromMIWithMMO( 520 const MachineBasicBlock::iterator &MI) const { 521 assert(MI->getNumMemOperands() > 0); 522 523 SyncScope::ID SSID = SyncScope::SingleThread; 524 AtomicOrdering Ordering = AtomicOrdering::NotAtomic; 525 AtomicOrdering FailureOrdering = AtomicOrdering::NotAtomic; 526 SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::NONE; 527 bool IsNonTemporal = true; 528 529 // Validator should check whether or not MMOs cover the entire set of 530 // locations accessed by the memory instruction. 531 for (const auto &MMO : MI->memoperands()) { 532 IsNonTemporal &= MMO->isNonTemporal(); 533 InstrAddrSpace |= 534 toSIAtomicAddrSpace(MMO->getPointerInfo().getAddrSpace()); 535 AtomicOrdering OpOrdering = MMO->getOrdering(); 536 if (OpOrdering != AtomicOrdering::NotAtomic) { 537 const auto &IsSyncScopeInclusion = 538 MMI->isSyncScopeInclusion(SSID, MMO->getSyncScopeID()); 539 if (!IsSyncScopeInclusion) { 540 reportUnsupported(MI, 541 "Unsupported non-inclusive atomic synchronization scope"); 542 return None; 543 } 544 545 SSID = IsSyncScopeInclusion.getValue() ? SSID : MMO->getSyncScopeID(); 546 Ordering = 547 isStrongerThan(Ordering, OpOrdering) ? 548 Ordering : MMO->getOrdering(); 549 assert(MMO->getFailureOrdering() != AtomicOrdering::Release && 550 MMO->getFailureOrdering() != AtomicOrdering::AcquireRelease); 551 FailureOrdering = 552 isStrongerThan(FailureOrdering, MMO->getFailureOrdering()) ? 553 FailureOrdering : MMO->getFailureOrdering(); 554 } 555 } 556 557 SIAtomicScope Scope = SIAtomicScope::NONE; 558 SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE; 559 bool IsCrossAddressSpaceOrdering = false; 560 if (Ordering != AtomicOrdering::NotAtomic) { 561 auto ScopeOrNone = toSIAtomicScope(SSID, InstrAddrSpace); 562 if (!ScopeOrNone) { 563 reportUnsupported(MI, "Unsupported atomic synchronization scope"); 564 return None; 565 } 566 std::tie(Scope, OrderingAddrSpace, IsCrossAddressSpaceOrdering) = 567 ScopeOrNone.getValue(); 568 if ((OrderingAddrSpace == SIAtomicAddrSpace::NONE) || 569 ((OrderingAddrSpace & SIAtomicAddrSpace::ATOMIC) != OrderingAddrSpace)) { 570 reportUnsupported(MI, "Unsupported atomic address space"); 571 return None; 572 } 573 } 574 return SIMemOpInfo(Ordering, Scope, OrderingAddrSpace, InstrAddrSpace, 575 IsCrossAddressSpaceOrdering, FailureOrdering, IsNonTemporal); 576 } 577 578 Optional<SIMemOpInfo> SIMemOpAccess::getLoadInfo( 579 const MachineBasicBlock::iterator &MI) const { 580 assert(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic); 581 582 if (!(MI->mayLoad() && !MI->mayStore())) 583 return None; 584 585 // Be conservative if there are no memory operands. 586 if (MI->getNumMemOperands() == 0) 587 return SIMemOpInfo(); 588 589 return constructFromMIWithMMO(MI); 590 } 591 592 Optional<SIMemOpInfo> SIMemOpAccess::getStoreInfo( 593 const MachineBasicBlock::iterator &MI) const { 594 assert(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic); 595 596 if (!(!MI->mayLoad() && MI->mayStore())) 597 return None; 598 599 // Be conservative if there are no memory operands. 600 if (MI->getNumMemOperands() == 0) 601 return SIMemOpInfo(); 602 603 return constructFromMIWithMMO(MI); 604 } 605 606 Optional<SIMemOpInfo> SIMemOpAccess::getAtomicFenceInfo( 607 const MachineBasicBlock::iterator &MI) const { 608 assert(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic); 609 610 if (MI->getOpcode() != AMDGPU::ATOMIC_FENCE) 611 return None; 612 613 AtomicOrdering Ordering = 614 static_cast<AtomicOrdering>(MI->getOperand(0).getImm()); 615 616 SyncScope::ID SSID = static_cast<SyncScope::ID>(MI->getOperand(1).getImm()); 617 auto ScopeOrNone = toSIAtomicScope(SSID, SIAtomicAddrSpace::ATOMIC); 618 if (!ScopeOrNone) { 619 reportUnsupported(MI, "Unsupported atomic synchronization scope"); 620 return None; 621 } 622 623 SIAtomicScope Scope = SIAtomicScope::NONE; 624 SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE; 625 bool IsCrossAddressSpaceOrdering = false; 626 std::tie(Scope, OrderingAddrSpace, IsCrossAddressSpaceOrdering) = 627 ScopeOrNone.getValue(); 628 629 if ((OrderingAddrSpace == SIAtomicAddrSpace::NONE) || 630 ((OrderingAddrSpace & SIAtomicAddrSpace::ATOMIC) != OrderingAddrSpace)) { 631 reportUnsupported(MI, "Unsupported atomic address space"); 632 return None; 633 } 634 635 return SIMemOpInfo(Ordering, Scope, OrderingAddrSpace, SIAtomicAddrSpace::ATOMIC, 636 IsCrossAddressSpaceOrdering); 637 } 638 639 Optional<SIMemOpInfo> SIMemOpAccess::getAtomicCmpxchgOrRmwInfo( 640 const MachineBasicBlock::iterator &MI) const { 641 assert(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic); 642 643 if (!(MI->mayLoad() && MI->mayStore())) 644 return None; 645 646 // Be conservative if there are no memory operands. 647 if (MI->getNumMemOperands() == 0) 648 return SIMemOpInfo(); 649 650 return constructFromMIWithMMO(MI); 651 } 652 653 SICacheControl::SICacheControl(const GCNSubtarget &ST) { 654 TII = ST.getInstrInfo(); 655 IV = getIsaVersion(ST.getCPU()); 656 InsertCacheInv = !ST.isAmdPalOS(); 657 } 658 659 /* static */ 660 std::unique_ptr<SICacheControl> SICacheControl::create(const GCNSubtarget &ST) { 661 GCNSubtarget::Generation Generation = ST.getGeneration(); 662 if (Generation <= AMDGPUSubtarget::SOUTHERN_ISLANDS) 663 return std::make_unique<SIGfx6CacheControl>(ST); 664 if (Generation < AMDGPUSubtarget::GFX10) 665 return std::make_unique<SIGfx7CacheControl>(ST); 666 return std::make_unique<SIGfx10CacheControl>(ST, ST.isCuModeEnabled()); 667 } 668 669 bool SIGfx6CacheControl::enableLoadCacheBypass( 670 const MachineBasicBlock::iterator &MI, 671 SIAtomicScope Scope, 672 SIAtomicAddrSpace AddrSpace) const { 673 assert(MI->mayLoad() && !MI->mayStore()); 674 bool Changed = false; 675 676 if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { 677 /// TODO: Do not set glc for rmw atomic operations as they 678 /// implicitly bypass the L1 cache. 679 680 switch (Scope) { 681 case SIAtomicScope::SYSTEM: 682 case SIAtomicScope::AGENT: 683 Changed |= enableGLCBit(MI); 684 break; 685 case SIAtomicScope::WORKGROUP: 686 case SIAtomicScope::WAVEFRONT: 687 case SIAtomicScope::SINGLETHREAD: 688 // No cache to bypass. 689 break; 690 default: 691 llvm_unreachable("Unsupported synchronization scope"); 692 } 693 } 694 695 /// The scratch address space does not need the global memory caches 696 /// to be bypassed as all memory operations by the same thread are 697 /// sequentially consistent, and no other thread can access scratch 698 /// memory. 699 700 /// Other address spaces do not hava a cache. 701 702 return Changed; 703 } 704 705 bool SIGfx6CacheControl::enableNonTemporal( 706 const MachineBasicBlock::iterator &MI) const { 707 assert(MI->mayLoad() ^ MI->mayStore()); 708 bool Changed = false; 709 710 /// TODO: Do not enableGLCBit if rmw atomic. 711 Changed |= enableGLCBit(MI); 712 Changed |= enableSLCBit(MI); 713 714 return Changed; 715 } 716 717 bool SIGfx6CacheControl::insertCacheInvalidate(MachineBasicBlock::iterator &MI, 718 SIAtomicScope Scope, 719 SIAtomicAddrSpace AddrSpace, 720 Position Pos) const { 721 if (!InsertCacheInv) 722 return false; 723 724 bool Changed = false; 725 726 MachineBasicBlock &MBB = *MI->getParent(); 727 DebugLoc DL = MI->getDebugLoc(); 728 729 if (Pos == Position::AFTER) 730 ++MI; 731 732 if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { 733 switch (Scope) { 734 case SIAtomicScope::SYSTEM: 735 case SIAtomicScope::AGENT: 736 BuildMI(MBB, MI, DL, TII->get(AMDGPU::BUFFER_WBINVL1)); 737 Changed = true; 738 break; 739 case SIAtomicScope::WORKGROUP: 740 case SIAtomicScope::WAVEFRONT: 741 case SIAtomicScope::SINGLETHREAD: 742 // No cache to invalidate. 743 break; 744 default: 745 llvm_unreachable("Unsupported synchronization scope"); 746 } 747 } 748 749 /// The scratch address space does not need the global memory cache 750 /// to be flushed as all memory operations by the same thread are 751 /// sequentially consistent, and no other thread can access scratch 752 /// memory. 753 754 /// Other address spaces do not hava a cache. 755 756 if (Pos == Position::AFTER) 757 --MI; 758 759 return Changed; 760 } 761 762 bool SIGfx6CacheControl::insertWait(MachineBasicBlock::iterator &MI, 763 SIAtomicScope Scope, 764 SIAtomicAddrSpace AddrSpace, 765 SIMemOp Op, 766 bool IsCrossAddrSpaceOrdering, 767 Position Pos) const { 768 bool Changed = false; 769 770 MachineBasicBlock &MBB = *MI->getParent(); 771 DebugLoc DL = MI->getDebugLoc(); 772 773 if (Pos == Position::AFTER) 774 ++MI; 775 776 bool VMCnt = false; 777 bool LGKMCnt = false; 778 779 if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { 780 switch (Scope) { 781 case SIAtomicScope::SYSTEM: 782 case SIAtomicScope::AGENT: 783 VMCnt |= true; 784 break; 785 case SIAtomicScope::WORKGROUP: 786 case SIAtomicScope::WAVEFRONT: 787 case SIAtomicScope::SINGLETHREAD: 788 // The L1 cache keeps all memory operations in order for 789 // wavefronts in the same work-group. 790 break; 791 default: 792 llvm_unreachable("Unsupported synchronization scope"); 793 } 794 } 795 796 if ((AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) { 797 switch (Scope) { 798 case SIAtomicScope::SYSTEM: 799 case SIAtomicScope::AGENT: 800 case SIAtomicScope::WORKGROUP: 801 // If no cross address space ordering then an LDS waitcnt is not 802 // needed as LDS operations for all waves are executed in a 803 // total global ordering as observed by all waves. Required if 804 // also synchronizing with global/GDS memory as LDS operations 805 // could be reordered with respect to later global/GDS memory 806 // operations of the same wave. 807 LGKMCnt |= IsCrossAddrSpaceOrdering; 808 break; 809 case SIAtomicScope::WAVEFRONT: 810 case SIAtomicScope::SINGLETHREAD: 811 // The LDS keeps all memory operations in order for 812 // the same wavesfront. 813 break; 814 default: 815 llvm_unreachable("Unsupported synchronization scope"); 816 } 817 } 818 819 if ((AddrSpace & SIAtomicAddrSpace::GDS) != SIAtomicAddrSpace::NONE) { 820 switch (Scope) { 821 case SIAtomicScope::SYSTEM: 822 case SIAtomicScope::AGENT: 823 // If no cross address space ordering then an GDS waitcnt is not 824 // needed as GDS operations for all waves are executed in a 825 // total global ordering as observed by all waves. Required if 826 // also synchronizing with global/LDS memory as GDS operations 827 // could be reordered with respect to later global/LDS memory 828 // operations of the same wave. 829 LGKMCnt |= IsCrossAddrSpaceOrdering; 830 break; 831 case SIAtomicScope::WORKGROUP: 832 case SIAtomicScope::WAVEFRONT: 833 case SIAtomicScope::SINGLETHREAD: 834 // The GDS keeps all memory operations in order for 835 // the same work-group. 836 break; 837 default: 838 llvm_unreachable("Unsupported synchronization scope"); 839 } 840 } 841 842 if (VMCnt || LGKMCnt) { 843 unsigned WaitCntImmediate = 844 AMDGPU::encodeWaitcnt(IV, 845 VMCnt ? 0 : getVmcntBitMask(IV), 846 getExpcntBitMask(IV), 847 LGKMCnt ? 0 : getLgkmcntBitMask(IV)); 848 BuildMI(MBB, MI, DL, TII->get(AMDGPU::S_WAITCNT)).addImm(WaitCntImmediate); 849 Changed = true; 850 } 851 852 if (Pos == Position::AFTER) 853 --MI; 854 855 return Changed; 856 } 857 858 bool SIGfx7CacheControl::insertCacheInvalidate(MachineBasicBlock::iterator &MI, 859 SIAtomicScope Scope, 860 SIAtomicAddrSpace AddrSpace, 861 Position Pos) const { 862 if (!InsertCacheInv) 863 return false; 864 865 bool Changed = false; 866 867 MachineBasicBlock &MBB = *MI->getParent(); 868 DebugLoc DL = MI->getDebugLoc(); 869 870 const GCNSubtarget &STM = MBB.getParent()->getSubtarget<GCNSubtarget>(); 871 872 const unsigned Flush = STM.isAmdPalOS() || STM.isMesa3DOS() 873 ? AMDGPU::BUFFER_WBINVL1 874 : AMDGPU::BUFFER_WBINVL1_VOL; 875 876 if (Pos == Position::AFTER) 877 ++MI; 878 879 if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { 880 switch (Scope) { 881 case SIAtomicScope::SYSTEM: 882 case SIAtomicScope::AGENT: 883 BuildMI(MBB, MI, DL, TII->get(Flush)); 884 Changed = true; 885 break; 886 case SIAtomicScope::WORKGROUP: 887 case SIAtomicScope::WAVEFRONT: 888 case SIAtomicScope::SINGLETHREAD: 889 // No cache to invalidate. 890 break; 891 default: 892 llvm_unreachable("Unsupported synchronization scope"); 893 } 894 } 895 896 /// The scratch address space does not need the global memory cache 897 /// to be flushed as all memory operations by the same thread are 898 /// sequentially consistent, and no other thread can access scratch 899 /// memory. 900 901 /// Other address spaces do not hava a cache. 902 903 if (Pos == Position::AFTER) 904 --MI; 905 906 return Changed; 907 } 908 909 bool SIGfx10CacheControl::enableLoadCacheBypass( 910 const MachineBasicBlock::iterator &MI, 911 SIAtomicScope Scope, 912 SIAtomicAddrSpace AddrSpace) const { 913 assert(MI->mayLoad() && !MI->mayStore()); 914 bool Changed = false; 915 916 if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { 917 /// TODO Do not set glc for rmw atomic operations as they 918 /// implicitly bypass the L0/L1 caches. 919 920 switch (Scope) { 921 case SIAtomicScope::SYSTEM: 922 case SIAtomicScope::AGENT: 923 Changed |= enableGLCBit(MI); 924 Changed |= enableDLCBit(MI); 925 break; 926 case SIAtomicScope::WORKGROUP: 927 // In WGP mode the waves of a work-group can be executing on either CU of 928 // the WGP. Therefore need to bypass the L0 which is per CU. Otherwise in 929 // CU mode and all waves of a work-group are on the same CU, and so the 930 // L0 does not need to be bypassed. 931 if (!CuMode) Changed |= enableGLCBit(MI); 932 break; 933 case SIAtomicScope::WAVEFRONT: 934 case SIAtomicScope::SINGLETHREAD: 935 // No cache to bypass. 936 break; 937 default: 938 llvm_unreachable("Unsupported synchronization scope"); 939 } 940 } 941 942 /// The scratch address space does not need the global memory caches 943 /// to be bypassed as all memory operations by the same thread are 944 /// sequentially consistent, and no other thread can access scratch 945 /// memory. 946 947 /// Other address spaces do not hava a cache. 948 949 return Changed; 950 } 951 952 bool SIGfx10CacheControl::enableNonTemporal( 953 const MachineBasicBlock::iterator &MI) const { 954 assert(MI->mayLoad() ^ MI->mayStore()); 955 bool Changed = false; 956 957 Changed |= enableSLCBit(MI); 958 /// TODO for store (non-rmw atomic) instructions also enableGLCBit(MI) 959 960 return Changed; 961 } 962 963 bool SIGfx10CacheControl::insertCacheInvalidate(MachineBasicBlock::iterator &MI, 964 SIAtomicScope Scope, 965 SIAtomicAddrSpace AddrSpace, 966 Position Pos) const { 967 if (!InsertCacheInv) 968 return false; 969 970 bool Changed = false; 971 972 MachineBasicBlock &MBB = *MI->getParent(); 973 DebugLoc DL = MI->getDebugLoc(); 974 975 if (Pos == Position::AFTER) 976 ++MI; 977 978 if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { 979 switch (Scope) { 980 case SIAtomicScope::SYSTEM: 981 case SIAtomicScope::AGENT: 982 BuildMI(MBB, MI, DL, TII->get(AMDGPU::BUFFER_GL0_INV)); 983 BuildMI(MBB, MI, DL, TII->get(AMDGPU::BUFFER_GL1_INV)); 984 Changed = true; 985 break; 986 case SIAtomicScope::WORKGROUP: 987 // In WGP mode the waves of a work-group can be executing on either CU of 988 // the WGP. Therefore need to invalidate the L0 which is per CU. Otherwise 989 // in CU mode and all waves of a work-group are on the same CU, and so the 990 // L0 does not need to be invalidated. 991 if (!CuMode) { 992 BuildMI(MBB, MI, DL, TII->get(AMDGPU::BUFFER_GL0_INV)); 993 Changed = true; 994 } 995 break; 996 case SIAtomicScope::WAVEFRONT: 997 case SIAtomicScope::SINGLETHREAD: 998 // No cache to invalidate. 999 break; 1000 default: 1001 llvm_unreachable("Unsupported synchronization scope"); 1002 } 1003 } 1004 1005 /// The scratch address space does not need the global memory cache 1006 /// to be flushed as all memory operations by the same thread are 1007 /// sequentially consistent, and no other thread can access scratch 1008 /// memory. 1009 1010 /// Other address spaces do not hava a cache. 1011 1012 if (Pos == Position::AFTER) 1013 --MI; 1014 1015 return Changed; 1016 } 1017 1018 bool SIGfx10CacheControl::insertWait(MachineBasicBlock::iterator &MI, 1019 SIAtomicScope Scope, 1020 SIAtomicAddrSpace AddrSpace, 1021 SIMemOp Op, 1022 bool IsCrossAddrSpaceOrdering, 1023 Position Pos) const { 1024 bool Changed = false; 1025 1026 MachineBasicBlock &MBB = *MI->getParent(); 1027 DebugLoc DL = MI->getDebugLoc(); 1028 1029 if (Pos == Position::AFTER) 1030 ++MI; 1031 1032 bool VMCnt = false; 1033 bool VSCnt = false; 1034 bool LGKMCnt = false; 1035 1036 if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { 1037 switch (Scope) { 1038 case SIAtomicScope::SYSTEM: 1039 case SIAtomicScope::AGENT: 1040 if ((Op & SIMemOp::LOAD) != SIMemOp::NONE) 1041 VMCnt |= true; 1042 if ((Op & SIMemOp::STORE) != SIMemOp::NONE) 1043 VSCnt |= true; 1044 break; 1045 case SIAtomicScope::WORKGROUP: 1046 // In WGP mode the waves of a work-group can be executing on either CU of 1047 // the WGP. Therefore need to wait for operations to complete to ensure 1048 // they are visible to waves in the other CU as the L0 is per CU. 1049 // Otherwise in CU mode and all waves of a work-group are on the same CU 1050 // which shares the same L0. 1051 if (!CuMode) { 1052 if ((Op & SIMemOp::LOAD) != SIMemOp::NONE) 1053 VMCnt |= true; 1054 if ((Op & SIMemOp::STORE) != SIMemOp::NONE) 1055 VSCnt |= true; 1056 } 1057 break; 1058 case SIAtomicScope::WAVEFRONT: 1059 case SIAtomicScope::SINGLETHREAD: 1060 // The L0 cache keeps all memory operations in order for 1061 // work-items in the same wavefront. 1062 break; 1063 default: 1064 llvm_unreachable("Unsupported synchronization scope"); 1065 } 1066 } 1067 1068 if ((AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) { 1069 switch (Scope) { 1070 case SIAtomicScope::SYSTEM: 1071 case SIAtomicScope::AGENT: 1072 case SIAtomicScope::WORKGROUP: 1073 // If no cross address space ordering then an LDS waitcnt is not 1074 // needed as LDS operations for all waves are executed in a 1075 // total global ordering as observed by all waves. Required if 1076 // also synchronizing with global/GDS memory as LDS operations 1077 // could be reordered with respect to later global/GDS memory 1078 // operations of the same wave. 1079 LGKMCnt |= IsCrossAddrSpaceOrdering; 1080 break; 1081 case SIAtomicScope::WAVEFRONT: 1082 case SIAtomicScope::SINGLETHREAD: 1083 // The LDS keeps all memory operations in order for 1084 // the same wavesfront. 1085 break; 1086 default: 1087 llvm_unreachable("Unsupported synchronization scope"); 1088 } 1089 } 1090 1091 if ((AddrSpace & SIAtomicAddrSpace::GDS) != SIAtomicAddrSpace::NONE) { 1092 switch (Scope) { 1093 case SIAtomicScope::SYSTEM: 1094 case SIAtomicScope::AGENT: 1095 // If no cross address space ordering then an GDS waitcnt is not 1096 // needed as GDS operations for all waves are executed in a 1097 // total global ordering as observed by all waves. Required if 1098 // also synchronizing with global/LDS memory as GDS operations 1099 // could be reordered with respect to later global/LDS memory 1100 // operations of the same wave. 1101 LGKMCnt |= IsCrossAddrSpaceOrdering; 1102 break; 1103 case SIAtomicScope::WORKGROUP: 1104 case SIAtomicScope::WAVEFRONT: 1105 case SIAtomicScope::SINGLETHREAD: 1106 // The GDS keeps all memory operations in order for 1107 // the same work-group. 1108 break; 1109 default: 1110 llvm_unreachable("Unsupported synchronization scope"); 1111 } 1112 } 1113 1114 if (VMCnt || LGKMCnt) { 1115 unsigned WaitCntImmediate = 1116 AMDGPU::encodeWaitcnt(IV, 1117 VMCnt ? 0 : getVmcntBitMask(IV), 1118 getExpcntBitMask(IV), 1119 LGKMCnt ? 0 : getLgkmcntBitMask(IV)); 1120 BuildMI(MBB, MI, DL, TII->get(AMDGPU::S_WAITCNT)).addImm(WaitCntImmediate); 1121 Changed = true; 1122 } 1123 1124 if (VSCnt) { 1125 BuildMI(MBB, MI, DL, TII->get(AMDGPU::S_WAITCNT_VSCNT)) 1126 .addReg(AMDGPU::SGPR_NULL, RegState::Undef) 1127 .addImm(0); 1128 Changed = true; 1129 } 1130 1131 if (Pos == Position::AFTER) 1132 --MI; 1133 1134 return Changed; 1135 } 1136 1137 bool SIMemoryLegalizer::removeAtomicPseudoMIs() { 1138 if (AtomicPseudoMIs.empty()) 1139 return false; 1140 1141 for (auto &MI : AtomicPseudoMIs) 1142 MI->eraseFromParent(); 1143 1144 AtomicPseudoMIs.clear(); 1145 return true; 1146 } 1147 1148 bool SIMemoryLegalizer::expandLoad(const SIMemOpInfo &MOI, 1149 MachineBasicBlock::iterator &MI) { 1150 assert(MI->mayLoad() && !MI->mayStore()); 1151 1152 bool Changed = false; 1153 1154 if (MOI.isAtomic()) { 1155 if (MOI.getOrdering() == AtomicOrdering::Monotonic || 1156 MOI.getOrdering() == AtomicOrdering::Acquire || 1157 MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) { 1158 Changed |= CC->enableLoadCacheBypass(MI, MOI.getScope(), 1159 MOI.getOrderingAddrSpace()); 1160 } 1161 1162 if (MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) 1163 Changed |= CC->insertWait(MI, MOI.getScope(), 1164 MOI.getOrderingAddrSpace(), 1165 SIMemOp::LOAD | SIMemOp::STORE, 1166 MOI.getIsCrossAddressSpaceOrdering(), 1167 Position::BEFORE); 1168 1169 if (MOI.getOrdering() == AtomicOrdering::Acquire || 1170 MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) { 1171 Changed |= CC->insertWait(MI, MOI.getScope(), 1172 MOI.getInstrAddrSpace(), 1173 SIMemOp::LOAD, 1174 MOI.getIsCrossAddressSpaceOrdering(), 1175 Position::AFTER); 1176 Changed |= CC->insertCacheInvalidate(MI, MOI.getScope(), 1177 MOI.getOrderingAddrSpace(), 1178 Position::AFTER); 1179 } 1180 1181 return Changed; 1182 } 1183 1184 // Atomic instructions do not have the nontemporal attribute. 1185 if (MOI.isNonTemporal()) { 1186 Changed |= CC->enableNonTemporal(MI); 1187 return Changed; 1188 } 1189 1190 return Changed; 1191 } 1192 1193 bool SIMemoryLegalizer::expandStore(const SIMemOpInfo &MOI, 1194 MachineBasicBlock::iterator &MI) { 1195 assert(!MI->mayLoad() && MI->mayStore()); 1196 1197 bool Changed = false; 1198 1199 if (MOI.isAtomic()) { 1200 if (MOI.getOrdering() == AtomicOrdering::Release || 1201 MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) 1202 Changed |= CC->insertWait(MI, MOI.getScope(), 1203 MOI.getOrderingAddrSpace(), 1204 SIMemOp::LOAD | SIMemOp::STORE, 1205 MOI.getIsCrossAddressSpaceOrdering(), 1206 Position::BEFORE); 1207 1208 return Changed; 1209 } 1210 1211 // Atomic instructions do not have the nontemporal attribute. 1212 if (MOI.isNonTemporal()) { 1213 Changed |= CC->enableNonTemporal(MI); 1214 return Changed; 1215 } 1216 1217 return Changed; 1218 } 1219 1220 bool SIMemoryLegalizer::expandAtomicFence(const SIMemOpInfo &MOI, 1221 MachineBasicBlock::iterator &MI) { 1222 assert(MI->getOpcode() == AMDGPU::ATOMIC_FENCE); 1223 1224 AtomicPseudoMIs.push_back(MI); 1225 bool Changed = false; 1226 1227 if (MOI.isAtomic()) { 1228 if (MOI.getOrdering() == AtomicOrdering::Acquire || 1229 MOI.getOrdering() == AtomicOrdering::Release || 1230 MOI.getOrdering() == AtomicOrdering::AcquireRelease || 1231 MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) 1232 /// TODO: This relies on a barrier always generating a waitcnt 1233 /// for LDS to ensure it is not reordered with the completion of 1234 /// the proceeding LDS operations. If barrier had a memory 1235 /// ordering and memory scope, then library does not need to 1236 /// generate a fence. Could add support in this file for 1237 /// barrier. SIInsertWaitcnt.cpp could then stop unconditionally 1238 /// adding waitcnt before a S_BARRIER. 1239 Changed |= CC->insertWait(MI, MOI.getScope(), 1240 MOI.getOrderingAddrSpace(), 1241 SIMemOp::LOAD | SIMemOp::STORE, 1242 MOI.getIsCrossAddressSpaceOrdering(), 1243 Position::BEFORE); 1244 1245 if (MOI.getOrdering() == AtomicOrdering::Acquire || 1246 MOI.getOrdering() == AtomicOrdering::AcquireRelease || 1247 MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) 1248 Changed |= CC->insertCacheInvalidate(MI, MOI.getScope(), 1249 MOI.getOrderingAddrSpace(), 1250 Position::BEFORE); 1251 1252 return Changed; 1253 } 1254 1255 return Changed; 1256 } 1257 1258 bool SIMemoryLegalizer::expandAtomicCmpxchgOrRmw(const SIMemOpInfo &MOI, 1259 MachineBasicBlock::iterator &MI) { 1260 assert(MI->mayLoad() && MI->mayStore()); 1261 1262 bool Changed = false; 1263 1264 if (MOI.isAtomic()) { 1265 if (MOI.getOrdering() == AtomicOrdering::Release || 1266 MOI.getOrdering() == AtomicOrdering::AcquireRelease || 1267 MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent || 1268 MOI.getFailureOrdering() == AtomicOrdering::SequentiallyConsistent) 1269 Changed |= CC->insertWait(MI, MOI.getScope(), 1270 MOI.getOrderingAddrSpace(), 1271 SIMemOp::LOAD | SIMemOp::STORE, 1272 MOI.getIsCrossAddressSpaceOrdering(), 1273 Position::BEFORE); 1274 1275 if (MOI.getOrdering() == AtomicOrdering::Acquire || 1276 MOI.getOrdering() == AtomicOrdering::AcquireRelease || 1277 MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent || 1278 MOI.getFailureOrdering() == AtomicOrdering::Acquire || 1279 MOI.getFailureOrdering() == AtomicOrdering::SequentiallyConsistent) { 1280 Changed |= CC->insertWait(MI, MOI.getScope(), 1281 MOI.getOrderingAddrSpace(), 1282 isAtomicRet(*MI) ? SIMemOp::LOAD : 1283 SIMemOp::STORE, 1284 MOI.getIsCrossAddressSpaceOrdering(), 1285 Position::AFTER); 1286 Changed |= CC->insertCacheInvalidate(MI, MOI.getScope(), 1287 MOI.getOrderingAddrSpace(), 1288 Position::AFTER); 1289 } 1290 1291 return Changed; 1292 } 1293 1294 return Changed; 1295 } 1296 1297 bool SIMemoryLegalizer::runOnMachineFunction(MachineFunction &MF) { 1298 bool Changed = false; 1299 1300 SIMemOpAccess MOA(MF); 1301 CC = SICacheControl::create(MF.getSubtarget<GCNSubtarget>()); 1302 1303 for (auto &MBB : MF) { 1304 for (auto MI = MBB.begin(); MI != MBB.end(); ++MI) { 1305 1306 if (MI->getOpcode() == TargetOpcode::BUNDLE && MI->mayLoadOrStore()) { 1307 MachineBasicBlock::instr_iterator II(MI->getIterator()); 1308 for (MachineBasicBlock::instr_iterator I = ++II, E = MBB.instr_end(); 1309 I != E && I->isBundledWithPred(); ++I) { 1310 I->unbundleFromPred(); 1311 for (MachineOperand &MO : I->operands()) 1312 if (MO.isReg()) 1313 MO.setIsInternalRead(false); 1314 } 1315 1316 MI->eraseFromParent(); 1317 MI = II->getIterator(); 1318 } 1319 1320 if (!(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic)) 1321 continue; 1322 1323 if (const auto &MOI = MOA.getLoadInfo(MI)) 1324 Changed |= expandLoad(MOI.getValue(), MI); 1325 else if (const auto &MOI = MOA.getStoreInfo(MI)) 1326 Changed |= expandStore(MOI.getValue(), MI); 1327 else if (const auto &MOI = MOA.getAtomicFenceInfo(MI)) 1328 Changed |= expandAtomicFence(MOI.getValue(), MI); 1329 else if (const auto &MOI = MOA.getAtomicCmpxchgOrRmwInfo(MI)) 1330 Changed |= expandAtomicCmpxchgOrRmw(MOI.getValue(), MI); 1331 } 1332 } 1333 1334 Changed |= removeAtomicPseudoMIs(); 1335 return Changed; 1336 } 1337 1338 INITIALIZE_PASS(SIMemoryLegalizer, DEBUG_TYPE, PASS_NAME, false, false) 1339 1340 char SIMemoryLegalizer::ID = 0; 1341 char &llvm::SIMemoryLegalizerID = SIMemoryLegalizer::ID; 1342 1343 FunctionPass *llvm::createSIMemoryLegalizerPass() { 1344 return new SIMemoryLegalizer(); 1345 } 1346