1 //===- MachineCopyPropagation.cpp - Machine Copy Propagation Pass ---------===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 //
9 // This is an extremely simple MachineInstr-level copy propagation pass.
10 //
11 // This pass forwards the source of COPYs to the users of their destinations
12 // when doing so is legal. For example:
13 //
14 // %reg1 = COPY %reg0
15 // ...
16 // ... = OP %reg1
17 //
18 // If
19 // - %reg0 has not been clobbered by the time of the use of %reg1
20 // - the register class constraints are satisfied
21 // - the COPY def is the only value that reaches OP
22 // then this pass replaces the above with:
23 //
24 // %reg1 = COPY %reg0
25 // ...
26 // ... = OP %reg0
27 //
28 // This pass also removes some redundant COPYs. For example:
29 //
30 // %R1 = COPY %R0
31 // ... // No clobber of %R1
32 // %R0 = COPY %R1 <<< Removed
33 //
34 // or
35 //
36 // %R1 = COPY %R0
37 // ... // No clobber of %R0
38 // %R1 = COPY %R0 <<< Removed
39 //
40 // or
41 //
42 // $R0 = OP ...
43 // ... // No read/clobber of $R0 and $R1
44 // $R1 = COPY $R0 // $R0 is killed
45 // Replace $R0 with $R1 and remove the COPY
46 // $R1 = OP ...
47 // ...
48 //
49 //===----------------------------------------------------------------------===//
50
51 #include "llvm/ADT/DenseMap.h"
52 #include "llvm/ADT/STLExtras.h"
53 #include "llvm/ADT/SetVector.h"
54 #include "llvm/ADT/SmallSet.h"
55 #include "llvm/ADT/SmallVector.h"
56 #include "llvm/ADT/Statistic.h"
57 #include "llvm/ADT/iterator_range.h"
58 #include "llvm/CodeGen/MachineBasicBlock.h"
59 #include "llvm/CodeGen/MachineFunction.h"
60 #include "llvm/CodeGen/MachineFunctionPass.h"
61 #include "llvm/CodeGen/MachineInstr.h"
62 #include "llvm/CodeGen/MachineOperand.h"
63 #include "llvm/CodeGen/MachineRegisterInfo.h"
64 #include "llvm/CodeGen/TargetInstrInfo.h"
65 #include "llvm/CodeGen/TargetRegisterInfo.h"
66 #include "llvm/CodeGen/TargetSubtargetInfo.h"
67 #include "llvm/InitializePasses.h"
68 #include "llvm/MC/MCRegister.h"
69 #include "llvm/MC/MCRegisterInfo.h"
70 #include "llvm/Pass.h"
71 #include "llvm/Support/Debug.h"
72 #include "llvm/Support/DebugCounter.h"
73 #include "llvm/Support/raw_ostream.h"
74 #include <cassert>
75 #include <iterator>
76
77 using namespace llvm;
78
79 #define DEBUG_TYPE "machine-cp"
80
81 STATISTIC(NumDeletes, "Number of dead copies deleted");
82 STATISTIC(NumCopyForwards, "Number of copy uses forwarded");
83 STATISTIC(NumCopyBackwardPropagated, "Number of copy defs backward propagated");
84 STATISTIC(SpillageChainsLength, "Length of spillage chains");
85 STATISTIC(NumSpillageChains, "Number of spillage chains");
86 DEBUG_COUNTER(FwdCounter, "machine-cp-fwd",
87 "Controls which register COPYs are forwarded");
88
89 static cl::opt<bool> MCPUseCopyInstr("mcp-use-is-copy-instr", cl::init(false),
90 cl::Hidden);
91 static cl::opt<cl::boolOrDefault>
92 EnableSpillageCopyElimination("enable-spill-copy-elim", cl::Hidden);
93
94 namespace {
95
isCopyInstr(const MachineInstr & MI,const TargetInstrInfo & TII,bool UseCopyInstr)96 static std::optional<DestSourcePair> isCopyInstr(const MachineInstr &MI,
97 const TargetInstrInfo &TII,
98 bool UseCopyInstr) {
99 if (UseCopyInstr)
100 return TII.isCopyInstr(MI);
101
102 if (MI.isCopy())
103 return std::optional<DestSourcePair>(
104 DestSourcePair{MI.getOperand(0), MI.getOperand(1)});
105
106 return std::nullopt;
107 }
108
109 class CopyTracker {
110 struct CopyInfo {
111 MachineInstr *MI, *LastSeenUseInCopy;
112 SmallVector<MCRegister, 4> DefRegs;
113 bool Avail;
114 };
115
116 DenseMap<MCRegUnit, CopyInfo> Copies;
117
118 public:
119 /// Mark all of the given registers and their subregisters as unavailable for
120 /// copying.
markRegsUnavailable(ArrayRef<MCRegister> Regs,const TargetRegisterInfo & TRI)121 void markRegsUnavailable(ArrayRef<MCRegister> Regs,
122 const TargetRegisterInfo &TRI) {
123 for (MCRegister Reg : Regs) {
124 // Source of copy is no longer available for propagation.
125 for (MCRegUnit Unit : TRI.regunits(Reg)) {
126 auto CI = Copies.find(Unit);
127 if (CI != Copies.end())
128 CI->second.Avail = false;
129 }
130 }
131 }
132
133 /// Remove register from copy maps.
invalidateRegister(MCRegister Reg,const TargetRegisterInfo & TRI,const TargetInstrInfo & TII,bool UseCopyInstr)134 void invalidateRegister(MCRegister Reg, const TargetRegisterInfo &TRI,
135 const TargetInstrInfo &TII, bool UseCopyInstr) {
136 // Since Reg might be a subreg of some registers, only invalidate Reg is not
137 // enough. We have to find the COPY defines Reg or registers defined by Reg
138 // and invalidate all of them. Similarly, we must invalidate all of the
139 // the subregisters used in the source of the COPY.
140 SmallSet<MCRegUnit, 8> RegUnitsToInvalidate;
141 auto InvalidateCopy = [&](MachineInstr *MI) {
142 std::optional<DestSourcePair> CopyOperands =
143 isCopyInstr(*MI, TII, UseCopyInstr);
144 assert(CopyOperands && "Expect copy");
145
146 auto Dest = TRI.regunits(CopyOperands->Destination->getReg().asMCReg());
147 auto Src = TRI.regunits(CopyOperands->Source->getReg().asMCReg());
148 RegUnitsToInvalidate.insert(Dest.begin(), Dest.end());
149 RegUnitsToInvalidate.insert(Src.begin(), Src.end());
150 };
151
152 for (MCRegUnit Unit : TRI.regunits(Reg)) {
153 auto I = Copies.find(Unit);
154 if (I != Copies.end()) {
155 if (MachineInstr *MI = I->second.MI)
156 InvalidateCopy(MI);
157 if (MachineInstr *MI = I->second.LastSeenUseInCopy)
158 InvalidateCopy(MI);
159 }
160 }
161 for (MCRegUnit Unit : RegUnitsToInvalidate)
162 Copies.erase(Unit);
163 }
164
165 /// Clobber a single register, removing it from the tracker's copy maps.
clobberRegister(MCRegister Reg,const TargetRegisterInfo & TRI,const TargetInstrInfo & TII,bool UseCopyInstr)166 void clobberRegister(MCRegister Reg, const TargetRegisterInfo &TRI,
167 const TargetInstrInfo &TII, bool UseCopyInstr) {
168 for (MCRegUnit Unit : TRI.regunits(Reg)) {
169 auto I = Copies.find(Unit);
170 if (I != Copies.end()) {
171 // When we clobber the source of a copy, we need to clobber everything
172 // it defined.
173 markRegsUnavailable(I->second.DefRegs, TRI);
174 // When we clobber the destination of a copy, we need to clobber the
175 // whole register it defined.
176 if (MachineInstr *MI = I->second.MI) {
177 std::optional<DestSourcePair> CopyOperands =
178 isCopyInstr(*MI, TII, UseCopyInstr);
179
180 MCRegister Def = CopyOperands->Destination->getReg().asMCReg();
181 MCRegister Src = CopyOperands->Source->getReg().asMCReg();
182
183 markRegsUnavailable(Def, TRI);
184
185 // Since we clobber the destination of a copy, the semantic of Src's
186 // "DefRegs" to contain Def is no longer effectual. We will also need
187 // to remove the record from the copy maps that indicates Src defined
188 // Def. Failing to do so might cause the target to miss some
189 // opportunities to further eliminate redundant copy instructions.
190 // Consider the following sequence during the
191 // ForwardCopyPropagateBlock procedure:
192 // L1: r0 = COPY r9 <- TrackMI
193 // L2: r0 = COPY r8 <- TrackMI (Remove r9 defined r0 from tracker)
194 // L3: use r0 <- Remove L2 from MaybeDeadCopies
195 // L4: early-clobber r9 <- Clobber r9 (L2 is still valid in tracker)
196 // L5: r0 = COPY r8 <- Remove NopCopy
197 for (MCRegUnit SrcUnit : TRI.regunits(Src)) {
198 auto SrcCopy = Copies.find(SrcUnit);
199 if (SrcCopy != Copies.end() && SrcCopy->second.LastSeenUseInCopy) {
200 // If SrcCopy defines multiple values, we only need
201 // to erase the record for Def in DefRegs.
202 for (auto itr = SrcCopy->second.DefRegs.begin();
203 itr != SrcCopy->second.DefRegs.end(); itr++) {
204 if (*itr == Def) {
205 SrcCopy->second.DefRegs.erase(itr);
206 // If DefReg becomes empty after removal, we can remove the
207 // SrcCopy from the tracker's copy maps. We only remove those
208 // entries solely record the Def is defined by Src. If an
209 // entry also contains the definition record of other Def'
210 // registers, it cannot be cleared.
211 if (SrcCopy->second.DefRegs.empty() && !SrcCopy->second.MI) {
212 Copies.erase(SrcCopy);
213 }
214 break;
215 }
216 }
217 }
218 }
219 }
220 // Now we can erase the copy.
221 Copies.erase(I);
222 }
223 }
224 }
225
226 /// Add this copy's registers into the tracker's copy maps.
trackCopy(MachineInstr * MI,const TargetRegisterInfo & TRI,const TargetInstrInfo & TII,bool UseCopyInstr)227 void trackCopy(MachineInstr *MI, const TargetRegisterInfo &TRI,
228 const TargetInstrInfo &TII, bool UseCopyInstr) {
229 std::optional<DestSourcePair> CopyOperands =
230 isCopyInstr(*MI, TII, UseCopyInstr);
231 assert(CopyOperands && "Tracking non-copy?");
232
233 MCRegister Src = CopyOperands->Source->getReg().asMCReg();
234 MCRegister Def = CopyOperands->Destination->getReg().asMCReg();
235
236 // Remember Def is defined by the copy.
237 for (MCRegUnit Unit : TRI.regunits(Def))
238 Copies[Unit] = {MI, nullptr, {}, true};
239
240 // Remember source that's copied to Def. Once it's clobbered, then
241 // it's no longer available for copy propagation.
242 for (MCRegUnit Unit : TRI.regunits(Src)) {
243 auto I = Copies.insert({Unit, {nullptr, nullptr, {}, false}});
244 auto &Copy = I.first->second;
245 if (!is_contained(Copy.DefRegs, Def))
246 Copy.DefRegs.push_back(Def);
247 Copy.LastSeenUseInCopy = MI;
248 }
249 }
250
hasAnyCopies()251 bool hasAnyCopies() {
252 return !Copies.empty();
253 }
254
findCopyForUnit(MCRegUnit RegUnit,const TargetRegisterInfo & TRI,bool MustBeAvailable=false)255 MachineInstr *findCopyForUnit(MCRegUnit RegUnit,
256 const TargetRegisterInfo &TRI,
257 bool MustBeAvailable = false) {
258 auto CI = Copies.find(RegUnit);
259 if (CI == Copies.end())
260 return nullptr;
261 if (MustBeAvailable && !CI->second.Avail)
262 return nullptr;
263 return CI->second.MI;
264 }
265
findCopyDefViaUnit(MCRegUnit RegUnit,const TargetRegisterInfo & TRI)266 MachineInstr *findCopyDefViaUnit(MCRegUnit RegUnit,
267 const TargetRegisterInfo &TRI) {
268 auto CI = Copies.find(RegUnit);
269 if (CI == Copies.end())
270 return nullptr;
271 if (CI->second.DefRegs.size() != 1)
272 return nullptr;
273 MCRegUnit RU = *TRI.regunits(CI->second.DefRegs[0]).begin();
274 return findCopyForUnit(RU, TRI, true);
275 }
276
findAvailBackwardCopy(MachineInstr & I,MCRegister Reg,const TargetRegisterInfo & TRI,const TargetInstrInfo & TII,bool UseCopyInstr)277 MachineInstr *findAvailBackwardCopy(MachineInstr &I, MCRegister Reg,
278 const TargetRegisterInfo &TRI,
279 const TargetInstrInfo &TII,
280 bool UseCopyInstr) {
281 MCRegUnit RU = *TRI.regunits(Reg).begin();
282 MachineInstr *AvailCopy = findCopyDefViaUnit(RU, TRI);
283
284 if (!AvailCopy)
285 return nullptr;
286
287 std::optional<DestSourcePair> CopyOperands =
288 isCopyInstr(*AvailCopy, TII, UseCopyInstr);
289 Register AvailSrc = CopyOperands->Source->getReg();
290 Register AvailDef = CopyOperands->Destination->getReg();
291 if (!TRI.isSubRegisterEq(AvailSrc, Reg))
292 return nullptr;
293
294 for (const MachineInstr &MI :
295 make_range(AvailCopy->getReverseIterator(), I.getReverseIterator()))
296 for (const MachineOperand &MO : MI.operands())
297 if (MO.isRegMask())
298 // FIXME: Shall we simultaneously invalidate AvailSrc or AvailDef?
299 if (MO.clobbersPhysReg(AvailSrc) || MO.clobbersPhysReg(AvailDef))
300 return nullptr;
301
302 return AvailCopy;
303 }
304
findAvailCopy(MachineInstr & DestCopy,MCRegister Reg,const TargetRegisterInfo & TRI,const TargetInstrInfo & TII,bool UseCopyInstr)305 MachineInstr *findAvailCopy(MachineInstr &DestCopy, MCRegister Reg,
306 const TargetRegisterInfo &TRI,
307 const TargetInstrInfo &TII, bool UseCopyInstr) {
308 // We check the first RegUnit here, since we'll only be interested in the
309 // copy if it copies the entire register anyway.
310 MCRegUnit RU = *TRI.regunits(Reg).begin();
311 MachineInstr *AvailCopy =
312 findCopyForUnit(RU, TRI, /*MustBeAvailable=*/true);
313
314 if (!AvailCopy)
315 return nullptr;
316
317 std::optional<DestSourcePair> CopyOperands =
318 isCopyInstr(*AvailCopy, TII, UseCopyInstr);
319 Register AvailSrc = CopyOperands->Source->getReg();
320 Register AvailDef = CopyOperands->Destination->getReg();
321 if (!TRI.isSubRegisterEq(AvailDef, Reg))
322 return nullptr;
323
324 // Check that the available copy isn't clobbered by any regmasks between
325 // itself and the destination.
326 for (const MachineInstr &MI :
327 make_range(AvailCopy->getIterator(), DestCopy.getIterator()))
328 for (const MachineOperand &MO : MI.operands())
329 if (MO.isRegMask())
330 if (MO.clobbersPhysReg(AvailSrc) || MO.clobbersPhysReg(AvailDef))
331 return nullptr;
332
333 return AvailCopy;
334 }
335
336 // Find last COPY that defines Reg before Current MachineInstr.
findLastSeenDefInCopy(const MachineInstr & Current,MCRegister Reg,const TargetRegisterInfo & TRI,const TargetInstrInfo & TII,bool UseCopyInstr)337 MachineInstr *findLastSeenDefInCopy(const MachineInstr &Current,
338 MCRegister Reg,
339 const TargetRegisterInfo &TRI,
340 const TargetInstrInfo &TII,
341 bool UseCopyInstr) {
342 MCRegUnit RU = *TRI.regunits(Reg).begin();
343 auto CI = Copies.find(RU);
344 if (CI == Copies.end() || !CI->second.Avail)
345 return nullptr;
346
347 MachineInstr *DefCopy = CI->second.MI;
348 std::optional<DestSourcePair> CopyOperands =
349 isCopyInstr(*DefCopy, TII, UseCopyInstr);
350 Register Def = CopyOperands->Destination->getReg();
351 if (!TRI.isSubRegisterEq(Def, Reg))
352 return nullptr;
353
354 for (const MachineInstr &MI :
355 make_range(static_cast<const MachineInstr *>(DefCopy)->getIterator(),
356 Current.getIterator()))
357 for (const MachineOperand &MO : MI.operands())
358 if (MO.isRegMask())
359 if (MO.clobbersPhysReg(Def)) {
360 LLVM_DEBUG(dbgs() << "MCP: Removed tracking of "
361 << printReg(Def, &TRI) << "\n");
362 return nullptr;
363 }
364
365 return DefCopy;
366 }
367
368 // Find last COPY that uses Reg.
findLastSeenUseInCopy(MCRegister Reg,const TargetRegisterInfo & TRI)369 MachineInstr *findLastSeenUseInCopy(MCRegister Reg,
370 const TargetRegisterInfo &TRI) {
371 MCRegUnit RU = *TRI.regunits(Reg).begin();
372 auto CI = Copies.find(RU);
373 if (CI == Copies.end())
374 return nullptr;
375 return CI->second.LastSeenUseInCopy;
376 }
377
clear()378 void clear() {
379 Copies.clear();
380 }
381 };
382
383 class MachineCopyPropagation : public MachineFunctionPass {
384 const TargetRegisterInfo *TRI = nullptr;
385 const TargetInstrInfo *TII = nullptr;
386 const MachineRegisterInfo *MRI = nullptr;
387
388 // Return true if this is a copy instruction and false otherwise.
389 bool UseCopyInstr;
390
391 public:
392 static char ID; // Pass identification, replacement for typeid
393
MachineCopyPropagation(bool CopyInstr=false)394 MachineCopyPropagation(bool CopyInstr = false)
395 : MachineFunctionPass(ID), UseCopyInstr(CopyInstr || MCPUseCopyInstr) {
396 initializeMachineCopyPropagationPass(*PassRegistry::getPassRegistry());
397 }
398
getAnalysisUsage(AnalysisUsage & AU) const399 void getAnalysisUsage(AnalysisUsage &AU) const override {
400 AU.setPreservesCFG();
401 MachineFunctionPass::getAnalysisUsage(AU);
402 }
403
404 bool runOnMachineFunction(MachineFunction &MF) override;
405
getRequiredProperties() const406 MachineFunctionProperties getRequiredProperties() const override {
407 return MachineFunctionProperties().set(
408 MachineFunctionProperties::Property::NoVRegs);
409 }
410
411 private:
412 typedef enum { DebugUse = false, RegularUse = true } DebugType;
413
414 void ReadRegister(MCRegister Reg, MachineInstr &Reader, DebugType DT);
415 void readSuccessorLiveIns(const MachineBasicBlock &MBB);
416 void ForwardCopyPropagateBlock(MachineBasicBlock &MBB);
417 void BackwardCopyPropagateBlock(MachineBasicBlock &MBB);
418 void EliminateSpillageCopies(MachineBasicBlock &MBB);
419 bool eraseIfRedundant(MachineInstr &Copy, MCRegister Src, MCRegister Def);
420 void forwardUses(MachineInstr &MI);
421 void propagateDefs(MachineInstr &MI);
422 bool isForwardableRegClassCopy(const MachineInstr &Copy,
423 const MachineInstr &UseI, unsigned UseIdx);
424 bool isBackwardPropagatableRegClassCopy(const MachineInstr &Copy,
425 const MachineInstr &UseI,
426 unsigned UseIdx);
427 bool hasImplicitOverlap(const MachineInstr &MI, const MachineOperand &Use);
428 bool hasOverlappingMultipleDef(const MachineInstr &MI,
429 const MachineOperand &MODef, Register Def);
430
431 /// Candidates for deletion.
432 SmallSetVector<MachineInstr *, 8> MaybeDeadCopies;
433
434 /// Multimap tracking debug users in current BB
435 DenseMap<MachineInstr *, SmallSet<MachineInstr *, 2>> CopyDbgUsers;
436
437 CopyTracker Tracker;
438
439 bool Changed = false;
440 };
441
442 } // end anonymous namespace
443
444 char MachineCopyPropagation::ID = 0;
445
446 char &llvm::MachineCopyPropagationID = MachineCopyPropagation::ID;
447
448 INITIALIZE_PASS(MachineCopyPropagation, DEBUG_TYPE,
449 "Machine Copy Propagation Pass", false, false)
450
ReadRegister(MCRegister Reg,MachineInstr & Reader,DebugType DT)451 void MachineCopyPropagation::ReadRegister(MCRegister Reg, MachineInstr &Reader,
452 DebugType DT) {
453 // If 'Reg' is defined by a copy, the copy is no longer a candidate
454 // for elimination. If a copy is "read" by a debug user, record the user
455 // for propagation.
456 for (MCRegUnit Unit : TRI->regunits(Reg)) {
457 if (MachineInstr *Copy = Tracker.findCopyForUnit(Unit, *TRI)) {
458 if (DT == RegularUse) {
459 LLVM_DEBUG(dbgs() << "MCP: Copy is used - not dead: "; Copy->dump());
460 MaybeDeadCopies.remove(Copy);
461 } else {
462 CopyDbgUsers[Copy].insert(&Reader);
463 }
464 }
465 }
466 }
467
readSuccessorLiveIns(const MachineBasicBlock & MBB)468 void MachineCopyPropagation::readSuccessorLiveIns(
469 const MachineBasicBlock &MBB) {
470 if (MaybeDeadCopies.empty())
471 return;
472
473 // If a copy result is livein to a successor, it is not dead.
474 for (const MachineBasicBlock *Succ : MBB.successors()) {
475 for (const auto &LI : Succ->liveins()) {
476 for (MCRegUnit Unit : TRI->regunits(LI.PhysReg)) {
477 if (MachineInstr *Copy = Tracker.findCopyForUnit(Unit, *TRI))
478 MaybeDeadCopies.remove(Copy);
479 }
480 }
481 }
482 }
483
484 /// Return true if \p PreviousCopy did copy register \p Src to register \p Def.
485 /// This fact may have been obscured by sub register usage or may not be true at
486 /// all even though Src and Def are subregisters of the registers used in
487 /// PreviousCopy. e.g.
488 /// isNopCopy("ecx = COPY eax", AX, CX) == true
489 /// isNopCopy("ecx = COPY eax", AH, CL) == false
isNopCopy(const MachineInstr & PreviousCopy,MCRegister Src,MCRegister Def,const TargetRegisterInfo * TRI,const TargetInstrInfo * TII,bool UseCopyInstr)490 static bool isNopCopy(const MachineInstr &PreviousCopy, MCRegister Src,
491 MCRegister Def, const TargetRegisterInfo *TRI,
492 const TargetInstrInfo *TII, bool UseCopyInstr) {
493
494 std::optional<DestSourcePair> CopyOperands =
495 isCopyInstr(PreviousCopy, *TII, UseCopyInstr);
496 MCRegister PreviousSrc = CopyOperands->Source->getReg().asMCReg();
497 MCRegister PreviousDef = CopyOperands->Destination->getReg().asMCReg();
498 if (Src == PreviousSrc && Def == PreviousDef)
499 return true;
500 if (!TRI->isSubRegister(PreviousSrc, Src))
501 return false;
502 unsigned SubIdx = TRI->getSubRegIndex(PreviousSrc, Src);
503 return SubIdx == TRI->getSubRegIndex(PreviousDef, Def);
504 }
505
506 /// Remove instruction \p Copy if there exists a previous copy that copies the
507 /// register \p Src to the register \p Def; This may happen indirectly by
508 /// copying the super registers.
eraseIfRedundant(MachineInstr & Copy,MCRegister Src,MCRegister Def)509 bool MachineCopyPropagation::eraseIfRedundant(MachineInstr &Copy,
510 MCRegister Src, MCRegister Def) {
511 // Avoid eliminating a copy from/to a reserved registers as we cannot predict
512 // the value (Example: The sparc zero register is writable but stays zero).
513 if (MRI->isReserved(Src) || MRI->isReserved(Def))
514 return false;
515
516 // Search for an existing copy.
517 MachineInstr *PrevCopy =
518 Tracker.findAvailCopy(Copy, Def, *TRI, *TII, UseCopyInstr);
519 if (!PrevCopy)
520 return false;
521
522 auto PrevCopyOperands = isCopyInstr(*PrevCopy, *TII, UseCopyInstr);
523 // Check that the existing copy uses the correct sub registers.
524 if (PrevCopyOperands->Destination->isDead())
525 return false;
526 if (!isNopCopy(*PrevCopy, Src, Def, TRI, TII, UseCopyInstr))
527 return false;
528
529 LLVM_DEBUG(dbgs() << "MCP: copy is a NOP, removing: "; Copy.dump());
530
531 // Copy was redundantly redefining either Src or Def. Remove earlier kill
532 // flags between Copy and PrevCopy because the value will be reused now.
533 std::optional<DestSourcePair> CopyOperands =
534 isCopyInstr(Copy, *TII, UseCopyInstr);
535 assert(CopyOperands);
536
537 Register CopyDef = CopyOperands->Destination->getReg();
538 assert(CopyDef == Src || CopyDef == Def);
539 for (MachineInstr &MI :
540 make_range(PrevCopy->getIterator(), Copy.getIterator()))
541 MI.clearRegisterKills(CopyDef, TRI);
542
543 // Clear undef flag from remaining copy if needed.
544 if (!CopyOperands->Source->isUndef()) {
545 PrevCopy->getOperand(PrevCopyOperands->Source->getOperandNo())
546 .setIsUndef(false);
547 }
548
549 Copy.eraseFromParent();
550 Changed = true;
551 ++NumDeletes;
552 return true;
553 }
554
isBackwardPropagatableRegClassCopy(const MachineInstr & Copy,const MachineInstr & UseI,unsigned UseIdx)555 bool MachineCopyPropagation::isBackwardPropagatableRegClassCopy(
556 const MachineInstr &Copy, const MachineInstr &UseI, unsigned UseIdx) {
557 std::optional<DestSourcePair> CopyOperands =
558 isCopyInstr(Copy, *TII, UseCopyInstr);
559 Register Def = CopyOperands->Destination->getReg();
560
561 if (const TargetRegisterClass *URC =
562 UseI.getRegClassConstraint(UseIdx, TII, TRI))
563 return URC->contains(Def);
564
565 // We don't process further if UseI is a COPY, since forward copy propagation
566 // should handle that.
567 return false;
568 }
569
570 /// Decide whether we should forward the source of \param Copy to its use in
571 /// \param UseI based on the physical register class constraints of the opcode
572 /// and avoiding introducing more cross-class COPYs.
isForwardableRegClassCopy(const MachineInstr & Copy,const MachineInstr & UseI,unsigned UseIdx)573 bool MachineCopyPropagation::isForwardableRegClassCopy(const MachineInstr &Copy,
574 const MachineInstr &UseI,
575 unsigned UseIdx) {
576 std::optional<DestSourcePair> CopyOperands =
577 isCopyInstr(Copy, *TII, UseCopyInstr);
578 Register CopySrcReg = CopyOperands->Source->getReg();
579
580 // If the new register meets the opcode register constraints, then allow
581 // forwarding.
582 if (const TargetRegisterClass *URC =
583 UseI.getRegClassConstraint(UseIdx, TII, TRI))
584 return URC->contains(CopySrcReg);
585
586 auto UseICopyOperands = isCopyInstr(UseI, *TII, UseCopyInstr);
587 if (!UseICopyOperands)
588 return false;
589
590 /// COPYs don't have register class constraints, so if the user instruction
591 /// is a COPY, we just try to avoid introducing additional cross-class
592 /// COPYs. For example:
593 ///
594 /// RegClassA = COPY RegClassB // Copy parameter
595 /// ...
596 /// RegClassB = COPY RegClassA // UseI parameter
597 ///
598 /// which after forwarding becomes
599 ///
600 /// RegClassA = COPY RegClassB
601 /// ...
602 /// RegClassB = COPY RegClassB
603 ///
604 /// so we have reduced the number of cross-class COPYs and potentially
605 /// introduced a nop COPY that can be removed.
606
607 // Allow forwarding if src and dst belong to any common class, so long as they
608 // don't belong to any (possibly smaller) common class that requires copies to
609 // go via a different class.
610 Register UseDstReg = UseICopyOperands->Destination->getReg();
611 bool Found = false;
612 bool IsCrossClass = false;
613 for (const TargetRegisterClass *RC : TRI->regclasses()) {
614 if (RC->contains(CopySrcReg) && RC->contains(UseDstReg)) {
615 Found = true;
616 if (TRI->getCrossCopyRegClass(RC) != RC) {
617 IsCrossClass = true;
618 break;
619 }
620 }
621 }
622 if (!Found)
623 return false;
624 if (!IsCrossClass)
625 return true;
626 // The forwarded copy would be cross-class. Only do this if the original copy
627 // was also cross-class.
628 Register CopyDstReg = CopyOperands->Destination->getReg();
629 for (const TargetRegisterClass *RC : TRI->regclasses()) {
630 if (RC->contains(CopySrcReg) && RC->contains(CopyDstReg) &&
631 TRI->getCrossCopyRegClass(RC) != RC)
632 return true;
633 }
634 return false;
635 }
636
637 /// Check that \p MI does not have implicit uses that overlap with it's \p Use
638 /// operand (the register being replaced), since these can sometimes be
639 /// implicitly tied to other operands. For example, on AMDGPU:
640 ///
641 /// V_MOVRELS_B32_e32 %VGPR2, %M0<imp-use>, %EXEC<imp-use>, %VGPR2_VGPR3_VGPR4_VGPR5<imp-use>
642 ///
643 /// the %VGPR2 is implicitly tied to the larger reg operand, but we have no
644 /// way of knowing we need to update the latter when updating the former.
hasImplicitOverlap(const MachineInstr & MI,const MachineOperand & Use)645 bool MachineCopyPropagation::hasImplicitOverlap(const MachineInstr &MI,
646 const MachineOperand &Use) {
647 for (const MachineOperand &MIUse : MI.uses())
648 if (&MIUse != &Use && MIUse.isReg() && MIUse.isImplicit() &&
649 MIUse.isUse() && TRI->regsOverlap(Use.getReg(), MIUse.getReg()))
650 return true;
651
652 return false;
653 }
654
655 /// For an MI that has multiple definitions, check whether \p MI has
656 /// a definition that overlaps with another of its definitions.
657 /// For example, on ARM: umull r9, r9, lr, r0
658 /// The umull instruction is unpredictable unless RdHi and RdLo are different.
hasOverlappingMultipleDef(const MachineInstr & MI,const MachineOperand & MODef,Register Def)659 bool MachineCopyPropagation::hasOverlappingMultipleDef(
660 const MachineInstr &MI, const MachineOperand &MODef, Register Def) {
661 for (const MachineOperand &MIDef : MI.all_defs()) {
662 if ((&MIDef != &MODef) && MIDef.isReg() &&
663 TRI->regsOverlap(Def, MIDef.getReg()))
664 return true;
665 }
666
667 return false;
668 }
669
670 /// Look for available copies whose destination register is used by \p MI and
671 /// replace the use in \p MI with the copy's source register.
forwardUses(MachineInstr & MI)672 void MachineCopyPropagation::forwardUses(MachineInstr &MI) {
673 if (!Tracker.hasAnyCopies())
674 return;
675
676 // Look for non-tied explicit vreg uses that have an active COPY
677 // instruction that defines the physical register allocated to them.
678 // Replace the vreg with the source of the active COPY.
679 for (unsigned OpIdx = 0, OpEnd = MI.getNumOperands(); OpIdx < OpEnd;
680 ++OpIdx) {
681 MachineOperand &MOUse = MI.getOperand(OpIdx);
682 // Don't forward into undef use operands since doing so can cause problems
683 // with the machine verifier, since it doesn't treat undef reads as reads,
684 // so we can end up with a live range that ends on an undef read, leading to
685 // an error that the live range doesn't end on a read of the live range
686 // register.
687 if (!MOUse.isReg() || MOUse.isTied() || MOUse.isUndef() || MOUse.isDef() ||
688 MOUse.isImplicit())
689 continue;
690
691 if (!MOUse.getReg())
692 continue;
693
694 // Check that the register is marked 'renamable' so we know it is safe to
695 // rename it without violating any constraints that aren't expressed in the
696 // IR (e.g. ABI or opcode requirements).
697 if (!MOUse.isRenamable())
698 continue;
699
700 MachineInstr *Copy = Tracker.findAvailCopy(MI, MOUse.getReg().asMCReg(),
701 *TRI, *TII, UseCopyInstr);
702 if (!Copy)
703 continue;
704
705 std::optional<DestSourcePair> CopyOperands =
706 isCopyInstr(*Copy, *TII, UseCopyInstr);
707 Register CopyDstReg = CopyOperands->Destination->getReg();
708 const MachineOperand &CopySrc = *CopyOperands->Source;
709 Register CopySrcReg = CopySrc.getReg();
710
711 Register ForwardedReg = CopySrcReg;
712 // MI might use a sub-register of the Copy destination, in which case the
713 // forwarded register is the matching sub-register of the Copy source.
714 if (MOUse.getReg() != CopyDstReg) {
715 unsigned SubRegIdx = TRI->getSubRegIndex(CopyDstReg, MOUse.getReg());
716 assert(SubRegIdx &&
717 "MI source is not a sub-register of Copy destination");
718 ForwardedReg = TRI->getSubReg(CopySrcReg, SubRegIdx);
719 if (!ForwardedReg) {
720 LLVM_DEBUG(dbgs() << "MCP: Copy source does not have sub-register "
721 << TRI->getSubRegIndexName(SubRegIdx) << '\n');
722 continue;
723 }
724 }
725
726 // Don't forward COPYs of reserved regs unless they are constant.
727 if (MRI->isReserved(CopySrcReg) && !MRI->isConstantPhysReg(CopySrcReg))
728 continue;
729
730 if (!isForwardableRegClassCopy(*Copy, MI, OpIdx))
731 continue;
732
733 if (hasImplicitOverlap(MI, MOUse))
734 continue;
735
736 // Check that the instruction is not a copy that partially overwrites the
737 // original copy source that we are about to use. The tracker mechanism
738 // cannot cope with that.
739 if (isCopyInstr(MI, *TII, UseCopyInstr) &&
740 MI.modifiesRegister(CopySrcReg, TRI) &&
741 !MI.definesRegister(CopySrcReg, /*TRI=*/nullptr)) {
742 LLVM_DEBUG(dbgs() << "MCP: Copy source overlap with dest in " << MI);
743 continue;
744 }
745
746 if (!DebugCounter::shouldExecute(FwdCounter)) {
747 LLVM_DEBUG(dbgs() << "MCP: Skipping forwarding due to debug counter:\n "
748 << MI);
749 continue;
750 }
751
752 LLVM_DEBUG(dbgs() << "MCP: Replacing " << printReg(MOUse.getReg(), TRI)
753 << "\n with " << printReg(ForwardedReg, TRI)
754 << "\n in " << MI << " from " << *Copy);
755
756 MOUse.setReg(ForwardedReg);
757
758 if (!CopySrc.isRenamable())
759 MOUse.setIsRenamable(false);
760 MOUse.setIsUndef(CopySrc.isUndef());
761
762 LLVM_DEBUG(dbgs() << "MCP: After replacement: " << MI << "\n");
763
764 // Clear kill markers that may have been invalidated.
765 for (MachineInstr &KMI :
766 make_range(Copy->getIterator(), std::next(MI.getIterator())))
767 KMI.clearRegisterKills(CopySrcReg, TRI);
768
769 ++NumCopyForwards;
770 Changed = true;
771 }
772 }
773
ForwardCopyPropagateBlock(MachineBasicBlock & MBB)774 void MachineCopyPropagation::ForwardCopyPropagateBlock(MachineBasicBlock &MBB) {
775 LLVM_DEBUG(dbgs() << "MCP: ForwardCopyPropagateBlock " << MBB.getName()
776 << "\n");
777
778 for (MachineInstr &MI : llvm::make_early_inc_range(MBB)) {
779 // Analyze copies (which don't overlap themselves).
780 std::optional<DestSourcePair> CopyOperands =
781 isCopyInstr(MI, *TII, UseCopyInstr);
782 if (CopyOperands) {
783
784 Register RegSrc = CopyOperands->Source->getReg();
785 Register RegDef = CopyOperands->Destination->getReg();
786
787 if (!TRI->regsOverlap(RegDef, RegSrc)) {
788 assert(RegDef.isPhysical() && RegSrc.isPhysical() &&
789 "MachineCopyPropagation should be run after register allocation!");
790
791 MCRegister Def = RegDef.asMCReg();
792 MCRegister Src = RegSrc.asMCReg();
793
794 // The two copies cancel out and the source of the first copy
795 // hasn't been overridden, eliminate the second one. e.g.
796 // %ecx = COPY %eax
797 // ... nothing clobbered eax.
798 // %eax = COPY %ecx
799 // =>
800 // %ecx = COPY %eax
801 //
802 // or
803 //
804 // %ecx = COPY %eax
805 // ... nothing clobbered eax.
806 // %ecx = COPY %eax
807 // =>
808 // %ecx = COPY %eax
809 if (eraseIfRedundant(MI, Def, Src) || eraseIfRedundant(MI, Src, Def))
810 continue;
811
812 forwardUses(MI);
813
814 // Src may have been changed by forwardUses()
815 CopyOperands = isCopyInstr(MI, *TII, UseCopyInstr);
816 Src = CopyOperands->Source->getReg().asMCReg();
817
818 // If Src is defined by a previous copy, the previous copy cannot be
819 // eliminated.
820 ReadRegister(Src, MI, RegularUse);
821 for (const MachineOperand &MO : MI.implicit_operands()) {
822 if (!MO.isReg() || !MO.readsReg())
823 continue;
824 MCRegister Reg = MO.getReg().asMCReg();
825 if (!Reg)
826 continue;
827 ReadRegister(Reg, MI, RegularUse);
828 }
829
830 LLVM_DEBUG(dbgs() << "MCP: Copy is a deletion candidate: "; MI.dump());
831
832 // Copy is now a candidate for deletion.
833 if (!MRI->isReserved(Def))
834 MaybeDeadCopies.insert(&MI);
835
836 // If 'Def' is previously source of another copy, then this earlier copy's
837 // source is no longer available. e.g.
838 // %xmm9 = copy %xmm2
839 // ...
840 // %xmm2 = copy %xmm0
841 // ...
842 // %xmm2 = copy %xmm9
843 Tracker.clobberRegister(Def, *TRI, *TII, UseCopyInstr);
844 for (const MachineOperand &MO : MI.implicit_operands()) {
845 if (!MO.isReg() || !MO.isDef())
846 continue;
847 MCRegister Reg = MO.getReg().asMCReg();
848 if (!Reg)
849 continue;
850 Tracker.clobberRegister(Reg, *TRI, *TII, UseCopyInstr);
851 }
852
853 Tracker.trackCopy(&MI, *TRI, *TII, UseCopyInstr);
854
855 continue;
856 }
857 }
858
859 // Clobber any earlyclobber regs first.
860 for (const MachineOperand &MO : MI.operands())
861 if (MO.isReg() && MO.isEarlyClobber()) {
862 MCRegister Reg = MO.getReg().asMCReg();
863 // If we have a tied earlyclobber, that means it is also read by this
864 // instruction, so we need to make sure we don't remove it as dead
865 // later.
866 if (MO.isTied())
867 ReadRegister(Reg, MI, RegularUse);
868 Tracker.clobberRegister(Reg, *TRI, *TII, UseCopyInstr);
869 }
870
871 forwardUses(MI);
872
873 // Not a copy.
874 SmallVector<Register, 4> Defs;
875 const MachineOperand *RegMask = nullptr;
876 for (const MachineOperand &MO : MI.operands()) {
877 if (MO.isRegMask())
878 RegMask = &MO;
879 if (!MO.isReg())
880 continue;
881 Register Reg = MO.getReg();
882 if (!Reg)
883 continue;
884
885 assert(!Reg.isVirtual() &&
886 "MachineCopyPropagation should be run after register allocation!");
887
888 if (MO.isDef() && !MO.isEarlyClobber()) {
889 Defs.push_back(Reg.asMCReg());
890 continue;
891 } else if (MO.readsReg())
892 ReadRegister(Reg.asMCReg(), MI, MO.isDebug() ? DebugUse : RegularUse);
893 }
894
895 // The instruction has a register mask operand which means that it clobbers
896 // a large set of registers. Treat clobbered registers the same way as
897 // defined registers.
898 if (RegMask) {
899 // Erase any MaybeDeadCopies whose destination register is clobbered.
900 for (SmallSetVector<MachineInstr *, 8>::iterator DI =
901 MaybeDeadCopies.begin();
902 DI != MaybeDeadCopies.end();) {
903 MachineInstr *MaybeDead = *DI;
904 std::optional<DestSourcePair> CopyOperands =
905 isCopyInstr(*MaybeDead, *TII, UseCopyInstr);
906 MCRegister Reg = CopyOperands->Destination->getReg().asMCReg();
907 assert(!MRI->isReserved(Reg));
908
909 if (!RegMask->clobbersPhysReg(Reg)) {
910 ++DI;
911 continue;
912 }
913
914 LLVM_DEBUG(dbgs() << "MCP: Removing copy due to regmask clobbering: ";
915 MaybeDead->dump());
916
917 // Make sure we invalidate any entries in the copy maps before erasing
918 // the instruction.
919 Tracker.clobberRegister(Reg, *TRI, *TII, UseCopyInstr);
920
921 // erase() will return the next valid iterator pointing to the next
922 // element after the erased one.
923 DI = MaybeDeadCopies.erase(DI);
924 MaybeDead->eraseFromParent();
925 Changed = true;
926 ++NumDeletes;
927 }
928 }
929
930 // Any previous copy definition or reading the Defs is no longer available.
931 for (MCRegister Reg : Defs)
932 Tracker.clobberRegister(Reg, *TRI, *TII, UseCopyInstr);
933 }
934
935 bool TracksLiveness = MRI->tracksLiveness();
936
937 // If liveness is tracked, we can use the live-in lists to know which
938 // copies aren't dead.
939 if (TracksLiveness)
940 readSuccessorLiveIns(MBB);
941
942 // If MBB doesn't have succesor, delete copies whose defs are not used.
943 // If MBB does have successors, we can only delete copies if we are able to
944 // use liveness information from successors to confirm they are really dead.
945 if (MBB.succ_empty() || TracksLiveness) {
946 for (MachineInstr *MaybeDead : MaybeDeadCopies) {
947 LLVM_DEBUG(dbgs() << "MCP: Removing copy due to no live-out succ: ";
948 MaybeDead->dump());
949
950 std::optional<DestSourcePair> CopyOperands =
951 isCopyInstr(*MaybeDead, *TII, UseCopyInstr);
952 assert(CopyOperands);
953
954 Register SrcReg = CopyOperands->Source->getReg();
955 Register DestReg = CopyOperands->Destination->getReg();
956 assert(!MRI->isReserved(DestReg));
957
958 // Update matching debug values, if any.
959 SmallVector<MachineInstr *> MaybeDeadDbgUsers(
960 CopyDbgUsers[MaybeDead].begin(), CopyDbgUsers[MaybeDead].end());
961 MRI->updateDbgUsersToReg(DestReg.asMCReg(), SrcReg.asMCReg(),
962 MaybeDeadDbgUsers);
963
964 MaybeDead->eraseFromParent();
965 Changed = true;
966 ++NumDeletes;
967 }
968 }
969
970 MaybeDeadCopies.clear();
971 CopyDbgUsers.clear();
972 Tracker.clear();
973 }
974
isBackwardPropagatableCopy(const DestSourcePair & CopyOperands,const MachineRegisterInfo & MRI)975 static bool isBackwardPropagatableCopy(const DestSourcePair &CopyOperands,
976 const MachineRegisterInfo &MRI) {
977 Register Def = CopyOperands.Destination->getReg();
978 Register Src = CopyOperands.Source->getReg();
979
980 if (!Def || !Src)
981 return false;
982
983 if (MRI.isReserved(Def) || MRI.isReserved(Src))
984 return false;
985
986 return CopyOperands.Source->isRenamable() && CopyOperands.Source->isKill();
987 }
988
propagateDefs(MachineInstr & MI)989 void MachineCopyPropagation::propagateDefs(MachineInstr &MI) {
990 if (!Tracker.hasAnyCopies())
991 return;
992
993 for (unsigned OpIdx = 0, OpEnd = MI.getNumOperands(); OpIdx != OpEnd;
994 ++OpIdx) {
995 MachineOperand &MODef = MI.getOperand(OpIdx);
996
997 if (!MODef.isReg() || MODef.isUse())
998 continue;
999
1000 // Ignore non-trivial cases.
1001 if (MODef.isTied() || MODef.isUndef() || MODef.isImplicit())
1002 continue;
1003
1004 if (!MODef.getReg())
1005 continue;
1006
1007 // We only handle if the register comes from a vreg.
1008 if (!MODef.isRenamable())
1009 continue;
1010
1011 MachineInstr *Copy = Tracker.findAvailBackwardCopy(
1012 MI, MODef.getReg().asMCReg(), *TRI, *TII, UseCopyInstr);
1013 if (!Copy)
1014 continue;
1015
1016 std::optional<DestSourcePair> CopyOperands =
1017 isCopyInstr(*Copy, *TII, UseCopyInstr);
1018 Register Def = CopyOperands->Destination->getReg();
1019 Register Src = CopyOperands->Source->getReg();
1020
1021 if (MODef.getReg() != Src)
1022 continue;
1023
1024 if (!isBackwardPropagatableRegClassCopy(*Copy, MI, OpIdx))
1025 continue;
1026
1027 if (hasImplicitOverlap(MI, MODef))
1028 continue;
1029
1030 if (hasOverlappingMultipleDef(MI, MODef, Def))
1031 continue;
1032
1033 LLVM_DEBUG(dbgs() << "MCP: Replacing " << printReg(MODef.getReg(), TRI)
1034 << "\n with " << printReg(Def, TRI) << "\n in "
1035 << MI << " from " << *Copy);
1036
1037 MODef.setReg(Def);
1038 MODef.setIsRenamable(CopyOperands->Destination->isRenamable());
1039
1040 LLVM_DEBUG(dbgs() << "MCP: After replacement: " << MI << "\n");
1041 MaybeDeadCopies.insert(Copy);
1042 Changed = true;
1043 ++NumCopyBackwardPropagated;
1044 }
1045 }
1046
BackwardCopyPropagateBlock(MachineBasicBlock & MBB)1047 void MachineCopyPropagation::BackwardCopyPropagateBlock(
1048 MachineBasicBlock &MBB) {
1049 LLVM_DEBUG(dbgs() << "MCP: BackwardCopyPropagateBlock " << MBB.getName()
1050 << "\n");
1051
1052 for (MachineInstr &MI : llvm::make_early_inc_range(llvm::reverse(MBB))) {
1053 // Ignore non-trivial COPYs.
1054 std::optional<DestSourcePair> CopyOperands =
1055 isCopyInstr(MI, *TII, UseCopyInstr);
1056 if (CopyOperands && MI.getNumOperands() == 2) {
1057 Register DefReg = CopyOperands->Destination->getReg();
1058 Register SrcReg = CopyOperands->Source->getReg();
1059
1060 if (!TRI->regsOverlap(DefReg, SrcReg)) {
1061 // Unlike forward cp, we don't invoke propagateDefs here,
1062 // just let forward cp do COPY-to-COPY propagation.
1063 if (isBackwardPropagatableCopy(*CopyOperands, *MRI)) {
1064 Tracker.invalidateRegister(SrcReg.asMCReg(), *TRI, *TII,
1065 UseCopyInstr);
1066 Tracker.invalidateRegister(DefReg.asMCReg(), *TRI, *TII,
1067 UseCopyInstr);
1068 Tracker.trackCopy(&MI, *TRI, *TII, UseCopyInstr);
1069 continue;
1070 }
1071 }
1072 }
1073
1074 // Invalidate any earlyclobber regs first.
1075 for (const MachineOperand &MO : MI.operands())
1076 if (MO.isReg() && MO.isEarlyClobber()) {
1077 MCRegister Reg = MO.getReg().asMCReg();
1078 if (!Reg)
1079 continue;
1080 Tracker.invalidateRegister(Reg, *TRI, *TII, UseCopyInstr);
1081 }
1082
1083 propagateDefs(MI);
1084 for (const MachineOperand &MO : MI.operands()) {
1085 if (!MO.isReg())
1086 continue;
1087
1088 if (!MO.getReg())
1089 continue;
1090
1091 if (MO.isDef())
1092 Tracker.invalidateRegister(MO.getReg().asMCReg(), *TRI, *TII,
1093 UseCopyInstr);
1094
1095 if (MO.readsReg()) {
1096 if (MO.isDebug()) {
1097 // Check if the register in the debug instruction is utilized
1098 // in a copy instruction, so we can update the debug info if the
1099 // register is changed.
1100 for (MCRegUnit Unit : TRI->regunits(MO.getReg().asMCReg())) {
1101 if (auto *Copy = Tracker.findCopyDefViaUnit(Unit, *TRI)) {
1102 CopyDbgUsers[Copy].insert(&MI);
1103 }
1104 }
1105 } else {
1106 Tracker.invalidateRegister(MO.getReg().asMCReg(), *TRI, *TII,
1107 UseCopyInstr);
1108 }
1109 }
1110 }
1111 }
1112
1113 for (auto *Copy : MaybeDeadCopies) {
1114 std::optional<DestSourcePair> CopyOperands =
1115 isCopyInstr(*Copy, *TII, UseCopyInstr);
1116 Register Src = CopyOperands->Source->getReg();
1117 Register Def = CopyOperands->Destination->getReg();
1118 SmallVector<MachineInstr *> MaybeDeadDbgUsers(CopyDbgUsers[Copy].begin(),
1119 CopyDbgUsers[Copy].end());
1120
1121 MRI->updateDbgUsersToReg(Src.asMCReg(), Def.asMCReg(), MaybeDeadDbgUsers);
1122 Copy->eraseFromParent();
1123 ++NumDeletes;
1124 }
1125
1126 MaybeDeadCopies.clear();
1127 CopyDbgUsers.clear();
1128 Tracker.clear();
1129 }
1130
printSpillReloadChain(DenseMap<MachineInstr *,SmallVector<MachineInstr * >> & SpillChain,DenseMap<MachineInstr *,SmallVector<MachineInstr * >> & ReloadChain,MachineInstr * Leader)1131 static void LLVM_ATTRIBUTE_UNUSED printSpillReloadChain(
1132 DenseMap<MachineInstr *, SmallVector<MachineInstr *>> &SpillChain,
1133 DenseMap<MachineInstr *, SmallVector<MachineInstr *>> &ReloadChain,
1134 MachineInstr *Leader) {
1135 auto &SC = SpillChain[Leader];
1136 auto &RC = ReloadChain[Leader];
1137 for (auto I = SC.rbegin(), E = SC.rend(); I != E; ++I)
1138 (*I)->dump();
1139 for (MachineInstr *MI : RC)
1140 MI->dump();
1141 }
1142
1143 // Remove spill-reload like copy chains. For example
1144 // r0 = COPY r1
1145 // r1 = COPY r2
1146 // r2 = COPY r3
1147 // r3 = COPY r4
1148 // <def-use r4>
1149 // r4 = COPY r3
1150 // r3 = COPY r2
1151 // r2 = COPY r1
1152 // r1 = COPY r0
1153 // will be folded into
1154 // r0 = COPY r1
1155 // r1 = COPY r4
1156 // <def-use r4>
1157 // r4 = COPY r1
1158 // r1 = COPY r0
1159 // TODO: Currently we don't track usage of r0 outside the chain, so we
1160 // conservatively keep its value as it was before the rewrite.
1161 //
1162 // The algorithm is trying to keep
1163 // property#1: No Def of spill COPY in the chain is used or defined until the
1164 // paired reload COPY in the chain uses the Def.
1165 //
1166 // property#2: NO Source of COPY in the chain is used or defined until the next
1167 // COPY in the chain defines the Source, except the innermost spill-reload
1168 // pair.
1169 //
1170 // The algorithm is conducted by checking every COPY inside the MBB, assuming
1171 // the COPY is a reload COPY, then try to find paired spill COPY by searching
1172 // the COPY defines the Src of the reload COPY backward. If such pair is found,
1173 // it either belongs to an existing chain or a new chain depends on
1174 // last available COPY uses the Def of the reload COPY.
1175 // Implementation notes, we use CopyTracker::findLastDefCopy(Reg, ...) to find
1176 // out last COPY that defines Reg; we use CopyTracker::findLastUseCopy(Reg, ...)
1177 // to find out last COPY that uses Reg. When we are encountered with a Non-COPY
1178 // instruction, we check registers in the operands of this instruction. If this
1179 // Reg is defined by a COPY, we untrack this Reg via
1180 // CopyTracker::clobberRegister(Reg, ...).
EliminateSpillageCopies(MachineBasicBlock & MBB)1181 void MachineCopyPropagation::EliminateSpillageCopies(MachineBasicBlock &MBB) {
1182 // ChainLeader maps MI inside a spill-reload chain to its innermost reload COPY.
1183 // Thus we can track if a MI belongs to an existing spill-reload chain.
1184 DenseMap<MachineInstr *, MachineInstr *> ChainLeader;
1185 // SpillChain maps innermost reload COPY of a spill-reload chain to a sequence
1186 // of COPYs that forms spills of a spill-reload chain.
1187 // ReloadChain maps innermost reload COPY of a spill-reload chain to a
1188 // sequence of COPYs that forms reloads of a spill-reload chain.
1189 DenseMap<MachineInstr *, SmallVector<MachineInstr *>> SpillChain, ReloadChain;
1190 // If a COPY's Source has use or def until next COPY defines the Source,
1191 // we put the COPY in this set to keep property#2.
1192 DenseSet<const MachineInstr *> CopySourceInvalid;
1193
1194 auto TryFoldSpillageCopies =
1195 [&, this](const SmallVectorImpl<MachineInstr *> &SC,
1196 const SmallVectorImpl<MachineInstr *> &RC) {
1197 assert(SC.size() == RC.size() && "Spill-reload should be paired");
1198
1199 // We need at least 3 pairs of copies for the transformation to apply,
1200 // because the first outermost pair cannot be removed since we don't
1201 // recolor outside of the chain and that we need at least one temporary
1202 // spill slot to shorten the chain. If we only have a chain of two
1203 // pairs, we already have the shortest sequence this code can handle:
1204 // the outermost pair for the temporary spill slot, and the pair that
1205 // use that temporary spill slot for the other end of the chain.
1206 // TODO: We might be able to simplify to one spill-reload pair if collecting
1207 // more infomation about the outermost COPY.
1208 if (SC.size() <= 2)
1209 return;
1210
1211 // If violate property#2, we don't fold the chain.
1212 for (const MachineInstr *Spill : drop_begin(SC))
1213 if (CopySourceInvalid.count(Spill))
1214 return;
1215
1216 for (const MachineInstr *Reload : drop_end(RC))
1217 if (CopySourceInvalid.count(Reload))
1218 return;
1219
1220 auto CheckCopyConstraint = [this](Register Def, Register Src) {
1221 for (const TargetRegisterClass *RC : TRI->regclasses()) {
1222 if (RC->contains(Def) && RC->contains(Src))
1223 return true;
1224 }
1225 return false;
1226 };
1227
1228 auto UpdateReg = [](MachineInstr *MI, const MachineOperand *Old,
1229 const MachineOperand *New) {
1230 for (MachineOperand &MO : MI->operands()) {
1231 if (&MO == Old)
1232 MO.setReg(New->getReg());
1233 }
1234 };
1235
1236 std::optional<DestSourcePair> InnerMostSpillCopy =
1237 isCopyInstr(*SC[0], *TII, UseCopyInstr);
1238 std::optional<DestSourcePair> OuterMostSpillCopy =
1239 isCopyInstr(*SC.back(), *TII, UseCopyInstr);
1240 std::optional<DestSourcePair> InnerMostReloadCopy =
1241 isCopyInstr(*RC[0], *TII, UseCopyInstr);
1242 std::optional<DestSourcePair> OuterMostReloadCopy =
1243 isCopyInstr(*RC.back(), *TII, UseCopyInstr);
1244 if (!CheckCopyConstraint(OuterMostSpillCopy->Source->getReg(),
1245 InnerMostSpillCopy->Source->getReg()) ||
1246 !CheckCopyConstraint(InnerMostReloadCopy->Destination->getReg(),
1247 OuterMostReloadCopy->Destination->getReg()))
1248 return;
1249
1250 SpillageChainsLength += SC.size() + RC.size();
1251 NumSpillageChains += 1;
1252 UpdateReg(SC[0], InnerMostSpillCopy->Destination,
1253 OuterMostSpillCopy->Source);
1254 UpdateReg(RC[0], InnerMostReloadCopy->Source,
1255 OuterMostReloadCopy->Destination);
1256
1257 for (size_t I = 1; I < SC.size() - 1; ++I) {
1258 SC[I]->eraseFromParent();
1259 RC[I]->eraseFromParent();
1260 NumDeletes += 2;
1261 }
1262 };
1263
1264 auto IsFoldableCopy = [this](const MachineInstr &MaybeCopy) {
1265 if (MaybeCopy.getNumImplicitOperands() > 0)
1266 return false;
1267 std::optional<DestSourcePair> CopyOperands =
1268 isCopyInstr(MaybeCopy, *TII, UseCopyInstr);
1269 if (!CopyOperands)
1270 return false;
1271 Register Src = CopyOperands->Source->getReg();
1272 Register Def = CopyOperands->Destination->getReg();
1273 return Src && Def && !TRI->regsOverlap(Src, Def) &&
1274 CopyOperands->Source->isRenamable() &&
1275 CopyOperands->Destination->isRenamable();
1276 };
1277
1278 auto IsSpillReloadPair = [&, this](const MachineInstr &Spill,
1279 const MachineInstr &Reload) {
1280 if (!IsFoldableCopy(Spill) || !IsFoldableCopy(Reload))
1281 return false;
1282 std::optional<DestSourcePair> SpillCopy =
1283 isCopyInstr(Spill, *TII, UseCopyInstr);
1284 std::optional<DestSourcePair> ReloadCopy =
1285 isCopyInstr(Reload, *TII, UseCopyInstr);
1286 if (!SpillCopy || !ReloadCopy)
1287 return false;
1288 return SpillCopy->Source->getReg() == ReloadCopy->Destination->getReg() &&
1289 SpillCopy->Destination->getReg() == ReloadCopy->Source->getReg();
1290 };
1291
1292 auto IsChainedCopy = [&, this](const MachineInstr &Prev,
1293 const MachineInstr &Current) {
1294 if (!IsFoldableCopy(Prev) || !IsFoldableCopy(Current))
1295 return false;
1296 std::optional<DestSourcePair> PrevCopy =
1297 isCopyInstr(Prev, *TII, UseCopyInstr);
1298 std::optional<DestSourcePair> CurrentCopy =
1299 isCopyInstr(Current, *TII, UseCopyInstr);
1300 if (!PrevCopy || !CurrentCopy)
1301 return false;
1302 return PrevCopy->Source->getReg() == CurrentCopy->Destination->getReg();
1303 };
1304
1305 for (MachineInstr &MI : llvm::make_early_inc_range(MBB)) {
1306 std::optional<DestSourcePair> CopyOperands =
1307 isCopyInstr(MI, *TII, UseCopyInstr);
1308
1309 // Update track information via non-copy instruction.
1310 SmallSet<Register, 8> RegsToClobber;
1311 if (!CopyOperands) {
1312 for (const MachineOperand &MO : MI.operands()) {
1313 if (!MO.isReg())
1314 continue;
1315 Register Reg = MO.getReg();
1316 if (!Reg)
1317 continue;
1318 MachineInstr *LastUseCopy =
1319 Tracker.findLastSeenUseInCopy(Reg.asMCReg(), *TRI);
1320 if (LastUseCopy) {
1321 LLVM_DEBUG(dbgs() << "MCP: Copy source of\n");
1322 LLVM_DEBUG(LastUseCopy->dump());
1323 LLVM_DEBUG(dbgs() << "might be invalidated by\n");
1324 LLVM_DEBUG(MI.dump());
1325 CopySourceInvalid.insert(LastUseCopy);
1326 }
1327 // Must be noted Tracker.clobberRegister(Reg, ...) removes tracking of
1328 // Reg, i.e, COPY that defines Reg is removed from the mapping as well
1329 // as marking COPYs that uses Reg unavailable.
1330 // We don't invoke CopyTracker::clobberRegister(Reg, ...) if Reg is not
1331 // defined by a previous COPY, since we don't want to make COPYs uses
1332 // Reg unavailable.
1333 if (Tracker.findLastSeenDefInCopy(MI, Reg.asMCReg(), *TRI, *TII,
1334 UseCopyInstr))
1335 // Thus we can keep the property#1.
1336 RegsToClobber.insert(Reg);
1337 }
1338 for (Register Reg : RegsToClobber) {
1339 Tracker.clobberRegister(Reg, *TRI, *TII, UseCopyInstr);
1340 LLVM_DEBUG(dbgs() << "MCP: Removed tracking of " << printReg(Reg, TRI)
1341 << "\n");
1342 }
1343 continue;
1344 }
1345
1346 Register Src = CopyOperands->Source->getReg();
1347 Register Def = CopyOperands->Destination->getReg();
1348 // Check if we can find a pair spill-reload copy.
1349 LLVM_DEBUG(dbgs() << "MCP: Searching paired spill for reload: ");
1350 LLVM_DEBUG(MI.dump());
1351 MachineInstr *MaybeSpill =
1352 Tracker.findLastSeenDefInCopy(MI, Src.asMCReg(), *TRI, *TII, UseCopyInstr);
1353 bool MaybeSpillIsChained = ChainLeader.count(MaybeSpill);
1354 if (!MaybeSpillIsChained && MaybeSpill &&
1355 IsSpillReloadPair(*MaybeSpill, MI)) {
1356 // Check if we already have an existing chain. Now we have a
1357 // spill-reload pair.
1358 // L2: r2 = COPY r3
1359 // L5: r3 = COPY r2
1360 // Looking for a valid COPY before L5 which uses r3.
1361 // This can be serverial cases.
1362 // Case #1:
1363 // No COPY is found, which can be r3 is def-use between (L2, L5), we
1364 // create a new chain for L2 and L5.
1365 // Case #2:
1366 // L2: r2 = COPY r3
1367 // L5: r3 = COPY r2
1368 // Such COPY is found and is L2, we create a new chain for L2 and L5.
1369 // Case #3:
1370 // L2: r2 = COPY r3
1371 // L3: r1 = COPY r3
1372 // L5: r3 = COPY r2
1373 // we create a new chain for L2 and L5.
1374 // Case #4:
1375 // L2: r2 = COPY r3
1376 // L3: r1 = COPY r3
1377 // L4: r3 = COPY r1
1378 // L5: r3 = COPY r2
1379 // Such COPY won't be found since L4 defines r3. we create a new chain
1380 // for L2 and L5.
1381 // Case #5:
1382 // L2: r2 = COPY r3
1383 // L3: r3 = COPY r1
1384 // L4: r1 = COPY r3
1385 // L5: r3 = COPY r2
1386 // COPY is found and is L4 which belongs to an existing chain, we add
1387 // L2 and L5 to this chain.
1388 LLVM_DEBUG(dbgs() << "MCP: Found spill: ");
1389 LLVM_DEBUG(MaybeSpill->dump());
1390 MachineInstr *MaybePrevReload =
1391 Tracker.findLastSeenUseInCopy(Def.asMCReg(), *TRI);
1392 auto Leader = ChainLeader.find(MaybePrevReload);
1393 MachineInstr *L = nullptr;
1394 if (Leader == ChainLeader.end() ||
1395 (MaybePrevReload && !IsChainedCopy(*MaybePrevReload, MI))) {
1396 L = &MI;
1397 assert(!SpillChain.count(L) &&
1398 "SpillChain should not have contained newly found chain");
1399 } else {
1400 assert(MaybePrevReload &&
1401 "Found a valid leader through nullptr should not happend");
1402 L = Leader->second;
1403 assert(SpillChain[L].size() > 0 &&
1404 "Existing chain's length should be larger than zero");
1405 }
1406 assert(!ChainLeader.count(&MI) && !ChainLeader.count(MaybeSpill) &&
1407 "Newly found paired spill-reload should not belong to any chain "
1408 "at this point");
1409 ChainLeader.insert({MaybeSpill, L});
1410 ChainLeader.insert({&MI, L});
1411 SpillChain[L].push_back(MaybeSpill);
1412 ReloadChain[L].push_back(&MI);
1413 LLVM_DEBUG(dbgs() << "MCP: Chain " << L << " now is:\n");
1414 LLVM_DEBUG(printSpillReloadChain(SpillChain, ReloadChain, L));
1415 } else if (MaybeSpill && !MaybeSpillIsChained) {
1416 // MaybeSpill is unable to pair with MI. That's to say adding MI makes
1417 // the chain invalid.
1418 // The COPY defines Src is no longer considered as a candidate of a
1419 // valid chain. Since we expect the Def of a spill copy isn't used by
1420 // any COPY instruction until a reload copy. For example:
1421 // L1: r1 = COPY r2
1422 // L2: r3 = COPY r1
1423 // If we later have
1424 // L1: r1 = COPY r2
1425 // L2: r3 = COPY r1
1426 // L3: r2 = COPY r1
1427 // L1 and L3 can't be a valid spill-reload pair.
1428 // Thus we keep the property#1.
1429 LLVM_DEBUG(dbgs() << "MCP: Not paired spill-reload:\n");
1430 LLVM_DEBUG(MaybeSpill->dump());
1431 LLVM_DEBUG(MI.dump());
1432 Tracker.clobberRegister(Src.asMCReg(), *TRI, *TII, UseCopyInstr);
1433 LLVM_DEBUG(dbgs() << "MCP: Removed tracking of " << printReg(Src, TRI)
1434 << "\n");
1435 }
1436 Tracker.trackCopy(&MI, *TRI, *TII, UseCopyInstr);
1437 }
1438
1439 for (auto I = SpillChain.begin(), E = SpillChain.end(); I != E; ++I) {
1440 auto &SC = I->second;
1441 assert(ReloadChain.count(I->first) &&
1442 "Reload chain of the same leader should exist");
1443 auto &RC = ReloadChain[I->first];
1444 TryFoldSpillageCopies(SC, RC);
1445 }
1446
1447 MaybeDeadCopies.clear();
1448 CopyDbgUsers.clear();
1449 Tracker.clear();
1450 }
1451
runOnMachineFunction(MachineFunction & MF)1452 bool MachineCopyPropagation::runOnMachineFunction(MachineFunction &MF) {
1453 if (skipFunction(MF.getFunction()))
1454 return false;
1455
1456 bool isSpillageCopyElimEnabled = false;
1457 switch (EnableSpillageCopyElimination) {
1458 case cl::BOU_UNSET:
1459 isSpillageCopyElimEnabled =
1460 MF.getSubtarget().enableSpillageCopyElimination();
1461 break;
1462 case cl::BOU_TRUE:
1463 isSpillageCopyElimEnabled = true;
1464 break;
1465 case cl::BOU_FALSE:
1466 isSpillageCopyElimEnabled = false;
1467 break;
1468 }
1469
1470 Changed = false;
1471
1472 TRI = MF.getSubtarget().getRegisterInfo();
1473 TII = MF.getSubtarget().getInstrInfo();
1474 MRI = &MF.getRegInfo();
1475
1476 for (MachineBasicBlock &MBB : MF) {
1477 if (isSpillageCopyElimEnabled)
1478 EliminateSpillageCopies(MBB);
1479 BackwardCopyPropagateBlock(MBB);
1480 ForwardCopyPropagateBlock(MBB);
1481 }
1482
1483 return Changed;
1484 }
1485
1486 MachineFunctionPass *
createMachineCopyPropagationPass(bool UseCopyInstr=false)1487 llvm::createMachineCopyPropagationPass(bool UseCopyInstr = false) {
1488 return new MachineCopyPropagation(UseCopyInstr);
1489 }
1490