xref: /freebsd/contrib/llvm-project/llvm/lib/Target/AArch64/AArch64InstrInfo.h (revision 47ef2a131091508e049ab10cad7f91a3c1342cd9)
1 //===- AArch64InstrInfo.h - AArch64 Instruction Information -----*- C++ -*-===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 //
9 // This file contains the AArch64 implementation of the TargetInstrInfo class.
10 //
11 //===----------------------------------------------------------------------===//
12 
13 #ifndef LLVM_LIB_TARGET_AARCH64_AARCH64INSTRINFO_H
14 #define LLVM_LIB_TARGET_AARCH64_AARCH64INSTRINFO_H
15 
16 #include "AArch64.h"
17 #include "AArch64RegisterInfo.h"
18 #include "llvm/CodeGen/TargetInstrInfo.h"
19 #include "llvm/Support/TypeSize.h"
20 #include <optional>
21 
22 #define GET_INSTRINFO_HEADER
23 #include "AArch64GenInstrInfo.inc"
24 
25 namespace llvm {
26 
27 class AArch64Subtarget;
28 
29 static const MachineMemOperand::Flags MOSuppressPair =
30     MachineMemOperand::MOTargetFlag1;
31 static const MachineMemOperand::Flags MOStridedAccess =
32     MachineMemOperand::MOTargetFlag2;
33 
34 #define FALKOR_STRIDED_ACCESS_MD "falkor.strided.access"
35 
36 // AArch64 MachineCombiner patterns
37 enum AArch64MachineCombinerPattern : unsigned {
38   // These are patterns used to reduce the length of dependence chain.
39   SUBADD_OP1 = MachineCombinerPattern::TARGET_PATTERN_START,
40   SUBADD_OP2,
41 
42   // These are multiply-add patterns matched by the AArch64 machine combiner.
43   MULADDW_OP1,
44   MULADDW_OP2,
45   MULSUBW_OP1,
46   MULSUBW_OP2,
47   MULADDWI_OP1,
48   MULSUBWI_OP1,
49   MULADDX_OP1,
50   MULADDX_OP2,
51   MULSUBX_OP1,
52   MULSUBX_OP2,
53   MULADDXI_OP1,
54   MULSUBXI_OP1,
55   // NEON integers vectors
56   MULADDv8i8_OP1,
57   MULADDv8i8_OP2,
58   MULADDv16i8_OP1,
59   MULADDv16i8_OP2,
60   MULADDv4i16_OP1,
61   MULADDv4i16_OP2,
62   MULADDv8i16_OP1,
63   MULADDv8i16_OP2,
64   MULADDv2i32_OP1,
65   MULADDv2i32_OP2,
66   MULADDv4i32_OP1,
67   MULADDv4i32_OP2,
68 
69   MULSUBv8i8_OP1,
70   MULSUBv8i8_OP2,
71   MULSUBv16i8_OP1,
72   MULSUBv16i8_OP2,
73   MULSUBv4i16_OP1,
74   MULSUBv4i16_OP2,
75   MULSUBv8i16_OP1,
76   MULSUBv8i16_OP2,
77   MULSUBv2i32_OP1,
78   MULSUBv2i32_OP2,
79   MULSUBv4i32_OP1,
80   MULSUBv4i32_OP2,
81 
82   MULADDv4i16_indexed_OP1,
83   MULADDv4i16_indexed_OP2,
84   MULADDv8i16_indexed_OP1,
85   MULADDv8i16_indexed_OP2,
86   MULADDv2i32_indexed_OP1,
87   MULADDv2i32_indexed_OP2,
88   MULADDv4i32_indexed_OP1,
89   MULADDv4i32_indexed_OP2,
90 
91   MULSUBv4i16_indexed_OP1,
92   MULSUBv4i16_indexed_OP2,
93   MULSUBv8i16_indexed_OP1,
94   MULSUBv8i16_indexed_OP2,
95   MULSUBv2i32_indexed_OP1,
96   MULSUBv2i32_indexed_OP2,
97   MULSUBv4i32_indexed_OP1,
98   MULSUBv4i32_indexed_OP2,
99 
100   // Floating Point
101   FMULADDH_OP1,
102   FMULADDH_OP2,
103   FMULSUBH_OP1,
104   FMULSUBH_OP2,
105   FMULADDS_OP1,
106   FMULADDS_OP2,
107   FMULSUBS_OP1,
108   FMULSUBS_OP2,
109   FMULADDD_OP1,
110   FMULADDD_OP2,
111   FMULSUBD_OP1,
112   FMULSUBD_OP2,
113   FNMULSUBH_OP1,
114   FNMULSUBS_OP1,
115   FNMULSUBD_OP1,
116   FMLAv1i32_indexed_OP1,
117   FMLAv1i32_indexed_OP2,
118   FMLAv1i64_indexed_OP1,
119   FMLAv1i64_indexed_OP2,
120   FMLAv4f16_OP1,
121   FMLAv4f16_OP2,
122   FMLAv8f16_OP1,
123   FMLAv8f16_OP2,
124   FMLAv2f32_OP2,
125   FMLAv2f32_OP1,
126   FMLAv2f64_OP1,
127   FMLAv2f64_OP2,
128   FMLAv4i16_indexed_OP1,
129   FMLAv4i16_indexed_OP2,
130   FMLAv8i16_indexed_OP1,
131   FMLAv8i16_indexed_OP2,
132   FMLAv2i32_indexed_OP1,
133   FMLAv2i32_indexed_OP2,
134   FMLAv2i64_indexed_OP1,
135   FMLAv2i64_indexed_OP2,
136   FMLAv4f32_OP1,
137   FMLAv4f32_OP2,
138   FMLAv4i32_indexed_OP1,
139   FMLAv4i32_indexed_OP2,
140   FMLSv1i32_indexed_OP2,
141   FMLSv1i64_indexed_OP2,
142   FMLSv4f16_OP1,
143   FMLSv4f16_OP2,
144   FMLSv8f16_OP1,
145   FMLSv8f16_OP2,
146   FMLSv2f32_OP1,
147   FMLSv2f32_OP2,
148   FMLSv2f64_OP1,
149   FMLSv2f64_OP2,
150   FMLSv4i16_indexed_OP1,
151   FMLSv4i16_indexed_OP2,
152   FMLSv8i16_indexed_OP1,
153   FMLSv8i16_indexed_OP2,
154   FMLSv2i32_indexed_OP1,
155   FMLSv2i32_indexed_OP2,
156   FMLSv2i64_indexed_OP1,
157   FMLSv2i64_indexed_OP2,
158   FMLSv4f32_OP1,
159   FMLSv4f32_OP2,
160   FMLSv4i32_indexed_OP1,
161   FMLSv4i32_indexed_OP2,
162 
163   FMULv2i32_indexed_OP1,
164   FMULv2i32_indexed_OP2,
165   FMULv2i64_indexed_OP1,
166   FMULv2i64_indexed_OP2,
167   FMULv4i16_indexed_OP1,
168   FMULv4i16_indexed_OP2,
169   FMULv4i32_indexed_OP1,
170   FMULv4i32_indexed_OP2,
171   FMULv8i16_indexed_OP1,
172   FMULv8i16_indexed_OP2,
173 
174   FNMADD,
175 };
176 class AArch64InstrInfo final : public AArch64GenInstrInfo {
177   const AArch64RegisterInfo RI;
178   const AArch64Subtarget &Subtarget;
179 
180 public:
181   explicit AArch64InstrInfo(const AArch64Subtarget &STI);
182 
183   /// getRegisterInfo - TargetInstrInfo is a superset of MRegister info.  As
184   /// such, whenever a client has an instance of instruction info, it should
185   /// always be able to get register info as well (through this method).
186   const AArch64RegisterInfo &getRegisterInfo() const { return RI; }
187 
188   unsigned getInstSizeInBytes(const MachineInstr &MI) const override;
189 
190   bool isAsCheapAsAMove(const MachineInstr &MI) const override;
191 
192   bool isCoalescableExtInstr(const MachineInstr &MI, Register &SrcReg,
193                              Register &DstReg, unsigned &SubIdx) const override;
194 
195   bool
196   areMemAccessesTriviallyDisjoint(const MachineInstr &MIa,
197                                   const MachineInstr &MIb) const override;
198 
199   Register isLoadFromStackSlot(const MachineInstr &MI,
200                                int &FrameIndex) const override;
201   Register isStoreToStackSlot(const MachineInstr &MI,
202                               int &FrameIndex) const override;
203 
204   /// Does this instruction set its full destination register to zero?
205   static bool isGPRZero(const MachineInstr &MI);
206 
207   /// Does this instruction rename a GPR without modifying bits?
208   static bool isGPRCopy(const MachineInstr &MI);
209 
210   /// Does this instruction rename an FPR without modifying bits?
211   static bool isFPRCopy(const MachineInstr &MI);
212 
213   /// Return true if pairing the given load or store is hinted to be
214   /// unprofitable.
215   static bool isLdStPairSuppressed(const MachineInstr &MI);
216 
217   /// Return true if the given load or store is a strided memory access.
218   static bool isStridedAccess(const MachineInstr &MI);
219 
220   /// Return true if it has an unscaled load/store offset.
221   static bool hasUnscaledLdStOffset(unsigned Opc);
222   static bool hasUnscaledLdStOffset(MachineInstr &MI) {
223     return hasUnscaledLdStOffset(MI.getOpcode());
224   }
225 
226   /// Returns the unscaled load/store for the scaled load/store opcode,
227   /// if there is a corresponding unscaled variant available.
228   static std::optional<unsigned> getUnscaledLdSt(unsigned Opc);
229 
230   /// Scaling factor for (scaled or unscaled) load or store.
231   static int getMemScale(unsigned Opc);
232   static int getMemScale(const MachineInstr &MI) {
233     return getMemScale(MI.getOpcode());
234   }
235 
236   /// Returns whether the instruction is a pre-indexed load.
237   static bool isPreLd(const MachineInstr &MI);
238 
239   /// Returns whether the instruction is a pre-indexed store.
240   static bool isPreSt(const MachineInstr &MI);
241 
242   /// Returns whether the instruction is a pre-indexed load/store.
243   static bool isPreLdSt(const MachineInstr &MI);
244 
245   /// Returns whether the instruction is a paired load/store.
246   static bool isPairedLdSt(const MachineInstr &MI);
247 
248   /// Returns the base register operator of a load/store.
249   static const MachineOperand &getLdStBaseOp(const MachineInstr &MI);
250 
251   /// Returns the immediate offset operator of a load/store.
252   static const MachineOperand &getLdStOffsetOp(const MachineInstr &MI);
253 
254   /// Returns whether the physical register is FP or NEON.
255   static bool isFpOrNEON(Register Reg);
256 
257   /// Returns whether the instruction is FP or NEON.
258   static bool isFpOrNEON(const MachineInstr &MI);
259 
260   /// Returns whether the instruction is in H form (16 bit operands)
261   static bool isHForm(const MachineInstr &MI);
262 
263   /// Returns whether the instruction is in Q form (128 bit operands)
264   static bool isQForm(const MachineInstr &MI);
265 
266   /// Returns whether the instruction can be compatible with non-zero BTYPE.
267   static bool hasBTISemantics(const MachineInstr &MI);
268 
269   /// Returns the index for the immediate for a given instruction.
270   static unsigned getLoadStoreImmIdx(unsigned Opc);
271 
272   /// Return true if pairing the given load or store may be paired with another.
273   static bool isPairableLdStInst(const MachineInstr &MI);
274 
275   /// Returns true if MI is one of the TCRETURN* instructions.
276   static bool isTailCallReturnInst(const MachineInstr &MI);
277 
278   /// Return the opcode that set flags when possible.  The caller is
279   /// responsible for ensuring the opc has a flag setting equivalent.
280   static unsigned convertToFlagSettingOpc(unsigned Opc);
281 
282   /// Return true if this is a load/store that can be potentially paired/merged.
283   bool isCandidateToMergeOrPair(const MachineInstr &MI) const;
284 
285   /// Hint that pairing the given load or store is unprofitable.
286   static void suppressLdStPair(MachineInstr &MI);
287 
288   std::optional<ExtAddrMode>
289   getAddrModeFromMemoryOp(const MachineInstr &MemI,
290                           const TargetRegisterInfo *TRI) const override;
291 
292   bool canFoldIntoAddrMode(const MachineInstr &MemI, Register Reg,
293                            const MachineInstr &AddrI,
294                            ExtAddrMode &AM) const override;
295 
296   MachineInstr *emitLdStWithAddr(MachineInstr &MemI,
297                                  const ExtAddrMode &AM) const override;
298 
299   bool getMemOperandsWithOffsetWidth(
300       const MachineInstr &MI, SmallVectorImpl<const MachineOperand *> &BaseOps,
301       int64_t &Offset, bool &OffsetIsScalable, LocationSize &Width,
302       const TargetRegisterInfo *TRI) const override;
303 
304   /// If \p OffsetIsScalable is set to 'true', the offset is scaled by `vscale`.
305   /// This is true for some SVE instructions like ldr/str that have a
306   /// 'reg + imm' addressing mode where the immediate is an index to the
307   /// scalable vector located at 'reg + imm * vscale x #bytes'.
308   bool getMemOperandWithOffsetWidth(const MachineInstr &MI,
309                                     const MachineOperand *&BaseOp,
310                                     int64_t &Offset, bool &OffsetIsScalable,
311                                     TypeSize &Width,
312                                     const TargetRegisterInfo *TRI) const;
313 
314   /// Return the immediate offset of the base register in a load/store \p LdSt.
315   MachineOperand &getMemOpBaseRegImmOfsOffsetOperand(MachineInstr &LdSt) const;
316 
317   /// Returns true if opcode \p Opc is a memory operation. If it is, set
318   /// \p Scale, \p Width, \p MinOffset, and \p MaxOffset accordingly.
319   ///
320   /// For unscaled instructions, \p Scale is set to 1.
321   static bool getMemOpInfo(unsigned Opcode, TypeSize &Scale, TypeSize &Width,
322                            int64_t &MinOffset, int64_t &MaxOffset);
323 
324   bool shouldClusterMemOps(ArrayRef<const MachineOperand *> BaseOps1,
325                            int64_t Offset1, bool OffsetIsScalable1,
326                            ArrayRef<const MachineOperand *> BaseOps2,
327                            int64_t Offset2, bool OffsetIsScalable2,
328                            unsigned ClusterSize,
329                            unsigned NumBytes) const override;
330 
331   void copyPhysRegTuple(MachineBasicBlock &MBB, MachineBasicBlock::iterator I,
332                         const DebugLoc &DL, MCRegister DestReg,
333                         MCRegister SrcReg, bool KillSrc, unsigned Opcode,
334                         llvm::ArrayRef<unsigned> Indices) const;
335   void copyGPRRegTuple(MachineBasicBlock &MBB, MachineBasicBlock::iterator I,
336                        DebugLoc DL, unsigned DestReg, unsigned SrcReg,
337                        bool KillSrc, unsigned Opcode, unsigned ZeroReg,
338                        llvm::ArrayRef<unsigned> Indices) const;
339   void copyPhysReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator I,
340                    const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg,
341                    bool KillSrc) const override;
342 
343   void storeRegToStackSlot(MachineBasicBlock &MBB,
344                            MachineBasicBlock::iterator MBBI, Register SrcReg,
345                            bool isKill, int FrameIndex,
346                            const TargetRegisterClass *RC,
347                            const TargetRegisterInfo *TRI,
348                            Register VReg) const override;
349 
350   void loadRegFromStackSlot(MachineBasicBlock &MBB,
351                             MachineBasicBlock::iterator MBBI, Register DestReg,
352                             int FrameIndex, const TargetRegisterClass *RC,
353                             const TargetRegisterInfo *TRI,
354                             Register VReg) const override;
355 
356   // This tells target independent code that it is okay to pass instructions
357   // with subreg operands to foldMemoryOperandImpl.
358   bool isSubregFoldable() const override { return true; }
359 
360   using TargetInstrInfo::foldMemoryOperandImpl;
361   MachineInstr *
362   foldMemoryOperandImpl(MachineFunction &MF, MachineInstr &MI,
363                         ArrayRef<unsigned> Ops,
364                         MachineBasicBlock::iterator InsertPt, int FrameIndex,
365                         LiveIntervals *LIS = nullptr,
366                         VirtRegMap *VRM = nullptr) const override;
367 
368   /// \returns true if a branch from an instruction with opcode \p BranchOpc
369   ///  bytes is capable of jumping to a position \p BrOffset bytes away.
370   bool isBranchOffsetInRange(unsigned BranchOpc,
371                              int64_t BrOffset) const override;
372 
373   MachineBasicBlock *getBranchDestBlock(const MachineInstr &MI) const override;
374 
375   void insertIndirectBranch(MachineBasicBlock &MBB,
376                             MachineBasicBlock &NewDestBB,
377                             MachineBasicBlock &RestoreBB, const DebugLoc &DL,
378                             int64_t BrOffset, RegScavenger *RS) const override;
379 
380   bool analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB,
381                      MachineBasicBlock *&FBB,
382                      SmallVectorImpl<MachineOperand> &Cond,
383                      bool AllowModify = false) const override;
384   bool analyzeBranchPredicate(MachineBasicBlock &MBB,
385                               MachineBranchPredicate &MBP,
386                               bool AllowModify) const override;
387   unsigned removeBranch(MachineBasicBlock &MBB,
388                         int *BytesRemoved = nullptr) const override;
389   unsigned insertBranch(MachineBasicBlock &MBB, MachineBasicBlock *TBB,
390                         MachineBasicBlock *FBB, ArrayRef<MachineOperand> Cond,
391                         const DebugLoc &DL,
392                         int *BytesAdded = nullptr) const override;
393 
394   std::unique_ptr<TargetInstrInfo::PipelinerLoopInfo>
395   analyzeLoopForPipelining(MachineBasicBlock *LoopBB) const override;
396 
397   bool
398   reverseBranchCondition(SmallVectorImpl<MachineOperand> &Cond) const override;
399   bool canInsertSelect(const MachineBasicBlock &, ArrayRef<MachineOperand> Cond,
400                        Register, Register, Register, int &, int &,
401                        int &) const override;
402   void insertSelect(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI,
403                     const DebugLoc &DL, Register DstReg,
404                     ArrayRef<MachineOperand> Cond, Register TrueReg,
405                     Register FalseReg) const override;
406 
407   void insertNoop(MachineBasicBlock &MBB,
408                   MachineBasicBlock::iterator MI) const override;
409 
410   MCInst getNop() const override;
411 
412   bool isSchedulingBoundary(const MachineInstr &MI,
413                             const MachineBasicBlock *MBB,
414                             const MachineFunction &MF) const override;
415 
416   /// analyzeCompare - For a comparison instruction, return the source registers
417   /// in SrcReg and SrcReg2, and the value it compares against in CmpValue.
418   /// Return true if the comparison instruction can be analyzed.
419   bool analyzeCompare(const MachineInstr &MI, Register &SrcReg,
420                       Register &SrcReg2, int64_t &CmpMask,
421                       int64_t &CmpValue) const override;
422   /// optimizeCompareInstr - Convert the instruction supplying the argument to
423   /// the comparison into one that sets the zero bit in the flags register.
424   bool optimizeCompareInstr(MachineInstr &CmpInstr, Register SrcReg,
425                             Register SrcReg2, int64_t CmpMask, int64_t CmpValue,
426                             const MachineRegisterInfo *MRI) const override;
427   bool optimizeCondBranch(MachineInstr &MI) const override;
428 
429   CombinerObjective getCombinerObjective(unsigned Pattern) const override;
430   /// Return true when a code sequence can improve throughput. It
431   /// should be called only for instructions in loops.
432   /// \param Pattern - combiner pattern
433   bool isThroughputPattern(unsigned Pattern) const override;
434   /// Return true when there is potentially a faster code sequence
435   /// for an instruction chain ending in ``Root``. All potential patterns are
436   /// listed in the ``Patterns`` array.
437   bool getMachineCombinerPatterns(MachineInstr &Root,
438                                   SmallVectorImpl<unsigned> &Patterns,
439                                   bool DoRegPressureReduce) const override;
440   /// Return true when Inst is associative and commutative so that it can be
441   /// reassociated. If Invert is true, then the inverse of Inst operation must
442   /// be checked.
443   bool isAssociativeAndCommutative(const MachineInstr &Inst,
444                                    bool Invert) const override;
445   /// When getMachineCombinerPatterns() finds patterns, this function generates
446   /// the instructions that could replace the original code sequence
447   void genAlternativeCodeSequence(
448       MachineInstr &Root, unsigned Pattern,
449       SmallVectorImpl<MachineInstr *> &InsInstrs,
450       SmallVectorImpl<MachineInstr *> &DelInstrs,
451       DenseMap<unsigned, unsigned> &InstrIdxForVirtReg) const override;
452   /// AArch64 supports MachineCombiner.
453   bool useMachineCombiner() const override;
454 
455   bool expandPostRAPseudo(MachineInstr &MI) const override;
456 
457   std::pair<unsigned, unsigned>
458   decomposeMachineOperandsTargetFlags(unsigned TF) const override;
459   ArrayRef<std::pair<unsigned, const char *>>
460   getSerializableDirectMachineOperandTargetFlags() const override;
461   ArrayRef<std::pair<unsigned, const char *>>
462   getSerializableBitmaskMachineOperandTargetFlags() const override;
463   ArrayRef<std::pair<MachineMemOperand::Flags, const char *>>
464   getSerializableMachineMemOperandTargetFlags() const override;
465 
466   bool isFunctionSafeToOutlineFrom(MachineFunction &MF,
467                                    bool OutlineFromLinkOnceODRs) const override;
468   std::optional<outliner::OutlinedFunction> getOutliningCandidateInfo(
469       std::vector<outliner::Candidate> &RepeatedSequenceLocs) const override;
470   void mergeOutliningCandidateAttributes(
471       Function &F, std::vector<outliner::Candidate> &Candidates) const override;
472   outliner::InstrType
473   getOutliningTypeImpl(MachineBasicBlock::iterator &MIT, unsigned Flags) const override;
474   SmallVector<
475       std::pair<MachineBasicBlock::iterator, MachineBasicBlock::iterator>>
476   getOutlinableRanges(MachineBasicBlock &MBB, unsigned &Flags) const override;
477   void buildOutlinedFrame(MachineBasicBlock &MBB, MachineFunction &MF,
478                           const outliner::OutlinedFunction &OF) const override;
479   MachineBasicBlock::iterator
480   insertOutlinedCall(Module &M, MachineBasicBlock &MBB,
481                      MachineBasicBlock::iterator &It, MachineFunction &MF,
482                      outliner::Candidate &C) const override;
483   bool shouldOutlineFromFunctionByDefault(MachineFunction &MF) const override;
484 
485   void buildClearRegister(Register Reg, MachineBasicBlock &MBB,
486                           MachineBasicBlock::iterator Iter, DebugLoc &DL,
487                           bool AllowSideEffects = true) const override;
488 
489   /// Returns the vector element size (B, H, S or D) of an SVE opcode.
490   uint64_t getElementSizeForOpcode(unsigned Opc) const;
491   /// Returns true if the opcode is for an SVE instruction that sets the
492   /// condition codes as if it's results had been fed to a PTEST instruction
493   /// along with the same general predicate.
494   bool isPTestLikeOpcode(unsigned Opc) const;
495   /// Returns true if the opcode is for an SVE WHILE## instruction.
496   bool isWhileOpcode(unsigned Opc) const;
497   /// Returns true if the instruction has a shift by immediate that can be
498   /// executed in one cycle less.
499   static bool isFalkorShiftExtFast(const MachineInstr &MI);
500   /// Return true if the instructions is a SEH instruciton used for unwinding
501   /// on Windows.
502   static bool isSEHInstruction(const MachineInstr &MI);
503 
504   std::optional<RegImmPair> isAddImmediate(const MachineInstr &MI,
505                                            Register Reg) const override;
506 
507   bool isFunctionSafeToSplit(const MachineFunction &MF) const override;
508 
509   bool isMBBSafeToSplitToCold(const MachineBasicBlock &MBB) const override;
510 
511   std::optional<ParamLoadedValue>
512   describeLoadedValue(const MachineInstr &MI, Register Reg) const override;
513 
514   unsigned int getTailDuplicateSize(CodeGenOptLevel OptLevel) const override;
515 
516   bool isExtendLikelyToBeFolded(MachineInstr &ExtMI,
517                                 MachineRegisterInfo &MRI) const override;
518 
519   static void decomposeStackOffsetForFrameOffsets(const StackOffset &Offset,
520                                                   int64_t &NumBytes,
521                                                   int64_t &NumPredicateVectors,
522                                                   int64_t &NumDataVectors);
523   static void decomposeStackOffsetForDwarfOffsets(const StackOffset &Offset,
524                                                   int64_t &ByteSized,
525                                                   int64_t &VGSized);
526 
527   // Return true if address of the form BaseReg + Scale * ScaledReg + Offset can
528   // be used for a load/store of NumBytes. BaseReg is always present and
529   // implicit.
530   bool isLegalAddressingMode(unsigned NumBytes, int64_t Offset,
531                              unsigned Scale) const;
532 
533   // Decrement the SP, issuing probes along the way. `TargetReg` is the new top
534   // of the stack. `FrameSetup` is passed as true, if the allocation is a part
535   // of constructing the activation frame of a function.
536   MachineBasicBlock::iterator probedStackAlloc(MachineBasicBlock::iterator MBBI,
537                                                Register TargetReg,
538                                                bool FrameSetup) const;
539 
540 #define GET_INSTRINFO_HELPER_DECLS
541 #include "AArch64GenInstrInfo.inc"
542 
543 protected:
544   /// If the specific machine instruction is an instruction that moves/copies
545   /// value from one register to another register return destination and source
546   /// registers as machine operands.
547   std::optional<DestSourcePair>
548   isCopyInstrImpl(const MachineInstr &MI) const override;
549   std::optional<DestSourcePair>
550   isCopyLikeInstrImpl(const MachineInstr &MI) const override;
551 
552 private:
553   unsigned getInstBundleLength(const MachineInstr &MI) const;
554 
555   /// Sets the offsets on outlined instructions in \p MBB which use SP
556   /// so that they will be valid post-outlining.
557   ///
558   /// \param MBB A \p MachineBasicBlock in an outlined function.
559   void fixupPostOutline(MachineBasicBlock &MBB) const;
560 
561   void instantiateCondBranch(MachineBasicBlock &MBB, const DebugLoc &DL,
562                              MachineBasicBlock *TBB,
563                              ArrayRef<MachineOperand> Cond) const;
564   bool substituteCmpToZero(MachineInstr &CmpInstr, unsigned SrcReg,
565                            const MachineRegisterInfo &MRI) const;
566   bool removeCmpToZeroOrOne(MachineInstr &CmpInstr, unsigned SrcReg,
567                             int CmpValue, const MachineRegisterInfo &MRI) const;
568 
569   /// Returns an unused general-purpose register which can be used for
570   /// constructing an outlined call if one exists. Returns 0 otherwise.
571   Register findRegisterToSaveLRTo(outliner::Candidate &C) const;
572 
573   /// Remove a ptest of a predicate-generating operation that already sets, or
574   /// can be made to set, the condition codes in an identical manner
575   bool optimizePTestInstr(MachineInstr *PTest, unsigned MaskReg,
576                           unsigned PredReg,
577                           const MachineRegisterInfo *MRI) const;
578   std::optional<unsigned>
579   canRemovePTestInstr(MachineInstr *PTest, MachineInstr *Mask,
580                       MachineInstr *Pred, const MachineRegisterInfo *MRI) const;
581 };
582 
583 struct UsedNZCV {
584   bool N = false;
585   bool Z = false;
586   bool C = false;
587   bool V = false;
588 
589   UsedNZCV() = default;
590 
591   UsedNZCV &operator|=(const UsedNZCV &UsedFlags) {
592     this->N |= UsedFlags.N;
593     this->Z |= UsedFlags.Z;
594     this->C |= UsedFlags.C;
595     this->V |= UsedFlags.V;
596     return *this;
597   }
598 };
599 
600 /// \returns Conditions flags used after \p CmpInstr in its MachineBB if  NZCV
601 /// flags are not alive in successors of the same \p CmpInstr and \p MI parent.
602 /// \returns std::nullopt otherwise.
603 ///
604 /// Collect instructions using that flags in \p CCUseInstrs if provided.
605 std::optional<UsedNZCV>
606 examineCFlagsUse(MachineInstr &MI, MachineInstr &CmpInstr,
607                  const TargetRegisterInfo &TRI,
608                  SmallVectorImpl<MachineInstr *> *CCUseInstrs = nullptr);
609 
610 /// Return true if there is an instruction /after/ \p DefMI and before \p UseMI
611 /// which either reads or clobbers NZCV.
612 bool isNZCVTouchedInInstructionRange(const MachineInstr &DefMI,
613                                      const MachineInstr &UseMI,
614                                      const TargetRegisterInfo *TRI);
615 
616 MCCFIInstruction createDefCFA(const TargetRegisterInfo &TRI, unsigned FrameReg,
617                               unsigned Reg, const StackOffset &Offset,
618                               bool LastAdjustmentWasScalable = true);
619 MCCFIInstruction createCFAOffset(const TargetRegisterInfo &MRI, unsigned Reg,
620                                  const StackOffset &OffsetFromDefCFA);
621 
622 /// emitFrameOffset - Emit instructions as needed to set DestReg to SrcReg
623 /// plus Offset.  This is intended to be used from within the prolog/epilog
624 /// insertion (PEI) pass, where a virtual scratch register may be allocated
625 /// if necessary, to be replaced by the scavenger at the end of PEI.
626 void emitFrameOffset(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI,
627                      const DebugLoc &DL, unsigned DestReg, unsigned SrcReg,
628                      StackOffset Offset, const TargetInstrInfo *TII,
629                      MachineInstr::MIFlag = MachineInstr::NoFlags,
630                      bool SetNZCV = false, bool NeedsWinCFI = false,
631                      bool *HasWinCFI = nullptr, bool EmitCFAOffset = false,
632                      StackOffset InitialOffset = {},
633                      unsigned FrameReg = AArch64::SP);
634 
635 /// rewriteAArch64FrameIndex - Rewrite MI to access 'Offset' bytes from the
636 /// FP. Return false if the offset could not be handled directly in MI, and
637 /// return the left-over portion by reference.
638 bool rewriteAArch64FrameIndex(MachineInstr &MI, unsigned FrameRegIdx,
639                               unsigned FrameReg, StackOffset &Offset,
640                               const AArch64InstrInfo *TII);
641 
642 /// Use to report the frame offset status in isAArch64FrameOffsetLegal.
643 enum AArch64FrameOffsetStatus {
644   AArch64FrameOffsetCannotUpdate = 0x0, ///< Offset cannot apply.
645   AArch64FrameOffsetIsLegal = 0x1,      ///< Offset is legal.
646   AArch64FrameOffsetCanUpdate = 0x2     ///< Offset can apply, at least partly.
647 };
648 
649 /// Check if the @p Offset is a valid frame offset for @p MI.
650 /// The returned value reports the validity of the frame offset for @p MI.
651 /// It uses the values defined by AArch64FrameOffsetStatus for that.
652 /// If result == AArch64FrameOffsetCannotUpdate, @p MI cannot be updated to
653 /// use an offset.eq
654 /// If result & AArch64FrameOffsetIsLegal, @p Offset can completely be
655 /// rewritten in @p MI.
656 /// If result & AArch64FrameOffsetCanUpdate, @p Offset contains the
657 /// amount that is off the limit of the legal offset.
658 /// If set, @p OutUseUnscaledOp will contain the whether @p MI should be
659 /// turned into an unscaled operator, which opcode is in @p OutUnscaledOp.
660 /// If set, @p EmittableOffset contains the amount that can be set in @p MI
661 /// (possibly with @p OutUnscaledOp if OutUseUnscaledOp is true) and that
662 /// is a legal offset.
663 int isAArch64FrameOffsetLegal(const MachineInstr &MI, StackOffset &Offset,
664                               bool *OutUseUnscaledOp = nullptr,
665                               unsigned *OutUnscaledOp = nullptr,
666                               int64_t *EmittableOffset = nullptr);
667 
668 static inline bool isUncondBranchOpcode(int Opc) { return Opc == AArch64::B; }
669 
670 static inline bool isCondBranchOpcode(int Opc) {
671   switch (Opc) {
672   case AArch64::Bcc:
673   case AArch64::CBZW:
674   case AArch64::CBZX:
675   case AArch64::CBNZW:
676   case AArch64::CBNZX:
677   case AArch64::TBZW:
678   case AArch64::TBZX:
679   case AArch64::TBNZW:
680   case AArch64::TBNZX:
681     return true;
682   default:
683     return false;
684   }
685 }
686 
687 static inline bool isIndirectBranchOpcode(int Opc) {
688   switch (Opc) {
689   case AArch64::BR:
690   case AArch64::BRAA:
691   case AArch64::BRAB:
692   case AArch64::BRAAZ:
693   case AArch64::BRABZ:
694     return true;
695   }
696   return false;
697 }
698 
699 static inline bool isPTrueOpcode(unsigned Opc) {
700   switch (Opc) {
701   case AArch64::PTRUE_B:
702   case AArch64::PTRUE_H:
703   case AArch64::PTRUE_S:
704   case AArch64::PTRUE_D:
705     return true;
706   default:
707     return false;
708   }
709 }
710 
711 /// Return opcode to be used for indirect calls.
712 unsigned getBLRCallOpcode(const MachineFunction &MF);
713 
714 /// Return XPAC opcode to be used for a ptrauth strip using the given key.
715 static inline unsigned getXPACOpcodeForKey(AArch64PACKey::ID K) {
716   using namespace AArch64PACKey;
717   switch (K) {
718   case IA: case IB: return AArch64::XPACI;
719   case DA: case DB: return AArch64::XPACD;
720   }
721   llvm_unreachable("Unhandled AArch64PACKey::ID enum");
722 }
723 
724 /// Return AUT opcode to be used for a ptrauth auth using the given key, or its
725 /// AUT*Z variant that doesn't take a discriminator operand, using zero instead.
726 static inline unsigned getAUTOpcodeForKey(AArch64PACKey::ID K, bool Zero) {
727   using namespace AArch64PACKey;
728   switch (K) {
729   case IA: return Zero ? AArch64::AUTIZA : AArch64::AUTIA;
730   case IB: return Zero ? AArch64::AUTIZB : AArch64::AUTIB;
731   case DA: return Zero ? AArch64::AUTDZA : AArch64::AUTDA;
732   case DB: return Zero ? AArch64::AUTDZB : AArch64::AUTDB;
733   }
734   llvm_unreachable("Unhandled AArch64PACKey::ID enum");
735 }
736 
737 /// Return PAC opcode to be used for a ptrauth sign using the given key, or its
738 /// PAC*Z variant that doesn't take a discriminator operand, using zero instead.
739 static inline unsigned getPACOpcodeForKey(AArch64PACKey::ID K, bool Zero) {
740   using namespace AArch64PACKey;
741   switch (K) {
742   case IA: return Zero ? AArch64::PACIZA : AArch64::PACIA;
743   case IB: return Zero ? AArch64::PACIZB : AArch64::PACIB;
744   case DA: return Zero ? AArch64::PACDZA : AArch64::PACDA;
745   case DB: return Zero ? AArch64::PACDZB : AArch64::PACDB;
746   }
747   llvm_unreachable("Unhandled AArch64PACKey::ID enum");
748 }
749 
750 // struct TSFlags {
751 #define TSFLAG_ELEMENT_SIZE_TYPE(X)      (X)        // 3-bits
752 #define TSFLAG_DESTRUCTIVE_INST_TYPE(X) ((X) << 3)  // 4-bits
753 #define TSFLAG_FALSE_LANE_TYPE(X)       ((X) << 7)  // 2-bits
754 #define TSFLAG_INSTR_FLAGS(X)           ((X) << 9)  // 2-bits
755 #define TSFLAG_SME_MATRIX_TYPE(X)       ((X) << 11) // 3-bits
756 // }
757 
758 namespace AArch64 {
759 
760 enum ElementSizeType {
761   ElementSizeMask = TSFLAG_ELEMENT_SIZE_TYPE(0x7),
762   ElementSizeNone = TSFLAG_ELEMENT_SIZE_TYPE(0x0),
763   ElementSizeB    = TSFLAG_ELEMENT_SIZE_TYPE(0x1),
764   ElementSizeH    = TSFLAG_ELEMENT_SIZE_TYPE(0x2),
765   ElementSizeS    = TSFLAG_ELEMENT_SIZE_TYPE(0x3),
766   ElementSizeD    = TSFLAG_ELEMENT_SIZE_TYPE(0x4),
767 };
768 
769 enum DestructiveInstType {
770   DestructiveInstTypeMask       = TSFLAG_DESTRUCTIVE_INST_TYPE(0xf),
771   NotDestructive                = TSFLAG_DESTRUCTIVE_INST_TYPE(0x0),
772   DestructiveOther              = TSFLAG_DESTRUCTIVE_INST_TYPE(0x1),
773   DestructiveUnary              = TSFLAG_DESTRUCTIVE_INST_TYPE(0x2),
774   DestructiveBinaryImm          = TSFLAG_DESTRUCTIVE_INST_TYPE(0x3),
775   DestructiveBinaryShImmUnpred  = TSFLAG_DESTRUCTIVE_INST_TYPE(0x4),
776   DestructiveBinary             = TSFLAG_DESTRUCTIVE_INST_TYPE(0x5),
777   DestructiveBinaryComm         = TSFLAG_DESTRUCTIVE_INST_TYPE(0x6),
778   DestructiveBinaryCommWithRev  = TSFLAG_DESTRUCTIVE_INST_TYPE(0x7),
779   DestructiveTernaryCommWithRev = TSFLAG_DESTRUCTIVE_INST_TYPE(0x8),
780   DestructiveUnaryPassthru      = TSFLAG_DESTRUCTIVE_INST_TYPE(0x9),
781 };
782 
783 enum FalseLaneType {
784   FalseLanesMask  = TSFLAG_FALSE_LANE_TYPE(0x3),
785   FalseLanesZero  = TSFLAG_FALSE_LANE_TYPE(0x1),
786   FalseLanesUndef = TSFLAG_FALSE_LANE_TYPE(0x2),
787 };
788 
789 // NOTE: This is a bit field.
790 static const uint64_t InstrFlagIsWhile     = TSFLAG_INSTR_FLAGS(0x1);
791 static const uint64_t InstrFlagIsPTestLike = TSFLAG_INSTR_FLAGS(0x2);
792 
793 enum SMEMatrixType {
794   SMEMatrixTypeMask = TSFLAG_SME_MATRIX_TYPE(0x7),
795   SMEMatrixNone     = TSFLAG_SME_MATRIX_TYPE(0x0),
796   SMEMatrixTileB    = TSFLAG_SME_MATRIX_TYPE(0x1),
797   SMEMatrixTileH    = TSFLAG_SME_MATRIX_TYPE(0x2),
798   SMEMatrixTileS    = TSFLAG_SME_MATRIX_TYPE(0x3),
799   SMEMatrixTileD    = TSFLAG_SME_MATRIX_TYPE(0x4),
800   SMEMatrixTileQ    = TSFLAG_SME_MATRIX_TYPE(0x5),
801   SMEMatrixArray    = TSFLAG_SME_MATRIX_TYPE(0x6),
802 };
803 
804 #undef TSFLAG_ELEMENT_SIZE_TYPE
805 #undef TSFLAG_DESTRUCTIVE_INST_TYPE
806 #undef TSFLAG_FALSE_LANE_TYPE
807 #undef TSFLAG_INSTR_FLAGS
808 #undef TSFLAG_SME_MATRIX_TYPE
809 
810 int getSVEPseudoMap(uint16_t Opcode);
811 int getSVERevInstr(uint16_t Opcode);
812 int getSVENonRevInstr(uint16_t Opcode);
813 
814 int getSMEPseudoMap(uint16_t Opcode);
815 }
816 
817 } // end namespace llvm
818 
819 #endif
820