xref: /freebsd/contrib/llvm-project/llvm/lib/Target/AArch64/AArch64ISelLowering.h (revision 8bcb0991864975618c09697b1aca10683346d9f0)
10b57cec5SDimitry Andric //==-- AArch64ISelLowering.h - AArch64 DAG Lowering Interface ----*- C++ -*-==//
20b57cec5SDimitry Andric //
30b57cec5SDimitry Andric // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
40b57cec5SDimitry Andric // See https://llvm.org/LICENSE.txt for license information.
50b57cec5SDimitry Andric // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
60b57cec5SDimitry Andric //
70b57cec5SDimitry Andric //===----------------------------------------------------------------------===//
80b57cec5SDimitry Andric //
90b57cec5SDimitry Andric // This file defines the interfaces that AArch64 uses to lower LLVM code into a
100b57cec5SDimitry Andric // selection DAG.
110b57cec5SDimitry Andric //
120b57cec5SDimitry Andric //===----------------------------------------------------------------------===//
130b57cec5SDimitry Andric 
140b57cec5SDimitry Andric #ifndef LLVM_LIB_TARGET_AARCH64_AARCH64ISELLOWERING_H
150b57cec5SDimitry Andric #define LLVM_LIB_TARGET_AARCH64_AARCH64ISELLOWERING_H
160b57cec5SDimitry Andric 
170b57cec5SDimitry Andric #include "AArch64.h"
180b57cec5SDimitry Andric #include "llvm/CodeGen/CallingConvLower.h"
190b57cec5SDimitry Andric #include "llvm/CodeGen/SelectionDAG.h"
200b57cec5SDimitry Andric #include "llvm/CodeGen/TargetLowering.h"
210b57cec5SDimitry Andric #include "llvm/IR/CallingConv.h"
220b57cec5SDimitry Andric #include "llvm/IR/Instruction.h"
230b57cec5SDimitry Andric 
240b57cec5SDimitry Andric namespace llvm {
250b57cec5SDimitry Andric 
260b57cec5SDimitry Andric namespace AArch64ISD {
270b57cec5SDimitry Andric 
280b57cec5SDimitry Andric enum NodeType : unsigned {
290b57cec5SDimitry Andric   FIRST_NUMBER = ISD::BUILTIN_OP_END,
300b57cec5SDimitry Andric   WrapperLarge, // 4-instruction MOVZ/MOVK sequence for 64-bit addresses.
310b57cec5SDimitry Andric   CALL,         // Function call.
320b57cec5SDimitry Andric 
330b57cec5SDimitry Andric   // Produces the full sequence of instructions for getting the thread pointer
340b57cec5SDimitry Andric   // offset of a variable into X0, using the TLSDesc model.
350b57cec5SDimitry Andric   TLSDESC_CALLSEQ,
360b57cec5SDimitry Andric   ADRP,     // Page address of a TargetGlobalAddress operand.
370b57cec5SDimitry Andric   ADR,      // ADR
380b57cec5SDimitry Andric   ADDlow,   // Add the low 12 bits of a TargetGlobalAddress operand.
390b57cec5SDimitry Andric   LOADgot,  // Load from automatically generated descriptor (e.g. Global
400b57cec5SDimitry Andric             // Offset Table, TLS record).
410b57cec5SDimitry Andric   RET_FLAG, // Return with a flag operand. Operand 0 is the chain operand.
420b57cec5SDimitry Andric   BRCOND,   // Conditional branch instruction; "b.cond".
430b57cec5SDimitry Andric   CSEL,
440b57cec5SDimitry Andric   FCSEL, // Conditional move instruction.
450b57cec5SDimitry Andric   CSINV, // Conditional select invert.
460b57cec5SDimitry Andric   CSNEG, // Conditional select negate.
470b57cec5SDimitry Andric   CSINC, // Conditional select increment.
480b57cec5SDimitry Andric 
490b57cec5SDimitry Andric   // Pointer to the thread's local storage area. Materialised from TPIDR_EL0 on
500b57cec5SDimitry Andric   // ELF.
510b57cec5SDimitry Andric   THREAD_POINTER,
520b57cec5SDimitry Andric   ADC,
530b57cec5SDimitry Andric   SBC, // adc, sbc instructions
540b57cec5SDimitry Andric 
550b57cec5SDimitry Andric   // Arithmetic instructions which write flags.
560b57cec5SDimitry Andric   ADDS,
570b57cec5SDimitry Andric   SUBS,
580b57cec5SDimitry Andric   ADCS,
590b57cec5SDimitry Andric   SBCS,
600b57cec5SDimitry Andric   ANDS,
610b57cec5SDimitry Andric 
620b57cec5SDimitry Andric   // Conditional compares. Operands: left,right,falsecc,cc,flags
630b57cec5SDimitry Andric   CCMP,
640b57cec5SDimitry Andric   CCMN,
650b57cec5SDimitry Andric   FCCMP,
660b57cec5SDimitry Andric 
670b57cec5SDimitry Andric   // Floating point comparison
680b57cec5SDimitry Andric   FCMP,
690b57cec5SDimitry Andric 
700b57cec5SDimitry Andric   // Scalar extract
710b57cec5SDimitry Andric   EXTR,
720b57cec5SDimitry Andric 
730b57cec5SDimitry Andric   // Scalar-to-vector duplication
740b57cec5SDimitry Andric   DUP,
750b57cec5SDimitry Andric   DUPLANE8,
760b57cec5SDimitry Andric   DUPLANE16,
770b57cec5SDimitry Andric   DUPLANE32,
780b57cec5SDimitry Andric   DUPLANE64,
790b57cec5SDimitry Andric 
800b57cec5SDimitry Andric   // Vector immedate moves
810b57cec5SDimitry Andric   MOVI,
820b57cec5SDimitry Andric   MOVIshift,
830b57cec5SDimitry Andric   MOVIedit,
840b57cec5SDimitry Andric   MOVImsl,
850b57cec5SDimitry Andric   FMOV,
860b57cec5SDimitry Andric   MVNIshift,
870b57cec5SDimitry Andric   MVNImsl,
880b57cec5SDimitry Andric 
890b57cec5SDimitry Andric   // Vector immediate ops
900b57cec5SDimitry Andric   BICi,
910b57cec5SDimitry Andric   ORRi,
920b57cec5SDimitry Andric 
930b57cec5SDimitry Andric   // Vector bit select: similar to ISD::VSELECT but not all bits within an
940b57cec5SDimitry Andric   // element must be identical.
950b57cec5SDimitry Andric   BSL,
960b57cec5SDimitry Andric 
970b57cec5SDimitry Andric   // Vector arithmetic negation
980b57cec5SDimitry Andric   NEG,
990b57cec5SDimitry Andric 
1000b57cec5SDimitry Andric   // Vector shuffles
1010b57cec5SDimitry Andric   ZIP1,
1020b57cec5SDimitry Andric   ZIP2,
1030b57cec5SDimitry Andric   UZP1,
1040b57cec5SDimitry Andric   UZP2,
1050b57cec5SDimitry Andric   TRN1,
1060b57cec5SDimitry Andric   TRN2,
1070b57cec5SDimitry Andric   REV16,
1080b57cec5SDimitry Andric   REV32,
1090b57cec5SDimitry Andric   REV64,
1100b57cec5SDimitry Andric   EXT,
1110b57cec5SDimitry Andric 
1120b57cec5SDimitry Andric   // Vector shift by scalar
1130b57cec5SDimitry Andric   VSHL,
1140b57cec5SDimitry Andric   VLSHR,
1150b57cec5SDimitry Andric   VASHR,
1160b57cec5SDimitry Andric 
1170b57cec5SDimitry Andric   // Vector shift by scalar (again)
1180b57cec5SDimitry Andric   SQSHL_I,
1190b57cec5SDimitry Andric   UQSHL_I,
1200b57cec5SDimitry Andric   SQSHLU_I,
1210b57cec5SDimitry Andric   SRSHR_I,
1220b57cec5SDimitry Andric   URSHR_I,
1230b57cec5SDimitry Andric 
1240b57cec5SDimitry Andric   // Vector comparisons
1250b57cec5SDimitry Andric   CMEQ,
1260b57cec5SDimitry Andric   CMGE,
1270b57cec5SDimitry Andric   CMGT,
1280b57cec5SDimitry Andric   CMHI,
1290b57cec5SDimitry Andric   CMHS,
1300b57cec5SDimitry Andric   FCMEQ,
1310b57cec5SDimitry Andric   FCMGE,
1320b57cec5SDimitry Andric   FCMGT,
1330b57cec5SDimitry Andric 
1340b57cec5SDimitry Andric   // Vector zero comparisons
1350b57cec5SDimitry Andric   CMEQz,
1360b57cec5SDimitry Andric   CMGEz,
1370b57cec5SDimitry Andric   CMGTz,
1380b57cec5SDimitry Andric   CMLEz,
1390b57cec5SDimitry Andric   CMLTz,
1400b57cec5SDimitry Andric   FCMEQz,
1410b57cec5SDimitry Andric   FCMGEz,
1420b57cec5SDimitry Andric   FCMGTz,
1430b57cec5SDimitry Andric   FCMLEz,
1440b57cec5SDimitry Andric   FCMLTz,
1450b57cec5SDimitry Andric 
1460b57cec5SDimitry Andric   // Vector across-lanes addition
1470b57cec5SDimitry Andric   // Only the lower result lane is defined.
1480b57cec5SDimitry Andric   SADDV,
1490b57cec5SDimitry Andric   UADDV,
1500b57cec5SDimitry Andric 
1510b57cec5SDimitry Andric   // Vector across-lanes min/max
1520b57cec5SDimitry Andric   // Only the lower result lane is defined.
1530b57cec5SDimitry Andric   SMINV,
1540b57cec5SDimitry Andric   UMINV,
1550b57cec5SDimitry Andric   SMAXV,
1560b57cec5SDimitry Andric   UMAXV,
1570b57cec5SDimitry Andric 
1580b57cec5SDimitry Andric   // Vector bitwise negation
1590b57cec5SDimitry Andric   NOT,
1600b57cec5SDimitry Andric 
1610b57cec5SDimitry Andric   // Vector bitwise selection
1620b57cec5SDimitry Andric   BIT,
1630b57cec5SDimitry Andric 
1640b57cec5SDimitry Andric   // Compare-and-branch
1650b57cec5SDimitry Andric   CBZ,
1660b57cec5SDimitry Andric   CBNZ,
1670b57cec5SDimitry Andric   TBZ,
1680b57cec5SDimitry Andric   TBNZ,
1690b57cec5SDimitry Andric 
1700b57cec5SDimitry Andric   // Tail calls
1710b57cec5SDimitry Andric   TC_RETURN,
1720b57cec5SDimitry Andric 
1730b57cec5SDimitry Andric   // Custom prefetch handling
1740b57cec5SDimitry Andric   PREFETCH,
1750b57cec5SDimitry Andric 
1760b57cec5SDimitry Andric   // {s|u}int to FP within a FP register.
1770b57cec5SDimitry Andric   SITOF,
1780b57cec5SDimitry Andric   UITOF,
1790b57cec5SDimitry Andric 
1800b57cec5SDimitry Andric   /// Natural vector cast. ISD::BITCAST is not natural in the big-endian
1810b57cec5SDimitry Andric   /// world w.r.t vectors; which causes additional REV instructions to be
1820b57cec5SDimitry Andric   /// generated to compensate for the byte-swapping. But sometimes we do
1830b57cec5SDimitry Andric   /// need to re-interpret the data in SIMD vector registers in big-endian
1840b57cec5SDimitry Andric   /// mode without emitting such REV instructions.
1850b57cec5SDimitry Andric   NVCAST,
1860b57cec5SDimitry Andric 
1870b57cec5SDimitry Andric   SMULL,
1880b57cec5SDimitry Andric   UMULL,
1890b57cec5SDimitry Andric 
1900b57cec5SDimitry Andric   // Reciprocal estimates and steps.
1910b57cec5SDimitry Andric   FRECPE, FRECPS,
1920b57cec5SDimitry Andric   FRSQRTE, FRSQRTS,
1930b57cec5SDimitry Andric 
194*8bcb0991SDimitry Andric   SUNPKHI,
195*8bcb0991SDimitry Andric   SUNPKLO,
196*8bcb0991SDimitry Andric   UUNPKHI,
197*8bcb0991SDimitry Andric   UUNPKLO,
198*8bcb0991SDimitry Andric 
1990b57cec5SDimitry Andric   // NEON Load/Store with post-increment base updates
2000b57cec5SDimitry Andric   LD2post = ISD::FIRST_TARGET_MEMORY_OPCODE,
2010b57cec5SDimitry Andric   LD3post,
2020b57cec5SDimitry Andric   LD4post,
2030b57cec5SDimitry Andric   ST2post,
2040b57cec5SDimitry Andric   ST3post,
2050b57cec5SDimitry Andric   ST4post,
2060b57cec5SDimitry Andric   LD1x2post,
2070b57cec5SDimitry Andric   LD1x3post,
2080b57cec5SDimitry Andric   LD1x4post,
2090b57cec5SDimitry Andric   ST1x2post,
2100b57cec5SDimitry Andric   ST1x3post,
2110b57cec5SDimitry Andric   ST1x4post,
2120b57cec5SDimitry Andric   LD1DUPpost,
2130b57cec5SDimitry Andric   LD2DUPpost,
2140b57cec5SDimitry Andric   LD3DUPpost,
2150b57cec5SDimitry Andric   LD4DUPpost,
2160b57cec5SDimitry Andric   LD1LANEpost,
2170b57cec5SDimitry Andric   LD2LANEpost,
2180b57cec5SDimitry Andric   LD3LANEpost,
2190b57cec5SDimitry Andric   LD4LANEpost,
2200b57cec5SDimitry Andric   ST2LANEpost,
2210b57cec5SDimitry Andric   ST3LANEpost,
2220b57cec5SDimitry Andric   ST4LANEpost,
2230b57cec5SDimitry Andric 
2240b57cec5SDimitry Andric   STG,
2250b57cec5SDimitry Andric   STZG,
2260b57cec5SDimitry Andric   ST2G,
2270b57cec5SDimitry Andric   STZ2G
2280b57cec5SDimitry Andric 
2290b57cec5SDimitry Andric };
2300b57cec5SDimitry Andric 
2310b57cec5SDimitry Andric } // end namespace AArch64ISD
2320b57cec5SDimitry Andric 
2330b57cec5SDimitry Andric namespace {
2340b57cec5SDimitry Andric 
2350b57cec5SDimitry Andric // Any instruction that defines a 32-bit result zeros out the high half of the
2360b57cec5SDimitry Andric // register. Truncate can be lowered to EXTRACT_SUBREG. CopyFromReg may
2370b57cec5SDimitry Andric // be copying from a truncate. But any other 32-bit operation will zero-extend
2380b57cec5SDimitry Andric // up to 64 bits.
2390b57cec5SDimitry Andric // FIXME: X86 also checks for CMOV here. Do we need something similar?
2400b57cec5SDimitry Andric static inline bool isDef32(const SDNode &N) {
2410b57cec5SDimitry Andric   unsigned Opc = N.getOpcode();
2420b57cec5SDimitry Andric   return Opc != ISD::TRUNCATE && Opc != TargetOpcode::EXTRACT_SUBREG &&
2430b57cec5SDimitry Andric          Opc != ISD::CopyFromReg;
2440b57cec5SDimitry Andric }
2450b57cec5SDimitry Andric 
2460b57cec5SDimitry Andric } // end anonymous namespace
2470b57cec5SDimitry Andric 
2480b57cec5SDimitry Andric class AArch64Subtarget;
2490b57cec5SDimitry Andric class AArch64TargetMachine;
2500b57cec5SDimitry Andric 
2510b57cec5SDimitry Andric class AArch64TargetLowering : public TargetLowering {
2520b57cec5SDimitry Andric public:
2530b57cec5SDimitry Andric   explicit AArch64TargetLowering(const TargetMachine &TM,
2540b57cec5SDimitry Andric                                  const AArch64Subtarget &STI);
2550b57cec5SDimitry Andric 
2560b57cec5SDimitry Andric   /// Selects the correct CCAssignFn for a given CallingConvention value.
2570b57cec5SDimitry Andric   CCAssignFn *CCAssignFnForCall(CallingConv::ID CC, bool IsVarArg) const;
2580b57cec5SDimitry Andric 
2590b57cec5SDimitry Andric   /// Selects the correct CCAssignFn for a given CallingConvention value.
2600b57cec5SDimitry Andric   CCAssignFn *CCAssignFnForReturn(CallingConv::ID CC) const;
2610b57cec5SDimitry Andric 
2620b57cec5SDimitry Andric   /// Determine which of the bits specified in Mask are known to be either zero
2630b57cec5SDimitry Andric   /// or one and return them in the KnownZero/KnownOne bitsets.
2640b57cec5SDimitry Andric   void computeKnownBitsForTargetNode(const SDValue Op, KnownBits &Known,
2650b57cec5SDimitry Andric                                      const APInt &DemandedElts,
2660b57cec5SDimitry Andric                                      const SelectionDAG &DAG,
2670b57cec5SDimitry Andric                                      unsigned Depth = 0) const override;
2680b57cec5SDimitry Andric 
269*8bcb0991SDimitry Andric   MVT getPointerTy(const DataLayout &DL, uint32_t AS = 0) const override {
270*8bcb0991SDimitry Andric     // Returning i64 unconditionally here (i.e. even for ILP32) means that the
271*8bcb0991SDimitry Andric     // *DAG* representation of pointers will always be 64-bits. They will be
272*8bcb0991SDimitry Andric     // truncated and extended when transferred to memory, but the 64-bit DAG
273*8bcb0991SDimitry Andric     // allows us to use AArch64's addressing modes much more easily.
274*8bcb0991SDimitry Andric     return MVT::getIntegerVT(64);
275*8bcb0991SDimitry Andric   }
276*8bcb0991SDimitry Andric 
2770b57cec5SDimitry Andric   bool targetShrinkDemandedConstant(SDValue Op, const APInt &Demanded,
2780b57cec5SDimitry Andric                                     TargetLoweringOpt &TLO) const override;
2790b57cec5SDimitry Andric 
2800b57cec5SDimitry Andric   MVT getScalarShiftAmountTy(const DataLayout &DL, EVT) const override;
2810b57cec5SDimitry Andric 
2820b57cec5SDimitry Andric   /// Returns true if the target allows unaligned memory accesses of the
2830b57cec5SDimitry Andric   /// specified type.
2840b57cec5SDimitry Andric   bool allowsMisalignedMemoryAccesses(
2850b57cec5SDimitry Andric       EVT VT, unsigned AddrSpace = 0, unsigned Align = 1,
2860b57cec5SDimitry Andric       MachineMemOperand::Flags Flags = MachineMemOperand::MONone,
2870b57cec5SDimitry Andric       bool *Fast = nullptr) const override;
288*8bcb0991SDimitry Andric   /// LLT variant.
289*8bcb0991SDimitry Andric   bool allowsMisalignedMemoryAccesses(
290*8bcb0991SDimitry Andric     LLT Ty, unsigned AddrSpace, unsigned Align, MachineMemOperand::Flags Flags,
291*8bcb0991SDimitry Andric     bool *Fast = nullptr) const override;
2920b57cec5SDimitry Andric 
2930b57cec5SDimitry Andric   /// Provide custom lowering hooks for some operations.
2940b57cec5SDimitry Andric   SDValue LowerOperation(SDValue Op, SelectionDAG &DAG) const override;
2950b57cec5SDimitry Andric 
2960b57cec5SDimitry Andric   const char *getTargetNodeName(unsigned Opcode) const override;
2970b57cec5SDimitry Andric 
2980b57cec5SDimitry Andric   SDValue PerformDAGCombine(SDNode *N, DAGCombinerInfo &DCI) const override;
2990b57cec5SDimitry Andric 
3000b57cec5SDimitry Andric   /// Returns true if a cast between SrcAS and DestAS is a noop.
3010b57cec5SDimitry Andric   bool isNoopAddrSpaceCast(unsigned SrcAS, unsigned DestAS) const override {
3020b57cec5SDimitry Andric     // Addrspacecasts are always noops.
3030b57cec5SDimitry Andric     return true;
3040b57cec5SDimitry Andric   }
3050b57cec5SDimitry Andric 
3060b57cec5SDimitry Andric   /// This method returns a target specific FastISel object, or null if the
3070b57cec5SDimitry Andric   /// target does not support "fast" ISel.
3080b57cec5SDimitry Andric   FastISel *createFastISel(FunctionLoweringInfo &funcInfo,
3090b57cec5SDimitry Andric                            const TargetLibraryInfo *libInfo) const override;
3100b57cec5SDimitry Andric 
3110b57cec5SDimitry Andric   bool isOffsetFoldingLegal(const GlobalAddressSDNode *GA) const override;
3120b57cec5SDimitry Andric 
3130b57cec5SDimitry Andric   bool isFPImmLegal(const APFloat &Imm, EVT VT,
3140b57cec5SDimitry Andric                     bool ForCodeSize) const override;
3150b57cec5SDimitry Andric 
3160b57cec5SDimitry Andric   /// Return true if the given shuffle mask can be codegen'd directly, or if it
3170b57cec5SDimitry Andric   /// should be stack expanded.
3180b57cec5SDimitry Andric   bool isShuffleMaskLegal(ArrayRef<int> M, EVT VT) const override;
3190b57cec5SDimitry Andric 
3200b57cec5SDimitry Andric   /// Return the ISD::SETCC ValueType.
3210b57cec5SDimitry Andric   EVT getSetCCResultType(const DataLayout &DL, LLVMContext &Context,
3220b57cec5SDimitry Andric                          EVT VT) const override;
3230b57cec5SDimitry Andric 
3240b57cec5SDimitry Andric   SDValue ReconstructShuffle(SDValue Op, SelectionDAG &DAG) const;
3250b57cec5SDimitry Andric 
3260b57cec5SDimitry Andric   MachineBasicBlock *EmitF128CSEL(MachineInstr &MI,
3270b57cec5SDimitry Andric                                   MachineBasicBlock *BB) const;
3280b57cec5SDimitry Andric 
3290b57cec5SDimitry Andric   MachineBasicBlock *EmitLoweredCatchRet(MachineInstr &MI,
3300b57cec5SDimitry Andric                                            MachineBasicBlock *BB) const;
3310b57cec5SDimitry Andric 
3320b57cec5SDimitry Andric   MachineBasicBlock *EmitLoweredCatchPad(MachineInstr &MI,
3330b57cec5SDimitry Andric                                          MachineBasicBlock *BB) const;
3340b57cec5SDimitry Andric 
3350b57cec5SDimitry Andric   MachineBasicBlock *
3360b57cec5SDimitry Andric   EmitInstrWithCustomInserter(MachineInstr &MI,
3370b57cec5SDimitry Andric                               MachineBasicBlock *MBB) const override;
3380b57cec5SDimitry Andric 
3390b57cec5SDimitry Andric   bool getTgtMemIntrinsic(IntrinsicInfo &Info, const CallInst &I,
3400b57cec5SDimitry Andric                           MachineFunction &MF,
3410b57cec5SDimitry Andric                           unsigned Intrinsic) const override;
3420b57cec5SDimitry Andric 
3430b57cec5SDimitry Andric   bool shouldReduceLoadWidth(SDNode *Load, ISD::LoadExtType ExtTy,
3440b57cec5SDimitry Andric                              EVT NewVT) const override;
3450b57cec5SDimitry Andric 
3460b57cec5SDimitry Andric   bool isTruncateFree(Type *Ty1, Type *Ty2) const override;
3470b57cec5SDimitry Andric   bool isTruncateFree(EVT VT1, EVT VT2) const override;
3480b57cec5SDimitry Andric 
3490b57cec5SDimitry Andric   bool isProfitableToHoist(Instruction *I) const override;
3500b57cec5SDimitry Andric 
3510b57cec5SDimitry Andric   bool isZExtFree(Type *Ty1, Type *Ty2) const override;
3520b57cec5SDimitry Andric   bool isZExtFree(EVT VT1, EVT VT2) const override;
3530b57cec5SDimitry Andric   bool isZExtFree(SDValue Val, EVT VT2) const override;
3540b57cec5SDimitry Andric 
3550b57cec5SDimitry Andric   bool shouldSinkOperands(Instruction *I,
3560b57cec5SDimitry Andric                           SmallVectorImpl<Use *> &Ops) const override;
3570b57cec5SDimitry Andric 
3580b57cec5SDimitry Andric   bool hasPairedLoad(EVT LoadedType, unsigned &RequiredAligment) const override;
3590b57cec5SDimitry Andric 
3600b57cec5SDimitry Andric   unsigned getMaxSupportedInterleaveFactor() const override { return 4; }
3610b57cec5SDimitry Andric 
3620b57cec5SDimitry Andric   bool lowerInterleavedLoad(LoadInst *LI,
3630b57cec5SDimitry Andric                             ArrayRef<ShuffleVectorInst *> Shuffles,
3640b57cec5SDimitry Andric                             ArrayRef<unsigned> Indices,
3650b57cec5SDimitry Andric                             unsigned Factor) const override;
3660b57cec5SDimitry Andric   bool lowerInterleavedStore(StoreInst *SI, ShuffleVectorInst *SVI,
3670b57cec5SDimitry Andric                              unsigned Factor) const override;
3680b57cec5SDimitry Andric 
3690b57cec5SDimitry Andric   bool isLegalAddImmediate(int64_t) const override;
3700b57cec5SDimitry Andric   bool isLegalICmpImmediate(int64_t) const override;
3710b57cec5SDimitry Andric 
3720b57cec5SDimitry Andric   bool shouldConsiderGEPOffsetSplit() const override;
3730b57cec5SDimitry Andric 
3740b57cec5SDimitry Andric   EVT getOptimalMemOpType(uint64_t Size, unsigned DstAlign, unsigned SrcAlign,
3750b57cec5SDimitry Andric                           bool IsMemset, bool ZeroMemset, bool MemcpyStrSrc,
3760b57cec5SDimitry Andric                           const AttributeList &FuncAttributes) const override;
3770b57cec5SDimitry Andric 
378*8bcb0991SDimitry Andric   LLT getOptimalMemOpLLT(uint64_t Size, unsigned DstAlign, unsigned SrcAlign,
379*8bcb0991SDimitry Andric                           bool IsMemset, bool ZeroMemset, bool MemcpyStrSrc,
380*8bcb0991SDimitry Andric                           const AttributeList &FuncAttributes) const override;
381*8bcb0991SDimitry Andric 
3820b57cec5SDimitry Andric   /// Return true if the addressing mode represented by AM is legal for this
3830b57cec5SDimitry Andric   /// target, for a load/store of the specified type.
3840b57cec5SDimitry Andric   bool isLegalAddressingMode(const DataLayout &DL, const AddrMode &AM, Type *Ty,
3850b57cec5SDimitry Andric                              unsigned AS,
3860b57cec5SDimitry Andric                              Instruction *I = nullptr) const override;
3870b57cec5SDimitry Andric 
3880b57cec5SDimitry Andric   /// Return the cost of the scaling factor used in the addressing
3890b57cec5SDimitry Andric   /// mode represented by AM for this target, for a load/store
3900b57cec5SDimitry Andric   /// of the specified type.
3910b57cec5SDimitry Andric   /// If the AM is supported, the return value must be >= 0.
3920b57cec5SDimitry Andric   /// If the AM is not supported, it returns a negative value.
3930b57cec5SDimitry Andric   int getScalingFactorCost(const DataLayout &DL, const AddrMode &AM, Type *Ty,
3940b57cec5SDimitry Andric                            unsigned AS) const override;
3950b57cec5SDimitry Andric 
3960b57cec5SDimitry Andric   /// Return true if an FMA operation is faster than a pair of fmul and fadd
3970b57cec5SDimitry Andric   /// instructions. fmuladd intrinsics will be expanded to FMAs when this method
3980b57cec5SDimitry Andric   /// returns true, otherwise fmuladd is expanded to fmul + fadd.
3990b57cec5SDimitry Andric   bool isFMAFasterThanFMulAndFAdd(EVT VT) const override;
4000b57cec5SDimitry Andric 
4010b57cec5SDimitry Andric   const MCPhysReg *getScratchRegisters(CallingConv::ID CC) const override;
4020b57cec5SDimitry Andric 
4030b57cec5SDimitry Andric   /// Returns false if N is a bit extraction pattern of (X >> C) & Mask.
4040b57cec5SDimitry Andric   bool isDesirableToCommuteWithShift(const SDNode *N,
4050b57cec5SDimitry Andric                                      CombineLevel Level) const override;
4060b57cec5SDimitry Andric 
4070b57cec5SDimitry Andric   /// Returns true if it is beneficial to convert a load of a constant
4080b57cec5SDimitry Andric   /// to just the constant itself.
4090b57cec5SDimitry Andric   bool shouldConvertConstantLoadToIntImm(const APInt &Imm,
4100b57cec5SDimitry Andric                                          Type *Ty) const override;
4110b57cec5SDimitry Andric 
4120b57cec5SDimitry Andric   /// Return true if EXTRACT_SUBVECTOR is cheap for this result type
4130b57cec5SDimitry Andric   /// with this index.
4140b57cec5SDimitry Andric   bool isExtractSubvectorCheap(EVT ResVT, EVT SrcVT,
4150b57cec5SDimitry Andric                                unsigned Index) const override;
4160b57cec5SDimitry Andric 
4170b57cec5SDimitry Andric   Value *emitLoadLinked(IRBuilder<> &Builder, Value *Addr,
4180b57cec5SDimitry Andric                         AtomicOrdering Ord) const override;
4190b57cec5SDimitry Andric   Value *emitStoreConditional(IRBuilder<> &Builder, Value *Val,
4200b57cec5SDimitry Andric                               Value *Addr, AtomicOrdering Ord) const override;
4210b57cec5SDimitry Andric 
4220b57cec5SDimitry Andric   void emitAtomicCmpXchgNoStoreLLBalance(IRBuilder<> &Builder) const override;
4230b57cec5SDimitry Andric 
4240b57cec5SDimitry Andric   TargetLoweringBase::AtomicExpansionKind
4250b57cec5SDimitry Andric   shouldExpandAtomicLoadInIR(LoadInst *LI) const override;
4260b57cec5SDimitry Andric   bool shouldExpandAtomicStoreInIR(StoreInst *SI) const override;
4270b57cec5SDimitry Andric   TargetLoweringBase::AtomicExpansionKind
4280b57cec5SDimitry Andric   shouldExpandAtomicRMWInIR(AtomicRMWInst *AI) const override;
4290b57cec5SDimitry Andric 
4300b57cec5SDimitry Andric   TargetLoweringBase::AtomicExpansionKind
4310b57cec5SDimitry Andric   shouldExpandAtomicCmpXchgInIR(AtomicCmpXchgInst *AI) const override;
4320b57cec5SDimitry Andric 
4330b57cec5SDimitry Andric   bool useLoadStackGuardNode() const override;
4340b57cec5SDimitry Andric   TargetLoweringBase::LegalizeTypeAction
4350b57cec5SDimitry Andric   getPreferredVectorAction(MVT VT) const override;
4360b57cec5SDimitry Andric 
4370b57cec5SDimitry Andric   /// If the target has a standard location for the stack protector cookie,
4380b57cec5SDimitry Andric   /// returns the address of that location. Otherwise, returns nullptr.
4390b57cec5SDimitry Andric   Value *getIRStackGuard(IRBuilder<> &IRB) const override;
4400b57cec5SDimitry Andric 
4410b57cec5SDimitry Andric   void insertSSPDeclarations(Module &M) const override;
4420b57cec5SDimitry Andric   Value *getSDagStackGuard(const Module &M) const override;
4430b57cec5SDimitry Andric   Function *getSSPStackGuardCheck(const Module &M) const override;
4440b57cec5SDimitry Andric 
4450b57cec5SDimitry Andric   /// If the target has a standard location for the unsafe stack pointer,
4460b57cec5SDimitry Andric   /// returns the address of that location. Otherwise, returns nullptr.
4470b57cec5SDimitry Andric   Value *getSafeStackPointerLocation(IRBuilder<> &IRB) const override;
4480b57cec5SDimitry Andric 
4490b57cec5SDimitry Andric   /// If a physical register, this returns the register that receives the
4500b57cec5SDimitry Andric   /// exception address on entry to an EH pad.
4510b57cec5SDimitry Andric   unsigned
4520b57cec5SDimitry Andric   getExceptionPointerRegister(const Constant *PersonalityFn) const override {
4530b57cec5SDimitry Andric     // FIXME: This is a guess. Has this been defined yet?
4540b57cec5SDimitry Andric     return AArch64::X0;
4550b57cec5SDimitry Andric   }
4560b57cec5SDimitry Andric 
4570b57cec5SDimitry Andric   /// If a physical register, this returns the register that receives the
4580b57cec5SDimitry Andric   /// exception typeid on entry to a landing pad.
4590b57cec5SDimitry Andric   unsigned
4600b57cec5SDimitry Andric   getExceptionSelectorRegister(const Constant *PersonalityFn) const override {
4610b57cec5SDimitry Andric     // FIXME: This is a guess. Has this been defined yet?
4620b57cec5SDimitry Andric     return AArch64::X1;
4630b57cec5SDimitry Andric   }
4640b57cec5SDimitry Andric 
4650b57cec5SDimitry Andric   bool isIntDivCheap(EVT VT, AttributeList Attr) const override;
4660b57cec5SDimitry Andric 
4670b57cec5SDimitry Andric   bool canMergeStoresTo(unsigned AddressSpace, EVT MemVT,
4680b57cec5SDimitry Andric                         const SelectionDAG &DAG) const override {
4690b57cec5SDimitry Andric     // Do not merge to float value size (128 bytes) if no implicit
4700b57cec5SDimitry Andric     // float attribute is set.
4710b57cec5SDimitry Andric 
4720b57cec5SDimitry Andric     bool NoFloat = DAG.getMachineFunction().getFunction().hasFnAttribute(
4730b57cec5SDimitry Andric         Attribute::NoImplicitFloat);
4740b57cec5SDimitry Andric 
4750b57cec5SDimitry Andric     if (NoFloat)
4760b57cec5SDimitry Andric       return (MemVT.getSizeInBits() <= 64);
4770b57cec5SDimitry Andric     return true;
4780b57cec5SDimitry Andric   }
4790b57cec5SDimitry Andric 
4800b57cec5SDimitry Andric   bool isCheapToSpeculateCttz() const override {
4810b57cec5SDimitry Andric     return true;
4820b57cec5SDimitry Andric   }
4830b57cec5SDimitry Andric 
4840b57cec5SDimitry Andric   bool isCheapToSpeculateCtlz() const override {
4850b57cec5SDimitry Andric     return true;
4860b57cec5SDimitry Andric   }
4870b57cec5SDimitry Andric 
4880b57cec5SDimitry Andric   bool isMaskAndCmp0FoldingBeneficial(const Instruction &AndI) const override;
4890b57cec5SDimitry Andric 
4900b57cec5SDimitry Andric   bool hasAndNotCompare(SDValue V) const override {
4910b57cec5SDimitry Andric     // We can use bics for any scalar.
4920b57cec5SDimitry Andric     return V.getValueType().isScalarInteger();
4930b57cec5SDimitry Andric   }
4940b57cec5SDimitry Andric 
4950b57cec5SDimitry Andric   bool hasAndNot(SDValue Y) const override {
4960b57cec5SDimitry Andric     EVT VT = Y.getValueType();
4970b57cec5SDimitry Andric 
4980b57cec5SDimitry Andric     if (!VT.isVector())
4990b57cec5SDimitry Andric       return hasAndNotCompare(Y);
5000b57cec5SDimitry Andric 
5010b57cec5SDimitry Andric     return VT.getSizeInBits() >= 64; // vector 'bic'
5020b57cec5SDimitry Andric   }
5030b57cec5SDimitry Andric 
504*8bcb0991SDimitry Andric   bool shouldProduceAndByConstByHoistingConstFromShiftsLHSOfAnd(
505*8bcb0991SDimitry Andric       SDValue X, ConstantSDNode *XC, ConstantSDNode *CC, SDValue Y,
506*8bcb0991SDimitry Andric       unsigned OldShiftOpcode, unsigned NewShiftOpcode,
507*8bcb0991SDimitry Andric       SelectionDAG &DAG) const override;
508*8bcb0991SDimitry Andric 
5090b57cec5SDimitry Andric   bool shouldExpandShift(SelectionDAG &DAG, SDNode *N) const override;
5100b57cec5SDimitry Andric 
5110b57cec5SDimitry Andric   bool shouldTransformSignedTruncationCheck(EVT XVT,
5120b57cec5SDimitry Andric                                             unsigned KeptBits) const override {
5130b57cec5SDimitry Andric     // For vectors, we don't have a preference..
5140b57cec5SDimitry Andric     if (XVT.isVector())
5150b57cec5SDimitry Andric       return false;
5160b57cec5SDimitry Andric 
5170b57cec5SDimitry Andric     auto VTIsOk = [](EVT VT) -> bool {
5180b57cec5SDimitry Andric       return VT == MVT::i8 || VT == MVT::i16 || VT == MVT::i32 ||
5190b57cec5SDimitry Andric              VT == MVT::i64;
5200b57cec5SDimitry Andric     };
5210b57cec5SDimitry Andric 
5220b57cec5SDimitry Andric     // We are ok with KeptBitsVT being byte/word/dword, what SXT supports.
5230b57cec5SDimitry Andric     // XVT will be larger than KeptBitsVT.
5240b57cec5SDimitry Andric     MVT KeptBitsVT = MVT::getIntegerVT(KeptBits);
5250b57cec5SDimitry Andric     return VTIsOk(XVT) && VTIsOk(KeptBitsVT);
5260b57cec5SDimitry Andric   }
5270b57cec5SDimitry Andric 
5280b57cec5SDimitry Andric   bool preferIncOfAddToSubOfNot(EVT VT) const override;
5290b57cec5SDimitry Andric 
5300b57cec5SDimitry Andric   bool hasBitPreservingFPLogic(EVT VT) const override {
5310b57cec5SDimitry Andric     // FIXME: Is this always true? It should be true for vectors at least.
5320b57cec5SDimitry Andric     return VT == MVT::f32 || VT == MVT::f64;
5330b57cec5SDimitry Andric   }
5340b57cec5SDimitry Andric 
5350b57cec5SDimitry Andric   bool supportSplitCSR(MachineFunction *MF) const override {
5360b57cec5SDimitry Andric     return MF->getFunction().getCallingConv() == CallingConv::CXX_FAST_TLS &&
5370b57cec5SDimitry Andric            MF->getFunction().hasFnAttribute(Attribute::NoUnwind);
5380b57cec5SDimitry Andric   }
5390b57cec5SDimitry Andric   void initializeSplitCSR(MachineBasicBlock *Entry) const override;
5400b57cec5SDimitry Andric   void insertCopiesSplitCSR(
5410b57cec5SDimitry Andric       MachineBasicBlock *Entry,
5420b57cec5SDimitry Andric       const SmallVectorImpl<MachineBasicBlock *> &Exits) const override;
5430b57cec5SDimitry Andric 
5440b57cec5SDimitry Andric   bool supportSwiftError() const override {
5450b57cec5SDimitry Andric     return true;
5460b57cec5SDimitry Andric   }
5470b57cec5SDimitry Andric 
5480b57cec5SDimitry Andric   /// Enable aggressive FMA fusion on targets that want it.
5490b57cec5SDimitry Andric   bool enableAggressiveFMAFusion(EVT VT) const override;
5500b57cec5SDimitry Andric 
5510b57cec5SDimitry Andric   /// Returns the size of the platform's va_list object.
5520b57cec5SDimitry Andric   unsigned getVaListSizeInBits(const DataLayout &DL) const override;
5530b57cec5SDimitry Andric 
5540b57cec5SDimitry Andric   /// Returns true if \p VecTy is a legal interleaved access type. This
5550b57cec5SDimitry Andric   /// function checks the vector element type and the overall width of the
5560b57cec5SDimitry Andric   /// vector.
5570b57cec5SDimitry Andric   bool isLegalInterleavedAccessType(VectorType *VecTy,
5580b57cec5SDimitry Andric                                     const DataLayout &DL) const;
5590b57cec5SDimitry Andric 
5600b57cec5SDimitry Andric   /// Returns the number of interleaved accesses that will be generated when
5610b57cec5SDimitry Andric   /// lowering accesses of the given type.
5620b57cec5SDimitry Andric   unsigned getNumInterleavedAccesses(VectorType *VecTy,
5630b57cec5SDimitry Andric                                      const DataLayout &DL) const;
5640b57cec5SDimitry Andric 
5650b57cec5SDimitry Andric   MachineMemOperand::Flags getMMOFlags(const Instruction &I) const override;
5660b57cec5SDimitry Andric 
5670b57cec5SDimitry Andric   bool functionArgumentNeedsConsecutiveRegisters(Type *Ty,
5680b57cec5SDimitry Andric                                                  CallingConv::ID CallConv,
5690b57cec5SDimitry Andric                                                  bool isVarArg) const override;
5700b57cec5SDimitry Andric   /// Used for exception handling on Win64.
5710b57cec5SDimitry Andric   bool needsFixedCatchObjects() const override;
5720b57cec5SDimitry Andric private:
5730b57cec5SDimitry Andric   /// Keep a pointer to the AArch64Subtarget around so that we can
5740b57cec5SDimitry Andric   /// make the right decision when generating code for different targets.
5750b57cec5SDimitry Andric   const AArch64Subtarget *Subtarget;
5760b57cec5SDimitry Andric 
5770b57cec5SDimitry Andric   bool isExtFreeImpl(const Instruction *Ext) const override;
5780b57cec5SDimitry Andric 
5790b57cec5SDimitry Andric   void addTypeForNEON(MVT VT, MVT PromotedBitwiseVT);
5800b57cec5SDimitry Andric   void addDRTypeForNEON(MVT VT);
5810b57cec5SDimitry Andric   void addQRTypeForNEON(MVT VT);
5820b57cec5SDimitry Andric 
5830b57cec5SDimitry Andric   SDValue LowerFormalArguments(SDValue Chain, CallingConv::ID CallConv,
5840b57cec5SDimitry Andric                                bool isVarArg,
5850b57cec5SDimitry Andric                                const SmallVectorImpl<ISD::InputArg> &Ins,
5860b57cec5SDimitry Andric                                const SDLoc &DL, SelectionDAG &DAG,
5870b57cec5SDimitry Andric                                SmallVectorImpl<SDValue> &InVals) const override;
5880b57cec5SDimitry Andric 
5890b57cec5SDimitry Andric   SDValue LowerCall(CallLoweringInfo & /*CLI*/,
5900b57cec5SDimitry Andric                     SmallVectorImpl<SDValue> &InVals) const override;
5910b57cec5SDimitry Andric 
5920b57cec5SDimitry Andric   SDValue LowerCallResult(SDValue Chain, SDValue InFlag,
5930b57cec5SDimitry Andric                           CallingConv::ID CallConv, bool isVarArg,
5940b57cec5SDimitry Andric                           const SmallVectorImpl<ISD::InputArg> &Ins,
5950b57cec5SDimitry Andric                           const SDLoc &DL, SelectionDAG &DAG,
5960b57cec5SDimitry Andric                           SmallVectorImpl<SDValue> &InVals, bool isThisReturn,
5970b57cec5SDimitry Andric                           SDValue ThisVal) const;
5980b57cec5SDimitry Andric 
5990b57cec5SDimitry Andric   SDValue LowerSTORE(SDValue Op, SelectionDAG &DAG) const;
6000b57cec5SDimitry Andric 
6010b57cec5SDimitry Andric   SDValue LowerINTRINSIC_WO_CHAIN(SDValue Op, SelectionDAG &DAG) const;
6020b57cec5SDimitry Andric 
6030b57cec5SDimitry Andric   bool isEligibleForTailCallOptimization(
6040b57cec5SDimitry Andric       SDValue Callee, CallingConv::ID CalleeCC, bool isVarArg,
6050b57cec5SDimitry Andric       const SmallVectorImpl<ISD::OutputArg> &Outs,
6060b57cec5SDimitry Andric       const SmallVectorImpl<SDValue> &OutVals,
6070b57cec5SDimitry Andric       const SmallVectorImpl<ISD::InputArg> &Ins, SelectionDAG &DAG) const;
6080b57cec5SDimitry Andric 
6090b57cec5SDimitry Andric   /// Finds the incoming stack arguments which overlap the given fixed stack
6100b57cec5SDimitry Andric   /// object and incorporates their load into the current chain. This prevents
6110b57cec5SDimitry Andric   /// an upcoming store from clobbering the stack argument before it's used.
6120b57cec5SDimitry Andric   SDValue addTokenForArgument(SDValue Chain, SelectionDAG &DAG,
6130b57cec5SDimitry Andric                               MachineFrameInfo &MFI, int ClobberedFI) const;
6140b57cec5SDimitry Andric 
6150b57cec5SDimitry Andric   bool DoesCalleeRestoreStack(CallingConv::ID CallCC, bool TailCallOpt) const;
6160b57cec5SDimitry Andric 
6170b57cec5SDimitry Andric   void saveVarArgRegisters(CCState &CCInfo, SelectionDAG &DAG, const SDLoc &DL,
6180b57cec5SDimitry Andric                            SDValue &Chain) const;
6190b57cec5SDimitry Andric 
6200b57cec5SDimitry Andric   bool CanLowerReturn(CallingConv::ID CallConv, MachineFunction &MF,
6210b57cec5SDimitry Andric                       bool isVarArg,
6220b57cec5SDimitry Andric                       const SmallVectorImpl<ISD::OutputArg> &Outs,
6230b57cec5SDimitry Andric                       LLVMContext &Context) const override;
6240b57cec5SDimitry Andric 
6250b57cec5SDimitry Andric   SDValue LowerReturn(SDValue Chain, CallingConv::ID CallConv, bool isVarArg,
6260b57cec5SDimitry Andric                       const SmallVectorImpl<ISD::OutputArg> &Outs,
6270b57cec5SDimitry Andric                       const SmallVectorImpl<SDValue> &OutVals, const SDLoc &DL,
6280b57cec5SDimitry Andric                       SelectionDAG &DAG) const override;
6290b57cec5SDimitry Andric 
6300b57cec5SDimitry Andric   SDValue getTargetNode(GlobalAddressSDNode *N, EVT Ty, SelectionDAG &DAG,
6310b57cec5SDimitry Andric                         unsigned Flag) const;
6320b57cec5SDimitry Andric   SDValue getTargetNode(JumpTableSDNode *N, EVT Ty, SelectionDAG &DAG,
6330b57cec5SDimitry Andric                         unsigned Flag) const;
6340b57cec5SDimitry Andric   SDValue getTargetNode(ConstantPoolSDNode *N, EVT Ty, SelectionDAG &DAG,
6350b57cec5SDimitry Andric                         unsigned Flag) const;
6360b57cec5SDimitry Andric   SDValue getTargetNode(BlockAddressSDNode *N, EVT Ty, SelectionDAG &DAG,
6370b57cec5SDimitry Andric                         unsigned Flag) const;
6380b57cec5SDimitry Andric   template <class NodeTy>
6390b57cec5SDimitry Andric   SDValue getGOT(NodeTy *N, SelectionDAG &DAG, unsigned Flags = 0) const;
6400b57cec5SDimitry Andric   template <class NodeTy>
6410b57cec5SDimitry Andric   SDValue getAddrLarge(NodeTy *N, SelectionDAG &DAG, unsigned Flags = 0) const;
6420b57cec5SDimitry Andric   template <class NodeTy>
6430b57cec5SDimitry Andric   SDValue getAddr(NodeTy *N, SelectionDAG &DAG, unsigned Flags = 0) const;
6440b57cec5SDimitry Andric   template <class NodeTy>
6450b57cec5SDimitry Andric   SDValue getAddrTiny(NodeTy *N, SelectionDAG &DAG, unsigned Flags = 0) const;
6460b57cec5SDimitry Andric   SDValue LowerADDROFRETURNADDR(SDValue Op, SelectionDAG &DAG) const;
6470b57cec5SDimitry Andric   SDValue LowerGlobalAddress(SDValue Op, SelectionDAG &DAG) const;
6480b57cec5SDimitry Andric   SDValue LowerGlobalTLSAddress(SDValue Op, SelectionDAG &DAG) const;
6490b57cec5SDimitry Andric   SDValue LowerDarwinGlobalTLSAddress(SDValue Op, SelectionDAG &DAG) const;
6500b57cec5SDimitry Andric   SDValue LowerELFGlobalTLSAddress(SDValue Op, SelectionDAG &DAG) const;
6510b57cec5SDimitry Andric   SDValue LowerELFTLSDescCallSeq(SDValue SymAddr, const SDLoc &DL,
6520b57cec5SDimitry Andric                                  SelectionDAG &DAG) const;
6530b57cec5SDimitry Andric   SDValue LowerWindowsGlobalTLSAddress(SDValue Op, SelectionDAG &DAG) const;
6540b57cec5SDimitry Andric   SDValue LowerSETCC(SDValue Op, SelectionDAG &DAG) const;
6550b57cec5SDimitry Andric   SDValue LowerBR_CC(SDValue Op, SelectionDAG &DAG) const;
6560b57cec5SDimitry Andric   SDValue LowerSELECT(SDValue Op, SelectionDAG &DAG) const;
6570b57cec5SDimitry Andric   SDValue LowerSELECT_CC(SDValue Op, SelectionDAG &DAG) const;
6580b57cec5SDimitry Andric   SDValue LowerSELECT_CC(ISD::CondCode CC, SDValue LHS, SDValue RHS,
6590b57cec5SDimitry Andric                          SDValue TVal, SDValue FVal, const SDLoc &dl,
6600b57cec5SDimitry Andric                          SelectionDAG &DAG) const;
6610b57cec5SDimitry Andric   SDValue LowerJumpTable(SDValue Op, SelectionDAG &DAG) const;
6620b57cec5SDimitry Andric   SDValue LowerBR_JT(SDValue Op, SelectionDAG &DAG) const;
6630b57cec5SDimitry Andric   SDValue LowerConstantPool(SDValue Op, SelectionDAG &DAG) const;
6640b57cec5SDimitry Andric   SDValue LowerBlockAddress(SDValue Op, SelectionDAG &DAG) const;
6650b57cec5SDimitry Andric   SDValue LowerAAPCS_VASTART(SDValue Op, SelectionDAG &DAG) const;
6660b57cec5SDimitry Andric   SDValue LowerDarwin_VASTART(SDValue Op, SelectionDAG &DAG) const;
6670b57cec5SDimitry Andric   SDValue LowerWin64_VASTART(SDValue Op, SelectionDAG &DAG) const;
6680b57cec5SDimitry Andric   SDValue LowerVASTART(SDValue Op, SelectionDAG &DAG) const;
6690b57cec5SDimitry Andric   SDValue LowerVACOPY(SDValue Op, SelectionDAG &DAG) const;
6700b57cec5SDimitry Andric   SDValue LowerVAARG(SDValue Op, SelectionDAG &DAG) const;
6710b57cec5SDimitry Andric   SDValue LowerFRAMEADDR(SDValue Op, SelectionDAG &DAG) const;
6720b57cec5SDimitry Andric   SDValue LowerSPONENTRY(SDValue Op, SelectionDAG &DAG) const;
6730b57cec5SDimitry Andric   SDValue LowerRETURNADDR(SDValue Op, SelectionDAG &DAG) const;
6740b57cec5SDimitry Andric   SDValue LowerFLT_ROUNDS_(SDValue Op, SelectionDAG &DAG) const;
6750b57cec5SDimitry Andric   SDValue LowerINSERT_VECTOR_ELT(SDValue Op, SelectionDAG &DAG) const;
6760b57cec5SDimitry Andric   SDValue LowerEXTRACT_VECTOR_ELT(SDValue Op, SelectionDAG &DAG) const;
6770b57cec5SDimitry Andric   SDValue LowerSCALAR_TO_VECTOR(SDValue Op, SelectionDAG &DAG) const;
6780b57cec5SDimitry Andric   SDValue LowerBUILD_VECTOR(SDValue Op, SelectionDAG &DAG) const;
6790b57cec5SDimitry Andric   SDValue LowerVECTOR_SHUFFLE(SDValue Op, SelectionDAG &DAG) const;
680*8bcb0991SDimitry Andric   SDValue LowerSPLAT_VECTOR(SDValue Op, SelectionDAG &DAG) const;
6810b57cec5SDimitry Andric   SDValue LowerEXTRACT_SUBVECTOR(SDValue Op, SelectionDAG &DAG) const;
6820b57cec5SDimitry Andric   SDValue LowerVectorSRA_SRL_SHL(SDValue Op, SelectionDAG &DAG) const;
6830b57cec5SDimitry Andric   SDValue LowerShiftLeftParts(SDValue Op, SelectionDAG &DAG) const;
6840b57cec5SDimitry Andric   SDValue LowerShiftRightParts(SDValue Op, SelectionDAG &DAG) const;
6850b57cec5SDimitry Andric   SDValue LowerVSETCC(SDValue Op, SelectionDAG &DAG) const;
6860b57cec5SDimitry Andric   SDValue LowerCTPOP(SDValue Op, SelectionDAG &DAG) const;
6870b57cec5SDimitry Andric   SDValue LowerF128Call(SDValue Op, SelectionDAG &DAG,
6880b57cec5SDimitry Andric                         RTLIB::Libcall Call) const;
6890b57cec5SDimitry Andric   SDValue LowerFCOPYSIGN(SDValue Op, SelectionDAG &DAG) const;
6900b57cec5SDimitry Andric   SDValue LowerFP_EXTEND(SDValue Op, SelectionDAG &DAG) const;
6910b57cec5SDimitry Andric   SDValue LowerFP_ROUND(SDValue Op, SelectionDAG &DAG) const;
6920b57cec5SDimitry Andric   SDValue LowerVectorFP_TO_INT(SDValue Op, SelectionDAG &DAG) const;
6930b57cec5SDimitry Andric   SDValue LowerFP_TO_INT(SDValue Op, SelectionDAG &DAG) const;
6940b57cec5SDimitry Andric   SDValue LowerINT_TO_FP(SDValue Op, SelectionDAG &DAG) const;
6950b57cec5SDimitry Andric   SDValue LowerVectorOR(SDValue Op, SelectionDAG &DAG) const;
6960b57cec5SDimitry Andric   SDValue LowerCONCAT_VECTORS(SDValue Op, SelectionDAG &DAG) const;
6970b57cec5SDimitry Andric   SDValue LowerFSINCOS(SDValue Op, SelectionDAG &DAG) const;
6980b57cec5SDimitry Andric   SDValue LowerVECREDUCE(SDValue Op, SelectionDAG &DAG) const;
6990b57cec5SDimitry Andric   SDValue LowerATOMIC_LOAD_SUB(SDValue Op, SelectionDAG &DAG) const;
7000b57cec5SDimitry Andric   SDValue LowerATOMIC_LOAD_AND(SDValue Op, SelectionDAG &DAG) const;
7010b57cec5SDimitry Andric   SDValue LowerDYNAMIC_STACKALLOC(SDValue Op, SelectionDAG &DAG) const;
7020b57cec5SDimitry Andric   SDValue LowerWindowsDYNAMIC_STACKALLOC(SDValue Op, SDValue Chain,
7030b57cec5SDimitry Andric                                          SDValue &Size,
7040b57cec5SDimitry Andric                                          SelectionDAG &DAG) const;
7050b57cec5SDimitry Andric 
7060b57cec5SDimitry Andric   SDValue BuildSDIVPow2(SDNode *N, const APInt &Divisor, SelectionDAG &DAG,
7070b57cec5SDimitry Andric                         SmallVectorImpl<SDNode *> &Created) const override;
7080b57cec5SDimitry Andric   SDValue getSqrtEstimate(SDValue Operand, SelectionDAG &DAG, int Enabled,
7090b57cec5SDimitry Andric                           int &ExtraSteps, bool &UseOneConst,
7100b57cec5SDimitry Andric                           bool Reciprocal) const override;
7110b57cec5SDimitry Andric   SDValue getRecipEstimate(SDValue Operand, SelectionDAG &DAG, int Enabled,
7120b57cec5SDimitry Andric                            int &ExtraSteps) const override;
7130b57cec5SDimitry Andric   unsigned combineRepeatedFPDivisors() const override;
7140b57cec5SDimitry Andric 
7150b57cec5SDimitry Andric   ConstraintType getConstraintType(StringRef Constraint) const override;
716*8bcb0991SDimitry Andric   Register getRegisterByName(const char* RegName, EVT VT,
717*8bcb0991SDimitry Andric                              const MachineFunction &MF) const override;
7180b57cec5SDimitry Andric 
7190b57cec5SDimitry Andric   /// Examine constraint string and operand type and determine a weight value.
7200b57cec5SDimitry Andric   /// The operand object must already have been set up with the operand type.
7210b57cec5SDimitry Andric   ConstraintWeight
7220b57cec5SDimitry Andric   getSingleConstraintMatchWeight(AsmOperandInfo &info,
7230b57cec5SDimitry Andric                                  const char *constraint) const override;
7240b57cec5SDimitry Andric 
7250b57cec5SDimitry Andric   std::pair<unsigned, const TargetRegisterClass *>
7260b57cec5SDimitry Andric   getRegForInlineAsmConstraint(const TargetRegisterInfo *TRI,
7270b57cec5SDimitry Andric                                StringRef Constraint, MVT VT) const override;
7280b57cec5SDimitry Andric 
7290b57cec5SDimitry Andric   const char *LowerXConstraint(EVT ConstraintVT) const override;
7300b57cec5SDimitry Andric 
7310b57cec5SDimitry Andric   void LowerAsmOperandForConstraint(SDValue Op, std::string &Constraint,
7320b57cec5SDimitry Andric                                     std::vector<SDValue> &Ops,
7330b57cec5SDimitry Andric                                     SelectionDAG &DAG) const override;
7340b57cec5SDimitry Andric 
7350b57cec5SDimitry Andric   unsigned getInlineAsmMemConstraint(StringRef ConstraintCode) const override {
7360b57cec5SDimitry Andric     if (ConstraintCode == "Q")
7370b57cec5SDimitry Andric       return InlineAsm::Constraint_Q;
7380b57cec5SDimitry Andric     // FIXME: clang has code for 'Ump', 'Utf', 'Usa', and 'Ush' but these are
7390b57cec5SDimitry Andric     //        followed by llvm_unreachable so we'll leave them unimplemented in
7400b57cec5SDimitry Andric     //        the backend for now.
7410b57cec5SDimitry Andric     return TargetLowering::getInlineAsmMemConstraint(ConstraintCode);
7420b57cec5SDimitry Andric   }
7430b57cec5SDimitry Andric 
7440b57cec5SDimitry Andric   bool isUsedByReturnOnly(SDNode *N, SDValue &Chain) const override;
7450b57cec5SDimitry Andric   bool mayBeEmittedAsTailCall(const CallInst *CI) const override;
7460b57cec5SDimitry Andric   bool getIndexedAddressParts(SDNode *Op, SDValue &Base, SDValue &Offset,
7470b57cec5SDimitry Andric                               ISD::MemIndexedMode &AM, bool &IsInc,
7480b57cec5SDimitry Andric                               SelectionDAG &DAG) const;
7490b57cec5SDimitry Andric   bool getPreIndexedAddressParts(SDNode *N, SDValue &Base, SDValue &Offset,
7500b57cec5SDimitry Andric                                  ISD::MemIndexedMode &AM,
7510b57cec5SDimitry Andric                                  SelectionDAG &DAG) const override;
7520b57cec5SDimitry Andric   bool getPostIndexedAddressParts(SDNode *N, SDNode *Op, SDValue &Base,
7530b57cec5SDimitry Andric                                   SDValue &Offset, ISD::MemIndexedMode &AM,
7540b57cec5SDimitry Andric                                   SelectionDAG &DAG) const override;
7550b57cec5SDimitry Andric 
7560b57cec5SDimitry Andric   void ReplaceNodeResults(SDNode *N, SmallVectorImpl<SDValue> &Results,
7570b57cec5SDimitry Andric                           SelectionDAG &DAG) const override;
7580b57cec5SDimitry Andric 
7590b57cec5SDimitry Andric   bool shouldNormalizeToSelectSequence(LLVMContext &, EVT) const override;
7600b57cec5SDimitry Andric 
7610b57cec5SDimitry Andric   void finalizeLowering(MachineFunction &MF) const override;
7620b57cec5SDimitry Andric };
7630b57cec5SDimitry Andric 
7640b57cec5SDimitry Andric namespace AArch64 {
7650b57cec5SDimitry Andric FastISel *createFastISel(FunctionLoweringInfo &funcInfo,
7660b57cec5SDimitry Andric                          const TargetLibraryInfo *libInfo);
7670b57cec5SDimitry Andric } // end namespace AArch64
7680b57cec5SDimitry Andric 
7690b57cec5SDimitry Andric } // end namespace llvm
7700b57cec5SDimitry Andric 
7710b57cec5SDimitry Andric #endif
772