1 //===-- ARMExpandPseudoInsts.cpp - Expand pseudo instructions -------------===// 2 // 3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. 4 // See https://llvm.org/LICENSE.txt for license information. 5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception 6 // 7 //===----------------------------------------------------------------------===// 8 // 9 // This file contains a pass that expands pseudo instructions into target 10 // instructions to allow proper scheduling, if-conversion, and other late 11 // optimizations. This pass should be run after register allocation but before 12 // the post-regalloc scheduling pass. 13 // 14 //===----------------------------------------------------------------------===// 15 16 #include "ARM.h" 17 #include "ARMBaseInstrInfo.h" 18 #include "ARMBaseRegisterInfo.h" 19 #include "ARMConstantPoolValue.h" 20 #include "ARMMachineFunctionInfo.h" 21 #include "ARMSubtarget.h" 22 #include "MCTargetDesc/ARMAddressingModes.h" 23 #include "llvm/CodeGen/LivePhysRegs.h" 24 #include "llvm/CodeGen/MachineFrameInfo.h" 25 #include "llvm/CodeGen/MachineFunctionPass.h" 26 #include "llvm/MC/MCAsmInfo.h" 27 #include "llvm/Support/Debug.h" 28 29 using namespace llvm; 30 31 #define DEBUG_TYPE "arm-pseudo" 32 33 static cl::opt<bool> 34 VerifyARMPseudo("verify-arm-pseudo-expand", cl::Hidden, 35 cl::desc("Verify machine code after expanding ARM pseudos")); 36 37 #define ARM_EXPAND_PSEUDO_NAME "ARM pseudo instruction expansion pass" 38 39 namespace { 40 class ARMExpandPseudo : public MachineFunctionPass { 41 public: 42 static char ID; 43 ARMExpandPseudo() : MachineFunctionPass(ID) {} 44 45 const ARMBaseInstrInfo *TII; 46 const TargetRegisterInfo *TRI; 47 const ARMSubtarget *STI; 48 ARMFunctionInfo *AFI; 49 50 bool runOnMachineFunction(MachineFunction &Fn) override; 51 52 MachineFunctionProperties getRequiredProperties() const override { 53 return MachineFunctionProperties().set( 54 MachineFunctionProperties::Property::NoVRegs); 55 } 56 57 StringRef getPassName() const override { 58 return ARM_EXPAND_PSEUDO_NAME; 59 } 60 61 private: 62 void TransferImpOps(MachineInstr &OldMI, 63 MachineInstrBuilder &UseMI, MachineInstrBuilder &DefMI); 64 bool ExpandMI(MachineBasicBlock &MBB, 65 MachineBasicBlock::iterator MBBI, 66 MachineBasicBlock::iterator &NextMBBI); 67 bool ExpandMBB(MachineBasicBlock &MBB); 68 void ExpandVLD(MachineBasicBlock::iterator &MBBI); 69 void ExpandVST(MachineBasicBlock::iterator &MBBI); 70 void ExpandLaneOp(MachineBasicBlock::iterator &MBBI); 71 void ExpandVTBL(MachineBasicBlock::iterator &MBBI, 72 unsigned Opc, bool IsExt); 73 void ExpandMQQPRLoadStore(MachineBasicBlock::iterator &MBBI); 74 void ExpandMOV32BitImm(MachineBasicBlock &MBB, 75 MachineBasicBlock::iterator &MBBI); 76 void CMSEClearGPRegs(MachineBasicBlock &MBB, 77 MachineBasicBlock::iterator MBBI, const DebugLoc &DL, 78 const SmallVectorImpl<unsigned> &ClearRegs, 79 unsigned ClobberReg); 80 MachineBasicBlock &CMSEClearFPRegs(MachineBasicBlock &MBB, 81 MachineBasicBlock::iterator MBBI); 82 MachineBasicBlock &CMSEClearFPRegsV8(MachineBasicBlock &MBB, 83 MachineBasicBlock::iterator MBBI, 84 const BitVector &ClearRegs); 85 MachineBasicBlock &CMSEClearFPRegsV81(MachineBasicBlock &MBB, 86 MachineBasicBlock::iterator MBBI, 87 const BitVector &ClearRegs); 88 void CMSESaveClearFPRegs(MachineBasicBlock &MBB, 89 MachineBasicBlock::iterator MBBI, DebugLoc &DL, 90 const LivePhysRegs &LiveRegs, 91 SmallVectorImpl<unsigned> &AvailableRegs); 92 void CMSESaveClearFPRegsV8(MachineBasicBlock &MBB, 93 MachineBasicBlock::iterator MBBI, DebugLoc &DL, 94 const LivePhysRegs &LiveRegs, 95 SmallVectorImpl<unsigned> &ScratchRegs); 96 void CMSESaveClearFPRegsV81(MachineBasicBlock &MBB, 97 MachineBasicBlock::iterator MBBI, DebugLoc &DL, 98 const LivePhysRegs &LiveRegs); 99 void CMSERestoreFPRegs(MachineBasicBlock &MBB, 100 MachineBasicBlock::iterator MBBI, DebugLoc &DL, 101 SmallVectorImpl<unsigned> &AvailableRegs); 102 void CMSERestoreFPRegsV8(MachineBasicBlock &MBB, 103 MachineBasicBlock::iterator MBBI, DebugLoc &DL, 104 SmallVectorImpl<unsigned> &AvailableRegs); 105 void CMSERestoreFPRegsV81(MachineBasicBlock &MBB, 106 MachineBasicBlock::iterator MBBI, DebugLoc &DL, 107 SmallVectorImpl<unsigned> &AvailableRegs); 108 bool ExpandCMP_SWAP(MachineBasicBlock &MBB, 109 MachineBasicBlock::iterator MBBI, unsigned LdrexOp, 110 unsigned StrexOp, unsigned UxtOp, 111 MachineBasicBlock::iterator &NextMBBI); 112 113 bool ExpandCMP_SWAP_64(MachineBasicBlock &MBB, 114 MachineBasicBlock::iterator MBBI, 115 MachineBasicBlock::iterator &NextMBBI); 116 }; 117 char ARMExpandPseudo::ID = 0; 118 } 119 120 INITIALIZE_PASS(ARMExpandPseudo, DEBUG_TYPE, ARM_EXPAND_PSEUDO_NAME, false, 121 false) 122 123 /// TransferImpOps - Transfer implicit operands on the pseudo instruction to 124 /// the instructions created from the expansion. 125 void ARMExpandPseudo::TransferImpOps(MachineInstr &OldMI, 126 MachineInstrBuilder &UseMI, 127 MachineInstrBuilder &DefMI) { 128 const MCInstrDesc &Desc = OldMI.getDesc(); 129 for (const MachineOperand &MO : 130 llvm::drop_begin(OldMI.operands(), Desc.getNumOperands())) { 131 assert(MO.isReg() && MO.getReg()); 132 if (MO.isUse()) 133 UseMI.add(MO); 134 else 135 DefMI.add(MO); 136 } 137 } 138 139 namespace { 140 // Constants for register spacing in NEON load/store instructions. 141 // For quad-register load-lane and store-lane pseudo instructors, the 142 // spacing is initially assumed to be EvenDblSpc, and that is changed to 143 // OddDblSpc depending on the lane number operand. 144 enum NEONRegSpacing { 145 SingleSpc, 146 SingleLowSpc , // Single spacing, low registers, three and four vectors. 147 SingleHighQSpc, // Single spacing, high registers, four vectors. 148 SingleHighTSpc, // Single spacing, high registers, three vectors. 149 EvenDblSpc, 150 OddDblSpc 151 }; 152 153 // Entries for NEON load/store information table. The table is sorted by 154 // PseudoOpc for fast binary-search lookups. 155 struct NEONLdStTableEntry { 156 uint16_t PseudoOpc; 157 uint16_t RealOpc; 158 bool IsLoad; 159 bool isUpdating; 160 bool hasWritebackOperand; 161 uint8_t RegSpacing; // One of type NEONRegSpacing 162 uint8_t NumRegs; // D registers loaded or stored 163 uint8_t RegElts; // elements per D register; used for lane ops 164 // FIXME: Temporary flag to denote whether the real instruction takes 165 // a single register (like the encoding) or all of the registers in 166 // the list (like the asm syntax and the isel DAG). When all definitions 167 // are converted to take only the single encoded register, this will 168 // go away. 169 bool copyAllListRegs; 170 171 // Comparison methods for binary search of the table. 172 bool operator<(const NEONLdStTableEntry &TE) const { 173 return PseudoOpc < TE.PseudoOpc; 174 } 175 friend bool operator<(const NEONLdStTableEntry &TE, unsigned PseudoOpc) { 176 return TE.PseudoOpc < PseudoOpc; 177 } 178 friend bool LLVM_ATTRIBUTE_UNUSED operator<(unsigned PseudoOpc, 179 const NEONLdStTableEntry &TE) { 180 return PseudoOpc < TE.PseudoOpc; 181 } 182 }; 183 } 184 185 static const NEONLdStTableEntry NEONLdStTable[] = { 186 { ARM::VLD1LNq16Pseudo, ARM::VLD1LNd16, true, false, false, EvenDblSpc, 1, 4 ,true}, 187 { ARM::VLD1LNq16Pseudo_UPD, ARM::VLD1LNd16_UPD, true, true, true, EvenDblSpc, 1, 4 ,true}, 188 { ARM::VLD1LNq32Pseudo, ARM::VLD1LNd32, true, false, false, EvenDblSpc, 1, 2 ,true}, 189 { ARM::VLD1LNq32Pseudo_UPD, ARM::VLD1LNd32_UPD, true, true, true, EvenDblSpc, 1, 2 ,true}, 190 { ARM::VLD1LNq8Pseudo, ARM::VLD1LNd8, true, false, false, EvenDblSpc, 1, 8 ,true}, 191 { ARM::VLD1LNq8Pseudo_UPD, ARM::VLD1LNd8_UPD, true, true, true, EvenDblSpc, 1, 8 ,true}, 192 193 { ARM::VLD1d16QPseudo, ARM::VLD1d16Q, true, false, false, SingleSpc, 4, 4 ,false}, 194 { ARM::VLD1d16QPseudoWB_fixed, ARM::VLD1d16Qwb_fixed, true, true, false, SingleSpc, 4, 4 ,false}, 195 { ARM::VLD1d16QPseudoWB_register, ARM::VLD1d16Qwb_register, true, true, true, SingleSpc, 4, 4 ,false}, 196 { ARM::VLD1d16TPseudo, ARM::VLD1d16T, true, false, false, SingleSpc, 3, 4 ,false}, 197 { ARM::VLD1d16TPseudoWB_fixed, ARM::VLD1d16Twb_fixed, true, true, false, SingleSpc, 3, 4 ,false}, 198 { ARM::VLD1d16TPseudoWB_register, ARM::VLD1d16Twb_register, true, true, true, SingleSpc, 3, 4 ,false}, 199 200 { ARM::VLD1d32QPseudo, ARM::VLD1d32Q, true, false, false, SingleSpc, 4, 2 ,false}, 201 { ARM::VLD1d32QPseudoWB_fixed, ARM::VLD1d32Qwb_fixed, true, true, false, SingleSpc, 4, 2 ,false}, 202 { ARM::VLD1d32QPseudoWB_register, ARM::VLD1d32Qwb_register, true, true, true, SingleSpc, 4, 2 ,false}, 203 { ARM::VLD1d32TPseudo, ARM::VLD1d32T, true, false, false, SingleSpc, 3, 2 ,false}, 204 { ARM::VLD1d32TPseudoWB_fixed, ARM::VLD1d32Twb_fixed, true, true, false, SingleSpc, 3, 2 ,false}, 205 { ARM::VLD1d32TPseudoWB_register, ARM::VLD1d32Twb_register, true, true, true, SingleSpc, 3, 2 ,false}, 206 207 { ARM::VLD1d64QPseudo, ARM::VLD1d64Q, true, false, false, SingleSpc, 4, 1 ,false}, 208 { ARM::VLD1d64QPseudoWB_fixed, ARM::VLD1d64Qwb_fixed, true, true, false, SingleSpc, 4, 1 ,false}, 209 { ARM::VLD1d64QPseudoWB_register, ARM::VLD1d64Qwb_register, true, true, true, SingleSpc, 4, 1 ,false}, 210 { ARM::VLD1d64TPseudo, ARM::VLD1d64T, true, false, false, SingleSpc, 3, 1 ,false}, 211 { ARM::VLD1d64TPseudoWB_fixed, ARM::VLD1d64Twb_fixed, true, true, false, SingleSpc, 3, 1 ,false}, 212 { ARM::VLD1d64TPseudoWB_register, ARM::VLD1d64Twb_register, true, true, true, SingleSpc, 3, 1 ,false}, 213 214 { ARM::VLD1d8QPseudo, ARM::VLD1d8Q, true, false, false, SingleSpc, 4, 8 ,false}, 215 { ARM::VLD1d8QPseudoWB_fixed, ARM::VLD1d8Qwb_fixed, true, true, false, SingleSpc, 4, 8 ,false}, 216 { ARM::VLD1d8QPseudoWB_register, ARM::VLD1d8Qwb_register, true, true, true, SingleSpc, 4, 8 ,false}, 217 { ARM::VLD1d8TPseudo, ARM::VLD1d8T, true, false, false, SingleSpc, 3, 8 ,false}, 218 { ARM::VLD1d8TPseudoWB_fixed, ARM::VLD1d8Twb_fixed, true, true, false, SingleSpc, 3, 8 ,false}, 219 { ARM::VLD1d8TPseudoWB_register, ARM::VLD1d8Twb_register, true, true, true, SingleSpc, 3, 8 ,false}, 220 221 { ARM::VLD1q16HighQPseudo, ARM::VLD1d16Q, true, false, false, SingleHighQSpc, 4, 4 ,false}, 222 { ARM::VLD1q16HighQPseudo_UPD, ARM::VLD1d16Qwb_fixed, true, true, true, SingleHighQSpc, 4, 4 ,false}, 223 { ARM::VLD1q16HighTPseudo, ARM::VLD1d16T, true, false, false, SingleHighTSpc, 3, 4 ,false}, 224 { ARM::VLD1q16HighTPseudo_UPD, ARM::VLD1d16Twb_fixed, true, true, true, SingleHighTSpc, 3, 4 ,false}, 225 { ARM::VLD1q16LowQPseudo_UPD, ARM::VLD1d16Qwb_fixed, true, true, true, SingleLowSpc, 4, 4 ,false}, 226 { ARM::VLD1q16LowTPseudo_UPD, ARM::VLD1d16Twb_fixed, true, true, true, SingleLowSpc, 3, 4 ,false}, 227 228 { ARM::VLD1q32HighQPseudo, ARM::VLD1d32Q, true, false, false, SingleHighQSpc, 4, 2 ,false}, 229 { ARM::VLD1q32HighQPseudo_UPD, ARM::VLD1d32Qwb_fixed, true, true, true, SingleHighQSpc, 4, 2 ,false}, 230 { ARM::VLD1q32HighTPseudo, ARM::VLD1d32T, true, false, false, SingleHighTSpc, 3, 2 ,false}, 231 { ARM::VLD1q32HighTPseudo_UPD, ARM::VLD1d32Twb_fixed, true, true, true, SingleHighTSpc, 3, 2 ,false}, 232 { ARM::VLD1q32LowQPseudo_UPD, ARM::VLD1d32Qwb_fixed, true, true, true, SingleLowSpc, 4, 2 ,false}, 233 { ARM::VLD1q32LowTPseudo_UPD, ARM::VLD1d32Twb_fixed, true, true, true, SingleLowSpc, 3, 2 ,false}, 234 235 { ARM::VLD1q64HighQPseudo, ARM::VLD1d64Q, true, false, false, SingleHighQSpc, 4, 1 ,false}, 236 { ARM::VLD1q64HighQPseudo_UPD, ARM::VLD1d64Qwb_fixed, true, true, true, SingleHighQSpc, 4, 1 ,false}, 237 { ARM::VLD1q64HighTPseudo, ARM::VLD1d64T, true, false, false, SingleHighTSpc, 3, 1 ,false}, 238 { ARM::VLD1q64HighTPseudo_UPD, ARM::VLD1d64Twb_fixed, true, true, true, SingleHighTSpc, 3, 1 ,false}, 239 { ARM::VLD1q64LowQPseudo_UPD, ARM::VLD1d64Qwb_fixed, true, true, true, SingleLowSpc, 4, 1 ,false}, 240 { ARM::VLD1q64LowTPseudo_UPD, ARM::VLD1d64Twb_fixed, true, true, true, SingleLowSpc, 3, 1 ,false}, 241 242 { ARM::VLD1q8HighQPseudo, ARM::VLD1d8Q, true, false, false, SingleHighQSpc, 4, 8 ,false}, 243 { ARM::VLD1q8HighQPseudo_UPD, ARM::VLD1d8Qwb_fixed, true, true, true, SingleHighQSpc, 4, 8 ,false}, 244 { ARM::VLD1q8HighTPseudo, ARM::VLD1d8T, true, false, false, SingleHighTSpc, 3, 8 ,false}, 245 { ARM::VLD1q8HighTPseudo_UPD, ARM::VLD1d8Twb_fixed, true, true, true, SingleHighTSpc, 3, 8 ,false}, 246 { ARM::VLD1q8LowQPseudo_UPD, ARM::VLD1d8Qwb_fixed, true, true, true, SingleLowSpc, 4, 8 ,false}, 247 { ARM::VLD1q8LowTPseudo_UPD, ARM::VLD1d8Twb_fixed, true, true, true, SingleLowSpc, 3, 8 ,false}, 248 249 { ARM::VLD2DUPq16EvenPseudo, ARM::VLD2DUPd16x2, true, false, false, EvenDblSpc, 2, 4 ,false}, 250 { ARM::VLD2DUPq16OddPseudo, ARM::VLD2DUPd16x2, true, false, false, OddDblSpc, 2, 4 ,false}, 251 { ARM::VLD2DUPq16OddPseudoWB_fixed, ARM::VLD2DUPd16x2wb_fixed, true, true, false, OddDblSpc, 2, 4 ,false}, 252 { ARM::VLD2DUPq16OddPseudoWB_register, ARM::VLD2DUPd16x2wb_register, true, true, true, OddDblSpc, 2, 4 ,false}, 253 { ARM::VLD2DUPq32EvenPseudo, ARM::VLD2DUPd32x2, true, false, false, EvenDblSpc, 2, 2 ,false}, 254 { ARM::VLD2DUPq32OddPseudo, ARM::VLD2DUPd32x2, true, false, false, OddDblSpc, 2, 2 ,false}, 255 { ARM::VLD2DUPq32OddPseudoWB_fixed, ARM::VLD2DUPd32x2wb_fixed, true, true, false, OddDblSpc, 2, 2 ,false}, 256 { ARM::VLD2DUPq32OddPseudoWB_register, ARM::VLD2DUPd32x2wb_register, true, true, true, OddDblSpc, 2, 2 ,false}, 257 { ARM::VLD2DUPq8EvenPseudo, ARM::VLD2DUPd8x2, true, false, false, EvenDblSpc, 2, 8 ,false}, 258 { ARM::VLD2DUPq8OddPseudo, ARM::VLD2DUPd8x2, true, false, false, OddDblSpc, 2, 8 ,false}, 259 { ARM::VLD2DUPq8OddPseudoWB_fixed, ARM::VLD2DUPd8x2wb_fixed, true, true, false, OddDblSpc, 2, 8 ,false}, 260 { ARM::VLD2DUPq8OddPseudoWB_register, ARM::VLD2DUPd8x2wb_register, true, true, true, OddDblSpc, 2, 8 ,false}, 261 262 { ARM::VLD2LNd16Pseudo, ARM::VLD2LNd16, true, false, false, SingleSpc, 2, 4 ,true}, 263 { ARM::VLD2LNd16Pseudo_UPD, ARM::VLD2LNd16_UPD, true, true, true, SingleSpc, 2, 4 ,true}, 264 { ARM::VLD2LNd32Pseudo, ARM::VLD2LNd32, true, false, false, SingleSpc, 2, 2 ,true}, 265 { ARM::VLD2LNd32Pseudo_UPD, ARM::VLD2LNd32_UPD, true, true, true, SingleSpc, 2, 2 ,true}, 266 { ARM::VLD2LNd8Pseudo, ARM::VLD2LNd8, true, false, false, SingleSpc, 2, 8 ,true}, 267 { ARM::VLD2LNd8Pseudo_UPD, ARM::VLD2LNd8_UPD, true, true, true, SingleSpc, 2, 8 ,true}, 268 { ARM::VLD2LNq16Pseudo, ARM::VLD2LNq16, true, false, false, EvenDblSpc, 2, 4 ,true}, 269 { ARM::VLD2LNq16Pseudo_UPD, ARM::VLD2LNq16_UPD, true, true, true, EvenDblSpc, 2, 4 ,true}, 270 { ARM::VLD2LNq32Pseudo, ARM::VLD2LNq32, true, false, false, EvenDblSpc, 2, 2 ,true}, 271 { ARM::VLD2LNq32Pseudo_UPD, ARM::VLD2LNq32_UPD, true, true, true, EvenDblSpc, 2, 2 ,true}, 272 273 { ARM::VLD2q16Pseudo, ARM::VLD2q16, true, false, false, SingleSpc, 4, 4 ,false}, 274 { ARM::VLD2q16PseudoWB_fixed, ARM::VLD2q16wb_fixed, true, true, false, SingleSpc, 4, 4 ,false}, 275 { ARM::VLD2q16PseudoWB_register, ARM::VLD2q16wb_register, true, true, true, SingleSpc, 4, 4 ,false}, 276 { ARM::VLD2q32Pseudo, ARM::VLD2q32, true, false, false, SingleSpc, 4, 2 ,false}, 277 { ARM::VLD2q32PseudoWB_fixed, ARM::VLD2q32wb_fixed, true, true, false, SingleSpc, 4, 2 ,false}, 278 { ARM::VLD2q32PseudoWB_register, ARM::VLD2q32wb_register, true, true, true, SingleSpc, 4, 2 ,false}, 279 { ARM::VLD2q8Pseudo, ARM::VLD2q8, true, false, false, SingleSpc, 4, 8 ,false}, 280 { ARM::VLD2q8PseudoWB_fixed, ARM::VLD2q8wb_fixed, true, true, false, SingleSpc, 4, 8 ,false}, 281 { ARM::VLD2q8PseudoWB_register, ARM::VLD2q8wb_register, true, true, true, SingleSpc, 4, 8 ,false}, 282 283 { ARM::VLD3DUPd16Pseudo, ARM::VLD3DUPd16, true, false, false, SingleSpc, 3, 4,true}, 284 { ARM::VLD3DUPd16Pseudo_UPD, ARM::VLD3DUPd16_UPD, true, true, true, SingleSpc, 3, 4,true}, 285 { ARM::VLD3DUPd32Pseudo, ARM::VLD3DUPd32, true, false, false, SingleSpc, 3, 2,true}, 286 { ARM::VLD3DUPd32Pseudo_UPD, ARM::VLD3DUPd32_UPD, true, true, true, SingleSpc, 3, 2,true}, 287 { ARM::VLD3DUPd8Pseudo, ARM::VLD3DUPd8, true, false, false, SingleSpc, 3, 8,true}, 288 { ARM::VLD3DUPd8Pseudo_UPD, ARM::VLD3DUPd8_UPD, true, true, true, SingleSpc, 3, 8,true}, 289 { ARM::VLD3DUPq16EvenPseudo, ARM::VLD3DUPq16, true, false, false, EvenDblSpc, 3, 4 ,true}, 290 { ARM::VLD3DUPq16OddPseudo, ARM::VLD3DUPq16, true, false, false, OddDblSpc, 3, 4 ,true}, 291 { ARM::VLD3DUPq16OddPseudo_UPD, ARM::VLD3DUPq16_UPD, true, true, true, OddDblSpc, 3, 4 ,true}, 292 { ARM::VLD3DUPq32EvenPseudo, ARM::VLD3DUPq32, true, false, false, EvenDblSpc, 3, 2 ,true}, 293 { ARM::VLD3DUPq32OddPseudo, ARM::VLD3DUPq32, true, false, false, OddDblSpc, 3, 2 ,true}, 294 { ARM::VLD3DUPq32OddPseudo_UPD, ARM::VLD3DUPq32_UPD, true, true, true, OddDblSpc, 3, 2 ,true}, 295 { ARM::VLD3DUPq8EvenPseudo, ARM::VLD3DUPq8, true, false, false, EvenDblSpc, 3, 8 ,true}, 296 { ARM::VLD3DUPq8OddPseudo, ARM::VLD3DUPq8, true, false, false, OddDblSpc, 3, 8 ,true}, 297 { ARM::VLD3DUPq8OddPseudo_UPD, ARM::VLD3DUPq8_UPD, true, true, true, OddDblSpc, 3, 8 ,true}, 298 299 { ARM::VLD3LNd16Pseudo, ARM::VLD3LNd16, true, false, false, SingleSpc, 3, 4 ,true}, 300 { ARM::VLD3LNd16Pseudo_UPD, ARM::VLD3LNd16_UPD, true, true, true, SingleSpc, 3, 4 ,true}, 301 { ARM::VLD3LNd32Pseudo, ARM::VLD3LNd32, true, false, false, SingleSpc, 3, 2 ,true}, 302 { ARM::VLD3LNd32Pseudo_UPD, ARM::VLD3LNd32_UPD, true, true, true, SingleSpc, 3, 2 ,true}, 303 { ARM::VLD3LNd8Pseudo, ARM::VLD3LNd8, true, false, false, SingleSpc, 3, 8 ,true}, 304 { ARM::VLD3LNd8Pseudo_UPD, ARM::VLD3LNd8_UPD, true, true, true, SingleSpc, 3, 8 ,true}, 305 { ARM::VLD3LNq16Pseudo, ARM::VLD3LNq16, true, false, false, EvenDblSpc, 3, 4 ,true}, 306 { ARM::VLD3LNq16Pseudo_UPD, ARM::VLD3LNq16_UPD, true, true, true, EvenDblSpc, 3, 4 ,true}, 307 { ARM::VLD3LNq32Pseudo, ARM::VLD3LNq32, true, false, false, EvenDblSpc, 3, 2 ,true}, 308 { ARM::VLD3LNq32Pseudo_UPD, ARM::VLD3LNq32_UPD, true, true, true, EvenDblSpc, 3, 2 ,true}, 309 310 { ARM::VLD3d16Pseudo, ARM::VLD3d16, true, false, false, SingleSpc, 3, 4 ,true}, 311 { ARM::VLD3d16Pseudo_UPD, ARM::VLD3d16_UPD, true, true, true, SingleSpc, 3, 4 ,true}, 312 { ARM::VLD3d32Pseudo, ARM::VLD3d32, true, false, false, SingleSpc, 3, 2 ,true}, 313 { ARM::VLD3d32Pseudo_UPD, ARM::VLD3d32_UPD, true, true, true, SingleSpc, 3, 2 ,true}, 314 { ARM::VLD3d8Pseudo, ARM::VLD3d8, true, false, false, SingleSpc, 3, 8 ,true}, 315 { ARM::VLD3d8Pseudo_UPD, ARM::VLD3d8_UPD, true, true, true, SingleSpc, 3, 8 ,true}, 316 317 { ARM::VLD3q16Pseudo_UPD, ARM::VLD3q16_UPD, true, true, true, EvenDblSpc, 3, 4 ,true}, 318 { ARM::VLD3q16oddPseudo, ARM::VLD3q16, true, false, false, OddDblSpc, 3, 4 ,true}, 319 { ARM::VLD3q16oddPseudo_UPD, ARM::VLD3q16_UPD, true, true, true, OddDblSpc, 3, 4 ,true}, 320 { ARM::VLD3q32Pseudo_UPD, ARM::VLD3q32_UPD, true, true, true, EvenDblSpc, 3, 2 ,true}, 321 { ARM::VLD3q32oddPseudo, ARM::VLD3q32, true, false, false, OddDblSpc, 3, 2 ,true}, 322 { ARM::VLD3q32oddPseudo_UPD, ARM::VLD3q32_UPD, true, true, true, OddDblSpc, 3, 2 ,true}, 323 { ARM::VLD3q8Pseudo_UPD, ARM::VLD3q8_UPD, true, true, true, EvenDblSpc, 3, 8 ,true}, 324 { ARM::VLD3q8oddPseudo, ARM::VLD3q8, true, false, false, OddDblSpc, 3, 8 ,true}, 325 { ARM::VLD3q8oddPseudo_UPD, ARM::VLD3q8_UPD, true, true, true, OddDblSpc, 3, 8 ,true}, 326 327 { ARM::VLD4DUPd16Pseudo, ARM::VLD4DUPd16, true, false, false, SingleSpc, 4, 4,true}, 328 { ARM::VLD4DUPd16Pseudo_UPD, ARM::VLD4DUPd16_UPD, true, true, true, SingleSpc, 4, 4,true}, 329 { ARM::VLD4DUPd32Pseudo, ARM::VLD4DUPd32, true, false, false, SingleSpc, 4, 2,true}, 330 { ARM::VLD4DUPd32Pseudo_UPD, ARM::VLD4DUPd32_UPD, true, true, true, SingleSpc, 4, 2,true}, 331 { ARM::VLD4DUPd8Pseudo, ARM::VLD4DUPd8, true, false, false, SingleSpc, 4, 8,true}, 332 { ARM::VLD4DUPd8Pseudo_UPD, ARM::VLD4DUPd8_UPD, true, true, true, SingleSpc, 4, 8,true}, 333 { ARM::VLD4DUPq16EvenPseudo, ARM::VLD4DUPq16, true, false, false, EvenDblSpc, 4, 4 ,true}, 334 { ARM::VLD4DUPq16OddPseudo, ARM::VLD4DUPq16, true, false, false, OddDblSpc, 4, 4 ,true}, 335 { ARM::VLD4DUPq16OddPseudo_UPD, ARM::VLD4DUPq16_UPD, true, true, true, OddDblSpc, 4, 4 ,true}, 336 { ARM::VLD4DUPq32EvenPseudo, ARM::VLD4DUPq32, true, false, false, EvenDblSpc, 4, 2 ,true}, 337 { ARM::VLD4DUPq32OddPseudo, ARM::VLD4DUPq32, true, false, false, OddDblSpc, 4, 2 ,true}, 338 { ARM::VLD4DUPq32OddPseudo_UPD, ARM::VLD4DUPq32_UPD, true, true, true, OddDblSpc, 4, 2 ,true}, 339 { ARM::VLD4DUPq8EvenPseudo, ARM::VLD4DUPq8, true, false, false, EvenDblSpc, 4, 8 ,true}, 340 { ARM::VLD4DUPq8OddPseudo, ARM::VLD4DUPq8, true, false, false, OddDblSpc, 4, 8 ,true}, 341 { ARM::VLD4DUPq8OddPseudo_UPD, ARM::VLD4DUPq8_UPD, true, true, true, OddDblSpc, 4, 8 ,true}, 342 343 { ARM::VLD4LNd16Pseudo, ARM::VLD4LNd16, true, false, false, SingleSpc, 4, 4 ,true}, 344 { ARM::VLD4LNd16Pseudo_UPD, ARM::VLD4LNd16_UPD, true, true, true, SingleSpc, 4, 4 ,true}, 345 { ARM::VLD4LNd32Pseudo, ARM::VLD4LNd32, true, false, false, SingleSpc, 4, 2 ,true}, 346 { ARM::VLD4LNd32Pseudo_UPD, ARM::VLD4LNd32_UPD, true, true, true, SingleSpc, 4, 2 ,true}, 347 { ARM::VLD4LNd8Pseudo, ARM::VLD4LNd8, true, false, false, SingleSpc, 4, 8 ,true}, 348 { ARM::VLD4LNd8Pseudo_UPD, ARM::VLD4LNd8_UPD, true, true, true, SingleSpc, 4, 8 ,true}, 349 { ARM::VLD4LNq16Pseudo, ARM::VLD4LNq16, true, false, false, EvenDblSpc, 4, 4 ,true}, 350 { ARM::VLD4LNq16Pseudo_UPD, ARM::VLD4LNq16_UPD, true, true, true, EvenDblSpc, 4, 4 ,true}, 351 { ARM::VLD4LNq32Pseudo, ARM::VLD4LNq32, true, false, false, EvenDblSpc, 4, 2 ,true}, 352 { ARM::VLD4LNq32Pseudo_UPD, ARM::VLD4LNq32_UPD, true, true, true, EvenDblSpc, 4, 2 ,true}, 353 354 { ARM::VLD4d16Pseudo, ARM::VLD4d16, true, false, false, SingleSpc, 4, 4 ,true}, 355 { ARM::VLD4d16Pseudo_UPD, ARM::VLD4d16_UPD, true, true, true, SingleSpc, 4, 4 ,true}, 356 { ARM::VLD4d32Pseudo, ARM::VLD4d32, true, false, false, SingleSpc, 4, 2 ,true}, 357 { ARM::VLD4d32Pseudo_UPD, ARM::VLD4d32_UPD, true, true, true, SingleSpc, 4, 2 ,true}, 358 { ARM::VLD4d8Pseudo, ARM::VLD4d8, true, false, false, SingleSpc, 4, 8 ,true}, 359 { ARM::VLD4d8Pseudo_UPD, ARM::VLD4d8_UPD, true, true, true, SingleSpc, 4, 8 ,true}, 360 361 { ARM::VLD4q16Pseudo_UPD, ARM::VLD4q16_UPD, true, true, true, EvenDblSpc, 4, 4 ,true}, 362 { ARM::VLD4q16oddPseudo, ARM::VLD4q16, true, false, false, OddDblSpc, 4, 4 ,true}, 363 { ARM::VLD4q16oddPseudo_UPD, ARM::VLD4q16_UPD, true, true, true, OddDblSpc, 4, 4 ,true}, 364 { ARM::VLD4q32Pseudo_UPD, ARM::VLD4q32_UPD, true, true, true, EvenDblSpc, 4, 2 ,true}, 365 { ARM::VLD4q32oddPseudo, ARM::VLD4q32, true, false, false, OddDblSpc, 4, 2 ,true}, 366 { ARM::VLD4q32oddPseudo_UPD, ARM::VLD4q32_UPD, true, true, true, OddDblSpc, 4, 2 ,true}, 367 { ARM::VLD4q8Pseudo_UPD, ARM::VLD4q8_UPD, true, true, true, EvenDblSpc, 4, 8 ,true}, 368 { ARM::VLD4q8oddPseudo, ARM::VLD4q8, true, false, false, OddDblSpc, 4, 8 ,true}, 369 { ARM::VLD4q8oddPseudo_UPD, ARM::VLD4q8_UPD, true, true, true, OddDblSpc, 4, 8 ,true}, 370 371 { ARM::VST1LNq16Pseudo, ARM::VST1LNd16, false, false, false, EvenDblSpc, 1, 4 ,true}, 372 { ARM::VST1LNq16Pseudo_UPD, ARM::VST1LNd16_UPD, false, true, true, EvenDblSpc, 1, 4 ,true}, 373 { ARM::VST1LNq32Pseudo, ARM::VST1LNd32, false, false, false, EvenDblSpc, 1, 2 ,true}, 374 { ARM::VST1LNq32Pseudo_UPD, ARM::VST1LNd32_UPD, false, true, true, EvenDblSpc, 1, 2 ,true}, 375 { ARM::VST1LNq8Pseudo, ARM::VST1LNd8, false, false, false, EvenDblSpc, 1, 8 ,true}, 376 { ARM::VST1LNq8Pseudo_UPD, ARM::VST1LNd8_UPD, false, true, true, EvenDblSpc, 1, 8 ,true}, 377 378 { ARM::VST1d16QPseudo, ARM::VST1d16Q, false, false, false, SingleSpc, 4, 4 ,false}, 379 { ARM::VST1d16QPseudoWB_fixed, ARM::VST1d16Qwb_fixed, false, true, false, SingleSpc, 4, 4 ,false}, 380 { ARM::VST1d16QPseudoWB_register, ARM::VST1d16Qwb_register, false, true, true, SingleSpc, 4, 4 ,false}, 381 { ARM::VST1d16TPseudo, ARM::VST1d16T, false, false, false, SingleSpc, 3, 4 ,false}, 382 { ARM::VST1d16TPseudoWB_fixed, ARM::VST1d16Twb_fixed, false, true, false, SingleSpc, 3, 4 ,false}, 383 { ARM::VST1d16TPseudoWB_register, ARM::VST1d16Twb_register, false, true, true, SingleSpc, 3, 4 ,false}, 384 385 { ARM::VST1d32QPseudo, ARM::VST1d32Q, false, false, false, SingleSpc, 4, 2 ,false}, 386 { ARM::VST1d32QPseudoWB_fixed, ARM::VST1d32Qwb_fixed, false, true, false, SingleSpc, 4, 2 ,false}, 387 { ARM::VST1d32QPseudoWB_register, ARM::VST1d32Qwb_register, false, true, true, SingleSpc, 4, 2 ,false}, 388 { ARM::VST1d32TPseudo, ARM::VST1d32T, false, false, false, SingleSpc, 3, 2 ,false}, 389 { ARM::VST1d32TPseudoWB_fixed, ARM::VST1d32Twb_fixed, false, true, false, SingleSpc, 3, 2 ,false}, 390 { ARM::VST1d32TPseudoWB_register, ARM::VST1d32Twb_register, false, true, true, SingleSpc, 3, 2 ,false}, 391 392 { ARM::VST1d64QPseudo, ARM::VST1d64Q, false, false, false, SingleSpc, 4, 1 ,false}, 393 { ARM::VST1d64QPseudoWB_fixed, ARM::VST1d64Qwb_fixed, false, true, false, SingleSpc, 4, 1 ,false}, 394 { ARM::VST1d64QPseudoWB_register, ARM::VST1d64Qwb_register, false, true, true, SingleSpc, 4, 1 ,false}, 395 { ARM::VST1d64TPseudo, ARM::VST1d64T, false, false, false, SingleSpc, 3, 1 ,false}, 396 { ARM::VST1d64TPseudoWB_fixed, ARM::VST1d64Twb_fixed, false, true, false, SingleSpc, 3, 1 ,false}, 397 { ARM::VST1d64TPseudoWB_register, ARM::VST1d64Twb_register, false, true, true, SingleSpc, 3, 1 ,false}, 398 399 { ARM::VST1d8QPseudo, ARM::VST1d8Q, false, false, false, SingleSpc, 4, 8 ,false}, 400 { ARM::VST1d8QPseudoWB_fixed, ARM::VST1d8Qwb_fixed, false, true, false, SingleSpc, 4, 8 ,false}, 401 { ARM::VST1d8QPseudoWB_register, ARM::VST1d8Qwb_register, false, true, true, SingleSpc, 4, 8 ,false}, 402 { ARM::VST1d8TPseudo, ARM::VST1d8T, false, false, false, SingleSpc, 3, 8 ,false}, 403 { ARM::VST1d8TPseudoWB_fixed, ARM::VST1d8Twb_fixed, false, true, false, SingleSpc, 3, 8 ,false}, 404 { ARM::VST1d8TPseudoWB_register, ARM::VST1d8Twb_register, false, true, true, SingleSpc, 3, 8 ,false}, 405 406 { ARM::VST1q16HighQPseudo, ARM::VST1d16Q, false, false, false, SingleHighQSpc, 4, 4 ,false}, 407 { ARM::VST1q16HighQPseudo_UPD, ARM::VST1d16Qwb_fixed, false, true, true, SingleHighQSpc, 4, 8 ,false}, 408 { ARM::VST1q16HighTPseudo, ARM::VST1d16T, false, false, false, SingleHighTSpc, 3, 4 ,false}, 409 { ARM::VST1q16HighTPseudo_UPD, ARM::VST1d16Twb_fixed, false, true, true, SingleHighTSpc, 3, 4 ,false}, 410 { ARM::VST1q16LowQPseudo_UPD, ARM::VST1d16Qwb_fixed, false, true, true, SingleLowSpc, 4, 4 ,false}, 411 { ARM::VST1q16LowTPseudo_UPD, ARM::VST1d16Twb_fixed, false, true, true, SingleLowSpc, 3, 4 ,false}, 412 413 { ARM::VST1q32HighQPseudo, ARM::VST1d32Q, false, false, false, SingleHighQSpc, 4, 2 ,false}, 414 { ARM::VST1q32HighQPseudo_UPD, ARM::VST1d32Qwb_fixed, false, true, true, SingleHighQSpc, 4, 8 ,false}, 415 { ARM::VST1q32HighTPseudo, ARM::VST1d32T, false, false, false, SingleHighTSpc, 3, 2 ,false}, 416 { ARM::VST1q32HighTPseudo_UPD, ARM::VST1d32Twb_fixed, false, true, true, SingleHighTSpc, 3, 2 ,false}, 417 { ARM::VST1q32LowQPseudo_UPD, ARM::VST1d32Qwb_fixed, false, true, true, SingleLowSpc, 4, 2 ,false}, 418 { ARM::VST1q32LowTPseudo_UPD, ARM::VST1d32Twb_fixed, false, true, true, SingleLowSpc, 3, 2 ,false}, 419 420 { ARM::VST1q64HighQPseudo, ARM::VST1d64Q, false, false, false, SingleHighQSpc, 4, 1 ,false}, 421 { ARM::VST1q64HighQPseudo_UPD, ARM::VST1d64Qwb_fixed, false, true, true, SingleHighQSpc, 4, 8 ,false}, 422 { ARM::VST1q64HighTPseudo, ARM::VST1d64T, false, false, false, SingleHighTSpc, 3, 1 ,false}, 423 { ARM::VST1q64HighTPseudo_UPD, ARM::VST1d64Twb_fixed, false, true, true, SingleHighTSpc, 3, 1 ,false}, 424 { ARM::VST1q64LowQPseudo_UPD, ARM::VST1d64Qwb_fixed, false, true, true, SingleLowSpc, 4, 1 ,false}, 425 { ARM::VST1q64LowTPseudo_UPD, ARM::VST1d64Twb_fixed, false, true, true, SingleLowSpc, 3, 1 ,false}, 426 427 { ARM::VST1q8HighQPseudo, ARM::VST1d8Q, false, false, false, SingleHighQSpc, 4, 8 ,false}, 428 { ARM::VST1q8HighQPseudo_UPD, ARM::VST1d8Qwb_fixed, false, true, true, SingleHighQSpc, 4, 8 ,false}, 429 { ARM::VST1q8HighTPseudo, ARM::VST1d8T, false, false, false, SingleHighTSpc, 3, 8 ,false}, 430 { ARM::VST1q8HighTPseudo_UPD, ARM::VST1d8Twb_fixed, false, true, true, SingleHighTSpc, 3, 8 ,false}, 431 { ARM::VST1q8LowQPseudo_UPD, ARM::VST1d8Qwb_fixed, false, true, true, SingleLowSpc, 4, 8 ,false}, 432 { ARM::VST1q8LowTPseudo_UPD, ARM::VST1d8Twb_fixed, false, true, true, SingleLowSpc, 3, 8 ,false}, 433 434 { ARM::VST2LNd16Pseudo, ARM::VST2LNd16, false, false, false, SingleSpc, 2, 4 ,true}, 435 { ARM::VST2LNd16Pseudo_UPD, ARM::VST2LNd16_UPD, false, true, true, SingleSpc, 2, 4 ,true}, 436 { ARM::VST2LNd32Pseudo, ARM::VST2LNd32, false, false, false, SingleSpc, 2, 2 ,true}, 437 { ARM::VST2LNd32Pseudo_UPD, ARM::VST2LNd32_UPD, false, true, true, SingleSpc, 2, 2 ,true}, 438 { ARM::VST2LNd8Pseudo, ARM::VST2LNd8, false, false, false, SingleSpc, 2, 8 ,true}, 439 { ARM::VST2LNd8Pseudo_UPD, ARM::VST2LNd8_UPD, false, true, true, SingleSpc, 2, 8 ,true}, 440 { ARM::VST2LNq16Pseudo, ARM::VST2LNq16, false, false, false, EvenDblSpc, 2, 4,true}, 441 { ARM::VST2LNq16Pseudo_UPD, ARM::VST2LNq16_UPD, false, true, true, EvenDblSpc, 2, 4,true}, 442 { ARM::VST2LNq32Pseudo, ARM::VST2LNq32, false, false, false, EvenDblSpc, 2, 2,true}, 443 { ARM::VST2LNq32Pseudo_UPD, ARM::VST2LNq32_UPD, false, true, true, EvenDblSpc, 2, 2,true}, 444 445 { ARM::VST2q16Pseudo, ARM::VST2q16, false, false, false, SingleSpc, 4, 4 ,false}, 446 { ARM::VST2q16PseudoWB_fixed, ARM::VST2q16wb_fixed, false, true, false, SingleSpc, 4, 4 ,false}, 447 { ARM::VST2q16PseudoWB_register, ARM::VST2q16wb_register, false, true, true, SingleSpc, 4, 4 ,false}, 448 { ARM::VST2q32Pseudo, ARM::VST2q32, false, false, false, SingleSpc, 4, 2 ,false}, 449 { ARM::VST2q32PseudoWB_fixed, ARM::VST2q32wb_fixed, false, true, false, SingleSpc, 4, 2 ,false}, 450 { ARM::VST2q32PseudoWB_register, ARM::VST2q32wb_register, false, true, true, SingleSpc, 4, 2 ,false}, 451 { ARM::VST2q8Pseudo, ARM::VST2q8, false, false, false, SingleSpc, 4, 8 ,false}, 452 { ARM::VST2q8PseudoWB_fixed, ARM::VST2q8wb_fixed, false, true, false, SingleSpc, 4, 8 ,false}, 453 { ARM::VST2q8PseudoWB_register, ARM::VST2q8wb_register, false, true, true, SingleSpc, 4, 8 ,false}, 454 455 { ARM::VST3LNd16Pseudo, ARM::VST3LNd16, false, false, false, SingleSpc, 3, 4 ,true}, 456 { ARM::VST3LNd16Pseudo_UPD, ARM::VST3LNd16_UPD, false, true, true, SingleSpc, 3, 4 ,true}, 457 { ARM::VST3LNd32Pseudo, ARM::VST3LNd32, false, false, false, SingleSpc, 3, 2 ,true}, 458 { ARM::VST3LNd32Pseudo_UPD, ARM::VST3LNd32_UPD, false, true, true, SingleSpc, 3, 2 ,true}, 459 { ARM::VST3LNd8Pseudo, ARM::VST3LNd8, false, false, false, SingleSpc, 3, 8 ,true}, 460 { ARM::VST3LNd8Pseudo_UPD, ARM::VST3LNd8_UPD, false, true, true, SingleSpc, 3, 8 ,true}, 461 { ARM::VST3LNq16Pseudo, ARM::VST3LNq16, false, false, false, EvenDblSpc, 3, 4,true}, 462 { ARM::VST3LNq16Pseudo_UPD, ARM::VST3LNq16_UPD, false, true, true, EvenDblSpc, 3, 4,true}, 463 { ARM::VST3LNq32Pseudo, ARM::VST3LNq32, false, false, false, EvenDblSpc, 3, 2,true}, 464 { ARM::VST3LNq32Pseudo_UPD, ARM::VST3LNq32_UPD, false, true, true, EvenDblSpc, 3, 2,true}, 465 466 { ARM::VST3d16Pseudo, ARM::VST3d16, false, false, false, SingleSpc, 3, 4 ,true}, 467 { ARM::VST3d16Pseudo_UPD, ARM::VST3d16_UPD, false, true, true, SingleSpc, 3, 4 ,true}, 468 { ARM::VST3d32Pseudo, ARM::VST3d32, false, false, false, SingleSpc, 3, 2 ,true}, 469 { ARM::VST3d32Pseudo_UPD, ARM::VST3d32_UPD, false, true, true, SingleSpc, 3, 2 ,true}, 470 { ARM::VST3d8Pseudo, ARM::VST3d8, false, false, false, SingleSpc, 3, 8 ,true}, 471 { ARM::VST3d8Pseudo_UPD, ARM::VST3d8_UPD, false, true, true, SingleSpc, 3, 8 ,true}, 472 473 { ARM::VST3q16Pseudo_UPD, ARM::VST3q16_UPD, false, true, true, EvenDblSpc, 3, 4 ,true}, 474 { ARM::VST3q16oddPseudo, ARM::VST3q16, false, false, false, OddDblSpc, 3, 4 ,true}, 475 { ARM::VST3q16oddPseudo_UPD, ARM::VST3q16_UPD, false, true, true, OddDblSpc, 3, 4 ,true}, 476 { ARM::VST3q32Pseudo_UPD, ARM::VST3q32_UPD, false, true, true, EvenDblSpc, 3, 2 ,true}, 477 { ARM::VST3q32oddPseudo, ARM::VST3q32, false, false, false, OddDblSpc, 3, 2 ,true}, 478 { ARM::VST3q32oddPseudo_UPD, ARM::VST3q32_UPD, false, true, true, OddDblSpc, 3, 2 ,true}, 479 { ARM::VST3q8Pseudo_UPD, ARM::VST3q8_UPD, false, true, true, EvenDblSpc, 3, 8 ,true}, 480 { ARM::VST3q8oddPseudo, ARM::VST3q8, false, false, false, OddDblSpc, 3, 8 ,true}, 481 { ARM::VST3q8oddPseudo_UPD, ARM::VST3q8_UPD, false, true, true, OddDblSpc, 3, 8 ,true}, 482 483 { ARM::VST4LNd16Pseudo, ARM::VST4LNd16, false, false, false, SingleSpc, 4, 4 ,true}, 484 { ARM::VST4LNd16Pseudo_UPD, ARM::VST4LNd16_UPD, false, true, true, SingleSpc, 4, 4 ,true}, 485 { ARM::VST4LNd32Pseudo, ARM::VST4LNd32, false, false, false, SingleSpc, 4, 2 ,true}, 486 { ARM::VST4LNd32Pseudo_UPD, ARM::VST4LNd32_UPD, false, true, true, SingleSpc, 4, 2 ,true}, 487 { ARM::VST4LNd8Pseudo, ARM::VST4LNd8, false, false, false, SingleSpc, 4, 8 ,true}, 488 { ARM::VST4LNd8Pseudo_UPD, ARM::VST4LNd8_UPD, false, true, true, SingleSpc, 4, 8 ,true}, 489 { ARM::VST4LNq16Pseudo, ARM::VST4LNq16, false, false, false, EvenDblSpc, 4, 4,true}, 490 { ARM::VST4LNq16Pseudo_UPD, ARM::VST4LNq16_UPD, false, true, true, EvenDblSpc, 4, 4,true}, 491 { ARM::VST4LNq32Pseudo, ARM::VST4LNq32, false, false, false, EvenDblSpc, 4, 2,true}, 492 { ARM::VST4LNq32Pseudo_UPD, ARM::VST4LNq32_UPD, false, true, true, EvenDblSpc, 4, 2,true}, 493 494 { ARM::VST4d16Pseudo, ARM::VST4d16, false, false, false, SingleSpc, 4, 4 ,true}, 495 { ARM::VST4d16Pseudo_UPD, ARM::VST4d16_UPD, false, true, true, SingleSpc, 4, 4 ,true}, 496 { ARM::VST4d32Pseudo, ARM::VST4d32, false, false, false, SingleSpc, 4, 2 ,true}, 497 { ARM::VST4d32Pseudo_UPD, ARM::VST4d32_UPD, false, true, true, SingleSpc, 4, 2 ,true}, 498 { ARM::VST4d8Pseudo, ARM::VST4d8, false, false, false, SingleSpc, 4, 8 ,true}, 499 { ARM::VST4d8Pseudo_UPD, ARM::VST4d8_UPD, false, true, true, SingleSpc, 4, 8 ,true}, 500 501 { ARM::VST4q16Pseudo_UPD, ARM::VST4q16_UPD, false, true, true, EvenDblSpc, 4, 4 ,true}, 502 { ARM::VST4q16oddPseudo, ARM::VST4q16, false, false, false, OddDblSpc, 4, 4 ,true}, 503 { ARM::VST4q16oddPseudo_UPD, ARM::VST4q16_UPD, false, true, true, OddDblSpc, 4, 4 ,true}, 504 { ARM::VST4q32Pseudo_UPD, ARM::VST4q32_UPD, false, true, true, EvenDblSpc, 4, 2 ,true}, 505 { ARM::VST4q32oddPseudo, ARM::VST4q32, false, false, false, OddDblSpc, 4, 2 ,true}, 506 { ARM::VST4q32oddPseudo_UPD, ARM::VST4q32_UPD, false, true, true, OddDblSpc, 4, 2 ,true}, 507 { ARM::VST4q8Pseudo_UPD, ARM::VST4q8_UPD, false, true, true, EvenDblSpc, 4, 8 ,true}, 508 { ARM::VST4q8oddPseudo, ARM::VST4q8, false, false, false, OddDblSpc, 4, 8 ,true}, 509 { ARM::VST4q8oddPseudo_UPD, ARM::VST4q8_UPD, false, true, true, OddDblSpc, 4, 8 ,true} 510 }; 511 512 /// LookupNEONLdSt - Search the NEONLdStTable for information about a NEON 513 /// load or store pseudo instruction. 514 static const NEONLdStTableEntry *LookupNEONLdSt(unsigned Opcode) { 515 #ifndef NDEBUG 516 // Make sure the table is sorted. 517 static std::atomic<bool> TableChecked(false); 518 if (!TableChecked.load(std::memory_order_relaxed)) { 519 assert(llvm::is_sorted(NEONLdStTable) && "NEONLdStTable is not sorted!"); 520 TableChecked.store(true, std::memory_order_relaxed); 521 } 522 #endif 523 524 auto I = llvm::lower_bound(NEONLdStTable, Opcode); 525 if (I != std::end(NEONLdStTable) && I->PseudoOpc == Opcode) 526 return I; 527 return nullptr; 528 } 529 530 /// GetDSubRegs - Get 4 D subregisters of a Q, QQ, or QQQQ register, 531 /// corresponding to the specified register spacing. Not all of the results 532 /// are necessarily valid, e.g., a Q register only has 2 D subregisters. 533 static void GetDSubRegs(unsigned Reg, NEONRegSpacing RegSpc, 534 const TargetRegisterInfo *TRI, unsigned &D0, 535 unsigned &D1, unsigned &D2, unsigned &D3) { 536 if (RegSpc == SingleSpc || RegSpc == SingleLowSpc) { 537 D0 = TRI->getSubReg(Reg, ARM::dsub_0); 538 D1 = TRI->getSubReg(Reg, ARM::dsub_1); 539 D2 = TRI->getSubReg(Reg, ARM::dsub_2); 540 D3 = TRI->getSubReg(Reg, ARM::dsub_3); 541 } else if (RegSpc == SingleHighQSpc) { 542 D0 = TRI->getSubReg(Reg, ARM::dsub_4); 543 D1 = TRI->getSubReg(Reg, ARM::dsub_5); 544 D2 = TRI->getSubReg(Reg, ARM::dsub_6); 545 D3 = TRI->getSubReg(Reg, ARM::dsub_7); 546 } else if (RegSpc == SingleHighTSpc) { 547 D0 = TRI->getSubReg(Reg, ARM::dsub_3); 548 D1 = TRI->getSubReg(Reg, ARM::dsub_4); 549 D2 = TRI->getSubReg(Reg, ARM::dsub_5); 550 D3 = TRI->getSubReg(Reg, ARM::dsub_6); 551 } else if (RegSpc == EvenDblSpc) { 552 D0 = TRI->getSubReg(Reg, ARM::dsub_0); 553 D1 = TRI->getSubReg(Reg, ARM::dsub_2); 554 D2 = TRI->getSubReg(Reg, ARM::dsub_4); 555 D3 = TRI->getSubReg(Reg, ARM::dsub_6); 556 } else { 557 assert(RegSpc == OddDblSpc && "unknown register spacing"); 558 D0 = TRI->getSubReg(Reg, ARM::dsub_1); 559 D1 = TRI->getSubReg(Reg, ARM::dsub_3); 560 D2 = TRI->getSubReg(Reg, ARM::dsub_5); 561 D3 = TRI->getSubReg(Reg, ARM::dsub_7); 562 } 563 } 564 565 /// ExpandVLD - Translate VLD pseudo instructions with Q, QQ or QQQQ register 566 /// operands to real VLD instructions with D register operands. 567 void ARMExpandPseudo::ExpandVLD(MachineBasicBlock::iterator &MBBI) { 568 MachineInstr &MI = *MBBI; 569 MachineBasicBlock &MBB = *MI.getParent(); 570 LLVM_DEBUG(dbgs() << "Expanding: "; MI.dump()); 571 572 const NEONLdStTableEntry *TableEntry = LookupNEONLdSt(MI.getOpcode()); 573 assert(TableEntry && TableEntry->IsLoad && "NEONLdStTable lookup failed"); 574 NEONRegSpacing RegSpc = (NEONRegSpacing)TableEntry->RegSpacing; 575 unsigned NumRegs = TableEntry->NumRegs; 576 577 MachineInstrBuilder MIB = BuildMI(MBB, MBBI, MI.getDebugLoc(), 578 TII->get(TableEntry->RealOpc)); 579 unsigned OpIdx = 0; 580 581 bool DstIsDead = MI.getOperand(OpIdx).isDead(); 582 Register DstReg = MI.getOperand(OpIdx++).getReg(); 583 584 bool IsVLD2DUP = TableEntry->RealOpc == ARM::VLD2DUPd8x2 || 585 TableEntry->RealOpc == ARM::VLD2DUPd16x2 || 586 TableEntry->RealOpc == ARM::VLD2DUPd32x2 || 587 TableEntry->RealOpc == ARM::VLD2DUPd8x2wb_fixed || 588 TableEntry->RealOpc == ARM::VLD2DUPd16x2wb_fixed || 589 TableEntry->RealOpc == ARM::VLD2DUPd32x2wb_fixed || 590 TableEntry->RealOpc == ARM::VLD2DUPd8x2wb_register || 591 TableEntry->RealOpc == ARM::VLD2DUPd16x2wb_register || 592 TableEntry->RealOpc == ARM::VLD2DUPd32x2wb_register; 593 594 if (IsVLD2DUP) { 595 unsigned SubRegIndex; 596 if (RegSpc == EvenDblSpc) { 597 SubRegIndex = ARM::dsub_0; 598 } else { 599 assert(RegSpc == OddDblSpc && "Unexpected spacing!"); 600 SubRegIndex = ARM::dsub_1; 601 } 602 Register SubReg = TRI->getSubReg(DstReg, SubRegIndex); 603 unsigned DstRegPair = TRI->getMatchingSuperReg(SubReg, ARM::dsub_0, 604 &ARM::DPairSpcRegClass); 605 MIB.addReg(DstRegPair, RegState::Define | getDeadRegState(DstIsDead)); 606 } else { 607 unsigned D0, D1, D2, D3; 608 GetDSubRegs(DstReg, RegSpc, TRI, D0, D1, D2, D3); 609 MIB.addReg(D0, RegState::Define | getDeadRegState(DstIsDead)); 610 if (NumRegs > 1 && TableEntry->copyAllListRegs) 611 MIB.addReg(D1, RegState::Define | getDeadRegState(DstIsDead)); 612 if (NumRegs > 2 && TableEntry->copyAllListRegs) 613 MIB.addReg(D2, RegState::Define | getDeadRegState(DstIsDead)); 614 if (NumRegs > 3 && TableEntry->copyAllListRegs) 615 MIB.addReg(D3, RegState::Define | getDeadRegState(DstIsDead)); 616 } 617 618 if (TableEntry->isUpdating) 619 MIB.add(MI.getOperand(OpIdx++)); 620 621 // Copy the addrmode6 operands. 622 MIB.add(MI.getOperand(OpIdx++)); 623 MIB.add(MI.getOperand(OpIdx++)); 624 625 // Copy the am6offset operand. 626 if (TableEntry->hasWritebackOperand) { 627 // TODO: The writing-back pseudo instructions we translate here are all 628 // defined to take am6offset nodes that are capable to represent both fixed 629 // and register forms. Some real instructions, however, do not rely on 630 // am6offset and have separate definitions for such forms. When this is the 631 // case, fixed forms do not take any offset nodes, so here we skip them for 632 // such instructions. Once all real and pseudo writing-back instructions are 633 // rewritten without use of am6offset nodes, this code will go away. 634 const MachineOperand &AM6Offset = MI.getOperand(OpIdx++); 635 if (TableEntry->RealOpc == ARM::VLD1d8Qwb_fixed || 636 TableEntry->RealOpc == ARM::VLD1d16Qwb_fixed || 637 TableEntry->RealOpc == ARM::VLD1d32Qwb_fixed || 638 TableEntry->RealOpc == ARM::VLD1d64Qwb_fixed || 639 TableEntry->RealOpc == ARM::VLD1d8Twb_fixed || 640 TableEntry->RealOpc == ARM::VLD1d16Twb_fixed || 641 TableEntry->RealOpc == ARM::VLD1d32Twb_fixed || 642 TableEntry->RealOpc == ARM::VLD1d64Twb_fixed || 643 TableEntry->RealOpc == ARM::VLD2DUPd8x2wb_fixed || 644 TableEntry->RealOpc == ARM::VLD2DUPd16x2wb_fixed || 645 TableEntry->RealOpc == ARM::VLD2DUPd32x2wb_fixed) { 646 assert(AM6Offset.getReg() == 0 && 647 "A fixed writing-back pseudo instruction provides an offset " 648 "register!"); 649 } else { 650 MIB.add(AM6Offset); 651 } 652 } 653 654 // For an instruction writing double-spaced subregs, the pseudo instruction 655 // has an extra operand that is a use of the super-register. Record the 656 // operand index and skip over it. 657 unsigned SrcOpIdx = 0; 658 if (!IsVLD2DUP) { 659 if (RegSpc == EvenDblSpc || RegSpc == OddDblSpc || 660 RegSpc == SingleLowSpc || RegSpc == SingleHighQSpc || 661 RegSpc == SingleHighTSpc) 662 SrcOpIdx = OpIdx++; 663 } 664 665 // Copy the predicate operands. 666 MIB.add(MI.getOperand(OpIdx++)); 667 MIB.add(MI.getOperand(OpIdx++)); 668 669 // Copy the super-register source operand used for double-spaced subregs over 670 // to the new instruction as an implicit operand. 671 if (SrcOpIdx != 0) { 672 MachineOperand MO = MI.getOperand(SrcOpIdx); 673 MO.setImplicit(true); 674 MIB.add(MO); 675 } 676 // Add an implicit def for the super-register. 677 MIB.addReg(DstReg, RegState::ImplicitDefine | getDeadRegState(DstIsDead)); 678 TransferImpOps(MI, MIB, MIB); 679 680 // Transfer memoperands. 681 MIB.cloneMemRefs(MI); 682 MI.eraseFromParent(); 683 LLVM_DEBUG(dbgs() << "To: "; MIB.getInstr()->dump();); 684 } 685 686 /// ExpandVST - Translate VST pseudo instructions with Q, QQ or QQQQ register 687 /// operands to real VST instructions with D register operands. 688 void ARMExpandPseudo::ExpandVST(MachineBasicBlock::iterator &MBBI) { 689 MachineInstr &MI = *MBBI; 690 MachineBasicBlock &MBB = *MI.getParent(); 691 LLVM_DEBUG(dbgs() << "Expanding: "; MI.dump()); 692 693 const NEONLdStTableEntry *TableEntry = LookupNEONLdSt(MI.getOpcode()); 694 assert(TableEntry && !TableEntry->IsLoad && "NEONLdStTable lookup failed"); 695 NEONRegSpacing RegSpc = (NEONRegSpacing)TableEntry->RegSpacing; 696 unsigned NumRegs = TableEntry->NumRegs; 697 698 MachineInstrBuilder MIB = BuildMI(MBB, MBBI, MI.getDebugLoc(), 699 TII->get(TableEntry->RealOpc)); 700 unsigned OpIdx = 0; 701 if (TableEntry->isUpdating) 702 MIB.add(MI.getOperand(OpIdx++)); 703 704 // Copy the addrmode6 operands. 705 MIB.add(MI.getOperand(OpIdx++)); 706 MIB.add(MI.getOperand(OpIdx++)); 707 708 if (TableEntry->hasWritebackOperand) { 709 // TODO: The writing-back pseudo instructions we translate here are all 710 // defined to take am6offset nodes that are capable to represent both fixed 711 // and register forms. Some real instructions, however, do not rely on 712 // am6offset and have separate definitions for such forms. When this is the 713 // case, fixed forms do not take any offset nodes, so here we skip them for 714 // such instructions. Once all real and pseudo writing-back instructions are 715 // rewritten without use of am6offset nodes, this code will go away. 716 const MachineOperand &AM6Offset = MI.getOperand(OpIdx++); 717 if (TableEntry->RealOpc == ARM::VST1d8Qwb_fixed || 718 TableEntry->RealOpc == ARM::VST1d16Qwb_fixed || 719 TableEntry->RealOpc == ARM::VST1d32Qwb_fixed || 720 TableEntry->RealOpc == ARM::VST1d64Qwb_fixed || 721 TableEntry->RealOpc == ARM::VST1d8Twb_fixed || 722 TableEntry->RealOpc == ARM::VST1d16Twb_fixed || 723 TableEntry->RealOpc == ARM::VST1d32Twb_fixed || 724 TableEntry->RealOpc == ARM::VST1d64Twb_fixed) { 725 assert(AM6Offset.getReg() == 0 && 726 "A fixed writing-back pseudo instruction provides an offset " 727 "register!"); 728 } else { 729 MIB.add(AM6Offset); 730 } 731 } 732 733 bool SrcIsKill = MI.getOperand(OpIdx).isKill(); 734 bool SrcIsUndef = MI.getOperand(OpIdx).isUndef(); 735 Register SrcReg = MI.getOperand(OpIdx++).getReg(); 736 unsigned D0, D1, D2, D3; 737 GetDSubRegs(SrcReg, RegSpc, TRI, D0, D1, D2, D3); 738 MIB.addReg(D0, getUndefRegState(SrcIsUndef)); 739 if (NumRegs > 1 && TableEntry->copyAllListRegs) 740 MIB.addReg(D1, getUndefRegState(SrcIsUndef)); 741 if (NumRegs > 2 && TableEntry->copyAllListRegs) 742 MIB.addReg(D2, getUndefRegState(SrcIsUndef)); 743 if (NumRegs > 3 && TableEntry->copyAllListRegs) 744 MIB.addReg(D3, getUndefRegState(SrcIsUndef)); 745 746 // Copy the predicate operands. 747 MIB.add(MI.getOperand(OpIdx++)); 748 MIB.add(MI.getOperand(OpIdx++)); 749 750 if (SrcIsKill && !SrcIsUndef) // Add an implicit kill for the super-reg. 751 MIB->addRegisterKilled(SrcReg, TRI, true); 752 else if (!SrcIsUndef) 753 MIB.addReg(SrcReg, RegState::Implicit); // Add implicit uses for src reg. 754 TransferImpOps(MI, MIB, MIB); 755 756 // Transfer memoperands. 757 MIB.cloneMemRefs(MI); 758 MI.eraseFromParent(); 759 LLVM_DEBUG(dbgs() << "To: "; MIB.getInstr()->dump();); 760 } 761 762 /// ExpandLaneOp - Translate VLD*LN and VST*LN instructions with Q, QQ or QQQQ 763 /// register operands to real instructions with D register operands. 764 void ARMExpandPseudo::ExpandLaneOp(MachineBasicBlock::iterator &MBBI) { 765 MachineInstr &MI = *MBBI; 766 MachineBasicBlock &MBB = *MI.getParent(); 767 LLVM_DEBUG(dbgs() << "Expanding: "; MI.dump()); 768 769 const NEONLdStTableEntry *TableEntry = LookupNEONLdSt(MI.getOpcode()); 770 assert(TableEntry && "NEONLdStTable lookup failed"); 771 NEONRegSpacing RegSpc = (NEONRegSpacing)TableEntry->RegSpacing; 772 unsigned NumRegs = TableEntry->NumRegs; 773 unsigned RegElts = TableEntry->RegElts; 774 775 MachineInstrBuilder MIB = BuildMI(MBB, MBBI, MI.getDebugLoc(), 776 TII->get(TableEntry->RealOpc)); 777 unsigned OpIdx = 0; 778 // The lane operand is always the 3rd from last operand, before the 2 779 // predicate operands. 780 unsigned Lane = MI.getOperand(MI.getDesc().getNumOperands() - 3).getImm(); 781 782 // Adjust the lane and spacing as needed for Q registers. 783 assert(RegSpc != OddDblSpc && "unexpected register spacing for VLD/VST-lane"); 784 if (RegSpc == EvenDblSpc && Lane >= RegElts) { 785 RegSpc = OddDblSpc; 786 Lane -= RegElts; 787 } 788 assert(Lane < RegElts && "out of range lane for VLD/VST-lane"); 789 790 unsigned D0 = 0, D1 = 0, D2 = 0, D3 = 0; 791 unsigned DstReg = 0; 792 bool DstIsDead = false; 793 if (TableEntry->IsLoad) { 794 DstIsDead = MI.getOperand(OpIdx).isDead(); 795 DstReg = MI.getOperand(OpIdx++).getReg(); 796 GetDSubRegs(DstReg, RegSpc, TRI, D0, D1, D2, D3); 797 MIB.addReg(D0, RegState::Define | getDeadRegState(DstIsDead)); 798 if (NumRegs > 1) 799 MIB.addReg(D1, RegState::Define | getDeadRegState(DstIsDead)); 800 if (NumRegs > 2) 801 MIB.addReg(D2, RegState::Define | getDeadRegState(DstIsDead)); 802 if (NumRegs > 3) 803 MIB.addReg(D3, RegState::Define | getDeadRegState(DstIsDead)); 804 } 805 806 if (TableEntry->isUpdating) 807 MIB.add(MI.getOperand(OpIdx++)); 808 809 // Copy the addrmode6 operands. 810 MIB.add(MI.getOperand(OpIdx++)); 811 MIB.add(MI.getOperand(OpIdx++)); 812 // Copy the am6offset operand. 813 if (TableEntry->hasWritebackOperand) 814 MIB.add(MI.getOperand(OpIdx++)); 815 816 // Grab the super-register source. 817 MachineOperand MO = MI.getOperand(OpIdx++); 818 if (!TableEntry->IsLoad) 819 GetDSubRegs(MO.getReg(), RegSpc, TRI, D0, D1, D2, D3); 820 821 // Add the subregs as sources of the new instruction. 822 unsigned SrcFlags = (getUndefRegState(MO.isUndef()) | 823 getKillRegState(MO.isKill())); 824 MIB.addReg(D0, SrcFlags); 825 if (NumRegs > 1) 826 MIB.addReg(D1, SrcFlags); 827 if (NumRegs > 2) 828 MIB.addReg(D2, SrcFlags); 829 if (NumRegs > 3) 830 MIB.addReg(D3, SrcFlags); 831 832 // Add the lane number operand. 833 MIB.addImm(Lane); 834 OpIdx += 1; 835 836 // Copy the predicate operands. 837 MIB.add(MI.getOperand(OpIdx++)); 838 MIB.add(MI.getOperand(OpIdx++)); 839 840 // Copy the super-register source to be an implicit source. 841 MO.setImplicit(true); 842 MIB.add(MO); 843 if (TableEntry->IsLoad) 844 // Add an implicit def for the super-register. 845 MIB.addReg(DstReg, RegState::ImplicitDefine | getDeadRegState(DstIsDead)); 846 TransferImpOps(MI, MIB, MIB); 847 // Transfer memoperands. 848 MIB.cloneMemRefs(MI); 849 MI.eraseFromParent(); 850 } 851 852 /// ExpandVTBL - Translate VTBL and VTBX pseudo instructions with Q or QQ 853 /// register operands to real instructions with D register operands. 854 void ARMExpandPseudo::ExpandVTBL(MachineBasicBlock::iterator &MBBI, 855 unsigned Opc, bool IsExt) { 856 MachineInstr &MI = *MBBI; 857 MachineBasicBlock &MBB = *MI.getParent(); 858 LLVM_DEBUG(dbgs() << "Expanding: "; MI.dump()); 859 860 MachineInstrBuilder MIB = BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(Opc)); 861 unsigned OpIdx = 0; 862 863 // Transfer the destination register operand. 864 MIB.add(MI.getOperand(OpIdx++)); 865 if (IsExt) { 866 MachineOperand VdSrc(MI.getOperand(OpIdx++)); 867 MIB.add(VdSrc); 868 } 869 870 bool SrcIsKill = MI.getOperand(OpIdx).isKill(); 871 Register SrcReg = MI.getOperand(OpIdx++).getReg(); 872 unsigned D0, D1, D2, D3; 873 GetDSubRegs(SrcReg, SingleSpc, TRI, D0, D1, D2, D3); 874 MIB.addReg(D0); 875 876 // Copy the other source register operand. 877 MachineOperand VmSrc(MI.getOperand(OpIdx++)); 878 MIB.add(VmSrc); 879 880 // Copy the predicate operands. 881 MIB.add(MI.getOperand(OpIdx++)); 882 MIB.add(MI.getOperand(OpIdx++)); 883 884 // Add an implicit kill and use for the super-reg. 885 MIB.addReg(SrcReg, RegState::Implicit | getKillRegState(SrcIsKill)); 886 TransferImpOps(MI, MIB, MIB); 887 MI.eraseFromParent(); 888 LLVM_DEBUG(dbgs() << "To: "; MIB.getInstr()->dump();); 889 } 890 891 void ARMExpandPseudo::ExpandMQQPRLoadStore(MachineBasicBlock::iterator &MBBI) { 892 MachineInstr &MI = *MBBI; 893 MachineBasicBlock &MBB = *MI.getParent(); 894 unsigned NewOpc = 895 MI.getOpcode() == ARM::MQQPRStore || MI.getOpcode() == ARM::MQQQQPRStore 896 ? ARM::VSTMDIA 897 : ARM::VLDMDIA; 898 MachineInstrBuilder MIB = 899 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(NewOpc)); 900 901 unsigned Flags = getKillRegState(MI.getOperand(0).isKill()) | 902 getDefRegState(MI.getOperand(0).isDef()); 903 Register SrcReg = MI.getOperand(0).getReg(); 904 905 // Copy the destination register. 906 MIB.add(MI.getOperand(1)); 907 MIB.add(predOps(ARMCC::AL)); 908 MIB.addReg(TRI->getSubReg(SrcReg, ARM::dsub_0), Flags); 909 MIB.addReg(TRI->getSubReg(SrcReg, ARM::dsub_1), Flags); 910 MIB.addReg(TRI->getSubReg(SrcReg, ARM::dsub_2), Flags); 911 MIB.addReg(TRI->getSubReg(SrcReg, ARM::dsub_3), Flags); 912 if (MI.getOpcode() == ARM::MQQQQPRStore || 913 MI.getOpcode() == ARM::MQQQQPRLoad) { 914 MIB.addReg(TRI->getSubReg(SrcReg, ARM::dsub_4), Flags); 915 MIB.addReg(TRI->getSubReg(SrcReg, ARM::dsub_5), Flags); 916 MIB.addReg(TRI->getSubReg(SrcReg, ARM::dsub_6), Flags); 917 MIB.addReg(TRI->getSubReg(SrcReg, ARM::dsub_7), Flags); 918 } 919 920 if (NewOpc == ARM::VSTMDIA) 921 MIB.addReg(SrcReg, RegState::Implicit); 922 923 TransferImpOps(MI, MIB, MIB); 924 MIB.cloneMemRefs(MI); 925 MI.eraseFromParent(); 926 } 927 928 static bool IsAnAddressOperand(const MachineOperand &MO) { 929 // This check is overly conservative. Unless we are certain that the machine 930 // operand is not a symbol reference, we return that it is a symbol reference. 931 // This is important as the load pair may not be split up Windows. 932 switch (MO.getType()) { 933 case MachineOperand::MO_Register: 934 case MachineOperand::MO_Immediate: 935 case MachineOperand::MO_CImmediate: 936 case MachineOperand::MO_FPImmediate: 937 case MachineOperand::MO_ShuffleMask: 938 return false; 939 case MachineOperand::MO_MachineBasicBlock: 940 return true; 941 case MachineOperand::MO_FrameIndex: 942 return false; 943 case MachineOperand::MO_ConstantPoolIndex: 944 case MachineOperand::MO_TargetIndex: 945 case MachineOperand::MO_JumpTableIndex: 946 case MachineOperand::MO_ExternalSymbol: 947 case MachineOperand::MO_GlobalAddress: 948 case MachineOperand::MO_BlockAddress: 949 return true; 950 case MachineOperand::MO_RegisterMask: 951 case MachineOperand::MO_RegisterLiveOut: 952 return false; 953 case MachineOperand::MO_Metadata: 954 case MachineOperand::MO_MCSymbol: 955 return true; 956 case MachineOperand::MO_DbgInstrRef: 957 case MachineOperand::MO_CFIIndex: 958 return false; 959 case MachineOperand::MO_IntrinsicID: 960 case MachineOperand::MO_Predicate: 961 llvm_unreachable("should not exist post-isel"); 962 } 963 llvm_unreachable("unhandled machine operand type"); 964 } 965 966 static MachineOperand makeImplicit(const MachineOperand &MO) { 967 MachineOperand NewMO = MO; 968 NewMO.setImplicit(); 969 return NewMO; 970 } 971 972 void ARMExpandPseudo::ExpandMOV32BitImm(MachineBasicBlock &MBB, 973 MachineBasicBlock::iterator &MBBI) { 974 MachineInstr &MI = *MBBI; 975 unsigned Opcode = MI.getOpcode(); 976 Register PredReg; 977 ARMCC::CondCodes Pred = getInstrPredicate(MI, PredReg); 978 Register DstReg = MI.getOperand(0).getReg(); 979 bool DstIsDead = MI.getOperand(0).isDead(); 980 bool isCC = Opcode == ARM::MOVCCi32imm || Opcode == ARM::t2MOVCCi32imm; 981 const MachineOperand &MO = MI.getOperand(isCC ? 2 : 1); 982 bool RequiresBundling = STI->isTargetWindows() && IsAnAddressOperand(MO); 983 MachineInstrBuilder LO16, HI16; 984 LLVM_DEBUG(dbgs() << "Expanding: "; MI.dump()); 985 986 if (!STI->hasV6T2Ops() && 987 (Opcode == ARM::MOVi32imm || Opcode == ARM::MOVCCi32imm)) { 988 // FIXME Windows CE supports older ARM CPUs 989 assert(!STI->isTargetWindows() && "Windows on ARM requires ARMv7+"); 990 991 assert (MO.isImm() && "MOVi32imm w/ non-immediate source operand!"); 992 unsigned ImmVal = (unsigned)MO.getImm(); 993 unsigned SOImmValV1 = 0, SOImmValV2 = 0; 994 995 if (ARM_AM::isSOImmTwoPartVal(ImmVal)) { // Expand into a movi + orr. 996 LO16 = BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(ARM::MOVi), DstReg); 997 HI16 = BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(ARM::ORRri)) 998 .addReg(DstReg, RegState::Define | getDeadRegState(DstIsDead)) 999 .addReg(DstReg); 1000 SOImmValV1 = ARM_AM::getSOImmTwoPartFirst(ImmVal); 1001 SOImmValV2 = ARM_AM::getSOImmTwoPartSecond(ImmVal); 1002 } else { // Expand into a mvn + sub. 1003 LO16 = BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(ARM::MVNi), DstReg); 1004 HI16 = BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(ARM::SUBri)) 1005 .addReg(DstReg, RegState::Define | getDeadRegState(DstIsDead)) 1006 .addReg(DstReg); 1007 SOImmValV1 = ARM_AM::getSOImmTwoPartFirst(-ImmVal); 1008 SOImmValV2 = ARM_AM::getSOImmTwoPartSecond(-ImmVal); 1009 SOImmValV1 = ~(-SOImmValV1); 1010 } 1011 1012 unsigned MIFlags = MI.getFlags(); 1013 LO16 = LO16.addImm(SOImmValV1); 1014 HI16 = HI16.addImm(SOImmValV2); 1015 LO16.cloneMemRefs(MI); 1016 HI16.cloneMemRefs(MI); 1017 LO16.setMIFlags(MIFlags); 1018 HI16.setMIFlags(MIFlags); 1019 LO16.addImm(Pred).addReg(PredReg).add(condCodeOp()); 1020 HI16.addImm(Pred).addReg(PredReg).add(condCodeOp()); 1021 if (isCC) 1022 LO16.add(makeImplicit(MI.getOperand(1))); 1023 TransferImpOps(MI, LO16, HI16); 1024 MI.eraseFromParent(); 1025 return; 1026 } 1027 1028 unsigned LO16Opc = 0; 1029 unsigned HI16Opc = 0; 1030 unsigned MIFlags = MI.getFlags(); 1031 if (Opcode == ARM::t2MOVi32imm || Opcode == ARM::t2MOVCCi32imm) { 1032 LO16Opc = ARM::t2MOVi16; 1033 HI16Opc = ARM::t2MOVTi16; 1034 } else { 1035 LO16Opc = ARM::MOVi16; 1036 HI16Opc = ARM::MOVTi16; 1037 } 1038 1039 LO16 = BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(LO16Opc), DstReg); 1040 HI16 = BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(HI16Opc)) 1041 .addReg(DstReg, RegState::Define | getDeadRegState(DstIsDead)) 1042 .addReg(DstReg); 1043 1044 LO16.setMIFlags(MIFlags); 1045 HI16.setMIFlags(MIFlags); 1046 1047 switch (MO.getType()) { 1048 case MachineOperand::MO_Immediate: { 1049 unsigned Imm = MO.getImm(); 1050 unsigned Lo16 = Imm & 0xffff; 1051 unsigned Hi16 = (Imm >> 16) & 0xffff; 1052 LO16 = LO16.addImm(Lo16); 1053 HI16 = HI16.addImm(Hi16); 1054 break; 1055 } 1056 case MachineOperand::MO_ExternalSymbol: { 1057 const char *ES = MO.getSymbolName(); 1058 unsigned TF = MO.getTargetFlags(); 1059 LO16 = LO16.addExternalSymbol(ES, TF | ARMII::MO_LO16); 1060 HI16 = HI16.addExternalSymbol(ES, TF | ARMII::MO_HI16); 1061 break; 1062 } 1063 default: { 1064 const GlobalValue *GV = MO.getGlobal(); 1065 unsigned TF = MO.getTargetFlags(); 1066 LO16 = LO16.addGlobalAddress(GV, MO.getOffset(), TF | ARMII::MO_LO16); 1067 HI16 = HI16.addGlobalAddress(GV, MO.getOffset(), TF | ARMII::MO_HI16); 1068 break; 1069 } 1070 } 1071 1072 LO16.cloneMemRefs(MI); 1073 HI16.cloneMemRefs(MI); 1074 LO16.addImm(Pred).addReg(PredReg); 1075 HI16.addImm(Pred).addReg(PredReg); 1076 1077 if (RequiresBundling) 1078 finalizeBundle(MBB, LO16->getIterator(), MBBI->getIterator()); 1079 1080 if (isCC) 1081 LO16.add(makeImplicit(MI.getOperand(1))); 1082 TransferImpOps(MI, LO16, HI16); 1083 MI.eraseFromParent(); 1084 LLVM_DEBUG(dbgs() << "To: "; LO16.getInstr()->dump();); 1085 LLVM_DEBUG(dbgs() << "And: "; HI16.getInstr()->dump();); 1086 } 1087 1088 // The size of the area, accessed by that VLSTM/VLLDM 1089 // S0-S31 + FPSCR + 8 more bytes (VPR + pad, or just pad) 1090 static const int CMSE_FP_SAVE_SIZE = 136; 1091 1092 static void determineGPRegsToClear(const MachineInstr &MI, 1093 const std::initializer_list<unsigned> &Regs, 1094 SmallVectorImpl<unsigned> &ClearRegs) { 1095 SmallVector<unsigned, 4> OpRegs; 1096 for (const MachineOperand &Op : MI.operands()) { 1097 if (!Op.isReg() || !Op.isUse()) 1098 continue; 1099 OpRegs.push_back(Op.getReg()); 1100 } 1101 llvm::sort(OpRegs); 1102 1103 std::set_difference(Regs.begin(), Regs.end(), OpRegs.begin(), OpRegs.end(), 1104 std::back_inserter(ClearRegs)); 1105 } 1106 1107 void ARMExpandPseudo::CMSEClearGPRegs( 1108 MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, 1109 const DebugLoc &DL, const SmallVectorImpl<unsigned> &ClearRegs, 1110 unsigned ClobberReg) { 1111 1112 if (STI->hasV8_1MMainlineOps()) { 1113 // Clear the registers using the CLRM instruction. 1114 MachineInstrBuilder CLRM = 1115 BuildMI(MBB, MBBI, DL, TII->get(ARM::t2CLRM)).add(predOps(ARMCC::AL)); 1116 for (unsigned R : ClearRegs) 1117 CLRM.addReg(R, RegState::Define); 1118 CLRM.addReg(ARM::APSR, RegState::Define); 1119 CLRM.addReg(ARM::CPSR, RegState::Define | RegState::Implicit); 1120 } else { 1121 // Clear the registers and flags by copying ClobberReg into them. 1122 // (Baseline can't do a high register clear in one instruction). 1123 for (unsigned Reg : ClearRegs) { 1124 if (Reg == ClobberReg) 1125 continue; 1126 BuildMI(MBB, MBBI, DL, TII->get(ARM::tMOVr), Reg) 1127 .addReg(ClobberReg) 1128 .add(predOps(ARMCC::AL)); 1129 } 1130 1131 BuildMI(MBB, MBBI, DL, TII->get(ARM::t2MSR_M)) 1132 .addImm(STI->hasDSP() ? 0xc00 : 0x800) 1133 .addReg(ClobberReg) 1134 .add(predOps(ARMCC::AL)); 1135 } 1136 } 1137 1138 // Find which FP registers need to be cleared. The parameter `ClearRegs` is 1139 // initialised with all elements set to true, and this function resets all the 1140 // bits, which correspond to register uses. Returns true if any floating point 1141 // register is defined, false otherwise. 1142 static bool determineFPRegsToClear(const MachineInstr &MI, 1143 BitVector &ClearRegs) { 1144 bool DefFP = false; 1145 for (const MachineOperand &Op : MI.operands()) { 1146 if (!Op.isReg()) 1147 continue; 1148 1149 Register Reg = Op.getReg(); 1150 if (Op.isDef()) { 1151 if ((Reg >= ARM::Q0 && Reg <= ARM::Q7) || 1152 (Reg >= ARM::D0 && Reg <= ARM::D15) || 1153 (Reg >= ARM::S0 && Reg <= ARM::S31)) 1154 DefFP = true; 1155 continue; 1156 } 1157 1158 if (Reg >= ARM::Q0 && Reg <= ARM::Q7) { 1159 int R = Reg - ARM::Q0; 1160 ClearRegs.reset(R * 4, (R + 1) * 4); 1161 } else if (Reg >= ARM::D0 && Reg <= ARM::D15) { 1162 int R = Reg - ARM::D0; 1163 ClearRegs.reset(R * 2, (R + 1) * 2); 1164 } else if (Reg >= ARM::S0 && Reg <= ARM::S31) { 1165 ClearRegs[Reg - ARM::S0] = false; 1166 } 1167 } 1168 return DefFP; 1169 } 1170 1171 MachineBasicBlock & 1172 ARMExpandPseudo::CMSEClearFPRegs(MachineBasicBlock &MBB, 1173 MachineBasicBlock::iterator MBBI) { 1174 BitVector ClearRegs(16, true); 1175 (void)determineFPRegsToClear(*MBBI, ClearRegs); 1176 1177 if (STI->hasV8_1MMainlineOps()) 1178 return CMSEClearFPRegsV81(MBB, MBBI, ClearRegs); 1179 else 1180 return CMSEClearFPRegsV8(MBB, MBBI, ClearRegs); 1181 } 1182 1183 // Clear the FP registers for v8.0-M, by copying over the content 1184 // of LR. Uses R12 as a scratch register. 1185 MachineBasicBlock & 1186 ARMExpandPseudo::CMSEClearFPRegsV8(MachineBasicBlock &MBB, 1187 MachineBasicBlock::iterator MBBI, 1188 const BitVector &ClearRegs) { 1189 if (!STI->hasFPRegs()) 1190 return MBB; 1191 1192 auto &RetI = *MBBI; 1193 const DebugLoc &DL = RetI.getDebugLoc(); 1194 1195 // If optimising for minimum size, clear FP registers unconditionally. 1196 // Otherwise, check the CONTROL.SFPA (Secure Floating-Point Active) bit and 1197 // don't clear them if they belong to the non-secure state. 1198 MachineBasicBlock *ClearBB, *DoneBB; 1199 if (STI->hasMinSize()) { 1200 ClearBB = DoneBB = &MBB; 1201 } else { 1202 MachineFunction *MF = MBB.getParent(); 1203 ClearBB = MF->CreateMachineBasicBlock(MBB.getBasicBlock()); 1204 DoneBB = MF->CreateMachineBasicBlock(MBB.getBasicBlock()); 1205 1206 MF->insert(++MBB.getIterator(), ClearBB); 1207 MF->insert(++ClearBB->getIterator(), DoneBB); 1208 1209 DoneBB->splice(DoneBB->end(), &MBB, MBBI, MBB.end()); 1210 DoneBB->transferSuccessors(&MBB); 1211 MBB.addSuccessor(ClearBB); 1212 MBB.addSuccessor(DoneBB); 1213 ClearBB->addSuccessor(DoneBB); 1214 1215 // At the new basic blocks we need to have live-in the registers, used 1216 // for the return value as well as LR, used to clear registers. 1217 for (const MachineOperand &Op : RetI.operands()) { 1218 if (!Op.isReg()) 1219 continue; 1220 Register Reg = Op.getReg(); 1221 if (Reg == ARM::NoRegister || Reg == ARM::LR) 1222 continue; 1223 assert(Reg.isPhysical() && "Unallocated register"); 1224 ClearBB->addLiveIn(Reg); 1225 DoneBB->addLiveIn(Reg); 1226 } 1227 ClearBB->addLiveIn(ARM::LR); 1228 DoneBB->addLiveIn(ARM::LR); 1229 1230 // Read the CONTROL register. 1231 BuildMI(MBB, MBB.end(), DL, TII->get(ARM::t2MRS_M), ARM::R12) 1232 .addImm(20) 1233 .add(predOps(ARMCC::AL)); 1234 // Check bit 3 (SFPA). 1235 BuildMI(MBB, MBB.end(), DL, TII->get(ARM::t2TSTri)) 1236 .addReg(ARM::R12) 1237 .addImm(8) 1238 .add(predOps(ARMCC::AL)); 1239 // If SFPA is clear, jump over ClearBB to DoneBB. 1240 BuildMI(MBB, MBB.end(), DL, TII->get(ARM::tBcc)) 1241 .addMBB(DoneBB) 1242 .addImm(ARMCC::EQ) 1243 .addReg(ARM::CPSR, RegState::Kill); 1244 } 1245 1246 // Emit the clearing sequence 1247 for (unsigned D = 0; D < 8; D++) { 1248 // Attempt to clear as double 1249 if (ClearRegs[D * 2 + 0] && ClearRegs[D * 2 + 1]) { 1250 unsigned Reg = ARM::D0 + D; 1251 BuildMI(ClearBB, DL, TII->get(ARM::VMOVDRR), Reg) 1252 .addReg(ARM::LR) 1253 .addReg(ARM::LR) 1254 .add(predOps(ARMCC::AL)); 1255 } else { 1256 // Clear first part as single 1257 if (ClearRegs[D * 2 + 0]) { 1258 unsigned Reg = ARM::S0 + D * 2; 1259 BuildMI(ClearBB, DL, TII->get(ARM::VMOVSR), Reg) 1260 .addReg(ARM::LR) 1261 .add(predOps(ARMCC::AL)); 1262 } 1263 // Clear second part as single 1264 if (ClearRegs[D * 2 + 1]) { 1265 unsigned Reg = ARM::S0 + D * 2 + 1; 1266 BuildMI(ClearBB, DL, TII->get(ARM::VMOVSR), Reg) 1267 .addReg(ARM::LR) 1268 .add(predOps(ARMCC::AL)); 1269 } 1270 } 1271 } 1272 1273 // Clear FPSCR bits 0-4, 7, 28-31 1274 // The other bits are program global according to the AAPCS 1275 BuildMI(ClearBB, DL, TII->get(ARM::VMRS), ARM::R12) 1276 .add(predOps(ARMCC::AL)); 1277 BuildMI(ClearBB, DL, TII->get(ARM::t2BICri), ARM::R12) 1278 .addReg(ARM::R12) 1279 .addImm(0x0000009F) 1280 .add(predOps(ARMCC::AL)) 1281 .add(condCodeOp()); 1282 BuildMI(ClearBB, DL, TII->get(ARM::t2BICri), ARM::R12) 1283 .addReg(ARM::R12) 1284 .addImm(0xF0000000) 1285 .add(predOps(ARMCC::AL)) 1286 .add(condCodeOp()); 1287 BuildMI(ClearBB, DL, TII->get(ARM::VMSR)) 1288 .addReg(ARM::R12) 1289 .add(predOps(ARMCC::AL)); 1290 1291 return *DoneBB; 1292 } 1293 1294 MachineBasicBlock & 1295 ARMExpandPseudo::CMSEClearFPRegsV81(MachineBasicBlock &MBB, 1296 MachineBasicBlock::iterator MBBI, 1297 const BitVector &ClearRegs) { 1298 auto &RetI = *MBBI; 1299 1300 // Emit a sequence of VSCCLRM <sreglist> instructions, one instruction for 1301 // each contiguous sequence of S-registers. 1302 int Start = -1, End = -1; 1303 for (int S = 0, E = ClearRegs.size(); S != E; ++S) { 1304 if (ClearRegs[S] && S == End + 1) { 1305 End = S; // extend range 1306 continue; 1307 } 1308 // Emit current range. 1309 if (Start < End) { 1310 MachineInstrBuilder VSCCLRM = 1311 BuildMI(MBB, MBBI, RetI.getDebugLoc(), TII->get(ARM::VSCCLRMS)) 1312 .add(predOps(ARMCC::AL)); 1313 while (++Start <= End) 1314 VSCCLRM.addReg(ARM::S0 + Start, RegState::Define); 1315 VSCCLRM.addReg(ARM::VPR, RegState::Define); 1316 } 1317 Start = End = S; 1318 } 1319 // Emit last range. 1320 if (Start < End) { 1321 MachineInstrBuilder VSCCLRM = 1322 BuildMI(MBB, MBBI, RetI.getDebugLoc(), TII->get(ARM::VSCCLRMS)) 1323 .add(predOps(ARMCC::AL)); 1324 while (++Start <= End) 1325 VSCCLRM.addReg(ARM::S0 + Start, RegState::Define); 1326 VSCCLRM.addReg(ARM::VPR, RegState::Define); 1327 } 1328 1329 return MBB; 1330 } 1331 1332 void ARMExpandPseudo::CMSESaveClearFPRegs( 1333 MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, DebugLoc &DL, 1334 const LivePhysRegs &LiveRegs, SmallVectorImpl<unsigned> &ScratchRegs) { 1335 if (STI->hasV8_1MMainlineOps()) 1336 CMSESaveClearFPRegsV81(MBB, MBBI, DL, LiveRegs); 1337 else if (STI->hasV8MMainlineOps()) 1338 CMSESaveClearFPRegsV8(MBB, MBBI, DL, LiveRegs, ScratchRegs); 1339 } 1340 1341 // Save and clear FP registers if present 1342 void ARMExpandPseudo::CMSESaveClearFPRegsV8( 1343 MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, DebugLoc &DL, 1344 const LivePhysRegs &LiveRegs, SmallVectorImpl<unsigned> &ScratchRegs) { 1345 1346 // Store an available register for FPSCR clearing 1347 assert(!ScratchRegs.empty()); 1348 unsigned SpareReg = ScratchRegs.front(); 1349 1350 // save space on stack for VLSTM 1351 BuildMI(MBB, MBBI, DL, TII->get(ARM::tSUBspi), ARM::SP) 1352 .addReg(ARM::SP) 1353 .addImm(CMSE_FP_SAVE_SIZE >> 2) 1354 .add(predOps(ARMCC::AL)); 1355 1356 // Use ScratchRegs to store the fp regs 1357 std::vector<std::tuple<unsigned, unsigned, unsigned>> ClearedFPRegs; 1358 std::vector<unsigned> NonclearedFPRegs; 1359 for (const MachineOperand &Op : MBBI->operands()) { 1360 if (Op.isReg() && Op.isUse()) { 1361 Register Reg = Op.getReg(); 1362 assert(!ARM::DPRRegClass.contains(Reg) || 1363 ARM::DPR_VFP2RegClass.contains(Reg)); 1364 assert(!ARM::QPRRegClass.contains(Reg)); 1365 if (ARM::DPR_VFP2RegClass.contains(Reg)) { 1366 if (ScratchRegs.size() >= 2) { 1367 unsigned SaveReg2 = ScratchRegs.pop_back_val(); 1368 unsigned SaveReg1 = ScratchRegs.pop_back_val(); 1369 ClearedFPRegs.emplace_back(Reg, SaveReg1, SaveReg2); 1370 1371 // Save the fp register to the normal registers 1372 BuildMI(MBB, MBBI, DL, TII->get(ARM::VMOVRRD)) 1373 .addReg(SaveReg1, RegState::Define) 1374 .addReg(SaveReg2, RegState::Define) 1375 .addReg(Reg) 1376 .add(predOps(ARMCC::AL)); 1377 } else { 1378 NonclearedFPRegs.push_back(Reg); 1379 } 1380 } else if (ARM::SPRRegClass.contains(Reg)) { 1381 if (ScratchRegs.size() >= 1) { 1382 unsigned SaveReg = ScratchRegs.pop_back_val(); 1383 ClearedFPRegs.emplace_back(Reg, SaveReg, 0); 1384 1385 // Save the fp register to the normal registers 1386 BuildMI(MBB, MBBI, DL, TII->get(ARM::VMOVRS), SaveReg) 1387 .addReg(Reg) 1388 .add(predOps(ARMCC::AL)); 1389 } else { 1390 NonclearedFPRegs.push_back(Reg); 1391 } 1392 } 1393 } 1394 } 1395 1396 bool passesFPReg = (!NonclearedFPRegs.empty() || !ClearedFPRegs.empty()); 1397 1398 if (passesFPReg) 1399 assert(STI->hasFPRegs() && "Subtarget needs fpregs"); 1400 1401 // Lazy store all fp registers to the stack. 1402 // This executes as NOP in the absence of floating-point support. 1403 MachineInstrBuilder VLSTM = BuildMI(MBB, MBBI, DL, TII->get(ARM::VLSTM)) 1404 .addReg(ARM::SP) 1405 .add(predOps(ARMCC::AL)); 1406 for (auto R : {ARM::VPR, ARM::FPSCR, ARM::FPSCR_NZCV, ARM::Q0, ARM::Q1, 1407 ARM::Q2, ARM::Q3, ARM::Q4, ARM::Q5, ARM::Q6, ARM::Q7}) 1408 VLSTM.addReg(R, RegState::Implicit | 1409 (LiveRegs.contains(R) ? 0 : RegState::Undef)); 1410 1411 // Restore all arguments 1412 for (const auto &Regs : ClearedFPRegs) { 1413 unsigned Reg, SaveReg1, SaveReg2; 1414 std::tie(Reg, SaveReg1, SaveReg2) = Regs; 1415 if (ARM::DPR_VFP2RegClass.contains(Reg)) 1416 BuildMI(MBB, MBBI, DL, TII->get(ARM::VMOVDRR), Reg) 1417 .addReg(SaveReg1) 1418 .addReg(SaveReg2) 1419 .add(predOps(ARMCC::AL)); 1420 else if (ARM::SPRRegClass.contains(Reg)) 1421 BuildMI(MBB, MBBI, DL, TII->get(ARM::VMOVSR), Reg) 1422 .addReg(SaveReg1) 1423 .add(predOps(ARMCC::AL)); 1424 } 1425 1426 for (unsigned Reg : NonclearedFPRegs) { 1427 if (ARM::DPR_VFP2RegClass.contains(Reg)) { 1428 if (STI->isLittle()) { 1429 BuildMI(MBB, MBBI, DL, TII->get(ARM::VLDRD), Reg) 1430 .addReg(ARM::SP) 1431 .addImm((Reg - ARM::D0) * 2) 1432 .add(predOps(ARMCC::AL)); 1433 } else { 1434 // For big-endian targets we need to load the two subregisters of Reg 1435 // manually because VLDRD would load them in wrong order 1436 unsigned SReg0 = TRI->getSubReg(Reg, ARM::ssub_0); 1437 BuildMI(MBB, MBBI, DL, TII->get(ARM::VLDRS), SReg0) 1438 .addReg(ARM::SP) 1439 .addImm((Reg - ARM::D0) * 2) 1440 .add(predOps(ARMCC::AL)); 1441 BuildMI(MBB, MBBI, DL, TII->get(ARM::VLDRS), SReg0 + 1) 1442 .addReg(ARM::SP) 1443 .addImm((Reg - ARM::D0) * 2 + 1) 1444 .add(predOps(ARMCC::AL)); 1445 } 1446 } else if (ARM::SPRRegClass.contains(Reg)) { 1447 BuildMI(MBB, MBBI, DL, TII->get(ARM::VLDRS), Reg) 1448 .addReg(ARM::SP) 1449 .addImm(Reg - ARM::S0) 1450 .add(predOps(ARMCC::AL)); 1451 } 1452 } 1453 // restore FPSCR from stack and clear bits 0-4, 7, 28-31 1454 // The other bits are program global according to the AAPCS 1455 if (passesFPReg) { 1456 BuildMI(MBB, MBBI, DL, TII->get(ARM::tLDRspi), SpareReg) 1457 .addReg(ARM::SP) 1458 .addImm(0x10) 1459 .add(predOps(ARMCC::AL)); 1460 BuildMI(MBB, MBBI, DL, TII->get(ARM::t2BICri), SpareReg) 1461 .addReg(SpareReg) 1462 .addImm(0x0000009F) 1463 .add(predOps(ARMCC::AL)) 1464 .add(condCodeOp()); 1465 BuildMI(MBB, MBBI, DL, TII->get(ARM::t2BICri), SpareReg) 1466 .addReg(SpareReg) 1467 .addImm(0xF0000000) 1468 .add(predOps(ARMCC::AL)) 1469 .add(condCodeOp()); 1470 BuildMI(MBB, MBBI, DL, TII->get(ARM::VMSR)) 1471 .addReg(SpareReg) 1472 .add(predOps(ARMCC::AL)); 1473 // The ldr must happen after a floating point instruction. To prevent the 1474 // post-ra scheduler to mess with the order, we create a bundle. 1475 finalizeBundle(MBB, VLSTM->getIterator(), MBBI->getIterator()); 1476 } 1477 } 1478 1479 void ARMExpandPseudo::CMSESaveClearFPRegsV81(MachineBasicBlock &MBB, 1480 MachineBasicBlock::iterator MBBI, 1481 DebugLoc &DL, 1482 const LivePhysRegs &LiveRegs) { 1483 BitVector ClearRegs(32, true); 1484 bool DefFP = determineFPRegsToClear(*MBBI, ClearRegs); 1485 1486 // If the instruction does not write to a FP register and no elements were 1487 // removed from the set, then no FP registers were used to pass 1488 // arguments/returns. 1489 if (!DefFP && ClearRegs.count() == ClearRegs.size()) { 1490 // save space on stack for VLSTM 1491 BuildMI(MBB, MBBI, DL, TII->get(ARM::tSUBspi), ARM::SP) 1492 .addReg(ARM::SP) 1493 .addImm(CMSE_FP_SAVE_SIZE >> 2) 1494 .add(predOps(ARMCC::AL)); 1495 1496 // Lazy store all FP registers to the stack 1497 MachineInstrBuilder VLSTM = BuildMI(MBB, MBBI, DL, TII->get(ARM::VLSTM)) 1498 .addReg(ARM::SP) 1499 .add(predOps(ARMCC::AL)); 1500 for (auto R : {ARM::VPR, ARM::FPSCR, ARM::FPSCR_NZCV, ARM::Q0, ARM::Q1, 1501 ARM::Q2, ARM::Q3, ARM::Q4, ARM::Q5, ARM::Q6, ARM::Q7}) 1502 VLSTM.addReg(R, RegState::Implicit | 1503 (LiveRegs.contains(R) ? 0 : RegState::Undef)); 1504 } else { 1505 // Push all the callee-saved registers (s16-s31). 1506 MachineInstrBuilder VPUSH = 1507 BuildMI(MBB, MBBI, DL, TII->get(ARM::VSTMSDB_UPD), ARM::SP) 1508 .addReg(ARM::SP) 1509 .add(predOps(ARMCC::AL)); 1510 for (int Reg = ARM::S16; Reg <= ARM::S31; ++Reg) 1511 VPUSH.addReg(Reg); 1512 1513 // Clear FP registers with a VSCCLRM. 1514 (void)CMSEClearFPRegsV81(MBB, MBBI, ClearRegs); 1515 1516 // Save floating-point context. 1517 BuildMI(MBB, MBBI, DL, TII->get(ARM::VSTR_FPCXTS_pre), ARM::SP) 1518 .addReg(ARM::SP) 1519 .addImm(-8) 1520 .add(predOps(ARMCC::AL)); 1521 } 1522 } 1523 1524 // Restore FP registers if present 1525 void ARMExpandPseudo::CMSERestoreFPRegs( 1526 MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, DebugLoc &DL, 1527 SmallVectorImpl<unsigned> &AvailableRegs) { 1528 if (STI->hasV8_1MMainlineOps()) 1529 CMSERestoreFPRegsV81(MBB, MBBI, DL, AvailableRegs); 1530 else if (STI->hasV8MMainlineOps()) 1531 CMSERestoreFPRegsV8(MBB, MBBI, DL, AvailableRegs); 1532 } 1533 1534 void ARMExpandPseudo::CMSERestoreFPRegsV8( 1535 MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, DebugLoc &DL, 1536 SmallVectorImpl<unsigned> &AvailableRegs) { 1537 1538 // Keep a scratch register for the mitigation sequence. 1539 unsigned ScratchReg = ARM::NoRegister; 1540 if (STI->fixCMSE_CVE_2021_35465()) 1541 ScratchReg = AvailableRegs.pop_back_val(); 1542 1543 // Use AvailableRegs to store the fp regs 1544 std::vector<std::tuple<unsigned, unsigned, unsigned>> ClearedFPRegs; 1545 std::vector<unsigned> NonclearedFPRegs; 1546 for (const MachineOperand &Op : MBBI->operands()) { 1547 if (Op.isReg() && Op.isDef()) { 1548 Register Reg = Op.getReg(); 1549 assert(!ARM::DPRRegClass.contains(Reg) || 1550 ARM::DPR_VFP2RegClass.contains(Reg)); 1551 assert(!ARM::QPRRegClass.contains(Reg)); 1552 if (ARM::DPR_VFP2RegClass.contains(Reg)) { 1553 if (AvailableRegs.size() >= 2) { 1554 unsigned SaveReg2 = AvailableRegs.pop_back_val(); 1555 unsigned SaveReg1 = AvailableRegs.pop_back_val(); 1556 ClearedFPRegs.emplace_back(Reg, SaveReg1, SaveReg2); 1557 1558 // Save the fp register to the normal registers 1559 BuildMI(MBB, MBBI, DL, TII->get(ARM::VMOVRRD)) 1560 .addReg(SaveReg1, RegState::Define) 1561 .addReg(SaveReg2, RegState::Define) 1562 .addReg(Reg) 1563 .add(predOps(ARMCC::AL)); 1564 } else { 1565 NonclearedFPRegs.push_back(Reg); 1566 } 1567 } else if (ARM::SPRRegClass.contains(Reg)) { 1568 if (AvailableRegs.size() >= 1) { 1569 unsigned SaveReg = AvailableRegs.pop_back_val(); 1570 ClearedFPRegs.emplace_back(Reg, SaveReg, 0); 1571 1572 // Save the fp register to the normal registers 1573 BuildMI(MBB, MBBI, DL, TII->get(ARM::VMOVRS), SaveReg) 1574 .addReg(Reg) 1575 .add(predOps(ARMCC::AL)); 1576 } else { 1577 NonclearedFPRegs.push_back(Reg); 1578 } 1579 } 1580 } 1581 } 1582 1583 bool returnsFPReg = (!NonclearedFPRegs.empty() || !ClearedFPRegs.empty()); 1584 1585 if (returnsFPReg) 1586 assert(STI->hasFPRegs() && "Subtarget needs fpregs"); 1587 1588 // Push FP regs that cannot be restored via normal registers on the stack 1589 for (unsigned Reg : NonclearedFPRegs) { 1590 if (ARM::DPR_VFP2RegClass.contains(Reg)) 1591 BuildMI(MBB, MBBI, DL, TII->get(ARM::VSTRD)) 1592 .addReg(Reg) 1593 .addReg(ARM::SP) 1594 .addImm((Reg - ARM::D0) * 2) 1595 .add(predOps(ARMCC::AL)); 1596 else if (ARM::SPRRegClass.contains(Reg)) 1597 BuildMI(MBB, MBBI, DL, TII->get(ARM::VSTRS)) 1598 .addReg(Reg) 1599 .addReg(ARM::SP) 1600 .addImm(Reg - ARM::S0) 1601 .add(predOps(ARMCC::AL)); 1602 } 1603 1604 // Lazy load fp regs from stack. 1605 // This executes as NOP in the absence of floating-point support. 1606 MachineInstrBuilder VLLDM = BuildMI(MBB, MBBI, DL, TII->get(ARM::VLLDM)) 1607 .addReg(ARM::SP) 1608 .add(predOps(ARMCC::AL)); 1609 1610 if (STI->fixCMSE_CVE_2021_35465()) { 1611 auto Bundler = MIBundleBuilder(MBB, VLLDM); 1612 // Read the CONTROL register. 1613 Bundler.append(BuildMI(*MBB.getParent(), DL, TII->get(ARM::t2MRS_M)) 1614 .addReg(ScratchReg, RegState::Define) 1615 .addImm(20) 1616 .add(predOps(ARMCC::AL))); 1617 // Check bit 3 (SFPA). 1618 Bundler.append(BuildMI(*MBB.getParent(), DL, TII->get(ARM::t2TSTri)) 1619 .addReg(ScratchReg) 1620 .addImm(8) 1621 .add(predOps(ARMCC::AL))); 1622 // Emit the IT block. 1623 Bundler.append(BuildMI(*MBB.getParent(), DL, TII->get(ARM::t2IT)) 1624 .addImm(ARMCC::NE) 1625 .addImm(8)); 1626 // If SFPA is clear jump over to VLLDM, otherwise execute an instruction 1627 // which has no functional effect apart from causing context creation: 1628 // vmovne s0, s0. In the absence of FPU we emit .inst.w 0xeeb00a40, 1629 // which is defined as NOP if not executed. 1630 if (STI->hasFPRegs()) 1631 Bundler.append(BuildMI(*MBB.getParent(), DL, TII->get(ARM::VMOVS)) 1632 .addReg(ARM::S0, RegState::Define) 1633 .addReg(ARM::S0, RegState::Undef) 1634 .add(predOps(ARMCC::NE))); 1635 else 1636 Bundler.append(BuildMI(*MBB.getParent(), DL, TII->get(ARM::INLINEASM)) 1637 .addExternalSymbol(".inst.w 0xeeb00a40") 1638 .addImm(InlineAsm::Extra_HasSideEffects)); 1639 finalizeBundle(MBB, Bundler.begin(), Bundler.end()); 1640 } 1641 1642 // Restore all FP registers via normal registers 1643 for (const auto &Regs : ClearedFPRegs) { 1644 unsigned Reg, SaveReg1, SaveReg2; 1645 std::tie(Reg, SaveReg1, SaveReg2) = Regs; 1646 if (ARM::DPR_VFP2RegClass.contains(Reg)) 1647 BuildMI(MBB, MBBI, DL, TII->get(ARM::VMOVDRR), Reg) 1648 .addReg(SaveReg1) 1649 .addReg(SaveReg2) 1650 .add(predOps(ARMCC::AL)); 1651 else if (ARM::SPRRegClass.contains(Reg)) 1652 BuildMI(MBB, MBBI, DL, TII->get(ARM::VMOVSR), Reg) 1653 .addReg(SaveReg1) 1654 .add(predOps(ARMCC::AL)); 1655 } 1656 1657 // Pop the stack space 1658 BuildMI(MBB, MBBI, DL, TII->get(ARM::tADDspi), ARM::SP) 1659 .addReg(ARM::SP) 1660 .addImm(CMSE_FP_SAVE_SIZE >> 2) 1661 .add(predOps(ARMCC::AL)); 1662 } 1663 1664 static bool definesOrUsesFPReg(const MachineInstr &MI) { 1665 for (const MachineOperand &Op : MI.operands()) { 1666 if (!Op.isReg()) 1667 continue; 1668 Register Reg = Op.getReg(); 1669 if ((Reg >= ARM::Q0 && Reg <= ARM::Q7) || 1670 (Reg >= ARM::D0 && Reg <= ARM::D15) || 1671 (Reg >= ARM::S0 && Reg <= ARM::S31)) 1672 return true; 1673 } 1674 return false; 1675 } 1676 1677 void ARMExpandPseudo::CMSERestoreFPRegsV81( 1678 MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, DebugLoc &DL, 1679 SmallVectorImpl<unsigned> &AvailableRegs) { 1680 if (!definesOrUsesFPReg(*MBBI)) { 1681 if (STI->fixCMSE_CVE_2021_35465()) { 1682 BuildMI(MBB, MBBI, DL, TII->get(ARM::VSCCLRMS)) 1683 .add(predOps(ARMCC::AL)) 1684 .addReg(ARM::VPR, RegState::Define); 1685 } 1686 1687 // Load FP registers from stack. 1688 BuildMI(MBB, MBBI, DL, TII->get(ARM::VLLDM)) 1689 .addReg(ARM::SP) 1690 .add(predOps(ARMCC::AL)); 1691 1692 // Pop the stack space 1693 BuildMI(MBB, MBBI, DL, TII->get(ARM::tADDspi), ARM::SP) 1694 .addReg(ARM::SP) 1695 .addImm(CMSE_FP_SAVE_SIZE >> 2) 1696 .add(predOps(ARMCC::AL)); 1697 } else { 1698 // Restore the floating point context. 1699 BuildMI(MBB, MBBI, MBBI->getDebugLoc(), TII->get(ARM::VLDR_FPCXTS_post), 1700 ARM::SP) 1701 .addReg(ARM::SP) 1702 .addImm(8) 1703 .add(predOps(ARMCC::AL)); 1704 1705 // Pop all the callee-saved registers (s16-s31). 1706 MachineInstrBuilder VPOP = 1707 BuildMI(MBB, MBBI, DL, TII->get(ARM::VLDMSIA_UPD), ARM::SP) 1708 .addReg(ARM::SP) 1709 .add(predOps(ARMCC::AL)); 1710 for (int Reg = ARM::S16; Reg <= ARM::S31; ++Reg) 1711 VPOP.addReg(Reg, RegState::Define); 1712 } 1713 } 1714 1715 /// Expand a CMP_SWAP pseudo-inst to an ldrex/strex loop as simply as 1716 /// possible. This only gets used at -O0 so we don't care about efficiency of 1717 /// the generated code. 1718 bool ARMExpandPseudo::ExpandCMP_SWAP(MachineBasicBlock &MBB, 1719 MachineBasicBlock::iterator MBBI, 1720 unsigned LdrexOp, unsigned StrexOp, 1721 unsigned UxtOp, 1722 MachineBasicBlock::iterator &NextMBBI) { 1723 bool IsThumb = STI->isThumb(); 1724 bool IsThumb1Only = STI->isThumb1Only(); 1725 MachineInstr &MI = *MBBI; 1726 DebugLoc DL = MI.getDebugLoc(); 1727 const MachineOperand &Dest = MI.getOperand(0); 1728 Register TempReg = MI.getOperand(1).getReg(); 1729 // Duplicating undef operands into 2 instructions does not guarantee the same 1730 // value on both; However undef should be replaced by xzr anyway. 1731 assert(!MI.getOperand(2).isUndef() && "cannot handle undef"); 1732 Register AddrReg = MI.getOperand(2).getReg(); 1733 Register DesiredReg = MI.getOperand(3).getReg(); 1734 Register NewReg = MI.getOperand(4).getReg(); 1735 1736 if (IsThumb) { 1737 assert(STI->hasV8MBaselineOps() && 1738 "CMP_SWAP not expected to be custom expanded for Thumb1"); 1739 assert((UxtOp == 0 || UxtOp == ARM::tUXTB || UxtOp == ARM::tUXTH) && 1740 "ARMv8-M.baseline does not have t2UXTB/t2UXTH"); 1741 assert((UxtOp == 0 || ARM::tGPRRegClass.contains(DesiredReg)) && 1742 "DesiredReg used for UXT op must be tGPR"); 1743 } 1744 1745 MachineFunction *MF = MBB.getParent(); 1746 auto LoadCmpBB = MF->CreateMachineBasicBlock(MBB.getBasicBlock()); 1747 auto StoreBB = MF->CreateMachineBasicBlock(MBB.getBasicBlock()); 1748 auto DoneBB = MF->CreateMachineBasicBlock(MBB.getBasicBlock()); 1749 1750 MF->insert(++MBB.getIterator(), LoadCmpBB); 1751 MF->insert(++LoadCmpBB->getIterator(), StoreBB); 1752 MF->insert(++StoreBB->getIterator(), DoneBB); 1753 1754 if (UxtOp) { 1755 MachineInstrBuilder MIB = 1756 BuildMI(MBB, MBBI, DL, TII->get(UxtOp), DesiredReg) 1757 .addReg(DesiredReg, RegState::Kill); 1758 if (!IsThumb) 1759 MIB.addImm(0); 1760 MIB.add(predOps(ARMCC::AL)); 1761 } 1762 1763 // .Lloadcmp: 1764 // ldrex rDest, [rAddr] 1765 // cmp rDest, rDesired 1766 // bne .Ldone 1767 1768 MachineInstrBuilder MIB; 1769 MIB = BuildMI(LoadCmpBB, DL, TII->get(LdrexOp), Dest.getReg()); 1770 MIB.addReg(AddrReg); 1771 if (LdrexOp == ARM::t2LDREX) 1772 MIB.addImm(0); // a 32-bit Thumb ldrex (only) allows an offset. 1773 MIB.add(predOps(ARMCC::AL)); 1774 1775 unsigned CMPrr = IsThumb ? ARM::tCMPhir : ARM::CMPrr; 1776 BuildMI(LoadCmpBB, DL, TII->get(CMPrr)) 1777 .addReg(Dest.getReg(), getKillRegState(Dest.isDead())) 1778 .addReg(DesiredReg) 1779 .add(predOps(ARMCC::AL)); 1780 unsigned Bcc = IsThumb ? ARM::tBcc : ARM::Bcc; 1781 BuildMI(LoadCmpBB, DL, TII->get(Bcc)) 1782 .addMBB(DoneBB) 1783 .addImm(ARMCC::NE) 1784 .addReg(ARM::CPSR, RegState::Kill); 1785 LoadCmpBB->addSuccessor(DoneBB); 1786 LoadCmpBB->addSuccessor(StoreBB); 1787 1788 // .Lstore: 1789 // strex rTempReg, rNew, [rAddr] 1790 // cmp rTempReg, #0 1791 // bne .Lloadcmp 1792 MIB = BuildMI(StoreBB, DL, TII->get(StrexOp), TempReg) 1793 .addReg(NewReg) 1794 .addReg(AddrReg); 1795 if (StrexOp == ARM::t2STREX) 1796 MIB.addImm(0); // a 32-bit Thumb strex (only) allows an offset. 1797 MIB.add(predOps(ARMCC::AL)); 1798 1799 unsigned CMPri = 1800 IsThumb ? (IsThumb1Only ? ARM::tCMPi8 : ARM::t2CMPri) : ARM::CMPri; 1801 BuildMI(StoreBB, DL, TII->get(CMPri)) 1802 .addReg(TempReg, RegState::Kill) 1803 .addImm(0) 1804 .add(predOps(ARMCC::AL)); 1805 BuildMI(StoreBB, DL, TII->get(Bcc)) 1806 .addMBB(LoadCmpBB) 1807 .addImm(ARMCC::NE) 1808 .addReg(ARM::CPSR, RegState::Kill); 1809 StoreBB->addSuccessor(LoadCmpBB); 1810 StoreBB->addSuccessor(DoneBB); 1811 1812 DoneBB->splice(DoneBB->end(), &MBB, MI, MBB.end()); 1813 DoneBB->transferSuccessors(&MBB); 1814 1815 MBB.addSuccessor(LoadCmpBB); 1816 1817 NextMBBI = MBB.end(); 1818 MI.eraseFromParent(); 1819 1820 // Recompute livein lists. 1821 LivePhysRegs LiveRegs; 1822 computeAndAddLiveIns(LiveRegs, *DoneBB); 1823 computeAndAddLiveIns(LiveRegs, *StoreBB); 1824 computeAndAddLiveIns(LiveRegs, *LoadCmpBB); 1825 // Do an extra pass around the loop to get loop carried registers right. 1826 StoreBB->clearLiveIns(); 1827 computeAndAddLiveIns(LiveRegs, *StoreBB); 1828 LoadCmpBB->clearLiveIns(); 1829 computeAndAddLiveIns(LiveRegs, *LoadCmpBB); 1830 1831 return true; 1832 } 1833 1834 /// ARM's ldrexd/strexd take a consecutive register pair (represented as a 1835 /// single GPRPair register), Thumb's take two separate registers so we need to 1836 /// extract the subregs from the pair. 1837 static void addExclusiveRegPair(MachineInstrBuilder &MIB, MachineOperand &Reg, 1838 unsigned Flags, bool IsThumb, 1839 const TargetRegisterInfo *TRI) { 1840 if (IsThumb) { 1841 Register RegLo = TRI->getSubReg(Reg.getReg(), ARM::gsub_0); 1842 Register RegHi = TRI->getSubReg(Reg.getReg(), ARM::gsub_1); 1843 MIB.addReg(RegLo, Flags); 1844 MIB.addReg(RegHi, Flags); 1845 } else 1846 MIB.addReg(Reg.getReg(), Flags); 1847 } 1848 1849 /// Expand a 64-bit CMP_SWAP to an ldrexd/strexd loop. 1850 bool ARMExpandPseudo::ExpandCMP_SWAP_64(MachineBasicBlock &MBB, 1851 MachineBasicBlock::iterator MBBI, 1852 MachineBasicBlock::iterator &NextMBBI) { 1853 bool IsThumb = STI->isThumb(); 1854 assert(!STI->isThumb1Only() && "CMP_SWAP_64 unsupported under Thumb1!"); 1855 MachineInstr &MI = *MBBI; 1856 DebugLoc DL = MI.getDebugLoc(); 1857 MachineOperand &Dest = MI.getOperand(0); 1858 Register TempReg = MI.getOperand(1).getReg(); 1859 // Duplicating undef operands into 2 instructions does not guarantee the same 1860 // value on both; However undef should be replaced by xzr anyway. 1861 assert(!MI.getOperand(2).isUndef() && "cannot handle undef"); 1862 Register AddrReg = MI.getOperand(2).getReg(); 1863 Register DesiredReg = MI.getOperand(3).getReg(); 1864 MachineOperand New = MI.getOperand(4); 1865 New.setIsKill(false); 1866 1867 Register DestLo = TRI->getSubReg(Dest.getReg(), ARM::gsub_0); 1868 Register DestHi = TRI->getSubReg(Dest.getReg(), ARM::gsub_1); 1869 Register DesiredLo = TRI->getSubReg(DesiredReg, ARM::gsub_0); 1870 Register DesiredHi = TRI->getSubReg(DesiredReg, ARM::gsub_1); 1871 1872 MachineFunction *MF = MBB.getParent(); 1873 auto LoadCmpBB = MF->CreateMachineBasicBlock(MBB.getBasicBlock()); 1874 auto StoreBB = MF->CreateMachineBasicBlock(MBB.getBasicBlock()); 1875 auto DoneBB = MF->CreateMachineBasicBlock(MBB.getBasicBlock()); 1876 1877 MF->insert(++MBB.getIterator(), LoadCmpBB); 1878 MF->insert(++LoadCmpBB->getIterator(), StoreBB); 1879 MF->insert(++StoreBB->getIterator(), DoneBB); 1880 1881 // .Lloadcmp: 1882 // ldrexd rDestLo, rDestHi, [rAddr] 1883 // cmp rDestLo, rDesiredLo 1884 // sbcs dead rTempReg, rDestHi, rDesiredHi 1885 // bne .Ldone 1886 unsigned LDREXD = IsThumb ? ARM::t2LDREXD : ARM::LDREXD; 1887 MachineInstrBuilder MIB; 1888 MIB = BuildMI(LoadCmpBB, DL, TII->get(LDREXD)); 1889 addExclusiveRegPair(MIB, Dest, RegState::Define, IsThumb, TRI); 1890 MIB.addReg(AddrReg).add(predOps(ARMCC::AL)); 1891 1892 unsigned CMPrr = IsThumb ? ARM::tCMPhir : ARM::CMPrr; 1893 BuildMI(LoadCmpBB, DL, TII->get(CMPrr)) 1894 .addReg(DestLo, getKillRegState(Dest.isDead())) 1895 .addReg(DesiredLo) 1896 .add(predOps(ARMCC::AL)); 1897 1898 BuildMI(LoadCmpBB, DL, TII->get(CMPrr)) 1899 .addReg(DestHi, getKillRegState(Dest.isDead())) 1900 .addReg(DesiredHi) 1901 .addImm(ARMCC::EQ).addReg(ARM::CPSR, RegState::Kill); 1902 1903 unsigned Bcc = IsThumb ? ARM::tBcc : ARM::Bcc; 1904 BuildMI(LoadCmpBB, DL, TII->get(Bcc)) 1905 .addMBB(DoneBB) 1906 .addImm(ARMCC::NE) 1907 .addReg(ARM::CPSR, RegState::Kill); 1908 LoadCmpBB->addSuccessor(DoneBB); 1909 LoadCmpBB->addSuccessor(StoreBB); 1910 1911 // .Lstore: 1912 // strexd rTempReg, rNewLo, rNewHi, [rAddr] 1913 // cmp rTempReg, #0 1914 // bne .Lloadcmp 1915 unsigned STREXD = IsThumb ? ARM::t2STREXD : ARM::STREXD; 1916 MIB = BuildMI(StoreBB, DL, TII->get(STREXD), TempReg); 1917 unsigned Flags = getKillRegState(New.isDead()); 1918 addExclusiveRegPair(MIB, New, Flags, IsThumb, TRI); 1919 MIB.addReg(AddrReg).add(predOps(ARMCC::AL)); 1920 1921 unsigned CMPri = IsThumb ? ARM::t2CMPri : ARM::CMPri; 1922 BuildMI(StoreBB, DL, TII->get(CMPri)) 1923 .addReg(TempReg, RegState::Kill) 1924 .addImm(0) 1925 .add(predOps(ARMCC::AL)); 1926 BuildMI(StoreBB, DL, TII->get(Bcc)) 1927 .addMBB(LoadCmpBB) 1928 .addImm(ARMCC::NE) 1929 .addReg(ARM::CPSR, RegState::Kill); 1930 StoreBB->addSuccessor(LoadCmpBB); 1931 StoreBB->addSuccessor(DoneBB); 1932 1933 DoneBB->splice(DoneBB->end(), &MBB, MI, MBB.end()); 1934 DoneBB->transferSuccessors(&MBB); 1935 1936 MBB.addSuccessor(LoadCmpBB); 1937 1938 NextMBBI = MBB.end(); 1939 MI.eraseFromParent(); 1940 1941 // Recompute livein lists. 1942 LivePhysRegs LiveRegs; 1943 computeAndAddLiveIns(LiveRegs, *DoneBB); 1944 computeAndAddLiveIns(LiveRegs, *StoreBB); 1945 computeAndAddLiveIns(LiveRegs, *LoadCmpBB); 1946 // Do an extra pass around the loop to get loop carried registers right. 1947 StoreBB->clearLiveIns(); 1948 computeAndAddLiveIns(LiveRegs, *StoreBB); 1949 LoadCmpBB->clearLiveIns(); 1950 computeAndAddLiveIns(LiveRegs, *LoadCmpBB); 1951 1952 return true; 1953 } 1954 1955 static void CMSEPushCalleeSaves(const TargetInstrInfo &TII, 1956 MachineBasicBlock &MBB, 1957 MachineBasicBlock::iterator MBBI, int JumpReg, 1958 const LivePhysRegs &LiveRegs, bool Thumb1Only) { 1959 const DebugLoc &DL = MBBI->getDebugLoc(); 1960 if (Thumb1Only) { // push Lo and Hi regs separately 1961 MachineInstrBuilder PushMIB = 1962 BuildMI(MBB, MBBI, DL, TII.get(ARM::tPUSH)).add(predOps(ARMCC::AL)); 1963 for (int Reg = ARM::R4; Reg < ARM::R8; ++Reg) { 1964 PushMIB.addReg( 1965 Reg, Reg == JumpReg || LiveRegs.contains(Reg) ? 0 : RegState::Undef); 1966 } 1967 1968 // Thumb1 can only tPUSH low regs, so we copy the high regs to the low 1969 // regs that we just saved and push the low regs again, taking care to 1970 // not clobber JumpReg. If JumpReg is one of the low registers, push first 1971 // the values of r9-r11, and then r8. That would leave them ordered in 1972 // memory, and allow us to later pop them with a single instructions. 1973 // FIXME: Could also use any of r0-r3 that are free (including in the 1974 // first PUSH above). 1975 for (int LoReg = ARM::R7, HiReg = ARM::R11; LoReg >= ARM::R4; --LoReg) { 1976 if (JumpReg == LoReg) 1977 continue; 1978 BuildMI(MBB, MBBI, DL, TII.get(ARM::tMOVr), LoReg) 1979 .addReg(HiReg, LiveRegs.contains(HiReg) ? 0 : RegState::Undef) 1980 .add(predOps(ARMCC::AL)); 1981 --HiReg; 1982 } 1983 MachineInstrBuilder PushMIB2 = 1984 BuildMI(MBB, MBBI, DL, TII.get(ARM::tPUSH)).add(predOps(ARMCC::AL)); 1985 for (int Reg = ARM::R4; Reg < ARM::R8; ++Reg) { 1986 if (Reg == JumpReg) 1987 continue; 1988 PushMIB2.addReg(Reg, RegState::Kill); 1989 } 1990 1991 // If we couldn't use a low register for temporary storage (because it was 1992 // the JumpReg), use r4 or r5, whichever is not JumpReg. It has already been 1993 // saved. 1994 if (JumpReg >= ARM::R4 && JumpReg <= ARM::R7) { 1995 int LoReg = JumpReg == ARM::R4 ? ARM::R5 : ARM::R4; 1996 BuildMI(MBB, MBBI, DL, TII.get(ARM::tMOVr), LoReg) 1997 .addReg(ARM::R8, LiveRegs.contains(ARM::R8) ? 0 : RegState::Undef) 1998 .add(predOps(ARMCC::AL)); 1999 BuildMI(MBB, MBBI, DL, TII.get(ARM::tPUSH)) 2000 .add(predOps(ARMCC::AL)) 2001 .addReg(LoReg, RegState::Kill); 2002 } 2003 } else { // push Lo and Hi registers with a single instruction 2004 MachineInstrBuilder PushMIB = 2005 BuildMI(MBB, MBBI, DL, TII.get(ARM::t2STMDB_UPD), ARM::SP) 2006 .addReg(ARM::SP) 2007 .add(predOps(ARMCC::AL)); 2008 for (int Reg = ARM::R4; Reg < ARM::R12; ++Reg) { 2009 PushMIB.addReg( 2010 Reg, Reg == JumpReg || LiveRegs.contains(Reg) ? 0 : RegState::Undef); 2011 } 2012 } 2013 } 2014 2015 static void CMSEPopCalleeSaves(const TargetInstrInfo &TII, 2016 MachineBasicBlock &MBB, 2017 MachineBasicBlock::iterator MBBI, int JumpReg, 2018 bool Thumb1Only) { 2019 const DebugLoc &DL = MBBI->getDebugLoc(); 2020 if (Thumb1Only) { 2021 MachineInstrBuilder PopMIB = 2022 BuildMI(MBB, MBBI, DL, TII.get(ARM::tPOP)).add(predOps(ARMCC::AL)); 2023 for (int R = 0; R < 4; ++R) { 2024 PopMIB.addReg(ARM::R4 + R, RegState::Define); 2025 BuildMI(MBB, MBBI, DL, TII.get(ARM::tMOVr), ARM::R8 + R) 2026 .addReg(ARM::R4 + R, RegState::Kill) 2027 .add(predOps(ARMCC::AL)); 2028 } 2029 MachineInstrBuilder PopMIB2 = 2030 BuildMI(MBB, MBBI, DL, TII.get(ARM::tPOP)).add(predOps(ARMCC::AL)); 2031 for (int R = 0; R < 4; ++R) 2032 PopMIB2.addReg(ARM::R4 + R, RegState::Define); 2033 } else { // pop Lo and Hi registers with a single instruction 2034 MachineInstrBuilder PopMIB = 2035 BuildMI(MBB, MBBI, DL, TII.get(ARM::t2LDMIA_UPD), ARM::SP) 2036 .addReg(ARM::SP) 2037 .add(predOps(ARMCC::AL)); 2038 for (int Reg = ARM::R4; Reg < ARM::R12; ++Reg) 2039 PopMIB.addReg(Reg, RegState::Define); 2040 } 2041 } 2042 2043 bool ARMExpandPseudo::ExpandMI(MachineBasicBlock &MBB, 2044 MachineBasicBlock::iterator MBBI, 2045 MachineBasicBlock::iterator &NextMBBI) { 2046 MachineInstr &MI = *MBBI; 2047 unsigned Opcode = MI.getOpcode(); 2048 switch (Opcode) { 2049 default: 2050 return false; 2051 2052 case ARM::VBSPd: 2053 case ARM::VBSPq: { 2054 Register DstReg = MI.getOperand(0).getReg(); 2055 if (DstReg == MI.getOperand(3).getReg()) { 2056 // Expand to VBIT 2057 unsigned NewOpc = Opcode == ARM::VBSPd ? ARM::VBITd : ARM::VBITq; 2058 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(NewOpc)) 2059 .add(MI.getOperand(0)) 2060 .add(MI.getOperand(3)) 2061 .add(MI.getOperand(2)) 2062 .add(MI.getOperand(1)) 2063 .addImm(MI.getOperand(4).getImm()) 2064 .add(MI.getOperand(5)); 2065 } else if (DstReg == MI.getOperand(2).getReg()) { 2066 // Expand to VBIF 2067 unsigned NewOpc = Opcode == ARM::VBSPd ? ARM::VBIFd : ARM::VBIFq; 2068 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(NewOpc)) 2069 .add(MI.getOperand(0)) 2070 .add(MI.getOperand(2)) 2071 .add(MI.getOperand(3)) 2072 .add(MI.getOperand(1)) 2073 .addImm(MI.getOperand(4).getImm()) 2074 .add(MI.getOperand(5)); 2075 } else { 2076 // Expand to VBSL 2077 unsigned NewOpc = Opcode == ARM::VBSPd ? ARM::VBSLd : ARM::VBSLq; 2078 if (DstReg == MI.getOperand(1).getReg()) { 2079 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(NewOpc)) 2080 .add(MI.getOperand(0)) 2081 .add(MI.getOperand(1)) 2082 .add(MI.getOperand(2)) 2083 .add(MI.getOperand(3)) 2084 .addImm(MI.getOperand(4).getImm()) 2085 .add(MI.getOperand(5)); 2086 } else { 2087 // Use move to satisfy constraints 2088 unsigned MoveOpc = Opcode == ARM::VBSPd ? ARM::VORRd : ARM::VORRq; 2089 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(MoveOpc)) 2090 .addReg(DstReg, 2091 RegState::Define | 2092 getRenamableRegState(MI.getOperand(0).isRenamable())) 2093 .add(MI.getOperand(1)) 2094 .add(MI.getOperand(1)) 2095 .addImm(MI.getOperand(4).getImm()) 2096 .add(MI.getOperand(5)); 2097 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(NewOpc)) 2098 .add(MI.getOperand(0)) 2099 .addReg(DstReg, 2100 RegState::Kill | 2101 getRenamableRegState(MI.getOperand(0).isRenamable())) 2102 .add(MI.getOperand(2)) 2103 .add(MI.getOperand(3)) 2104 .addImm(MI.getOperand(4).getImm()) 2105 .add(MI.getOperand(5)); 2106 } 2107 } 2108 MI.eraseFromParent(); 2109 return true; 2110 } 2111 2112 case ARM::TCRETURNdi: 2113 case ARM::TCRETURNri: { 2114 MachineBasicBlock::iterator MBBI = MBB.getLastNonDebugInstr(); 2115 if (MBBI->getOpcode() == ARM::SEH_EpilogEnd) 2116 MBBI--; 2117 if (MBBI->getOpcode() == ARM::SEH_Nop_Ret) 2118 MBBI--; 2119 assert(MBBI->isReturn() && 2120 "Can only insert epilog into returning blocks"); 2121 unsigned RetOpcode = MBBI->getOpcode(); 2122 DebugLoc dl = MBBI->getDebugLoc(); 2123 const ARMBaseInstrInfo &TII = *static_cast<const ARMBaseInstrInfo *>( 2124 MBB.getParent()->getSubtarget().getInstrInfo()); 2125 2126 // Tail call return: adjust the stack pointer and jump to callee. 2127 MBBI = MBB.getLastNonDebugInstr(); 2128 if (MBBI->getOpcode() == ARM::SEH_EpilogEnd) 2129 MBBI--; 2130 if (MBBI->getOpcode() == ARM::SEH_Nop_Ret) 2131 MBBI--; 2132 MachineOperand &JumpTarget = MBBI->getOperand(0); 2133 2134 // Jump to label or value in register. 2135 if (RetOpcode == ARM::TCRETURNdi) { 2136 MachineFunction *MF = MBB.getParent(); 2137 bool NeedsWinCFI = MF->getTarget().getMCAsmInfo()->usesWindowsCFI() && 2138 MF->getFunction().needsUnwindTableEntry(); 2139 unsigned TCOpcode = 2140 STI->isThumb() 2141 ? ((STI->isTargetMachO() || NeedsWinCFI) ? ARM::tTAILJMPd 2142 : ARM::tTAILJMPdND) 2143 : ARM::TAILJMPd; 2144 MachineInstrBuilder MIB = BuildMI(MBB, MBBI, dl, TII.get(TCOpcode)); 2145 if (JumpTarget.isGlobal()) 2146 MIB.addGlobalAddress(JumpTarget.getGlobal(), JumpTarget.getOffset(), 2147 JumpTarget.getTargetFlags()); 2148 else { 2149 assert(JumpTarget.isSymbol()); 2150 MIB.addExternalSymbol(JumpTarget.getSymbolName(), 2151 JumpTarget.getTargetFlags()); 2152 } 2153 2154 // Add the default predicate in Thumb mode. 2155 if (STI->isThumb()) 2156 MIB.add(predOps(ARMCC::AL)); 2157 } else if (RetOpcode == ARM::TCRETURNri) { 2158 unsigned Opcode = 2159 STI->isThumb() ? ARM::tTAILJMPr 2160 : (STI->hasV4TOps() ? ARM::TAILJMPr : ARM::TAILJMPr4); 2161 BuildMI(MBB, MBBI, dl, 2162 TII.get(Opcode)) 2163 .addReg(JumpTarget.getReg(), RegState::Kill); 2164 } 2165 2166 auto NewMI = std::prev(MBBI); 2167 for (unsigned i = 2, e = MBBI->getNumOperands(); i != e; ++i) 2168 NewMI->addOperand(MBBI->getOperand(i)); 2169 2170 2171 // Update call site info and delete the pseudo instruction TCRETURN. 2172 if (MI.isCandidateForCallSiteEntry()) 2173 MI.getMF()->moveCallSiteInfo(&MI, &*NewMI); 2174 MBB.erase(MBBI); 2175 2176 MBBI = NewMI; 2177 return true; 2178 } 2179 case ARM::tBXNS_RET: { 2180 // For v8.0-M.Main we need to authenticate LR before clearing FPRs, which 2181 // uses R12 as a scratch register. 2182 if (!STI->hasV8_1MMainlineOps() && AFI->shouldSignReturnAddress()) 2183 BuildMI(MBB, MBBI, DebugLoc(), TII->get(ARM::t2AUT)); 2184 2185 MachineBasicBlock &AfterBB = CMSEClearFPRegs(MBB, MBBI); 2186 2187 if (STI->hasV8_1MMainlineOps()) { 2188 // Restore the non-secure floating point context. 2189 BuildMI(MBB, MBBI, MBBI->getDebugLoc(), 2190 TII->get(ARM::VLDR_FPCXTNS_post), ARM::SP) 2191 .addReg(ARM::SP) 2192 .addImm(4) 2193 .add(predOps(ARMCC::AL)); 2194 2195 if (AFI->shouldSignReturnAddress()) 2196 BuildMI(AfterBB, AfterBB.end(), DebugLoc(), TII->get(ARM::t2AUT)); 2197 } 2198 2199 // Clear all GPR that are not a use of the return instruction. 2200 assert(llvm::all_of(MBBI->operands(), [](const MachineOperand &Op) { 2201 return !Op.isReg() || Op.getReg() != ARM::R12; 2202 })); 2203 SmallVector<unsigned, 5> ClearRegs; 2204 determineGPRegsToClear( 2205 *MBBI, {ARM::R0, ARM::R1, ARM::R2, ARM::R3, ARM::R12}, ClearRegs); 2206 CMSEClearGPRegs(AfterBB, AfterBB.end(), MBBI->getDebugLoc(), ClearRegs, 2207 ARM::LR); 2208 2209 MachineInstrBuilder NewMI = 2210 BuildMI(AfterBB, AfterBB.end(), MBBI->getDebugLoc(), 2211 TII->get(ARM::tBXNS)) 2212 .addReg(ARM::LR) 2213 .add(predOps(ARMCC::AL)); 2214 for (const MachineOperand &Op : MI.operands()) 2215 NewMI->addOperand(Op); 2216 MI.eraseFromParent(); 2217 return true; 2218 } 2219 case ARM::tBLXNS_CALL: { 2220 DebugLoc DL = MBBI->getDebugLoc(); 2221 Register JumpReg = MBBI->getOperand(0).getReg(); 2222 2223 // Figure out which registers are live at the point immediately before the 2224 // call. When we indiscriminately push a set of registers, the live 2225 // registers are added as ordinary use operands, whereas dead registers 2226 // are "undef". 2227 LivePhysRegs LiveRegs(*TRI); 2228 LiveRegs.addLiveOuts(MBB); 2229 for (const MachineInstr &MI : make_range(MBB.rbegin(), MBBI.getReverse())) 2230 LiveRegs.stepBackward(MI); 2231 LiveRegs.stepBackward(*MBBI); 2232 2233 CMSEPushCalleeSaves(*TII, MBB, MBBI, JumpReg, LiveRegs, 2234 AFI->isThumb1OnlyFunction()); 2235 2236 SmallVector<unsigned, 16> ClearRegs; 2237 determineGPRegsToClear(*MBBI, 2238 {ARM::R0, ARM::R1, ARM::R2, ARM::R3, ARM::R4, 2239 ARM::R5, ARM::R6, ARM::R7, ARM::R8, ARM::R9, 2240 ARM::R10, ARM::R11, ARM::R12}, 2241 ClearRegs); 2242 auto OriginalClearRegs = ClearRegs; 2243 2244 // Get the first cleared register as a scratch (to use later with tBIC). 2245 // We need to use the first so we can ensure it is a low register. 2246 unsigned ScratchReg = ClearRegs.front(); 2247 2248 // Clear LSB of JumpReg 2249 if (AFI->isThumb2Function()) { 2250 BuildMI(MBB, MBBI, DL, TII->get(ARM::t2BICri), JumpReg) 2251 .addReg(JumpReg) 2252 .addImm(1) 2253 .add(predOps(ARMCC::AL)) 2254 .add(condCodeOp()); 2255 } else { 2256 // We need to use an extra register to cope with 8M Baseline, 2257 // since we have saved all of the registers we are ok to trash a non 2258 // argument register here. 2259 BuildMI(MBB, MBBI, DL, TII->get(ARM::tMOVi8), ScratchReg) 2260 .add(condCodeOp()) 2261 .addImm(1) 2262 .add(predOps(ARMCC::AL)); 2263 BuildMI(MBB, MBBI, DL, TII->get(ARM::tBIC), JumpReg) 2264 .addReg(ARM::CPSR, RegState::Define) 2265 .addReg(JumpReg) 2266 .addReg(ScratchReg) 2267 .add(predOps(ARMCC::AL)); 2268 } 2269 2270 CMSESaveClearFPRegs(MBB, MBBI, DL, LiveRegs, 2271 ClearRegs); // save+clear FP regs with ClearRegs 2272 CMSEClearGPRegs(MBB, MBBI, DL, ClearRegs, JumpReg); 2273 2274 const MachineInstrBuilder NewCall = 2275 BuildMI(MBB, MBBI, DL, TII->get(ARM::tBLXNSr)) 2276 .add(predOps(ARMCC::AL)) 2277 .addReg(JumpReg, RegState::Kill); 2278 2279 for (const MachineOperand &MO : llvm::drop_begin(MI.operands())) 2280 NewCall->addOperand(MO); 2281 if (MI.isCandidateForCallSiteEntry()) 2282 MI.getMF()->moveCallSiteInfo(&MI, NewCall.getInstr()); 2283 2284 CMSERestoreFPRegs(MBB, MBBI, DL, OriginalClearRegs); // restore FP registers 2285 2286 CMSEPopCalleeSaves(*TII, MBB, MBBI, JumpReg, AFI->isThumb1OnlyFunction()); 2287 2288 MI.eraseFromParent(); 2289 return true; 2290 } 2291 case ARM::VMOVHcc: 2292 case ARM::VMOVScc: 2293 case ARM::VMOVDcc: { 2294 unsigned newOpc = Opcode != ARM::VMOVDcc ? ARM::VMOVS : ARM::VMOVD; 2295 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(newOpc), 2296 MI.getOperand(1).getReg()) 2297 .add(MI.getOperand(2)) 2298 .addImm(MI.getOperand(3).getImm()) // 'pred' 2299 .add(MI.getOperand(4)) 2300 .add(makeImplicit(MI.getOperand(1))); 2301 2302 MI.eraseFromParent(); 2303 return true; 2304 } 2305 case ARM::t2MOVCCr: 2306 case ARM::MOVCCr: { 2307 unsigned Opc = AFI->isThumbFunction() ? ARM::t2MOVr : ARM::MOVr; 2308 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(Opc), 2309 MI.getOperand(1).getReg()) 2310 .add(MI.getOperand(2)) 2311 .addImm(MI.getOperand(3).getImm()) // 'pred' 2312 .add(MI.getOperand(4)) 2313 .add(condCodeOp()) // 's' bit 2314 .add(makeImplicit(MI.getOperand(1))); 2315 2316 MI.eraseFromParent(); 2317 return true; 2318 } 2319 case ARM::MOVCCsi: { 2320 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(ARM::MOVsi), 2321 (MI.getOperand(1).getReg())) 2322 .add(MI.getOperand(2)) 2323 .addImm(MI.getOperand(3).getImm()) 2324 .addImm(MI.getOperand(4).getImm()) // 'pred' 2325 .add(MI.getOperand(5)) 2326 .add(condCodeOp()) // 's' bit 2327 .add(makeImplicit(MI.getOperand(1))); 2328 2329 MI.eraseFromParent(); 2330 return true; 2331 } 2332 case ARM::MOVCCsr: { 2333 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(ARM::MOVsr), 2334 (MI.getOperand(1).getReg())) 2335 .add(MI.getOperand(2)) 2336 .add(MI.getOperand(3)) 2337 .addImm(MI.getOperand(4).getImm()) 2338 .addImm(MI.getOperand(5).getImm()) // 'pred' 2339 .add(MI.getOperand(6)) 2340 .add(condCodeOp()) // 's' bit 2341 .add(makeImplicit(MI.getOperand(1))); 2342 2343 MI.eraseFromParent(); 2344 return true; 2345 } 2346 case ARM::t2MOVCCi16: 2347 case ARM::MOVCCi16: { 2348 unsigned NewOpc = AFI->isThumbFunction() ? ARM::t2MOVi16 : ARM::MOVi16; 2349 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(NewOpc), 2350 MI.getOperand(1).getReg()) 2351 .addImm(MI.getOperand(2).getImm()) 2352 .addImm(MI.getOperand(3).getImm()) // 'pred' 2353 .add(MI.getOperand(4)) 2354 .add(makeImplicit(MI.getOperand(1))); 2355 MI.eraseFromParent(); 2356 return true; 2357 } 2358 case ARM::t2MOVCCi: 2359 case ARM::MOVCCi: { 2360 unsigned Opc = AFI->isThumbFunction() ? ARM::t2MOVi : ARM::MOVi; 2361 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(Opc), 2362 MI.getOperand(1).getReg()) 2363 .addImm(MI.getOperand(2).getImm()) 2364 .addImm(MI.getOperand(3).getImm()) // 'pred' 2365 .add(MI.getOperand(4)) 2366 .add(condCodeOp()) // 's' bit 2367 .add(makeImplicit(MI.getOperand(1))); 2368 2369 MI.eraseFromParent(); 2370 return true; 2371 } 2372 case ARM::t2MVNCCi: 2373 case ARM::MVNCCi: { 2374 unsigned Opc = AFI->isThumbFunction() ? ARM::t2MVNi : ARM::MVNi; 2375 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(Opc), 2376 MI.getOperand(1).getReg()) 2377 .addImm(MI.getOperand(2).getImm()) 2378 .addImm(MI.getOperand(3).getImm()) // 'pred' 2379 .add(MI.getOperand(4)) 2380 .add(condCodeOp()) // 's' bit 2381 .add(makeImplicit(MI.getOperand(1))); 2382 2383 MI.eraseFromParent(); 2384 return true; 2385 } 2386 case ARM::t2MOVCClsl: 2387 case ARM::t2MOVCClsr: 2388 case ARM::t2MOVCCasr: 2389 case ARM::t2MOVCCror: { 2390 unsigned NewOpc; 2391 switch (Opcode) { 2392 case ARM::t2MOVCClsl: NewOpc = ARM::t2LSLri; break; 2393 case ARM::t2MOVCClsr: NewOpc = ARM::t2LSRri; break; 2394 case ARM::t2MOVCCasr: NewOpc = ARM::t2ASRri; break; 2395 case ARM::t2MOVCCror: NewOpc = ARM::t2RORri; break; 2396 default: llvm_unreachable("unexpeced conditional move"); 2397 } 2398 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(NewOpc), 2399 MI.getOperand(1).getReg()) 2400 .add(MI.getOperand(2)) 2401 .addImm(MI.getOperand(3).getImm()) 2402 .addImm(MI.getOperand(4).getImm()) // 'pred' 2403 .add(MI.getOperand(5)) 2404 .add(condCodeOp()) // 's' bit 2405 .add(makeImplicit(MI.getOperand(1))); 2406 MI.eraseFromParent(); 2407 return true; 2408 } 2409 case ARM::Int_eh_sjlj_dispatchsetup: { 2410 MachineFunction &MF = *MI.getParent()->getParent(); 2411 const ARMBaseInstrInfo *AII = 2412 static_cast<const ARMBaseInstrInfo*>(TII); 2413 const ARMBaseRegisterInfo &RI = AII->getRegisterInfo(); 2414 // For functions using a base pointer, we rematerialize it (via the frame 2415 // pointer) here since eh.sjlj.setjmp and eh.sjlj.longjmp don't do it 2416 // for us. Otherwise, expand to nothing. 2417 if (RI.hasBasePointer(MF)) { 2418 int32_t NumBytes = AFI->getFramePtrSpillOffset(); 2419 Register FramePtr = RI.getFrameRegister(MF); 2420 assert(MF.getSubtarget().getFrameLowering()->hasFP(MF) && 2421 "base pointer without frame pointer?"); 2422 2423 if (AFI->isThumb2Function()) { 2424 emitT2RegPlusImmediate(MBB, MBBI, MI.getDebugLoc(), ARM::R6, 2425 FramePtr, -NumBytes, ARMCC::AL, 0, *TII); 2426 } else if (AFI->isThumbFunction()) { 2427 emitThumbRegPlusImmediate(MBB, MBBI, MI.getDebugLoc(), ARM::R6, 2428 FramePtr, -NumBytes, *TII, RI); 2429 } else { 2430 emitARMRegPlusImmediate(MBB, MBBI, MI.getDebugLoc(), ARM::R6, 2431 FramePtr, -NumBytes, ARMCC::AL, 0, 2432 *TII); 2433 } 2434 // If there's dynamic realignment, adjust for it. 2435 if (RI.hasStackRealignment(MF)) { 2436 MachineFrameInfo &MFI = MF.getFrameInfo(); 2437 Align MaxAlign = MFI.getMaxAlign(); 2438 assert (!AFI->isThumb1OnlyFunction()); 2439 // Emit bic r6, r6, MaxAlign 2440 assert(MaxAlign <= Align(256) && 2441 "The BIC instruction cannot encode " 2442 "immediates larger than 256 with all lower " 2443 "bits set."); 2444 unsigned bicOpc = AFI->isThumbFunction() ? 2445 ARM::t2BICri : ARM::BICri; 2446 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(bicOpc), ARM::R6) 2447 .addReg(ARM::R6, RegState::Kill) 2448 .addImm(MaxAlign.value() - 1) 2449 .add(predOps(ARMCC::AL)) 2450 .add(condCodeOp()); 2451 } 2452 } 2453 MI.eraseFromParent(); 2454 return true; 2455 } 2456 2457 case ARM::MOVsrl_flag: 2458 case ARM::MOVsra_flag: { 2459 // These are just fancy MOVs instructions. 2460 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(ARM::MOVsi), 2461 MI.getOperand(0).getReg()) 2462 .add(MI.getOperand(1)) 2463 .addImm(ARM_AM::getSORegOpc( 2464 (Opcode == ARM::MOVsrl_flag ? ARM_AM::lsr : ARM_AM::asr), 1)) 2465 .add(predOps(ARMCC::AL)) 2466 .addReg(ARM::CPSR, RegState::Define); 2467 MI.eraseFromParent(); 2468 return true; 2469 } 2470 case ARM::RRX: { 2471 // This encodes as "MOVs Rd, Rm, rrx 2472 MachineInstrBuilder MIB = 2473 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(ARM::MOVsi), 2474 MI.getOperand(0).getReg()) 2475 .add(MI.getOperand(1)) 2476 .addImm(ARM_AM::getSORegOpc(ARM_AM::rrx, 0)) 2477 .add(predOps(ARMCC::AL)) 2478 .add(condCodeOp()); 2479 TransferImpOps(MI, MIB, MIB); 2480 MI.eraseFromParent(); 2481 return true; 2482 } 2483 case ARM::tTPsoft: 2484 case ARM::TPsoft: { 2485 const bool Thumb = Opcode == ARM::tTPsoft; 2486 2487 MachineInstrBuilder MIB; 2488 MachineFunction *MF = MBB.getParent(); 2489 if (STI->genLongCalls()) { 2490 MachineConstantPool *MCP = MF->getConstantPool(); 2491 unsigned PCLabelID = AFI->createPICLabelUId(); 2492 MachineConstantPoolValue *CPV = 2493 ARMConstantPoolSymbol::Create(MF->getFunction().getContext(), 2494 "__aeabi_read_tp", PCLabelID, 0); 2495 Register Reg = MI.getOperand(0).getReg(); 2496 MIB = 2497 BuildMI(MBB, MBBI, MI.getDebugLoc(), 2498 TII->get(Thumb ? ARM::tLDRpci : ARM::LDRi12), Reg) 2499 .addConstantPoolIndex(MCP->getConstantPoolIndex(CPV, Align(4))); 2500 if (!Thumb) 2501 MIB.addImm(0); 2502 MIB.add(predOps(ARMCC::AL)); 2503 2504 MIB = 2505 BuildMI(MBB, MBBI, MI.getDebugLoc(), 2506 TII->get(Thumb ? gettBLXrOpcode(*MF) : getBLXOpcode(*MF))); 2507 if (Thumb) 2508 MIB.add(predOps(ARMCC::AL)); 2509 MIB.addReg(Reg, RegState::Kill); 2510 } else { 2511 MIB = BuildMI(MBB, MBBI, MI.getDebugLoc(), 2512 TII->get(Thumb ? ARM::tBL : ARM::BL)); 2513 if (Thumb) 2514 MIB.add(predOps(ARMCC::AL)); 2515 MIB.addExternalSymbol("__aeabi_read_tp", 0); 2516 } 2517 2518 MIB.cloneMemRefs(MI); 2519 TransferImpOps(MI, MIB, MIB); 2520 // Update the call site info. 2521 if (MI.isCandidateForCallSiteEntry()) 2522 MF->moveCallSiteInfo(&MI, &*MIB); 2523 MI.eraseFromParent(); 2524 return true; 2525 } 2526 case ARM::tLDRpci_pic: 2527 case ARM::t2LDRpci_pic: { 2528 unsigned NewLdOpc = (Opcode == ARM::tLDRpci_pic) 2529 ? ARM::tLDRpci : ARM::t2LDRpci; 2530 Register DstReg = MI.getOperand(0).getReg(); 2531 bool DstIsDead = MI.getOperand(0).isDead(); 2532 MachineInstrBuilder MIB1 = 2533 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(NewLdOpc), DstReg) 2534 .add(MI.getOperand(1)) 2535 .add(predOps(ARMCC::AL)); 2536 MIB1.cloneMemRefs(MI); 2537 MachineInstrBuilder MIB2 = 2538 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(ARM::tPICADD)) 2539 .addReg(DstReg, RegState::Define | getDeadRegState(DstIsDead)) 2540 .addReg(DstReg) 2541 .add(MI.getOperand(2)); 2542 TransferImpOps(MI, MIB1, MIB2); 2543 MI.eraseFromParent(); 2544 return true; 2545 } 2546 2547 case ARM::LDRLIT_ga_abs: 2548 case ARM::LDRLIT_ga_pcrel: 2549 case ARM::LDRLIT_ga_pcrel_ldr: 2550 case ARM::tLDRLIT_ga_abs: 2551 case ARM::t2LDRLIT_ga_pcrel: 2552 case ARM::tLDRLIT_ga_pcrel: { 2553 Register DstReg = MI.getOperand(0).getReg(); 2554 bool DstIsDead = MI.getOperand(0).isDead(); 2555 const MachineOperand &MO1 = MI.getOperand(1); 2556 auto Flags = MO1.getTargetFlags(); 2557 const GlobalValue *GV = MO1.getGlobal(); 2558 bool IsARM = Opcode != ARM::tLDRLIT_ga_pcrel && 2559 Opcode != ARM::tLDRLIT_ga_abs && 2560 Opcode != ARM::t2LDRLIT_ga_pcrel; 2561 bool IsPIC = 2562 Opcode != ARM::LDRLIT_ga_abs && Opcode != ARM::tLDRLIT_ga_abs; 2563 unsigned LDRLITOpc = IsARM ? ARM::LDRi12 : ARM::tLDRpci; 2564 if (Opcode == ARM::t2LDRLIT_ga_pcrel) 2565 LDRLITOpc = ARM::t2LDRpci; 2566 unsigned PICAddOpc = 2567 IsARM 2568 ? (Opcode == ARM::LDRLIT_ga_pcrel_ldr ? ARM::PICLDR : ARM::PICADD) 2569 : ARM::tPICADD; 2570 2571 // We need a new const-pool entry to load from. 2572 MachineConstantPool *MCP = MBB.getParent()->getConstantPool(); 2573 unsigned ARMPCLabelIndex = 0; 2574 MachineConstantPoolValue *CPV; 2575 2576 if (IsPIC) { 2577 unsigned PCAdj = IsARM ? 8 : 4; 2578 auto Modifier = (Flags & ARMII::MO_GOT) 2579 ? ARMCP::GOT_PREL 2580 : ARMCP::no_modifier; 2581 ARMPCLabelIndex = AFI->createPICLabelUId(); 2582 CPV = ARMConstantPoolConstant::Create( 2583 GV, ARMPCLabelIndex, ARMCP::CPValue, PCAdj, Modifier, 2584 /*AddCurrentAddr*/ Modifier == ARMCP::GOT_PREL); 2585 } else 2586 CPV = ARMConstantPoolConstant::Create(GV, ARMCP::no_modifier); 2587 2588 MachineInstrBuilder MIB = 2589 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(LDRLITOpc), DstReg) 2590 .addConstantPoolIndex(MCP->getConstantPoolIndex(CPV, Align(4))); 2591 if (IsARM) 2592 MIB.addImm(0); 2593 MIB.add(predOps(ARMCC::AL)); 2594 2595 if (IsPIC) { 2596 MachineInstrBuilder MIB = 2597 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(PICAddOpc)) 2598 .addReg(DstReg, RegState::Define | getDeadRegState(DstIsDead)) 2599 .addReg(DstReg) 2600 .addImm(ARMPCLabelIndex); 2601 2602 if (IsARM) 2603 MIB.add(predOps(ARMCC::AL)); 2604 } 2605 2606 MI.eraseFromParent(); 2607 return true; 2608 } 2609 case ARM::MOV_ga_pcrel: 2610 case ARM::MOV_ga_pcrel_ldr: 2611 case ARM::t2MOV_ga_pcrel: { 2612 // Expand into movw + movw. Also "add pc" / ldr [pc] in PIC mode. 2613 unsigned LabelId = AFI->createPICLabelUId(); 2614 Register DstReg = MI.getOperand(0).getReg(); 2615 bool DstIsDead = MI.getOperand(0).isDead(); 2616 const MachineOperand &MO1 = MI.getOperand(1); 2617 const GlobalValue *GV = MO1.getGlobal(); 2618 unsigned TF = MO1.getTargetFlags(); 2619 bool isARM = Opcode != ARM::t2MOV_ga_pcrel; 2620 unsigned LO16Opc = isARM ? ARM::MOVi16_ga_pcrel : ARM::t2MOVi16_ga_pcrel; 2621 unsigned HI16Opc = isARM ? ARM::MOVTi16_ga_pcrel :ARM::t2MOVTi16_ga_pcrel; 2622 unsigned LO16TF = TF | ARMII::MO_LO16; 2623 unsigned HI16TF = TF | ARMII::MO_HI16; 2624 unsigned PICAddOpc = isARM 2625 ? (Opcode == ARM::MOV_ga_pcrel_ldr ? ARM::PICLDR : ARM::PICADD) 2626 : ARM::tPICADD; 2627 MachineInstrBuilder MIB1 = BuildMI(MBB, MBBI, MI.getDebugLoc(), 2628 TII->get(LO16Opc), DstReg) 2629 .addGlobalAddress(GV, MO1.getOffset(), TF | LO16TF) 2630 .addImm(LabelId); 2631 2632 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(HI16Opc), DstReg) 2633 .addReg(DstReg) 2634 .addGlobalAddress(GV, MO1.getOffset(), TF | HI16TF) 2635 .addImm(LabelId); 2636 2637 MachineInstrBuilder MIB3 = BuildMI(MBB, MBBI, MI.getDebugLoc(), 2638 TII->get(PICAddOpc)) 2639 .addReg(DstReg, RegState::Define | getDeadRegState(DstIsDead)) 2640 .addReg(DstReg).addImm(LabelId); 2641 if (isARM) { 2642 MIB3.add(predOps(ARMCC::AL)); 2643 if (Opcode == ARM::MOV_ga_pcrel_ldr) 2644 MIB3.cloneMemRefs(MI); 2645 } 2646 TransferImpOps(MI, MIB1, MIB3); 2647 MI.eraseFromParent(); 2648 return true; 2649 } 2650 2651 case ARM::MOVi32imm: 2652 case ARM::MOVCCi32imm: 2653 case ARM::t2MOVi32imm: 2654 case ARM::t2MOVCCi32imm: 2655 ExpandMOV32BitImm(MBB, MBBI); 2656 return true; 2657 2658 case ARM::SUBS_PC_LR: { 2659 MachineInstrBuilder MIB = 2660 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(ARM::SUBri), ARM::PC) 2661 .addReg(ARM::LR) 2662 .add(MI.getOperand(0)) 2663 .add(MI.getOperand(1)) 2664 .add(MI.getOperand(2)) 2665 .addReg(ARM::CPSR, RegState::Undef); 2666 TransferImpOps(MI, MIB, MIB); 2667 MI.eraseFromParent(); 2668 return true; 2669 } 2670 case ARM::VLDMQIA: { 2671 unsigned NewOpc = ARM::VLDMDIA; 2672 MachineInstrBuilder MIB = 2673 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(NewOpc)); 2674 unsigned OpIdx = 0; 2675 2676 // Grab the Q register destination. 2677 bool DstIsDead = MI.getOperand(OpIdx).isDead(); 2678 Register DstReg = MI.getOperand(OpIdx++).getReg(); 2679 2680 // Copy the source register. 2681 MIB.add(MI.getOperand(OpIdx++)); 2682 2683 // Copy the predicate operands. 2684 MIB.add(MI.getOperand(OpIdx++)); 2685 MIB.add(MI.getOperand(OpIdx++)); 2686 2687 // Add the destination operands (D subregs). 2688 Register D0 = TRI->getSubReg(DstReg, ARM::dsub_0); 2689 Register D1 = TRI->getSubReg(DstReg, ARM::dsub_1); 2690 MIB.addReg(D0, RegState::Define | getDeadRegState(DstIsDead)) 2691 .addReg(D1, RegState::Define | getDeadRegState(DstIsDead)); 2692 2693 // Add an implicit def for the super-register. 2694 MIB.addReg(DstReg, RegState::ImplicitDefine | getDeadRegState(DstIsDead)); 2695 TransferImpOps(MI, MIB, MIB); 2696 MIB.cloneMemRefs(MI); 2697 MI.eraseFromParent(); 2698 return true; 2699 } 2700 2701 case ARM::VSTMQIA: { 2702 unsigned NewOpc = ARM::VSTMDIA; 2703 MachineInstrBuilder MIB = 2704 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(NewOpc)); 2705 unsigned OpIdx = 0; 2706 2707 // Grab the Q register source. 2708 bool SrcIsKill = MI.getOperand(OpIdx).isKill(); 2709 Register SrcReg = MI.getOperand(OpIdx++).getReg(); 2710 2711 // Copy the destination register. 2712 MachineOperand Dst(MI.getOperand(OpIdx++)); 2713 MIB.add(Dst); 2714 2715 // Copy the predicate operands. 2716 MIB.add(MI.getOperand(OpIdx++)); 2717 MIB.add(MI.getOperand(OpIdx++)); 2718 2719 // Add the source operands (D subregs). 2720 Register D0 = TRI->getSubReg(SrcReg, ARM::dsub_0); 2721 Register D1 = TRI->getSubReg(SrcReg, ARM::dsub_1); 2722 MIB.addReg(D0, SrcIsKill ? RegState::Kill : 0) 2723 .addReg(D1, SrcIsKill ? RegState::Kill : 0); 2724 2725 if (SrcIsKill) // Add an implicit kill for the Q register. 2726 MIB->addRegisterKilled(SrcReg, TRI, true); 2727 2728 TransferImpOps(MI, MIB, MIB); 2729 MIB.cloneMemRefs(MI); 2730 MI.eraseFromParent(); 2731 return true; 2732 } 2733 2734 case ARM::VLD2q8Pseudo: 2735 case ARM::VLD2q16Pseudo: 2736 case ARM::VLD2q32Pseudo: 2737 case ARM::VLD2q8PseudoWB_fixed: 2738 case ARM::VLD2q16PseudoWB_fixed: 2739 case ARM::VLD2q32PseudoWB_fixed: 2740 case ARM::VLD2q8PseudoWB_register: 2741 case ARM::VLD2q16PseudoWB_register: 2742 case ARM::VLD2q32PseudoWB_register: 2743 case ARM::VLD3d8Pseudo: 2744 case ARM::VLD3d16Pseudo: 2745 case ARM::VLD3d32Pseudo: 2746 case ARM::VLD1d8TPseudo: 2747 case ARM::VLD1d8TPseudoWB_fixed: 2748 case ARM::VLD1d8TPseudoWB_register: 2749 case ARM::VLD1d16TPseudo: 2750 case ARM::VLD1d16TPseudoWB_fixed: 2751 case ARM::VLD1d16TPseudoWB_register: 2752 case ARM::VLD1d32TPseudo: 2753 case ARM::VLD1d32TPseudoWB_fixed: 2754 case ARM::VLD1d32TPseudoWB_register: 2755 case ARM::VLD1d64TPseudo: 2756 case ARM::VLD1d64TPseudoWB_fixed: 2757 case ARM::VLD1d64TPseudoWB_register: 2758 case ARM::VLD3d8Pseudo_UPD: 2759 case ARM::VLD3d16Pseudo_UPD: 2760 case ARM::VLD3d32Pseudo_UPD: 2761 case ARM::VLD3q8Pseudo_UPD: 2762 case ARM::VLD3q16Pseudo_UPD: 2763 case ARM::VLD3q32Pseudo_UPD: 2764 case ARM::VLD3q8oddPseudo: 2765 case ARM::VLD3q16oddPseudo: 2766 case ARM::VLD3q32oddPseudo: 2767 case ARM::VLD3q8oddPseudo_UPD: 2768 case ARM::VLD3q16oddPseudo_UPD: 2769 case ARM::VLD3q32oddPseudo_UPD: 2770 case ARM::VLD4d8Pseudo: 2771 case ARM::VLD4d16Pseudo: 2772 case ARM::VLD4d32Pseudo: 2773 case ARM::VLD1d8QPseudo: 2774 case ARM::VLD1d8QPseudoWB_fixed: 2775 case ARM::VLD1d8QPseudoWB_register: 2776 case ARM::VLD1d16QPseudo: 2777 case ARM::VLD1d16QPseudoWB_fixed: 2778 case ARM::VLD1d16QPseudoWB_register: 2779 case ARM::VLD1d32QPseudo: 2780 case ARM::VLD1d32QPseudoWB_fixed: 2781 case ARM::VLD1d32QPseudoWB_register: 2782 case ARM::VLD1d64QPseudo: 2783 case ARM::VLD1d64QPseudoWB_fixed: 2784 case ARM::VLD1d64QPseudoWB_register: 2785 case ARM::VLD1q8HighQPseudo: 2786 case ARM::VLD1q8HighQPseudo_UPD: 2787 case ARM::VLD1q8LowQPseudo_UPD: 2788 case ARM::VLD1q8HighTPseudo: 2789 case ARM::VLD1q8HighTPseudo_UPD: 2790 case ARM::VLD1q8LowTPseudo_UPD: 2791 case ARM::VLD1q16HighQPseudo: 2792 case ARM::VLD1q16HighQPseudo_UPD: 2793 case ARM::VLD1q16LowQPseudo_UPD: 2794 case ARM::VLD1q16HighTPseudo: 2795 case ARM::VLD1q16HighTPseudo_UPD: 2796 case ARM::VLD1q16LowTPseudo_UPD: 2797 case ARM::VLD1q32HighQPseudo: 2798 case ARM::VLD1q32HighQPseudo_UPD: 2799 case ARM::VLD1q32LowQPseudo_UPD: 2800 case ARM::VLD1q32HighTPseudo: 2801 case ARM::VLD1q32HighTPseudo_UPD: 2802 case ARM::VLD1q32LowTPseudo_UPD: 2803 case ARM::VLD1q64HighQPseudo: 2804 case ARM::VLD1q64HighQPseudo_UPD: 2805 case ARM::VLD1q64LowQPseudo_UPD: 2806 case ARM::VLD1q64HighTPseudo: 2807 case ARM::VLD1q64HighTPseudo_UPD: 2808 case ARM::VLD1q64LowTPseudo_UPD: 2809 case ARM::VLD4d8Pseudo_UPD: 2810 case ARM::VLD4d16Pseudo_UPD: 2811 case ARM::VLD4d32Pseudo_UPD: 2812 case ARM::VLD4q8Pseudo_UPD: 2813 case ARM::VLD4q16Pseudo_UPD: 2814 case ARM::VLD4q32Pseudo_UPD: 2815 case ARM::VLD4q8oddPseudo: 2816 case ARM::VLD4q16oddPseudo: 2817 case ARM::VLD4q32oddPseudo: 2818 case ARM::VLD4q8oddPseudo_UPD: 2819 case ARM::VLD4q16oddPseudo_UPD: 2820 case ARM::VLD4q32oddPseudo_UPD: 2821 case ARM::VLD3DUPd8Pseudo: 2822 case ARM::VLD3DUPd16Pseudo: 2823 case ARM::VLD3DUPd32Pseudo: 2824 case ARM::VLD3DUPd8Pseudo_UPD: 2825 case ARM::VLD3DUPd16Pseudo_UPD: 2826 case ARM::VLD3DUPd32Pseudo_UPD: 2827 case ARM::VLD4DUPd8Pseudo: 2828 case ARM::VLD4DUPd16Pseudo: 2829 case ARM::VLD4DUPd32Pseudo: 2830 case ARM::VLD4DUPd8Pseudo_UPD: 2831 case ARM::VLD4DUPd16Pseudo_UPD: 2832 case ARM::VLD4DUPd32Pseudo_UPD: 2833 case ARM::VLD2DUPq8EvenPseudo: 2834 case ARM::VLD2DUPq8OddPseudo: 2835 case ARM::VLD2DUPq16EvenPseudo: 2836 case ARM::VLD2DUPq16OddPseudo: 2837 case ARM::VLD2DUPq32EvenPseudo: 2838 case ARM::VLD2DUPq32OddPseudo: 2839 case ARM::VLD2DUPq8OddPseudoWB_fixed: 2840 case ARM::VLD2DUPq8OddPseudoWB_register: 2841 case ARM::VLD2DUPq16OddPseudoWB_fixed: 2842 case ARM::VLD2DUPq16OddPseudoWB_register: 2843 case ARM::VLD2DUPq32OddPseudoWB_fixed: 2844 case ARM::VLD2DUPq32OddPseudoWB_register: 2845 case ARM::VLD3DUPq8EvenPseudo: 2846 case ARM::VLD3DUPq8OddPseudo: 2847 case ARM::VLD3DUPq16EvenPseudo: 2848 case ARM::VLD3DUPq16OddPseudo: 2849 case ARM::VLD3DUPq32EvenPseudo: 2850 case ARM::VLD3DUPq32OddPseudo: 2851 case ARM::VLD3DUPq8OddPseudo_UPD: 2852 case ARM::VLD3DUPq16OddPseudo_UPD: 2853 case ARM::VLD3DUPq32OddPseudo_UPD: 2854 case ARM::VLD4DUPq8EvenPseudo: 2855 case ARM::VLD4DUPq8OddPseudo: 2856 case ARM::VLD4DUPq16EvenPseudo: 2857 case ARM::VLD4DUPq16OddPseudo: 2858 case ARM::VLD4DUPq32EvenPseudo: 2859 case ARM::VLD4DUPq32OddPseudo: 2860 case ARM::VLD4DUPq8OddPseudo_UPD: 2861 case ARM::VLD4DUPq16OddPseudo_UPD: 2862 case ARM::VLD4DUPq32OddPseudo_UPD: 2863 ExpandVLD(MBBI); 2864 return true; 2865 2866 case ARM::VST2q8Pseudo: 2867 case ARM::VST2q16Pseudo: 2868 case ARM::VST2q32Pseudo: 2869 case ARM::VST2q8PseudoWB_fixed: 2870 case ARM::VST2q16PseudoWB_fixed: 2871 case ARM::VST2q32PseudoWB_fixed: 2872 case ARM::VST2q8PseudoWB_register: 2873 case ARM::VST2q16PseudoWB_register: 2874 case ARM::VST2q32PseudoWB_register: 2875 case ARM::VST3d8Pseudo: 2876 case ARM::VST3d16Pseudo: 2877 case ARM::VST3d32Pseudo: 2878 case ARM::VST1d8TPseudo: 2879 case ARM::VST1d8TPseudoWB_fixed: 2880 case ARM::VST1d8TPseudoWB_register: 2881 case ARM::VST1d16TPseudo: 2882 case ARM::VST1d16TPseudoWB_fixed: 2883 case ARM::VST1d16TPseudoWB_register: 2884 case ARM::VST1d32TPseudo: 2885 case ARM::VST1d32TPseudoWB_fixed: 2886 case ARM::VST1d32TPseudoWB_register: 2887 case ARM::VST1d64TPseudo: 2888 case ARM::VST1d64TPseudoWB_fixed: 2889 case ARM::VST1d64TPseudoWB_register: 2890 case ARM::VST3d8Pseudo_UPD: 2891 case ARM::VST3d16Pseudo_UPD: 2892 case ARM::VST3d32Pseudo_UPD: 2893 case ARM::VST3q8Pseudo_UPD: 2894 case ARM::VST3q16Pseudo_UPD: 2895 case ARM::VST3q32Pseudo_UPD: 2896 case ARM::VST3q8oddPseudo: 2897 case ARM::VST3q16oddPseudo: 2898 case ARM::VST3q32oddPseudo: 2899 case ARM::VST3q8oddPseudo_UPD: 2900 case ARM::VST3q16oddPseudo_UPD: 2901 case ARM::VST3q32oddPseudo_UPD: 2902 case ARM::VST4d8Pseudo: 2903 case ARM::VST4d16Pseudo: 2904 case ARM::VST4d32Pseudo: 2905 case ARM::VST1d8QPseudo: 2906 case ARM::VST1d8QPseudoWB_fixed: 2907 case ARM::VST1d8QPseudoWB_register: 2908 case ARM::VST1d16QPseudo: 2909 case ARM::VST1d16QPseudoWB_fixed: 2910 case ARM::VST1d16QPseudoWB_register: 2911 case ARM::VST1d32QPseudo: 2912 case ARM::VST1d32QPseudoWB_fixed: 2913 case ARM::VST1d32QPseudoWB_register: 2914 case ARM::VST1d64QPseudo: 2915 case ARM::VST1d64QPseudoWB_fixed: 2916 case ARM::VST1d64QPseudoWB_register: 2917 case ARM::VST4d8Pseudo_UPD: 2918 case ARM::VST4d16Pseudo_UPD: 2919 case ARM::VST4d32Pseudo_UPD: 2920 case ARM::VST1q8HighQPseudo: 2921 case ARM::VST1q8LowQPseudo_UPD: 2922 case ARM::VST1q8HighTPseudo: 2923 case ARM::VST1q8LowTPseudo_UPD: 2924 case ARM::VST1q16HighQPseudo: 2925 case ARM::VST1q16LowQPseudo_UPD: 2926 case ARM::VST1q16HighTPseudo: 2927 case ARM::VST1q16LowTPseudo_UPD: 2928 case ARM::VST1q32HighQPseudo: 2929 case ARM::VST1q32LowQPseudo_UPD: 2930 case ARM::VST1q32HighTPseudo: 2931 case ARM::VST1q32LowTPseudo_UPD: 2932 case ARM::VST1q64HighQPseudo: 2933 case ARM::VST1q64LowQPseudo_UPD: 2934 case ARM::VST1q64HighTPseudo: 2935 case ARM::VST1q64LowTPseudo_UPD: 2936 case ARM::VST1q8HighTPseudo_UPD: 2937 case ARM::VST1q16HighTPseudo_UPD: 2938 case ARM::VST1q32HighTPseudo_UPD: 2939 case ARM::VST1q64HighTPseudo_UPD: 2940 case ARM::VST1q8HighQPseudo_UPD: 2941 case ARM::VST1q16HighQPseudo_UPD: 2942 case ARM::VST1q32HighQPseudo_UPD: 2943 case ARM::VST1q64HighQPseudo_UPD: 2944 case ARM::VST4q8Pseudo_UPD: 2945 case ARM::VST4q16Pseudo_UPD: 2946 case ARM::VST4q32Pseudo_UPD: 2947 case ARM::VST4q8oddPseudo: 2948 case ARM::VST4q16oddPseudo: 2949 case ARM::VST4q32oddPseudo: 2950 case ARM::VST4q8oddPseudo_UPD: 2951 case ARM::VST4q16oddPseudo_UPD: 2952 case ARM::VST4q32oddPseudo_UPD: 2953 ExpandVST(MBBI); 2954 return true; 2955 2956 case ARM::VLD1LNq8Pseudo: 2957 case ARM::VLD1LNq16Pseudo: 2958 case ARM::VLD1LNq32Pseudo: 2959 case ARM::VLD1LNq8Pseudo_UPD: 2960 case ARM::VLD1LNq16Pseudo_UPD: 2961 case ARM::VLD1LNq32Pseudo_UPD: 2962 case ARM::VLD2LNd8Pseudo: 2963 case ARM::VLD2LNd16Pseudo: 2964 case ARM::VLD2LNd32Pseudo: 2965 case ARM::VLD2LNq16Pseudo: 2966 case ARM::VLD2LNq32Pseudo: 2967 case ARM::VLD2LNd8Pseudo_UPD: 2968 case ARM::VLD2LNd16Pseudo_UPD: 2969 case ARM::VLD2LNd32Pseudo_UPD: 2970 case ARM::VLD2LNq16Pseudo_UPD: 2971 case ARM::VLD2LNq32Pseudo_UPD: 2972 case ARM::VLD3LNd8Pseudo: 2973 case ARM::VLD3LNd16Pseudo: 2974 case ARM::VLD3LNd32Pseudo: 2975 case ARM::VLD3LNq16Pseudo: 2976 case ARM::VLD3LNq32Pseudo: 2977 case ARM::VLD3LNd8Pseudo_UPD: 2978 case ARM::VLD3LNd16Pseudo_UPD: 2979 case ARM::VLD3LNd32Pseudo_UPD: 2980 case ARM::VLD3LNq16Pseudo_UPD: 2981 case ARM::VLD3LNq32Pseudo_UPD: 2982 case ARM::VLD4LNd8Pseudo: 2983 case ARM::VLD4LNd16Pseudo: 2984 case ARM::VLD4LNd32Pseudo: 2985 case ARM::VLD4LNq16Pseudo: 2986 case ARM::VLD4LNq32Pseudo: 2987 case ARM::VLD4LNd8Pseudo_UPD: 2988 case ARM::VLD4LNd16Pseudo_UPD: 2989 case ARM::VLD4LNd32Pseudo_UPD: 2990 case ARM::VLD4LNq16Pseudo_UPD: 2991 case ARM::VLD4LNq32Pseudo_UPD: 2992 case ARM::VST1LNq8Pseudo: 2993 case ARM::VST1LNq16Pseudo: 2994 case ARM::VST1LNq32Pseudo: 2995 case ARM::VST1LNq8Pseudo_UPD: 2996 case ARM::VST1LNq16Pseudo_UPD: 2997 case ARM::VST1LNq32Pseudo_UPD: 2998 case ARM::VST2LNd8Pseudo: 2999 case ARM::VST2LNd16Pseudo: 3000 case ARM::VST2LNd32Pseudo: 3001 case ARM::VST2LNq16Pseudo: 3002 case ARM::VST2LNq32Pseudo: 3003 case ARM::VST2LNd8Pseudo_UPD: 3004 case ARM::VST2LNd16Pseudo_UPD: 3005 case ARM::VST2LNd32Pseudo_UPD: 3006 case ARM::VST2LNq16Pseudo_UPD: 3007 case ARM::VST2LNq32Pseudo_UPD: 3008 case ARM::VST3LNd8Pseudo: 3009 case ARM::VST3LNd16Pseudo: 3010 case ARM::VST3LNd32Pseudo: 3011 case ARM::VST3LNq16Pseudo: 3012 case ARM::VST3LNq32Pseudo: 3013 case ARM::VST3LNd8Pseudo_UPD: 3014 case ARM::VST3LNd16Pseudo_UPD: 3015 case ARM::VST3LNd32Pseudo_UPD: 3016 case ARM::VST3LNq16Pseudo_UPD: 3017 case ARM::VST3LNq32Pseudo_UPD: 3018 case ARM::VST4LNd8Pseudo: 3019 case ARM::VST4LNd16Pseudo: 3020 case ARM::VST4LNd32Pseudo: 3021 case ARM::VST4LNq16Pseudo: 3022 case ARM::VST4LNq32Pseudo: 3023 case ARM::VST4LNd8Pseudo_UPD: 3024 case ARM::VST4LNd16Pseudo_UPD: 3025 case ARM::VST4LNd32Pseudo_UPD: 3026 case ARM::VST4LNq16Pseudo_UPD: 3027 case ARM::VST4LNq32Pseudo_UPD: 3028 ExpandLaneOp(MBBI); 3029 return true; 3030 3031 case ARM::VTBL3Pseudo: ExpandVTBL(MBBI, ARM::VTBL3, false); return true; 3032 case ARM::VTBL4Pseudo: ExpandVTBL(MBBI, ARM::VTBL4, false); return true; 3033 case ARM::VTBX3Pseudo: ExpandVTBL(MBBI, ARM::VTBX3, true); return true; 3034 case ARM::VTBX4Pseudo: ExpandVTBL(MBBI, ARM::VTBX4, true); return true; 3035 3036 case ARM::MQQPRLoad: 3037 case ARM::MQQPRStore: 3038 case ARM::MQQQQPRLoad: 3039 case ARM::MQQQQPRStore: 3040 ExpandMQQPRLoadStore(MBBI); 3041 return true; 3042 3043 case ARM::tCMP_SWAP_8: 3044 assert(STI->isThumb()); 3045 return ExpandCMP_SWAP(MBB, MBBI, ARM::t2LDREXB, ARM::t2STREXB, ARM::tUXTB, 3046 NextMBBI); 3047 case ARM::tCMP_SWAP_16: 3048 assert(STI->isThumb()); 3049 return ExpandCMP_SWAP(MBB, MBBI, ARM::t2LDREXH, ARM::t2STREXH, ARM::tUXTH, 3050 NextMBBI); 3051 case ARM::tCMP_SWAP_32: 3052 assert(STI->isThumb()); 3053 return ExpandCMP_SWAP(MBB, MBBI, ARM::t2LDREX, ARM::t2STREX, 0, NextMBBI); 3054 3055 case ARM::CMP_SWAP_8: 3056 assert(!STI->isThumb()); 3057 return ExpandCMP_SWAP(MBB, MBBI, ARM::LDREXB, ARM::STREXB, ARM::UXTB, 3058 NextMBBI); 3059 case ARM::CMP_SWAP_16: 3060 assert(!STI->isThumb()); 3061 return ExpandCMP_SWAP(MBB, MBBI, ARM::LDREXH, ARM::STREXH, ARM::UXTH, 3062 NextMBBI); 3063 case ARM::CMP_SWAP_32: 3064 assert(!STI->isThumb()); 3065 return ExpandCMP_SWAP(MBB, MBBI, ARM::LDREX, ARM::STREX, 0, NextMBBI); 3066 3067 case ARM::CMP_SWAP_64: 3068 return ExpandCMP_SWAP_64(MBB, MBBI, NextMBBI); 3069 3070 case ARM::tBL_PUSHLR: 3071 case ARM::BL_PUSHLR: { 3072 const bool Thumb = Opcode == ARM::tBL_PUSHLR; 3073 Register Reg = MI.getOperand(0).getReg(); 3074 assert(Reg == ARM::LR && "expect LR register!"); 3075 MachineInstrBuilder MIB; 3076 if (Thumb) { 3077 // push {lr} 3078 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(ARM::tPUSH)) 3079 .add(predOps(ARMCC::AL)) 3080 .addReg(Reg); 3081 3082 // bl __gnu_mcount_nc 3083 MIB = BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(ARM::tBL)); 3084 } else { 3085 // stmdb sp!, {lr} 3086 BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(ARM::STMDB_UPD)) 3087 .addReg(ARM::SP, RegState::Define) 3088 .addReg(ARM::SP) 3089 .add(predOps(ARMCC::AL)) 3090 .addReg(Reg); 3091 3092 // bl __gnu_mcount_nc 3093 MIB = BuildMI(MBB, MBBI, MI.getDebugLoc(), TII->get(ARM::BL)); 3094 } 3095 MIB.cloneMemRefs(MI); 3096 for (const MachineOperand &MO : llvm::drop_begin(MI.operands())) 3097 MIB.add(MO); 3098 MI.eraseFromParent(); 3099 return true; 3100 } 3101 case ARM::t2CALL_BTI: { 3102 MachineFunction &MF = *MI.getMF(); 3103 MachineInstrBuilder MIB = 3104 BuildMI(MF, MI.getDebugLoc(), TII->get(ARM::tBL)); 3105 MIB.cloneMemRefs(MI); 3106 for (unsigned i = 0; i < MI.getNumOperands(); ++i) 3107 MIB.add(MI.getOperand(i)); 3108 if (MI.isCandidateForCallSiteEntry()) 3109 MF.moveCallSiteInfo(&MI, MIB.getInstr()); 3110 MIBundleBuilder Bundler(MBB, MI); 3111 Bundler.append(MIB); 3112 Bundler.append(BuildMI(MF, MI.getDebugLoc(), TII->get(ARM::t2BTI))); 3113 finalizeBundle(MBB, Bundler.begin(), Bundler.end()); 3114 MI.eraseFromParent(); 3115 return true; 3116 } 3117 case ARM::LOADDUAL: 3118 case ARM::STOREDUAL: { 3119 Register PairReg = MI.getOperand(0).getReg(); 3120 3121 MachineInstrBuilder MIB = 3122 BuildMI(MBB, MBBI, MI.getDebugLoc(), 3123 TII->get(Opcode == ARM::LOADDUAL ? ARM::LDRD : ARM::STRD)) 3124 .addReg(TRI->getSubReg(PairReg, ARM::gsub_0), 3125 Opcode == ARM::LOADDUAL ? RegState::Define : 0) 3126 .addReg(TRI->getSubReg(PairReg, ARM::gsub_1), 3127 Opcode == ARM::LOADDUAL ? RegState::Define : 0); 3128 for (const MachineOperand &MO : llvm::drop_begin(MI.operands())) 3129 MIB.add(MO); 3130 MIB.add(predOps(ARMCC::AL)); 3131 MIB.cloneMemRefs(MI); 3132 MI.eraseFromParent(); 3133 return true; 3134 } 3135 } 3136 } 3137 3138 bool ARMExpandPseudo::ExpandMBB(MachineBasicBlock &MBB) { 3139 bool Modified = false; 3140 3141 MachineBasicBlock::iterator MBBI = MBB.begin(), E = MBB.end(); 3142 while (MBBI != E) { 3143 MachineBasicBlock::iterator NMBBI = std::next(MBBI); 3144 Modified |= ExpandMI(MBB, MBBI, NMBBI); 3145 MBBI = NMBBI; 3146 } 3147 3148 return Modified; 3149 } 3150 3151 bool ARMExpandPseudo::runOnMachineFunction(MachineFunction &MF) { 3152 STI = &MF.getSubtarget<ARMSubtarget>(); 3153 TII = STI->getInstrInfo(); 3154 TRI = STI->getRegisterInfo(); 3155 AFI = MF.getInfo<ARMFunctionInfo>(); 3156 3157 LLVM_DEBUG(dbgs() << "********** ARM EXPAND PSEUDO INSTRUCTIONS **********\n" 3158 << "********** Function: " << MF.getName() << '\n'); 3159 3160 bool Modified = false; 3161 for (MachineBasicBlock &MBB : MF) 3162 Modified |= ExpandMBB(MBB); 3163 if (VerifyARMPseudo) 3164 MF.verify(this, "After expanding ARM pseudo instructions."); 3165 3166 LLVM_DEBUG(dbgs() << "***************************************************\n"); 3167 return Modified; 3168 } 3169 3170 /// createARMExpandPseudoPass - returns an instance of the pseudo instruction 3171 /// expansion pass. 3172 FunctionPass *llvm::createARMExpandPseudoPass() { 3173 return new ARMExpandPseudo(); 3174 } 3175