10b57cec5SDimitry Andric //===-- RISCVISelDAGToDAG.cpp - A dag to dag inst selector for RISCV ------===// 20b57cec5SDimitry Andric // 30b57cec5SDimitry Andric // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. 40b57cec5SDimitry Andric // See https://llvm.org/LICENSE.txt for license information. 50b57cec5SDimitry Andric // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception 60b57cec5SDimitry Andric // 70b57cec5SDimitry Andric //===----------------------------------------------------------------------===// 80b57cec5SDimitry Andric // 90b57cec5SDimitry Andric // This file defines an instruction selector for the RISCV target. 100b57cec5SDimitry Andric // 110b57cec5SDimitry Andric //===----------------------------------------------------------------------===// 120b57cec5SDimitry Andric 135ffd83dbSDimitry Andric #include "RISCVISelDAGToDAG.h" 140b57cec5SDimitry Andric #include "MCTargetDesc/RISCVMCTargetDesc.h" 15e8d8bef9SDimitry Andric #include "MCTargetDesc/RISCVMatInt.h" 16fe6060f1SDimitry Andric #include "RISCVISelLowering.h" 17fe6060f1SDimitry Andric #include "RISCVMachineFunctionInfo.h" 180b57cec5SDimitry Andric #include "llvm/CodeGen/MachineFrameInfo.h" 19e8d8bef9SDimitry Andric #include "llvm/IR/IntrinsicsRISCV.h" 205ffd83dbSDimitry Andric #include "llvm/Support/Alignment.h" 210b57cec5SDimitry Andric #include "llvm/Support/Debug.h" 22fe6060f1SDimitry Andric #include "llvm/Support/KnownBits.h" 230b57cec5SDimitry Andric #include "llvm/Support/MathExtras.h" 240b57cec5SDimitry Andric #include "llvm/Support/raw_ostream.h" 255ffd83dbSDimitry Andric 260b57cec5SDimitry Andric using namespace llvm; 270b57cec5SDimitry Andric 280b57cec5SDimitry Andric #define DEBUG_TYPE "riscv-isel" 290b57cec5SDimitry Andric 30fe6060f1SDimitry Andric namespace llvm { 31fe6060f1SDimitry Andric namespace RISCV { 32fe6060f1SDimitry Andric #define GET_RISCVVSSEGTable_IMPL 33fe6060f1SDimitry Andric #define GET_RISCVVLSEGTable_IMPL 34fe6060f1SDimitry Andric #define GET_RISCVVLXSEGTable_IMPL 35fe6060f1SDimitry Andric #define GET_RISCVVSXSEGTable_IMPL 36fe6060f1SDimitry Andric #define GET_RISCVVLETable_IMPL 37fe6060f1SDimitry Andric #define GET_RISCVVSETable_IMPL 38fe6060f1SDimitry Andric #define GET_RISCVVLXTable_IMPL 39fe6060f1SDimitry Andric #define GET_RISCVVSXTable_IMPL 40fe6060f1SDimitry Andric #include "RISCVGenSearchableTables.inc" 41fe6060f1SDimitry Andric } // namespace RISCV 42fe6060f1SDimitry Andric } // namespace llvm 43fe6060f1SDimitry Andric 44fe6060f1SDimitry Andric void RISCVDAGToDAGISel::PreprocessISelDAG() { 45fe6060f1SDimitry Andric for (SelectionDAG::allnodes_iterator I = CurDAG->allnodes_begin(), 46fe6060f1SDimitry Andric E = CurDAG->allnodes_end(); 47fe6060f1SDimitry Andric I != E;) { 48fe6060f1SDimitry Andric SDNode *N = &*I++; // Preincrement iterator to avoid invalidation issues. 49fe6060f1SDimitry Andric 50fe6060f1SDimitry Andric // Lower SPLAT_VECTOR_SPLIT_I64 to two scalar stores and a stride 0 vector 51fe6060f1SDimitry Andric // load. Done after lowering and combining so that we have a chance to 52fe6060f1SDimitry Andric // optimize this to VMV_V_X_VL when the upper bits aren't needed. 53fe6060f1SDimitry Andric if (N->getOpcode() != RISCVISD::SPLAT_VECTOR_SPLIT_I64_VL) 54fe6060f1SDimitry Andric continue; 55fe6060f1SDimitry Andric 56fe6060f1SDimitry Andric assert(N->getNumOperands() == 3 && "Unexpected number of operands"); 57fe6060f1SDimitry Andric MVT VT = N->getSimpleValueType(0); 58fe6060f1SDimitry Andric SDValue Lo = N->getOperand(0); 59fe6060f1SDimitry Andric SDValue Hi = N->getOperand(1); 60fe6060f1SDimitry Andric SDValue VL = N->getOperand(2); 61fe6060f1SDimitry Andric assert(VT.getVectorElementType() == MVT::i64 && VT.isScalableVector() && 62fe6060f1SDimitry Andric Lo.getValueType() == MVT::i32 && Hi.getValueType() == MVT::i32 && 63fe6060f1SDimitry Andric "Unexpected VTs!"); 64fe6060f1SDimitry Andric MachineFunction &MF = CurDAG->getMachineFunction(); 65fe6060f1SDimitry Andric RISCVMachineFunctionInfo *FuncInfo = MF.getInfo<RISCVMachineFunctionInfo>(); 66fe6060f1SDimitry Andric SDLoc DL(N); 67fe6060f1SDimitry Andric 68fe6060f1SDimitry Andric // We use the same frame index we use for moving two i32s into 64-bit FPR. 69fe6060f1SDimitry Andric // This is an analogous operation. 70fe6060f1SDimitry Andric int FI = FuncInfo->getMoveF64FrameIndex(MF); 71fe6060f1SDimitry Andric MachinePointerInfo MPI = MachinePointerInfo::getFixedStack(MF, FI); 72fe6060f1SDimitry Andric const TargetLowering &TLI = CurDAG->getTargetLoweringInfo(); 73fe6060f1SDimitry Andric SDValue StackSlot = 74fe6060f1SDimitry Andric CurDAG->getFrameIndex(FI, TLI.getPointerTy(CurDAG->getDataLayout())); 75fe6060f1SDimitry Andric 76fe6060f1SDimitry Andric SDValue Chain = CurDAG->getEntryNode(); 77fe6060f1SDimitry Andric Lo = CurDAG->getStore(Chain, DL, Lo, StackSlot, MPI, Align(8)); 78fe6060f1SDimitry Andric 79fe6060f1SDimitry Andric SDValue OffsetSlot = 80fe6060f1SDimitry Andric CurDAG->getMemBasePlusOffset(StackSlot, TypeSize::Fixed(4), DL); 81fe6060f1SDimitry Andric Hi = CurDAG->getStore(Chain, DL, Hi, OffsetSlot, MPI.getWithOffset(4), 82fe6060f1SDimitry Andric Align(8)); 83fe6060f1SDimitry Andric 84fe6060f1SDimitry Andric Chain = CurDAG->getNode(ISD::TokenFactor, DL, MVT::Other, Lo, Hi); 85fe6060f1SDimitry Andric 86fe6060f1SDimitry Andric SDVTList VTs = CurDAG->getVTList({VT, MVT::Other}); 87fe6060f1SDimitry Andric SDValue IntID = 88fe6060f1SDimitry Andric CurDAG->getTargetConstant(Intrinsic::riscv_vlse, DL, MVT::i64); 89fe6060f1SDimitry Andric SDValue Ops[] = {Chain, IntID, StackSlot, 90fe6060f1SDimitry Andric CurDAG->getRegister(RISCV::X0, MVT::i64), VL}; 91fe6060f1SDimitry Andric 92fe6060f1SDimitry Andric SDValue Result = CurDAG->getMemIntrinsicNode( 93fe6060f1SDimitry Andric ISD::INTRINSIC_W_CHAIN, DL, VTs, Ops, MVT::i64, MPI, Align(8), 94fe6060f1SDimitry Andric MachineMemOperand::MOLoad); 95fe6060f1SDimitry Andric 96fe6060f1SDimitry Andric // We're about to replace all uses of the SPLAT_VECTOR_SPLIT_I64 with the 97fe6060f1SDimitry Andric // vlse we created. This will cause general havok on the dag because 98fe6060f1SDimitry Andric // anything below the conversion could be folded into other existing nodes. 99fe6060f1SDimitry Andric // To avoid invalidating 'I', back it up to the convert node. 100fe6060f1SDimitry Andric --I; 101fe6060f1SDimitry Andric CurDAG->ReplaceAllUsesOfValueWith(SDValue(N, 0), Result); 102fe6060f1SDimitry Andric 103fe6060f1SDimitry Andric // Now that we did that, the node is dead. Increment the iterator to the 104fe6060f1SDimitry Andric // next node to process, then delete N. 105fe6060f1SDimitry Andric ++I; 106fe6060f1SDimitry Andric CurDAG->DeleteNode(N); 107fe6060f1SDimitry Andric } 108fe6060f1SDimitry Andric } 109fe6060f1SDimitry Andric 1100b57cec5SDimitry Andric void RISCVDAGToDAGISel::PostprocessISelDAG() { 111*349cc55cSDimitry Andric SelectionDAG::allnodes_iterator Position = CurDAG->allnodes_end(); 112*349cc55cSDimitry Andric 113*349cc55cSDimitry Andric bool MadeChange = false; 114*349cc55cSDimitry Andric while (Position != CurDAG->allnodes_begin()) { 115*349cc55cSDimitry Andric SDNode *N = &*--Position; 116*349cc55cSDimitry Andric // Skip dead nodes and any non-machine opcodes. 117*349cc55cSDimitry Andric if (N->use_empty() || !N->isMachineOpcode()) 118*349cc55cSDimitry Andric continue; 119*349cc55cSDimitry Andric 120*349cc55cSDimitry Andric MadeChange |= doPeepholeSExtW(N); 121*349cc55cSDimitry Andric MadeChange |= doPeepholeLoadStoreADDI(N); 122*349cc55cSDimitry Andric } 123*349cc55cSDimitry Andric 124*349cc55cSDimitry Andric if (MadeChange) 125*349cc55cSDimitry Andric CurDAG->RemoveDeadNodes(); 1260b57cec5SDimitry Andric } 1270b57cec5SDimitry Andric 1280b57cec5SDimitry Andric static SDNode *selectImm(SelectionDAG *CurDAG, const SDLoc &DL, int64_t Imm, 129fe6060f1SDimitry Andric const RISCVSubtarget &Subtarget) { 130fe6060f1SDimitry Andric MVT XLenVT = Subtarget.getXLenVT(); 131fe6060f1SDimitry Andric RISCVMatInt::InstSeq Seq = 132fe6060f1SDimitry Andric RISCVMatInt::generateInstSeq(Imm, Subtarget.getFeatureBits()); 1330b57cec5SDimitry Andric 1348bcb0991SDimitry Andric SDNode *Result = nullptr; 1350b57cec5SDimitry Andric SDValue SrcReg = CurDAG->getRegister(RISCV::X0, XLenVT); 1360b57cec5SDimitry Andric for (RISCVMatInt::Inst &Inst : Seq) { 1370b57cec5SDimitry Andric SDValue SDImm = CurDAG->getTargetConstant(Inst.Imm, DL, XLenVT); 1380b57cec5SDimitry Andric if (Inst.Opc == RISCV::LUI) 1390b57cec5SDimitry Andric Result = CurDAG->getMachineNode(RISCV::LUI, DL, XLenVT, SDImm); 140fe6060f1SDimitry Andric else if (Inst.Opc == RISCV::ADDUW) 141fe6060f1SDimitry Andric Result = CurDAG->getMachineNode(RISCV::ADDUW, DL, XLenVT, SrcReg, 142fe6060f1SDimitry Andric CurDAG->getRegister(RISCV::X0, XLenVT)); 143*349cc55cSDimitry Andric else if (Inst.Opc == RISCV::SH1ADD || Inst.Opc == RISCV::SH2ADD || 144*349cc55cSDimitry Andric Inst.Opc == RISCV::SH3ADD) 145*349cc55cSDimitry Andric Result = CurDAG->getMachineNode(Inst.Opc, DL, XLenVT, SrcReg, SrcReg); 1460b57cec5SDimitry Andric else 1470b57cec5SDimitry Andric Result = CurDAG->getMachineNode(Inst.Opc, DL, XLenVT, SrcReg, SDImm); 1480b57cec5SDimitry Andric 1490b57cec5SDimitry Andric // Only the first instruction has X0 as its source. 1500b57cec5SDimitry Andric SrcReg = SDValue(Result, 0); 1510b57cec5SDimitry Andric } 1520b57cec5SDimitry Andric 1530b57cec5SDimitry Andric return Result; 1540b57cec5SDimitry Andric } 1550b57cec5SDimitry Andric 156e8d8bef9SDimitry Andric static SDValue createTupleImpl(SelectionDAG &CurDAG, ArrayRef<SDValue> Regs, 157e8d8bef9SDimitry Andric unsigned RegClassID, unsigned SubReg0) { 158e8d8bef9SDimitry Andric assert(Regs.size() >= 2 && Regs.size() <= 8); 159e8d8bef9SDimitry Andric 160e8d8bef9SDimitry Andric SDLoc DL(Regs[0]); 161e8d8bef9SDimitry Andric SmallVector<SDValue, 8> Ops; 162e8d8bef9SDimitry Andric 163e8d8bef9SDimitry Andric Ops.push_back(CurDAG.getTargetConstant(RegClassID, DL, MVT::i32)); 164e8d8bef9SDimitry Andric 165e8d8bef9SDimitry Andric for (unsigned I = 0; I < Regs.size(); ++I) { 166e8d8bef9SDimitry Andric Ops.push_back(Regs[I]); 167e8d8bef9SDimitry Andric Ops.push_back(CurDAG.getTargetConstant(SubReg0 + I, DL, MVT::i32)); 168e8d8bef9SDimitry Andric } 169e8d8bef9SDimitry Andric SDNode *N = 170e8d8bef9SDimitry Andric CurDAG.getMachineNode(TargetOpcode::REG_SEQUENCE, DL, MVT::Untyped, Ops); 171e8d8bef9SDimitry Andric return SDValue(N, 0); 172e8d8bef9SDimitry Andric } 173e8d8bef9SDimitry Andric 174e8d8bef9SDimitry Andric static SDValue createM1Tuple(SelectionDAG &CurDAG, ArrayRef<SDValue> Regs, 175e8d8bef9SDimitry Andric unsigned NF) { 176e8d8bef9SDimitry Andric static const unsigned RegClassIDs[] = { 177e8d8bef9SDimitry Andric RISCV::VRN2M1RegClassID, RISCV::VRN3M1RegClassID, RISCV::VRN4M1RegClassID, 178e8d8bef9SDimitry Andric RISCV::VRN5M1RegClassID, RISCV::VRN6M1RegClassID, RISCV::VRN7M1RegClassID, 179e8d8bef9SDimitry Andric RISCV::VRN8M1RegClassID}; 180e8d8bef9SDimitry Andric 181e8d8bef9SDimitry Andric return createTupleImpl(CurDAG, Regs, RegClassIDs[NF - 2], RISCV::sub_vrm1_0); 182e8d8bef9SDimitry Andric } 183e8d8bef9SDimitry Andric 184e8d8bef9SDimitry Andric static SDValue createM2Tuple(SelectionDAG &CurDAG, ArrayRef<SDValue> Regs, 185e8d8bef9SDimitry Andric unsigned NF) { 186e8d8bef9SDimitry Andric static const unsigned RegClassIDs[] = {RISCV::VRN2M2RegClassID, 187e8d8bef9SDimitry Andric RISCV::VRN3M2RegClassID, 188e8d8bef9SDimitry Andric RISCV::VRN4M2RegClassID}; 189e8d8bef9SDimitry Andric 190e8d8bef9SDimitry Andric return createTupleImpl(CurDAG, Regs, RegClassIDs[NF - 2], RISCV::sub_vrm2_0); 191e8d8bef9SDimitry Andric } 192e8d8bef9SDimitry Andric 193e8d8bef9SDimitry Andric static SDValue createM4Tuple(SelectionDAG &CurDAG, ArrayRef<SDValue> Regs, 194e8d8bef9SDimitry Andric unsigned NF) { 195e8d8bef9SDimitry Andric return createTupleImpl(CurDAG, Regs, RISCV::VRN2M4RegClassID, 196e8d8bef9SDimitry Andric RISCV::sub_vrm4_0); 197e8d8bef9SDimitry Andric } 198e8d8bef9SDimitry Andric 199e8d8bef9SDimitry Andric static SDValue createTuple(SelectionDAG &CurDAG, ArrayRef<SDValue> Regs, 200fe6060f1SDimitry Andric unsigned NF, RISCVII::VLMUL LMUL) { 201e8d8bef9SDimitry Andric switch (LMUL) { 202e8d8bef9SDimitry Andric default: 203e8d8bef9SDimitry Andric llvm_unreachable("Invalid LMUL."); 204fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_F8: 205fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_F4: 206fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_F2: 207fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_1: 208e8d8bef9SDimitry Andric return createM1Tuple(CurDAG, Regs, NF); 209fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_2: 210e8d8bef9SDimitry Andric return createM2Tuple(CurDAG, Regs, NF); 211fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_4: 212e8d8bef9SDimitry Andric return createM4Tuple(CurDAG, Regs, NF); 213e8d8bef9SDimitry Andric } 214e8d8bef9SDimitry Andric } 215e8d8bef9SDimitry Andric 216fe6060f1SDimitry Andric void RISCVDAGToDAGISel::addVectorLoadStoreOperands( 217fe6060f1SDimitry Andric SDNode *Node, unsigned Log2SEW, const SDLoc &DL, unsigned CurOp, 218fe6060f1SDimitry Andric bool IsMasked, bool IsStridedOrIndexed, SmallVectorImpl<SDValue> &Operands, 219*349cc55cSDimitry Andric bool IsLoad, MVT *IndexVT) { 220fe6060f1SDimitry Andric SDValue Chain = Node->getOperand(0); 221fe6060f1SDimitry Andric SDValue Glue; 222fe6060f1SDimitry Andric 223fe6060f1SDimitry Andric SDValue Base; 224fe6060f1SDimitry Andric SelectBaseAddr(Node->getOperand(CurOp++), Base); 225fe6060f1SDimitry Andric Operands.push_back(Base); // Base pointer. 226fe6060f1SDimitry Andric 227fe6060f1SDimitry Andric if (IsStridedOrIndexed) { 228fe6060f1SDimitry Andric Operands.push_back(Node->getOperand(CurOp++)); // Index. 229fe6060f1SDimitry Andric if (IndexVT) 230fe6060f1SDimitry Andric *IndexVT = Operands.back()->getSimpleValueType(0); 231fe6060f1SDimitry Andric } 232fe6060f1SDimitry Andric 233fe6060f1SDimitry Andric if (IsMasked) { 234fe6060f1SDimitry Andric // Mask needs to be copied to V0. 235fe6060f1SDimitry Andric SDValue Mask = Node->getOperand(CurOp++); 236fe6060f1SDimitry Andric Chain = CurDAG->getCopyToReg(Chain, DL, RISCV::V0, Mask, SDValue()); 237fe6060f1SDimitry Andric Glue = Chain.getValue(1); 238fe6060f1SDimitry Andric Operands.push_back(CurDAG->getRegister(RISCV::V0, Mask.getValueType())); 239fe6060f1SDimitry Andric } 240fe6060f1SDimitry Andric SDValue VL; 241fe6060f1SDimitry Andric selectVLOp(Node->getOperand(CurOp++), VL); 242fe6060f1SDimitry Andric Operands.push_back(VL); 243fe6060f1SDimitry Andric 244fe6060f1SDimitry Andric MVT XLenVT = Subtarget->getXLenVT(); 245fe6060f1SDimitry Andric SDValue SEWOp = CurDAG->getTargetConstant(Log2SEW, DL, XLenVT); 246fe6060f1SDimitry Andric Operands.push_back(SEWOp); 247fe6060f1SDimitry Andric 248*349cc55cSDimitry Andric // Masked load has the tail policy argument. 249*349cc55cSDimitry Andric if (IsMasked && IsLoad) { 250*349cc55cSDimitry Andric // Policy must be a constant. 251*349cc55cSDimitry Andric uint64_t Policy = Node->getConstantOperandVal(CurOp++); 252*349cc55cSDimitry Andric SDValue PolicyOp = CurDAG->getTargetConstant(Policy, DL, XLenVT); 253*349cc55cSDimitry Andric Operands.push_back(PolicyOp); 254*349cc55cSDimitry Andric } 255*349cc55cSDimitry Andric 256fe6060f1SDimitry Andric Operands.push_back(Chain); // Chain. 257fe6060f1SDimitry Andric if (Glue) 258fe6060f1SDimitry Andric Operands.push_back(Glue); 259fe6060f1SDimitry Andric } 260fe6060f1SDimitry Andric 261fe6060f1SDimitry Andric void RISCVDAGToDAGISel::selectVLSEG(SDNode *Node, bool IsMasked, 262e8d8bef9SDimitry Andric bool IsStrided) { 263e8d8bef9SDimitry Andric SDLoc DL(Node); 264e8d8bef9SDimitry Andric unsigned NF = Node->getNumValues() - 1; 265fe6060f1SDimitry Andric MVT VT = Node->getSimpleValueType(0); 266fe6060f1SDimitry Andric unsigned Log2SEW = Log2_32(VT.getScalarSizeInBits()); 267fe6060f1SDimitry Andric RISCVII::VLMUL LMUL = RISCVTargetLowering::getLMUL(VT); 268e8d8bef9SDimitry Andric 269fe6060f1SDimitry Andric unsigned CurOp = 2; 270fe6060f1SDimitry Andric SmallVector<SDValue, 8> Operands; 271fe6060f1SDimitry Andric if (IsMasked) { 272fe6060f1SDimitry Andric SmallVector<SDValue, 8> Regs(Node->op_begin() + CurOp, 273fe6060f1SDimitry Andric Node->op_begin() + CurOp + NF); 274e8d8bef9SDimitry Andric SDValue MaskedOff = createTuple(*CurDAG, Regs, NF, LMUL); 275e8d8bef9SDimitry Andric Operands.push_back(MaskedOff); 276fe6060f1SDimitry Andric CurOp += NF; 277e8d8bef9SDimitry Andric } 278fe6060f1SDimitry Andric 279fe6060f1SDimitry Andric addVectorLoadStoreOperands(Node, Log2SEW, DL, CurOp, IsMasked, IsStrided, 280*349cc55cSDimitry Andric Operands, /*IsLoad=*/true); 281fe6060f1SDimitry Andric 282fe6060f1SDimitry Andric const RISCV::VLSEGPseudo *P = 283fe6060f1SDimitry Andric RISCV::getVLSEGPseudo(NF, IsMasked, IsStrided, /*FF*/ false, Log2SEW, 284fe6060f1SDimitry Andric static_cast<unsigned>(LMUL)); 285fe6060f1SDimitry Andric MachineSDNode *Load = 286e8d8bef9SDimitry Andric CurDAG->getMachineNode(P->Pseudo, DL, MVT::Untyped, MVT::Other, Operands); 287fe6060f1SDimitry Andric 288fe6060f1SDimitry Andric if (auto *MemOp = dyn_cast<MemSDNode>(Node)) 289fe6060f1SDimitry Andric CurDAG->setNodeMemRefs(Load, {MemOp->getMemOperand()}); 290fe6060f1SDimitry Andric 291e8d8bef9SDimitry Andric SDValue SuperReg = SDValue(Load, 0); 292fe6060f1SDimitry Andric for (unsigned I = 0; I < NF; ++I) { 293fe6060f1SDimitry Andric unsigned SubRegIdx = RISCVTargetLowering::getSubregIndexByMVT(VT, I); 294e8d8bef9SDimitry Andric ReplaceUses(SDValue(Node, I), 295fe6060f1SDimitry Andric CurDAG->getTargetExtractSubreg(SubRegIdx, DL, VT, SuperReg)); 296fe6060f1SDimitry Andric } 297e8d8bef9SDimitry Andric 298e8d8bef9SDimitry Andric ReplaceUses(SDValue(Node, NF), SDValue(Load, 1)); 299e8d8bef9SDimitry Andric CurDAG->RemoveDeadNode(Node); 300e8d8bef9SDimitry Andric } 301e8d8bef9SDimitry Andric 302fe6060f1SDimitry Andric void RISCVDAGToDAGISel::selectVLSEGFF(SDNode *Node, bool IsMasked) { 303e8d8bef9SDimitry Andric SDLoc DL(Node); 304fe6060f1SDimitry Andric unsigned NF = Node->getNumValues() - 2; // Do not count VL and Chain. 305fe6060f1SDimitry Andric MVT VT = Node->getSimpleValueType(0); 306e8d8bef9SDimitry Andric MVT XLenVT = Subtarget->getXLenVT(); 307fe6060f1SDimitry Andric unsigned Log2SEW = Log2_32(VT.getScalarSizeInBits()); 308fe6060f1SDimitry Andric RISCVII::VLMUL LMUL = RISCVTargetLowering::getLMUL(VT); 309e8d8bef9SDimitry Andric 310fe6060f1SDimitry Andric unsigned CurOp = 2; 311e8d8bef9SDimitry Andric SmallVector<SDValue, 7> Operands; 312fe6060f1SDimitry Andric if (IsMasked) { 313fe6060f1SDimitry Andric SmallVector<SDValue, 8> Regs(Node->op_begin() + CurOp, 314fe6060f1SDimitry Andric Node->op_begin() + CurOp + NF); 315e8d8bef9SDimitry Andric SDValue MaskedOff = createTuple(*CurDAG, Regs, NF, LMUL); 316fe6060f1SDimitry Andric Operands.push_back(MaskedOff); 317fe6060f1SDimitry Andric CurOp += NF; 318fe6060f1SDimitry Andric } 319e8d8bef9SDimitry Andric 320fe6060f1SDimitry Andric addVectorLoadStoreOperands(Node, Log2SEW, DL, CurOp, IsMasked, 321*349cc55cSDimitry Andric /*IsStridedOrIndexed*/ false, Operands, 322*349cc55cSDimitry Andric /*IsLoad=*/true); 323fe6060f1SDimitry Andric 324fe6060f1SDimitry Andric const RISCV::VLSEGPseudo *P = 325fe6060f1SDimitry Andric RISCV::getVLSEGPseudo(NF, IsMasked, /*Strided*/ false, /*FF*/ true, 326fe6060f1SDimitry Andric Log2SEW, static_cast<unsigned>(LMUL)); 327fe6060f1SDimitry Andric MachineSDNode *Load = CurDAG->getMachineNode(P->Pseudo, DL, MVT::Untyped, 328fe6060f1SDimitry Andric MVT::Other, MVT::Glue, Operands); 329fe6060f1SDimitry Andric SDNode *ReadVL = CurDAG->getMachineNode(RISCV::PseudoReadVL, DL, XLenVT, 330fe6060f1SDimitry Andric /*Glue*/ SDValue(Load, 2)); 331fe6060f1SDimitry Andric 332fe6060f1SDimitry Andric if (auto *MemOp = dyn_cast<MemSDNode>(Node)) 333fe6060f1SDimitry Andric CurDAG->setNodeMemRefs(Load, {MemOp->getMemOperand()}); 334fe6060f1SDimitry Andric 335e8d8bef9SDimitry Andric SDValue SuperReg = SDValue(Load, 0); 336fe6060f1SDimitry Andric for (unsigned I = 0; I < NF; ++I) { 337fe6060f1SDimitry Andric unsigned SubRegIdx = RISCVTargetLowering::getSubregIndexByMVT(VT, I); 338e8d8bef9SDimitry Andric ReplaceUses(SDValue(Node, I), 339fe6060f1SDimitry Andric CurDAG->getTargetExtractSubreg(SubRegIdx, DL, VT, SuperReg)); 340fe6060f1SDimitry Andric } 341fe6060f1SDimitry Andric 342fe6060f1SDimitry Andric ReplaceUses(SDValue(Node, NF), SDValue(ReadVL, 0)); // VL 343fe6060f1SDimitry Andric ReplaceUses(SDValue(Node, NF + 1), SDValue(Load, 1)); // Chain 344fe6060f1SDimitry Andric CurDAG->RemoveDeadNode(Node); 345fe6060f1SDimitry Andric } 346fe6060f1SDimitry Andric 347fe6060f1SDimitry Andric void RISCVDAGToDAGISel::selectVLXSEG(SDNode *Node, bool IsMasked, 348fe6060f1SDimitry Andric bool IsOrdered) { 349fe6060f1SDimitry Andric SDLoc DL(Node); 350fe6060f1SDimitry Andric unsigned NF = Node->getNumValues() - 1; 351fe6060f1SDimitry Andric MVT VT = Node->getSimpleValueType(0); 352fe6060f1SDimitry Andric unsigned Log2SEW = Log2_32(VT.getScalarSizeInBits()); 353fe6060f1SDimitry Andric RISCVII::VLMUL LMUL = RISCVTargetLowering::getLMUL(VT); 354fe6060f1SDimitry Andric 355fe6060f1SDimitry Andric unsigned CurOp = 2; 356fe6060f1SDimitry Andric SmallVector<SDValue, 8> Operands; 357fe6060f1SDimitry Andric if (IsMasked) { 358fe6060f1SDimitry Andric SmallVector<SDValue, 8> Regs(Node->op_begin() + CurOp, 359fe6060f1SDimitry Andric Node->op_begin() + CurOp + NF); 360fe6060f1SDimitry Andric SDValue MaskedOff = createTuple(*CurDAG, Regs, NF, LMUL); 361fe6060f1SDimitry Andric Operands.push_back(MaskedOff); 362fe6060f1SDimitry Andric CurOp += NF; 363fe6060f1SDimitry Andric } 364fe6060f1SDimitry Andric 365fe6060f1SDimitry Andric MVT IndexVT; 366fe6060f1SDimitry Andric addVectorLoadStoreOperands(Node, Log2SEW, DL, CurOp, IsMasked, 367*349cc55cSDimitry Andric /*IsStridedOrIndexed*/ true, Operands, 368*349cc55cSDimitry Andric /*IsLoad=*/true, &IndexVT); 369fe6060f1SDimitry Andric 370fe6060f1SDimitry Andric assert(VT.getVectorElementCount() == IndexVT.getVectorElementCount() && 371fe6060f1SDimitry Andric "Element count mismatch"); 372fe6060f1SDimitry Andric 373fe6060f1SDimitry Andric RISCVII::VLMUL IndexLMUL = RISCVTargetLowering::getLMUL(IndexVT); 374fe6060f1SDimitry Andric unsigned IndexLog2EEW = Log2_32(IndexVT.getScalarSizeInBits()); 375fe6060f1SDimitry Andric const RISCV::VLXSEGPseudo *P = RISCV::getVLXSEGPseudo( 376fe6060f1SDimitry Andric NF, IsMasked, IsOrdered, IndexLog2EEW, static_cast<unsigned>(LMUL), 377fe6060f1SDimitry Andric static_cast<unsigned>(IndexLMUL)); 378fe6060f1SDimitry Andric MachineSDNode *Load = 379fe6060f1SDimitry Andric CurDAG->getMachineNode(P->Pseudo, DL, MVT::Untyped, MVT::Other, Operands); 380fe6060f1SDimitry Andric 381fe6060f1SDimitry Andric if (auto *MemOp = dyn_cast<MemSDNode>(Node)) 382fe6060f1SDimitry Andric CurDAG->setNodeMemRefs(Load, {MemOp->getMemOperand()}); 383fe6060f1SDimitry Andric 384fe6060f1SDimitry Andric SDValue SuperReg = SDValue(Load, 0); 385fe6060f1SDimitry Andric for (unsigned I = 0; I < NF; ++I) { 386fe6060f1SDimitry Andric unsigned SubRegIdx = RISCVTargetLowering::getSubregIndexByMVT(VT, I); 387fe6060f1SDimitry Andric ReplaceUses(SDValue(Node, I), 388fe6060f1SDimitry Andric CurDAG->getTargetExtractSubreg(SubRegIdx, DL, VT, SuperReg)); 389fe6060f1SDimitry Andric } 390e8d8bef9SDimitry Andric 391e8d8bef9SDimitry Andric ReplaceUses(SDValue(Node, NF), SDValue(Load, 1)); 392e8d8bef9SDimitry Andric CurDAG->RemoveDeadNode(Node); 393e8d8bef9SDimitry Andric } 394e8d8bef9SDimitry Andric 395fe6060f1SDimitry Andric void RISCVDAGToDAGISel::selectVSSEG(SDNode *Node, bool IsMasked, 396e8d8bef9SDimitry Andric bool IsStrided) { 397e8d8bef9SDimitry Andric SDLoc DL(Node); 398e8d8bef9SDimitry Andric unsigned NF = Node->getNumOperands() - 4; 399e8d8bef9SDimitry Andric if (IsStrided) 400e8d8bef9SDimitry Andric NF--; 401fe6060f1SDimitry Andric if (IsMasked) 402e8d8bef9SDimitry Andric NF--; 403fe6060f1SDimitry Andric MVT VT = Node->getOperand(2)->getSimpleValueType(0); 404fe6060f1SDimitry Andric unsigned Log2SEW = Log2_32(VT.getScalarSizeInBits()); 405fe6060f1SDimitry Andric RISCVII::VLMUL LMUL = RISCVTargetLowering::getLMUL(VT); 406e8d8bef9SDimitry Andric SmallVector<SDValue, 8> Regs(Node->op_begin() + 2, Node->op_begin() + 2 + NF); 407e8d8bef9SDimitry Andric SDValue StoreVal = createTuple(*CurDAG, Regs, NF, LMUL); 408fe6060f1SDimitry Andric 409fe6060f1SDimitry Andric SmallVector<SDValue, 8> Operands; 410e8d8bef9SDimitry Andric Operands.push_back(StoreVal); 411fe6060f1SDimitry Andric unsigned CurOp = 2 + NF; 412fe6060f1SDimitry Andric 413fe6060f1SDimitry Andric addVectorLoadStoreOperands(Node, Log2SEW, DL, CurOp, IsMasked, IsStrided, 414fe6060f1SDimitry Andric Operands); 415fe6060f1SDimitry Andric 416fe6060f1SDimitry Andric const RISCV::VSSEGPseudo *P = RISCV::getVSSEGPseudo( 417fe6060f1SDimitry Andric NF, IsMasked, IsStrided, Log2SEW, static_cast<unsigned>(LMUL)); 418fe6060f1SDimitry Andric MachineSDNode *Store = 419e8d8bef9SDimitry Andric CurDAG->getMachineNode(P->Pseudo, DL, Node->getValueType(0), Operands); 420fe6060f1SDimitry Andric 421fe6060f1SDimitry Andric if (auto *MemOp = dyn_cast<MemSDNode>(Node)) 422fe6060f1SDimitry Andric CurDAG->setNodeMemRefs(Store, {MemOp->getMemOperand()}); 423fe6060f1SDimitry Andric 424e8d8bef9SDimitry Andric ReplaceNode(Node, Store); 425e8d8bef9SDimitry Andric } 426e8d8bef9SDimitry Andric 427fe6060f1SDimitry Andric void RISCVDAGToDAGISel::selectVSXSEG(SDNode *Node, bool IsMasked, 428fe6060f1SDimitry Andric bool IsOrdered) { 429e8d8bef9SDimitry Andric SDLoc DL(Node); 430e8d8bef9SDimitry Andric unsigned NF = Node->getNumOperands() - 5; 431fe6060f1SDimitry Andric if (IsMasked) 432fe6060f1SDimitry Andric --NF; 433fe6060f1SDimitry Andric MVT VT = Node->getOperand(2)->getSimpleValueType(0); 434fe6060f1SDimitry Andric unsigned Log2SEW = Log2_32(VT.getScalarSizeInBits()); 435fe6060f1SDimitry Andric RISCVII::VLMUL LMUL = RISCVTargetLowering::getLMUL(VT); 436e8d8bef9SDimitry Andric SmallVector<SDValue, 8> Regs(Node->op_begin() + 2, Node->op_begin() + 2 + NF); 437e8d8bef9SDimitry Andric SDValue StoreVal = createTuple(*CurDAG, Regs, NF, LMUL); 438e8d8bef9SDimitry Andric 439fe6060f1SDimitry Andric SmallVector<SDValue, 8> Operands; 440fe6060f1SDimitry Andric Operands.push_back(StoreVal); 441fe6060f1SDimitry Andric unsigned CurOp = 2 + NF; 442fe6060f1SDimitry Andric 443fe6060f1SDimitry Andric MVT IndexVT; 444fe6060f1SDimitry Andric addVectorLoadStoreOperands(Node, Log2SEW, DL, CurOp, IsMasked, 445*349cc55cSDimitry Andric /*IsStridedOrIndexed*/ true, Operands, 446*349cc55cSDimitry Andric /*IsLoad=*/false, &IndexVT); 447fe6060f1SDimitry Andric 448fe6060f1SDimitry Andric assert(VT.getVectorElementCount() == IndexVT.getVectorElementCount() && 449fe6060f1SDimitry Andric "Element count mismatch"); 450fe6060f1SDimitry Andric 451fe6060f1SDimitry Andric RISCVII::VLMUL IndexLMUL = RISCVTargetLowering::getLMUL(IndexVT); 452fe6060f1SDimitry Andric unsigned IndexLog2EEW = Log2_32(IndexVT.getScalarSizeInBits()); 453fe6060f1SDimitry Andric const RISCV::VSXSEGPseudo *P = RISCV::getVSXSEGPseudo( 454fe6060f1SDimitry Andric NF, IsMasked, IsOrdered, IndexLog2EEW, static_cast<unsigned>(LMUL), 455e8d8bef9SDimitry Andric static_cast<unsigned>(IndexLMUL)); 456fe6060f1SDimitry Andric MachineSDNode *Store = 457e8d8bef9SDimitry Andric CurDAG->getMachineNode(P->Pseudo, DL, Node->getValueType(0), Operands); 458fe6060f1SDimitry Andric 459fe6060f1SDimitry Andric if (auto *MemOp = dyn_cast<MemSDNode>(Node)) 460fe6060f1SDimitry Andric CurDAG->setNodeMemRefs(Store, {MemOp->getMemOperand()}); 461fe6060f1SDimitry Andric 462e8d8bef9SDimitry Andric ReplaceNode(Node, Store); 463e8d8bef9SDimitry Andric } 464e8d8bef9SDimitry Andric 4650b57cec5SDimitry Andric 4660b57cec5SDimitry Andric void RISCVDAGToDAGISel::Select(SDNode *Node) { 4670b57cec5SDimitry Andric // If we have a custom node, we have already selected. 4680b57cec5SDimitry Andric if (Node->isMachineOpcode()) { 4690b57cec5SDimitry Andric LLVM_DEBUG(dbgs() << "== "; Node->dump(CurDAG); dbgs() << "\n"); 4700b57cec5SDimitry Andric Node->setNodeId(-1); 4710b57cec5SDimitry Andric return; 4720b57cec5SDimitry Andric } 4730b57cec5SDimitry Andric 4740b57cec5SDimitry Andric // Instruction Selection not handled by the auto-generated tablegen selection 4750b57cec5SDimitry Andric // should be handled here. 4760b57cec5SDimitry Andric unsigned Opcode = Node->getOpcode(); 4770b57cec5SDimitry Andric MVT XLenVT = Subtarget->getXLenVT(); 4780b57cec5SDimitry Andric SDLoc DL(Node); 479fe6060f1SDimitry Andric MVT VT = Node->getSimpleValueType(0); 4800b57cec5SDimitry Andric 4810b57cec5SDimitry Andric switch (Opcode) { 4820b57cec5SDimitry Andric case ISD::Constant: { 483fe6060f1SDimitry Andric auto *ConstNode = cast<ConstantSDNode>(Node); 484*349cc55cSDimitry Andric if (VT == XLenVT && ConstNode->isZero()) { 485e8d8bef9SDimitry Andric SDValue New = 486e8d8bef9SDimitry Andric CurDAG->getCopyFromReg(CurDAG->getEntryNode(), DL, RISCV::X0, XLenVT); 4870b57cec5SDimitry Andric ReplaceNode(Node, New.getNode()); 4880b57cec5SDimitry Andric return; 4890b57cec5SDimitry Andric } 490*349cc55cSDimitry Andric int64_t Imm = ConstNode->getSExtValue(); 491*349cc55cSDimitry Andric // If the upper XLen-16 bits are not used, try to convert this to a simm12 492*349cc55cSDimitry Andric // by sign extending bit 15. 493*349cc55cSDimitry Andric if (isUInt<16>(Imm) && isInt<12>(SignExtend64(Imm, 16)) && 494*349cc55cSDimitry Andric hasAllHUsers(Node)) 495*349cc55cSDimitry Andric Imm = SignExtend64(Imm, 16); 496*349cc55cSDimitry Andric // If the upper 32-bits are not used try to convert this into a simm32 by 497*349cc55cSDimitry Andric // sign extending bit 32. 498*349cc55cSDimitry Andric if (!isInt<32>(Imm) && isUInt<32>(Imm) && hasAllWUsers(Node)) 499*349cc55cSDimitry Andric Imm = SignExtend64(Imm, 32); 500*349cc55cSDimitry Andric 501*349cc55cSDimitry Andric ReplaceNode(Node, selectImm(CurDAG, DL, Imm, *Subtarget)); 5020b57cec5SDimitry Andric return; 5030b57cec5SDimitry Andric } 5040b57cec5SDimitry Andric case ISD::FrameIndex: { 5050b57cec5SDimitry Andric SDValue Imm = CurDAG->getTargetConstant(0, DL, XLenVT); 5060b57cec5SDimitry Andric int FI = cast<FrameIndexSDNode>(Node)->getIndex(); 5070b57cec5SDimitry Andric SDValue TFI = CurDAG->getTargetFrameIndex(FI, VT); 5080b57cec5SDimitry Andric ReplaceNode(Node, CurDAG->getMachineNode(RISCV::ADDI, DL, VT, TFI, Imm)); 5090b57cec5SDimitry Andric return; 5100b57cec5SDimitry Andric } 511fe6060f1SDimitry Andric case ISD::SRL: { 512fe6060f1SDimitry Andric // We don't need this transform if zext.h is supported. 513fe6060f1SDimitry Andric if (Subtarget->hasStdExtZbb() || Subtarget->hasStdExtZbp()) 514fe6060f1SDimitry Andric break; 515fe6060f1SDimitry Andric // Optimize (srl (and X, 0xffff), C) -> 516fe6060f1SDimitry Andric // (srli (slli X, (XLen-16), (XLen-16) + C) 517fe6060f1SDimitry Andric // Taking into account that the 0xffff may have had lower bits unset by 518fe6060f1SDimitry Andric // SimplifyDemandedBits. This avoids materializing the 0xffff immediate. 519fe6060f1SDimitry Andric // This pattern occurs when type legalizing i16 right shifts. 520fe6060f1SDimitry Andric // FIXME: This could be extended to other AND masks. 521fe6060f1SDimitry Andric auto *N1C = dyn_cast<ConstantSDNode>(Node->getOperand(1)); 522fe6060f1SDimitry Andric if (N1C) { 523fe6060f1SDimitry Andric uint64_t ShAmt = N1C->getZExtValue(); 524fe6060f1SDimitry Andric SDValue N0 = Node->getOperand(0); 525fe6060f1SDimitry Andric if (ShAmt < 16 && N0.getOpcode() == ISD::AND && N0.hasOneUse() && 526fe6060f1SDimitry Andric isa<ConstantSDNode>(N0.getOperand(1))) { 527fe6060f1SDimitry Andric uint64_t Mask = N0.getConstantOperandVal(1); 528fe6060f1SDimitry Andric Mask |= maskTrailingOnes<uint64_t>(ShAmt); 529fe6060f1SDimitry Andric if (Mask == 0xffff) { 530fe6060f1SDimitry Andric unsigned LShAmt = Subtarget->getXLen() - 16; 531fe6060f1SDimitry Andric SDNode *SLLI = 532fe6060f1SDimitry Andric CurDAG->getMachineNode(RISCV::SLLI, DL, VT, N0->getOperand(0), 533fe6060f1SDimitry Andric CurDAG->getTargetConstant(LShAmt, DL, VT)); 534fe6060f1SDimitry Andric SDNode *SRLI = CurDAG->getMachineNode( 535fe6060f1SDimitry Andric RISCV::SRLI, DL, VT, SDValue(SLLI, 0), 536fe6060f1SDimitry Andric CurDAG->getTargetConstant(LShAmt + ShAmt, DL, VT)); 537fe6060f1SDimitry Andric ReplaceNode(Node, SRLI); 538fe6060f1SDimitry Andric return; 539fe6060f1SDimitry Andric } 540fe6060f1SDimitry Andric } 541fe6060f1SDimitry Andric } 542fe6060f1SDimitry Andric 543fe6060f1SDimitry Andric break; 544fe6060f1SDimitry Andric } 545fe6060f1SDimitry Andric case ISD::AND: { 546fe6060f1SDimitry Andric auto *N1C = dyn_cast<ConstantSDNode>(Node->getOperand(1)); 547fe6060f1SDimitry Andric if (!N1C) 548fe6060f1SDimitry Andric break; 549fe6060f1SDimitry Andric 550fe6060f1SDimitry Andric SDValue N0 = Node->getOperand(0); 551fe6060f1SDimitry Andric 552fe6060f1SDimitry Andric bool LeftShift = N0.getOpcode() == ISD::SHL; 553fe6060f1SDimitry Andric if (!LeftShift && N0.getOpcode() != ISD::SRL) 554fe6060f1SDimitry Andric break; 555fe6060f1SDimitry Andric 556fe6060f1SDimitry Andric auto *C = dyn_cast<ConstantSDNode>(N0.getOperand(1)); 557fe6060f1SDimitry Andric if (!C) 558fe6060f1SDimitry Andric break; 559fe6060f1SDimitry Andric uint64_t C2 = C->getZExtValue(); 560fe6060f1SDimitry Andric unsigned XLen = Subtarget->getXLen(); 561fe6060f1SDimitry Andric if (!C2 || C2 >= XLen) 562fe6060f1SDimitry Andric break; 563fe6060f1SDimitry Andric 564fe6060f1SDimitry Andric uint64_t C1 = N1C->getZExtValue(); 565fe6060f1SDimitry Andric 566fe6060f1SDimitry Andric // Keep track of whether this is a andi, zext.h, or zext.w. 567fe6060f1SDimitry Andric bool ZExtOrANDI = isInt<12>(N1C->getSExtValue()); 568fe6060f1SDimitry Andric if (C1 == UINT64_C(0xFFFF) && 569fe6060f1SDimitry Andric (Subtarget->hasStdExtZbb() || Subtarget->hasStdExtZbp())) 570fe6060f1SDimitry Andric ZExtOrANDI = true; 571fe6060f1SDimitry Andric if (C1 == UINT64_C(0xFFFFFFFF) && Subtarget->hasStdExtZba()) 572fe6060f1SDimitry Andric ZExtOrANDI = true; 573fe6060f1SDimitry Andric 574fe6060f1SDimitry Andric // Clear irrelevant bits in the mask. 575fe6060f1SDimitry Andric if (LeftShift) 576fe6060f1SDimitry Andric C1 &= maskTrailingZeros<uint64_t>(C2); 577fe6060f1SDimitry Andric else 578fe6060f1SDimitry Andric C1 &= maskTrailingOnes<uint64_t>(XLen - C2); 579fe6060f1SDimitry Andric 580fe6060f1SDimitry Andric // Some transforms should only be done if the shift has a single use or 581fe6060f1SDimitry Andric // the AND would become (srli (slli X, 32), 32) 582fe6060f1SDimitry Andric bool OneUseOrZExtW = N0.hasOneUse() || C1 == UINT64_C(0xFFFFFFFF); 583fe6060f1SDimitry Andric 584fe6060f1SDimitry Andric SDValue X = N0.getOperand(0); 585fe6060f1SDimitry Andric 586fe6060f1SDimitry Andric // Turn (and (srl x, c2) c1) -> (srli (slli x, c3-c2), c3) if c1 is a mask 587fe6060f1SDimitry Andric // with c3 leading zeros. 588fe6060f1SDimitry Andric if (!LeftShift && isMask_64(C1)) { 589fe6060f1SDimitry Andric uint64_t C3 = XLen - (64 - countLeadingZeros(C1)); 590fe6060f1SDimitry Andric if (C2 < C3) { 591fe6060f1SDimitry Andric // If the number of leading zeros is C2+32 this can be SRLIW. 592fe6060f1SDimitry Andric if (C2 + 32 == C3) { 593fe6060f1SDimitry Andric SDNode *SRLIW = 594fe6060f1SDimitry Andric CurDAG->getMachineNode(RISCV::SRLIW, DL, XLenVT, X, 595fe6060f1SDimitry Andric CurDAG->getTargetConstant(C2, DL, XLenVT)); 596fe6060f1SDimitry Andric ReplaceNode(Node, SRLIW); 597fe6060f1SDimitry Andric return; 598fe6060f1SDimitry Andric } 599fe6060f1SDimitry Andric 600fe6060f1SDimitry Andric // (and (srl (sexti32 Y), c2), c1) -> (srliw (sraiw Y, 31), c3 - 32) if 601fe6060f1SDimitry Andric // c1 is a mask with c3 leading zeros and c2 >= 32 and c3-c2==1. 602fe6060f1SDimitry Andric // 603fe6060f1SDimitry Andric // This pattern occurs when (i32 (srl (sra 31), c3 - 32)) is type 604fe6060f1SDimitry Andric // legalized and goes through DAG combine. 605fe6060f1SDimitry Andric SDValue Y; 606fe6060f1SDimitry Andric if (C2 >= 32 && (C3 - C2) == 1 && N0.hasOneUse() && 607fe6060f1SDimitry Andric selectSExti32(X, Y)) { 608fe6060f1SDimitry Andric SDNode *SRAIW = 609fe6060f1SDimitry Andric CurDAG->getMachineNode(RISCV::SRAIW, DL, XLenVT, Y, 610fe6060f1SDimitry Andric CurDAG->getTargetConstant(31, DL, XLenVT)); 611fe6060f1SDimitry Andric SDNode *SRLIW = CurDAG->getMachineNode( 612fe6060f1SDimitry Andric RISCV::SRLIW, DL, XLenVT, SDValue(SRAIW, 0), 613fe6060f1SDimitry Andric CurDAG->getTargetConstant(C3 - 32, DL, XLenVT)); 614fe6060f1SDimitry Andric ReplaceNode(Node, SRLIW); 615fe6060f1SDimitry Andric return; 616fe6060f1SDimitry Andric } 617fe6060f1SDimitry Andric 618fe6060f1SDimitry Andric // (srli (slli x, c3-c2), c3). 619fe6060f1SDimitry Andric if (OneUseOrZExtW && !ZExtOrANDI) { 620fe6060f1SDimitry Andric SDNode *SLLI = CurDAG->getMachineNode( 621fe6060f1SDimitry Andric RISCV::SLLI, DL, XLenVT, X, 622fe6060f1SDimitry Andric CurDAG->getTargetConstant(C3 - C2, DL, XLenVT)); 623fe6060f1SDimitry Andric SDNode *SRLI = 624fe6060f1SDimitry Andric CurDAG->getMachineNode(RISCV::SRLI, DL, XLenVT, SDValue(SLLI, 0), 625fe6060f1SDimitry Andric CurDAG->getTargetConstant(C3, DL, XLenVT)); 626fe6060f1SDimitry Andric ReplaceNode(Node, SRLI); 627fe6060f1SDimitry Andric return; 628fe6060f1SDimitry Andric } 629fe6060f1SDimitry Andric } 630fe6060f1SDimitry Andric } 631fe6060f1SDimitry Andric 632*349cc55cSDimitry Andric // Turn (and (shl x, c2), c1) -> (srli (slli c2+c3), c3) if c1 is a mask 633fe6060f1SDimitry Andric // shifted by c2 bits with c3 leading zeros. 634fe6060f1SDimitry Andric if (LeftShift && isShiftedMask_64(C1)) { 635fe6060f1SDimitry Andric uint64_t C3 = XLen - (64 - countLeadingZeros(C1)); 636fe6060f1SDimitry Andric 637fe6060f1SDimitry Andric if (C2 + C3 < XLen && 638fe6060f1SDimitry Andric C1 == (maskTrailingOnes<uint64_t>(XLen - (C2 + C3)) << C2)) { 639fe6060f1SDimitry Andric // Use slli.uw when possible. 640fe6060f1SDimitry Andric if ((XLen - (C2 + C3)) == 32 && Subtarget->hasStdExtZba()) { 641fe6060f1SDimitry Andric SDNode *SLLIUW = 642fe6060f1SDimitry Andric CurDAG->getMachineNode(RISCV::SLLIUW, DL, XLenVT, X, 643fe6060f1SDimitry Andric CurDAG->getTargetConstant(C2, DL, XLenVT)); 644fe6060f1SDimitry Andric ReplaceNode(Node, SLLIUW); 645fe6060f1SDimitry Andric return; 646fe6060f1SDimitry Andric } 647fe6060f1SDimitry Andric 648fe6060f1SDimitry Andric // (srli (slli c2+c3), c3) 649fe6060f1SDimitry Andric if (OneUseOrZExtW && !ZExtOrANDI) { 650fe6060f1SDimitry Andric SDNode *SLLI = CurDAG->getMachineNode( 651fe6060f1SDimitry Andric RISCV::SLLI, DL, XLenVT, X, 652fe6060f1SDimitry Andric CurDAG->getTargetConstant(C2 + C3, DL, XLenVT)); 653fe6060f1SDimitry Andric SDNode *SRLI = 654fe6060f1SDimitry Andric CurDAG->getMachineNode(RISCV::SRLI, DL, XLenVT, SDValue(SLLI, 0), 655fe6060f1SDimitry Andric CurDAG->getTargetConstant(C3, DL, XLenVT)); 656fe6060f1SDimitry Andric ReplaceNode(Node, SRLI); 657fe6060f1SDimitry Andric return; 658fe6060f1SDimitry Andric } 659fe6060f1SDimitry Andric } 660fe6060f1SDimitry Andric } 661fe6060f1SDimitry Andric 662*349cc55cSDimitry Andric // Turn (and (shr x, c2), c1) -> (slli (srli x, c2+c3), c3) if c1 is a 663*349cc55cSDimitry Andric // shifted mask with c2 leading zeros and c3 trailing zeros. 664*349cc55cSDimitry Andric if (!LeftShift && isShiftedMask_64(C1)) { 665*349cc55cSDimitry Andric uint64_t Leading = XLen - (64 - countLeadingZeros(C1)); 666*349cc55cSDimitry Andric uint64_t C3 = countTrailingZeros(C1); 667*349cc55cSDimitry Andric if (Leading == C2 && C2 + C3 < XLen && OneUseOrZExtW && !ZExtOrANDI) { 668*349cc55cSDimitry Andric SDNode *SRLI = CurDAG->getMachineNode( 669*349cc55cSDimitry Andric RISCV::SRLI, DL, XLenVT, X, 670*349cc55cSDimitry Andric CurDAG->getTargetConstant(C2 + C3, DL, XLenVT)); 671*349cc55cSDimitry Andric SDNode *SLLI = 672*349cc55cSDimitry Andric CurDAG->getMachineNode(RISCV::SLLI, DL, XLenVT, SDValue(SRLI, 0), 673*349cc55cSDimitry Andric CurDAG->getTargetConstant(C3, DL, XLenVT)); 674*349cc55cSDimitry Andric ReplaceNode(Node, SLLI); 675*349cc55cSDimitry Andric return; 676*349cc55cSDimitry Andric } 677*349cc55cSDimitry Andric // If the leading zero count is C2+32, we can use SRLIW instead of SRLI. 678*349cc55cSDimitry Andric if (Leading > 32 && (Leading - 32) == C2 && C2 + C3 < 32 && 679*349cc55cSDimitry Andric OneUseOrZExtW && !ZExtOrANDI) { 680*349cc55cSDimitry Andric SDNode *SRLIW = CurDAG->getMachineNode( 681*349cc55cSDimitry Andric RISCV::SRLIW, DL, XLenVT, X, 682*349cc55cSDimitry Andric CurDAG->getTargetConstant(C2 + C3, DL, XLenVT)); 683*349cc55cSDimitry Andric SDNode *SLLI = 684*349cc55cSDimitry Andric CurDAG->getMachineNode(RISCV::SLLI, DL, XLenVT, SDValue(SRLIW, 0), 685*349cc55cSDimitry Andric CurDAG->getTargetConstant(C3, DL, XLenVT)); 686*349cc55cSDimitry Andric ReplaceNode(Node, SLLI); 687*349cc55cSDimitry Andric return; 688*349cc55cSDimitry Andric } 689*349cc55cSDimitry Andric } 690*349cc55cSDimitry Andric 691*349cc55cSDimitry Andric // Turn (and (shl x, c2), c1) -> (slli (srli x, c3-c2), c3) if c1 is a 692*349cc55cSDimitry Andric // shifted mask with no leading zeros and c3 trailing zeros. 693*349cc55cSDimitry Andric if (LeftShift && isShiftedMask_64(C1)) { 694*349cc55cSDimitry Andric uint64_t Leading = XLen - (64 - countLeadingZeros(C1)); 695*349cc55cSDimitry Andric uint64_t C3 = countTrailingZeros(C1); 696*349cc55cSDimitry Andric if (Leading == 0 && C2 < C3 && OneUseOrZExtW && !ZExtOrANDI) { 697*349cc55cSDimitry Andric SDNode *SRLI = CurDAG->getMachineNode( 698*349cc55cSDimitry Andric RISCV::SRLI, DL, XLenVT, X, 699*349cc55cSDimitry Andric CurDAG->getTargetConstant(C3 - C2, DL, XLenVT)); 700*349cc55cSDimitry Andric SDNode *SLLI = 701*349cc55cSDimitry Andric CurDAG->getMachineNode(RISCV::SLLI, DL, XLenVT, SDValue(SRLI, 0), 702*349cc55cSDimitry Andric CurDAG->getTargetConstant(C3, DL, XLenVT)); 703*349cc55cSDimitry Andric ReplaceNode(Node, SLLI); 704*349cc55cSDimitry Andric return; 705*349cc55cSDimitry Andric } 706*349cc55cSDimitry Andric // If we have (32-C2) leading zeros, we can use SRLIW instead of SRLI. 707*349cc55cSDimitry Andric if (C2 < C3 && Leading + C2 == 32 && OneUseOrZExtW && !ZExtOrANDI) { 708*349cc55cSDimitry Andric SDNode *SRLIW = CurDAG->getMachineNode( 709*349cc55cSDimitry Andric RISCV::SRLIW, DL, XLenVT, X, 710*349cc55cSDimitry Andric CurDAG->getTargetConstant(C3 - C2, DL, XLenVT)); 711*349cc55cSDimitry Andric SDNode *SLLI = 712*349cc55cSDimitry Andric CurDAG->getMachineNode(RISCV::SLLI, DL, XLenVT, SDValue(SRLIW, 0), 713*349cc55cSDimitry Andric CurDAG->getTargetConstant(C3, DL, XLenVT)); 714*349cc55cSDimitry Andric ReplaceNode(Node, SLLI); 715*349cc55cSDimitry Andric return; 716*349cc55cSDimitry Andric } 717*349cc55cSDimitry Andric } 718*349cc55cSDimitry Andric 719fe6060f1SDimitry Andric break; 720fe6060f1SDimitry Andric } 721fe6060f1SDimitry Andric case ISD::INTRINSIC_WO_CHAIN: { 722fe6060f1SDimitry Andric unsigned IntNo = Node->getConstantOperandVal(0); 723fe6060f1SDimitry Andric switch (IntNo) { 724fe6060f1SDimitry Andric // By default we do not custom select any intrinsic. 725fe6060f1SDimitry Andric default: 726fe6060f1SDimitry Andric break; 727fe6060f1SDimitry Andric case Intrinsic::riscv_vmsgeu: 728fe6060f1SDimitry Andric case Intrinsic::riscv_vmsge: { 729fe6060f1SDimitry Andric SDValue Src1 = Node->getOperand(1); 730fe6060f1SDimitry Andric SDValue Src2 = Node->getOperand(2); 731fe6060f1SDimitry Andric // Only custom select scalar second operand. 732fe6060f1SDimitry Andric if (Src2.getValueType() != XLenVT) 733fe6060f1SDimitry Andric break; 734fe6060f1SDimitry Andric // Small constants are handled with patterns. 735fe6060f1SDimitry Andric if (auto *C = dyn_cast<ConstantSDNode>(Src2)) { 736fe6060f1SDimitry Andric int64_t CVal = C->getSExtValue(); 737fe6060f1SDimitry Andric if (CVal >= -15 && CVal <= 16) 738fe6060f1SDimitry Andric break; 739fe6060f1SDimitry Andric } 740fe6060f1SDimitry Andric bool IsUnsigned = IntNo == Intrinsic::riscv_vmsgeu; 741fe6060f1SDimitry Andric MVT Src1VT = Src1.getSimpleValueType(); 742fe6060f1SDimitry Andric unsigned VMSLTOpcode, VMNANDOpcode; 743fe6060f1SDimitry Andric switch (RISCVTargetLowering::getLMUL(Src1VT)) { 744fe6060f1SDimitry Andric default: 745fe6060f1SDimitry Andric llvm_unreachable("Unexpected LMUL!"); 746fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_F8: 747fe6060f1SDimitry Andric VMSLTOpcode = 748fe6060f1SDimitry Andric IsUnsigned ? RISCV::PseudoVMSLTU_VX_MF8 : RISCV::PseudoVMSLT_VX_MF8; 749fe6060f1SDimitry Andric VMNANDOpcode = RISCV::PseudoVMNAND_MM_MF8; 750fe6060f1SDimitry Andric break; 751fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_F4: 752fe6060f1SDimitry Andric VMSLTOpcode = 753fe6060f1SDimitry Andric IsUnsigned ? RISCV::PseudoVMSLTU_VX_MF4 : RISCV::PseudoVMSLT_VX_MF4; 754fe6060f1SDimitry Andric VMNANDOpcode = RISCV::PseudoVMNAND_MM_MF4; 755fe6060f1SDimitry Andric break; 756fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_F2: 757fe6060f1SDimitry Andric VMSLTOpcode = 758fe6060f1SDimitry Andric IsUnsigned ? RISCV::PseudoVMSLTU_VX_MF2 : RISCV::PseudoVMSLT_VX_MF2; 759fe6060f1SDimitry Andric VMNANDOpcode = RISCV::PseudoVMNAND_MM_MF2; 760fe6060f1SDimitry Andric break; 761fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_1: 762fe6060f1SDimitry Andric VMSLTOpcode = 763fe6060f1SDimitry Andric IsUnsigned ? RISCV::PseudoVMSLTU_VX_M1 : RISCV::PseudoVMSLT_VX_M1; 764fe6060f1SDimitry Andric VMNANDOpcode = RISCV::PseudoVMNAND_MM_M1; 765fe6060f1SDimitry Andric break; 766fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_2: 767fe6060f1SDimitry Andric VMSLTOpcode = 768fe6060f1SDimitry Andric IsUnsigned ? RISCV::PseudoVMSLTU_VX_M2 : RISCV::PseudoVMSLT_VX_M2; 769fe6060f1SDimitry Andric VMNANDOpcode = RISCV::PseudoVMNAND_MM_M2; 770fe6060f1SDimitry Andric break; 771fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_4: 772fe6060f1SDimitry Andric VMSLTOpcode = 773fe6060f1SDimitry Andric IsUnsigned ? RISCV::PseudoVMSLTU_VX_M4 : RISCV::PseudoVMSLT_VX_M4; 774fe6060f1SDimitry Andric VMNANDOpcode = RISCV::PseudoVMNAND_MM_M4; 775fe6060f1SDimitry Andric break; 776fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_8: 777fe6060f1SDimitry Andric VMSLTOpcode = 778fe6060f1SDimitry Andric IsUnsigned ? RISCV::PseudoVMSLTU_VX_M8 : RISCV::PseudoVMSLT_VX_M8; 779fe6060f1SDimitry Andric VMNANDOpcode = RISCV::PseudoVMNAND_MM_M8; 780fe6060f1SDimitry Andric break; 781fe6060f1SDimitry Andric } 782fe6060f1SDimitry Andric SDValue SEW = CurDAG->getTargetConstant( 783fe6060f1SDimitry Andric Log2_32(Src1VT.getScalarSizeInBits()), DL, XLenVT); 784fe6060f1SDimitry Andric SDValue VL; 785fe6060f1SDimitry Andric selectVLOp(Node->getOperand(3), VL); 786fe6060f1SDimitry Andric 787fe6060f1SDimitry Andric // Expand to 788fe6060f1SDimitry Andric // vmslt{u}.vx vd, va, x; vmnand.mm vd, vd, vd 789fe6060f1SDimitry Andric SDValue Cmp = SDValue( 790fe6060f1SDimitry Andric CurDAG->getMachineNode(VMSLTOpcode, DL, VT, {Src1, Src2, VL, SEW}), 791fe6060f1SDimitry Andric 0); 792fe6060f1SDimitry Andric ReplaceNode(Node, CurDAG->getMachineNode(VMNANDOpcode, DL, VT, 793fe6060f1SDimitry Andric {Cmp, Cmp, VL, SEW})); 794fe6060f1SDimitry Andric return; 795fe6060f1SDimitry Andric } 796fe6060f1SDimitry Andric case Intrinsic::riscv_vmsgeu_mask: 797fe6060f1SDimitry Andric case Intrinsic::riscv_vmsge_mask: { 798fe6060f1SDimitry Andric SDValue Src1 = Node->getOperand(2); 799fe6060f1SDimitry Andric SDValue Src2 = Node->getOperand(3); 800fe6060f1SDimitry Andric // Only custom select scalar second operand. 801fe6060f1SDimitry Andric if (Src2.getValueType() != XLenVT) 802fe6060f1SDimitry Andric break; 803fe6060f1SDimitry Andric // Small constants are handled with patterns. 804fe6060f1SDimitry Andric if (auto *C = dyn_cast<ConstantSDNode>(Src2)) { 805fe6060f1SDimitry Andric int64_t CVal = C->getSExtValue(); 806fe6060f1SDimitry Andric if (CVal >= -15 && CVal <= 16) 807fe6060f1SDimitry Andric break; 808fe6060f1SDimitry Andric } 809fe6060f1SDimitry Andric bool IsUnsigned = IntNo == Intrinsic::riscv_vmsgeu_mask; 810fe6060f1SDimitry Andric MVT Src1VT = Src1.getSimpleValueType(); 811*349cc55cSDimitry Andric unsigned VMSLTOpcode, VMSLTMaskOpcode, VMXOROpcode, VMANDNOpcode; 812fe6060f1SDimitry Andric switch (RISCVTargetLowering::getLMUL(Src1VT)) { 813fe6060f1SDimitry Andric default: 814fe6060f1SDimitry Andric llvm_unreachable("Unexpected LMUL!"); 815fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_F8: 816fe6060f1SDimitry Andric VMSLTOpcode = 817fe6060f1SDimitry Andric IsUnsigned ? RISCV::PseudoVMSLTU_VX_MF8 : RISCV::PseudoVMSLT_VX_MF8; 818fe6060f1SDimitry Andric VMSLTMaskOpcode = IsUnsigned ? RISCV::PseudoVMSLTU_VX_MF8_MASK 819fe6060f1SDimitry Andric : RISCV::PseudoVMSLT_VX_MF8_MASK; 820fe6060f1SDimitry Andric break; 821fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_F4: 822fe6060f1SDimitry Andric VMSLTOpcode = 823fe6060f1SDimitry Andric IsUnsigned ? RISCV::PseudoVMSLTU_VX_MF4 : RISCV::PseudoVMSLT_VX_MF4; 824fe6060f1SDimitry Andric VMSLTMaskOpcode = IsUnsigned ? RISCV::PseudoVMSLTU_VX_MF4_MASK 825fe6060f1SDimitry Andric : RISCV::PseudoVMSLT_VX_MF4_MASK; 826fe6060f1SDimitry Andric break; 827fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_F2: 828fe6060f1SDimitry Andric VMSLTOpcode = 829fe6060f1SDimitry Andric IsUnsigned ? RISCV::PseudoVMSLTU_VX_MF2 : RISCV::PseudoVMSLT_VX_MF2; 830fe6060f1SDimitry Andric VMSLTMaskOpcode = IsUnsigned ? RISCV::PseudoVMSLTU_VX_MF2_MASK 831fe6060f1SDimitry Andric : RISCV::PseudoVMSLT_VX_MF2_MASK; 832fe6060f1SDimitry Andric break; 833fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_1: 834fe6060f1SDimitry Andric VMSLTOpcode = 835fe6060f1SDimitry Andric IsUnsigned ? RISCV::PseudoVMSLTU_VX_M1 : RISCV::PseudoVMSLT_VX_M1; 836fe6060f1SDimitry Andric VMSLTMaskOpcode = IsUnsigned ? RISCV::PseudoVMSLTU_VX_M1_MASK 837fe6060f1SDimitry Andric : RISCV::PseudoVMSLT_VX_M1_MASK; 838fe6060f1SDimitry Andric break; 839fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_2: 840fe6060f1SDimitry Andric VMSLTOpcode = 841fe6060f1SDimitry Andric IsUnsigned ? RISCV::PseudoVMSLTU_VX_M2 : RISCV::PseudoVMSLT_VX_M2; 842fe6060f1SDimitry Andric VMSLTMaskOpcode = IsUnsigned ? RISCV::PseudoVMSLTU_VX_M2_MASK 843fe6060f1SDimitry Andric : RISCV::PseudoVMSLT_VX_M2_MASK; 844fe6060f1SDimitry Andric break; 845fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_4: 846fe6060f1SDimitry Andric VMSLTOpcode = 847fe6060f1SDimitry Andric IsUnsigned ? RISCV::PseudoVMSLTU_VX_M4 : RISCV::PseudoVMSLT_VX_M4; 848fe6060f1SDimitry Andric VMSLTMaskOpcode = IsUnsigned ? RISCV::PseudoVMSLTU_VX_M4_MASK 849fe6060f1SDimitry Andric : RISCV::PseudoVMSLT_VX_M4_MASK; 850fe6060f1SDimitry Andric break; 851fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_8: 852fe6060f1SDimitry Andric VMSLTOpcode = 853fe6060f1SDimitry Andric IsUnsigned ? RISCV::PseudoVMSLTU_VX_M8 : RISCV::PseudoVMSLT_VX_M8; 854fe6060f1SDimitry Andric VMSLTMaskOpcode = IsUnsigned ? RISCV::PseudoVMSLTU_VX_M8_MASK 855fe6060f1SDimitry Andric : RISCV::PseudoVMSLT_VX_M8_MASK; 856fe6060f1SDimitry Andric break; 857fe6060f1SDimitry Andric } 858fe6060f1SDimitry Andric // Mask operations use the LMUL from the mask type. 859fe6060f1SDimitry Andric switch (RISCVTargetLowering::getLMUL(VT)) { 860fe6060f1SDimitry Andric default: 861fe6060f1SDimitry Andric llvm_unreachable("Unexpected LMUL!"); 862fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_F8: 863fe6060f1SDimitry Andric VMXOROpcode = RISCV::PseudoVMXOR_MM_MF8; 864*349cc55cSDimitry Andric VMANDNOpcode = RISCV::PseudoVMANDN_MM_MF8; 865fe6060f1SDimitry Andric break; 866fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_F4: 867fe6060f1SDimitry Andric VMXOROpcode = RISCV::PseudoVMXOR_MM_MF4; 868*349cc55cSDimitry Andric VMANDNOpcode = RISCV::PseudoVMANDN_MM_MF4; 869fe6060f1SDimitry Andric break; 870fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_F2: 871fe6060f1SDimitry Andric VMXOROpcode = RISCV::PseudoVMXOR_MM_MF2; 872*349cc55cSDimitry Andric VMANDNOpcode = RISCV::PseudoVMANDN_MM_MF2; 873fe6060f1SDimitry Andric break; 874fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_1: 875fe6060f1SDimitry Andric VMXOROpcode = RISCV::PseudoVMXOR_MM_M1; 876*349cc55cSDimitry Andric VMANDNOpcode = RISCV::PseudoVMANDN_MM_M1; 877fe6060f1SDimitry Andric break; 878fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_2: 879fe6060f1SDimitry Andric VMXOROpcode = RISCV::PseudoVMXOR_MM_M2; 880*349cc55cSDimitry Andric VMANDNOpcode = RISCV::PseudoVMANDN_MM_M2; 881fe6060f1SDimitry Andric break; 882fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_4: 883fe6060f1SDimitry Andric VMXOROpcode = RISCV::PseudoVMXOR_MM_M4; 884*349cc55cSDimitry Andric VMANDNOpcode = RISCV::PseudoVMANDN_MM_M4; 885fe6060f1SDimitry Andric break; 886fe6060f1SDimitry Andric case RISCVII::VLMUL::LMUL_8: 887fe6060f1SDimitry Andric VMXOROpcode = RISCV::PseudoVMXOR_MM_M8; 888*349cc55cSDimitry Andric VMANDNOpcode = RISCV::PseudoVMANDN_MM_M8; 889fe6060f1SDimitry Andric break; 890fe6060f1SDimitry Andric } 891fe6060f1SDimitry Andric SDValue SEW = CurDAG->getTargetConstant( 892fe6060f1SDimitry Andric Log2_32(Src1VT.getScalarSizeInBits()), DL, XLenVT); 893fe6060f1SDimitry Andric SDValue MaskSEW = CurDAG->getTargetConstant(0, DL, XLenVT); 894fe6060f1SDimitry Andric SDValue VL; 895fe6060f1SDimitry Andric selectVLOp(Node->getOperand(5), VL); 896fe6060f1SDimitry Andric SDValue MaskedOff = Node->getOperand(1); 897fe6060f1SDimitry Andric SDValue Mask = Node->getOperand(4); 898fe6060f1SDimitry Andric // If the MaskedOff value and the Mask are the same value use 899*349cc55cSDimitry Andric // vmslt{u}.vx vt, va, x; vmandn.mm vd, vd, vt 900fe6060f1SDimitry Andric // This avoids needing to copy v0 to vd before starting the next sequence. 901fe6060f1SDimitry Andric if (Mask == MaskedOff) { 902fe6060f1SDimitry Andric SDValue Cmp = SDValue( 903fe6060f1SDimitry Andric CurDAG->getMachineNode(VMSLTOpcode, DL, VT, {Src1, Src2, VL, SEW}), 904fe6060f1SDimitry Andric 0); 905*349cc55cSDimitry Andric ReplaceNode(Node, CurDAG->getMachineNode(VMANDNOpcode, DL, VT, 906fe6060f1SDimitry Andric {Mask, Cmp, VL, MaskSEW})); 907fe6060f1SDimitry Andric return; 908fe6060f1SDimitry Andric } 909fe6060f1SDimitry Andric 910fe6060f1SDimitry Andric // Mask needs to be copied to V0. 911fe6060f1SDimitry Andric SDValue Chain = CurDAG->getCopyToReg(CurDAG->getEntryNode(), DL, 912fe6060f1SDimitry Andric RISCV::V0, Mask, SDValue()); 913fe6060f1SDimitry Andric SDValue Glue = Chain.getValue(1); 914fe6060f1SDimitry Andric SDValue V0 = CurDAG->getRegister(RISCV::V0, VT); 915fe6060f1SDimitry Andric 916fe6060f1SDimitry Andric // Otherwise use 917fe6060f1SDimitry Andric // vmslt{u}.vx vd, va, x, v0.t; vmxor.mm vd, vd, v0 918fe6060f1SDimitry Andric SDValue Cmp = SDValue( 919fe6060f1SDimitry Andric CurDAG->getMachineNode(VMSLTMaskOpcode, DL, VT, 920fe6060f1SDimitry Andric {MaskedOff, Src1, Src2, V0, VL, SEW, Glue}), 921fe6060f1SDimitry Andric 0); 922fe6060f1SDimitry Andric ReplaceNode(Node, CurDAG->getMachineNode(VMXOROpcode, DL, VT, 923fe6060f1SDimitry Andric {Cmp, Mask, VL, MaskSEW})); 924fe6060f1SDimitry Andric return; 925fe6060f1SDimitry Andric } 926fe6060f1SDimitry Andric } 927fe6060f1SDimitry Andric break; 928fe6060f1SDimitry Andric } 929e8d8bef9SDimitry Andric case ISD::INTRINSIC_W_CHAIN: { 930e8d8bef9SDimitry Andric unsigned IntNo = cast<ConstantSDNode>(Node->getOperand(1))->getZExtValue(); 931e8d8bef9SDimitry Andric switch (IntNo) { 932e8d8bef9SDimitry Andric // By default we do not custom select any intrinsic. 933e8d8bef9SDimitry Andric default: 9340b57cec5SDimitry Andric break; 9350b57cec5SDimitry Andric 936fe6060f1SDimitry Andric case Intrinsic::riscv_vsetvli: 937fe6060f1SDimitry Andric case Intrinsic::riscv_vsetvlimax: { 938*349cc55cSDimitry Andric if (!Subtarget->hasVInstructions()) 939e8d8bef9SDimitry Andric break; 940e8d8bef9SDimitry Andric 941fe6060f1SDimitry Andric bool VLMax = IntNo == Intrinsic::riscv_vsetvlimax; 942fe6060f1SDimitry Andric unsigned Offset = VLMax ? 2 : 3; 943e8d8bef9SDimitry Andric 944fe6060f1SDimitry Andric assert(Node->getNumOperands() == Offset + 2 && 945fe6060f1SDimitry Andric "Unexpected number of operands"); 946fe6060f1SDimitry Andric 947fe6060f1SDimitry Andric unsigned SEW = 948fe6060f1SDimitry Andric RISCVVType::decodeVSEW(Node->getConstantOperandVal(Offset) & 0x7); 949fe6060f1SDimitry Andric RISCVII::VLMUL VLMul = static_cast<RISCVII::VLMUL>( 950fe6060f1SDimitry Andric Node->getConstantOperandVal(Offset + 1) & 0x7); 951e8d8bef9SDimitry Andric 952e8d8bef9SDimitry Andric unsigned VTypeI = RISCVVType::encodeVTYPE( 953fe6060f1SDimitry Andric VLMul, SEW, /*TailAgnostic*/ true, /*MaskAgnostic*/ false); 954e8d8bef9SDimitry Andric SDValue VTypeIOp = CurDAG->getTargetConstant(VTypeI, DL, XLenVT); 955e8d8bef9SDimitry Andric 956fe6060f1SDimitry Andric SDValue VLOperand; 957*349cc55cSDimitry Andric unsigned Opcode = RISCV::PseudoVSETVLI; 958fe6060f1SDimitry Andric if (VLMax) { 959fe6060f1SDimitry Andric VLOperand = CurDAG->getRegister(RISCV::X0, XLenVT); 960*349cc55cSDimitry Andric Opcode = RISCV::PseudoVSETVLIX0; 961fe6060f1SDimitry Andric } else { 962fe6060f1SDimitry Andric VLOperand = Node->getOperand(2); 963fe6060f1SDimitry Andric 964e8d8bef9SDimitry Andric if (auto *C = dyn_cast<ConstantSDNode>(VLOperand)) { 965d409305fSDimitry Andric uint64_t AVL = C->getZExtValue(); 966d409305fSDimitry Andric if (isUInt<5>(AVL)) { 967d409305fSDimitry Andric SDValue VLImm = CurDAG->getTargetConstant(AVL, DL, XLenVT); 968fe6060f1SDimitry Andric ReplaceNode( 969fe6060f1SDimitry Andric Node, CurDAG->getMachineNode(RISCV::PseudoVSETIVLI, DL, XLenVT, 970d409305fSDimitry Andric MVT::Other, VLImm, VTypeIOp, 971d409305fSDimitry Andric /* Chain */ Node->getOperand(0))); 972d409305fSDimitry Andric return; 973e8d8bef9SDimitry Andric } 974e8d8bef9SDimitry Andric } 975e8d8bef9SDimitry Andric } 976e8d8bef9SDimitry Andric 977e8d8bef9SDimitry Andric ReplaceNode(Node, 978*349cc55cSDimitry Andric CurDAG->getMachineNode(Opcode, DL, XLenVT, 979e8d8bef9SDimitry Andric MVT::Other, VLOperand, VTypeIOp, 980e8d8bef9SDimitry Andric /* Chain */ Node->getOperand(0))); 981e8d8bef9SDimitry Andric return; 982e8d8bef9SDimitry Andric } 983e8d8bef9SDimitry Andric case Intrinsic::riscv_vlseg2: 984e8d8bef9SDimitry Andric case Intrinsic::riscv_vlseg3: 985e8d8bef9SDimitry Andric case Intrinsic::riscv_vlseg4: 986e8d8bef9SDimitry Andric case Intrinsic::riscv_vlseg5: 987e8d8bef9SDimitry Andric case Intrinsic::riscv_vlseg6: 988e8d8bef9SDimitry Andric case Intrinsic::riscv_vlseg7: 989e8d8bef9SDimitry Andric case Intrinsic::riscv_vlseg8: { 990fe6060f1SDimitry Andric selectVLSEG(Node, /*IsMasked*/ false, /*IsStrided*/ false); 991e8d8bef9SDimitry Andric return; 992e8d8bef9SDimitry Andric } 993e8d8bef9SDimitry Andric case Intrinsic::riscv_vlseg2_mask: 994e8d8bef9SDimitry Andric case Intrinsic::riscv_vlseg3_mask: 995e8d8bef9SDimitry Andric case Intrinsic::riscv_vlseg4_mask: 996e8d8bef9SDimitry Andric case Intrinsic::riscv_vlseg5_mask: 997e8d8bef9SDimitry Andric case Intrinsic::riscv_vlseg6_mask: 998e8d8bef9SDimitry Andric case Intrinsic::riscv_vlseg7_mask: 999e8d8bef9SDimitry Andric case Intrinsic::riscv_vlseg8_mask: { 1000fe6060f1SDimitry Andric selectVLSEG(Node, /*IsMasked*/ true, /*IsStrided*/ false); 1001e8d8bef9SDimitry Andric return; 1002e8d8bef9SDimitry Andric } 1003e8d8bef9SDimitry Andric case Intrinsic::riscv_vlsseg2: 1004e8d8bef9SDimitry Andric case Intrinsic::riscv_vlsseg3: 1005e8d8bef9SDimitry Andric case Intrinsic::riscv_vlsseg4: 1006e8d8bef9SDimitry Andric case Intrinsic::riscv_vlsseg5: 1007e8d8bef9SDimitry Andric case Intrinsic::riscv_vlsseg6: 1008e8d8bef9SDimitry Andric case Intrinsic::riscv_vlsseg7: 1009e8d8bef9SDimitry Andric case Intrinsic::riscv_vlsseg8: { 1010fe6060f1SDimitry Andric selectVLSEG(Node, /*IsMasked*/ false, /*IsStrided*/ true); 1011e8d8bef9SDimitry Andric return; 1012e8d8bef9SDimitry Andric } 1013e8d8bef9SDimitry Andric case Intrinsic::riscv_vlsseg2_mask: 1014e8d8bef9SDimitry Andric case Intrinsic::riscv_vlsseg3_mask: 1015e8d8bef9SDimitry Andric case Intrinsic::riscv_vlsseg4_mask: 1016e8d8bef9SDimitry Andric case Intrinsic::riscv_vlsseg5_mask: 1017e8d8bef9SDimitry Andric case Intrinsic::riscv_vlsseg6_mask: 1018e8d8bef9SDimitry Andric case Intrinsic::riscv_vlsseg7_mask: 1019e8d8bef9SDimitry Andric case Intrinsic::riscv_vlsseg8_mask: { 1020fe6060f1SDimitry Andric selectVLSEG(Node, /*IsMasked*/ true, /*IsStrided*/ true); 1021e8d8bef9SDimitry Andric return; 1022e8d8bef9SDimitry Andric } 1023e8d8bef9SDimitry Andric case Intrinsic::riscv_vloxseg2: 1024e8d8bef9SDimitry Andric case Intrinsic::riscv_vloxseg3: 1025e8d8bef9SDimitry Andric case Intrinsic::riscv_vloxseg4: 1026e8d8bef9SDimitry Andric case Intrinsic::riscv_vloxseg5: 1027e8d8bef9SDimitry Andric case Intrinsic::riscv_vloxseg6: 1028e8d8bef9SDimitry Andric case Intrinsic::riscv_vloxseg7: 1029e8d8bef9SDimitry Andric case Intrinsic::riscv_vloxseg8: 1030fe6060f1SDimitry Andric selectVLXSEG(Node, /*IsMasked*/ false, /*IsOrdered*/ true); 1031fe6060f1SDimitry Andric return; 1032e8d8bef9SDimitry Andric case Intrinsic::riscv_vluxseg2: 1033e8d8bef9SDimitry Andric case Intrinsic::riscv_vluxseg3: 1034e8d8bef9SDimitry Andric case Intrinsic::riscv_vluxseg4: 1035e8d8bef9SDimitry Andric case Intrinsic::riscv_vluxseg5: 1036e8d8bef9SDimitry Andric case Intrinsic::riscv_vluxseg6: 1037e8d8bef9SDimitry Andric case Intrinsic::riscv_vluxseg7: 1038fe6060f1SDimitry Andric case Intrinsic::riscv_vluxseg8: 1039fe6060f1SDimitry Andric selectVLXSEG(Node, /*IsMasked*/ false, /*IsOrdered*/ false); 1040e8d8bef9SDimitry Andric return; 1041e8d8bef9SDimitry Andric case Intrinsic::riscv_vloxseg2_mask: 1042e8d8bef9SDimitry Andric case Intrinsic::riscv_vloxseg3_mask: 1043e8d8bef9SDimitry Andric case Intrinsic::riscv_vloxseg4_mask: 1044e8d8bef9SDimitry Andric case Intrinsic::riscv_vloxseg5_mask: 1045e8d8bef9SDimitry Andric case Intrinsic::riscv_vloxseg6_mask: 1046e8d8bef9SDimitry Andric case Intrinsic::riscv_vloxseg7_mask: 1047e8d8bef9SDimitry Andric case Intrinsic::riscv_vloxseg8_mask: 1048fe6060f1SDimitry Andric selectVLXSEG(Node, /*IsMasked*/ true, /*IsOrdered*/ true); 1049fe6060f1SDimitry Andric return; 1050e8d8bef9SDimitry Andric case Intrinsic::riscv_vluxseg2_mask: 1051e8d8bef9SDimitry Andric case Intrinsic::riscv_vluxseg3_mask: 1052e8d8bef9SDimitry Andric case Intrinsic::riscv_vluxseg4_mask: 1053e8d8bef9SDimitry Andric case Intrinsic::riscv_vluxseg5_mask: 1054e8d8bef9SDimitry Andric case Intrinsic::riscv_vluxseg6_mask: 1055e8d8bef9SDimitry Andric case Intrinsic::riscv_vluxseg7_mask: 1056fe6060f1SDimitry Andric case Intrinsic::riscv_vluxseg8_mask: 1057fe6060f1SDimitry Andric selectVLXSEG(Node, /*IsMasked*/ true, /*IsOrdered*/ false); 1058fe6060f1SDimitry Andric return; 1059fe6060f1SDimitry Andric case Intrinsic::riscv_vlseg8ff: 1060fe6060f1SDimitry Andric case Intrinsic::riscv_vlseg7ff: 1061fe6060f1SDimitry Andric case Intrinsic::riscv_vlseg6ff: 1062fe6060f1SDimitry Andric case Intrinsic::riscv_vlseg5ff: 1063fe6060f1SDimitry Andric case Intrinsic::riscv_vlseg4ff: 1064fe6060f1SDimitry Andric case Intrinsic::riscv_vlseg3ff: 1065fe6060f1SDimitry Andric case Intrinsic::riscv_vlseg2ff: { 1066fe6060f1SDimitry Andric selectVLSEGFF(Node, /*IsMasked*/ false); 1067fe6060f1SDimitry Andric return; 1068fe6060f1SDimitry Andric } 1069fe6060f1SDimitry Andric case Intrinsic::riscv_vlseg8ff_mask: 1070fe6060f1SDimitry Andric case Intrinsic::riscv_vlseg7ff_mask: 1071fe6060f1SDimitry Andric case Intrinsic::riscv_vlseg6ff_mask: 1072fe6060f1SDimitry Andric case Intrinsic::riscv_vlseg5ff_mask: 1073fe6060f1SDimitry Andric case Intrinsic::riscv_vlseg4ff_mask: 1074fe6060f1SDimitry Andric case Intrinsic::riscv_vlseg3ff_mask: 1075fe6060f1SDimitry Andric case Intrinsic::riscv_vlseg2ff_mask: { 1076fe6060f1SDimitry Andric selectVLSEGFF(Node, /*IsMasked*/ true); 1077fe6060f1SDimitry Andric return; 1078fe6060f1SDimitry Andric } 1079fe6060f1SDimitry Andric case Intrinsic::riscv_vloxei: 1080fe6060f1SDimitry Andric case Intrinsic::riscv_vloxei_mask: 1081fe6060f1SDimitry Andric case Intrinsic::riscv_vluxei: 1082fe6060f1SDimitry Andric case Intrinsic::riscv_vluxei_mask: { 1083fe6060f1SDimitry Andric bool IsMasked = IntNo == Intrinsic::riscv_vloxei_mask || 1084fe6060f1SDimitry Andric IntNo == Intrinsic::riscv_vluxei_mask; 1085fe6060f1SDimitry Andric bool IsOrdered = IntNo == Intrinsic::riscv_vloxei || 1086fe6060f1SDimitry Andric IntNo == Intrinsic::riscv_vloxei_mask; 1087fe6060f1SDimitry Andric 1088fe6060f1SDimitry Andric MVT VT = Node->getSimpleValueType(0); 1089fe6060f1SDimitry Andric unsigned Log2SEW = Log2_32(VT.getScalarSizeInBits()); 1090fe6060f1SDimitry Andric 1091fe6060f1SDimitry Andric unsigned CurOp = 2; 1092fe6060f1SDimitry Andric SmallVector<SDValue, 8> Operands; 1093fe6060f1SDimitry Andric if (IsMasked) 1094fe6060f1SDimitry Andric Operands.push_back(Node->getOperand(CurOp++)); 1095fe6060f1SDimitry Andric 1096fe6060f1SDimitry Andric MVT IndexVT; 1097fe6060f1SDimitry Andric addVectorLoadStoreOperands(Node, Log2SEW, DL, CurOp, IsMasked, 1098fe6060f1SDimitry Andric /*IsStridedOrIndexed*/ true, Operands, 1099*349cc55cSDimitry Andric /*IsLoad=*/true, &IndexVT); 1100fe6060f1SDimitry Andric 1101fe6060f1SDimitry Andric assert(VT.getVectorElementCount() == IndexVT.getVectorElementCount() && 1102fe6060f1SDimitry Andric "Element count mismatch"); 1103fe6060f1SDimitry Andric 1104fe6060f1SDimitry Andric RISCVII::VLMUL LMUL = RISCVTargetLowering::getLMUL(VT); 1105fe6060f1SDimitry Andric RISCVII::VLMUL IndexLMUL = RISCVTargetLowering::getLMUL(IndexVT); 1106fe6060f1SDimitry Andric unsigned IndexLog2EEW = Log2_32(IndexVT.getScalarSizeInBits()); 1107fe6060f1SDimitry Andric const RISCV::VLX_VSXPseudo *P = RISCV::getVLXPseudo( 1108fe6060f1SDimitry Andric IsMasked, IsOrdered, IndexLog2EEW, static_cast<unsigned>(LMUL), 1109fe6060f1SDimitry Andric static_cast<unsigned>(IndexLMUL)); 1110fe6060f1SDimitry Andric MachineSDNode *Load = 1111fe6060f1SDimitry Andric CurDAG->getMachineNode(P->Pseudo, DL, Node->getVTList(), Operands); 1112fe6060f1SDimitry Andric 1113fe6060f1SDimitry Andric if (auto *MemOp = dyn_cast<MemSDNode>(Node)) 1114fe6060f1SDimitry Andric CurDAG->setNodeMemRefs(Load, {MemOp->getMemOperand()}); 1115fe6060f1SDimitry Andric 1116fe6060f1SDimitry Andric ReplaceNode(Node, Load); 1117fe6060f1SDimitry Andric return; 1118fe6060f1SDimitry Andric } 1119*349cc55cSDimitry Andric case Intrinsic::riscv_vlm: 1120fe6060f1SDimitry Andric case Intrinsic::riscv_vle: 1121fe6060f1SDimitry Andric case Intrinsic::riscv_vle_mask: 1122fe6060f1SDimitry Andric case Intrinsic::riscv_vlse: 1123fe6060f1SDimitry Andric case Intrinsic::riscv_vlse_mask: { 1124fe6060f1SDimitry Andric bool IsMasked = IntNo == Intrinsic::riscv_vle_mask || 1125fe6060f1SDimitry Andric IntNo == Intrinsic::riscv_vlse_mask; 1126fe6060f1SDimitry Andric bool IsStrided = 1127fe6060f1SDimitry Andric IntNo == Intrinsic::riscv_vlse || IntNo == Intrinsic::riscv_vlse_mask; 1128fe6060f1SDimitry Andric 1129fe6060f1SDimitry Andric MVT VT = Node->getSimpleValueType(0); 1130fe6060f1SDimitry Andric unsigned Log2SEW = Log2_32(VT.getScalarSizeInBits()); 1131fe6060f1SDimitry Andric 1132fe6060f1SDimitry Andric unsigned CurOp = 2; 1133fe6060f1SDimitry Andric SmallVector<SDValue, 8> Operands; 1134fe6060f1SDimitry Andric if (IsMasked) 1135fe6060f1SDimitry Andric Operands.push_back(Node->getOperand(CurOp++)); 1136fe6060f1SDimitry Andric 1137fe6060f1SDimitry Andric addVectorLoadStoreOperands(Node, Log2SEW, DL, CurOp, IsMasked, IsStrided, 1138*349cc55cSDimitry Andric Operands, /*IsLoad=*/true); 1139fe6060f1SDimitry Andric 1140fe6060f1SDimitry Andric RISCVII::VLMUL LMUL = RISCVTargetLowering::getLMUL(VT); 1141fe6060f1SDimitry Andric const RISCV::VLEPseudo *P = 1142fe6060f1SDimitry Andric RISCV::getVLEPseudo(IsMasked, IsStrided, /*FF*/ false, Log2SEW, 1143fe6060f1SDimitry Andric static_cast<unsigned>(LMUL)); 1144fe6060f1SDimitry Andric MachineSDNode *Load = 1145fe6060f1SDimitry Andric CurDAG->getMachineNode(P->Pseudo, DL, Node->getVTList(), Operands); 1146fe6060f1SDimitry Andric 1147fe6060f1SDimitry Andric if (auto *MemOp = dyn_cast<MemSDNode>(Node)) 1148fe6060f1SDimitry Andric CurDAG->setNodeMemRefs(Load, {MemOp->getMemOperand()}); 1149fe6060f1SDimitry Andric 1150fe6060f1SDimitry Andric ReplaceNode(Node, Load); 1151fe6060f1SDimitry Andric return; 1152fe6060f1SDimitry Andric } 1153fe6060f1SDimitry Andric case Intrinsic::riscv_vleff: 1154fe6060f1SDimitry Andric case Intrinsic::riscv_vleff_mask: { 1155fe6060f1SDimitry Andric bool IsMasked = IntNo == Intrinsic::riscv_vleff_mask; 1156fe6060f1SDimitry Andric 1157fe6060f1SDimitry Andric MVT VT = Node->getSimpleValueType(0); 1158fe6060f1SDimitry Andric unsigned Log2SEW = Log2_32(VT.getScalarSizeInBits()); 1159fe6060f1SDimitry Andric 1160fe6060f1SDimitry Andric unsigned CurOp = 2; 1161fe6060f1SDimitry Andric SmallVector<SDValue, 7> Operands; 1162fe6060f1SDimitry Andric if (IsMasked) 1163fe6060f1SDimitry Andric Operands.push_back(Node->getOperand(CurOp++)); 1164fe6060f1SDimitry Andric 1165fe6060f1SDimitry Andric addVectorLoadStoreOperands(Node, Log2SEW, DL, CurOp, IsMasked, 1166*349cc55cSDimitry Andric /*IsStridedOrIndexed*/ false, Operands, 1167*349cc55cSDimitry Andric /*IsLoad=*/true); 1168fe6060f1SDimitry Andric 1169fe6060f1SDimitry Andric RISCVII::VLMUL LMUL = RISCVTargetLowering::getLMUL(VT); 1170fe6060f1SDimitry Andric const RISCV::VLEPseudo *P = 1171fe6060f1SDimitry Andric RISCV::getVLEPseudo(IsMasked, /*Strided*/ false, /*FF*/ true, Log2SEW, 1172fe6060f1SDimitry Andric static_cast<unsigned>(LMUL)); 1173fe6060f1SDimitry Andric MachineSDNode *Load = 1174fe6060f1SDimitry Andric CurDAG->getMachineNode(P->Pseudo, DL, Node->getValueType(0), 1175fe6060f1SDimitry Andric MVT::Other, MVT::Glue, Operands); 1176fe6060f1SDimitry Andric SDNode *ReadVL = CurDAG->getMachineNode(RISCV::PseudoReadVL, DL, XLenVT, 1177fe6060f1SDimitry Andric /*Glue*/ SDValue(Load, 2)); 1178fe6060f1SDimitry Andric 1179fe6060f1SDimitry Andric if (auto *MemOp = dyn_cast<MemSDNode>(Node)) 1180fe6060f1SDimitry Andric CurDAG->setNodeMemRefs(Load, {MemOp->getMemOperand()}); 1181fe6060f1SDimitry Andric 1182fe6060f1SDimitry Andric ReplaceUses(SDValue(Node, 0), SDValue(Load, 0)); 1183fe6060f1SDimitry Andric ReplaceUses(SDValue(Node, 1), SDValue(ReadVL, 0)); // VL 1184fe6060f1SDimitry Andric ReplaceUses(SDValue(Node, 2), SDValue(Load, 1)); // Chain 1185fe6060f1SDimitry Andric CurDAG->RemoveDeadNode(Node); 11860b57cec5SDimitry Andric return; 11870b57cec5SDimitry Andric } 11880b57cec5SDimitry Andric } 11890b57cec5SDimitry Andric break; 11900b57cec5SDimitry Andric } 1191e8d8bef9SDimitry Andric case ISD::INTRINSIC_VOID: { 1192e8d8bef9SDimitry Andric unsigned IntNo = cast<ConstantSDNode>(Node->getOperand(1))->getZExtValue(); 1193e8d8bef9SDimitry Andric switch (IntNo) { 1194e8d8bef9SDimitry Andric case Intrinsic::riscv_vsseg2: 1195e8d8bef9SDimitry Andric case Intrinsic::riscv_vsseg3: 1196e8d8bef9SDimitry Andric case Intrinsic::riscv_vsseg4: 1197e8d8bef9SDimitry Andric case Intrinsic::riscv_vsseg5: 1198e8d8bef9SDimitry Andric case Intrinsic::riscv_vsseg6: 1199e8d8bef9SDimitry Andric case Intrinsic::riscv_vsseg7: 1200e8d8bef9SDimitry Andric case Intrinsic::riscv_vsseg8: { 1201fe6060f1SDimitry Andric selectVSSEG(Node, /*IsMasked*/ false, /*IsStrided*/ false); 12020b57cec5SDimitry Andric return; 12030b57cec5SDimitry Andric } 1204e8d8bef9SDimitry Andric case Intrinsic::riscv_vsseg2_mask: 1205e8d8bef9SDimitry Andric case Intrinsic::riscv_vsseg3_mask: 1206e8d8bef9SDimitry Andric case Intrinsic::riscv_vsseg4_mask: 1207e8d8bef9SDimitry Andric case Intrinsic::riscv_vsseg5_mask: 1208e8d8bef9SDimitry Andric case Intrinsic::riscv_vsseg6_mask: 1209e8d8bef9SDimitry Andric case Intrinsic::riscv_vsseg7_mask: 1210e8d8bef9SDimitry Andric case Intrinsic::riscv_vsseg8_mask: { 1211fe6060f1SDimitry Andric selectVSSEG(Node, /*IsMasked*/ true, /*IsStrided*/ false); 1212e8d8bef9SDimitry Andric return; 1213e8d8bef9SDimitry Andric } 1214e8d8bef9SDimitry Andric case Intrinsic::riscv_vssseg2: 1215e8d8bef9SDimitry Andric case Intrinsic::riscv_vssseg3: 1216e8d8bef9SDimitry Andric case Intrinsic::riscv_vssseg4: 1217e8d8bef9SDimitry Andric case Intrinsic::riscv_vssseg5: 1218e8d8bef9SDimitry Andric case Intrinsic::riscv_vssseg6: 1219e8d8bef9SDimitry Andric case Intrinsic::riscv_vssseg7: 1220e8d8bef9SDimitry Andric case Intrinsic::riscv_vssseg8: { 1221fe6060f1SDimitry Andric selectVSSEG(Node, /*IsMasked*/ false, /*IsStrided*/ true); 1222e8d8bef9SDimitry Andric return; 1223e8d8bef9SDimitry Andric } 1224e8d8bef9SDimitry Andric case Intrinsic::riscv_vssseg2_mask: 1225e8d8bef9SDimitry Andric case Intrinsic::riscv_vssseg3_mask: 1226e8d8bef9SDimitry Andric case Intrinsic::riscv_vssseg4_mask: 1227e8d8bef9SDimitry Andric case Intrinsic::riscv_vssseg5_mask: 1228e8d8bef9SDimitry Andric case Intrinsic::riscv_vssseg6_mask: 1229e8d8bef9SDimitry Andric case Intrinsic::riscv_vssseg7_mask: 1230e8d8bef9SDimitry Andric case Intrinsic::riscv_vssseg8_mask: { 1231fe6060f1SDimitry Andric selectVSSEG(Node, /*IsMasked*/ true, /*IsStrided*/ true); 1232e8d8bef9SDimitry Andric return; 1233e8d8bef9SDimitry Andric } 1234e8d8bef9SDimitry Andric case Intrinsic::riscv_vsoxseg2: 1235e8d8bef9SDimitry Andric case Intrinsic::riscv_vsoxseg3: 1236e8d8bef9SDimitry Andric case Intrinsic::riscv_vsoxseg4: 1237e8d8bef9SDimitry Andric case Intrinsic::riscv_vsoxseg5: 1238e8d8bef9SDimitry Andric case Intrinsic::riscv_vsoxseg6: 1239e8d8bef9SDimitry Andric case Intrinsic::riscv_vsoxseg7: 1240e8d8bef9SDimitry Andric case Intrinsic::riscv_vsoxseg8: 1241fe6060f1SDimitry Andric selectVSXSEG(Node, /*IsMasked*/ false, /*IsOrdered*/ true); 1242fe6060f1SDimitry Andric return; 1243e8d8bef9SDimitry Andric case Intrinsic::riscv_vsuxseg2: 1244e8d8bef9SDimitry Andric case Intrinsic::riscv_vsuxseg3: 1245e8d8bef9SDimitry Andric case Intrinsic::riscv_vsuxseg4: 1246e8d8bef9SDimitry Andric case Intrinsic::riscv_vsuxseg5: 1247e8d8bef9SDimitry Andric case Intrinsic::riscv_vsuxseg6: 1248e8d8bef9SDimitry Andric case Intrinsic::riscv_vsuxseg7: 1249fe6060f1SDimitry Andric case Intrinsic::riscv_vsuxseg8: 1250fe6060f1SDimitry Andric selectVSXSEG(Node, /*IsMasked*/ false, /*IsOrdered*/ false); 1251e8d8bef9SDimitry Andric return; 1252e8d8bef9SDimitry Andric case Intrinsic::riscv_vsoxseg2_mask: 1253e8d8bef9SDimitry Andric case Intrinsic::riscv_vsoxseg3_mask: 1254e8d8bef9SDimitry Andric case Intrinsic::riscv_vsoxseg4_mask: 1255e8d8bef9SDimitry Andric case Intrinsic::riscv_vsoxseg5_mask: 1256e8d8bef9SDimitry Andric case Intrinsic::riscv_vsoxseg6_mask: 1257e8d8bef9SDimitry Andric case Intrinsic::riscv_vsoxseg7_mask: 1258e8d8bef9SDimitry Andric case Intrinsic::riscv_vsoxseg8_mask: 1259fe6060f1SDimitry Andric selectVSXSEG(Node, /*IsMasked*/ true, /*IsOrdered*/ true); 1260fe6060f1SDimitry Andric return; 1261e8d8bef9SDimitry Andric case Intrinsic::riscv_vsuxseg2_mask: 1262e8d8bef9SDimitry Andric case Intrinsic::riscv_vsuxseg3_mask: 1263e8d8bef9SDimitry Andric case Intrinsic::riscv_vsuxseg4_mask: 1264e8d8bef9SDimitry Andric case Intrinsic::riscv_vsuxseg5_mask: 1265e8d8bef9SDimitry Andric case Intrinsic::riscv_vsuxseg6_mask: 1266e8d8bef9SDimitry Andric case Intrinsic::riscv_vsuxseg7_mask: 1267fe6060f1SDimitry Andric case Intrinsic::riscv_vsuxseg8_mask: 1268fe6060f1SDimitry Andric selectVSXSEG(Node, /*IsMasked*/ true, /*IsOrdered*/ false); 1269fe6060f1SDimitry Andric return; 1270fe6060f1SDimitry Andric case Intrinsic::riscv_vsoxei: 1271fe6060f1SDimitry Andric case Intrinsic::riscv_vsoxei_mask: 1272fe6060f1SDimitry Andric case Intrinsic::riscv_vsuxei: 1273fe6060f1SDimitry Andric case Intrinsic::riscv_vsuxei_mask: { 1274fe6060f1SDimitry Andric bool IsMasked = IntNo == Intrinsic::riscv_vsoxei_mask || 1275fe6060f1SDimitry Andric IntNo == Intrinsic::riscv_vsuxei_mask; 1276fe6060f1SDimitry Andric bool IsOrdered = IntNo == Intrinsic::riscv_vsoxei || 1277fe6060f1SDimitry Andric IntNo == Intrinsic::riscv_vsoxei_mask; 1278fe6060f1SDimitry Andric 1279fe6060f1SDimitry Andric MVT VT = Node->getOperand(2)->getSimpleValueType(0); 1280fe6060f1SDimitry Andric unsigned Log2SEW = Log2_32(VT.getScalarSizeInBits()); 1281fe6060f1SDimitry Andric 1282fe6060f1SDimitry Andric unsigned CurOp = 2; 1283fe6060f1SDimitry Andric SmallVector<SDValue, 8> Operands; 1284fe6060f1SDimitry Andric Operands.push_back(Node->getOperand(CurOp++)); // Store value. 1285fe6060f1SDimitry Andric 1286fe6060f1SDimitry Andric MVT IndexVT; 1287fe6060f1SDimitry Andric addVectorLoadStoreOperands(Node, Log2SEW, DL, CurOp, IsMasked, 1288fe6060f1SDimitry Andric /*IsStridedOrIndexed*/ true, Operands, 1289*349cc55cSDimitry Andric /*IsLoad=*/false, &IndexVT); 1290fe6060f1SDimitry Andric 1291fe6060f1SDimitry Andric assert(VT.getVectorElementCount() == IndexVT.getVectorElementCount() && 1292fe6060f1SDimitry Andric "Element count mismatch"); 1293fe6060f1SDimitry Andric 1294fe6060f1SDimitry Andric RISCVII::VLMUL LMUL = RISCVTargetLowering::getLMUL(VT); 1295fe6060f1SDimitry Andric RISCVII::VLMUL IndexLMUL = RISCVTargetLowering::getLMUL(IndexVT); 1296fe6060f1SDimitry Andric unsigned IndexLog2EEW = Log2_32(IndexVT.getScalarSizeInBits()); 1297fe6060f1SDimitry Andric const RISCV::VLX_VSXPseudo *P = RISCV::getVSXPseudo( 1298fe6060f1SDimitry Andric IsMasked, IsOrdered, IndexLog2EEW, static_cast<unsigned>(LMUL), 1299fe6060f1SDimitry Andric static_cast<unsigned>(IndexLMUL)); 1300fe6060f1SDimitry Andric MachineSDNode *Store = 1301fe6060f1SDimitry Andric CurDAG->getMachineNode(P->Pseudo, DL, Node->getVTList(), Operands); 1302fe6060f1SDimitry Andric 1303fe6060f1SDimitry Andric if (auto *MemOp = dyn_cast<MemSDNode>(Node)) 1304fe6060f1SDimitry Andric CurDAG->setNodeMemRefs(Store, {MemOp->getMemOperand()}); 1305fe6060f1SDimitry Andric 1306fe6060f1SDimitry Andric ReplaceNode(Node, Store); 1307fe6060f1SDimitry Andric return; 1308fe6060f1SDimitry Andric } 1309*349cc55cSDimitry Andric case Intrinsic::riscv_vsm: 1310fe6060f1SDimitry Andric case Intrinsic::riscv_vse: 1311fe6060f1SDimitry Andric case Intrinsic::riscv_vse_mask: 1312fe6060f1SDimitry Andric case Intrinsic::riscv_vsse: 1313fe6060f1SDimitry Andric case Intrinsic::riscv_vsse_mask: { 1314fe6060f1SDimitry Andric bool IsMasked = IntNo == Intrinsic::riscv_vse_mask || 1315fe6060f1SDimitry Andric IntNo == Intrinsic::riscv_vsse_mask; 1316fe6060f1SDimitry Andric bool IsStrided = 1317fe6060f1SDimitry Andric IntNo == Intrinsic::riscv_vsse || IntNo == Intrinsic::riscv_vsse_mask; 1318fe6060f1SDimitry Andric 1319fe6060f1SDimitry Andric MVT VT = Node->getOperand(2)->getSimpleValueType(0); 1320fe6060f1SDimitry Andric unsigned Log2SEW = Log2_32(VT.getScalarSizeInBits()); 1321fe6060f1SDimitry Andric 1322fe6060f1SDimitry Andric unsigned CurOp = 2; 1323fe6060f1SDimitry Andric SmallVector<SDValue, 8> Operands; 1324fe6060f1SDimitry Andric Operands.push_back(Node->getOperand(CurOp++)); // Store value. 1325fe6060f1SDimitry Andric 1326fe6060f1SDimitry Andric addVectorLoadStoreOperands(Node, Log2SEW, DL, CurOp, IsMasked, IsStrided, 1327fe6060f1SDimitry Andric Operands); 1328fe6060f1SDimitry Andric 1329fe6060f1SDimitry Andric RISCVII::VLMUL LMUL = RISCVTargetLowering::getLMUL(VT); 1330fe6060f1SDimitry Andric const RISCV::VSEPseudo *P = RISCV::getVSEPseudo( 1331fe6060f1SDimitry Andric IsMasked, IsStrided, Log2SEW, static_cast<unsigned>(LMUL)); 1332fe6060f1SDimitry Andric MachineSDNode *Store = 1333fe6060f1SDimitry Andric CurDAG->getMachineNode(P->Pseudo, DL, Node->getVTList(), Operands); 1334fe6060f1SDimitry Andric if (auto *MemOp = dyn_cast<MemSDNode>(Node)) 1335fe6060f1SDimitry Andric CurDAG->setNodeMemRefs(Store, {MemOp->getMemOperand()}); 1336fe6060f1SDimitry Andric 1337fe6060f1SDimitry Andric ReplaceNode(Node, Store); 1338e8d8bef9SDimitry Andric return; 1339e8d8bef9SDimitry Andric } 1340e8d8bef9SDimitry Andric } 1341e8d8bef9SDimitry Andric break; 1342e8d8bef9SDimitry Andric } 1343fe6060f1SDimitry Andric case ISD::BITCAST: { 1344fe6060f1SDimitry Andric MVT SrcVT = Node->getOperand(0).getSimpleValueType(); 1345fe6060f1SDimitry Andric // Just drop bitcasts between vectors if both are fixed or both are 1346fe6060f1SDimitry Andric // scalable. 1347fe6060f1SDimitry Andric if ((VT.isScalableVector() && SrcVT.isScalableVector()) || 1348fe6060f1SDimitry Andric (VT.isFixedLengthVector() && SrcVT.isFixedLengthVector())) { 1349fe6060f1SDimitry Andric ReplaceUses(SDValue(Node, 0), Node->getOperand(0)); 1350fe6060f1SDimitry Andric CurDAG->RemoveDeadNode(Node); 1351e8d8bef9SDimitry Andric return; 1352e8d8bef9SDimitry Andric } 1353fe6060f1SDimitry Andric break; 1354fe6060f1SDimitry Andric } 1355fe6060f1SDimitry Andric case ISD::INSERT_SUBVECTOR: { 1356fe6060f1SDimitry Andric SDValue V = Node->getOperand(0); 1357fe6060f1SDimitry Andric SDValue SubV = Node->getOperand(1); 1358fe6060f1SDimitry Andric SDLoc DL(SubV); 1359fe6060f1SDimitry Andric auto Idx = Node->getConstantOperandVal(2); 1360fe6060f1SDimitry Andric MVT SubVecVT = SubV.getSimpleValueType(); 1361fe6060f1SDimitry Andric 1362fe6060f1SDimitry Andric const RISCVTargetLowering &TLI = *Subtarget->getTargetLowering(); 1363fe6060f1SDimitry Andric MVT SubVecContainerVT = SubVecVT; 1364fe6060f1SDimitry Andric // Establish the correct scalable-vector types for any fixed-length type. 1365fe6060f1SDimitry Andric if (SubVecVT.isFixedLengthVector()) 1366fe6060f1SDimitry Andric SubVecContainerVT = TLI.getContainerForFixedLengthVector(SubVecVT); 1367fe6060f1SDimitry Andric if (VT.isFixedLengthVector()) 1368fe6060f1SDimitry Andric VT = TLI.getContainerForFixedLengthVector(VT); 1369fe6060f1SDimitry Andric 1370fe6060f1SDimitry Andric const auto *TRI = Subtarget->getRegisterInfo(); 1371fe6060f1SDimitry Andric unsigned SubRegIdx; 1372fe6060f1SDimitry Andric std::tie(SubRegIdx, Idx) = 1373fe6060f1SDimitry Andric RISCVTargetLowering::decomposeSubvectorInsertExtractToSubRegs( 1374fe6060f1SDimitry Andric VT, SubVecContainerVT, Idx, TRI); 1375fe6060f1SDimitry Andric 1376fe6060f1SDimitry Andric // If the Idx hasn't been completely eliminated then this is a subvector 1377fe6060f1SDimitry Andric // insert which doesn't naturally align to a vector register. These must 1378fe6060f1SDimitry Andric // be handled using instructions to manipulate the vector registers. 1379fe6060f1SDimitry Andric if (Idx != 0) 1380fe6060f1SDimitry Andric break; 1381fe6060f1SDimitry Andric 1382fe6060f1SDimitry Andric RISCVII::VLMUL SubVecLMUL = RISCVTargetLowering::getLMUL(SubVecContainerVT); 1383fe6060f1SDimitry Andric bool IsSubVecPartReg = SubVecLMUL == RISCVII::VLMUL::LMUL_F2 || 1384fe6060f1SDimitry Andric SubVecLMUL == RISCVII::VLMUL::LMUL_F4 || 1385fe6060f1SDimitry Andric SubVecLMUL == RISCVII::VLMUL::LMUL_F8; 1386fe6060f1SDimitry Andric (void)IsSubVecPartReg; // Silence unused variable warning without asserts. 1387fe6060f1SDimitry Andric assert((!IsSubVecPartReg || V.isUndef()) && 1388fe6060f1SDimitry Andric "Expecting lowering to have created legal INSERT_SUBVECTORs when " 1389fe6060f1SDimitry Andric "the subvector is smaller than a full-sized register"); 1390fe6060f1SDimitry Andric 1391fe6060f1SDimitry Andric // If we haven't set a SubRegIdx, then we must be going between 1392fe6060f1SDimitry Andric // equally-sized LMUL groups (e.g. VR -> VR). This can be done as a copy. 1393fe6060f1SDimitry Andric if (SubRegIdx == RISCV::NoSubRegister) { 1394fe6060f1SDimitry Andric unsigned InRegClassID = RISCVTargetLowering::getRegClassIDForVecVT(VT); 1395fe6060f1SDimitry Andric assert(RISCVTargetLowering::getRegClassIDForVecVT(SubVecContainerVT) == 1396fe6060f1SDimitry Andric InRegClassID && 1397fe6060f1SDimitry Andric "Unexpected subvector extraction"); 1398fe6060f1SDimitry Andric SDValue RC = CurDAG->getTargetConstant(InRegClassID, DL, XLenVT); 1399fe6060f1SDimitry Andric SDNode *NewNode = CurDAG->getMachineNode(TargetOpcode::COPY_TO_REGCLASS, 1400fe6060f1SDimitry Andric DL, VT, SubV, RC); 1401fe6060f1SDimitry Andric ReplaceNode(Node, NewNode); 1402fe6060f1SDimitry Andric return; 1403fe6060f1SDimitry Andric } 1404fe6060f1SDimitry Andric 1405fe6060f1SDimitry Andric SDValue Insert = CurDAG->getTargetInsertSubreg(SubRegIdx, DL, VT, V, SubV); 1406fe6060f1SDimitry Andric ReplaceNode(Node, Insert.getNode()); 1407fe6060f1SDimitry Andric return; 1408fe6060f1SDimitry Andric } 1409fe6060f1SDimitry Andric case ISD::EXTRACT_SUBVECTOR: { 1410fe6060f1SDimitry Andric SDValue V = Node->getOperand(0); 1411fe6060f1SDimitry Andric auto Idx = Node->getConstantOperandVal(1); 1412fe6060f1SDimitry Andric MVT InVT = V.getSimpleValueType(); 1413fe6060f1SDimitry Andric SDLoc DL(V); 1414fe6060f1SDimitry Andric 1415fe6060f1SDimitry Andric const RISCVTargetLowering &TLI = *Subtarget->getTargetLowering(); 1416fe6060f1SDimitry Andric MVT SubVecContainerVT = VT; 1417fe6060f1SDimitry Andric // Establish the correct scalable-vector types for any fixed-length type. 1418fe6060f1SDimitry Andric if (VT.isFixedLengthVector()) 1419fe6060f1SDimitry Andric SubVecContainerVT = TLI.getContainerForFixedLengthVector(VT); 1420fe6060f1SDimitry Andric if (InVT.isFixedLengthVector()) 1421fe6060f1SDimitry Andric InVT = TLI.getContainerForFixedLengthVector(InVT); 1422fe6060f1SDimitry Andric 1423fe6060f1SDimitry Andric const auto *TRI = Subtarget->getRegisterInfo(); 1424fe6060f1SDimitry Andric unsigned SubRegIdx; 1425fe6060f1SDimitry Andric std::tie(SubRegIdx, Idx) = 1426fe6060f1SDimitry Andric RISCVTargetLowering::decomposeSubvectorInsertExtractToSubRegs( 1427fe6060f1SDimitry Andric InVT, SubVecContainerVT, Idx, TRI); 1428fe6060f1SDimitry Andric 1429fe6060f1SDimitry Andric // If the Idx hasn't been completely eliminated then this is a subvector 1430fe6060f1SDimitry Andric // extract which doesn't naturally align to a vector register. These must 1431fe6060f1SDimitry Andric // be handled using instructions to manipulate the vector registers. 1432fe6060f1SDimitry Andric if (Idx != 0) 1433fe6060f1SDimitry Andric break; 1434fe6060f1SDimitry Andric 1435fe6060f1SDimitry Andric // If we haven't set a SubRegIdx, then we must be going between 1436fe6060f1SDimitry Andric // equally-sized LMUL types (e.g. VR -> VR). This can be done as a copy. 1437fe6060f1SDimitry Andric if (SubRegIdx == RISCV::NoSubRegister) { 1438fe6060f1SDimitry Andric unsigned InRegClassID = RISCVTargetLowering::getRegClassIDForVecVT(InVT); 1439fe6060f1SDimitry Andric assert(RISCVTargetLowering::getRegClassIDForVecVT(SubVecContainerVT) == 1440fe6060f1SDimitry Andric InRegClassID && 1441fe6060f1SDimitry Andric "Unexpected subvector extraction"); 1442fe6060f1SDimitry Andric SDValue RC = CurDAG->getTargetConstant(InRegClassID, DL, XLenVT); 1443fe6060f1SDimitry Andric SDNode *NewNode = 1444fe6060f1SDimitry Andric CurDAG->getMachineNode(TargetOpcode::COPY_TO_REGCLASS, DL, VT, V, RC); 1445fe6060f1SDimitry Andric ReplaceNode(Node, NewNode); 1446fe6060f1SDimitry Andric return; 1447fe6060f1SDimitry Andric } 1448fe6060f1SDimitry Andric 1449fe6060f1SDimitry Andric SDValue Extract = CurDAG->getTargetExtractSubreg(SubRegIdx, DL, VT, V); 1450fe6060f1SDimitry Andric ReplaceNode(Node, Extract.getNode()); 1451fe6060f1SDimitry Andric return; 1452fe6060f1SDimitry Andric } 1453fe6060f1SDimitry Andric case RISCVISD::VMV_V_X_VL: 1454fe6060f1SDimitry Andric case RISCVISD::VFMV_V_F_VL: { 1455fe6060f1SDimitry Andric // Try to match splat of a scalar load to a strided load with stride of x0. 1456fe6060f1SDimitry Andric SDValue Src = Node->getOperand(0); 1457fe6060f1SDimitry Andric auto *Ld = dyn_cast<LoadSDNode>(Src); 1458fe6060f1SDimitry Andric if (!Ld) 1459fe6060f1SDimitry Andric break; 1460fe6060f1SDimitry Andric EVT MemVT = Ld->getMemoryVT(); 1461fe6060f1SDimitry Andric // The memory VT should be the same size as the element type. 1462fe6060f1SDimitry Andric if (MemVT.getStoreSize() != VT.getVectorElementType().getStoreSize()) 1463fe6060f1SDimitry Andric break; 1464fe6060f1SDimitry Andric if (!IsProfitableToFold(Src, Node, Node) || 1465fe6060f1SDimitry Andric !IsLegalToFold(Src, Node, Node, TM.getOptLevel())) 1466fe6060f1SDimitry Andric break; 1467fe6060f1SDimitry Andric 1468fe6060f1SDimitry Andric SDValue VL; 1469fe6060f1SDimitry Andric selectVLOp(Node->getOperand(1), VL); 1470fe6060f1SDimitry Andric 1471fe6060f1SDimitry Andric unsigned Log2SEW = Log2_32(VT.getScalarSizeInBits()); 1472fe6060f1SDimitry Andric SDValue SEW = CurDAG->getTargetConstant(Log2SEW, DL, XLenVT); 1473fe6060f1SDimitry Andric 1474fe6060f1SDimitry Andric SDValue Operands[] = {Ld->getBasePtr(), 1475fe6060f1SDimitry Andric CurDAG->getRegister(RISCV::X0, XLenVT), VL, SEW, 1476fe6060f1SDimitry Andric Ld->getChain()}; 1477fe6060f1SDimitry Andric 1478fe6060f1SDimitry Andric RISCVII::VLMUL LMUL = RISCVTargetLowering::getLMUL(VT); 1479fe6060f1SDimitry Andric const RISCV::VLEPseudo *P = RISCV::getVLEPseudo( 1480fe6060f1SDimitry Andric /*IsMasked*/ false, /*IsStrided*/ true, /*FF*/ false, Log2SEW, 1481fe6060f1SDimitry Andric static_cast<unsigned>(LMUL)); 1482fe6060f1SDimitry Andric MachineSDNode *Load = 1483fe6060f1SDimitry Andric CurDAG->getMachineNode(P->Pseudo, DL, Node->getVTList(), Operands); 1484fe6060f1SDimitry Andric 1485fe6060f1SDimitry Andric if (auto *MemOp = dyn_cast<MemSDNode>(Node)) 1486fe6060f1SDimitry Andric CurDAG->setNodeMemRefs(Load, {MemOp->getMemOperand()}); 1487fe6060f1SDimitry Andric 1488fe6060f1SDimitry Andric ReplaceNode(Node, Load); 1489e8d8bef9SDimitry Andric return; 1490e8d8bef9SDimitry Andric } 1491e8d8bef9SDimitry Andric } 14920b57cec5SDimitry Andric 14930b57cec5SDimitry Andric // Select the default instruction. 14940b57cec5SDimitry Andric SelectCode(Node); 14950b57cec5SDimitry Andric } 14960b57cec5SDimitry Andric 14970b57cec5SDimitry Andric bool RISCVDAGToDAGISel::SelectInlineAsmMemoryOperand( 14980b57cec5SDimitry Andric const SDValue &Op, unsigned ConstraintID, std::vector<SDValue> &OutOps) { 14990b57cec5SDimitry Andric switch (ConstraintID) { 15000b57cec5SDimitry Andric case InlineAsm::Constraint_m: 15010b57cec5SDimitry Andric // We just support simple memory operands that have a single address 15020b57cec5SDimitry Andric // operand and need no special handling. 15030b57cec5SDimitry Andric OutOps.push_back(Op); 15040b57cec5SDimitry Andric return false; 15050b57cec5SDimitry Andric case InlineAsm::Constraint_A: 15060b57cec5SDimitry Andric OutOps.push_back(Op); 15070b57cec5SDimitry Andric return false; 15080b57cec5SDimitry Andric default: 15090b57cec5SDimitry Andric break; 15100b57cec5SDimitry Andric } 15110b57cec5SDimitry Andric 15120b57cec5SDimitry Andric return true; 15130b57cec5SDimitry Andric } 15140b57cec5SDimitry Andric 15150b57cec5SDimitry Andric bool RISCVDAGToDAGISel::SelectAddrFI(SDValue Addr, SDValue &Base) { 1516fe6060f1SDimitry Andric if (auto *FIN = dyn_cast<FrameIndexSDNode>(Addr)) { 15170b57cec5SDimitry Andric Base = CurDAG->getTargetFrameIndex(FIN->getIndex(), Subtarget->getXLenVT()); 15180b57cec5SDimitry Andric return true; 15190b57cec5SDimitry Andric } 15200b57cec5SDimitry Andric return false; 15210b57cec5SDimitry Andric } 15220b57cec5SDimitry Andric 1523fe6060f1SDimitry Andric bool RISCVDAGToDAGISel::SelectBaseAddr(SDValue Addr, SDValue &Base) { 1524fe6060f1SDimitry Andric // If this is FrameIndex, select it directly. Otherwise just let it get 1525fe6060f1SDimitry Andric // selected to a register independently. 1526fe6060f1SDimitry Andric if (auto *FIN = dyn_cast<FrameIndexSDNode>(Addr)) 1527fe6060f1SDimitry Andric Base = CurDAG->getTargetFrameIndex(FIN->getIndex(), Subtarget->getXLenVT()); 1528fe6060f1SDimitry Andric else 1529fe6060f1SDimitry Andric Base = Addr; 1530fe6060f1SDimitry Andric return true; 1531e8d8bef9SDimitry Andric } 1532e8d8bef9SDimitry Andric 1533fe6060f1SDimitry Andric bool RISCVDAGToDAGISel::selectShiftMask(SDValue N, unsigned ShiftWidth, 1534fe6060f1SDimitry Andric SDValue &ShAmt) { 1535fe6060f1SDimitry Andric // Shift instructions on RISCV only read the lower 5 or 6 bits of the shift 1536fe6060f1SDimitry Andric // amount. If there is an AND on the shift amount, we can bypass it if it 1537fe6060f1SDimitry Andric // doesn't affect any of those bits. 1538fe6060f1SDimitry Andric if (N.getOpcode() == ISD::AND && isa<ConstantSDNode>(N.getOperand(1))) { 1539fe6060f1SDimitry Andric const APInt &AndMask = N->getConstantOperandAPInt(1); 1540979e22ffSDimitry Andric 1541fe6060f1SDimitry Andric // Since the max shift amount is a power of 2 we can subtract 1 to make a 1542fe6060f1SDimitry Andric // mask that covers the bits needed to represent all shift amounts. 1543fe6060f1SDimitry Andric assert(isPowerOf2_32(ShiftWidth) && "Unexpected max shift amount!"); 1544fe6060f1SDimitry Andric APInt ShMask(AndMask.getBitWidth(), ShiftWidth - 1); 1545e8d8bef9SDimitry Andric 1546fe6060f1SDimitry Andric if (ShMask.isSubsetOf(AndMask)) { 1547fe6060f1SDimitry Andric ShAmt = N.getOperand(0); 1548fe6060f1SDimitry Andric return true; 1549e8d8bef9SDimitry Andric } 1550e8d8bef9SDimitry Andric 1551fe6060f1SDimitry Andric // SimplifyDemandedBits may have optimized the mask so try restoring any 1552fe6060f1SDimitry Andric // bits that are known zero. 1553fe6060f1SDimitry Andric KnownBits Known = CurDAG->computeKnownBits(N->getOperand(0)); 1554fe6060f1SDimitry Andric if (ShMask.isSubsetOf(AndMask | Known.Zero)) { 1555fe6060f1SDimitry Andric ShAmt = N.getOperand(0); 1556fe6060f1SDimitry Andric return true; 1557fe6060f1SDimitry Andric } 1558fe6060f1SDimitry Andric } 1559fe6060f1SDimitry Andric 1560fe6060f1SDimitry Andric ShAmt = N; 1561fe6060f1SDimitry Andric return true; 1562fe6060f1SDimitry Andric } 1563fe6060f1SDimitry Andric 1564fe6060f1SDimitry Andric bool RISCVDAGToDAGISel::selectSExti32(SDValue N, SDValue &Val) { 1565fe6060f1SDimitry Andric if (N.getOpcode() == ISD::SIGN_EXTEND_INREG && 1566fe6060f1SDimitry Andric cast<VTSDNode>(N.getOperand(1))->getVT() == MVT::i32) { 1567fe6060f1SDimitry Andric Val = N.getOperand(0); 1568fe6060f1SDimitry Andric return true; 1569fe6060f1SDimitry Andric } 1570fe6060f1SDimitry Andric MVT VT = N.getSimpleValueType(); 1571fe6060f1SDimitry Andric if (CurDAG->ComputeNumSignBits(N) > (VT.getSizeInBits() - 32)) { 1572fe6060f1SDimitry Andric Val = N; 1573fe6060f1SDimitry Andric return true; 1574fe6060f1SDimitry Andric } 1575fe6060f1SDimitry Andric 1576fe6060f1SDimitry Andric return false; 1577fe6060f1SDimitry Andric } 1578fe6060f1SDimitry Andric 1579fe6060f1SDimitry Andric bool RISCVDAGToDAGISel::selectZExti32(SDValue N, SDValue &Val) { 1580fe6060f1SDimitry Andric if (N.getOpcode() == ISD::AND) { 1581fe6060f1SDimitry Andric auto *C = dyn_cast<ConstantSDNode>(N.getOperand(1)); 1582fe6060f1SDimitry Andric if (C && C->getZExtValue() == UINT64_C(0xFFFFFFFF)) { 1583fe6060f1SDimitry Andric Val = N.getOperand(0); 1584fe6060f1SDimitry Andric return true; 1585fe6060f1SDimitry Andric } 1586fe6060f1SDimitry Andric } 1587fe6060f1SDimitry Andric MVT VT = N.getSimpleValueType(); 1588fe6060f1SDimitry Andric APInt Mask = APInt::getHighBitsSet(VT.getSizeInBits(), 32); 1589fe6060f1SDimitry Andric if (CurDAG->MaskedValueIsZero(N, Mask)) { 1590fe6060f1SDimitry Andric Val = N; 1591fe6060f1SDimitry Andric return true; 1592fe6060f1SDimitry Andric } 1593fe6060f1SDimitry Andric 1594fe6060f1SDimitry Andric return false; 1595fe6060f1SDimitry Andric } 1596fe6060f1SDimitry Andric 1597*349cc55cSDimitry Andric // Return true if all users of this SDNode* only consume the lower \p Bits. 1598*349cc55cSDimitry Andric // This can be used to form W instructions for add/sub/mul/shl even when the 1599*349cc55cSDimitry Andric // root isn't a sext_inreg. This can allow the ADDW/SUBW/MULW/SLLIW to CSE if 1600*349cc55cSDimitry Andric // SimplifyDemandedBits has made it so some users see a sext_inreg and some 1601*349cc55cSDimitry Andric // don't. The sext_inreg+add/sub/mul/shl will get selected, but still leave 1602*349cc55cSDimitry Andric // the add/sub/mul/shl to become non-W instructions. By checking the users we 1603*349cc55cSDimitry Andric // may be able to use a W instruction and CSE with the other instruction if 1604*349cc55cSDimitry Andric // this has happened. We could try to detect that the CSE opportunity exists 1605*349cc55cSDimitry Andric // before doing this, but that would be more complicated. 1606*349cc55cSDimitry Andric // TODO: Does this need to look through AND/OR/XOR to their users to find more 1607*349cc55cSDimitry Andric // opportunities. 1608*349cc55cSDimitry Andric bool RISCVDAGToDAGISel::hasAllNBitUsers(SDNode *Node, unsigned Bits) const { 1609*349cc55cSDimitry Andric assert((Node->getOpcode() == ISD::ADD || Node->getOpcode() == ISD::SUB || 1610*349cc55cSDimitry Andric Node->getOpcode() == ISD::MUL || Node->getOpcode() == ISD::SHL || 1611*349cc55cSDimitry Andric Node->getOpcode() == ISD::SRL || 1612*349cc55cSDimitry Andric Node->getOpcode() == ISD::SIGN_EXTEND_INREG || 1613*349cc55cSDimitry Andric isa<ConstantSDNode>(Node)) && 1614*349cc55cSDimitry Andric "Unexpected opcode"); 1615*349cc55cSDimitry Andric 1616*349cc55cSDimitry Andric for (auto UI = Node->use_begin(), UE = Node->use_end(); UI != UE; ++UI) { 1617*349cc55cSDimitry Andric SDNode *User = *UI; 1618*349cc55cSDimitry Andric // Users of this node should have already been instruction selected 1619*349cc55cSDimitry Andric if (!User->isMachineOpcode()) 1620*349cc55cSDimitry Andric return false; 1621*349cc55cSDimitry Andric 1622*349cc55cSDimitry Andric // TODO: Add more opcodes? 1623*349cc55cSDimitry Andric switch (User->getMachineOpcode()) { 1624*349cc55cSDimitry Andric default: 1625*349cc55cSDimitry Andric return false; 1626*349cc55cSDimitry Andric case RISCV::ADDW: 1627*349cc55cSDimitry Andric case RISCV::ADDIW: 1628*349cc55cSDimitry Andric case RISCV::SUBW: 1629*349cc55cSDimitry Andric case RISCV::MULW: 1630*349cc55cSDimitry Andric case RISCV::SLLW: 1631*349cc55cSDimitry Andric case RISCV::SLLIW: 1632*349cc55cSDimitry Andric case RISCV::SRAW: 1633*349cc55cSDimitry Andric case RISCV::SRAIW: 1634*349cc55cSDimitry Andric case RISCV::SRLW: 1635*349cc55cSDimitry Andric case RISCV::SRLIW: 1636*349cc55cSDimitry Andric case RISCV::DIVW: 1637*349cc55cSDimitry Andric case RISCV::DIVUW: 1638*349cc55cSDimitry Andric case RISCV::REMW: 1639*349cc55cSDimitry Andric case RISCV::REMUW: 1640*349cc55cSDimitry Andric case RISCV::ROLW: 1641*349cc55cSDimitry Andric case RISCV::RORW: 1642*349cc55cSDimitry Andric case RISCV::RORIW: 1643*349cc55cSDimitry Andric case RISCV::CLZW: 1644*349cc55cSDimitry Andric case RISCV::CTZW: 1645*349cc55cSDimitry Andric case RISCV::CPOPW: 1646*349cc55cSDimitry Andric case RISCV::SLLIUW: 1647*349cc55cSDimitry Andric case RISCV::FCVT_H_W: 1648*349cc55cSDimitry Andric case RISCV::FCVT_H_WU: 1649*349cc55cSDimitry Andric case RISCV::FCVT_S_W: 1650*349cc55cSDimitry Andric case RISCV::FCVT_S_WU: 1651*349cc55cSDimitry Andric case RISCV::FCVT_D_W: 1652*349cc55cSDimitry Andric case RISCV::FCVT_D_WU: 1653*349cc55cSDimitry Andric if (Bits < 32) 1654*349cc55cSDimitry Andric return false; 1655*349cc55cSDimitry Andric break; 1656*349cc55cSDimitry Andric case RISCV::SLLI: 1657*349cc55cSDimitry Andric // SLLI only uses the lower (XLen - ShAmt) bits. 1658*349cc55cSDimitry Andric if (Bits < Subtarget->getXLen() - User->getConstantOperandVal(1)) 1659*349cc55cSDimitry Andric return false; 1660*349cc55cSDimitry Andric break; 1661*349cc55cSDimitry Andric case RISCV::ADDUW: 1662*349cc55cSDimitry Andric case RISCV::SH1ADDUW: 1663*349cc55cSDimitry Andric case RISCV::SH2ADDUW: 1664*349cc55cSDimitry Andric case RISCV::SH3ADDUW: 1665*349cc55cSDimitry Andric // The first operand to add.uw/shXadd.uw is implicitly zero extended from 1666*349cc55cSDimitry Andric // 32 bits. 1667*349cc55cSDimitry Andric if (UI.getOperandNo() != 0 || Bits < 32) 1668*349cc55cSDimitry Andric return false; 1669*349cc55cSDimitry Andric break; 1670*349cc55cSDimitry Andric case RISCV::SB: 1671*349cc55cSDimitry Andric if (UI.getOperandNo() != 0 || Bits < 8) 1672*349cc55cSDimitry Andric return false; 1673*349cc55cSDimitry Andric break; 1674*349cc55cSDimitry Andric case RISCV::SH: 1675*349cc55cSDimitry Andric if (UI.getOperandNo() != 0 || Bits < 16) 1676*349cc55cSDimitry Andric return false; 1677*349cc55cSDimitry Andric break; 1678*349cc55cSDimitry Andric case RISCV::SW: 1679*349cc55cSDimitry Andric if (UI.getOperandNo() != 0 || Bits < 32) 1680*349cc55cSDimitry Andric return false; 1681*349cc55cSDimitry Andric break; 1682*349cc55cSDimitry Andric } 1683*349cc55cSDimitry Andric } 1684*349cc55cSDimitry Andric 1685*349cc55cSDimitry Andric return true; 1686*349cc55cSDimitry Andric } 1687*349cc55cSDimitry Andric 1688fe6060f1SDimitry Andric // Select VL as a 5 bit immediate or a value that will become a register. This 1689fe6060f1SDimitry Andric // allows us to choose betwen VSETIVLI or VSETVLI later. 1690d409305fSDimitry Andric bool RISCVDAGToDAGISel::selectVLOp(SDValue N, SDValue &VL) { 1691d409305fSDimitry Andric auto *C = dyn_cast<ConstantSDNode>(N); 1692fe6060f1SDimitry Andric if (C && isUInt<5>(C->getZExtValue())) 1693fe6060f1SDimitry Andric VL = CurDAG->getTargetConstant(C->getZExtValue(), SDLoc(N), 1694fe6060f1SDimitry Andric N->getValueType(0)); 1695d409305fSDimitry Andric else 1696d409305fSDimitry Andric VL = N; 1697d409305fSDimitry Andric 1698d409305fSDimitry Andric return true; 1699d409305fSDimitry Andric } 1700d409305fSDimitry Andric 1701e8d8bef9SDimitry Andric bool RISCVDAGToDAGISel::selectVSplat(SDValue N, SDValue &SplatVal) { 1702e8d8bef9SDimitry Andric if (N.getOpcode() != ISD::SPLAT_VECTOR && 1703fe6060f1SDimitry Andric N.getOpcode() != RISCVISD::SPLAT_VECTOR_I64 && 1704fe6060f1SDimitry Andric N.getOpcode() != RISCVISD::VMV_V_X_VL) 1705e8d8bef9SDimitry Andric return false; 1706e8d8bef9SDimitry Andric SplatVal = N.getOperand(0); 1707979e22ffSDimitry Andric return true; 1708979e22ffSDimitry Andric } 1709e8d8bef9SDimitry Andric 1710fe6060f1SDimitry Andric using ValidateFn = bool (*)(int64_t); 1711fe6060f1SDimitry Andric 1712fe6060f1SDimitry Andric static bool selectVSplatSimmHelper(SDValue N, SDValue &SplatVal, 1713fe6060f1SDimitry Andric SelectionDAG &DAG, 1714fe6060f1SDimitry Andric const RISCVSubtarget &Subtarget, 1715fe6060f1SDimitry Andric ValidateFn ValidateImm) { 1716e8d8bef9SDimitry Andric if ((N.getOpcode() != ISD::SPLAT_VECTOR && 1717fe6060f1SDimitry Andric N.getOpcode() != RISCVISD::SPLAT_VECTOR_I64 && 1718fe6060f1SDimitry Andric N.getOpcode() != RISCVISD::VMV_V_X_VL) || 1719e8d8bef9SDimitry Andric !isa<ConstantSDNode>(N.getOperand(0))) 1720979e22ffSDimitry Andric return false; 1721e8d8bef9SDimitry Andric 1722e8d8bef9SDimitry Andric int64_t SplatImm = cast<ConstantSDNode>(N.getOperand(0))->getSExtValue(); 1723e8d8bef9SDimitry Andric 1724fe6060f1SDimitry Andric // ISD::SPLAT_VECTOR, RISCVISD::SPLAT_VECTOR_I64 and RISCVISD::VMV_V_X_VL 1725fe6060f1SDimitry Andric // share semantics when the operand type is wider than the resulting vector 1726fe6060f1SDimitry Andric // element type: an implicit truncation first takes place. Therefore, perform 1727fe6060f1SDimitry Andric // a manual truncation/sign-extension in order to ignore any truncated bits 1728fe6060f1SDimitry Andric // and catch any zero-extended immediate. 1729e8d8bef9SDimitry Andric // For example, we wish to match (i8 -1) -> (XLenVT 255) as a simm5 by first 1730e8d8bef9SDimitry Andric // sign-extending to (XLenVT -1). 1731fe6060f1SDimitry Andric MVT XLenVT = Subtarget.getXLenVT(); 1732e8d8bef9SDimitry Andric assert(XLenVT == N.getOperand(0).getSimpleValueType() && 1733e8d8bef9SDimitry Andric "Unexpected splat operand type"); 1734fe6060f1SDimitry Andric MVT EltVT = N.getSimpleValueType().getVectorElementType(); 1735fe6060f1SDimitry Andric if (EltVT.bitsLT(XLenVT)) 1736e8d8bef9SDimitry Andric SplatImm = SignExtend64(SplatImm, EltVT.getSizeInBits()); 1737979e22ffSDimitry Andric 1738fe6060f1SDimitry Andric if (!ValidateImm(SplatImm)) 1739e8d8bef9SDimitry Andric return false; 1740979e22ffSDimitry Andric 1741fe6060f1SDimitry Andric SplatVal = DAG.getTargetConstant(SplatImm, SDLoc(N), XLenVT); 1742979e22ffSDimitry Andric return true; 1743979e22ffSDimitry Andric } 1744e8d8bef9SDimitry Andric 1745fe6060f1SDimitry Andric bool RISCVDAGToDAGISel::selectVSplatSimm5(SDValue N, SDValue &SplatVal) { 1746fe6060f1SDimitry Andric return selectVSplatSimmHelper(N, SplatVal, *CurDAG, *Subtarget, 1747fe6060f1SDimitry Andric [](int64_t Imm) { return isInt<5>(Imm); }); 1748fe6060f1SDimitry Andric } 1749fe6060f1SDimitry Andric 1750fe6060f1SDimitry Andric bool RISCVDAGToDAGISel::selectVSplatSimm5Plus1(SDValue N, SDValue &SplatVal) { 1751fe6060f1SDimitry Andric return selectVSplatSimmHelper( 1752fe6060f1SDimitry Andric N, SplatVal, *CurDAG, *Subtarget, 1753fe6060f1SDimitry Andric [](int64_t Imm) { return (isInt<5>(Imm) && Imm != -16) || Imm == 16; }); 1754fe6060f1SDimitry Andric } 1755fe6060f1SDimitry Andric 1756fe6060f1SDimitry Andric bool RISCVDAGToDAGISel::selectVSplatSimm5Plus1NonZero(SDValue N, 1757fe6060f1SDimitry Andric SDValue &SplatVal) { 1758fe6060f1SDimitry Andric return selectVSplatSimmHelper( 1759fe6060f1SDimitry Andric N, SplatVal, *CurDAG, *Subtarget, [](int64_t Imm) { 1760fe6060f1SDimitry Andric return Imm != 0 && ((isInt<5>(Imm) && Imm != -16) || Imm == 16); 1761fe6060f1SDimitry Andric }); 1762fe6060f1SDimitry Andric } 1763fe6060f1SDimitry Andric 1764e8d8bef9SDimitry Andric bool RISCVDAGToDAGISel::selectVSplatUimm5(SDValue N, SDValue &SplatVal) { 1765e8d8bef9SDimitry Andric if ((N.getOpcode() != ISD::SPLAT_VECTOR && 1766fe6060f1SDimitry Andric N.getOpcode() != RISCVISD::SPLAT_VECTOR_I64 && 1767fe6060f1SDimitry Andric N.getOpcode() != RISCVISD::VMV_V_X_VL) || 1768e8d8bef9SDimitry Andric !isa<ConstantSDNode>(N.getOperand(0))) 1769979e22ffSDimitry Andric return false; 1770979e22ffSDimitry Andric 1771e8d8bef9SDimitry Andric int64_t SplatImm = cast<ConstantSDNode>(N.getOperand(0))->getSExtValue(); 1772979e22ffSDimitry Andric 1773e8d8bef9SDimitry Andric if (!isUInt<5>(SplatImm)) 1774e8d8bef9SDimitry Andric return false; 1775e8d8bef9SDimitry Andric 1776e8d8bef9SDimitry Andric SplatVal = 1777e8d8bef9SDimitry Andric CurDAG->getTargetConstant(SplatImm, SDLoc(N), Subtarget->getXLenVT()); 1778e8d8bef9SDimitry Andric 1779979e22ffSDimitry Andric return true; 1780979e22ffSDimitry Andric } 1781979e22ffSDimitry Andric 1782fe6060f1SDimitry Andric bool RISCVDAGToDAGISel::selectRVVSimm5(SDValue N, unsigned Width, 1783fe6060f1SDimitry Andric SDValue &Imm) { 1784fe6060f1SDimitry Andric if (auto *C = dyn_cast<ConstantSDNode>(N)) { 1785fe6060f1SDimitry Andric int64_t ImmVal = SignExtend64(C->getSExtValue(), Width); 1786fe6060f1SDimitry Andric 1787fe6060f1SDimitry Andric if (!isInt<5>(ImmVal)) 1788fe6060f1SDimitry Andric return false; 1789fe6060f1SDimitry Andric 1790fe6060f1SDimitry Andric Imm = CurDAG->getTargetConstant(ImmVal, SDLoc(N), Subtarget->getXLenVT()); 1791fe6060f1SDimitry Andric return true; 1792fe6060f1SDimitry Andric } 1793fe6060f1SDimitry Andric 1794fe6060f1SDimitry Andric return false; 1795fe6060f1SDimitry Andric } 1796fe6060f1SDimitry Andric 17970b57cec5SDimitry Andric // Merge an ADDI into the offset of a load/store instruction where possible. 17985ffd83dbSDimitry Andric // (load (addi base, off1), off2) -> (load base, off1+off2) 17995ffd83dbSDimitry Andric // (store val, (addi base, off1), off2) -> (store val, base, off1+off2) 18005ffd83dbSDimitry Andric // This is possible when off1+off2 fits a 12-bit immediate. 1801*349cc55cSDimitry Andric bool RISCVDAGToDAGISel::doPeepholeLoadStoreADDI(SDNode *N) { 18020b57cec5SDimitry Andric int OffsetOpIdx; 18030b57cec5SDimitry Andric int BaseOpIdx; 18040b57cec5SDimitry Andric 18050b57cec5SDimitry Andric // Only attempt this optimisation for I-type loads and S-type stores. 18060b57cec5SDimitry Andric switch (N->getMachineOpcode()) { 18070b57cec5SDimitry Andric default: 1808*349cc55cSDimitry Andric return false; 18090b57cec5SDimitry Andric case RISCV::LB: 18100b57cec5SDimitry Andric case RISCV::LH: 18110b57cec5SDimitry Andric case RISCV::LW: 18120b57cec5SDimitry Andric case RISCV::LBU: 18130b57cec5SDimitry Andric case RISCV::LHU: 18140b57cec5SDimitry Andric case RISCV::LWU: 18150b57cec5SDimitry Andric case RISCV::LD: 1816e8d8bef9SDimitry Andric case RISCV::FLH: 18170b57cec5SDimitry Andric case RISCV::FLW: 18180b57cec5SDimitry Andric case RISCV::FLD: 18190b57cec5SDimitry Andric BaseOpIdx = 0; 18200b57cec5SDimitry Andric OffsetOpIdx = 1; 18210b57cec5SDimitry Andric break; 18220b57cec5SDimitry Andric case RISCV::SB: 18230b57cec5SDimitry Andric case RISCV::SH: 18240b57cec5SDimitry Andric case RISCV::SW: 18250b57cec5SDimitry Andric case RISCV::SD: 1826e8d8bef9SDimitry Andric case RISCV::FSH: 18270b57cec5SDimitry Andric case RISCV::FSW: 18280b57cec5SDimitry Andric case RISCV::FSD: 18290b57cec5SDimitry Andric BaseOpIdx = 1; 18300b57cec5SDimitry Andric OffsetOpIdx = 2; 18310b57cec5SDimitry Andric break; 18320b57cec5SDimitry Andric } 18330b57cec5SDimitry Andric 18345ffd83dbSDimitry Andric if (!isa<ConstantSDNode>(N->getOperand(OffsetOpIdx))) 1835*349cc55cSDimitry Andric return false; 18360b57cec5SDimitry Andric 18370b57cec5SDimitry Andric SDValue Base = N->getOperand(BaseOpIdx); 18380b57cec5SDimitry Andric 18390b57cec5SDimitry Andric // If the base is an ADDI, we can merge it in to the load/store. 18400b57cec5SDimitry Andric if (!Base.isMachineOpcode() || Base.getMachineOpcode() != RISCV::ADDI) 1841*349cc55cSDimitry Andric return false; 18420b57cec5SDimitry Andric 18430b57cec5SDimitry Andric SDValue ImmOperand = Base.getOperand(1); 18445ffd83dbSDimitry Andric uint64_t Offset2 = N->getConstantOperandVal(OffsetOpIdx); 18450b57cec5SDimitry Andric 1846fe6060f1SDimitry Andric if (auto *Const = dyn_cast<ConstantSDNode>(ImmOperand)) { 18475ffd83dbSDimitry Andric int64_t Offset1 = Const->getSExtValue(); 18485ffd83dbSDimitry Andric int64_t CombinedOffset = Offset1 + Offset2; 18495ffd83dbSDimitry Andric if (!isInt<12>(CombinedOffset)) 1850*349cc55cSDimitry Andric return false; 18515ffd83dbSDimitry Andric ImmOperand = CurDAG->getTargetConstant(CombinedOffset, SDLoc(ImmOperand), 18525ffd83dbSDimitry Andric ImmOperand.getValueType()); 1853fe6060f1SDimitry Andric } else if (auto *GA = dyn_cast<GlobalAddressSDNode>(ImmOperand)) { 18545ffd83dbSDimitry Andric // If the off1 in (addi base, off1) is a global variable's address (its 18555ffd83dbSDimitry Andric // low part, really), then we can rely on the alignment of that variable 18565ffd83dbSDimitry Andric // to provide a margin of safety before off1 can overflow the 12 bits. 18575ffd83dbSDimitry Andric // Check if off2 falls within that margin; if so off1+off2 can't overflow. 18585ffd83dbSDimitry Andric const DataLayout &DL = CurDAG->getDataLayout(); 18595ffd83dbSDimitry Andric Align Alignment = GA->getGlobal()->getPointerAlignment(DL); 18605ffd83dbSDimitry Andric if (Offset2 != 0 && Alignment <= Offset2) 1861*349cc55cSDimitry Andric return false; 18625ffd83dbSDimitry Andric int64_t Offset1 = GA->getOffset(); 18635ffd83dbSDimitry Andric int64_t CombinedOffset = Offset1 + Offset2; 18640b57cec5SDimitry Andric ImmOperand = CurDAG->getTargetGlobalAddress( 18650b57cec5SDimitry Andric GA->getGlobal(), SDLoc(ImmOperand), ImmOperand.getValueType(), 18665ffd83dbSDimitry Andric CombinedOffset, GA->getTargetFlags()); 1867fe6060f1SDimitry Andric } else if (auto *CP = dyn_cast<ConstantPoolSDNode>(ImmOperand)) { 18685ffd83dbSDimitry Andric // Ditto. 18695ffd83dbSDimitry Andric Align Alignment = CP->getAlign(); 18705ffd83dbSDimitry Andric if (Offset2 != 0 && Alignment <= Offset2) 1871*349cc55cSDimitry Andric return false; 18725ffd83dbSDimitry Andric int64_t Offset1 = CP->getOffset(); 18735ffd83dbSDimitry Andric int64_t CombinedOffset = Offset1 + Offset2; 18745ffd83dbSDimitry Andric ImmOperand = CurDAG->getTargetConstantPool( 18755ffd83dbSDimitry Andric CP->getConstVal(), ImmOperand.getValueType(), CP->getAlign(), 18765ffd83dbSDimitry Andric CombinedOffset, CP->getTargetFlags()); 18770b57cec5SDimitry Andric } else { 1878*349cc55cSDimitry Andric return false; 18790b57cec5SDimitry Andric } 18800b57cec5SDimitry Andric 18810b57cec5SDimitry Andric LLVM_DEBUG(dbgs() << "Folding add-immediate into mem-op:\nBase: "); 18820b57cec5SDimitry Andric LLVM_DEBUG(Base->dump(CurDAG)); 18830b57cec5SDimitry Andric LLVM_DEBUG(dbgs() << "\nN: "); 18840b57cec5SDimitry Andric LLVM_DEBUG(N->dump(CurDAG)); 18850b57cec5SDimitry Andric LLVM_DEBUG(dbgs() << "\n"); 18860b57cec5SDimitry Andric 18870b57cec5SDimitry Andric // Modify the offset operand of the load/store. 18880b57cec5SDimitry Andric if (BaseOpIdx == 0) // Load 18890b57cec5SDimitry Andric CurDAG->UpdateNodeOperands(N, Base.getOperand(0), ImmOperand, 18900b57cec5SDimitry Andric N->getOperand(2)); 18910b57cec5SDimitry Andric else // Store 18920b57cec5SDimitry Andric CurDAG->UpdateNodeOperands(N, N->getOperand(0), Base.getOperand(0), 18930b57cec5SDimitry Andric ImmOperand, N->getOperand(3)); 18940b57cec5SDimitry Andric 1895*349cc55cSDimitry Andric return true; 18960b57cec5SDimitry Andric } 1897*349cc55cSDimitry Andric 1898*349cc55cSDimitry Andric // Try to remove sext.w if the input is a W instruction or can be made into 1899*349cc55cSDimitry Andric // a W instruction cheaply. 1900*349cc55cSDimitry Andric bool RISCVDAGToDAGISel::doPeepholeSExtW(SDNode *N) { 1901*349cc55cSDimitry Andric // Look for the sext.w pattern, addiw rd, rs1, 0. 1902*349cc55cSDimitry Andric if (N->getMachineOpcode() != RISCV::ADDIW || 1903*349cc55cSDimitry Andric !isNullConstant(N->getOperand(1))) 1904*349cc55cSDimitry Andric return false; 1905*349cc55cSDimitry Andric 1906*349cc55cSDimitry Andric SDValue N0 = N->getOperand(0); 1907*349cc55cSDimitry Andric if (!N0.isMachineOpcode()) 1908*349cc55cSDimitry Andric return false; 1909*349cc55cSDimitry Andric 1910*349cc55cSDimitry Andric switch (N0.getMachineOpcode()) { 1911*349cc55cSDimitry Andric default: 1912*349cc55cSDimitry Andric break; 1913*349cc55cSDimitry Andric case RISCV::ADD: 1914*349cc55cSDimitry Andric case RISCV::ADDI: 1915*349cc55cSDimitry Andric case RISCV::SUB: 1916*349cc55cSDimitry Andric case RISCV::MUL: 1917*349cc55cSDimitry Andric case RISCV::SLLI: { 1918*349cc55cSDimitry Andric // Convert sext.w+add/sub/mul to their W instructions. This will create 1919*349cc55cSDimitry Andric // a new independent instruction. This improves latency. 1920*349cc55cSDimitry Andric unsigned Opc; 1921*349cc55cSDimitry Andric switch (N0.getMachineOpcode()) { 1922*349cc55cSDimitry Andric default: 1923*349cc55cSDimitry Andric llvm_unreachable("Unexpected opcode!"); 1924*349cc55cSDimitry Andric case RISCV::ADD: Opc = RISCV::ADDW; break; 1925*349cc55cSDimitry Andric case RISCV::ADDI: Opc = RISCV::ADDIW; break; 1926*349cc55cSDimitry Andric case RISCV::SUB: Opc = RISCV::SUBW; break; 1927*349cc55cSDimitry Andric case RISCV::MUL: Opc = RISCV::MULW; break; 1928*349cc55cSDimitry Andric case RISCV::SLLI: Opc = RISCV::SLLIW; break; 1929*349cc55cSDimitry Andric } 1930*349cc55cSDimitry Andric 1931*349cc55cSDimitry Andric SDValue N00 = N0.getOperand(0); 1932*349cc55cSDimitry Andric SDValue N01 = N0.getOperand(1); 1933*349cc55cSDimitry Andric 1934*349cc55cSDimitry Andric // Shift amount needs to be uimm5. 1935*349cc55cSDimitry Andric if (N0.getMachineOpcode() == RISCV::SLLI && 1936*349cc55cSDimitry Andric !isUInt<5>(cast<ConstantSDNode>(N01)->getSExtValue())) 1937*349cc55cSDimitry Andric break; 1938*349cc55cSDimitry Andric 1939*349cc55cSDimitry Andric SDNode *Result = 1940*349cc55cSDimitry Andric CurDAG->getMachineNode(Opc, SDLoc(N), N->getValueType(0), 1941*349cc55cSDimitry Andric N00, N01); 1942*349cc55cSDimitry Andric ReplaceUses(N, Result); 1943*349cc55cSDimitry Andric return true; 1944*349cc55cSDimitry Andric } 1945*349cc55cSDimitry Andric case RISCV::ADDW: 1946*349cc55cSDimitry Andric case RISCV::ADDIW: 1947*349cc55cSDimitry Andric case RISCV::SUBW: 1948*349cc55cSDimitry Andric case RISCV::MULW: 1949*349cc55cSDimitry Andric case RISCV::SLLIW: 1950*349cc55cSDimitry Andric // Result is already sign extended just remove the sext.w. 1951*349cc55cSDimitry Andric // NOTE: We only handle the nodes that are selected with hasAllWUsers. 1952*349cc55cSDimitry Andric ReplaceUses(N, N0.getNode()); 1953*349cc55cSDimitry Andric return true; 1954*349cc55cSDimitry Andric } 1955*349cc55cSDimitry Andric 1956*349cc55cSDimitry Andric return false; 19570b57cec5SDimitry Andric } 19580b57cec5SDimitry Andric 19590b57cec5SDimitry Andric // This pass converts a legalized DAG into a RISCV-specific DAG, ready 19600b57cec5SDimitry Andric // for instruction scheduling. 19610b57cec5SDimitry Andric FunctionPass *llvm::createRISCVISelDag(RISCVTargetMachine &TM) { 19620b57cec5SDimitry Andric return new RISCVDAGToDAGISel(TM); 19630b57cec5SDimitry Andric } 1964