1 //===-- lib/CodeGen/GlobalISel/InlineAsmLowering.cpp ----------------------===// 2 // 3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. 4 // See https://llvm.org/LICENSE.txt for license information. 5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception 6 // 7 //===----------------------------------------------------------------------===// 8 /// 9 /// \file 10 /// This file implements the lowering from LLVM IR inline asm to MIR INLINEASM 11 /// 12 //===----------------------------------------------------------------------===// 13 14 #include "llvm/CodeGen/GlobalISel/InlineAsmLowering.h" 15 #include "llvm/CodeGen/GlobalISel/MachineIRBuilder.h" 16 #include "llvm/CodeGen/MachineOperand.h" 17 #include "llvm/CodeGen/MachineRegisterInfo.h" 18 #include "llvm/CodeGen/TargetLowering.h" 19 #include "llvm/IR/Module.h" 20 21 #define DEBUG_TYPE "inline-asm-lowering" 22 23 using namespace llvm; 24 25 void InlineAsmLowering::anchor() {} 26 27 namespace { 28 29 /// GISelAsmOperandInfo - This contains information for each constraint that we 30 /// are lowering. 31 class GISelAsmOperandInfo : public TargetLowering::AsmOperandInfo { 32 public: 33 /// Regs - If this is a register or register class operand, this 34 /// contains the set of assigned registers corresponding to the operand. 35 SmallVector<Register, 1> Regs; 36 37 explicit GISelAsmOperandInfo(const TargetLowering::AsmOperandInfo &Info) 38 : TargetLowering::AsmOperandInfo(Info) {} 39 }; 40 41 using GISelAsmOperandInfoVector = SmallVector<GISelAsmOperandInfo, 16>; 42 43 class ExtraFlags { 44 unsigned Flags = 0; 45 46 public: 47 explicit ExtraFlags(const CallBase &CB) { 48 const InlineAsm *IA = cast<InlineAsm>(CB.getCalledOperand()); 49 if (IA->hasSideEffects()) 50 Flags |= InlineAsm::Extra_HasSideEffects; 51 if (IA->isAlignStack()) 52 Flags |= InlineAsm::Extra_IsAlignStack; 53 if (CB.isConvergent()) 54 Flags |= InlineAsm::Extra_IsConvergent; 55 Flags |= IA->getDialect() * InlineAsm::Extra_AsmDialect; 56 } 57 58 void update(const TargetLowering::AsmOperandInfo &OpInfo) { 59 // Ideally, we would only check against memory constraints. However, the 60 // meaning of an Other constraint can be target-specific and we can't easily 61 // reason about it. Therefore, be conservative and set MayLoad/MayStore 62 // for Other constraints as well. 63 if (OpInfo.ConstraintType == TargetLowering::C_Memory || 64 OpInfo.ConstraintType == TargetLowering::C_Other) { 65 if (OpInfo.Type == InlineAsm::isInput) 66 Flags |= InlineAsm::Extra_MayLoad; 67 else if (OpInfo.Type == InlineAsm::isOutput) 68 Flags |= InlineAsm::Extra_MayStore; 69 else if (OpInfo.Type == InlineAsm::isClobber) 70 Flags |= (InlineAsm::Extra_MayLoad | InlineAsm::Extra_MayStore); 71 } 72 } 73 74 unsigned get() const { return Flags; } 75 }; 76 77 } // namespace 78 79 /// Assign virtual/physical registers for the specified register operand. 80 static void getRegistersForValue(MachineFunction &MF, 81 MachineIRBuilder &MIRBuilder, 82 GISelAsmOperandInfo &OpInfo, 83 GISelAsmOperandInfo &RefOpInfo) { 84 85 const TargetLowering &TLI = *MF.getSubtarget().getTargetLowering(); 86 const TargetRegisterInfo &TRI = *MF.getSubtarget().getRegisterInfo(); 87 88 // No work to do for memory operations. 89 if (OpInfo.ConstraintType == TargetLowering::C_Memory) 90 return; 91 92 // If this is a constraint for a single physreg, or a constraint for a 93 // register class, find it. 94 Register AssignedReg; 95 const TargetRegisterClass *RC; 96 std::tie(AssignedReg, RC) = TLI.getRegForInlineAsmConstraint( 97 &TRI, RefOpInfo.ConstraintCode, RefOpInfo.ConstraintVT); 98 // RC is unset only on failure. Return immediately. 99 if (!RC) 100 return; 101 102 // No need to allocate a matching input constraint since the constraint it's 103 // matching to has already been allocated. 104 if (OpInfo.isMatchingInputConstraint()) 105 return; 106 107 // Initialize NumRegs. 108 unsigned NumRegs = 1; 109 if (OpInfo.ConstraintVT != MVT::Other) 110 NumRegs = 111 TLI.getNumRegisters(MF.getFunction().getContext(), OpInfo.ConstraintVT); 112 113 // If this is a constraint for a specific physical register, but the type of 114 // the operand requires more than one register to be passed, we allocate the 115 // required amount of physical registers, starting from the selected physical 116 // register. 117 // For this, first retrieve a register iterator for the given register class 118 TargetRegisterClass::iterator I = RC->begin(); 119 MachineRegisterInfo &RegInfo = MF.getRegInfo(); 120 121 // Advance the iterator to the assigned register (if set) 122 if (AssignedReg) { 123 for (; *I != AssignedReg; ++I) 124 assert(I != RC->end() && "AssignedReg should be a member of provided RC"); 125 } 126 127 // Finally, assign the registers. If the AssignedReg isn't set, create virtual 128 // registers with the provided register class 129 for (; NumRegs; --NumRegs, ++I) { 130 assert(I != RC->end() && "Ran out of registers to allocate!"); 131 Register R = AssignedReg ? Register(*I) : RegInfo.createVirtualRegister(RC); 132 OpInfo.Regs.push_back(R); 133 } 134 } 135 136 /// Return an integer indicating how general CT is. 137 static unsigned getConstraintGenerality(TargetLowering::ConstraintType CT) { 138 switch (CT) { 139 case TargetLowering::C_Immediate: 140 case TargetLowering::C_Other: 141 case TargetLowering::C_Unknown: 142 return 0; 143 case TargetLowering::C_Register: 144 return 1; 145 case TargetLowering::C_RegisterClass: 146 return 2; 147 case TargetLowering::C_Memory: 148 case TargetLowering::C_Address: 149 return 3; 150 } 151 llvm_unreachable("Invalid constraint type"); 152 } 153 154 static void chooseConstraint(TargetLowering::AsmOperandInfo &OpInfo, 155 const TargetLowering *TLI) { 156 assert(OpInfo.Codes.size() > 1 && "Doesn't have multiple constraint options"); 157 unsigned BestIdx = 0; 158 TargetLowering::ConstraintType BestType = TargetLowering::C_Unknown; 159 int BestGenerality = -1; 160 161 // Loop over the options, keeping track of the most general one. 162 for (unsigned i = 0, e = OpInfo.Codes.size(); i != e; ++i) { 163 TargetLowering::ConstraintType CType = 164 TLI->getConstraintType(OpInfo.Codes[i]); 165 166 // Indirect 'other' or 'immediate' constraints are not allowed. 167 if (OpInfo.isIndirect && !(CType == TargetLowering::C_Memory || 168 CType == TargetLowering::C_Register || 169 CType == TargetLowering::C_RegisterClass)) 170 continue; 171 172 // If this is an 'other' or 'immediate' constraint, see if the operand is 173 // valid for it. For example, on X86 we might have an 'rI' constraint. If 174 // the operand is an integer in the range [0..31] we want to use I (saving a 175 // load of a register), otherwise we must use 'r'. 176 if (CType == TargetLowering::C_Other || 177 CType == TargetLowering::C_Immediate) { 178 assert(OpInfo.Codes[i].size() == 1 && 179 "Unhandled multi-letter 'other' constraint"); 180 // FIXME: prefer immediate constraints if the target allows it 181 } 182 183 // Things with matching constraints can only be registers, per gcc 184 // documentation. This mainly affects "g" constraints. 185 if (CType == TargetLowering::C_Memory && OpInfo.hasMatchingInput()) 186 continue; 187 188 // This constraint letter is more general than the previous one, use it. 189 int Generality = getConstraintGenerality(CType); 190 if (Generality > BestGenerality) { 191 BestType = CType; 192 BestIdx = i; 193 BestGenerality = Generality; 194 } 195 } 196 197 OpInfo.ConstraintCode = OpInfo.Codes[BestIdx]; 198 OpInfo.ConstraintType = BestType; 199 } 200 201 static void computeConstraintToUse(const TargetLowering *TLI, 202 TargetLowering::AsmOperandInfo &OpInfo) { 203 assert(!OpInfo.Codes.empty() && "Must have at least one constraint"); 204 205 // Single-letter constraints ('r') are very common. 206 if (OpInfo.Codes.size() == 1) { 207 OpInfo.ConstraintCode = OpInfo.Codes[0]; 208 OpInfo.ConstraintType = TLI->getConstraintType(OpInfo.ConstraintCode); 209 } else { 210 chooseConstraint(OpInfo, TLI); 211 } 212 213 // 'X' matches anything. 214 if (OpInfo.ConstraintCode == "X" && OpInfo.CallOperandVal) { 215 // Labels and constants are handled elsewhere ('X' is the only thing 216 // that matches labels). For Functions, the type here is the type of 217 // the result, which is not what we want to look at; leave them alone. 218 Value *Val = OpInfo.CallOperandVal; 219 if (isa<BasicBlock>(Val) || isa<ConstantInt>(Val) || isa<Function>(Val)) 220 return; 221 222 // Otherwise, try to resolve it to something we know about by looking at 223 // the actual operand type. 224 if (const char *Repl = TLI->LowerXConstraint(OpInfo.ConstraintVT)) { 225 OpInfo.ConstraintCode = Repl; 226 OpInfo.ConstraintType = TLI->getConstraintType(OpInfo.ConstraintCode); 227 } 228 } 229 } 230 231 static unsigned getNumOpRegs(const MachineInstr &I, unsigned OpIdx) { 232 unsigned Flag = I.getOperand(OpIdx).getImm(); 233 return InlineAsm::getNumOperandRegisters(Flag); 234 } 235 236 static bool buildAnyextOrCopy(Register Dst, Register Src, 237 MachineIRBuilder &MIRBuilder) { 238 const TargetRegisterInfo *TRI = 239 MIRBuilder.getMF().getSubtarget().getRegisterInfo(); 240 MachineRegisterInfo *MRI = MIRBuilder.getMRI(); 241 242 auto SrcTy = MRI->getType(Src); 243 if (!SrcTy.isValid()) { 244 LLVM_DEBUG(dbgs() << "Source type for copy is not valid\n"); 245 return false; 246 } 247 unsigned SrcSize = TRI->getRegSizeInBits(Src, *MRI); 248 unsigned DstSize = TRI->getRegSizeInBits(Dst, *MRI); 249 250 if (DstSize < SrcSize) { 251 LLVM_DEBUG(dbgs() << "Input can't fit in destination reg class\n"); 252 return false; 253 } 254 255 // Attempt to anyext small scalar sources. 256 if (DstSize > SrcSize) { 257 if (!SrcTy.isScalar()) { 258 LLVM_DEBUG(dbgs() << "Can't extend non-scalar input to size of" 259 "destination register class\n"); 260 return false; 261 } 262 Src = MIRBuilder.buildAnyExt(LLT::scalar(DstSize), Src).getReg(0); 263 } 264 265 MIRBuilder.buildCopy(Dst, Src); 266 return true; 267 } 268 269 bool InlineAsmLowering::lowerInlineAsm( 270 MachineIRBuilder &MIRBuilder, const CallBase &Call, 271 std::function<ArrayRef<Register>(const Value &Val)> GetOrCreateVRegs) 272 const { 273 const InlineAsm *IA = cast<InlineAsm>(Call.getCalledOperand()); 274 275 /// ConstraintOperands - Information about all of the constraints. 276 GISelAsmOperandInfoVector ConstraintOperands; 277 278 MachineFunction &MF = MIRBuilder.getMF(); 279 const Function &F = MF.getFunction(); 280 const DataLayout &DL = F.getParent()->getDataLayout(); 281 const TargetRegisterInfo *TRI = MF.getSubtarget().getRegisterInfo(); 282 283 MachineRegisterInfo *MRI = MIRBuilder.getMRI(); 284 285 TargetLowering::AsmOperandInfoVector TargetConstraints = 286 TLI->ParseConstraints(DL, TRI, Call); 287 288 ExtraFlags ExtraInfo(Call); 289 unsigned ArgNo = 0; // ArgNo - The argument of the CallInst. 290 unsigned ResNo = 0; // ResNo - The result number of the next output. 291 for (auto &T : TargetConstraints) { 292 ConstraintOperands.push_back(GISelAsmOperandInfo(T)); 293 GISelAsmOperandInfo &OpInfo = ConstraintOperands.back(); 294 295 // Compute the value type for each operand. 296 if (OpInfo.hasArg()) { 297 OpInfo.CallOperandVal = const_cast<Value *>(Call.getArgOperand(ArgNo)); 298 299 if (isa<BasicBlock>(OpInfo.CallOperandVal)) { 300 LLVM_DEBUG(dbgs() << "Basic block input operands not supported yet\n"); 301 return false; 302 } 303 304 Type *OpTy = OpInfo.CallOperandVal->getType(); 305 306 // If this is an indirect operand, the operand is a pointer to the 307 // accessed type. 308 if (OpInfo.isIndirect) { 309 OpTy = Call.getParamElementType(ArgNo); 310 assert(OpTy && "Indirect operand must have elementtype attribute"); 311 } 312 313 // FIXME: Support aggregate input operands 314 if (!OpTy->isSingleValueType()) { 315 LLVM_DEBUG( 316 dbgs() << "Aggregate input operands are not supported yet\n"); 317 return false; 318 } 319 320 OpInfo.ConstraintVT = 321 TLI->getAsmOperandValueType(DL, OpTy, true).getSimpleVT(); 322 ++ArgNo; 323 } else if (OpInfo.Type == InlineAsm::isOutput && !OpInfo.isIndirect) { 324 assert(!Call.getType()->isVoidTy() && "Bad inline asm!"); 325 if (StructType *STy = dyn_cast<StructType>(Call.getType())) { 326 OpInfo.ConstraintVT = 327 TLI->getSimpleValueType(DL, STy->getElementType(ResNo)); 328 } else { 329 assert(ResNo == 0 && "Asm only has one result!"); 330 OpInfo.ConstraintVT = 331 TLI->getAsmOperandValueType(DL, Call.getType()).getSimpleVT(); 332 } 333 ++ResNo; 334 } else { 335 assert(OpInfo.Type != InlineAsm::isLabel && 336 "GlobalISel currently doesn't support callbr"); 337 OpInfo.ConstraintVT = MVT::Other; 338 } 339 340 if (OpInfo.ConstraintVT == MVT::i64x8) 341 return false; 342 343 // Compute the constraint code and ConstraintType to use. 344 computeConstraintToUse(TLI, OpInfo); 345 346 // The selected constraint type might expose new sideeffects 347 ExtraInfo.update(OpInfo); 348 } 349 350 // At this point, all operand types are decided. 351 // Create the MachineInstr, but don't insert it yet since input 352 // operands still need to insert instructions before this one 353 auto Inst = MIRBuilder.buildInstrNoInsert(TargetOpcode::INLINEASM) 354 .addExternalSymbol(IA->getAsmString().c_str()) 355 .addImm(ExtraInfo.get()); 356 357 // Starting from this operand: flag followed by register(s) will be added as 358 // operands to Inst for each constraint. Used for matching input constraints. 359 unsigned StartIdx = Inst->getNumOperands(); 360 361 // Collects the output operands for later processing 362 GISelAsmOperandInfoVector OutputOperands; 363 364 for (auto &OpInfo : ConstraintOperands) { 365 GISelAsmOperandInfo &RefOpInfo = 366 OpInfo.isMatchingInputConstraint() 367 ? ConstraintOperands[OpInfo.getMatchedOperand()] 368 : OpInfo; 369 370 // Assign registers for register operands 371 getRegistersForValue(MF, MIRBuilder, OpInfo, RefOpInfo); 372 373 switch (OpInfo.Type) { 374 case InlineAsm::isOutput: 375 if (OpInfo.ConstraintType == TargetLowering::C_Memory) { 376 unsigned ConstraintID = 377 TLI->getInlineAsmMemConstraint(OpInfo.ConstraintCode); 378 assert(ConstraintID != InlineAsm::Constraint_Unknown && 379 "Failed to convert memory constraint code to constraint id."); 380 381 // Add information to the INLINEASM instruction to know about this 382 // output. 383 unsigned OpFlags = InlineAsm::getFlagWord(InlineAsm::Kind_Mem, 1); 384 OpFlags = InlineAsm::getFlagWordForMem(OpFlags, ConstraintID); 385 Inst.addImm(OpFlags); 386 ArrayRef<Register> SourceRegs = 387 GetOrCreateVRegs(*OpInfo.CallOperandVal); 388 assert( 389 SourceRegs.size() == 1 && 390 "Expected the memory output to fit into a single virtual register"); 391 Inst.addReg(SourceRegs[0]); 392 } else { 393 // Otherwise, this outputs to a register (directly for C_Register / 394 // C_RegisterClass/C_Other. 395 assert(OpInfo.ConstraintType == TargetLowering::C_Register || 396 OpInfo.ConstraintType == TargetLowering::C_RegisterClass || 397 OpInfo.ConstraintType == TargetLowering::C_Other); 398 399 // Find a register that we can use. 400 if (OpInfo.Regs.empty()) { 401 LLVM_DEBUG(dbgs() 402 << "Couldn't allocate output register for constraint\n"); 403 return false; 404 } 405 406 // Add information to the INLINEASM instruction to know that this 407 // register is set. 408 unsigned Flag = InlineAsm::getFlagWord( 409 OpInfo.isEarlyClobber ? InlineAsm::Kind_RegDefEarlyClobber 410 : InlineAsm::Kind_RegDef, 411 OpInfo.Regs.size()); 412 if (OpInfo.Regs.front().isVirtual()) { 413 // Put the register class of the virtual registers in the flag word. 414 // That way, later passes can recompute register class constraints for 415 // inline assembly as well as normal instructions. Don't do this for 416 // tied operands that can use the regclass information from the def. 417 const TargetRegisterClass *RC = MRI->getRegClass(OpInfo.Regs.front()); 418 Flag = InlineAsm::getFlagWordForRegClass(Flag, RC->getID()); 419 } 420 421 Inst.addImm(Flag); 422 423 for (Register Reg : OpInfo.Regs) { 424 Inst.addReg(Reg, 425 RegState::Define | getImplRegState(Reg.isPhysical()) | 426 (OpInfo.isEarlyClobber ? RegState::EarlyClobber : 0)); 427 } 428 429 // Remember this output operand for later processing 430 OutputOperands.push_back(OpInfo); 431 } 432 433 break; 434 case InlineAsm::isInput: 435 case InlineAsm::isLabel: { 436 if (OpInfo.isMatchingInputConstraint()) { 437 unsigned DefIdx = OpInfo.getMatchedOperand(); 438 // Find operand with register def that corresponds to DefIdx. 439 unsigned InstFlagIdx = StartIdx; 440 for (unsigned i = 0; i < DefIdx; ++i) 441 InstFlagIdx += getNumOpRegs(*Inst, InstFlagIdx) + 1; 442 assert(getNumOpRegs(*Inst, InstFlagIdx) == 1 && "Wrong flag"); 443 444 unsigned MatchedOperandFlag = Inst->getOperand(InstFlagIdx).getImm(); 445 if (InlineAsm::isMemKind(MatchedOperandFlag)) { 446 LLVM_DEBUG(dbgs() << "Matching input constraint to mem operand not " 447 "supported. This should be target specific.\n"); 448 return false; 449 } 450 if (!InlineAsm::isRegDefKind(MatchedOperandFlag) && 451 !InlineAsm::isRegDefEarlyClobberKind(MatchedOperandFlag)) { 452 LLVM_DEBUG(dbgs() << "Unknown matching constraint\n"); 453 return false; 454 } 455 456 // We want to tie input to register in next operand. 457 unsigned DefRegIdx = InstFlagIdx + 1; 458 Register Def = Inst->getOperand(DefRegIdx).getReg(); 459 460 ArrayRef<Register> SrcRegs = GetOrCreateVRegs(*OpInfo.CallOperandVal); 461 assert(SrcRegs.size() == 1 && "Single register is expected here"); 462 463 // When Def is physreg: use given input. 464 Register In = SrcRegs[0]; 465 // When Def is vreg: copy input to new vreg with same reg class as Def. 466 if (Def.isVirtual()) { 467 In = MRI->createVirtualRegister(MRI->getRegClass(Def)); 468 if (!buildAnyextOrCopy(In, SrcRegs[0], MIRBuilder)) 469 return false; 470 } 471 472 // Add Flag and input register operand (In) to Inst. Tie In to Def. 473 unsigned UseFlag = InlineAsm::getFlagWord(InlineAsm::Kind_RegUse, 1); 474 unsigned Flag = InlineAsm::getFlagWordForMatchingOp(UseFlag, DefIdx); 475 Inst.addImm(Flag); 476 Inst.addReg(In); 477 Inst->tieOperands(DefRegIdx, Inst->getNumOperands() - 1); 478 break; 479 } 480 481 if (OpInfo.ConstraintType == TargetLowering::C_Other && 482 OpInfo.isIndirect) { 483 LLVM_DEBUG(dbgs() << "Indirect input operands with unknown constraint " 484 "not supported yet\n"); 485 return false; 486 } 487 488 if (OpInfo.ConstraintType == TargetLowering::C_Immediate || 489 OpInfo.ConstraintType == TargetLowering::C_Other) { 490 491 std::vector<MachineOperand> Ops; 492 if (!lowerAsmOperandForConstraint(OpInfo.CallOperandVal, 493 OpInfo.ConstraintCode, Ops, 494 MIRBuilder)) { 495 LLVM_DEBUG(dbgs() << "Don't support constraint: " 496 << OpInfo.ConstraintCode << " yet\n"); 497 return false; 498 } 499 500 assert(Ops.size() > 0 && 501 "Expected constraint to be lowered to at least one operand"); 502 503 // Add information to the INLINEASM node to know about this input. 504 unsigned OpFlags = 505 InlineAsm::getFlagWord(InlineAsm::Kind_Imm, Ops.size()); 506 Inst.addImm(OpFlags); 507 Inst.add(Ops); 508 break; 509 } 510 511 if (OpInfo.ConstraintType == TargetLowering::C_Memory) { 512 513 if (!OpInfo.isIndirect) { 514 LLVM_DEBUG(dbgs() 515 << "Cannot indirectify memory input operands yet\n"); 516 return false; 517 } 518 519 assert(OpInfo.isIndirect && "Operand must be indirect to be a mem!"); 520 521 unsigned ConstraintID = 522 TLI->getInlineAsmMemConstraint(OpInfo.ConstraintCode); 523 unsigned OpFlags = InlineAsm::getFlagWord(InlineAsm::Kind_Mem, 1); 524 OpFlags = InlineAsm::getFlagWordForMem(OpFlags, ConstraintID); 525 Inst.addImm(OpFlags); 526 ArrayRef<Register> SourceRegs = 527 GetOrCreateVRegs(*OpInfo.CallOperandVal); 528 assert( 529 SourceRegs.size() == 1 && 530 "Expected the memory input to fit into a single virtual register"); 531 Inst.addReg(SourceRegs[0]); 532 break; 533 } 534 535 assert((OpInfo.ConstraintType == TargetLowering::C_RegisterClass || 536 OpInfo.ConstraintType == TargetLowering::C_Register) && 537 "Unknown constraint type!"); 538 539 if (OpInfo.isIndirect) { 540 LLVM_DEBUG(dbgs() << "Can't handle indirect register inputs yet " 541 "for constraint '" 542 << OpInfo.ConstraintCode << "'\n"); 543 return false; 544 } 545 546 // Copy the input into the appropriate registers. 547 if (OpInfo.Regs.empty()) { 548 LLVM_DEBUG( 549 dbgs() 550 << "Couldn't allocate input register for register constraint\n"); 551 return false; 552 } 553 554 unsigned NumRegs = OpInfo.Regs.size(); 555 ArrayRef<Register> SourceRegs = GetOrCreateVRegs(*OpInfo.CallOperandVal); 556 assert(NumRegs == SourceRegs.size() && 557 "Expected the number of input registers to match the number of " 558 "source registers"); 559 560 if (NumRegs > 1) { 561 LLVM_DEBUG(dbgs() << "Input operands with multiple input registers are " 562 "not supported yet\n"); 563 return false; 564 } 565 566 unsigned Flag = InlineAsm::getFlagWord(InlineAsm::Kind_RegUse, NumRegs); 567 if (OpInfo.Regs.front().isVirtual()) { 568 // Put the register class of the virtual registers in the flag word. 569 const TargetRegisterClass *RC = MRI->getRegClass(OpInfo.Regs.front()); 570 Flag = InlineAsm::getFlagWordForRegClass(Flag, RC->getID()); 571 } 572 Inst.addImm(Flag); 573 if (!buildAnyextOrCopy(OpInfo.Regs[0], SourceRegs[0], MIRBuilder)) 574 return false; 575 Inst.addReg(OpInfo.Regs[0]); 576 break; 577 } 578 579 case InlineAsm::isClobber: { 580 581 unsigned NumRegs = OpInfo.Regs.size(); 582 if (NumRegs > 0) { 583 unsigned Flag = 584 InlineAsm::getFlagWord(InlineAsm::Kind_Clobber, NumRegs); 585 Inst.addImm(Flag); 586 587 for (Register Reg : OpInfo.Regs) { 588 Inst.addReg(Reg, RegState::Define | RegState::EarlyClobber | 589 getImplRegState(Reg.isPhysical())); 590 } 591 } 592 break; 593 } 594 } 595 } 596 597 if (const MDNode *SrcLoc = Call.getMetadata("srcloc")) 598 Inst.addMetadata(SrcLoc); 599 600 // All inputs are handled, insert the instruction now 601 MIRBuilder.insertInstr(Inst); 602 603 // Finally, copy the output operands into the output registers 604 ArrayRef<Register> ResRegs = GetOrCreateVRegs(Call); 605 if (ResRegs.size() != OutputOperands.size()) { 606 LLVM_DEBUG(dbgs() << "Expected the number of output registers to match the " 607 "number of destination registers\n"); 608 return false; 609 } 610 for (unsigned int i = 0, e = ResRegs.size(); i < e; i++) { 611 GISelAsmOperandInfo &OpInfo = OutputOperands[i]; 612 613 if (OpInfo.Regs.empty()) 614 continue; 615 616 switch (OpInfo.ConstraintType) { 617 case TargetLowering::C_Register: 618 case TargetLowering::C_RegisterClass: { 619 if (OpInfo.Regs.size() > 1) { 620 LLVM_DEBUG(dbgs() << "Output operands with multiple defining " 621 "registers are not supported yet\n"); 622 return false; 623 } 624 625 Register SrcReg = OpInfo.Regs[0]; 626 unsigned SrcSize = TRI->getRegSizeInBits(SrcReg, *MRI); 627 LLT ResTy = MRI->getType(ResRegs[i]); 628 if (ResTy.isScalar() && ResTy.getSizeInBits() < SrcSize) { 629 // First copy the non-typed virtual register into a generic virtual 630 // register 631 Register Tmp1Reg = 632 MRI->createGenericVirtualRegister(LLT::scalar(SrcSize)); 633 MIRBuilder.buildCopy(Tmp1Reg, SrcReg); 634 // Need to truncate the result of the register 635 MIRBuilder.buildTrunc(ResRegs[i], Tmp1Reg); 636 } else if (ResTy.getSizeInBits() == SrcSize) { 637 MIRBuilder.buildCopy(ResRegs[i], SrcReg); 638 } else { 639 LLVM_DEBUG(dbgs() << "Unhandled output operand with " 640 "mismatched register size\n"); 641 return false; 642 } 643 644 break; 645 } 646 case TargetLowering::C_Immediate: 647 case TargetLowering::C_Other: 648 LLVM_DEBUG( 649 dbgs() << "Cannot lower target specific output constraints yet\n"); 650 return false; 651 case TargetLowering::C_Memory: 652 break; // Already handled. 653 case TargetLowering::C_Address: 654 break; // Silence warning. 655 case TargetLowering::C_Unknown: 656 LLVM_DEBUG(dbgs() << "Unexpected unknown constraint\n"); 657 return false; 658 } 659 } 660 661 return true; 662 } 663 664 bool InlineAsmLowering::lowerAsmOperandForConstraint( 665 Value *Val, StringRef Constraint, std::vector<MachineOperand> &Ops, 666 MachineIRBuilder &MIRBuilder) const { 667 if (Constraint.size() > 1) 668 return false; 669 670 char ConstraintLetter = Constraint[0]; 671 switch (ConstraintLetter) { 672 default: 673 return false; 674 case 'i': // Simple Integer or Relocatable Constant 675 case 'n': // immediate integer with a known value. 676 if (ConstantInt *CI = dyn_cast<ConstantInt>(Val)) { 677 assert(CI->getBitWidth() <= 64 && 678 "expected immediate to fit into 64-bits"); 679 // Boolean constants should be zero-extended, others are sign-extended 680 bool IsBool = CI->getBitWidth() == 1; 681 int64_t ExtVal = IsBool ? CI->getZExtValue() : CI->getSExtValue(); 682 Ops.push_back(MachineOperand::CreateImm(ExtVal)); 683 return true; 684 } 685 return false; 686 } 687 } 688