1 //===-- X86AsmParser.cpp - Parse X86 assembly to MCInst instructions ------===// 2 // 3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. 4 // See https://llvm.org/LICENSE.txt for license information. 5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception 6 // 7 //===----------------------------------------------------------------------===// 8 9 #include "MCTargetDesc/X86BaseInfo.h" 10 #include "MCTargetDesc/X86IntelInstPrinter.h" 11 #include "MCTargetDesc/X86MCExpr.h" 12 #include "MCTargetDesc/X86MCTargetDesc.h" 13 #include "MCTargetDesc/X86TargetStreamer.h" 14 #include "TargetInfo/X86TargetInfo.h" 15 #include "X86AsmParserCommon.h" 16 #include "X86Operand.h" 17 #include "llvm/ADT/STLExtras.h" 18 #include "llvm/ADT/SmallString.h" 19 #include "llvm/ADT/SmallVector.h" 20 #include "llvm/ADT/StringSwitch.h" 21 #include "llvm/ADT/Twine.h" 22 #include "llvm/MC/MCContext.h" 23 #include "llvm/MC/MCExpr.h" 24 #include "llvm/MC/MCInst.h" 25 #include "llvm/MC/MCInstrInfo.h" 26 #include "llvm/MC/MCParser/MCAsmLexer.h" 27 #include "llvm/MC/MCParser/MCAsmParser.h" 28 #include "llvm/MC/MCParser/MCParsedAsmOperand.h" 29 #include "llvm/MC/MCParser/MCTargetAsmParser.h" 30 #include "llvm/MC/MCRegisterInfo.h" 31 #include "llvm/MC/MCSection.h" 32 #include "llvm/MC/MCStreamer.h" 33 #include "llvm/MC/MCSubtargetInfo.h" 34 #include "llvm/MC/MCSymbol.h" 35 #include "llvm/MC/TargetRegistry.h" 36 #include "llvm/Support/CommandLine.h" 37 #include "llvm/Support/Compiler.h" 38 #include "llvm/Support/SourceMgr.h" 39 #include "llvm/Support/raw_ostream.h" 40 #include <algorithm> 41 #include <memory> 42 43 using namespace llvm; 44 45 static cl::opt<bool> LVIInlineAsmHardening( 46 "x86-experimental-lvi-inline-asm-hardening", 47 cl::desc("Harden inline assembly code that may be vulnerable to Load Value" 48 " Injection (LVI). This feature is experimental."), cl::Hidden); 49 50 static bool checkScale(unsigned Scale, StringRef &ErrMsg) { 51 if (Scale != 1 && Scale != 2 && Scale != 4 && Scale != 8) { 52 ErrMsg = "scale factor in address must be 1, 2, 4 or 8"; 53 return true; 54 } 55 return false; 56 } 57 58 namespace { 59 60 static const char OpPrecedence[] = { 61 0, // IC_OR 62 1, // IC_XOR 63 2, // IC_AND 64 4, // IC_LSHIFT 65 4, // IC_RSHIFT 66 5, // IC_PLUS 67 5, // IC_MINUS 68 6, // IC_MULTIPLY 69 6, // IC_DIVIDE 70 6, // IC_MOD 71 7, // IC_NOT 72 8, // IC_NEG 73 9, // IC_RPAREN 74 10, // IC_LPAREN 75 0, // IC_IMM 76 0, // IC_REGISTER 77 3, // IC_EQ 78 3, // IC_NE 79 3, // IC_LT 80 3, // IC_LE 81 3, // IC_GT 82 3 // IC_GE 83 }; 84 85 class X86AsmParser : public MCTargetAsmParser { 86 ParseInstructionInfo *InstInfo; 87 bool Code16GCC; 88 unsigned ForcedDataPrefix = 0; 89 90 enum VEXEncoding { 91 VEXEncoding_Default, 92 VEXEncoding_VEX, 93 VEXEncoding_VEX2, 94 VEXEncoding_VEX3, 95 VEXEncoding_EVEX, 96 }; 97 98 VEXEncoding ForcedVEXEncoding = VEXEncoding_Default; 99 100 enum DispEncoding { 101 DispEncoding_Default, 102 DispEncoding_Disp8, 103 DispEncoding_Disp32, 104 }; 105 106 DispEncoding ForcedDispEncoding = DispEncoding_Default; 107 108 private: 109 SMLoc consumeToken() { 110 MCAsmParser &Parser = getParser(); 111 SMLoc Result = Parser.getTok().getLoc(); 112 Parser.Lex(); 113 return Result; 114 } 115 116 X86TargetStreamer &getTargetStreamer() { 117 assert(getParser().getStreamer().getTargetStreamer() && 118 "do not have a target streamer"); 119 MCTargetStreamer &TS = *getParser().getStreamer().getTargetStreamer(); 120 return static_cast<X86TargetStreamer &>(TS); 121 } 122 123 unsigned MatchInstruction(const OperandVector &Operands, MCInst &Inst, 124 uint64_t &ErrorInfo, FeatureBitset &MissingFeatures, 125 bool matchingInlineAsm, unsigned VariantID = 0) { 126 // In Code16GCC mode, match as 32-bit. 127 if (Code16GCC) 128 SwitchMode(X86::Is32Bit); 129 unsigned rv = MatchInstructionImpl(Operands, Inst, ErrorInfo, 130 MissingFeatures, matchingInlineAsm, 131 VariantID); 132 if (Code16GCC) 133 SwitchMode(X86::Is16Bit); 134 return rv; 135 } 136 137 enum InfixCalculatorTok { 138 IC_OR = 0, 139 IC_XOR, 140 IC_AND, 141 IC_LSHIFT, 142 IC_RSHIFT, 143 IC_PLUS, 144 IC_MINUS, 145 IC_MULTIPLY, 146 IC_DIVIDE, 147 IC_MOD, 148 IC_NOT, 149 IC_NEG, 150 IC_RPAREN, 151 IC_LPAREN, 152 IC_IMM, 153 IC_REGISTER, 154 IC_EQ, 155 IC_NE, 156 IC_LT, 157 IC_LE, 158 IC_GT, 159 IC_GE 160 }; 161 162 enum IntelOperatorKind { 163 IOK_INVALID = 0, 164 IOK_LENGTH, 165 IOK_SIZE, 166 IOK_TYPE, 167 }; 168 169 enum MasmOperatorKind { 170 MOK_INVALID = 0, 171 MOK_LENGTHOF, 172 MOK_SIZEOF, 173 MOK_TYPE, 174 }; 175 176 class InfixCalculator { 177 typedef std::pair< InfixCalculatorTok, int64_t > ICToken; 178 SmallVector<InfixCalculatorTok, 4> InfixOperatorStack; 179 SmallVector<ICToken, 4> PostfixStack; 180 181 bool isUnaryOperator(InfixCalculatorTok Op) const { 182 return Op == IC_NEG || Op == IC_NOT; 183 } 184 185 public: 186 int64_t popOperand() { 187 assert (!PostfixStack.empty() && "Poped an empty stack!"); 188 ICToken Op = PostfixStack.pop_back_val(); 189 if (!(Op.first == IC_IMM || Op.first == IC_REGISTER)) 190 return -1; // The invalid Scale value will be caught later by checkScale 191 return Op.second; 192 } 193 void pushOperand(InfixCalculatorTok Op, int64_t Val = 0) { 194 assert ((Op == IC_IMM || Op == IC_REGISTER) && 195 "Unexpected operand!"); 196 PostfixStack.push_back(std::make_pair(Op, Val)); 197 } 198 199 void popOperator() { InfixOperatorStack.pop_back(); } 200 void pushOperator(InfixCalculatorTok Op) { 201 // Push the new operator if the stack is empty. 202 if (InfixOperatorStack.empty()) { 203 InfixOperatorStack.push_back(Op); 204 return; 205 } 206 207 // Push the new operator if it has a higher precedence than the operator 208 // on the top of the stack or the operator on the top of the stack is a 209 // left parentheses. 210 unsigned Idx = InfixOperatorStack.size() - 1; 211 InfixCalculatorTok StackOp = InfixOperatorStack[Idx]; 212 if (OpPrecedence[Op] > OpPrecedence[StackOp] || StackOp == IC_LPAREN) { 213 InfixOperatorStack.push_back(Op); 214 return; 215 } 216 217 // The operator on the top of the stack has higher precedence than the 218 // new operator. 219 unsigned ParenCount = 0; 220 while (true) { 221 // Nothing to process. 222 if (InfixOperatorStack.empty()) 223 break; 224 225 Idx = InfixOperatorStack.size() - 1; 226 StackOp = InfixOperatorStack[Idx]; 227 if (!(OpPrecedence[StackOp] >= OpPrecedence[Op] || ParenCount)) 228 break; 229 230 // If we have an even parentheses count and we see a left parentheses, 231 // then stop processing. 232 if (!ParenCount && StackOp == IC_LPAREN) 233 break; 234 235 if (StackOp == IC_RPAREN) { 236 ++ParenCount; 237 InfixOperatorStack.pop_back(); 238 } else if (StackOp == IC_LPAREN) { 239 --ParenCount; 240 InfixOperatorStack.pop_back(); 241 } else { 242 InfixOperatorStack.pop_back(); 243 PostfixStack.push_back(std::make_pair(StackOp, 0)); 244 } 245 } 246 // Push the new operator. 247 InfixOperatorStack.push_back(Op); 248 } 249 250 int64_t execute() { 251 // Push any remaining operators onto the postfix stack. 252 while (!InfixOperatorStack.empty()) { 253 InfixCalculatorTok StackOp = InfixOperatorStack.pop_back_val(); 254 if (StackOp != IC_LPAREN && StackOp != IC_RPAREN) 255 PostfixStack.push_back(std::make_pair(StackOp, 0)); 256 } 257 258 if (PostfixStack.empty()) 259 return 0; 260 261 SmallVector<ICToken, 16> OperandStack; 262 for (unsigned i = 0, e = PostfixStack.size(); i != e; ++i) { 263 ICToken Op = PostfixStack[i]; 264 if (Op.first == IC_IMM || Op.first == IC_REGISTER) { 265 OperandStack.push_back(Op); 266 } else if (isUnaryOperator(Op.first)) { 267 assert (OperandStack.size() > 0 && "Too few operands."); 268 ICToken Operand = OperandStack.pop_back_val(); 269 assert (Operand.first == IC_IMM && 270 "Unary operation with a register!"); 271 switch (Op.first) { 272 default: 273 report_fatal_error("Unexpected operator!"); 274 break; 275 case IC_NEG: 276 OperandStack.push_back(std::make_pair(IC_IMM, -Operand.second)); 277 break; 278 case IC_NOT: 279 OperandStack.push_back(std::make_pair(IC_IMM, ~Operand.second)); 280 break; 281 } 282 } else { 283 assert (OperandStack.size() > 1 && "Too few operands."); 284 int64_t Val; 285 ICToken Op2 = OperandStack.pop_back_val(); 286 ICToken Op1 = OperandStack.pop_back_val(); 287 switch (Op.first) { 288 default: 289 report_fatal_error("Unexpected operator!"); 290 break; 291 case IC_PLUS: 292 Val = Op1.second + Op2.second; 293 OperandStack.push_back(std::make_pair(IC_IMM, Val)); 294 break; 295 case IC_MINUS: 296 Val = Op1.second - Op2.second; 297 OperandStack.push_back(std::make_pair(IC_IMM, Val)); 298 break; 299 case IC_MULTIPLY: 300 assert (Op1.first == IC_IMM && Op2.first == IC_IMM && 301 "Multiply operation with an immediate and a register!"); 302 Val = Op1.second * Op2.second; 303 OperandStack.push_back(std::make_pair(IC_IMM, Val)); 304 break; 305 case IC_DIVIDE: 306 assert (Op1.first == IC_IMM && Op2.first == IC_IMM && 307 "Divide operation with an immediate and a register!"); 308 assert (Op2.second != 0 && "Division by zero!"); 309 Val = Op1.second / Op2.second; 310 OperandStack.push_back(std::make_pair(IC_IMM, Val)); 311 break; 312 case IC_MOD: 313 assert (Op1.first == IC_IMM && Op2.first == IC_IMM && 314 "Modulo operation with an immediate and a register!"); 315 Val = Op1.second % Op2.second; 316 OperandStack.push_back(std::make_pair(IC_IMM, Val)); 317 break; 318 case IC_OR: 319 assert (Op1.first == IC_IMM && Op2.first == IC_IMM && 320 "Or operation with an immediate and a register!"); 321 Val = Op1.second | Op2.second; 322 OperandStack.push_back(std::make_pair(IC_IMM, Val)); 323 break; 324 case IC_XOR: 325 assert(Op1.first == IC_IMM && Op2.first == IC_IMM && 326 "Xor operation with an immediate and a register!"); 327 Val = Op1.second ^ Op2.second; 328 OperandStack.push_back(std::make_pair(IC_IMM, Val)); 329 break; 330 case IC_AND: 331 assert (Op1.first == IC_IMM && Op2.first == IC_IMM && 332 "And operation with an immediate and a register!"); 333 Val = Op1.second & Op2.second; 334 OperandStack.push_back(std::make_pair(IC_IMM, Val)); 335 break; 336 case IC_LSHIFT: 337 assert (Op1.first == IC_IMM && Op2.first == IC_IMM && 338 "Left shift operation with an immediate and a register!"); 339 Val = Op1.second << Op2.second; 340 OperandStack.push_back(std::make_pair(IC_IMM, Val)); 341 break; 342 case IC_RSHIFT: 343 assert (Op1.first == IC_IMM && Op2.first == IC_IMM && 344 "Right shift operation with an immediate and a register!"); 345 Val = Op1.second >> Op2.second; 346 OperandStack.push_back(std::make_pair(IC_IMM, Val)); 347 break; 348 case IC_EQ: 349 assert(Op1.first == IC_IMM && Op2.first == IC_IMM && 350 "Equals operation with an immediate and a register!"); 351 Val = (Op1.second == Op2.second) ? -1 : 0; 352 OperandStack.push_back(std::make_pair(IC_IMM, Val)); 353 break; 354 case IC_NE: 355 assert(Op1.first == IC_IMM && Op2.first == IC_IMM && 356 "Not-equals operation with an immediate and a register!"); 357 Val = (Op1.second != Op2.second) ? -1 : 0; 358 OperandStack.push_back(std::make_pair(IC_IMM, Val)); 359 break; 360 case IC_LT: 361 assert(Op1.first == IC_IMM && Op2.first == IC_IMM && 362 "Less-than operation with an immediate and a register!"); 363 Val = (Op1.second < Op2.second) ? -1 : 0; 364 OperandStack.push_back(std::make_pair(IC_IMM, Val)); 365 break; 366 case IC_LE: 367 assert(Op1.first == IC_IMM && Op2.first == IC_IMM && 368 "Less-than-or-equal operation with an immediate and a " 369 "register!"); 370 Val = (Op1.second <= Op2.second) ? -1 : 0; 371 OperandStack.push_back(std::make_pair(IC_IMM, Val)); 372 break; 373 case IC_GT: 374 assert(Op1.first == IC_IMM && Op2.first == IC_IMM && 375 "Greater-than operation with an immediate and a register!"); 376 Val = (Op1.second > Op2.second) ? -1 : 0; 377 OperandStack.push_back(std::make_pair(IC_IMM, Val)); 378 break; 379 case IC_GE: 380 assert(Op1.first == IC_IMM && Op2.first == IC_IMM && 381 "Greater-than-or-equal operation with an immediate and a " 382 "register!"); 383 Val = (Op1.second >= Op2.second) ? -1 : 0; 384 OperandStack.push_back(std::make_pair(IC_IMM, Val)); 385 break; 386 } 387 } 388 } 389 assert (OperandStack.size() == 1 && "Expected a single result."); 390 return OperandStack.pop_back_val().second; 391 } 392 }; 393 394 enum IntelExprState { 395 IES_INIT, 396 IES_OR, 397 IES_XOR, 398 IES_AND, 399 IES_EQ, 400 IES_NE, 401 IES_LT, 402 IES_LE, 403 IES_GT, 404 IES_GE, 405 IES_LSHIFT, 406 IES_RSHIFT, 407 IES_PLUS, 408 IES_MINUS, 409 IES_OFFSET, 410 IES_CAST, 411 IES_NOT, 412 IES_MULTIPLY, 413 IES_DIVIDE, 414 IES_MOD, 415 IES_LBRAC, 416 IES_RBRAC, 417 IES_LPAREN, 418 IES_RPAREN, 419 IES_REGISTER, 420 IES_INTEGER, 421 IES_IDENTIFIER, 422 IES_ERROR 423 }; 424 425 class IntelExprStateMachine { 426 IntelExprState State = IES_INIT, PrevState = IES_ERROR; 427 unsigned BaseReg = 0, IndexReg = 0, TmpReg = 0, Scale = 0; 428 int64_t Imm = 0; 429 const MCExpr *Sym = nullptr; 430 StringRef SymName; 431 InfixCalculator IC; 432 InlineAsmIdentifierInfo Info; 433 short BracCount = 0; 434 bool MemExpr = false; 435 bool OffsetOperator = false; 436 bool AttachToOperandIdx = false; 437 bool IsPIC = false; 438 SMLoc OffsetOperatorLoc; 439 AsmTypeInfo CurType; 440 441 bool setSymRef(const MCExpr *Val, StringRef ID, StringRef &ErrMsg) { 442 if (Sym) { 443 ErrMsg = "cannot use more than one symbol in memory operand"; 444 return true; 445 } 446 Sym = Val; 447 SymName = ID; 448 return false; 449 } 450 451 public: 452 IntelExprStateMachine() = default; 453 454 void addImm(int64_t imm) { Imm += imm; } 455 short getBracCount() const { return BracCount; } 456 bool isMemExpr() const { return MemExpr; } 457 bool isOffsetOperator() const { return OffsetOperator; } 458 SMLoc getOffsetLoc() const { return OffsetOperatorLoc; } 459 unsigned getBaseReg() const { return BaseReg; } 460 unsigned getIndexReg() const { return IndexReg; } 461 unsigned getScale() const { return Scale; } 462 const MCExpr *getSym() const { return Sym; } 463 StringRef getSymName() const { return SymName; } 464 StringRef getType() const { return CurType.Name; } 465 unsigned getSize() const { return CurType.Size; } 466 unsigned getElementSize() const { return CurType.ElementSize; } 467 unsigned getLength() const { return CurType.Length; } 468 int64_t getImm() { return Imm + IC.execute(); } 469 bool isValidEndState() const { 470 return State == IES_RBRAC || State == IES_INTEGER; 471 } 472 473 // Is the intel expression appended after an operand index. 474 // [OperandIdx][Intel Expression] 475 // This is neccessary for checking if it is an independent 476 // intel expression at back end when parse inline asm. 477 void setAppendAfterOperand() { AttachToOperandIdx = true; } 478 479 bool isPIC() const { return IsPIC; } 480 void setPIC() { IsPIC = true; } 481 482 bool hadError() const { return State == IES_ERROR; } 483 const InlineAsmIdentifierInfo &getIdentifierInfo() const { return Info; } 484 485 bool regsUseUpError(StringRef &ErrMsg) { 486 // This case mostly happen in inline asm, e.g. Arr[BaseReg + IndexReg] 487 // can not intruduce additional register in inline asm in PIC model. 488 if (IsPIC && AttachToOperandIdx) 489 ErrMsg = "Don't use 2 or more regs for mem offset in PIC model!"; 490 else 491 ErrMsg = "BaseReg/IndexReg already set!"; 492 return true; 493 } 494 495 void onOr() { 496 IntelExprState CurrState = State; 497 switch (State) { 498 default: 499 State = IES_ERROR; 500 break; 501 case IES_INTEGER: 502 case IES_RPAREN: 503 case IES_REGISTER: 504 State = IES_OR; 505 IC.pushOperator(IC_OR); 506 break; 507 } 508 PrevState = CurrState; 509 } 510 void onXor() { 511 IntelExprState CurrState = State; 512 switch (State) { 513 default: 514 State = IES_ERROR; 515 break; 516 case IES_INTEGER: 517 case IES_RPAREN: 518 case IES_REGISTER: 519 State = IES_XOR; 520 IC.pushOperator(IC_XOR); 521 break; 522 } 523 PrevState = CurrState; 524 } 525 void onAnd() { 526 IntelExprState CurrState = State; 527 switch (State) { 528 default: 529 State = IES_ERROR; 530 break; 531 case IES_INTEGER: 532 case IES_RPAREN: 533 case IES_REGISTER: 534 State = IES_AND; 535 IC.pushOperator(IC_AND); 536 break; 537 } 538 PrevState = CurrState; 539 } 540 void onEq() { 541 IntelExprState CurrState = State; 542 switch (State) { 543 default: 544 State = IES_ERROR; 545 break; 546 case IES_INTEGER: 547 case IES_RPAREN: 548 case IES_REGISTER: 549 State = IES_EQ; 550 IC.pushOperator(IC_EQ); 551 break; 552 } 553 PrevState = CurrState; 554 } 555 void onNE() { 556 IntelExprState CurrState = State; 557 switch (State) { 558 default: 559 State = IES_ERROR; 560 break; 561 case IES_INTEGER: 562 case IES_RPAREN: 563 case IES_REGISTER: 564 State = IES_NE; 565 IC.pushOperator(IC_NE); 566 break; 567 } 568 PrevState = CurrState; 569 } 570 void onLT() { 571 IntelExprState CurrState = State; 572 switch (State) { 573 default: 574 State = IES_ERROR; 575 break; 576 case IES_INTEGER: 577 case IES_RPAREN: 578 case IES_REGISTER: 579 State = IES_LT; 580 IC.pushOperator(IC_LT); 581 break; 582 } 583 PrevState = CurrState; 584 } 585 void onLE() { 586 IntelExprState CurrState = State; 587 switch (State) { 588 default: 589 State = IES_ERROR; 590 break; 591 case IES_INTEGER: 592 case IES_RPAREN: 593 case IES_REGISTER: 594 State = IES_LE; 595 IC.pushOperator(IC_LE); 596 break; 597 } 598 PrevState = CurrState; 599 } 600 void onGT() { 601 IntelExprState CurrState = State; 602 switch (State) { 603 default: 604 State = IES_ERROR; 605 break; 606 case IES_INTEGER: 607 case IES_RPAREN: 608 case IES_REGISTER: 609 State = IES_GT; 610 IC.pushOperator(IC_GT); 611 break; 612 } 613 PrevState = CurrState; 614 } 615 void onGE() { 616 IntelExprState CurrState = State; 617 switch (State) { 618 default: 619 State = IES_ERROR; 620 break; 621 case IES_INTEGER: 622 case IES_RPAREN: 623 case IES_REGISTER: 624 State = IES_GE; 625 IC.pushOperator(IC_GE); 626 break; 627 } 628 PrevState = CurrState; 629 } 630 void onLShift() { 631 IntelExprState CurrState = State; 632 switch (State) { 633 default: 634 State = IES_ERROR; 635 break; 636 case IES_INTEGER: 637 case IES_RPAREN: 638 case IES_REGISTER: 639 State = IES_LSHIFT; 640 IC.pushOperator(IC_LSHIFT); 641 break; 642 } 643 PrevState = CurrState; 644 } 645 void onRShift() { 646 IntelExprState CurrState = State; 647 switch (State) { 648 default: 649 State = IES_ERROR; 650 break; 651 case IES_INTEGER: 652 case IES_RPAREN: 653 case IES_REGISTER: 654 State = IES_RSHIFT; 655 IC.pushOperator(IC_RSHIFT); 656 break; 657 } 658 PrevState = CurrState; 659 } 660 bool onPlus(StringRef &ErrMsg) { 661 IntelExprState CurrState = State; 662 switch (State) { 663 default: 664 State = IES_ERROR; 665 break; 666 case IES_INTEGER: 667 case IES_RPAREN: 668 case IES_REGISTER: 669 case IES_OFFSET: 670 State = IES_PLUS; 671 IC.pushOperator(IC_PLUS); 672 if (CurrState == IES_REGISTER && PrevState != IES_MULTIPLY) { 673 // If we already have a BaseReg, then assume this is the IndexReg with 674 // no explicit scale. 675 if (!BaseReg) { 676 BaseReg = TmpReg; 677 } else { 678 if (IndexReg) 679 return regsUseUpError(ErrMsg); 680 IndexReg = TmpReg; 681 Scale = 0; 682 } 683 } 684 break; 685 } 686 PrevState = CurrState; 687 return false; 688 } 689 bool onMinus(StringRef &ErrMsg) { 690 IntelExprState CurrState = State; 691 switch (State) { 692 default: 693 State = IES_ERROR; 694 break; 695 case IES_OR: 696 case IES_XOR: 697 case IES_AND: 698 case IES_EQ: 699 case IES_NE: 700 case IES_LT: 701 case IES_LE: 702 case IES_GT: 703 case IES_GE: 704 case IES_LSHIFT: 705 case IES_RSHIFT: 706 case IES_PLUS: 707 case IES_NOT: 708 case IES_MULTIPLY: 709 case IES_DIVIDE: 710 case IES_MOD: 711 case IES_LPAREN: 712 case IES_RPAREN: 713 case IES_LBRAC: 714 case IES_RBRAC: 715 case IES_INTEGER: 716 case IES_REGISTER: 717 case IES_INIT: 718 case IES_OFFSET: 719 State = IES_MINUS; 720 // push minus operator if it is not a negate operator 721 if (CurrState == IES_REGISTER || CurrState == IES_RPAREN || 722 CurrState == IES_INTEGER || CurrState == IES_RBRAC || 723 CurrState == IES_OFFSET) 724 IC.pushOperator(IC_MINUS); 725 else if (PrevState == IES_REGISTER && CurrState == IES_MULTIPLY) { 726 // We have negate operator for Scale: it's illegal 727 ErrMsg = "Scale can't be negative"; 728 return true; 729 } else 730 IC.pushOperator(IC_NEG); 731 if (CurrState == IES_REGISTER && PrevState != IES_MULTIPLY) { 732 // If we already have a BaseReg, then assume this is the IndexReg with 733 // no explicit scale. 734 if (!BaseReg) { 735 BaseReg = TmpReg; 736 } else { 737 if (IndexReg) 738 return regsUseUpError(ErrMsg); 739 IndexReg = TmpReg; 740 Scale = 0; 741 } 742 } 743 break; 744 } 745 PrevState = CurrState; 746 return false; 747 } 748 void onNot() { 749 IntelExprState CurrState = State; 750 switch (State) { 751 default: 752 State = IES_ERROR; 753 break; 754 case IES_OR: 755 case IES_XOR: 756 case IES_AND: 757 case IES_EQ: 758 case IES_NE: 759 case IES_LT: 760 case IES_LE: 761 case IES_GT: 762 case IES_GE: 763 case IES_LSHIFT: 764 case IES_RSHIFT: 765 case IES_PLUS: 766 case IES_MINUS: 767 case IES_NOT: 768 case IES_MULTIPLY: 769 case IES_DIVIDE: 770 case IES_MOD: 771 case IES_LPAREN: 772 case IES_LBRAC: 773 case IES_INIT: 774 State = IES_NOT; 775 IC.pushOperator(IC_NOT); 776 break; 777 } 778 PrevState = CurrState; 779 } 780 bool onRegister(unsigned Reg, StringRef &ErrMsg) { 781 IntelExprState CurrState = State; 782 switch (State) { 783 default: 784 State = IES_ERROR; 785 break; 786 case IES_PLUS: 787 case IES_LPAREN: 788 case IES_LBRAC: 789 State = IES_REGISTER; 790 TmpReg = Reg; 791 IC.pushOperand(IC_REGISTER); 792 break; 793 case IES_MULTIPLY: 794 // Index Register - Scale * Register 795 if (PrevState == IES_INTEGER) { 796 if (IndexReg) 797 return regsUseUpError(ErrMsg); 798 State = IES_REGISTER; 799 IndexReg = Reg; 800 // Get the scale and replace the 'Scale * Register' with '0'. 801 Scale = IC.popOperand(); 802 if (checkScale(Scale, ErrMsg)) 803 return true; 804 IC.pushOperand(IC_IMM); 805 IC.popOperator(); 806 } else { 807 State = IES_ERROR; 808 } 809 break; 810 } 811 PrevState = CurrState; 812 return false; 813 } 814 bool onIdentifierExpr(const MCExpr *SymRef, StringRef SymRefName, 815 const InlineAsmIdentifierInfo &IDInfo, 816 const AsmTypeInfo &Type, bool ParsingMSInlineAsm, 817 StringRef &ErrMsg) { 818 // InlineAsm: Treat an enum value as an integer 819 if (ParsingMSInlineAsm) 820 if (IDInfo.isKind(InlineAsmIdentifierInfo::IK_EnumVal)) 821 return onInteger(IDInfo.Enum.EnumVal, ErrMsg); 822 // Treat a symbolic constant like an integer 823 if (auto *CE = dyn_cast<MCConstantExpr>(SymRef)) 824 return onInteger(CE->getValue(), ErrMsg); 825 PrevState = State; 826 switch (State) { 827 default: 828 State = IES_ERROR; 829 break; 830 case IES_CAST: 831 case IES_PLUS: 832 case IES_MINUS: 833 case IES_NOT: 834 case IES_INIT: 835 case IES_LBRAC: 836 case IES_LPAREN: 837 if (setSymRef(SymRef, SymRefName, ErrMsg)) 838 return true; 839 MemExpr = true; 840 State = IES_INTEGER; 841 IC.pushOperand(IC_IMM); 842 if (ParsingMSInlineAsm) 843 Info = IDInfo; 844 setTypeInfo(Type); 845 break; 846 } 847 return false; 848 } 849 bool onInteger(int64_t TmpInt, StringRef &ErrMsg) { 850 IntelExprState CurrState = State; 851 switch (State) { 852 default: 853 State = IES_ERROR; 854 break; 855 case IES_PLUS: 856 case IES_MINUS: 857 case IES_NOT: 858 case IES_OR: 859 case IES_XOR: 860 case IES_AND: 861 case IES_EQ: 862 case IES_NE: 863 case IES_LT: 864 case IES_LE: 865 case IES_GT: 866 case IES_GE: 867 case IES_LSHIFT: 868 case IES_RSHIFT: 869 case IES_DIVIDE: 870 case IES_MOD: 871 case IES_MULTIPLY: 872 case IES_LPAREN: 873 case IES_INIT: 874 case IES_LBRAC: 875 State = IES_INTEGER; 876 if (PrevState == IES_REGISTER && CurrState == IES_MULTIPLY) { 877 // Index Register - Register * Scale 878 if (IndexReg) 879 return regsUseUpError(ErrMsg); 880 IndexReg = TmpReg; 881 Scale = TmpInt; 882 if (checkScale(Scale, ErrMsg)) 883 return true; 884 // Get the scale and replace the 'Register * Scale' with '0'. 885 IC.popOperator(); 886 } else { 887 IC.pushOperand(IC_IMM, TmpInt); 888 } 889 break; 890 } 891 PrevState = CurrState; 892 return false; 893 } 894 void onStar() { 895 PrevState = State; 896 switch (State) { 897 default: 898 State = IES_ERROR; 899 break; 900 case IES_INTEGER: 901 case IES_REGISTER: 902 case IES_RPAREN: 903 State = IES_MULTIPLY; 904 IC.pushOperator(IC_MULTIPLY); 905 break; 906 } 907 } 908 void onDivide() { 909 PrevState = State; 910 switch (State) { 911 default: 912 State = IES_ERROR; 913 break; 914 case IES_INTEGER: 915 case IES_RPAREN: 916 State = IES_DIVIDE; 917 IC.pushOperator(IC_DIVIDE); 918 break; 919 } 920 } 921 void onMod() { 922 PrevState = State; 923 switch (State) { 924 default: 925 State = IES_ERROR; 926 break; 927 case IES_INTEGER: 928 case IES_RPAREN: 929 State = IES_MOD; 930 IC.pushOperator(IC_MOD); 931 break; 932 } 933 } 934 bool onLBrac() { 935 if (BracCount) 936 return true; 937 PrevState = State; 938 switch (State) { 939 default: 940 State = IES_ERROR; 941 break; 942 case IES_RBRAC: 943 case IES_INTEGER: 944 case IES_RPAREN: 945 State = IES_PLUS; 946 IC.pushOperator(IC_PLUS); 947 CurType.Length = 1; 948 CurType.Size = CurType.ElementSize; 949 break; 950 case IES_INIT: 951 case IES_CAST: 952 assert(!BracCount && "BracCount should be zero on parsing's start"); 953 State = IES_LBRAC; 954 break; 955 } 956 MemExpr = true; 957 BracCount++; 958 return false; 959 } 960 bool onRBrac(StringRef &ErrMsg) { 961 IntelExprState CurrState = State; 962 switch (State) { 963 default: 964 State = IES_ERROR; 965 break; 966 case IES_INTEGER: 967 case IES_OFFSET: 968 case IES_REGISTER: 969 case IES_RPAREN: 970 if (BracCount-- != 1) { 971 ErrMsg = "unexpected bracket encountered"; 972 return true; 973 } 974 State = IES_RBRAC; 975 if (CurrState == IES_REGISTER && PrevState != IES_MULTIPLY) { 976 // If we already have a BaseReg, then assume this is the IndexReg with 977 // no explicit scale. 978 if (!BaseReg) { 979 BaseReg = TmpReg; 980 } else { 981 if (IndexReg) 982 return regsUseUpError(ErrMsg); 983 IndexReg = TmpReg; 984 Scale = 0; 985 } 986 } 987 break; 988 } 989 PrevState = CurrState; 990 return false; 991 } 992 void onLParen() { 993 IntelExprState CurrState = State; 994 switch (State) { 995 default: 996 State = IES_ERROR; 997 break; 998 case IES_PLUS: 999 case IES_MINUS: 1000 case IES_NOT: 1001 case IES_OR: 1002 case IES_XOR: 1003 case IES_AND: 1004 case IES_EQ: 1005 case IES_NE: 1006 case IES_LT: 1007 case IES_LE: 1008 case IES_GT: 1009 case IES_GE: 1010 case IES_LSHIFT: 1011 case IES_RSHIFT: 1012 case IES_MULTIPLY: 1013 case IES_DIVIDE: 1014 case IES_MOD: 1015 case IES_LPAREN: 1016 case IES_INIT: 1017 case IES_LBRAC: 1018 State = IES_LPAREN; 1019 IC.pushOperator(IC_LPAREN); 1020 break; 1021 } 1022 PrevState = CurrState; 1023 } 1024 void onRParen() { 1025 PrevState = State; 1026 switch (State) { 1027 default: 1028 State = IES_ERROR; 1029 break; 1030 case IES_INTEGER: 1031 case IES_OFFSET: 1032 case IES_REGISTER: 1033 case IES_RBRAC: 1034 case IES_RPAREN: 1035 State = IES_RPAREN; 1036 IC.pushOperator(IC_RPAREN); 1037 break; 1038 } 1039 } 1040 bool onOffset(const MCExpr *Val, SMLoc OffsetLoc, StringRef ID, 1041 const InlineAsmIdentifierInfo &IDInfo, 1042 bool ParsingMSInlineAsm, StringRef &ErrMsg) { 1043 PrevState = State; 1044 switch (State) { 1045 default: 1046 ErrMsg = "unexpected offset operator expression"; 1047 return true; 1048 case IES_PLUS: 1049 case IES_INIT: 1050 case IES_LBRAC: 1051 if (setSymRef(Val, ID, ErrMsg)) 1052 return true; 1053 OffsetOperator = true; 1054 OffsetOperatorLoc = OffsetLoc; 1055 State = IES_OFFSET; 1056 // As we cannot yet resolve the actual value (offset), we retain 1057 // the requested semantics by pushing a '0' to the operands stack 1058 IC.pushOperand(IC_IMM); 1059 if (ParsingMSInlineAsm) { 1060 Info = IDInfo; 1061 } 1062 break; 1063 } 1064 return false; 1065 } 1066 void onCast(AsmTypeInfo Info) { 1067 PrevState = State; 1068 switch (State) { 1069 default: 1070 State = IES_ERROR; 1071 break; 1072 case IES_LPAREN: 1073 setTypeInfo(Info); 1074 State = IES_CAST; 1075 break; 1076 } 1077 } 1078 void setTypeInfo(AsmTypeInfo Type) { CurType = Type; } 1079 }; 1080 1081 bool Error(SMLoc L, const Twine &Msg, SMRange Range = std::nullopt, 1082 bool MatchingInlineAsm = false) { 1083 MCAsmParser &Parser = getParser(); 1084 if (MatchingInlineAsm) { 1085 if (!getLexer().isAtStartOfStatement()) 1086 Parser.eatToEndOfStatement(); 1087 return false; 1088 } 1089 return Parser.Error(L, Msg, Range); 1090 } 1091 1092 bool MatchRegisterByName(MCRegister &RegNo, StringRef RegName, SMLoc StartLoc, 1093 SMLoc EndLoc); 1094 bool ParseRegister(MCRegister &RegNo, SMLoc &StartLoc, SMLoc &EndLoc, 1095 bool RestoreOnFailure); 1096 1097 std::unique_ptr<X86Operand> DefaultMemSIOperand(SMLoc Loc); 1098 std::unique_ptr<X86Operand> DefaultMemDIOperand(SMLoc Loc); 1099 bool IsSIReg(unsigned Reg); 1100 unsigned GetSIDIForRegClass(unsigned RegClassID, unsigned Reg, bool IsSIReg); 1101 void 1102 AddDefaultSrcDestOperands(OperandVector &Operands, 1103 std::unique_ptr<llvm::MCParsedAsmOperand> &&Src, 1104 std::unique_ptr<llvm::MCParsedAsmOperand> &&Dst); 1105 bool VerifyAndAdjustOperands(OperandVector &OrigOperands, 1106 OperandVector &FinalOperands); 1107 bool parseOperand(OperandVector &Operands, StringRef Name); 1108 bool parseATTOperand(OperandVector &Operands); 1109 bool parseIntelOperand(OperandVector &Operands, StringRef Name); 1110 bool ParseIntelOffsetOperator(const MCExpr *&Val, StringRef &ID, 1111 InlineAsmIdentifierInfo &Info, SMLoc &End); 1112 bool ParseIntelDotOperator(IntelExprStateMachine &SM, SMLoc &End); 1113 unsigned IdentifyIntelInlineAsmOperator(StringRef Name); 1114 unsigned ParseIntelInlineAsmOperator(unsigned OpKind); 1115 unsigned IdentifyMasmOperator(StringRef Name); 1116 bool ParseMasmOperator(unsigned OpKind, int64_t &Val); 1117 bool ParseRoundingModeOp(SMLoc Start, OperandVector &Operands); 1118 bool ParseIntelNamedOperator(StringRef Name, IntelExprStateMachine &SM, 1119 bool &ParseError, SMLoc &End); 1120 bool ParseMasmNamedOperator(StringRef Name, IntelExprStateMachine &SM, 1121 bool &ParseError, SMLoc &End); 1122 void RewriteIntelExpression(IntelExprStateMachine &SM, SMLoc Start, 1123 SMLoc End); 1124 bool ParseIntelExpression(IntelExprStateMachine &SM, SMLoc &End); 1125 bool ParseIntelInlineAsmIdentifier(const MCExpr *&Val, StringRef &Identifier, 1126 InlineAsmIdentifierInfo &Info, 1127 bool IsUnevaluatedOperand, SMLoc &End, 1128 bool IsParsingOffsetOperator = false); 1129 void tryParseOperandIdx(AsmToken::TokenKind PrevTK, 1130 IntelExprStateMachine &SM); 1131 1132 bool ParseMemOperand(unsigned SegReg, const MCExpr *Disp, SMLoc StartLoc, 1133 SMLoc EndLoc, OperandVector &Operands); 1134 1135 X86::CondCode ParseConditionCode(StringRef CCode); 1136 1137 bool ParseIntelMemoryOperandSize(unsigned &Size); 1138 bool CreateMemForMSInlineAsm(unsigned SegReg, const MCExpr *Disp, 1139 unsigned BaseReg, unsigned IndexReg, 1140 unsigned Scale, SMLoc Start, SMLoc End, 1141 unsigned Size, StringRef Identifier, 1142 const InlineAsmIdentifierInfo &Info, 1143 OperandVector &Operands); 1144 1145 bool parseDirectiveArch(); 1146 bool parseDirectiveNops(SMLoc L); 1147 bool parseDirectiveEven(SMLoc L); 1148 bool ParseDirectiveCode(StringRef IDVal, SMLoc L); 1149 1150 /// CodeView FPO data directives. 1151 bool parseDirectiveFPOProc(SMLoc L); 1152 bool parseDirectiveFPOSetFrame(SMLoc L); 1153 bool parseDirectiveFPOPushReg(SMLoc L); 1154 bool parseDirectiveFPOStackAlloc(SMLoc L); 1155 bool parseDirectiveFPOStackAlign(SMLoc L); 1156 bool parseDirectiveFPOEndPrologue(SMLoc L); 1157 bool parseDirectiveFPOEndProc(SMLoc L); 1158 1159 /// SEH directives. 1160 bool parseSEHRegisterNumber(unsigned RegClassID, MCRegister &RegNo); 1161 bool parseDirectiveSEHPushReg(SMLoc); 1162 bool parseDirectiveSEHSetFrame(SMLoc); 1163 bool parseDirectiveSEHSaveReg(SMLoc); 1164 bool parseDirectiveSEHSaveXMM(SMLoc); 1165 bool parseDirectiveSEHPushFrame(SMLoc); 1166 1167 unsigned checkTargetMatchPredicate(MCInst &Inst) override; 1168 1169 bool validateInstruction(MCInst &Inst, const OperandVector &Ops); 1170 bool processInstruction(MCInst &Inst, const OperandVector &Ops); 1171 1172 // Load Value Injection (LVI) Mitigations for machine code 1173 void emitWarningForSpecialLVIInstruction(SMLoc Loc); 1174 void applyLVICFIMitigation(MCInst &Inst, MCStreamer &Out); 1175 void applyLVILoadHardeningMitigation(MCInst &Inst, MCStreamer &Out); 1176 1177 /// Wrapper around MCStreamer::emitInstruction(). Possibly adds 1178 /// instrumentation around Inst. 1179 void emitInstruction(MCInst &Inst, OperandVector &Operands, MCStreamer &Out); 1180 1181 bool MatchAndEmitInstruction(SMLoc IDLoc, unsigned &Opcode, 1182 OperandVector &Operands, MCStreamer &Out, 1183 uint64_t &ErrorInfo, 1184 bool MatchingInlineAsm) override; 1185 1186 void MatchFPUWaitAlias(SMLoc IDLoc, X86Operand &Op, OperandVector &Operands, 1187 MCStreamer &Out, bool MatchingInlineAsm); 1188 1189 bool ErrorMissingFeature(SMLoc IDLoc, const FeatureBitset &MissingFeatures, 1190 bool MatchingInlineAsm); 1191 1192 bool MatchAndEmitATTInstruction(SMLoc IDLoc, unsigned &Opcode, 1193 OperandVector &Operands, MCStreamer &Out, 1194 uint64_t &ErrorInfo, 1195 bool MatchingInlineAsm); 1196 1197 bool MatchAndEmitIntelInstruction(SMLoc IDLoc, unsigned &Opcode, 1198 OperandVector &Operands, MCStreamer &Out, 1199 uint64_t &ErrorInfo, 1200 bool MatchingInlineAsm); 1201 1202 bool OmitRegisterFromClobberLists(unsigned RegNo) override; 1203 1204 /// Parses AVX512 specific operand primitives: masked registers ({%k<NUM>}, {z}) 1205 /// and memory broadcasting ({1to<NUM>}) primitives, updating Operands vector if required. 1206 /// return false if no parsing errors occurred, true otherwise. 1207 bool HandleAVX512Operand(OperandVector &Operands); 1208 1209 bool ParseZ(std::unique_ptr<X86Operand> &Z, const SMLoc &StartLoc); 1210 1211 bool is64BitMode() const { 1212 // FIXME: Can tablegen auto-generate this? 1213 return getSTI().getFeatureBits()[X86::Is64Bit]; 1214 } 1215 bool is32BitMode() const { 1216 // FIXME: Can tablegen auto-generate this? 1217 return getSTI().getFeatureBits()[X86::Is32Bit]; 1218 } 1219 bool is16BitMode() const { 1220 // FIXME: Can tablegen auto-generate this? 1221 return getSTI().getFeatureBits()[X86::Is16Bit]; 1222 } 1223 void SwitchMode(unsigned mode) { 1224 MCSubtargetInfo &STI = copySTI(); 1225 FeatureBitset AllModes({X86::Is64Bit, X86::Is32Bit, X86::Is16Bit}); 1226 FeatureBitset OldMode = STI.getFeatureBits() & AllModes; 1227 FeatureBitset FB = ComputeAvailableFeatures( 1228 STI.ToggleFeature(OldMode.flip(mode))); 1229 setAvailableFeatures(FB); 1230 1231 assert(FeatureBitset({mode}) == (STI.getFeatureBits() & AllModes)); 1232 } 1233 1234 unsigned getPointerWidth() { 1235 if (is16BitMode()) return 16; 1236 if (is32BitMode()) return 32; 1237 if (is64BitMode()) return 64; 1238 llvm_unreachable("invalid mode"); 1239 } 1240 1241 bool isParsingIntelSyntax() { 1242 return getParser().getAssemblerDialect(); 1243 } 1244 1245 /// @name Auto-generated Matcher Functions 1246 /// { 1247 1248 #define GET_ASSEMBLER_HEADER 1249 #include "X86GenAsmMatcher.inc" 1250 1251 /// } 1252 1253 public: 1254 enum X86MatchResultTy { 1255 Match_Unsupported = FIRST_TARGET_MATCH_RESULT_TY, 1256 #define GET_OPERAND_DIAGNOSTIC_TYPES 1257 #include "X86GenAsmMatcher.inc" 1258 }; 1259 1260 X86AsmParser(const MCSubtargetInfo &sti, MCAsmParser &Parser, 1261 const MCInstrInfo &mii, const MCTargetOptions &Options) 1262 : MCTargetAsmParser(Options, sti, mii), InstInfo(nullptr), 1263 Code16GCC(false) { 1264 1265 Parser.addAliasForDirective(".word", ".2byte"); 1266 1267 // Initialize the set of available features. 1268 setAvailableFeatures(ComputeAvailableFeatures(getSTI().getFeatureBits())); 1269 } 1270 1271 bool parseRegister(MCRegister &RegNo, SMLoc &StartLoc, 1272 SMLoc &EndLoc) override; 1273 OperandMatchResultTy tryParseRegister(MCRegister &RegNo, SMLoc &StartLoc, 1274 SMLoc &EndLoc) override; 1275 1276 bool parsePrimaryExpr(const MCExpr *&Res, SMLoc &EndLoc) override; 1277 1278 bool ParseInstruction(ParseInstructionInfo &Info, StringRef Name, 1279 SMLoc NameLoc, OperandVector &Operands) override; 1280 1281 bool ParseDirective(AsmToken DirectiveID) override; 1282 }; 1283 } // end anonymous namespace 1284 1285 /// @name Auto-generated Match Functions 1286 /// { 1287 1288 static unsigned MatchRegisterName(StringRef Name); 1289 1290 /// } 1291 1292 static bool CheckBaseRegAndIndexRegAndScale(unsigned BaseReg, unsigned IndexReg, 1293 unsigned Scale, bool Is64BitMode, 1294 StringRef &ErrMsg) { 1295 // If we have both a base register and an index register make sure they are 1296 // both 64-bit or 32-bit registers. 1297 // To support VSIB, IndexReg can be 128-bit or 256-bit registers. 1298 1299 if (BaseReg != 0 && 1300 !(BaseReg == X86::RIP || BaseReg == X86::EIP || 1301 X86MCRegisterClasses[X86::GR16RegClassID].contains(BaseReg) || 1302 X86MCRegisterClasses[X86::GR32RegClassID].contains(BaseReg) || 1303 X86MCRegisterClasses[X86::GR64RegClassID].contains(BaseReg))) { 1304 ErrMsg = "invalid base+index expression"; 1305 return true; 1306 } 1307 1308 if (IndexReg != 0 && 1309 !(IndexReg == X86::EIZ || IndexReg == X86::RIZ || 1310 X86MCRegisterClasses[X86::GR16RegClassID].contains(IndexReg) || 1311 X86MCRegisterClasses[X86::GR32RegClassID].contains(IndexReg) || 1312 X86MCRegisterClasses[X86::GR64RegClassID].contains(IndexReg) || 1313 X86MCRegisterClasses[X86::VR128XRegClassID].contains(IndexReg) || 1314 X86MCRegisterClasses[X86::VR256XRegClassID].contains(IndexReg) || 1315 X86MCRegisterClasses[X86::VR512RegClassID].contains(IndexReg))) { 1316 ErrMsg = "invalid base+index expression"; 1317 return true; 1318 } 1319 1320 if (((BaseReg == X86::RIP || BaseReg == X86::EIP) && IndexReg != 0) || 1321 IndexReg == X86::EIP || IndexReg == X86::RIP || 1322 IndexReg == X86::ESP || IndexReg == X86::RSP) { 1323 ErrMsg = "invalid base+index expression"; 1324 return true; 1325 } 1326 1327 // Check for use of invalid 16-bit registers. Only BX/BP/SI/DI are allowed, 1328 // and then only in non-64-bit modes. 1329 if (X86MCRegisterClasses[X86::GR16RegClassID].contains(BaseReg) && 1330 (Is64BitMode || (BaseReg != X86::BX && BaseReg != X86::BP && 1331 BaseReg != X86::SI && BaseReg != X86::DI))) { 1332 ErrMsg = "invalid 16-bit base register"; 1333 return true; 1334 } 1335 1336 if (BaseReg == 0 && 1337 X86MCRegisterClasses[X86::GR16RegClassID].contains(IndexReg)) { 1338 ErrMsg = "16-bit memory operand may not include only index register"; 1339 return true; 1340 } 1341 1342 if (BaseReg != 0 && IndexReg != 0) { 1343 if (X86MCRegisterClasses[X86::GR64RegClassID].contains(BaseReg) && 1344 (X86MCRegisterClasses[X86::GR16RegClassID].contains(IndexReg) || 1345 X86MCRegisterClasses[X86::GR32RegClassID].contains(IndexReg) || 1346 IndexReg == X86::EIZ)) { 1347 ErrMsg = "base register is 64-bit, but index register is not"; 1348 return true; 1349 } 1350 if (X86MCRegisterClasses[X86::GR32RegClassID].contains(BaseReg) && 1351 (X86MCRegisterClasses[X86::GR16RegClassID].contains(IndexReg) || 1352 X86MCRegisterClasses[X86::GR64RegClassID].contains(IndexReg) || 1353 IndexReg == X86::RIZ)) { 1354 ErrMsg = "base register is 32-bit, but index register is not"; 1355 return true; 1356 } 1357 if (X86MCRegisterClasses[X86::GR16RegClassID].contains(BaseReg)) { 1358 if (X86MCRegisterClasses[X86::GR32RegClassID].contains(IndexReg) || 1359 X86MCRegisterClasses[X86::GR64RegClassID].contains(IndexReg)) { 1360 ErrMsg = "base register is 16-bit, but index register is not"; 1361 return true; 1362 } 1363 if ((BaseReg != X86::BX && BaseReg != X86::BP) || 1364 (IndexReg != X86::SI && IndexReg != X86::DI)) { 1365 ErrMsg = "invalid 16-bit base/index register combination"; 1366 return true; 1367 } 1368 } 1369 } 1370 1371 // RIP/EIP-relative addressing is only supported in 64-bit mode. 1372 if (!Is64BitMode && BaseReg != 0 && 1373 (BaseReg == X86::RIP || BaseReg == X86::EIP)) { 1374 ErrMsg = "IP-relative addressing requires 64-bit mode"; 1375 return true; 1376 } 1377 1378 return checkScale(Scale, ErrMsg); 1379 } 1380 1381 bool X86AsmParser::MatchRegisterByName(MCRegister &RegNo, StringRef RegName, 1382 SMLoc StartLoc, SMLoc EndLoc) { 1383 // If we encounter a %, ignore it. This code handles registers with and 1384 // without the prefix, unprefixed registers can occur in cfi directives. 1385 RegName.consume_front("%"); 1386 1387 RegNo = MatchRegisterName(RegName); 1388 1389 // If the match failed, try the register name as lowercase. 1390 if (RegNo == 0) 1391 RegNo = MatchRegisterName(RegName.lower()); 1392 1393 // The "flags" and "mxcsr" registers cannot be referenced directly. 1394 // Treat it as an identifier instead. 1395 if (isParsingMSInlineAsm() && isParsingIntelSyntax() && 1396 (RegNo == X86::EFLAGS || RegNo == X86::MXCSR)) 1397 RegNo = 0; 1398 1399 if (!is64BitMode()) { 1400 // FIXME: This should be done using Requires<Not64BitMode> and 1401 // Requires<In64BitMode> so "eiz" usage in 64-bit instructions can be also 1402 // checked. 1403 if (RegNo == X86::RIZ || RegNo == X86::RIP || 1404 X86MCRegisterClasses[X86::GR64RegClassID].contains(RegNo) || 1405 X86II::isX86_64NonExtLowByteReg(RegNo) || 1406 X86II::isX86_64ExtendedReg(RegNo)) { 1407 return Error(StartLoc, 1408 "register %" + RegName + " is only available in 64-bit mode", 1409 SMRange(StartLoc, EndLoc)); 1410 } 1411 } 1412 1413 // If this is "db[0-15]", match it as an alias 1414 // for dr[0-15]. 1415 if (RegNo == 0 && RegName.startswith("db")) { 1416 if (RegName.size() == 3) { 1417 switch (RegName[2]) { 1418 case '0': 1419 RegNo = X86::DR0; 1420 break; 1421 case '1': 1422 RegNo = X86::DR1; 1423 break; 1424 case '2': 1425 RegNo = X86::DR2; 1426 break; 1427 case '3': 1428 RegNo = X86::DR3; 1429 break; 1430 case '4': 1431 RegNo = X86::DR4; 1432 break; 1433 case '5': 1434 RegNo = X86::DR5; 1435 break; 1436 case '6': 1437 RegNo = X86::DR6; 1438 break; 1439 case '7': 1440 RegNo = X86::DR7; 1441 break; 1442 case '8': 1443 RegNo = X86::DR8; 1444 break; 1445 case '9': 1446 RegNo = X86::DR9; 1447 break; 1448 } 1449 } else if (RegName.size() == 4 && RegName[2] == '1') { 1450 switch (RegName[3]) { 1451 case '0': 1452 RegNo = X86::DR10; 1453 break; 1454 case '1': 1455 RegNo = X86::DR11; 1456 break; 1457 case '2': 1458 RegNo = X86::DR12; 1459 break; 1460 case '3': 1461 RegNo = X86::DR13; 1462 break; 1463 case '4': 1464 RegNo = X86::DR14; 1465 break; 1466 case '5': 1467 RegNo = X86::DR15; 1468 break; 1469 } 1470 } 1471 } 1472 1473 if (RegNo == 0) { 1474 if (isParsingIntelSyntax()) 1475 return true; 1476 return Error(StartLoc, "invalid register name", SMRange(StartLoc, EndLoc)); 1477 } 1478 return false; 1479 } 1480 1481 bool X86AsmParser::ParseRegister(MCRegister &RegNo, SMLoc &StartLoc, 1482 SMLoc &EndLoc, bool RestoreOnFailure) { 1483 MCAsmParser &Parser = getParser(); 1484 MCAsmLexer &Lexer = getLexer(); 1485 RegNo = 0; 1486 1487 SmallVector<AsmToken, 5> Tokens; 1488 auto OnFailure = [RestoreOnFailure, &Lexer, &Tokens]() { 1489 if (RestoreOnFailure) { 1490 while (!Tokens.empty()) { 1491 Lexer.UnLex(Tokens.pop_back_val()); 1492 } 1493 } 1494 }; 1495 1496 const AsmToken &PercentTok = Parser.getTok(); 1497 StartLoc = PercentTok.getLoc(); 1498 1499 // If we encounter a %, ignore it. This code handles registers with and 1500 // without the prefix, unprefixed registers can occur in cfi directives. 1501 if (!isParsingIntelSyntax() && PercentTok.is(AsmToken::Percent)) { 1502 Tokens.push_back(PercentTok); 1503 Parser.Lex(); // Eat percent token. 1504 } 1505 1506 const AsmToken &Tok = Parser.getTok(); 1507 EndLoc = Tok.getEndLoc(); 1508 1509 if (Tok.isNot(AsmToken::Identifier)) { 1510 OnFailure(); 1511 if (isParsingIntelSyntax()) return true; 1512 return Error(StartLoc, "invalid register name", 1513 SMRange(StartLoc, EndLoc)); 1514 } 1515 1516 if (MatchRegisterByName(RegNo, Tok.getString(), StartLoc, EndLoc)) { 1517 OnFailure(); 1518 return true; 1519 } 1520 1521 // Parse "%st" as "%st(0)" and "%st(1)", which is multiple tokens. 1522 if (RegNo == X86::ST0) { 1523 Tokens.push_back(Tok); 1524 Parser.Lex(); // Eat 'st' 1525 1526 // Check to see if we have '(4)' after %st. 1527 if (Lexer.isNot(AsmToken::LParen)) 1528 return false; 1529 // Lex the paren. 1530 Tokens.push_back(Parser.getTok()); 1531 Parser.Lex(); 1532 1533 const AsmToken &IntTok = Parser.getTok(); 1534 if (IntTok.isNot(AsmToken::Integer)) { 1535 OnFailure(); 1536 return Error(IntTok.getLoc(), "expected stack index"); 1537 } 1538 switch (IntTok.getIntVal()) { 1539 case 0: RegNo = X86::ST0; break; 1540 case 1: RegNo = X86::ST1; break; 1541 case 2: RegNo = X86::ST2; break; 1542 case 3: RegNo = X86::ST3; break; 1543 case 4: RegNo = X86::ST4; break; 1544 case 5: RegNo = X86::ST5; break; 1545 case 6: RegNo = X86::ST6; break; 1546 case 7: RegNo = X86::ST7; break; 1547 default: 1548 OnFailure(); 1549 return Error(IntTok.getLoc(), "invalid stack index"); 1550 } 1551 1552 // Lex IntTok 1553 Tokens.push_back(IntTok); 1554 Parser.Lex(); 1555 if (Lexer.isNot(AsmToken::RParen)) { 1556 OnFailure(); 1557 return Error(Parser.getTok().getLoc(), "expected ')'"); 1558 } 1559 1560 EndLoc = Parser.getTok().getEndLoc(); 1561 Parser.Lex(); // Eat ')' 1562 return false; 1563 } 1564 1565 EndLoc = Parser.getTok().getEndLoc(); 1566 1567 if (RegNo == 0) { 1568 OnFailure(); 1569 if (isParsingIntelSyntax()) return true; 1570 return Error(StartLoc, "invalid register name", 1571 SMRange(StartLoc, EndLoc)); 1572 } 1573 1574 Parser.Lex(); // Eat identifier token. 1575 return false; 1576 } 1577 1578 bool X86AsmParser::parseRegister(MCRegister &RegNo, SMLoc &StartLoc, 1579 SMLoc &EndLoc) { 1580 return ParseRegister(RegNo, StartLoc, EndLoc, /*RestoreOnFailure=*/false); 1581 } 1582 1583 OperandMatchResultTy X86AsmParser::tryParseRegister(MCRegister &RegNo, 1584 SMLoc &StartLoc, 1585 SMLoc &EndLoc) { 1586 bool Result = 1587 ParseRegister(RegNo, StartLoc, EndLoc, /*RestoreOnFailure=*/true); 1588 bool PendingErrors = getParser().hasPendingError(); 1589 getParser().clearPendingErrors(); 1590 if (PendingErrors) 1591 return MatchOperand_ParseFail; 1592 if (Result) 1593 return MatchOperand_NoMatch; 1594 return MatchOperand_Success; 1595 } 1596 1597 std::unique_ptr<X86Operand> X86AsmParser::DefaultMemSIOperand(SMLoc Loc) { 1598 bool Parse32 = is32BitMode() || Code16GCC; 1599 unsigned Basereg = is64BitMode() ? X86::RSI : (Parse32 ? X86::ESI : X86::SI); 1600 const MCExpr *Disp = MCConstantExpr::create(0, getContext()); 1601 return X86Operand::CreateMem(getPointerWidth(), /*SegReg=*/0, Disp, 1602 /*BaseReg=*/Basereg, /*IndexReg=*/0, /*Scale=*/1, 1603 Loc, Loc, 0); 1604 } 1605 1606 std::unique_ptr<X86Operand> X86AsmParser::DefaultMemDIOperand(SMLoc Loc) { 1607 bool Parse32 = is32BitMode() || Code16GCC; 1608 unsigned Basereg = is64BitMode() ? X86::RDI : (Parse32 ? X86::EDI : X86::DI); 1609 const MCExpr *Disp = MCConstantExpr::create(0, getContext()); 1610 return X86Operand::CreateMem(getPointerWidth(), /*SegReg=*/0, Disp, 1611 /*BaseReg=*/Basereg, /*IndexReg=*/0, /*Scale=*/1, 1612 Loc, Loc, 0); 1613 } 1614 1615 bool X86AsmParser::IsSIReg(unsigned Reg) { 1616 switch (Reg) { 1617 default: llvm_unreachable("Only (R|E)SI and (R|E)DI are expected!"); 1618 case X86::RSI: 1619 case X86::ESI: 1620 case X86::SI: 1621 return true; 1622 case X86::RDI: 1623 case X86::EDI: 1624 case X86::DI: 1625 return false; 1626 } 1627 } 1628 1629 unsigned X86AsmParser::GetSIDIForRegClass(unsigned RegClassID, unsigned Reg, 1630 bool IsSIReg) { 1631 switch (RegClassID) { 1632 default: llvm_unreachable("Unexpected register class"); 1633 case X86::GR64RegClassID: 1634 return IsSIReg ? X86::RSI : X86::RDI; 1635 case X86::GR32RegClassID: 1636 return IsSIReg ? X86::ESI : X86::EDI; 1637 case X86::GR16RegClassID: 1638 return IsSIReg ? X86::SI : X86::DI; 1639 } 1640 } 1641 1642 void X86AsmParser::AddDefaultSrcDestOperands( 1643 OperandVector& Operands, std::unique_ptr<llvm::MCParsedAsmOperand> &&Src, 1644 std::unique_ptr<llvm::MCParsedAsmOperand> &&Dst) { 1645 if (isParsingIntelSyntax()) { 1646 Operands.push_back(std::move(Dst)); 1647 Operands.push_back(std::move(Src)); 1648 } 1649 else { 1650 Operands.push_back(std::move(Src)); 1651 Operands.push_back(std::move(Dst)); 1652 } 1653 } 1654 1655 bool X86AsmParser::VerifyAndAdjustOperands(OperandVector &OrigOperands, 1656 OperandVector &FinalOperands) { 1657 1658 if (OrigOperands.size() > 1) { 1659 // Check if sizes match, OrigOperands also contains the instruction name 1660 assert(OrigOperands.size() == FinalOperands.size() + 1 && 1661 "Operand size mismatch"); 1662 1663 SmallVector<std::pair<SMLoc, std::string>, 2> Warnings; 1664 // Verify types match 1665 int RegClassID = -1; 1666 for (unsigned int i = 0; i < FinalOperands.size(); ++i) { 1667 X86Operand &OrigOp = static_cast<X86Operand &>(*OrigOperands[i + 1]); 1668 X86Operand &FinalOp = static_cast<X86Operand &>(*FinalOperands[i]); 1669 1670 if (FinalOp.isReg() && 1671 (!OrigOp.isReg() || FinalOp.getReg() != OrigOp.getReg())) 1672 // Return false and let a normal complaint about bogus operands happen 1673 return false; 1674 1675 if (FinalOp.isMem()) { 1676 1677 if (!OrigOp.isMem()) 1678 // Return false and let a normal complaint about bogus operands happen 1679 return false; 1680 1681 unsigned OrigReg = OrigOp.Mem.BaseReg; 1682 unsigned FinalReg = FinalOp.Mem.BaseReg; 1683 1684 // If we've already encounterd a register class, make sure all register 1685 // bases are of the same register class 1686 if (RegClassID != -1 && 1687 !X86MCRegisterClasses[RegClassID].contains(OrigReg)) { 1688 return Error(OrigOp.getStartLoc(), 1689 "mismatching source and destination index registers"); 1690 } 1691 1692 if (X86MCRegisterClasses[X86::GR64RegClassID].contains(OrigReg)) 1693 RegClassID = X86::GR64RegClassID; 1694 else if (X86MCRegisterClasses[X86::GR32RegClassID].contains(OrigReg)) 1695 RegClassID = X86::GR32RegClassID; 1696 else if (X86MCRegisterClasses[X86::GR16RegClassID].contains(OrigReg)) 1697 RegClassID = X86::GR16RegClassID; 1698 else 1699 // Unexpected register class type 1700 // Return false and let a normal complaint about bogus operands happen 1701 return false; 1702 1703 bool IsSI = IsSIReg(FinalReg); 1704 FinalReg = GetSIDIForRegClass(RegClassID, FinalReg, IsSI); 1705 1706 if (FinalReg != OrigReg) { 1707 std::string RegName = IsSI ? "ES:(R|E)SI" : "ES:(R|E)DI"; 1708 Warnings.push_back(std::make_pair( 1709 OrigOp.getStartLoc(), 1710 "memory operand is only for determining the size, " + RegName + 1711 " will be used for the location")); 1712 } 1713 1714 FinalOp.Mem.Size = OrigOp.Mem.Size; 1715 FinalOp.Mem.SegReg = OrigOp.Mem.SegReg; 1716 FinalOp.Mem.BaseReg = FinalReg; 1717 } 1718 } 1719 1720 // Produce warnings only if all the operands passed the adjustment - prevent 1721 // legal cases like "movsd (%rax), %xmm0" mistakenly produce warnings 1722 for (auto &WarningMsg : Warnings) { 1723 Warning(WarningMsg.first, WarningMsg.second); 1724 } 1725 1726 // Remove old operands 1727 for (unsigned int i = 0; i < FinalOperands.size(); ++i) 1728 OrigOperands.pop_back(); 1729 } 1730 // OrigOperands.append(FinalOperands.begin(), FinalOperands.end()); 1731 for (unsigned int i = 0; i < FinalOperands.size(); ++i) 1732 OrigOperands.push_back(std::move(FinalOperands[i])); 1733 1734 return false; 1735 } 1736 1737 bool X86AsmParser::parseOperand(OperandVector &Operands, StringRef Name) { 1738 if (isParsingIntelSyntax()) 1739 return parseIntelOperand(Operands, Name); 1740 1741 return parseATTOperand(Operands); 1742 } 1743 1744 bool X86AsmParser::CreateMemForMSInlineAsm( 1745 unsigned SegReg, const MCExpr *Disp, unsigned BaseReg, unsigned IndexReg, 1746 unsigned Scale, SMLoc Start, SMLoc End, unsigned Size, StringRef Identifier, 1747 const InlineAsmIdentifierInfo &Info, OperandVector &Operands) { 1748 // If we found a decl other than a VarDecl, then assume it is a FuncDecl or 1749 // some other label reference. 1750 if (Info.isKind(InlineAsmIdentifierInfo::IK_Label)) { 1751 // Insert an explicit size if the user didn't have one. 1752 if (!Size) { 1753 Size = getPointerWidth(); 1754 InstInfo->AsmRewrites->emplace_back(AOK_SizeDirective, Start, 1755 /*Len=*/0, Size); 1756 } 1757 // Create an absolute memory reference in order to match against 1758 // instructions taking a PC relative operand. 1759 Operands.push_back(X86Operand::CreateMem(getPointerWidth(), Disp, Start, 1760 End, Size, Identifier, 1761 Info.Label.Decl)); 1762 return false; 1763 } 1764 // We either have a direct symbol reference, or an offset from a symbol. The 1765 // parser always puts the symbol on the LHS, so look there for size 1766 // calculation purposes. 1767 unsigned FrontendSize = 0; 1768 void *Decl = nullptr; 1769 bool IsGlobalLV = false; 1770 if (Info.isKind(InlineAsmIdentifierInfo::IK_Var)) { 1771 // Size is in terms of bits in this context. 1772 FrontendSize = Info.Var.Type * 8; 1773 Decl = Info.Var.Decl; 1774 IsGlobalLV = Info.Var.IsGlobalLV; 1775 } 1776 // It is widely common for MS InlineAsm to use a global variable and one/two 1777 // registers in a mmory expression, and though unaccessible via rip/eip. 1778 if (IsGlobalLV && (BaseReg || IndexReg)) { 1779 Operands.push_back(X86Operand::CreateMem(getPointerWidth(), Disp, Start, 1780 End, Size, Identifier, Decl, 0, 1781 BaseReg && IndexReg)); 1782 return false; 1783 } 1784 // Otherwise, we set the base register to a non-zero value 1785 // if we don't know the actual value at this time. This is necessary to 1786 // get the matching correct in some cases. 1787 BaseReg = BaseReg ? BaseReg : 1; 1788 Operands.push_back(X86Operand::CreateMem( 1789 getPointerWidth(), SegReg, Disp, BaseReg, IndexReg, Scale, Start, End, 1790 Size, 1791 /*DefaultBaseReg=*/X86::RIP, Identifier, Decl, FrontendSize)); 1792 return false; 1793 } 1794 1795 // Some binary bitwise operators have a named synonymous 1796 // Query a candidate string for being such a named operator 1797 // and if so - invoke the appropriate handler 1798 bool X86AsmParser::ParseIntelNamedOperator(StringRef Name, 1799 IntelExprStateMachine &SM, 1800 bool &ParseError, SMLoc &End) { 1801 // A named operator should be either lower or upper case, but not a mix... 1802 // except in MASM, which uses full case-insensitivity. 1803 if (Name.compare(Name.lower()) && Name.compare(Name.upper()) && 1804 !getParser().isParsingMasm()) 1805 return false; 1806 if (Name.equals_insensitive("not")) { 1807 SM.onNot(); 1808 } else if (Name.equals_insensitive("or")) { 1809 SM.onOr(); 1810 } else if (Name.equals_insensitive("shl")) { 1811 SM.onLShift(); 1812 } else if (Name.equals_insensitive("shr")) { 1813 SM.onRShift(); 1814 } else if (Name.equals_insensitive("xor")) { 1815 SM.onXor(); 1816 } else if (Name.equals_insensitive("and")) { 1817 SM.onAnd(); 1818 } else if (Name.equals_insensitive("mod")) { 1819 SM.onMod(); 1820 } else if (Name.equals_insensitive("offset")) { 1821 SMLoc OffsetLoc = getTok().getLoc(); 1822 const MCExpr *Val = nullptr; 1823 StringRef ID; 1824 InlineAsmIdentifierInfo Info; 1825 ParseError = ParseIntelOffsetOperator(Val, ID, Info, End); 1826 if (ParseError) 1827 return true; 1828 StringRef ErrMsg; 1829 ParseError = 1830 SM.onOffset(Val, OffsetLoc, ID, Info, isParsingMSInlineAsm(), ErrMsg); 1831 if (ParseError) 1832 return Error(SMLoc::getFromPointer(Name.data()), ErrMsg); 1833 } else { 1834 return false; 1835 } 1836 if (!Name.equals_insensitive("offset")) 1837 End = consumeToken(); 1838 return true; 1839 } 1840 bool X86AsmParser::ParseMasmNamedOperator(StringRef Name, 1841 IntelExprStateMachine &SM, 1842 bool &ParseError, SMLoc &End) { 1843 if (Name.equals_insensitive("eq")) { 1844 SM.onEq(); 1845 } else if (Name.equals_insensitive("ne")) { 1846 SM.onNE(); 1847 } else if (Name.equals_insensitive("lt")) { 1848 SM.onLT(); 1849 } else if (Name.equals_insensitive("le")) { 1850 SM.onLE(); 1851 } else if (Name.equals_insensitive("gt")) { 1852 SM.onGT(); 1853 } else if (Name.equals_insensitive("ge")) { 1854 SM.onGE(); 1855 } else { 1856 return false; 1857 } 1858 End = consumeToken(); 1859 return true; 1860 } 1861 1862 // Check if current intel expression append after an operand. 1863 // Like: [Operand][Intel Expression] 1864 void X86AsmParser::tryParseOperandIdx(AsmToken::TokenKind PrevTK, 1865 IntelExprStateMachine &SM) { 1866 if (PrevTK != AsmToken::RBrac) 1867 return; 1868 1869 SM.setAppendAfterOperand(); 1870 } 1871 1872 bool X86AsmParser::ParseIntelExpression(IntelExprStateMachine &SM, SMLoc &End) { 1873 MCAsmParser &Parser = getParser(); 1874 StringRef ErrMsg; 1875 1876 AsmToken::TokenKind PrevTK = AsmToken::Error; 1877 1878 if (getContext().getObjectFileInfo()->isPositionIndependent()) 1879 SM.setPIC(); 1880 1881 bool Done = false; 1882 while (!Done) { 1883 // Get a fresh reference on each loop iteration in case the previous 1884 // iteration moved the token storage during UnLex(). 1885 const AsmToken &Tok = Parser.getTok(); 1886 1887 bool UpdateLocLex = true; 1888 AsmToken::TokenKind TK = getLexer().getKind(); 1889 1890 switch (TK) { 1891 default: 1892 if ((Done = SM.isValidEndState())) 1893 break; 1894 return Error(Tok.getLoc(), "unknown token in expression"); 1895 case AsmToken::Error: 1896 return Error(getLexer().getErrLoc(), getLexer().getErr()); 1897 break; 1898 case AsmToken::EndOfStatement: 1899 Done = true; 1900 break; 1901 case AsmToken::Real: 1902 // DotOperator: [ebx].0 1903 UpdateLocLex = false; 1904 if (ParseIntelDotOperator(SM, End)) 1905 return true; 1906 break; 1907 case AsmToken::Dot: 1908 if (!Parser.isParsingMasm()) { 1909 if ((Done = SM.isValidEndState())) 1910 break; 1911 return Error(Tok.getLoc(), "unknown token in expression"); 1912 } 1913 // MASM allows spaces around the dot operator (e.g., "var . x") 1914 Lex(); 1915 UpdateLocLex = false; 1916 if (ParseIntelDotOperator(SM, End)) 1917 return true; 1918 break; 1919 case AsmToken::Dollar: 1920 if (!Parser.isParsingMasm()) { 1921 if ((Done = SM.isValidEndState())) 1922 break; 1923 return Error(Tok.getLoc(), "unknown token in expression"); 1924 } 1925 [[fallthrough]]; 1926 case AsmToken::String: { 1927 if (Parser.isParsingMasm()) { 1928 // MASM parsers handle strings in expressions as constants. 1929 SMLoc ValueLoc = Tok.getLoc(); 1930 int64_t Res; 1931 const MCExpr *Val; 1932 if (Parser.parsePrimaryExpr(Val, End, nullptr)) 1933 return true; 1934 UpdateLocLex = false; 1935 if (!Val->evaluateAsAbsolute(Res, getStreamer().getAssemblerPtr())) 1936 return Error(ValueLoc, "expected absolute value"); 1937 if (SM.onInteger(Res, ErrMsg)) 1938 return Error(ValueLoc, ErrMsg); 1939 break; 1940 } 1941 [[fallthrough]]; 1942 } 1943 case AsmToken::At: 1944 case AsmToken::Identifier: { 1945 SMLoc IdentLoc = Tok.getLoc(); 1946 StringRef Identifier = Tok.getString(); 1947 UpdateLocLex = false; 1948 if (Parser.isParsingMasm()) { 1949 size_t DotOffset = Identifier.find_first_of('.'); 1950 if (DotOffset != StringRef::npos) { 1951 consumeToken(); 1952 StringRef LHS = Identifier.slice(0, DotOffset); 1953 StringRef Dot = Identifier.slice(DotOffset, DotOffset + 1); 1954 StringRef RHS = Identifier.slice(DotOffset + 1, StringRef::npos); 1955 if (!RHS.empty()) { 1956 getLexer().UnLex(AsmToken(AsmToken::Identifier, RHS)); 1957 } 1958 getLexer().UnLex(AsmToken(AsmToken::Dot, Dot)); 1959 if (!LHS.empty()) { 1960 getLexer().UnLex(AsmToken(AsmToken::Identifier, LHS)); 1961 } 1962 break; 1963 } 1964 } 1965 // (MASM only) <TYPE> PTR operator 1966 if (Parser.isParsingMasm()) { 1967 const AsmToken &NextTok = getLexer().peekTok(); 1968 if (NextTok.is(AsmToken::Identifier) && 1969 NextTok.getIdentifier().equals_insensitive("ptr")) { 1970 AsmTypeInfo Info; 1971 if (Parser.lookUpType(Identifier, Info)) 1972 return Error(Tok.getLoc(), "unknown type"); 1973 SM.onCast(Info); 1974 // Eat type and PTR. 1975 consumeToken(); 1976 End = consumeToken(); 1977 break; 1978 } 1979 } 1980 // Register, or (MASM only) <register>.<field> 1981 MCRegister Reg; 1982 if (Tok.is(AsmToken::Identifier)) { 1983 if (!ParseRegister(Reg, IdentLoc, End, /*RestoreOnFailure=*/true)) { 1984 if (SM.onRegister(Reg, ErrMsg)) 1985 return Error(IdentLoc, ErrMsg); 1986 break; 1987 } 1988 if (Parser.isParsingMasm()) { 1989 const std::pair<StringRef, StringRef> IDField = 1990 Tok.getString().split('.'); 1991 const StringRef ID = IDField.first, Field = IDField.second; 1992 SMLoc IDEndLoc = SMLoc::getFromPointer(ID.data() + ID.size()); 1993 if (!Field.empty() && 1994 !MatchRegisterByName(Reg, ID, IdentLoc, IDEndLoc)) { 1995 if (SM.onRegister(Reg, ErrMsg)) 1996 return Error(IdentLoc, ErrMsg); 1997 1998 AsmFieldInfo Info; 1999 SMLoc FieldStartLoc = SMLoc::getFromPointer(Field.data()); 2000 if (Parser.lookUpField(Field, Info)) 2001 return Error(FieldStartLoc, "unknown offset"); 2002 else if (SM.onPlus(ErrMsg)) 2003 return Error(getTok().getLoc(), ErrMsg); 2004 else if (SM.onInteger(Info.Offset, ErrMsg)) 2005 return Error(IdentLoc, ErrMsg); 2006 SM.setTypeInfo(Info.Type); 2007 2008 End = consumeToken(); 2009 break; 2010 } 2011 } 2012 } 2013 // Operator synonymous ("not", "or" etc.) 2014 bool ParseError = false; 2015 if (ParseIntelNamedOperator(Identifier, SM, ParseError, End)) { 2016 if (ParseError) 2017 return true; 2018 break; 2019 } 2020 if (Parser.isParsingMasm() && 2021 ParseMasmNamedOperator(Identifier, SM, ParseError, End)) { 2022 if (ParseError) 2023 return true; 2024 break; 2025 } 2026 // Symbol reference, when parsing assembly content 2027 InlineAsmIdentifierInfo Info; 2028 AsmFieldInfo FieldInfo; 2029 const MCExpr *Val; 2030 if (isParsingMSInlineAsm() || Parser.isParsingMasm()) { 2031 // MS Dot Operator expression 2032 if (Identifier.count('.') && 2033 (PrevTK == AsmToken::RBrac || PrevTK == AsmToken::RParen)) { 2034 if (ParseIntelDotOperator(SM, End)) 2035 return true; 2036 break; 2037 } 2038 } 2039 if (isParsingMSInlineAsm()) { 2040 // MS InlineAsm operators (TYPE/LENGTH/SIZE) 2041 if (unsigned OpKind = IdentifyIntelInlineAsmOperator(Identifier)) { 2042 if (int64_t Val = ParseIntelInlineAsmOperator(OpKind)) { 2043 if (SM.onInteger(Val, ErrMsg)) 2044 return Error(IdentLoc, ErrMsg); 2045 } else { 2046 return true; 2047 } 2048 break; 2049 } 2050 // MS InlineAsm identifier 2051 // Call parseIdentifier() to combine @ with the identifier behind it. 2052 if (TK == AsmToken::At && Parser.parseIdentifier(Identifier)) 2053 return Error(IdentLoc, "expected identifier"); 2054 if (ParseIntelInlineAsmIdentifier(Val, Identifier, Info, false, End)) 2055 return true; 2056 else if (SM.onIdentifierExpr(Val, Identifier, Info, FieldInfo.Type, 2057 true, ErrMsg)) 2058 return Error(IdentLoc, ErrMsg); 2059 break; 2060 } 2061 if (Parser.isParsingMasm()) { 2062 if (unsigned OpKind = IdentifyMasmOperator(Identifier)) { 2063 int64_t Val; 2064 if (ParseMasmOperator(OpKind, Val)) 2065 return true; 2066 if (SM.onInteger(Val, ErrMsg)) 2067 return Error(IdentLoc, ErrMsg); 2068 break; 2069 } 2070 if (!getParser().lookUpType(Identifier, FieldInfo.Type)) { 2071 // Field offset immediate; <TYPE>.<field specification> 2072 Lex(); // eat type 2073 bool EndDot = parseOptionalToken(AsmToken::Dot); 2074 while (EndDot || (getTok().is(AsmToken::Identifier) && 2075 getTok().getString().startswith("."))) { 2076 getParser().parseIdentifier(Identifier); 2077 if (!EndDot) 2078 Identifier.consume_front("."); 2079 EndDot = Identifier.consume_back("."); 2080 if (getParser().lookUpField(FieldInfo.Type.Name, Identifier, 2081 FieldInfo)) { 2082 SMLoc IDEnd = 2083 SMLoc::getFromPointer(Identifier.data() + Identifier.size()); 2084 return Error(IdentLoc, "Unable to lookup field reference!", 2085 SMRange(IdentLoc, IDEnd)); 2086 } 2087 if (!EndDot) 2088 EndDot = parseOptionalToken(AsmToken::Dot); 2089 } 2090 if (SM.onInteger(FieldInfo.Offset, ErrMsg)) 2091 return Error(IdentLoc, ErrMsg); 2092 break; 2093 } 2094 } 2095 if (getParser().parsePrimaryExpr(Val, End, &FieldInfo.Type)) { 2096 return Error(Tok.getLoc(), "Unexpected identifier!"); 2097 } else if (SM.onIdentifierExpr(Val, Identifier, Info, FieldInfo.Type, 2098 false, ErrMsg)) { 2099 return Error(IdentLoc, ErrMsg); 2100 } 2101 break; 2102 } 2103 case AsmToken::Integer: { 2104 // Look for 'b' or 'f' following an Integer as a directional label 2105 SMLoc Loc = getTok().getLoc(); 2106 int64_t IntVal = getTok().getIntVal(); 2107 End = consumeToken(); 2108 UpdateLocLex = false; 2109 if (getLexer().getKind() == AsmToken::Identifier) { 2110 StringRef IDVal = getTok().getString(); 2111 if (IDVal == "f" || IDVal == "b") { 2112 MCSymbol *Sym = 2113 getContext().getDirectionalLocalSymbol(IntVal, IDVal == "b"); 2114 MCSymbolRefExpr::VariantKind Variant = MCSymbolRefExpr::VK_None; 2115 const MCExpr *Val = 2116 MCSymbolRefExpr::create(Sym, Variant, getContext()); 2117 if (IDVal == "b" && Sym->isUndefined()) 2118 return Error(Loc, "invalid reference to undefined symbol"); 2119 StringRef Identifier = Sym->getName(); 2120 InlineAsmIdentifierInfo Info; 2121 AsmTypeInfo Type; 2122 if (SM.onIdentifierExpr(Val, Identifier, Info, Type, 2123 isParsingMSInlineAsm(), ErrMsg)) 2124 return Error(Loc, ErrMsg); 2125 End = consumeToken(); 2126 } else { 2127 if (SM.onInteger(IntVal, ErrMsg)) 2128 return Error(Loc, ErrMsg); 2129 } 2130 } else { 2131 if (SM.onInteger(IntVal, ErrMsg)) 2132 return Error(Loc, ErrMsg); 2133 } 2134 break; 2135 } 2136 case AsmToken::Plus: 2137 if (SM.onPlus(ErrMsg)) 2138 return Error(getTok().getLoc(), ErrMsg); 2139 break; 2140 case AsmToken::Minus: 2141 if (SM.onMinus(ErrMsg)) 2142 return Error(getTok().getLoc(), ErrMsg); 2143 break; 2144 case AsmToken::Tilde: SM.onNot(); break; 2145 case AsmToken::Star: SM.onStar(); break; 2146 case AsmToken::Slash: SM.onDivide(); break; 2147 case AsmToken::Percent: SM.onMod(); break; 2148 case AsmToken::Pipe: SM.onOr(); break; 2149 case AsmToken::Caret: SM.onXor(); break; 2150 case AsmToken::Amp: SM.onAnd(); break; 2151 case AsmToken::LessLess: 2152 SM.onLShift(); break; 2153 case AsmToken::GreaterGreater: 2154 SM.onRShift(); break; 2155 case AsmToken::LBrac: 2156 if (SM.onLBrac()) 2157 return Error(Tok.getLoc(), "unexpected bracket encountered"); 2158 tryParseOperandIdx(PrevTK, SM); 2159 break; 2160 case AsmToken::RBrac: 2161 if (SM.onRBrac(ErrMsg)) { 2162 return Error(Tok.getLoc(), ErrMsg); 2163 } 2164 break; 2165 case AsmToken::LParen: SM.onLParen(); break; 2166 case AsmToken::RParen: SM.onRParen(); break; 2167 } 2168 if (SM.hadError()) 2169 return Error(Tok.getLoc(), "unknown token in expression"); 2170 2171 if (!Done && UpdateLocLex) 2172 End = consumeToken(); 2173 2174 PrevTK = TK; 2175 } 2176 return false; 2177 } 2178 2179 void X86AsmParser::RewriteIntelExpression(IntelExprStateMachine &SM, 2180 SMLoc Start, SMLoc End) { 2181 SMLoc Loc = Start; 2182 unsigned ExprLen = End.getPointer() - Start.getPointer(); 2183 // Skip everything before a symbol displacement (if we have one) 2184 if (SM.getSym() && !SM.isOffsetOperator()) { 2185 StringRef SymName = SM.getSymName(); 2186 if (unsigned Len = SymName.data() - Start.getPointer()) 2187 InstInfo->AsmRewrites->emplace_back(AOK_Skip, Start, Len); 2188 Loc = SMLoc::getFromPointer(SymName.data() + SymName.size()); 2189 ExprLen = End.getPointer() - (SymName.data() + SymName.size()); 2190 // If we have only a symbol than there's no need for complex rewrite, 2191 // simply skip everything after it 2192 if (!(SM.getBaseReg() || SM.getIndexReg() || SM.getImm())) { 2193 if (ExprLen) 2194 InstInfo->AsmRewrites->emplace_back(AOK_Skip, Loc, ExprLen); 2195 return; 2196 } 2197 } 2198 // Build an Intel Expression rewrite 2199 StringRef BaseRegStr; 2200 StringRef IndexRegStr; 2201 StringRef OffsetNameStr; 2202 if (SM.getBaseReg()) 2203 BaseRegStr = X86IntelInstPrinter::getRegisterName(SM.getBaseReg()); 2204 if (SM.getIndexReg()) 2205 IndexRegStr = X86IntelInstPrinter::getRegisterName(SM.getIndexReg()); 2206 if (SM.isOffsetOperator()) 2207 OffsetNameStr = SM.getSymName(); 2208 // Emit it 2209 IntelExpr Expr(BaseRegStr, IndexRegStr, SM.getScale(), OffsetNameStr, 2210 SM.getImm(), SM.isMemExpr()); 2211 InstInfo->AsmRewrites->emplace_back(Loc, ExprLen, Expr); 2212 } 2213 2214 // Inline assembly may use variable names with namespace alias qualifiers. 2215 bool X86AsmParser::ParseIntelInlineAsmIdentifier( 2216 const MCExpr *&Val, StringRef &Identifier, InlineAsmIdentifierInfo &Info, 2217 bool IsUnevaluatedOperand, SMLoc &End, bool IsParsingOffsetOperator) { 2218 MCAsmParser &Parser = getParser(); 2219 assert(isParsingMSInlineAsm() && "Expected to be parsing inline assembly."); 2220 Val = nullptr; 2221 2222 StringRef LineBuf(Identifier.data()); 2223 SemaCallback->LookupInlineAsmIdentifier(LineBuf, Info, IsUnevaluatedOperand); 2224 2225 const AsmToken &Tok = Parser.getTok(); 2226 SMLoc Loc = Tok.getLoc(); 2227 2228 // Advance the token stream until the end of the current token is 2229 // after the end of what the frontend claimed. 2230 const char *EndPtr = Tok.getLoc().getPointer() + LineBuf.size(); 2231 do { 2232 End = Tok.getEndLoc(); 2233 getLexer().Lex(); 2234 } while (End.getPointer() < EndPtr); 2235 Identifier = LineBuf; 2236 2237 // The frontend should end parsing on an assembler token boundary, unless it 2238 // failed parsing. 2239 assert((End.getPointer() == EndPtr || 2240 Info.isKind(InlineAsmIdentifierInfo::IK_Invalid)) && 2241 "frontend claimed part of a token?"); 2242 2243 // If the identifier lookup was unsuccessful, assume that we are dealing with 2244 // a label. 2245 if (Info.isKind(InlineAsmIdentifierInfo::IK_Invalid)) { 2246 StringRef InternalName = 2247 SemaCallback->LookupInlineAsmLabel(Identifier, getSourceManager(), 2248 Loc, false); 2249 assert(InternalName.size() && "We should have an internal name here."); 2250 // Push a rewrite for replacing the identifier name with the internal name, 2251 // unless we are parsing the operand of an offset operator 2252 if (!IsParsingOffsetOperator) 2253 InstInfo->AsmRewrites->emplace_back(AOK_Label, Loc, Identifier.size(), 2254 InternalName); 2255 else 2256 Identifier = InternalName; 2257 } else if (Info.isKind(InlineAsmIdentifierInfo::IK_EnumVal)) 2258 return false; 2259 // Create the symbol reference. 2260 MCSymbol *Sym = getContext().getOrCreateSymbol(Identifier); 2261 MCSymbolRefExpr::VariantKind Variant = MCSymbolRefExpr::VK_None; 2262 Val = MCSymbolRefExpr::create(Sym, Variant, getParser().getContext()); 2263 return false; 2264 } 2265 2266 //ParseRoundingModeOp - Parse AVX-512 rounding mode operand 2267 bool X86AsmParser::ParseRoundingModeOp(SMLoc Start, OperandVector &Operands) { 2268 MCAsmParser &Parser = getParser(); 2269 const AsmToken &Tok = Parser.getTok(); 2270 // Eat "{" and mark the current place. 2271 const SMLoc consumedToken = consumeToken(); 2272 if (Tok.isNot(AsmToken::Identifier)) 2273 return Error(Tok.getLoc(), "Expected an identifier after {"); 2274 if (Tok.getIdentifier().startswith("r")){ 2275 int rndMode = StringSwitch<int>(Tok.getIdentifier()) 2276 .Case("rn", X86::STATIC_ROUNDING::TO_NEAREST_INT) 2277 .Case("rd", X86::STATIC_ROUNDING::TO_NEG_INF) 2278 .Case("ru", X86::STATIC_ROUNDING::TO_POS_INF) 2279 .Case("rz", X86::STATIC_ROUNDING::TO_ZERO) 2280 .Default(-1); 2281 if (-1 == rndMode) 2282 return Error(Tok.getLoc(), "Invalid rounding mode."); 2283 Parser.Lex(); // Eat "r*" of r*-sae 2284 if (!getLexer().is(AsmToken::Minus)) 2285 return Error(Tok.getLoc(), "Expected - at this point"); 2286 Parser.Lex(); // Eat "-" 2287 Parser.Lex(); // Eat the sae 2288 if (!getLexer().is(AsmToken::RCurly)) 2289 return Error(Tok.getLoc(), "Expected } at this point"); 2290 SMLoc End = Tok.getEndLoc(); 2291 Parser.Lex(); // Eat "}" 2292 const MCExpr *RndModeOp = 2293 MCConstantExpr::create(rndMode, Parser.getContext()); 2294 Operands.push_back(X86Operand::CreateImm(RndModeOp, Start, End)); 2295 return false; 2296 } 2297 if(Tok.getIdentifier().equals("sae")){ 2298 Parser.Lex(); // Eat the sae 2299 if (!getLexer().is(AsmToken::RCurly)) 2300 return Error(Tok.getLoc(), "Expected } at this point"); 2301 Parser.Lex(); // Eat "}" 2302 Operands.push_back(X86Operand::CreateToken("{sae}", consumedToken)); 2303 return false; 2304 } 2305 return Error(Tok.getLoc(), "unknown token in expression"); 2306 } 2307 2308 /// Parse the '.' operator. 2309 bool X86AsmParser::ParseIntelDotOperator(IntelExprStateMachine &SM, 2310 SMLoc &End) { 2311 const AsmToken &Tok = getTok(); 2312 AsmFieldInfo Info; 2313 2314 // Drop the optional '.'. 2315 StringRef DotDispStr = Tok.getString(); 2316 if (DotDispStr.startswith(".")) 2317 DotDispStr = DotDispStr.drop_front(1); 2318 StringRef TrailingDot; 2319 2320 // .Imm gets lexed as a real. 2321 if (Tok.is(AsmToken::Real)) { 2322 APInt DotDisp; 2323 DotDispStr.getAsInteger(10, DotDisp); 2324 Info.Offset = DotDisp.getZExtValue(); 2325 } else if ((isParsingMSInlineAsm() || getParser().isParsingMasm()) && 2326 Tok.is(AsmToken::Identifier)) { 2327 if (DotDispStr.endswith(".")) { 2328 TrailingDot = DotDispStr.substr(DotDispStr.size() - 1); 2329 DotDispStr = DotDispStr.drop_back(1); 2330 } 2331 const std::pair<StringRef, StringRef> BaseMember = DotDispStr.split('.'); 2332 const StringRef Base = BaseMember.first, Member = BaseMember.second; 2333 if (getParser().lookUpField(SM.getType(), DotDispStr, Info) && 2334 getParser().lookUpField(SM.getSymName(), DotDispStr, Info) && 2335 getParser().lookUpField(DotDispStr, Info) && 2336 (!SemaCallback || 2337 SemaCallback->LookupInlineAsmField(Base, Member, Info.Offset))) 2338 return Error(Tok.getLoc(), "Unable to lookup field reference!"); 2339 } else { 2340 return Error(Tok.getLoc(), "Unexpected token type!"); 2341 } 2342 2343 // Eat the DotExpression and update End 2344 End = SMLoc::getFromPointer(DotDispStr.data()); 2345 const char *DotExprEndLoc = DotDispStr.data() + DotDispStr.size(); 2346 while (Tok.getLoc().getPointer() < DotExprEndLoc) 2347 Lex(); 2348 if (!TrailingDot.empty()) 2349 getLexer().UnLex(AsmToken(AsmToken::Dot, TrailingDot)); 2350 SM.addImm(Info.Offset); 2351 SM.setTypeInfo(Info.Type); 2352 return false; 2353 } 2354 2355 /// Parse the 'offset' operator. 2356 /// This operator is used to specify the location of a given operand 2357 bool X86AsmParser::ParseIntelOffsetOperator(const MCExpr *&Val, StringRef &ID, 2358 InlineAsmIdentifierInfo &Info, 2359 SMLoc &End) { 2360 // Eat offset, mark start of identifier. 2361 SMLoc Start = Lex().getLoc(); 2362 ID = getTok().getString(); 2363 if (!isParsingMSInlineAsm()) { 2364 if ((getTok().isNot(AsmToken::Identifier) && 2365 getTok().isNot(AsmToken::String)) || 2366 getParser().parsePrimaryExpr(Val, End, nullptr)) 2367 return Error(Start, "unexpected token!"); 2368 } else if (ParseIntelInlineAsmIdentifier(Val, ID, Info, false, End, true)) { 2369 return Error(Start, "unable to lookup expression"); 2370 } else if (Info.isKind(InlineAsmIdentifierInfo::IK_EnumVal)) { 2371 return Error(Start, "offset operator cannot yet handle constants"); 2372 } 2373 return false; 2374 } 2375 2376 // Query a candidate string for being an Intel assembly operator 2377 // Report back its kind, or IOK_INVALID if does not evaluated as a known one 2378 unsigned X86AsmParser::IdentifyIntelInlineAsmOperator(StringRef Name) { 2379 return StringSwitch<unsigned>(Name) 2380 .Cases("TYPE","type",IOK_TYPE) 2381 .Cases("SIZE","size",IOK_SIZE) 2382 .Cases("LENGTH","length",IOK_LENGTH) 2383 .Default(IOK_INVALID); 2384 } 2385 2386 /// Parse the 'LENGTH', 'TYPE' and 'SIZE' operators. The LENGTH operator 2387 /// returns the number of elements in an array. It returns the value 1 for 2388 /// non-array variables. The SIZE operator returns the size of a C or C++ 2389 /// variable. A variable's size is the product of its LENGTH and TYPE. The 2390 /// TYPE operator returns the size of a C or C++ type or variable. If the 2391 /// variable is an array, TYPE returns the size of a single element. 2392 unsigned X86AsmParser::ParseIntelInlineAsmOperator(unsigned OpKind) { 2393 MCAsmParser &Parser = getParser(); 2394 const AsmToken &Tok = Parser.getTok(); 2395 Parser.Lex(); // Eat operator. 2396 2397 const MCExpr *Val = nullptr; 2398 InlineAsmIdentifierInfo Info; 2399 SMLoc Start = Tok.getLoc(), End; 2400 StringRef Identifier = Tok.getString(); 2401 if (ParseIntelInlineAsmIdentifier(Val, Identifier, Info, 2402 /*IsUnevaluatedOperand=*/true, End)) 2403 return 0; 2404 2405 if (!Info.isKind(InlineAsmIdentifierInfo::IK_Var)) { 2406 Error(Start, "unable to lookup expression"); 2407 return 0; 2408 } 2409 2410 unsigned CVal = 0; 2411 switch(OpKind) { 2412 default: llvm_unreachable("Unexpected operand kind!"); 2413 case IOK_LENGTH: CVal = Info.Var.Length; break; 2414 case IOK_SIZE: CVal = Info.Var.Size; break; 2415 case IOK_TYPE: CVal = Info.Var.Type; break; 2416 } 2417 2418 return CVal; 2419 } 2420 2421 // Query a candidate string for being an Intel assembly operator 2422 // Report back its kind, or IOK_INVALID if does not evaluated as a known one 2423 unsigned X86AsmParser::IdentifyMasmOperator(StringRef Name) { 2424 return StringSwitch<unsigned>(Name.lower()) 2425 .Case("type", MOK_TYPE) 2426 .Cases("size", "sizeof", MOK_SIZEOF) 2427 .Cases("length", "lengthof", MOK_LENGTHOF) 2428 .Default(MOK_INVALID); 2429 } 2430 2431 /// Parse the 'LENGTHOF', 'SIZEOF', and 'TYPE' operators. The LENGTHOF operator 2432 /// returns the number of elements in an array. It returns the value 1 for 2433 /// non-array variables. The SIZEOF operator returns the size of a type or 2434 /// variable in bytes. A variable's size is the product of its LENGTH and TYPE. 2435 /// The TYPE operator returns the size of a variable. If the variable is an 2436 /// array, TYPE returns the size of a single element. 2437 bool X86AsmParser::ParseMasmOperator(unsigned OpKind, int64_t &Val) { 2438 MCAsmParser &Parser = getParser(); 2439 SMLoc OpLoc = Parser.getTok().getLoc(); 2440 Parser.Lex(); // Eat operator. 2441 2442 Val = 0; 2443 if (OpKind == MOK_SIZEOF || OpKind == MOK_TYPE) { 2444 // Check for SIZEOF(<type>) and TYPE(<type>). 2445 bool InParens = Parser.getTok().is(AsmToken::LParen); 2446 const AsmToken &IDTok = InParens ? getLexer().peekTok() : Parser.getTok(); 2447 AsmTypeInfo Type; 2448 if (IDTok.is(AsmToken::Identifier) && 2449 !Parser.lookUpType(IDTok.getIdentifier(), Type)) { 2450 Val = Type.Size; 2451 2452 // Eat tokens. 2453 if (InParens) 2454 parseToken(AsmToken::LParen); 2455 parseToken(AsmToken::Identifier); 2456 if (InParens) 2457 parseToken(AsmToken::RParen); 2458 } 2459 } 2460 2461 if (!Val) { 2462 IntelExprStateMachine SM; 2463 SMLoc End, Start = Parser.getTok().getLoc(); 2464 if (ParseIntelExpression(SM, End)) 2465 return true; 2466 2467 switch (OpKind) { 2468 default: 2469 llvm_unreachable("Unexpected operand kind!"); 2470 case MOK_SIZEOF: 2471 Val = SM.getSize(); 2472 break; 2473 case MOK_LENGTHOF: 2474 Val = SM.getLength(); 2475 break; 2476 case MOK_TYPE: 2477 Val = SM.getElementSize(); 2478 break; 2479 } 2480 2481 if (!Val) 2482 return Error(OpLoc, "expression has unknown type", SMRange(Start, End)); 2483 } 2484 2485 return false; 2486 } 2487 2488 bool X86AsmParser::ParseIntelMemoryOperandSize(unsigned &Size) { 2489 Size = StringSwitch<unsigned>(getTok().getString()) 2490 .Cases("BYTE", "byte", 8) 2491 .Cases("WORD", "word", 16) 2492 .Cases("DWORD", "dword", 32) 2493 .Cases("FLOAT", "float", 32) 2494 .Cases("LONG", "long", 32) 2495 .Cases("FWORD", "fword", 48) 2496 .Cases("DOUBLE", "double", 64) 2497 .Cases("QWORD", "qword", 64) 2498 .Cases("MMWORD","mmword", 64) 2499 .Cases("XWORD", "xword", 80) 2500 .Cases("TBYTE", "tbyte", 80) 2501 .Cases("XMMWORD", "xmmword", 128) 2502 .Cases("YMMWORD", "ymmword", 256) 2503 .Cases("ZMMWORD", "zmmword", 512) 2504 .Default(0); 2505 if (Size) { 2506 const AsmToken &Tok = Lex(); // Eat operand size (e.g., byte, word). 2507 if (!(Tok.getString().equals("PTR") || Tok.getString().equals("ptr"))) 2508 return Error(Tok.getLoc(), "Expected 'PTR' or 'ptr' token!"); 2509 Lex(); // Eat ptr. 2510 } 2511 return false; 2512 } 2513 2514 bool X86AsmParser::parseIntelOperand(OperandVector &Operands, StringRef Name) { 2515 MCAsmParser &Parser = getParser(); 2516 const AsmToken &Tok = Parser.getTok(); 2517 SMLoc Start, End; 2518 2519 // Parse optional Size directive. 2520 unsigned Size; 2521 if (ParseIntelMemoryOperandSize(Size)) 2522 return true; 2523 bool PtrInOperand = bool(Size); 2524 2525 Start = Tok.getLoc(); 2526 2527 // Rounding mode operand. 2528 if (getLexer().is(AsmToken::LCurly)) 2529 return ParseRoundingModeOp(Start, Operands); 2530 2531 // Register operand. 2532 MCRegister RegNo; 2533 if (Tok.is(AsmToken::Identifier) && !parseRegister(RegNo, Start, End)) { 2534 if (RegNo == X86::RIP) 2535 return Error(Start, "rip can only be used as a base register"); 2536 // A Register followed by ':' is considered a segment override 2537 if (Tok.isNot(AsmToken::Colon)) { 2538 if (PtrInOperand) 2539 return Error(Start, "expected memory operand after 'ptr', " 2540 "found register operand instead"); 2541 Operands.push_back(X86Operand::CreateReg(RegNo, Start, End)); 2542 return false; 2543 } 2544 // An alleged segment override. check if we have a valid segment register 2545 if (!X86MCRegisterClasses[X86::SEGMENT_REGRegClassID].contains(RegNo)) 2546 return Error(Start, "invalid segment register"); 2547 // Eat ':' and update Start location 2548 Start = Lex().getLoc(); 2549 } 2550 2551 // Immediates and Memory 2552 IntelExprStateMachine SM; 2553 if (ParseIntelExpression(SM, End)) 2554 return true; 2555 2556 if (isParsingMSInlineAsm()) 2557 RewriteIntelExpression(SM, Start, Tok.getLoc()); 2558 2559 int64_t Imm = SM.getImm(); 2560 const MCExpr *Disp = SM.getSym(); 2561 const MCExpr *ImmDisp = MCConstantExpr::create(Imm, getContext()); 2562 if (Disp && Imm) 2563 Disp = MCBinaryExpr::createAdd(Disp, ImmDisp, getContext()); 2564 if (!Disp) 2565 Disp = ImmDisp; 2566 2567 // RegNo != 0 specifies a valid segment register, 2568 // and we are parsing a segment override 2569 if (!SM.isMemExpr() && !RegNo) { 2570 if (isParsingMSInlineAsm() && SM.isOffsetOperator()) { 2571 const InlineAsmIdentifierInfo &Info = SM.getIdentifierInfo(); 2572 if (Info.isKind(InlineAsmIdentifierInfo::IK_Var)) { 2573 // Disp includes the address of a variable; make sure this is recorded 2574 // for later handling. 2575 Operands.push_back(X86Operand::CreateImm(Disp, Start, End, 2576 SM.getSymName(), Info.Var.Decl, 2577 Info.Var.IsGlobalLV)); 2578 return false; 2579 } 2580 } 2581 2582 Operands.push_back(X86Operand::CreateImm(Disp, Start, End)); 2583 return false; 2584 } 2585 2586 StringRef ErrMsg; 2587 unsigned BaseReg = SM.getBaseReg(); 2588 unsigned IndexReg = SM.getIndexReg(); 2589 if (IndexReg && BaseReg == X86::RIP) 2590 BaseReg = 0; 2591 unsigned Scale = SM.getScale(); 2592 if (!PtrInOperand) 2593 Size = SM.getElementSize() << 3; 2594 2595 if (Scale == 0 && BaseReg != X86::ESP && BaseReg != X86::RSP && 2596 (IndexReg == X86::ESP || IndexReg == X86::RSP)) 2597 std::swap(BaseReg, IndexReg); 2598 2599 // If BaseReg is a vector register and IndexReg is not, swap them unless 2600 // Scale was specified in which case it would be an error. 2601 if (Scale == 0 && 2602 !(X86MCRegisterClasses[X86::VR128XRegClassID].contains(IndexReg) || 2603 X86MCRegisterClasses[X86::VR256XRegClassID].contains(IndexReg) || 2604 X86MCRegisterClasses[X86::VR512RegClassID].contains(IndexReg)) && 2605 (X86MCRegisterClasses[X86::VR128XRegClassID].contains(BaseReg) || 2606 X86MCRegisterClasses[X86::VR256XRegClassID].contains(BaseReg) || 2607 X86MCRegisterClasses[X86::VR512RegClassID].contains(BaseReg))) 2608 std::swap(BaseReg, IndexReg); 2609 2610 if (Scale != 0 && 2611 X86MCRegisterClasses[X86::GR16RegClassID].contains(IndexReg)) 2612 return Error(Start, "16-bit addresses cannot have a scale"); 2613 2614 // If there was no explicit scale specified, change it to 1. 2615 if (Scale == 0) 2616 Scale = 1; 2617 2618 // If this is a 16-bit addressing mode with the base and index in the wrong 2619 // order, swap them so CheckBaseRegAndIndexRegAndScale doesn't fail. It is 2620 // shared with att syntax where order matters. 2621 if ((BaseReg == X86::SI || BaseReg == X86::DI) && 2622 (IndexReg == X86::BX || IndexReg == X86::BP)) 2623 std::swap(BaseReg, IndexReg); 2624 2625 if ((BaseReg || IndexReg) && 2626 CheckBaseRegAndIndexRegAndScale(BaseReg, IndexReg, Scale, is64BitMode(), 2627 ErrMsg)) 2628 return Error(Start, ErrMsg); 2629 if (isParsingMSInlineAsm()) 2630 return CreateMemForMSInlineAsm(RegNo, Disp, BaseReg, IndexReg, Scale, Start, 2631 End, Size, SM.getSymName(), 2632 SM.getIdentifierInfo(), Operands); 2633 2634 // When parsing x64 MS-style assembly, all non-absolute references to a named 2635 // variable default to RIP-relative. 2636 unsigned DefaultBaseReg = X86::NoRegister; 2637 bool MaybeDirectBranchDest = true; 2638 2639 if (Parser.isParsingMasm()) { 2640 bool IsUnconditionalBranch = 2641 Name.equals_insensitive("jmp") || Name.equals_insensitive("call"); 2642 if (is64BitMode() && SM.getElementSize() > 0) { 2643 DefaultBaseReg = X86::RIP; 2644 } 2645 if (IsUnconditionalBranch) { 2646 if (PtrInOperand) { 2647 MaybeDirectBranchDest = false; 2648 if (is64BitMode()) 2649 DefaultBaseReg = X86::RIP; 2650 } else if (!BaseReg && !IndexReg && Disp && 2651 Disp->getKind() == MCExpr::SymbolRef) { 2652 if (is64BitMode()) { 2653 if (SM.getSize() == 8) { 2654 MaybeDirectBranchDest = false; 2655 DefaultBaseReg = X86::RIP; 2656 } 2657 } else { 2658 if (SM.getSize() == 4 || SM.getSize() == 2) 2659 MaybeDirectBranchDest = false; 2660 } 2661 } 2662 } 2663 } 2664 2665 if ((BaseReg || IndexReg || RegNo || DefaultBaseReg != X86::NoRegister)) 2666 Operands.push_back(X86Operand::CreateMem( 2667 getPointerWidth(), RegNo, Disp, BaseReg, IndexReg, Scale, Start, End, 2668 Size, DefaultBaseReg, /*SymName=*/StringRef(), /*OpDecl=*/nullptr, 2669 /*FrontendSize=*/0, /*UseUpRegs=*/false, MaybeDirectBranchDest)); 2670 else 2671 Operands.push_back(X86Operand::CreateMem( 2672 getPointerWidth(), Disp, Start, End, Size, /*SymName=*/StringRef(), 2673 /*OpDecl=*/nullptr, /*FrontendSize=*/0, /*UseUpRegs=*/false, 2674 MaybeDirectBranchDest)); 2675 return false; 2676 } 2677 2678 bool X86AsmParser::parseATTOperand(OperandVector &Operands) { 2679 MCAsmParser &Parser = getParser(); 2680 switch (getLexer().getKind()) { 2681 case AsmToken::Dollar: { 2682 // $42 or $ID -> immediate. 2683 SMLoc Start = Parser.getTok().getLoc(), End; 2684 Parser.Lex(); 2685 const MCExpr *Val; 2686 // This is an immediate, so we should not parse a register. Do a precheck 2687 // for '%' to supercede intra-register parse errors. 2688 SMLoc L = Parser.getTok().getLoc(); 2689 if (check(getLexer().is(AsmToken::Percent), L, 2690 "expected immediate expression") || 2691 getParser().parseExpression(Val, End) || 2692 check(isa<X86MCExpr>(Val), L, "expected immediate expression")) 2693 return true; 2694 Operands.push_back(X86Operand::CreateImm(Val, Start, End)); 2695 return false; 2696 } 2697 case AsmToken::LCurly: { 2698 SMLoc Start = Parser.getTok().getLoc(); 2699 return ParseRoundingModeOp(Start, Operands); 2700 } 2701 default: { 2702 // This a memory operand or a register. We have some parsing complications 2703 // as a '(' may be part of an immediate expression or the addressing mode 2704 // block. This is complicated by the fact that an assembler-level variable 2705 // may refer either to a register or an immediate expression. 2706 2707 SMLoc Loc = Parser.getTok().getLoc(), EndLoc; 2708 const MCExpr *Expr = nullptr; 2709 unsigned Reg = 0; 2710 if (getLexer().isNot(AsmToken::LParen)) { 2711 // No '(' so this is either a displacement expression or a register. 2712 if (Parser.parseExpression(Expr, EndLoc)) 2713 return true; 2714 if (auto *RE = dyn_cast<X86MCExpr>(Expr)) { 2715 // Segment Register. Reset Expr and copy value to register. 2716 Expr = nullptr; 2717 Reg = RE->getRegNo(); 2718 2719 // Check the register. 2720 if (Reg == X86::EIZ || Reg == X86::RIZ) 2721 return Error( 2722 Loc, "%eiz and %riz can only be used as index registers", 2723 SMRange(Loc, EndLoc)); 2724 if (Reg == X86::RIP) 2725 return Error(Loc, "%rip can only be used as a base register", 2726 SMRange(Loc, EndLoc)); 2727 // Return register that are not segment prefixes immediately. 2728 if (!Parser.parseOptionalToken(AsmToken::Colon)) { 2729 Operands.push_back(X86Operand::CreateReg(Reg, Loc, EndLoc)); 2730 return false; 2731 } 2732 if (!X86MCRegisterClasses[X86::SEGMENT_REGRegClassID].contains(Reg)) 2733 return Error(Loc, "invalid segment register"); 2734 // Accept a '*' absolute memory reference after the segment. Place it 2735 // before the full memory operand. 2736 if (getLexer().is(AsmToken::Star)) 2737 Operands.push_back(X86Operand::CreateToken("*", consumeToken())); 2738 } 2739 } 2740 // This is a Memory operand. 2741 return ParseMemOperand(Reg, Expr, Loc, EndLoc, Operands); 2742 } 2743 } 2744 } 2745 2746 // X86::COND_INVALID if not a recognized condition code or alternate mnemonic, 2747 // otherwise the EFLAGS Condition Code enumerator. 2748 X86::CondCode X86AsmParser::ParseConditionCode(StringRef CC) { 2749 return StringSwitch<X86::CondCode>(CC) 2750 .Case("o", X86::COND_O) // Overflow 2751 .Case("no", X86::COND_NO) // No Overflow 2752 .Cases("b", "nae", X86::COND_B) // Below/Neither Above nor Equal 2753 .Cases("ae", "nb", X86::COND_AE) // Above or Equal/Not Below 2754 .Cases("e", "z", X86::COND_E) // Equal/Zero 2755 .Cases("ne", "nz", X86::COND_NE) // Not Equal/Not Zero 2756 .Cases("be", "na", X86::COND_BE) // Below or Equal/Not Above 2757 .Cases("a", "nbe", X86::COND_A) // Above/Neither Below nor Equal 2758 .Case("s", X86::COND_S) // Sign 2759 .Case("ns", X86::COND_NS) // No Sign 2760 .Cases("p", "pe", X86::COND_P) // Parity/Parity Even 2761 .Cases("np", "po", X86::COND_NP) // No Parity/Parity Odd 2762 .Cases("l", "nge", X86::COND_L) // Less/Neither Greater nor Equal 2763 .Cases("ge", "nl", X86::COND_GE) // Greater or Equal/Not Less 2764 .Cases("le", "ng", X86::COND_LE) // Less or Equal/Not Greater 2765 .Cases("g", "nle", X86::COND_G) // Greater/Neither Less nor Equal 2766 .Default(X86::COND_INVALID); 2767 } 2768 2769 // true on failure, false otherwise 2770 // If no {z} mark was found - Parser doesn't advance 2771 bool X86AsmParser::ParseZ(std::unique_ptr<X86Operand> &Z, 2772 const SMLoc &StartLoc) { 2773 MCAsmParser &Parser = getParser(); 2774 // Assuming we are just pass the '{' mark, quering the next token 2775 // Searched for {z}, but none was found. Return false, as no parsing error was 2776 // encountered 2777 if (!(getLexer().is(AsmToken::Identifier) && 2778 (getLexer().getTok().getIdentifier() == "z"))) 2779 return false; 2780 Parser.Lex(); // Eat z 2781 // Query and eat the '}' mark 2782 if (!getLexer().is(AsmToken::RCurly)) 2783 return Error(getLexer().getLoc(), "Expected } at this point"); 2784 Parser.Lex(); // Eat '}' 2785 // Assign Z with the {z} mark operand 2786 Z = X86Operand::CreateToken("{z}", StartLoc); 2787 return false; 2788 } 2789 2790 // true on failure, false otherwise 2791 bool X86AsmParser::HandleAVX512Operand(OperandVector &Operands) { 2792 MCAsmParser &Parser = getParser(); 2793 if (getLexer().is(AsmToken::LCurly)) { 2794 // Eat "{" and mark the current place. 2795 const SMLoc consumedToken = consumeToken(); 2796 // Distinguish {1to<NUM>} from {%k<NUM>}. 2797 if(getLexer().is(AsmToken::Integer)) { 2798 // Parse memory broadcasting ({1to<NUM>}). 2799 if (getLexer().getTok().getIntVal() != 1) 2800 return TokError("Expected 1to<NUM> at this point"); 2801 StringRef Prefix = getLexer().getTok().getString(); 2802 Parser.Lex(); // Eat first token of 1to8 2803 if (!getLexer().is(AsmToken::Identifier)) 2804 return TokError("Expected 1to<NUM> at this point"); 2805 // Recognize only reasonable suffixes. 2806 SmallVector<char, 5> BroadcastVector; 2807 StringRef BroadcastString = (Prefix + getLexer().getTok().getIdentifier()) 2808 .toStringRef(BroadcastVector); 2809 if (!BroadcastString.startswith("1to")) 2810 return TokError("Expected 1to<NUM> at this point"); 2811 const char *BroadcastPrimitive = 2812 StringSwitch<const char *>(BroadcastString) 2813 .Case("1to2", "{1to2}") 2814 .Case("1to4", "{1to4}") 2815 .Case("1to8", "{1to8}") 2816 .Case("1to16", "{1to16}") 2817 .Case("1to32", "{1to32}") 2818 .Default(nullptr); 2819 if (!BroadcastPrimitive) 2820 return TokError("Invalid memory broadcast primitive."); 2821 Parser.Lex(); // Eat trailing token of 1toN 2822 if (!getLexer().is(AsmToken::RCurly)) 2823 return TokError("Expected } at this point"); 2824 Parser.Lex(); // Eat "}" 2825 Operands.push_back(X86Operand::CreateToken(BroadcastPrimitive, 2826 consumedToken)); 2827 // No AVX512 specific primitives can pass 2828 // after memory broadcasting, so return. 2829 return false; 2830 } else { 2831 // Parse either {k}{z}, {z}{k}, {k} or {z} 2832 // last one have no meaning, but GCC accepts it 2833 // Currently, we're just pass a '{' mark 2834 std::unique_ptr<X86Operand> Z; 2835 if (ParseZ(Z, consumedToken)) 2836 return true; 2837 // Reaching here means that parsing of the allegadly '{z}' mark yielded 2838 // no errors. 2839 // Query for the need of further parsing for a {%k<NUM>} mark 2840 if (!Z || getLexer().is(AsmToken::LCurly)) { 2841 SMLoc StartLoc = Z ? consumeToken() : consumedToken; 2842 // Parse an op-mask register mark ({%k<NUM>}), which is now to be 2843 // expected 2844 MCRegister RegNo; 2845 SMLoc RegLoc; 2846 if (!parseRegister(RegNo, RegLoc, StartLoc) && 2847 X86MCRegisterClasses[X86::VK1RegClassID].contains(RegNo)) { 2848 if (RegNo == X86::K0) 2849 return Error(RegLoc, "Register k0 can't be used as write mask"); 2850 if (!getLexer().is(AsmToken::RCurly)) 2851 return Error(getLexer().getLoc(), "Expected } at this point"); 2852 Operands.push_back(X86Operand::CreateToken("{", StartLoc)); 2853 Operands.push_back( 2854 X86Operand::CreateReg(RegNo, StartLoc, StartLoc)); 2855 Operands.push_back(X86Operand::CreateToken("}", consumeToken())); 2856 } else 2857 return Error(getLexer().getLoc(), 2858 "Expected an op-mask register at this point"); 2859 // {%k<NUM>} mark is found, inquire for {z} 2860 if (getLexer().is(AsmToken::LCurly) && !Z) { 2861 // Have we've found a parsing error, or found no (expected) {z} mark 2862 // - report an error 2863 if (ParseZ(Z, consumeToken()) || !Z) 2864 return Error(getLexer().getLoc(), 2865 "Expected a {z} mark at this point"); 2866 2867 } 2868 // '{z}' on its own is meaningless, hence should be ignored. 2869 // on the contrary - have it been accompanied by a K register, 2870 // allow it. 2871 if (Z) 2872 Operands.push_back(std::move(Z)); 2873 } 2874 } 2875 } 2876 return false; 2877 } 2878 2879 /// ParseMemOperand: 'seg : disp(basereg, indexreg, scale)'. The '%ds:' prefix 2880 /// has already been parsed if present. disp may be provided as well. 2881 bool X86AsmParser::ParseMemOperand(unsigned SegReg, const MCExpr *Disp, 2882 SMLoc StartLoc, SMLoc EndLoc, 2883 OperandVector &Operands) { 2884 MCAsmParser &Parser = getParser(); 2885 SMLoc Loc; 2886 // Based on the initial passed values, we may be in any of these cases, we are 2887 // in one of these cases (with current position (*)): 2888 2889 // 1. seg : * disp (base-index-scale-expr) 2890 // 2. seg : *(disp) (base-index-scale-expr) 2891 // 3. seg : *(base-index-scale-expr) 2892 // 4. disp *(base-index-scale-expr) 2893 // 5. *(disp) (base-index-scale-expr) 2894 // 6. *(base-index-scale-expr) 2895 // 7. disp * 2896 // 8. *(disp) 2897 2898 // If we do not have an displacement yet, check if we're in cases 4 or 6 by 2899 // checking if the first object after the parenthesis is a register (or an 2900 // identifier referring to a register) and parse the displacement or default 2901 // to 0 as appropriate. 2902 auto isAtMemOperand = [this]() { 2903 if (this->getLexer().isNot(AsmToken::LParen)) 2904 return false; 2905 AsmToken Buf[2]; 2906 StringRef Id; 2907 auto TokCount = this->getLexer().peekTokens(Buf, true); 2908 if (TokCount == 0) 2909 return false; 2910 switch (Buf[0].getKind()) { 2911 case AsmToken::Percent: 2912 case AsmToken::Comma: 2913 return true; 2914 // These lower cases are doing a peekIdentifier. 2915 case AsmToken::At: 2916 case AsmToken::Dollar: 2917 if ((TokCount > 1) && 2918 (Buf[1].is(AsmToken::Identifier) || Buf[1].is(AsmToken::String)) && 2919 (Buf[0].getLoc().getPointer() + 1 == Buf[1].getLoc().getPointer())) 2920 Id = StringRef(Buf[0].getLoc().getPointer(), 2921 Buf[1].getIdentifier().size() + 1); 2922 break; 2923 case AsmToken::Identifier: 2924 case AsmToken::String: 2925 Id = Buf[0].getIdentifier(); 2926 break; 2927 default: 2928 return false; 2929 } 2930 // We have an ID. Check if it is bound to a register. 2931 if (!Id.empty()) { 2932 MCSymbol *Sym = this->getContext().getOrCreateSymbol(Id); 2933 if (Sym->isVariable()) { 2934 auto V = Sym->getVariableValue(/*SetUsed*/ false); 2935 return isa<X86MCExpr>(V); 2936 } 2937 } 2938 return false; 2939 }; 2940 2941 if (!Disp) { 2942 // Parse immediate if we're not at a mem operand yet. 2943 if (!isAtMemOperand()) { 2944 if (Parser.parseTokenLoc(Loc) || Parser.parseExpression(Disp, EndLoc)) 2945 return true; 2946 assert(!isa<X86MCExpr>(Disp) && "Expected non-register here."); 2947 } else { 2948 // Disp is implicitly zero if we haven't parsed it yet. 2949 Disp = MCConstantExpr::create(0, Parser.getContext()); 2950 } 2951 } 2952 2953 // We are now either at the end of the operand or at the '(' at the start of a 2954 // base-index-scale-expr. 2955 2956 if (!parseOptionalToken(AsmToken::LParen)) { 2957 if (SegReg == 0) 2958 Operands.push_back( 2959 X86Operand::CreateMem(getPointerWidth(), Disp, StartLoc, EndLoc)); 2960 else 2961 Operands.push_back(X86Operand::CreateMem(getPointerWidth(), SegReg, Disp, 2962 0, 0, 1, StartLoc, EndLoc)); 2963 return false; 2964 } 2965 2966 // If we reached here, then eat the '(' and Process 2967 // the rest of the memory operand. 2968 unsigned BaseReg = 0, IndexReg = 0, Scale = 1; 2969 SMLoc BaseLoc = getLexer().getLoc(); 2970 const MCExpr *E; 2971 StringRef ErrMsg; 2972 2973 // Parse BaseReg if one is provided. 2974 if (getLexer().isNot(AsmToken::Comma) && getLexer().isNot(AsmToken::RParen)) { 2975 if (Parser.parseExpression(E, EndLoc) || 2976 check(!isa<X86MCExpr>(E), BaseLoc, "expected register here")) 2977 return true; 2978 2979 // Check the register. 2980 BaseReg = cast<X86MCExpr>(E)->getRegNo(); 2981 if (BaseReg == X86::EIZ || BaseReg == X86::RIZ) 2982 return Error(BaseLoc, "eiz and riz can only be used as index registers", 2983 SMRange(BaseLoc, EndLoc)); 2984 } 2985 2986 if (parseOptionalToken(AsmToken::Comma)) { 2987 // Following the comma we should have either an index register, or a scale 2988 // value. We don't support the later form, but we want to parse it 2989 // correctly. 2990 // 2991 // Even though it would be completely consistent to support syntax like 2992 // "1(%eax,,1)", the assembler doesn't. Use "eiz" or "riz" for this. 2993 if (getLexer().isNot(AsmToken::RParen)) { 2994 if (Parser.parseTokenLoc(Loc) || Parser.parseExpression(E, EndLoc)) 2995 return true; 2996 2997 if (!isa<X86MCExpr>(E)) { 2998 // We've parsed an unexpected Scale Value instead of an index 2999 // register. Interpret it as an absolute. 3000 int64_t ScaleVal; 3001 if (!E->evaluateAsAbsolute(ScaleVal, getStreamer().getAssemblerPtr())) 3002 return Error(Loc, "expected absolute expression"); 3003 if (ScaleVal != 1) 3004 Warning(Loc, "scale factor without index register is ignored"); 3005 Scale = 1; 3006 } else { // IndexReg Found. 3007 IndexReg = cast<X86MCExpr>(E)->getRegNo(); 3008 3009 if (BaseReg == X86::RIP) 3010 return Error(Loc, 3011 "%rip as base register can not have an index register"); 3012 if (IndexReg == X86::RIP) 3013 return Error(Loc, "%rip is not allowed as an index register"); 3014 3015 if (parseOptionalToken(AsmToken::Comma)) { 3016 // Parse the scale amount: 3017 // ::= ',' [scale-expression] 3018 3019 // A scale amount without an index is ignored. 3020 if (getLexer().isNot(AsmToken::RParen)) { 3021 int64_t ScaleVal; 3022 if (Parser.parseTokenLoc(Loc) || 3023 Parser.parseAbsoluteExpression(ScaleVal)) 3024 return Error(Loc, "expected scale expression"); 3025 Scale = (unsigned)ScaleVal; 3026 // Validate the scale amount. 3027 if (X86MCRegisterClasses[X86::GR16RegClassID].contains(BaseReg) && 3028 Scale != 1) 3029 return Error(Loc, "scale factor in 16-bit address must be 1"); 3030 if (checkScale(Scale, ErrMsg)) 3031 return Error(Loc, ErrMsg); 3032 } 3033 } 3034 } 3035 } 3036 } 3037 3038 // Ok, we've eaten the memory operand, verify we have a ')' and eat it too. 3039 if (parseToken(AsmToken::RParen, "unexpected token in memory operand")) 3040 return true; 3041 3042 // This is to support otherwise illegal operand (%dx) found in various 3043 // unofficial manuals examples (e.g. "out[s]?[bwl]? %al, (%dx)") and must now 3044 // be supported. Mark such DX variants separately fix only in special cases. 3045 if (BaseReg == X86::DX && IndexReg == 0 && Scale == 1 && SegReg == 0 && 3046 isa<MCConstantExpr>(Disp) && 3047 cast<MCConstantExpr>(Disp)->getValue() == 0) { 3048 Operands.push_back(X86Operand::CreateDXReg(BaseLoc, BaseLoc)); 3049 return false; 3050 } 3051 3052 if (CheckBaseRegAndIndexRegAndScale(BaseReg, IndexReg, Scale, is64BitMode(), 3053 ErrMsg)) 3054 return Error(BaseLoc, ErrMsg); 3055 3056 if (SegReg || BaseReg || IndexReg) 3057 Operands.push_back(X86Operand::CreateMem(getPointerWidth(), SegReg, Disp, 3058 BaseReg, IndexReg, Scale, StartLoc, 3059 EndLoc)); 3060 else 3061 Operands.push_back( 3062 X86Operand::CreateMem(getPointerWidth(), Disp, StartLoc, EndLoc)); 3063 return false; 3064 } 3065 3066 // Parse either a standard primary expression or a register. 3067 bool X86AsmParser::parsePrimaryExpr(const MCExpr *&Res, SMLoc &EndLoc) { 3068 MCAsmParser &Parser = getParser(); 3069 // See if this is a register first. 3070 if (getTok().is(AsmToken::Percent) || 3071 (isParsingIntelSyntax() && getTok().is(AsmToken::Identifier) && 3072 MatchRegisterName(Parser.getTok().getString()))) { 3073 SMLoc StartLoc = Parser.getTok().getLoc(); 3074 MCRegister RegNo; 3075 if (parseRegister(RegNo, StartLoc, EndLoc)) 3076 return true; 3077 Res = X86MCExpr::create(RegNo, Parser.getContext()); 3078 return false; 3079 } 3080 return Parser.parsePrimaryExpr(Res, EndLoc, nullptr); 3081 } 3082 3083 bool X86AsmParser::ParseInstruction(ParseInstructionInfo &Info, StringRef Name, 3084 SMLoc NameLoc, OperandVector &Operands) { 3085 MCAsmParser &Parser = getParser(); 3086 InstInfo = &Info; 3087 3088 // Reset the forced VEX encoding. 3089 ForcedVEXEncoding = VEXEncoding_Default; 3090 ForcedDispEncoding = DispEncoding_Default; 3091 3092 // Parse pseudo prefixes. 3093 while (true) { 3094 if (Name == "{") { 3095 if (getLexer().isNot(AsmToken::Identifier)) 3096 return Error(Parser.getTok().getLoc(), "Unexpected token after '{'"); 3097 std::string Prefix = Parser.getTok().getString().lower(); 3098 Parser.Lex(); // Eat identifier. 3099 if (getLexer().isNot(AsmToken::RCurly)) 3100 return Error(Parser.getTok().getLoc(), "Expected '}'"); 3101 Parser.Lex(); // Eat curly. 3102 3103 if (Prefix == "vex") 3104 ForcedVEXEncoding = VEXEncoding_VEX; 3105 else if (Prefix == "vex2") 3106 ForcedVEXEncoding = VEXEncoding_VEX2; 3107 else if (Prefix == "vex3") 3108 ForcedVEXEncoding = VEXEncoding_VEX3; 3109 else if (Prefix == "evex") 3110 ForcedVEXEncoding = VEXEncoding_EVEX; 3111 else if (Prefix == "disp8") 3112 ForcedDispEncoding = DispEncoding_Disp8; 3113 else if (Prefix == "disp32") 3114 ForcedDispEncoding = DispEncoding_Disp32; 3115 else 3116 return Error(NameLoc, "unknown prefix"); 3117 3118 NameLoc = Parser.getTok().getLoc(); 3119 if (getLexer().is(AsmToken::LCurly)) { 3120 Parser.Lex(); 3121 Name = "{"; 3122 } else { 3123 if (getLexer().isNot(AsmToken::Identifier)) 3124 return Error(Parser.getTok().getLoc(), "Expected identifier"); 3125 // FIXME: The mnemonic won't match correctly if its not in lower case. 3126 Name = Parser.getTok().getString(); 3127 Parser.Lex(); 3128 } 3129 continue; 3130 } 3131 // Parse MASM style pseudo prefixes. 3132 if (isParsingMSInlineAsm()) { 3133 if (Name.equals_insensitive("vex")) 3134 ForcedVEXEncoding = VEXEncoding_VEX; 3135 else if (Name.equals_insensitive("vex2")) 3136 ForcedVEXEncoding = VEXEncoding_VEX2; 3137 else if (Name.equals_insensitive("vex3")) 3138 ForcedVEXEncoding = VEXEncoding_VEX3; 3139 else if (Name.equals_insensitive("evex")) 3140 ForcedVEXEncoding = VEXEncoding_EVEX; 3141 3142 if (ForcedVEXEncoding != VEXEncoding_Default) { 3143 if (getLexer().isNot(AsmToken::Identifier)) 3144 return Error(Parser.getTok().getLoc(), "Expected identifier"); 3145 // FIXME: The mnemonic won't match correctly if its not in lower case. 3146 Name = Parser.getTok().getString(); 3147 NameLoc = Parser.getTok().getLoc(); 3148 Parser.Lex(); 3149 } 3150 } 3151 break; 3152 } 3153 3154 // Support the suffix syntax for overriding displacement size as well. 3155 if (Name.consume_back(".d32")) { 3156 ForcedDispEncoding = DispEncoding_Disp32; 3157 } else if (Name.consume_back(".d8")) { 3158 ForcedDispEncoding = DispEncoding_Disp8; 3159 } 3160 3161 StringRef PatchedName = Name; 3162 3163 // Hack to skip "short" following Jcc. 3164 if (isParsingIntelSyntax() && 3165 (PatchedName == "jmp" || PatchedName == "jc" || PatchedName == "jnc" || 3166 PatchedName == "jcxz" || PatchedName == "jecxz" || 3167 (PatchedName.startswith("j") && 3168 ParseConditionCode(PatchedName.substr(1)) != X86::COND_INVALID))) { 3169 StringRef NextTok = Parser.getTok().getString(); 3170 if (Parser.isParsingMasm() ? NextTok.equals_insensitive("short") 3171 : NextTok == "short") { 3172 SMLoc NameEndLoc = 3173 NameLoc.getFromPointer(NameLoc.getPointer() + Name.size()); 3174 // Eat the short keyword. 3175 Parser.Lex(); 3176 // MS and GAS ignore the short keyword; they both determine the jmp type 3177 // based on the distance of the label. (NASM does emit different code with 3178 // and without "short," though.) 3179 InstInfo->AsmRewrites->emplace_back(AOK_Skip, NameEndLoc, 3180 NextTok.size() + 1); 3181 } 3182 } 3183 3184 // FIXME: Hack to recognize setneb as setne. 3185 if (PatchedName.startswith("set") && PatchedName.endswith("b") && 3186 PatchedName != "setb" && PatchedName != "setnb") 3187 PatchedName = PatchedName.substr(0, Name.size()-1); 3188 3189 unsigned ComparisonPredicate = ~0U; 3190 3191 // FIXME: Hack to recognize cmp<comparison code>{sh,ss,sd,ph,ps,pd}. 3192 if ((PatchedName.startswith("cmp") || PatchedName.startswith("vcmp")) && 3193 (PatchedName.endswith("ss") || PatchedName.endswith("sd") || 3194 PatchedName.endswith("sh") || PatchedName.endswith("ph") || 3195 PatchedName.endswith("ps") || PatchedName.endswith("pd"))) { 3196 bool IsVCMP = PatchedName[0] == 'v'; 3197 unsigned CCIdx = IsVCMP ? 4 : 3; 3198 unsigned CC = StringSwitch<unsigned>( 3199 PatchedName.slice(CCIdx, PatchedName.size() - 2)) 3200 .Case("eq", 0x00) 3201 .Case("eq_oq", 0x00) 3202 .Case("lt", 0x01) 3203 .Case("lt_os", 0x01) 3204 .Case("le", 0x02) 3205 .Case("le_os", 0x02) 3206 .Case("unord", 0x03) 3207 .Case("unord_q", 0x03) 3208 .Case("neq", 0x04) 3209 .Case("neq_uq", 0x04) 3210 .Case("nlt", 0x05) 3211 .Case("nlt_us", 0x05) 3212 .Case("nle", 0x06) 3213 .Case("nle_us", 0x06) 3214 .Case("ord", 0x07) 3215 .Case("ord_q", 0x07) 3216 /* AVX only from here */ 3217 .Case("eq_uq", 0x08) 3218 .Case("nge", 0x09) 3219 .Case("nge_us", 0x09) 3220 .Case("ngt", 0x0A) 3221 .Case("ngt_us", 0x0A) 3222 .Case("false", 0x0B) 3223 .Case("false_oq", 0x0B) 3224 .Case("neq_oq", 0x0C) 3225 .Case("ge", 0x0D) 3226 .Case("ge_os", 0x0D) 3227 .Case("gt", 0x0E) 3228 .Case("gt_os", 0x0E) 3229 .Case("true", 0x0F) 3230 .Case("true_uq", 0x0F) 3231 .Case("eq_os", 0x10) 3232 .Case("lt_oq", 0x11) 3233 .Case("le_oq", 0x12) 3234 .Case("unord_s", 0x13) 3235 .Case("neq_us", 0x14) 3236 .Case("nlt_uq", 0x15) 3237 .Case("nle_uq", 0x16) 3238 .Case("ord_s", 0x17) 3239 .Case("eq_us", 0x18) 3240 .Case("nge_uq", 0x19) 3241 .Case("ngt_uq", 0x1A) 3242 .Case("false_os", 0x1B) 3243 .Case("neq_os", 0x1C) 3244 .Case("ge_oq", 0x1D) 3245 .Case("gt_oq", 0x1E) 3246 .Case("true_us", 0x1F) 3247 .Default(~0U); 3248 if (CC != ~0U && (IsVCMP || CC < 8) && 3249 (IsVCMP || PatchedName.back() != 'h')) { 3250 if (PatchedName.endswith("ss")) 3251 PatchedName = IsVCMP ? "vcmpss" : "cmpss"; 3252 else if (PatchedName.endswith("sd")) 3253 PatchedName = IsVCMP ? "vcmpsd" : "cmpsd"; 3254 else if (PatchedName.endswith("ps")) 3255 PatchedName = IsVCMP ? "vcmpps" : "cmpps"; 3256 else if (PatchedName.endswith("pd")) 3257 PatchedName = IsVCMP ? "vcmppd" : "cmppd"; 3258 else if (PatchedName.endswith("sh")) 3259 PatchedName = "vcmpsh"; 3260 else if (PatchedName.endswith("ph")) 3261 PatchedName = "vcmpph"; 3262 else 3263 llvm_unreachable("Unexpected suffix!"); 3264 3265 ComparisonPredicate = CC; 3266 } 3267 } 3268 3269 // FIXME: Hack to recognize vpcmp<comparison code>{ub,uw,ud,uq,b,w,d,q}. 3270 if (PatchedName.startswith("vpcmp") && 3271 (PatchedName.back() == 'b' || PatchedName.back() == 'w' || 3272 PatchedName.back() == 'd' || PatchedName.back() == 'q')) { 3273 unsigned SuffixSize = PatchedName.drop_back().back() == 'u' ? 2 : 1; 3274 unsigned CC = StringSwitch<unsigned>( 3275 PatchedName.slice(5, PatchedName.size() - SuffixSize)) 3276 .Case("eq", 0x0) // Only allowed on unsigned. Checked below. 3277 .Case("lt", 0x1) 3278 .Case("le", 0x2) 3279 //.Case("false", 0x3) // Not a documented alias. 3280 .Case("neq", 0x4) 3281 .Case("nlt", 0x5) 3282 .Case("nle", 0x6) 3283 //.Case("true", 0x7) // Not a documented alias. 3284 .Default(~0U); 3285 if (CC != ~0U && (CC != 0 || SuffixSize == 2)) { 3286 switch (PatchedName.back()) { 3287 default: llvm_unreachable("Unexpected character!"); 3288 case 'b': PatchedName = SuffixSize == 2 ? "vpcmpub" : "vpcmpb"; break; 3289 case 'w': PatchedName = SuffixSize == 2 ? "vpcmpuw" : "vpcmpw"; break; 3290 case 'd': PatchedName = SuffixSize == 2 ? "vpcmpud" : "vpcmpd"; break; 3291 case 'q': PatchedName = SuffixSize == 2 ? "vpcmpuq" : "vpcmpq"; break; 3292 } 3293 // Set up the immediate to push into the operands later. 3294 ComparisonPredicate = CC; 3295 } 3296 } 3297 3298 // FIXME: Hack to recognize vpcom<comparison code>{ub,uw,ud,uq,b,w,d,q}. 3299 if (PatchedName.startswith("vpcom") && 3300 (PatchedName.back() == 'b' || PatchedName.back() == 'w' || 3301 PatchedName.back() == 'd' || PatchedName.back() == 'q')) { 3302 unsigned SuffixSize = PatchedName.drop_back().back() == 'u' ? 2 : 1; 3303 unsigned CC = StringSwitch<unsigned>( 3304 PatchedName.slice(5, PatchedName.size() - SuffixSize)) 3305 .Case("lt", 0x0) 3306 .Case("le", 0x1) 3307 .Case("gt", 0x2) 3308 .Case("ge", 0x3) 3309 .Case("eq", 0x4) 3310 .Case("neq", 0x5) 3311 .Case("false", 0x6) 3312 .Case("true", 0x7) 3313 .Default(~0U); 3314 if (CC != ~0U) { 3315 switch (PatchedName.back()) { 3316 default: llvm_unreachable("Unexpected character!"); 3317 case 'b': PatchedName = SuffixSize == 2 ? "vpcomub" : "vpcomb"; break; 3318 case 'w': PatchedName = SuffixSize == 2 ? "vpcomuw" : "vpcomw"; break; 3319 case 'd': PatchedName = SuffixSize == 2 ? "vpcomud" : "vpcomd"; break; 3320 case 'q': PatchedName = SuffixSize == 2 ? "vpcomuq" : "vpcomq"; break; 3321 } 3322 // Set up the immediate to push into the operands later. 3323 ComparisonPredicate = CC; 3324 } 3325 } 3326 3327 3328 // Determine whether this is an instruction prefix. 3329 // FIXME: 3330 // Enhance prefixes integrity robustness. for example, following forms 3331 // are currently tolerated: 3332 // repz repnz <insn> ; GAS errors for the use of two similar prefixes 3333 // lock addq %rax, %rbx ; Destination operand must be of memory type 3334 // xacquire <insn> ; xacquire must be accompanied by 'lock' 3335 bool IsPrefix = 3336 StringSwitch<bool>(Name) 3337 .Cases("cs", "ds", "es", "fs", "gs", "ss", true) 3338 .Cases("rex64", "data32", "data16", "addr32", "addr16", true) 3339 .Cases("xacquire", "xrelease", true) 3340 .Cases("acquire", "release", isParsingIntelSyntax()) 3341 .Default(false); 3342 3343 auto isLockRepeatNtPrefix = [](StringRef N) { 3344 return StringSwitch<bool>(N) 3345 .Cases("lock", "rep", "repe", "repz", "repne", "repnz", "notrack", true) 3346 .Default(false); 3347 }; 3348 3349 bool CurlyAsEndOfStatement = false; 3350 3351 unsigned Flags = X86::IP_NO_PREFIX; 3352 while (isLockRepeatNtPrefix(Name.lower())) { 3353 unsigned Prefix = 3354 StringSwitch<unsigned>(Name) 3355 .Cases("lock", "lock", X86::IP_HAS_LOCK) 3356 .Cases("rep", "repe", "repz", X86::IP_HAS_REPEAT) 3357 .Cases("repne", "repnz", X86::IP_HAS_REPEAT_NE) 3358 .Cases("notrack", "notrack", X86::IP_HAS_NOTRACK) 3359 .Default(X86::IP_NO_PREFIX); // Invalid prefix (impossible) 3360 Flags |= Prefix; 3361 if (getLexer().is(AsmToken::EndOfStatement)) { 3362 // We don't have real instr with the given prefix 3363 // let's use the prefix as the instr. 3364 // TODO: there could be several prefixes one after another 3365 Flags = X86::IP_NO_PREFIX; 3366 break; 3367 } 3368 // FIXME: The mnemonic won't match correctly if its not in lower case. 3369 Name = Parser.getTok().getString(); 3370 Parser.Lex(); // eat the prefix 3371 // Hack: we could have something like "rep # some comment" or 3372 // "lock; cmpxchg16b $1" or "lock\0A\09incl" or "lock/incl" 3373 while (Name.startswith(";") || Name.startswith("\n") || 3374 Name.startswith("#") || Name.startswith("\t") || 3375 Name.startswith("/")) { 3376 // FIXME: The mnemonic won't match correctly if its not in lower case. 3377 Name = Parser.getTok().getString(); 3378 Parser.Lex(); // go to next prefix or instr 3379 } 3380 } 3381 3382 if (Flags) 3383 PatchedName = Name; 3384 3385 // Hacks to handle 'data16' and 'data32' 3386 if (PatchedName == "data16" && is16BitMode()) { 3387 return Error(NameLoc, "redundant data16 prefix"); 3388 } 3389 if (PatchedName == "data32") { 3390 if (is32BitMode()) 3391 return Error(NameLoc, "redundant data32 prefix"); 3392 if (is64BitMode()) 3393 return Error(NameLoc, "'data32' is not supported in 64-bit mode"); 3394 // Hack to 'data16' for the table lookup. 3395 PatchedName = "data16"; 3396 3397 if (getLexer().isNot(AsmToken::EndOfStatement)) { 3398 StringRef Next = Parser.getTok().getString(); 3399 getLexer().Lex(); 3400 // data32 effectively changes the instruction suffix. 3401 // TODO Generalize. 3402 if (Next == "callw") 3403 Next = "calll"; 3404 if (Next == "ljmpw") 3405 Next = "ljmpl"; 3406 3407 Name = Next; 3408 PatchedName = Name; 3409 ForcedDataPrefix = X86::Is32Bit; 3410 IsPrefix = false; 3411 } 3412 } 3413 3414 Operands.push_back(X86Operand::CreateToken(PatchedName, NameLoc)); 3415 3416 // Push the immediate if we extracted one from the mnemonic. 3417 if (ComparisonPredicate != ~0U && !isParsingIntelSyntax()) { 3418 const MCExpr *ImmOp = MCConstantExpr::create(ComparisonPredicate, 3419 getParser().getContext()); 3420 Operands.push_back(X86Operand::CreateImm(ImmOp, NameLoc, NameLoc)); 3421 } 3422 3423 // This does the actual operand parsing. Don't parse any more if we have a 3424 // prefix juxtaposed with an operation like "lock incl 4(%rax)", because we 3425 // just want to parse the "lock" as the first instruction and the "incl" as 3426 // the next one. 3427 if (getLexer().isNot(AsmToken::EndOfStatement) && !IsPrefix) { 3428 // Parse '*' modifier. 3429 if (getLexer().is(AsmToken::Star)) 3430 Operands.push_back(X86Operand::CreateToken("*", consumeToken())); 3431 3432 // Read the operands. 3433 while (true) { 3434 if (parseOperand(Operands, Name)) 3435 return true; 3436 if (HandleAVX512Operand(Operands)) 3437 return true; 3438 3439 // check for comma and eat it 3440 if (getLexer().is(AsmToken::Comma)) 3441 Parser.Lex(); 3442 else 3443 break; 3444 } 3445 3446 // In MS inline asm curly braces mark the beginning/end of a block, 3447 // therefore they should be interepreted as end of statement 3448 CurlyAsEndOfStatement = 3449 isParsingIntelSyntax() && isParsingMSInlineAsm() && 3450 (getLexer().is(AsmToken::LCurly) || getLexer().is(AsmToken::RCurly)); 3451 if (getLexer().isNot(AsmToken::EndOfStatement) && !CurlyAsEndOfStatement) 3452 return TokError("unexpected token in argument list"); 3453 } 3454 3455 // Push the immediate if we extracted one from the mnemonic. 3456 if (ComparisonPredicate != ~0U && isParsingIntelSyntax()) { 3457 const MCExpr *ImmOp = MCConstantExpr::create(ComparisonPredicate, 3458 getParser().getContext()); 3459 Operands.push_back(X86Operand::CreateImm(ImmOp, NameLoc, NameLoc)); 3460 } 3461 3462 // Consume the EndOfStatement or the prefix separator Slash 3463 if (getLexer().is(AsmToken::EndOfStatement) || 3464 (IsPrefix && getLexer().is(AsmToken::Slash))) 3465 Parser.Lex(); 3466 else if (CurlyAsEndOfStatement) 3467 // Add an actual EndOfStatement before the curly brace 3468 Info.AsmRewrites->emplace_back(AOK_EndOfStatement, 3469 getLexer().getTok().getLoc(), 0); 3470 3471 // This is for gas compatibility and cannot be done in td. 3472 // Adding "p" for some floating point with no argument. 3473 // For example: fsub --> fsubp 3474 bool IsFp = 3475 Name == "fsub" || Name == "fdiv" || Name == "fsubr" || Name == "fdivr"; 3476 if (IsFp && Operands.size() == 1) { 3477 const char *Repl = StringSwitch<const char *>(Name) 3478 .Case("fsub", "fsubp") 3479 .Case("fdiv", "fdivp") 3480 .Case("fsubr", "fsubrp") 3481 .Case("fdivr", "fdivrp"); 3482 static_cast<X86Operand &>(*Operands[0]).setTokenValue(Repl); 3483 } 3484 3485 if ((Name == "mov" || Name == "movw" || Name == "movl") && 3486 (Operands.size() == 3)) { 3487 X86Operand &Op1 = (X86Operand &)*Operands[1]; 3488 X86Operand &Op2 = (X86Operand &)*Operands[2]; 3489 SMLoc Loc = Op1.getEndLoc(); 3490 // Moving a 32 or 16 bit value into a segment register has the same 3491 // behavior. Modify such instructions to always take shorter form. 3492 if (Op1.isReg() && Op2.isReg() && 3493 X86MCRegisterClasses[X86::SEGMENT_REGRegClassID].contains( 3494 Op2.getReg()) && 3495 (X86MCRegisterClasses[X86::GR16RegClassID].contains(Op1.getReg()) || 3496 X86MCRegisterClasses[X86::GR32RegClassID].contains(Op1.getReg()))) { 3497 // Change instruction name to match new instruction. 3498 if (Name != "mov" && Name[3] == (is16BitMode() ? 'l' : 'w')) { 3499 Name = is16BitMode() ? "movw" : "movl"; 3500 Operands[0] = X86Operand::CreateToken(Name, NameLoc); 3501 } 3502 // Select the correct equivalent 16-/32-bit source register. 3503 unsigned Reg = 3504 getX86SubSuperRegisterOrZero(Op1.getReg(), is16BitMode() ? 16 : 32); 3505 Operands[1] = X86Operand::CreateReg(Reg, Loc, Loc); 3506 } 3507 } 3508 3509 // This is a terrible hack to handle "out[s]?[bwl]? %al, (%dx)" -> 3510 // "outb %al, %dx". Out doesn't take a memory form, but this is a widely 3511 // documented form in various unofficial manuals, so a lot of code uses it. 3512 if ((Name == "outb" || Name == "outsb" || Name == "outw" || Name == "outsw" || 3513 Name == "outl" || Name == "outsl" || Name == "out" || Name == "outs") && 3514 Operands.size() == 3) { 3515 X86Operand &Op = (X86Operand &)*Operands.back(); 3516 if (Op.isDXReg()) 3517 Operands.back() = X86Operand::CreateReg(X86::DX, Op.getStartLoc(), 3518 Op.getEndLoc()); 3519 } 3520 // Same hack for "in[s]?[bwl]? (%dx), %al" -> "inb %dx, %al". 3521 if ((Name == "inb" || Name == "insb" || Name == "inw" || Name == "insw" || 3522 Name == "inl" || Name == "insl" || Name == "in" || Name == "ins") && 3523 Operands.size() == 3) { 3524 X86Operand &Op = (X86Operand &)*Operands[1]; 3525 if (Op.isDXReg()) 3526 Operands[1] = X86Operand::CreateReg(X86::DX, Op.getStartLoc(), 3527 Op.getEndLoc()); 3528 } 3529 3530 SmallVector<std::unique_ptr<MCParsedAsmOperand>, 2> TmpOperands; 3531 bool HadVerifyError = false; 3532 3533 // Append default arguments to "ins[bwld]" 3534 if (Name.startswith("ins") && 3535 (Operands.size() == 1 || Operands.size() == 3) && 3536 (Name == "insb" || Name == "insw" || Name == "insl" || Name == "insd" || 3537 Name == "ins")) { 3538 3539 AddDefaultSrcDestOperands(TmpOperands, 3540 X86Operand::CreateReg(X86::DX, NameLoc, NameLoc), 3541 DefaultMemDIOperand(NameLoc)); 3542 HadVerifyError = VerifyAndAdjustOperands(Operands, TmpOperands); 3543 } 3544 3545 // Append default arguments to "outs[bwld]" 3546 if (Name.startswith("outs") && 3547 (Operands.size() == 1 || Operands.size() == 3) && 3548 (Name == "outsb" || Name == "outsw" || Name == "outsl" || 3549 Name == "outsd" || Name == "outs")) { 3550 AddDefaultSrcDestOperands(TmpOperands, DefaultMemSIOperand(NameLoc), 3551 X86Operand::CreateReg(X86::DX, NameLoc, NameLoc)); 3552 HadVerifyError = VerifyAndAdjustOperands(Operands, TmpOperands); 3553 } 3554 3555 // Transform "lods[bwlq]" into "lods[bwlq] ($SIREG)" for appropriate 3556 // values of $SIREG according to the mode. It would be nice if this 3557 // could be achieved with InstAlias in the tables. 3558 if (Name.startswith("lods") && 3559 (Operands.size() == 1 || Operands.size() == 2) && 3560 (Name == "lods" || Name == "lodsb" || Name == "lodsw" || 3561 Name == "lodsl" || Name == "lodsd" || Name == "lodsq")) { 3562 TmpOperands.push_back(DefaultMemSIOperand(NameLoc)); 3563 HadVerifyError = VerifyAndAdjustOperands(Operands, TmpOperands); 3564 } 3565 3566 // Transform "stos[bwlq]" into "stos[bwlq] ($DIREG)" for appropriate 3567 // values of $DIREG according to the mode. It would be nice if this 3568 // could be achieved with InstAlias in the tables. 3569 if (Name.startswith("stos") && 3570 (Operands.size() == 1 || Operands.size() == 2) && 3571 (Name == "stos" || Name == "stosb" || Name == "stosw" || 3572 Name == "stosl" || Name == "stosd" || Name == "stosq")) { 3573 TmpOperands.push_back(DefaultMemDIOperand(NameLoc)); 3574 HadVerifyError = VerifyAndAdjustOperands(Operands, TmpOperands); 3575 } 3576 3577 // Transform "scas[bwlq]" into "scas[bwlq] ($DIREG)" for appropriate 3578 // values of $DIREG according to the mode. It would be nice if this 3579 // could be achieved with InstAlias in the tables. 3580 if (Name.startswith("scas") && 3581 (Operands.size() == 1 || Operands.size() == 2) && 3582 (Name == "scas" || Name == "scasb" || Name == "scasw" || 3583 Name == "scasl" || Name == "scasd" || Name == "scasq")) { 3584 TmpOperands.push_back(DefaultMemDIOperand(NameLoc)); 3585 HadVerifyError = VerifyAndAdjustOperands(Operands, TmpOperands); 3586 } 3587 3588 // Add default SI and DI operands to "cmps[bwlq]". 3589 if (Name.startswith("cmps") && 3590 (Operands.size() == 1 || Operands.size() == 3) && 3591 (Name == "cmps" || Name == "cmpsb" || Name == "cmpsw" || 3592 Name == "cmpsl" || Name == "cmpsd" || Name == "cmpsq")) { 3593 AddDefaultSrcDestOperands(TmpOperands, DefaultMemDIOperand(NameLoc), 3594 DefaultMemSIOperand(NameLoc)); 3595 HadVerifyError = VerifyAndAdjustOperands(Operands, TmpOperands); 3596 } 3597 3598 // Add default SI and DI operands to "movs[bwlq]". 3599 if (((Name.startswith("movs") && 3600 (Name == "movs" || Name == "movsb" || Name == "movsw" || 3601 Name == "movsl" || Name == "movsd" || Name == "movsq")) || 3602 (Name.startswith("smov") && 3603 (Name == "smov" || Name == "smovb" || Name == "smovw" || 3604 Name == "smovl" || Name == "smovd" || Name == "smovq"))) && 3605 (Operands.size() == 1 || Operands.size() == 3)) { 3606 if (Name == "movsd" && Operands.size() == 1 && !isParsingIntelSyntax()) 3607 Operands.back() = X86Operand::CreateToken("movsl", NameLoc); 3608 AddDefaultSrcDestOperands(TmpOperands, DefaultMemSIOperand(NameLoc), 3609 DefaultMemDIOperand(NameLoc)); 3610 HadVerifyError = VerifyAndAdjustOperands(Operands, TmpOperands); 3611 } 3612 3613 // Check if we encountered an error for one the string insturctions 3614 if (HadVerifyError) { 3615 return HadVerifyError; 3616 } 3617 3618 // Transforms "xlat mem8" into "xlatb" 3619 if ((Name == "xlat" || Name == "xlatb") && Operands.size() == 2) { 3620 X86Operand &Op1 = static_cast<X86Operand &>(*Operands[1]); 3621 if (Op1.isMem8()) { 3622 Warning(Op1.getStartLoc(), "memory operand is only for determining the " 3623 "size, (R|E)BX will be used for the location"); 3624 Operands.pop_back(); 3625 static_cast<X86Operand &>(*Operands[0]).setTokenValue("xlatb"); 3626 } 3627 } 3628 3629 if (Flags) 3630 Operands.push_back(X86Operand::CreatePrefix(Flags, NameLoc, NameLoc)); 3631 return false; 3632 } 3633 3634 bool X86AsmParser::processInstruction(MCInst &Inst, const OperandVector &Ops) { 3635 const MCRegisterInfo *MRI = getContext().getRegisterInfo(); 3636 3637 switch (Inst.getOpcode()) { 3638 default: return false; 3639 case X86::JMP_1: 3640 // {disp32} forces a larger displacement as if the instruction was relaxed. 3641 // NOTE: 16-bit mode uses 16-bit displacement even though it says {disp32}. 3642 // This matches GNU assembler. 3643 if (ForcedDispEncoding == DispEncoding_Disp32) { 3644 Inst.setOpcode(is16BitMode() ? X86::JMP_2 : X86::JMP_4); 3645 return true; 3646 } 3647 3648 return false; 3649 case X86::JCC_1: 3650 // {disp32} forces a larger displacement as if the instruction was relaxed. 3651 // NOTE: 16-bit mode uses 16-bit displacement even though it says {disp32}. 3652 // This matches GNU assembler. 3653 if (ForcedDispEncoding == DispEncoding_Disp32) { 3654 Inst.setOpcode(is16BitMode() ? X86::JCC_2 : X86::JCC_4); 3655 return true; 3656 } 3657 3658 return false; 3659 case X86::VMOVZPQILo2PQIrr: 3660 case X86::VMOVAPDrr: 3661 case X86::VMOVAPDYrr: 3662 case X86::VMOVAPSrr: 3663 case X86::VMOVAPSYrr: 3664 case X86::VMOVDQArr: 3665 case X86::VMOVDQAYrr: 3666 case X86::VMOVDQUrr: 3667 case X86::VMOVDQUYrr: 3668 case X86::VMOVUPDrr: 3669 case X86::VMOVUPDYrr: 3670 case X86::VMOVUPSrr: 3671 case X86::VMOVUPSYrr: { 3672 // We can get a smaller encoding by using VEX.R instead of VEX.B if one of 3673 // the registers is extended, but other isn't. 3674 if (ForcedVEXEncoding == VEXEncoding_VEX3 || 3675 MRI->getEncodingValue(Inst.getOperand(0).getReg()) >= 8 || 3676 MRI->getEncodingValue(Inst.getOperand(1).getReg()) < 8) 3677 return false; 3678 3679 unsigned NewOpc; 3680 switch (Inst.getOpcode()) { 3681 default: llvm_unreachable("Invalid opcode"); 3682 case X86::VMOVZPQILo2PQIrr: NewOpc = X86::VMOVPQI2QIrr; break; 3683 case X86::VMOVAPDrr: NewOpc = X86::VMOVAPDrr_REV; break; 3684 case X86::VMOVAPDYrr: NewOpc = X86::VMOVAPDYrr_REV; break; 3685 case X86::VMOVAPSrr: NewOpc = X86::VMOVAPSrr_REV; break; 3686 case X86::VMOVAPSYrr: NewOpc = X86::VMOVAPSYrr_REV; break; 3687 case X86::VMOVDQArr: NewOpc = X86::VMOVDQArr_REV; break; 3688 case X86::VMOVDQAYrr: NewOpc = X86::VMOVDQAYrr_REV; break; 3689 case X86::VMOVDQUrr: NewOpc = X86::VMOVDQUrr_REV; break; 3690 case X86::VMOVDQUYrr: NewOpc = X86::VMOVDQUYrr_REV; break; 3691 case X86::VMOVUPDrr: NewOpc = X86::VMOVUPDrr_REV; break; 3692 case X86::VMOVUPDYrr: NewOpc = X86::VMOVUPDYrr_REV; break; 3693 case X86::VMOVUPSrr: NewOpc = X86::VMOVUPSrr_REV; break; 3694 case X86::VMOVUPSYrr: NewOpc = X86::VMOVUPSYrr_REV; break; 3695 } 3696 Inst.setOpcode(NewOpc); 3697 return true; 3698 } 3699 case X86::VMOVSDrr: 3700 case X86::VMOVSSrr: { 3701 // We can get a smaller encoding by using VEX.R instead of VEX.B if one of 3702 // the registers is extended, but other isn't. 3703 if (ForcedVEXEncoding == VEXEncoding_VEX3 || 3704 MRI->getEncodingValue(Inst.getOperand(0).getReg()) >= 8 || 3705 MRI->getEncodingValue(Inst.getOperand(2).getReg()) < 8) 3706 return false; 3707 3708 unsigned NewOpc; 3709 switch (Inst.getOpcode()) { 3710 default: llvm_unreachable("Invalid opcode"); 3711 case X86::VMOVSDrr: NewOpc = X86::VMOVSDrr_REV; break; 3712 case X86::VMOVSSrr: NewOpc = X86::VMOVSSrr_REV; break; 3713 } 3714 Inst.setOpcode(NewOpc); 3715 return true; 3716 } 3717 case X86::RCR8ri: case X86::RCR16ri: case X86::RCR32ri: case X86::RCR64ri: 3718 case X86::RCL8ri: case X86::RCL16ri: case X86::RCL32ri: case X86::RCL64ri: 3719 case X86::ROR8ri: case X86::ROR16ri: case X86::ROR32ri: case X86::ROR64ri: 3720 case X86::ROL8ri: case X86::ROL16ri: case X86::ROL32ri: case X86::ROL64ri: 3721 case X86::SAR8ri: case X86::SAR16ri: case X86::SAR32ri: case X86::SAR64ri: 3722 case X86::SHR8ri: case X86::SHR16ri: case X86::SHR32ri: case X86::SHR64ri: 3723 case X86::SHL8ri: case X86::SHL16ri: case X86::SHL32ri: case X86::SHL64ri: { 3724 // Optimize s{hr,ar,hl} $1, <op> to "shift <op>". Similar for rotate. 3725 // FIXME: It would be great if we could just do this with an InstAlias. 3726 if (!Inst.getOperand(2).isImm() || Inst.getOperand(2).getImm() != 1) 3727 return false; 3728 3729 unsigned NewOpc; 3730 switch (Inst.getOpcode()) { 3731 default: llvm_unreachable("Invalid opcode"); 3732 case X86::RCR8ri: NewOpc = X86::RCR8r1; break; 3733 case X86::RCR16ri: NewOpc = X86::RCR16r1; break; 3734 case X86::RCR32ri: NewOpc = X86::RCR32r1; break; 3735 case X86::RCR64ri: NewOpc = X86::RCR64r1; break; 3736 case X86::RCL8ri: NewOpc = X86::RCL8r1; break; 3737 case X86::RCL16ri: NewOpc = X86::RCL16r1; break; 3738 case X86::RCL32ri: NewOpc = X86::RCL32r1; break; 3739 case X86::RCL64ri: NewOpc = X86::RCL64r1; break; 3740 case X86::ROR8ri: NewOpc = X86::ROR8r1; break; 3741 case X86::ROR16ri: NewOpc = X86::ROR16r1; break; 3742 case X86::ROR32ri: NewOpc = X86::ROR32r1; break; 3743 case X86::ROR64ri: NewOpc = X86::ROR64r1; break; 3744 case X86::ROL8ri: NewOpc = X86::ROL8r1; break; 3745 case X86::ROL16ri: NewOpc = X86::ROL16r1; break; 3746 case X86::ROL32ri: NewOpc = X86::ROL32r1; break; 3747 case X86::ROL64ri: NewOpc = X86::ROL64r1; break; 3748 case X86::SAR8ri: NewOpc = X86::SAR8r1; break; 3749 case X86::SAR16ri: NewOpc = X86::SAR16r1; break; 3750 case X86::SAR32ri: NewOpc = X86::SAR32r1; break; 3751 case X86::SAR64ri: NewOpc = X86::SAR64r1; break; 3752 case X86::SHR8ri: NewOpc = X86::SHR8r1; break; 3753 case X86::SHR16ri: NewOpc = X86::SHR16r1; break; 3754 case X86::SHR32ri: NewOpc = X86::SHR32r1; break; 3755 case X86::SHR64ri: NewOpc = X86::SHR64r1; break; 3756 case X86::SHL8ri: NewOpc = X86::SHL8r1; break; 3757 case X86::SHL16ri: NewOpc = X86::SHL16r1; break; 3758 case X86::SHL32ri: NewOpc = X86::SHL32r1; break; 3759 case X86::SHL64ri: NewOpc = X86::SHL64r1; break; 3760 } 3761 3762 MCInst TmpInst; 3763 TmpInst.setOpcode(NewOpc); 3764 TmpInst.addOperand(Inst.getOperand(0)); 3765 TmpInst.addOperand(Inst.getOperand(1)); 3766 Inst = TmpInst; 3767 return true; 3768 } 3769 case X86::RCR8mi: case X86::RCR16mi: case X86::RCR32mi: case X86::RCR64mi: 3770 case X86::RCL8mi: case X86::RCL16mi: case X86::RCL32mi: case X86::RCL64mi: 3771 case X86::ROR8mi: case X86::ROR16mi: case X86::ROR32mi: case X86::ROR64mi: 3772 case X86::ROL8mi: case X86::ROL16mi: case X86::ROL32mi: case X86::ROL64mi: 3773 case X86::SAR8mi: case X86::SAR16mi: case X86::SAR32mi: case X86::SAR64mi: 3774 case X86::SHR8mi: case X86::SHR16mi: case X86::SHR32mi: case X86::SHR64mi: 3775 case X86::SHL8mi: case X86::SHL16mi: case X86::SHL32mi: case X86::SHL64mi: { 3776 // Optimize s{hr,ar,hl} $1, <op> to "shift <op>". Similar for rotate. 3777 // FIXME: It would be great if we could just do this with an InstAlias. 3778 if (!Inst.getOperand(X86::AddrNumOperands).isImm() || 3779 Inst.getOperand(X86::AddrNumOperands).getImm() != 1) 3780 return false; 3781 3782 unsigned NewOpc; 3783 switch (Inst.getOpcode()) { 3784 default: llvm_unreachable("Invalid opcode"); 3785 case X86::RCR8mi: NewOpc = X86::RCR8m1; break; 3786 case X86::RCR16mi: NewOpc = X86::RCR16m1; break; 3787 case X86::RCR32mi: NewOpc = X86::RCR32m1; break; 3788 case X86::RCR64mi: NewOpc = X86::RCR64m1; break; 3789 case X86::RCL8mi: NewOpc = X86::RCL8m1; break; 3790 case X86::RCL16mi: NewOpc = X86::RCL16m1; break; 3791 case X86::RCL32mi: NewOpc = X86::RCL32m1; break; 3792 case X86::RCL64mi: NewOpc = X86::RCL64m1; break; 3793 case X86::ROR8mi: NewOpc = X86::ROR8m1; break; 3794 case X86::ROR16mi: NewOpc = X86::ROR16m1; break; 3795 case X86::ROR32mi: NewOpc = X86::ROR32m1; break; 3796 case X86::ROR64mi: NewOpc = X86::ROR64m1; break; 3797 case X86::ROL8mi: NewOpc = X86::ROL8m1; break; 3798 case X86::ROL16mi: NewOpc = X86::ROL16m1; break; 3799 case X86::ROL32mi: NewOpc = X86::ROL32m1; break; 3800 case X86::ROL64mi: NewOpc = X86::ROL64m1; break; 3801 case X86::SAR8mi: NewOpc = X86::SAR8m1; break; 3802 case X86::SAR16mi: NewOpc = X86::SAR16m1; break; 3803 case X86::SAR32mi: NewOpc = X86::SAR32m1; break; 3804 case X86::SAR64mi: NewOpc = X86::SAR64m1; break; 3805 case X86::SHR8mi: NewOpc = X86::SHR8m1; break; 3806 case X86::SHR16mi: NewOpc = X86::SHR16m1; break; 3807 case X86::SHR32mi: NewOpc = X86::SHR32m1; break; 3808 case X86::SHR64mi: NewOpc = X86::SHR64m1; break; 3809 case X86::SHL8mi: NewOpc = X86::SHL8m1; break; 3810 case X86::SHL16mi: NewOpc = X86::SHL16m1; break; 3811 case X86::SHL32mi: NewOpc = X86::SHL32m1; break; 3812 case X86::SHL64mi: NewOpc = X86::SHL64m1; break; 3813 } 3814 3815 MCInst TmpInst; 3816 TmpInst.setOpcode(NewOpc); 3817 for (int i = 0; i != X86::AddrNumOperands; ++i) 3818 TmpInst.addOperand(Inst.getOperand(i)); 3819 Inst = TmpInst; 3820 return true; 3821 } 3822 case X86::INT: { 3823 // Transforms "int $3" into "int3" as a size optimization. We can't write an 3824 // instalias with an immediate operand yet. 3825 if (!Inst.getOperand(0).isImm() || Inst.getOperand(0).getImm() != 3) 3826 return false; 3827 3828 MCInst TmpInst; 3829 TmpInst.setOpcode(X86::INT3); 3830 Inst = TmpInst; 3831 return true; 3832 } 3833 } 3834 } 3835 3836 bool X86AsmParser::validateInstruction(MCInst &Inst, const OperandVector &Ops) { 3837 using namespace X86; 3838 const MCRegisterInfo *MRI = getContext().getRegisterInfo(); 3839 unsigned Opcode = Inst.getOpcode(); 3840 uint64_t TSFlags = MII.get(Opcode).TSFlags; 3841 if (isVFCMADDCPH(Opcode) || isVFCMADDCSH(Opcode) || isVFMADDCPH(Opcode) || 3842 isVFMADDCSH(Opcode)) { 3843 unsigned Dest = Inst.getOperand(0).getReg(); 3844 for (unsigned i = 2; i < Inst.getNumOperands(); i++) 3845 if (Inst.getOperand(i).isReg() && Dest == Inst.getOperand(i).getReg()) 3846 return Warning(Ops[0]->getStartLoc(), "Destination register should be " 3847 "distinct from source registers"); 3848 } else if (isVFCMULCPH(Opcode) || isVFCMULCSH(Opcode) || isVFMULCPH(Opcode) || 3849 isVFMULCSH(Opcode)) { 3850 unsigned Dest = Inst.getOperand(0).getReg(); 3851 // The mask variants have different operand list. Scan from the third 3852 // operand to avoid emitting incorrect warning. 3853 // VFMULCPHZrr Dest, Src1, Src2 3854 // VFMULCPHZrrk Dest, Dest, Mask, Src1, Src2 3855 // VFMULCPHZrrkz Dest, Mask, Src1, Src2 3856 for (unsigned i = TSFlags & X86II::EVEX_K ? 2 : 1; 3857 i < Inst.getNumOperands(); i++) 3858 if (Inst.getOperand(i).isReg() && Dest == Inst.getOperand(i).getReg()) 3859 return Warning(Ops[0]->getStartLoc(), "Destination register should be " 3860 "distinct from source registers"); 3861 } else if (isV4FMADDPS(Opcode) || isV4FMADDSS(Opcode) || 3862 isV4FNMADDPS(Opcode) || isV4FNMADDSS(Opcode) || 3863 isVP4DPWSSDS(Opcode) || isVP4DPWSSD(Opcode)) { 3864 unsigned Src2 = Inst.getOperand(Inst.getNumOperands() - 3865 X86::AddrNumOperands - 1).getReg(); 3866 unsigned Src2Enc = MRI->getEncodingValue(Src2); 3867 if (Src2Enc % 4 != 0) { 3868 StringRef RegName = X86IntelInstPrinter::getRegisterName(Src2); 3869 unsigned GroupStart = (Src2Enc / 4) * 4; 3870 unsigned GroupEnd = GroupStart + 3; 3871 return Warning(Ops[0]->getStartLoc(), 3872 "source register '" + RegName + "' implicitly denotes '" + 3873 RegName.take_front(3) + Twine(GroupStart) + "' to '" + 3874 RegName.take_front(3) + Twine(GroupEnd) + 3875 "' source group"); 3876 } 3877 } else if (isVGATHERDPD(Opcode) || isVGATHERDPS(Opcode) || 3878 isVGATHERQPD(Opcode) || isVGATHERQPS(Opcode) || 3879 isVPGATHERDD(Opcode) || isVPGATHERDQ(Opcode) || 3880 isVPGATHERQD(Opcode) || isVPGATHERQQ(Opcode)) { 3881 bool HasEVEX = (TSFlags & X86II::EncodingMask) == X86II::EVEX; 3882 if (HasEVEX) { 3883 unsigned Dest = MRI->getEncodingValue(Inst.getOperand(0).getReg()); 3884 unsigned Index = MRI->getEncodingValue( 3885 Inst.getOperand(4 + X86::AddrIndexReg).getReg()); 3886 if (Dest == Index) 3887 return Warning(Ops[0]->getStartLoc(), "index and destination registers " 3888 "should be distinct"); 3889 } else { 3890 unsigned Dest = MRI->getEncodingValue(Inst.getOperand(0).getReg()); 3891 unsigned Mask = MRI->getEncodingValue(Inst.getOperand(1).getReg()); 3892 unsigned Index = MRI->getEncodingValue( 3893 Inst.getOperand(3 + X86::AddrIndexReg).getReg()); 3894 if (Dest == Mask || Dest == Index || Mask == Index) 3895 return Warning(Ops[0]->getStartLoc(), "mask, index, and destination " 3896 "registers should be distinct"); 3897 } 3898 } 3899 3900 // Check that we aren't mixing AH/BH/CH/DH with REX prefix. We only need to 3901 // check this with the legacy encoding, VEX/EVEX/XOP don't use REX. 3902 if ((TSFlags & X86II::EncodingMask) == 0) { 3903 MCPhysReg HReg = X86::NoRegister; 3904 bool UsesRex = TSFlags & X86II::REX_W; 3905 unsigned NumOps = Inst.getNumOperands(); 3906 for (unsigned i = 0; i != NumOps; ++i) { 3907 const MCOperand &MO = Inst.getOperand(i); 3908 if (!MO.isReg()) 3909 continue; 3910 unsigned Reg = MO.getReg(); 3911 if (Reg == X86::AH || Reg == X86::BH || Reg == X86::CH || Reg == X86::DH) 3912 HReg = Reg; 3913 if (X86II::isX86_64NonExtLowByteReg(Reg) || 3914 X86II::isX86_64ExtendedReg(Reg)) 3915 UsesRex = true; 3916 } 3917 3918 if (UsesRex && HReg != X86::NoRegister) { 3919 StringRef RegName = X86IntelInstPrinter::getRegisterName(HReg); 3920 return Error(Ops[0]->getStartLoc(), 3921 "can't encode '" + RegName + "' in an instruction requiring " 3922 "REX prefix"); 3923 } 3924 } 3925 3926 return false; 3927 } 3928 3929 static const char *getSubtargetFeatureName(uint64_t Val); 3930 3931 void X86AsmParser::emitWarningForSpecialLVIInstruction(SMLoc Loc) { 3932 Warning(Loc, "Instruction may be vulnerable to LVI and " 3933 "requires manual mitigation"); 3934 Note(SMLoc(), "See https://software.intel.com/" 3935 "security-software-guidance/insights/" 3936 "deep-dive-load-value-injection#specialinstructions" 3937 " for more information"); 3938 } 3939 3940 /// RET instructions and also instructions that indirect calls/jumps from memory 3941 /// combine a load and a branch within a single instruction. To mitigate these 3942 /// instructions against LVI, they must be decomposed into separate load and 3943 /// branch instructions, with an LFENCE in between. For more details, see: 3944 /// - X86LoadValueInjectionRetHardening.cpp 3945 /// - X86LoadValueInjectionIndirectThunks.cpp 3946 /// - https://software.intel.com/security-software-guidance/insights/deep-dive-load-value-injection 3947 /// 3948 /// Returns `true` if a mitigation was applied or warning was emitted. 3949 void X86AsmParser::applyLVICFIMitigation(MCInst &Inst, MCStreamer &Out) { 3950 // Information on control-flow instructions that require manual mitigation can 3951 // be found here: 3952 // https://software.intel.com/security-software-guidance/insights/deep-dive-load-value-injection#specialinstructions 3953 switch (Inst.getOpcode()) { 3954 case X86::RET16: 3955 case X86::RET32: 3956 case X86::RET64: 3957 case X86::RETI16: 3958 case X86::RETI32: 3959 case X86::RETI64: { 3960 MCInst ShlInst, FenceInst; 3961 bool Parse32 = is32BitMode() || Code16GCC; 3962 unsigned Basereg = 3963 is64BitMode() ? X86::RSP : (Parse32 ? X86::ESP : X86::SP); 3964 const MCExpr *Disp = MCConstantExpr::create(0, getContext()); 3965 auto ShlMemOp = X86Operand::CreateMem(getPointerWidth(), /*SegReg=*/0, Disp, 3966 /*BaseReg=*/Basereg, /*IndexReg=*/0, 3967 /*Scale=*/1, SMLoc{}, SMLoc{}, 0); 3968 ShlInst.setOpcode(X86::SHL64mi); 3969 ShlMemOp->addMemOperands(ShlInst, 5); 3970 ShlInst.addOperand(MCOperand::createImm(0)); 3971 FenceInst.setOpcode(X86::LFENCE); 3972 Out.emitInstruction(ShlInst, getSTI()); 3973 Out.emitInstruction(FenceInst, getSTI()); 3974 return; 3975 } 3976 case X86::JMP16m: 3977 case X86::JMP32m: 3978 case X86::JMP64m: 3979 case X86::CALL16m: 3980 case X86::CALL32m: 3981 case X86::CALL64m: 3982 emitWarningForSpecialLVIInstruction(Inst.getLoc()); 3983 return; 3984 } 3985 } 3986 3987 /// To mitigate LVI, every instruction that performs a load can be followed by 3988 /// an LFENCE instruction to squash any potential mis-speculation. There are 3989 /// some instructions that require additional considerations, and may requre 3990 /// manual mitigation. For more details, see: 3991 /// https://software.intel.com/security-software-guidance/insights/deep-dive-load-value-injection 3992 /// 3993 /// Returns `true` if a mitigation was applied or warning was emitted. 3994 void X86AsmParser::applyLVILoadHardeningMitigation(MCInst &Inst, 3995 MCStreamer &Out) { 3996 auto Opcode = Inst.getOpcode(); 3997 auto Flags = Inst.getFlags(); 3998 if ((Flags & X86::IP_HAS_REPEAT) || (Flags & X86::IP_HAS_REPEAT_NE)) { 3999 // Information on REP string instructions that require manual mitigation can 4000 // be found here: 4001 // https://software.intel.com/security-software-guidance/insights/deep-dive-load-value-injection#specialinstructions 4002 switch (Opcode) { 4003 case X86::CMPSB: 4004 case X86::CMPSW: 4005 case X86::CMPSL: 4006 case X86::CMPSQ: 4007 case X86::SCASB: 4008 case X86::SCASW: 4009 case X86::SCASL: 4010 case X86::SCASQ: 4011 emitWarningForSpecialLVIInstruction(Inst.getLoc()); 4012 return; 4013 } 4014 } else if (Opcode == X86::REP_PREFIX || Opcode == X86::REPNE_PREFIX) { 4015 // If a REP instruction is found on its own line, it may or may not be 4016 // followed by a vulnerable instruction. Emit a warning just in case. 4017 emitWarningForSpecialLVIInstruction(Inst.getLoc()); 4018 return; 4019 } 4020 4021 const MCInstrDesc &MCID = MII.get(Inst.getOpcode()); 4022 4023 // Can't mitigate after terminators or calls. A control flow change may have 4024 // already occurred. 4025 if (MCID.isTerminator() || MCID.isCall()) 4026 return; 4027 4028 // LFENCE has the mayLoad property, don't double fence. 4029 if (MCID.mayLoad() && Inst.getOpcode() != X86::LFENCE) { 4030 MCInst FenceInst; 4031 FenceInst.setOpcode(X86::LFENCE); 4032 Out.emitInstruction(FenceInst, getSTI()); 4033 } 4034 } 4035 4036 void X86AsmParser::emitInstruction(MCInst &Inst, OperandVector &Operands, 4037 MCStreamer &Out) { 4038 if (LVIInlineAsmHardening && 4039 getSTI().getFeatureBits()[X86::FeatureLVIControlFlowIntegrity]) 4040 applyLVICFIMitigation(Inst, Out); 4041 4042 Out.emitInstruction(Inst, getSTI()); 4043 4044 if (LVIInlineAsmHardening && 4045 getSTI().getFeatureBits()[X86::FeatureLVILoadHardening]) 4046 applyLVILoadHardeningMitigation(Inst, Out); 4047 } 4048 4049 bool X86AsmParser::MatchAndEmitInstruction(SMLoc IDLoc, unsigned &Opcode, 4050 OperandVector &Operands, 4051 MCStreamer &Out, uint64_t &ErrorInfo, 4052 bool MatchingInlineAsm) { 4053 if (isParsingIntelSyntax()) 4054 return MatchAndEmitIntelInstruction(IDLoc, Opcode, Operands, Out, ErrorInfo, 4055 MatchingInlineAsm); 4056 return MatchAndEmitATTInstruction(IDLoc, Opcode, Operands, Out, ErrorInfo, 4057 MatchingInlineAsm); 4058 } 4059 4060 void X86AsmParser::MatchFPUWaitAlias(SMLoc IDLoc, X86Operand &Op, 4061 OperandVector &Operands, MCStreamer &Out, 4062 bool MatchingInlineAsm) { 4063 // FIXME: This should be replaced with a real .td file alias mechanism. 4064 // Also, MatchInstructionImpl should actually *do* the EmitInstruction 4065 // call. 4066 const char *Repl = StringSwitch<const char *>(Op.getToken()) 4067 .Case("finit", "fninit") 4068 .Case("fsave", "fnsave") 4069 .Case("fstcw", "fnstcw") 4070 .Case("fstcww", "fnstcw") 4071 .Case("fstenv", "fnstenv") 4072 .Case("fstsw", "fnstsw") 4073 .Case("fstsww", "fnstsw") 4074 .Case("fclex", "fnclex") 4075 .Default(nullptr); 4076 if (Repl) { 4077 MCInst Inst; 4078 Inst.setOpcode(X86::WAIT); 4079 Inst.setLoc(IDLoc); 4080 if (!MatchingInlineAsm) 4081 emitInstruction(Inst, Operands, Out); 4082 Operands[0] = X86Operand::CreateToken(Repl, IDLoc); 4083 } 4084 } 4085 4086 bool X86AsmParser::ErrorMissingFeature(SMLoc IDLoc, 4087 const FeatureBitset &MissingFeatures, 4088 bool MatchingInlineAsm) { 4089 assert(MissingFeatures.any() && "Unknown missing feature!"); 4090 SmallString<126> Msg; 4091 raw_svector_ostream OS(Msg); 4092 OS << "instruction requires:"; 4093 for (unsigned i = 0, e = MissingFeatures.size(); i != e; ++i) { 4094 if (MissingFeatures[i]) 4095 OS << ' ' << getSubtargetFeatureName(i); 4096 } 4097 return Error(IDLoc, OS.str(), SMRange(), MatchingInlineAsm); 4098 } 4099 4100 static unsigned getPrefixes(OperandVector &Operands) { 4101 unsigned Result = 0; 4102 X86Operand &Prefix = static_cast<X86Operand &>(*Operands.back()); 4103 if (Prefix.isPrefix()) { 4104 Result = Prefix.getPrefix(); 4105 Operands.pop_back(); 4106 } 4107 return Result; 4108 } 4109 4110 unsigned X86AsmParser::checkTargetMatchPredicate(MCInst &Inst) { 4111 unsigned Opc = Inst.getOpcode(); 4112 const MCInstrDesc &MCID = MII.get(Opc); 4113 4114 if (ForcedVEXEncoding == VEXEncoding_EVEX && 4115 (MCID.TSFlags & X86II::EncodingMask) != X86II::EVEX) 4116 return Match_Unsupported; 4117 4118 if ((ForcedVEXEncoding == VEXEncoding_VEX || 4119 ForcedVEXEncoding == VEXEncoding_VEX2 || 4120 ForcedVEXEncoding == VEXEncoding_VEX3) && 4121 (MCID.TSFlags & X86II::EncodingMask) != X86II::VEX) 4122 return Match_Unsupported; 4123 4124 // These instructions are only available with {vex}, {vex2} or {vex3} prefix 4125 if (MCID.TSFlags & X86II::ExplicitVEXPrefix && 4126 (ForcedVEXEncoding != VEXEncoding_VEX && 4127 ForcedVEXEncoding != VEXEncoding_VEX2 && 4128 ForcedVEXEncoding != VEXEncoding_VEX3)) 4129 return Match_Unsupported; 4130 4131 return Match_Success; 4132 } 4133 4134 bool X86AsmParser::MatchAndEmitATTInstruction(SMLoc IDLoc, unsigned &Opcode, 4135 OperandVector &Operands, 4136 MCStreamer &Out, 4137 uint64_t &ErrorInfo, 4138 bool MatchingInlineAsm) { 4139 assert(!Operands.empty() && "Unexpect empty operand list!"); 4140 assert((*Operands[0]).isToken() && "Leading operand should always be a mnemonic!"); 4141 SMRange EmptyRange = std::nullopt; 4142 4143 // First, handle aliases that expand to multiple instructions. 4144 MatchFPUWaitAlias(IDLoc, static_cast<X86Operand &>(*Operands[0]), Operands, 4145 Out, MatchingInlineAsm); 4146 X86Operand &Op = static_cast<X86Operand &>(*Operands[0]); 4147 unsigned Prefixes = getPrefixes(Operands); 4148 4149 MCInst Inst; 4150 4151 // If VEX/EVEX encoding is forced, we need to pass the USE_* flag to the 4152 // encoder and printer. 4153 if (ForcedVEXEncoding == VEXEncoding_VEX) 4154 Prefixes |= X86::IP_USE_VEX; 4155 else if (ForcedVEXEncoding == VEXEncoding_VEX2) 4156 Prefixes |= X86::IP_USE_VEX2; 4157 else if (ForcedVEXEncoding == VEXEncoding_VEX3) 4158 Prefixes |= X86::IP_USE_VEX3; 4159 else if (ForcedVEXEncoding == VEXEncoding_EVEX) 4160 Prefixes |= X86::IP_USE_EVEX; 4161 4162 // Set encoded flags for {disp8} and {disp32}. 4163 if (ForcedDispEncoding == DispEncoding_Disp8) 4164 Prefixes |= X86::IP_USE_DISP8; 4165 else if (ForcedDispEncoding == DispEncoding_Disp32) 4166 Prefixes |= X86::IP_USE_DISP32; 4167 4168 if (Prefixes) 4169 Inst.setFlags(Prefixes); 4170 4171 // In 16-bit mode, if data32 is specified, temporarily switch to 32-bit mode 4172 // when matching the instruction. 4173 if (ForcedDataPrefix == X86::Is32Bit) 4174 SwitchMode(X86::Is32Bit); 4175 // First, try a direct match. 4176 FeatureBitset MissingFeatures; 4177 unsigned OriginalError = MatchInstruction(Operands, Inst, ErrorInfo, 4178 MissingFeatures, MatchingInlineAsm, 4179 isParsingIntelSyntax()); 4180 if (ForcedDataPrefix == X86::Is32Bit) { 4181 SwitchMode(X86::Is16Bit); 4182 ForcedDataPrefix = 0; 4183 } 4184 switch (OriginalError) { 4185 default: llvm_unreachable("Unexpected match result!"); 4186 case Match_Success: 4187 if (!MatchingInlineAsm && validateInstruction(Inst, Operands)) 4188 return true; 4189 // Some instructions need post-processing to, for example, tweak which 4190 // encoding is selected. Loop on it while changes happen so the 4191 // individual transformations can chain off each other. 4192 if (!MatchingInlineAsm) 4193 while (processInstruction(Inst, Operands)) 4194 ; 4195 4196 Inst.setLoc(IDLoc); 4197 if (!MatchingInlineAsm) 4198 emitInstruction(Inst, Operands, Out); 4199 Opcode = Inst.getOpcode(); 4200 return false; 4201 case Match_InvalidImmUnsignedi4: { 4202 SMLoc ErrorLoc = ((X86Operand &)*Operands[ErrorInfo]).getStartLoc(); 4203 if (ErrorLoc == SMLoc()) 4204 ErrorLoc = IDLoc; 4205 return Error(ErrorLoc, "immediate must be an integer in range [0, 15]", 4206 EmptyRange, MatchingInlineAsm); 4207 } 4208 case Match_MissingFeature: 4209 return ErrorMissingFeature(IDLoc, MissingFeatures, MatchingInlineAsm); 4210 case Match_InvalidOperand: 4211 case Match_MnemonicFail: 4212 case Match_Unsupported: 4213 break; 4214 } 4215 if (Op.getToken().empty()) { 4216 Error(IDLoc, "instruction must have size higher than 0", EmptyRange, 4217 MatchingInlineAsm); 4218 return true; 4219 } 4220 4221 // FIXME: Ideally, we would only attempt suffix matches for things which are 4222 // valid prefixes, and we could just infer the right unambiguous 4223 // type. However, that requires substantially more matcher support than the 4224 // following hack. 4225 4226 // Change the operand to point to a temporary token. 4227 StringRef Base = Op.getToken(); 4228 SmallString<16> Tmp; 4229 Tmp += Base; 4230 Tmp += ' '; 4231 Op.setTokenValue(Tmp); 4232 4233 // If this instruction starts with an 'f', then it is a floating point stack 4234 // instruction. These come in up to three forms for 32-bit, 64-bit, and 4235 // 80-bit floating point, which use the suffixes s,l,t respectively. 4236 // 4237 // Otherwise, we assume that this may be an integer instruction, which comes 4238 // in 8/16/32/64-bit forms using the b,w,l,q suffixes respectively. 4239 const char *Suffixes = Base[0] != 'f' ? "bwlq" : "slt\0"; 4240 // MemSize corresponding to Suffixes. { 8, 16, 32, 64 } { 32, 64, 80, 0 } 4241 const char *MemSize = Base[0] != 'f' ? "\x08\x10\x20\x40" : "\x20\x40\x50\0"; 4242 4243 // Check for the various suffix matches. 4244 uint64_t ErrorInfoIgnore; 4245 FeatureBitset ErrorInfoMissingFeatures; // Init suppresses compiler warnings. 4246 unsigned Match[4]; 4247 4248 // Some instruction like VPMULDQ is NOT the variant of VPMULD but a new one. 4249 // So we should make sure the suffix matcher only works for memory variant 4250 // that has the same size with the suffix. 4251 // FIXME: This flag is a workaround for legacy instructions that didn't 4252 // declare non suffix variant assembly. 4253 bool HasVectorReg = false; 4254 X86Operand *MemOp = nullptr; 4255 for (const auto &Op : Operands) { 4256 X86Operand *X86Op = static_cast<X86Operand *>(Op.get()); 4257 if (X86Op->isVectorReg()) 4258 HasVectorReg = true; 4259 else if (X86Op->isMem()) { 4260 MemOp = X86Op; 4261 assert(MemOp->Mem.Size == 0 && "Memory size always 0 under ATT syntax"); 4262 // Have we found an unqualified memory operand, 4263 // break. IA allows only one memory operand. 4264 break; 4265 } 4266 } 4267 4268 for (unsigned I = 0, E = std::size(Match); I != E; ++I) { 4269 Tmp.back() = Suffixes[I]; 4270 if (MemOp && HasVectorReg) 4271 MemOp->Mem.Size = MemSize[I]; 4272 Match[I] = Match_MnemonicFail; 4273 if (MemOp || !HasVectorReg) { 4274 Match[I] = 4275 MatchInstruction(Operands, Inst, ErrorInfoIgnore, MissingFeatures, 4276 MatchingInlineAsm, isParsingIntelSyntax()); 4277 // If this returned as a missing feature failure, remember that. 4278 if (Match[I] == Match_MissingFeature) 4279 ErrorInfoMissingFeatures = MissingFeatures; 4280 } 4281 } 4282 4283 // Restore the old token. 4284 Op.setTokenValue(Base); 4285 4286 // If exactly one matched, then we treat that as a successful match (and the 4287 // instruction will already have been filled in correctly, since the failing 4288 // matches won't have modified it). 4289 unsigned NumSuccessfulMatches = llvm::count(Match, Match_Success); 4290 if (NumSuccessfulMatches == 1) { 4291 if (!MatchingInlineAsm && validateInstruction(Inst, Operands)) 4292 return true; 4293 // Some instructions need post-processing to, for example, tweak which 4294 // encoding is selected. Loop on it while changes happen so the 4295 // individual transformations can chain off each other. 4296 if (!MatchingInlineAsm) 4297 while (processInstruction(Inst, Operands)) 4298 ; 4299 4300 Inst.setLoc(IDLoc); 4301 if (!MatchingInlineAsm) 4302 emitInstruction(Inst, Operands, Out); 4303 Opcode = Inst.getOpcode(); 4304 return false; 4305 } 4306 4307 // Otherwise, the match failed, try to produce a decent error message. 4308 4309 // If we had multiple suffix matches, then identify this as an ambiguous 4310 // match. 4311 if (NumSuccessfulMatches > 1) { 4312 char MatchChars[4]; 4313 unsigned NumMatches = 0; 4314 for (unsigned I = 0, E = std::size(Match); I != E; ++I) 4315 if (Match[I] == Match_Success) 4316 MatchChars[NumMatches++] = Suffixes[I]; 4317 4318 SmallString<126> Msg; 4319 raw_svector_ostream OS(Msg); 4320 OS << "ambiguous instructions require an explicit suffix (could be "; 4321 for (unsigned i = 0; i != NumMatches; ++i) { 4322 if (i != 0) 4323 OS << ", "; 4324 if (i + 1 == NumMatches) 4325 OS << "or "; 4326 OS << "'" << Base << MatchChars[i] << "'"; 4327 } 4328 OS << ")"; 4329 Error(IDLoc, OS.str(), EmptyRange, MatchingInlineAsm); 4330 return true; 4331 } 4332 4333 // Okay, we know that none of the variants matched successfully. 4334 4335 // If all of the instructions reported an invalid mnemonic, then the original 4336 // mnemonic was invalid. 4337 if (llvm::count(Match, Match_MnemonicFail) == 4) { 4338 if (OriginalError == Match_MnemonicFail) 4339 return Error(IDLoc, "invalid instruction mnemonic '" + Base + "'", 4340 Op.getLocRange(), MatchingInlineAsm); 4341 4342 if (OriginalError == Match_Unsupported) 4343 return Error(IDLoc, "unsupported instruction", EmptyRange, 4344 MatchingInlineAsm); 4345 4346 assert(OriginalError == Match_InvalidOperand && "Unexpected error"); 4347 // Recover location info for the operand if we know which was the problem. 4348 if (ErrorInfo != ~0ULL) { 4349 if (ErrorInfo >= Operands.size()) 4350 return Error(IDLoc, "too few operands for instruction", EmptyRange, 4351 MatchingInlineAsm); 4352 4353 X86Operand &Operand = (X86Operand &)*Operands[ErrorInfo]; 4354 if (Operand.getStartLoc().isValid()) { 4355 SMRange OperandRange = Operand.getLocRange(); 4356 return Error(Operand.getStartLoc(), "invalid operand for instruction", 4357 OperandRange, MatchingInlineAsm); 4358 } 4359 } 4360 4361 return Error(IDLoc, "invalid operand for instruction", EmptyRange, 4362 MatchingInlineAsm); 4363 } 4364 4365 // If one instruction matched as unsupported, report this as unsupported. 4366 if (llvm::count(Match, Match_Unsupported) == 1) { 4367 return Error(IDLoc, "unsupported instruction", EmptyRange, 4368 MatchingInlineAsm); 4369 } 4370 4371 // If one instruction matched with a missing feature, report this as a 4372 // missing feature. 4373 if (llvm::count(Match, Match_MissingFeature) == 1) { 4374 ErrorInfo = Match_MissingFeature; 4375 return ErrorMissingFeature(IDLoc, ErrorInfoMissingFeatures, 4376 MatchingInlineAsm); 4377 } 4378 4379 // If one instruction matched with an invalid operand, report this as an 4380 // operand failure. 4381 if (llvm::count(Match, Match_InvalidOperand) == 1) { 4382 return Error(IDLoc, "invalid operand for instruction", EmptyRange, 4383 MatchingInlineAsm); 4384 } 4385 4386 // If all of these were an outright failure, report it in a useless way. 4387 Error(IDLoc, "unknown use of instruction mnemonic without a size suffix", 4388 EmptyRange, MatchingInlineAsm); 4389 return true; 4390 } 4391 4392 bool X86AsmParser::MatchAndEmitIntelInstruction(SMLoc IDLoc, unsigned &Opcode, 4393 OperandVector &Operands, 4394 MCStreamer &Out, 4395 uint64_t &ErrorInfo, 4396 bool MatchingInlineAsm) { 4397 assert(!Operands.empty() && "Unexpect empty operand list!"); 4398 assert((*Operands[0]).isToken() && "Leading operand should always be a mnemonic!"); 4399 StringRef Mnemonic = (static_cast<X86Operand &>(*Operands[0])).getToken(); 4400 SMRange EmptyRange = std::nullopt; 4401 StringRef Base = (static_cast<X86Operand &>(*Operands[0])).getToken(); 4402 unsigned Prefixes = getPrefixes(Operands); 4403 4404 // First, handle aliases that expand to multiple instructions. 4405 MatchFPUWaitAlias(IDLoc, static_cast<X86Operand &>(*Operands[0]), Operands, Out, MatchingInlineAsm); 4406 X86Operand &Op = static_cast<X86Operand &>(*Operands[0]); 4407 4408 MCInst Inst; 4409 4410 // If VEX/EVEX encoding is forced, we need to pass the USE_* flag to the 4411 // encoder and printer. 4412 if (ForcedVEXEncoding == VEXEncoding_VEX) 4413 Prefixes |= X86::IP_USE_VEX; 4414 else if (ForcedVEXEncoding == VEXEncoding_VEX2) 4415 Prefixes |= X86::IP_USE_VEX2; 4416 else if (ForcedVEXEncoding == VEXEncoding_VEX3) 4417 Prefixes |= X86::IP_USE_VEX3; 4418 else if (ForcedVEXEncoding == VEXEncoding_EVEX) 4419 Prefixes |= X86::IP_USE_EVEX; 4420 4421 // Set encoded flags for {disp8} and {disp32}. 4422 if (ForcedDispEncoding == DispEncoding_Disp8) 4423 Prefixes |= X86::IP_USE_DISP8; 4424 else if (ForcedDispEncoding == DispEncoding_Disp32) 4425 Prefixes |= X86::IP_USE_DISP32; 4426 4427 if (Prefixes) 4428 Inst.setFlags(Prefixes); 4429 4430 // Find one unsized memory operand, if present. 4431 X86Operand *UnsizedMemOp = nullptr; 4432 for (const auto &Op : Operands) { 4433 X86Operand *X86Op = static_cast<X86Operand *>(Op.get()); 4434 if (X86Op->isMemUnsized()) { 4435 UnsizedMemOp = X86Op; 4436 // Have we found an unqualified memory operand, 4437 // break. IA allows only one memory operand. 4438 break; 4439 } 4440 } 4441 4442 // Allow some instructions to have implicitly pointer-sized operands. This is 4443 // compatible with gas. 4444 if (UnsizedMemOp) { 4445 static const char *const PtrSizedInstrs[] = {"call", "jmp", "push"}; 4446 for (const char *Instr : PtrSizedInstrs) { 4447 if (Mnemonic == Instr) { 4448 UnsizedMemOp->Mem.Size = getPointerWidth(); 4449 break; 4450 } 4451 } 4452 } 4453 4454 SmallVector<unsigned, 8> Match; 4455 FeatureBitset ErrorInfoMissingFeatures; 4456 FeatureBitset MissingFeatures; 4457 4458 // If unsized push has immediate operand we should default the default pointer 4459 // size for the size. 4460 if (Mnemonic == "push" && Operands.size() == 2) { 4461 auto *X86Op = static_cast<X86Operand *>(Operands[1].get()); 4462 if (X86Op->isImm()) { 4463 // If it's not a constant fall through and let remainder take care of it. 4464 const auto *CE = dyn_cast<MCConstantExpr>(X86Op->getImm()); 4465 unsigned Size = getPointerWidth(); 4466 if (CE && 4467 (isIntN(Size, CE->getValue()) || isUIntN(Size, CE->getValue()))) { 4468 SmallString<16> Tmp; 4469 Tmp += Base; 4470 Tmp += (is64BitMode()) 4471 ? "q" 4472 : (is32BitMode()) ? "l" : (is16BitMode()) ? "w" : " "; 4473 Op.setTokenValue(Tmp); 4474 // Do match in ATT mode to allow explicit suffix usage. 4475 Match.push_back(MatchInstruction(Operands, Inst, ErrorInfo, 4476 MissingFeatures, MatchingInlineAsm, 4477 false /*isParsingIntelSyntax()*/)); 4478 Op.setTokenValue(Base); 4479 } 4480 } 4481 } 4482 4483 // If an unsized memory operand is present, try to match with each memory 4484 // operand size. In Intel assembly, the size is not part of the instruction 4485 // mnemonic. 4486 if (UnsizedMemOp && UnsizedMemOp->isMemUnsized()) { 4487 static const unsigned MopSizes[] = {8, 16, 32, 64, 80, 128, 256, 512}; 4488 for (unsigned Size : MopSizes) { 4489 UnsizedMemOp->Mem.Size = Size; 4490 uint64_t ErrorInfoIgnore; 4491 unsigned LastOpcode = Inst.getOpcode(); 4492 unsigned M = MatchInstruction(Operands, Inst, ErrorInfoIgnore, 4493 MissingFeatures, MatchingInlineAsm, 4494 isParsingIntelSyntax()); 4495 if (Match.empty() || LastOpcode != Inst.getOpcode()) 4496 Match.push_back(M); 4497 4498 // If this returned as a missing feature failure, remember that. 4499 if (Match.back() == Match_MissingFeature) 4500 ErrorInfoMissingFeatures = MissingFeatures; 4501 } 4502 4503 // Restore the size of the unsized memory operand if we modified it. 4504 UnsizedMemOp->Mem.Size = 0; 4505 } 4506 4507 // If we haven't matched anything yet, this is not a basic integer or FPU 4508 // operation. There shouldn't be any ambiguity in our mnemonic table, so try 4509 // matching with the unsized operand. 4510 if (Match.empty()) { 4511 Match.push_back(MatchInstruction( 4512 Operands, Inst, ErrorInfo, MissingFeatures, MatchingInlineAsm, 4513 isParsingIntelSyntax())); 4514 // If this returned as a missing feature failure, remember that. 4515 if (Match.back() == Match_MissingFeature) 4516 ErrorInfoMissingFeatures = MissingFeatures; 4517 } 4518 4519 // Restore the size of the unsized memory operand if we modified it. 4520 if (UnsizedMemOp) 4521 UnsizedMemOp->Mem.Size = 0; 4522 4523 // If it's a bad mnemonic, all results will be the same. 4524 if (Match.back() == Match_MnemonicFail) { 4525 return Error(IDLoc, "invalid instruction mnemonic '" + Mnemonic + "'", 4526 Op.getLocRange(), MatchingInlineAsm); 4527 } 4528 4529 unsigned NumSuccessfulMatches = llvm::count(Match, Match_Success); 4530 4531 // If matching was ambiguous and we had size information from the frontend, 4532 // try again with that. This handles cases like "movxz eax, m8/m16". 4533 if (UnsizedMemOp && NumSuccessfulMatches > 1 && 4534 UnsizedMemOp->getMemFrontendSize()) { 4535 UnsizedMemOp->Mem.Size = UnsizedMemOp->getMemFrontendSize(); 4536 unsigned M = MatchInstruction( 4537 Operands, Inst, ErrorInfo, MissingFeatures, MatchingInlineAsm, 4538 isParsingIntelSyntax()); 4539 if (M == Match_Success) 4540 NumSuccessfulMatches = 1; 4541 4542 // Add a rewrite that encodes the size information we used from the 4543 // frontend. 4544 InstInfo->AsmRewrites->emplace_back( 4545 AOK_SizeDirective, UnsizedMemOp->getStartLoc(), 4546 /*Len=*/0, UnsizedMemOp->getMemFrontendSize()); 4547 } 4548 4549 // If exactly one matched, then we treat that as a successful match (and the 4550 // instruction will already have been filled in correctly, since the failing 4551 // matches won't have modified it). 4552 if (NumSuccessfulMatches == 1) { 4553 if (!MatchingInlineAsm && validateInstruction(Inst, Operands)) 4554 return true; 4555 // Some instructions need post-processing to, for example, tweak which 4556 // encoding is selected. Loop on it while changes happen so the individual 4557 // transformations can chain off each other. 4558 if (!MatchingInlineAsm) 4559 while (processInstruction(Inst, Operands)) 4560 ; 4561 Inst.setLoc(IDLoc); 4562 if (!MatchingInlineAsm) 4563 emitInstruction(Inst, Operands, Out); 4564 Opcode = Inst.getOpcode(); 4565 return false; 4566 } else if (NumSuccessfulMatches > 1) { 4567 assert(UnsizedMemOp && 4568 "multiple matches only possible with unsized memory operands"); 4569 return Error(UnsizedMemOp->getStartLoc(), 4570 "ambiguous operand size for instruction '" + Mnemonic + "\'", 4571 UnsizedMemOp->getLocRange()); 4572 } 4573 4574 // If one instruction matched as unsupported, report this as unsupported. 4575 if (llvm::count(Match, Match_Unsupported) == 1) { 4576 return Error(IDLoc, "unsupported instruction", EmptyRange, 4577 MatchingInlineAsm); 4578 } 4579 4580 // If one instruction matched with a missing feature, report this as a 4581 // missing feature. 4582 if (llvm::count(Match, Match_MissingFeature) == 1) { 4583 ErrorInfo = Match_MissingFeature; 4584 return ErrorMissingFeature(IDLoc, ErrorInfoMissingFeatures, 4585 MatchingInlineAsm); 4586 } 4587 4588 // If one instruction matched with an invalid operand, report this as an 4589 // operand failure. 4590 if (llvm::count(Match, Match_InvalidOperand) == 1) { 4591 return Error(IDLoc, "invalid operand for instruction", EmptyRange, 4592 MatchingInlineAsm); 4593 } 4594 4595 if (llvm::count(Match, Match_InvalidImmUnsignedi4) == 1) { 4596 SMLoc ErrorLoc = ((X86Operand &)*Operands[ErrorInfo]).getStartLoc(); 4597 if (ErrorLoc == SMLoc()) 4598 ErrorLoc = IDLoc; 4599 return Error(ErrorLoc, "immediate must be an integer in range [0, 15]", 4600 EmptyRange, MatchingInlineAsm); 4601 } 4602 4603 // If all of these were an outright failure, report it in a useless way. 4604 return Error(IDLoc, "unknown instruction mnemonic", EmptyRange, 4605 MatchingInlineAsm); 4606 } 4607 4608 bool X86AsmParser::OmitRegisterFromClobberLists(unsigned RegNo) { 4609 return X86MCRegisterClasses[X86::SEGMENT_REGRegClassID].contains(RegNo); 4610 } 4611 4612 bool X86AsmParser::ParseDirective(AsmToken DirectiveID) { 4613 MCAsmParser &Parser = getParser(); 4614 StringRef IDVal = DirectiveID.getIdentifier(); 4615 if (IDVal.startswith(".arch")) 4616 return parseDirectiveArch(); 4617 if (IDVal.startswith(".code")) 4618 return ParseDirectiveCode(IDVal, DirectiveID.getLoc()); 4619 else if (IDVal.startswith(".att_syntax")) { 4620 if (getLexer().isNot(AsmToken::EndOfStatement)) { 4621 if (Parser.getTok().getString() == "prefix") 4622 Parser.Lex(); 4623 else if (Parser.getTok().getString() == "noprefix") 4624 return Error(DirectiveID.getLoc(), "'.att_syntax noprefix' is not " 4625 "supported: registers must have a " 4626 "'%' prefix in .att_syntax"); 4627 } 4628 getParser().setAssemblerDialect(0); 4629 return false; 4630 } else if (IDVal.startswith(".intel_syntax")) { 4631 getParser().setAssemblerDialect(1); 4632 if (getLexer().isNot(AsmToken::EndOfStatement)) { 4633 if (Parser.getTok().getString() == "noprefix") 4634 Parser.Lex(); 4635 else if (Parser.getTok().getString() == "prefix") 4636 return Error(DirectiveID.getLoc(), "'.intel_syntax prefix' is not " 4637 "supported: registers must not have " 4638 "a '%' prefix in .intel_syntax"); 4639 } 4640 return false; 4641 } else if (IDVal == ".nops") 4642 return parseDirectiveNops(DirectiveID.getLoc()); 4643 else if (IDVal == ".even") 4644 return parseDirectiveEven(DirectiveID.getLoc()); 4645 else if (IDVal == ".cv_fpo_proc") 4646 return parseDirectiveFPOProc(DirectiveID.getLoc()); 4647 else if (IDVal == ".cv_fpo_setframe") 4648 return parseDirectiveFPOSetFrame(DirectiveID.getLoc()); 4649 else if (IDVal == ".cv_fpo_pushreg") 4650 return parseDirectiveFPOPushReg(DirectiveID.getLoc()); 4651 else if (IDVal == ".cv_fpo_stackalloc") 4652 return parseDirectiveFPOStackAlloc(DirectiveID.getLoc()); 4653 else if (IDVal == ".cv_fpo_stackalign") 4654 return parseDirectiveFPOStackAlign(DirectiveID.getLoc()); 4655 else if (IDVal == ".cv_fpo_endprologue") 4656 return parseDirectiveFPOEndPrologue(DirectiveID.getLoc()); 4657 else if (IDVal == ".cv_fpo_endproc") 4658 return parseDirectiveFPOEndProc(DirectiveID.getLoc()); 4659 else if (IDVal == ".seh_pushreg" || 4660 (Parser.isParsingMasm() && IDVal.equals_insensitive(".pushreg"))) 4661 return parseDirectiveSEHPushReg(DirectiveID.getLoc()); 4662 else if (IDVal == ".seh_setframe" || 4663 (Parser.isParsingMasm() && IDVal.equals_insensitive(".setframe"))) 4664 return parseDirectiveSEHSetFrame(DirectiveID.getLoc()); 4665 else if (IDVal == ".seh_savereg" || 4666 (Parser.isParsingMasm() && IDVal.equals_insensitive(".savereg"))) 4667 return parseDirectiveSEHSaveReg(DirectiveID.getLoc()); 4668 else if (IDVal == ".seh_savexmm" || 4669 (Parser.isParsingMasm() && IDVal.equals_insensitive(".savexmm128"))) 4670 return parseDirectiveSEHSaveXMM(DirectiveID.getLoc()); 4671 else if (IDVal == ".seh_pushframe" || 4672 (Parser.isParsingMasm() && IDVal.equals_insensitive(".pushframe"))) 4673 return parseDirectiveSEHPushFrame(DirectiveID.getLoc()); 4674 4675 return true; 4676 } 4677 4678 bool X86AsmParser::parseDirectiveArch() { 4679 // Ignore .arch for now. 4680 getParser().parseStringToEndOfStatement(); 4681 return false; 4682 } 4683 4684 /// parseDirectiveNops 4685 /// ::= .nops size[, control] 4686 bool X86AsmParser::parseDirectiveNops(SMLoc L) { 4687 int64_t NumBytes = 0, Control = 0; 4688 SMLoc NumBytesLoc, ControlLoc; 4689 const MCSubtargetInfo& STI = getSTI(); 4690 NumBytesLoc = getTok().getLoc(); 4691 if (getParser().checkForValidSection() || 4692 getParser().parseAbsoluteExpression(NumBytes)) 4693 return true; 4694 4695 if (parseOptionalToken(AsmToken::Comma)) { 4696 ControlLoc = getTok().getLoc(); 4697 if (getParser().parseAbsoluteExpression(Control)) 4698 return true; 4699 } 4700 if (getParser().parseEOL()) 4701 return true; 4702 4703 if (NumBytes <= 0) { 4704 Error(NumBytesLoc, "'.nops' directive with non-positive size"); 4705 return false; 4706 } 4707 4708 if (Control < 0) { 4709 Error(ControlLoc, "'.nops' directive with negative NOP size"); 4710 return false; 4711 } 4712 4713 /// Emit nops 4714 getParser().getStreamer().emitNops(NumBytes, Control, L, STI); 4715 4716 return false; 4717 } 4718 4719 /// parseDirectiveEven 4720 /// ::= .even 4721 bool X86AsmParser::parseDirectiveEven(SMLoc L) { 4722 if (parseEOL()) 4723 return false; 4724 4725 const MCSection *Section = getStreamer().getCurrentSectionOnly(); 4726 if (!Section) { 4727 getStreamer().initSections(false, getSTI()); 4728 Section = getStreamer().getCurrentSectionOnly(); 4729 } 4730 if (Section->useCodeAlign()) 4731 getStreamer().emitCodeAlignment(Align(2), &getSTI(), 0); 4732 else 4733 getStreamer().emitValueToAlignment(Align(2), 0, 1, 0); 4734 return false; 4735 } 4736 4737 /// ParseDirectiveCode 4738 /// ::= .code16 | .code32 | .code64 4739 bool X86AsmParser::ParseDirectiveCode(StringRef IDVal, SMLoc L) { 4740 MCAsmParser &Parser = getParser(); 4741 Code16GCC = false; 4742 if (IDVal == ".code16") { 4743 Parser.Lex(); 4744 if (!is16BitMode()) { 4745 SwitchMode(X86::Is16Bit); 4746 getParser().getStreamer().emitAssemblerFlag(MCAF_Code16); 4747 } 4748 } else if (IDVal == ".code16gcc") { 4749 // .code16gcc parses as if in 32-bit mode, but emits code in 16-bit mode. 4750 Parser.Lex(); 4751 Code16GCC = true; 4752 if (!is16BitMode()) { 4753 SwitchMode(X86::Is16Bit); 4754 getParser().getStreamer().emitAssemblerFlag(MCAF_Code16); 4755 } 4756 } else if (IDVal == ".code32") { 4757 Parser.Lex(); 4758 if (!is32BitMode()) { 4759 SwitchMode(X86::Is32Bit); 4760 getParser().getStreamer().emitAssemblerFlag(MCAF_Code32); 4761 } 4762 } else if (IDVal == ".code64") { 4763 Parser.Lex(); 4764 if (!is64BitMode()) { 4765 SwitchMode(X86::Is64Bit); 4766 getParser().getStreamer().emitAssemblerFlag(MCAF_Code64); 4767 } 4768 } else { 4769 Error(L, "unknown directive " + IDVal); 4770 return false; 4771 } 4772 4773 return false; 4774 } 4775 4776 // .cv_fpo_proc foo 4777 bool X86AsmParser::parseDirectiveFPOProc(SMLoc L) { 4778 MCAsmParser &Parser = getParser(); 4779 StringRef ProcName; 4780 int64_t ParamsSize; 4781 if (Parser.parseIdentifier(ProcName)) 4782 return Parser.TokError("expected symbol name"); 4783 if (Parser.parseIntToken(ParamsSize, "expected parameter byte count")) 4784 return true; 4785 if (!isUIntN(32, ParamsSize)) 4786 return Parser.TokError("parameters size out of range"); 4787 if (parseEOL()) 4788 return true; 4789 MCSymbol *ProcSym = getContext().getOrCreateSymbol(ProcName); 4790 return getTargetStreamer().emitFPOProc(ProcSym, ParamsSize, L); 4791 } 4792 4793 // .cv_fpo_setframe ebp 4794 bool X86AsmParser::parseDirectiveFPOSetFrame(SMLoc L) { 4795 MCRegister Reg; 4796 SMLoc DummyLoc; 4797 if (parseRegister(Reg, DummyLoc, DummyLoc) || parseEOL()) 4798 return true; 4799 return getTargetStreamer().emitFPOSetFrame(Reg, L); 4800 } 4801 4802 // .cv_fpo_pushreg ebx 4803 bool X86AsmParser::parseDirectiveFPOPushReg(SMLoc L) { 4804 MCRegister Reg; 4805 SMLoc DummyLoc; 4806 if (parseRegister(Reg, DummyLoc, DummyLoc) || parseEOL()) 4807 return true; 4808 return getTargetStreamer().emitFPOPushReg(Reg, L); 4809 } 4810 4811 // .cv_fpo_stackalloc 20 4812 bool X86AsmParser::parseDirectiveFPOStackAlloc(SMLoc L) { 4813 MCAsmParser &Parser = getParser(); 4814 int64_t Offset; 4815 if (Parser.parseIntToken(Offset, "expected offset") || parseEOL()) 4816 return true; 4817 return getTargetStreamer().emitFPOStackAlloc(Offset, L); 4818 } 4819 4820 // .cv_fpo_stackalign 8 4821 bool X86AsmParser::parseDirectiveFPOStackAlign(SMLoc L) { 4822 MCAsmParser &Parser = getParser(); 4823 int64_t Offset; 4824 if (Parser.parseIntToken(Offset, "expected offset") || parseEOL()) 4825 return true; 4826 return getTargetStreamer().emitFPOStackAlign(Offset, L); 4827 } 4828 4829 // .cv_fpo_endprologue 4830 bool X86AsmParser::parseDirectiveFPOEndPrologue(SMLoc L) { 4831 MCAsmParser &Parser = getParser(); 4832 if (Parser.parseEOL()) 4833 return true; 4834 return getTargetStreamer().emitFPOEndPrologue(L); 4835 } 4836 4837 // .cv_fpo_endproc 4838 bool X86AsmParser::parseDirectiveFPOEndProc(SMLoc L) { 4839 MCAsmParser &Parser = getParser(); 4840 if (Parser.parseEOL()) 4841 return true; 4842 return getTargetStreamer().emitFPOEndProc(L); 4843 } 4844 4845 bool X86AsmParser::parseSEHRegisterNumber(unsigned RegClassID, 4846 MCRegister &RegNo) { 4847 SMLoc startLoc = getLexer().getLoc(); 4848 const MCRegisterInfo *MRI = getContext().getRegisterInfo(); 4849 4850 // Try parsing the argument as a register first. 4851 if (getLexer().getTok().isNot(AsmToken::Integer)) { 4852 SMLoc endLoc; 4853 if (parseRegister(RegNo, startLoc, endLoc)) 4854 return true; 4855 4856 if (!X86MCRegisterClasses[RegClassID].contains(RegNo)) { 4857 return Error(startLoc, 4858 "register is not supported for use with this directive"); 4859 } 4860 } else { 4861 // Otherwise, an integer number matching the encoding of the desired 4862 // register may appear. 4863 int64_t EncodedReg; 4864 if (getParser().parseAbsoluteExpression(EncodedReg)) 4865 return true; 4866 4867 // The SEH register number is the same as the encoding register number. Map 4868 // from the encoding back to the LLVM register number. 4869 RegNo = 0; 4870 for (MCPhysReg Reg : X86MCRegisterClasses[RegClassID]) { 4871 if (MRI->getEncodingValue(Reg) == EncodedReg) { 4872 RegNo = Reg; 4873 break; 4874 } 4875 } 4876 if (RegNo == 0) { 4877 return Error(startLoc, 4878 "incorrect register number for use with this directive"); 4879 } 4880 } 4881 4882 return false; 4883 } 4884 4885 bool X86AsmParser::parseDirectiveSEHPushReg(SMLoc Loc) { 4886 MCRegister Reg; 4887 if (parseSEHRegisterNumber(X86::GR64RegClassID, Reg)) 4888 return true; 4889 4890 if (getLexer().isNot(AsmToken::EndOfStatement)) 4891 return TokError("expected end of directive"); 4892 4893 getParser().Lex(); 4894 getStreamer().emitWinCFIPushReg(Reg, Loc); 4895 return false; 4896 } 4897 4898 bool X86AsmParser::parseDirectiveSEHSetFrame(SMLoc Loc) { 4899 MCRegister Reg; 4900 int64_t Off; 4901 if (parseSEHRegisterNumber(X86::GR64RegClassID, Reg)) 4902 return true; 4903 if (getLexer().isNot(AsmToken::Comma)) 4904 return TokError("you must specify a stack pointer offset"); 4905 4906 getParser().Lex(); 4907 if (getParser().parseAbsoluteExpression(Off)) 4908 return true; 4909 4910 if (getLexer().isNot(AsmToken::EndOfStatement)) 4911 return TokError("expected end of directive"); 4912 4913 getParser().Lex(); 4914 getStreamer().emitWinCFISetFrame(Reg, Off, Loc); 4915 return false; 4916 } 4917 4918 bool X86AsmParser::parseDirectiveSEHSaveReg(SMLoc Loc) { 4919 MCRegister Reg; 4920 int64_t Off; 4921 if (parseSEHRegisterNumber(X86::GR64RegClassID, Reg)) 4922 return true; 4923 if (getLexer().isNot(AsmToken::Comma)) 4924 return TokError("you must specify an offset on the stack"); 4925 4926 getParser().Lex(); 4927 if (getParser().parseAbsoluteExpression(Off)) 4928 return true; 4929 4930 if (getLexer().isNot(AsmToken::EndOfStatement)) 4931 return TokError("expected end of directive"); 4932 4933 getParser().Lex(); 4934 getStreamer().emitWinCFISaveReg(Reg, Off, Loc); 4935 return false; 4936 } 4937 4938 bool X86AsmParser::parseDirectiveSEHSaveXMM(SMLoc Loc) { 4939 MCRegister Reg; 4940 int64_t Off; 4941 if (parseSEHRegisterNumber(X86::VR128XRegClassID, Reg)) 4942 return true; 4943 if (getLexer().isNot(AsmToken::Comma)) 4944 return TokError("you must specify an offset on the stack"); 4945 4946 getParser().Lex(); 4947 if (getParser().parseAbsoluteExpression(Off)) 4948 return true; 4949 4950 if (getLexer().isNot(AsmToken::EndOfStatement)) 4951 return TokError("expected end of directive"); 4952 4953 getParser().Lex(); 4954 getStreamer().emitWinCFISaveXMM(Reg, Off, Loc); 4955 return false; 4956 } 4957 4958 bool X86AsmParser::parseDirectiveSEHPushFrame(SMLoc Loc) { 4959 bool Code = false; 4960 StringRef CodeID; 4961 if (getLexer().is(AsmToken::At)) { 4962 SMLoc startLoc = getLexer().getLoc(); 4963 getParser().Lex(); 4964 if (!getParser().parseIdentifier(CodeID)) { 4965 if (CodeID != "code") 4966 return Error(startLoc, "expected @code"); 4967 Code = true; 4968 } 4969 } 4970 4971 if (getLexer().isNot(AsmToken::EndOfStatement)) 4972 return TokError("expected end of directive"); 4973 4974 getParser().Lex(); 4975 getStreamer().emitWinCFIPushFrame(Code, Loc); 4976 return false; 4977 } 4978 4979 // Force static initialization. 4980 extern "C" LLVM_EXTERNAL_VISIBILITY void LLVMInitializeX86AsmParser() { 4981 RegisterMCAsmParser<X86AsmParser> X(getTheX86_32Target()); 4982 RegisterMCAsmParser<X86AsmParser> Y(getTheX86_64Target()); 4983 } 4984 4985 #define GET_REGISTER_MATCHER 4986 #define GET_MATCHER_IMPLEMENTATION 4987 #define GET_SUBTARGET_FEATURE_NAME 4988 #include "X86GenAsmMatcher.inc" 4989