11db9f3b2SDimitry Andric //===- X86CompressEVEX.cpp ------------------------------------------------===//
21db9f3b2SDimitry Andric //
31db9f3b2SDimitry Andric // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
41db9f3b2SDimitry Andric // See https://llvm.org/LICENSE.txt for license information.
51db9f3b2SDimitry Andric // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
61db9f3b2SDimitry Andric //
71db9f3b2SDimitry Andric //===----------------------------------------------------------------------===//
81db9f3b2SDimitry Andric //
91db9f3b2SDimitry Andric // This pass compresses instructions from EVEX space to legacy/VEX/EVEX space
101db9f3b2SDimitry Andric // when possible in order to reduce code size or facilitate HW decoding.
111db9f3b2SDimitry Andric //
121db9f3b2SDimitry Andric // Possible compression:
131db9f3b2SDimitry Andric // a. AVX512 instruction (EVEX) -> AVX instruction (VEX)
141db9f3b2SDimitry Andric // b. Promoted instruction (EVEX) -> pre-promotion instruction (legacy/VEX)
151db9f3b2SDimitry Andric // c. NDD (EVEX) -> non-NDD (legacy)
161db9f3b2SDimitry Andric // d. NF_ND (EVEX) -> NF (EVEX)
17*0fca6ea1SDimitry Andric // e. NonNF (EVEX) -> NF (EVEX)
181db9f3b2SDimitry Andric //
191db9f3b2SDimitry Andric // Compression a, b and c can always reduce code size, with some exceptions
201db9f3b2SDimitry Andric // such as promoted 16-bit CRC32 which is as long as the legacy version.
211db9f3b2SDimitry Andric //
221db9f3b2SDimitry Andric // legacy:
231db9f3b2SDimitry Andric // crc32w %si, %eax ## encoding: [0x66,0xf2,0x0f,0x38,0xf1,0xc6]
241db9f3b2SDimitry Andric // promoted:
251db9f3b2SDimitry Andric // crc32w %si, %eax ## encoding: [0x62,0xf4,0x7d,0x08,0xf1,0xc6]
261db9f3b2SDimitry Andric //
271db9f3b2SDimitry Andric // From performance perspective, these should be same (same uops and same EXE
281db9f3b2SDimitry Andric // ports). From a FMV perspective, an older legacy encoding is preferred b/c it
291db9f3b2SDimitry Andric // can execute in more places (broader HW install base). So we will still do
301db9f3b2SDimitry Andric // the compression.
311db9f3b2SDimitry Andric //
321db9f3b2SDimitry Andric // Compression d can help hardware decode (HW may skip reading the NDD
331db9f3b2SDimitry Andric // register) although the instruction length remains unchanged.
34*0fca6ea1SDimitry Andric //
35*0fca6ea1SDimitry Andric // Compression e can help hardware skip updating EFLAGS although the instruction
36*0fca6ea1SDimitry Andric // length remains unchanged.
371db9f3b2SDimitry Andric //===----------------------------------------------------------------------===//
381db9f3b2SDimitry Andric
391db9f3b2SDimitry Andric #include "MCTargetDesc/X86BaseInfo.h"
401db9f3b2SDimitry Andric #include "MCTargetDesc/X86InstComments.h"
411db9f3b2SDimitry Andric #include "X86.h"
421db9f3b2SDimitry Andric #include "X86InstrInfo.h"
431db9f3b2SDimitry Andric #include "X86Subtarget.h"
441db9f3b2SDimitry Andric #include "llvm/ADT/StringRef.h"
451db9f3b2SDimitry Andric #include "llvm/CodeGen/MachineFunction.h"
461db9f3b2SDimitry Andric #include "llvm/CodeGen/MachineFunctionPass.h"
471db9f3b2SDimitry Andric #include "llvm/CodeGen/MachineInstr.h"
481db9f3b2SDimitry Andric #include "llvm/CodeGen/MachineOperand.h"
491db9f3b2SDimitry Andric #include "llvm/MC/MCInstrDesc.h"
501db9f3b2SDimitry Andric #include "llvm/Pass.h"
511db9f3b2SDimitry Andric #include <atomic>
521db9f3b2SDimitry Andric #include <cassert>
531db9f3b2SDimitry Andric #include <cstdint>
541db9f3b2SDimitry Andric
551db9f3b2SDimitry Andric using namespace llvm;
561db9f3b2SDimitry Andric
571db9f3b2SDimitry Andric #define COMP_EVEX_DESC "Compressing EVEX instrs when possible"
581db9f3b2SDimitry Andric #define COMP_EVEX_NAME "x86-compress-evex"
591db9f3b2SDimitry Andric
601db9f3b2SDimitry Andric #define DEBUG_TYPE COMP_EVEX_NAME
611db9f3b2SDimitry Andric
621db9f3b2SDimitry Andric namespace {
63*0fca6ea1SDimitry Andric // Including the generated EVEX compression tables.
64*0fca6ea1SDimitry Andric #define GET_X86_COMPRESS_EVEX_TABLE
65*0fca6ea1SDimitry Andric #include "X86GenInstrMapping.inc"
661db9f3b2SDimitry Andric
671db9f3b2SDimitry Andric class CompressEVEXPass : public MachineFunctionPass {
681db9f3b2SDimitry Andric public:
691db9f3b2SDimitry Andric static char ID;
CompressEVEXPass()701db9f3b2SDimitry Andric CompressEVEXPass() : MachineFunctionPass(ID) {}
getPassName() const711db9f3b2SDimitry Andric StringRef getPassName() const override { return COMP_EVEX_DESC; }
721db9f3b2SDimitry Andric
731db9f3b2SDimitry Andric bool runOnMachineFunction(MachineFunction &MF) override;
741db9f3b2SDimitry Andric
751db9f3b2SDimitry Andric // This pass runs after regalloc and doesn't support VReg operands.
getRequiredProperties() const761db9f3b2SDimitry Andric MachineFunctionProperties getRequiredProperties() const override {
771db9f3b2SDimitry Andric return MachineFunctionProperties().set(
781db9f3b2SDimitry Andric MachineFunctionProperties::Property::NoVRegs);
791db9f3b2SDimitry Andric }
801db9f3b2SDimitry Andric };
811db9f3b2SDimitry Andric
821db9f3b2SDimitry Andric } // end anonymous namespace
831db9f3b2SDimitry Andric
841db9f3b2SDimitry Andric char CompressEVEXPass::ID = 0;
851db9f3b2SDimitry Andric
usesExtendedRegister(const MachineInstr & MI)861db9f3b2SDimitry Andric static bool usesExtendedRegister(const MachineInstr &MI) {
871db9f3b2SDimitry Andric auto isHiRegIdx = [](unsigned Reg) {
881db9f3b2SDimitry Andric // Check for XMM register with indexes between 16 - 31.
891db9f3b2SDimitry Andric if (Reg >= X86::XMM16 && Reg <= X86::XMM31)
901db9f3b2SDimitry Andric return true;
911db9f3b2SDimitry Andric // Check for YMM register with indexes between 16 - 31.
921db9f3b2SDimitry Andric if (Reg >= X86::YMM16 && Reg <= X86::YMM31)
931db9f3b2SDimitry Andric return true;
941db9f3b2SDimitry Andric // Check for GPR with indexes between 16 - 31.
951db9f3b2SDimitry Andric if (X86II::isApxExtendedReg(Reg))
961db9f3b2SDimitry Andric return true;
971db9f3b2SDimitry Andric return false;
981db9f3b2SDimitry Andric };
991db9f3b2SDimitry Andric
1001db9f3b2SDimitry Andric // Check that operands are not ZMM regs or
1011db9f3b2SDimitry Andric // XMM/YMM regs with hi indexes between 16 - 31.
1021db9f3b2SDimitry Andric for (const MachineOperand &MO : MI.explicit_operands()) {
1031db9f3b2SDimitry Andric if (!MO.isReg())
1041db9f3b2SDimitry Andric continue;
1051db9f3b2SDimitry Andric
1061db9f3b2SDimitry Andric Register Reg = MO.getReg();
1071db9f3b2SDimitry Andric assert(!X86II::isZMMReg(Reg) &&
1081db9f3b2SDimitry Andric "ZMM instructions should not be in the EVEX->VEX tables");
1091db9f3b2SDimitry Andric if (isHiRegIdx(Reg))
1101db9f3b2SDimitry Andric return true;
1111db9f3b2SDimitry Andric }
1121db9f3b2SDimitry Andric
1131db9f3b2SDimitry Andric return false;
1141db9f3b2SDimitry Andric }
1151db9f3b2SDimitry Andric
1161db9f3b2SDimitry Andric // Do any custom cleanup needed to finalize the conversion.
performCustomAdjustments(MachineInstr & MI,unsigned NewOpc)1171db9f3b2SDimitry Andric static bool performCustomAdjustments(MachineInstr &MI, unsigned NewOpc) {
1181db9f3b2SDimitry Andric (void)NewOpc;
1191db9f3b2SDimitry Andric unsigned Opc = MI.getOpcode();
1201db9f3b2SDimitry Andric switch (Opc) {
1211db9f3b2SDimitry Andric case X86::VALIGNDZ128rri:
1221db9f3b2SDimitry Andric case X86::VALIGNDZ128rmi:
1231db9f3b2SDimitry Andric case X86::VALIGNQZ128rri:
1241db9f3b2SDimitry Andric case X86::VALIGNQZ128rmi: {
1251db9f3b2SDimitry Andric assert((NewOpc == X86::VPALIGNRrri || NewOpc == X86::VPALIGNRrmi) &&
1261db9f3b2SDimitry Andric "Unexpected new opcode!");
1271db9f3b2SDimitry Andric unsigned Scale =
1281db9f3b2SDimitry Andric (Opc == X86::VALIGNQZ128rri || Opc == X86::VALIGNQZ128rmi) ? 8 : 4;
1291db9f3b2SDimitry Andric MachineOperand &Imm = MI.getOperand(MI.getNumExplicitOperands() - 1);
1301db9f3b2SDimitry Andric Imm.setImm(Imm.getImm() * Scale);
1311db9f3b2SDimitry Andric break;
1321db9f3b2SDimitry Andric }
1331db9f3b2SDimitry Andric case X86::VSHUFF32X4Z256rmi:
1341db9f3b2SDimitry Andric case X86::VSHUFF32X4Z256rri:
1351db9f3b2SDimitry Andric case X86::VSHUFF64X2Z256rmi:
1361db9f3b2SDimitry Andric case X86::VSHUFF64X2Z256rri:
1371db9f3b2SDimitry Andric case X86::VSHUFI32X4Z256rmi:
1381db9f3b2SDimitry Andric case X86::VSHUFI32X4Z256rri:
1391db9f3b2SDimitry Andric case X86::VSHUFI64X2Z256rmi:
1401db9f3b2SDimitry Andric case X86::VSHUFI64X2Z256rri: {
1411db9f3b2SDimitry Andric assert((NewOpc == X86::VPERM2F128rr || NewOpc == X86::VPERM2I128rr ||
1421db9f3b2SDimitry Andric NewOpc == X86::VPERM2F128rm || NewOpc == X86::VPERM2I128rm) &&
1431db9f3b2SDimitry Andric "Unexpected new opcode!");
1441db9f3b2SDimitry Andric MachineOperand &Imm = MI.getOperand(MI.getNumExplicitOperands() - 1);
1451db9f3b2SDimitry Andric int64_t ImmVal = Imm.getImm();
1461db9f3b2SDimitry Andric // Set bit 5, move bit 1 to bit 4, copy bit 0.
1471db9f3b2SDimitry Andric Imm.setImm(0x20 | ((ImmVal & 2) << 3) | (ImmVal & 1));
1481db9f3b2SDimitry Andric break;
1491db9f3b2SDimitry Andric }
1501db9f3b2SDimitry Andric case X86::VRNDSCALEPDZ128rri:
1511db9f3b2SDimitry Andric case X86::VRNDSCALEPDZ128rmi:
1521db9f3b2SDimitry Andric case X86::VRNDSCALEPSZ128rri:
1531db9f3b2SDimitry Andric case X86::VRNDSCALEPSZ128rmi:
1541db9f3b2SDimitry Andric case X86::VRNDSCALEPDZ256rri:
1551db9f3b2SDimitry Andric case X86::VRNDSCALEPDZ256rmi:
1561db9f3b2SDimitry Andric case X86::VRNDSCALEPSZ256rri:
1571db9f3b2SDimitry Andric case X86::VRNDSCALEPSZ256rmi:
1581db9f3b2SDimitry Andric case X86::VRNDSCALESDZr:
1591db9f3b2SDimitry Andric case X86::VRNDSCALESDZm:
1601db9f3b2SDimitry Andric case X86::VRNDSCALESSZr:
1611db9f3b2SDimitry Andric case X86::VRNDSCALESSZm:
1621db9f3b2SDimitry Andric case X86::VRNDSCALESDZr_Int:
1631db9f3b2SDimitry Andric case X86::VRNDSCALESDZm_Int:
1641db9f3b2SDimitry Andric case X86::VRNDSCALESSZr_Int:
1651db9f3b2SDimitry Andric case X86::VRNDSCALESSZm_Int:
1661db9f3b2SDimitry Andric const MachineOperand &Imm = MI.getOperand(MI.getNumExplicitOperands() - 1);
1671db9f3b2SDimitry Andric int64_t ImmVal = Imm.getImm();
1681db9f3b2SDimitry Andric // Ensure that only bits 3:0 of the immediate are used.
1691db9f3b2SDimitry Andric if ((ImmVal & 0xf) != ImmVal)
1701db9f3b2SDimitry Andric return false;
1711db9f3b2SDimitry Andric break;
1721db9f3b2SDimitry Andric }
1731db9f3b2SDimitry Andric
1741db9f3b2SDimitry Andric return true;
1751db9f3b2SDimitry Andric }
1761db9f3b2SDimitry Andric
CompressEVEXImpl(MachineInstr & MI,const X86Subtarget & ST)177*0fca6ea1SDimitry Andric static bool CompressEVEXImpl(MachineInstr &MI, const X86Subtarget &ST) {
178*0fca6ea1SDimitry Andric uint64_t TSFlags = MI.getDesc().TSFlags;
179*0fca6ea1SDimitry Andric
180*0fca6ea1SDimitry Andric // Check for EVEX instructions only.
181*0fca6ea1SDimitry Andric if ((TSFlags & X86II::EncodingMask) != X86II::EVEX)
182*0fca6ea1SDimitry Andric return false;
183*0fca6ea1SDimitry Andric
184*0fca6ea1SDimitry Andric // Instructions with mask or 512-bit vector can't be converted to VEX.
185*0fca6ea1SDimitry Andric if (TSFlags & (X86II::EVEX_K | X86II::EVEX_L2))
186*0fca6ea1SDimitry Andric return false;
187*0fca6ea1SDimitry Andric
188*0fca6ea1SDimitry Andric auto IsRedundantNewDataDest = [&](unsigned &Opc) {
1897a6dacacSDimitry Andric // $rbx = ADD64rr_ND $rbx, $rax / $rbx = ADD64rr_ND $rax, $rbx
1907a6dacacSDimitry Andric // ->
1917a6dacacSDimitry Andric // $rbx = ADD64rr $rbx, $rax
1927a6dacacSDimitry Andric const MCInstrDesc &Desc = MI.getDesc();
1937a6dacacSDimitry Andric Register Reg0 = MI.getOperand(0).getReg();
1947a6dacacSDimitry Andric const MachineOperand &Op1 = MI.getOperand(1);
195*0fca6ea1SDimitry Andric if (!Op1.isReg() || X86::getFirstAddrOperandIdx(MI) == 1 ||
196*0fca6ea1SDimitry Andric X86::isCFCMOVCC(MI.getOpcode()))
1977a6dacacSDimitry Andric return false;
1987a6dacacSDimitry Andric Register Reg1 = Op1.getReg();
1997a6dacacSDimitry Andric if (Reg1 == Reg0)
2007a6dacacSDimitry Andric return true;
2017a6dacacSDimitry Andric
2027a6dacacSDimitry Andric // Op1 and Op2 may be commutable for ND instructions.
2037a6dacacSDimitry Andric if (!Desc.isCommutable() || Desc.getNumOperands() < 3 ||
2047a6dacacSDimitry Andric !MI.getOperand(2).isReg() || MI.getOperand(2).getReg() != Reg0)
2057a6dacacSDimitry Andric return false;
2067a6dacacSDimitry Andric // Opcode may change after commute, e.g. SHRD -> SHLD
2077a6dacacSDimitry Andric ST.getInstrInfo()->commuteInstruction(MI, false, 1, 2);
208*0fca6ea1SDimitry Andric Opc = MI.getOpcode();
2097a6dacacSDimitry Andric return true;
210*0fca6ea1SDimitry Andric };
2111db9f3b2SDimitry Andric
2121db9f3b2SDimitry Andric // EVEX_B has several meanings.
2131db9f3b2SDimitry Andric // AVX512:
2141db9f3b2SDimitry Andric // register form: rounding control or SAE
2151db9f3b2SDimitry Andric // memory form: broadcast
2161db9f3b2SDimitry Andric //
2171db9f3b2SDimitry Andric // APX:
2181db9f3b2SDimitry Andric // MAP4: NDD
2191db9f3b2SDimitry Andric //
2201db9f3b2SDimitry Andric // For AVX512 cases, EVEX prefix is needed in order to carry this information
2211db9f3b2SDimitry Andric // thus preventing the transformation to VEX encoding.
2227a6dacacSDimitry Andric bool IsND = X86II::hasNewDataDest(TSFlags);
223*0fca6ea1SDimitry Andric if (TSFlags & X86II::EVEX_B && !IsND)
2241db9f3b2SDimitry Andric return false;
2251db9f3b2SDimitry Andric unsigned Opc = MI.getOpcode();
226*0fca6ea1SDimitry Andric // MOVBE*rr is special because it has semantic of NDD but not set EVEX_B.
227*0fca6ea1SDimitry Andric bool IsNDLike = IsND || Opc == X86::MOVBE32rr || Opc == X86::MOVBE64rr;
228*0fca6ea1SDimitry Andric bool IsRedundantNDD = IsNDLike ? IsRedundantNewDataDest(Opc) : false;
2291db9f3b2SDimitry Andric
230*0fca6ea1SDimitry Andric auto GetCompressedOpc = [&](unsigned Opc) -> unsigned {
231*0fca6ea1SDimitry Andric ArrayRef<X86TableEntry> Table = ArrayRef(X86CompressEVEXTable);
232*0fca6ea1SDimitry Andric const auto I = llvm::lower_bound(Table, Opc);
233*0fca6ea1SDimitry Andric if (I == Table.end() || I->OldOpc != Opc)
234*0fca6ea1SDimitry Andric return 0;
235*0fca6ea1SDimitry Andric
2367a6dacacSDimitry Andric if (usesExtendedRegister(MI) || !checkPredicate(I->NewOpc, &ST) ||
2371db9f3b2SDimitry Andric !performCustomAdjustments(MI, I->NewOpc))
238*0fca6ea1SDimitry Andric return 0;
239*0fca6ea1SDimitry Andric return I->NewOpc;
240*0fca6ea1SDimitry Andric };
241*0fca6ea1SDimitry Andric // NonNF -> NF only if it's not a compressible NDD instruction and eflags is
242*0fca6ea1SDimitry Andric // dead.
243*0fca6ea1SDimitry Andric unsigned NewOpc = IsRedundantNDD
244*0fca6ea1SDimitry Andric ? X86::getNonNDVariant(Opc)
245*0fca6ea1SDimitry Andric : ((IsNDLike && ST.hasNF() &&
246*0fca6ea1SDimitry Andric MI.registerDefIsDead(X86::EFLAGS, /*TRI=*/nullptr))
247*0fca6ea1SDimitry Andric ? X86::getNFVariant(Opc)
248*0fca6ea1SDimitry Andric : GetCompressedOpc(Opc));
2491db9f3b2SDimitry Andric
250*0fca6ea1SDimitry Andric if (!NewOpc)
251*0fca6ea1SDimitry Andric return false;
252*0fca6ea1SDimitry Andric
253*0fca6ea1SDimitry Andric const MCInstrDesc &NewDesc = ST.getInstrInfo()->get(NewOpc);
2541db9f3b2SDimitry Andric MI.setDesc(NewDesc);
2557a6dacacSDimitry Andric unsigned AsmComment;
2567a6dacacSDimitry Andric switch (NewDesc.TSFlags & X86II::EncodingMask) {
2577a6dacacSDimitry Andric case X86II::LEGACY:
2587a6dacacSDimitry Andric AsmComment = X86::AC_EVEX_2_LEGACY;
2597a6dacacSDimitry Andric break;
2607a6dacacSDimitry Andric case X86II::VEX:
2617a6dacacSDimitry Andric AsmComment = X86::AC_EVEX_2_VEX;
2627a6dacacSDimitry Andric break;
2637a6dacacSDimitry Andric case X86II::EVEX:
2647a6dacacSDimitry Andric AsmComment = X86::AC_EVEX_2_EVEX;
2657a6dacacSDimitry Andric assert(IsND && (NewDesc.TSFlags & X86II::EVEX_NF) &&
2667a6dacacSDimitry Andric "Unknown EVEX2EVEX compression");
2677a6dacacSDimitry Andric break;
2687a6dacacSDimitry Andric default:
2697a6dacacSDimitry Andric llvm_unreachable("Unknown EVEX compression");
2707a6dacacSDimitry Andric }
2711db9f3b2SDimitry Andric MI.setAsmPrinterFlag(AsmComment);
272*0fca6ea1SDimitry Andric if (IsRedundantNDD)
2737a6dacacSDimitry Andric MI.tieOperands(0, 1);
2747a6dacacSDimitry Andric
2751db9f3b2SDimitry Andric return true;
2761db9f3b2SDimitry Andric }
2771db9f3b2SDimitry Andric
runOnMachineFunction(MachineFunction & MF)2781db9f3b2SDimitry Andric bool CompressEVEXPass::runOnMachineFunction(MachineFunction &MF) {
2791db9f3b2SDimitry Andric #ifndef NDEBUG
2801db9f3b2SDimitry Andric // Make sure the tables are sorted.
2811db9f3b2SDimitry Andric static std::atomic<bool> TableChecked(false);
2821db9f3b2SDimitry Andric if (!TableChecked.load(std::memory_order_relaxed)) {
2831db9f3b2SDimitry Andric assert(llvm::is_sorted(X86CompressEVEXTable) &&
2841db9f3b2SDimitry Andric "X86CompressEVEXTable is not sorted!");
2851db9f3b2SDimitry Andric TableChecked.store(true, std::memory_order_relaxed);
2861db9f3b2SDimitry Andric }
2871db9f3b2SDimitry Andric #endif
2881db9f3b2SDimitry Andric const X86Subtarget &ST = MF.getSubtarget<X86Subtarget>();
2891db9f3b2SDimitry Andric if (!ST.hasAVX512() && !ST.hasEGPR() && !ST.hasNDD())
2901db9f3b2SDimitry Andric return false;
2911db9f3b2SDimitry Andric
2921db9f3b2SDimitry Andric bool Changed = false;
2931db9f3b2SDimitry Andric
2941db9f3b2SDimitry Andric for (MachineBasicBlock &MBB : MF) {
2951db9f3b2SDimitry Andric // Traverse the basic block.
2961db9f3b2SDimitry Andric for (MachineInstr &MI : MBB)
2971db9f3b2SDimitry Andric Changed |= CompressEVEXImpl(MI, ST);
2981db9f3b2SDimitry Andric }
2991db9f3b2SDimitry Andric
3001db9f3b2SDimitry Andric return Changed;
3011db9f3b2SDimitry Andric }
3021db9f3b2SDimitry Andric
INITIALIZE_PASS(CompressEVEXPass,COMP_EVEX_NAME,COMP_EVEX_DESC,false,false)3031db9f3b2SDimitry Andric INITIALIZE_PASS(CompressEVEXPass, COMP_EVEX_NAME, COMP_EVEX_DESC, false, false)
3041db9f3b2SDimitry Andric
3051db9f3b2SDimitry Andric FunctionPass *llvm::createX86CompressEVEXPass() {
3061db9f3b2SDimitry Andric return new CompressEVEXPass();
3071db9f3b2SDimitry Andric }
308