xref: /freebsd/contrib/llvm-project/llvm/lib/CodeGen/StackMaps.cpp (revision c2de0116c80176829406289da3b79b6e70855ea4)
1  //===- StackMaps.cpp ------------------------------------------------------===//
2  //
3  // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4  // See https://llvm.org/LICENSE.txt for license information.
5  // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6  //
7  //===----------------------------------------------------------------------===//
8  
9  #include "llvm/CodeGen/StackMaps.h"
10  #include "llvm/ADT/DenseMapInfo.h"
11  #include "llvm/ADT/STLExtras.h"
12  #include "llvm/ADT/Twine.h"
13  #include "llvm/CodeGen/AsmPrinter.h"
14  #include "llvm/CodeGen/MachineFrameInfo.h"
15  #include "llvm/CodeGen/MachineFunction.h"
16  #include "llvm/CodeGen/MachineInstr.h"
17  #include "llvm/CodeGen/MachineOperand.h"
18  #include "llvm/CodeGen/TargetOpcodes.h"
19  #include "llvm/CodeGen/TargetRegisterInfo.h"
20  #include "llvm/CodeGen/TargetSubtargetInfo.h"
21  #include "llvm/IR/DataLayout.h"
22  #include "llvm/MC/MCContext.h"
23  #include "llvm/MC/MCExpr.h"
24  #include "llvm/MC/MCObjectFileInfo.h"
25  #include "llvm/MC/MCRegisterInfo.h"
26  #include "llvm/MC/MCStreamer.h"
27  #include "llvm/Support/CommandLine.h"
28  #include "llvm/Support/Debug.h"
29  #include "llvm/Support/ErrorHandling.h"
30  #include "llvm/Support/MathExtras.h"
31  #include "llvm/Support/raw_ostream.h"
32  #include <algorithm>
33  #include <cassert>
34  #include <cstdint>
35  #include <iterator>
36  #include <utility>
37  
38  using namespace llvm;
39  
40  #define DEBUG_TYPE "stackmaps"
41  
42  static cl::opt<int> StackMapVersion(
43      "stackmap-version", cl::init(3), cl::Hidden,
44      cl::desc("Specify the stackmap encoding version (default = 3)"));
45  
46  const char *StackMaps::WSMP = "Stack Maps: ";
47  
48  static uint64_t getConstMetaVal(const MachineInstr &MI, unsigned Idx) {
49    assert(MI.getOperand(Idx).isImm() &&
50           MI.getOperand(Idx).getImm() == StackMaps::ConstantOp);
51    const auto &MO = MI.getOperand(Idx + 1);
52    assert(MO.isImm());
53    return MO.getImm();
54  }
55  
56  StackMapOpers::StackMapOpers(const MachineInstr *MI)
57    : MI(MI) {
58    assert(getVarIdx() <= MI->getNumOperands() &&
59           "invalid stackmap definition");
60  }
61  
62  PatchPointOpers::PatchPointOpers(const MachineInstr *MI)
63      : MI(MI), HasDef(MI->getOperand(0).isReg() && MI->getOperand(0).isDef() &&
64                       !MI->getOperand(0).isImplicit()) {
65  #ifndef NDEBUG
66    unsigned CheckStartIdx = 0, e = MI->getNumOperands();
67    while (CheckStartIdx < e && MI->getOperand(CheckStartIdx).isReg() &&
68           MI->getOperand(CheckStartIdx).isDef() &&
69           !MI->getOperand(CheckStartIdx).isImplicit())
70      ++CheckStartIdx;
71  
72    assert(getMetaIdx() == CheckStartIdx &&
73           "Unexpected additional definition in Patchpoint intrinsic.");
74  #endif
75  }
76  
77  unsigned PatchPointOpers::getNextScratchIdx(unsigned StartIdx) const {
78    if (!StartIdx)
79      StartIdx = getVarIdx();
80  
81    // Find the next scratch register (implicit def and early clobber)
82    unsigned ScratchIdx = StartIdx, e = MI->getNumOperands();
83    while (ScratchIdx < e &&
84           !(MI->getOperand(ScratchIdx).isReg() &&
85             MI->getOperand(ScratchIdx).isDef() &&
86             MI->getOperand(ScratchIdx).isImplicit() &&
87             MI->getOperand(ScratchIdx).isEarlyClobber()))
88      ++ScratchIdx;
89  
90    assert(ScratchIdx != e && "No scratch register available");
91    return ScratchIdx;
92  }
93  
94  unsigned StatepointOpers::getNumGcMapEntriesIdx() {
95    // Take index of num of allocas and skip all allocas records.
96    unsigned CurIdx = getNumAllocaIdx();
97    unsigned NumAllocas = getConstMetaVal(*MI, CurIdx - 1);
98    CurIdx++;
99    while (NumAllocas--)
100      CurIdx = StackMaps::getNextMetaArgIdx(MI, CurIdx);
101    return CurIdx + 1; // skip <StackMaps::ConstantOp>
102  }
103  
104  unsigned StatepointOpers::getNumAllocaIdx() {
105    // Take index of num of gc ptrs and skip all gc ptr records.
106    unsigned CurIdx = getNumGCPtrIdx();
107    unsigned NumGCPtrs = getConstMetaVal(*MI, CurIdx - 1);
108    CurIdx++;
109    while (NumGCPtrs--)
110      CurIdx = StackMaps::getNextMetaArgIdx(MI, CurIdx);
111    return CurIdx + 1; // skip <StackMaps::ConstantOp>
112  }
113  
114  unsigned StatepointOpers::getNumGCPtrIdx() {
115    // Take index of num of deopt args and skip all deopt records.
116    unsigned CurIdx = getNumDeoptArgsIdx();
117    unsigned NumDeoptArgs = getConstMetaVal(*MI, CurIdx - 1);
118    CurIdx++;
119    while (NumDeoptArgs--) {
120      CurIdx = StackMaps::getNextMetaArgIdx(MI, CurIdx);
121    }
122    return CurIdx + 1; // skip <StackMaps::ConstantOp>
123  }
124  
125  int StatepointOpers::getFirstGCPtrIdx() {
126    unsigned NumGCPtrsIdx = getNumGCPtrIdx();
127    unsigned NumGCPtrs = getConstMetaVal(*MI, NumGCPtrsIdx - 1);
128    if (NumGCPtrs == 0)
129      return -1;
130    ++NumGCPtrsIdx; // skip <num gc ptrs>
131    assert(NumGCPtrsIdx < MI->getNumOperands());
132    return (int)NumGCPtrsIdx;
133  }
134  
135  unsigned StatepointOpers::getGCPointerMap(
136      SmallVectorImpl<std::pair<unsigned, unsigned>> &GCMap) {
137    unsigned CurIdx = getNumGcMapEntriesIdx();
138    unsigned GCMapSize = getConstMetaVal(*MI, CurIdx - 1);
139    CurIdx++;
140    for (unsigned N = 0; N < GCMapSize; ++N) {
141      unsigned B = MI->getOperand(CurIdx++).getImm();
142      unsigned D = MI->getOperand(CurIdx++).getImm();
143      GCMap.push_back(std::make_pair(B, D));
144    }
145  
146    return GCMapSize;
147  }
148  
149  bool StatepointOpers::isFoldableReg(Register Reg) const {
150    unsigned FoldableAreaStart = getVarIdx();
151    for (const MachineOperand &MO : MI->uses()) {
152      if (MO.getOperandNo() >= FoldableAreaStart)
153        break;
154      if (MO.isReg() && MO.getReg() == Reg)
155        return false;
156    }
157    return true;
158  }
159  
160  bool StatepointOpers::isFoldableReg(const MachineInstr *MI, Register Reg) {
161    if (MI->getOpcode() != TargetOpcode::STATEPOINT)
162      return false;
163    return StatepointOpers(MI).isFoldableReg(Reg);
164  }
165  
166  StackMaps::StackMaps(AsmPrinter &AP) : AP(AP) {
167    if (StackMapVersion != 3)
168      llvm_unreachable("Unsupported stackmap version!");
169  }
170  
171  unsigned StackMaps::getNextMetaArgIdx(const MachineInstr *MI, unsigned CurIdx) {
172    assert(CurIdx < MI->getNumOperands() && "Bad meta arg index");
173    const auto &MO = MI->getOperand(CurIdx);
174    if (MO.isImm()) {
175      switch (MO.getImm()) {
176      default:
177        llvm_unreachable("Unrecognized operand type.");
178      case StackMaps::DirectMemRefOp:
179        CurIdx += 2;
180        break;
181      case StackMaps::IndirectMemRefOp:
182        CurIdx += 3;
183        break;
184      case StackMaps::ConstantOp:
185        ++CurIdx;
186        break;
187      }
188    }
189    ++CurIdx;
190    assert(CurIdx < MI->getNumOperands() && "points past operand list");
191    return CurIdx;
192  }
193  
194  /// Go up the super-register chain until we hit a valid dwarf register number.
195  static unsigned getDwarfRegNum(unsigned Reg, const TargetRegisterInfo *TRI) {
196    int RegNum;
197    for (MCPhysReg SR : TRI->superregs_inclusive(Reg)) {
198      RegNum = TRI->getDwarfRegNum(SR, false);
199      if (RegNum >= 0)
200        break;
201    }
202  
203    assert(RegNum >= 0 && isUInt<16>(RegNum) && "Invalid Dwarf register number.");
204    return (unsigned)RegNum;
205  }
206  
207  MachineInstr::const_mop_iterator
208  StackMaps::parseOperand(MachineInstr::const_mop_iterator MOI,
209                          MachineInstr::const_mop_iterator MOE, LocationVec &Locs,
210                          LiveOutVec &LiveOuts) {
211    const TargetRegisterInfo *TRI = AP.MF->getSubtarget().getRegisterInfo();
212    if (MOI->isImm()) {
213      switch (MOI->getImm()) {
214      default:
215        llvm_unreachable("Unrecognized operand type.");
216      case StackMaps::DirectMemRefOp: {
217        auto &DL = AP.MF->getDataLayout();
218  
219        unsigned Size = DL.getPointerSizeInBits();
220        assert((Size % 8) == 0 && "Need pointer size in bytes.");
221        Size /= 8;
222        Register Reg = (++MOI)->getReg();
223        int64_t Imm = (++MOI)->getImm();
224        Locs.emplace_back(StackMaps::Location::Direct, Size,
225                          getDwarfRegNum(Reg, TRI), Imm);
226        break;
227      }
228      case StackMaps::IndirectMemRefOp: {
229        int64_t Size = (++MOI)->getImm();
230        assert(Size > 0 && "Need a valid size for indirect memory locations.");
231        Register Reg = (++MOI)->getReg();
232        int64_t Imm = (++MOI)->getImm();
233        Locs.emplace_back(StackMaps::Location::Indirect, Size,
234                          getDwarfRegNum(Reg, TRI), Imm);
235        break;
236      }
237      case StackMaps::ConstantOp: {
238        ++MOI;
239        assert(MOI->isImm() && "Expected constant operand.");
240        int64_t Imm = MOI->getImm();
241        if (isInt<32>(Imm)) {
242          Locs.emplace_back(Location::Constant, sizeof(int64_t), 0, Imm);
243        } else {
244          // ConstPool is intentionally a MapVector of 'uint64_t's (as
245          // opposed to 'int64_t's).  We should never be in a situation
246          // where we have to insert either the tombstone or the empty
247          // keys into a map, and for a DenseMap<uint64_t, T> these are
248          // (uint64_t)0 and (uint64_t)-1.  They can be and are
249          // represented using 32 bit integers.
250          assert((uint64_t)Imm != DenseMapInfo<uint64_t>::getEmptyKey() &&
251                 (uint64_t)Imm != DenseMapInfo<uint64_t>::getTombstoneKey() &&
252                 "empty and tombstone keys should fit in 32 bits!");
253          auto Result = ConstPool.insert(std::make_pair(Imm, Imm));
254          Locs.emplace_back(Location::ConstantIndex, sizeof(int64_t), 0,
255                            Result.first - ConstPool.begin());
256        }
257        break;
258      }
259      }
260      return ++MOI;
261    }
262  
263    // The physical register number will ultimately be encoded as a DWARF regno.
264    // The stack map also records the size of a spill slot that can hold the
265    // register content. (The runtime can track the actual size of the data type
266    // if it needs to.)
267    if (MOI->isReg()) {
268      // Skip implicit registers (this includes our scratch registers)
269      if (MOI->isImplicit())
270        return ++MOI;
271  
272      if (MOI->isUndef()) {
273        // Record `undef` register as constant. Use same value as ISel uses.
274        Locs.emplace_back(Location::Constant, sizeof(int64_t), 0, 0xFEFEFEFE);
275        return ++MOI;
276      }
277  
278      assert(MOI->getReg().isPhysical() &&
279             "Virtreg operands should have been rewritten before now.");
280      const TargetRegisterClass *RC = TRI->getMinimalPhysRegClass(MOI->getReg());
281      assert(!MOI->getSubReg() && "Physical subreg still around.");
282  
283      unsigned Offset = 0;
284      unsigned DwarfRegNum = getDwarfRegNum(MOI->getReg(), TRI);
285      unsigned LLVMRegNum = *TRI->getLLVMRegNum(DwarfRegNum, false);
286      unsigned SubRegIdx = TRI->getSubRegIndex(LLVMRegNum, MOI->getReg());
287      if (SubRegIdx)
288        Offset = TRI->getSubRegIdxOffset(SubRegIdx);
289  
290      Locs.emplace_back(Location::Register, TRI->getSpillSize(*RC),
291                        DwarfRegNum, Offset);
292      return ++MOI;
293    }
294  
295    if (MOI->isRegLiveOut())
296      LiveOuts = parseRegisterLiveOutMask(MOI->getRegLiveOut());
297  
298    return ++MOI;
299  }
300  
301  void StackMaps::print(raw_ostream &OS) {
302    const TargetRegisterInfo *TRI =
303        AP.MF ? AP.MF->getSubtarget().getRegisterInfo() : nullptr;
304    OS << WSMP << "callsites:\n";
305    for (const auto &CSI : CSInfos) {
306      const LocationVec &CSLocs = CSI.Locations;
307      const LiveOutVec &LiveOuts = CSI.LiveOuts;
308  
309      OS << WSMP << "callsite " << CSI.ID << "\n";
310      OS << WSMP << "  has " << CSLocs.size() << " locations\n";
311  
312      unsigned Idx = 0;
313      for (const auto &Loc : CSLocs) {
314        OS << WSMP << "\t\tLoc " << Idx << ": ";
315        switch (Loc.Type) {
316        case Location::Unprocessed:
317          OS << "<Unprocessed operand>";
318          break;
319        case Location::Register:
320          OS << "Register ";
321          if (TRI)
322            OS << printReg(Loc.Reg, TRI);
323          else
324            OS << Loc.Reg;
325          break;
326        case Location::Direct:
327          OS << "Direct ";
328          if (TRI)
329            OS << printReg(Loc.Reg, TRI);
330          else
331            OS << Loc.Reg;
332          if (Loc.Offset)
333            OS << " + " << Loc.Offset;
334          break;
335        case Location::Indirect:
336          OS << "Indirect ";
337          if (TRI)
338            OS << printReg(Loc.Reg, TRI);
339          else
340            OS << Loc.Reg;
341          OS << "+" << Loc.Offset;
342          break;
343        case Location::Constant:
344          OS << "Constant " << Loc.Offset;
345          break;
346        case Location::ConstantIndex:
347          OS << "Constant Index " << Loc.Offset;
348          break;
349        }
350        OS << "\t[encoding: .byte " << Loc.Type << ", .byte 0"
351           << ", .short " << Loc.Size << ", .short " << Loc.Reg << ", .short 0"
352           << ", .int " << Loc.Offset << "]\n";
353        Idx++;
354      }
355  
356      OS << WSMP << "\thas " << LiveOuts.size() << " live-out registers\n";
357  
358      Idx = 0;
359      for (const auto &LO : LiveOuts) {
360        OS << WSMP << "\t\tLO " << Idx << ": ";
361        if (TRI)
362          OS << printReg(LO.Reg, TRI);
363        else
364          OS << LO.Reg;
365        OS << "\t[encoding: .short " << LO.DwarfRegNum << ", .byte 0, .byte "
366           << LO.Size << "]\n";
367        Idx++;
368      }
369    }
370  }
371  
372  /// Create a live-out register record for the given register Reg.
373  StackMaps::LiveOutReg
374  StackMaps::createLiveOutReg(unsigned Reg, const TargetRegisterInfo *TRI) const {
375    unsigned DwarfRegNum = getDwarfRegNum(Reg, TRI);
376    unsigned Size = TRI->getSpillSize(*TRI->getMinimalPhysRegClass(Reg));
377    return LiveOutReg(Reg, DwarfRegNum, Size);
378  }
379  
380  /// Parse the register live-out mask and return a vector of live-out registers
381  /// that need to be recorded in the stackmap.
382  StackMaps::LiveOutVec
383  StackMaps::parseRegisterLiveOutMask(const uint32_t *Mask) const {
384    assert(Mask && "No register mask specified");
385    const TargetRegisterInfo *TRI = AP.MF->getSubtarget().getRegisterInfo();
386    LiveOutVec LiveOuts;
387  
388    // Create a LiveOutReg for each bit that is set in the register mask.
389    for (unsigned Reg = 0, NumRegs = TRI->getNumRegs(); Reg != NumRegs; ++Reg)
390      if ((Mask[Reg / 32] >> (Reg % 32)) & 1)
391        LiveOuts.push_back(createLiveOutReg(Reg, TRI));
392  
393    // We don't need to keep track of a register if its super-register is already
394    // in the list. Merge entries that refer to the same dwarf register and use
395    // the maximum size that needs to be spilled.
396  
397    llvm::sort(LiveOuts, [](const LiveOutReg &LHS, const LiveOutReg &RHS) {
398      // Only sort by the dwarf register number.
399      return LHS.DwarfRegNum < RHS.DwarfRegNum;
400    });
401  
402    for (auto I = LiveOuts.begin(), E = LiveOuts.end(); I != E; ++I) {
403      for (auto *II = std::next(I); II != E; ++II) {
404        if (I->DwarfRegNum != II->DwarfRegNum) {
405          // Skip all the now invalid entries.
406          I = --II;
407          break;
408        }
409        I->Size = std::max(I->Size, II->Size);
410        if (I->Reg && TRI->isSuperRegister(I->Reg, II->Reg))
411          I->Reg = II->Reg;
412        II->Reg = 0; // mark for deletion.
413      }
414    }
415  
416    llvm::erase_if(LiveOuts, [](const LiveOutReg &LO) { return LO.Reg == 0; });
417  
418    return LiveOuts;
419  }
420  
421  // See statepoint MI format description in StatepointOpers' class comment
422  // in include/llvm/CodeGen/StackMaps.h
423  void StackMaps::parseStatepointOpers(const MachineInstr &MI,
424                                       MachineInstr::const_mop_iterator MOI,
425                                       MachineInstr::const_mop_iterator MOE,
426                                       LocationVec &Locations,
427                                       LiveOutVec &LiveOuts) {
428    LLVM_DEBUG(dbgs() << "record statepoint : " << MI << "\n");
429    StatepointOpers SO(&MI);
430    MOI = parseOperand(MOI, MOE, Locations, LiveOuts); // CC
431    MOI = parseOperand(MOI, MOE, Locations, LiveOuts); // Flags
432    MOI = parseOperand(MOI, MOE, Locations, LiveOuts); // Num Deopts
433  
434    // Record Deopt Args.
435    unsigned NumDeoptArgs = Locations.back().Offset;
436    assert(Locations.back().Type == Location::Constant);
437    assert(NumDeoptArgs == SO.getNumDeoptArgs());
438  
439    while (NumDeoptArgs--)
440      MOI = parseOperand(MOI, MOE, Locations, LiveOuts);
441  
442    // Record gc base/derived pairs
443    assert(MOI->isImm() && MOI->getImm() == StackMaps::ConstantOp);
444    ++MOI;
445    assert(MOI->isImm());
446    unsigned NumGCPointers = MOI->getImm();
447    ++MOI;
448    if (NumGCPointers) {
449      // Map logical index of GC ptr to MI operand index.
450      SmallVector<unsigned, 8> GCPtrIndices;
451      unsigned GCPtrIdx = (unsigned)SO.getFirstGCPtrIdx();
452      assert((int)GCPtrIdx != -1);
453      assert(MOI - MI.operands_begin() == GCPtrIdx + 0LL);
454      while (NumGCPointers--) {
455        GCPtrIndices.push_back(GCPtrIdx);
456        GCPtrIdx = StackMaps::getNextMetaArgIdx(&MI, GCPtrIdx);
457      }
458  
459      SmallVector<std::pair<unsigned, unsigned>, 8> GCPairs;
460      unsigned NumGCPairs = SO.getGCPointerMap(GCPairs);
461      (void)NumGCPairs;
462      LLVM_DEBUG(dbgs() << "NumGCPairs = " << NumGCPairs << "\n");
463  
464      auto MOB = MI.operands_begin();
465      for (auto &P : GCPairs) {
466        assert(P.first < GCPtrIndices.size() && "base pointer index not found");
467        assert(P.second < GCPtrIndices.size() &&
468               "derived pointer index not found");
469        unsigned BaseIdx = GCPtrIndices[P.first];
470        unsigned DerivedIdx = GCPtrIndices[P.second];
471        LLVM_DEBUG(dbgs() << "Base : " << BaseIdx << " Derived : " << DerivedIdx
472                          << "\n");
473        (void)parseOperand(MOB + BaseIdx, MOE, Locations, LiveOuts);
474        (void)parseOperand(MOB + DerivedIdx, MOE, Locations, LiveOuts);
475      }
476  
477      MOI = MOB + GCPtrIdx;
478    }
479  
480    // Record gc allocas
481    assert(MOI < MOE);
482    assert(MOI->isImm() && MOI->getImm() == StackMaps::ConstantOp);
483    ++MOI;
484    unsigned NumAllocas = MOI->getImm();
485    ++MOI;
486    while (NumAllocas--) {
487      MOI = parseOperand(MOI, MOE, Locations, LiveOuts);
488      assert(MOI < MOE);
489    }
490  }
491  
492  void StackMaps::recordStackMapOpers(const MCSymbol &MILabel,
493                                      const MachineInstr &MI, uint64_t ID,
494                                      MachineInstr::const_mop_iterator MOI,
495                                      MachineInstr::const_mop_iterator MOE,
496                                      bool recordResult) {
497    MCContext &OutContext = AP.OutStreamer->getContext();
498  
499    LocationVec Locations;
500    LiveOutVec LiveOuts;
501  
502    if (recordResult) {
503      assert(PatchPointOpers(&MI).hasDef() && "Stackmap has no return value.");
504      parseOperand(MI.operands_begin(), std::next(MI.operands_begin()), Locations,
505                   LiveOuts);
506    }
507  
508    // Parse operands.
509    if (MI.getOpcode() == TargetOpcode::STATEPOINT)
510      parseStatepointOpers(MI, MOI, MOE, Locations, LiveOuts);
511    else
512      while (MOI != MOE)
513        MOI = parseOperand(MOI, MOE, Locations, LiveOuts);
514  
515    // Create an expression to calculate the offset of the callsite from function
516    // entry.
517    const MCExpr *CSOffsetExpr = MCBinaryExpr::createSub(
518        MCSymbolRefExpr::create(&MILabel, OutContext),
519        MCSymbolRefExpr::create(AP.CurrentFnSymForSize, OutContext), OutContext);
520  
521    CSInfos.emplace_back(CSOffsetExpr, ID, std::move(Locations),
522                         std::move(LiveOuts));
523  
524    // Record the stack size of the current function and update callsite count.
525    const MachineFrameInfo &MFI = AP.MF->getFrameInfo();
526    const TargetRegisterInfo *RegInfo = AP.MF->getSubtarget().getRegisterInfo();
527    bool HasDynamicFrameSize =
528        MFI.hasVarSizedObjects() || RegInfo->hasStackRealignment(*(AP.MF));
529    uint64_t FrameSize = HasDynamicFrameSize ? UINT64_MAX : MFI.getStackSize();
530  
531    auto CurrentIt = FnInfos.find(AP.CurrentFnSym);
532    if (CurrentIt != FnInfos.end())
533      CurrentIt->second.RecordCount++;
534    else
535      FnInfos.insert(std::make_pair(AP.CurrentFnSym, FunctionInfo(FrameSize)));
536  }
537  
538  void StackMaps::recordStackMap(const MCSymbol &L, const MachineInstr &MI) {
539    assert(MI.getOpcode() == TargetOpcode::STACKMAP && "expected stackmap");
540  
541    StackMapOpers opers(&MI);
542    const int64_t ID = MI.getOperand(PatchPointOpers::IDPos).getImm();
543    recordStackMapOpers(L, MI, ID, std::next(MI.operands_begin(),
544                                             opers.getVarIdx()),
545                        MI.operands_end());
546  }
547  
548  void StackMaps::recordPatchPoint(const MCSymbol &L, const MachineInstr &MI) {
549    assert(MI.getOpcode() == TargetOpcode::PATCHPOINT && "expected patchpoint");
550  
551    PatchPointOpers opers(&MI);
552    const int64_t ID = opers.getID();
553    auto MOI = std::next(MI.operands_begin(), opers.getStackMapStartIdx());
554    recordStackMapOpers(L, MI, ID, MOI, MI.operands_end(),
555                        opers.isAnyReg() && opers.hasDef());
556  
557  #ifndef NDEBUG
558    // verify anyregcc
559    auto &Locations = CSInfos.back().Locations;
560    if (opers.isAnyReg()) {
561      unsigned NArgs = opers.getNumCallArgs();
562      for (unsigned i = 0, e = (opers.hasDef() ? NArgs + 1 : NArgs); i != e; ++i)
563        assert(Locations[i].Type == Location::Register &&
564               "anyreg arg must be in reg.");
565    }
566  #endif
567  }
568  
569  void StackMaps::recordStatepoint(const MCSymbol &L, const MachineInstr &MI) {
570    assert(MI.getOpcode() == TargetOpcode::STATEPOINT && "expected statepoint");
571  
572    StatepointOpers opers(&MI);
573    const unsigned StartIdx = opers.getVarIdx();
574    recordStackMapOpers(L, MI, opers.getID(), MI.operands_begin() + StartIdx,
575                        MI.operands_end(), false);
576  }
577  
578  /// Emit the stackmap header.
579  ///
580  /// Header {
581  ///   uint8  : Stack Map Version (currently 3)
582  ///   uint8  : Reserved (expected to be 0)
583  ///   uint16 : Reserved (expected to be 0)
584  /// }
585  /// uint32 : NumFunctions
586  /// uint32 : NumConstants
587  /// uint32 : NumRecords
588  void StackMaps::emitStackmapHeader(MCStreamer &OS) {
589    // Header.
590    OS.emitIntValue(StackMapVersion, 1); // Version.
591    OS.emitIntValue(0, 1);               // Reserved.
592    OS.emitInt16(0);                     // Reserved.
593  
594    // Num functions.
595    LLVM_DEBUG(dbgs() << WSMP << "#functions = " << FnInfos.size() << '\n');
596    OS.emitInt32(FnInfos.size());
597    // Num constants.
598    LLVM_DEBUG(dbgs() << WSMP << "#constants = " << ConstPool.size() << '\n');
599    OS.emitInt32(ConstPool.size());
600    // Num callsites.
601    LLVM_DEBUG(dbgs() << WSMP << "#callsites = " << CSInfos.size() << '\n');
602    OS.emitInt32(CSInfos.size());
603  }
604  
605  /// Emit the function frame record for each function.
606  ///
607  /// StkSizeRecord[NumFunctions] {
608  ///   uint64 : Function Address
609  ///   uint64 : Stack Size
610  ///   uint64 : Record Count
611  /// }
612  void StackMaps::emitFunctionFrameRecords(MCStreamer &OS) {
613    // Function Frame records.
614    LLVM_DEBUG(dbgs() << WSMP << "functions:\n");
615    for (auto const &FR : FnInfos) {
616      LLVM_DEBUG(dbgs() << WSMP << "function addr: " << FR.first
617                        << " frame size: " << FR.second.StackSize
618                        << " callsite count: " << FR.second.RecordCount << '\n');
619      OS.emitSymbolValue(FR.first, 8);
620      OS.emitIntValue(FR.second.StackSize, 8);
621      OS.emitIntValue(FR.second.RecordCount, 8);
622    }
623  }
624  
625  /// Emit the constant pool.
626  ///
627  /// int64  : Constants[NumConstants]
628  void StackMaps::emitConstantPoolEntries(MCStreamer &OS) {
629    // Constant pool entries.
630    LLVM_DEBUG(dbgs() << WSMP << "constants:\n");
631    for (const auto &ConstEntry : ConstPool) {
632      LLVM_DEBUG(dbgs() << WSMP << ConstEntry.second << '\n');
633      OS.emitIntValue(ConstEntry.second, 8);
634    }
635  }
636  
637  /// Emit the callsite info for each callsite.
638  ///
639  /// StkMapRecord[NumRecords] {
640  ///   uint64 : PatchPoint ID
641  ///   uint32 : Instruction Offset
642  ///   uint16 : Reserved (record flags)
643  ///   uint16 : NumLocations
644  ///   Location[NumLocations] {
645  ///     uint8  : Register | Direct | Indirect | Constant | ConstantIndex
646  ///     uint8  : Size in Bytes
647  ///     uint16 : Dwarf RegNum
648  ///     int32  : Offset
649  ///   }
650  ///   uint16 : Padding
651  ///   uint16 : NumLiveOuts
652  ///   LiveOuts[NumLiveOuts] {
653  ///     uint16 : Dwarf RegNum
654  ///     uint8  : Reserved
655  ///     uint8  : Size in Bytes
656  ///   }
657  ///   uint32 : Padding (only if required to align to 8 byte)
658  /// }
659  ///
660  /// Location Encoding, Type, Value:
661  ///   0x1, Register, Reg                 (value in register)
662  ///   0x2, Direct, Reg + Offset          (frame index)
663  ///   0x3, Indirect, [Reg + Offset]      (spilled value)
664  ///   0x4, Constant, Offset              (small constant)
665  ///   0x5, ConstIndex, Constants[Offset] (large constant)
666  void StackMaps::emitCallsiteEntries(MCStreamer &OS) {
667    LLVM_DEBUG(print(dbgs()));
668    // Callsite entries.
669    for (const auto &CSI : CSInfos) {
670      const LocationVec &CSLocs = CSI.Locations;
671      const LiveOutVec &LiveOuts = CSI.LiveOuts;
672  
673      // Verify stack map entry. It's better to communicate a problem to the
674      // runtime than crash in case of in-process compilation. Currently, we do
675      // simple overflow checks, but we may eventually communicate other
676      // compilation errors this way.
677      if (CSLocs.size() > UINT16_MAX || LiveOuts.size() > UINT16_MAX) {
678        OS.emitIntValue(UINT64_MAX, 8); // Invalid ID.
679        OS.emitValue(CSI.CSOffsetExpr, 4);
680        OS.emitInt16(0); // Reserved.
681        OS.emitInt16(0); // 0 locations.
682        OS.emitInt16(0); // padding.
683        OS.emitInt16(0); // 0 live-out registers.
684        OS.emitInt32(0); // padding.
685        continue;
686      }
687  
688      OS.emitIntValue(CSI.ID, 8);
689      OS.emitValue(CSI.CSOffsetExpr, 4);
690  
691      // Reserved for flags.
692      OS.emitInt16(0);
693      OS.emitInt16(CSLocs.size());
694  
695      for (const auto &Loc : CSLocs) {
696        OS.emitIntValue(Loc.Type, 1);
697        OS.emitIntValue(0, 1);  // Reserved
698        OS.emitInt16(Loc.Size);
699        OS.emitInt16(Loc.Reg);
700        OS.emitInt16(0); // Reserved
701        OS.emitInt32(Loc.Offset);
702      }
703  
704      // Emit alignment to 8 byte.
705      OS.emitValueToAlignment(Align(8));
706  
707      // Num live-out registers and padding to align to 4 byte.
708      OS.emitInt16(0);
709      OS.emitInt16(LiveOuts.size());
710  
711      for (const auto &LO : LiveOuts) {
712        OS.emitInt16(LO.DwarfRegNum);
713        OS.emitIntValue(0, 1);
714        OS.emitIntValue(LO.Size, 1);
715      }
716      // Emit alignment to 8 byte.
717      OS.emitValueToAlignment(Align(8));
718    }
719  }
720  
721  /// Serialize the stackmap data.
722  void StackMaps::serializeToStackMapSection() {
723    (void)WSMP;
724    // Bail out if there's no stack map data.
725    assert((!CSInfos.empty() || ConstPool.empty()) &&
726           "Expected empty constant pool too!");
727    assert((!CSInfos.empty() || FnInfos.empty()) &&
728           "Expected empty function record too!");
729    if (CSInfos.empty())
730      return;
731  
732    MCContext &OutContext = AP.OutStreamer->getContext();
733    MCStreamer &OS = *AP.OutStreamer;
734  
735    // Create the section.
736    MCSection *StackMapSection =
737        OutContext.getObjectFileInfo()->getStackMapSection();
738    OS.switchSection(StackMapSection);
739  
740    // Emit a dummy symbol to force section inclusion.
741    OS.emitLabel(OutContext.getOrCreateSymbol(Twine("__LLVM_StackMaps")));
742  
743    // Serialize data.
744    LLVM_DEBUG(dbgs() << "********** Stack Map Output **********\n");
745    emitStackmapHeader(OS);
746    emitFunctionFrameRecords(OS);
747    emitConstantPoolEntries(OS);
748    emitCallsiteEntries(OS);
749    OS.addBlankLine();
750  
751    // Clean up.
752    CSInfos.clear();
753    ConstPool.clear();
754  }
755