//===- X86_64.cpp ---------------------------------------------------------===// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception // //===----------------------------------------------------------------------===// #include "InputFiles.h" #include "Symbols.h" #include "SyntheticSections.h" #include "Target.h" #include "lld/Common/ErrorHandler.h" #include "llvm/BinaryFormat/MachO.h" #include "llvm/Support/Endian.h" using namespace llvm::MachO; using namespace llvm::support::endian; using namespace lld; using namespace lld::macho; namespace { struct X86_64 : TargetInfo { X86_64(); int64_t getEmbeddedAddend(MemoryBufferRef, uint64_t offset, const relocation_info) const override; void relocateOne(uint8_t *loc, const Reloc &, uint64_t va, uint64_t relocVA) const override; void writeStub(uint8_t *buf, const Symbol &) const override; void writeStubHelperHeader(uint8_t *buf) const override; void writeStubHelperEntry(uint8_t *buf, const DylibSymbol &, uint64_t entryAddr) const override; void relaxGotLoad(uint8_t *loc, uint8_t type) const override; const RelocAttrs &getRelocAttrs(uint8_t type) const override; uint64_t getPageSize() const override { return 4 * 1024; } }; } // namespace const RelocAttrs &X86_64::getRelocAttrs(uint8_t type) const { static const std::array relocAttrsArray{{ #define B(x) RelocAttrBits::x {"UNSIGNED", B(UNSIGNED) | B(ABSOLUTE) | B(EXTERN) | B(LOCAL) | B(BYTE4) | B(BYTE8)}, {"SIGNED", B(PCREL) | B(EXTERN) | B(LOCAL) | B(BYTE4)}, {"BRANCH", B(PCREL) | B(EXTERN) | B(BRANCH) | B(BYTE4)}, {"GOT_LOAD", B(PCREL) | B(EXTERN) | B(GOT) | B(LOAD) | B(BYTE4)}, {"GOT", B(PCREL) | B(EXTERN) | B(GOT) | B(POINTER) | B(BYTE4)}, {"SUBTRACTOR", B(SUBTRAHEND) | B(EXTERN) | B(BYTE4) | B(BYTE8)}, {"SIGNED_1", B(PCREL) | B(EXTERN) | B(LOCAL) | B(BYTE4)}, {"SIGNED_2", B(PCREL) | B(EXTERN) | B(LOCAL) | B(BYTE4)}, {"SIGNED_4", B(PCREL) | B(EXTERN) | B(LOCAL) | B(BYTE4)}, {"TLV", B(PCREL) | B(EXTERN) | B(TLV) | B(LOAD) | B(BYTE4)}, #undef B }}; assert(type < relocAttrsArray.size() && "invalid relocation type"); if (type >= relocAttrsArray.size()) return invalidRelocAttrs; return relocAttrsArray[type]; } static int pcrelOffset(uint8_t type) { switch (type) { case X86_64_RELOC_SIGNED_1: return 1; case X86_64_RELOC_SIGNED_2: return 2; case X86_64_RELOC_SIGNED_4: return 4; default: return 0; } } int64_t X86_64::getEmbeddedAddend(MemoryBufferRef mb, uint64_t offset, relocation_info rel) const { auto *buf = reinterpret_cast(mb.getBufferStart()); const uint8_t *loc = buf + offset + rel.r_address; switch (rel.r_length) { case 2: return static_cast(read32le(loc)) + pcrelOffset(rel.r_type); case 3: return read64le(loc) + pcrelOffset(rel.r_type); default: llvm_unreachable("invalid r_length"); } } void X86_64::relocateOne(uint8_t *loc, const Reloc &r, uint64_t value, uint64_t relocVA) const { if (r.pcrel) { uint64_t pc = relocVA + 4 + pcrelOffset(r.type); value -= pc; } switch (r.length) { case 2: if (r.type == X86_64_RELOC_UNSIGNED) checkUInt(r, value, 32); else checkInt(r, value, 32); write32le(loc, value); break; case 3: write64le(loc, value); break; default: llvm_unreachable("invalid r_length"); } } // The following methods emit a number of assembly sequences with RIP-relative // addressing. Note that RIP-relative addressing on X86-64 has the RIP pointing // to the next instruction, not the current instruction, so we always have to // account for the current instruction's size when calculating offsets. // writeRipRelative helps with that. // // bufAddr: The virtual address corresponding to buf[0]. // bufOff: The offset within buf of the next instruction. // destAddr: The destination address that the current instruction references. static void writeRipRelative(SymbolDiagnostic d, uint8_t *buf, uint64_t bufAddr, uint64_t bufOff, uint64_t destAddr) { uint64_t rip = bufAddr + bufOff; checkInt(d, destAddr - rip, 32); // For the instructions we care about, the RIP-relative address is always // stored in the last 4 bytes of the instruction. write32le(buf + bufOff - 4, destAddr - rip); } static constexpr uint8_t stub[] = { 0xff, 0x25, 0, 0, 0, 0, // jmpq *__la_symbol_ptr(%rip) }; void X86_64::writeStub(uint8_t *buf, const Symbol &sym) const { memcpy(buf, stub, 2); // just copy the two nonzero bytes uint64_t stubAddr = in.stubs->addr + sym.stubsIndex * sizeof(stub); writeRipRelative({&sym, "stub"}, buf, stubAddr, sizeof(stub), in.lazyPointers->addr + sym.stubsIndex * LP64::wordSize); } static constexpr uint8_t stubHelperHeader[] = { 0x4c, 0x8d, 0x1d, 0, 0, 0, 0, // 0x0: leaq ImageLoaderCache(%rip), %r11 0x41, 0x53, // 0x7: pushq %r11 0xff, 0x25, 0, 0, 0, 0, // 0x9: jmpq *dyld_stub_binder@GOT(%rip) 0x90, // 0xf: nop }; void X86_64::writeStubHelperHeader(uint8_t *buf) const { memcpy(buf, stubHelperHeader, sizeof(stubHelperHeader)); SymbolDiagnostic d = {nullptr, "stub helper header"}; writeRipRelative(d, buf, in.stubHelper->addr, 7, in.imageLoaderCache->getVA()); writeRipRelative(d, buf, in.stubHelper->addr, 0xf, in.got->addr + in.stubHelper->stubBinder->gotIndex * LP64::wordSize); } static constexpr uint8_t stubHelperEntry[] = { 0x68, 0, 0, 0, 0, // 0x0: pushq 0xe9, 0, 0, 0, 0, // 0x5: jmp <__stub_helper> }; void X86_64::writeStubHelperEntry(uint8_t *buf, const DylibSymbol &sym, uint64_t entryAddr) const { memcpy(buf, stubHelperEntry, sizeof(stubHelperEntry)); write32le(buf + 1, sym.lazyBindOffset); writeRipRelative({&sym, "stub helper"}, buf, entryAddr, sizeof(stubHelperEntry), in.stubHelper->addr); } void X86_64::relaxGotLoad(uint8_t *loc, uint8_t type) const { // Convert MOVQ to LEAQ if (loc[-2] != 0x8b) error(getRelocAttrs(type).name + " reloc requires MOVQ instruction"); loc[-2] = 0x8d; } X86_64::X86_64() : TargetInfo(LP64()) { cpuType = CPU_TYPE_X86_64; cpuSubtype = CPU_SUBTYPE_X86_64_ALL; stubSize = sizeof(stub); stubHelperHeaderSize = sizeof(stubHelperHeader); stubHelperEntrySize = sizeof(stubHelperEntry); } TargetInfo *macho::createX86_64TargetInfo() { static X86_64 t; return &t; }