//===- MemoryMapper.cpp - Cross-process memory mapper ------------*- C++ -*-==// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception // //===----------------------------------------------------------------------===// #include "llvm/ExecutionEngine/Orc/MemoryMapper.h" #include "llvm/ExecutionEngine/Orc/Shared/OrcRTBridge.h" #include "llvm/Support/WindowsError.h" #include #if defined(LLVM_ON_UNIX) && !defined(__ANDROID__) #include #include #include #elif defined(_WIN32) #include #endif namespace llvm { namespace orc { MemoryMapper::~MemoryMapper() {} InProcessMemoryMapper::InProcessMemoryMapper(size_t PageSize) : PageSize(PageSize) {} Expected> InProcessMemoryMapper::Create() { auto PageSize = sys::Process::getPageSize(); if (!PageSize) return PageSize.takeError(); return std::make_unique(*PageSize); } void InProcessMemoryMapper::reserve(size_t NumBytes, OnReservedFunction OnReserved) { std::error_code EC; auto MB = sys::Memory::allocateMappedMemory( NumBytes, nullptr, sys::Memory::MF_READ | sys::Memory::MF_WRITE, EC); if (EC) return OnReserved(errorCodeToError(EC)); { std::lock_guard Lock(Mutex); Reservations[MB.base()].Size = MB.allocatedSize(); } OnReserved( ExecutorAddrRange(ExecutorAddr::fromPtr(MB.base()), MB.allocatedSize())); } char *InProcessMemoryMapper::prepare(ExecutorAddr Addr, size_t ContentSize) { return Addr.toPtr(); } void InProcessMemoryMapper::initialize(MemoryMapper::AllocInfo &AI, OnInitializedFunction OnInitialized) { ExecutorAddr MinAddr(~0ULL); ExecutorAddr MaxAddr(0); // FIXME: Release finalize lifetime segments. for (auto &Segment : AI.Segments) { auto Base = AI.MappingBase + Segment.Offset; auto Size = Segment.ContentSize + Segment.ZeroFillSize; if (Base < MinAddr) MinAddr = Base; if (Base + Size > MaxAddr) MaxAddr = Base + Size; std::memset((Base + Segment.ContentSize).toPtr(), 0, Segment.ZeroFillSize); if (auto EC = sys::Memory::protectMappedMemory( {Base.toPtr(), Size}, toSysMemoryProtectionFlags(Segment.AG.getMemProt()))) { return OnInitialized(errorCodeToError(EC)); } if ((Segment.AG.getMemProt() & MemProt::Exec) == MemProt::Exec) sys::Memory::InvalidateInstructionCache(Base.toPtr(), Size); } auto DeinitializeActions = shared::runFinalizeActions(AI.Actions); if (!DeinitializeActions) return OnInitialized(DeinitializeActions.takeError()); { std::lock_guard Lock(Mutex); // This is the maximum range whose permission have been possibly modified Allocations[MinAddr].Size = MaxAddr - MinAddr; Allocations[MinAddr].DeinitializationActions = std::move(*DeinitializeActions); Reservations[AI.MappingBase.toPtr()].Allocations.push_back(MinAddr); } OnInitialized(MinAddr); } void InProcessMemoryMapper::deinitialize( ArrayRef Bases, MemoryMapper::OnDeinitializedFunction OnDeinitialized) { Error AllErr = Error::success(); { std::lock_guard Lock(Mutex); for (auto Base : llvm::reverse(Bases)) { if (Error Err = shared::runDeallocActions( Allocations[Base].DeinitializationActions)) { AllErr = joinErrors(std::move(AllErr), std::move(Err)); } // Reset protections to read/write so the area can be reused if (auto EC = sys::Memory::protectMappedMemory( {Base.toPtr(), Allocations[Base].Size}, sys::Memory::ProtectionFlags::MF_READ | sys::Memory::ProtectionFlags::MF_WRITE)) { AllErr = joinErrors(std::move(AllErr), errorCodeToError(EC)); } Allocations.erase(Base); } } OnDeinitialized(std::move(AllErr)); } void InProcessMemoryMapper::release(ArrayRef Bases, OnReleasedFunction OnReleased) { Error Err = Error::success(); for (auto Base : Bases) { std::vector AllocAddrs; size_t Size; { std::lock_guard Lock(Mutex); auto &R = Reservations[Base.toPtr()]; Size = R.Size; AllocAddrs.swap(R.Allocations); } // deinitialize sub allocations std::promise P; auto F = P.get_future(); deinitialize(AllocAddrs, [&](Error Err) { P.set_value(std::move(Err)); }); if (Error E = F.get()) { Err = joinErrors(std::move(Err), std::move(E)); } // free the memory auto MB = sys::MemoryBlock(Base.toPtr(), Size); auto EC = sys::Memory::releaseMappedMemory(MB); if (EC) { Err = joinErrors(std::move(Err), errorCodeToError(EC)); } std::lock_guard Lock(Mutex); Reservations.erase(Base.toPtr()); } OnReleased(std::move(Err)); } InProcessMemoryMapper::~InProcessMemoryMapper() { std::vector ReservationAddrs; { std::lock_guard Lock(Mutex); ReservationAddrs.reserve(Reservations.size()); for (const auto &R : Reservations) { ReservationAddrs.push_back(ExecutorAddr::fromPtr(R.getFirst())); } } std::promise P; auto F = P.get_future(); release(ReservationAddrs, [&](Error Err) { P.set_value(std::move(Err)); }); cantFail(F.get()); } // SharedMemoryMapper SharedMemoryMapper::SharedMemoryMapper(ExecutorProcessControl &EPC, SymbolAddrs SAs, size_t PageSize) : EPC(EPC), SAs(SAs), PageSize(PageSize) { #if (!defined(LLVM_ON_UNIX) || defined(__ANDROID__)) && !defined(_WIN32) llvm_unreachable("SharedMemoryMapper is not supported on this platform yet"); #endif } Expected> SharedMemoryMapper::Create(ExecutorProcessControl &EPC, SymbolAddrs SAs) { #if (defined(LLVM_ON_UNIX) && !defined(__ANDROID__)) || defined(_WIN32) auto PageSize = sys::Process::getPageSize(); if (!PageSize) return PageSize.takeError(); return std::make_unique(EPC, SAs, *PageSize); #else return make_error( "SharedMemoryMapper is not supported on this platform yet", inconvertibleErrorCode()); #endif } void SharedMemoryMapper::reserve(size_t NumBytes, OnReservedFunction OnReserved) { #if (defined(LLVM_ON_UNIX) && !defined(__ANDROID__)) || defined(_WIN32) EPC.callSPSWrapperAsync< rt::SPSExecutorSharedMemoryMapperServiceReserveSignature>( SAs.Reserve, [this, NumBytes, OnReserved = std::move(OnReserved)]( Error SerializationErr, Expected> Result) mutable { if (SerializationErr) { cantFail(Result.takeError()); return OnReserved(std::move(SerializationErr)); } if (!Result) return OnReserved(Result.takeError()); ExecutorAddr RemoteAddr; std::string SharedMemoryName; std::tie(RemoteAddr, SharedMemoryName) = std::move(*Result); void *LocalAddr = nullptr; #if defined(LLVM_ON_UNIX) int SharedMemoryFile = shm_open(SharedMemoryName.c_str(), O_RDWR, 0700); if (SharedMemoryFile < 0) { return OnReserved(errorCodeToError( std::error_code(errno, std::generic_category()))); } // this prevents other processes from accessing it by name shm_unlink(SharedMemoryName.c_str()); LocalAddr = mmap(nullptr, NumBytes, PROT_READ | PROT_WRITE, MAP_SHARED, SharedMemoryFile, 0); if (LocalAddr == MAP_FAILED) { return OnReserved(errorCodeToError( std::error_code(errno, std::generic_category()))); } close(SharedMemoryFile); #elif defined(_WIN32) std::wstring WideSharedMemoryName(SharedMemoryName.begin(), SharedMemoryName.end()); HANDLE SharedMemoryFile = OpenFileMappingW( FILE_MAP_ALL_ACCESS, FALSE, WideSharedMemoryName.c_str()); if (!SharedMemoryFile) return OnReserved(errorCodeToError(mapWindowsError(GetLastError()))); LocalAddr = MapViewOfFile(SharedMemoryFile, FILE_MAP_ALL_ACCESS, 0, 0, 0); if (!LocalAddr) { CloseHandle(SharedMemoryFile); return OnReserved(errorCodeToError(mapWindowsError(GetLastError()))); } CloseHandle(SharedMemoryFile); #endif { std::lock_guard Lock(Mutex); Reservations.insert({RemoteAddr, {LocalAddr, NumBytes}}); } OnReserved(ExecutorAddrRange(RemoteAddr, NumBytes)); }, SAs.Instance, static_cast(NumBytes)); #else OnReserved(make_error( "SharedMemoryMapper is not supported on this platform yet", inconvertibleErrorCode())); #endif } char *SharedMemoryMapper::prepare(ExecutorAddr Addr, size_t ContentSize) { auto R = Reservations.upper_bound(Addr); assert(R != Reservations.begin() && "Attempt to prepare unreserved range"); R--; ExecutorAddrDiff Offset = Addr - R->first; return static_cast(R->second.LocalAddr) + Offset; } void SharedMemoryMapper::initialize(MemoryMapper::AllocInfo &AI, OnInitializedFunction OnInitialized) { auto Reservation = Reservations.upper_bound(AI.MappingBase); assert(Reservation != Reservations.begin() && "Attempt to initialize unreserved range"); Reservation--; auto AllocationOffset = AI.MappingBase - Reservation->first; tpctypes::SharedMemoryFinalizeRequest FR; AI.Actions.swap(FR.Actions); FR.Segments.reserve(AI.Segments.size()); for (auto Segment : AI.Segments) { char *Base = static_cast(Reservation->second.LocalAddr) + AllocationOffset + Segment.Offset; std::memset(Base + Segment.ContentSize, 0, Segment.ZeroFillSize); tpctypes::SharedMemorySegFinalizeRequest SegReq; SegReq.RAG = {Segment.AG.getMemProt(), Segment.AG.getMemLifetimePolicy() == MemLifetimePolicy::Finalize}; SegReq.Addr = AI.MappingBase + Segment.Offset; SegReq.Size = Segment.ContentSize + Segment.ZeroFillSize; FR.Segments.push_back(SegReq); } EPC.callSPSWrapperAsync< rt::SPSExecutorSharedMemoryMapperServiceInitializeSignature>( SAs.Initialize, [OnInitialized = std::move(OnInitialized)]( Error SerializationErr, Expected Result) mutable { if (SerializationErr) { cantFail(Result.takeError()); return OnInitialized(std::move(SerializationErr)); } OnInitialized(std::move(Result)); }, SAs.Instance, Reservation->first, std::move(FR)); } void SharedMemoryMapper::deinitialize( ArrayRef Allocations, MemoryMapper::OnDeinitializedFunction OnDeinitialized) { EPC.callSPSWrapperAsync< rt::SPSExecutorSharedMemoryMapperServiceDeinitializeSignature>( SAs.Deinitialize, [OnDeinitialized = std::move(OnDeinitialized)](Error SerializationErr, Error Result) mutable { if (SerializationErr) { cantFail(std::move(Result)); return OnDeinitialized(std::move(SerializationErr)); } OnDeinitialized(std::move(Result)); }, SAs.Instance, Allocations); } void SharedMemoryMapper::release(ArrayRef Bases, OnReleasedFunction OnReleased) { #if (defined(LLVM_ON_UNIX) && !defined(__ANDROID__)) || defined(_WIN32) Error Err = Error::success(); { std::lock_guard Lock(Mutex); for (auto Base : Bases) { #if defined(LLVM_ON_UNIX) if (munmap(Reservations[Base].LocalAddr, Reservations[Base].Size) != 0) Err = joinErrors(std::move(Err), errorCodeToError(std::error_code( errno, std::generic_category()))); #elif defined(_WIN32) if (!UnmapViewOfFile(Reservations[Base].LocalAddr)) Err = joinErrors(std::move(Err), errorCodeToError(mapWindowsError(GetLastError()))); #endif Reservations.erase(Base); } } EPC.callSPSWrapperAsync< rt::SPSExecutorSharedMemoryMapperServiceReleaseSignature>( SAs.Release, [OnReleased = std::move(OnReleased), Err = std::move(Err)](Error SerializationErr, Error Result) mutable { if (SerializationErr) { cantFail(std::move(Result)); return OnReleased( joinErrors(std::move(Err), std::move(SerializationErr))); } return OnReleased(joinErrors(std::move(Err), std::move(Result))); }, SAs.Instance, Bases); #else OnReleased(make_error( "SharedMemoryMapper is not supported on this platform yet", inconvertibleErrorCode())); #endif } SharedMemoryMapper::~SharedMemoryMapper() { std::lock_guard Lock(Mutex); for (const auto &R : Reservations) { #if defined(LLVM_ON_UNIX) && !defined(__ANDROID__) munmap(R.second.LocalAddr, R.second.Size); #elif defined(_WIN32) UnmapViewOfFile(R.second.LocalAddr); #else (void)R; #endif } } } // namespace orc } // namespace llvm