1 //=-- lsan_common_mac.cpp -------------------------------------------------===// 2 // 3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. 4 // See https://llvm.org/LICENSE.txt for license information. 5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception 6 // 7 //===----------------------------------------------------------------------===// 8 // 9 // This file is a part of LeakSanitizer. 10 // Implementation of common leak checking functionality. Darwin-specific code. 11 // 12 //===----------------------------------------------------------------------===// 13 14 #include "sanitizer_common/sanitizer_platform.h" 15 #include "sanitizer_common/sanitizer_libc.h" 16 #include "lsan_common.h" 17 18 #if CAN_SANITIZE_LEAKS && SANITIZER_APPLE 19 20 # include <mach/mach.h> 21 # include <mach/vm_statistics.h> 22 # include <pthread.h> 23 24 # include "lsan_allocator.h" 25 # include "sanitizer_common/sanitizer_allocator_internal.h" 26 namespace __lsan { 27 28 class ThreadContextLsanBase; 29 30 enum class SeenRegion { 31 None = 0, 32 AllocOnce = 1 << 0, 33 LibDispatch = 1 << 1, 34 Foundation = 1 << 2, 35 All = AllocOnce | LibDispatch | Foundation 36 }; 37 38 inline SeenRegion operator|(SeenRegion left, SeenRegion right) { 39 return static_cast<SeenRegion>(static_cast<int>(left) | 40 static_cast<int>(right)); 41 } 42 43 inline SeenRegion &operator|=(SeenRegion &left, const SeenRegion &right) { 44 left = left | right; 45 return left; 46 } 47 48 struct RegionScanState { 49 SeenRegion seen_regions = SeenRegion::None; 50 bool in_libdispatch = false; 51 }; 52 53 typedef struct { 54 int disable_counter; 55 ThreadContextLsanBase *current_thread; 56 AllocatorCache cache; 57 } thread_local_data_t; 58 59 static pthread_key_t key; 60 static pthread_once_t key_once = PTHREAD_ONCE_INIT; 61 62 // The main thread destructor requires the current thread, 63 // so we can't destroy it until it's been used and reset. 64 void restore_tid_data(void *ptr) { 65 thread_local_data_t *data = (thread_local_data_t *)ptr; 66 if (data->current_thread) 67 pthread_setspecific(key, data); 68 } 69 70 static void make_tls_key() { 71 CHECK_EQ(pthread_key_create(&key, restore_tid_data), 0); 72 } 73 74 static thread_local_data_t *get_tls_val(bool alloc) { 75 pthread_once(&key_once, make_tls_key); 76 77 thread_local_data_t *ptr = (thread_local_data_t *)pthread_getspecific(key); 78 if (ptr == NULL && alloc) { 79 ptr = (thread_local_data_t *)InternalAlloc(sizeof(*ptr)); 80 ptr->disable_counter = 0; 81 ptr->current_thread = nullptr; 82 ptr->cache = AllocatorCache(); 83 pthread_setspecific(key, ptr); 84 } 85 86 return ptr; 87 } 88 89 bool DisabledInThisThread() { 90 thread_local_data_t *data = get_tls_val(false); 91 return data ? data->disable_counter > 0 : false; 92 } 93 94 void DisableInThisThread() { ++get_tls_val(true)->disable_counter; } 95 96 void EnableInThisThread() { 97 int *disable_counter = &get_tls_val(true)->disable_counter; 98 if (*disable_counter == 0) { 99 DisableCounterUnderflow(); 100 } 101 --*disable_counter; 102 } 103 104 ThreadContextLsanBase *GetCurrentThread() { 105 thread_local_data_t *data = get_tls_val(false); 106 return data ? data->current_thread : nullptr; 107 } 108 109 void SetCurrentThread(ThreadContextLsanBase *tctx) { 110 get_tls_val(true)->current_thread = tctx; 111 } 112 113 AllocatorCache *GetAllocatorCache() { return &get_tls_val(true)->cache; } 114 115 LoadedModule *GetLinker() { return nullptr; } 116 117 // Required on Linux for initialization of TLS behavior, but should not be 118 // required on Darwin. 119 void InitializePlatformSpecificModules() {} 120 121 // Sections which can't contain contain global pointers. This list errs on the 122 // side of caution to avoid false positives, at the expense of performance. 123 // 124 // Other potentially safe sections include: 125 // __all_image_info, __crash_info, __const, __got, __interpose, __objc_msg_break 126 // 127 // Sections which definitely cannot be included here are: 128 // __objc_data, __objc_const, __data, __bss, __common, __thread_data, 129 // __thread_bss, __thread_vars, __objc_opt_rw, __objc_opt_ptrs 130 static const char *kSkippedSecNames[] = { 131 "__cfstring", "__la_symbol_ptr", "__mod_init_func", 132 "__mod_term_func", "__nl_symbol_ptr", "__objc_classlist", 133 "__objc_classrefs", "__objc_imageinfo", "__objc_nlclslist", 134 "__objc_protolist", "__objc_selrefs", "__objc_superrefs"}; 135 136 // Scans global variables for heap pointers. 137 void ProcessGlobalRegions(Frontier *frontier) { 138 for (auto name : kSkippedSecNames) 139 CHECK(internal_strnlen(name, kMaxSegName + 1) <= kMaxSegName); 140 141 MemoryMappingLayout memory_mapping(false); 142 InternalMmapVector<LoadedModule> modules; 143 modules.reserve(128); 144 memory_mapping.DumpListOfModules(&modules); 145 for (uptr i = 0; i < modules.size(); ++i) { 146 // Even when global scanning is disabled, we still need to scan 147 // system libraries for stashed pointers 148 if (!flags()->use_globals && modules[i].instrumented()) continue; 149 150 for (const __sanitizer::LoadedModule::AddressRange &range : 151 modules[i].ranges()) { 152 // Sections storing global variables are writable and non-executable 153 if (range.executable || !range.writable) continue; 154 155 for (auto name : kSkippedSecNames) { 156 if (!internal_strcmp(range.name, name)) continue; 157 } 158 159 ScanGlobalRange(range.beg, range.end, frontier); 160 } 161 } 162 } 163 164 void ProcessPlatformSpecificAllocations(Frontier *frontier) { 165 vm_address_t address = 0; 166 kern_return_t err = KERN_SUCCESS; 167 168 InternalMmapVector<Region> mapped_regions; 169 bool use_root_regions = flags()->use_root_regions && HasRootRegions(); 170 171 RegionScanState scan_state; 172 while (err == KERN_SUCCESS) { 173 vm_size_t size = 0; 174 unsigned depth = 1; 175 struct vm_region_submap_info_64 info; 176 mach_msg_type_number_t count = VM_REGION_SUBMAP_INFO_COUNT_64; 177 err = vm_region_recurse_64(mach_task_self(), &address, &size, &depth, 178 (vm_region_info_t)&info, &count); 179 180 uptr end_address = address + size; 181 if (info.user_tag == VM_MEMORY_OS_ALLOC_ONCE) { 182 // libxpc stashes some pointers in the Kernel Alloc Once page, 183 // make sure not to report those as leaks. 184 scan_state.seen_regions |= SeenRegion::AllocOnce; 185 ScanRangeForPointers(address, end_address, frontier, "GLOBAL", 186 kReachable); 187 } else if (info.user_tag == VM_MEMORY_FOUNDATION) { 188 // Objective-C block trampolines use the Foundation region. 189 scan_state.seen_regions |= SeenRegion::Foundation; 190 ScanRangeForPointers(address, end_address, frontier, "GLOBAL", 191 kReachable); 192 } else if (info.user_tag == VM_MEMORY_LIBDISPATCH) { 193 // Dispatch continuations use the libdispatch region. Empirically, there 194 // can be more than one region with this tag, so we'll optimistically 195 // assume that they're continguous. Otherwise, we would need to scan every 196 // region to ensure we find them all. 197 scan_state.in_libdispatch = true; 198 ScanRangeForPointers(address, end_address, frontier, "GLOBAL", 199 kReachable); 200 } else if (scan_state.in_libdispatch) { 201 scan_state.seen_regions |= SeenRegion::LibDispatch; 202 scan_state.in_libdispatch = false; 203 } 204 205 // Recursing over the full memory map is very slow, break out 206 // early if we don't need the full iteration. 207 if (scan_state.seen_regions == SeenRegion::All && !use_root_regions) { 208 break; 209 } 210 211 // This additional root region scan is required on Darwin in order to 212 // detect root regions contained within mmap'd memory regions, because 213 // the Darwin implementation of sanitizer_procmaps traverses images 214 // as loaded by dyld, and not the complete set of all memory regions. 215 // 216 // TODO(fjricci) - remove this once sanitizer_procmaps_mac has the same 217 // behavior as sanitizer_procmaps_linux and traverses all memory regions 218 if (use_root_regions && (info.protection & kProtectionRead)) 219 mapped_regions.push_back({address, end_address}); 220 221 address = end_address; 222 } 223 ScanRootRegions(frontier, mapped_regions); 224 } 225 226 // On darwin, we can intercept _exit gracefully, and return a failing exit code 227 // if required at that point. Calling Die() here is undefined behavior and 228 // causes rare race conditions. 229 void HandleLeaks() {} 230 231 void LockStuffAndStopTheWorld(StopTheWorldCallback callback, 232 CheckForLeaksParam *argument) { 233 ScopedStopTheWorldLock lock; 234 StopTheWorld(callback, argument); 235 } 236 237 } // namespace __lsan 238 239 #endif // CAN_SANITIZE_LEAKS && SANITIZER_APPLE 240