xref: /freebsd/contrib/llvm-project/compiler-rt/lib/tsan/rtl/tsan_dense_alloc.h (revision a521f2116473fbd8c09db395518f060a27d02334)
1 //===-- tsan_dense_alloc.h --------------------------------------*- C++ -*-===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 //
9 // This file is a part of ThreadSanitizer (TSan), a race detector.
10 //
11 // A DenseSlabAlloc is a freelist-based allocator of fixed-size objects.
12 // DenseSlabAllocCache is a thread-local cache for DenseSlabAlloc.
13 // The only difference with traditional slab allocators is that DenseSlabAlloc
14 // allocates/free indices of objects and provide a functionality to map
15 // the index onto the real pointer. The index is u32, that is, 2 times smaller
16 // than uptr (hense the Dense prefix).
17 //===----------------------------------------------------------------------===//
18 #ifndef TSAN_DENSE_ALLOC_H
19 #define TSAN_DENSE_ALLOC_H
20 
21 #include "sanitizer_common/sanitizer_common.h"
22 #include "tsan_defs.h"
23 #include "tsan_mutex.h"
24 
25 namespace __tsan {
26 
27 class DenseSlabAllocCache {
28   static const uptr kSize = 128;
29   typedef u32 IndexT;
30   uptr pos;
31   IndexT cache[kSize];
32   template<typename T, uptr kL1Size, uptr kL2Size> friend class DenseSlabAlloc;
33 };
34 
35 template<typename T, uptr kL1Size, uptr kL2Size>
36 class DenseSlabAlloc {
37  public:
38   typedef DenseSlabAllocCache Cache;
39   typedef typename Cache::IndexT IndexT;
40 
41   explicit DenseSlabAlloc(const char *name) {
42     // Check that kL1Size and kL2Size are sane.
43     CHECK_EQ(kL1Size & (kL1Size - 1), 0);
44     CHECK_EQ(kL2Size & (kL2Size - 1), 0);
45     CHECK_GE(1ull << (sizeof(IndexT) * 8), kL1Size * kL2Size);
46     // Check that it makes sense to use the dense alloc.
47     CHECK_GE(sizeof(T), sizeof(IndexT));
48     internal_memset(map_, 0, sizeof(map_));
49     freelist_ = 0;
50     fillpos_ = 0;
51     name_ = name;
52   }
53 
54   ~DenseSlabAlloc() {
55     for (uptr i = 0; i < kL1Size; i++) {
56       if (map_[i] != 0)
57         UnmapOrDie(map_[i], kL2Size * sizeof(T));
58     }
59   }
60 
61   IndexT Alloc(Cache *c) {
62     if (c->pos == 0)
63       Refill(c);
64     return c->cache[--c->pos];
65   }
66 
67   void Free(Cache *c, IndexT idx) {
68     DCHECK_NE(idx, 0);
69     if (c->pos == Cache::kSize)
70       Drain(c);
71     c->cache[c->pos++] = idx;
72   }
73 
74   T *Map(IndexT idx) {
75     DCHECK_NE(idx, 0);
76     DCHECK_LE(idx, kL1Size * kL2Size);
77     return &map_[idx / kL2Size][idx % kL2Size];
78   }
79 
80   void FlushCache(Cache *c) {
81     SpinMutexLock lock(&mtx_);
82     while (c->pos) {
83       IndexT idx = c->cache[--c->pos];
84       *(IndexT*)Map(idx) = freelist_;
85       freelist_ = idx;
86     }
87   }
88 
89   void InitCache(Cache *c) {
90     c->pos = 0;
91     internal_memset(c->cache, 0, sizeof(c->cache));
92   }
93 
94  private:
95   T *map_[kL1Size];
96   SpinMutex mtx_;
97   IndexT freelist_;
98   uptr fillpos_;
99   const char *name_;
100 
101   void Refill(Cache *c) {
102     SpinMutexLock lock(&mtx_);
103     if (freelist_ == 0) {
104       if (fillpos_ == kL1Size) {
105         Printf("ThreadSanitizer: %s overflow (%zu*%zu). Dying.\n",
106             name_, kL1Size, kL2Size);
107         Die();
108       }
109       VPrintf(2, "ThreadSanitizer: growing %s: %zu out of %zu*%zu\n",
110           name_, fillpos_, kL1Size, kL2Size);
111       T *batch = (T*)MmapOrDie(kL2Size * sizeof(T), name_);
112       // Reserve 0 as invalid index.
113       IndexT start = fillpos_ == 0 ? 1 : 0;
114       for (IndexT i = start; i < kL2Size; i++) {
115         new(batch + i) T;
116         *(IndexT*)(batch + i) = i + 1 + fillpos_ * kL2Size;
117       }
118       *(IndexT*)(batch + kL2Size - 1) = 0;
119       freelist_ = fillpos_ * kL2Size + start;
120       map_[fillpos_++] = batch;
121     }
122     for (uptr i = 0; i < Cache::kSize / 2 && freelist_ != 0; i++) {
123       IndexT idx = freelist_;
124       c->cache[c->pos++] = idx;
125       freelist_ = *(IndexT*)Map(idx);
126     }
127   }
128 
129   void Drain(Cache *c) {
130     SpinMutexLock lock(&mtx_);
131     for (uptr i = 0; i < Cache::kSize / 2; i++) {
132       IndexT idx = c->cache[--c->pos];
133       *(IndexT*)Map(idx) = freelist_;
134       freelist_ = idx;
135     }
136   }
137 };
138 
139 }  // namespace __tsan
140 
141 #endif  // TSAN_DENSE_ALLOC_H
142