1 //===- llvm/ADT/SmallPtrSet.cpp - 'Normally small' pointer set ------------===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 //
9 // This file implements the SmallPtrSet class. See SmallPtrSet.h for an
10 // overview of the algorithm.
11 //
12 //===----------------------------------------------------------------------===//
13
14 #include "llvm/ADT/SmallPtrSet.h"
15 #include "llvm/ADT/DenseMapInfo.h"
16 #include "llvm/Support/MathExtras.h"
17 #include "llvm/Support/MemAlloc.h"
18 #include <algorithm>
19 #include <cassert>
20 #include <cstdlib>
21
22 using namespace llvm;
23
shrink_and_clear()24 void SmallPtrSetImplBase::shrink_and_clear() {
25 assert(!isSmall() && "Can't shrink a small set!");
26 free(CurArray);
27
28 // Reduce the number of buckets.
29 unsigned Size = size();
30 CurArraySize = Size > 16 ? 1 << (Log2_32_Ceil(Size) + 1) : 32;
31 NumNonEmpty = NumTombstones = 0;
32
33 // Install the new array. Clear all the buckets to empty.
34 CurArray = (const void**)safe_malloc(sizeof(void*) * CurArraySize);
35
36 memset(CurArray, -1, CurArraySize*sizeof(void*));
37 }
38
39 std::pair<const void *const *, bool>
insert_imp_big(const void * Ptr)40 SmallPtrSetImplBase::insert_imp_big(const void *Ptr) {
41 if (LLVM_UNLIKELY(size() * 4 >= CurArraySize * 3)) {
42 // If more than 3/4 of the array is full, grow.
43 Grow(CurArraySize < 64 ? 128 : CurArraySize * 2);
44 } else if (LLVM_UNLIKELY(CurArraySize - NumNonEmpty < CurArraySize / 8)) {
45 // If fewer of 1/8 of the array is empty (meaning that many are filled with
46 // tombstones), rehash.
47 Grow(CurArraySize);
48 }
49
50 // Okay, we know we have space. Find a hash bucket.
51 const void **Bucket = const_cast<const void**>(FindBucketFor(Ptr));
52 if (*Bucket == Ptr)
53 return std::make_pair(Bucket, false); // Already inserted, good.
54
55 // Otherwise, insert it!
56 if (*Bucket == getTombstoneMarker())
57 --NumTombstones;
58 else
59 ++NumNonEmpty; // Track density.
60 *Bucket = Ptr;
61 incrementEpoch();
62 return std::make_pair(Bucket, true);
63 }
64
FindBucketFor(const void * Ptr) const65 const void * const *SmallPtrSetImplBase::FindBucketFor(const void *Ptr) const {
66 unsigned Bucket = DenseMapInfo<void *>::getHashValue(Ptr) & (CurArraySize-1);
67 unsigned ArraySize = CurArraySize;
68 unsigned ProbeAmt = 1;
69 const void *const *Array = CurArray;
70 const void *const *Tombstone = nullptr;
71 while (true) {
72 // If we found an empty bucket, the pointer doesn't exist in the set.
73 // Return a tombstone if we've seen one so far, or the empty bucket if
74 // not.
75 if (LLVM_LIKELY(Array[Bucket] == getEmptyMarker()))
76 return Tombstone ? Tombstone : Array+Bucket;
77
78 // Found Ptr's bucket?
79 if (LLVM_LIKELY(Array[Bucket] == Ptr))
80 return Array+Bucket;
81
82 // If this is a tombstone, remember it. If Ptr ends up not in the set, we
83 // prefer to return it than something that would require more probing.
84 if (Array[Bucket] == getTombstoneMarker() && !Tombstone)
85 Tombstone = Array+Bucket; // Remember the first tombstone found.
86
87 // It's a hash collision or a tombstone. Reprobe.
88 Bucket = (Bucket + ProbeAmt++) & (ArraySize-1);
89 }
90 }
91
92 /// Grow - Allocate a larger backing store for the buckets and move it over.
93 ///
Grow(unsigned NewSize)94 void SmallPtrSetImplBase::Grow(unsigned NewSize) {
95 const void **OldBuckets = CurArray;
96 const void **OldEnd = EndPointer();
97 bool WasSmall = isSmall();
98
99 // Install the new array. Clear all the buckets to empty.
100 const void **NewBuckets = (const void**) safe_malloc(sizeof(void*) * NewSize);
101
102 // Reset member only if memory was allocated successfully
103 CurArray = NewBuckets;
104 CurArraySize = NewSize;
105 memset(CurArray, -1, NewSize*sizeof(void*));
106
107 // Copy over all valid entries.
108 for (const void **BucketPtr = OldBuckets; BucketPtr != OldEnd; ++BucketPtr) {
109 // Copy over the element if it is valid.
110 const void *Elt = *BucketPtr;
111 if (Elt != getTombstoneMarker() && Elt != getEmptyMarker())
112 *const_cast<void**>(FindBucketFor(Elt)) = const_cast<void*>(Elt);
113 }
114
115 if (!WasSmall)
116 free(OldBuckets);
117 NumNonEmpty -= NumTombstones;
118 NumTombstones = 0;
119 }
120
SmallPtrSetImplBase(const void ** SmallStorage,const SmallPtrSetImplBase & that)121 SmallPtrSetImplBase::SmallPtrSetImplBase(const void **SmallStorage,
122 const SmallPtrSetImplBase &that) {
123 SmallArray = SmallStorage;
124
125 // If we're becoming small, prepare to insert into our stack space
126 if (that.isSmall()) {
127 CurArray = SmallArray;
128 // Otherwise, allocate new heap space (unless we were the same size)
129 } else {
130 CurArray = (const void**)safe_malloc(sizeof(void*) * that.CurArraySize);
131 }
132
133 // Copy over the that array.
134 CopyHelper(that);
135 }
136
SmallPtrSetImplBase(const void ** SmallStorage,unsigned SmallSize,SmallPtrSetImplBase && that)137 SmallPtrSetImplBase::SmallPtrSetImplBase(const void **SmallStorage,
138 unsigned SmallSize,
139 SmallPtrSetImplBase &&that) {
140 SmallArray = SmallStorage;
141 MoveHelper(SmallSize, std::move(that));
142 }
143
CopyFrom(const SmallPtrSetImplBase & RHS)144 void SmallPtrSetImplBase::CopyFrom(const SmallPtrSetImplBase &RHS) {
145 assert(&RHS != this && "Self-copy should be handled by the caller.");
146
147 if (isSmall() && RHS.isSmall())
148 assert(CurArraySize == RHS.CurArraySize &&
149 "Cannot assign sets with different small sizes");
150
151 // If we're becoming small, prepare to insert into our stack space
152 if (RHS.isSmall()) {
153 if (!isSmall())
154 free(CurArray);
155 CurArray = SmallArray;
156 // Otherwise, allocate new heap space (unless we were the same size)
157 } else if (CurArraySize != RHS.CurArraySize) {
158 if (isSmall())
159 CurArray = (const void**)safe_malloc(sizeof(void*) * RHS.CurArraySize);
160 else {
161 const void **T = (const void**)safe_realloc(CurArray,
162 sizeof(void*) * RHS.CurArraySize);
163 CurArray = T;
164 }
165 }
166
167 CopyHelper(RHS);
168 }
169
CopyHelper(const SmallPtrSetImplBase & RHS)170 void SmallPtrSetImplBase::CopyHelper(const SmallPtrSetImplBase &RHS) {
171 // Copy over the new array size
172 CurArraySize = RHS.CurArraySize;
173
174 // Copy over the contents from the other set
175 std::copy(RHS.CurArray, RHS.EndPointer(), CurArray);
176
177 NumNonEmpty = RHS.NumNonEmpty;
178 NumTombstones = RHS.NumTombstones;
179 }
180
MoveFrom(unsigned SmallSize,SmallPtrSetImplBase && RHS)181 void SmallPtrSetImplBase::MoveFrom(unsigned SmallSize,
182 SmallPtrSetImplBase &&RHS) {
183 if (!isSmall())
184 free(CurArray);
185 MoveHelper(SmallSize, std::move(RHS));
186 }
187
MoveHelper(unsigned SmallSize,SmallPtrSetImplBase && RHS)188 void SmallPtrSetImplBase::MoveHelper(unsigned SmallSize,
189 SmallPtrSetImplBase &&RHS) {
190 assert(&RHS != this && "Self-move should be handled by the caller.");
191
192 if (RHS.isSmall()) {
193 // Copy a small RHS rather than moving.
194 CurArray = SmallArray;
195 std::copy(RHS.CurArray, RHS.CurArray + RHS.NumNonEmpty, CurArray);
196 } else {
197 CurArray = RHS.CurArray;
198 RHS.CurArray = RHS.SmallArray;
199 }
200
201 // Copy the rest of the trivial members.
202 CurArraySize = RHS.CurArraySize;
203 NumNonEmpty = RHS.NumNonEmpty;
204 NumTombstones = RHS.NumTombstones;
205
206 // Make the RHS small and empty.
207 RHS.CurArraySize = SmallSize;
208 assert(RHS.CurArray == RHS.SmallArray);
209 RHS.NumNonEmpty = 0;
210 RHS.NumTombstones = 0;
211 }
212
swap(SmallPtrSetImplBase & RHS)213 void SmallPtrSetImplBase::swap(SmallPtrSetImplBase &RHS) {
214 if (this == &RHS) return;
215
216 // We can only avoid copying elements if neither set is small.
217 if (!this->isSmall() && !RHS.isSmall()) {
218 std::swap(this->CurArray, RHS.CurArray);
219 std::swap(this->CurArraySize, RHS.CurArraySize);
220 std::swap(this->NumNonEmpty, RHS.NumNonEmpty);
221 std::swap(this->NumTombstones, RHS.NumTombstones);
222 return;
223 }
224
225 // FIXME: From here on we assume that both sets have the same small size.
226
227 // If only RHS is small, copy the small elements into LHS and move the pointer
228 // from LHS to RHS.
229 if (!this->isSmall() && RHS.isSmall()) {
230 assert(RHS.CurArray == RHS.SmallArray);
231 std::copy(RHS.CurArray, RHS.CurArray + RHS.NumNonEmpty, this->SmallArray);
232 std::swap(RHS.CurArraySize, this->CurArraySize);
233 std::swap(this->NumNonEmpty, RHS.NumNonEmpty);
234 std::swap(this->NumTombstones, RHS.NumTombstones);
235 RHS.CurArray = this->CurArray;
236 this->CurArray = this->SmallArray;
237 return;
238 }
239
240 // If only LHS is small, copy the small elements into RHS and move the pointer
241 // from RHS to LHS.
242 if (this->isSmall() && !RHS.isSmall()) {
243 assert(this->CurArray == this->SmallArray);
244 std::copy(this->CurArray, this->CurArray + this->NumNonEmpty,
245 RHS.SmallArray);
246 std::swap(RHS.CurArraySize, this->CurArraySize);
247 std::swap(RHS.NumNonEmpty, this->NumNonEmpty);
248 std::swap(RHS.NumTombstones, this->NumTombstones);
249 this->CurArray = RHS.CurArray;
250 RHS.CurArray = RHS.SmallArray;
251 return;
252 }
253
254 // Both a small, just swap the small elements.
255 assert(this->isSmall() && RHS.isSmall());
256 unsigned MinNonEmpty = std::min(this->NumNonEmpty, RHS.NumNonEmpty);
257 std::swap_ranges(this->SmallArray, this->SmallArray + MinNonEmpty,
258 RHS.SmallArray);
259 if (this->NumNonEmpty > MinNonEmpty) {
260 std::copy(this->SmallArray + MinNonEmpty,
261 this->SmallArray + this->NumNonEmpty,
262 RHS.SmallArray + MinNonEmpty);
263 } else {
264 std::copy(RHS.SmallArray + MinNonEmpty, RHS.SmallArray + RHS.NumNonEmpty,
265 this->SmallArray + MinNonEmpty);
266 }
267 assert(this->CurArraySize == RHS.CurArraySize);
268 std::swap(this->NumNonEmpty, RHS.NumNonEmpty);
269 std::swap(this->NumTombstones, RHS.NumTombstones);
270 }
271