1 /*-
2 * SPDX-License-Identifier: BSD-2-Clause
3 *
4 * Copyright (c) 2017, Jeffrey Roberson <jeff@freebsd.org>
5 * All rights reserved.
6 *
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
9 * are met:
10 * 1. Redistributions of source code must retain the above copyright
11 * notice unmodified, this list of conditions, and the following
12 * disclaimer.
13 * 2. Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in the
15 * documentation and/or other materials provided with the distribution.
16 *
17 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
18 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
19 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
20 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
21 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
22 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
26 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27 *
28 */
29
30 #include <sys/cdefs.h>
31 #include "opt_vm.h"
32
33 #include <sys/param.h>
34 #include <sys/systm.h>
35 #include <sys/bitset.h>
36 #include <sys/domainset.h>
37 #include <sys/proc.h>
38 #include <sys/lock.h>
39 #include <sys/mutex.h>
40 #include <sys/malloc.h>
41 #include <sys/rwlock.h>
42 #include <sys/pctrie.h>
43 #include <sys/vmmeter.h>
44
45 #include <vm/vm.h>
46 #include <vm/vm_param.h>
47 #include <vm/vm_domainset.h>
48 #include <vm/vm_object.h>
49 #include <vm/vm_page.h>
50 #include <vm/vm_phys.h>
51
52 #ifdef NUMA
53 /*
54 * Iterators are written such that the first nowait pass has as short a
55 * codepath as possible to eliminate bloat from the allocator. It is
56 * assumed that most allocations are successful.
57 */
58
59 static int vm_domainset_default_stride = 64;
60
61 /*
62 * Determine which policy is to be used for this allocation.
63 */
64 static void
vm_domainset_iter_init(struct vm_domainset_iter * di,struct domainset * ds,int * iter,struct vm_object * obj,vm_pindex_t pindex)65 vm_domainset_iter_init(struct vm_domainset_iter *di, struct domainset *ds,
66 int *iter, struct vm_object *obj, vm_pindex_t pindex)
67 {
68
69 di->di_domain = ds;
70 di->di_iter = iter;
71 di->di_policy = ds->ds_policy;
72 DOMAINSET_COPY(&ds->ds_mask, &di->di_valid_mask);
73 if (di->di_policy == DOMAINSET_POLICY_INTERLEAVE) {
74 #if VM_NRESERVLEVEL > 0
75 if (vm_object_reserv(obj)) {
76 /*
77 * Color the pindex so we end up on the correct
78 * reservation boundary.
79 */
80 pindex += obj->pg_color;
81 #if VM_NRESERVLEVEL > 1
82 pindex >>= VM_LEVEL_1_ORDER;
83 #endif
84 pindex >>= VM_LEVEL_0_ORDER;
85 } else
86 #endif
87 pindex /= vm_domainset_default_stride;
88 /*
89 * Offset pindex so the first page of each object does
90 * not end up in domain 0.
91 */
92 if (obj != NULL)
93 pindex += (((uintptr_t)obj) / sizeof(*obj));
94 di->di_offset = pindex;
95 }
96 /* Skip domains below min on the first pass. */
97 di->di_minskip = true;
98 }
99
100 static void
vm_domainset_iter_rr(struct vm_domainset_iter * di,int * domain)101 vm_domainset_iter_rr(struct vm_domainset_iter *di, int *domain)
102 {
103
104 *domain = di->di_domain->ds_order[
105 ++(*di->di_iter) % di->di_domain->ds_cnt];
106 }
107
108 static void
vm_domainset_iter_prefer(struct vm_domainset_iter * di,int * domain)109 vm_domainset_iter_prefer(struct vm_domainset_iter *di, int *domain)
110 {
111 int d;
112
113 do {
114 d = di->di_domain->ds_order[
115 ++(*di->di_iter) % di->di_domain->ds_cnt];
116 } while (d == di->di_domain->ds_prefer);
117 *domain = d;
118 }
119
120 static void
vm_domainset_iter_interleave(struct vm_domainset_iter * di,int * domain)121 vm_domainset_iter_interleave(struct vm_domainset_iter *di, int *domain)
122 {
123 int d;
124
125 d = di->di_offset % di->di_domain->ds_cnt;
126 *di->di_iter = d;
127 *domain = di->di_domain->ds_order[d];
128 }
129
130 static void
vm_domainset_iter_next(struct vm_domainset_iter * di,int * domain)131 vm_domainset_iter_next(struct vm_domainset_iter *di, int *domain)
132 {
133
134 KASSERT(di->di_n > 0, ("%s: Invalid n %d", __func__, di->di_n));
135 switch (di->di_policy) {
136 case DOMAINSET_POLICY_FIRSTTOUCH:
137 /*
138 * To prevent impossible allocations we convert an invalid
139 * first-touch to round-robin.
140 */
141 /* FALLTHROUGH */
142 case DOMAINSET_POLICY_INTERLEAVE:
143 /* FALLTHROUGH */
144 case DOMAINSET_POLICY_ROUNDROBIN:
145 vm_domainset_iter_rr(di, domain);
146 break;
147 case DOMAINSET_POLICY_PREFER:
148 vm_domainset_iter_prefer(di, domain);
149 break;
150 default:
151 panic("%s: Unknown policy %d", __func__, di->di_policy);
152 }
153 KASSERT(*domain < vm_ndomains,
154 ("%s: Invalid domain %d", __func__, *domain));
155 }
156
157 static void
vm_domainset_iter_first(struct vm_domainset_iter * di,int * domain)158 vm_domainset_iter_first(struct vm_domainset_iter *di, int *domain)
159 {
160
161 switch (di->di_policy) {
162 case DOMAINSET_POLICY_FIRSTTOUCH:
163 *domain = PCPU_GET(domain);
164 if (DOMAINSET_ISSET(*domain, &di->di_valid_mask)) {
165 /*
166 * Add an extra iteration because we will visit the
167 * current domain a second time in the rr iterator.
168 */
169 di->di_n = di->di_domain->ds_cnt + 1;
170 break;
171 }
172 /*
173 * To prevent impossible allocations we convert an invalid
174 * first-touch to round-robin.
175 */
176 /* FALLTHROUGH */
177 case DOMAINSET_POLICY_ROUNDROBIN:
178 di->di_n = di->di_domain->ds_cnt;
179 vm_domainset_iter_rr(di, domain);
180 break;
181 case DOMAINSET_POLICY_PREFER:
182 *domain = di->di_domain->ds_prefer;
183 di->di_n = di->di_domain->ds_cnt;
184 break;
185 case DOMAINSET_POLICY_INTERLEAVE:
186 vm_domainset_iter_interleave(di, domain);
187 di->di_n = di->di_domain->ds_cnt;
188 break;
189 default:
190 panic("%s: Unknown policy %d", __func__, di->di_policy);
191 }
192 KASSERT(di->di_n > 0, ("%s: Invalid n %d", __func__, di->di_n));
193 KASSERT(*domain < vm_ndomains,
194 ("%s: Invalid domain %d", __func__, *domain));
195 }
196
197 void
vm_domainset_iter_page_init(struct vm_domainset_iter * di,struct vm_object * obj,vm_pindex_t pindex,int * domain,int * req,struct pctrie_iter * pages)198 vm_domainset_iter_page_init(struct vm_domainset_iter *di, struct vm_object *obj,
199 vm_pindex_t pindex, int *domain, int *req, struct pctrie_iter *pages)
200 {
201 struct domainset_ref *dr;
202
203 /*
204 * Object policy takes precedence over thread policy. The policies
205 * are immutable and unsynchronized. Updates can race but pointer
206 * loads are assumed to be atomic.
207 */
208 if (obj != NULL && obj->domain.dr_policy != NULL)
209 dr = &obj->domain;
210 else
211 dr = &curthread->td_domain;
212 vm_domainset_iter_init(di, dr->dr_policy, &dr->dr_iter, obj, pindex);
213 di->di_flags = *req;
214 *req = (di->di_flags & ~(VM_ALLOC_WAITOK | VM_ALLOC_WAITFAIL)) |
215 VM_ALLOC_NOWAIT;
216 vm_domainset_iter_first(di, domain);
217 if (vm_page_count_min_domain(*domain))
218 vm_domainset_iter_page(di, obj, domain, pages);
219 }
220
221 int
vm_domainset_iter_page(struct vm_domainset_iter * di,struct vm_object * obj,int * domain,struct pctrie_iter * pages)222 vm_domainset_iter_page(struct vm_domainset_iter *di, struct vm_object *obj,
223 int *domain, struct pctrie_iter *pages)
224 {
225 if (__predict_false(DOMAINSET_EMPTY(&di->di_valid_mask)))
226 return (ENOMEM);
227
228 /* If there are more domains to visit we run the iterator. */
229 while (--di->di_n != 0) {
230 vm_domainset_iter_next(di, domain);
231 if (DOMAINSET_ISSET(*domain, &di->di_valid_mask) &&
232 (!di->di_minskip || !vm_page_count_min_domain(*domain)))
233 return (0);
234 }
235
236 /* If we skipped domains below min restart the search. */
237 if (di->di_minskip) {
238 di->di_minskip = false;
239 vm_domainset_iter_first(di, domain);
240 return (0);
241 }
242
243 /* If we visited all domains and this was a NOWAIT we return error. */
244 if ((di->di_flags & (VM_ALLOC_WAITOK | VM_ALLOC_WAITFAIL)) == 0)
245 return (ENOMEM);
246
247 /* Wait for one of the domains to accumulate some free pages. */
248 if (obj != NULL) {
249 VM_OBJECT_WUNLOCK(obj);
250 if (pages != NULL)
251 pctrie_iter_reset(pages);
252 }
253 vm_wait_doms(&di->di_valid_mask, 0);
254 if (obj != NULL)
255 VM_OBJECT_WLOCK(obj);
256 if ((di->di_flags & VM_ALLOC_WAITFAIL) != 0)
257 return (ENOMEM);
258
259 /* Restart the search. */
260 vm_domainset_iter_first(di, domain);
261
262 return (0);
263 }
264
265 static void
_vm_domainset_iter_policy_init(struct vm_domainset_iter * di,int * domain,int * flags)266 _vm_domainset_iter_policy_init(struct vm_domainset_iter *di, int *domain,
267 int *flags)
268 {
269
270 di->di_flags = *flags;
271 *flags = (di->di_flags & ~M_WAITOK) | M_NOWAIT;
272 vm_domainset_iter_first(di, domain);
273 if (vm_page_count_min_domain(*domain))
274 vm_domainset_iter_policy(di, domain);
275 }
276
277 void
vm_domainset_iter_policy_init(struct vm_domainset_iter * di,struct domainset * ds,int * domain,int * flags)278 vm_domainset_iter_policy_init(struct vm_domainset_iter *di,
279 struct domainset *ds, int *domain, int *flags)
280 {
281
282 vm_domainset_iter_init(di, ds, &curthread->td_domain.dr_iter, NULL, 0);
283 _vm_domainset_iter_policy_init(di, domain, flags);
284 }
285
286 void
vm_domainset_iter_policy_ref_init(struct vm_domainset_iter * di,struct domainset_ref * dr,int * domain,int * flags)287 vm_domainset_iter_policy_ref_init(struct vm_domainset_iter *di,
288 struct domainset_ref *dr, int *domain, int *flags)
289 {
290
291 vm_domainset_iter_init(di, dr->dr_policy, &dr->dr_iter, NULL, 0);
292 _vm_domainset_iter_policy_init(di, domain, flags);
293 }
294
295 int
vm_domainset_iter_policy(struct vm_domainset_iter * di,int * domain)296 vm_domainset_iter_policy(struct vm_domainset_iter *di, int *domain)
297 {
298 if (DOMAINSET_EMPTY(&di->di_valid_mask))
299 return (ENOMEM);
300
301 /* If there are more domains to visit we run the iterator. */
302 while (--di->di_n != 0) {
303 vm_domainset_iter_next(di, domain);
304 if (DOMAINSET_ISSET(*domain, &di->di_valid_mask) &&
305 (!di->di_minskip || !vm_page_count_min_domain(*domain)))
306 return (0);
307 }
308
309 /* If we skipped domains below min restart the search. */
310 if (di->di_minskip) {
311 di->di_minskip = false;
312 vm_domainset_iter_first(di, domain);
313 return (0);
314 }
315
316 /* If we visited all domains and this was a NOWAIT we return error. */
317 if ((di->di_flags & M_WAITOK) == 0)
318 return (ENOMEM);
319
320 /* Wait for one of the domains to accumulate some free pages. */
321 vm_wait_doms(&di->di_valid_mask, 0);
322
323 /* Restart the search. */
324 vm_domainset_iter_first(di, domain);
325
326 return (0);
327 }
328
329 void
vm_domainset_iter_ignore(struct vm_domainset_iter * di,int domain)330 vm_domainset_iter_ignore(struct vm_domainset_iter *di, int domain)
331 {
332 KASSERT(DOMAINSET_ISSET(domain, &di->di_valid_mask),
333 ("%s: domain %d not present in di_valid_mask for di %p",
334 __func__, domain, di));
335 DOMAINSET_CLR(domain, &di->di_valid_mask);
336 }
337
338 #else /* !NUMA */
339
340 int
vm_domainset_iter_page(struct vm_domainset_iter * di,struct vm_object * obj,int * domain,struct pctrie_iter * pages)341 vm_domainset_iter_page(struct vm_domainset_iter *di, struct vm_object *obj,
342 int *domain, struct pctrie_iter *pages)
343 {
344
345 return (EJUSTRETURN);
346 }
347
348 void
vm_domainset_iter_page_init(struct vm_domainset_iter * di,struct vm_object * obj,vm_pindex_t pindex,int * domain,int * flags,struct pctrie_iter * pages)349 vm_domainset_iter_page_init(struct vm_domainset_iter *di, struct vm_object *obj,
350 vm_pindex_t pindex, int *domain, int *flags, struct pctrie_iter *pages)
351 {
352
353 *domain = 0;
354 }
355
356 int
vm_domainset_iter_policy(struct vm_domainset_iter * di,int * domain)357 vm_domainset_iter_policy(struct vm_domainset_iter *di, int *domain)
358 {
359
360 return (EJUSTRETURN);
361 }
362
363 void
vm_domainset_iter_policy_init(struct vm_domainset_iter * di,struct domainset * ds,int * domain,int * flags)364 vm_domainset_iter_policy_init(struct vm_domainset_iter *di,
365 struct domainset *ds, int *domain, int *flags)
366 {
367
368 *domain = 0;
369 }
370
371 void
vm_domainset_iter_policy_ref_init(struct vm_domainset_iter * di,struct domainset_ref * dr,int * domain,int * flags)372 vm_domainset_iter_policy_ref_init(struct vm_domainset_iter *di,
373 struct domainset_ref *dr, int *domain, int *flags)
374 {
375
376 *domain = 0;
377 }
378
379 void
vm_domainset_iter_ignore(struct vm_domainset_iter * di __unused,int domain __unused)380 vm_domainset_iter_ignore(struct vm_domainset_iter *di __unused,
381 int domain __unused)
382 {
383 }
384
385 #endif /* NUMA */
386