xref: /linux/kernel/bpf/bpf_lru_list.c (revision 23c48a124b469cee2eb0c75e6d22d366d1caa118)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /* Copyright (c) 2016 Facebook
3  */
4 #include <linux/cpumask.h>
5 #include <linux/spinlock.h>
6 #include <linux/percpu.h>
7 
8 #include "bpf_lru_list.h"
9 
10 #define LOCAL_FREE_TARGET		(128)
11 #define LOCAL_NR_SCANS			LOCAL_FREE_TARGET
12 
13 #define PERCPU_FREE_TARGET		(4)
14 #define PERCPU_NR_SCANS			PERCPU_FREE_TARGET
15 
16 /* Helpers to get the local list index */
17 #define LOCAL_LIST_IDX(t)	((t) - BPF_LOCAL_LIST_T_OFFSET)
18 #define LOCAL_FREE_LIST_IDX	LOCAL_LIST_IDX(BPF_LRU_LOCAL_LIST_T_FREE)
19 #define LOCAL_PENDING_LIST_IDX	LOCAL_LIST_IDX(BPF_LRU_LOCAL_LIST_T_PENDING)
20 #define IS_LOCAL_LIST_TYPE(t)	((t) >= BPF_LOCAL_LIST_T_OFFSET)
21 
22 static int get_next_cpu(int cpu)
23 {
24 	cpu = cpumask_next(cpu, cpu_possible_mask);
25 	if (cpu >= nr_cpu_ids)
26 		cpu = cpumask_first(cpu_possible_mask);
27 	return cpu;
28 }
29 
30 /* Local list helpers */
31 static struct list_head *local_free_list(struct bpf_lru_locallist *loc_l)
32 {
33 	return &loc_l->lists[LOCAL_FREE_LIST_IDX];
34 }
35 
36 static struct list_head *local_pending_list(struct bpf_lru_locallist *loc_l)
37 {
38 	return &loc_l->lists[LOCAL_PENDING_LIST_IDX];
39 }
40 
41 /* bpf_lru_node helpers */
42 static bool bpf_lru_node_is_ref(const struct bpf_lru_node *node)
43 {
44 	return node->ref;
45 }
46 
47 static void bpf_lru_list_count_inc(struct bpf_lru_list *l,
48 				   enum bpf_lru_list_type type)
49 {
50 	if (type < NR_BPF_LRU_LIST_COUNT)
51 		l->counts[type]++;
52 }
53 
54 static void bpf_lru_list_count_dec(struct bpf_lru_list *l,
55 				   enum bpf_lru_list_type type)
56 {
57 	if (type < NR_BPF_LRU_LIST_COUNT)
58 		l->counts[type]--;
59 }
60 
61 static void __bpf_lru_node_move_to_free(struct bpf_lru_list *l,
62 					struct bpf_lru_node *node,
63 					struct list_head *free_list,
64 					enum bpf_lru_list_type tgt_free_type)
65 {
66 	if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type)))
67 		return;
68 
69 	/* If the removing node is the next_inactive_rotation candidate,
70 	 * move the next_inactive_rotation pointer also.
71 	 */
72 	if (&node->list == l->next_inactive_rotation)
73 		l->next_inactive_rotation = l->next_inactive_rotation->prev;
74 
75 	bpf_lru_list_count_dec(l, node->type);
76 
77 	node->type = tgt_free_type;
78 	list_move(&node->list, free_list);
79 }
80 
81 /* Move nodes from local list to the LRU list */
82 static void __bpf_lru_node_move_in(struct bpf_lru_list *l,
83 				   struct bpf_lru_node *node,
84 				   enum bpf_lru_list_type tgt_type)
85 {
86 	if (WARN_ON_ONCE(!IS_LOCAL_LIST_TYPE(node->type)) ||
87 	    WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(tgt_type)))
88 		return;
89 
90 	bpf_lru_list_count_inc(l, tgt_type);
91 	node->type = tgt_type;
92 	node->ref = 0;
93 	list_move(&node->list, &l->lists[tgt_type]);
94 }
95 
96 /* Move nodes between or within active and inactive list (like
97  * active to inactive, inactive to active or tail of active back to
98  * the head of active).
99  */
100 static void __bpf_lru_node_move(struct bpf_lru_list *l,
101 				struct bpf_lru_node *node,
102 				enum bpf_lru_list_type tgt_type)
103 {
104 	if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type)) ||
105 	    WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(tgt_type)))
106 		return;
107 
108 	if (node->type != tgt_type) {
109 		bpf_lru_list_count_dec(l, node->type);
110 		bpf_lru_list_count_inc(l, tgt_type);
111 		node->type = tgt_type;
112 	}
113 	node->ref = 0;
114 
115 	/* If the moving node is the next_inactive_rotation candidate,
116 	 * move the next_inactive_rotation pointer also.
117 	 */
118 	if (&node->list == l->next_inactive_rotation)
119 		l->next_inactive_rotation = l->next_inactive_rotation->prev;
120 
121 	list_move(&node->list, &l->lists[tgt_type]);
122 }
123 
124 static bool bpf_lru_list_inactive_low(const struct bpf_lru_list *l)
125 {
126 	return l->counts[BPF_LRU_LIST_T_INACTIVE] <
127 		l->counts[BPF_LRU_LIST_T_ACTIVE];
128 }
129 
130 /* Rotate the active list:
131  * 1. Start from tail
132  * 2. If the node has the ref bit set, it will be rotated
133  *    back to the head of active list with the ref bit cleared.
134  *    Give this node one more chance to survive in the active list.
135  * 3. If the ref bit is not set, move it to the head of the
136  *    inactive list.
137  * 4. It will at most scan nr_scans nodes
138  */
139 static void __bpf_lru_list_rotate_active(struct bpf_lru *lru,
140 					 struct bpf_lru_list *l)
141 {
142 	struct list_head *active = &l->lists[BPF_LRU_LIST_T_ACTIVE];
143 	struct bpf_lru_node *node, *tmp_node, *first_node;
144 	unsigned int i = 0;
145 
146 	first_node = list_first_entry(active, struct bpf_lru_node, list);
147 	list_for_each_entry_safe_reverse(node, tmp_node, active, list) {
148 		if (bpf_lru_node_is_ref(node))
149 			__bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE);
150 		else
151 			__bpf_lru_node_move(l, node, BPF_LRU_LIST_T_INACTIVE);
152 
153 		if (++i == lru->nr_scans || node == first_node)
154 			break;
155 	}
156 }
157 
158 /* Rotate the inactive list.  It starts from the next_inactive_rotation
159  * 1. If the node has ref bit set, it will be moved to the head
160  *    of active list with the ref bit cleared.
161  * 2. If the node does not have ref bit set, it will leave it
162  *    at its current location (i.e. do nothing) so that it can
163  *    be considered during the next inactive_shrink.
164  * 3. It will at most scan nr_scans nodes
165  */
166 static void __bpf_lru_list_rotate_inactive(struct bpf_lru *lru,
167 					   struct bpf_lru_list *l)
168 {
169 	struct list_head *inactive = &l->lists[BPF_LRU_LIST_T_INACTIVE];
170 	struct list_head *cur, *last, *next = inactive;
171 	struct bpf_lru_node *node;
172 	unsigned int i = 0;
173 
174 	if (list_empty(inactive))
175 		return;
176 
177 	last = l->next_inactive_rotation->next;
178 	if (last == inactive)
179 		last = last->next;
180 
181 	cur = l->next_inactive_rotation;
182 	while (i < lru->nr_scans) {
183 		if (cur == inactive) {
184 			cur = cur->prev;
185 			continue;
186 		}
187 
188 		node = list_entry(cur, struct bpf_lru_node, list);
189 		next = cur->prev;
190 		if (bpf_lru_node_is_ref(node))
191 			__bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE);
192 		if (cur == last)
193 			break;
194 		cur = next;
195 		i++;
196 	}
197 
198 	l->next_inactive_rotation = next;
199 }
200 
201 /* Shrink the inactive list.  It starts from the tail of the
202  * inactive list and only move the nodes without the ref bit
203  * set to the designated free list.
204  */
205 static unsigned int
206 __bpf_lru_list_shrink_inactive(struct bpf_lru *lru,
207 			       struct bpf_lru_list *l,
208 			       unsigned int tgt_nshrink,
209 			       struct list_head *free_list,
210 			       enum bpf_lru_list_type tgt_free_type)
211 {
212 	struct list_head *inactive = &l->lists[BPF_LRU_LIST_T_INACTIVE];
213 	struct bpf_lru_node *node, *tmp_node;
214 	unsigned int nshrinked = 0;
215 	unsigned int i = 0;
216 
217 	list_for_each_entry_safe_reverse(node, tmp_node, inactive, list) {
218 		if (bpf_lru_node_is_ref(node)) {
219 			__bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE);
220 		} else if (lru->del_from_htab(lru->del_arg, node)) {
221 			__bpf_lru_node_move_to_free(l, node, free_list,
222 						    tgt_free_type);
223 			if (++nshrinked == tgt_nshrink)
224 				break;
225 		}
226 
227 		if (++i == lru->nr_scans)
228 			break;
229 	}
230 
231 	return nshrinked;
232 }
233 
234 /* 1. Rotate the active list (if needed)
235  * 2. Always rotate the inactive list
236  */
237 static void __bpf_lru_list_rotate(struct bpf_lru *lru, struct bpf_lru_list *l)
238 {
239 	if (bpf_lru_list_inactive_low(l))
240 		__bpf_lru_list_rotate_active(lru, l);
241 
242 	__bpf_lru_list_rotate_inactive(lru, l);
243 }
244 
245 /* Calls __bpf_lru_list_shrink_inactive() to shrink some
246  * ref-bit-cleared nodes and move them to the designated
247  * free list.
248  *
249  * If it cannot get a free node after calling
250  * __bpf_lru_list_shrink_inactive().  It will just remove
251  * one node from either inactive or active list without
252  * honoring the ref-bit.  It prefers inactive list to active
253  * list in this situation.
254  */
255 static unsigned int __bpf_lru_list_shrink(struct bpf_lru *lru,
256 					  struct bpf_lru_list *l,
257 					  unsigned int tgt_nshrink,
258 					  struct list_head *free_list,
259 					  enum bpf_lru_list_type tgt_free_type)
260 
261 {
262 	struct bpf_lru_node *node, *tmp_node;
263 	struct list_head *force_shrink_list;
264 	unsigned int nshrinked;
265 
266 	nshrinked = __bpf_lru_list_shrink_inactive(lru, l, tgt_nshrink,
267 						   free_list, tgt_free_type);
268 	if (nshrinked)
269 		return nshrinked;
270 
271 	/* Do a force shrink by ignoring the reference bit */
272 	if (!list_empty(&l->lists[BPF_LRU_LIST_T_INACTIVE]))
273 		force_shrink_list = &l->lists[BPF_LRU_LIST_T_INACTIVE];
274 	else
275 		force_shrink_list = &l->lists[BPF_LRU_LIST_T_ACTIVE];
276 
277 	list_for_each_entry_safe_reverse(node, tmp_node, force_shrink_list,
278 					 list) {
279 		if (lru->del_from_htab(lru->del_arg, node)) {
280 			__bpf_lru_node_move_to_free(l, node, free_list,
281 						    tgt_free_type);
282 			return 1;
283 		}
284 	}
285 
286 	return 0;
287 }
288 
289 /* Flush the nodes from the local pending list to the LRU list */
290 static void __local_list_flush(struct bpf_lru_list *l,
291 			       struct bpf_lru_locallist *loc_l)
292 {
293 	struct bpf_lru_node *node, *tmp_node;
294 
295 	list_for_each_entry_safe_reverse(node, tmp_node,
296 					 local_pending_list(loc_l), list) {
297 		if (bpf_lru_node_is_ref(node))
298 			__bpf_lru_node_move_in(l, node, BPF_LRU_LIST_T_ACTIVE);
299 		else
300 			__bpf_lru_node_move_in(l, node,
301 					       BPF_LRU_LIST_T_INACTIVE);
302 	}
303 }
304 
305 static void bpf_lru_list_push_free(struct bpf_lru_list *l,
306 				   struct bpf_lru_node *node)
307 {
308 	unsigned long flags;
309 
310 	if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type)))
311 		return;
312 
313 	raw_spin_lock_irqsave(&l->lock, flags);
314 	__bpf_lru_node_move(l, node, BPF_LRU_LIST_T_FREE);
315 	raw_spin_unlock_irqrestore(&l->lock, flags);
316 }
317 
318 static void bpf_lru_list_pop_free_to_local(struct bpf_lru *lru,
319 					   struct bpf_lru_locallist *loc_l)
320 {
321 	struct bpf_lru_list *l = &lru->common_lru.lru_list;
322 	struct bpf_lru_node *node, *tmp_node;
323 	unsigned int nfree = 0;
324 
325 	raw_spin_lock(&l->lock);
326 
327 	__local_list_flush(l, loc_l);
328 
329 	__bpf_lru_list_rotate(lru, l);
330 
331 	list_for_each_entry_safe(node, tmp_node, &l->lists[BPF_LRU_LIST_T_FREE],
332 				 list) {
333 		__bpf_lru_node_move_to_free(l, node, local_free_list(loc_l),
334 					    BPF_LRU_LOCAL_LIST_T_FREE);
335 		if (++nfree == LOCAL_FREE_TARGET)
336 			break;
337 	}
338 
339 	if (nfree < LOCAL_FREE_TARGET)
340 		__bpf_lru_list_shrink(lru, l, LOCAL_FREE_TARGET - nfree,
341 				      local_free_list(loc_l),
342 				      BPF_LRU_LOCAL_LIST_T_FREE);
343 
344 	raw_spin_unlock(&l->lock);
345 }
346 
347 static void __local_list_add_pending(struct bpf_lru *lru,
348 				     struct bpf_lru_locallist *loc_l,
349 				     int cpu,
350 				     struct bpf_lru_node *node,
351 				     u32 hash)
352 {
353 	*(u32 *)((void *)node + lru->hash_offset) = hash;
354 	node->cpu = cpu;
355 	node->type = BPF_LRU_LOCAL_LIST_T_PENDING;
356 	node->ref = 0;
357 	list_add(&node->list, local_pending_list(loc_l));
358 }
359 
360 static struct bpf_lru_node *
361 __local_list_pop_free(struct bpf_lru_locallist *loc_l)
362 {
363 	struct bpf_lru_node *node;
364 
365 	node = list_first_entry_or_null(local_free_list(loc_l),
366 					struct bpf_lru_node,
367 					list);
368 	if (node)
369 		list_del(&node->list);
370 
371 	return node;
372 }
373 
374 static struct bpf_lru_node *
375 __local_list_pop_pending(struct bpf_lru *lru, struct bpf_lru_locallist *loc_l)
376 {
377 	struct bpf_lru_node *node;
378 	bool force = false;
379 
380 ignore_ref:
381 	/* Get from the tail (i.e. older element) of the pending list. */
382 	list_for_each_entry_reverse(node, local_pending_list(loc_l),
383 				    list) {
384 		if ((!bpf_lru_node_is_ref(node) || force) &&
385 		    lru->del_from_htab(lru->del_arg, node)) {
386 			list_del(&node->list);
387 			return node;
388 		}
389 	}
390 
391 	if (!force) {
392 		force = true;
393 		goto ignore_ref;
394 	}
395 
396 	return NULL;
397 }
398 
399 static struct bpf_lru_node *bpf_percpu_lru_pop_free(struct bpf_lru *lru,
400 						    u32 hash)
401 {
402 	struct list_head *free_list;
403 	struct bpf_lru_node *node = NULL;
404 	struct bpf_lru_list *l;
405 	unsigned long flags;
406 	int cpu = raw_smp_processor_id();
407 
408 	l = per_cpu_ptr(lru->percpu_lru, cpu);
409 
410 	raw_spin_lock_irqsave(&l->lock, flags);
411 
412 	__bpf_lru_list_rotate(lru, l);
413 
414 	free_list = &l->lists[BPF_LRU_LIST_T_FREE];
415 	if (list_empty(free_list))
416 		__bpf_lru_list_shrink(lru, l, PERCPU_FREE_TARGET, free_list,
417 				      BPF_LRU_LIST_T_FREE);
418 
419 	if (!list_empty(free_list)) {
420 		node = list_first_entry(free_list, struct bpf_lru_node, list);
421 		*(u32 *)((void *)node + lru->hash_offset) = hash;
422 		node->ref = 0;
423 		__bpf_lru_node_move(l, node, BPF_LRU_LIST_T_INACTIVE);
424 	}
425 
426 	raw_spin_unlock_irqrestore(&l->lock, flags);
427 
428 	return node;
429 }
430 
431 static struct bpf_lru_node *bpf_common_lru_pop_free(struct bpf_lru *lru,
432 						    u32 hash)
433 {
434 	struct bpf_lru_locallist *loc_l, *steal_loc_l;
435 	struct bpf_common_lru *clru = &lru->common_lru;
436 	struct bpf_lru_node *node;
437 	int steal, first_steal;
438 	unsigned long flags;
439 	int cpu = raw_smp_processor_id();
440 
441 	loc_l = per_cpu_ptr(clru->local_list, cpu);
442 
443 	raw_spin_lock_irqsave(&loc_l->lock, flags);
444 
445 	node = __local_list_pop_free(loc_l);
446 	if (!node) {
447 		bpf_lru_list_pop_free_to_local(lru, loc_l);
448 		node = __local_list_pop_free(loc_l);
449 	}
450 
451 	if (node)
452 		__local_list_add_pending(lru, loc_l, cpu, node, hash);
453 
454 	raw_spin_unlock_irqrestore(&loc_l->lock, flags);
455 
456 	if (node)
457 		return node;
458 
459 	/* No free nodes found from the local free list and
460 	 * the global LRU list.
461 	 *
462 	 * Steal from the local free/pending list of the
463 	 * current CPU and remote CPU in RR.  It starts
464 	 * with the loc_l->next_steal CPU.
465 	 */
466 
467 	first_steal = loc_l->next_steal;
468 	steal = first_steal;
469 	do {
470 		steal_loc_l = per_cpu_ptr(clru->local_list, steal);
471 
472 		raw_spin_lock_irqsave(&steal_loc_l->lock, flags);
473 
474 		node = __local_list_pop_free(steal_loc_l);
475 		if (!node)
476 			node = __local_list_pop_pending(lru, steal_loc_l);
477 
478 		raw_spin_unlock_irqrestore(&steal_loc_l->lock, flags);
479 
480 		steal = get_next_cpu(steal);
481 	} while (!node && steal != first_steal);
482 
483 	loc_l->next_steal = steal;
484 
485 	if (node) {
486 		raw_spin_lock_irqsave(&loc_l->lock, flags);
487 		__local_list_add_pending(lru, loc_l, cpu, node, hash);
488 		raw_spin_unlock_irqrestore(&loc_l->lock, flags);
489 	}
490 
491 	return node;
492 }
493 
494 struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash)
495 {
496 	if (lru->percpu)
497 		return bpf_percpu_lru_pop_free(lru, hash);
498 	else
499 		return bpf_common_lru_pop_free(lru, hash);
500 }
501 
502 static void bpf_common_lru_push_free(struct bpf_lru *lru,
503 				     struct bpf_lru_node *node)
504 {
505 	u8 node_type = READ_ONCE(node->type);
506 	unsigned long flags;
507 
508 	if (WARN_ON_ONCE(node_type == BPF_LRU_LIST_T_FREE) ||
509 	    WARN_ON_ONCE(node_type == BPF_LRU_LOCAL_LIST_T_FREE))
510 		return;
511 
512 	if (node_type == BPF_LRU_LOCAL_LIST_T_PENDING) {
513 		struct bpf_lru_locallist *loc_l;
514 
515 		loc_l = per_cpu_ptr(lru->common_lru.local_list, node->cpu);
516 
517 		raw_spin_lock_irqsave(&loc_l->lock, flags);
518 
519 		if (unlikely(node->type != BPF_LRU_LOCAL_LIST_T_PENDING)) {
520 			raw_spin_unlock_irqrestore(&loc_l->lock, flags);
521 			goto check_lru_list;
522 		}
523 
524 		node->type = BPF_LRU_LOCAL_LIST_T_FREE;
525 		node->ref = 0;
526 		list_move(&node->list, local_free_list(loc_l));
527 
528 		raw_spin_unlock_irqrestore(&loc_l->lock, flags);
529 		return;
530 	}
531 
532 check_lru_list:
533 	bpf_lru_list_push_free(&lru->common_lru.lru_list, node);
534 }
535 
536 static void bpf_percpu_lru_push_free(struct bpf_lru *lru,
537 				     struct bpf_lru_node *node)
538 {
539 	struct bpf_lru_list *l;
540 	unsigned long flags;
541 
542 	l = per_cpu_ptr(lru->percpu_lru, node->cpu);
543 
544 	raw_spin_lock_irqsave(&l->lock, flags);
545 
546 	__bpf_lru_node_move(l, node, BPF_LRU_LIST_T_FREE);
547 
548 	raw_spin_unlock_irqrestore(&l->lock, flags);
549 }
550 
551 void bpf_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node)
552 {
553 	if (lru->percpu)
554 		bpf_percpu_lru_push_free(lru, node);
555 	else
556 		bpf_common_lru_push_free(lru, node);
557 }
558 
559 static void bpf_common_lru_populate(struct bpf_lru *lru, void *buf,
560 				    u32 node_offset, u32 elem_size,
561 				    u32 nr_elems)
562 {
563 	struct bpf_lru_list *l = &lru->common_lru.lru_list;
564 	u32 i;
565 
566 	for (i = 0; i < nr_elems; i++) {
567 		struct bpf_lru_node *node;
568 
569 		node = (struct bpf_lru_node *)(buf + node_offset);
570 		node->type = BPF_LRU_LIST_T_FREE;
571 		node->ref = 0;
572 		list_add(&node->list, &l->lists[BPF_LRU_LIST_T_FREE]);
573 		buf += elem_size;
574 	}
575 }
576 
577 static void bpf_percpu_lru_populate(struct bpf_lru *lru, void *buf,
578 				    u32 node_offset, u32 elem_size,
579 				    u32 nr_elems)
580 {
581 	u32 i, pcpu_entries;
582 	int cpu;
583 	struct bpf_lru_list *l;
584 
585 	pcpu_entries = nr_elems / num_possible_cpus();
586 
587 	i = 0;
588 
589 	for_each_possible_cpu(cpu) {
590 		struct bpf_lru_node *node;
591 
592 		l = per_cpu_ptr(lru->percpu_lru, cpu);
593 again:
594 		node = (struct bpf_lru_node *)(buf + node_offset);
595 		node->cpu = cpu;
596 		node->type = BPF_LRU_LIST_T_FREE;
597 		node->ref = 0;
598 		list_add(&node->list, &l->lists[BPF_LRU_LIST_T_FREE]);
599 		i++;
600 		buf += elem_size;
601 		if (i == nr_elems)
602 			break;
603 		if (i % pcpu_entries)
604 			goto again;
605 	}
606 }
607 
608 void bpf_lru_populate(struct bpf_lru *lru, void *buf, u32 node_offset,
609 		      u32 elem_size, u32 nr_elems)
610 {
611 	if (lru->percpu)
612 		bpf_percpu_lru_populate(lru, buf, node_offset, elem_size,
613 					nr_elems);
614 	else
615 		bpf_common_lru_populate(lru, buf, node_offset, elem_size,
616 					nr_elems);
617 }
618 
619 static void bpf_lru_locallist_init(struct bpf_lru_locallist *loc_l, int cpu)
620 {
621 	int i;
622 
623 	for (i = 0; i < NR_BPF_LRU_LOCAL_LIST_T; i++)
624 		INIT_LIST_HEAD(&loc_l->lists[i]);
625 
626 	loc_l->next_steal = cpu;
627 
628 	raw_spin_lock_init(&loc_l->lock);
629 }
630 
631 static void bpf_lru_list_init(struct bpf_lru_list *l)
632 {
633 	int i;
634 
635 	for (i = 0; i < NR_BPF_LRU_LIST_T; i++)
636 		INIT_LIST_HEAD(&l->lists[i]);
637 
638 	for (i = 0; i < NR_BPF_LRU_LIST_COUNT; i++)
639 		l->counts[i] = 0;
640 
641 	l->next_inactive_rotation = &l->lists[BPF_LRU_LIST_T_INACTIVE];
642 
643 	raw_spin_lock_init(&l->lock);
644 }
645 
646 int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset,
647 		 del_from_htab_func del_from_htab, void *del_arg)
648 {
649 	int cpu;
650 
651 	if (percpu) {
652 		lru->percpu_lru = alloc_percpu(struct bpf_lru_list);
653 		if (!lru->percpu_lru)
654 			return -ENOMEM;
655 
656 		for_each_possible_cpu(cpu) {
657 			struct bpf_lru_list *l;
658 
659 			l = per_cpu_ptr(lru->percpu_lru, cpu);
660 			bpf_lru_list_init(l);
661 		}
662 		lru->nr_scans = PERCPU_NR_SCANS;
663 	} else {
664 		struct bpf_common_lru *clru = &lru->common_lru;
665 
666 		clru->local_list = alloc_percpu(struct bpf_lru_locallist);
667 		if (!clru->local_list)
668 			return -ENOMEM;
669 
670 		for_each_possible_cpu(cpu) {
671 			struct bpf_lru_locallist *loc_l;
672 
673 			loc_l = per_cpu_ptr(clru->local_list, cpu);
674 			bpf_lru_locallist_init(loc_l, cpu);
675 		}
676 
677 		bpf_lru_list_init(&clru->lru_list);
678 		lru->nr_scans = LOCAL_NR_SCANS;
679 	}
680 
681 	lru->percpu = percpu;
682 	lru->del_from_htab = del_from_htab;
683 	lru->del_arg = del_arg;
684 	lru->hash_offset = hash_offset;
685 
686 	return 0;
687 }
688 
689 void bpf_lru_destroy(struct bpf_lru *lru)
690 {
691 	if (lru->percpu)
692 		free_percpu(lru->percpu_lru);
693 	else
694 		free_percpu(lru->common_lru.local_list);
695 }
696