1 /* Copyright (c) 2016 Facebook 2 * 3 * This program is free software; you can redistribute it and/or 4 * modify it under the terms of version 2 of the GNU General Public 5 * License as published by the Free Software Foundation. 6 */ 7 #include <linux/cpumask.h> 8 #include <linux/spinlock.h> 9 #include <linux/percpu.h> 10 11 #include "bpf_lru_list.h" 12 13 #define LOCAL_FREE_TARGET (128) 14 #define LOCAL_NR_SCANS LOCAL_FREE_TARGET 15 16 #define PERCPU_FREE_TARGET (4) 17 #define PERCPU_NR_SCANS PERCPU_FREE_TARGET 18 19 /* Helpers to get the local list index */ 20 #define LOCAL_LIST_IDX(t) ((t) - BPF_LOCAL_LIST_T_OFFSET) 21 #define LOCAL_FREE_LIST_IDX LOCAL_LIST_IDX(BPF_LRU_LOCAL_LIST_T_FREE) 22 #define LOCAL_PENDING_LIST_IDX LOCAL_LIST_IDX(BPF_LRU_LOCAL_LIST_T_PENDING) 23 #define IS_LOCAL_LIST_TYPE(t) ((t) >= BPF_LOCAL_LIST_T_OFFSET) 24 25 static int get_next_cpu(int cpu) 26 { 27 cpu = cpumask_next(cpu, cpu_possible_mask); 28 if (cpu >= nr_cpu_ids) 29 cpu = cpumask_first(cpu_possible_mask); 30 return cpu; 31 } 32 33 /* Local list helpers */ 34 static struct list_head *local_free_list(struct bpf_lru_locallist *loc_l) 35 { 36 return &loc_l->lists[LOCAL_FREE_LIST_IDX]; 37 } 38 39 static struct list_head *local_pending_list(struct bpf_lru_locallist *loc_l) 40 { 41 return &loc_l->lists[LOCAL_PENDING_LIST_IDX]; 42 } 43 44 /* bpf_lru_node helpers */ 45 static bool bpf_lru_node_is_ref(const struct bpf_lru_node *node) 46 { 47 return node->ref; 48 } 49 50 static void bpf_lru_list_count_inc(struct bpf_lru_list *l, 51 enum bpf_lru_list_type type) 52 { 53 if (type < NR_BPF_LRU_LIST_COUNT) 54 l->counts[type]++; 55 } 56 57 static void bpf_lru_list_count_dec(struct bpf_lru_list *l, 58 enum bpf_lru_list_type type) 59 { 60 if (type < NR_BPF_LRU_LIST_COUNT) 61 l->counts[type]--; 62 } 63 64 static void __bpf_lru_node_move_to_free(struct bpf_lru_list *l, 65 struct bpf_lru_node *node, 66 struct list_head *free_list, 67 enum bpf_lru_list_type tgt_free_type) 68 { 69 if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type))) 70 return; 71 72 /* If the removing node is the next_inactive_rotation candidate, 73 * move the next_inactive_rotation pointer also. 74 */ 75 if (&node->list == l->next_inactive_rotation) 76 l->next_inactive_rotation = l->next_inactive_rotation->prev; 77 78 bpf_lru_list_count_dec(l, node->type); 79 80 node->type = tgt_free_type; 81 list_move(&node->list, free_list); 82 } 83 84 /* Move nodes from local list to the LRU list */ 85 static void __bpf_lru_node_move_in(struct bpf_lru_list *l, 86 struct bpf_lru_node *node, 87 enum bpf_lru_list_type tgt_type) 88 { 89 if (WARN_ON_ONCE(!IS_LOCAL_LIST_TYPE(node->type)) || 90 WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(tgt_type))) 91 return; 92 93 bpf_lru_list_count_inc(l, tgt_type); 94 node->type = tgt_type; 95 node->ref = 0; 96 list_move(&node->list, &l->lists[tgt_type]); 97 } 98 99 /* Move nodes between or within active and inactive list (like 100 * active to inactive, inactive to active or tail of active back to 101 * the head of active). 102 */ 103 static void __bpf_lru_node_move(struct bpf_lru_list *l, 104 struct bpf_lru_node *node, 105 enum bpf_lru_list_type tgt_type) 106 { 107 if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type)) || 108 WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(tgt_type))) 109 return; 110 111 if (node->type != tgt_type) { 112 bpf_lru_list_count_dec(l, node->type); 113 bpf_lru_list_count_inc(l, tgt_type); 114 node->type = tgt_type; 115 } 116 node->ref = 0; 117 118 /* If the moving node is the next_inactive_rotation candidate, 119 * move the next_inactive_rotation pointer also. 120 */ 121 if (&node->list == l->next_inactive_rotation) 122 l->next_inactive_rotation = l->next_inactive_rotation->prev; 123 124 list_move(&node->list, &l->lists[tgt_type]); 125 } 126 127 static bool bpf_lru_list_inactive_low(const struct bpf_lru_list *l) 128 { 129 return l->counts[BPF_LRU_LIST_T_INACTIVE] < 130 l->counts[BPF_LRU_LIST_T_ACTIVE]; 131 } 132 133 /* Rotate the active list: 134 * 1. Start from tail 135 * 2. If the node has the ref bit set, it will be rotated 136 * back to the head of active list with the ref bit cleared. 137 * Give this node one more chance to survive in the active list. 138 * 3. If the ref bit is not set, move it to the head of the 139 * inactive list. 140 * 4. It will at most scan nr_scans nodes 141 */ 142 static void __bpf_lru_list_rotate_active(struct bpf_lru *lru, 143 struct bpf_lru_list *l) 144 { 145 struct list_head *active = &l->lists[BPF_LRU_LIST_T_ACTIVE]; 146 struct bpf_lru_node *node, *tmp_node, *first_node; 147 unsigned int i = 0; 148 149 first_node = list_first_entry(active, struct bpf_lru_node, list); 150 list_for_each_entry_safe_reverse(node, tmp_node, active, list) { 151 if (bpf_lru_node_is_ref(node)) 152 __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE); 153 else 154 __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_INACTIVE); 155 156 if (++i == lru->nr_scans || node == first_node) 157 break; 158 } 159 } 160 161 /* Rotate the inactive list. It starts from the next_inactive_rotation 162 * 1. If the node has ref bit set, it will be moved to the head 163 * of active list with the ref bit cleared. 164 * 2. If the node does not have ref bit set, it will leave it 165 * at its current location (i.e. do nothing) so that it can 166 * be considered during the next inactive_shrink. 167 * 3. It will at most scan nr_scans nodes 168 */ 169 static void __bpf_lru_list_rotate_inactive(struct bpf_lru *lru, 170 struct bpf_lru_list *l) 171 { 172 struct list_head *inactive = &l->lists[BPF_LRU_LIST_T_INACTIVE]; 173 struct list_head *cur, *last, *next = inactive; 174 struct bpf_lru_node *node; 175 unsigned int i = 0; 176 177 if (list_empty(inactive)) 178 return; 179 180 last = l->next_inactive_rotation->next; 181 if (last == inactive) 182 last = last->next; 183 184 cur = l->next_inactive_rotation; 185 while (i < lru->nr_scans) { 186 if (cur == inactive) { 187 cur = cur->prev; 188 continue; 189 } 190 191 node = list_entry(cur, struct bpf_lru_node, list); 192 next = cur->prev; 193 if (bpf_lru_node_is_ref(node)) 194 __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE); 195 if (cur == last) 196 break; 197 cur = next; 198 i++; 199 } 200 201 l->next_inactive_rotation = next; 202 } 203 204 /* Shrink the inactive list. It starts from the tail of the 205 * inactive list and only move the nodes without the ref bit 206 * set to the designated free list. 207 */ 208 static unsigned int 209 __bpf_lru_list_shrink_inactive(struct bpf_lru *lru, 210 struct bpf_lru_list *l, 211 unsigned int tgt_nshrink, 212 struct list_head *free_list, 213 enum bpf_lru_list_type tgt_free_type) 214 { 215 struct list_head *inactive = &l->lists[BPF_LRU_LIST_T_INACTIVE]; 216 struct bpf_lru_node *node, *tmp_node; 217 unsigned int nshrinked = 0; 218 unsigned int i = 0; 219 220 list_for_each_entry_safe_reverse(node, tmp_node, inactive, list) { 221 if (bpf_lru_node_is_ref(node)) { 222 __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE); 223 } else if (lru->del_from_htab(lru->del_arg, node)) { 224 __bpf_lru_node_move_to_free(l, node, free_list, 225 tgt_free_type); 226 if (++nshrinked == tgt_nshrink) 227 break; 228 } 229 230 if (++i == lru->nr_scans) 231 break; 232 } 233 234 return nshrinked; 235 } 236 237 /* 1. Rotate the active list (if needed) 238 * 2. Always rotate the inactive list 239 */ 240 static void __bpf_lru_list_rotate(struct bpf_lru *lru, struct bpf_lru_list *l) 241 { 242 if (bpf_lru_list_inactive_low(l)) 243 __bpf_lru_list_rotate_active(lru, l); 244 245 __bpf_lru_list_rotate_inactive(lru, l); 246 } 247 248 /* Calls __bpf_lru_list_shrink_inactive() to shrink some 249 * ref-bit-cleared nodes and move them to the designated 250 * free list. 251 * 252 * If it cannot get a free node after calling 253 * __bpf_lru_list_shrink_inactive(). It will just remove 254 * one node from either inactive or active list without 255 * honoring the ref-bit. It prefers inactive list to active 256 * list in this situation. 257 */ 258 static unsigned int __bpf_lru_list_shrink(struct bpf_lru *lru, 259 struct bpf_lru_list *l, 260 unsigned int tgt_nshrink, 261 struct list_head *free_list, 262 enum bpf_lru_list_type tgt_free_type) 263 264 { 265 struct bpf_lru_node *node, *tmp_node; 266 struct list_head *force_shrink_list; 267 unsigned int nshrinked; 268 269 nshrinked = __bpf_lru_list_shrink_inactive(lru, l, tgt_nshrink, 270 free_list, tgt_free_type); 271 if (nshrinked) 272 return nshrinked; 273 274 /* Do a force shrink by ignoring the reference bit */ 275 if (!list_empty(&l->lists[BPF_LRU_LIST_T_INACTIVE])) 276 force_shrink_list = &l->lists[BPF_LRU_LIST_T_INACTIVE]; 277 else 278 force_shrink_list = &l->lists[BPF_LRU_LIST_T_ACTIVE]; 279 280 list_for_each_entry_safe_reverse(node, tmp_node, force_shrink_list, 281 list) { 282 if (lru->del_from_htab(lru->del_arg, node)) { 283 __bpf_lru_node_move_to_free(l, node, free_list, 284 tgt_free_type); 285 return 1; 286 } 287 } 288 289 return 0; 290 } 291 292 /* Flush the nodes from the local pending list to the LRU list */ 293 static void __local_list_flush(struct bpf_lru_list *l, 294 struct bpf_lru_locallist *loc_l) 295 { 296 struct bpf_lru_node *node, *tmp_node; 297 298 list_for_each_entry_safe_reverse(node, tmp_node, 299 local_pending_list(loc_l), list) { 300 if (bpf_lru_node_is_ref(node)) 301 __bpf_lru_node_move_in(l, node, BPF_LRU_LIST_T_ACTIVE); 302 else 303 __bpf_lru_node_move_in(l, node, 304 BPF_LRU_LIST_T_INACTIVE); 305 } 306 } 307 308 static void bpf_lru_list_push_free(struct bpf_lru_list *l, 309 struct bpf_lru_node *node) 310 { 311 unsigned long flags; 312 313 if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type))) 314 return; 315 316 raw_spin_lock_irqsave(&l->lock, flags); 317 __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_FREE); 318 raw_spin_unlock_irqrestore(&l->lock, flags); 319 } 320 321 static void bpf_lru_list_pop_free_to_local(struct bpf_lru *lru, 322 struct bpf_lru_locallist *loc_l) 323 { 324 struct bpf_lru_list *l = &lru->common_lru.lru_list; 325 struct bpf_lru_node *node, *tmp_node; 326 unsigned int nfree = 0; 327 328 raw_spin_lock(&l->lock); 329 330 __local_list_flush(l, loc_l); 331 332 __bpf_lru_list_rotate(lru, l); 333 334 list_for_each_entry_safe(node, tmp_node, &l->lists[BPF_LRU_LIST_T_FREE], 335 list) { 336 __bpf_lru_node_move_to_free(l, node, local_free_list(loc_l), 337 BPF_LRU_LOCAL_LIST_T_FREE); 338 if (++nfree == LOCAL_FREE_TARGET) 339 break; 340 } 341 342 if (nfree < LOCAL_FREE_TARGET) 343 __bpf_lru_list_shrink(lru, l, LOCAL_FREE_TARGET - nfree, 344 local_free_list(loc_l), 345 BPF_LRU_LOCAL_LIST_T_FREE); 346 347 raw_spin_unlock(&l->lock); 348 } 349 350 static void __local_list_add_pending(struct bpf_lru *lru, 351 struct bpf_lru_locallist *loc_l, 352 int cpu, 353 struct bpf_lru_node *node, 354 u32 hash) 355 { 356 *(u32 *)((void *)node + lru->hash_offset) = hash; 357 node->cpu = cpu; 358 node->type = BPF_LRU_LOCAL_LIST_T_PENDING; 359 node->ref = 0; 360 list_add(&node->list, local_pending_list(loc_l)); 361 } 362 363 static struct bpf_lru_node * 364 __local_list_pop_free(struct bpf_lru_locallist *loc_l) 365 { 366 struct bpf_lru_node *node; 367 368 node = list_first_entry_or_null(local_free_list(loc_l), 369 struct bpf_lru_node, 370 list); 371 if (node) 372 list_del(&node->list); 373 374 return node; 375 } 376 377 static struct bpf_lru_node * 378 __local_list_pop_pending(struct bpf_lru *lru, struct bpf_lru_locallist *loc_l) 379 { 380 struct bpf_lru_node *node; 381 bool force = false; 382 383 ignore_ref: 384 /* Get from the tail (i.e. older element) of the pending list. */ 385 list_for_each_entry_reverse(node, local_pending_list(loc_l), 386 list) { 387 if ((!bpf_lru_node_is_ref(node) || force) && 388 lru->del_from_htab(lru->del_arg, node)) { 389 list_del(&node->list); 390 return node; 391 } 392 } 393 394 if (!force) { 395 force = true; 396 goto ignore_ref; 397 } 398 399 return NULL; 400 } 401 402 static struct bpf_lru_node *bpf_percpu_lru_pop_free(struct bpf_lru *lru, 403 u32 hash) 404 { 405 struct list_head *free_list; 406 struct bpf_lru_node *node = NULL; 407 struct bpf_lru_list *l; 408 unsigned long flags; 409 int cpu = raw_smp_processor_id(); 410 411 l = per_cpu_ptr(lru->percpu_lru, cpu); 412 413 raw_spin_lock_irqsave(&l->lock, flags); 414 415 __bpf_lru_list_rotate(lru, l); 416 417 free_list = &l->lists[BPF_LRU_LIST_T_FREE]; 418 if (list_empty(free_list)) 419 __bpf_lru_list_shrink(lru, l, PERCPU_FREE_TARGET, free_list, 420 BPF_LRU_LIST_T_FREE); 421 422 if (!list_empty(free_list)) { 423 node = list_first_entry(free_list, struct bpf_lru_node, list); 424 *(u32 *)((void *)node + lru->hash_offset) = hash; 425 node->ref = 0; 426 __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_INACTIVE); 427 } 428 429 raw_spin_unlock_irqrestore(&l->lock, flags); 430 431 return node; 432 } 433 434 static struct bpf_lru_node *bpf_common_lru_pop_free(struct bpf_lru *lru, 435 u32 hash) 436 { 437 struct bpf_lru_locallist *loc_l, *steal_loc_l; 438 struct bpf_common_lru *clru = &lru->common_lru; 439 struct bpf_lru_node *node; 440 int steal, first_steal; 441 unsigned long flags; 442 int cpu = raw_smp_processor_id(); 443 444 loc_l = per_cpu_ptr(clru->local_list, cpu); 445 446 raw_spin_lock_irqsave(&loc_l->lock, flags); 447 448 node = __local_list_pop_free(loc_l); 449 if (!node) { 450 bpf_lru_list_pop_free_to_local(lru, loc_l); 451 node = __local_list_pop_free(loc_l); 452 } 453 454 if (node) 455 __local_list_add_pending(lru, loc_l, cpu, node, hash); 456 457 raw_spin_unlock_irqrestore(&loc_l->lock, flags); 458 459 if (node) 460 return node; 461 462 /* No free nodes found from the local free list and 463 * the global LRU list. 464 * 465 * Steal from the local free/pending list of the 466 * current CPU and remote CPU in RR. It starts 467 * with the loc_l->next_steal CPU. 468 */ 469 470 first_steal = loc_l->next_steal; 471 steal = first_steal; 472 do { 473 steal_loc_l = per_cpu_ptr(clru->local_list, steal); 474 475 raw_spin_lock_irqsave(&steal_loc_l->lock, flags); 476 477 node = __local_list_pop_free(steal_loc_l); 478 if (!node) 479 node = __local_list_pop_pending(lru, steal_loc_l); 480 481 raw_spin_unlock_irqrestore(&steal_loc_l->lock, flags); 482 483 steal = get_next_cpu(steal); 484 } while (!node && steal != first_steal); 485 486 loc_l->next_steal = steal; 487 488 if (node) { 489 raw_spin_lock_irqsave(&loc_l->lock, flags); 490 __local_list_add_pending(lru, loc_l, cpu, node, hash); 491 raw_spin_unlock_irqrestore(&loc_l->lock, flags); 492 } 493 494 return node; 495 } 496 497 struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash) 498 { 499 if (lru->percpu) 500 return bpf_percpu_lru_pop_free(lru, hash); 501 else 502 return bpf_common_lru_pop_free(lru, hash); 503 } 504 505 static void bpf_common_lru_push_free(struct bpf_lru *lru, 506 struct bpf_lru_node *node) 507 { 508 unsigned long flags; 509 510 if (WARN_ON_ONCE(node->type == BPF_LRU_LIST_T_FREE) || 511 WARN_ON_ONCE(node->type == BPF_LRU_LOCAL_LIST_T_FREE)) 512 return; 513 514 if (node->type == BPF_LRU_LOCAL_LIST_T_PENDING) { 515 struct bpf_lru_locallist *loc_l; 516 517 loc_l = per_cpu_ptr(lru->common_lru.local_list, node->cpu); 518 519 raw_spin_lock_irqsave(&loc_l->lock, flags); 520 521 if (unlikely(node->type != BPF_LRU_LOCAL_LIST_T_PENDING)) { 522 raw_spin_unlock_irqrestore(&loc_l->lock, flags); 523 goto check_lru_list; 524 } 525 526 node->type = BPF_LRU_LOCAL_LIST_T_FREE; 527 node->ref = 0; 528 list_move(&node->list, local_free_list(loc_l)); 529 530 raw_spin_unlock_irqrestore(&loc_l->lock, flags); 531 return; 532 } 533 534 check_lru_list: 535 bpf_lru_list_push_free(&lru->common_lru.lru_list, node); 536 } 537 538 static void bpf_percpu_lru_push_free(struct bpf_lru *lru, 539 struct bpf_lru_node *node) 540 { 541 struct bpf_lru_list *l; 542 unsigned long flags; 543 544 l = per_cpu_ptr(lru->percpu_lru, node->cpu); 545 546 raw_spin_lock_irqsave(&l->lock, flags); 547 548 __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_FREE); 549 550 raw_spin_unlock_irqrestore(&l->lock, flags); 551 } 552 553 void bpf_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node) 554 { 555 if (lru->percpu) 556 bpf_percpu_lru_push_free(lru, node); 557 else 558 bpf_common_lru_push_free(lru, node); 559 } 560 561 static void bpf_common_lru_populate(struct bpf_lru *lru, void *buf, 562 u32 node_offset, u32 elem_size, 563 u32 nr_elems) 564 { 565 struct bpf_lru_list *l = &lru->common_lru.lru_list; 566 u32 i; 567 568 for (i = 0; i < nr_elems; i++) { 569 struct bpf_lru_node *node; 570 571 node = (struct bpf_lru_node *)(buf + node_offset); 572 node->type = BPF_LRU_LIST_T_FREE; 573 node->ref = 0; 574 list_add(&node->list, &l->lists[BPF_LRU_LIST_T_FREE]); 575 buf += elem_size; 576 } 577 } 578 579 static void bpf_percpu_lru_populate(struct bpf_lru *lru, void *buf, 580 u32 node_offset, u32 elem_size, 581 u32 nr_elems) 582 { 583 u32 i, pcpu_entries; 584 int cpu; 585 struct bpf_lru_list *l; 586 587 pcpu_entries = nr_elems / num_possible_cpus(); 588 589 i = 0; 590 591 for_each_possible_cpu(cpu) { 592 struct bpf_lru_node *node; 593 594 l = per_cpu_ptr(lru->percpu_lru, cpu); 595 again: 596 node = (struct bpf_lru_node *)(buf + node_offset); 597 node->cpu = cpu; 598 node->type = BPF_LRU_LIST_T_FREE; 599 node->ref = 0; 600 list_add(&node->list, &l->lists[BPF_LRU_LIST_T_FREE]); 601 i++; 602 buf += elem_size; 603 if (i == nr_elems) 604 break; 605 if (i % pcpu_entries) 606 goto again; 607 } 608 } 609 610 void bpf_lru_populate(struct bpf_lru *lru, void *buf, u32 node_offset, 611 u32 elem_size, u32 nr_elems) 612 { 613 if (lru->percpu) 614 bpf_percpu_lru_populate(lru, buf, node_offset, elem_size, 615 nr_elems); 616 else 617 bpf_common_lru_populate(lru, buf, node_offset, elem_size, 618 nr_elems); 619 } 620 621 static void bpf_lru_locallist_init(struct bpf_lru_locallist *loc_l, int cpu) 622 { 623 int i; 624 625 for (i = 0; i < NR_BPF_LRU_LOCAL_LIST_T; i++) 626 INIT_LIST_HEAD(&loc_l->lists[i]); 627 628 loc_l->next_steal = cpu; 629 630 raw_spin_lock_init(&loc_l->lock); 631 } 632 633 static void bpf_lru_list_init(struct bpf_lru_list *l) 634 { 635 int i; 636 637 for (i = 0; i < NR_BPF_LRU_LIST_T; i++) 638 INIT_LIST_HEAD(&l->lists[i]); 639 640 for (i = 0; i < NR_BPF_LRU_LIST_COUNT; i++) 641 l->counts[i] = 0; 642 643 l->next_inactive_rotation = &l->lists[BPF_LRU_LIST_T_INACTIVE]; 644 645 raw_spin_lock_init(&l->lock); 646 } 647 648 int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset, 649 del_from_htab_func del_from_htab, void *del_arg) 650 { 651 int cpu; 652 653 if (percpu) { 654 lru->percpu_lru = alloc_percpu(struct bpf_lru_list); 655 if (!lru->percpu_lru) 656 return -ENOMEM; 657 658 for_each_possible_cpu(cpu) { 659 struct bpf_lru_list *l; 660 661 l = per_cpu_ptr(lru->percpu_lru, cpu); 662 bpf_lru_list_init(l); 663 } 664 lru->nr_scans = PERCPU_NR_SCANS; 665 } else { 666 struct bpf_common_lru *clru = &lru->common_lru; 667 668 clru->local_list = alloc_percpu(struct bpf_lru_locallist); 669 if (!clru->local_list) 670 return -ENOMEM; 671 672 for_each_possible_cpu(cpu) { 673 struct bpf_lru_locallist *loc_l; 674 675 loc_l = per_cpu_ptr(clru->local_list, cpu); 676 bpf_lru_locallist_init(loc_l, cpu); 677 } 678 679 bpf_lru_list_init(&clru->lru_list); 680 lru->nr_scans = LOCAL_NR_SCANS; 681 } 682 683 lru->percpu = percpu; 684 lru->del_from_htab = del_from_htab; 685 lru->del_arg = del_arg; 686 lru->hash_offset = hash_offset; 687 688 return 0; 689 } 690 691 void bpf_lru_destroy(struct bpf_lru *lru) 692 { 693 if (lru->percpu) 694 free_percpu(lru->percpu_lru); 695 else 696 free_percpu(lru->common_lru.local_list); 697 } 698