| /linux/mm/ |
| H A D | list_lru.c | 22 static inline bool list_lru_memcg_aware(struct list_lru *lru) in list_lru_memcg_aware() argument 24 return lru->memcg_aware; in list_lru_memcg_aware() 27 static void list_lru_register(struct list_lru *lru) in list_lru_register() argument 29 if (!list_lru_memcg_aware(lru)) in list_lru_register() 33 list_add(&lru->list, &memcg_list_lrus); in list_lru_register() 37 static void list_lru_unregister(struct list_lru *lru) in list_lru_unregister() argument 39 if (!list_lru_memcg_aware(lru)) in list_lru_unregister() 43 list_del(&lru->list); in list_lru_unregister() 47 static int lru_shrinker_id(struct list_lru *lru) in lru_shrinker_id() argument 49 return lru in lru_shrinker_id() 53 list_lru_from_memcg_idx(struct list_lru * lru,int nid,int idx) list_lru_from_memcg_idx() argument 80 lock_list_lru_of_memcg(struct list_lru * lru,int nid,struct mem_cgroup * memcg,bool irq,bool skip_empty) lock_list_lru_of_memcg() argument 113 list_lru_register(struct list_lru * lru) list_lru_register() argument 117 list_lru_unregister(struct list_lru * lru) list_lru_unregister() argument 121 lru_shrinker_id(struct list_lru * lru) lru_shrinker_id() argument 126 list_lru_memcg_aware(struct list_lru * lru) list_lru_memcg_aware() argument 132 list_lru_from_memcg_idx(struct list_lru * lru,int nid,int idx) list_lru_from_memcg_idx() argument 138 lock_list_lru_of_memcg(struct list_lru * lru,int nid,struct mem_cgroup * memcg,bool irq,bool skip_empty) lock_list_lru_of_memcg() argument 161 list_lru_add(struct list_lru * lru,struct list_head * item,int nid,struct mem_cgroup * memcg) list_lru_add() argument 183 list_lru_add_obj(struct list_lru * lru,struct list_head * item) list_lru_add_obj() argument 201 list_lru_del(struct list_lru * lru,struct list_head * item,int nid,struct mem_cgroup * memcg) list_lru_del() argument 220 list_lru_del_obj(struct list_lru * lru,struct list_head * item) list_lru_del_obj() argument 252 list_lru_count_one(struct list_lru * lru,int nid,struct mem_cgroup * memcg) list_lru_count_one() argument 270 list_lru_count_node(struct list_lru * lru,int nid) list_lru_count_node() argument 280 __list_lru_walk_one(struct list_lru * lru,int nid,struct mem_cgroup * memcg,list_lru_walk_cb isolate,void * cb_arg,unsigned long * nr_to_walk,bool irq_off) __list_lru_walk_one() argument 337 list_lru_walk_one(struct list_lru * lru,int nid,struct mem_cgroup * memcg,list_lru_walk_cb isolate,void * cb_arg,unsigned long * nr_to_walk) list_lru_walk_one() argument 347 list_lru_walk_one_irq(struct list_lru * lru,int nid,struct mem_cgroup * memcg,list_lru_walk_cb isolate,void * cb_arg,unsigned long * nr_to_walk) list_lru_walk_one_irq() argument 355 list_lru_walk_node(struct list_lru * lru,int nid,list_lru_walk_cb isolate,void * cb_arg,unsigned long * nr_to_walk) list_lru_walk_node() argument 393 init_one_lru(struct list_lru * lru,struct list_lru_one * l) init_one_lru() argument 405 memcg_init_list_lru_one(struct list_lru * lru,gfp_t gfp) memcg_init_list_lru_one() argument 420 memcg_init_list_lru(struct list_lru * lru,bool memcg_aware) memcg_init_list_lru() argument 427 memcg_destroy_list_lru(struct list_lru * lru) memcg_destroy_list_lru() argument 443 memcg_reparent_list_lru_one(struct list_lru * lru,int nid,struct list_lru_one * src,struct mem_cgroup * dst_memcg) memcg_reparent_list_lru_one() argument 469 struct list_lru *lru; memcg_reparent_list_lrus() local 506 memcg_list_lru_allocated(struct mem_cgroup * memcg,struct list_lru * lru) memcg_list_lru_allocated() argument 513 memcg_list_lru_alloc(struct mem_cgroup * memcg,struct list_lru * lru,gfp_t gfp) memcg_list_lru_alloc() argument 565 memcg_init_list_lru(struct list_lru * lru,bool memcg_aware) memcg_init_list_lru() argument 569 memcg_destroy_list_lru(struct list_lru * lru) memcg_destroy_list_lru() argument 574 __list_lru_init(struct list_lru * lru,bool memcg_aware,struct shrinker * shrinker) __list_lru_init() argument 602 list_lru_destroy(struct list_lru * lru) list_lru_destroy() argument [all...] |
| H A D | balloon_compaction.c | |
| H A D | page_reporting.c | 183 list_for_each_entry_safe(page, next, list, lru) { in page_reporting_cycle() 218 if (!list_is_first(&page->lru, list)) in page_reporting_cycle() 219 list_rotate_to_front(&page->lru, list); in page_reporting_cycle() 243 next = list_first_entry(list, struct page, lru); in page_reporting_cycle() 251 if (!list_entry_is_head(next, list, lru) && !list_is_first(&next->lru, list)) in page_reporting_cycle() 252 list_rotate_to_front(&next->lru, list); in page_reporting_cycle()
|
| H A D | hugetlb_vmemmap.c | 198 list_for_each_entry_safe(page, next, list, lru) 210 list_del(&walk->vmemmap_head->lru); in vmemmap_remap_pte() 228 list_add(&page->lru, walk->vmemmap_pages); in vmemmap_remap_pte() 238 page = list_first_entry(walk->vmemmap_pages, struct page, lru); 239 list_del(&page->lru); 349 list_add(&page->lru, list); in vmemmap_remap_free() 355 list_for_each_entry_safe(page, next, list, lru) in vmemmap_remap_free() 462 list_for_each_entry_safe(folio, t_folio, folio_list, lru) { in __hugetlb_vmemmap_restore_folio() 471 list_move(&folio->lru, non_hvo_folios); in __hugetlb_vmemmap_restore_folio() 562 list_add(&vmemmap_head->lru, vmemmap_page in vmemmap_should_optimize_folio() [all...] |
| H A D | migrate.c | 254 * built from lru, balloon, hugetlbfs page. See isolate_migratepages_range() 262 list_for_each_entry_safe(folio, folio2, l, lru) { in putback_movable_pages() 267 list_del(&folio->lru); in putback_movable_pages() 294 list_add(&folio->lru, list); in isolate_folio_to_list() 1176 list_move_tail(&src->lru, ret); 1368 prev = dst->lru.prev; in migrate_folio_move() 1369 list_del(&dst->lru); in migrate_folio_move() 1427 list_del(&src->lru); in migrate_folio_move() 1441 list_add(&dst->lru, prev); 1567 list_move_tail(&src->lru, re in try_split_folio() [all...] |
| H A D | vmscan.c | 187 if ((_folio)->lru.prev != _base) { \ 190 prev = lru_to_folio(&(_folio->lru)); \ 389 * @lruvec: lru vector 390 * @lru: lru to use 393 unsigned long lruvec_lru_size(struct lruvec *lruvec, enum lru_list lru, int zone_idx) in zone_reclaimable_pages() 401 size += mem_cgroup_get_zone_lru_size(lruvec, lru, zid); in zone_reclaimable_pages() 403 size += zone_page_state(zone, NR_ZONE_LRU_BASE + lru); in zone_reclaimable_pages() 943 * They could be mistakenly treated as file lru. So further anon in folio_check_references() 1087 list_del(&folio->lru); in shrink_folio_list() 412 lruvec_lru_size(struct lruvec * lruvec,enum lru_list lru,int zone_idx) lruvec_lru_size() argument 1676 update_lru_sizes(struct lruvec * lruvec,enum lru_list lru,unsigned long * nr_zone_taken) update_lru_sizes() argument 1713 isolate_lru_folios(unsigned long nr_to_scan,struct lruvec * lruvec,struct list_head * dst,unsigned long * nr_scanned,struct scan_control * sc,enum lru_list lru) isolate_lru_folios() argument 1979 shrink_inactive_list(unsigned long nr_to_scan,struct lruvec * lruvec,struct scan_control * sc,enum lru_list lru) shrink_inactive_list() argument 2101 shrink_active_list(unsigned long nr_to_scan,struct lruvec * lruvec,struct scan_control * sc,enum lru_list lru) shrink_active_list() argument 2249 shrink_list(enum lru_list lru,unsigned long nr_to_scan,struct lruvec * lruvec,struct scan_control * sc) shrink_list() argument 2536 enum lru_list lru; get_scan_count() local 3309 enum lru_list lru = type * LRU_INACTIVE_FILE; reset_batch_size() local 3997 enum lru_list lru = type * LRU_INACTIVE_FILE; inc_max_seq() local 5089 enum lru_list lru; state_is_valid() local 5109 enum lru_list lru; fill_evictable() local 5777 enum lru_list lru; shrink_lruvec() local [all...] |
| H A D | pgtable-generic.c | 175 INIT_LIST_HEAD(&pgtable->lru); in pgtable_trans_huge_deposit() 177 list_add(&pgtable->lru, &pmd_huge_pte(mm, pmdp)->lru); in pgtable_trans_huge_deposit() 192 pmd_huge_pte(mm, pmdp) = list_first_entry_or_null(&pgtable->lru, in pgtable_trans_huge_withdraw() 193 struct page, lru); in pgtable_trans_huge_withdraw() 195 list_del(&pgtable->lru); in pgtable_trans_huge_withdraw()
|
| H A D | mmzone.c | 77 enum lru_list lru; in lruvec_init() local 83 for_each_lru(lru) in lruvec_init() 84 INIT_LIST_HEAD(&lruvec->lists[lru]); in lruvec_init()
|
| H A D | zpdesc.h | 34 struct list_head lru; member 55 ZPDESC_MATCH(lru, lru);
|
| /linux/include/linux/ |
| H A D | mm_inline.h | 34 enum lru_list lru, enum zone_type zid, in page_is_file_lru() 42 mod_lruvec_state(lruvec, NR_LRU_BASE + lru, nr_pages); in __update_lru_size() 44 NR_ZONE_LRU_BASE + lru, nr_pages); in __update_lru_size() 48 enum lru_list lru, enum zone_type zid, in __update_lru_size() 51 __update_lru_size(lruvec, lru, zid, nr_pages); 53 mem_cgroup_update_lru_size(lruvec, lru, zid, nr_pages); in update_lru_size() argument 58 * __folio_clear_lru_flags - Clear page lru flags before releasing a page. in update_lru_size() 59 * @folio: The folio that was on lru and now has a zero reference. in update_lru_size() 84 enum lru_list lru; 91 lru in folio_lru_list() 39 __update_lru_size(struct lruvec * lruvec,enum lru_list lru,enum zone_type zid,long nr_pages) __update_lru_size() argument 89 enum lru_list lru; folio_lru_list() local 180 enum lru_list lru = type * LRU_INACTIVE_FILE; lru_gen_update_size() local 343 enum lru_list lru = folio_lru_list(folio); lruvec_add_folio() local 357 enum lru_list lru = folio_lru_list(folio); lruvec_add_folio_tail() local 371 enum lru_list lru = folio_lru_list(folio); lruvec_del_folio() local [all...] |
| H A D | mmzone.h | 244 NR_ISOLATED_ANON, /* Temporary isolated pages from anon lru */ 245 NR_ISOLATED_FILE, /* Temporary isolated pages from file lru */ 403 #define for_each_lru(lru) for (lru = 0; lru < NR_LRU_LISTS; lru++) 405 #define for_each_evictable_lru(lru) for (lru = 0; lru <= LRU_ACTIVE_FILE; lru 333 for_each_lru(lru) global() argument 335 for_each_evictable_lru(lru) global() argument 337 is_file_lru(enum lru_list lru) is_file_lru() argument 342 is_active_lru(enum lru_list lru) is_active_lru() argument [all...] |
| /linux/samples/bpf/ |
| H A D | test_lru_dist.c | 99 static void pfect_lru_init(struct pfect_lru *lru, unsigned int lru_size, in pfect_lru_init() argument 102 lru->map_fd = bpf_map_create(BPF_MAP_TYPE_HASH, NULL, in pfect_lru_init() 106 assert(lru->map_fd != -1); in pfect_lru_init() 108 lru->free_nodes = malloc(lru_size * sizeof(struct pfect_lru_node)); in pfect_lru_init() 109 assert(lru->free_nodes); in pfect_lru_init() 111 INIT_LIST_HEAD(&lru->list); in pfect_lru_init() 112 lru->cur_size = 0; in pfect_lru_init() 113 lru->lru_size = lru_size; in pfect_lru_init() 114 lru->nr_unique = lru->nr_misses = lru->total = 0; in pfect_lru_init() 117 static void pfect_lru_destroy(struct pfect_lru *lru) in pfect_lru_destroy() argument [all …]
|
| /linux/net/ceph/ |
| H A D | pagelist.c | 32 struct page *page = list_entry(pl->head.prev, struct page, lru); in ceph_pagelist_unmap_tail() 45 lru); in ceph_pagelist_release() 46 list_del(&page->lru); in ceph_pagelist_release() 61 page = list_first_entry(&pl->free_list, struct page, lru); in ceph_pagelist_addpage() 62 list_del(&page->lru); in ceph_pagelist_addpage() 69 list_add_tail(&page->lru, &pl->head); in ceph_pagelist_addpage() 113 list_add_tail(&page->lru, &pl->free_list); in ceph_pagelist_reserve() 125 struct page, lru); in ceph_pagelist_free_reserve() 126 list_del(&page->lru); in ceph_pagelist_free_reserve()
|
| /linux/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
| H A D | nv50.c | 38 struct list_head lru; member 53 struct list_head lru; member 143 eobj = list_first_entry_or_null(&imem->lru, typeof(*eobj), lru); in nv50_instobj_kmap() 149 list_del_init(&eobj->lru); in nv50_instobj_kmap() 205 if (likely(iobj->lru.next) && iobj->map) { in nv50_instobj_release() 206 BUG_ON(!list_empty(&iobj->lru)); in nv50_instobj_release() 207 list_add_tail(&iobj->lru, &imem->lru); in nv50_instobj_release() 249 if (likely(iobj->lru.next)) in nv50_instobj_acquire() 250 list_del_init(&iobj->lru); in nv50_instobj_acquire() 275 if (likely(iobj->lru.next)) { in nv50_instobj_boot() [all …]
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_gem.c | 630 * Move folios to appropriate lru and release the folios, decrementing the 1543 * @lru: The LRU to initialize in drm_gem_lru_remove_locked() 1547 drm_gem_lru_init(struct drm_gem_lru *lru, struct mutex *lock) 1549 lru->lock = lock; 1550 lru->count = 0; 1551 INIT_LIST_HEAD(&lru->list); 1558 obj->lru->count -= obj->size >> PAGE_SHIFT; in drm_gem_lru_remove() 1559 WARN_ON(obj->lru->count < 0); in drm_gem_lru_remove() 1561 obj->lru = NULL; in drm_gem_lru_remove() 1574 struct drm_gem_lru *lru 1529 drm_gem_lru_init(struct drm_gem_lru * lru,struct mutex * lock) drm_gem_lru_init() argument 1556 struct drm_gem_lru *lru = obj->lru; drm_gem_lru_remove() local 1576 drm_gem_lru_move_tail_locked(struct drm_gem_lru * lru,struct drm_gem_object * obj) drm_gem_lru_move_tail_locked() argument 1600 drm_gem_lru_move_tail(struct drm_gem_lru * lru,struct drm_gem_object * obj) drm_gem_lru_move_tail() argument 1625 drm_gem_lru_scan(struct drm_gem_lru * lru,unsigned int nr_to_scan,unsigned long * remaining,bool (* shrink)(struct drm_gem_object * obj,struct ww_acquire_ctx * ticket),struct ww_acquire_ctx * ticket) drm_gem_lru_scan() argument [all...] |
| /linux/arch/mips/mm/ |
| H A D | cerr-sb1.c | 326 uint8_t lru; in extract_ic() local 348 lru = (taghi >> 14) & 0xff; in extract_ic() 352 (lru & 0x3), in extract_ic() 353 ((lru >> 2) & 0x3), in extract_ic() 354 ((lru >> 4) & 0x3), in extract_ic() 355 ((lru >> 6) & 0x3)); in extract_ic() 481 uint8_t ecc, lru; in extract_dc() local 503 lru = (taghi >> 14) & 0xff; in extract_dc() 507 (lru & 0x3), in extract_dc() 508 ((lru >> 2) & 0x3), in extract_dc() [all …]
|
| /linux/kernel/bpf/ |
| H A D | bpf_lru_list.h | 72 int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset, 74 void bpf_lru_populate(struct bpf_lru *lru, void *buf, u32 node_offset, 76 void bpf_lru_destroy(struct bpf_lru *lru); 77 struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash); 78 void bpf_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node);
|
| /linux/drivers/android/binder/ |
| H A D | page_range.rs | 12 // 3. lru spinlock 193 /// * Free. The page is None. The `lru` element is not queued. 194 /// * Available. The page is Some. The `lru` element is queued to the shrinker's lru. 195 /// * Used. The page is Some. The `lru` element is not queued. 200 lru: bindings::list_head, field 246 /// Add this page to the lru list, if not already in the list. 253 let lru_ptr = unsafe { &raw mut (*me).lru }; in list_lru_add() 254 // SAFETY: The lru pointer is valid, and we're not using it with any other lru lis in list_lru_add() 320 let lru = &raw mut (*info).lru; register_with_vma() localVariable [all...] |
| /linux/drivers/xen/ |
| H A D | balloon.c | 154 list_add_tail(&page->lru, &ballooned_pages); in balloon_append() 157 list_add(&page->lru, &ballooned_pages); in balloon_append() 173 page = list_entry(ballooned_pages.next, struct page, lru); in balloon_retrieve() 176 list_del(&page->lru); in balloon_retrieve() 191 struct list_head *next = page->lru.next; in balloon_next_page() 194 return list_entry(next, struct page, lru); in balloon_next_page() 399 page = list_first_entry_or_null(&ballooned_pages, struct page, lru); in increase_reservation() 453 list_add(&page->lru, &pages); in decrease_reservation() 470 list_for_each_entry_safe(page, tmp, &pages, lru) { in decrease_reservation() 475 list_del(&page->lru); in decrease_reservation()
|
| /linux/drivers/md/ |
| H A D | dm-writecache.c | 83 struct list_head lru; member 110 struct list_head lru; member 668 list_add(&ins->lru, &wc->lru); in writecache_insert_entry() 674 list_del(&e->lru); in writecache_unlink() 695 list_add_tail(&e->lru, &wc->freelist); in writecache_add_to_freelist() 736 e = container_of(wc->freelist.next, struct wc_entry, lru); in writecache_pop_from_freelist() 739 list_del(&e->lru); in writecache_pop_from_freelist() 775 wc->lru.next = LIST_POISON1; in writecache_poison_lists() 776 wc->lru in writecache_poison_lists() [all...] |
| /linux/kernel/power/ |
| H A D | wakelock.c | 32 struct list_head lru; member 95 list_add(&wl->lru, &wakelocks_lru_list); in wakelocks_lru_add() 100 list_move(&wl->lru, &wakelocks_lru_list); in wakelocks_lru_most_recent() 111 list_for_each_entry_safe_reverse(wl, aux, &wakelocks_lru_list, lru) { in __wakelocks_gc() 126 list_del(&wl->lru); in __wakelocks_gc()
|
| /linux/include/trace/events/ |
| H A D | pagemap.h | 37 __field(enum lru_list, lru ) 44 __entry->lru = folio_lru_list(folio); 52 __entry->lru,
|
| /linux/drivers/gpu/drm/msm/ |
| H A D | msm_gem.c | 152 drm_gem_lru_move_tail_locked(&priv->lru.pinned, obj); in update_lru_active() 154 drm_gem_lru_move_tail_locked(&priv->lru.willneed, obj); in update_lru_active() 158 drm_gem_lru_move_tail_locked(&priv->lru.dontneed, obj); in update_lru_active() 172 drm_gem_lru_move_tail_locked(&priv->lru.unbacked, obj); in update_lru_locked() 182 mutex_lock(&priv->lru.lock); in update_lru() 184 mutex_unlock(&priv->lru.lock); in update_lru() 281 * Update the pin count of the object, call under lru.lock 290 drm_gem_lru_move_tail_locked(&priv->lru.pinned, obj); in msm_gem_pin_obj_locked() 297 mutex_lock(&priv->lru.lock); in pin_obj_locked() 299 mutex_unlock(&priv->lru in pin_obj_locked() [all...] |
| /linux/drivers/android/ |
| H A D | binder_alloc.h | 67 struct list_head lru; member 78 return &mdata->lru; in page_to_lru() 125 struct list_lru_one *lru,
|
| /linux/drivers/misc/ |
| H A D | vmw_balloon.c | 670 struct page, lru); in vmballoon_alloc_page_list() 671 list_del(&page->lru); in vmballoon_alloc_page_list() 685 list_add(&page->lru, &ctl->pages); in vmballoon_alloc_page_list() 863 list_for_each_entry(page, &ctl->pages, lru) in vmballoon_lock() 895 list_move(&page->lru, &ctl->refused_pages); in vmballoon_lock() 920 list_for_each_entry_safe(page, tmp, page_list, lru) { in vmballoon_release_page_list() 921 list_del(&page->lru); in vmballoon_release_page_list() 1006 list_for_each_entry(page, pages, lru) { in vmballoon_enqueue_page_list() 1050 list_for_each_entry_safe(page, tmp, &b->huge_pages, lru) { in vmballoon_dequeue_page_list() 1053 list_move(&page->lru, pages); in vmballoon_dequeue_page_list() [all …]
|