Home
last modified time | relevance | path

Searched refs:lru (Results 1 – 25 of 142) sorted by relevance

123456

/linux/mm/
H A Dlist_lru.c22 static inline bool list_lru_memcg_aware(struct list_lru *lru) in list_lru_memcg_aware() argument
24 return lru->memcg_aware; in list_lru_memcg_aware()
27 static void list_lru_register(struct list_lru *lru) in list_lru_register() argument
29 if (!list_lru_memcg_aware(lru)) in list_lru_register()
33 list_add(&lru->list, &memcg_list_lrus); in list_lru_register()
37 static void list_lru_unregister(struct list_lru *lru) in list_lru_unregister() argument
39 if (!list_lru_memcg_aware(lru)) in list_lru_unregister()
43 list_del(&lru->list); in list_lru_unregister()
47 static int lru_shrinker_id(struct list_lru *lru) in lru_shrinker_id() argument
49 return lru->shrinker_id; in lru_shrinker_id()
[all …]
H A Dballoon_compaction.c48 list_for_each_entry_safe(page, tmp, pages, lru) { in balloon_page_list_enqueue()
49 list_del(&page->lru); in balloon_page_list_enqueue()
84 list_for_each_entry_safe(page, tmp, &b_dev_info->pages, lru) { in balloon_page_list_dequeue()
104 list_add(&page->lru, pages); in balloon_page_list_dequeue()
200 return list_first_entry(&pages, struct page, lru); in balloon_page_dequeue()
213 list_del(&page->lru); in balloon_page_isolate()
226 list_add(&page->lru, &b_dev_info->pages); in balloon_page_putback()
H A Dhugetlb_vmemmap.c198 list_for_each_entry_safe(page, next, list, lru) in free_vmemmap_page_list()
216 list_del(&walk->reuse_page->lru); in vmemmap_remap_pte()
227 list_add(&page->lru, walk->vmemmap_pages); in vmemmap_remap_pte()
259 page = list_first_entry(walk->vmemmap_pages, struct page, lru); in vmemmap_restore_pte()
260 list_del(&page->lru); in vmemmap_restore_pte()
340 list_add(&walk.reuse_page->lru, vmemmap_pages); in vmemmap_remap_free()
392 list_add(&page->lru, list); in alloc_vmemmap_page_list()
397 list_for_each_entry_safe(page, next, list, lru) in alloc_vmemmap_page_list()
508 list_for_each_entry_safe(folio, t_folio, folio_list, lru) { in hugetlb_vmemmap_restore_folios()
518 list_move(&folio->lru, non_hvo_folio in hugetlb_vmemmap_restore_folios()
[all...]
H A Dpage_reporting.c183 list_for_each_entry_safe(page, next, list, lru) { in page_reporting_cycle()
218 if (!list_is_first(&page->lru, list)) in page_reporting_cycle()
219 list_rotate_to_front(&page->lru, list); in page_reporting_cycle()
243 next = list_first_entry(list, struct page, lru); in page_reporting_cycle()
251 if (!list_entry_is_head(next, list, lru) && !list_is_first(&next->lru, list)) in page_reporting_cycle()
252 list_rotate_to_front(&next->lru, list); in page_reporting_cycle()
H A Dvmscan.c180 if ((_folio)->lru.prev != _base) { \
183 prev = lru_to_folio(&(_folio->lru)); \
362 * @lruvec: lru vector
363 * @lru: lru to use
366 static unsigned long lruvec_lru_size(struct lruvec *lruvec, enum lru_list lru, in lruvec_lru_size() argument
379 size += mem_cgroup_get_zone_lru_size(lruvec, lru, zid); in lruvec_lru_size()
381 size += zone_page_state(zone, NR_ZONE_LRU_BASE + lru); in lruvec_lru_size()
896 * They could be mistakenly treated as file lru. So further anon in folio_check_dirty_writeback()
1040 list_del(&folio->lru); in shrink_folio_list()
1559 update_lru_sizes(struct lruvec * lruvec,enum lru_list lru,unsigned long * nr_zone_taken) update_lru_sizes() argument
1615 isolate_lru_folios(unsigned long nr_to_scan,struct lruvec * lruvec,struct list_head * dst,unsigned long * nr_scanned,struct scan_control * sc,enum lru_list lru) isolate_lru_folios() argument
1882 shrink_inactive_list(unsigned long nr_to_scan,struct lruvec * lruvec,struct scan_control * sc,enum lru_list lru) shrink_inactive_list() argument
2001 shrink_active_list(unsigned long nr_to_scan,struct lruvec * lruvec,struct scan_control * sc,enum lru_list lru) shrink_active_list() argument
2152 shrink_list(enum lru_list lru,unsigned long nr_to_scan,struct lruvec * lruvec,struct scan_control * sc) shrink_list() argument
2345 enum lru_list lru; get_scan_count() local
3178 enum lru_list lru = type * LRU_INACTIVE_FILE; reset_batch_size() local
3805 enum lru_list lru = type * LRU_INACTIVE_FILE; inc_max_seq() local
4957 enum lru_list lru; state_is_valid() local
4977 enum lru_list lru; fill_evictable() local
5646 enum lru_list lru; shrink_lruvec() local
[all...]
H A Dmigrate.c144 * built from lru, balloon, hugetlbfs page. See isolate_migratepages_range()
152 list_for_each_entry_safe(folio, folio2, l, lru) { in putback_movable_pages()
157 list_del(&folio->lru); in putback_movable_pages()
159 * We isolated non-lru movable folio so here we can use in putback_movable_pages()
986 * In case of non-lru page, it could be released after in move_to_new_folio()
1076 list_move_tail(&src->lru, ret); in migrate_folio_undo_src()
1128 list_del(&src->lru); in migrate_folio_unmap()
1284 prev = dst->lru.prev; in migrate_folio_move()
1285 list_del(&dst->lru); in migrate_folio_move()
1324 list_del(&src->lru); in migrate_folio_move()
[all...]
H A Dpgtable-generic.c172 INIT_LIST_HEAD(&pgtable->lru); in pgtable_trans_huge_deposit()
174 list_add(&pgtable->lru, &pmd_huge_pte(mm, pmdp)->lru); in pgtable_trans_huge_deposit()
189 pmd_huge_pte(mm, pmdp) = list_first_entry_or_null(&pgtable->lru, in pgtable_trans_huge_withdraw()
190 struct page, lru); in pgtable_trans_huge_withdraw()
192 list_del(&pgtable->lru); in pgtable_trans_huge_withdraw()
/linux/include/linux/
H A Dlist_lru.h47 struct list_lru_one lru; member
61 void list_lru_destroy(struct list_lru *lru);
62 int __list_lru_init(struct list_lru *lru, bool memcg_aware,
65 #define list_lru_init(lru) \ argument
66 __list_lru_init((lru), false, NULL, NULL)
67 #define list_lru_init_memcg(lru, shrinker) \ argument
68 __list_lru_init((lru), true, NULL, shrinker)
70 int memcg_list_lru_alloc(struct mem_cgroup *memcg, struct list_lru *lru,
92 bool list_lru_add(struct list_lru *lru, struct list_head *item, int nid,
106 bool list_lru_add_obj(struct list_lru *lru, struct list_head *item);
[all …]
H A Dmm_inline.h39 enum lru_list lru, enum zone_type zid, in __update_lru_size() argument
47 __mod_lruvec_state(lruvec, NR_LRU_BASE + lru, nr_pages); in __update_lru_size()
49 NR_ZONE_LRU_BASE + lru, nr_pages); in __update_lru_size()
53 enum lru_list lru, enum zone_type zid, in update_lru_size() argument
56 __update_lru_size(lruvec, lru, zid, nr_pages); in update_lru_size()
58 mem_cgroup_update_lru_size(lruvec, lru, zid, nr_pages); in update_lru_size()
89 enum lru_list lru; in folio_lru_list() local
96 lru = folio_is_file_lru(folio) ? LRU_INACTIVE_FILE : LRU_INACTIVE_ANON; in folio_lru_list()
98 lru += LRU_ACTIVE; in folio_lru_list()
100 return lru; in folio_lru_list()
[all …]
H A Dballoon_compaction.h97 list_add(&page->lru, &balloon->pages); in balloon_page_insert()
118 list_del(&page->lru); in balloon_page_delete()
141 list_add(&page->lru, &balloon->pages); in balloon_page_insert()
147 list_del(&page->lru); in balloon_page_delete()
166 list_add(&page->lru, pages); in balloon_page_push()
178 struct page *page = list_first_entry_or_null(pages, struct page, lru); in balloon_page_pop()
183 list_del(&page->lru); in balloon_page_pop()
/linux/kernel/bpf/
H A Dbpf_lru_list.c144 static void __bpf_lru_list_rotate_active(struct bpf_lru *lru, in __bpf_lru_list_rotate_active() argument
158 if (++i == lru->nr_scans || node == first_node) in __bpf_lru_list_rotate_active()
171 static void __bpf_lru_list_rotate_inactive(struct bpf_lru *lru, in __bpf_lru_list_rotate_inactive() argument
187 while (i < lru->nr_scans) { in __bpf_lru_list_rotate_inactive()
211 __bpf_lru_list_shrink_inactive(struct bpf_lru *lru, in __bpf_lru_list_shrink_inactive() argument
225 } else if (lru->del_from_htab(lru->del_arg, node)) { in __bpf_lru_list_shrink_inactive()
232 if (++i == lru->nr_scans) in __bpf_lru_list_shrink_inactive()
242 static void __bpf_lru_list_rotate(struct bpf_lru *lru, struct bpf_lru_list *l) in __bpf_lru_list_rotate() argument
245 __bpf_lru_list_rotate_active(lru, l); in __bpf_lru_list_rotate()
247 __bpf_lru_list_rotate_inactive(lru, l); in __bpf_lru_list_rotate()
[all …]
H A Dbpf_lru_list.h71 int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset,
73 void bpf_lru_populate(struct bpf_lru *lru, void *buf, u32 node_offset,
75 void bpf_lru_destroy(struct bpf_lru *lru);
76 struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash);
77 void bpf_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node);
/linux/samples/bpf/
H A Dtest_lru_dist.c99 static void pfect_lru_init(struct pfect_lru *lru, unsigned int lru_size, in pfect_lru_init() argument
102 lru->map_fd = bpf_map_create(BPF_MAP_TYPE_HASH, NULL, in pfect_lru_init()
106 assert(lru->map_fd != -1); in pfect_lru_init()
108 lru->free_nodes = malloc(lru_size * sizeof(struct pfect_lru_node)); in pfect_lru_init()
109 assert(lru->free_nodes); in pfect_lru_init()
111 INIT_LIST_HEAD(&lru->list); in pfect_lru_init()
112 lru->cur_size = 0; in pfect_lru_init()
113 lru->lru_size = lru_size; in pfect_lru_init()
114 lru->nr_unique = lru->nr_misses = lru->total = 0; in pfect_lru_init()
117 static void pfect_lru_destroy(struct pfect_lru *lru) in pfect_lru_destroy() argument
[all …]
/linux/net/ceph/
H A Dpagelist.c32 struct page *page = list_entry(pl->head.prev, struct page, lru); in ceph_pagelist_unmap_tail()
45 lru); in ceph_pagelist_release()
46 list_del(&page->lru); in ceph_pagelist_release()
61 page = list_first_entry(&pl->free_list, struct page, lru); in ceph_pagelist_addpage()
62 list_del(&page->lru); in ceph_pagelist_addpage()
69 list_add_tail(&page->lru, &pl->head); in ceph_pagelist_addpage()
113 list_add_tail(&page->lru, &pl->free_list); in ceph_pagelist_reserve()
125 struct page, lru); in ceph_pagelist_free_reserve()
126 list_del(&page->lru); in ceph_pagelist_free_reserve()
159 page = list_entry(pl->head.prev, struct page, lru); in ceph_pagelist_truncate()
[all …]
/linux/fs/nfs/
H A Dnfs42xattr.c71 struct list_head lru; member
82 struct list_head lru; member
130 struct list_lru *lru; in nfs4_xattr_entry_lru_add() local
132 lru = (entry->flags & NFS4_XATTR_ENTRY_EXTVAL) ? in nfs4_xattr_entry_lru_add()
135 return list_lru_add_obj(lru, &entry->lru); in nfs4_xattr_entry_lru_add()
141 struct list_lru *lru; in nfs4_xattr_entry_lru_del() local
143 lru = (entry->flags & NFS4_XATTR_ENTRY_EXTVAL) ? in nfs4_xattr_entry_lru_del()
146 return list_lru_del_obj(lru, &entry->lru); in nfs4_xattr_entry_lru_del()
239 INIT_LIST_HEAD(&entry->lru); in nfs4_xattr_alloc_entry()
261 if (WARN_ON(!list_empty(&entry->lru))) in nfs4_xattr_free_entry_cb()
[all …]
/linux/drivers/gpu/drm/
H A Ddrm_gem.c1347 drm_gem_lru_init(struct drm_gem_lru *lru, struct mutex *lock) in drm_gem_lru_init() argument
1349 lru->lock = lock; in drm_gem_lru_init()
1350 lru->count = 0; in drm_gem_lru_init()
1351 INIT_LIST_HEAD(&lru->list); in drm_gem_lru_init()
1358 obj->lru->count -= obj->size >> PAGE_SHIFT; in drm_gem_lru_remove_locked()
1359 WARN_ON(obj->lru->count < 0); in drm_gem_lru_remove_locked()
1361 obj->lru = NULL; in drm_gem_lru_remove_locked()
1374 struct drm_gem_lru *lru = obj->lru; in drm_gem_lru_remove() local
1376 if (!lru) in drm_gem_lru_remove()
1379 mutex_lock(lru->lock); in drm_gem_lru_remove()
[all …]
/linux/drivers/gpu/drm/msm/
H A Dmsm_gem_shrinker.c38 unsigned count = priv->lru.dontneed.count; in msm_gem_shrinker_count()
41 count += priv->lru.willneed.count; in msm_gem_shrinker_count()
104 struct drm_gem_lru *lru; in msm_gem_shrinker_scan() member
111 { &priv->lru.dontneed, purge, true }, in msm_gem_shrinker_scan()
112 { &priv->lru.willneed, evict, can_swap() }, in msm_gem_shrinker_scan()
113 { &priv->lru.dontneed, active_purge, can_block(sc) }, in msm_gem_shrinker_scan()
114 { &priv->lru.willneed, active_evict, can_swap() && can_block(sc) }, in msm_gem_shrinker_scan()
124 drm_gem_lru_scan(stages[i].lru, nr, in msm_gem_shrinker_scan()
183 &priv->lru.dontneed, in msm_gem_shrinker_vmap()
184 &priv->lru.willneed, in msm_gem_shrinker_vmap()
[all …]
/linux/drivers/gpu/drm/ttm/
H A Dttm_resource.c73 list_bulk_move_tail(&man->lru[j], &pos->first->lru, in ttm_lru_bulk_move_tail()
74 &pos->last->lru); in ttm_lru_bulk_move_tail()
93 pos->first = list_next_entry(res, lru); in ttm_lru_bulk_move_pos_tail()
94 list_move(&res->lru, &pos->last->lru); in ttm_lru_bulk_move_pos_tail()
125 pos->first = list_next_entry(res, lru); in ttm_lru_bulk_move_del()
127 pos->last = list_prev_entry(res, lru); in ttm_lru_bulk_move_del()
129 list_move(&res->lru, &pos->last->lru); in ttm_lru_bulk_move_del()
158 list_move_tail(&res->lru, &bdev->pinned); in ttm_resource_move_to_lru_tail()
169 list_move_tail(&res->lru, &man->lru[bo->priority]); in ttm_resource_move_to_lru_tail()
200 list_add_tail(&res->lru, &bo->bdev->pinned); in ttm_resource_init()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/
H A Dnv50.c38 struct list_head lru; member
53 struct list_head lru; member
143 eobj = list_first_entry_or_null(&imem->lru, typeof(*eobj), lru); in nv50_instobj_kmap()
149 list_del_init(&eobj->lru); in nv50_instobj_kmap()
205 if (likely(iobj->lru.next) && iobj->map) { in nv50_instobj_release()
206 BUG_ON(!list_empty(&iobj->lru)); in nv50_instobj_release()
207 list_add_tail(&iobj->lru, &imem->lru); in nv50_instobj_release()
249 if (likely(iobj->lru.next)) in nv50_instobj_acquire()
250 list_del_init(&iobj->lru); in nv50_instobj_acquire()
275 if (likely(iobj->lru.next)) { in nv50_instobj_boot()
[all …]
/linux/drivers/md/
H A Ddm-bufio.c86 struct lru *lru; member
92 struct lru { struct
101 static void lru_init(struct lru *lru) in lru_init() argument
103 lru->cursor = NULL; in lru_init()
104 lru->count = 0; in lru_init()
105 INIT_LIST_HEAD(&lru->iterators); in lru_init()
108 static void lru_destroy(struct lru *lru) in lru_destroy() argument
110 WARN_ON_ONCE(lru->cursor); in lru_destroy()
111 WARN_ON_ONCE(!list_empty(&lru->iterators)); in lru_destroy()
117 static void lru_insert(struct lru *lru, struct lru_entry *le) in lru_insert() argument
[all …]
/linux/arch/mips/mm/
H A Dcerr-sb1.c326 uint8_t lru; in extract_ic() local
348 lru = (taghi >> 14) & 0xff; in extract_ic()
352 (lru & 0x3), in extract_ic()
353 ((lru >> 2) & 0x3), in extract_ic()
354 ((lru >> 4) & 0x3), in extract_ic()
355 ((lru >> 6) & 0x3)); in extract_ic()
481 uint8_t ecc, lru; in extract_dc() local
503 lru = (taghi >> 14) & 0xff; in extract_dc()
507 (lru & 0x3), in extract_dc()
508 ((lru >> 2) & 0x3), in extract_dc()
[all …]
/linux/fs/bcachefs/
H A Dlru.c9 #include "lru.h"
21 "lru entry at time=0"); in bch2_lru_invalid()
29 const struct bch_lru *lru = bkey_s_c_to_lru(k).v; in bch2_lru_to_text() local
31 prt_printf(out, "idx %llu", le64_to_cpu(lru->idx)); in bch2_lru_to_text()
34 void bch2_lru_pos_to_text(struct printbuf *out, struct bpos lru) in bch2_lru_pos_to_text() argument
37 lru_pos_id(lru), in bch2_lru_pos_to_text()
38 lru_pos_time(lru), in bch2_lru_pos_to_text()
39 u64_to_bucket(lru.offset).inode, in bch2_lru_pos_to_text()
40 u64_to_bucket(lru.offset).offset); in bch2_lru_pos_to_text()
99 "lru ke in bch2_check_lru_key()
[all...]
/linux/fs/proc/
H A Dmeminfo.c42 int lru; in meminfo_proc_show() local
53 for (lru = LRU_BASE; lru < NR_LRU_LISTS; lru++) in meminfo_proc_show()
54 pages[lru] = global_node_page_state(NR_LRU_BASE + lru); in meminfo_proc_show()
/linux/drivers/xen/
H A Dballoon.c153 list_add_tail(&page->lru, &ballooned_pages); in balloon_append()
156 list_add(&page->lru, &ballooned_pages); in balloon_append()
170 page = list_entry(ballooned_pages.next, struct page, lru); in balloon_retrieve()
173 list_del(&page->lru); in balloon_retrieve()
186 struct list_head *next = page->lru.next; in balloon_next_page()
189 return list_entry(next, struct page, lru); in balloon_next_page()
394 page = list_first_entry_or_null(&ballooned_pages, struct page, lru); in increase_reservation()
444 list_add(&page->lru, &pages); in decrease_reservation()
461 list_for_each_entry_safe(page, tmp, &pages, lru) { in decrease_reservation()
466 list_del(&page->lru); in decrease_reservation()
/linux/kernel/power/
H A Dwakelock.c32 struct list_head lru; member
92 list_add(&wl->lru, &wakelocks_lru_list); in wakelocks_lru_add()
97 list_move(&wl->lru, &wakelocks_lru_list); in wakelocks_lru_most_recent()
108 list_for_each_entry_safe_reverse(wl, aux, &wakelocks_lru_list, lru) { in __wakelocks_gc()
123 list_del(&wl->lru); in __wakelocks_gc()

123456