Home
last modified time | relevance | path

Searched refs:lru (Results 1 – 25 of 114) sorted by relevance

12345

/linux/mm/
H A Dlist_lru.c22 static inline bool list_lru_memcg_aware(struct list_lru *lru) in list_lru_memcg_aware() argument
24 return lru->memcg_aware; in list_lru_memcg_aware()
27 static void list_lru_register(struct list_lru *lru) in list_lru_register() argument
29 if (!list_lru_memcg_aware(lru)) in list_lru_register()
33 list_add(&lru->list, &memcg_list_lrus); in list_lru_register()
37 static void list_lru_unregister(struct list_lru *lru) in list_lru_unregister() argument
39 if (!list_lru_memcg_aware(lru)) in list_lru_unregister()
43 list_del(&lru->list); in list_lru_unregister()
47 static int lru_shrinker_id(struct list_lru *lru) in lru_shrinker_id() argument
49 return lru->shrinker_id; in lru_shrinker_id()
[all …]
H A Dballoon_compaction.c
H A Dpage_reporting.c183 list_for_each_entry_safe(page, next, list, lru) { in page_reporting_cycle()
218 if (!list_is_first(&page->lru, list)) in page_reporting_cycle()
219 list_rotate_to_front(&page->lru, list); in page_reporting_cycle()
243 next = list_first_entry(list, struct page, lru); in page_reporting_cycle()
251 if (!list_entry_is_head(next, list, lru) && !list_is_first(&next->lru, list)) in page_reporting_cycle()
252 list_rotate_to_front(&next->lru, list); in page_reporting_cycle()
H A Dhugetlb_vmemmap.c204 list_for_each_entry_safe(page, next, list, lru) in free_vmemmap_page_list()
222 list_del(&walk->reuse_page->lru); in vmemmap_remap_pte()
233 list_add(&page->lru, walk->vmemmap_pages); in vmemmap_remap_pte()
265 page = list_first_entry(walk->vmemmap_pages, struct page, lru); in vmemmap_restore_pte()
266 list_del(&page->lru); in vmemmap_restore_pte()
346 list_add(&walk.reuse_page->lru, vmemmap_pages); in vmemmap_remap_free()
400 list_add(&page->lru, list); in alloc_vmemmap_page_list()
406 list_for_each_entry_safe(page, next, list, lru) in alloc_vmemmap_page_list()
527 list_for_each_entry_safe(folio, t_folio, folio_list, lru) { in hugetlb_vmemmap_restore_folios()
539 list_move(&folio->lru, non_hvo_folio in hugetlb_vmemmap_restore_folios()
[all...]
H A Dmigrate.c262 list_for_each_entry_safe(folio, folio2, l, lru) { in putback_movable_pages()
267 list_del(&folio->lru); in putback_movable_pages()
294 list_add(&folio->lru, list); in isolate_folio_to_list()
1174 list_move_tail(&src->lru, ret); in migrate_folio_undo_src()
1364 prev = dst->lru.prev; in migrate_folio_move()
1365 list_del(&dst->lru); in migrate_folio_move()
1408 list_del(&src->lru); in migrate_folio_move()
1422 list_add(&dst->lru, prev); in migrate_folio_move()
1548 list_move_tail(&src->lru, ret); in unmap_and_move_huge_page()
1577 list_move_tail(&folio->lru, split_folios); in try_split_folio()
[all …]
H A Dvmscan.c187 if ((_folio)->lru.prev != _base) { \
190 prev = lru_to_folio(&(_folio->lru)); \
412 static unsigned long lruvec_lru_size(struct lruvec *lruvec, enum lru_list lru, in lruvec_lru_size() argument
421 size += mem_cgroup_get_zone_lru_size(lruvec, lru, zid); in lruvec_lru_size()
423 size += zone_page_state(zone, NR_ZONE_LRU_BASE + lru); in lruvec_lru_size()
1112 list_del(&folio->lru); in shrink_folio_list()
1251 list_add_tail(&folio->lru, folio_list); in shrink_folio_list()
1276 list_add(&folio->lru, &demote_folios); in shrink_folio_list()
1569 list_add(&folio->lru, &ret_folios); in shrink_folio_list()
1633 list_for_each_entry_safe(folio, next, folio_list, lru) { in reclaim_clean_pages_from_list()
[all …]
H A Dmmzone.c77 enum lru_list lru; in lruvec_init() local
83 for_each_lru(lru) in lruvec_init()
84 INIT_LIST_HEAD(&lruvec->lists[lru]); in lruvec_init()
H A Dpgtable-generic.c173 INIT_LIST_HEAD(&pgtable->lru); in pgtable_trans_huge_deposit()
175 list_add(&pgtable->lru, &pmd_huge_pte(mm, pmdp)->lru); in pgtable_trans_huge_deposit()
190 pmd_huge_pte(mm, pmdp) = list_first_entry_or_null(&pgtable->lru, in pgtable_trans_huge_withdraw()
191 struct page, lru); in pgtable_trans_huge_withdraw()
193 list_del(&pgtable->lru); in pgtable_trans_huge_withdraw()
H A Dzpdesc.h34 struct list_head lru; member
55 ZPDESC_MATCH(lru, lru);
/linux/include/linux/
H A Dmm_inline.h39 enum lru_list lru, enum zone_type zid, in __update_lru_size() argument
47 mod_lruvec_state(lruvec, NR_LRU_BASE + lru, nr_pages); in __update_lru_size()
49 NR_ZONE_LRU_BASE + lru, nr_pages); in __update_lru_size()
53 enum lru_list lru, enum zone_type zid, in update_lru_size() argument
56 __update_lru_size(lruvec, lru, zid, nr_pages); in update_lru_size()
58 mem_cgroup_update_lru_size(lruvec, lru, zid, nr_pages); in update_lru_size()
89 enum lru_list lru; in folio_lru_list() local
96 lru = folio_is_file_lru(folio) ? LRU_INACTIVE_FILE : LRU_INACTIVE_ANON; in folio_lru_list()
98 lru += LRU_ACTIVE; in folio_lru_list()
100 return lru; in folio_lru_list()
[all …]
H A Dmmzone.h333 #define for_each_lru(lru) for (lru = 0; lru < NR_LRU_LISTS; lru++) argument
335 #define for_each_evictable_lru(lru) for (lru = 0; lru <= LRU_ACTIVE_FILE; lru++) argument
337 static inline bool is_file_lru(enum lru_list lru) in is_file_lru() argument
339 return (lru == LRU_INACTIVE_FILE || lru == LRU_ACTIVE_FILE); in is_file_lru()
342 static inline bool is_active_lru(enum lru_list lru) in is_active_lru() argument
344 return (lru == LRU_ACTIVE_ANON || lru == LRU_ACTIVE_FILE); in is_active_lru()
/linux/samples/bpf/
H A Dtest_lru_dist.c99 static void pfect_lru_init(struct pfect_lru *lru, unsigned int lru_size, in pfect_lru_init() argument
102 lru->map_fd = bpf_map_create(BPF_MAP_TYPE_HASH, NULL, in pfect_lru_init()
106 assert(lru->map_fd != -1); in pfect_lru_init()
108 lru->free_nodes = malloc(lru_size * sizeof(struct pfect_lru_node)); in pfect_lru_init()
109 assert(lru->free_nodes); in pfect_lru_init()
111 INIT_LIST_HEAD(&lru->list); in pfect_lru_init()
112 lru->cur_size = 0; in pfect_lru_init()
113 lru->lru_size = lru_size; in pfect_lru_init()
114 lru->nr_unique = lru->nr_misses = lru->total = 0; in pfect_lru_init()
117 static void pfect_lru_destroy(struct pfect_lru *lru) in pfect_lru_destroy() argument
[all …]
/linux/net/ceph/
H A Dpagelist.c32 struct page *page = list_entry(pl->head.prev, struct page, lru); in ceph_pagelist_unmap_tail()
45 lru); in ceph_pagelist_release()
46 list_del(&page->lru); in ceph_pagelist_release()
61 page = list_first_entry(&pl->free_list, struct page, lru); in ceph_pagelist_addpage()
62 list_del(&page->lru); in ceph_pagelist_addpage()
69 list_add_tail(&page->lru, &pl->head); in ceph_pagelist_addpage()
113 list_add_tail(&page->lru, &pl->free_list); in ceph_pagelist_reserve()
125 struct page, lru); in ceph_pagelist_free_reserve()
126 list_del(&page->lru); in ceph_pagelist_free_reserve()
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/
H A Dnv50.c38 struct list_head lru; member
53 struct list_head lru; member
143 eobj = list_first_entry_or_null(&imem->lru, typeof(*eobj), lru); in nv50_instobj_kmap()
149 list_del_init(&eobj->lru); in nv50_instobj_kmap()
205 if (likely(iobj->lru.next) && iobj->map) { in nv50_instobj_release()
206 BUG_ON(!list_empty(&iobj->lru)); in nv50_instobj_release()
207 list_add_tail(&iobj->lru, &imem->lru); in nv50_instobj_release()
249 if (likely(iobj->lru.next)) in nv50_instobj_acquire()
250 list_del_init(&iobj->lru); in nv50_instobj_acquire()
275 if (likely(iobj->lru.next)) { in nv50_instobj_boot()
[all …]
/linux/drivers/gpu/drm/
H A Ddrm_gem.c1529 drm_gem_lru_init(struct drm_gem_lru *lru, struct mutex *lock) in drm_gem_lru_init() argument
1531 lru->lock = lock; in drm_gem_lru_init()
1532 lru->count = 0; in drm_gem_lru_init()
1533 INIT_LIST_HEAD(&lru->list); in drm_gem_lru_init()
1540 obj->lru->count -= obj->size >> PAGE_SHIFT; in drm_gem_lru_remove_locked()
1541 WARN_ON(obj->lru->count < 0); in drm_gem_lru_remove_locked()
1543 obj->lru = NULL; in drm_gem_lru_remove_locked()
1556 struct drm_gem_lru *lru = obj->lru; in drm_gem_lru_remove() local
1558 if (!lru) in drm_gem_lru_remove()
1561 mutex_lock(lru->lock); in drm_gem_lru_remove()
[all …]
/linux/arch/mips/mm/
H A Dcerr-sb1.c326 uint8_t lru; in extract_ic() local
348 lru = (taghi >> 14) & 0xff; in extract_ic()
352 (lru & 0x3), in extract_ic()
353 ((lru >> 2) & 0x3), in extract_ic()
354 ((lru >> 4) & 0x3), in extract_ic()
355 ((lru >> 6) & 0x3)); in extract_ic()
481 uint8_t ecc, lru; in extract_dc() local
503 lru = (taghi >> 14) & 0xff; in extract_dc()
507 (lru & 0x3), in extract_dc()
508 ((lru >> 2) & 0x3), in extract_dc()
[all …]
/linux/kernel/bpf/
H A Dbpf_lru_list.h72 int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset,
74 void bpf_lru_populate(struct bpf_lru *lru, void *buf, u32 node_offset,
76 void bpf_lru_destroy(struct bpf_lru *lru);
77 struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash);
78 void bpf_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node);
/linux/drivers/android/binder/
H A Dpage_range.rs174 lru: bindings::list_head, field
227 let lru_ptr = unsafe { &raw mut (*me).lru }; in list_lru_add()
239 let lru_ptr = unsafe { &raw mut (*me).lru }; in list_lru_del()
294 let lru = &raw mut (*info).lru; in register_with_vma() localVariable
295 (&raw mut (*lru).next).write(lru); in register_with_vma()
296 (&raw mut (*lru).prev).write(lru); in register_with_vma()
660 lru: *mut bindings::list_lru_one, in rust_shrink_free_page()
699 unsafe { bindings::list_lru_isolate(lru, item) }; in rust_shrink_free_page()
718 unsafe { bindings::spin_unlock(&raw mut (*lru).lock) }; in rust_shrink_free_page()
/linux/drivers/xen/
H A Dballoon.c154 list_add_tail(&page->lru, &ballooned_pages); in balloon_append()
157 list_add(&page->lru, &ballooned_pages); in balloon_append()
173 page = list_entry(ballooned_pages.next, struct page, lru); in balloon_retrieve()
176 list_del(&page->lru); in balloon_retrieve()
191 struct list_head *next = page->lru.next; in balloon_next_page()
194 return list_entry(next, struct page, lru); in balloon_next_page()
399 page = list_first_entry_or_null(&ballooned_pages, struct page, lru); in increase_reservation()
453 list_add(&page->lru, &pages); in decrease_reservation()
470 list_for_each_entry_safe(page, tmp, &pages, lru) { in decrease_reservation()
475 list_del(&page->lru); in decrease_reservation()
/linux/drivers/md/
H A Ddm-writecache.c83 struct list_head lru; member
110 struct list_head lru; member
668 list_add(&ins->lru, &wc->lru); in writecache_insert_entry()
674 list_del(&e->lru); in writecache_unlink()
695 list_add_tail(&e->lru, &wc->freelist); in writecache_add_to_freelist()
736 e = container_of(wc->freelist.next, struct wc_entry, lru); in writecache_pop_from_freelist()
739 list_del(&e->lru); in writecache_pop_from_freelist()
775 wc->lru.next = LIST_POISON1; in writecache_poison_lists()
776 wc->lru.prev = LIST_POISON2; in writecache_poison_lists()
801 if (list_empty(&wc->lru)) in writecache_flush()
[all …]
/linux/kernel/power/
H A Dwakelock.c32 struct list_head lru; member
95 list_add(&wl->lru, &wakelocks_lru_list); in wakelocks_lru_add()
100 list_move(&wl->lru, &wakelocks_lru_list); in wakelocks_lru_most_recent()
111 list_for_each_entry_safe_reverse(wl, aux, &wakelocks_lru_list, lru) { in __wakelocks_gc()
126 list_del(&wl->lru); in __wakelocks_gc()
/linux/include/trace/events/
H A Dpagemap.h37 __field(enum lru_list, lru )
44 __entry->lru = folio_lru_list(folio);
52 __entry->lru,
/linux/drivers/gpu/drm/msm/
H A Dmsm_gem.c152 drm_gem_lru_move_tail_locked(&priv->lru.pinned, obj); in update_lru_active()
154 drm_gem_lru_move_tail_locked(&priv->lru.willneed, obj); in update_lru_active()
158 drm_gem_lru_move_tail_locked(&priv->lru.dontneed, obj); in update_lru_active()
172 drm_gem_lru_move_tail_locked(&priv->lru.unbacked, obj); in update_lru_locked()
182 mutex_lock(&priv->lru.lock); in update_lru()
184 mutex_unlock(&priv->lru.lock); in update_lru()
290 drm_gem_lru_move_tail_locked(&priv->lru.pinned, obj); in msm_gem_pin_obj_locked()
297 mutex_lock(&priv->lru.lock); in pin_obj_locked()
299 mutex_unlock(&priv->lru.lock); in pin_obj_locked()
495 mutex_lock(&priv->lru.lock); in msm_gem_unpin_locked()
[all …]
/linux/drivers/android/
H A Dbinder_alloc.h67 struct list_head lru; member
78 return &mdata->lru; in page_to_lru()
125 struct list_lru_one *lru,
/linux/drivers/misc/
H A Dvmw_balloon.c670 struct page, lru); in vmballoon_alloc_page_list()
671 list_del(&page->lru); in vmballoon_alloc_page_list()
685 list_add(&page->lru, &ctl->pages); in vmballoon_alloc_page_list()
863 list_for_each_entry(page, &ctl->pages, lru) in vmballoon_lock()
895 list_move(&page->lru, &ctl->refused_pages); in vmballoon_lock()
920 list_for_each_entry_safe(page, tmp, page_list, lru) { in vmballoon_release_page_list()
921 list_del(&page->lru); in vmballoon_release_page_list()
1006 list_for_each_entry(page, pages, lru) { in vmballoon_enqueue_page_list()
1050 list_for_each_entry_safe(page, tmp, &b->huge_pages, lru) { in vmballoon_dequeue_page_list()
1053 list_move(&page->lru, pages); in vmballoon_dequeue_page_list()
[all …]

12345