| /linux/mm/damon/ |
| H A D | ops-common.c | 330 static unsigned int damon_migrate_folio_list(struct list_head *folio_list, in damon_migrate_folio_list() argument 339 while (!list_empty(folio_list)) { in damon_migrate_folio_list() 344 folio = lru_to_folio(folio_list); in damon_migrate_folio_list() 367 list_splice_init(&migrate_folios, folio_list); in damon_migrate_folio_list() 371 list_splice(&ret_folios, folio_list); in damon_migrate_folio_list() 373 while (!list_empty(folio_list)) { in damon_migrate_folio_list() 374 folio = lru_to_folio(folio_list); in damon_migrate_folio_list() 382 unsigned long damon_migrate_pages(struct list_head *folio_list, int target_nid) in damon_migrate_pages() argument 389 if (list_empty(folio_list)) in damon_migrate_pages() 398 nid = folio_nid(lru_to_folio(folio_list)); in damon_migrate_pages() [all …]
|
| H A D | paddr.c | 156 LIST_HEAD(folio_list); in damon_pa_pageout() 196 list_add(&folio->lru, &folio_list); in damon_pa_pageout() 203 applied = reclaim_pages(&folio_list); in damon_pa_pageout() 262 LIST_HEAD(folio_list); in damon_pa_migrate() 280 list_add(&folio->lru, &folio_list); in damon_pa_migrate() 285 applied = damon_migrate_pages(&folio_list, s->target_nid); in damon_pa_migrate()
|
| H A D | ops-common.h | 23 unsigned long damon_migrate_pages(struct list_head *folio_list, int target_nid);
|
| H A D | vaddr.c | 807 LIST_HEAD(folio_list); in damos_va_migrate()
|
| /linux/mm/ |
| H A D | hugetlb_vmemmap.c | 508 * @folio_list: list of folios. 516 * non-processed folios will remain on folio_list. 519 struct list_head *folio_list, in hugetlb_vmemmap_restore_folios() argument 527 list_for_each_entry_safe(folio, t_folio, folio_list, lru) { in hugetlb_vmemmap_restore_folios() 654 struct list_head *folio_list, in __hugetlb_vmemmap_optimize_folios() argument 663 list_for_each_entry(folio, folio_list, lru) { in __hugetlb_vmemmap_optimize_folios() 709 list_for_each_entry(folio, folio_list, lru) { in __hugetlb_vmemmap_optimize_folios() 737 void hugetlb_vmemmap_optimize_folios(struct hstate *h, struct list_head *folio_list) in hugetlb_vmemmap_optimize_folios() argument 739 __hugetlb_vmemmap_optimize_folios(h, folio_list, false); in hugetlb_vmemmap_optimize_folios() 742 void hugetlb_vmemmap_optimize_bootmem_folios(struct hstate *h, struct list_head *folio_list) in hugetlb_vmemmap_optimize_bootmem_folios() argument [all...] |
| H A D | hugetlb.c | 1635 struct list_head *folio_list, in bulk_vmemmap_restore_error() argument 1667 list_for_each_entry_safe(folio, t_folio, folio_list, lru) in bulk_vmemmap_restore_error() 1686 struct list_head *folio_list) in update_and_free_pages_bulk() argument 1698 ret = hugetlb_vmemmap_restore_folios(h, folio_list, &non_hvo_folios); in update_and_free_pages_bulk() 1700 bulk_vmemmap_restore_error(h, folio_list, &non_hvo_folios); in update_and_free_pages_bulk() 1711 VM_WARN_ON(!list_empty(folio_list)); in update_and_free_pages_bulk() 1928 struct list_head *folio_list) in prep_and_add_allocated_folios() argument 1934 hugetlb_vmemmap_optimize_folios(h, folio_list); in prep_and_add_allocated_folios() 1938 list_for_each_entry_safe(folio, tmp_f, folio_list, lru) { in prep_and_add_allocated_folios() 3247 struct list_head *folio_list) in prep_and_add_bootmem_folios() argument [all...] |
| H A D | vmscan.c | 638 struct swap_iocb **plug, struct list_head *folio_list) in writeout() argument 650 res = shmem_writeout(folio, plug, folio_list); in writeout() 674 struct swap_iocb **plug, struct list_head *folio_list) in pageout() argument 699 return writeout(folio, mapping, plug, folio_list); in pageout() 1083 static unsigned int shrink_folio_list(struct list_head *folio_list, in shrink_folio_list() argument 1102 while (!list_empty(folio_list)) { in shrink_folio_list() 1111 folio = lru_to_folio(folio_list); in shrink_folio_list() 1251 list_add_tail(&folio->lru, folio_list); in shrink_folio_list() 1303 split_folio_to_list(folio, folio_list)) in shrink_folio_list() 1312 if (split_folio_to_list(folio, folio_list)) in shrink_folio_list() [all …]
|
| H A D | hugetlb_internal.h | 101 struct list_head *folio_list);
|
| H A D | madvise.c | 364 LIST_HEAD(folio_list); in madvise_cold_or_pageout_pte_range() 435 list_add(&folio->lru, &folio_list); in madvise_cold_or_pageout_pte_range() 442 reclaim_pages(&folio_list); in madvise_cold_or_pageout_pte_range() 550 list_add(&folio->lru, &folio_list); in madvise_cold_or_pageout_pte_range() 561 reclaim_pages(&folio_list); in madvise_cold_or_pageout_pte_range()
|
| H A D | internal.h | 1343 unsigned long reclaim_pages(struct list_head *folio_list); 1345 struct list_head *folio_list);
|
| H A D | shmem.c | 1591 struct list_head *folio_list) in shmem_writeout() argument 1630 if (split_folio_to_list(folio, folio_list)) in shmem_writeout()
|
| /linux/include/linux/ |
| H A D | shmem_fs.h | 126 struct list_head *folio_list);
|