| /linux/mm/ |
| H A D | swap_state.c | 411 struct folio *new_folio = NULL; in __read_swap_cache_async() local 439 if (!new_folio) { in __read_swap_cache_async() 440 new_folio = folio_alloc_mpol(gfp_mask, 0, mpol, ilx, numa_node_id()); in __read_swap_cache_async() 441 if (!new_folio) in __read_swap_cache_async() 478 __folio_set_locked(new_folio); in __read_swap_cache_async() 479 __folio_set_swapbacked(new_folio); in __read_swap_cache_async() 481 if (mem_cgroup_swapin_charge_folio(new_folio, NULL, gfp_mask, entry)) in __read_swap_cache_async() 484 swap_cache_add_folio(new_folio, entry, &shadow); in __read_swap_cache_async() 488 workingset_refault(new_folio, shadow); in __read_swap_cache_async() 491 folio_add_lru(new_folio); in __read_swap_cache_async() [all …]
|
| H A D | hugetlb.c | 2726 struct folio *new_folio = NULL; in alloc_and_dissolve_hugetlb_folio() local 2763 if (!new_folio) { in alloc_and_dissolve_hugetlb_folio() 2766 new_folio = alloc_fresh_hugetlb_folio(h, gfp_mask, in alloc_and_dissolve_hugetlb_folio() 2768 if (!new_folio) in alloc_and_dissolve_hugetlb_folio() 2786 account_new_hugetlb_folio(h, new_folio); in alloc_and_dissolve_hugetlb_folio() 2787 enqueue_hugetlb_folio(h, new_folio); in alloc_and_dissolve_hugetlb_folio() 2800 if (new_folio) in alloc_and_dissolve_hugetlb_folio() 2801 update_and_free_hugetlb_folio(h, new_folio, false); in alloc_and_dissolve_hugetlb_folio() 3990 struct folio *new_folio = (struct folio *)page; in demote_free_hugetlb_folios() local 3995 new_folio->mapping = NULL; in demote_free_hugetlb_folios() [all …]
|
| H A D | huge_memory.c | 3442 static void lru_add_split_folio(struct folio *folio, struct folio *new_folio, in lru_add_split_folio() argument 3445 VM_BUG_ON_FOLIO(folio_test_lru(new_folio), folio); in lru_add_split_folio() 3454 folio_get(new_folio); in lru_add_split_folio() 3455 list_add_tail(&new_folio->lru, list); in lru_add_split_folio() 3460 new_folio->mlock_count = 0; in lru_add_split_folio() 3462 list_add_tail(&new_folio->lru, &folio->lru); in lru_add_split_folio() 3463 folio_set_lru(new_folio); in lru_add_split_folio() 3504 struct folio *new_folio = (struct folio *)new_head; in __split_folio_to_order() local 3506 VM_BUG_ON_PAGE(atomic_read(&new_folio->_mapcount) != -1, new_head); in __split_folio_to_order() 3521 new_folio->flags.f &= ~PAGE_FLAGS_CHECK_AT_PREP; in __split_folio_to_order() [all …]
|
| H A D | khugepaged.c | 1863 struct folio *folio, *tmp, *new_folio; in collapse_file() local 1873 result = alloc_charge_folio(&new_folio, mm, cc); in collapse_file() 1879 __folio_set_locked(new_folio); in collapse_file() 1881 __folio_set_swapbacked(new_folio); in collapse_file() 1882 new_folio->index = start; in collapse_file() 1883 new_folio->mapping = mapping; in collapse_file() 2112 dst = folio_page(new_folio, 0); in collapse_file() 2198 lruvec_stat_mod_folio(new_folio, NR_SHMEM_THPS, HPAGE_PMD_NR); in collapse_file() 2200 lruvec_stat_mod_folio(new_folio, NR_FILE_THPS, HPAGE_PMD_NR); in collapse_file() 2203 lruvec_stat_mod_folio(new_folio, NR_FILE_PAGES, nr_none); in collapse_file() [all …]
|
| H A D | memory.c | 1051 struct folio *new_folio; in copy_present_page() local 1054 new_folio = *prealloc; in copy_present_page() 1055 if (!new_folio) in copy_present_page() 1063 if (copy_mc_user_highpage(&new_folio->page, page, addr, src_vma)) in copy_present_page() 1067 __folio_mark_uptodate(new_folio); in copy_present_page() 1068 folio_add_new_anon_rmap(new_folio, dst_vma, addr, RMAP_EXCLUSIVE); in copy_present_page() 1069 folio_add_lru_vma(new_folio, dst_vma); in copy_present_page() 1073 pte = folio_mk_pte(new_folio, dst_vma->vm_page_prot); in copy_present_page() 1188 struct folio *new_folio; in folio_prealloc() local 1191 new_folio = vma_alloc_zeroed_movable_folio(vma, addr); in folio_prealloc() [all …]
|
| H A D | ksm.c | 3109 struct folio *new_folio; in ksm_might_need_to_copy() local 3129 new_folio = vma_alloc_folio(GFP_HIGHUSER_MOVABLE, 0, vma, addr); in ksm_might_need_to_copy() 3130 if (new_folio && in ksm_might_need_to_copy() 3131 mem_cgroup_charge(new_folio, vma->vm_mm, GFP_KERNEL)) { in ksm_might_need_to_copy() 3132 folio_put(new_folio); in ksm_might_need_to_copy() 3133 new_folio = NULL; in ksm_might_need_to_copy() 3135 if (new_folio) { in ksm_might_need_to_copy() 3136 if (copy_mc_user_highpage(folio_page(new_folio, 0), page, in ksm_might_need_to_copy() 3138 folio_put(new_folio); in ksm_might_need_to_copy() 3141 folio_set_dirty(new_folio); in ksm_might_need_to_copy() [all …]
|
| /linux/include/linux/ |
| H A D | memremap.h | 245 struct folio *new_folio) in zone_device_private_split_cb() argument 249 if (new_folio) { in zone_device_private_split_cb() 250 new_folio->pgmap = original_folio->pgmap; in zone_device_private_split_cb() 251 new_folio->page.mapping = in zone_device_private_split_cb() 256 new_folio); in zone_device_private_split_cb() 296 struct folio *new_folio) in zone_device_private_split_cb() argument
|
| H A D | hugetlb_cgroup.h | 157 struct folio *new_folio); 266 struct folio *new_folio) in hugetlb_cgroup_migrate() argument
|
| H A D | hugetlb.h | 161 void move_hugetlb_state(struct folio *old_folio, struct folio *new_folio, int reason); 431 struct folio *new_folio, int reason) in move_hugetlb_state() argument
|
| /linux/include/trace/events/ |
| H A D | huge_memory.h | 202 TP_PROTO(struct mm_struct *mm, struct folio *new_folio, pgoff_t index, 205 TP_ARGS(mm, new_folio, index, addr, is_shmem, file, nr, result), 219 __entry->hpfn = new_folio ? folio_pfn(new_folio) : -1;
|
| /linux/fs/f2fs/ |
| H A D | namei.c | 919 struct folio *old_folio, *new_folio = NULL; in f2fs_rename() local 1002 &new_folio); in f2fs_rename() 1004 if (IS_ERR(new_folio)) in f2fs_rename() 1005 err = PTR_ERR(new_folio); in f2fs_rename() 1017 f2fs_set_link(new_dir, new_entry, new_folio, old_inode); in f2fs_rename() 1018 new_folio = NULL; in f2fs_rename() 1097 f2fs_folio_put(new_folio, false); in f2fs_rename() 1115 struct folio *old_folio, *new_folio; in f2fs_cross_rename() local 1150 new_entry = f2fs_find_entry(new_dir, &new_dentry->d_name, &new_folio); in f2fs_cross_rename() 1152 if (IS_ERR(new_folio)) in f2fs_cross_rename() [all …]
|
| /linux/fs/minix/ |
| H A D | namei.c | 218 struct folio *new_folio; in minix_rename() local 237 new_de = minix_find_entry(new_dentry, &new_folio); in minix_rename() 240 err = minix_set_link(new_de, new_folio, old_inode); in minix_rename() 241 folio_release_kmap(new_folio, new_de); in minix_rename()
|
| /linux/fs/ |
| H A D | dax.c | 403 struct folio *new_folio = (struct folio *)page; in dax_folio_put() local 408 new_folio->mapping = NULL; in dax_folio_put() 413 new_folio->pgmap = pgmap; in dax_folio_put() 414 new_folio->share = 0; in dax_folio_put() 415 WARN_ON_ONCE(folio_ref_count(new_folio)); in dax_folio_put()
|