/linux/mm/ |
H A D | hugetlb.c | 3013 struct folio *new_folio = NULL; in alloc_and_dissolve_hugetlb_folio() 3020 * Freed from under us. Drop new_folio too. in alloc_and_dissolve_hugetlb_folio() 3045 if (!new_folio) { in alloc_and_dissolve_hugetlb_folio() 3047 new_folio = alloc_buddy_hugetlb_folio(h, gfp_mask, nid, in alloc_and_dissolve_hugetlb_folio() 3049 if (!new_folio) in alloc_and_dissolve_hugetlb_folio() 3051 __prep_new_hugetlb_folio(h, new_folio); in alloc_and_dissolve_hugetlb_folio() 3059 * and enqueue_hugetlb_folio() for new_folio. The counters will in alloc_and_dissolve_hugetlb_folio() 3065 * Ref count on new_folio is already zero as it was dropped in alloc_and_dissolve_hugetlb_folio() 3069 enqueue_hugetlb_folio(h, new_folio); in alloc_and_dissolve_hugetlb_folio() 3082 if (new_folio) { in alloc_and_dissolve_hugetlb_folio() 3018 struct folio *new_folio = NULL; alloc_and_dissolve_hugetlb_folio() local 5357 hugetlb_install_folio(struct vm_area_struct * vma,pte_t * ptep,unsigned long addr,struct folio * new_folio,pte_t old,unsigned long sz) hugetlb_install_folio() argument 5488 struct folio *new_folio; copy_hugetlb_page_range() local 5935 struct folio *new_folio; hugetlb_wp() local 6211 bool new_folio, new_pagecache_folio = false; hugetlb_no_page() local 7649 move_hugetlb_state(struct folio * old_folio,struct folio * new_folio,int reason) move_hugetlb_state() argument [all...] |
H A D | memory.c | 907 struct folio *new_folio; in copy_present_page() 910 new_folio = *prealloc; in copy_present_page() 911 if (!new_folio) in copy_present_page() 919 copy_user_highpage(&new_folio->page, page, addr, src_vma); in copy_present_page() 920 __folio_mark_uptodate(new_folio); in copy_present_page() 921 folio_add_new_anon_rmap(new_folio, dst_vma, addr); in copy_present_page() 922 folio_add_lru_vma(new_folio, dst_vma); in copy_present_page() 926 pte = mk_pte(&new_folio->page, dst_vma->vm_page_prot); in copy_present_page() 1045 struct folio *new_folio; in folio_prealloc() 1048 new_folio in folio_prealloc() 905 struct folio *new_folio; copy_present_page() local 1043 struct folio *new_folio; folio_prealloc() local 3247 struct folio *new_folio = NULL; wp_page_copy() local [all...] |
H A D | ksm.c | 3064 struct folio *new_folio; in ksm_might_need_to_copy() 3084 new_folio = vma_alloc_folio(GFP_HIGHUSER_MOVABLE, 0, vma, addr, false); in ksm_might_need_to_copy() 3085 if (new_folio && in ksm_might_need_to_copy() 3086 mem_cgroup_charge(new_folio, vma->vm_mm, GFP_KERNEL)) { in ksm_might_need_to_copy() 3087 folio_put(new_folio); in ksm_might_need_to_copy() 3088 new_folio = NULL; in ksm_might_need_to_copy() 3090 if (new_folio) { in ksm_might_need_to_copy() 3091 if (copy_mc_user_highpage(folio_page(new_folio, 0), page, in ksm_might_need_to_copy() 3093 folio_put(new_folio); in ksm_might_need_to_copy() 3097 folio_set_dirty(new_folio); in ksm_might_need_to_copy() 3057 struct folio *new_folio; ksm_might_need_to_copy() local [all...] |
H A D | hugetlb_cgroup.c | 871 void hugetlb_cgroup_migrate(struct folio *old_folio, struct folio *new_folio) in hugetlb_cgroup_migrate() argument 887 set_hugetlb_cgroup(new_folio, h_cg); in hugetlb_cgroup_migrate() 888 set_hugetlb_cgroup_rsvd(new_folio, h_cg_rsvd); in hugetlb_cgroup_migrate() 889 list_move(&new_folio->lru, &h->hugepage_activelist); in hugetlb_cgroup_migrate()
|
H A D | huge_memory.c | 2734 * Careful: new_folio is not a "real" folio before we cleared PageTail. in unmap_folio() 2737 struct folio *new_folio = (struct folio *)page_tail; in unmap_folio() 2787 new_folio->swap.val = folio->swap.val + tail; in lru_add_page_tail() 2801 folio_set_large_rmappable(new_folio); in __split_huge_page_tail() 2807 folio_nr_pages(new_folio) : 0)); in __split_huge_page_tail() 2810 folio_set_young(new_folio); in __split_huge_page_tail() 2812 folio_set_idle(new_folio); in __split_huge_page_tail() 2814 folio_xchg_last_cpupid(new_folio, folio_last_cpupid(folio)); in __split_huge_page_tail() 2876 struct folio *new_folio = (struct folio *)head; in __split_huge_page_tail() 2878 folio_set_order(new_folio, new_orde in __split_huge_page_tail() 2802 struct folio *new_folio = (struct folio *)page_tail; __split_huge_page_tail() local 2941 struct folio *new_folio = (struct folio *)head; __split_huge_page() local 2983 struct folio *new_folio = page_folio(subpage); __split_huge_page() local [all...] |
/linux/fs/ext2/ |
H A D | namei.c | 354 struct folio *new_folio; in ext2_rename() local 362 &new_folio); in ext2_rename() 367 err = ext2_set_link(new_dir, new_de, new_folio, old_inode, true); in ext2_rename() 368 folio_release_kmap(new_folio, new_de); in ext2_rename()
|
/linux/include/linux/ |
H A D | hugetlb_cgroup.h | 157 struct folio *new_folio); 266 struct folio *new_folio) in hugetlb_cgroup_migrate() argument
|
H A D | hugetlb.h | 170 void move_hugetlb_state(struct folio *old_folio, struct folio *new_folio, int reason); 447 struct folio *new_folio, int reason) in isolate_hugetlb() 465 move_hugetlb_state(struct folio * old_folio,struct folio * new_folio,int reason) move_hugetlb_state() argument
|
/linux/fs/nilfs2/ |
H A D | namei.c | 378 struct folio *new_folio; in nilfs_rename() local 386 new_de = nilfs_find_entry(new_dir, &new_dentry->d_name, &new_folio); in nilfs_rename() 389 nilfs_set_link(new_dir, new_de, new_folio, old_inode); in nilfs_rename() 390 folio_release_kmap(new_folio, new_de); in nilfs_rename()
|
/linux/fs/jbd2/ |
H A D | journal.c | 336 struct folio *new_folio; in jbd2_journal_write_metadata_buffer() local 365 new_folio = virt_to_folio(jh_in->b_frozen_data); in jbd2_journal_write_metadata_buffer() 366 new_offset = offset_in_folio(new_folio, jh_in->b_frozen_data); in jbd2_journal_write_metadata_buffer() 368 new_folio = jh2bh(jh_in)->b_folio; in jbd2_journal_write_metadata_buffer() 369 new_offset = offset_in_folio(new_folio, jh2bh(jh_in)->b_data); in jbd2_journal_write_metadata_buffer() 372 mapped_data = kmap_local_folio(new_folio, new_offset); in jbd2_journal_write_metadata_buffer() 411 memcpy_from_folio(tmp, new_folio, new_offset, bh_in->b_size); in jbd2_journal_write_metadata_buffer() 413 new_folio = virt_to_folio(tmp); in jbd2_journal_write_metadata_buffer() 414 new_offset = offset_in_folio(new_folio, tmp); in jbd2_journal_write_metadata_buffer() 430 mapped_data = kmap_local_folio(new_folio, new_offse in jbd2_journal_write_metadata_buffer() [all...] |
/linux/kernel/events/ |
H A D | uprobes.c | 157 struct folio *new_folio; in __replace_page() local 167 new_folio = page_folio(new_page); in __replace_page() 168 err = mem_cgroup_charge(new_folio, vma->vm_mm, GFP_KERNEL); in __replace_page() 183 folio_get(new_folio); in __replace_page() 184 folio_add_new_anon_rmap(new_folio, vma, addr); in __replace_page() 185 folio_add_lru_vma(new_folio, vma); in __replace_page()
|