Home
last modified time | relevance | path

Searched refs:page_folio (Results 1 – 25 of 137) sorted by relevance

123456

/linux/mm/
H A Dfolio-compat.c16 return folio_unlock(page_folio(page)); in unlock_page()
22 return folio_end_writeback(page_folio(page)); in end_page_writeback()
28 return folio_wait_writeback(page_folio(page)); in wait_on_page_writeback()
34 folio_mark_accessed(page_folio(page)); in mark_page_accessed()
40 folio_start_writeback(page_folio(page)); in set_page_writeback()
46 return folio_mark_dirty(page_folio(page)); in set_page_dirty()
52 return folio_mark_dirty_lock(page_folio(page)); in set_page_dirty_lock()
58 return folio_clear_dirty_for_io(page_folio(page)); in clear_page_dirty_for_io()
65 return folio_redirty_for_writepage(wbc, page_folio(page)); in redirty_page_for_writepage()
72 return filemap_add_folio(mapping, page_folio(pag in add_to_page_cache_lru()
[all...]
H A Dmigrate_device.c86 struct folio *fault_folio = fault_page ? page_folio(fault_page) : NULL; in migrate_vma_split_folio()
103 new_fault_folio = fault_page ? page_folio(fault_page) : NULL; in migrate_vma_split_folio()
258 page_folio(migrate->fault_page) : NULL; in migrate_vma_collect_pmd()
312 folio = page_folio(page); in migrate_vma_collect_pmd()
355 folio = page ? page_folio(page) : NULL; in migrate_vma_collect_pmd()
391 folio = page_folio(page); in migrate_vma_collect_pmd()
539 struct folio *folio = page_folio(page); in migrate_vma_check_page()
571 page_folio(fault_page) : NULL; in migrate_device_unmap()
589 folio = page_folio(page); in migrate_device_unmap()
641 folio = page_folio(page); in migrate_device_unmap()
[all …]
H A Dmemory-failure.c187 ret = dissolve_free_hugetlb_folio(page_folio(page)); in __page_handle_poison()
334 shake_folio(page_folio(page)); in shake_page()
400 tk->size_shift = folio_shift(page_folio(p)); in __add_to_kill()
984 count -= folio_nr_pages(page_folio(p)); in has_extra_refcount()
1022 struct folio *folio = page_folio(p); in me_pagecache_clean()
1080 struct folio *folio = page_folio(p); in me_pagecache_dirty()
1118 struct folio *folio = page_folio(p); in me_swapcache_dirty()
1140 struct folio *folio = page_folio(p); in me_swapcache_clean()
1162 struct folio *folio = page_folio(p); in me_huge_page()
1356 struct folio *folio = page_folio(page); in __get_hwpoison_page()
[all …]
H A Dhwpoison-inject.c22 struct folio *folio = page_folio(p); in hwpoison_filter_dev()
116 folio = page_folio(p); in hwpoison_inject()
H A Dgup.c56 folio = page_folio(page); in sanity_check_pinned_pages()
79 folio = page_folio(page); in try_get_folio()
87 * could be that between calling page_folio() and the refcount in try_get_folio()
94 if (unlikely(page_folio(page) != folio)) { in try_get_folio()
188 gup_put_folio(page_folio(page), 1, FOLL_PIN); in unpin_user_page()
236 struct folio *folio = page_folio(next); in gup_folio_range_next()
250 struct folio *folio = page_folio(list[i]); in gup_folio_next()
254 if (page_folio(list[nr]) != folio) in gup_folio_next()
541 return page_folio(page); in try_grab_folio_fast()
674 ret = try_grab_folio(page_folio(pag in follow_huge_pud()
[all...]
H A Dmmu_gather.c68 folio_remove_rmap_ptes(page_folio(page), page, nr_pages, in tlb_flush_rmap_batch()
178 VM_WARN_ON_ONCE(page_folio(page) != page_folio(page + nr_pages - 1)); in __tlb_remove_folio_pages_size()
H A Dpage_idle.c42 folio = page_folio(page); in page_idle_get_folio()
45 if (unlikely(page_folio(page) != folio || !folio_test_lru(folio))) { in page_idle_get_folio()
/linux/arch/sh/mm/
H A Dcache.c64 struct folio *folio = page_folio(page); in copy_to_user_page()
85 struct folio *folio = page_folio(page); in copy_from_user_page()
102 struct folio *src = page_folio(from); in copy_user_highpage()
150 struct folio *folio = page_folio(pfn_to_page(pfn)); in __update_cache()
160 struct folio *folio = page_folio(page); in __flush_anon_page()
247 cacheop_on_each_cpu(local_flush_icache_folio, page_folio(page), 1); in flush_icache_pages()
H A Dkmap.c30 struct folio *folio = page_folio(page); in kmap_coherent()
/linux/include/linux/
H A Dmemremap.h176 folio_is_device_private(page_folio(page)); in is_device_private_page()
201 folio_is_pci_p2pdma(page_folio(page)); in is_pci_p2pdma_page()
212 return folio_is_device_coherent(page_folio(page)); in is_device_coherent_page()
223 return folio_is_fsdax(page_folio(page)); in is_fsdax_page()
H A Dpage-flags.h306 #define page_folio(p) (_Generic((p), \ macro
729 unsigned long flags = (unsigned long)page_folio(page)->mapping; in PageAnonNotKsm()
736 return folio_test_anon(page_folio(page)); in PageAnon()
804 return folio_test_uptodate(page_folio(page)); in PageUptodate()
1079 return folio_test_hugetlb(page_folio(page)); in PAGE_TYPE_OPS()
1093 folio = page_folio(page); in is_page_hwpoison()
/linux/lib/
H A Dtest_hmm.c583 mdevice->free_folios = page_folio(page); in dmirror_allocate_chunk()
665 zone_device_folio_init(page_folio(dpage), in dmirror_devmem_alloc_page()
666 page_pgmap(folio_page(page_folio(dpage), 0)), in dmirror_devmem_alloc_page()
723 folio = page_folio(spage); in dmirror_migrate_alloc_and_copy()
763 struct folio *folio = page_folio(dpage); in dmirror_migrate_alloc_and_copy()
860 nr = folio_nr_pages(page_folio(dpage)); in dmirror_migrate_finalize_and_map()
867 VM_WARN_ON(folio_nr_pages(page_folio(rpage)) != nr); in dmirror_migrate_finalize_and_map()
870 entry = folio_page(page_folio(rpage), i); in dmirror_migrate_finalize_and_map()
969 order = folio_order(page_folio(spage)); in dmirror_devmem_fault_alloc_and_copy()
1405 order = folio_order(page_folio(spage)); in dmirror_device_evict_chunk()
[all …]
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_dmem.c205 sfolio = page_folio(vmf->page); in nouveau_dmem_migrate_to_ram()
255 dfolio = page_folio(dpage); in nouveau_dmem_migrate_to_ram()
360 drm->dmem->free_folios = page_folio(page); in nouveau_dmem_chunk_alloc()
417 folio = page_folio(page); in nouveau_dmem_page_alloc_locked()
423 folio = page_folio(page); in nouveau_dmem_page_alloc_locked()
492 struct folio *folio = page_folio( in nouveau_dmem_evict_chunk()
511 page_folio(migrate_pfn_to_page(src_pfns[i])), in nouveau_dmem_evict_chunk()
512 page_folio(dpage), in nouveau_dmem_evict_chunk()
753 if (drm->dmem->migrate.copy_func(drm, folio_nr_pages(page_folio(spage)), in nouveau_dmem_migrate_copy_one()
770 if (folio_order(page_folio(dpage))) in nouveau_dmem_migrate_copy_one()
[all …]
/linux/arch/csky/abiv1/
H A Dcacheflush.c40 flush_dcache_folio(page_folio(page)); in flush_dcache_page()
58 folio = page_folio(pfn_to_page(pfn)); in update_mmu_cache_range()
/linux/arch/arm64/mm/
H A Dcopypage.c21 struct folio *src = page_folio(from); in copy_highpage()
22 struct folio *dst = page_folio(to); in copy_highpage()
H A Dflush.c54 struct folio *folio = page_folio(pte_page(pte)); in __sync_icache_dcache()
79 flush_dcache_folio(page_folio(page)); in flush_dcache_page()
/linux/arch/xtensa/mm/
H A Dcache.c90 struct folio *folio = page_folio(page); in clear_user_highpage()
105 struct folio *folio = page_folio(dst); in copy_user_highpage()
226 folio = page_folio(pfn_to_page(pfn)); in update_mmu_cache_range()
/linux/arch/mips/mm/
H A Dinit.c85 BUG_ON(folio_test_dcache_dirty(page_folio(page))); in __kmap_pgprot()
166 struct folio *src = page_folio(from); in copy_user_highpage()
192 struct folio *folio = page_folio(page); in copy_to_user_page()
212 struct folio *folio = page_folio(page); in copy_from_user_page()
H A Dcache.c130 struct folio *folio = page_folio(page); in __flush_anon_page()
157 folio = page_folio(pfn_to_page(pfn)); in __update_cache()
/linux/drivers/gpu/drm/
H A Ddrm_pagemap.c243 folio = page_folio(page); in drm_pagemap_migrate_map_pages()
561 unsigned long nr_pages = src_page ? NR_PAGES(folio_order(page_folio(src_page))) : 1; in drm_pagemap_migrate_to_devmem()
653 unsigned long nr_pages = page ? NR_PAGES(folio_order(page_folio(page))) : 1; in drm_pagemap_migrate_to_devmem()
723 order = folio_order(page_folio(src_page)); in drm_pagemap_migrate_populate_ram_pfn()
753 WARN_ON_ONCE(!folio_trylock(page_folio(page))); in drm_pagemap_migrate_populate_ram_pfn()
755 order = folio_order(page_folio(page)); in drm_pagemap_migrate_populate_ram_pfn()
775 order = folio_order(page_folio(page)); in drm_pagemap_migrate_populate_ram_pfn()
/linux/fs/f2fs/
H A Dcompress.c89 struct folio *folio = page_folio(page); in f2fs_set_compressed_page()
140 return page_folio(ctx->rpages[0]); in f2fs_compress_control_folio()
596 folio = page_folio(page); in f2fs_compress_free_page()
847 unsigned long pgidx = page_folio(pages[index])->index; in f2fs_all_cluster_page_ready()
861 struct folio *folio = page_folio(pages[index + i]); in f2fs_all_cluster_page_ready()
884 if (page_folio(page)->index >= nr_pages) in cluster_has_invalid_data()
1070 if (folio_test_writeback(page_folio(cc->rpages[i]))) in cancel_cluster_writeback()
1200 struct folio *folio = page_folio(cc.rpages[0]); in f2fs_compress_write_end()
1244 struct folio *folio = page_folio(rpages[i]); in f2fs_truncate_partial_cluster()
1330 folio = page_folio(cc->rpages[last_index]); in f2fs_write_compressed_pages()
[all …]
/linux/fs/erofs/
H A Dzdata.c603 folio = page_folio(pcl->compressed_bvecs[i].page); in erofs_try_to_free_all_cached_folios()
633 if (bvec->page && page_folio(bvec->page) == folio) { in z_erofs_cache_release_folio()
868 folio_get(page_folio((struct page *)ptr)); in z_erofs_pcluster_begin()
1103 return !page_folio(page)->mapping && !z_erofs_is_shortlived_page(page); in z_erofs_page_is_invalidated()
1187 erofs_onlinefolio_end(page_folio(bvi->bvec.page), err, true); in z_erofs_fill_other_copies()
1238 erofs_folio_is_managed(EROFS_SB(be->sb), page_folio(page))) { in z_erofs_parse_in_bvecs()
1330 folio_put(page_folio(pcl->compressed_bvecs[0].page)); in z_erofs_decompress_pcluster()
1338 if (erofs_folio_is_managed(sbi, page_folio(page))) { in z_erofs_decompress_pcluster()
1359 erofs_onlinefolio_end(page_folio(page), err, true); in z_erofs_decompress_pcluster()
1513 folio = page_folio(zb in z_erofs_fill_bio_vec()
[all...]
/linux/arch/arm/mm/
H A Dcopypage-xscale.c87 struct folio *src = page_folio(from); in xscale_mc_copy_user_highpage()
H A Dcopypage-v4mc.c67 struct folio *src = page_folio(from); in v4_mc_copy_user_highpage()
/linux/arch/nios2/mm/
H A Dcacheflush.c205 flush_dcache_folio(page_folio(page)); in flush_dcache_page()
229 folio = page_folio(pfn_to_page(pfn)); in update_mmu_cache_range()

123456