| /linux/mm/ |
| H A D | folio-compat.c | 16 return folio_unlock(page_folio(page)); in unlock_page() 22 return folio_end_writeback(page_folio(page)); in end_page_writeback() 28 return folio_wait_writeback(page_folio(page)); in wait_on_page_writeback() 34 folio_mark_accessed(page_folio(page)); in mark_page_accessed() 40 folio_start_writeback(page_folio(page)); in set_page_writeback() 46 return folio_mark_dirty(page_folio(page)); in set_page_dirty() 52 return folio_mark_dirty_lock(page_folio(page)); in set_page_dirty_lock() 58 return folio_clear_dirty_for_io(page_folio(page)); in clear_page_dirty_for_io() 65 return folio_redirty_for_writepage(wbc, page_folio(page)); in redirty_page_for_writepage() 72 return filemap_add_folio(mapping, page_folio(pag in add_to_page_cache_lru() [all...] |
| H A D | migrate_device.c | 86 struct folio *fault_folio = fault_page ? page_folio(fault_page) : NULL; in migrate_vma_split_folio() 103 new_fault_folio = fault_page ? page_folio(fault_page) : NULL; in migrate_vma_split_folio() 258 page_folio(migrate->fault_page) : NULL; in migrate_vma_collect_pmd() 312 folio = page_folio(page); in migrate_vma_collect_pmd() 355 folio = page ? page_folio(page) : NULL; in migrate_vma_collect_pmd() 391 folio = page_folio(page); in migrate_vma_collect_pmd() 539 struct folio *folio = page_folio(page); in migrate_vma_check_page() 571 page_folio(fault_page) : NULL; in migrate_device_unmap() 589 folio = page_folio(page); in migrate_device_unmap() 641 folio = page_folio(page); in migrate_device_unmap() [all …]
|
| H A D | memory-failure.c | 187 ret = dissolve_free_hugetlb_folio(page_folio(page)); in __page_handle_poison() 334 shake_folio(page_folio(page)); in shake_page() 400 tk->size_shift = folio_shift(page_folio(p)); in __add_to_kill() 984 count -= folio_nr_pages(page_folio(p)); in has_extra_refcount() 1022 struct folio *folio = page_folio(p); in me_pagecache_clean() 1080 struct folio *folio = page_folio(p); in me_pagecache_dirty() 1118 struct folio *folio = page_folio(p); in me_swapcache_dirty() 1140 struct folio *folio = page_folio(p); in me_swapcache_clean() 1162 struct folio *folio = page_folio(p); in me_huge_page() 1356 struct folio *folio = page_folio(page); in __get_hwpoison_page() [all …]
|
| H A D | hwpoison-inject.c | 22 struct folio *folio = page_folio(p); in hwpoison_filter_dev() 116 folio = page_folio(p); in hwpoison_inject()
|
| H A D | gup.c | 56 folio = page_folio(page); in sanity_check_pinned_pages() 79 folio = page_folio(page); in try_get_folio() 87 * could be that between calling page_folio() and the refcount in try_get_folio() 94 if (unlikely(page_folio(page) != folio)) { in try_get_folio() 188 gup_put_folio(page_folio(page), 1, FOLL_PIN); in unpin_user_page() 236 struct folio *folio = page_folio(next); in gup_folio_range_next() 250 struct folio *folio = page_folio(list[i]); in gup_folio_next() 254 if (page_folio(list[nr]) != folio) in gup_folio_next() 541 return page_folio(page); in try_grab_folio_fast() 674 ret = try_grab_folio(page_folio(pag in follow_huge_pud() [all...] |
| H A D | mmu_gather.c | 68 folio_remove_rmap_ptes(page_folio(page), page, nr_pages, in tlb_flush_rmap_batch() 178 VM_WARN_ON_ONCE(page_folio(page) != page_folio(page + nr_pages - 1)); in __tlb_remove_folio_pages_size()
|
| H A D | page_idle.c | 42 folio = page_folio(page); in page_idle_get_folio() 45 if (unlikely(page_folio(page) != folio || !folio_test_lru(folio))) { in page_idle_get_folio()
|
| /linux/arch/sh/mm/ |
| H A D | cache.c | 64 struct folio *folio = page_folio(page); in copy_to_user_page() 85 struct folio *folio = page_folio(page); in copy_from_user_page() 102 struct folio *src = page_folio(from); in copy_user_highpage() 150 struct folio *folio = page_folio(pfn_to_page(pfn)); in __update_cache() 160 struct folio *folio = page_folio(page); in __flush_anon_page() 247 cacheop_on_each_cpu(local_flush_icache_folio, page_folio(page), 1); in flush_icache_pages()
|
| H A D | kmap.c | 30 struct folio *folio = page_folio(page); in kmap_coherent()
|
| /linux/include/linux/ |
| H A D | memremap.h | 176 folio_is_device_private(page_folio(page)); in is_device_private_page() 201 folio_is_pci_p2pdma(page_folio(page)); in is_pci_p2pdma_page() 212 return folio_is_device_coherent(page_folio(page)); in is_device_coherent_page() 223 return folio_is_fsdax(page_folio(page)); in is_fsdax_page()
|
| H A D | page-flags.h | 306 #define page_folio(p) (_Generic((p), \ macro 729 unsigned long flags = (unsigned long)page_folio(page)->mapping; in PageAnonNotKsm() 736 return folio_test_anon(page_folio(page)); in PageAnon() 804 return folio_test_uptodate(page_folio(page)); in PageUptodate() 1079 return folio_test_hugetlb(page_folio(page)); in PAGE_TYPE_OPS() 1093 folio = page_folio(page); in is_page_hwpoison()
|
| /linux/lib/ |
| H A D | test_hmm.c | 583 mdevice->free_folios = page_folio(page); in dmirror_allocate_chunk() 665 zone_device_folio_init(page_folio(dpage), in dmirror_devmem_alloc_page() 666 page_pgmap(folio_page(page_folio(dpage), 0)), in dmirror_devmem_alloc_page() 723 folio = page_folio(spage); in dmirror_migrate_alloc_and_copy() 763 struct folio *folio = page_folio(dpage); in dmirror_migrate_alloc_and_copy() 860 nr = folio_nr_pages(page_folio(dpage)); in dmirror_migrate_finalize_and_map() 867 VM_WARN_ON(folio_nr_pages(page_folio(rpage)) != nr); in dmirror_migrate_finalize_and_map() 870 entry = folio_page(page_folio(rpage), i); in dmirror_migrate_finalize_and_map() 969 order = folio_order(page_folio(spage)); in dmirror_devmem_fault_alloc_and_copy() 1405 order = folio_order(page_folio(spage)); in dmirror_device_evict_chunk() [all …]
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_dmem.c | 205 sfolio = page_folio(vmf->page); in nouveau_dmem_migrate_to_ram() 255 dfolio = page_folio(dpage); in nouveau_dmem_migrate_to_ram() 360 drm->dmem->free_folios = page_folio(page); in nouveau_dmem_chunk_alloc() 417 folio = page_folio(page); in nouveau_dmem_page_alloc_locked() 423 folio = page_folio(page); in nouveau_dmem_page_alloc_locked() 492 struct folio *folio = page_folio( in nouveau_dmem_evict_chunk() 511 page_folio(migrate_pfn_to_page(src_pfns[i])), in nouveau_dmem_evict_chunk() 512 page_folio(dpage), in nouveau_dmem_evict_chunk() 753 if (drm->dmem->migrate.copy_func(drm, folio_nr_pages(page_folio(spage)), in nouveau_dmem_migrate_copy_one() 770 if (folio_order(page_folio(dpage))) in nouveau_dmem_migrate_copy_one() [all …]
|
| /linux/arch/csky/abiv1/ |
| H A D | cacheflush.c | 40 flush_dcache_folio(page_folio(page)); in flush_dcache_page() 58 folio = page_folio(pfn_to_page(pfn)); in update_mmu_cache_range()
|
| /linux/arch/arm64/mm/ |
| H A D | copypage.c | 21 struct folio *src = page_folio(from); in copy_highpage() 22 struct folio *dst = page_folio(to); in copy_highpage()
|
| H A D | flush.c | 54 struct folio *folio = page_folio(pte_page(pte)); in __sync_icache_dcache() 79 flush_dcache_folio(page_folio(page)); in flush_dcache_page()
|
| /linux/arch/xtensa/mm/ |
| H A D | cache.c | 90 struct folio *folio = page_folio(page); in clear_user_highpage() 105 struct folio *folio = page_folio(dst); in copy_user_highpage() 226 folio = page_folio(pfn_to_page(pfn)); in update_mmu_cache_range()
|
| /linux/arch/mips/mm/ |
| H A D | init.c | 85 BUG_ON(folio_test_dcache_dirty(page_folio(page))); in __kmap_pgprot() 166 struct folio *src = page_folio(from); in copy_user_highpage() 192 struct folio *folio = page_folio(page); in copy_to_user_page() 212 struct folio *folio = page_folio(page); in copy_from_user_page()
|
| H A D | cache.c | 130 struct folio *folio = page_folio(page); in __flush_anon_page() 157 folio = page_folio(pfn_to_page(pfn)); in __update_cache()
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_pagemap.c | 243 folio = page_folio(page); in drm_pagemap_migrate_map_pages() 561 unsigned long nr_pages = src_page ? NR_PAGES(folio_order(page_folio(src_page))) : 1; in drm_pagemap_migrate_to_devmem() 653 unsigned long nr_pages = page ? NR_PAGES(folio_order(page_folio(page))) : 1; in drm_pagemap_migrate_to_devmem() 723 order = folio_order(page_folio(src_page)); in drm_pagemap_migrate_populate_ram_pfn() 753 WARN_ON_ONCE(!folio_trylock(page_folio(page))); in drm_pagemap_migrate_populate_ram_pfn() 755 order = folio_order(page_folio(page)); in drm_pagemap_migrate_populate_ram_pfn() 775 order = folio_order(page_folio(page)); in drm_pagemap_migrate_populate_ram_pfn()
|
| /linux/fs/f2fs/ |
| H A D | compress.c | 89 struct folio *folio = page_folio(page); in f2fs_set_compressed_page() 140 return page_folio(ctx->rpages[0]); in f2fs_compress_control_folio() 596 folio = page_folio(page); in f2fs_compress_free_page() 847 unsigned long pgidx = page_folio(pages[index])->index; in f2fs_all_cluster_page_ready() 861 struct folio *folio = page_folio(pages[index + i]); in f2fs_all_cluster_page_ready() 884 if (page_folio(page)->index >= nr_pages) in cluster_has_invalid_data() 1070 if (folio_test_writeback(page_folio(cc->rpages[i]))) in cancel_cluster_writeback() 1200 struct folio *folio = page_folio(cc.rpages[0]); in f2fs_compress_write_end() 1244 struct folio *folio = page_folio(rpages[i]); in f2fs_truncate_partial_cluster() 1330 folio = page_folio(cc->rpages[last_index]); in f2fs_write_compressed_pages() [all …]
|
| /linux/fs/erofs/ |
| H A D | zdata.c | 603 folio = page_folio(pcl->compressed_bvecs[i].page); in erofs_try_to_free_all_cached_folios() 633 if (bvec->page && page_folio(bvec->page) == folio) { in z_erofs_cache_release_folio() 868 folio_get(page_folio((struct page *)ptr)); in z_erofs_pcluster_begin() 1103 return !page_folio(page)->mapping && !z_erofs_is_shortlived_page(page); in z_erofs_page_is_invalidated() 1187 erofs_onlinefolio_end(page_folio(bvi->bvec.page), err, true); in z_erofs_fill_other_copies() 1238 erofs_folio_is_managed(EROFS_SB(be->sb), page_folio(page))) { in z_erofs_parse_in_bvecs() 1330 folio_put(page_folio(pcl->compressed_bvecs[0].page)); in z_erofs_decompress_pcluster() 1338 if (erofs_folio_is_managed(sbi, page_folio(page))) { in z_erofs_decompress_pcluster() 1359 erofs_onlinefolio_end(page_folio(page), err, true); in z_erofs_decompress_pcluster() 1513 folio = page_folio(zb in z_erofs_fill_bio_vec() [all...] |
| /linux/arch/arm/mm/ |
| H A D | copypage-xscale.c | 87 struct folio *src = page_folio(from); in xscale_mc_copy_user_highpage()
|
| H A D | copypage-v4mc.c | 67 struct folio *src = page_folio(from); in v4_mc_copy_user_highpage()
|
| /linux/arch/nios2/mm/ |
| H A D | cacheflush.c | 205 flush_dcache_folio(page_folio(page)); in flush_dcache_page() 229 folio = page_folio(pfn_to_page(pfn)); in update_mmu_cache_range()
|