| /linux/arch/arm64/mm/ |
| H A D | copypage.c | 35 from != folio_page(src, 0)) in copy_highpage() 48 kfrom = page_address(folio_page(src, i)); in copy_highpage() 49 kto = page_address(folio_page(dst, i)); in copy_highpage()
|
| /linux/mm/ |
| H A D | secretmem.c | 75 err = set_direct_map_invalid_noflush(folio_page(folio, 0)); in secretmem_fault() 90 set_direct_map_default_noflush(folio_page(folio, 0)); in secretmem_fault() 154 set_direct_map_default_noflush(folio_page(folio, 0)); in secretmem_free_folio()
|
| H A D | util.c | 727 copy_highpage(folio_page(dst, i), folio_page(src, i)); in folio_copy() 741 if (copy_mc_highpage(folio_page(dst, i), folio_page(src, i))) in folio_mc_copy() 1132 flush_dcache_page(folio_page(folio, i)); in flush_dcache_folio()
|
| H A D | truncate.c | 254 split_at = folio_page(folio, PAGE_ALIGN_DOWN(offset) / PAGE_SIZE); in truncate_inode_partial_folio() 266 split_at2 = folio_page(folio, in truncate_inode_partial_folio()
|
| H A D | page_io.c | 213 entry = page_swap_entry(folio_page(folio, i)); in swap_zeromap_folio_set() 231 entry = page_swap_entry(folio_page(folio, i)); in swap_zeromap_folio_clear()
|
| H A D | khugepaged.c | 786 struct page *page = folio_page(folio, i); in __collapse_huge_page_copy() 1558 if (folio_page(folio, i) != page) in try_collapse_pte_mapped_thp() 1614 if (folio_page(folio, i) != page) in try_collapse_pte_mapped_thp() 2104 dst = folio_page(new_folio, 0); in collapse_file() 2115 if (copy_mc_highpage(dst, folio_page(folio, i)) > 0) { in collapse_file()
|
| H A D | memremap.c | 429 __ClearPageAnonExclusive(folio_page(folio, i)); in free_zone_device_folio()
|
| H A D | rmap.c | 1656 struct page *page = folio_page(folio, i); in folio_add_new_anon_rmap() 2097 subpage = folio_page(folio, pfn - folio_pfn(folio)); in try_to_unmap_one() 2472 subpage = folio_page(folio, pfn - folio_pfn(folio)); in try_to_migrate_one() 2503 subpage = folio_page(folio, pfn - folio_pfn(folio)); in try_to_migrate_one()
|
| H A D | slab.h | 153 #define slab_page(s) folio_page(slab_folio(s), 0)
|
| H A D | migrate_device.c | 210 unsigned long pfn = page_to_pfn(folio_page(folio, 0)); in migrate_vma_collect_huge_pmd() 217 ret = set_pmd_migration_entry(&pvmw, folio_page(folio, 0)); in migrate_vma_collect_huge_pmd()
|
| /linux/lib/ |
| H A D | test_hmm.c | 600 *ppage = folio_page(mdevice->free_folios, 0); in dmirror_allocate_chunk() 643 rpage = folio_page(folio_alloc(GFP_HIGHUSER, order), 0); in dmirror_devmem_alloc_page() 650 dpage = folio_page(mdevice->free_folios, 0); in dmirror_devmem_alloc_page() 666 page_pgmap(folio_page(page_folio(dpage), 0)), in dmirror_devmem_alloc_page() 870 entry = folio_page(page_folio(rpage), i); in dmirror_migrate_finalize_and_map() 980 dpage = folio_page(vma_alloc_folio(GFP_HIGHUSER_MOVABLE, in dmirror_devmem_fault_alloc_and_copy() 1408 dpage = folio_page(folio_alloc(GFP_HIGHUSER_MOVABLE, in dmirror_device_evict_chunk() 1439 if (dmirror_page_to_chunk(folio_page(folio, 0)) == devmem) in dmirror_remove_free_pages() 1691 struct page *rpage = BACKING_PAGE(folio_page(head, 0)); in dmirror_devmem_folio_split() 1712 rpage_tail = folio_page(rfolio, offset); in dmirror_devmem_folio_split() [all …]
|
| H A D | scatterlist.c | 1302 sg_set_page(sg, folio_page(folio, 0), part, offset); in extract_folioq_to_sg() 1363 sg_set_page(sg, folio_page(folio, 0), len, offset); in extract_xarray_to_sg()
|
| /linux/fs/freevxfs/ |
| H A D | vxfs_immed.c | 38 memcpy_to_page(folio_page(folio, i), 0, src, PAGE_SIZE); in vxfs_immed_read_folio()
|
| /linux/arch/arm/mm/ |
| H A D | flush.c | 220 void *addr = kmap_high_get(folio_page(folio, i)); in __flush_dcache_folio() 223 kunmap_high(folio_page(folio, i)); in __flush_dcache_folio()
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_dmem.c | 160 struct page *dpage = folio_page(dfolio, 0); in nouveau_dmem_copy_folio() 161 struct page *spage = folio_page(sfolio, 0); in nouveau_dmem_copy_folio() 242 dpage = folio_page(vma_alloc_folio(GFP_HIGHUSER | __GFP_ZERO, in nouveau_dmem_migrate_to_ram() 428 zone_device_folio_init(folio, page_pgmap(folio_page(folio, 0)), order); in nouveau_dmem_page_alloc_locked() 497 dpage = folio_page( in nouveau_dmem_evict_chunk()
|
| /linux/arch/openrisc/mm/ |
| H A D | cache.c | 97 sync_icache_dcache(folio_page(folio, nr)); in update_cache()
|
| /linux/arch/mips/include/asm/ |
| H A D | cacheflush.h | 59 __flush_dcache_folio_pages(folio, folio_page(folio, 0), in flush_dcache_folio()
|
| /linux/include/linux/ |
| H A D | highmem-internal.h | 86 const struct page *page = folio_page(folio, offset / PAGE_SIZE); in kmap_local_folio()
|
| H A D | nfs_page.h | 195 return folio_page(folio, pgbase >> PAGE_SHIFT); in nfs_page_to_page()
|
| H A D | pagemap.h | 983 return folio_page(folio, index & (folio_nr_pages(folio) - 1)); in folio_file_page() 1488 array[i++] = folio_page(folio, 0); in __readahead_batch()
|
| /linux/virt/kvm/ |
| H A D | guest_memfd.c | 417 clear_highpage(folio_page(folio, 0)); in kvm_gmem_fault_user_mapping() 528 struct page *page = folio_page(folio, 0); in kvm_gmem_free_folio() 805 clear_highpage(folio_page(folio, 0)); in kvm_gmem_get_pfn()
|
| /linux/fs/btrfs/ |
| H A D | tree-checker.c | 68 dump_page(folio_page(eb->folios[0], 0), "eb page dump"); in generic_err() 96 dump_page(folio_page(eb->folios[0], 0), "eb page dump"); in file_extent_err() 157 dump_page(folio_page(eb->folios[0], 0), "eb page dump"); in dir_item_err() 673 dump_page(folio_page(eb->folios[0], 0), "eb page dump"); in block_group_err() 1070 dump_page(folio_page(eb->folios[0], 0), "eb page dump"); in dev_item_err() 1343 dump_page(folio_page(eb->folios[0], 0), "eb page dump"); in extent_err()
|
| /linux/fs/crypto/ |
| H A D | crypto.c | 274 struct page *page = folio_page(folio, i >> PAGE_SHIFT); in fscrypt_decrypt_pagecache_blocks()
|
| /linux/drivers/dma-buf/ |
| H A D | udmabuf.c | 123 pages[pg] = folio_page(ubuf->folios[pg], in vmap_udmabuf()
|
| /linux/fs/ecryptfs/ |
| H A D | crypto.c | 376 folio_page(folio, 0), folio->index, in ecryptfs_encrypt_page() 449 struct page *page = folio_page(folio, 0); in ecryptfs_decrypt_page()
|