Home
last modified time | relevance | path

Searched refs:folio_page (Results 1 – 25 of 53) sorted by relevance

123

/linux/arch/arm64/mm/
H A Dcopypage.c35 from != folio_page(src, 0)) in copy_highpage()
48 kfrom = page_address(folio_page(src, i)); in copy_highpage()
49 kto = page_address(folio_page(dst, i)); in copy_highpage()
/linux/mm/
H A Dsecretmem.c75 err = set_direct_map_invalid_noflush(folio_page(folio, 0)); in secretmem_fault()
90 set_direct_map_default_noflush(folio_page(folio, 0)); in secretmem_fault()
154 set_direct_map_default_noflush(folio_page(folio, 0)); in secretmem_free_folio()
H A Dutil.c727 copy_highpage(folio_page(dst, i), folio_page(src, i)); in folio_copy()
741 if (copy_mc_highpage(folio_page(dst, i), folio_page(src, i))) in folio_mc_copy()
1132 flush_dcache_page(folio_page(folio, i)); in flush_dcache_folio()
H A Dtruncate.c254 split_at = folio_page(folio, PAGE_ALIGN_DOWN(offset) / PAGE_SIZE); in truncate_inode_partial_folio()
266 split_at2 = folio_page(folio, in truncate_inode_partial_folio()
H A Dpage_io.c213 entry = page_swap_entry(folio_page(folio, i)); in swap_zeromap_folio_set()
231 entry = page_swap_entry(folio_page(folio, i)); in swap_zeromap_folio_clear()
H A Dkhugepaged.c786 struct page *page = folio_page(folio, i); in __collapse_huge_page_copy()
1558 if (folio_page(folio, i) != page) in try_collapse_pte_mapped_thp()
1614 if (folio_page(folio, i) != page) in try_collapse_pte_mapped_thp()
2104 dst = folio_page(new_folio, 0); in collapse_file()
2115 if (copy_mc_highpage(dst, folio_page(folio, i)) > 0) { in collapse_file()
H A Dmemremap.c429 __ClearPageAnonExclusive(folio_page(folio, i)); in free_zone_device_folio()
H A Drmap.c1656 struct page *page = folio_page(folio, i); in folio_add_new_anon_rmap()
2097 subpage = folio_page(folio, pfn - folio_pfn(folio)); in try_to_unmap_one()
2472 subpage = folio_page(folio, pfn - folio_pfn(folio)); in try_to_migrate_one()
2503 subpage = folio_page(folio, pfn - folio_pfn(folio)); in try_to_migrate_one()
H A Dslab.h153 #define slab_page(s) folio_page(slab_folio(s), 0)
H A Dmigrate_device.c210 unsigned long pfn = page_to_pfn(folio_page(folio, 0)); in migrate_vma_collect_huge_pmd()
217 ret = set_pmd_migration_entry(&pvmw, folio_page(folio, 0)); in migrate_vma_collect_huge_pmd()
/linux/lib/
H A Dtest_hmm.c600 *ppage = folio_page(mdevice->free_folios, 0); in dmirror_allocate_chunk()
643 rpage = folio_page(folio_alloc(GFP_HIGHUSER, order), 0); in dmirror_devmem_alloc_page()
650 dpage = folio_page(mdevice->free_folios, 0); in dmirror_devmem_alloc_page()
666 page_pgmap(folio_page(page_folio(dpage), 0)), in dmirror_devmem_alloc_page()
870 entry = folio_page(page_folio(rpage), i); in dmirror_migrate_finalize_and_map()
980 dpage = folio_page(vma_alloc_folio(GFP_HIGHUSER_MOVABLE, in dmirror_devmem_fault_alloc_and_copy()
1408 dpage = folio_page(folio_alloc(GFP_HIGHUSER_MOVABLE, in dmirror_device_evict_chunk()
1439 if (dmirror_page_to_chunk(folio_page(folio, 0)) == devmem) in dmirror_remove_free_pages()
1691 struct page *rpage = BACKING_PAGE(folio_page(head, 0)); in dmirror_devmem_folio_split()
1712 rpage_tail = folio_page(rfolio, offset); in dmirror_devmem_folio_split()
[all …]
H A Dscatterlist.c1302 sg_set_page(sg, folio_page(folio, 0), part, offset); in extract_folioq_to_sg()
1363 sg_set_page(sg, folio_page(folio, 0), len, offset); in extract_xarray_to_sg()
/linux/fs/freevxfs/
H A Dvxfs_immed.c38 memcpy_to_page(folio_page(folio, i), 0, src, PAGE_SIZE); in vxfs_immed_read_folio()
/linux/arch/arm/mm/
H A Dflush.c220 void *addr = kmap_high_get(folio_page(folio, i)); in __flush_dcache_folio()
223 kunmap_high(folio_page(folio, i)); in __flush_dcache_folio()
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_dmem.c160 struct page *dpage = folio_page(dfolio, 0); in nouveau_dmem_copy_folio()
161 struct page *spage = folio_page(sfolio, 0); in nouveau_dmem_copy_folio()
242 dpage = folio_page(vma_alloc_folio(GFP_HIGHUSER | __GFP_ZERO, in nouveau_dmem_migrate_to_ram()
428 zone_device_folio_init(folio, page_pgmap(folio_page(folio, 0)), order); in nouveau_dmem_page_alloc_locked()
497 dpage = folio_page( in nouveau_dmem_evict_chunk()
/linux/arch/openrisc/mm/
H A Dcache.c97 sync_icache_dcache(folio_page(folio, nr)); in update_cache()
/linux/arch/mips/include/asm/
H A Dcacheflush.h59 __flush_dcache_folio_pages(folio, folio_page(folio, 0), in flush_dcache_folio()
/linux/include/linux/
H A Dhighmem-internal.h86 const struct page *page = folio_page(folio, offset / PAGE_SIZE); in kmap_local_folio()
H A Dnfs_page.h195 return folio_page(folio, pgbase >> PAGE_SHIFT); in nfs_page_to_page()
H A Dpagemap.h983 return folio_page(folio, index & (folio_nr_pages(folio) - 1)); in folio_file_page()
1488 array[i++] = folio_page(folio, 0); in __readahead_batch()
/linux/virt/kvm/
H A Dguest_memfd.c417 clear_highpage(folio_page(folio, 0)); in kvm_gmem_fault_user_mapping()
528 struct page *page = folio_page(folio, 0); in kvm_gmem_free_folio()
805 clear_highpage(folio_page(folio, 0)); in kvm_gmem_get_pfn()
/linux/fs/btrfs/
H A Dtree-checker.c68 dump_page(folio_page(eb->folios[0], 0), "eb page dump"); in generic_err()
96 dump_page(folio_page(eb->folios[0], 0), "eb page dump"); in file_extent_err()
157 dump_page(folio_page(eb->folios[0], 0), "eb page dump"); in dir_item_err()
673 dump_page(folio_page(eb->folios[0], 0), "eb page dump"); in block_group_err()
1070 dump_page(folio_page(eb->folios[0], 0), "eb page dump"); in dev_item_err()
1343 dump_page(folio_page(eb->folios[0], 0), "eb page dump"); in extent_err()
/linux/fs/crypto/
H A Dcrypto.c274 struct page *page = folio_page(folio, i >> PAGE_SHIFT); in fscrypt_decrypt_pagecache_blocks()
/linux/drivers/dma-buf/
H A Dudmabuf.c123 pages[pg] = folio_page(ubuf->folios[pg], in vmap_udmabuf()
/linux/fs/ecryptfs/
H A Dcrypto.c376 folio_page(folio, 0), folio->index, in ecryptfs_encrypt_page()
449 struct page *page = folio_page(folio, 0); in ecryptfs_decrypt_page()

123