/linux/fs/hfsplus/ |
H A D | bnode.c | 214 struct page **src_page, **dst_page; in hfs_bnode_copy() local 226 src_page = src_node->page + (src >> PAGE_SHIFT); in hfs_bnode_copy() 233 memcpy_page(*dst_page, src, *src_page, src, l); in hfs_bnode_copy() 238 memcpy_page(*++dst_page, 0, *++src_page, 0, l); in hfs_bnode_copy() 246 src_ptr = kmap_local_page(*src_page) + src; in hfs_bnode_copy() 264 src_page++; in hfs_bnode_copy() 271 struct page **src_page, **dst_page; in hfs_bnode_move() local 286 src_page = node->page + (src >> PAGE_SHIFT); in hfs_bnode_move() 295 src_ptr = kmap_local_page(*src_page); in hfs_bnode_move() 302 src_page--; in hfs_bnode_move() [all …]
|
/linux/drivers/gpu/drm/ |
H A D | drm_pagemap.c | 446 struct page *page, *src_page; in drm_pagemap_migrate_populate_ram_pfn() local 451 src_page = migrate_pfn_to_page(src_mpfn[i]); in drm_pagemap_migrate_populate_ram_pfn() 452 if (!src_page) in drm_pagemap_migrate_populate_ram_pfn() 456 if (src_page->zone_device_data != in drm_pagemap_migrate_populate_ram_pfn()
|
/linux/io_uring/ |
H A D | zcrx.c | 963 static ssize_t io_copy_page(struct io_copy_cache *cc, struct page *src_page, in io_copy_page() argument 977 folio_test_partial_kmap(page_folio(src_page))) { in io_copy_page() 980 src_page = nth_page(src_page, src_offset / PAGE_SIZE); in io_copy_page() 987 src_addr = kmap_local_page(src_page) + src_offset; in io_copy_page() 1003 struct page *src_page, unsigned int src_offset, in io_zcrx_copy_chunk() argument 1028 n = io_copy_page(&cc, src_page, src_offset, len); in io_zcrx_copy_chunk()
|
/linux/fs/crypto/ |
H A D | crypto.c | 112 struct page *src_page, struct page *dest_page, in fscrypt_crypt_data_unit() argument 134 sg_set_page(&src, src_page, len, offs); in fscrypt_crypt_data_unit()
|
H A D | fscrypt_private.h | 339 struct page *src_page, struct page *dest_page,
|
/linux/arch/x86/kernel/cpu/sgx/ |
H A D | ioctl.c | 214 struct page *src_page; in __sgx_encl_add_page() local 225 ret = get_user_pages(src, 1, 0, &src_page); in __sgx_encl_add_page() 232 pginfo.contents = (unsigned long)kmap_local_page(src_page); in __sgx_encl_add_page() 237 put_page(src_page); in __sgx_encl_add_page()
|
/linux/fs/hfs/ |
H A D | bnode.c | 201 struct page *src_page, *dst_page; in hfs_bnode_copy() local 212 src_page = src_node->page[0]; in hfs_bnode_copy() 215 memcpy_page(dst_page, dst, src_page, src, len); in hfs_bnode_copy()
|
/linux/mm/ |
H A D | khugepaged.c | 725 struct page *src_page = pte_page(pteval); in __collapse_huge_page_copy_succeeded() local 727 src = page_folio(src_page); in __collapse_huge_page_copy_succeeded() 744 folio_remove_rmap_ptes(src, src_page, nr_ptes, vma); in __collapse_huge_page_copy_succeeded() 815 struct page *src_page; in __collapse_huge_page_copy() local 821 src_page = pte_page(pteval); in __collapse_huge_page_copy() 822 if (copy_mc_user_highpage(page, src_page, src_addr, vma) > 0) { in __collapse_huge_page_copy()
|
H A D | huge_memory.c | 1671 struct page *src_page; in follow_devmap_pmd() 1755 src_page = pmd_page(pmd); in copy_huge_pmd() 1756 VM_BUG_ON_PAGE(!PageHead(src_page), src_page); in copy_huge_pmd() 1757 src_folio = page_folio(src_page); in copy_huge_pmd() 1760 if (unlikely(folio_try_dup_anon_rmap_pmd(src_folio, src_page, dst_vma, src_vma))) { in copy_huge_pmd() 2542 struct page *src_page; in change_huge_pud() 2573 src_page = pmd_page(src_pmdval); in move_pages_huge_pmd() 2576 if (unlikely(!PageAnonExclusive(src_page))) { in move_pages_huge_pmd() 2581 src_folio = page_folio(src_page); in move_pages_huge_pmd() 1694 struct page *src_page; copy_huge_pmd() local 2565 struct page *src_page; move_pages_huge_pmd() local [all...] |
H A D | memory.c | 7067 struct page *src_page; in copy_user_gigantic_page() local 7072 src_page = folio_page(src, i); in copy_user_gigantic_page() 7075 if (copy_mc_user_highpage(dst_page, src_page, in copy_user_gigantic_page()
|
/linux/arch/x86/kvm/vmx/ |
H A D | tdx.c | 3160 struct page *src_page; in tdx_gmem_post_populate() local 3168 ret = get_user_pages_fast((unsigned long)src, 1, 0, &src_page); in tdx_gmem_post_populate() 3194 src_page, &entry, &level_state); in tdx_gmem_post_populate() 3215 put_page(src_page); in tdx_gmem_post_populate()
|