| /linux/drivers/gpu/drm/ |
| H A D | drm_pagemap.c | 570 struct page *src_page = migrate_pfn_to_page(migrate.src[i]); in drm_pagemap_migrate_to_devmem() local 571 unsigned long nr_pages = src_page ? NR_PAGES(folio_order(page_folio(src_page))) : 1; in drm_pagemap_migrate_to_devmem() 573 if (src_page && is_zone_device_page(src_page)) { in drm_pagemap_migrate_to_devmem() 574 if (page_pgmap(src_page) == pagemap) in drm_pagemap_migrate_to_devmem() 601 struct page *src_page = migrate_pfn_to_page(migrate.src[i]); in drm_pagemap_migrate_to_devmem() local 605 if (src_page && is_device_private_page(src_page)) { in drm_pagemap_migrate_to_devmem() 606 struct drm_pagemap_zdd *src_zdd = src_page->zone_device_data; in drm_pagemap_migrate_to_devmem() 608 if (page_pgmap(src_page) == pagemap && in drm_pagemap_migrate_to_devmem() 618 pages[i] = src_page; in drm_pagemap_migrate_to_devmem() 716 struct page *page = NULL, *src_page; in drm_pagemap_migrate_populate_ram_pfn() local [all …]
|
| /linux/crypto/ |
| H A D | scatterwalk.c | 115 struct page *src_page = sg_page(src); in memcpy_sglist() local 130 src_page += src_offset / PAGE_SIZE; in memcpy_sglist() 133 if (src_page != dst_page) { in memcpy_sglist() 136 src_page, src_oip, len); in memcpy_sglist() 152 src_virt = page_address(src_page) + src_offset; in memcpy_sglist()
|
| /linux/virt/kvm/ |
| H A D | guest_memfd.c | 825 struct file *file, gfn_t gfn, struct page *src_page, in __kvm_gmem_populate() argument 850 ret = post_populate(kvm, gfn, pfn, src_page, opaque); in __kvm_gmem_populate() 886 struct page *src_page = NULL; in kvm_gmem_populate() local 896 ret = get_user_pages_fast(uaddr, 1, 0, &src_page); in kvm_gmem_populate() 905 ret = __kvm_gmem_populate(kvm, slot, file, start_gfn + i, src_page, in kvm_gmem_populate() 908 if (src_page) in kvm_gmem_populate() 909 put_page(src_page); in kvm_gmem_populate()
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_blit.c | 369 u32 src_page = src_offset >> PAGE_SHIFT; in vmw_bo_cpu_blit_line() local 373 bool unmap_src = d->src_addr && (src_page != d->mapped_src || in vmw_bo_cpu_blit_line() 403 if (WARN_ON_ONCE(src_page >= d->src_num_pages)) in vmw_bo_cpu_blit_line() 407 kmap_atomic_prot(d->src_pages[src_page], in vmw_bo_cpu_blit_line() 412 d->mapped_src = src_page; in vmw_bo_cpu_blit_line()
|
| /linux/fs/crypto/ |
| H A D | crypto.c | 112 struct page *src_page, struct page *dest_page, in fscrypt_crypt_data_unit() argument 134 sg_set_page(&src, src_page, len, offs); in fscrypt_crypt_data_unit()
|
| H A D | fscrypt_private.h | 339 struct page *src_page, struct page *dest_page,
|
| /linux/lib/ |
| H A D | test_hmm.c | 715 struct page *src_page; in dmirror_migrate_alloc_and_copy() local 733 src_page = pfn_to_page(spfn + i); in dmirror_migrate_alloc_and_copy() 736 copy_highpage(rpage, src_page); in dmirror_migrate_alloc_and_copy() 770 struct page *src_page = in dmirror_migrate_alloc_and_copy() local 774 copy_highpage(dst_page, src_page); in dmirror_migrate_alloc_and_copy() 998 struct page *src_page; in dmirror_devmem_fault_alloc_and_copy() local 1012 src_page = pfn_to_page(page_to_pfn(spage) + i); in dmirror_devmem_fault_alloc_and_copy() 1016 copy_highpage(dst_page, src_page); in dmirror_devmem_fault_alloc_and_copy()
|
| /linux/arch/x86/kernel/cpu/sgx/ |
| H A D | ioctl.c | 214 struct page *src_page; in __sgx_encl_add_page() local 225 ret = get_user_pages(src, 1, 0, &src_page); in __sgx_encl_add_page() 232 pginfo.contents = (unsigned long)kmap_local_page(src_page); in __sgx_encl_add_page() 237 put_page(src_page); in __sgx_encl_add_page()
|
| /linux/io_uring/ |
| H A D | zcrx.c | 1241 static ssize_t io_copy_page(struct io_copy_cache *cc, struct page *src_page, in io_copy_page() argument 1255 folio_test_partial_kmap(page_folio(src_page))) { in io_copy_page() 1258 src_page += src_offset / PAGE_SIZE; in io_copy_page() 1265 src_addr = kmap_local_page(src_page) + src_offset; in io_copy_page() 1282 struct page *src_page, unsigned int src_offset, in io_zcrx_copy_chunk() argument 1303 n = io_copy_page(&cc, src_page, src_offset, len); in io_zcrx_copy_chunk()
|
| /linux/mm/ |
| H A D | khugepaged.c | 698 struct page *src_page = pte_page(pteval); in __collapse_huge_page_copy_succeeded() local 700 src = page_folio(src_page); in __collapse_huge_page_copy_succeeded() 717 folio_remove_rmap_ptes(src, src_page, nr_ptes, vma); in __collapse_huge_page_copy_succeeded() 788 struct page *src_page; in __collapse_huge_page_copy() local 794 src_page = pte_page(pteval); in __collapse_huge_page_copy() 795 if (copy_mc_user_highpage(page, src_page, src_addr, vma) > 0) { in __collapse_huge_page_copy()
|
| H A D | huge_memory.c | 1851 struct page *src_page; in copy_huge_pmd() local 1918 src_page = pmd_page(pmd); in copy_huge_pmd() 1919 VM_BUG_ON_PAGE(!PageHead(src_page), src_page); in copy_huge_pmd() 1920 src_folio = page_folio(src_page); in copy_huge_pmd() 1923 if (unlikely(folio_try_dup_anon_rmap_pmd(src_folio, src_page, dst_vma, src_vma))) { in copy_huge_pmd() 2708 struct page *src_page; in move_pages_huge_pmd() local 2738 src_page = pmd_page(src_pmdval); in move_pages_huge_pmd() 2741 if (unlikely(!PageAnonExclusive(src_page))) { in move_pages_huge_pmd() 2746 src_folio = page_folio(src_page); in move_pages_huge_pmd()
|
| H A D | memory.c | 7362 struct page *src_page; in copy_user_gigantic_page() local 7367 src_page = folio_page(src, i); in copy_user_gigantic_page() 7370 if (copy_mc_user_highpage(dst_page, src_page, in copy_user_gigantic_page()
|
| /linux/fs/ecryptfs/ |
| H A D | crypto.c | 300 struct page *src_page, in crypt_extent() argument 316 sg_set_page(&src_sg, src_page, extent_size, in crypt_extent()
|
| /linux/arch/x86/kvm/svm/ |
| H A D | sev.c | 2281 struct page *src_page, void *opaque) in sev_gmem_post_populate() argument 2290 if (WARN_ON_ONCE(sev_populate_args->type != KVM_SEV_SNP_PAGE_TYPE_ZERO && !src_page)) in sev_gmem_post_populate() 2301 if (src_page) { in sev_gmem_post_populate() 2302 void *src_vaddr = kmap_local_page(src_page); in sev_gmem_post_populate() 2338 void *src_vaddr = kmap_local_page(src_page); in sev_gmem_post_populate()
|
| /linux/arch/x86/kvm/vmx/ |
| H A D | tdx.c | 3120 struct page *src_page, void *_arg) in tdx_gmem_post_populate() argument 3131 if (!src_page) in tdx_gmem_post_populate() 3134 kvm_tdx->page_add_src = src_page; in tdx_gmem_post_populate()
|