Home
last modified time | relevance | path

Searched refs:src_page (Results 1 – 15 of 15) sorted by relevance

/linux/drivers/gpu/drm/
H A Ddrm_pagemap.c570 struct page *src_page = migrate_pfn_to_page(migrate.src[i]); in drm_pagemap_migrate_to_devmem() local
571 unsigned long nr_pages = src_page ? NR_PAGES(folio_order(page_folio(src_page))) : 1; in drm_pagemap_migrate_to_devmem()
573 if (src_page && is_zone_device_page(src_page)) { in drm_pagemap_migrate_to_devmem()
574 if (page_pgmap(src_page) == pagemap) in drm_pagemap_migrate_to_devmem()
601 struct page *src_page = migrate_pfn_to_page(migrate.src[i]); in drm_pagemap_migrate_to_devmem() local
605 if (src_page && is_device_private_page(src_page)) { in drm_pagemap_migrate_to_devmem()
606 struct drm_pagemap_zdd *src_zdd = src_page->zone_device_data; in drm_pagemap_migrate_to_devmem()
608 if (page_pgmap(src_page) == pagemap && in drm_pagemap_migrate_to_devmem()
618 pages[i] = src_page; in drm_pagemap_migrate_to_devmem()
716 struct page *page = NULL, *src_page; in drm_pagemap_migrate_populate_ram_pfn() local
[all …]
/linux/crypto/
H A Dscatterwalk.c115 struct page *src_page = sg_page(src); in memcpy_sglist() local
130 src_page += src_offset / PAGE_SIZE; in memcpy_sglist()
133 if (src_page != dst_page) { in memcpy_sglist()
136 src_page, src_oip, len); in memcpy_sglist()
152 src_virt = page_address(src_page) + src_offset; in memcpy_sglist()
/linux/virt/kvm/
H A Dguest_memfd.c825 struct file *file, gfn_t gfn, struct page *src_page, in __kvm_gmem_populate() argument
850 ret = post_populate(kvm, gfn, pfn, src_page, opaque); in __kvm_gmem_populate()
886 struct page *src_page = NULL; in kvm_gmem_populate() local
896 ret = get_user_pages_fast(uaddr, 1, 0, &src_page); in kvm_gmem_populate()
905 ret = __kvm_gmem_populate(kvm, slot, file, start_gfn + i, src_page, in kvm_gmem_populate()
908 if (src_page) in kvm_gmem_populate()
909 put_page(src_page); in kvm_gmem_populate()
/linux/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_blit.c369 u32 src_page = src_offset >> PAGE_SHIFT; in vmw_bo_cpu_blit_line() local
373 bool unmap_src = d->src_addr && (src_page != d->mapped_src || in vmw_bo_cpu_blit_line()
403 if (WARN_ON_ONCE(src_page >= d->src_num_pages)) in vmw_bo_cpu_blit_line()
407 kmap_atomic_prot(d->src_pages[src_page], in vmw_bo_cpu_blit_line()
412 d->mapped_src = src_page; in vmw_bo_cpu_blit_line()
/linux/fs/crypto/
H A Dcrypto.c112 struct page *src_page, struct page *dest_page, in fscrypt_crypt_data_unit() argument
134 sg_set_page(&src, src_page, len, offs); in fscrypt_crypt_data_unit()
H A Dfscrypt_private.h339 struct page *src_page, struct page *dest_page,
/linux/lib/
H A Dtest_hmm.c715 struct page *src_page; in dmirror_migrate_alloc_and_copy() local
733 src_page = pfn_to_page(spfn + i); in dmirror_migrate_alloc_and_copy()
736 copy_highpage(rpage, src_page); in dmirror_migrate_alloc_and_copy()
770 struct page *src_page = in dmirror_migrate_alloc_and_copy() local
774 copy_highpage(dst_page, src_page); in dmirror_migrate_alloc_and_copy()
998 struct page *src_page; in dmirror_devmem_fault_alloc_and_copy() local
1012 src_page = pfn_to_page(page_to_pfn(spage) + i); in dmirror_devmem_fault_alloc_and_copy()
1016 copy_highpage(dst_page, src_page); in dmirror_devmem_fault_alloc_and_copy()
/linux/arch/x86/kernel/cpu/sgx/
H A Dioctl.c214 struct page *src_page; in __sgx_encl_add_page() local
225 ret = get_user_pages(src, 1, 0, &src_page); in __sgx_encl_add_page()
232 pginfo.contents = (unsigned long)kmap_local_page(src_page); in __sgx_encl_add_page()
237 put_page(src_page); in __sgx_encl_add_page()
/linux/io_uring/
H A Dzcrx.c1241 static ssize_t io_copy_page(struct io_copy_cache *cc, struct page *src_page, in io_copy_page() argument
1255 folio_test_partial_kmap(page_folio(src_page))) { in io_copy_page()
1258 src_page += src_offset / PAGE_SIZE; in io_copy_page()
1265 src_addr = kmap_local_page(src_page) + src_offset; in io_copy_page()
1282 struct page *src_page, unsigned int src_offset, in io_zcrx_copy_chunk() argument
1303 n = io_copy_page(&cc, src_page, src_offset, len); in io_zcrx_copy_chunk()
/linux/mm/
H A Dkhugepaged.c698 struct page *src_page = pte_page(pteval); in __collapse_huge_page_copy_succeeded() local
700 src = page_folio(src_page); in __collapse_huge_page_copy_succeeded()
717 folio_remove_rmap_ptes(src, src_page, nr_ptes, vma); in __collapse_huge_page_copy_succeeded()
788 struct page *src_page; in __collapse_huge_page_copy() local
794 src_page = pte_page(pteval); in __collapse_huge_page_copy()
795 if (copy_mc_user_highpage(page, src_page, src_addr, vma) > 0) { in __collapse_huge_page_copy()
H A Dhuge_memory.c1851 struct page *src_page; in copy_huge_pmd() local
1918 src_page = pmd_page(pmd); in copy_huge_pmd()
1919 VM_BUG_ON_PAGE(!PageHead(src_page), src_page); in copy_huge_pmd()
1920 src_folio = page_folio(src_page); in copy_huge_pmd()
1923 if (unlikely(folio_try_dup_anon_rmap_pmd(src_folio, src_page, dst_vma, src_vma))) { in copy_huge_pmd()
2708 struct page *src_page; in move_pages_huge_pmd() local
2738 src_page = pmd_page(src_pmdval); in move_pages_huge_pmd()
2741 if (unlikely(!PageAnonExclusive(src_page))) { in move_pages_huge_pmd()
2746 src_folio = page_folio(src_page); in move_pages_huge_pmd()
H A Dmemory.c7362 struct page *src_page; in copy_user_gigantic_page() local
7367 src_page = folio_page(src, i); in copy_user_gigantic_page()
7370 if (copy_mc_user_highpage(dst_page, src_page, in copy_user_gigantic_page()
/linux/fs/ecryptfs/
H A Dcrypto.c300 struct page *src_page, in crypt_extent() argument
316 sg_set_page(&src_sg, src_page, extent_size, in crypt_extent()
/linux/arch/x86/kvm/svm/
H A Dsev.c2281 struct page *src_page, void *opaque) in sev_gmem_post_populate() argument
2290 if (WARN_ON_ONCE(sev_populate_args->type != KVM_SEV_SNP_PAGE_TYPE_ZERO && !src_page)) in sev_gmem_post_populate()
2301 if (src_page) { in sev_gmem_post_populate()
2302 void *src_vaddr = kmap_local_page(src_page); in sev_gmem_post_populate()
2338 void *src_vaddr = kmap_local_page(src_page); in sev_gmem_post_populate()
/linux/arch/x86/kvm/vmx/
H A Dtdx.c3120 struct page *src_page, void *_arg) in tdx_gmem_post_populate() argument
3131 if (!src_page) in tdx_gmem_post_populate()
3134 kvm_tdx->page_add_src = src_page; in tdx_gmem_post_populate()