| /linux/arch/loongarch/include/asm/ |
| H A D | page.h | 77 #define virt_to_page(kaddr) phys_to_page(__pa(kaddr)) macro 89 #define virt_to_page(kaddr) \ macro 98 #define virt_to_pfn(kaddr) page_to_pfn(virt_to_page(kaddr))
|
| H A D | kfence.h | 43 set_page_address(virt_to_page(kaddr), vaddr); in arch_kfence_init_pool()
|
| /linux/Documentation/arch/powerpc/ |
| H A D | vmemmap_dedup.rst | 23 +-----------+ ---virt_to_page---> +-----------+ mapping to +-----------+ 52 +-----------+ ---virt_to_page---> +-----------+ mapping to +-----------+ 81 +-----------+ ---virt_to_page---> +-----------+ mapping to +-----------+
|
| /linux/arch/arc/include/asm/ |
| H A D | pgtable-levels.h | 112 #define p4d_page(p4d) virt_to_page(p4d_pgtable(p4d)) 134 #define pud_page(pud) virt_to_page(pud_pgtable(pud)) 161 #define pmd_page(pmd) virt_to_page((void *)pmd_page_vaddr(pmd))
|
| /linux/arch/powerpc/mm/kasan/ |
| H A D | init_book3e_64.c | 19 return p4d_page(p4d) == virt_to_page(lm_alias(kasan_early_shadow_pud)); in kasan_pud_table() 24 return pud_page(pud) == virt_to_page(lm_alias(kasan_early_shadow_pmd)); in kasan_pmd_table() 29 return pmd_page(pmd) == virt_to_page(lm_alias(kasan_early_shadow_pte)); in kasan_pte_table()
|
| /linux/arch/xtensa/mm/ |
| H A D | kasan_init.c | 26 mk_pte(virt_to_page(kasan_early_shadow_page), in kasan_early_init() 89 mk_pte(virt_to_page(kasan_early_shadow_page), in kasan_init()
|
| /linux/fs/ubifs/ |
| H A D | crypto.c | 53 err = fscrypt_encrypt_block_inplace(inode, virt_to_page(p), pad_len, in ubifs_encrypt() 78 err = fscrypt_decrypt_block_inplace(inode, virt_to_page(&dn->data), in ubifs_decrypt()
|
| /linux/kernel/events/ |
| H A D | ring_buffer.c | 646 struct page *page = virt_to_page(rb->aux_pages[idx]); in rb_free_aux_page() 761 struct page *page = virt_to_page(rb->aux_pages[0]); in rb_alloc_aux() 813 return virt_to_page(rb->user_page); in __perf_mmap_to_page() 815 return virt_to_page(rb->data_pages[pgoff - 1]); in __perf_mmap_to_page() 833 struct page *page = virt_to_page(addr); in perf_mmap_free_page() 972 return virt_to_page(rb->aux_pages[aux_pgoff]); in perf_mmap_to_page()
|
| /linux/arch/sparc/kernel/ |
| H A D | leon_smp.c | 248 free_reserved_page(virt_to_page(&trapbase_cpu1[0])); in leon_smp_done() 251 free_reserved_page(virt_to_page(&trapbase_cpu2[0])); in leon_smp_done() 254 free_reserved_page(virt_to_page(&trapbase_cpu3[0])); in leon_smp_done()
|
| /linux/mm/kmsan/ |
| H A D | shadow.c | 76 return virt_to_page(vaddr); in virt_to_page_or_null() 289 shadow_p = virt_to_page((char *)shadow + addr); in kmsan_init_alloc_meta_for_range() 292 origin_p = virt_to_page((char *)origin + addr); in kmsan_init_alloc_meta_for_range()
|
| /linux/arch/nios2/mm/ |
| H A D | ioremap.c | 139 for (page = virt_to_page(t_addr); in ioremap() 140 page <= virt_to_page(t_end); page++) in ioremap()
|
| /linux/arch/arm/include/asm/vdso/ |
| H A D | vsyscall.h | 13 flush_dcache_page(virt_to_page(vdata)); in __arch_sync_vdso_time_data()
|
| /linux/arch/m68k/include/asm/ |
| H A D | sun3_pgtable.h | 107 #define pte_page(pte) virt_to_page(__pte_page(pte)) 109 #define pmd_page(pmd) virt_to_page((void *)pmd_page_vaddr(pmd))
|
| /linux/drivers/char/agp/ |
| H A D | uninorth-agp.c | 413 for (page = virt_to_page(table), i = 0; page <= virt_to_page(table_end); in uninorth_create_gatt_table() 467 for (page = virt_to_page(table); page <= virt_to_page(table_end); page++) in uninorth_free_gatt_table()
|
| H A D | generic.c | 930 for (page = virt_to_page(table); page <= virt_to_page(table_end); page++) in agp_generic_create_gatt_table() 949 for (page = virt_to_page(table); page <= virt_to_page(table_end); page++) in agp_generic_create_gatt_table() 1010 for (page = virt_to_page(table); page <= virt_to_page(table_end); page++) in agp_generic_free_gatt_table()
|
| /linux/arch/x86/xen/ |
| H A D | mmu_pv.c | 229 struct page *page = virt_to_page(ptr); in xen_page_pinned() 519 struct page *page = virt_to_page(pgd_page); in xen_get_user_pgd() 631 (*func)(mm, virt_to_page(pmd), PT_PMD); in xen_pud_walk() 649 (*func)(mm, virt_to_page(pud), PT_PUD); in xen_p4d_walk() 700 (*func)(mm, virt_to_page(pgd), PT_PGD); in __xen_pgd_walk() 806 xen_pin_page(mm, virt_to_page(user_pgd), PT_PGD); in __xen_pgd_pin() 863 SetPagePinned(virt_to_page(level3_user_vsyscall)); in xen_after_bootmem() 923 xen_unpin_page(mm, virt_to_page(user_pgd), PT_PGD); in __xen_pgd_unpin() 1108 ClearPagePinned(virt_to_page(__va(pa))); in xen_cleanmfnmap_free_pgtbl() 1461 struct page *page = virt_to_page(pgd); in xen_pgd_alloc() [all …]
|
| /linux/lib/ |
| H A D | test_free_pages.c | 20 struct page *page = virt_to_page((void *)addr); in test_free_pages()
|
| /linux/drivers/net/vmxnet3/ |
| H A D | vmxnet3_xdp.c | 153 page = virt_to_page(xdpf->data); in vmxnet3_xdp_xmit_frame() 271 page = virt_to_page(xdp->data_hard_start); in vmxnet3_run_xdp() 423 rbi->page = virt_to_page(new_data); in vmxnet3_process_xdp()
|
| /linux/arch/sparc/mm/ |
| H A D | iommu.c | 366 set_pte(ptep, mk_pte(virt_to_page(page), dvma_prot)); in sbus_iommu_alloc() 369 MKIOPTE(page_to_pfn(virt_to_page(page)), ioperm_noc); in sbus_iommu_alloc() 402 struct page *page = virt_to_page(cpu_addr); in sbus_iommu_free()
|
| /linux/net/sunrpc/xprtrdma/ |
| H A D | svc_rdma_backchannel.c | 91 get_page(virt_to_page(rqst->rq_buffer)); in svc_rdma_bc_sendto() 131 put_page(virt_to_page(rqst->rq_buffer)); in xprt_rdma_bc_free()
|
| /linux/arch/arm64/kernel/ |
| H A D | vdso.c | 196 aarch32_vectors_page = virt_to_page((void *)vdso_page); in aarch32_alloc_kuser_vdso_page() 214 aarch32_sig_page = virt_to_page(sigpage); in aarch32_alloc_sigpage()
|
| /linux/arch/csky/kernel/ |
| H A D | vdso.c | 32 pg = virt_to_page(vdso_start + (i << PAGE_SHIFT)); in vdso_init()
|
| /linux/kernel/power/ |
| H A D | snapshot.c | 197 while (res && swsusp_page_is_free(virt_to_page(res))) { in get_image_page() 199 swsusp_set_page_forbidden(virt_to_page(res)); in get_image_page() 204 swsusp_set_page_forbidden(virt_to_page(res)); in get_image_page() 205 swsusp_set_page_free(virt_to_page(res)); in get_image_page() 261 page = virt_to_page(addr); in free_image_page() 2542 pbe->copy_page = virt_to_page(kaddr); in get_highmem_page_buffer() 2692 if (!swsusp_page_is_free(virt_to_page(lp))) { in prepare_image() 2698 swsusp_set_page_forbidden(virt_to_page(lp)); in prepare_image() 2699 swsusp_set_page_free(virt_to_page(lp)); in prepare_image()
|
| /linux/arch/sh/kernel/vsyscall/ |
| H A D | vsyscall.c | 60 syscall_pages[0] = virt_to_page(syscall_page); in vsyscall_init()
|
| /linux/arch/csky/include/asm/ |
| H A D | page.h | 80 #define virt_to_page(x) (mem_map + MAP_NR(x)) macro
|