Home
last modified time | relevance | path

Searched refs:pgd_page (Results 1 – 8 of 8) sorted by relevance

/linux/arch/riscv/kvm/
H A Dmmu.c565 struct page *pgd_page; in kvm_riscv_mmu_alloc_pgd() local
572 pgd_page = alloc_pages(GFP_KERNEL | __GFP_ZERO, in kvm_riscv_mmu_alloc_pgd()
574 if (!pgd_page) in kvm_riscv_mmu_alloc_pgd()
576 kvm->arch.pgd = page_to_virt(pgd_page); in kvm_riscv_mmu_alloc_pgd()
577 kvm->arch.pgd_phys = page_to_phys(pgd_page); in kvm_riscv_mmu_alloc_pgd()
/linux/arch/riscv/include/asm/
H A Dpgtable-64.h390 static inline struct page *pgd_page(pgd_t pgd) in pgd_page() function
394 #define pgd_page(pgd) pgd_page(pgd) macro
/linux/arch/x86/include/asm/
H A Dinit.h15 int kernel_ident_mapping_init(struct x86_mapping_info *info, pgd_t *pgd_page,
/linux/include/asm-generic/
H A Dpgtable-nop4d.h43 #define pgd_page(pgd) (p4d_page((p4d_t){ pgd })) macro
/linux/mm/kasan/
H A Dinit.c35 return pgd_page(pgd) == virt_to_page(lm_alias(kasan_early_shadow_p4d)); in kasan_p4d_table()
340 p4d_free(&init_mm, (p4d_t *)page_to_virt(pgd_page(*pgd))); in kasan_free_p4d()
/linux/mm/
H A Dptdump.c38 if (pgd_page(val) == virt_to_page(lm_alias(kasan_early_shadow_p4d))) in ptdump_pgd_entry()
/linux/arch/x86/xen/
H A Dmmu_pv.c511 pgd_t *pgd_page = (pgd_t *)(((unsigned long)pgd) & PAGE_MASK); in xen_get_user_pgd() local
512 unsigned offset = pgd - pgd_page; in xen_get_user_pgd()
519 struct page *page = virt_to_page(pgd_page); in xen_get_user_pgd()
/linux/arch/arm64/include/asm/
H A Dpgtable.h1176 #define pgd_page(pgd) pfn_to_page(__phys_to_pfn(__pgd_to_phys(pgd))) macro