Searched refs:pgd_page (Results 1 – 8 of 8) sorted by relevance
565 struct page *pgd_page; in kvm_riscv_mmu_alloc_pgd() local572 pgd_page = alloc_pages(GFP_KERNEL | __GFP_ZERO, in kvm_riscv_mmu_alloc_pgd()574 if (!pgd_page) in kvm_riscv_mmu_alloc_pgd()576 kvm->arch.pgd = page_to_virt(pgd_page); in kvm_riscv_mmu_alloc_pgd()577 kvm->arch.pgd_phys = page_to_phys(pgd_page); in kvm_riscv_mmu_alloc_pgd()
390 static inline struct page *pgd_page(pgd_t pgd) in pgd_page() function394 #define pgd_page(pgd) pgd_page(pgd) macro
15 int kernel_ident_mapping_init(struct x86_mapping_info *info, pgd_t *pgd_page,
43 #define pgd_page(pgd) (p4d_page((p4d_t){ pgd })) macro
35 return pgd_page(pgd) == virt_to_page(lm_alias(kasan_early_shadow_p4d)); in kasan_p4d_table()340 p4d_free(&init_mm, (p4d_t *)page_to_virt(pgd_page(*pgd))); in kasan_free_p4d()
38 if (pgd_page(val) == virt_to_page(lm_alias(kasan_early_shadow_p4d))) in ptdump_pgd_entry()
511 pgd_t *pgd_page = (pgd_t *)(((unsigned long)pgd) & PAGE_MASK); in xen_get_user_pgd() local512 unsigned offset = pgd - pgd_page; in xen_get_user_pgd()519 struct page *page = virt_to_page(pgd_page); in xen_get_user_pgd()
1176 #define pgd_page(pgd) pfn_to_page(__phys_to_pfn(__pgd_to_phys(pgd))) macro