Lines Matching defs:gfn
286 gfn_t gfn = kvm_mmu_page_get_gfn(sp, spte_index(sptep)); in kvm_flush_remote_tlbs_sptep() local
291 static void mark_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, u64 gfn, in mark_mmio_spte()
676 gfn_t gfn, unsigned int access) in kvm_mmu_page_set_translation()
697 gfn_t gfn = kvm_mmu_page_get_gfn(sp, index); in kvm_mmu_page_set_access() local
706 static struct kvm_lpage_info *lpage_info_slot(gfn_t gfn, in lpage_info_slot()
724 gfn_t gfn, int count) in update_gfn_disallow_lpage_count()
738 void kvm_mmu_gfn_disallow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_disallow_lpage()
743 void kvm_mmu_gfn_allow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_allow_lpage()
752 gfn_t gfn; in account_shadowed() local
809 gfn_t gfn; in unaccount_shadowed() local
838 gfn_t gfn, in gfn_to_memslot_dirty_bitmap()
1017 static struct kvm_rmap_head *gfn_to_rmap(gfn_t gfn, int level, in gfn_to_rmap()
1031 gfn_t gfn; in rmap_remove() local
1319 struct kvm_memory_slot *slot, u64 gfn, in kvm_mmu_slot_gfn_write_protect()
1340 static bool kvm_vcpu_write_protect_gfn(struct kvm_vcpu *vcpu, u64 gfn) in kvm_vcpu_write_protect_gfn()
1363 gfn_t gfn; member
1525 u64 *spte, gfn_t gfn, unsigned int access) in __rmap_add()
1547 u64 *spte, gfn_t gfn, unsigned int access) in rmap_add()
1654 static unsigned kvm_page_table_hashfn(gfn_t gfn) in kvm_page_table_hashfn()
2059 gfn_t gfn, in kvm_mmu_find_shadow_page()
2144 gfn_t gfn, in kvm_mmu_alloc_shadow_page()
2181 gfn_t gfn, in __kvm_mmu_get_shadow_page()
2201 gfn_t gfn, in kvm_mmu_get_shadow_page()
2260 u64 *sptep, gfn_t gfn, in kvm_mmu_get_child_sp()
2725 gfn_t gfn, bool synchronizing, bool prefetch) in mmu_try_to_unsync_pages()
2826 u64 *sptep, unsigned int pte_access, gfn_t gfn, in mmu_set_spte()
2897 static bool kvm_mmu_prefetch_sptes(struct kvm_vcpu *vcpu, gfn_t gfn, u64 *sptep, in kvm_mmu_prefetch_sptes()
2938 gfn_t gfn = kvm_mmu_page_get_gfn(sp, spte_index(start)); in direct_pte_prefetch_many() local
3022 static int host_pfn_mapping_level(struct kvm *kvm, gfn_t gfn, in host_pfn_mapping_level()
3088 gfn_t gfn, int max_level, bool is_private) in __kvm_mmu_max_mapping_level()
3111 const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_max_mapping_level()
3220 static void kvm_send_hwpoison_signal(struct kvm_memory_slot *slot, gfn_t gfn) in kvm_send_hwpoison_signal()
3640 static hpa_t mmu_alloc_root(struct kvm_vcpu *vcpu, gfn_t gfn, int quadrant, in mmu_alloc_root()
4170 gfn_t gfn = get_mmio_spte_gfn(spte); in handle_mmio_page_fault() local
4917 static bool sync_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, in sync_mmio_spte()
5907 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_mmu_track_write() local
6681 gfn_t gfn; in shadow_mmu_get_sp_for_split() local
6712 gfn_t gfn; in shadow_mmu_split_huge_page() local
6754 gfn_t gfn; in shadow_mmu_try_split_huge_page() local
7044 gfn_t gfn = slot->base_gfn + i; in kvm_mmu_zap_memslot_pages_and_flush() local
7525 static bool hugepage_test_mixed(struct kvm_memory_slot *slot, gfn_t gfn, in hugepage_test_mixed()
7531 static void hugepage_clear_mixed(struct kvm_memory_slot *slot, gfn_t gfn, in hugepage_clear_mixed()
7537 static void hugepage_set_mixed(struct kvm_memory_slot *slot, gfn_t gfn, in hugepage_set_mixed()
7544 gfn_t gfn, int level, unsigned long attrs) in hugepage_has_attrs()
7585 gfn_t gfn = gfn_round_for_level(range->start, level); in kvm_arch_post_set_memory_attributes() local
7644 gfn_t gfn; in kvm_mmu_init_memslot_memory_attributes() local