Lines Matching refs:kvm_memory_slot
708 const struct kvm_memory_slot *slot, int level) in lpage_info_slot()
724 static void update_gfn_disallow_lpage_count(const struct kvm_memory_slot *slot, in update_gfn_disallow_lpage_count()
739 void kvm_mmu_gfn_disallow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_disallow_lpage()
744 void kvm_mmu_gfn_allow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_allow_lpage()
752 struct kvm_memory_slot *slot; in account_shadowed()
809 struct kvm_memory_slot *slot; in unaccount_shadowed()
838 static struct kvm_memory_slot *gfn_to_memslot_dirty_bitmap(struct kvm_vcpu *vcpu, in gfn_to_memslot_dirty_bitmap()
842 struct kvm_memory_slot *slot; in gfn_to_memslot_dirty_bitmap()
1172 const struct kvm_memory_slot *slot) in gfn_to_rmap()
1183 struct kvm_memory_slot *slot; in rmap_remove()
1356 const struct kvm_memory_slot *slot) in __rmap_clear_dirty()
1374 struct kvm_memory_slot *slot, in kvm_mmu_write_protect_pt_masked()
1397 struct kvm_memory_slot *slot, in kvm_mmu_clear_dirty_pt_masked()
1420 struct kvm_memory_slot *slot, in kvm_arch_mmu_enable_log_dirty_pt_masked()
1472 struct kvm_memory_slot *slot, u64 gfn, in kvm_mmu_slot_gfn_write_protect()
1495 struct kvm_memory_slot *slot; in kvm_vcpu_write_protect_gfn()
1502 const struct kvm_memory_slot *slot) in kvm_zap_rmap()
1509 const struct kvm_memory_slot *slot;
1534 const struct kvm_memory_slot *slot, in slot_rmap_walk_init()
1579 const struct kvm_memory_slot *slot);
1582 const struct kvm_memory_slot *slot, in __walk_slot_rmaps()
1615 const struct kvm_memory_slot *slot, in walk_slot_rmaps()
1626 const struct kvm_memory_slot *slot, in walk_slot_rmaps_4k()
1634 const struct kvm_memory_slot *slot, in __kvm_rmap_zap_gfn_range()
1677 const struct kvm_memory_slot *slot, in __rmap_add()
1699 static void rmap_add(struct kvm_vcpu *vcpu, const struct kvm_memory_slot *slot, in rmap_add()
2925 int mmu_try_to_unsync_pages(struct kvm *kvm, const struct kvm_memory_slot *slot, in mmu_try_to_unsync_pages()
3026 static int mmu_set_spte(struct kvm_vcpu *vcpu, struct kvm_memory_slot *slot, in mmu_set_spte()
3103 struct kvm_memory_slot *slot; in kvm_mmu_prefetch_sptes()
3225 const struct kvm_memory_slot *slot) in host_pfn_mapping_level()
3289 const struct kvm_memory_slot *slot, in __kvm_mmu_max_mapping_level()
3313 const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_max_mapping_level()
3323 struct kvm_memory_slot *slot = fault->slot; in kvm_mmu_hugepage_adjust()
3422 static void kvm_send_hwpoison_signal(struct kvm_memory_slot *slot, gfn_t gfn) in kvm_send_hwpoison_signal()
3937 struct kvm_memory_slot *slot; in mmu_first_shadow_root_alloc()
4621 struct kvm_memory_slot *slot = fault->slot; in kvm_mmu_faultin_pfn()
6784 const struct kvm_memory_slot *memslot; in kvm_rmap_zap_gfn_range()
6844 const struct kvm_memory_slot *slot) in slot_rmap_write_protect()
6850 const struct kvm_memory_slot *memslot, in kvm_mmu_slot_remove_write_access()
6948 const struct kvm_memory_slot *slot, in shadow_mmu_split_huge_page()
6994 const struct kvm_memory_slot *slot, in shadow_mmu_try_split_huge_page()
7034 const struct kvm_memory_slot *slot) in shadow_mmu_try_split_huge_pages()
7079 const struct kvm_memory_slot *slot, in kvm_shadow_mmu_try_split_huge_pages()
7098 const struct kvm_memory_slot *memslot, in kvm_mmu_try_split_huge_pages()
7117 const struct kvm_memory_slot *memslot, in kvm_mmu_slot_try_split_huge_pages()
7149 const struct kvm_memory_slot *slot) in kvm_mmu_zap_collapsible_spte()
7185 const struct kvm_memory_slot *slot) in kvm_rmap_zap_collapsible_sptes()
7197 const struct kvm_memory_slot *slot) in kvm_mmu_recover_huge_pages()
7213 const struct kvm_memory_slot *memslot) in kvm_mmu_slot_leaf_clear_dirty()
7272 struct kvm_memory_slot *slot, in kvm_mmu_zap_memslot_pages_and_flush()
7306 struct kvm_memory_slot *slot) in kvm_mmu_zap_memslot()
7330 struct kvm_memory_slot *slot) in kvm_arch_flush_shadow_memslot()
7588 struct kvm_memory_slot *slot; in kvm_recover_nx_huge_pages()
7745 static bool hugepage_test_mixed(struct kvm_memory_slot *slot, gfn_t gfn, in hugepage_test_mixed()
7751 static void hugepage_clear_mixed(struct kvm_memory_slot *slot, gfn_t gfn, in hugepage_clear_mixed()
7757 static void hugepage_set_mixed(struct kvm_memory_slot *slot, gfn_t gfn, in hugepage_set_mixed()
7766 struct kvm_memory_slot *slot = range->slot; in kvm_arch_pre_set_memory_attributes()
7826 static bool hugepage_has_attrs(struct kvm *kvm, struct kvm_memory_slot *slot, in hugepage_has_attrs()
7847 struct kvm_memory_slot *slot = range->slot; in kvm_arch_post_set_memory_attributes()
7911 struct kvm_memory_slot *slot) in kvm_mmu_init_memslot_memory_attributes()