Searched refs:kvm_mmu_page (Results 1 – 11 of 11) sorted by relevance
| /linux/arch/x86/kvm/mmu/ |
| H A D | tdp_mmu.c | 56 static void tdp_mmu_free_sp(struct kvm_mmu_page *sp) in tdp_mmu_free_sp() 73 struct kvm_mmu_page *sp = container_of(head, struct kvm_mmu_page, in tdp_mmu_free_sp_rcu_callback() 79 void kvm_tdp_mmu_put_root(struct kvm *kvm, struct kvm_mmu_page *root) in kvm_tdp_mmu_put_root() 97 static bool tdp_mmu_root_match(struct kvm_mmu_page *root, in tdp_mmu_root_match() 121 static struct kvm_mmu_page *tdp_mmu_next_root(struct kvm *kvm, in tdp_mmu_next_root() 122 struct kvm_mmu_page *prev_root, in tdp_mmu_next_root() 125 struct kvm_mmu_page *next_root; in tdp_mmu_next_root() 214 static struct kvm_mmu_page *tdp_mmu_alloc_sp(struct kvm_vcpu *vcpu) in tdp_mmu_alloc_sp() 216 struct kvm_mmu_page *sp; in tdp_mmu_alloc_sp() 224 static void tdp_mmu_init_sp(struct kvm_mmu_page *sp, tdp_ptep_t sptep, in tdp_mmu_init_sp() [all …]
|
| H A D | tdp_mmu.h | 15 __must_check static inline bool kvm_tdp_mmu_get_root(struct kvm_mmu_page *root) in kvm_tdp_mmu_get_root() 20 void kvm_tdp_mmu_put_root(struct kvm *kvm, struct kvm_mmu_page *root); 48 static inline struct kvm_mmu_page *tdp_mmu_get_root_for_fault(struct kvm_vcpu *vcpu, in tdp_mmu_get_root_for_fault() 57 static inline struct kvm_mmu_page *tdp_mmu_get_root(struct kvm_vcpu *vcpu, in tdp_mmu_get_root() 68 struct kvm_mmu_page *sp); 117 static inline bool is_tdp_mmu_page(struct kvm_mmu_page *sp) { return sp->tdp_mmu_page; } in is_tdp_mmu_page() 119 static inline bool is_tdp_mmu_page(struct kvm_mmu_page *sp) { return false; } in is_tdp_mmu_page()
|
| H A D | mmu_internal.h | 44 struct kvm_mmu_page { struct 150 static inline int kvm_mmu_page_as_id(struct kvm_mmu_page *sp) in kvm_mmu_page_as_id() argument 155 static inline bool is_mirror_sp(const struct kvm_mmu_page *sp) in is_mirror_sp() 160 static inline void kvm_mmu_alloc_external_spt(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp) in kvm_mmu_alloc_external_spt() 171 static inline gfn_t kvm_gfn_root_bits(const struct kvm *kvm, const struct kvm_mmu_page *root) in kvm_gfn_root_bits() 184 struct kvm_mmu_page *sp) in kvm_mmu_page_ad_need_write_protect() 409 void track_possible_nx_huge_page(struct kvm *kvm, struct kvm_mmu_page *sp, 411 void untrack_possible_nx_huge_page(struct kvm *kvm, struct kvm_mmu_page *sp,
|
| H A D | spte.h | 259 static inline struct kvm_mmu_page *to_shadow_page(hpa_t shadow_page) in to_shadow_page() 263 return (struct kvm_mmu_page *)page_private(page); in to_shadow_page() 266 static inline struct kvm_mmu_page *spte_to_child_sp(u64 spte) in spte_to_child_sp() 271 static inline struct kvm_mmu_page *sptep_to_sp(u64 *sptep) in sptep_to_sp() 276 static inline struct kvm_mmu_page *root_to_sp(hpa_t root) in root_to_sp() 295 struct kvm_mmu_page *root = root_to_sp(vcpu->arch.mmu->root.hpa); in kvm_vcpu_can_access_host_mmio() 321 static inline bool sp_ad_disabled(struct kvm_mmu_page *sp) in sp_ad_disabled() 543 bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
|
| H A D | mmu.c | 281 static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index); 286 struct kvm_mmu_page *sp = sptep_to_sp(sptep); in kvm_flush_remote_tlbs_sptep() 370 struct kvm_mmu_page *sp = sptep_to_sp(sptep); in count_spte_clear() 454 struct kvm_mmu_page *sp = sptep_to_sp(sptep); in __get_spte_lockless() 637 static bool sp_has_gptes(struct kvm_mmu_page *sp); 639 static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index) in kvm_mmu_page_get_gfn() 656 static u32 kvm_mmu_page_get_access(struct kvm_mmu_page *sp, int index) in kvm_mmu_page_get_access() 676 static void kvm_mmu_page_set_translation(struct kvm_mmu_page *sp, int index, in kvm_mmu_page_set_translation() 695 static void kvm_mmu_page_set_access(struct kvm_mmu_page *sp, int index, in kvm_mmu_page_set_access() 749 static void account_shadowed(struct kvm *kvm, struct kvm_mmu_page *s [all...] |
| H A D | mmutrace.h | 163 TP_PROTO(struct kvm_mmu_page *sp, bool created), 182 TP_PROTO(struct kvm_mmu_page *sp), 197 TP_PROTO(struct kvm_mmu_page *sp), 203 TP_PROTO(struct kvm_mmu_page *sp), 209 TP_PROTO(struct kvm_mmu_page *sp),
|
| H A D | paging_tmpl.h | 152 struct kvm_mmu_page *sp, u64 *spte, in FNAME() 532 FNAME(prefetch_gpte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in FNAME() 574 struct kvm_mmu_page *sp; in FNAME() 617 struct kvm_mmu_page *sp = NULL; in FNAME() 847 static gpa_t FNAME(get_level1_sp_gpa)(struct kvm_mmu_page *sp) in FNAME() 895 static int FNAME(sync_spte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, int i) in FNAME()
|
| H A D | spte.c | 143 struct kvm_mmu_page *root = root_to_sp(vcpu->arch.mmu->root.hpa); in kvm_track_host_mmio_mapping() 186 bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in make_spte()
|
| /linux/arch/x86/include/asm/ |
| H A D | kvm_host.h | 455 struct kvm_mmu_page; 473 struct kvm_mmu_page *sp, int i); 1377 * A list of kvm_mmu_page structs that, if zapped, could possibly be 1566 * - the link field of kvm_mmu_page structs used by the TDP MMU 1568 * - the possible_nx_huge_page_link field of kvm_mmu_page structs used
|
| /linux/Documentation/virt/kvm/x86/ |
| H A D | mmu.rst | 121 The principal data structure is the shadow page, 'struct kvm_mmu_page'. A 198 shadow page; it is also used to go back from a struct kvm_mmu_page
|
| /linux/arch/x86/kvm/vmx/ |
| H A D | vmx.c | 3391 struct kvm_mmu_page *root; in vmx_flush_tlb_guest()
|