/linux/arch/x86/kvm/mmu/ |
H A D | tdp_mmu.c | 198 gfn_t gfn, union kvm_mmu_page_role role) in tdp_mmu_init_sp() 290 static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, 346 gfn_t base_gfn = sp->gfn; in handle_removed_pt() 355 gfn_t gfn = base_gfn + i * KVM_PAGES_PER_HPAGE(level); in handle_removed_pt() 441 static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, in handle_changed_spte() 600 u64 old_spte, u64 new_spte, gfn_t gfn, int level) in tdp_mmu_set_spte() 692 static inline gfn_t tdp_mmu_max_gfn_exclusive(void) in tdp_mmu_max_gfn_exclusive() 708 gfn_t end = tdp_mmu_max_gfn_exclusive(); in __tdp_mmu_zap_root() 709 gfn_t start = 0; in __tdp_mmu_zap_root() 805 gfn_t start, gfn_t end, bool can_yield, bool flush) in tdp_mmu_zap_leafs() [all …]
|
H A D | page_track.h | 19 gfn_t gfn); 21 struct kvm_memory_slot *slot, gfn_t gfn); 24 const struct kvm_memory_slot *slot, gfn_t gfn);
|
H A D | page_track.c | 75 static void update_gfn_write_track(struct kvm_memory_slot *slot, gfn_t gfn, in update_gfn_write_track() 91 gfn_t gfn) in __kvm_write_track_add_gfn() 114 struct kvm_memory_slot *slot, gfn_t gfn) in __kvm_write_track_remove_gfn() 137 const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_gfn_is_write_tracked() 319 int kvm_write_track_add_gfn(struct kvm *kvm, gfn_t gfn) in kvm_write_track_add_gfn() 349 int kvm_write_track_remove_gfn(struct kvm *kvm, gfn_t gfn) in kvm_write_track_remove_gfn()
|
H A D | mmu.c | 280 static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index); 286 gfn_t gfn = kvm_mmu_page_get_gfn(sp, spte_index(sptep)); in kvm_flush_remote_tlbs_sptep() 300 static gfn_t get_mmio_spte_gfn(u64 spte) in get_mmio_spte_gfn() 631 static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index) in kvm_mmu_page_get_gfn() 669 gfn_t gfn, unsigned int access) in kvm_mmu_page_set_translation() 690 gfn_t gfn = kvm_mmu_page_get_gfn(sp, index); in kvm_mmu_page_set_access() 699 static struct kvm_lpage_info *lpage_info_slot(gfn_t gfn, in lpage_info_slot() 717 gfn_t gfn, int count) in update_gfn_disallow_lpage_count() 731 void kvm_mmu_gfn_disallow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_disallow_lpage() 736 void kvm_mmu_gfn_allow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_allow_lpage() [all …]
|
H A D | mmutrace.h | 213 TP_PROTO(u64 *sptep, gfn_t gfn, u64 spte), 218 __field(gfn_t, gfn) 236 TP_PROTO(u64 addr, gfn_t gfn, unsigned access), 241 __field(gfn_t, gfn) 336 TP_PROTO(int level, gfn_t gfn, u64 *sptep), 394 TP_PROTO(int as_id, gfn_t gfn, int level, u64 old_spte, u64 new_spte),
|
H A D | tdp_iter.c | 40 int min_level, gfn_t next_last_level_gfn) in tdp_iter_start()
|
H A D | spte.h | 521 unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn,
|
/linux/arch/x86/include/asm/ |
H A D | kvm_page_track.h | 43 void (*track_remove_region)(gfn_t gfn, unsigned long nr_pages, 52 int kvm_write_track_add_gfn(struct kvm *kvm, gfn_t gfn); 53 int kvm_write_track_remove_gfn(struct kvm *kvm, gfn_t gfn);
|
/linux/arch/x86/kvm/ |
H A D | kvm_onhyperv.h | 10 int hv_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, gfn_t nr_pages);
|
H A D | kvm_onhyperv.c | 95 int hv_flush_remote_tlbs_range(struct kvm *kvm, gfn_t start_gfn, gfn_t nr_pages) in hv_flush_remote_tlbs_range()
|
H A D | x86.c | 1028 gfn_t pdpt_gfn = cr3 >> PAGE_SHIFT; in load_pdptrs() 13270 static inline u32 kvm_async_pf_hash_fn(gfn_t gfn) in kvm_async_pf_hash_fn() 13282 static void kvm_add_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_add_async_pf_gfn() 13292 static u32 kvm_async_pf_gfn_slot(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_async_pf_gfn_slot() 13305 bool kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_find_async_pf_gfn() 13310 static void kvm_del_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_del_async_pf_gfn() 13606 int kvm_arch_gmem_prepare(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn, int max_order) in kvm_arch_gmem_prepare()
|
H A D | hyperv.c | 760 gfn_t msg_page_gfn; in synic_deliver_msg()
|
/linux/virt/kvm/ |
H A D | kvm_main.c | 316 void kvm_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, u64 nr_pages) in kvm_flush_remote_tlbs_range() 674 void kvm_mmu_invalidate_range_add(struct kvm *kvm, gfn_t start, gfn_t end) in kvm_mmu_invalidate_range_add() 1917 gfn_t start, gfn_t end) in kvm_check_memslot_overlap() 1944 gfn_t base_gfn; in __kvm_set_memory_region() 2200 gfn_t offset; in kvm_get_dirty_log_protect() 2269 gfn_t offset; in kvm_clear_dirty_log_protect() 2366 bool kvm_range_has_memory_attributes(struct kvm *kvm, gfn_t start, gfn_t end, in kvm_range_has_memory_attributes() 2461 static int kvm_vm_set_mem_attributes(struct kvm *kvm, gfn_t start, gfn_t end, in kvm_vm_set_mem_attributes() 2520 gfn_t start, end; in kvm_vm_ioctl_set_mem_attributes() 2546 struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn) in gfn_to_memslot() [all …]
|
H A D | guest_memfd.c | 33 gfn_t gfn = slot->base_gfn + index - slot->gmem.pgoff; in __kvm_gmem_prepare_folio() 57 gfn_t gfn, struct folio *folio) in kvm_gmem_prepare_folio() 305 static pgoff_t kvm_gmem_get_index(struct kvm_memory_slot *slot, gfn_t gfn) 480 BUILD_BUG_ON(sizeof(gfn_t) != sizeof(slot->gmem.pgoff)); in kvm_gmem_bind() 597 gfn_t gfn, kvm_pfn_t *pfn, struct page **page, in kvm_gmem_get_pfn() 632 long kvm_gmem_populate(struct kvm *kvm, gfn_t start_gfn, void __user *src, long npages, in kvm_gmem_populate() 659 gfn_t gfn = start_gfn + i; in kvm_gmem_populate()
|
H A D | pfncache.c | 299 gfn_t gfn = gpa_to_gfn(gpa); in __kvm_gpc_refresh()
|
/linux/drivers/gpu/drm/i915/gvt/ |
H A D | kvmgt.c | 92 gfn_t gfn; 100 gfn_t gfn; 111 static void kvmgt_page_track_remove_region(gfn_t gfn, unsigned long nr_pages, 228 static struct gvt_dma *__gvt_cache_find_gfn(struct intel_vgpu *vgpu, gfn_t gfn) in __gvt_cache_find_gfn() 246 static int __gvt_cache_add(struct intel_vgpu *vgpu, gfn_t gfn, in __gvt_cache_add() 349 __kvmgt_protect_table_find(struct intel_vgpu *info, gfn_t gfn) in __kvmgt_protect_table_find() 365 static bool kvmgt_gfn_is_write_protected(struct intel_vgpu *info, gfn_t gfn) in kvmgt_gfn_is_write_protected() 373 static void kvmgt_protect_table_add(struct intel_vgpu *info, gfn_t gfn) in kvmgt_protect_table_add() 388 static void kvmgt_protect_table_del(struct intel_vgpu *info, gfn_t gfn) in kvmgt_protect_table_del() 1604 static void kvmgt_page_track_remove_region(gfn_t gfn, unsigned long nr_pages, in kvmgt_page_track_remove_region()
|
/linux/include/linux/ |
H A D | kvm_types.h | 42 typedef u64 gfn_t; typedef
|
/linux/arch/mips/include/asm/ |
H A D | kvm_host.h | 808 bool kvm_mips_flush_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn); 809 int kvm_mips_mkclean_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn);
|
/linux/arch/riscv/kvm/ |
H A D | vcpu_exit.c | 19 gfn_t gfn; in gstage_page_fault()
|
H A D | vcpu_sbi_sta.c | 36 gfn_t gfn; in kvm_riscv_vcpu_record_steal_time()
|
/linux/arch/x86/kvm/vmx/ |
H A D | x86_ops.h | 106 u8 vmx_get_mt_mask(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio);
|
/linux/arch/x86/kvm/svm/ |
H A D | svm.h | 763 int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order); 787 static inline int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order) in sev_gmem_prepare()
|
/linux/arch/s390/kvm/ |
H A D | gaccess.c | 828 const gfn_t gfn = gpa_to_gfn(gpa); in access_guest_page() 846 gfn_t gfn; in access_guest_page_with_key() 1023 gfn_t gfn = gpa_to_gfn(gpa); in cmpxchg_guest_abs_with_key()
|
/linux/arch/powerpc/kvm/ |
H A D | booke.c | 1287 gfn_t gfn; in kvmppc_handle_exit() 1344 gfn_t gfn; in kvmppc_handle_exit()
|
/linux/arch/arm64/kvm/vgic/ |
H A D | vgic-its.c | 883 gfn_t gfn = gpa >> PAGE_SHIFT; in __is_visible_gfn_locked()
|