Lines Matching refs:gfn_t
316 void kvm_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, u64 nr_pages) in kvm_flush_remote_tlbs_range()
674 void kvm_mmu_invalidate_range_add(struct kvm *kvm, gfn_t start, gfn_t end) in kvm_mmu_invalidate_range_add()
1917 gfn_t start, gfn_t end) in kvm_check_memslot_overlap()
1944 gfn_t base_gfn; in __kvm_set_memory_region()
2200 gfn_t offset; in kvm_get_dirty_log_protect()
2269 gfn_t offset; in kvm_clear_dirty_log_protect()
2366 bool kvm_range_has_memory_attributes(struct kvm *kvm, gfn_t start, gfn_t end, in kvm_range_has_memory_attributes()
2461 static int kvm_vm_set_mem_attributes(struct kvm *kvm, gfn_t start, gfn_t end, in kvm_vm_set_mem_attributes()
2520 gfn_t start, end; in kvm_vm_ioctl_set_mem_attributes()
2546 struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn) in gfn_to_memslot()
2552 struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_vcpu_gfn_to_memslot()
2585 bool kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn) in kvm_is_visible_gfn()
2593 bool kvm_vcpu_is_visible_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_vcpu_is_visible_gfn()
2601 unsigned long kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_host_page_size()
2630 static unsigned long __gfn_to_hva_many(const struct kvm_memory_slot *slot, gfn_t gfn, in __gfn_to_hva_many()
2631 gfn_t *nr_pages, bool write) in __gfn_to_hva_many()
2645 static unsigned long gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn, in gfn_to_hva_many()
2646 gfn_t *nr_pages) in gfn_to_hva_many()
2652 gfn_t gfn) in gfn_to_hva_memslot()
2658 unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn) in gfn_to_hva()
2664 unsigned long kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_vcpu_gfn_to_hva()
2679 gfn_t gfn, bool *writable) in gfn_to_hva_memslot_prot()
2689 unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable) in gfn_to_hva_prot()
2696 unsigned long kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable) in kvm_vcpu_gfn_to_hva_prot()
2966 kvm_pfn_t __kvm_faultin_pfn(const struct kvm_memory_slot *slot, gfn_t gfn, in __kvm_faultin_pfn()
2988 int kvm_prefetch_pages(struct kvm_memory_slot *slot, gfn_t gfn, in kvm_prefetch_pages()
2992 gfn_t entry = 0; in kvm_prefetch_pages()
3012 struct page *__gfn_to_page(struct kvm *kvm, gfn_t gfn, bool write) in __gfn_to_page()
3027 int __kvm_vcpu_map(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map, in __kvm_vcpu_map()
3098 static int __kvm_read_guest_page(struct kvm_memory_slot *slot, gfn_t gfn, in __kvm_read_guest_page()
3116 int kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset, in kvm_read_guest_page()
3125 int kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data, in kvm_vcpu_read_guest_page()
3136 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_read_guest()
3156 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_vcpu_read_guest()
3174 static int __kvm_read_guest_atomic(struct kvm_memory_slot *slot, gfn_t gfn, in __kvm_read_guest_atomic()
3197 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_vcpu_read_guest_atomic()
3207 struct kvm_memory_slot *memslot, gfn_t gfn, in __kvm_write_guest_page()
3226 int kvm_write_guest_page(struct kvm *kvm, gfn_t gfn, in kvm_write_guest_page()
3235 int kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, in kvm_vcpu_write_guest_page()
3247 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_write_guest()
3268 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_vcpu_write_guest()
3291 gfn_t start_gfn = gpa >> PAGE_SHIFT; in __kvm_gfn_to_hva_cache_init()
3292 gfn_t end_gfn = (gpa + len - 1) >> PAGE_SHIFT; in __kvm_gfn_to_hva_cache_init()
3293 gfn_t nr_pages_needed = end_gfn - start_gfn + 1; in __kvm_gfn_to_hva_cache_init()
3294 gfn_t nr_pages_avail; in __kvm_gfn_to_hva_cache_init()
3413 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_clear_guest()
3432 gfn_t gfn) in mark_page_dirty_in_slot()
3455 void mark_page_dirty(struct kvm *kvm, gfn_t gfn) in mark_page_dirty()
3464 void kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_vcpu_mark_page_dirty()