| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | private_mem_conversions_test.c | 30 #define memcmp_g(gpa, pattern, size) \ argument 32 u8 *mem = (u8 *)gpa; \ 37 "Guest expected 0x%x at offset %lu (gpa 0x%lx), got 0x%x", \ 38 pattern, i, gpa + i, mem[i]); \ 41 static void memcmp_h(u8 *mem, gpa_t gpa, u8 pattern, size_t size) in memcmp_h() argument 47 "Host expected 0x%x at gpa 0x%lx, got 0x%x", in memcmp_h() 48 pattern, gpa + i, mem[i]); in memcmp_h() 53 * Execute KVM hypercall to map/unmap gpa range which will cause userspace exit 54 * to back/unback private memory. Subsequent accesses by guest to the gpa range 73 static void guest_sync_shared(gpa_t gpa, u6 argument 79 guest_sync_private(uint64_t gpa,uint64_t size,uint8_t pattern) guest_sync_private() argument 89 guest_map_mem(uint64_t gpa,uint64_t size,bool map_shared,bool do_fallocate) guest_map_mem() argument 101 guest_map_shared(uint64_t gpa,uint64_t size,bool do_fallocate) guest_map_shared() argument 106 guest_map_private(uint64_t gpa,uint64_t size,bool do_fallocate) guest_map_private() argument 137 uint64_t gpa = base_gpa + test_ranges[i].offset; guest_test_explicit_conversion() local 217 guest_punch_hole(uint64_t gpa,uint64_t size) guest_punch_hole() argument 242 uint64_t gpa = base_gpa + test_ranges[i].offset; guest_test_punch_hole() local 292 uint64_t gpa = run->hypercall.args[0]; handle_exit_hypercall() local 340 uint64_t gpa = uc.args[1]; __test_mem_conversions() local 405 uint64_t gpa = BASE_DATA_GPA + i * per_cpu_size; test_mem_conversions() local [all...] |
| /linux/tools/testing/selftests/kvm/ |
| H A D | mmu_stress_test.c | 25 gpa_t gpa; in guest_code() local 29 for (gpa = start_gpa; gpa < end_gpa; gpa += stride) in guest_code() 30 vcpu_arch_put_guest(*((volatile u64 *)gpa), gpa); in guest_code() 34 for (gpa = start_gpa; gpa < end_gpa; gpa += stride) in guest_code() 35 *((volatile u64 *)gpa); in guest_code() 209 uint64_t gpa, nr_bytes; spawn_workers() local 285 uint64_t max_gpa, gpa, slot_size, max_mem, i; main() local [all...] |
| H A D | pre_fault_memory_test.c | 36 gpa_t gpa; member 58 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS, data->gpa, in delete_slot_worker() 68 .gpa = base_gpa + offset, in pre_fault_memory() 74 .gpa = base_gpa, in pre_fault_memory() 164 gpa_t gpa, gva, alignment, guest_page_size; in __test_pre_fault_memory() local 177 gpa = (vm->max_gfn - TEST_NPAGES) * guest_page_size; in __test_pre_fault_memory() 179 gpa = align_down(gpa, alignment); in __test_pre_fault_memory() 180 gva = gpa & ((1ULL << (vm->va_bits - 1)) - 1); in __test_pre_fault_memory() 182 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS, gpa, TEST_SLO in __test_pre_fault_memory() [all...] |
| H A D | memslot_perf_test.c | 189 static void *vm_gpa2hva(struct vm_data *data, gpa_t gpa, u64 *rempages) in vm_gpa2hva() 196 TEST_ASSERT(gpa >= MEM_GPA, "Too low gpa to translate"); in vm_gpa2hva() 197 TEST_ASSERT(gpa < MEM_GPA + data->npages * guest_page_size, in vm_gpa2hva() 198 "Too high gpa to translate"); in vm_gpa2hva() 199 gpa -= MEM_GPA; in vm_gpa2hva() 201 gpage = gpa / guest_page_size; in vm_gpa2hva() 202 pgoffs = gpa % guest_page_size; in vm_gpa2hva() 215 "Asking for remaining pages in slot but gpa not page aligned"); in vm_gpa2hva() 335 gpa_t gpa; in prepare_vm() 188 vm_gpa2hva(struct vm_data * data,uint64_t gpa,uint64_t * rempages) vm_gpa2hva() argument 334 uint64_t gpa; prepare_vm() local 640 uint64_t gpa, ctr; test_memslot_do_unmap() local 665 uint64_t gpa; test_memslot_map_unmap_check() local [all...] |
| H A D | memslot_modification_stress_test.c | 61 gpa_t gpa; in add_remove_memslot() 68 gpa = memstress_args.gpa - pages * vm->page_size; in add_remove_memslot() 72 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS, gpa, in add_remove_memslot() 60 uint64_t gpa; add_remove_memslot() local
|
| /linux/arch/powerpc/kvm/ |
| H A D | book3s_64_mmu_radix.c | 149 u64 pte, base, gpa; in kvmppc_mmu_walk_radix_tree() local 204 gpa = pte & 0x01fffffffffff000ul; in kvmppc_mmu_walk_radix_tree() 205 if (gpa & ((1ul << offset) - 1)) in kvmppc_mmu_walk_radix_tree() 207 gpa |= eaddr & ((1ul << offset) - 1); in kvmppc_mmu_walk_radix_tree() 215 gpte->raddr = gpa; in kvmppc_mmu_walk_radix_tree() 423 void kvmppc_unmap_pte(struct kvm *kvm, pte_t *pte, unsigned long gpa, in kvmppc_unmap_pte() argument 430 unsigned long gfn = gpa >> PAGE_SHIFT; in kvmppc_unmap_pte() 434 old = kvmppc_radix_update_pte(kvm, pte, ~0UL, 0, gpa, shift); in kvmppc_unmap_pte() 435 kvmppc_radix_tlbie_page(kvm, gpa, shift, lpid); in kvmppc_unmap_pte() 454 gpa &= ~(page_size - 1); in kvmppc_unmap_pte() [all …]
|
| H A D | book3s_hv_uvmem.c | 234 unsigned long gpa; member 516 struct kvm *kvm, unsigned long gpa, struct page *fault_page) in __kvmppc_svm_page_out() argument 536 if (!kvmppc_gfn_is_uvmem_pfn(gpa >> page_shift, kvm, NULL)) in __kvmppc_svm_page_out() 569 gpa, 0, page_shift); in __kvmppc_svm_page_out() 589 struct kvm *kvm, unsigned long gpa, in kvmppc_svm_page_out() argument 595 ret = __kvmppc_svm_page_out(vma, start, end, page_shift, kvm, gpa, in kvmppc_svm_page_out() 645 PAGE_SHIFT, kvm, pvt->gpa, NULL)) in kvmppc_uvmem_drop_pages() 647 pvt->gpa, addr); in kvmppc_uvmem_drop_pages() 695 static struct page *kvmppc_uvmem_get_page(unsigned long gpa, struct kvm *kvm) in kvmppc_uvmem_get_page() argument 719 kvmppc_gfn_secure_uvmem_pfn(gpa >> PAGE_SHIFT, uvmem_pfn, kvm); in kvmppc_uvmem_get_page() [all …]
|
| /linux/virt/kvm/ |
| H A D | pfncache.c | 60 static bool kvm_gpc_is_valid_len(gpa_t gpa, unsigned long uhva, in kvm_gpc_is_valid_len() argument 63 unsigned long offset = kvm_is_error_gpa(gpa) ? offset_in_page(uhva) : in kvm_gpc_is_valid_len() 64 offset_in_page(gpa); in kvm_gpc_is_valid_len() 84 if (!kvm_is_error_gpa(gpc->gpa) && gpc->generation != slots->generation) in kvm_gpc_check() 90 if (!kvm_gpc_is_valid_len(gpc->gpa, gpc->uhva, len)) in kvm_gpc_check() 166 .gfn = gpa_to_gfn(gpc->gpa), in hva_to_pfn_retry() 177 * Invalidate the cache prior to dropping gpc->lock, the gpa=>uhva in hva_to_pfn_retry() 179 * different task may not fail the gpa/uhva/generation checks. in hva_to_pfn_retry() 256 static int __kvm_gpc_refresh(struct gfn_to_pfn_cache *gpc, gpa_t gpa, unsigned long uhva) in __kvm_gpc_refresh() 266 /* Either gpa o in __kvm_gpc_refresh() 248 __kvm_gpc_refresh(struct gfn_to_pfn_cache * gpc,gpa_t gpa,unsigned long uhva) __kvm_gpc_refresh() argument 389 __kvm_gpc_activate(struct gfn_to_pfn_cache * gpc,gpa_t gpa,unsigned long uhva,unsigned long len) __kvm_gpc_activate() argument 419 kvm_gpc_activate(struct gfn_to_pfn_cache * gpc,gpa_t gpa,unsigned long len) kvm_gpc_activate() argument [all...] |
| /linux/tools/testing/selftests/kvm/lib/ |
| H A D | memstress.c | 108 vcpu_args->gpa = args->gpa + (i * vcpu_memory_bytes); in memstress_setup_vcpus() 113 vcpu_args->gpa = args->gpa; in memstress_setup_vcpus() 118 pr_debug("Added VCPU %d with test mem gpa [%lx, %lx)\n", in memstress_setup_vcpus() 119 i, vcpu_args->gpa, vcpu_args->gpa + in memstress_setup_vcpus() 197 args->gpa = (region_end_gfn - guest_num_pages - 1) * args->guest_page_size; in memstress_create_vm() 198 args->gpa = align_down(args->gpa, backing_src_pages in memstress_create_vm() [all...] |
| H A D | kvm_util.c | 922 gpa_t gpa, u64 size, void *hva) in __vm_set_user_memory_region2() 927 .guest_phys_addr = gpa, in __vm_set_user_memory_region2() 936 gpa_t gpa, u64 size, void *hva) in vm_set_user_memory_region2() 938 int ret = __vm_set_user_memory_region(vm, slot, flags, gpa, size, hva); in vm_set_user_memory_region2() 949 gpa_t gpa, u64 size, void *hva, in vm_mem_add() 955 .guest_phys_addr = gpa, in vm_mem_add() 968 gpa_t gpa, u64 size, void *hva, in vm_mem_add() 971 int ret = __vm_set_user_memory_region2(vm, slot, flags, gpa, size, hva, in vm_mem_add() 981 gpa_t gpa, u32 slot, u64 npages, u32 flags, in vm_mem_add() 996 TEST_ASSERT((gpa in vm_mem_add() 889 __vm_set_user_memory_region(struct kvm_vm * vm,uint32_t slot,uint32_t flags,uint64_t gpa,uint64_t size,void * hva) __vm_set_user_memory_region() argument 903 vm_set_user_memory_region(struct kvm_vm * vm,uint32_t slot,uint32_t flags,uint64_t gpa,uint64_t size,void * hva) vm_set_user_memory_region() argument 916 __vm_set_user_memory_region2(struct kvm_vm * vm,uint32_t slot,uint32_t flags,uint64_t gpa,uint64_t size,void * hva,uint32_t guest_memfd,uint64_t guest_memfd_offset) __vm_set_user_memory_region2() argument 935 vm_set_user_memory_region2(struct kvm_vm * vm,uint32_t slot,uint32_t flags,uint64_t gpa,uint64_t size,void * hva,uint32_t guest_memfd,uint64_t guest_memfd_offset) vm_set_user_memory_region2() argument 948 vm_mem_add(struct kvm_vm * vm,enum vm_mem_backing_src_type src_type,uint64_t gpa,uint32_t slot,uint64_t npages,uint32_t flags,int guest_memfd,uint64_t guest_memfd_offset) vm_mem_add() argument 1118 vm_userspace_mem_region_add(struct kvm_vm * vm,enum vm_mem_backing_src_type src_type,uint64_t gpa,uint32_t slot,uint64_t npages,uint32_t flags) vm_userspace_mem_region_add() argument 1256 uint64_t gpa, len; vm_guest_mem_fallocate() local 1588 addr_gpa2hva(struct kvm_vm * vm,vm_paddr_t gpa) addr_gpa2hva() argument 1664 addr_gpa2alias(struct kvm_vm * vm,vm_paddr_t gpa) addr_gpa2alias() argument [all...] |
| /linux/arch/x86/include/asm/uv/ |
| H A D | uv_hub.h | 461 uv_gpa_in_mmr_space(unsigned long gpa) in uv_gpa_in_mmr_space() argument 463 return (gpa >> 62) == 0x3UL; in uv_gpa_in_mmr_space() 467 static inline unsigned long uv_gpa_to_soc_phys_ram(unsigned long gpa) in uv_gpa_to_soc_phys_ram() argument 475 gpa = ((gpa << uv_hub_info->m_shift) >> uv_hub_info->m_shift) | in uv_gpa_to_soc_phys_ram() 476 ((gpa >> uv_hub_info->n_lshift) << uv_hub_info->m_val); in uv_gpa_to_soc_phys_ram() 478 paddr = gpa & uv_hub_info->gpa_mask; in uv_gpa_to_soc_phys_ram() 485 static inline unsigned long uv_gpa_to_gnode(unsigned long gpa) in uv_gpa_to_gnode() argument 490 return gpa >> n_lshift; in uv_gpa_to_gnode() 492 return uv_gam_range(gpa)->nasid >> 1; in uv_gpa_to_gnode() 496 static inline int uv_gpa_to_pnode(unsigned long gpa) in uv_gpa_to_pnode() argument [all …]
|
| /linux/arch/s390/kvm/ |
| H A D | vsie.c | 687 static int pin_guest_page(struct kvm *kvm, gpa_t gpa, hpa_t *hpa) in pin_guest_page() argument 691 page = gfn_to_page(kvm, gpa_to_gfn(gpa)); in pin_guest_page() 694 *hpa = (hpa_t)page_to_phys(page) + (gpa & ~PAGE_MASK); in pin_guest_page() 699 static void unpin_guest_page(struct kvm *kvm, gpa_t gpa, hpa_t hpa) in unpin_guest_page() argument 703 mark_page_dirty(kvm, gpa_to_gfn(gpa)); in unpin_guest_page() 768 gpa_t gpa; in pin_blocks() local 771 gpa = READ_ONCE(scb_o->scaol) & ~0xfUL; in pin_blocks() 773 gpa |= (u64) READ_ONCE(scb_o->scaoh) << 32; in pin_blocks() 774 if (gpa) { in pin_blocks() 775 if (gpa < 2 * PAGE_SIZE) in pin_blocks() [all …]
|
| H A D | gaccess.c | 444 static int deref_table(struct kvm *kvm, unsigned long gpa, unsigned long *val) in deref_table() argument 446 return kvm_read_guest(kvm, gpa, val, sizeof(*val)); in deref_table() 471 unsigned long *gpa, const union asce asce, in guest_translate_gva() argument 630 *gpa = raddr.addr; in guest_translate_gva() 654 enum gacc_mode mode, gpa_t gpa) in vm_check_access_key_gpa() argument 660 r = dat_get_storage_key(kvm->arch.gmap->asce, gpa_to_gfn(gpa), &storage_key); in vm_check_access_key_gpa() 705 enum gacc_mode mode, union asce asce, gpa_t gpa, in vcpu_check_access_key_gpa() argument 719 r = dat_get_storage_key(vcpu->arch.gmap->asce, gpa_to_gfn(gpa), &storage_key); in vcpu_check_access_key_gpa() 784 unsigned long gpa; in guest_range_to_gpas() local 794 rc = guest_translate_gva(vcpu, ga, &gpa, asce, mode, &prot); in guest_range_to_gpas() [all …]
|
| /linux/arch/x86/kvm/mmu/ |
| H A D | page_track.h | 30 void __kvm_page_track_write(struct kvm *kvm, gpa_t gpa, const u8 *new, int bytes); 41 static inline void __kvm_page_track_write(struct kvm *kvm, gpa_t gpa, in __kvm_page_track_write() argument 50 static inline void kvm_page_track_write(struct kvm_vcpu *vcpu, gpa_t gpa, in kvm_page_track_write() argument 53 __kvm_page_track_write(vcpu->kvm, gpa, new, bytes); in kvm_page_track_write() 55 kvm_mmu_track_write(vcpu, gpa, new, bytes); in kvm_page_track_write()
|
| /linux/arch/riscv/kvm/ |
| H A D | mmu.c | 35 int kvm_riscv_mmu_ioremap(struct kvm *kvm, gpa_t gpa, phys_addr_t hpa, in kvm_riscv_mmu_ioremap() argument 51 end = (gpa + size + PAGE_SIZE - 1) & PAGE_MASK; in kvm_riscv_mmu_ioremap() 55 for (addr = gpa; addr < end; addr += PAGE_SIZE) { in kvm_riscv_mmu_ioremap() 82 void kvm_riscv_mmu_iounmap(struct kvm *kvm, gpa_t gpa, unsigned long size) in kvm_riscv_mmu_iounmap() argument 89 kvm_riscv_gstage_unmap_range(&gstage, gpa, size, false); in kvm_riscv_mmu_iounmap() 128 gpa_t gpa = slot->base_gfn << PAGE_SHIFT; in kvm_arch_flush_shadow_memslot() local 135 kvm_riscv_gstage_unmap_range(&gstage, gpa, size, false); in kvm_arch_flush_shadow_memslot() 398 kvm_pfn_t *hfnp, gpa_t *gpa) in transparent_hugepage_adjust() argument 414 *gpa &= PMD_MASK; in transparent_hugepage_adjust() 425 gpa_t gpa, unsigned long hva, bool is_write, in kvm_riscv_mmu_map() argument [all …]
|
| H A D | tlb.c | 24 gpa_t gpa, gpa_t gpsz, in kvm_riscv_local_hfence_gvma_vmid_gpa() argument 36 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa() 41 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa() 52 void kvm_riscv_local_hfence_gvma_gpa(gpa_t gpa, gpa_t gpsz, in kvm_riscv_local_hfence_gvma_gpa() argument 64 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa() 69 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa() 374 gpa_t gpa, gpa_t gpsz, in kvm_riscv_hfence_gvma_vmid_gpa() argument 382 data.addr = gpa; in kvm_riscv_hfence_gvma_vmid_gpa()
|
| /linux/drivers/gpu/drm/i915/gvt/ |
| H A D | page_track.c | 160 int intel_vgpu_page_track_handler(struct intel_vgpu *vgpu, u64 gpa, in intel_vgpu_page_track_handler() argument 166 page_track = intel_vgpu_find_page_track(vgpu, gpa >> PAGE_SHIFT); in intel_vgpu_page_track_handler() 172 intel_gvt_page_track_remove(vgpu, gpa >> PAGE_SHIFT); in intel_vgpu_page_track_handler() 174 ret = page_track->handler(page_track, gpa, data, bytes); in intel_vgpu_page_track_handler() 176 gvt_err("guest page write error, gpa %llx\n", gpa); in intel_vgpu_page_track_handler()
|
| /linux/arch/x86/kvm/vmx/ |
| H A D | common.h | 77 static inline bool vt_is_tdx_private_gpa(struct kvm *kvm, gpa_t gpa) in vt_is_tdx_private_gpa() argument 80 return !kvm_is_addr_direct(kvm, gpa); in vt_is_tdx_private_gpa() 83 static inline int __vmx_handle_ept_violation(struct kvm_vcpu *vcpu, gpa_t gpa, in __vmx_handle_ept_violation() argument 105 if (vt_is_tdx_private_gpa(vcpu->kvm, gpa)) in __vmx_handle_ept_violation() 108 return kvm_mmu_page_fault(vcpu, gpa, error_code, NULL, 0); in __vmx_handle_ept_violation()
|
| /linux/arch/loongarch/kvm/ |
| H A D | tlb.c | 24 void kvm_flush_tlb_gpa(struct kvm_vcpu *vcpu, unsigned long gpa) in kvm_flush_tlb_gpa() argument 27 gpa &= (PAGE_MASK << 1); in kvm_flush_tlb_gpa() 28 invtlb(INVTLB_GID_ADDR, read_csr_gstat() & CSR_GSTAT_GID, gpa); in kvm_flush_tlb_gpa()
|
| H A D | mmu.c | 528 gpa_t gpa = range->start << PAGE_SHIFT; in kvm_test_age_gfn() local 529 kvm_pte_t *ptep = kvm_populate_gpa(kvm, NULL, gpa, 0); in kvm_test_age_gfn() 552 static int kvm_map_page_fast(struct kvm_vcpu *vcpu, unsigned long gpa, bool write) in kvm_map_page_fast() argument 556 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_map_page_fast() 563 ptep = kvm_populate_gpa(kvm, NULL, gpa, 0); in kvm_map_page_fast() 772 static int kvm_map_page(struct kvm_vcpu *vcpu, unsigned long gpa, bool write) in kvm_map_page() argument 779 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_map_page() 787 err = kvm_map_page_fast(vcpu, gpa, write); in kvm_map_page() 875 ptep = kvm_populate_gpa(kvm, NULL, gpa, 0); in kvm_map_page() 887 ptep = kvm_populate_gpa(kvm, memcache, gpa, level); in kvm_map_page() [all …]
|
| H A D | vcpu.c | 156 gpa_t gpa; in kvm_update_stolen_time() local 162 gpa = vcpu->arch.st.guest_addr; in kvm_update_stolen_time() 163 if (!(gpa & KVM_STEAL_PHYS_VALID)) in kvm_update_stolen_time() 166 gpa &= KVM_STEAL_PHYS_MASK; in kvm_update_stolen_time() 168 if (slots->generation != ghc->generation || gpa != ghc->gpa) { in kvm_update_stolen_time() 169 if (kvm_gfn_to_hva_cache_init(vcpu->kvm, ghc, gpa, sizeof(*st))) { in kvm_update_stolen_time() 170 ghc->gpa = INVALID_GPA; in kvm_update_stolen_time() 198 mark_page_dirty_in_slot(vcpu->kvm, ghc->memslot, gpa_to_gfn(ghc->gpa)); in kvm_update_stolen_time() 1119 u64 gpa; in kvm_loongarch_pvtime_get_attr() local 1126 gpa = vcpu->arch.st.guest_addr; in kvm_loongarch_pvtime_get_attr() [all …]
|
| /linux/arch/riscv/include/asm/ |
| H A D | kvm_mmu.h | 11 int kvm_riscv_mmu_ioremap(struct kvm *kvm, gpa_t gpa, phys_addr_t hpa, 13 void kvm_riscv_mmu_iounmap(struct kvm *kvm, gpa_t gpa, unsigned long size); 15 gpa_t gpa, unsigned long hva, bool is_write,
|
| /linux/arch/x86/kvm/ |
| H A D | mmu.h | 103 void kvm_mmu_track_write(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new, 291 gpa_t translate_nested_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u64 access, 296 gpa_t gpa, u64 access, in kvm_translate_gpa() argument 300 return gpa; in kvm_translate_gpa() 301 return translate_nested_gpa(vcpu, gpa, access, exception); in kvm_translate_gpa() 314 static inline bool kvm_is_addr_direct(struct kvm *kvm, gpa_t gpa) in kvm_is_addr_direct() argument 318 return !gpa_direct_bits || (gpa & gpa_direct_bits); in kvm_is_addr_direct()
|
| /linux/tools/testing/selftests/kvm/include/ |
| H A D | kvm_util.h | 421 static inline void vm_set_memory_attributes(struct kvm_vm *vm, gpa_t gpa, in vm_mem_set_private() 426 .address = gpa, in vm_mem_set_shared() 442 static inline void vm_mem_set_private(struct kvm_vm *vm, gpa_t gpa, in vm_guest_mem_allocate() 445 vm_set_memory_attributes(vm, gpa, size, KVM_MEMORY_ATTRIBUTE_PRIVATE); 448 static inline void vm_mem_set_shared(struct kvm_vm *vm, gpa_t gpa, 451 vm_set_memory_attributes(vm, gpa, size, 0); 454 void vm_guest_mem_fallocate(struct kvm_vm *vm, gpa_t gpa, u64 size, 457 static inline void vm_guest_mem_punch_hole(struct kvm_vm *vm, gpa_t gpa, in kvm_vm_get_dirty_log() 460 vm_guest_mem_fallocate(vm, gpa, size, true); in kvm_vm_get_dirty_log() 463 static inline void vm_guest_mem_allocate(struct kvm_vm *vm, gpa_t gpa, in kvm_vm_clear_dirty_log() 397 vm_set_memory_attributes(struct kvm_vm * vm,uint64_t gpa,uint64_t size,uint64_t attributes) vm_set_memory_attributes() argument 418 vm_mem_set_private(struct kvm_vm * vm,uint64_t gpa,uint64_t size) vm_mem_set_private() argument 424 vm_mem_set_shared(struct kvm_vm * vm,uint64_t gpa,uint64_t size) vm_mem_set_shared() argument 433 vm_guest_mem_punch_hole(struct kvm_vm * vm,uint64_t gpa,uint64_t size) vm_guest_mem_punch_hole() argument 439 vm_guest_mem_allocate(struct kvm_vm * vm,uint64_t gpa,uint64_t size) vm_guest_mem_allocate() argument 719 vm_untag_gpa(struct kvm_vm * vm,vm_paddr_t gpa) vm_untag_gpa() argument [all...] |
| /linux/include/linux/ |
| H A D | kvm_host.h | 153 static inline bool kvm_is_error_gpa(gpa_t gpa) in kvm_is_error_gpa() argument 155 return gpa == INVALID_GPA; in kvm_is_error_gpa() 319 gpa_t gpa; member 1325 int kvm_read_guest(struct kvm *kvm, gpa_t gpa, void *data, unsigned long len); 1333 int kvm_write_guest(struct kvm *kvm, gpa_t gpa, const void *data, 1341 gpa_t gpa, unsigned long len); 1354 #define kvm_get_guest(kvm, gpa, v) \ argument 1356 gpa_t __gpa = gpa; \ 1376 #define kvm_put_guest(kvm, gpa, v) \ argument 1378 gpa_t __gpa = gpa; \ [all …]
|