| /linux/tools/testing/selftests/kvm/lib/loongarch/ |
| H A D | ucall.c | 12 gva_t *ucall_exit_mmio_addr; 16 gva_t mmio_gva = vm_unused_gva_gap(vm, vm->page_size, KVM_UTIL_MIN_VADDR); in ucall_arch_init() 22 write_guest_global(vm, ucall_exit_mmio_addr, (gva_t *)mmio_gva); in ucall_arch_init()
|
| H A D | processor.c | 16 static gva_t exception_handlers; 18 static u64 virt_pte_index(struct kvm_vm *vm, gva_t gva, int level) in virt_pte_index() 75 static u64 *virt_populate_pte(struct kvm_vm *vm, gva_t gva, int alloc) in virt_populate_pte() 109 gpa_t addr_arch_gva2gpa(struct kvm_vm *vm, gva_t gva) in addr_arch_gva2gpa() 119 void virt_arch_pg_map(struct kvm_vm *vm, gva_t gva, gpa_t gpa) in virt_arch_pg_map()
|
| /linux/tools/testing/selftests/kvm/lib/arm64/ |
| H A D | ucall.c | 9 gva_t *ucall_exit_mmio_addr; 13 gva_t mmio_gva = vm_unused_gva_gap(vm, vm->page_size, KVM_UTIL_MIN_VADDR); in ucall_arch_init() 19 write_guest_global(vm, ucall_exit_mmio_addr, (gva_t *)mmio_gva); in ucall_arch_init()
|
| H A D | processor.c | 22 static gva_t exception_handlers; 24 static u64 pgd_index(struct kvm_vm *vm, gva_t gva) in pgd_index() 32 static u64 pud_index(struct kvm_vm *vm, gva_t gva) in pud_index() 43 static u64 pmd_index(struct kvm_vm *vm, gva_t gva) in pmd_index() 54 static u64 pte_index(struct kvm_vm *vm, gva_t gva) in pte_index() 124 static void _virt_pg_map(struct kvm_vm *vm, gva_t gva, gpa_t gpa, in _virt_pg_map() 176 void virt_arch_pg_map(struct kvm_vm *vm, gva_t gva, gpa_t gpa) in virt_arch_pg_map() 183 u64 *virt_get_pte_hva_at_level(struct kvm_vm *vm, gva_t gva, int level) in virt_get_pte_hva_at_level() 227 u64 *virt_get_pte_hva(struct kvm_vm *vm, gva_t gva) in virt_get_pte_hva() 232 gpa_t addr_arch_gva2gpa(struct kvm_vm *vm, gva_t gva) in addr_arch_gva2gpa() [all …]
|
| /linux/arch/powerpc/kvm/ |
| H A D | book3s_32_mmu.c | 69 static int kvmppc_mmu_book3s_32_xlate_bat(struct kvm_vcpu *vcpu, gva_t eaddr, 75 static u32 find_sr(struct kvm_vcpu *vcpu, gva_t eaddr) in find_sr() 80 static u64 kvmppc_mmu_book3s_32_ea_to_vp(struct kvm_vcpu *vcpu, gva_t eaddr, in kvmppc_mmu_book3s_32_ea_to_vp() 94 u32 sre, gva_t eaddr, in kvmppc_mmu_book3s_32_get_pteg() 121 static u32 kvmppc_mmu_book3s_32_get_ptem(u32 sre, gva_t eaddr, bool primary) in kvmppc_mmu_book3s_32_get_ptem() 127 static int kvmppc_mmu_book3s_32_xlate_bat(struct kvm_vcpu *vcpu, gva_t eaddr, in kvmppc_mmu_book3s_32_xlate_bat() 182 static int kvmppc_mmu_book3s_32_xlate_pte(struct kvm_vcpu *vcpu, gva_t eaddr, in kvmppc_mmu_book3s_32_xlate_pte() 290 static int kvmppc_mmu_book3s_32_xlate(struct kvm_vcpu *vcpu, gva_t eaddr, in kvmppc_mmu_book3s_32_xlate()
|
| H A D | e500.h | 127 int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea); 128 int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea); 129 int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea); 154 static inline gva_t get_tlb_eaddr(const struct kvm_book3e_206_tlb_entry *tlbe) in get_tlb_eaddr() 165 static inline gva_t get_tlb_end(const struct kvm_book3e_206_tlb_entry *tlbe) in get_tlb_end()
|
| H A D | e500_mmu.c | 48 static int tlb0_set_base(gva_t addr, int sets, int ways) in tlb0_set_base() 58 static int gtlb0_set_base(struct kvmppc_vcpu_e500 *vcpu_e500, gva_t addr) in gtlb0_set_base() 81 gva_t eaddr, int tlbsel, unsigned int pid, int as) in kvmppc_e500_tlb_index() 127 gva_t eaddr, int as) in kvmppc_e500_deliver_tlb_miss() 155 gva_t eaddr; in kvmppc_recalc_tlb1map_range() 245 int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea) in kvmppc_e500_emul_tlbivax() 293 gva_t ea) in tlbilx_one() 307 int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea) in kvmppc_e500_emul_tlbilx() 341 int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea) in kvmppc_e500_emul_tlbsx() 454 gva_t eaddr, unsigned int pid, int as) in kvmppc_e500_tlb_search() [all …]
|
| H A D | book3s_64_mmu.c | 29 gva_t eaddr) in kvmppc_mmu_book3s_64_find_slbe() 73 static u64 kvmppc_slb_calc_vpn(struct kvmppc_slb *slb, gva_t eaddr) in kvmppc_slb_calc_vpn() 81 static u64 kvmppc_mmu_book3s_64_ea_to_vp(struct kvm_vcpu *vcpu, gva_t eaddr, in kvmppc_mmu_book3s_64_ea_to_vp() 109 static u32 kvmppc_mmu_book3s_64_get_page(struct kvmppc_slb *slbe, gva_t eaddr) in kvmppc_mmu_book3s_64_get_page() 117 struct kvmppc_slb *slbe, gva_t eaddr, in kvmppc_mmu_book3s_64_get_pteg() 155 static u64 kvmppc_mmu_book3s_64_get_avpn(struct kvmppc_slb *slbe, gva_t eaddr) in kvmppc_mmu_book3s_64_get_avpn() 191 static int kvmppc_mmu_book3s_64_xlate(struct kvm_vcpu *vcpu, gva_t eaddr, in kvmppc_mmu_book3s_64_xlate() 412 static int kvmppc_mmu_book3s_64_slbfee(struct kvm_vcpu *vcpu, gva_t eaddr, in kvmppc_mmu_book3s_64_slbfee()
|
| /linux/tools/testing/selftests/kvm/include/ |
| H A D | kvm_util.h | 115 gva_t handlers; 718 gva_t vm_unused_gva_gap(struct kvm_vm *vm, size_t sz, gva_t min_gva); 719 gva_t vm_alloc(struct kvm_vm *vm, size_t sz, gva_t min_gva); 720 gva_t __vm_alloc(struct kvm_vm *vm, size_t sz, gva_t min_gva, 722 gva_t vm_alloc_shared(struct kvm_vm *vm, size_t sz, gva_t min_gva, 724 gva_t vm_alloc_pages(struct kvm_vm *vm, int nr_pages); 725 gva_t __vm_alloc_page(struct kvm_vm *vm, enum kvm_mem_region_type type); 726 gva_t vm_alloc_page(struct kvm_vm *vm); 728 void virt_map(struct kvm_vm *vm, gva_t gva, gpa_t gpa, 731 void *addr_gva2hva(struct kvm_vm *vm, gva_t gva); [all …]
|
| /linux/tools/testing/selftests/kvm/include/x86/ |
| H A D | kvm_util_arch.h | 36 gva_t gdt; 37 gva_t tss; 38 gva_t idt;
|
| H A D | hyperv.h | 257 static inline u8 __hyperv_hypercall(u64 control, gva_t input_address, in __hyperv_hypercall() 258 gva_t output_address, in __hyperv_hypercall() 277 static inline void hyperv_hypercall(u64 control, gva_t input_address, in hyperv_hypercall() 278 gva_t output_address) in hyperv_hypercall() 350 gva_t *p_hv_pages_gva);
|
| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | nested_dirty_log_test.c | 50 static void l2_guest_code(gva_t base) in l2_guest_code() 52 gva_t page0 = TEST_GUEST_ADDR(base, 0); in l2_guest_code() 53 gva_t page1 = TEST_GUEST_ADDR(base, 1); in l2_guest_code() 146 gva_t gva = arg & ~(PAGE_SIZE - 1); in test_handle_ucall_sync() 201 gva_t nested_gva = 0; in test_dirty_log()
|
| H A D | hyperv_extended_hypercalls.c | 19 gva_t out_pg_gva) in guest_code() 38 gva_t hcall_out_page; in main() 39 gva_t hcall_in_page; in main()
|
| H A D | svm_nested_vmcb12_gpa.c | 77 gva_t nested_gva = 0; in test_invalid_vmcb12() 93 gva_t nested_gva = 0; in test_unmappable_vmcb12() 106 gva_t nested_gva = 0; in test_unmappable_vmcb12_vmexit()
|
| H A D | triple_fault_event_test.c | 75 gva_t vmx_pages_gva; in main() 81 gva_t svm_gva; in main()
|
| H A D | hyperv_tlb_flush.c | 64 gva_t hcall_gva; 66 gva_t test_pages; 67 gva_t test_pages_pte[NTEST_PAGES]; 71 static void worker_guest_code(gva_t test_data) in worker_guest_code() 199 static void sender_guest_code(gva_t test_data) in sender_guest_code() 584 gva_t test_data_page, gva; in main()
|
| H A D | cpuid_test.c | 143 struct kvm_cpuid2 *vcpu_alloc_cpuid(struct kvm_vm *vm, gva_t *p_gva, struct kvm_cpuid2 *cpuid) in vcpu_alloc_cpuid() 146 gva_t gva = vm_alloc(vm, size, KVM_UTIL_MIN_VADDR); in vcpu_alloc_cpuid() 220 gva_t cpuid_gva; in main()
|
| /linux/tools/testing/selftests/kvm/lib/ |
| H A D | ucall_common.c | 32 gva_t gva; in ucall_init() 99 ucall_arch_do_ucall((gva_t)uc->hva); in ucall_assert() 116 ucall_arch_do_ucall((gva_t)uc->hva); in ucall_fmt() 138 ucall_arch_do_ucall((gva_t)uc->hva); in ucall()
|
| H A D | elf.c | 160 gva_t seg_vstart = align_down(phdr.p_vaddr, vm->page_size); in kvm_vm_elf_load() 161 gva_t seg_vend = phdr.p_vaddr + phdr.p_memsz - 1; in kvm_vm_elf_load() 165 gva_t gva = __vm_alloc(vm, seg_size, seg_vstart, MEM_REGION_CODE); in kvm_vm_elf_load()
|
| H A D | kvm_util.c | 1374 gva_t vm_unused_gva_gap(struct kvm_vm *vm, size_t sz, gva_t min_gva) in vm_unused_gva_gap() 1439 static gva_t ____vm_alloc(struct kvm_vm *vm, size_t sz, gva_t min_gva, in ____vm_alloc() 1453 gva_t gva_start = vm_unused_gva_gap(vm, sz, min_gva); in ____vm_alloc() 1456 for (gva_t gva = gva_start; pages > 0; in ____vm_alloc() 1465 gva_t __vm_alloc(struct kvm_vm *vm, size_t sz, gva_t min_gva, in __vm_alloc() 1472 gva_t vm_alloc_shared(struct kvm_vm *vm, size_t sz, gva_t min_gva, in vm_alloc_shared() 1485 gva_t vm_alloc(struct kvm_vm *vm, size_t sz, gva_t min_gva) in vm_alloc() 1490 gva_t vm_alloc_pages(struct kvm_vm *vm, int nr_pages) in vm_alloc_pages() 1495 gva_t __vm_alloc_page(struct kvm_vm *vm, enum kvm_mem_region_type type) in __vm_alloc_page() 1500 gva_t vm_alloc_page(struct kvm_vm *vm) in vm_alloc_page() [all …]
|
| /linux/tools/testing/selftests/kvm/include/arm64/ |
| H A D | ucall.h | 13 extern gva_t *ucall_exit_mmio_addr; 15 static inline void ucall_arch_do_ucall(gva_t uc) in ucall_arch_do_ucall()
|
| /linux/tools/testing/selftests/kvm/include/loongarch/ |
| H A D | ucall.h | 13 extern gva_t *ucall_exit_mmio_addr; 15 static inline void ucall_arch_do_ucall(gva_t uc) in ucall_arch_do_ucall()
|
| /linux/tools/testing/selftests/kvm/lib/x86/ |
| H A D | processor.c | 24 gva_t exception_handlers; 210 u64 *parent_pte, gva_t gva, int level) in virt_get_pte() 226 gva_t gva, in virt_create_upper_pte() 259 void __virt_pg_map(struct kvm_vm *vm, struct kvm_mmu *mmu, gva_t gva, in __virt_pg_map() 318 void virt_arch_pg_map(struct kvm_vm *vm, gva_t gva, gpa_t gpa) in virt_arch_pg_map() 323 void virt_map_level(struct kvm_vm *vm, gva_t gva, gpa_t gpa, in virt_map_level() 359 gva_t gva, in __vm_get_page_table_entry() 400 u64 *vm_get_pte(struct kvm_vm *vm, gva_t gva) in vm_get_pte() 616 gpa_t addr_arch_gva2gpa(struct kvm_vm *vm, gva_t gva) in addr_arch_gva2gpa() 631 static void kvm_seg_set_tss_64bit(gva_t base, struct kvm_segment *segp) in kvm_seg_set_tss_64bit() [all …]
|
| H A D | hyperv.c | 79 gva_t *p_hv_pages_gva) in vcpu_alloc_hyperv_test_pages() 81 gva_t hv_pages_gva = vm_alloc_page(vm); in vcpu_alloc_hyperv_test_pages()
|
| /linux/tools/testing/selftests/kvm/s390/ |
| H A D | tprot.c | 210 gva_t guest_0_page; in main() 219 mprotect(addr_gva2hva(vm, (gva_t)pages), PAGE_SIZE * 2, PROT_READ); in main() 232 mprotect(addr_gva2hva(vm, (gva_t)0), PAGE_SIZE, PROT_READ); in main()
|