Home
last modified time | relevance | path

Searched refs:gva_t (Results 1 – 25 of 100) sorted by relevance

1234

/linux/tools/testing/selftests/kvm/lib/loongarch/
H A Ducall.c12 gva_t *ucall_exit_mmio_addr;
16 gva_t mmio_gva = vm_unused_gva_gap(vm, vm->page_size, KVM_UTIL_MIN_VADDR); in ucall_arch_init()
22 write_guest_global(vm, ucall_exit_mmio_addr, (gva_t *)mmio_gva); in ucall_arch_init()
H A Dprocessor.c16 static gva_t exception_handlers;
18 static u64 virt_pte_index(struct kvm_vm *vm, gva_t gva, int level) in virt_pte_index()
75 static u64 *virt_populate_pte(struct kvm_vm *vm, gva_t gva, int alloc) in virt_populate_pte()
109 gpa_t addr_arch_gva2gpa(struct kvm_vm *vm, gva_t gva) in addr_arch_gva2gpa()
119 void virt_arch_pg_map(struct kvm_vm *vm, gva_t gva, gpa_t gpa) in virt_arch_pg_map()
/linux/tools/testing/selftests/kvm/lib/arm64/
H A Ducall.c9 gva_t *ucall_exit_mmio_addr;
13 gva_t mmio_gva = vm_unused_gva_gap(vm, vm->page_size, KVM_UTIL_MIN_VADDR); in ucall_arch_init()
19 write_guest_global(vm, ucall_exit_mmio_addr, (gva_t *)mmio_gva); in ucall_arch_init()
H A Dprocessor.c22 static gva_t exception_handlers;
24 static u64 pgd_index(struct kvm_vm *vm, gva_t gva) in pgd_index()
32 static u64 pud_index(struct kvm_vm *vm, gva_t gva) in pud_index()
43 static u64 pmd_index(struct kvm_vm *vm, gva_t gva) in pmd_index()
54 static u64 pte_index(struct kvm_vm *vm, gva_t gva) in pte_index()
124 static void _virt_pg_map(struct kvm_vm *vm, gva_t gva, gpa_t gpa, in _virt_pg_map()
176 void virt_arch_pg_map(struct kvm_vm *vm, gva_t gva, gpa_t gpa) in virt_arch_pg_map()
183 u64 *virt_get_pte_hva_at_level(struct kvm_vm *vm, gva_t gva, int level) in virt_get_pte_hva_at_level()
227 u64 *virt_get_pte_hva(struct kvm_vm *vm, gva_t gva) in virt_get_pte_hva()
232 gpa_t addr_arch_gva2gpa(struct kvm_vm *vm, gva_t gva) in addr_arch_gva2gpa()
[all …]
/linux/arch/powerpc/kvm/
H A Dbook3s_32_mmu.c69 static int kvmppc_mmu_book3s_32_xlate_bat(struct kvm_vcpu *vcpu, gva_t eaddr,
75 static u32 find_sr(struct kvm_vcpu *vcpu, gva_t eaddr) in find_sr()
80 static u64 kvmppc_mmu_book3s_32_ea_to_vp(struct kvm_vcpu *vcpu, gva_t eaddr, in kvmppc_mmu_book3s_32_ea_to_vp()
94 u32 sre, gva_t eaddr, in kvmppc_mmu_book3s_32_get_pteg()
121 static u32 kvmppc_mmu_book3s_32_get_ptem(u32 sre, gva_t eaddr, bool primary) in kvmppc_mmu_book3s_32_get_ptem()
127 static int kvmppc_mmu_book3s_32_xlate_bat(struct kvm_vcpu *vcpu, gva_t eaddr, in kvmppc_mmu_book3s_32_xlate_bat()
182 static int kvmppc_mmu_book3s_32_xlate_pte(struct kvm_vcpu *vcpu, gva_t eaddr, in kvmppc_mmu_book3s_32_xlate_pte()
290 static int kvmppc_mmu_book3s_32_xlate(struct kvm_vcpu *vcpu, gva_t eaddr, in kvmppc_mmu_book3s_32_xlate()
H A De500.h127 int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea);
128 int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea);
129 int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea);
154 static inline gva_t get_tlb_eaddr(const struct kvm_book3e_206_tlb_entry *tlbe) in get_tlb_eaddr()
165 static inline gva_t get_tlb_end(const struct kvm_book3e_206_tlb_entry *tlbe) in get_tlb_end()
H A De500_mmu.c48 static int tlb0_set_base(gva_t addr, int sets, int ways) in tlb0_set_base()
58 static int gtlb0_set_base(struct kvmppc_vcpu_e500 *vcpu_e500, gva_t addr) in gtlb0_set_base()
81 gva_t eaddr, int tlbsel, unsigned int pid, int as) in kvmppc_e500_tlb_index()
127 gva_t eaddr, int as) in kvmppc_e500_deliver_tlb_miss()
155 gva_t eaddr; in kvmppc_recalc_tlb1map_range()
245 int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea) in kvmppc_e500_emul_tlbivax()
293 gva_t ea) in tlbilx_one()
307 int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea) in kvmppc_e500_emul_tlbilx()
341 int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea) in kvmppc_e500_emul_tlbsx()
454 gva_t eaddr, unsigned int pid, int as) in kvmppc_e500_tlb_search()
[all …]
H A Dbook3s_64_mmu.c29 gva_t eaddr) in kvmppc_mmu_book3s_64_find_slbe()
73 static u64 kvmppc_slb_calc_vpn(struct kvmppc_slb *slb, gva_t eaddr) in kvmppc_slb_calc_vpn()
81 static u64 kvmppc_mmu_book3s_64_ea_to_vp(struct kvm_vcpu *vcpu, gva_t eaddr, in kvmppc_mmu_book3s_64_ea_to_vp()
109 static u32 kvmppc_mmu_book3s_64_get_page(struct kvmppc_slb *slbe, gva_t eaddr) in kvmppc_mmu_book3s_64_get_page()
117 struct kvmppc_slb *slbe, gva_t eaddr, in kvmppc_mmu_book3s_64_get_pteg()
155 static u64 kvmppc_mmu_book3s_64_get_avpn(struct kvmppc_slb *slbe, gva_t eaddr) in kvmppc_mmu_book3s_64_get_avpn()
191 static int kvmppc_mmu_book3s_64_xlate(struct kvm_vcpu *vcpu, gva_t eaddr, in kvmppc_mmu_book3s_64_xlate()
412 static int kvmppc_mmu_book3s_64_slbfee(struct kvm_vcpu *vcpu, gva_t eaddr, in kvmppc_mmu_book3s_64_slbfee()
/linux/tools/testing/selftests/kvm/include/
H A Dkvm_util.h115 gva_t handlers;
718 gva_t vm_unused_gva_gap(struct kvm_vm *vm, size_t sz, gva_t min_gva);
719 gva_t vm_alloc(struct kvm_vm *vm, size_t sz, gva_t min_gva);
720 gva_t __vm_alloc(struct kvm_vm *vm, size_t sz, gva_t min_gva,
722 gva_t vm_alloc_shared(struct kvm_vm *vm, size_t sz, gva_t min_gva,
724 gva_t vm_alloc_pages(struct kvm_vm *vm, int nr_pages);
725 gva_t __vm_alloc_page(struct kvm_vm *vm, enum kvm_mem_region_type type);
726 gva_t vm_alloc_page(struct kvm_vm *vm);
728 void virt_map(struct kvm_vm *vm, gva_t gva, gpa_t gpa,
731 void *addr_gva2hva(struct kvm_vm *vm, gva_t gva);
[all …]
/linux/tools/testing/selftests/kvm/include/x86/
H A Dkvm_util_arch.h36 gva_t gdt;
37 gva_t tss;
38 gva_t idt;
H A Dhyperv.h257 static inline u8 __hyperv_hypercall(u64 control, gva_t input_address, in __hyperv_hypercall()
258 gva_t output_address, in __hyperv_hypercall()
277 static inline void hyperv_hypercall(u64 control, gva_t input_address, in hyperv_hypercall()
278 gva_t output_address) in hyperv_hypercall()
350 gva_t *p_hv_pages_gva);
/linux/tools/testing/selftests/kvm/x86/
H A Dnested_dirty_log_test.c50 static void l2_guest_code(gva_t base) in l2_guest_code()
52 gva_t page0 = TEST_GUEST_ADDR(base, 0); in l2_guest_code()
53 gva_t page1 = TEST_GUEST_ADDR(base, 1); in l2_guest_code()
146 gva_t gva = arg & ~(PAGE_SIZE - 1); in test_handle_ucall_sync()
201 gva_t nested_gva = 0; in test_dirty_log()
H A Dhyperv_extended_hypercalls.c19 gva_t out_pg_gva) in guest_code()
38 gva_t hcall_out_page; in main()
39 gva_t hcall_in_page; in main()
H A Dsvm_nested_vmcb12_gpa.c77 gva_t nested_gva = 0; in test_invalid_vmcb12()
93 gva_t nested_gva = 0; in test_unmappable_vmcb12()
106 gva_t nested_gva = 0; in test_unmappable_vmcb12_vmexit()
H A Dtriple_fault_event_test.c75 gva_t vmx_pages_gva; in main()
81 gva_t svm_gva; in main()
H A Dhyperv_tlb_flush.c64 gva_t hcall_gva;
66 gva_t test_pages;
67 gva_t test_pages_pte[NTEST_PAGES];
71 static void worker_guest_code(gva_t test_data) in worker_guest_code()
199 static void sender_guest_code(gva_t test_data) in sender_guest_code()
584 gva_t test_data_page, gva; in main()
H A Dcpuid_test.c143 struct kvm_cpuid2 *vcpu_alloc_cpuid(struct kvm_vm *vm, gva_t *p_gva, struct kvm_cpuid2 *cpuid) in vcpu_alloc_cpuid()
146 gva_t gva = vm_alloc(vm, size, KVM_UTIL_MIN_VADDR); in vcpu_alloc_cpuid()
220 gva_t cpuid_gva; in main()
/linux/tools/testing/selftests/kvm/lib/
H A Ducall_common.c32 gva_t gva; in ucall_init()
99 ucall_arch_do_ucall((gva_t)uc->hva); in ucall_assert()
116 ucall_arch_do_ucall((gva_t)uc->hva); in ucall_fmt()
138 ucall_arch_do_ucall((gva_t)uc->hva); in ucall()
H A Delf.c160 gva_t seg_vstart = align_down(phdr.p_vaddr, vm->page_size); in kvm_vm_elf_load()
161 gva_t seg_vend = phdr.p_vaddr + phdr.p_memsz - 1; in kvm_vm_elf_load()
165 gva_t gva = __vm_alloc(vm, seg_size, seg_vstart, MEM_REGION_CODE); in kvm_vm_elf_load()
H A Dkvm_util.c1374 gva_t vm_unused_gva_gap(struct kvm_vm *vm, size_t sz, gva_t min_gva) in vm_unused_gva_gap()
1439 static gva_t ____vm_alloc(struct kvm_vm *vm, size_t sz, gva_t min_gva, in ____vm_alloc()
1453 gva_t gva_start = vm_unused_gva_gap(vm, sz, min_gva); in ____vm_alloc()
1456 for (gva_t gva = gva_start; pages > 0; in ____vm_alloc()
1465 gva_t __vm_alloc(struct kvm_vm *vm, size_t sz, gva_t min_gva, in __vm_alloc()
1472 gva_t vm_alloc_shared(struct kvm_vm *vm, size_t sz, gva_t min_gva, in vm_alloc_shared()
1485 gva_t vm_alloc(struct kvm_vm *vm, size_t sz, gva_t min_gva) in vm_alloc()
1490 gva_t vm_alloc_pages(struct kvm_vm *vm, int nr_pages) in vm_alloc_pages()
1495 gva_t __vm_alloc_page(struct kvm_vm *vm, enum kvm_mem_region_type type) in __vm_alloc_page()
1500 gva_t vm_alloc_page(struct kvm_vm *vm) in vm_alloc_page()
[all …]
/linux/tools/testing/selftests/kvm/include/arm64/
H A Ducall.h13 extern gva_t *ucall_exit_mmio_addr;
15 static inline void ucall_arch_do_ucall(gva_t uc) in ucall_arch_do_ucall()
/linux/tools/testing/selftests/kvm/include/loongarch/
H A Ducall.h13 extern gva_t *ucall_exit_mmio_addr;
15 static inline void ucall_arch_do_ucall(gva_t uc) in ucall_arch_do_ucall()
/linux/tools/testing/selftests/kvm/lib/x86/
H A Dprocessor.c24 gva_t exception_handlers;
210 u64 *parent_pte, gva_t gva, int level) in virt_get_pte()
226 gva_t gva, in virt_create_upper_pte()
259 void __virt_pg_map(struct kvm_vm *vm, struct kvm_mmu *mmu, gva_t gva, in __virt_pg_map()
318 void virt_arch_pg_map(struct kvm_vm *vm, gva_t gva, gpa_t gpa) in virt_arch_pg_map()
323 void virt_map_level(struct kvm_vm *vm, gva_t gva, gpa_t gpa, in virt_map_level()
359 gva_t gva, in __vm_get_page_table_entry()
400 u64 *vm_get_pte(struct kvm_vm *vm, gva_t gva) in vm_get_pte()
616 gpa_t addr_arch_gva2gpa(struct kvm_vm *vm, gva_t gva) in addr_arch_gva2gpa()
631 static void kvm_seg_set_tss_64bit(gva_t base, struct kvm_segment *segp) in kvm_seg_set_tss_64bit()
[all …]
H A Dhyperv.c79 gva_t *p_hv_pages_gva) in vcpu_alloc_hyperv_test_pages()
81 gva_t hv_pages_gva = vm_alloc_page(vm); in vcpu_alloc_hyperv_test_pages()
/linux/tools/testing/selftests/kvm/s390/
H A Dtprot.c210 gva_t guest_0_page; in main()
219 mprotect(addr_gva2hva(vm, (gva_t)pages), PAGE_SIZE * 2, PROT_READ); in main()
232 mprotect(addr_gva2hva(vm, (gva_t)0), PAGE_SIZE, PROT_READ); in main()

1234