| /linux/tools/testing/selftests/kvm/lib/s390/ |
| H A D | processor.c | 50 void virt_arch_pg_map(struct kvm_vm *vm, uint64_t gva, uint64_t gpa) in virt_arch_pg_map() argument 55 TEST_ASSERT((gva % vm->page_size) == 0, in virt_arch_pg_map() 58 gva, vm->page_size); in virt_arch_pg_map() 60 (gva >> vm->page_shift)), in virt_arch_pg_map() 62 gva); in virt_arch_pg_map() 66 gva, vm->page_size); in virt_arch_pg_map() 70 gva, vm->max_gfn, vm->page_size); in virt_arch_pg_map() 75 idx = (gva >> (64 - 11 * ri)) & 0x7ffu; in virt_arch_pg_map() 82 idx = (gva >> PAGE_SHIFT) & 0x0ffu; /* page index */ in virt_arch_pg_map() 89 vm_paddr_t addr_arch_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva) in addr_arch_gva2gpa() argument [all...] |
| /linux/arch/riscv/kvm/ |
| H A D | tlb.c | 82 unsigned long gva, in kvm_riscv_local_hfence_vvma_asid_gva() argument 97 for (pos = gva; pos < (gva + gvsz); pos += BIT(order)) in kvm_riscv_local_hfence_vvma_asid_gva() 102 for (pos = gva; pos < (gva + gvsz); pos += BIT(order)) in kvm_riscv_local_hfence_vvma_asid_gva() 123 unsigned long gva, unsigned long gvsz, in kvm_riscv_local_hfence_vvma_gva() argument 137 for (pos = gva; pos < (gva + gvsz); pos += BIT(order)) in kvm_riscv_local_hfence_vvma_gva() 142 for (pos = gva; pos < (gva + gvsz); pos += BIT(order)) in kvm_riscv_local_hfence_vvma_gva() 402 unsigned long gva, unsigned long gvsz, in kvm_riscv_hfence_vvma_asid_gva() argument 411 data.addr = gva; in kvm_riscv_hfence_vvma_asid_gva() 433 unsigned long gva, unsigned long gvsz, in kvm_riscv_hfence_vvma_gva() argument 441 data.addr = gva; in kvm_riscv_hfence_vvma_gva()
|
| /linux/tools/testing/selftests/kvm/lib/arm64/ |
| H A D | processor.c | 24 static uint64_t pgd_index(struct kvm_vm *vm, vm_vaddr_t gva) in page_align() 29 return (gva >> shift) & mask; in pgd_index() argument 32 static uint64_t pud_index(struct kvm_vm *vm, vm_vaddr_t gva) in pgd_index() 40 return (gva >> shift) & mask; in pud_index() 43 static uint64_t pmd_index(struct kvm_vm *vm, vm_vaddr_t gva) in pud_index() 51 return (gva >> shift) & mask; in pmd_index() 54 static uint64_t pte_index(struct kvm_vm *vm, vm_vaddr_t gva) in pmd_index() 57 return (gva >> vm->page_shift) & mask; in pmd_index() 184 uint64_t *virt_get_pte_hva_at_level(struct kvm_vm *vm, vm_vaddr_t gva, int level) in virt_arch_pg_map() 191 ptep = addr_gpa2hva(vm, vm->mmu.pgd) + pgd_index(vm, gva) * in virt_get_pte_hva_at_level() 37 pud_index(struct kvm_vm * vm,vm_vaddr_t gva) pud_index() argument 48 pmd_index(struct kvm_vm * vm,vm_vaddr_t gva) pmd_index() argument 59 pte_index(struct kvm_vm * vm,vm_vaddr_t gva) pte_index() argument 189 virt_get_pte_hva_at_level(struct kvm_vm * vm,vm_vaddr_t gva,int level) virt_get_pte_hva_at_level() argument 233 virt_get_pte_hva(struct kvm_vm * vm,vm_vaddr_t gva) virt_get_pte_hva() argument 238 addr_arch_gva2gpa(struct kvm_vm * vm,vm_vaddr_t gva) addr_arch_gva2gpa() argument [all...] |
| /linux/drivers/infiniband/hw/mana/ |
| H A D | mr.c | 51 req.gva.dma_region_handle = mr_params->gva.dma_region_handle; in mana_ib_gd_create_mr() 52 req.gva.virtual_address = mr_params->gva.virtual_address; in mana_ib_gd_create_mr() 53 req.gva.access_flags = mr_params->gva.access_flags; in mana_ib_gd_create_mr() 172 mr_params.gva.dma_region_handle = dma_region_handle; in mana_ib_reg_user_mr() 173 mr_params.gva.virtual_address = iova; in mana_ib_reg_user_mr() 174 mr_params.gva.access_flags = in mana_ib_reg_user_mr() 247 mr_params.gva.dma_region_handle = dma_region_handle; in mana_ib_reg_user_mr_dmabuf() 248 mr_params.gva.virtual_address = iova; in mana_ib_reg_user_mr_dmabuf() 249 mr_params.gva.access_flags = in mana_ib_reg_user_mr_dmabuf()
|
| /linux/arch/riscv/include/asm/ |
| H A D | kvm_tlb.h | 43 unsigned long gva, 49 unsigned long gva, unsigned long gvsz, 71 unsigned long gva, unsigned long gvsz, 79 unsigned long gva, unsigned long gvsz,
|
| /linux/tools/testing/selftests/kvm/lib/loongarch/ |
| H A D | processor.c | 17 static uint64_t virt_pte_index(struct kvm_vm *vm, vm_vaddr_t gva, int level) in virt_pte_index() argument 24 return (gva >> shift) & mask; in virt_pte_index() 74 static uint64_t *virt_populate_pte(struct kvm_vm *vm, vm_vaddr_t gva, int alloc) in virt_populate_pte() argument 86 ptep = addr_gpa2hva(vm, child) + virt_pte_index(vm, gva, level) * 8; in virt_populate_pte() 100 ptep = addr_gpa2hva(vm, child) + virt_pte_index(vm, gva, level) * 8; in virt_populate_pte() 104 TEST_FAIL("No mapping for vm virtual address, gva: 0x%lx", gva); in virt_populate_pte() 108 vm_paddr_t addr_arch_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva) in addr_arch_gva2gpa() argument 112 ptep = virt_populate_pte(vm, gva, 0); in addr_arch_gva2gpa() 113 TEST_ASSERT(*ptep != 0, "Virtual address vaddr: 0x%lx not mapped\n", gva); in addr_arch_gva2gpa() [all...] |
| /linux/tools/testing/selftests/kvm/lib/ |
| H A D | memstress.c | 52 uint64_t gva; in memstress_guest_code() local 60 gva = vcpu_args->gva; in memstress_guest_code() 76 addr = gva + (page * args->guest_page_size); in memstress_guest_code() 104 vcpu_args->gva = guest_test_virt_mem + in memstress_setup_vcpus() 110 vcpu_args->gva = guest_test_virt_mem; in memstress_setup_vcpus()
|
| /linux/arch/x86/kvm/ |
| H A D | x86.h | 360 gva_t gva, gfn_t gfn, unsigned access) in vcpu_cache_mmio_info() argument 371 vcpu->arch.mmio_gva = mmu_is_nested(vcpu) ? 0 : gva & PAGE_MASK; in vcpu_cache_mmio_info() 388 static inline void vcpu_clear_mmio_info(struct kvm_vcpu *vcpu, gva_t gva) in vcpu_clear_mmio_info() argument 390 if (gva != MMIO_GVA_ANY && vcpu->arch.mmio_gva != (gva & PAGE_MASK)) in vcpu_clear_mmio_info() 396 static inline bool vcpu_match_mmio_gva(struct kvm_vcpu *vcpu, unsigned long gva) in vcpu_match_mmio_gva() argument 399 vcpu->arch.mmio_gva == (gva & PAGE_MASK)) in vcpu_match_mmio_gva() 470 void kvm_fixup_and_inject_pf_error(struct kvm_vcpu *vcpu, gva_t gva, u16 error_code); 656 int kvm_handle_invpcid(struct kvm_vcpu *vcpu, unsigned long type, gva_t gva);
|
| /linux/tools/testing/selftests/kvm/ |
| H A D | pre_fault_memory_test.c | 164 uint64_t gpa, gva, alignment, guest_page_size; in __test_pre_fault_memory() local 184 gva = gpa & ((1ULL << (vm->va_bits - 1)) - 1); in __test_pre_fault_memory() 188 virt_map(vm, gva, gpa, TEST_NPAGES); in __test_pre_fault_memory() 197 vcpu_args_set(vcpu, 1, gva); in __test_pre_fault_memory()
|
| H A D | access_tracking_perf_test.c | 126 static uint64_t lookup_pfn(int pagemap_fd, struct kvm_vm *vm, uint64_t gva) in lookup_pfn() 128 uint64_t hva = (uint64_t) addr_gva2hva(vm, gva); in lookup_pfn() 177 uint64_t base_gva = vcpu_args->gva; in pageidle_mark_vcpu_memory_idle() 196 uint64_t gva = base_gva + page * memstress_args.guest_page_size; in pageidle_mark_vcpu_memory_idle() 197 uint64_t pfn = lookup_pfn(pagemap_fd, vm, gva); in pageidle_mark_vcpu_memory_idle() 125 lookup_pfn(int pagemap_fd,struct kvm_vm * vm,uint64_t gva) lookup_pfn() argument 195 uint64_t gva = base_gva + page * memstress_args.guest_page_size; pageidle_mark_vcpu_memory_idle() local
|
| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | sev_smoke_test.c | 111 vm_vaddr_t gva; in test_sync_vmsa() local 118 gva = vm_vaddr_alloc_shared(vm, PAGE_SIZE, KVM_UTIL_MIN_VADDR, in test_sync_vmsa() 120 hva = addr_gva2hva(vm, gva); in test_sync_vmsa() 122 vcpu_args_set(vcpu, 1, gva); in test_sync_vmsa()
|
| H A D | cpuid_test.c | 146 vm_vaddr_t gva = vm_vaddr_alloc(vm, size, KVM_UTIL_MIN_VADDR); in vcpu_alloc_cpuid() local 147 struct kvm_cpuid2 *guest_cpuids = addr_gva2hva(vm, gva); in vcpu_alloc_cpuid() 151 *p_gva = gva; in vcpu_alloc_cpuid()
|
| H A D | hyperv_tlb_flush.c | 584 vm_vaddr_t test_data_page, gva; in main() local 620 gva = vm_vaddr_unused_gap(vm, NTEST_PAGES * PAGE_SIZE, KVM_UTIL_MIN_VADDR); in main() 624 virt_pg_map(vm, gva + PAGE_SIZE * i, gpa & PAGE_MASK); in main() 625 data->test_pages_pte[i] = gva + (gpa & ~PAGE_MASK); in main()
|
| /linux/arch/mips/kvm/ |
| H A D | tlb.c | 166 int kvm_vz_guest_tlb_lookup(struct kvm_vcpu *vcpu, unsigned long gva, in kvm_vz_guest_tlb_lookup() argument 184 write_gc0_entryhi((o_entryhi & 0x3ff) | (gva & ~0xfffl)); in kvm_vz_guest_tlb_lookup() 226 pa = entrylo[!!(gva & pagemaskbit)]; in kvm_vz_guest_tlb_lookup() 240 pa |= gva & ~(pagemask | pagemaskbit); in kvm_vz_guest_tlb_lookup()
|
| H A D | vz.c | 197 static gpa_t kvm_vz_gva_to_gpa_cb(gva_t gva) in kvm_vz_gva_to_gpa_cb() argument 200 return gva; in kvm_vz_gva_to_gpa_cb() 689 static int kvm_vz_gva_to_gpa(struct kvm_vcpu *vcpu, unsigned long gva, in kvm_vz_gva_to_gpa() argument 692 u32 gva32 = gva; in kvm_vz_gva_to_gpa() 695 if ((long)gva == (s32)gva32) { in kvm_vz_gva_to_gpa() 750 } else if ((gva & 0xc000000000000000) == 0x8000000000000000) { in kvm_vz_gva_to_gpa() 758 if (segctl & (1ull << (56 + ((gva >> 59) & 0x7)))) { in kvm_vz_gva_to_gpa() 772 *gpa = gva & 0x07ffffffffffffff; in kvm_vz_gva_to_gpa() 778 return kvm_vz_guest_tlb_lookup(vcpu, gva, gpa); in kvm_vz_gva_to_gpa()
|
| /linux/tools/testing/selftests/kvm/lib/riscv/ |
| H A D | processor.c | 55 static uint64_t pte_index(struct kvm_vm *vm, vm_vaddr_t gva, int level) 62 return (gva & pte_index_mask[level]) >> pte_index_shift[level]; in pte_index() 122 vm_paddr_t addr_arch_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva) in virt_arch_pg_map() 130 ptep = addr_gpa2hva(vm, vm->mmu.pgd) + pte_index(vm, gva, level) * 8; in addr_arch_gva2gpa() 137 pte_index(vm, gva, level) * 8; in addr_arch_gva2gpa() 143 return pte_addr(vm, *ptep) + (gva & (vm->page_size - 1)); in addr_arch_gva2gpa() 146 TEST_FAIL("No mapping for vm virtual address gva: 0x%lx level: %d", in addr_arch_gva2gpa() 147 gva, level); in addr_arch_gva2gpa() 59 pte_index(struct kvm_vm * vm,vm_vaddr_t gva,int level) pte_index() argument 126 addr_arch_gva2gpa(struct kvm_vm * vm,vm_vaddr_t gva) addr_arch_gva2gpa() argument
|
| /linux/tools/testing/selftests/kvm/arm64/ |
| H A D | sea_to_user.c | 203 run->arm_sea.gva, run->arm_sea.gpa); in run_vm() 216 TEST_ASSERT_EQ(run->arm_sea.gva, EINJ_GVA); in run_vm()
|
| /linux/tools/testing/selftests/kvm/include/ |
| H A D | memstress.h | 24 uint64_t gva; member
|
| H A D | kvm_util.h | 734 void *addr_gva2hva(struct kvm_vm *vm, vm_vaddr_t gva); 1239 * gva - VM virtual address 1247 * address given by @gva. 1249 vm_paddr_t addr_arch_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva); 1251 static inline vm_paddr_t addr_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva) in virt_dump() 1253 return addr_arch_gva2gpa(vm, gva); in virt_dump() 1228 addr_gva2gpa(struct kvm_vm * vm,vm_vaddr_t gva) addr_gva2gpa() argument
|
| /linux/tools/testing/selftests/kvm/include/arm64/ |
| H A D | processor.h | 182 uint64_t *virt_get_pte_hva_at_level(struct kvm_vm *vm, vm_vaddr_t gva, int level); in cpu_relax() 183 uint64_t *virt_get_pte_hva(struct kvm_vm *vm, vm_vaddr_t gva); in cpu_relax()
|
| /linux/arch/x86/include/asm/ |
| H A D | kvm_host.h | 1960 gva_t (*get_untagged_addr)(struct kvm_vcpu *vcpu, gva_t gva, unsigned int flags); 2292 gpa_t kvm_mmu_gva_to_gpa_read(struct kvm_vcpu *vcpu, gva_t gva, 2294 gpa_t kvm_mmu_gva_to_gpa_write(struct kvm_vcpu *vcpu, gva_t gva, 2296 gpa_t kvm_mmu_gva_to_gpa_system(struct kvm_vcpu *vcpu, gva_t gva, 2322 void kvm_mmu_invlpg(struct kvm_vcpu *vcpu, gva_t gva); 2325 void kvm_mmu_invpcid_gva(struct kvm_vcpu *vcpu, gva_t gva, unsigned long pcid);
|
| /linux/arch/arm64/kvm/ |
| H A D | nested.c | 22 u64 gva; member 966 va_start = vt->gva & ~(va_size - 1); in invalidate_vncr_va() 1328 vt->gva = va; in kvm_translate_vncr() 1365 if (read_vncr_el2(vcpu) != vt->gva) in kvm_vncr_tlb_lookup() 1469 if (read_vncr_el2(vcpu) != vt->gva) in kvm_map_l1_vncr()
|
| /linux/arch/x86/kvm/vmx/ |
| H A D | nested.c | 5346 gva_t gva; in nested_vmx_get_vmptr() local 5352 sizeof(*vmpointer), &gva)) { in nested_vmx_get_vmptr() 5357 r = kvm_read_guest_virt(vcpu, gva, vmpointer, sizeof(*vmpointer), &e); in nested_vmx_get_vmptr() 5641 gva_t gva = 0; in handle_vmread() local 5702 instr_info, true, len, &gva)) in handle_vmread() 5705 r = kvm_write_guest_virt_system(vcpu, gva, &value, len, &e); in handle_vmread() 5747 gva_t gva; in handle_vmwrite() local 5776 instr_info, false, len, &gva)) in handle_vmwrite() 5778 r = kvm_read_guest_virt(vcpu, gva, &value, len, &e); in handle_vmwrite() 5934 gva_t gva; in handle_vmptrst() local [all …]
|
| /linux/include/net/mana/ |
| H A D | gdma.h | 883 } gva; 908 } __packed gva; 852 } gva; global() member 871 } gva; global() member
|
| /linux/tools/testing/selftests/kvm/lib/x86/ |
| H A D | processor.c | 619 vm_paddr_t addr_arch_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva) in addr_arch_gva2gpa() argument 622 uint64_t *pte = __vm_get_page_table_entry(vm, &vm->mmu, gva, &level); in addr_arch_gva2gpa() 625 "Leaf PTE not PRESENT for gva: 0x%08lx", gva); in addr_arch_gva2gpa() 631 return vm_untag_gpa(vm, PTE_GET_PA(*pte)) | (gva & ~HUGEPAGE_MASK(level)); in addr_arch_gva2gpa()
|