Home
last modified time | relevance | path

Searched refs:vma_pagesize (Results 1 – 2 of 2) sorted by relevance

/linux/arch/riscv/kvm/
H A Dmmu.c438 unsigned long vma_pagesize, mmu_seq; in kvm_riscv_mmu_map() local
467 vma_pagesize = 1ULL << vma_pageshift; in kvm_riscv_mmu_map()
469 vma_pagesize = PAGE_SIZE; in kvm_riscv_mmu_map()
471 if (vma_pagesize == PMD_SIZE || vma_pagesize == PUD_SIZE) in kvm_riscv_mmu_map()
485 if (vma_pagesize != PUD_SIZE && in kvm_riscv_mmu_map()
486 vma_pagesize != PMD_SIZE && in kvm_riscv_mmu_map()
487 vma_pagesize != PAGE_SIZE) { in kvm_riscv_mmu_map()
488 kvm_err("Invalid VMA page size 0x%lx\n", vma_pagesize); in kvm_riscv_mmu_map()
515 if (!logging && (vma_pagesize == PAGE_SIZE)) in kvm_riscv_mmu_map()
516 vma_pagesize = transparent_hugepage_adjust(kvm, memslot, hva, &hfn, &gpa); in kvm_riscv_mmu_map()
[all …]
/linux/arch/arm64/kvm/
H A Dmmu.c1663 long vma_pagesize; member
1830 s2vi->vma_pagesize = BIT(kvm_s2_resolve_vma_size(s2fd, s2vi, vma)); in kvm_s2_fault_get_vma_info()
1837 s2vi->gfn = ALIGN_DOWN(s2fd->fault_ipa, s2vi->vma_pagesize) >> PAGE_SHIFT; in kvm_s2_fault_get_vma_info()
1868 return ALIGN_DOWN(ipa, s2vi->vma_pagesize) >> PAGE_SHIFT; in get_canonical_gfn()
1885 kvm_send_hwpoison_signal(s2fd->hva, __ffs(s2vi->vma_pagesize)); in kvm_s2_fault_pin_pfn()
2007 mapping_size = s2vi->vma_pagesize; in kvm_s2_fault_map()