| /linux/arch/powerpc/kvm/ |
| H A D | book3s_hv_rm_mmu.c | 69 unsigned long *rmap, long pte_index, int realmode) in kvmppc_add_revmap_chain() argument 84 tail->forw = pte_index; in kvmppc_add_revmap_chain() 85 head->back = pte_index; in kvmppc_add_revmap_chain() 87 rev->forw = rev->back = pte_index; in kvmppc_add_revmap_chain() 89 pte_index | KVMPPC_RMAP_PRESENT | KVMPPC_RMAP_HPT; in kvmppc_add_revmap_chain() 147 static void remove_revmap_chain(struct kvm *kvm, long pte_index, in remove_revmap_chain() argument 170 if (head == pte_index) { in remove_revmap_chain() 172 if (head == pte_index) in remove_revmap_chain() 185 long pte_index, unsigned long pteh, unsigned long ptel, in kvmppc_do_h_enter() argument 303 if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) in kvmppc_do_h_enter() [all …]
|
| H A D | book3s_pr_papr.c | 23 static unsigned long get_pteg_addr(struct kvm_vcpu *vcpu, long pte_index) in get_pteg_addr() argument 28 pte_index <<= 4; in get_pteg_addr() 29 pte_index &= ((1 << ((vcpu_book3s->sdr1 & 0x1f) + 11)) - 1) << 7 | 0x70; in get_pteg_addr() 31 pteg_addr |= pte_index; in get_pteg_addr() 39 long pte_index = kvmppc_get_gpr(vcpu, 5); in kvmppc_h_pr_enter() local 45 i = pte_index & 7; in kvmppc_h_pr_enter() 46 pte_index &= ~7UL; in kvmppc_h_pr_enter() 47 pteg_addr = get_pteg_addr(vcpu, pte_index); in kvmppc_h_pr_enter() 76 kvmppc_set_gpr(vcpu, 4, pte_index | i); in kvmppc_h_pr_enter() 89 unsigned long pte_index = kvmppc_get_gpr(vcpu, 5); in kvmppc_h_pr_remove() local [all …]
|
| H A D | book3s_64_mmu_hv.c | 48 long pte_index, unsigned long pteh, 293 long pte_index, unsigned long pteh, in kvmppc_virtmode_do_h_enter() argument 299 ret = kvmppc_do_h_enter(kvm, flags, pte_index, pteh, ptel, in kvmppc_virtmode_do_h_enter()
|
| /linux/tools/testing/selftests/kvm/lib/riscv/ |
| H A D | processor.c | 55 static uint64_t pte_index(struct kvm_vm *vm, vm_vaddr_t gva, int level) 97 ptep = addr_gpa2hva(vm, vm->mmu.pgd) + pte_index(vm, vaddr, level) * 8; in virt_arch_pg_map() 107 pte_index(vm, vaddr, level) * 8; in virt_arch_pg_map() 130 ptep = addr_gpa2hva(vm, vm->mmu.pgd) + pte_index(vm, gva, level) * 8; in addr_arch_gva2gpa() 137 pte_index(vm, gva, level) * 8; in addr_arch_gva2gpa() 59 static uint64_t pte_index(struct kvm_vm *vm, vm_vaddr_t gva, int level) pte_index() function
|
| /linux/drivers/iommu/ |
| H A D | rockchip-iommu.c | 538 u32 dte_index, pte_index, page_offset; in log_iova() local 550 pte_index = rk_iova_pte_index(iova); in log_iova() 563 pte_addr_phys = rk_ops->pt_address(dte) + (pte_index * 4); in log_iova() 575 &iova, dte_index, pte_index, page_offset); in log_iova() 830 u32 dte_index, pte_index; in rk_iommu_map() local 849 pte_index = rk_iova_pte_index(iova); in rk_iommu_map() 850 pte_addr = &page_table[pte_index]; in rk_iommu_map() 852 pte_dma = rk_ops->pt_address(dte_index) + pte_index * sizeof(u32); in rk_iommu_map()
|
| H A D | sun50i-iommu.c | 601 u32 pte_index; in sun50i_iommu_map() local 619 pte_index = sun50i_iova_get_pte_index(iova); in sun50i_iommu_map() 620 pte_addr = &page_table[pte_index]; in sun50i_iommu_map()
|
| /linux/arch/powerpc/include/asm/ |
| H A D | kvm_ppc.h | 830 long pte_index, unsigned long pteh, unsigned long ptel); 832 unsigned long pte_index, unsigned long avpn); 835 unsigned long pte_index, unsigned long avpn); 837 unsigned long pte_index); 839 unsigned long pte_index); 841 unsigned long pte_index);
|
| /linux/arch/x86/power/ |
| H A D | hibernate_32.c | 164 set_pte(pte + pte_index(restore_jump_address), in set_up_temporary_text_mapping()
|
| /linux/tools/testing/selftests/kvm/lib/arm64/ |
| H A D | processor.c | 54 static uint64_t pte_index(struct kvm_vm *vm, vm_vaddr_t gva) in pmd_index() 164 ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pte_index(vm, vaddr) * 8; in _virt_pg_map() 213 ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pte_index(vm, gva) * 8; in virt_get_pte_hva_at_level() 59 static uint64_t pte_index(struct kvm_vm *vm, vm_vaddr_t gva) pte_index() function
|
| /linux/arch/x86/boot/compressed/ |
| H A D | ident_map_64.c | 229 return pte + pte_index(__address); in split_large_pmd()
|
| /linux/include/linux/ |
| H A D | pgtable.h | 48 static inline unsigned long pte_index(unsigned long address) in pte_index() function 98 return (pte_t *)pmd_page_vaddr(*pmd) + pte_index(address); in pte_offset_kernel() 105 ((pte_t *)kmap_local_page(pmd_page(*(pmd))) + pte_index((address)))
|
| /linux/arch/x86/hyperv/ |
| H A D | hv_crash.c | 528 pte = hv_crash_ptpgs[3] + pte_index(addr) * sizeof(pte); in hv_crash_build_tramp_pt()
|
| /linux/arch/arm64/mm/ |
| H A D | mmu.c | 220 ptep += pte_index(addr); in alloc_init_cont_pte() 239 ptep += pte_index(next) - pte_index(addr); in alloc_init_cont_pte()
|
| /linux/arch/x86/mm/ |
| H A D | ioremap.c | 840 return &bm_pte[pte_index(addr)]; in early_ioremap_pte()
|
| /linux/mm/ |
| H A D | hugetlb_vmemmap.c | 126 struct page *page = head ? head + pte_index(addr) : in vmemmap_pmd_entry()
|
| H A D | memory.c | 2435 remaining_pages_total, PTRS_PER_PTE - pte_index(addr)); in insert_pages() 4654 if (can_swapin_thp(vmf, pte + pte_index(addr), 1 << order)) in alloc_swap_folio() 5169 if (pte_range_none(pte + pte_index(addr), 1 << order)) in alloc_anon_folio() 5628 pgoff_t pte_off = pte_index(vmf->address); in finish_fault() 5736 pgoff_t pte_off = pte_index(vmf->address); in do_fault_around()
|
| H A D | vmscan.c | 3527 for (i = pte_index(start), addr = start; addr != end; i++, addr += PAGE_SIZE) { in walk_pte_range()
|
| /linux/arch/arm64/include/asm/ |
| H A D | pgtable.h | 855 #define pte_offset_phys(dir,addr) (pmd_page_paddr(READ_ONCE(*(dir))) + pte_index(addr) * sizeof(pte…
|
| /linux/arch/x86/xen/ |
| H A D | mmu_pv.c | 1940 pte = native_make_pte(xen_read_phys_ulong(pa + pte_index(vaddr) * in xen_early_virt_to_phys()
|