| /linux/arch/x86/include/asm/ |
| H A D | kfence.h | 33 if (level != PG_LEVEL_4K) in arch_kfence_init_pool() 47 if (WARN_ON(!pte || level != PG_LEVEL_4K)) in kfence_protect_page()
|
| H A D | kvm_host.h | 157 #define KVM_NR_PAGE_SIZES (KVM_MAX_HUGEPAGE_LEVEL - PG_LEVEL_4K + 1)
|
| /linux/arch/x86/coco/tdx/ |
| H A D | tdx-shared.c | 24 case PG_LEVEL_4K: in try_accept_one() 64 accept_size = try_accept_one(start, len, PG_LEVEL_4K); in tdx_accept_memory()
|
| /linux/arch/x86/kvm/mmu/ |
| H A D | spte.c | 226 if (level > PG_LEVEL_4K && (pte_access & ACC_EXEC_MASK) && in make_spte() 239 if (level > PG_LEVEL_4K) in make_spte() 291 WARN_ON_ONCE(level > PG_LEVEL_4K); in make_spte() 351 if (role.level == PG_LEVEL_4K) { in make_small_spte() 370 KVM_BUG_ON(!is_shadow_present_pte(small_spte) || level == PG_LEVEL_4K, kvm); in make_huge_spte()
|
| H A D | paging_tmpl.h | 74 #define gpte_to_gfn(pte) gpte_to_gfn_lvl((pte), PG_LEVEL_4K) 163 if (FNAME(is_rsvd_bits_set)(vcpu->arch.mmu, gpte, PG_LEVEL_4K)) in FNAME() 295 gpte |= level - PG_LEVEL_4K - 1; in FNAME() 444 if (walker->level > PG_LEVEL_4K && is_cpuid_PSE36()) in FNAME() 556 if (level == PG_LEVEL_4K) { in FNAME() 581 if (sp->role.level > PG_LEVEL_4K) in FNAME() 851 WARN_ON_ONCE(sp->role.level != PG_LEVEL_4K); in FNAME()
|
| H A D | page_track.c | 80 index = gfn_to_index(gfn, slot->base_gfn, PG_LEVEL_4K); in update_gfn_write_track() 109 if (kvm_mmu_slot_gfn_write_protect(kvm, slot, gfn, PG_LEVEL_4K)) in __kvm_write_track_add_gfn() 147 index = gfn_to_index(gfn, slot->base_gfn, PG_LEVEL_4K); in kvm_gfn_is_write_tracked()
|
| H A D | mmu.c | 772 if (sp->role.level > PG_LEVEL_4K) in account_shadowed() 777 if (kvm_mmu_slot_gfn_write_protect(kvm, slot, gfn, PG_LEVEL_4K)) in account_shadowed() 778 kvm_flush_remote_tlbs_gfn(kvm, gfn, PG_LEVEL_4K); in account_shadowed() 820 if (sp->role.level > PG_LEVEL_4K) in unaccount_shadowed() 1183 return &slot->arch.rmap[level - PG_LEVEL_4K][idx]; in gfn_to_rmap() 1381 PG_LEVEL_4K, slot); in kvm_mmu_write_protect_pt_masked() 1404 PG_LEVEL_4K, slot); in kvm_mmu_clear_dirty_pt_masked() 1432 kvm_mmu_try_split_huge_pages(kvm, slot, start, end + 1, PG_LEVEL_4K); in kvm_arch_mmu_enable_log_dirty_pt_masked() 1491 return kvm_mmu_slot_gfn_write_protect(vcpu->kvm, slot, gfn, PG_LEVEL_4K); in kvm_vcpu_write_protect_gfn() 1623 return walk_slot_rmaps(kvm, slot, fn, PG_LEVEL_4K, PG_LEVEL_4K, flush_on_yield); in walk_slot_rmaps_4k() [all …]
|
| H A D | mmu_internal.h | 358 .req_level = PG_LEVEL_4K, in kvm_mmu_do_page_fault() 359 .goal_level = PG_LEVEL_4K, in kvm_mmu_do_page_fault()
|
| /linux/tools/testing/selftests/kvm/lib/x86/ |
| H A D | processor.c | 292 current_level > PG_LEVEL_4K; in __virt_pg_map() 301 pte = virt_get_pte(vm, mmu, pte, gva, PG_LEVEL_4K); in __virt_pg_map() 320 __virt_pg_map(vm, &vm->mmu, gva, gpa, PG_LEVEL_4K); in virt_arch_pg_map() 383 current_level > PG_LEVEL_4K; in __vm_get_page_table_entry() 390 return virt_get_pte(vm, mmu, pte, gva, PG_LEVEL_4K); in __vm_get_page_table_entry() 395 int level = PG_LEVEL_4K; in tdp_get_pte() 402 int level = PG_LEVEL_4K; in vm_get_pte() 515 __tdp_map(vm, l2_gpa, gpa, size, PG_LEVEL_4K); in tdp_map()
|
| /linux/arch/x86/xen/ |
| H A D | p2m.c | 240 BUG_ON(!ptep || level != PG_LEVEL_4K); in xen_build_mfn_list_list() 438 BUG_ON(!ptep || level != PG_LEVEL_4K); in get_phys_to_machine() 532 BUG_ON(!ptep || level != PG_LEVEL_4K); in xen_alloc_p2m_entry() 656 BUG_ON(!ptep || level != PG_LEVEL_4K); in __set_phys_to_machine()
|
| /linux/arch/x86/mm/ |
| H A D | mem_encrypt_amd.c | 223 case PG_LEVEL_4K: in pg_level_to_pfn() 412 if (level == PG_LEVEL_4K) { in early_set_memory_enc_dec()
|
| H A D | pti.c | 300 if (!pte || WARN_ON(level != PG_LEVEL_4K) || pte_none(*pte)) in pti_setup_vsyscall()
|
| H A D | kmmio.c | 174 case PG_LEVEL_4K: in clear_page_presence()
|
| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | nx_huge_pages_test.c | 145 virt_map_level(vm, HPAGE_GVA, HPAGE_GPA, nr_bytes, PG_LEVEL_4K); in run_test()
|
| /linux/arch/x86/kvm/vmx/ |
| H A D | capabilities.h | 340 return PG_LEVEL_4K; in ept_caps_to_lpage_level()
|
| /linux/drivers/crypto/ccp/ |
| H A D | sev-dev-tio.c | 477 PG_LEVEL_4K, 0, true)) in sla_alloc() 481 if (rmp_make_private(sla_to_pfn(ret), 0, PG_LEVEL_4K, 0, true)) in sla_alloc()
|
| H A D | sev-dev.c | 403 ret = rmp_make_shared(__phys_to_pfn(paddr), PG_LEVEL_4K); in snp_reclaim_pages() 426 rc = rmp_make_private(pfn, 0, PG_LEVEL_4K, 0, true); in rmp_mark_pages_firmware()
|
| /linux/arch/x86/mm/pat/ |
| H A D | set_memory.c | 124 direct_pages_count[PG_LEVEL_4K] << 2); in arch_report_meminfo() 581 if (lookup_address(start, &level) && (level != PG_LEVEL_4K)) in protect_kernel_text_ro() 765 *level = PG_LEVEL_4K; in lookup_address_in_pgd_attr() 1866 if (level == PG_LEVEL_4K) { in __change_page_attr()
|
| /linux/tools/testing/selftests/kvm/include/x86/ |
| H A D | processor.h | 1462 PG_LEVEL_4K, enumerator 1472 #define PG_SIZE_4K PG_LEVEL_SIZE(PG_LEVEL_4K)
|
| /linux/drivers/iommu/amd/ |
| H A D | init.c | 4052 if (level > PG_LEVEL_4K) { in iommu_page_make_shared() 4064 return rmp_make_shared(pfn, PG_LEVEL_4K); in iommu_page_make_shared()
|
| /linux/arch/x86/kvm/ |
| H A D | x86.c | 13768 kvm_mmu_slot_try_split_huge_pages(kvm, new, PG_LEVEL_4K); in kvm_mmu_slot_apply_flags() 13774 kvm_mmu_slot_remove_write_access(kvm, new, PG_LEVEL_4K); in kvm_mmu_slot_apply_flags()
|