Lines Matching refs:PG_LEVEL_4K

770 	if (sp->role.level > PG_LEVEL_4K)  in account_shadowed()
775 if (kvm_mmu_slot_gfn_write_protect(kvm, slot, gfn, PG_LEVEL_4K)) in account_shadowed()
776 kvm_flush_remote_tlbs_gfn(kvm, gfn, PG_LEVEL_4K); in account_shadowed()
816 if (sp->role.level > PG_LEVEL_4K) in unaccount_shadowed()
1177 return &slot->arch.rmap[level - PG_LEVEL_4K][idx]; in gfn_to_rmap()
1291 WARN_ON_ONCE(sp->role.level == PG_LEVEL_4K); in drop_large_spte()
1388 PG_LEVEL_4K, slot); in kvm_mmu_write_protect_pt_masked()
1411 PG_LEVEL_4K, slot); in kvm_mmu_clear_dirty_pt_masked()
1439 kvm_mmu_try_split_huge_pages(kvm, slot, start, end + 1, PG_LEVEL_4K); in kvm_arch_mmu_enable_log_dirty_pt_masked()
1498 return kvm_mmu_slot_gfn_write_protect(vcpu->kvm, slot, gfn, PG_LEVEL_4K); in kvm_vcpu_write_protect_gfn()
1630 return walk_slot_rmaps(kvm, slot, fn, PG_LEVEL_4K, PG_LEVEL_4K, flush_on_yield); in walk_slot_rmaps_4k()
1639 PG_LEVEL_4K, KVM_MAX_HUGEPAGE_LEVEL, in __kvm_rmap_zap_gfn_range()
1720 for (level = PG_LEVEL_4K; level <= KVM_MAX_HUGEPAGE_LEVEL; level++) { in kvm_rmap_age_gfn_range()
2143 if (level == PG_LEVEL_4K) in mmu_pages_next()
2165 WARN_ON_ONCE(level == PG_LEVEL_4K); in mmu_pages_first()
2278 if (role.level > PG_LEVEL_4K && sp->unsync) in kvm_mmu_find_shadow_page()
2452 WARN_ON_ONCE(role.level != PG_LEVEL_4K); in kvm_mmu_child_role()
2510 if (iterator->level < PG_LEVEL_4K) in shadow_walk_okay()
2660 if (parent->role.level == PG_LEVEL_4K) in mmu_zap_unsync_children()
2978 WARN_ON_ONCE(sp->role.level != PG_LEVEL_4K); in mmu_try_to_unsync_pages()
3058 if (level > PG_LEVEL_4K && !is_large_pte(*sptep)) { in mmu_set_spte()
3186 if (sp->role.level > PG_LEVEL_4K) in direct_pte_prefetch()
3227 int level = PG_LEVEL_4K; in host_pfn_mapping_level()
3296 for ( ; max_level > PG_LEVEL_4K; max_level--) { in __kvm_mmu_max_mapping_level()
3305 if (max_level == PG_LEVEL_4K) in __kvm_mmu_max_mapping_level()
3306 return PG_LEVEL_4K; in __kvm_mmu_max_mapping_level()
3328 if (unlikely(fault->max_level == PG_LEVEL_4K)) in kvm_mmu_hugepage_adjust()
3344 if (fault->req_level == PG_LEVEL_4K || fault->huge_page_disallowed) in kvm_mmu_hugepage_adjust()
3359 if (cur_level > PG_LEVEL_4K && in disallowed_hugepage_adjust()
3689 if (sp->role.level > PG_LEVEL_4K && in fast_page_fault()
4512 order != KVM_HPAGE_GFN_SHIFT(PG_LEVEL_4K)); in kvm_max_level_for_order()
4520 return PG_LEVEL_4K; in kvm_max_level_for_order()
4528 if (max_level == PG_LEVEL_4K) in kvm_max_private_mapping_level()
4529 return PG_LEVEL_4K; in kvm_max_private_mapping_level()
4532 if (max_level == PG_LEVEL_4K) in kvm_max_private_mapping_level()
4533 return PG_LEVEL_4K; in kvm_max_private_mapping_level()
4951 u8 level = PG_LEVEL_4K; in kvm_arch_vcpu_pre_fault_memory()
6076 if (sp->role.level == PG_LEVEL_4K) in detect_write_flooding()
6182 if (gentry && sp->role.level != PG_LEVEL_4K) in kvm_mmu_track_write()
7192 PG_LEVEL_4K, KVM_MAX_HUGEPAGE_LEVEL - 1, true)) in kvm_rmap_zap_collapsible_sptes()