| /linux/virt/kvm/ |
| H A D | pfncache.c | 75 struct kvm_memslots *slots = kvm_memslots(gpc->kvm); in kvm_gpc_check() 293 struct kvm_memslots *slots = kvm_memslots(gpc->kvm); in __kvm_gpc_refresh()
|
| H A D | kvm_main.c | 569 struct kvm_memslots *slots; in kvm_handle_hva_range() 947 static void kvm_free_memslots(struct kvm *kvm, struct kvm_memslots *slots) in kvm_free_memslots() 1101 struct kvm_memslots *slots; in kvm_create_vm() 1432 static struct kvm_memslots *kvm_get_inactive_memslots(struct kvm *kvm, int as_id) in kvm_get_inactive_memslots() 1434 struct kvm_memslots *active = __kvm_memslots(kvm, as_id); in kvm_get_inactive_memslots() 1460 static void kvm_insert_gfn_node(struct kvm_memslots *slots, in kvm_insert_gfn_node() 1485 static void kvm_erase_gfn_node(struct kvm_memslots *slots, in kvm_erase_gfn_node() 1491 static void kvm_replace_gfn_node(struct kvm_memslots *slots, in kvm_replace_gfn_node() 1517 struct kvm_memslots *slots = kvm_get_inactive_memslots(kvm, as_id); in kvm_replace_memslot() 1602 struct kvm_memslots *slots = kvm_get_inactive_memslots(kvm, as_id); in kvm_swap_active_memslots() [all …]
|
| /linux/include/linux/ |
| H A D | kvm_host.h | 753 struct kvm_memslots { struct 790 struct kvm_memslots __memslots[KVM_MAX_NR_ADDRESS_SPACES][2]; argument 792 struct kvm_memslots __rcu *memslots[KVM_MAX_NR_ADDRESS_SPACES]; 1079 static inline struct kvm_memslots *__kvm_memslots(struct kvm *kvm, int as_id) in __kvm_memslots() 1087 static inline struct kvm_memslots *kvm_memslots(struct kvm *kvm) in kvm_memslots() function 1092 static inline struct kvm_memslots *kvm_vcpu_memslots(struct kvm_vcpu *vcpu) in kvm_vcpu_memslots() 1099 static inline bool kvm_memslots_empty(struct kvm_memslots *slots) in kvm_memslots_empty() 1112 struct kvm_memory_slot *id_to_memslot(struct kvm_memslots *slots, int id) in id_to_memslot() 1127 struct kvm_memslots *slots; 1142 struct kvm_memslots *slots, in kvm_memslot_iter_start() [all …]
|
| /linux/arch/powerpc/kvm/ |
| H A D | book3s_hv_uvmem.c | 467 struct kvm_memslots *slots; in kvmppc_h_svm_init_start() 488 slots = kvm_memslots(kvm); in kvmppc_h_svm_init_start() 496 slots = kvm_memslots(kvm); in kvmppc_h_svm_init_start() 676 kvm_for_each_memslot(memslot, bkt, kvm_memslots(kvm)) in kvmppc_h_svm_init_abort() 832 struct kvm_memslots *slots; in kvmppc_h_svm_init_done() 842 slots = kvm_memslots(kvm); in kvmppc_h_svm_init_done()
|
| H A D | book3s_64_mmu_hv.c | 741 struct kvm_memslots *slots; in kvmppc_rmap_reset() 746 slots = kvm_memslots(kvm); in kvmppc_rmap_reset() 1250 __gfn_to_memslot(kvm_memslots(kvm), gfn); in resize_hpt_rehash_hpte()
|
| H A D | book3s_64_vio.c | 358 memslot = __gfn_to_memslot(kvm_memslots(kvm), gfn); in kvmppc_tce_to_ua()
|
| H A D | book3s_hv_nested.c | 825 kvm_for_each_memslot(memslot, bkt, kvm_memslots(kvm)) in kvmhv_release_all_nested()
|
| /linux/arch/arm64/kvm/ |
| H A D | mmu.c | 372 struct kvm_memslots *slots; in stage2_flush_vm() 379 slots = kvm_memslots(kvm); in stage2_flush_vm() 1079 struct kvm_memslots *slots; in stage2_unmap_vm() 1087 slots = kvm_memslots(kvm); in stage2_unmap_vm() 1242 struct kvm_memslots *slots = kvm_memslots(kvm); in kvm_mmu_wp_memory_region() 1270 struct kvm_memslots *slots; in kvm_mmu_split_memory_region() 1276 slots = kvm_memslots(kvm); in kvm_mmu_split_memory_region()
|
| /linux/arch/riscv/kvm/ |
| H A D | mmu.c | 21 struct kvm_memslots *slots = kvm_memslots(kvm); in mmu_wp_memory_region()
|
| /linux/arch/x86/kvm/ |
| H A D | debugfs.c | 94 struct kvm_memslots *slots; in kvm_mmu_rmaps_stat_show()
|
| H A D | x86.c | 3752 struct kvm_memslots *slots; in record_steal_time() 3768 slots = kvm_memslots(vcpu->kvm); in record_steal_time() 5244 struct kvm_memslots *slots; in kvm_steal_time_set_preempted() 5271 slots = kvm_memslots(vcpu->kvm); in kvm_steal_time_set_preempted() 13434 struct kvm_memslots *slots = kvm_memslots(kvm); in __x86_set_memory_region()
|
| /linux/arch/x86/kvm/mmu/ |
| H A D | page_track.c | 171 struct kvm_memslots *slots; in kvm_enable_external_write_tracking()
|
| H A D | mmu.c | 753 struct kvm_memslots *slots; in account_shadowed() 812 struct kvm_memslots *slots; in unaccount_shadowed() 1188 struct kvm_memslots *slots; in rmap_remove() 4015 struct kvm_memslots *slots; in mmu_first_shadow_root_alloc() 6912 struct kvm_memslots *slots; in kvm_rmap_zap_gfn_range()
|
| /linux/arch/loongarch/kvm/ |
| H A D | vcpu.c | 157 struct kvm_memslots *slots; in kvm_update_stolen_time() 167 slots = kvm_memslots(vcpu->kvm); in kvm_update_stolen_time() 1813 struct kvm_memslots *slots; in kvm_vcpu_set_pv_preempted() 1830 slots = kvm_memslots(vcpu->kvm); in kvm_vcpu_set_pv_preempted()
|
| /linux/Documentation/virt/kvm/x86/ |
| H A D | mmu.rst | 197 determines which of the kvm_memslots array was used to build this 476 kvm_memslots(kvm)->generation, and increased whenever guest memory info 487 Unfortunately, a single memory access might access kvm_memslots(kvm) multiple 493 returns; thus, bit 63 of kvm_memslots(kvm)->generation set to 1 only during a
|
| /linux/arch/s390/kvm/ |
| H A D | pv.c | 345 npages = kvm_s390_get_gfn_end(kvm_memslots(kvm)); in kvm_s390_pv_alloc_vm()
|
| H A D | kvm-s390.c | 1172 struct kvm_memslots *slots; in kvm_s390_vm_start_migration() 1179 slots = kvm_memslots(kvm); in kvm_s390_vm_start_migration()
|
| /linux/arch/x86/kvm/vmx/ |
| H A D | vmx.c | 6934 struct kvm_memslots *slots = kvm_memslots(kvm); in vmx_set_apic_access_page_addr()
|