| /linux/arch/x86/kvm/ |
| H A D | smm.c | 204 kvm_x86_call(get_gdt)(vcpu, &dt); in enter_smm_save_state_32() 208 kvm_x86_call(get_idt)(vcpu, &dt); in enter_smm_save_state_32() 224 smram->int_shadow = kvm_x86_call(get_interrupt_shadow)(vcpu); in enter_smm_save_state_32() 254 kvm_x86_call(get_idt)(vcpu, &dt); in enter_smm_save_state_64() 260 kvm_x86_call(get_gdt)(vcpu, &dt); in enter_smm_save_state_64() 271 smram->int_shadow = kvm_x86_call(get_interrupt_shadow)(vcpu); in enter_smm_save_state_64() 305 if (kvm_x86_call(enter_smm)(vcpu, &smram)) in enter_smm() 313 if (kvm_x86_call(get_nmi_mask)(vcpu)) in enter_smm() 316 kvm_x86_call(set_nmi_mask)(vcpu, true); in enter_smm() 321 kvm_x86_call(set_interrupt_shadow)(vcpu, 0); in enter_smm() [all …]
|
| H A D | kvm_cache_regs.h | 119 kvm_x86_call(cache_reg)(vcpu, reg); in kvm_register_read_raw() 159 kvm_x86_call(cache_reg)(vcpu, VCPU_EXREG_PDPTR); in kvm_pdptr_read() 174 kvm_x86_call(cache_reg)(vcpu, VCPU_EXREG_CR0); in kvm_read_cr0_bits() 196 kvm_x86_call(cache_reg)(vcpu, VCPU_EXREG_CR4); in kvm_read_cr4_bits() 211 kvm_x86_call(cache_reg)(vcpu, VCPU_EXREG_CR3); in kvm_read_cr3()
|
| H A D | x86.c | 1028 if (kvm_x86_call(get_cpl)(vcpu) <= required_cpl) in kvm_require_cpl() 1119 return kvm_x86_call(is_valid_cr0)(vcpu, cr0); in kvm_is_valid_cr0() 1183 kvm_x86_call(get_cs_db_l_bits)(vcpu, &cs_db, &cs_l); in kvm_set_cr0() 1200 kvm_x86_call(set_cr0)(vcpu, cr0); in kvm_set_cr0() 1313 if (kvm_x86_call(get_cpl)(vcpu) != 0 || in kvm_emulate_xsetbv() 1326 kvm_x86_call(is_valid_cr4)(vcpu, cr4); in kvm_is_valid_cr4() 1397 kvm_x86_call(set_cr4)(vcpu, cr4); in kvm_set_cr4() 1536 kvm_x86_call(set_dr7)(vcpu, dr7); in kvm_update_dr7() 1715 return kvm_x86_call(get_feature_msr)(index, data); in kvm_get_feature_msr() 1777 r = kvm_x86_call(set_efer)(vcpu, efer); in set_efer() [all …]
|
| H A D | irq.c | 118 return kvm_x86_call(protected_apic_has_interrupt)(v); in kvm_cpu_has_interrupt() 183 kvm_x86_call(migrate_timers)(vcpu); in __kvm_migrate_timers() 461 r = kvm_x86_call(pi_update_irte)(irqfd, irqfd->kvm, host_irq, irqfd->gsi, in kvm_pi_update_irte() 487 kvm_x86_call(pi_start_bypass)(kvm); in kvm_arch_irq_bypass_add_producer()
|
| H A D | mmu.h | 155 kvm_x86_call(load_mmu_pgd)(vcpu, root_hpa, in kvm_mmu_load_pgd() 191 unsigned long rflags = kvm_x86_call(get_rflags)(vcpu); in permission_fault()
|
| H A D | lapic.c | 765 kvm_x86_call(hwapic_isr_update)(apic->vcpu, vec); in apic_set_isr() 806 kvm_x86_call(hwapic_isr_update)(apic->vcpu, apic_find_highest_isr(apic)); in apic_clear_isr() 944 highest_irr = kvm_x86_call(sync_pir_to_irr)(apic->vcpu); in apic_has_interrupt_for_ppr() 1433 kvm_x86_call(deliver_interrupt)(apic, delivery_mode, in __apic_accept_irq() 2245 kvm_x86_call(cancel_hv_timer)(apic->vcpu); in cancel_hv_timer() 2262 if (kvm_x86_call(set_hv_timer)(vcpu, ktimer->tscdeadline, &expired)) in start_hv_timer() 2777 kvm_x86_call(set_virtual_apic_mode)(vcpu); in __kvm_apic_set_base() 2843 kvm_x86_call(hwapic_isr_update)(vcpu, apic_find_highest_isr(apic)); in kvm_apic_update_apicv() 2913 kvm_x86_call(apicv_pre_state_restore)(vcpu); in kvm_lapic_reset() 2975 kvm_x86_call(apicv_post_state_restore)(vcpu); in kvm_lapic_reset() [all …]
|
| H A D | smm.h | 145 if (!kvm_x86_call(has_emulated_msr)(vcpu->kvm, MSR_IA32_SMBASE)) in kvm_inject_smi()
|
| H A D | x86.h | 258 kvm_x86_call(get_cs_db_l_bits)(vcpu, &cs_db, &cs_l); in is_64_bit_mode()
|
| H A D | hyperv.c | 1421 kvm_x86_call(patch_hypercall)(vcpu, instructions + i); in kvm_hv_set_msr_pw() 1993 kvm_x86_call(flush_tlb_gva)(vcpu, gva + j * PAGE_SIZE); in kvm_hv_vcpu_flush_tlb() 2538 if (kvm_x86_call(get_cpl)(vcpu) != 0 || !is_protmode(vcpu)) { in kvm_hv_hypercall()
|
| H A D | xen.c | 1318 kvm_x86_call(patch_hypercall)(vcpu, instructions + 5); in kvm_xen_write_hypercall_page() 1707 cpl = kvm_x86_call(get_cpl)(vcpu); in kvm_xen_hypercall()
|
| H A D | cpuid.c | 460 kvm_x86_call(vcpu_after_set_cpuid)(vcpu); in kvm_vcpu_after_set_cpuid()
|
| /linux/arch/x86/include/asm/ |
| H A D | kvm_host.h | 2012 #define kvm_x86_call(func) static_call(kvm_x86_##func) macro 2038 !kvm_x86_call(flush_remote_tlbs)(kvm)) in kvm_arch_flush_remote_tlbs() 2051 return kvm_x86_call(flush_remote_tlbs_range)(kvm, gfn, nr_pages); in kvm_arch_flush_remote_tlbs_range() 2455 kvm_x86_call(vcpu_blocking)(vcpu); in kvm_arch_vcpu_blocking() 2460 kvm_x86_call(vcpu_unblocking)(vcpu); in kvm_arch_vcpu_unblocking()
|
| /linux/arch/x86/kvm/mmu/ |
| H A D | tdp_mmu.c | 378 kvm_x86_call(remove_external_spte)(kvm, gfn, level, old_spte); in remove_external_spte() 484 WARN_ON(kvm_x86_call(free_external_spt)(kvm, base_gfn, sp->role.level, in handle_removed_pt() 536 ret = kvm_x86_call(set_external_spte)(kvm, gfn, level, new_spte); in set_external_spte_present() 541 ret = kvm_x86_call(link_external_spt)(kvm, gfn, level, external_spt); in set_external_spte_present()
|
| H A D | spte.c | 243 spte |= kvm_x86_call(get_mt_mask)(vcpu, gfn, in make_spte()
|
| H A D | mmu.c | 3338 coco_level = kvm_x86_call(gmem_max_mapping_level)(kvm, pfn, is_private); in kvm_gmem_max_mapping_level() 6078 kvm_x86_call(flush_tlb_current)(vcpu); in kvm_mmu_load() 6537 kvm_x86_call(flush_tlb_gva)(vcpu, addr); in kvm_mmu_invalidate_addr()
|