Home
last modified time | relevance | path

Searched refs:efer (Results 1 – 25 of 29) sorted by relevance

12

/linux/arch/x86/realmode/
H A Dinit.c103 u64 efer; in setup_real_mode() local
151 rdmsrq(MSR_EFER, efer); in setup_real_mode()
152 trampoline_header->efer = efer & ~EFER_LMA; in setup_real_mode()
/linux/include/xen/interface/hvm/
H A Dhvm_vcpu.h35 uint64_t efer; member
89 uint64_t efer; member
/linux/arch/x86/kvm/
H A Dsmm.c82 CHECK_SMRAM64_OFFSET(efer, 0xFED0); in check_smram_offsets()
250 smram->efer = vcpu->arch.efer; in enter_smm_save_state_64()
536 if (__kvm_emulate_msr_write(vcpu, MSR_EFER, smstate->efer & ~EFER_LMA)) in rsm_load_state_64()
622 unsigned long cr4, efer; in emulator_leave_smm() local
630 efer = 0; in emulator_leave_smm()
631 __kvm_emulate_msr_write(vcpu, MSR_EFER, efer); in emulator_leave_smm()
H A Demulate.c733 u64 efer; in emulator_recalc_and_set_mode() local
738 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in emulator_recalc_and_set_mode()
742 if (efer & EFER_LMA) in emulator_recalc_and_set_mode()
750 if (efer & EFER_LMA) in emulator_recalc_and_set_mode()
759 if (efer & EFER_LMA) { in emulator_recalc_and_set_mode()
1450 u64 efer = 0; in get_descriptor_ptr() local
1452 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in get_descriptor_ptr()
1453 if (!(efer & EFER_LMA)) in get_descriptor_ptr()
1493 u64 efer = 0, cet = 0, ssp = 0; in emulator_is_ssp_invalid() local
1498 if (ctxt->ops->get_msr(ctxt, MSR_EFER, &efer)) in emulator_is_ssp_invalid()
[all …]
H A Dsmm.h104 u64 efer; member
H A Dx86.c1177 if ((vcpu->arch.efer & EFER_LME) && !is_paging(vcpu) && in kvm_set_cr0()
1188 if (!(vcpu->arch.efer & EFER_LME) && (cr0 & X86_CR0_PG) && in kvm_set_cr0()
1726 static bool __kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer) in __kvm_valid_efer() argument
1728 if (efer & EFER_AUTOIBRS && !guest_cpu_cap_has(vcpu, X86_FEATURE_AUTOIBRS)) in __kvm_valid_efer()
1731 if (efer & EFER_FFXSR && !guest_cpu_cap_has(vcpu, X86_FEATURE_FXSR_OPT)) in __kvm_valid_efer()
1734 if (efer & EFER_SVME && !guest_cpu_cap_has(vcpu, X86_FEATURE_SVM)) in __kvm_valid_efer()
1737 if (efer & (EFER_LME | EFER_LMA) && in __kvm_valid_efer()
1741 if (efer & EFER_NX && !guest_cpu_cap_has(vcpu, X86_FEATURE_NX)) in __kvm_valid_efer()
1747 bool kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer) in kvm_valid_efer() argument
1749 if (efer & efer_reserved_bits) in kvm_valid_efer()
[all …]
H A Dx86.h50 u64 efer; member
244 return !!(vcpu->arch.efer & EFER_LMA); in is_long_mode()
H A Dmmu.h88 unsigned long cr4, u64 efer, gpa_t nested_cr3);
/linux/arch/x86/kvm/svm/
H A Dnested.c97 svm->vmcb01.ptr->save.efer, in nested_svm_init_mmu_context()
373 if (CC(!(save->efer & EFER_SVME))) in __nested_vmcb_check_save()
388 if ((save->efer & EFER_LME) && (save->cr0 & X86_CR0_PG)) { in __nested_vmcb_check_save()
399 if (CC(!kvm_valid_efer(vcpu, save->efer))) in __nested_vmcb_check_save()
500 to->efer = from->efer; in __nested_copy_vmcb_save_to_cache()
684 svm_set_efer(vcpu, svm->nested.save.efer); in nested_vmcb02_prepare_save()
1043 vmcb01->save.efer = vcpu->arch.efer; in nested_svm_vmrun()
1088 to_save->efer = from_save->efer; in svm_copy_vmrun_state()
1156 vmcb12->save.efer = svm->vcpu.arch.efer; in nested_svm_vmexit()
1281 svm_set_efer(vcpu, vmcb01->save.efer); in nested_svm_vmexit()
[all …]
H A Dsvm.c203 int svm_set_efer(struct kvm_vcpu *vcpu, u64 efer) in svm_set_efer() argument
206 u64 old_efer = vcpu->arch.efer; in svm_set_efer()
207 vcpu->arch.efer = efer; in svm_set_efer()
211 efer |= EFER_NX; in svm_set_efer()
213 if (!(efer & EFER_LMA)) in svm_set_efer()
214 efer &= ~EFER_LME; in svm_set_efer()
217 if ((old_efer & EFER_SVME) != (efer & EFER_SVME)) { in svm_set_efer()
218 if (!(efer & EFER_SVME)) { in svm_set_efer()
236 vcpu->arch.efer = old_efer; in svm_set_efer()
249 svm->vmcb->save.efer = efer | EFER_SVME; in svm_set_efer()
[all …]
H A Dsvm.h143 u64 efer; member
721 int svm_set_efer(struct kvm_vcpu *vcpu, u64 efer);
/linux/arch/x86/hyperv/
H A Dhv_crash.c77 u64 efer; member
160 native_wrmsrq(MSR_EFER, ctxt->efer); in hv_crash_c_entry()
217 ctxt->efer = __rdmsr(MSR_EFER); in hv_hvcrash_ctxt_save()
H A Dhv_vtl.c165 input->vp_context.efer = native_rdmsrq(MSR_EFER); in hv_vtl_bringup_vcpu()
H A Divm.c331 vmsa->efer = native_read_msr(MSR_EFER); in hv_snp_boot_ap()
/linux/arch/x86/include/asm/
H A Drealmode.h52 u64 efer;
/linux/arch/x86/power/
H A Dcpu.c119 rdmsrq(MSR_EFER, ctxt->efer); in __save_processor_state()
212 wrmsrq(MSR_EFER, ctxt->efer); in __restore_processor_state()
/linux/arch/x86/kvm/vmx/
H A Dvmx.c1157 u64 guest_efer = vmx->vcpu.arch.efer; in update_transition_efer()
1182 (enable_ept && ((vmx->vcpu.arch.efer ^ kvm_host.efer) & EFER_NX))) { in update_transition_efer()
1185 if (guest_efer != kvm_host.efer) in update_transition_efer()
1186 add_atomic_switch_msr(vmx, MSR_EFER, guest_efer, kvm_host.efer); in update_transition_efer()
1199 guest_efer |= kvm_host.efer & ignore_bits; in update_transition_efer()
2011 (vmx->vcpu.arch.efer & EFER_SCE); in vmx_setup_uret_msrs()
3309 int vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer) in vmx_set_efer() argument
3317 vcpu->arch.efer = efer; in vmx_set_efer()
3319 if (efer & EFER_LMA) in vmx_set_efer()
3324 if (KVM_BUG_ON(efer & EFER_LMA, vcpu->kvm)) in vmx_set_efer()
[all …]
H A Dx86_ops.h71 int vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer);
H A Dnested.c2280 return vmx->vcpu.arch.efer | (EFER_LMA | EFER_LME); in nested_vmx_calc_efer()
2282 return vmx->vcpu.arch.efer & ~(EFER_LMA | EFER_LME); in nested_vmx_calc_efer()
2494 if (guest_efer != kvm_host.efer) in prepare_vmcs02_early()
2507 if (cpu_has_load_ia32_efer() && guest_efer != kvm_host.efer) in prepare_vmcs02_early()
2787 vcpu->arch.efer = nested_vmx_calc_efer(vmx, vmcs12); in prepare_vmcs02()
2789 vmx_set_efer(vcpu, vcpu->arch.efer); in prepare_vmcs02()
3125 !!(vcpu->arch.efer & EFER_LMA))) in nested_vmx_check_address_space_size()
4687 vmcs12->guest_ia32_efer = vcpu->arch.efer; in sync_vmcs02_to_vmcs12()
4768 vcpu->arch.efer = vmcs12->host_ia32_efer; in load_vmcs12_host_state()
4770 vcpu->arch.efer |= (EFER_LMA | EFER_LME); in load_vmcs12_host_state()
[all …]
H A Dmain.c416 static int vt_set_efer(struct kvm_vcpu *vcpu, u64 efer) in vt_set_efer() argument
421 return vmx_set_efer(vcpu, efer); in vt_set_efer()
/linux/tools/arch/x86/include/uapi/asm/
H A Dkvm.h156 __u64 efer; member
167 __u64 efer; member
/linux/drivers/parport/
H A Dparport_pc.c1114 static void decode_winbond(int efer, int key, int devid, int devrev, int oldid) in decode_winbond() argument
1159 efer, key, devid, devrev, oldid, type); in decode_winbond()
1162 show_parconfig_winbond(efer, key); in decode_winbond()
1165 static void decode_smsc(int efer, int key, int devid, int devrev) in decode_smsc() argument
1191 efer, key, devid, devrev, type); in decode_smsc()
1194 func(efer, key); in decode_smsc()
/linux/arch/x86/kvm/mmu/
H A Dmmu.c189 const u64 efer; member
214 BUILD_MMU_ROLE_REGS_ACCESSOR(efer, nx, EFER_NX);
215 BUILD_MMU_ROLE_REGS_ACCESSOR(efer, lma, EFER_LMA);
234 BUILD_MMU_ROLE_ACCESSOR(base, efer, nx);
235 BUILD_MMU_ROLE_ACCESSOR(ext, efer, lma);
252 .efer = vcpu->arch.efer, in vcpu_to_role_regs()
5871 unsigned long cr4, u64 efer, gpa_t nested_cr3) in kvm_init_shadow_npt_mmu() argument
5877 .efer = efer, in kvm_init_shadow_npt_mmu()
/linux/tools/testing/selftests/kvm/lib/x86/
H A Dprocessor.c141 fprintf(stream, "%*scr8: 0x%.16llx efer: 0x%.16llx " in sregs_dump()
144 sregs->cr8, sregs->efer, sregs->apic_base); in sregs_dump()
664 sregs.efer |= (EFER_LME | EFER_LMA | EFER_NX); in kvm_arch_vm_post_create()
/linux/include/hyperv/
H A Dhvhdk.h113 u64 efer; member

12