| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | set_sregs_test.c | 51 u64 cr4 = KVM_ALWAYS_ALLOWED_CR4; in calc_supported_cr4_feature_bits() local 54 cr4 |= X86_CR4_UMIP; in calc_supported_cr4_feature_bits() 56 cr4 |= X86_CR4_LA57; in calc_supported_cr4_feature_bits() 58 cr4 |= X86_CR4_VMXE; in calc_supported_cr4_feature_bits() 60 cr4 |= X86_CR4_SMXE; in calc_supported_cr4_feature_bits() 62 cr4 |= X86_CR4_FSGSBASE; in calc_supported_cr4_feature_bits() 64 cr4 |= X86_CR4_PCIDE; in calc_supported_cr4_feature_bits() 66 cr4 |= X86_CR4_OSXSAVE; in calc_supported_cr4_feature_bits() 68 cr4 |= X86_CR4_SMEP; in calc_supported_cr4_feature_bits() 70 cr4 |= X86_CR4_SMAP; in calc_supported_cr4_feature_bits() [all …]
|
| H A D | sync_regs_test.c | 138 __u64 *cr4 = &run->s.regs.sregs.cr4; in race_sregs_cr4() local 139 __u64 pae_enabled = *cr4; in race_sregs_cr4() 140 __u64 pae_disabled = *cr4 & ~X86_CR4_PAE; in race_sregs_cr4() 144 WRITE_ONCE(*cr4, pae_enabled); in race_sregs_cr4() 148 WRITE_ONCE(*cr4, pae_disabled); in race_sregs_cr4() 178 TEST_ASSERT((run->s.regs.sregs.cr4 & X86_CR4_PAE) && in race_sync_regs() 181 !!(run->s.regs.sregs.cr4 & X86_CR4_PAE), in race_sync_regs()
|
| H A D | cpuid_test.c | 172 sregs.cr4 ^= X86_CR4_OSXSAVE; in set_cpuid_after_run() 174 sregs.cr4 ^= X86_CR4_PKE; in set_cpuid_after_run()
|
| /linux/arch/x86/kvm/ |
| H A D | smm.c | 30 CHECK_SMRAM32_OFFSET(cr4, 0xFF14); in check_smram_offsets() 96 CHECK_SMRAM64_OFFSET(cr4, 0xFF48); in check_smram_offsets() 220 smram->cr4 = kvm_read_cr4(vcpu); in enter_smm_save_state_32() 245 smram->cr4 = kvm_read_cr4(vcpu); in enter_smm_save_state_64() 420 u64 cr0, u64 cr3, u64 cr4) in rsm_enter_protected_mode() argument 427 if (cr4 & X86_CR4_PCIDE) { in rsm_enter_protected_mode() 441 bad = kvm_set_cr4(vcpu, cr4 & ~X86_CR4_PCIDE); in rsm_enter_protected_mode() 449 if (cr4 & X86_CR4_PCIDE) { in rsm_enter_protected_mode() 450 bad = kvm_set_cr4(vcpu, cr4); in rsm_enter_protected_mode() 504 smstate->cr3, smstate->cr4); in rsm_load_state_32() [all …]
|
| H A D | smm.h | 31 u32 cr4; member 127 u64 cr4; member
|
| /linux/drivers/misc/lkdtm/ |
| H A D | bugs.c | 682 unsigned long cr4; in lkdtm_UNSET_SMEP() local 685 cr4 = native_read_cr4(); in lkdtm_UNSET_SMEP() 687 if ((cr4 & X86_CR4_SMEP) != X86_CR4_SMEP) { in lkdtm_UNSET_SMEP() 691 cr4 &= ~(X86_CR4_SMEP); in lkdtm_UNSET_SMEP() 694 native_write_cr4(cr4); in lkdtm_UNSET_SMEP() 695 if (cr4 == native_read_cr4()) { in lkdtm_UNSET_SMEP() 697 cr4 |= X86_CR4_SMEP; in lkdtm_UNSET_SMEP() 699 native_write_cr4(cr4); in lkdtm_UNSET_SMEP() 729 direct_write_cr4(cr4); in lkdtm_UNSET_SMEP() 734 cr4 |= X86_CR4_SMEP; in lkdtm_UNSET_SMEP() [all …]
|
| /linux/arch/x86/kernel/ |
| H A D | sev_verify_cbit.S | 35 movq %cr4, %rsi 40 movq %rdx, %cr4 71 movq %rsi, %cr4
|
| H A D | head_64.S | 222 movq %cr4, %rcx 227 movq %rcx, %cr4 233 movq %rcx, %cr4
|
| /linux/arch/x86/kernel/cpu/mtrr/ |
| H A D | cyrix.c | 135 static u32 cr4, ccr3; variable 143 cr4 = __read_cr4(); in prepare_set() 144 __write_cr4(cr4 & ~X86_CR4_PGE); in prepare_set() 176 __write_cr4(cr4); in post_set()
|
| /linux/arch/x86/power/ |
| H A D | hibernate_asm_64.S | 37 movq %rdx, %cr4; # turn off PGE 40 movq %rax, %cr4; # turn PGE back on 129 movq %rcx, %cr4; # turn off PGE 132 movq %rbx, %cr4; # turn PGE back on
|
| H A D | cpu.c | 128 ctxt->cr4 = __read_cr4(); in __save_processor_state() 208 if (ctxt->cr4) in __restore_processor_state() 209 __write_cr4(ctxt->cr4); in __restore_processor_state() 213 __write_cr4(ctxt->cr4); in __restore_processor_state() 245 if (ctxt->cr4 & X86_CR4_FRED) { in __restore_processor_state()
|
| /linux/arch/x86/mm/ |
| H A D | mem_encrypt_boot.S | 103 mov %cr4, %rdx 105 mov %rdx, %cr4 107 mov %rdx, %cr4
|
| /linux/tools/testing/selftests/kvm/lib/x86/ |
| H A D | vmx.c | 131 unsigned long cr4; in prepare_for_vmx_operation() local 143 __asm__ __volatile__("mov %%cr4, %0" : "=r"(cr4) : : "memory"); in prepare_for_vmx_operation() 144 cr4 &= rdmsr(MSR_IA32_VMX_CR4_FIXED1); in prepare_for_vmx_operation() 145 cr4 |= rdmsr(MSR_IA32_VMX_CR4_FIXED0); in prepare_for_vmx_operation() 147 cr4 |= X86_CR4_VMXE; in prepare_for_vmx_operation() 148 __asm__ __volatile__("mov %0, %%cr4" : : "r"(cr4) : "memory"); in prepare_for_vmx_operation()
|
| /linux/include/xen/interface/hvm/ |
| H A D | hvm_vcpu.h | 25 uint32_t cr4; member 88 uint64_t cr4; member
|
| /linux/arch/x86/platform/olpc/ |
| H A D | xo1-wakeup.S | 30 movl %eax, %cr4 64 movl %cr4, %edx
|
| /linux/arch/x86/kernel/cpu/ |
| H A D | common.c | 499 unsigned long newval, cr4 = this_cpu_read(cpu_tlbstate.cr4); in cr4_update_irqsoff() local 503 newval = (cr4 & ~clear) | set; in cr4_update_irqsoff() 504 if (newval != cr4) { in cr4_update_irqsoff() 505 this_cpu_write(cpu_tlbstate.cr4, newval); in cr4_update_irqsoff() 514 return this_cpu_read(cpu_tlbstate.cr4); in cr4_read_shadow() 520 unsigned long cr4 = __read_cr4(); in cr4_init() local 523 cr4 |= X86_CR4_PCIDE; in cr4_init() 525 cr4 = (cr4 & ~cr4_pinned_mask) | cr4_pinned_bits; in cr4_init() 527 __write_cr4(cr4); in cr4_init() 530 this_cpu_write(cpu_tlbstate.cr4, cr4); in cr4_init() [all …]
|
| /linux/arch/powerpc/kernel/ |
| H A D | cpu_setup_6xx.S | 366 cmplwi cr4,r3,0x8002 /* 7457 */ 374 cror 4*cr0+eq,4*cr0+eq,4*cr4+eq 437 cmplwi cr4,r3,0x8002 /* 7457 */ 445 cror 4*cr0+eq,4*cr0+eq,4*cr4+eq
|
| /linux/arch/x86/kvm/svm/ |
| H A D | nested.c | 96 kvm_init_shadow_npt_mmu(vcpu, X86_CR0_PG, svm->vmcb01.ptr->save.cr4, in nested_svm_init_mmu_context() 431 if (CC(!(save->cr4 & X86_CR4_PAE)) || in nested_vmcb_check_save() 442 if (CC(!__kvm_is_valid_cr4(vcpu, save->cr4))) in nested_vmcb_check_save() 549 to->cr4 = from->cr4; in __nested_copy_vmcb_save_to_cache() 756 svm_set_cr4(vcpu, svm->nested.save.cr4); in nested_vmcb02_prepare_save() 1140 vmcb01->save.cr4 = vcpu->arch.cr4; in nested_svm_vmrun() 1179 to_save->cr4 = from_save->cr4; in svm_copy_vmrun_state() 1237 vmcb12->save.cr4 = svm->vcpu.arch.cr4; in nested_svm_vmexit_update_vmcb12() 1393 svm_set_cr4(vcpu, vmcb01->save.cr4); in nested_svm_vmexit()
|
| /linux/scripts/gdb/linux/ |
| H A D | pgtable.py | 209 cr4 = gdb.parse_and_eval('$cr4') 210 page_levels = 5 if cr4 & (1 << 12) else 4
|
| /linux/arch/x86/hyperv/ |
| H A D | hv_crash.c | 61 ulong cr4; member 171 asm volatile("movq %0, %%cr4" : : "r"(hv_crash_ctxt.cr4)); in hv_crash_c_entry() 205 ctxt->cr4 = native_read_cr4(); in hv_hvcrash_ctxt_save()
|
| H A D | hv_trampoline.S | 49 movl %ecx, %cr4
|
| /linux/arch/x86/realmode/rm/ |
| H A D | trampoline_64.S | 157 movl %eax, %cr4 # Enable PAE mode 226 movq %cr4, %rax
|
| /linux/arch/x86/include/asm/ |
| H A D | realmode.h | 53 u32 cr4;
|
| /linux/arch/powerpc/boot/ |
| H A D | ppc_asm.h | 17 #define cr4 4 macro
|
| /linux/arch/x86/kvm/vmx/ |
| H A D | x86_ops.h | 69 void vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4); 70 bool vmx_is_valid_cr4(struct kvm_vcpu *vcpu, unsigned long cr4);
|