| /linux/arch/x86/power/ |
| H A D | hibernate_asm_64.S | 31 movq %r9, %cr3 38 movq %cr3, %rcx; # flush TLB 39 movq %rcx, %cr3 94 movq %cr3, %rax 125 movq %rax, %cr3 130 movq %cr3, %rcx; # flush TLB 131 movq %rcx, %cr3;
|
| H A D | hibernate.c | 61 unsigned long cr3; member 122 rdr->cr3 = restore_cr3 & ~CR3_PCID_MASK; in arch_hibernation_header_save() 144 restore_cr3 = rdr->cr3; in arch_hibernation_header_restore()
|
| H A D | cpu.c | 127 ctxt->cr3 = __read_cr3(); in __save_processor_state() 215 write_cr3(ctxt->cr3); in __restore_processor_state()
|
| /linux/scripts/gdb/linux/ |
| H A D | pgtable.py | 64 def __init__(self, cr3, page_levels): 65 self.cr3 = cr3 67 self.page_level_write_through = is_bit_defined_tupled(cr3, 3) 68 self.page_level_cache_disabled = is_bit_defined_tupled(cr3, 4) 69 self.next_entry_physical_address = cr3 & PHYSICAL_ADDRESS_MASK & page_mask() 77 cr3: 78 {'cr3 binary data': <30} {hex(self.cr3)} 208 cr3_data = gdb.parse_and_eval('$cr3') 66 __init__(self, cr3, page_levels) global() argument [all...] |
| /linux/arch/x86/kernel/ |
| H A D | sev_verify_cbit.S | 56 movq %cr3, %rcx 59 movq %rdi, %cr3 68 movq %rcx, %cr3
|
| H A D | head_64.S | 136 movq %rax, %cr3 196 movq %rax, %cr3
|
| /linux/arch/x86/platform/olpc/ |
| H A D | xo1-wakeup.S | 25 # Set up %cr3 27 movl %eax, %cr3 51 movl %cr3, %eax 52 movl %eax, %cr3
|
| /linux/arch/x86/entry/ |
| H A D | calling.h | 175 mov %cr3, \scratch_reg 177 mov \scratch_reg, %cr3 185 mov %cr3, \scratch_reg 213 mov \scratch_reg, %cr3 232 movq %cr3, \scratch_reg 243 movq \scratch_reg, %cr3 274 movq \save_reg, %cr3
|
| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | nested_invalid_cr3_test.c | 30 save_cr3 = svm->vmcb->save.cr3; in l1_svm_code() 31 svm->vmcb->save.cr3 = -1ull; in l1_svm_code() 36 svm->vmcb->save.cr3 = save_cr3; in l1_svm_code()
|
| /linux/arch/x86/kvm/ |
| H A D | smm.c | 57 CHECK_SMRAM32_OFFSET(cr3, 0xFFF8); in check_smram_offsets() 97 CHECK_SMRAM64_OFFSET(cr3, 0xFF50); in check_smram_offsets() 191 smram->cr3 = kvm_read_cr3(vcpu); in enter_smm_save_state_32() 244 smram->cr3 = kvm_read_cr3(vcpu); in enter_smm_save_state_64() 420 u64 cr0, u64 cr3, u64 cr4) in rsm_enter_protected_mode() argument 428 pcid = cr3 & 0xfff; in rsm_enter_protected_mode() 429 cr3 &= ~0xfff; in rsm_enter_protected_mode() 432 bad = kvm_set_cr3(vcpu, cr3); in rsm_enter_protected_mode() 454 bad = kvm_set_cr3(vcpu, cr3 | pcid); in rsm_enter_protected_mode() 504 smstate->cr3, smstate->cr4); in rsm_load_state_32() [all …]
|
| H A D | smm.h | 64 u32 cr3; member 128 u64 cr3; member
|
| H A D | mmu.h | 126 static inline unsigned long kvm_get_pcid(struct kvm_vcpu *vcpu, gpa_t cr3) in kvm_get_pcid() argument 131 ? cr3 & X86_CR3_PCID_MASK in kvm_get_pcid()
|
| H A D | tss.h | 13 u32 cr3; member
|
| /linux/include/xen/interface/hvm/ |
| H A D | hvm_vcpu.h | 24 uint32_t cr3; member 87 uint64_t cr3; member
|
| /linux/arch/x86/hyperv/ |
| H A D | hv_trampoline.S | 54 movl %eax, %cr3 90 movq %rbx, %cr3
|
| /linux/arch/x86/kvm/svm/ |
| H A D | nested.c | 61 u64 cr3 = svm->nested.ctl.nested_cr3; in nested_svm_get_tdp_pdptr() local 69 ret = kvm_vcpu_read_guest_page(vcpu, gpa_to_gfn(cr3), &pdpte, in nested_svm_get_tdp_pdptr() 70 (cr3 & GENMASK(11, 5)) + index * 8, 8); in nested_svm_get_tdp_pdptr() 433 CC(!kvm_vcpu_is_legal_cr3(vcpu, save->cr3))) in nested_vmcb_check_save() 550 to->cr3 = from->cr3; in __nested_copy_vmcb_save_to_cache() 673 static int nested_svm_load_cr3(struct kvm_vcpu *vcpu, unsigned long cr3, in nested_svm_load_cr3() argument 676 if (CC(!kvm_vcpu_is_legal_cr3(vcpu, cr3))) in nested_svm_load_cr3() 680 CC(!load_pdptrs(vcpu, cr3))) in nested_svm_load_cr3() 683 vcpu->arch.cr3 = cr3; in nested_svm_load_cr3() 689 kvm_mmu_new_pgd(vcpu, cr3); in nested_svm_load_cr3() [all …]
|
| /linux/include/trace/events/ |
| H A D | xen.h | 368 TP_PROTO(bool kernel, unsigned long cr3), 369 TP_ARGS(kernel, cr3), 372 __field(unsigned long, cr3) 375 __entry->cr3 = cr3), 377 __entry->kernel ? "kernel" : "user", __entry->cr3)
|
| /linux/arch/x86/kernel/acpi/ |
| H A D | wakeup_32.S | 27 movl %cr3, %eax 28 movl %eax, %cr3
|
| /linux/arch/x86/xen/ |
| H A D | mmu_pv.c | 1374 static void __xen_write_cr3(bool kernel, unsigned long cr3) in __xen_write_cr3() argument 1379 trace_xen_mmu_write_cr3(kernel, cr3); in __xen_write_cr3() 1381 if (cr3) in __xen_write_cr3() 1382 mfn = pfn_to_mfn(PFN_DOWN(cr3)); in __xen_write_cr3() 1394 this_cpu_write(xen_cr3, cr3); in __xen_write_cr3() 1398 xen_mc_callback(set_current_cr3, (void *)cr3); in __xen_write_cr3() 1401 static void xen_write_cr3(unsigned long cr3) in xen_write_cr3() argument 1403 pgd_t *user_pgd = xen_get_user_pgd(__va(cr3)); in xen_write_cr3() 1411 this_cpu_write(xen_cr3, cr3); in xen_write_cr3() 1413 __xen_write_cr3(true, cr3); in xen_write_cr3() [all …]
|
| /linux/arch/powerpc/kernel/ |
| H A D | cpu_setup_6xx.S | 365 cmplwi cr3,r3,0x8001 /* 7455 */ 373 cror 4*cr0+eq,4*cr0+eq,4*cr3+eq 436 cmplwi cr3,r3,0x8001 /* 7455 */ 444 cror 4*cr0+eq,4*cr0+eq,4*cr3+eq
|
| /linux/arch/x86/realmode/rm/ |
| H A D | trampoline_64.S | 161 movl %eax, %cr3 244 movq %rax, %cr3
|
| H A D | reboot.S | 103 movl %ecx, %cr3
|
| H A D | wakeup_asm.S | 112 movl %eax, %cr3
|
| /linux/arch/powerpc/boot/ |
| H A D | ppc_asm.h | 16 #define cr3 3 macro
|
| /linux/arch/x86/mm/ |
| H A D | mem_encrypt_boot.S | 100 mov %rdx, %cr3
|