Searched refs:MSR_SPE (Results 1 – 8 of 8) sorted by relevance
371 msr_check_and_set(MSR_SPE); in giveup_spe()373 msr_check_and_clear(MSR_SPE); in giveup_spe()381 msr_check_and_set(MSR_SPE); in enable_kernel_spe()383 if (current->thread.regs && (current->thread.regs->msr & MSR_SPE)) { in enable_kernel_spe()394 if (tsk->thread.regs->msr & MSR_SPE) { in flush_spe_to_thread()415 msr_all_available |= MSR_SPE; in init_msr_all_available()443 if (usermsr & MSR_SPE) in giveup_all()585 if (usermsr & MSR_SPE) in save_all()597 if (tsk->thread.regs->msr & MSR_SPE) in flush_all_to_thread()
830 oris r5,r5,MSR_SPE@h834 oris r9,r9,MSR_SPE@h850 oris r3,r3,MSR_SPE@h961 lis r3,MSR_SPE@h
322 msr |= MSR_SPE; in __unsafe_save_user_regs()545 regs_set_return_msr(regs, regs->msr & ~MSR_SPE); in restore_user_regs()546 if (msr & MSR_SPE) { in restore_user_regs()
106 msr_check_and_clear(MSR_SPE); in disable_kernel_spe()
27 #define MSR_SPE __MASK(MSR_SPE_LG) macro
84 #ifndef MSR_SPE85 #define MSR_SPE 0 macro
119 vcpu->arch.shadow_msr &= ~MSR_SPE; in kvmppc_vcpu_disable_spe()129 vcpu->arch.shadow_msr |= MSR_SPE; in kvmppc_vcpu_enable_spe()135 if (vcpu->arch.shared->msr & MSR_SPE) { in kvmppc_vcpu_sync_spe()136 if (!(vcpu->arch.shadow_msr & MSR_SPE)) in kvmppc_vcpu_sync_spe()138 } else if (vcpu->arch.shadow_msr & MSR_SPE) { in kvmppc_vcpu_sync_spe()1177 if (vcpu->arch.shared->msr & MSR_SPE) in kvmppc_handle_exit()
310 if (vcpu->arch.shadow_msr & MSR_SPE) in kvmppc_core_vcpu_put_e500()