| /linux/arch/arm64/kvm/hyp/vhe/ |
| H A D | switch.c | 87 guest_hcr = __vcpu_sys_reg(vcpu, HCR_EL2); in __compute_hcr() 92 va |= __vcpu_sys_reg(vcpu, VNCR_EL2) & GENMASK(PAGE_SHIFT - 1, 0); in __compute_hcr() 129 val = __vcpu_sys_reg(vcpu, CNTP_CVAL_EL0); in __activate_traps() 131 val = __vcpu_sys_reg(vcpu, CNTHP_CVAL_EL2); in __activate_traps() 242 cval = __vcpu_sys_reg(vcpu, CNTP_CVAL_EL0); in compute_emulated_cntx_ctl_el0() 243 ctl = __vcpu_sys_reg(vcpu, CNTP_CTL_EL0); in compute_emulated_cntx_ctl_el0() 247 cval = __vcpu_sys_reg(vcpu, CNTV_CVAL_EL0); in compute_emulated_cntx_ctl_el0() 248 ctl = __vcpu_sys_reg(vcpu, CNTV_CTL_EL0); in compute_emulated_cntx_ctl_el0() 289 val = __vcpu_sys_reg(vcpu, CNTP_CVAL_EL0); in kvm_hyp_handle_timer() 298 val = __vcpu_sys_reg(vcpu, CNTP_CVAL_EL0); in kvm_hyp_handle_timer() [all …]
|
| /linux/arch/arm64/kvm/ |
| H A D | pauth.c | 40 mod = __vcpu_sys_reg(vcpu, SP_EL2); in compute_pac() 170 ikey.lo = __vcpu_sys_reg(vcpu, APIBKEYLO_EL1); in kvm_auth_eretax() 171 ikey.hi = __vcpu_sys_reg(vcpu, APIBKEYHI_EL1); in kvm_auth_eretax() 176 ikey.lo = __vcpu_sys_reg(vcpu, APIAKEYLO_EL1); in kvm_auth_eretax() 177 ikey.hi = __vcpu_sys_reg(vcpu, APIAKEYHI_EL1); in kvm_auth_eretax()
|
| H A D | sys_regs.c | 315 val |= __vcpu_sys_reg(vcpu, reg) & ~CNTKCTL_VALID_BITS; in vcpu_read_sys_reg() 338 return __vcpu_sys_reg(vcpu, reg); in vcpu_read_sys_reg() 818 p->regval = __vcpu_sys_reg(vcpu, r->reg); in trap_oslsr_el1() 1026 return __vcpu_sys_reg(vcpu, r->reg); in reset_pmu_reg() 1034 return __vcpu_sys_reg(vcpu, r->reg); in reset_pmevcntr() 1046 return __vcpu_sys_reg(vcpu, r->reg); in reset_pmevtyper() 1054 return __vcpu_sys_reg(vcpu, r->reg); in reset_pmselr() 1070 return __vcpu_sys_reg(vcpu, r->reg); in reset_pmcr() 1075 u64 reg = __vcpu_sys_reg(vcpu, PMUSERENR_EL0); in check_pmu_access_disabled() 1143 p->regval = __vcpu_sys_reg(vcpu, PMSELR_EL0) in access_pmselr() [all …]
|
| H A D | trace_arm.h | 352 __entry->hcr_el2 = __vcpu_sys_reg(vcpu, HCR_EL2); 382 __entry->hcr_el2 = __vcpu_sys_reg(vcpu, HCR_EL2);
|
| H A D | at.c | 105 !(__vcpu_sys_reg(vcpu, HCRX_EL2) & HCRX_EL2_TCR2En)) in effective_tcr2() 149 hcr = __vcpu_sys_reg(vcpu, HCR_EL2); in setup_s1_walk() 804 if (__vcpu_sys_reg(vcpu, HCR_EL2) & HCR_FWB) { in compute_par_s12() 858 if ((__vcpu_sys_reg(vcpu, HCR_EL2) & HCR_CD) && in compute_par_s12() 890 (__vcpu_sys_reg(vcpu, HCR_EL2) & HCR_DC)) { in compute_par_s1()
|
| H A D | nested.c | 795 if (__vcpu_sys_reg(vcpu, HCR_EL2) & HCR_NV) in kvm_vcpu_load_hw_mmu() 1261 return (u64)sign_extend64(__vcpu_sys_reg(vcpu, VNCR_EL2), 48); in read_vncr_el2() 1922 u64 guest_mdcr = __vcpu_sys_reg(vcpu, MDCR_EL2); in kvm_nested_setup_mdcr_el2()
|
| H A D | vgic-sys-reg-v3.c | 310 *val = __vcpu_sys_reg(vcpu, r->reg); in get_gic_ich_reg()
|
| H A D | config.c | 1670 u64 nested = __vcpu_sys_reg(vcpu, reg); in __compute_fgt()
|
| H A D | arm.c | 603 val = __vcpu_sys_reg(vcpu, HCR_EL2); in vcpu_set_pauth_traps()
|
| /linux/arch/arm64/include/asm/ |
| H A D | kvm_emulate.h | 199 (__vcpu_sys_reg(vcpu, HCR_EL2) & HCR_E2H)); in vcpu_el2_e2h_is_set() 234 hcr = __vcpu_sys_reg(vcpu, HCR_EL2); in is_hyp_ctxt() 266 (__vcpu_sys_reg(vcpu, HCRX_EL2) & HCRX_EL2_TMEA); in vserror_state_is_nested() 324 u64 hcr_el2 = __vcpu_sys_reg(vcpu, HCR_EL2); in guest_hyp_wfx_traps_enabled() 506 return __vcpu_sys_reg(vcpu, MPIDR_EL1) & MPIDR_HWID_BITMASK; in kvm_vcpu_get_mpidr_aff() 620 u64 cptr = __vcpu_sys_reg(vcpu, CPTR_EL2); in vcpu_sanitised_cptr_el2()
|
| H A D | kvm_host.h | 1229 #define __vcpu_sys_reg(v,r) \ macro 1454 (!!(__vcpu_sys_reg(vcpu, OSLSR_EL1) & OSLSR_EL1_OSLK))
|
| /linux/arch/arm64/kvm/hyp/include/hyp/ |
| H A D | sysreg-sr.h | 370 write_sysreg(__vcpu_sys_reg(vcpu, DACR32_EL2), dacr32_el2); in __sysreg32_restore_state() 371 write_sysreg(__vcpu_sys_reg(vcpu, IFSR32_EL2), ifsr32_el2); in __sysreg32_restore_state() 374 write_sysreg(__vcpu_sys_reg(vcpu, DBGVCR32_EL2), dbgvcr32_el2); in __sysreg32_restore_state()
|
| /linux/arch/arm64/kvm/hyp/ |
| H A D | exception.c | 28 return __vcpu_sys_reg(vcpu, reg); in __vcpu_read_sys_reg()
|