| /linux/arch/riscv/include/asm/ |
| H A D | kvm_vcpu_fp.h | 15 struct kvm_cpu_context; 18 void __kvm_riscv_fp_f_save(struct kvm_cpu_context *context); 19 void __kvm_riscv_fp_f_restore(struct kvm_cpu_context *context); 20 void __kvm_riscv_fp_d_save(struct kvm_cpu_context *context); 21 void __kvm_riscv_fp_d_restore(struct kvm_cpu_context *context); 24 void kvm_riscv_vcpu_guest_fp_save(struct kvm_cpu_context *cntx, 26 void kvm_riscv_vcpu_guest_fp_restore(struct kvm_cpu_context *cntx, 28 void kvm_riscv_vcpu_host_fp_save(struct kvm_cpu_context *cntx); 29 void kvm_riscv_vcpu_host_fp_restore(struct kvm_cpu_context *cntx); 34 static inline void kvm_riscv_vcpu_guest_fp_save(struct kvm_cpu_context *cntx, in kvm_riscv_vcpu_guest_fp_save() [all …]
|
| H A D | kvm_vcpu_vector.h | 19 static __always_inline void __kvm_riscv_vector_save(struct kvm_cpu_context *context) in __kvm_riscv_vector_save() 24 static __always_inline void __kvm_riscv_vector_restore(struct kvm_cpu_context *context) in __kvm_riscv_vector_restore() 30 void kvm_riscv_vcpu_guest_vector_save(struct kvm_cpu_context *cntx, 32 void kvm_riscv_vcpu_guest_vector_restore(struct kvm_cpu_context *cntx, 34 void kvm_riscv_vcpu_host_vector_save(struct kvm_cpu_context *cntx); 35 void kvm_riscv_vcpu_host_vector_restore(struct kvm_cpu_context *cntx); 40 struct kvm_cpu_context; 46 static inline void kvm_riscv_vcpu_guest_vector_save(struct kvm_cpu_context *cntx, in kvm_riscv_vcpu_guest_vector_save() 51 static inline void kvm_riscv_vcpu_guest_vector_restore(struct kvm_cpu_context *cntx, in kvm_riscv_vcpu_guest_vector_restore() 56 static inline void kvm_riscv_vcpu_host_vector_save(struct kvm_cpu_context *cntx) in kvm_riscv_vcpu_host_vector_save() [all …]
|
| /linux/arch/arm64/kvm/hyp/nvhe/ |
| H A D | hyp-main.c | 27 void __kvm_hyp_host_forward_smc(struct kvm_cpu_context *host_ctxt); 169 static void handle___pkvm_vcpu_load(struct kvm_cpu_context *host_ctxt) in handle___pkvm_vcpu_load() 190 static void handle___pkvm_vcpu_put(struct kvm_cpu_context *host_ctxt) in handle___pkvm_vcpu_put() 198 static void handle___kvm_vcpu_run(struct kvm_cpu_context *host_ctxt) in handle___kvm_vcpu_run() 248 static void handle___pkvm_host_donate_guest(struct kvm_cpu_context *host_ctxt) in handle___pkvm_host_donate_guest() 268 static void handle___pkvm_host_share_guest(struct kvm_cpu_context *host_ctxt) in handle___pkvm_host_share_guest() 290 static void handle___pkvm_host_unshare_guest(struct kvm_cpu_context *host_ctxt) in handle___pkvm_host_unshare_guest() 308 static void handle___pkvm_host_relax_perms_guest(struct kvm_cpu_context *host_ctxt) in handle___pkvm_host_relax_perms_guest() 324 static void handle___pkvm_host_wrprotect_guest(struct kvm_cpu_context *host_ctxt) in handle___pkvm_host_wrprotect_guest() 342 static void handle___pkvm_host_test_clear_young_guest(struct kvm_cpu_context *host_ctxt) in handle___pkvm_host_test_clear_young_guest() [all …]
|
| H A D | tlb.c | 24 struct kvm_cpu_context *host_ctxt; in enter_vmid_context() 121 struct kvm_cpu_context *host_ctxt; in exit_vmid_context()
|
| H A D | setup.c | 289 struct kvm_cpu_context *host_ctxt = host_data_ptr(host_ctxt); in __pkvm_init_finalise()
|
| /linux/arch/arm64/kvm/hyp/include/hyp/ |
| H A D | sysreg-sr.h | 19 static inline bool ctxt_has_s1poe(struct kvm_cpu_context *ctxt); 21 static inline struct kvm_vcpu *ctxt_to_vcpu(struct kvm_cpu_context *ctxt) in ctxt_to_vcpu() 31 static inline bool ctxt_is_guest(struct kvm_cpu_context *ctxt) in ctxt_is_guest() 36 static inline u64 *ctxt_mdscr_el1(struct kvm_cpu_context *ctxt) in ctxt_mdscr_el1() 46 static inline u64 ctxt_midr_el1(struct kvm_cpu_context *ctxt) in ctxt_midr_el1() 57 static inline void __sysreg_save_common_state(struct kvm_cpu_context *ctxt) in __sysreg_save_common_state() 66 static inline void __sysreg_save_user_state(struct kvm_cpu_context *ctxt) in __sysreg_save_user_state() 72 static inline bool ctxt_has_mte(struct kvm_cpu_context *ctxt) in ctxt_has_mte() 79 static inline bool ctxt_has_s1pie(struct kvm_cpu_context *ctxt) in ctxt_has_s1pie() 90 static inline bool ctxt_has_tcrx(struct kvm_cpu_context *ctxt) in ctxt_has_tcrx() [all …]
|
| H A D | debug-sr.h | 107 struct kvm_cpu_context *ctxt) in __debug_save_state() 121 struct kvm_cpu_context *ctxt) in __debug_restore_state() 136 struct kvm_cpu_context *host_ctxt; in __debug_switch_to_guest_common() 137 struct kvm_cpu_context *guest_ctxt; in __debug_switch_to_guest_common() 155 struct kvm_cpu_context *host_ctxt; in __debug_switch_to_host_common() 156 struct kvm_cpu_context *guest_ctxt; in __debug_switch_to_host_common()
|
| /linux/arch/riscv/kvm/ |
| H A D | vcpu_fp.c | 21 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_fp_reset() 31 static void kvm_riscv_vcpu_fp_clean(struct kvm_cpu_context *cntx) in kvm_riscv_vcpu_fp_clean() 37 void kvm_riscv_vcpu_guest_fp_save(struct kvm_cpu_context *cntx, in kvm_riscv_vcpu_guest_fp_save() 49 void kvm_riscv_vcpu_guest_fp_restore(struct kvm_cpu_context *cntx, in kvm_riscv_vcpu_guest_fp_restore() 61 void kvm_riscv_vcpu_host_fp_save(struct kvm_cpu_context *cntx) in kvm_riscv_vcpu_host_fp_save() 70 void kvm_riscv_vcpu_host_fp_restore(struct kvm_cpu_context *cntx) in kvm_riscv_vcpu_host_fp_restore() 83 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_get_reg_fp() 132 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_set_reg_fp()
|
| H A D | vcpu_sbi_replace.c | 20 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_time_handler() 51 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_ipi_handler() 95 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_rfence_handler() 151 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_srst_handler()
|
| H A D | vcpu_sbi_hsm.c | 18 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_hsm_vcpu_start() 65 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_hsm_vcpu_get_status() 84 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_hsm_handler()
|
| H A D | vcpu_sbi.c | 131 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_forward_handler() 185 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_load_reset_state() 201 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_return() 591 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_ecall()
|
| H A D | vcpu_onereg.c | 202 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_get_reg_core() 238 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_set_reg_core() 771 const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in num_fp_f_regs() 800 const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in num_fp_d_regs() 891 const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in copy_vector_reg_indices()
|
| H A D | vcpu_sbi_system.c | 15 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_susp_handler()
|
| H A D | vcpu_sbi_base.c | 19 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_base_handler()
|
| H A D | vcpu_insn.c | 333 struct kvm_cpu_context *ct; in kvm_riscv_vcpu_virtual_insn() 381 struct kvm_cpu_context *ct = &vcpu->arch.guest_context; in kvm_riscv_vcpu_mmio_load() 509 struct kvm_cpu_context *ct = &vcpu->arch.guest_context; in kvm_riscv_vcpu_mmio_store()
|
| H A D | vcpu_sbi_v01.c | 24 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_v01_handler()
|
| H A D | vcpu.c | 60 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_context_reset() 753 struct kvm_cpu_context *gcntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_enter_exit() 754 struct kvm_cpu_context *hcntx = &vcpu->arch.host_context; in kvm_riscv_vcpu_enter_exit()
|
| H A D | vcpu_sbi_fwft.c | 328 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_fwft_handler()
|
| /linux/arch/arm64/kernel/ |
| H A D | asm-offsets.c | 108 DEFINE(CPU_USER_PT_REGS, offsetof(struct kvm_cpu_context, regs)); in main() 109 DEFINE(CPU_ELR_EL2, offsetof(struct kvm_cpu_context, sys_regs[ELR_EL2])); in main() 110 DEFINE(CPU_RGSR_EL1, offsetof(struct kvm_cpu_context, sys_regs[RGSR_EL1])); in main() 111 DEFINE(CPU_GCR_EL1, offsetof(struct kvm_cpu_context, sys_regs[GCR_EL1])); in main() 112 DEFINE(CPU_APIAKEYLO_EL1, offsetof(struct kvm_cpu_context, sys_regs[APIAKEYLO_EL1])); in main() 113 DEFINE(CPU_APIBKEYLO_EL1, offsetof(struct kvm_cpu_context, sys_regs[APIBKEYLO_EL1])); in main() 114 DEFINE(CPU_APDAKEYLO_EL1, offsetof(struct kvm_cpu_context, sys_regs[APDAKEYLO_EL1])); in main() 115 DEFINE(CPU_APDBKEYLO_EL1, offsetof(struct kvm_cpu_context, sys_regs[APDBKEYLO_EL1])); in main() 116 DEFINE(CPU_APGAKEYLO_EL1, offsetof(struct kvm_cpu_context, sys_regs[APGAKEYLO_EL1])); in main() 117 DEFINE(HOST_CONTEXT_VCPU, offsetof(struct kvm_cpu_context, __hyp_running_vcpu)); in main()
|
| /linux/arch/arm64/kvm/hyp/vhe/ |
| H A D | switch.c | 33 DEFINE_PER_CPU(struct kvm_cpu_context, kvm_hyp_ctxt); 574 struct kvm_cpu_context *host_ctxt; in __kvm_vcpu_run_vhe() 575 struct kvm_cpu_context *guest_ctxt; in __kvm_vcpu_run_vhe() 660 struct kvm_cpu_context *host_ctxt; in __hyp_call_panic()
|
| /linux/arch/arm64/kvm/hyp/include/nvhe/ |
| H A D | ffa.h | 15 bool kvm_host_ffa_handler(struct kvm_cpu_context *host_ctxt, u32 func_id);
|
| /linux/arch/arm64/include/asm/ |
| H A D | kvm_host.h | 717 struct kvm_cpu_context { struct 770 struct kvm_cpu_context host_ctxt; 861 struct kvm_cpu_context ctxt; 1187 static inline u64 *___ctxt_sys_reg(const struct kvm_cpu_context *ctxt, int r) in ___ctxt_sys_reg() 1210 const struct kvm_cpu_context *ctxt = &(v)->arch.ctxt; \ 1220 const struct kvm_cpu_context *ctxt = &(v)->arch.ctxt; \ 1231 const struct kvm_cpu_context *ctxt = &(v)->arch.ctxt; \ 1435 static inline void kvm_init_host_cpu_context(struct kvm_cpu_context *cpu_ctxt) in kvm_init_host_cpu_context()
|
| H A D | kvm_asm.h | 319 struct kvm_cpu_context; 320 void handle_trap(struct kvm_cpu_context *host_ctxt);
|
| H A D | kvm_emulate.h | 180 static inline bool vcpu_is_el2_ctxt(const struct kvm_cpu_context *ctxt) in vcpu_is_el2_ctxt()
|
| /linux/arch/arm64/kvm/ |
| H A D | arm.c | 117 DECLARE_KVM_NVHE_PER_CPU(struct kvm_cpu_context, kvm_hyp_ctxt); 617 struct kvm_cpu_context *ctxt; in vcpu_set_pauth_traps() 2662 struct kvm_cpu_context *hyp_ctxt; in pkvm_hyp_init_ptrauth()
|