Home
last modified time | relevance | path

Searched refs:guest_context (Results 1 – 16 of 16) sorted by relevance

/linux/arch/riscv/kvm/
H A Dvcpu_exit.c74 old_hstatus = csr_swap(CSR_HSTATUS, vcpu->arch.guest_context.hstatus); in kvm_riscv_vcpu_unpriv_read()
146 if (vcpu->arch.guest_context.sstatus & SR_SPP) in kvm_riscv_vcpu_trap_redirect()
166 vcpu->arch.guest_context.sepc = ncsr_read(CSR_VSTVEC); in kvm_riscv_vcpu_trap_redirect()
169 vcpu->arch.guest_context.sstatus |= SR_SPP; in kvm_riscv_vcpu_trap_redirect()
176 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) { in vcpu_redirect()
229 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) in kvm_riscv_vcpu_exit()
235 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) in kvm_riscv_vcpu_exit()
239 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) in kvm_riscv_vcpu_exit()
254 vcpu->arch.guest_context.sepc, in kvm_riscv_vcpu_exit()
255 vcpu->arch.guest_context.sstatus, in kvm_riscv_vcpu_exit()
[all …]
H A Dvcpu_insn.c36 utrap.sepc = vcpu->arch.guest_context.sepc; in truly_illegal_insn()
52 utrap.sepc = vcpu->arch.guest_context.sepc; in truly_virtual_insn()
86 kvm_vcpu_on_spin(vcpu, vcpu->arch.guest_context.sstatus & SR_SPP); in wrs_insn()
138 SET_RD(insn, &vcpu->arch.guest_context, in kvm_riscv_vcpu_csr_return()
142 vcpu->arch.guest_context.sepc += INSN_LEN(insn); in kvm_riscv_vcpu_csr_return()
152 ulong rs1_val = GET_RS1(insn, &vcpu->arch.guest_context); in csr_insn()
292 vcpu->arch.guest_context.sepc += INSN_LEN(insn); in system_opcode_insn()
337 ct = &vcpu->arch.guest_context; in kvm_riscv_vcpu_virtual_insn()
381 struct kvm_cpu_context *ct = &vcpu->arch.guest_context; in kvm_riscv_vcpu_mmio_load()
509 struct kvm_cpu_context *ct = &vcpu->arch.guest_context; in kvm_riscv_vcpu_mmio_store()
[all …]
H A Dvcpu_vector.c22 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_vector_reset()
78 vcpu->arch.guest_context.vector.datap = kzalloc(riscv_v_vsize, GFP_KERNEL); in kvm_riscv_vcpu_alloc_vector_context()
79 if (!vcpu->arch.guest_context.vector.datap) in kvm_riscv_vcpu_alloc_vector_context()
91 kfree(vcpu->arch.guest_context.vector.datap); in kvm_riscv_vcpu_free_vector_context()
101 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_vreg_addr()
182 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_set_reg_vector()
H A Dvcpu_sbi_replace.c20 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_time_handler()
51 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_ipi_handler()
95 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_rfence_handler()
151 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_srst_handler()
H A Dvcpu_sbi_hsm.c18 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_hsm_vcpu_start()
65 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_hsm_vcpu_get_status()
84 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_hsm_handler()
H A Dvcpu_fp.c20 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_fp_reset()
82 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_get_reg_fp()
131 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_set_reg_fp()
H A Dvcpu.c58 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_context_reset()
226 return (vcpu->arch.guest_context.sstatus & SR_SPP) ? true : false; in kvm_arch_vcpu_in_kernel()
232 return vcpu->arch.guest_context.sepc; in kvm_arch_vcpu_get_ip()
630 kvm_riscv_vcpu_guest_fp_restore(&vcpu->arch.guest_context, in kvm_arch_vcpu_load()
633 kvm_riscv_vcpu_guest_vector_restore(&vcpu->arch.guest_context, in kvm_arch_vcpu_load()
652 kvm_riscv_vcpu_guest_fp_save(&vcpu->arch.guest_context, in kvm_arch_vcpu_put()
657 kvm_riscv_vcpu_guest_vector_save(&vcpu->arch.guest_context, in kvm_arch_vcpu_put()
788 struct kvm_cpu_context *gcntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_enter_exit()
H A Dvcpu_sbi.c131 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_forward_handler()
185 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_load_reset_state()
201 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_return()
213 vcpu->arch.guest_context.sepc += 4; in kvm_riscv_vcpu_sbi_return()
591 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_ecall()
H A Dvcpu_sbi_system.c15 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_susp_handler()
H A Dtrace.h25 __entry->pc = vcpu->arch.guest_context.sepc;
H A Dvcpu_sbi_base.c19 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_base_handler()
H A Dvcpu_onereg.c442 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_get_reg_core()
478 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_set_reg_core()
1009 const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in num_fp_f_regs()
1038 const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in num_fp_d_regs()
1129 const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in copy_vector_reg_indices()
H A Dvcpu_sbi_v01.c24 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_v01_handler()
H A Dvcpu_sbi_fwft.c328 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_fwft_handler()
H A Daia_imsic.c889 vcpu->arch.guest_context.hstatus &= ~HSTATUS_VGEIN; in kvm_riscv_vcpu_aia_imsic_update()
891 vcpu->arch.guest_context.hstatus |= in kvm_riscv_vcpu_aia_imsic_update()
/linux/arch/riscv/include/asm/
H A Dkvm_host.h212 struct kvm_cpu_context guest_context; member