Searched refs:guest_context (Results 1 – 16 of 16) sorted by relevance
| /linux/arch/riscv/kvm/ |
| H A D | vcpu_exit.c | 74 old_hstatus = csr_swap(CSR_HSTATUS, vcpu->arch.guest_context.hstatus); in kvm_riscv_vcpu_unpriv_read() 146 if (vcpu->arch.guest_context.sstatus & SR_SPP) in kvm_riscv_vcpu_trap_redirect() 166 vcpu->arch.guest_context.sepc = ncsr_read(CSR_VSTVEC); in kvm_riscv_vcpu_trap_redirect() 169 vcpu->arch.guest_context.sstatus |= SR_SPP; in kvm_riscv_vcpu_trap_redirect() 176 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) { in vcpu_redirect() 229 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) in kvm_riscv_vcpu_exit() 235 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) in kvm_riscv_vcpu_exit() 239 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) in kvm_riscv_vcpu_exit() 254 vcpu->arch.guest_context.sepc, in kvm_riscv_vcpu_exit() 255 vcpu->arch.guest_context.sstatus, in kvm_riscv_vcpu_exit() [all …]
|
| H A D | vcpu_insn.c | 36 utrap.sepc = vcpu->arch.guest_context.sepc; in truly_illegal_insn() 52 utrap.sepc = vcpu->arch.guest_context.sepc; in truly_virtual_insn() 86 kvm_vcpu_on_spin(vcpu, vcpu->arch.guest_context.sstatus & SR_SPP); in wrs_insn() 138 SET_RD(insn, &vcpu->arch.guest_context, in kvm_riscv_vcpu_csr_return() 142 vcpu->arch.guest_context.sepc += INSN_LEN(insn); in kvm_riscv_vcpu_csr_return() 152 ulong rs1_val = GET_RS1(insn, &vcpu->arch.guest_context); in csr_insn() 292 vcpu->arch.guest_context.sepc += INSN_LEN(insn); in system_opcode_insn() 337 ct = &vcpu->arch.guest_context; in kvm_riscv_vcpu_virtual_insn() 381 struct kvm_cpu_context *ct = &vcpu->arch.guest_context; in kvm_riscv_vcpu_mmio_load() 509 struct kvm_cpu_context *ct = &vcpu->arch.guest_context; in kvm_riscv_vcpu_mmio_store() [all …]
|
| H A D | vcpu_vector.c | 22 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_vector_reset() 78 vcpu->arch.guest_context.vector.datap = kzalloc(riscv_v_vsize, GFP_KERNEL); in kvm_riscv_vcpu_alloc_vector_context() 79 if (!vcpu->arch.guest_context.vector.datap) in kvm_riscv_vcpu_alloc_vector_context() 91 kfree(vcpu->arch.guest_context.vector.datap); in kvm_riscv_vcpu_free_vector_context() 101 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_vreg_addr() 182 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_set_reg_vector()
|
| H A D | vcpu_sbi_replace.c | 20 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_time_handler() 51 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_ipi_handler() 95 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_rfence_handler() 151 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_srst_handler()
|
| H A D | vcpu_sbi_hsm.c | 18 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_hsm_vcpu_start() 65 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_hsm_vcpu_get_status() 84 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_hsm_handler()
|
| H A D | vcpu_fp.c | 20 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_fp_reset() 82 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_get_reg_fp() 131 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_set_reg_fp()
|
| H A D | vcpu.c | 58 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_context_reset() 226 return (vcpu->arch.guest_context.sstatus & SR_SPP) ? true : false; in kvm_arch_vcpu_in_kernel() 232 return vcpu->arch.guest_context.sepc; in kvm_arch_vcpu_get_ip() 630 kvm_riscv_vcpu_guest_fp_restore(&vcpu->arch.guest_context, in kvm_arch_vcpu_load() 633 kvm_riscv_vcpu_guest_vector_restore(&vcpu->arch.guest_context, in kvm_arch_vcpu_load() 652 kvm_riscv_vcpu_guest_fp_save(&vcpu->arch.guest_context, in kvm_arch_vcpu_put() 657 kvm_riscv_vcpu_guest_vector_save(&vcpu->arch.guest_context, in kvm_arch_vcpu_put() 788 struct kvm_cpu_context *gcntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_enter_exit()
|
| H A D | vcpu_sbi.c | 131 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_forward_handler() 185 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_load_reset_state() 201 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_return() 213 vcpu->arch.guest_context.sepc += 4; in kvm_riscv_vcpu_sbi_return() 591 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_ecall()
|
| H A D | vcpu_sbi_system.c | 15 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_susp_handler()
|
| H A D | trace.h | 25 __entry->pc = vcpu->arch.guest_context.sepc;
|
| H A D | vcpu_sbi_base.c | 19 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_base_handler()
|
| H A D | vcpu_onereg.c | 442 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_get_reg_core() 478 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_set_reg_core() 1009 const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in num_fp_f_regs() 1038 const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in num_fp_d_regs() 1129 const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in copy_vector_reg_indices()
|
| H A D | vcpu_sbi_v01.c | 24 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_v01_handler()
|
| H A D | vcpu_sbi_fwft.c | 328 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_fwft_handler()
|
| H A D | aia_imsic.c | 889 vcpu->arch.guest_context.hstatus &= ~HSTATUS_VGEIN; in kvm_riscv_vcpu_aia_imsic_update() 891 vcpu->arch.guest_context.hstatus |= in kvm_riscv_vcpu_aia_imsic_update()
|
| /linux/arch/riscv/include/asm/ |
| H A D | kvm_host.h | 212 struct kvm_cpu_context guest_context; member
|