Lines Matching refs:ulong

124 	((ulong *)((ulong)(regs) + REG_OFFSET(insn, pos)))
153 int (*func)(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn);
157 ulong insn) in truly_illegal_insn()
173 ulong insn) in truly_virtual_insn()
202 static int wfi_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) in wfi_insn()
209 static int wrs_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) in wrs_insn()
255 ulong insn; in kvm_riscv_vcpu_csr_return()
273 static int csr_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) in csr_insn()
278 ulong rs1_val = GET_RS1(insn, &vcpu->arch.guest_context); in csr_insn()
280 ulong val = 0, wr_mask = 0, new_val = 0; in csr_insn()
399 ulong insn) in system_opcode_insn()
518 shift = 8 * (sizeof(ulong) - len); in kvm_riscv_vcpu_mmio_load()
521 shift = 8 * (sizeof(ulong) - len); in kvm_riscv_vcpu_mmio_load()
524 shift = 8 * (sizeof(ulong) - len); in kvm_riscv_vcpu_mmio_load()
528 shift = 8 * (sizeof(ulong) - len); in kvm_riscv_vcpu_mmio_load()
534 shift = 8 * (sizeof(ulong) - len); in kvm_riscv_vcpu_mmio_load()
540 shift = 8 * (sizeof(ulong) - len); in kvm_riscv_vcpu_mmio_load()
545 shift = 8 * (sizeof(ulong) - len); in kvm_riscv_vcpu_mmio_load()
549 shift = 8 * (sizeof(ulong) - len); in kvm_riscv_vcpu_mmio_load()
554 shift = 8 * (sizeof(ulong) - len); in kvm_riscv_vcpu_mmio_load()
611 ulong data; in kvm_riscv_vcpu_mmio_store()
737 ulong insn; in kvm_riscv_vcpu_mmio_return()
756 (ulong)data8 << shift >> shift); in kvm_riscv_vcpu_mmio_return()
761 (ulong)data16 << shift >> shift); in kvm_riscv_vcpu_mmio_return()
766 (ulong)data32 << shift >> shift); in kvm_riscv_vcpu_mmio_return()
771 (ulong)data64 << shift >> shift); in kvm_riscv_vcpu_mmio_return()