xref: /linux/arch/riscv/kvm/vcpu_timer.c (revision 9958d30f38b96fb763a10d44d18ddad39127d5f4)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (C) 2019 Western Digital Corporation or its affiliates.
4  *
5  * Authors:
6  *     Atish Patra <atish.patra@wdc.com>
7  */
8 
9 #include <linux/errno.h>
10 #include <linux/err.h>
11 #include <linux/kvm_host.h>
12 #include <linux/uaccess.h>
13 #include <clocksource/timer-riscv.h>
14 #include <asm/csr.h>
15 #include <asm/delay.h>
16 #include <asm/kvm_vcpu_timer.h>
17 
18 static u64 kvm_riscv_current_cycles(struct kvm_guest_timer *gt)
19 {
20 	return get_cycles64() + gt->time_delta;
21 }
22 
23 static u64 kvm_riscv_delta_cycles2ns(u64 cycles,
24 				     struct kvm_guest_timer *gt,
25 				     struct kvm_vcpu_timer *t)
26 {
27 	unsigned long flags;
28 	u64 cycles_now, cycles_delta, delta_ns;
29 
30 	local_irq_save(flags);
31 	cycles_now = kvm_riscv_current_cycles(gt);
32 	if (cycles_now < cycles)
33 		cycles_delta = cycles - cycles_now;
34 	else
35 		cycles_delta = 0;
36 	delta_ns = (cycles_delta * gt->nsec_mult) >> gt->nsec_shift;
37 	local_irq_restore(flags);
38 
39 	return delta_ns;
40 }
41 
42 static enum hrtimer_restart kvm_riscv_vcpu_hrtimer_expired(struct hrtimer *h)
43 {
44 	u64 delta_ns;
45 	struct kvm_vcpu_timer *t = container_of(h, struct kvm_vcpu_timer, hrt);
46 	struct kvm_vcpu *vcpu = container_of(t, struct kvm_vcpu, arch.timer);
47 	struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
48 
49 	if (kvm_riscv_current_cycles(gt) < t->next_cycles) {
50 		delta_ns = kvm_riscv_delta_cycles2ns(t->next_cycles, gt, t);
51 		hrtimer_forward_now(&t->hrt, ktime_set(0, delta_ns));
52 		return HRTIMER_RESTART;
53 	}
54 
55 	t->next_set = false;
56 	kvm_riscv_vcpu_set_interrupt(vcpu, IRQ_VS_TIMER);
57 
58 	return HRTIMER_NORESTART;
59 }
60 
61 static int kvm_riscv_vcpu_timer_cancel(struct kvm_vcpu_timer *t)
62 {
63 	if (!t->init_done || !t->next_set)
64 		return -EINVAL;
65 
66 	hrtimer_cancel(&t->hrt);
67 	t->next_set = false;
68 
69 	return 0;
70 }
71 
72 int kvm_riscv_vcpu_timer_next_event(struct kvm_vcpu *vcpu, u64 ncycles)
73 {
74 	struct kvm_vcpu_timer *t = &vcpu->arch.timer;
75 	struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
76 	u64 delta_ns;
77 
78 	if (!t->init_done)
79 		return -EINVAL;
80 
81 	kvm_riscv_vcpu_unset_interrupt(vcpu, IRQ_VS_TIMER);
82 
83 	delta_ns = kvm_riscv_delta_cycles2ns(ncycles, gt, t);
84 	t->next_cycles = ncycles;
85 	hrtimer_start(&t->hrt, ktime_set(0, delta_ns), HRTIMER_MODE_REL);
86 	t->next_set = true;
87 
88 	return 0;
89 }
90 
91 int kvm_riscv_vcpu_get_reg_timer(struct kvm_vcpu *vcpu,
92 				 const struct kvm_one_reg *reg)
93 {
94 	struct kvm_vcpu_timer *t = &vcpu->arch.timer;
95 	struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
96 	u64 __user *uaddr = (u64 __user *)(unsigned long)reg->addr;
97 	unsigned long reg_num = reg->id & ~(KVM_REG_ARCH_MASK |
98 					    KVM_REG_SIZE_MASK |
99 					    KVM_REG_RISCV_TIMER);
100 	u64 reg_val;
101 
102 	if (KVM_REG_SIZE(reg->id) != sizeof(u64))
103 		return -EINVAL;
104 	if (reg_num >= sizeof(struct kvm_riscv_timer) / sizeof(u64))
105 		return -EINVAL;
106 
107 	switch (reg_num) {
108 	case KVM_REG_RISCV_TIMER_REG(frequency):
109 		reg_val = riscv_timebase;
110 		break;
111 	case KVM_REG_RISCV_TIMER_REG(time):
112 		reg_val = kvm_riscv_current_cycles(gt);
113 		break;
114 	case KVM_REG_RISCV_TIMER_REG(compare):
115 		reg_val = t->next_cycles;
116 		break;
117 	case KVM_REG_RISCV_TIMER_REG(state):
118 		reg_val = (t->next_set) ? KVM_RISCV_TIMER_STATE_ON :
119 					  KVM_RISCV_TIMER_STATE_OFF;
120 		break;
121 	default:
122 		return -EINVAL;
123 	}
124 
125 	if (copy_to_user(uaddr, &reg_val, KVM_REG_SIZE(reg->id)))
126 		return -EFAULT;
127 
128 	return 0;
129 }
130 
131 int kvm_riscv_vcpu_set_reg_timer(struct kvm_vcpu *vcpu,
132 				 const struct kvm_one_reg *reg)
133 {
134 	struct kvm_vcpu_timer *t = &vcpu->arch.timer;
135 	struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
136 	u64 __user *uaddr = (u64 __user *)(unsigned long)reg->addr;
137 	unsigned long reg_num = reg->id & ~(KVM_REG_ARCH_MASK |
138 					    KVM_REG_SIZE_MASK |
139 					    KVM_REG_RISCV_TIMER);
140 	u64 reg_val;
141 	int ret = 0;
142 
143 	if (KVM_REG_SIZE(reg->id) != sizeof(u64))
144 		return -EINVAL;
145 	if (reg_num >= sizeof(struct kvm_riscv_timer) / sizeof(u64))
146 		return -EINVAL;
147 
148 	if (copy_from_user(&reg_val, uaddr, KVM_REG_SIZE(reg->id)))
149 		return -EFAULT;
150 
151 	switch (reg_num) {
152 	case KVM_REG_RISCV_TIMER_REG(frequency):
153 		ret = -EOPNOTSUPP;
154 		break;
155 	case KVM_REG_RISCV_TIMER_REG(time):
156 		gt->time_delta = reg_val - get_cycles64();
157 		break;
158 	case KVM_REG_RISCV_TIMER_REG(compare):
159 		t->next_cycles = reg_val;
160 		break;
161 	case KVM_REG_RISCV_TIMER_REG(state):
162 		if (reg_val == KVM_RISCV_TIMER_STATE_ON)
163 			ret = kvm_riscv_vcpu_timer_next_event(vcpu, reg_val);
164 		else
165 			ret = kvm_riscv_vcpu_timer_cancel(t);
166 		break;
167 	default:
168 		ret = -EINVAL;
169 		break;
170 	}
171 
172 	return ret;
173 }
174 
175 int kvm_riscv_vcpu_timer_init(struct kvm_vcpu *vcpu)
176 {
177 	struct kvm_vcpu_timer *t = &vcpu->arch.timer;
178 
179 	if (t->init_done)
180 		return -EINVAL;
181 
182 	hrtimer_init(&t->hrt, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
183 	t->hrt.function = kvm_riscv_vcpu_hrtimer_expired;
184 	t->init_done = true;
185 	t->next_set = false;
186 
187 	return 0;
188 }
189 
190 int kvm_riscv_vcpu_timer_deinit(struct kvm_vcpu *vcpu)
191 {
192 	int ret;
193 
194 	ret = kvm_riscv_vcpu_timer_cancel(&vcpu->arch.timer);
195 	vcpu->arch.timer.init_done = false;
196 
197 	return ret;
198 }
199 
200 int kvm_riscv_vcpu_timer_reset(struct kvm_vcpu *vcpu)
201 {
202 	return kvm_riscv_vcpu_timer_cancel(&vcpu->arch.timer);
203 }
204 
205 void kvm_riscv_vcpu_timer_restore(struct kvm_vcpu *vcpu)
206 {
207 	struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
208 
209 #ifdef CONFIG_64BIT
210 	csr_write(CSR_HTIMEDELTA, gt->time_delta);
211 #else
212 	csr_write(CSR_HTIMEDELTA, (u32)(gt->time_delta));
213 	csr_write(CSR_HTIMEDELTAH, (u32)(gt->time_delta >> 32));
214 #endif
215 }
216 
217 int kvm_riscv_guest_timer_init(struct kvm *kvm)
218 {
219 	struct kvm_guest_timer *gt = &kvm->arch.timer;
220 
221 	riscv_cs_get_mult_shift(&gt->nsec_mult, &gt->nsec_shift);
222 	gt->time_delta = -get_cycles64();
223 
224 	return 0;
225 }
226