Lines Matching refs:tsc_timestamp

2783 static inline u64 vgettsc(struct pvclock_clock *clock, u64 *tsc_timestamp,  in vgettsc()  argument
2792 tsc_timestamp, &tsc_pg_val)) { in vgettsc()
2804 *tsc_timestamp = read_tsc(); in vgettsc()
2805 v = (*tsc_timestamp - clock->cycle_last) & in vgettsc()
2813 *tsc_timestamp = v = 0; in vgettsc()
2822 static int do_kvmclock_base(s64 *t, u64 *tsc_timestamp) in do_kvmclock_base() argument
2832 ns += vgettsc(&gtod->raw_clock, tsc_timestamp, &mode); in do_kvmclock_base()
2845 static int do_monotonic(s64 *t, u64 *tsc_timestamp) in do_monotonic() argument
2855 ns += vgettsc(&gtod->clock, tsc_timestamp, &mode); in do_monotonic()
2864 static int do_realtime(struct timespec64 *ts, u64 *tsc_timestamp) in do_realtime() argument
2875 ns += vgettsc(&gtod->clock, tsc_timestamp, &mode); in do_realtime()
2890 static bool kvm_get_time_and_clockread(s64 *kernel_ns, u64 *tsc_timestamp) in kvm_get_time_and_clockread() argument
2897 tsc_timestamp)); in kvm_get_time_and_clockread()
2904 bool kvm_get_monotonic_and_clockread(s64 *kernel_ns, u64 *tsc_timestamp) in kvm_get_monotonic_and_clockread() argument
2911 tsc_timestamp)); in kvm_get_monotonic_and_clockread()
2922 u64 *tsc_timestamp) in kvm_get_walltime_and_clockread() argument
2928 return gtod_is_based_on_tsc(do_realtime(ts, tsc_timestamp)); in kvm_get_walltime_and_clockread()
3087 hv_clock.tsc_timestamp = ka->master_cycle_now; in __get_kvmclock()
3180 u64 tsc_timestamp, host_tsc; in kvm_guest_time_update() local
3223 tsc_timestamp = kvm_read_l1_tsc(v, host_tsc); in kvm_guest_time_update()
3237 if (tsc > tsc_timestamp) { in kvm_guest_time_update()
3238 adjust_tsc_offset_guest(v, tsc - tsc_timestamp); in kvm_guest_time_update()
3239 tsc_timestamp = tsc; in kvm_guest_time_update()
3259 vcpu->hv_clock.tsc_timestamp = tsc_timestamp; in kvm_guest_time_update()
3261 vcpu->last_guest_tsc = tsc_timestamp; in kvm_guest_time_update()
3339 hv_clock.tsc_timestamp = ka->master_cycle_now; in kvm_get_wall_clock_epoch()