/* * This file is subject to the terms and conditions of the GNU General Public * License. See the file "COPYING" in the main directory of this archive * for more details. * * Copyright (C) 2007 by Ralf Baechle */ #include #include #include #include #include static u64 c0_hpt_read(struct clocksource *cs) { return read_c0_count(); } static struct clocksource clocksource_mips = { .name = "MIPS", .read = c0_hpt_read, .mask = CLOCKSOURCE_MASK(32), .flags = CLOCK_SOURCE_IS_CONTINUOUS | CLOCK_SOURCE_MUST_VERIFY | CLOCK_SOURCE_VERIFY_PERCPU, }; static u64 __maybe_unused notrace r4k_read_sched_clock(void) { return read_c0_count(); } static inline unsigned int rdhwr_count(void) { unsigned int count; __asm__ __volatile__( " .set push\n" " .set mips32r2\n" " rdhwr %0, $2\n" " .set pop\n" : "=r" (count)); return count; } static bool rdhwr_count_usable(void) { unsigned int prev, curr, i; /* * Older QEMUs have a broken implementation of RDHWR for the CP0 count * which always returns a constant value. Try to identify this and don't * use it in the VDSO if it is broken. This workaround can be removed * once the fix has been in QEMU stable for a reasonable amount of time. */ for (i = 0, prev = rdhwr_count(); i < 100; i++) { curr = rdhwr_count(); if (curr != prev) return true; prev = curr; } pr_warn("Not using R4K clocksource in VDSO due to broken RDHWR\n"); return false; } static inline __init bool count_can_be_sched_clock(void) { if (IS_ENABLED(CONFIG_CPU_FREQ)) return false; if (num_possible_cpus() > 1 && !IS_ENABLED(CONFIG_HAVE_UNSTABLE_SCHED_CLOCK)) return false; return true; } #ifdef CONFIG_CPU_FREQ static bool __read_mostly r4k_clock_unstable; static void r4k_clocksource_unstable(char *reason) { if (r4k_clock_unstable) return; r4k_clock_unstable = true; pr_info("R4K timer is unstable due to %s\n", reason); clocksource_mark_unstable(&clocksource_mips); } static int r4k_cpufreq_callback(struct notifier_block *nb, unsigned long val, void *data) { if (val == CPUFREQ_POSTCHANGE) r4k_clocksource_unstable("CPU frequency change"); return 0; } static struct notifier_block r4k_cpufreq_notifier = { .notifier_call = r4k_cpufreq_callback, }; static int __init r4k_register_cpufreq_notifier(void) { return cpufreq_register_notifier(&r4k_cpufreq_notifier, CPUFREQ_TRANSITION_NOTIFIER); } core_initcall(r4k_register_cpufreq_notifier); #endif /* !CONFIG_CPU_FREQ */ int __init init_r4k_clocksource(void) { if (!cpu_has_counter || !mips_hpt_frequency) return -ENXIO; /* Calculate a somewhat reasonable rating value */ clocksource_mips.rating = 200; clocksource_mips.rating += clamp(mips_hpt_frequency / 10000000, 0, 99); /* * R2 onwards makes the count accessible to user mode so it can be used * by the VDSO (HWREna is configured by configure_hwrena()). */ if (cpu_has_mips_r2_r6 && rdhwr_count_usable()) clocksource_mips.vdso_clock_mode = VDSO_CLOCKMODE_R4K; clocksource_register_hz(&clocksource_mips, mips_hpt_frequency); if (count_can_be_sched_clock()) sched_clock_register(r4k_read_sched_clock, 32, mips_hpt_frequency); return 0; }