1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3 * spu aware cpufreq governor for the cell processor
4 *
5 * © Copyright IBM Corporation 2006-2008
6 *
7 * Author: Christian Krafft <krafft@de.ibm.com>
8 */
9
10 #include <linux/cpufreq.h>
11 #include <linux/sched.h>
12 #include <linux/sched/loadavg.h>
13 #include <linux/module.h>
14 #include <linux/timer.h>
15 #include <linux/workqueue.h>
16 #include <linux/atomic.h>
17 #include <asm/machdep.h>
18 #include <asm/spu.h>
19
20 #define POLL_TIME 100000 /* in µs */
21 #define EXP 753 /* exp(-1) in fixed-point */
22
23 struct spu_gov_info_struct {
24 unsigned long busy_spus; /* fixed-point */
25 struct cpufreq_policy *policy;
26 struct delayed_work work;
27 unsigned int poll_int; /* µs */
28 };
29 static DEFINE_PER_CPU(struct spu_gov_info_struct, spu_gov_info);
30
calc_freq(struct spu_gov_info_struct * info)31 static int calc_freq(struct spu_gov_info_struct *info)
32 {
33 int cpu;
34 int busy_spus;
35
36 cpu = info->policy->cpu;
37 busy_spus = atomic_read(&cbe_spu_info[cpu_to_node(cpu)].busy_spus);
38
39 info->busy_spus = calc_load(info->busy_spus, EXP, busy_spus * FIXED_1);
40 pr_debug("cpu %d: busy_spus=%d, info->busy_spus=%ld\n",
41 cpu, busy_spus, info->busy_spus);
42
43 return info->policy->max * info->busy_spus / FIXED_1;
44 }
45
spu_gov_work(struct work_struct * work)46 static void spu_gov_work(struct work_struct *work)
47 {
48 struct spu_gov_info_struct *info;
49 int delay;
50 unsigned long target_freq;
51
52 info = container_of(work, struct spu_gov_info_struct, work.work);
53
54 /* after cancel_delayed_work_sync we unset info->policy */
55 BUG_ON(info->policy == NULL);
56
57 target_freq = calc_freq(info);
58 __cpufreq_driver_target(info->policy, target_freq, CPUFREQ_RELATION_H);
59
60 delay = usecs_to_jiffies(info->poll_int);
61 schedule_delayed_work_on(info->policy->cpu, &info->work, delay);
62 }
63
spu_gov_init_work(struct spu_gov_info_struct * info)64 static void spu_gov_init_work(struct spu_gov_info_struct *info)
65 {
66 int delay = usecs_to_jiffies(info->poll_int);
67 INIT_DEFERRABLE_WORK(&info->work, spu_gov_work);
68 schedule_delayed_work_on(info->policy->cpu, &info->work, delay);
69 }
70
spu_gov_cancel_work(struct spu_gov_info_struct * info)71 static void spu_gov_cancel_work(struct spu_gov_info_struct *info)
72 {
73 cancel_delayed_work_sync(&info->work);
74 }
75
spu_gov_start(struct cpufreq_policy * policy)76 static int spu_gov_start(struct cpufreq_policy *policy)
77 {
78 unsigned int cpu = policy->cpu;
79 struct spu_gov_info_struct *info = &per_cpu(spu_gov_info, cpu);
80 struct spu_gov_info_struct *affected_info;
81 int i;
82
83 if (!cpu_online(cpu)) {
84 printk(KERN_ERR "cpu %d is not online\n", cpu);
85 return -EINVAL;
86 }
87
88 if (!policy->cur) {
89 printk(KERN_ERR "no cpu specified in policy\n");
90 return -EINVAL;
91 }
92
93 /* initialize spu_gov_info for all affected cpus */
94 for_each_cpu(i, policy->cpus) {
95 affected_info = &per_cpu(spu_gov_info, i);
96 affected_info->policy = policy;
97 }
98
99 info->poll_int = POLL_TIME;
100
101 /* setup timer */
102 spu_gov_init_work(info);
103
104 return 0;
105 }
106
spu_gov_stop(struct cpufreq_policy * policy)107 static void spu_gov_stop(struct cpufreq_policy *policy)
108 {
109 unsigned int cpu = policy->cpu;
110 struct spu_gov_info_struct *info = &per_cpu(spu_gov_info, cpu);
111 int i;
112
113 /* cancel timer */
114 spu_gov_cancel_work(info);
115
116 /* clean spu_gov_info for all affected cpus */
117 for_each_cpu (i, policy->cpus) {
118 info = &per_cpu(spu_gov_info, i);
119 info->policy = NULL;
120 }
121 }
122
123 static struct cpufreq_governor spu_governor = {
124 .name = "spudemand",
125 .start = spu_gov_start,
126 .stop = spu_gov_stop,
127 .owner = THIS_MODULE,
128 };
129 cpufreq_governor_init(spu_governor);
130 cpufreq_governor_exit(spu_governor);
131
132 MODULE_DESCRIPTION("SPU-aware cpufreq governor for the cell processor");
133 MODULE_LICENSE("GPL");
134 MODULE_AUTHOR("Christian Krafft <krafft@de.ibm.com>");
135