xref: /linux/drivers/media/platform/qcom/venus/pm_helpers.c (revision fcab107abe1ab5be9dbe874baa722372da8f4f73)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (C) 2019 Linaro Ltd.
4  *
5  * Author: Stanimir Varbanov <stanimir.varbanov@linaro.org>
6  */
7 #include <linux/clk.h>
8 #include <linux/interconnect.h>
9 #include <linux/iopoll.h>
10 #include <linux/kernel.h>
11 #include <linux/pm_domain.h>
12 #include <linux/pm_opp.h>
13 #include <linux/pm_runtime.h>
14 #include <linux/reset.h>
15 #include <linux/types.h>
16 #include <media/v4l2-mem2mem.h>
17 
18 #include "core.h"
19 #include "hfi_parser.h"
20 #include "hfi_venus_io.h"
21 #include "pm_helpers.h"
22 #include "hfi_platform.h"
23 
24 static bool legacy_binding;
25 
26 static int core_clks_get(struct venus_core *core)
27 {
28 	const struct venus_resources *res = core->res;
29 	struct device *dev = core->dev;
30 	unsigned int i;
31 
32 	for (i = 0; i < res->clks_num; i++) {
33 		core->clks[i] = devm_clk_get(dev, res->clks[i]);
34 		if (IS_ERR(core->clks[i]))
35 			return PTR_ERR(core->clks[i]);
36 	}
37 
38 	return 0;
39 }
40 
41 static int core_clks_enable(struct venus_core *core)
42 {
43 	const struct venus_resources *res = core->res;
44 	const struct freq_tbl *freq_tbl = core->res->freq_tbl;
45 	unsigned int freq_tbl_size = core->res->freq_tbl_size;
46 	unsigned long freq;
47 	unsigned int i;
48 	int ret;
49 
50 	if (!freq_tbl)
51 		return -EINVAL;
52 
53 	freq = freq_tbl[freq_tbl_size - 1].freq;
54 
55 	for (i = 0; i < res->clks_num; i++) {
56 		if (IS_V6(core)) {
57 			ret = clk_set_rate(core->clks[i], freq);
58 			if (ret)
59 				goto err;
60 		}
61 
62 		ret = clk_prepare_enable(core->clks[i]);
63 		if (ret)
64 			goto err;
65 	}
66 
67 	return 0;
68 err:
69 	while (i--)
70 		clk_disable_unprepare(core->clks[i]);
71 
72 	return ret;
73 }
74 
75 static void core_clks_disable(struct venus_core *core)
76 {
77 	const struct venus_resources *res = core->res;
78 	unsigned int i = res->clks_num;
79 
80 	while (i--)
81 		clk_disable_unprepare(core->clks[i]);
82 }
83 
84 static int core_clks_set_rate(struct venus_core *core, unsigned long freq)
85 {
86 	int ret;
87 
88 	ret = dev_pm_opp_set_rate(core->dev, freq);
89 	if (ret)
90 		return ret;
91 
92 	ret = clk_set_rate(core->vcodec0_clks[0], freq);
93 	if (ret)
94 		return ret;
95 
96 	ret = clk_set_rate(core->vcodec1_clks[0], freq);
97 	if (ret)
98 		return ret;
99 
100 	return 0;
101 }
102 
103 static int vcodec_clks_get(struct venus_core *core, struct device *dev,
104 			   struct clk **clks, const char * const *id)
105 {
106 	const struct venus_resources *res = core->res;
107 	unsigned int i;
108 
109 	for (i = 0; i < res->vcodec_clks_num; i++) {
110 		if (!id[i])
111 			continue;
112 		clks[i] = devm_clk_get(dev, id[i]);
113 		if (IS_ERR(clks[i]))
114 			return PTR_ERR(clks[i]);
115 	}
116 
117 	return 0;
118 }
119 
120 static int vcodec_clks_enable(struct venus_core *core, struct clk **clks)
121 {
122 	const struct venus_resources *res = core->res;
123 	unsigned int i;
124 	int ret;
125 
126 	for (i = 0; i < res->vcodec_clks_num; i++) {
127 		ret = clk_prepare_enable(clks[i]);
128 		if (ret)
129 			goto err;
130 	}
131 
132 	return 0;
133 err:
134 	while (i--)
135 		clk_disable_unprepare(clks[i]);
136 
137 	return ret;
138 }
139 
140 static void vcodec_clks_disable(struct venus_core *core, struct clk **clks)
141 {
142 	const struct venus_resources *res = core->res;
143 	unsigned int i = res->vcodec_clks_num;
144 
145 	while (i--)
146 		clk_disable_unprepare(clks[i]);
147 }
148 
149 static u32 load_per_instance(struct venus_inst *inst)
150 {
151 	u32 mbs;
152 
153 	if (!inst || !(inst->state >= INST_INIT && inst->state < INST_STOP))
154 		return 0;
155 
156 	mbs = (ALIGN(inst->width, 16) / 16) * (ALIGN(inst->height, 16) / 16);
157 
158 	return mbs * inst->fps;
159 }
160 
161 static u32 load_per_type(struct venus_core *core, u32 session_type)
162 {
163 	struct venus_inst *inst = NULL;
164 	u32 mbs_per_sec = 0;
165 
166 	list_for_each_entry(inst, &core->instances, list) {
167 		if (inst->session_type != session_type)
168 			continue;
169 
170 		mbs_per_sec += load_per_instance(inst);
171 	}
172 
173 	return mbs_per_sec;
174 }
175 
176 static void mbs_to_bw(struct venus_inst *inst, u32 mbs, u32 *avg, u32 *peak)
177 {
178 	const struct venus_resources *res = inst->core->res;
179 	const struct bw_tbl *bw_tbl;
180 	unsigned int num_rows, i;
181 
182 	*avg = 0;
183 	*peak = 0;
184 
185 	if (mbs == 0)
186 		return;
187 
188 	if (inst->session_type == VIDC_SESSION_TYPE_ENC) {
189 		num_rows = res->bw_tbl_enc_size;
190 		bw_tbl = res->bw_tbl_enc;
191 	} else if (inst->session_type == VIDC_SESSION_TYPE_DEC) {
192 		num_rows = res->bw_tbl_dec_size;
193 		bw_tbl = res->bw_tbl_dec;
194 	} else {
195 		return;
196 	}
197 
198 	if (!bw_tbl || num_rows == 0)
199 		return;
200 
201 	for (i = 0; i < num_rows; i++) {
202 		if (i != 0 && mbs > bw_tbl[i].mbs_per_sec)
203 			break;
204 
205 		if (inst->dpb_fmt & HFI_COLOR_FORMAT_10_BIT_BASE) {
206 			*avg = bw_tbl[i].avg_10bit;
207 			*peak = bw_tbl[i].peak_10bit;
208 		} else {
209 			*avg = bw_tbl[i].avg;
210 			*peak = bw_tbl[i].peak;
211 		}
212 	}
213 }
214 
215 static int load_scale_bw(struct venus_core *core)
216 {
217 	struct venus_inst *inst = NULL;
218 	u32 mbs_per_sec, avg, peak, total_avg = 0, total_peak = 0;
219 
220 	list_for_each_entry(inst, &core->instances, list) {
221 		mbs_per_sec = load_per_instance(inst);
222 		mbs_to_bw(inst, mbs_per_sec, &avg, &peak);
223 		total_avg += avg;
224 		total_peak += peak;
225 	}
226 
227 	/*
228 	 * keep minimum bandwidth vote for "video-mem" path,
229 	 * so that clks can be disabled during vdec_session_release().
230 	 * Actual bandwidth drop will be done during device supend
231 	 * so that device can power down without any warnings.
232 	 */
233 
234 	if (!total_avg && !total_peak)
235 		total_avg = kbps_to_icc(1000);
236 
237 	dev_dbg(core->dev, VDBGL "total: avg_bw: %u, peak_bw: %u\n",
238 		total_avg, total_peak);
239 
240 	return icc_set_bw(core->video_path, total_avg, total_peak);
241 }
242 
243 static int load_scale_v1(struct venus_inst *inst)
244 {
245 	struct venus_core *core = inst->core;
246 	const struct freq_tbl *table = core->res->freq_tbl;
247 	unsigned int num_rows = core->res->freq_tbl_size;
248 	unsigned long freq = table[0].freq;
249 	struct device *dev = core->dev;
250 	u32 mbs_per_sec;
251 	unsigned int i;
252 	int ret = 0;
253 
254 	mutex_lock(&core->lock);
255 	mbs_per_sec = load_per_type(core, VIDC_SESSION_TYPE_ENC) +
256 		      load_per_type(core, VIDC_SESSION_TYPE_DEC);
257 
258 	if (mbs_per_sec > core->res->max_load)
259 		dev_warn(dev, "HW is overloaded, needed: %d max: %d\n",
260 			 mbs_per_sec, core->res->max_load);
261 
262 	if (!mbs_per_sec && num_rows > 1) {
263 		freq = table[num_rows - 1].freq;
264 		goto set_freq;
265 	}
266 
267 	for (i = 0; i < num_rows; i++) {
268 		if (mbs_per_sec > table[i].load)
269 			break;
270 		freq = table[i].freq;
271 	}
272 
273 set_freq:
274 
275 	ret = core_clks_set_rate(core, freq);
276 	if (ret) {
277 		dev_err(dev, "failed to set clock rate %lu (%d)\n",
278 			freq, ret);
279 		goto exit;
280 	}
281 
282 	ret = load_scale_bw(core);
283 	if (ret) {
284 		dev_err(dev, "failed to set bandwidth (%d)\n",
285 			ret);
286 		goto exit;
287 	}
288 
289 exit:
290 	mutex_unlock(&core->lock);
291 	return ret;
292 }
293 
294 static int core_get_v1(struct venus_core *core)
295 {
296 	int ret;
297 
298 	ret = core_clks_get(core);
299 	if (ret)
300 		return ret;
301 
302 	ret = devm_pm_opp_set_clkname(core->dev, "core");
303 	if (ret)
304 		return ret;
305 
306 	return 0;
307 }
308 
309 static void core_put_v1(struct venus_core *core)
310 {
311 }
312 
313 static int core_power_v1(struct venus_core *core, int on)
314 {
315 	int ret = 0;
316 
317 	if (on == POWER_ON)
318 		ret = core_clks_enable(core);
319 	else
320 		core_clks_disable(core);
321 
322 	return ret;
323 }
324 
325 static const struct venus_pm_ops pm_ops_v1 = {
326 	.core_get = core_get_v1,
327 	.core_put = core_put_v1,
328 	.core_power = core_power_v1,
329 	.load_scale = load_scale_v1,
330 };
331 
332 static void
333 vcodec_control_v3(struct venus_core *core, u32 session_type, bool enable)
334 {
335 	void __iomem *ctrl;
336 
337 	if (session_type == VIDC_SESSION_TYPE_DEC)
338 		ctrl = core->wrapper_base + WRAPPER_VDEC_VCODEC_POWER_CONTROL;
339 	else
340 		ctrl = core->wrapper_base + WRAPPER_VENC_VCODEC_POWER_CONTROL;
341 
342 	if (enable)
343 		writel(0, ctrl);
344 	else
345 		writel(1, ctrl);
346 }
347 
348 static int vdec_get_v3(struct device *dev)
349 {
350 	struct venus_core *core = dev_get_drvdata(dev);
351 
352 	return vcodec_clks_get(core, dev, core->vcodec0_clks,
353 			       core->res->vcodec0_clks);
354 }
355 
356 static int vdec_power_v3(struct device *dev, int on)
357 {
358 	struct venus_core *core = dev_get_drvdata(dev);
359 	int ret = 0;
360 
361 	vcodec_control_v3(core, VIDC_SESSION_TYPE_DEC, true);
362 
363 	if (on == POWER_ON)
364 		ret = vcodec_clks_enable(core, core->vcodec0_clks);
365 	else
366 		vcodec_clks_disable(core, core->vcodec0_clks);
367 
368 	vcodec_control_v3(core, VIDC_SESSION_TYPE_DEC, false);
369 
370 	return ret;
371 }
372 
373 static int venc_get_v3(struct device *dev)
374 {
375 	struct venus_core *core = dev_get_drvdata(dev);
376 
377 	return vcodec_clks_get(core, dev, core->vcodec1_clks,
378 			       core->res->vcodec1_clks);
379 }
380 
381 static int venc_power_v3(struct device *dev, int on)
382 {
383 	struct venus_core *core = dev_get_drvdata(dev);
384 	int ret = 0;
385 
386 	vcodec_control_v3(core, VIDC_SESSION_TYPE_ENC, true);
387 
388 	if (on == POWER_ON)
389 		ret = vcodec_clks_enable(core, core->vcodec1_clks);
390 	else
391 		vcodec_clks_disable(core, core->vcodec1_clks);
392 
393 	vcodec_control_v3(core, VIDC_SESSION_TYPE_ENC, false);
394 
395 	return ret;
396 }
397 
398 static const struct venus_pm_ops pm_ops_v3 = {
399 	.core_get = core_get_v1,
400 	.core_put = core_put_v1,
401 	.core_power = core_power_v1,
402 	.vdec_get = vdec_get_v3,
403 	.vdec_power = vdec_power_v3,
404 	.venc_get = venc_get_v3,
405 	.venc_power = venc_power_v3,
406 	.load_scale = load_scale_v1,
407 };
408 
409 static int vcodec_control_v4(struct venus_core *core, u32 coreid, bool enable)
410 {
411 	void __iomem *ctrl, *stat;
412 	u32 val;
413 	int ret;
414 
415 	ret = dev_pm_genpd_set_hwmode(core->pmdomains->pd_devs[coreid], !enable);
416 	if (ret == -EOPNOTSUPP) {
417 		core->hwmode_dev = false;
418 		goto legacy;
419 	}
420 
421 	core->hwmode_dev = true;
422 	return ret;
423 
424 legacy:
425 	if (coreid == VIDC_CORE_ID_1) {
426 		ctrl = core->wrapper_base + WRAPPER_VCODEC0_MMCC_POWER_CONTROL;
427 		stat = core->wrapper_base + WRAPPER_VCODEC0_MMCC_POWER_STATUS;
428 	} else {
429 		ctrl = core->wrapper_base + WRAPPER_VCODEC1_MMCC_POWER_CONTROL;
430 		stat = core->wrapper_base + WRAPPER_VCODEC1_MMCC_POWER_STATUS;
431 	}
432 
433 	if (enable) {
434 		writel(0, ctrl);
435 
436 		ret = readl_poll_timeout(stat, val, val & BIT(1), 1, 100);
437 		if (ret)
438 			return ret;
439 	} else {
440 		writel(1, ctrl);
441 
442 		ret = readl_poll_timeout(stat, val, !(val & BIT(1)), 1, 100);
443 		if (ret)
444 			return ret;
445 	}
446 
447 	return 0;
448 }
449 
450 static int poweroff_coreid(struct venus_core *core, unsigned int coreid_mask)
451 {
452 	int ret;
453 
454 	if (coreid_mask & VIDC_CORE_ID_1) {
455 		ret = vcodec_control_v4(core, VIDC_CORE_ID_1, true);
456 		if (ret)
457 			return ret;
458 
459 		vcodec_clks_disable(core, core->vcodec0_clks);
460 
461 		if (!core->hwmode_dev) {
462 			ret = vcodec_control_v4(core, VIDC_CORE_ID_1, false);
463 			if (ret)
464 				return ret;
465 		}
466 
467 		ret = pm_runtime_put_sync(core->pmdomains->pd_devs[1]);
468 		if (ret < 0)
469 			return ret;
470 	}
471 
472 	if (coreid_mask & VIDC_CORE_ID_2) {
473 		ret = vcodec_control_v4(core, VIDC_CORE_ID_2, true);
474 		if (ret)
475 			return ret;
476 
477 		vcodec_clks_disable(core, core->vcodec1_clks);
478 
479 		if (!core->hwmode_dev) {
480 			ret = vcodec_control_v4(core, VIDC_CORE_ID_2, false);
481 			if (ret)
482 				return ret;
483 		}
484 
485 		ret = pm_runtime_put_sync(core->pmdomains->pd_devs[2]);
486 		if (ret < 0)
487 			return ret;
488 	}
489 
490 	return 0;
491 }
492 
493 static int poweron_coreid(struct venus_core *core, unsigned int coreid_mask)
494 {
495 	int ret;
496 
497 	if (coreid_mask & VIDC_CORE_ID_1) {
498 		ret = pm_runtime_get_sync(core->pmdomains->pd_devs[1]);
499 		if (ret < 0)
500 			return ret;
501 
502 		ret = vcodec_control_v4(core, VIDC_CORE_ID_1, true);
503 		if (ret)
504 			return ret;
505 
506 		ret = vcodec_clks_enable(core, core->vcodec0_clks);
507 		if (ret)
508 			return ret;
509 
510 		ret = vcodec_control_v4(core, VIDC_CORE_ID_1, false);
511 		if (ret < 0)
512 			return ret;
513 	}
514 
515 	if (coreid_mask & VIDC_CORE_ID_2) {
516 		ret = pm_runtime_get_sync(core->pmdomains->pd_devs[2]);
517 		if (ret < 0)
518 			return ret;
519 
520 		ret = vcodec_control_v4(core, VIDC_CORE_ID_2, true);
521 		if (ret)
522 			return ret;
523 
524 		ret = vcodec_clks_enable(core, core->vcodec1_clks);
525 		if (ret)
526 			return ret;
527 
528 		ret = vcodec_control_v4(core, VIDC_CORE_ID_2, false);
529 		if (ret < 0)
530 			return ret;
531 	}
532 
533 	return 0;
534 }
535 
536 static inline int power_save_mode_enable(struct venus_inst *inst,
537 					 bool enable)
538 {
539 	struct venc_controls *enc_ctr = &inst->controls.enc;
540 	const u32 ptype = HFI_PROPERTY_CONFIG_VENC_PERF_MODE;
541 	u32 venc_mode;
542 	int ret = 0;
543 
544 	if (inst->session_type != VIDC_SESSION_TYPE_ENC)
545 		return 0;
546 
547 	if (enc_ctr->bitrate_mode == V4L2_MPEG_VIDEO_BITRATE_MODE_CQ)
548 		enable = false;
549 
550 	venc_mode = enable ? HFI_VENC_PERFMODE_POWER_SAVE :
551 		HFI_VENC_PERFMODE_MAX_QUALITY;
552 
553 	ret = hfi_session_set_property(inst, ptype, &venc_mode);
554 	if (ret)
555 		return ret;
556 
557 	inst->flags = enable ? inst->flags | VENUS_LOW_POWER :
558 		inst->flags & ~VENUS_LOW_POWER;
559 
560 	return ret;
561 }
562 
563 static int move_core_to_power_save_mode(struct venus_core *core,
564 					u32 core_id)
565 {
566 	struct venus_inst *inst = NULL;
567 
568 	mutex_lock(&core->lock);
569 	list_for_each_entry(inst, &core->instances, list) {
570 		if (inst->clk_data.core_id == core_id &&
571 		    inst->session_type == VIDC_SESSION_TYPE_ENC)
572 			power_save_mode_enable(inst, true);
573 	}
574 	mutex_unlock(&core->lock);
575 	return 0;
576 }
577 
578 static void
579 min_loaded_core(struct venus_inst *inst, u32 *min_coreid, u32 *min_load, bool low_power)
580 {
581 	u32 mbs_per_sec, load, core1_load = 0, core2_load = 0;
582 	u32 cores_max = core_num_max(inst);
583 	struct venus_core *core = inst->core;
584 	struct venus_inst *inst_pos;
585 	unsigned long vpp_freq;
586 	u32 coreid;
587 
588 	mutex_lock(&core->lock);
589 
590 	list_for_each_entry(inst_pos, &core->instances, list) {
591 		if (inst_pos == inst)
592 			continue;
593 
594 		if (inst_pos->state != INST_START)
595 			continue;
596 
597 		if (inst->session_type == VIDC_SESSION_TYPE_DEC)
598 			vpp_freq = inst_pos->clk_data.vpp_freq;
599 		else if (inst->session_type == VIDC_SESSION_TYPE_ENC)
600 			vpp_freq = low_power ? inst_pos->clk_data.low_power_freq :
601 				inst_pos->clk_data.vpp_freq;
602 		else
603 			continue;
604 
605 		coreid = inst_pos->clk_data.core_id;
606 
607 		mbs_per_sec = load_per_instance(inst_pos);
608 		load = mbs_per_sec * vpp_freq;
609 
610 		if ((coreid & VIDC_CORE_ID_3) == VIDC_CORE_ID_3) {
611 			core1_load += load / 2;
612 			core2_load += load / 2;
613 		} else if (coreid & VIDC_CORE_ID_1) {
614 			core1_load += load;
615 		} else if (coreid & VIDC_CORE_ID_2) {
616 			core2_load += load;
617 		}
618 	}
619 
620 	*min_coreid = core1_load <= core2_load ?
621 			VIDC_CORE_ID_1 : VIDC_CORE_ID_2;
622 	*min_load = min(core1_load, core2_load);
623 
624 	if (cores_max < VIDC_CORE_ID_2 || core->res->vcodec_num < 2) {
625 		*min_coreid = VIDC_CORE_ID_1;
626 		*min_load = core1_load;
627 	}
628 
629 	mutex_unlock(&core->lock);
630 }
631 
632 static int decide_core(struct venus_inst *inst)
633 {
634 	const u32 ptype = HFI_PROPERTY_CONFIG_VIDEOCORES_USAGE;
635 	struct venus_core *core = inst->core;
636 	u32 min_coreid, min_load, cur_inst_load;
637 	u32 min_lp_coreid, min_lp_load, cur_inst_lp_load;
638 	struct hfi_videocores_usage_type cu;
639 	unsigned long max_freq;
640 	int ret = 0;
641 
642 	if (legacy_binding) {
643 		if (inst->session_type == VIDC_SESSION_TYPE_DEC)
644 			cu.video_core_enable_mask = VIDC_CORE_ID_1;
645 		else
646 			cu.video_core_enable_mask = VIDC_CORE_ID_2;
647 
648 		goto done;
649 	}
650 
651 	if (inst->clk_data.core_id != VIDC_CORE_ID_DEFAULT)
652 		return 0;
653 
654 	cur_inst_load = load_per_instance(inst);
655 	cur_inst_load *= inst->clk_data.vpp_freq;
656 	/*TODO : divide this inst->load by work_route */
657 
658 	cur_inst_lp_load = load_per_instance(inst);
659 	cur_inst_lp_load *= inst->clk_data.low_power_freq;
660 	/*TODO : divide this inst->load by work_route */
661 
662 	max_freq = core->res->freq_tbl[0].freq;
663 
664 	min_loaded_core(inst, &min_coreid, &min_load, false);
665 	min_loaded_core(inst, &min_lp_coreid, &min_lp_load, true);
666 
667 	if (cur_inst_load + min_load <= max_freq) {
668 		inst->clk_data.core_id = min_coreid;
669 		cu.video_core_enable_mask = min_coreid;
670 	} else if (cur_inst_lp_load + min_load <= max_freq) {
671 		/* Move current instance to LP and return */
672 		inst->clk_data.core_id = min_coreid;
673 		cu.video_core_enable_mask = min_coreid;
674 		power_save_mode_enable(inst, true);
675 	} else if (cur_inst_lp_load + min_lp_load <= max_freq) {
676 		/* Move all instances to LP mode and return */
677 		inst->clk_data.core_id = min_lp_coreid;
678 		cu.video_core_enable_mask = min_lp_coreid;
679 		move_core_to_power_save_mode(core, min_lp_coreid);
680 	} else {
681 		dev_warn(core->dev, "HW can't support this load");
682 		return -EINVAL;
683 	}
684 
685 done:
686 	ret = hfi_session_set_property(inst, ptype, &cu);
687 	if (ret)
688 		return ret;
689 
690 	return ret;
691 }
692 
693 static int acquire_core(struct venus_inst *inst)
694 {
695 	struct venus_core *core = inst->core;
696 	unsigned int coreid_mask = 0;
697 
698 	if (inst->core_acquired)
699 		return 0;
700 
701 	inst->core_acquired = true;
702 
703 	if (inst->clk_data.core_id & VIDC_CORE_ID_1) {
704 		if (core->core0_usage_count++)
705 			return 0;
706 
707 		coreid_mask = VIDC_CORE_ID_1;
708 	}
709 
710 	if (inst->clk_data.core_id & VIDC_CORE_ID_2) {
711 		if (core->core1_usage_count++)
712 			return 0;
713 
714 		coreid_mask |= VIDC_CORE_ID_2;
715 	}
716 
717 	return poweron_coreid(core, coreid_mask);
718 }
719 
720 static int release_core(struct venus_inst *inst)
721 {
722 	struct venus_core *core = inst->core;
723 	unsigned int coreid_mask = 0;
724 	int ret;
725 
726 	if (!inst->core_acquired)
727 		return 0;
728 
729 	if (inst->clk_data.core_id & VIDC_CORE_ID_1) {
730 		if (--core->core0_usage_count)
731 			goto done;
732 
733 		coreid_mask = VIDC_CORE_ID_1;
734 	}
735 
736 	if (inst->clk_data.core_id & VIDC_CORE_ID_2) {
737 		if (--core->core1_usage_count)
738 			goto done;
739 
740 		coreid_mask |= VIDC_CORE_ID_2;
741 	}
742 
743 	ret = poweroff_coreid(core, coreid_mask);
744 	if (ret)
745 		return ret;
746 
747 done:
748 	inst->clk_data.core_id = VIDC_CORE_ID_DEFAULT;
749 	inst->core_acquired = false;
750 	return 0;
751 }
752 
753 static int coreid_power_v4(struct venus_inst *inst, int on)
754 {
755 	struct venus_core *core = inst->core;
756 	int ret;
757 
758 	if (legacy_binding)
759 		return 0;
760 
761 	if (on == POWER_ON) {
762 		ret = decide_core(inst);
763 		if (ret)
764 			return ret;
765 
766 		mutex_lock(&core->lock);
767 		ret = acquire_core(inst);
768 		mutex_unlock(&core->lock);
769 	} else {
770 		mutex_lock(&core->lock);
771 		ret = release_core(inst);
772 		mutex_unlock(&core->lock);
773 	}
774 
775 	return ret;
776 }
777 
778 static int vdec_get_v4(struct device *dev)
779 {
780 	struct venus_core *core = dev_get_drvdata(dev);
781 
782 	if (!legacy_binding)
783 		return 0;
784 
785 	return vcodec_clks_get(core, dev, core->vcodec0_clks,
786 			       core->res->vcodec0_clks);
787 }
788 
789 static void vdec_put_v4(struct device *dev)
790 {
791 	struct venus_core *core = dev_get_drvdata(dev);
792 	unsigned int i;
793 
794 	if (!legacy_binding)
795 		return;
796 
797 	for (i = 0; i < core->res->vcodec_clks_num; i++)
798 		core->vcodec0_clks[i] = NULL;
799 }
800 
801 static int vdec_power_v4(struct device *dev, int on)
802 {
803 	struct venus_core *core = dev_get_drvdata(dev);
804 	int ret;
805 
806 	if (!legacy_binding)
807 		return 0;
808 
809 	ret = vcodec_control_v4(core, VIDC_CORE_ID_1, true);
810 	if (ret)
811 		return ret;
812 
813 	if (on == POWER_ON)
814 		ret = vcodec_clks_enable(core, core->vcodec0_clks);
815 	else
816 		vcodec_clks_disable(core, core->vcodec0_clks);
817 
818 	ret = vcodec_control_v4(core, VIDC_CORE_ID_1, false);
819 
820 	return ret;
821 }
822 
823 static int venc_get_v4(struct device *dev)
824 {
825 	struct venus_core *core = dev_get_drvdata(dev);
826 
827 	if (!legacy_binding)
828 		return 0;
829 
830 	return vcodec_clks_get(core, dev, core->vcodec1_clks,
831 			       core->res->vcodec1_clks);
832 }
833 
834 static void venc_put_v4(struct device *dev)
835 {
836 	struct venus_core *core = dev_get_drvdata(dev);
837 	unsigned int i;
838 
839 	if (!legacy_binding)
840 		return;
841 
842 	for (i = 0; i < core->res->vcodec_clks_num; i++)
843 		core->vcodec1_clks[i] = NULL;
844 }
845 
846 static int venc_power_v4(struct device *dev, int on)
847 {
848 	struct venus_core *core = dev_get_drvdata(dev);
849 	int ret;
850 
851 	if (!legacy_binding)
852 		return 0;
853 
854 	ret = vcodec_control_v4(core, VIDC_CORE_ID_2, true);
855 	if (ret)
856 		return ret;
857 
858 	if (on == POWER_ON)
859 		ret = vcodec_clks_enable(core, core->vcodec1_clks);
860 	else
861 		vcodec_clks_disable(core, core->vcodec1_clks);
862 
863 	ret = vcodec_control_v4(core, VIDC_CORE_ID_2, false);
864 
865 	return ret;
866 }
867 
868 static int vcodec_domains_get(struct venus_core *core)
869 {
870 	int ret;
871 	struct device *dev = core->dev;
872 	const struct venus_resources *res = core->res;
873 	struct dev_pm_domain_attach_data vcodec_data = {
874 		.pd_names = res->vcodec_pmdomains,
875 		.num_pd_names = res->vcodec_pmdomains_num,
876 		.pd_flags = PD_FLAG_NO_DEV_LINK,
877 	};
878 	struct dev_pm_domain_attach_data opp_pd_data = {
879 		.pd_names = res->opp_pmdomain,
880 		.num_pd_names = 1,
881 		.pd_flags = PD_FLAG_DEV_LINK_ON | PD_FLAG_REQUIRED_OPP,
882 	};
883 
884 	if (!res->vcodec_pmdomains_num)
885 		goto skip_pmdomains;
886 
887 	ret = devm_pm_domain_attach_list(dev, &vcodec_data, &core->pmdomains);
888 	if (ret < 0)
889 		return ret;
890 
891 skip_pmdomains:
892 	if (!res->opp_pmdomain)
893 		return 0;
894 
895 	/* Attach the power domain for setting performance state */
896 	ret = devm_pm_domain_attach_list(dev, &opp_pd_data, &core->opp_pmdomain);
897 	if (ret < 0)
898 		return ret;
899 
900 	return 0;
901 }
902 
903 static int core_resets_reset(struct venus_core *core)
904 {
905 	const struct venus_resources *res = core->res;
906 	unsigned int i;
907 	int ret;
908 
909 	if (!res->resets_num)
910 		return 0;
911 
912 	for (i = 0; i < res->resets_num; i++) {
913 		ret = reset_control_assert(core->resets[i]);
914 		if (ret)
915 			goto err;
916 
917 		usleep_range(150, 250);
918 		ret = reset_control_deassert(core->resets[i]);
919 		if (ret)
920 			goto err;
921 	}
922 
923 err:
924 	return ret;
925 }
926 
927 static int core_resets_get(struct venus_core *core)
928 {
929 	struct device *dev = core->dev;
930 	const struct venus_resources *res = core->res;
931 	unsigned int i;
932 	int ret;
933 
934 	if (!res->resets_num)
935 		return 0;
936 
937 	for (i = 0; i < res->resets_num; i++) {
938 		core->resets[i] =
939 			devm_reset_control_get_exclusive(dev, res->resets[i]);
940 		if (IS_ERR(core->resets[i])) {
941 			ret = PTR_ERR(core->resets[i]);
942 			return ret;
943 		}
944 	}
945 
946 	return 0;
947 }
948 
949 static int core_get_v4(struct venus_core *core)
950 {
951 	struct device *dev = core->dev;
952 	const struct venus_resources *res = core->res;
953 	int ret;
954 
955 	ret = core_clks_get(core);
956 	if (ret)
957 		return ret;
958 
959 	if (!res->vcodec_pmdomains_num)
960 		legacy_binding = true;
961 
962 	dev_info(dev, "%s legacy binding\n", legacy_binding ? "" : "non");
963 
964 	ret = vcodec_clks_get(core, dev, core->vcodec0_clks, res->vcodec0_clks);
965 	if (ret)
966 		return ret;
967 
968 	ret = vcodec_clks_get(core, dev, core->vcodec1_clks, res->vcodec1_clks);
969 	if (ret)
970 		return ret;
971 
972 	ret = core_resets_get(core);
973 	if (ret)
974 		return ret;
975 
976 	if (legacy_binding)
977 		return 0;
978 
979 	ret = devm_pm_opp_set_clkname(dev, "core");
980 	if (ret)
981 		return ret;
982 
983 	ret = vcodec_domains_get(core);
984 	if (ret)
985 		return ret;
986 
987 	if (core->res->opp_pmdomain) {
988 		ret = devm_pm_opp_of_add_table(dev);
989 		if (ret && ret != -ENODEV) {
990 			dev_err(dev, "invalid OPP table in device tree\n");
991 			return ret;
992 		}
993 	}
994 
995 	return 0;
996 }
997 
998 static void core_put_v4(struct venus_core *core)
999 {
1000 }
1001 
1002 static int core_power_v4(struct venus_core *core, int on)
1003 {
1004 	struct device *dev = core->dev;
1005 	struct device *pmctrl = core->pmdomains ?
1006 			core->pmdomains->pd_devs[0] : NULL;
1007 	int ret = 0;
1008 
1009 	if (on == POWER_ON) {
1010 		if (pmctrl) {
1011 			ret = pm_runtime_resume_and_get(pmctrl);
1012 			if (ret < 0) {
1013 				return ret;
1014 			}
1015 		}
1016 
1017 		ret = core_resets_reset(core);
1018 		if (ret) {
1019 			if (pmctrl)
1020 				pm_runtime_put_sync(pmctrl);
1021 			return ret;
1022 		}
1023 
1024 		ret = core_clks_enable(core);
1025 		if (ret < 0 && pmctrl)
1026 			pm_runtime_put_sync(pmctrl);
1027 	} else {
1028 		/* Drop the performance state vote */
1029 		if (core->opp_pmdomain)
1030 			dev_pm_opp_set_rate(dev, 0);
1031 
1032 		core_clks_disable(core);
1033 
1034 		ret = core_resets_reset(core);
1035 
1036 		if (pmctrl)
1037 			pm_runtime_put_sync(pmctrl);
1038 	}
1039 
1040 	return ret;
1041 }
1042 
1043 static unsigned long calculate_inst_freq(struct venus_inst *inst,
1044 					 unsigned long filled_len)
1045 {
1046 	unsigned long vpp_freq_per_mb = 0, vpp_freq = 0, vsp_freq = 0;
1047 	u32 fps = (u32)inst->fps;
1048 	u32 mbs_per_sec;
1049 
1050 	mbs_per_sec = load_per_instance(inst);
1051 
1052 	if (inst->state != INST_START)
1053 		return 0;
1054 
1055 	if (inst->session_type == VIDC_SESSION_TYPE_ENC) {
1056 		vpp_freq_per_mb = inst->flags & VENUS_LOW_POWER ?
1057 			inst->clk_data.low_power_freq :
1058 			inst->clk_data.vpp_freq;
1059 
1060 		vpp_freq = mbs_per_sec * vpp_freq_per_mb;
1061 	} else {
1062 		vpp_freq = mbs_per_sec * inst->clk_data.vpp_freq;
1063 	}
1064 
1065 	/* 21 / 20 is overhead factor */
1066 	vpp_freq += vpp_freq / 20;
1067 	vsp_freq = mbs_per_sec * inst->clk_data.vsp_freq;
1068 
1069 	/* 10 / 7 is overhead factor */
1070 	if (inst->session_type == VIDC_SESSION_TYPE_ENC)
1071 		vsp_freq += (inst->controls.enc.bitrate * 10) / 7;
1072 	else
1073 		vsp_freq += ((fps * filled_len * 8) * 10) / 7;
1074 
1075 	return max(vpp_freq, vsp_freq);
1076 }
1077 
1078 static int load_scale_v4(struct venus_inst *inst)
1079 {
1080 	struct venus_core *core = inst->core;
1081 	const struct freq_tbl *table = core->res->freq_tbl;
1082 	unsigned int num_rows = core->res->freq_tbl_size;
1083 	struct device *dev = core->dev;
1084 	unsigned long freq = 0, freq_core1 = 0, freq_core2 = 0;
1085 	unsigned long filled_len = 0;
1086 	int i, ret = 0;
1087 
1088 	for (i = 0; i < inst->num_input_bufs; i++)
1089 		filled_len = max(filled_len, inst->payloads[i]);
1090 
1091 	if (inst->session_type == VIDC_SESSION_TYPE_DEC && !filled_len)
1092 		return ret;
1093 
1094 	freq = calculate_inst_freq(inst, filled_len);
1095 	inst->clk_data.freq = freq;
1096 
1097 	mutex_lock(&core->lock);
1098 	list_for_each_entry(inst, &core->instances, list) {
1099 		if (inst->clk_data.core_id == VIDC_CORE_ID_1) {
1100 			freq_core1 += inst->clk_data.freq;
1101 		} else if (inst->clk_data.core_id == VIDC_CORE_ID_2) {
1102 			freq_core2 += inst->clk_data.freq;
1103 		} else if (inst->clk_data.core_id == VIDC_CORE_ID_3) {
1104 			freq_core1 += inst->clk_data.freq;
1105 			freq_core2 += inst->clk_data.freq;
1106 		}
1107 	}
1108 
1109 	freq = max(freq_core1, freq_core2);
1110 
1111 	if (freq > table[0].freq) {
1112 		dev_dbg(dev, VDBGL "requested clock rate: %lu scaling clock rate : %lu\n",
1113 			freq, table[0].freq);
1114 
1115 		freq = table[0].freq;
1116 		goto set_freq;
1117 	}
1118 
1119 	for (i = num_rows - 1 ; i >= 0; i--) {
1120 		if (freq <= table[i].freq) {
1121 			freq = table[i].freq;
1122 			break;
1123 		}
1124 	}
1125 
1126 set_freq:
1127 
1128 	ret = core_clks_set_rate(core, freq);
1129 	if (ret) {
1130 		dev_err(dev, "failed to set clock rate %lu (%d)\n",
1131 			freq, ret);
1132 		goto exit;
1133 	}
1134 
1135 	ret = load_scale_bw(core);
1136 	if (ret) {
1137 		dev_err(dev, "failed to set bandwidth (%d)\n",
1138 			ret);
1139 		goto exit;
1140 	}
1141 
1142 exit:
1143 	mutex_unlock(&core->lock);
1144 	return ret;
1145 }
1146 
1147 static const struct venus_pm_ops pm_ops_v4 = {
1148 	.core_get = core_get_v4,
1149 	.core_put = core_put_v4,
1150 	.core_power = core_power_v4,
1151 	.vdec_get = vdec_get_v4,
1152 	.vdec_put = vdec_put_v4,
1153 	.vdec_power = vdec_power_v4,
1154 	.venc_get = venc_get_v4,
1155 	.venc_put = venc_put_v4,
1156 	.venc_power = venc_power_v4,
1157 	.coreid_power = coreid_power_v4,
1158 	.load_scale = load_scale_v4,
1159 };
1160 
1161 const struct venus_pm_ops *venus_pm_get(enum hfi_version version)
1162 {
1163 	switch (version) {
1164 	case HFI_VERSION_1XX:
1165 	default:
1166 		return &pm_ops_v1;
1167 	case HFI_VERSION_3XX:
1168 		return &pm_ops_v3;
1169 	case HFI_VERSION_4XX:
1170 	case HFI_VERSION_6XX:
1171 		return &pm_ops_v4;
1172 	}
1173 
1174 	return NULL;
1175 }
1176