1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * Hantro VPU codec driver
4 *
5 * Copyright (C) 2018 Collabora, Ltd.
6 * Copyright 2018 Google LLC.
7 * Tomasz Figa <tfiga@chromium.org>
8 *
9 * Based on s5p-mfc driver by Samsung Electronics Co., Ltd.
10 * Copyright (C) 2011 Samsung Electronics Co., Ltd.
11 */
12
13 #include <linux/clk.h>
14 #include <linux/module.h>
15 #include <linux/of.h>
16 #include <linux/of_platform.h>
17 #include <linux/platform_device.h>
18 #include <linux/pm.h>
19 #include <linux/pm_runtime.h>
20 #include <linux/slab.h>
21 #include <linux/videodev2.h>
22 #include <linux/workqueue.h>
23 #include <media/v4l2-event.h>
24 #include <media/v4l2-mem2mem.h>
25 #include <media/videobuf2-core.h>
26 #include <media/videobuf2-vmalloc.h>
27
28 #include "hantro_v4l2.h"
29 #include "hantro.h"
30 #include "hantro_hw.h"
31
32 #define DRIVER_NAME "hantro-vpu"
33
34 int hantro_debug;
35 module_param_named(debug, hantro_debug, int, 0644);
36 MODULE_PARM_DESC(debug,
37 "Debug level - higher value produces more verbose messages");
38
hantro_get_ctrl(struct hantro_ctx * ctx,u32 id)39 void *hantro_get_ctrl(struct hantro_ctx *ctx, u32 id)
40 {
41 struct v4l2_ctrl *ctrl;
42
43 ctrl = v4l2_ctrl_find(&ctx->ctrl_handler, id);
44 return ctrl ? ctrl->p_cur.p : NULL;
45 }
46
hantro_get_ref(struct hantro_ctx * ctx,u64 ts)47 dma_addr_t hantro_get_ref(struct hantro_ctx *ctx, u64 ts)
48 {
49 struct vb2_queue *q = v4l2_m2m_get_dst_vq(ctx->fh.m2m_ctx);
50 struct vb2_buffer *buf;
51
52 buf = vb2_find_buffer(q, ts);
53 if (!buf)
54 return 0;
55 return hantro_get_dec_buf_addr(ctx, buf);
56 }
57
58 static const struct v4l2_event hantro_eos_event = {
59 .type = V4L2_EVENT_EOS
60 };
61
hantro_job_finish_no_pm(struct hantro_dev * vpu,struct hantro_ctx * ctx,enum vb2_buffer_state result)62 static void hantro_job_finish_no_pm(struct hantro_dev *vpu,
63 struct hantro_ctx *ctx,
64 enum vb2_buffer_state result)
65 {
66 struct vb2_v4l2_buffer *src, *dst;
67
68 src = v4l2_m2m_next_src_buf(ctx->fh.m2m_ctx);
69 dst = v4l2_m2m_next_dst_buf(ctx->fh.m2m_ctx);
70
71 if (WARN_ON(!src))
72 return;
73 if (WARN_ON(!dst))
74 return;
75
76 src->sequence = ctx->sequence_out++;
77 dst->sequence = ctx->sequence_cap++;
78
79 if (v4l2_m2m_is_last_draining_src_buf(ctx->fh.m2m_ctx, src)) {
80 dst->flags |= V4L2_BUF_FLAG_LAST;
81 v4l2_event_queue_fh(&ctx->fh, &hantro_eos_event);
82 v4l2_m2m_mark_stopped(ctx->fh.m2m_ctx);
83 }
84
85 v4l2_m2m_buf_done_and_job_finish(ctx->dev->m2m_dev, ctx->fh.m2m_ctx,
86 result);
87 }
88
hantro_job_finish(struct hantro_dev * vpu,struct hantro_ctx * ctx,enum vb2_buffer_state result)89 static void hantro_job_finish(struct hantro_dev *vpu,
90 struct hantro_ctx *ctx,
91 enum vb2_buffer_state result)
92 {
93 pm_runtime_put_autosuspend(vpu->dev);
94
95 clk_bulk_disable(vpu->variant->num_clocks, vpu->clocks);
96
97 hantro_job_finish_no_pm(vpu, ctx, result);
98 }
99
hantro_irq_done(struct hantro_dev * vpu,enum vb2_buffer_state result)100 void hantro_irq_done(struct hantro_dev *vpu,
101 enum vb2_buffer_state result)
102 {
103 struct hantro_ctx *ctx =
104 v4l2_m2m_get_curr_priv(vpu->m2m_dev);
105
106 /*
107 * If cancel_delayed_work returns false
108 * the timeout expired. The watchdog is running,
109 * and will take care of finishing the job.
110 */
111 if (cancel_delayed_work(&vpu->watchdog_work)) {
112 if (result == VB2_BUF_STATE_DONE && ctx->codec_ops->done)
113 ctx->codec_ops->done(ctx);
114 hantro_job_finish(vpu, ctx, result);
115 }
116 }
117
hantro_watchdog(struct work_struct * work)118 void hantro_watchdog(struct work_struct *work)
119 {
120 struct hantro_dev *vpu;
121 struct hantro_ctx *ctx;
122
123 vpu = container_of(to_delayed_work(work),
124 struct hantro_dev, watchdog_work);
125 ctx = v4l2_m2m_get_curr_priv(vpu->m2m_dev);
126 if (ctx) {
127 vpu_err("frame processing timed out!\n");
128 if (ctx->codec_ops->reset)
129 ctx->codec_ops->reset(ctx);
130 hantro_job_finish(vpu, ctx, VB2_BUF_STATE_ERROR);
131 }
132 }
133
hantro_start_prepare_run(struct hantro_ctx * ctx)134 void hantro_start_prepare_run(struct hantro_ctx *ctx)
135 {
136 struct vb2_v4l2_buffer *src_buf;
137
138 src_buf = hantro_get_src_buf(ctx);
139 v4l2_ctrl_request_setup(src_buf->vb2_buf.req_obj.req,
140 &ctx->ctrl_handler);
141
142 if (!ctx->is_encoder && !ctx->dev->variant->late_postproc) {
143 if (hantro_needs_postproc(ctx, ctx->vpu_dst_fmt))
144 hantro_postproc_enable(ctx);
145 else
146 hantro_postproc_disable(ctx);
147 }
148 }
149
hantro_end_prepare_run(struct hantro_ctx * ctx)150 void hantro_end_prepare_run(struct hantro_ctx *ctx)
151 {
152 struct vb2_v4l2_buffer *src_buf;
153
154 if (!ctx->is_encoder && ctx->dev->variant->late_postproc) {
155 if (hantro_needs_postproc(ctx, ctx->vpu_dst_fmt))
156 hantro_postproc_enable(ctx);
157 else
158 hantro_postproc_disable(ctx);
159 }
160
161 src_buf = hantro_get_src_buf(ctx);
162 v4l2_ctrl_request_complete(src_buf->vb2_buf.req_obj.req,
163 &ctx->ctrl_handler);
164
165 /* Kick the watchdog. */
166 schedule_delayed_work(&ctx->dev->watchdog_work,
167 msecs_to_jiffies(2000));
168 }
169
device_run(void * priv)170 static void device_run(void *priv)
171 {
172 struct hantro_ctx *ctx = priv;
173 struct vb2_v4l2_buffer *src, *dst;
174 int ret;
175
176 src = hantro_get_src_buf(ctx);
177 dst = hantro_get_dst_buf(ctx);
178
179 ret = pm_runtime_resume_and_get(ctx->dev->dev);
180 if (ret < 0)
181 goto err_cancel_job;
182
183 ret = clk_bulk_enable(ctx->dev->variant->num_clocks, ctx->dev->clocks);
184 if (ret)
185 goto err_cancel_job;
186
187 v4l2_m2m_buf_copy_metadata(src, dst);
188
189 if (ctx->codec_ops->run(ctx))
190 goto err_cancel_job;
191
192 return;
193
194 err_cancel_job:
195 hantro_job_finish_no_pm(ctx->dev, ctx, VB2_BUF_STATE_ERROR);
196 }
197
198 static const struct v4l2_m2m_ops vpu_m2m_ops = {
199 .device_run = device_run,
200 };
201
202 static int
queue_init(void * priv,struct vb2_queue * src_vq,struct vb2_queue * dst_vq)203 queue_init(void *priv, struct vb2_queue *src_vq, struct vb2_queue *dst_vq)
204 {
205 struct hantro_ctx *ctx = priv;
206 int ret;
207
208 src_vq->type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
209 src_vq->io_modes = VB2_MMAP | VB2_DMABUF;
210 src_vq->drv_priv = ctx;
211 src_vq->ops = &hantro_queue_ops;
212 src_vq->mem_ops = &vb2_dma_contig_memops;
213
214 /*
215 * Driver does mostly sequential access, so sacrifice TLB efficiency
216 * for faster allocation. Also, no CPU access on the source queue,
217 * so no kernel mapping needed.
218 */
219 src_vq->dma_attrs = DMA_ATTR_ALLOC_SINGLE_PAGES |
220 DMA_ATTR_NO_KERNEL_MAPPING;
221 src_vq->buf_struct_size = sizeof(struct v4l2_m2m_buffer);
222 src_vq->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_COPY;
223 src_vq->lock = &ctx->dev->vpu_mutex;
224 src_vq->dev = ctx->dev->v4l2_dev.dev;
225 src_vq->supports_requests = true;
226
227 ret = vb2_queue_init(src_vq);
228 if (ret)
229 return ret;
230
231 dst_vq->bidirectional = true;
232 dst_vq->mem_ops = &vb2_dma_contig_memops;
233 dst_vq->dma_attrs = DMA_ATTR_ALLOC_SINGLE_PAGES;
234 /*
235 * The Kernel needs access to the JPEG destination buffer for the
236 * JPEG encoder to fill in the JPEG headers.
237 */
238 if (!ctx->is_encoder) {
239 dst_vq->dma_attrs |= DMA_ATTR_NO_KERNEL_MAPPING;
240 dst_vq->max_num_buffers = MAX_POSTPROC_BUFFERS;
241 }
242
243 dst_vq->type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
244 dst_vq->io_modes = VB2_MMAP | VB2_DMABUF;
245 dst_vq->drv_priv = ctx;
246 dst_vq->ops = &hantro_queue_ops;
247 dst_vq->buf_struct_size = sizeof(struct hantro_decoded_buffer);
248 dst_vq->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_COPY;
249 dst_vq->lock = &ctx->dev->vpu_mutex;
250 dst_vq->dev = ctx->dev->v4l2_dev.dev;
251
252 return vb2_queue_init(dst_vq);
253 }
254
hantro_try_ctrl(struct v4l2_ctrl * ctrl)255 static int hantro_try_ctrl(struct v4l2_ctrl *ctrl)
256 {
257 if (ctrl->id == V4L2_CID_STATELESS_H264_SPS) {
258 const struct v4l2_ctrl_h264_sps *sps = ctrl->p_new.p_h264_sps;
259
260 if (sps->chroma_format_idc > 1)
261 /* Only 4:0:0 and 4:2:0 are supported */
262 return -EINVAL;
263 if (sps->bit_depth_luma_minus8 != sps->bit_depth_chroma_minus8)
264 /* Luma and chroma bit depth mismatch */
265 return -EINVAL;
266 if (sps->bit_depth_luma_minus8 != 0)
267 /* Only 8-bit is supported */
268 return -EINVAL;
269 } else if (ctrl->id == V4L2_CID_STATELESS_HEVC_SPS) {
270 const struct v4l2_ctrl_hevc_sps *sps = ctrl->p_new.p_hevc_sps;
271
272 if (sps->bit_depth_luma_minus8 != 0 && sps->bit_depth_luma_minus8 != 2)
273 /* Only 8-bit and 10-bit are supported */
274 return -EINVAL;
275 } else if (ctrl->id == V4L2_CID_STATELESS_VP9_FRAME) {
276 const struct v4l2_ctrl_vp9_frame *dec_params = ctrl->p_new.p_vp9_frame;
277
278 /* We only support profile 0 */
279 if (dec_params->profile != 0)
280 return -EINVAL;
281 } else if (ctrl->id == V4L2_CID_STATELESS_AV1_SEQUENCE) {
282 const struct v4l2_ctrl_av1_sequence *sequence = ctrl->p_new.p_av1_sequence;
283
284 if (sequence->bit_depth != 8 && sequence->bit_depth != 10)
285 return -EINVAL;
286 }
287
288 return 0;
289 }
290
hantro_jpeg_s_ctrl(struct v4l2_ctrl * ctrl)291 static int hantro_jpeg_s_ctrl(struct v4l2_ctrl *ctrl)
292 {
293 struct hantro_ctx *ctx;
294
295 ctx = container_of(ctrl->handler,
296 struct hantro_ctx, ctrl_handler);
297
298 vpu_debug(1, "s_ctrl: id = %d, val = %d\n", ctrl->id, ctrl->val);
299
300 switch (ctrl->id) {
301 case V4L2_CID_JPEG_COMPRESSION_QUALITY:
302 ctx->jpeg_quality = ctrl->val;
303 break;
304 default:
305 return -EINVAL;
306 }
307
308 return 0;
309 }
310
hantro_vp9_s_ctrl(struct v4l2_ctrl * ctrl)311 static int hantro_vp9_s_ctrl(struct v4l2_ctrl *ctrl)
312 {
313 struct hantro_ctx *ctx;
314
315 ctx = container_of(ctrl->handler,
316 struct hantro_ctx, ctrl_handler);
317
318 switch (ctrl->id) {
319 case V4L2_CID_STATELESS_VP9_FRAME: {
320 int bit_depth = ctrl->p_new.p_vp9_frame->bit_depth;
321
322 if (ctx->bit_depth == bit_depth)
323 return 0;
324
325 return hantro_reset_raw_fmt(ctx, bit_depth, HANTRO_AUTO_POSTPROC);
326 }
327 default:
328 return -EINVAL;
329 }
330
331 return 0;
332 }
333
hantro_hevc_s_ctrl(struct v4l2_ctrl * ctrl)334 static int hantro_hevc_s_ctrl(struct v4l2_ctrl *ctrl)
335 {
336 struct hantro_ctx *ctx;
337
338 ctx = container_of(ctrl->handler,
339 struct hantro_ctx, ctrl_handler);
340
341 switch (ctrl->id) {
342 case V4L2_CID_STATELESS_HEVC_SPS: {
343 const struct v4l2_ctrl_hevc_sps *sps = ctrl->p_new.p_hevc_sps;
344 int bit_depth = sps->bit_depth_luma_minus8 + 8;
345
346 if (ctx->bit_depth == bit_depth)
347 return 0;
348
349 return hantro_reset_raw_fmt(ctx, bit_depth, HANTRO_AUTO_POSTPROC);
350 }
351 default:
352 return -EINVAL;
353 }
354
355 return 0;
356 }
357
hantro_av1_s_ctrl(struct v4l2_ctrl * ctrl)358 static int hantro_av1_s_ctrl(struct v4l2_ctrl *ctrl)
359 {
360 struct hantro_ctx *ctx;
361
362 ctx = container_of(ctrl->handler,
363 struct hantro_ctx, ctrl_handler);
364
365 switch (ctrl->id) {
366 case V4L2_CID_STATELESS_AV1_SEQUENCE:
367 {
368 int bit_depth = ctrl->p_new.p_av1_sequence->bit_depth;
369 bool need_postproc = HANTRO_AUTO_POSTPROC;
370
371 if (ctrl->p_new.p_av1_sequence->flags
372 & V4L2_AV1_SEQUENCE_FLAG_FILM_GRAIN_PARAMS_PRESENT)
373 need_postproc = HANTRO_FORCE_POSTPROC;
374
375 if (ctx->bit_depth == bit_depth &&
376 ctx->need_postproc == need_postproc)
377 return 0;
378
379 return hantro_reset_raw_fmt(ctx, bit_depth, need_postproc);
380 }
381 default:
382 return -EINVAL;
383 }
384
385 return 0;
386 }
387
388 static const struct v4l2_ctrl_ops hantro_ctrl_ops = {
389 .try_ctrl = hantro_try_ctrl,
390 };
391
392 static const struct v4l2_ctrl_ops hantro_jpeg_ctrl_ops = {
393 .s_ctrl = hantro_jpeg_s_ctrl,
394 };
395
396 static const struct v4l2_ctrl_ops hantro_vp9_ctrl_ops = {
397 .s_ctrl = hantro_vp9_s_ctrl,
398 };
399
400 static const struct v4l2_ctrl_ops hantro_hevc_ctrl_ops = {
401 .try_ctrl = hantro_try_ctrl,
402 .s_ctrl = hantro_hevc_s_ctrl,
403 };
404
405 static const struct v4l2_ctrl_ops hantro_av1_ctrl_ops = {
406 .try_ctrl = hantro_try_ctrl,
407 .s_ctrl = hantro_av1_s_ctrl,
408 };
409
410 #define HANTRO_JPEG_ACTIVE_MARKERS (V4L2_JPEG_ACTIVE_MARKER_APP0 | \
411 V4L2_JPEG_ACTIVE_MARKER_COM | \
412 V4L2_JPEG_ACTIVE_MARKER_DQT | \
413 V4L2_JPEG_ACTIVE_MARKER_DHT)
414
415 static const struct hantro_ctrl controls[] = {
416 {
417 .codec = HANTRO_JPEG_ENCODER,
418 .cfg = {
419 .id = V4L2_CID_JPEG_COMPRESSION_QUALITY,
420 .min = 5,
421 .max = 100,
422 .step = 1,
423 .def = 50,
424 .ops = &hantro_jpeg_ctrl_ops,
425 },
426 }, {
427 .codec = HANTRO_JPEG_ENCODER,
428 .cfg = {
429 .id = V4L2_CID_JPEG_ACTIVE_MARKER,
430 .max = HANTRO_JPEG_ACTIVE_MARKERS,
431 .def = HANTRO_JPEG_ACTIVE_MARKERS,
432 /*
433 * Changing the set of active markers/segments also
434 * messes up the alignment of the JPEG header, which
435 * is needed to allow the hardware to write directly
436 * to the output buffer. Implementing this introduces
437 * a lot of complexity for little gain, as the markers
438 * enabled is already the minimum required set.
439 */
440 .flags = V4L2_CTRL_FLAG_READ_ONLY,
441 },
442 }, {
443 .codec = HANTRO_MPEG2_DECODER,
444 .cfg = {
445 .id = V4L2_CID_STATELESS_MPEG2_SEQUENCE,
446 },
447 }, {
448 .codec = HANTRO_MPEG2_DECODER,
449 .cfg = {
450 .id = V4L2_CID_STATELESS_MPEG2_PICTURE,
451 },
452 }, {
453 .codec = HANTRO_MPEG2_DECODER,
454 .cfg = {
455 .id = V4L2_CID_STATELESS_MPEG2_QUANTISATION,
456 },
457 }, {
458 .codec = HANTRO_VP8_DECODER,
459 .cfg = {
460 .id = V4L2_CID_STATELESS_VP8_FRAME,
461 },
462 }, {
463 .codec = HANTRO_H264_DECODER,
464 .cfg = {
465 .id = V4L2_CID_STATELESS_H264_DECODE_PARAMS,
466 },
467 }, {
468 .codec = HANTRO_H264_DECODER,
469 .cfg = {
470 .id = V4L2_CID_STATELESS_H264_SPS,
471 .ops = &hantro_ctrl_ops,
472 },
473 }, {
474 .codec = HANTRO_H264_DECODER,
475 .cfg = {
476 .id = V4L2_CID_STATELESS_H264_PPS,
477 },
478 }, {
479 .codec = HANTRO_H264_DECODER,
480 .cfg = {
481 .id = V4L2_CID_STATELESS_H264_SCALING_MATRIX,
482 },
483 }, {
484 .codec = HANTRO_H264_DECODER,
485 .cfg = {
486 .id = V4L2_CID_STATELESS_H264_DECODE_MODE,
487 .min = V4L2_STATELESS_H264_DECODE_MODE_FRAME_BASED,
488 .def = V4L2_STATELESS_H264_DECODE_MODE_FRAME_BASED,
489 .max = V4L2_STATELESS_H264_DECODE_MODE_FRAME_BASED,
490 },
491 }, {
492 .codec = HANTRO_H264_DECODER,
493 .cfg = {
494 .id = V4L2_CID_STATELESS_H264_START_CODE,
495 .min = V4L2_STATELESS_H264_START_CODE_ANNEX_B,
496 .def = V4L2_STATELESS_H264_START_CODE_ANNEX_B,
497 .max = V4L2_STATELESS_H264_START_CODE_ANNEX_B,
498 },
499 }, {
500 .codec = HANTRO_H264_DECODER,
501 .cfg = {
502 .id = V4L2_CID_MPEG_VIDEO_H264_PROFILE,
503 .min = V4L2_MPEG_VIDEO_H264_PROFILE_BASELINE,
504 .max = V4L2_MPEG_VIDEO_H264_PROFILE_HIGH,
505 .menu_skip_mask =
506 BIT(V4L2_MPEG_VIDEO_H264_PROFILE_EXTENDED),
507 .def = V4L2_MPEG_VIDEO_H264_PROFILE_MAIN,
508 }
509 }, {
510 .codec = HANTRO_HEVC_DECODER,
511 .cfg = {
512 .id = V4L2_CID_STATELESS_HEVC_DECODE_MODE,
513 .min = V4L2_STATELESS_HEVC_DECODE_MODE_FRAME_BASED,
514 .max = V4L2_STATELESS_HEVC_DECODE_MODE_FRAME_BASED,
515 .def = V4L2_STATELESS_HEVC_DECODE_MODE_FRAME_BASED,
516 },
517 }, {
518 .codec = HANTRO_HEVC_DECODER,
519 .cfg = {
520 .id = V4L2_CID_STATELESS_HEVC_START_CODE,
521 .min = V4L2_STATELESS_HEVC_START_CODE_ANNEX_B,
522 .max = V4L2_STATELESS_HEVC_START_CODE_ANNEX_B,
523 .def = V4L2_STATELESS_HEVC_START_CODE_ANNEX_B,
524 },
525 }, {
526 .codec = HANTRO_HEVC_DECODER,
527 .cfg = {
528 .id = V4L2_CID_MPEG_VIDEO_HEVC_PROFILE,
529 .min = V4L2_MPEG_VIDEO_HEVC_PROFILE_MAIN,
530 .max = V4L2_MPEG_VIDEO_HEVC_PROFILE_MAIN_10,
531 .def = V4L2_MPEG_VIDEO_HEVC_PROFILE_MAIN,
532 },
533 }, {
534 .codec = HANTRO_HEVC_DECODER,
535 .cfg = {
536 .id = V4L2_CID_MPEG_VIDEO_HEVC_LEVEL,
537 .min = V4L2_MPEG_VIDEO_HEVC_LEVEL_1,
538 .max = V4L2_MPEG_VIDEO_HEVC_LEVEL_5_1,
539 },
540 }, {
541 .codec = HANTRO_HEVC_DECODER,
542 .cfg = {
543 .id = V4L2_CID_STATELESS_HEVC_SPS,
544 .ops = &hantro_hevc_ctrl_ops,
545 },
546 }, {
547 .codec = HANTRO_HEVC_DECODER,
548 .cfg = {
549 .id = V4L2_CID_STATELESS_HEVC_PPS,
550 },
551 }, {
552 .codec = HANTRO_HEVC_DECODER,
553 .cfg = {
554 .id = V4L2_CID_STATELESS_HEVC_DECODE_PARAMS,
555 },
556 }, {
557 .codec = HANTRO_HEVC_DECODER,
558 .cfg = {
559 .id = V4L2_CID_STATELESS_HEVC_SCALING_MATRIX,
560 },
561 }, {
562 .codec = HANTRO_VP9_DECODER,
563 .cfg = {
564 .id = V4L2_CID_STATELESS_VP9_FRAME,
565 .ops = &hantro_vp9_ctrl_ops,
566 },
567 }, {
568 .codec = HANTRO_VP9_DECODER,
569 .cfg = {
570 .id = V4L2_CID_STATELESS_VP9_COMPRESSED_HDR,
571 },
572 }, {
573 .codec = HANTRO_AV1_DECODER,
574 .cfg = {
575 .id = V4L2_CID_STATELESS_AV1_FRAME,
576 },
577 }, {
578 .codec = HANTRO_AV1_DECODER,
579 .cfg = {
580 .id = V4L2_CID_STATELESS_AV1_TILE_GROUP_ENTRY,
581 .dims = { V4L2_AV1_MAX_TILE_COUNT },
582 },
583 }, {
584 .codec = HANTRO_AV1_DECODER,
585 .cfg = {
586 .id = V4L2_CID_STATELESS_AV1_SEQUENCE,
587 .ops = &hantro_av1_ctrl_ops,
588 },
589 }, {
590 .codec = HANTRO_AV1_DECODER,
591 .cfg = {
592 .id = V4L2_CID_STATELESS_AV1_FILM_GRAIN,
593 },
594 },
595 };
596
hantro_ctrls_setup(struct hantro_dev * vpu,struct hantro_ctx * ctx,int allowed_codecs)597 static int hantro_ctrls_setup(struct hantro_dev *vpu,
598 struct hantro_ctx *ctx,
599 int allowed_codecs)
600 {
601 int i, num_ctrls = ARRAY_SIZE(controls);
602
603 v4l2_ctrl_handler_init(&ctx->ctrl_handler, num_ctrls);
604
605 for (i = 0; i < num_ctrls; i++) {
606 if (!(allowed_codecs & controls[i].codec))
607 continue;
608
609 v4l2_ctrl_new_custom(&ctx->ctrl_handler,
610 &controls[i].cfg, NULL);
611 if (ctx->ctrl_handler.error) {
612 vpu_err("Adding control (%d) failed %d\n",
613 controls[i].cfg.id,
614 ctx->ctrl_handler.error);
615 v4l2_ctrl_handler_free(&ctx->ctrl_handler);
616 return ctx->ctrl_handler.error;
617 }
618 }
619 return v4l2_ctrl_handler_setup(&ctx->ctrl_handler);
620 }
621
622 /*
623 * V4L2 file operations.
624 */
625
hantro_open(struct file * filp)626 static int hantro_open(struct file *filp)
627 {
628 struct hantro_dev *vpu = video_drvdata(filp);
629 struct video_device *vdev = video_devdata(filp);
630 struct hantro_func *func = hantro_vdev_to_func(vdev);
631 struct hantro_ctx *ctx;
632 int allowed_codecs, ret;
633
634 /*
635 * We do not need any extra locking here, because we operate only
636 * on local data here, except reading few fields from dev, which
637 * do not change through device's lifetime (which is guaranteed by
638 * reference on module from open()) and V4L2 internal objects (such
639 * as vdev and ctx->fh), which have proper locking done in respective
640 * helper functions used here.
641 */
642
643 ctx = kzalloc_obj(*ctx);
644 if (!ctx)
645 return -ENOMEM;
646
647 ctx->dev = vpu;
648 if (func->id == MEDIA_ENT_F_PROC_VIDEO_ENCODER) {
649 allowed_codecs = vpu->variant->codec & HANTRO_ENCODERS;
650 ctx->is_encoder = true;
651 } else if (func->id == MEDIA_ENT_F_PROC_VIDEO_DECODER) {
652 allowed_codecs = vpu->variant->codec & HANTRO_DECODERS;
653 ctx->is_encoder = false;
654 } else {
655 ret = -ENODEV;
656 goto err_ctx_free;
657 }
658
659 ctx->fh.m2m_ctx = v4l2_m2m_ctx_init(vpu->m2m_dev, ctx, queue_init);
660 if (IS_ERR(ctx->fh.m2m_ctx)) {
661 ret = PTR_ERR(ctx->fh.m2m_ctx);
662 goto err_ctx_free;
663 }
664
665 v4l2_fh_init(&ctx->fh, vdev);
666 v4l2_fh_add(&ctx->fh, filp);
667
668 hantro_reset_fmts(ctx);
669
670 ret = hantro_ctrls_setup(vpu, ctx, allowed_codecs);
671 if (ret) {
672 vpu_err("Failed to set up controls\n");
673 goto err_fh_free;
674 }
675 ctx->fh.ctrl_handler = &ctx->ctrl_handler;
676
677 return 0;
678
679 err_fh_free:
680 v4l2_fh_del(&ctx->fh, filp);
681 v4l2_fh_exit(&ctx->fh);
682 err_ctx_free:
683 kfree(ctx);
684 return ret;
685 }
686
hantro_release(struct file * filp)687 static int hantro_release(struct file *filp)
688 {
689 struct hantro_ctx *ctx = file_to_ctx(filp);
690
691 /*
692 * No need for extra locking because this was the last reference
693 * to this file.
694 */
695 v4l2_m2m_ctx_release(ctx->fh.m2m_ctx);
696 v4l2_fh_del(&ctx->fh, filp);
697 v4l2_fh_exit(&ctx->fh);
698 v4l2_ctrl_handler_free(&ctx->ctrl_handler);
699 kfree(ctx);
700
701 return 0;
702 }
703
704 static const struct v4l2_file_operations hantro_fops = {
705 .owner = THIS_MODULE,
706 .open = hantro_open,
707 .release = hantro_release,
708 .poll = v4l2_m2m_fop_poll,
709 .unlocked_ioctl = video_ioctl2,
710 .mmap = v4l2_m2m_fop_mmap,
711 };
712
713 static const struct of_device_id of_hantro_match[] = {
714 #ifdef CONFIG_VIDEO_HANTRO_ROCKCHIP
715 { .compatible = "rockchip,px30-vpu", .data = &px30_vpu_variant, },
716 { .compatible = "rockchip,rk3036-vpu", .data = &rk3036_vpu_variant, },
717 { .compatible = "rockchip,rk3066-vpu", .data = &rk3066_vpu_variant, },
718 { .compatible = "rockchip,rk3288-vpu", .data = &rk3288_vpu_variant, },
719 { .compatible = "rockchip,rk3328-vpu", .data = &rk3328_vpu_variant, },
720 { .compatible = "rockchip,rk3399-vpu", .data = &rk3399_vpu_variant, },
721 { .compatible = "rockchip,rk3568-vepu", .data = &rk3568_vepu_variant, },
722 { .compatible = "rockchip,rk3568-vpu", .data = &rk3568_vpu_variant, },
723 { .compatible = "rockchip,rk3588-vepu121", .data = &rk3568_vepu_variant, },
724 { .compatible = "rockchip,rk3588-av1-vpu", .data = &rk3588_vpu981_variant, },
725 #endif
726 #ifdef CONFIG_VIDEO_HANTRO_IMX8M
727 { .compatible = "nxp,imx8mm-vpu-g1", .data = &imx8mm_vpu_g1_variant, },
728 { .compatible = "nxp,imx8mq-vpu", .data = &imx8mq_vpu_variant, },
729 { .compatible = "nxp,imx8mq-vpu-g1", .data = &imx8mq_vpu_g1_variant },
730 { .compatible = "nxp,imx8mq-vpu-g2", .data = &imx8mq_vpu_g2_variant },
731 #endif
732 #ifdef CONFIG_VIDEO_HANTRO_SAMA5D4
733 { .compatible = "microchip,sama5d4-vdec", .data = &sama5d4_vdec_variant, },
734 #endif
735 #ifdef CONFIG_VIDEO_HANTRO_SUNXI
736 { .compatible = "allwinner,sun50i-h6-vpu-g2", .data = &sunxi_vpu_variant, },
737 #endif
738 #ifdef CONFIG_VIDEO_HANTRO_STM32MP25
739 { .compatible = "st,stm32mp25-vdec", .data = &stm32mp25_vdec_variant, },
740 { .compatible = "st,stm32mp25-venc", .data = &stm32mp25_venc_variant, },
741 #endif
742 { /* sentinel */ }
743 };
744 MODULE_DEVICE_TABLE(of, of_hantro_match);
745
hantro_register_entity(struct media_device * mdev,struct media_entity * entity,const char * entity_name,struct media_pad * pads,int num_pads,int function,struct video_device * vdev)746 static int hantro_register_entity(struct media_device *mdev,
747 struct media_entity *entity,
748 const char *entity_name,
749 struct media_pad *pads, int num_pads,
750 int function, struct video_device *vdev)
751 {
752 char *name;
753 int ret;
754
755 entity->obj_type = MEDIA_ENTITY_TYPE_BASE;
756 if (function == MEDIA_ENT_F_IO_V4L) {
757 entity->info.dev.major = VIDEO_MAJOR;
758 entity->info.dev.minor = vdev->minor;
759 }
760
761 name = devm_kasprintf(mdev->dev, GFP_KERNEL, "%s-%s", vdev->name,
762 entity_name);
763 if (!name)
764 return -ENOMEM;
765
766 entity->name = name;
767 entity->function = function;
768
769 ret = media_entity_pads_init(entity, num_pads, pads);
770 if (ret)
771 return ret;
772
773 ret = media_device_register_entity(mdev, entity);
774 if (ret)
775 return ret;
776
777 return 0;
778 }
779
hantro_attach_func(struct hantro_dev * vpu,struct hantro_func * func)780 static int hantro_attach_func(struct hantro_dev *vpu,
781 struct hantro_func *func)
782 {
783 struct media_device *mdev = &vpu->mdev;
784 struct media_link *link;
785 int ret;
786
787 /* Create the three encoder entities with their pads */
788 func->source_pad.flags = MEDIA_PAD_FL_SOURCE;
789 ret = hantro_register_entity(mdev, &func->vdev.entity, "source",
790 &func->source_pad, 1, MEDIA_ENT_F_IO_V4L,
791 &func->vdev);
792 if (ret)
793 return ret;
794
795 func->proc_pads[0].flags = MEDIA_PAD_FL_SINK;
796 func->proc_pads[1].flags = MEDIA_PAD_FL_SOURCE;
797 ret = hantro_register_entity(mdev, &func->proc, "proc",
798 func->proc_pads, 2, func->id,
799 &func->vdev);
800 if (ret)
801 goto err_rel_entity0;
802
803 func->sink_pad.flags = MEDIA_PAD_FL_SINK;
804 ret = hantro_register_entity(mdev, &func->sink, "sink",
805 &func->sink_pad, 1, MEDIA_ENT_F_IO_V4L,
806 &func->vdev);
807 if (ret)
808 goto err_rel_entity1;
809
810 /* Connect the three entities */
811 ret = media_create_pad_link(&func->vdev.entity, 0, &func->proc, 0,
812 MEDIA_LNK_FL_IMMUTABLE |
813 MEDIA_LNK_FL_ENABLED);
814 if (ret)
815 goto err_rel_entity2;
816
817 ret = media_create_pad_link(&func->proc, 1, &func->sink, 0,
818 MEDIA_LNK_FL_IMMUTABLE |
819 MEDIA_LNK_FL_ENABLED);
820 if (ret)
821 goto err_rm_links0;
822
823 /* Create video interface */
824 func->intf_devnode = media_devnode_create(mdev, MEDIA_INTF_T_V4L_VIDEO,
825 0, VIDEO_MAJOR,
826 func->vdev.minor);
827 if (!func->intf_devnode) {
828 ret = -ENOMEM;
829 goto err_rm_links1;
830 }
831
832 /* Connect the two DMA engines to the interface */
833 link = media_create_intf_link(&func->vdev.entity,
834 &func->intf_devnode->intf,
835 MEDIA_LNK_FL_IMMUTABLE |
836 MEDIA_LNK_FL_ENABLED);
837 if (!link) {
838 ret = -ENOMEM;
839 goto err_rm_devnode;
840 }
841
842 link = media_create_intf_link(&func->sink, &func->intf_devnode->intf,
843 MEDIA_LNK_FL_IMMUTABLE |
844 MEDIA_LNK_FL_ENABLED);
845 if (!link) {
846 ret = -ENOMEM;
847 goto err_rm_devnode;
848 }
849 return 0;
850
851 err_rm_devnode:
852 media_devnode_remove(func->intf_devnode);
853
854 err_rm_links1:
855 media_entity_remove_links(&func->sink);
856
857 err_rm_links0:
858 media_entity_remove_links(&func->proc);
859 media_entity_remove_links(&func->vdev.entity);
860
861 err_rel_entity2:
862 media_device_unregister_entity(&func->sink);
863
864 err_rel_entity1:
865 media_device_unregister_entity(&func->proc);
866
867 err_rel_entity0:
868 media_device_unregister_entity(&func->vdev.entity);
869 return ret;
870 }
871
hantro_detach_func(struct hantro_func * func)872 static void hantro_detach_func(struct hantro_func *func)
873 {
874 media_devnode_remove(func->intf_devnode);
875 media_entity_remove_links(&func->sink);
876 media_entity_remove_links(&func->proc);
877 media_entity_remove_links(&func->vdev.entity);
878 media_device_unregister_entity(&func->sink);
879 media_device_unregister_entity(&func->proc);
880 media_device_unregister_entity(&func->vdev.entity);
881 }
882
hantro_add_func(struct hantro_dev * vpu,unsigned int funcid)883 static int hantro_add_func(struct hantro_dev *vpu, unsigned int funcid)
884 {
885 const struct of_device_id *match;
886 struct hantro_func *func;
887 struct video_device *vfd;
888 int ret;
889
890 match = of_match_node(of_hantro_match, vpu->dev->of_node);
891 func = devm_kzalloc(vpu->dev, sizeof(*func), GFP_KERNEL);
892 if (!func) {
893 v4l2_err(&vpu->v4l2_dev, "Failed to allocate video device\n");
894 return -ENOMEM;
895 }
896
897 func->id = funcid;
898
899 vfd = &func->vdev;
900 vfd->fops = &hantro_fops;
901 vfd->release = video_device_release_empty;
902 vfd->lock = &vpu->vpu_mutex;
903 vfd->v4l2_dev = &vpu->v4l2_dev;
904 vfd->vfl_dir = VFL_DIR_M2M;
905 vfd->device_caps = V4L2_CAP_STREAMING | V4L2_CAP_VIDEO_M2M_MPLANE;
906 vfd->ioctl_ops = &hantro_ioctl_ops;
907 strscpy(vfd->name, match->compatible, sizeof(vfd->name));
908 strlcat(vfd->name, funcid == MEDIA_ENT_F_PROC_VIDEO_ENCODER ?
909 "-enc" : "-dec", sizeof(vfd->name));
910
911 if (funcid == MEDIA_ENT_F_PROC_VIDEO_ENCODER) {
912 vpu->encoder = func;
913 v4l2_disable_ioctl(vfd, VIDIOC_TRY_DECODER_CMD);
914 v4l2_disable_ioctl(vfd, VIDIOC_DECODER_CMD);
915 } else {
916 vpu->decoder = func;
917 v4l2_disable_ioctl(vfd, VIDIOC_TRY_ENCODER_CMD);
918 v4l2_disable_ioctl(vfd, VIDIOC_ENCODER_CMD);
919 v4l2_disable_ioctl(vfd, VIDIOC_G_SELECTION);
920 v4l2_disable_ioctl(vfd, VIDIOC_S_SELECTION);
921 }
922
923 video_set_drvdata(vfd, vpu);
924
925 ret = video_register_device(vfd, VFL_TYPE_VIDEO, -1);
926 if (ret) {
927 v4l2_err(&vpu->v4l2_dev, "Failed to register video device\n");
928 return ret;
929 }
930
931 ret = hantro_attach_func(vpu, func);
932 if (ret) {
933 v4l2_err(&vpu->v4l2_dev,
934 "Failed to attach functionality to the media device\n");
935 goto err_unreg_dev;
936 }
937
938 v4l2_info(&vpu->v4l2_dev, "registered %s as /dev/video%d\n", vfd->name,
939 vfd->num);
940
941 return 0;
942
943 err_unreg_dev:
944 video_unregister_device(vfd);
945 return ret;
946 }
947
hantro_add_enc_func(struct hantro_dev * vpu)948 static int hantro_add_enc_func(struct hantro_dev *vpu)
949 {
950 if (!vpu->variant->enc_fmts)
951 return 0;
952
953 return hantro_add_func(vpu, MEDIA_ENT_F_PROC_VIDEO_ENCODER);
954 }
955
hantro_add_dec_func(struct hantro_dev * vpu)956 static int hantro_add_dec_func(struct hantro_dev *vpu)
957 {
958 if (!vpu->variant->dec_fmts)
959 return 0;
960
961 return hantro_add_func(vpu, MEDIA_ENT_F_PROC_VIDEO_DECODER);
962 }
963
hantro_remove_func(struct hantro_dev * vpu,unsigned int funcid)964 static void hantro_remove_func(struct hantro_dev *vpu,
965 unsigned int funcid)
966 {
967 struct hantro_func *func;
968
969 if (funcid == MEDIA_ENT_F_PROC_VIDEO_ENCODER)
970 func = vpu->encoder;
971 else
972 func = vpu->decoder;
973
974 if (!func)
975 return;
976
977 hantro_detach_func(func);
978 video_unregister_device(&func->vdev);
979 }
980
hantro_remove_enc_func(struct hantro_dev * vpu)981 static void hantro_remove_enc_func(struct hantro_dev *vpu)
982 {
983 hantro_remove_func(vpu, MEDIA_ENT_F_PROC_VIDEO_ENCODER);
984 }
985
hantro_remove_dec_func(struct hantro_dev * vpu)986 static void hantro_remove_dec_func(struct hantro_dev *vpu)
987 {
988 hantro_remove_func(vpu, MEDIA_ENT_F_PROC_VIDEO_DECODER);
989 }
990
991 static const struct media_device_ops hantro_m2m_media_ops = {
992 .req_validate = vb2_request_validate,
993 .req_queue = v4l2_m2m_request_queue,
994 };
995
996 /*
997 * Some SoCs, like RK3588 have multiple identical Hantro cores, but the
998 * kernel is currently missing support for multi-core handling. Exposing
999 * separate devices for each core to userspace is bad, since that does
1000 * not allow scheduling tasks properly (and creates ABI). With this workaround
1001 * the driver will only probe for the first core and early exit for the other
1002 * cores. Once the driver gains multi-core support, the same technique
1003 * for detecting the main core can be used to cluster all cores together.
1004 */
hantro_disable_multicore(struct hantro_dev * vpu)1005 static int hantro_disable_multicore(struct hantro_dev *vpu)
1006 {
1007 struct device_node *node = NULL;
1008 const char *compatible;
1009 bool is_main_core;
1010 int ret;
1011
1012 /* Intentionally ignores the fallback strings */
1013 ret = of_property_read_string(vpu->dev->of_node, "compatible", &compatible);
1014 if (ret)
1015 return ret;
1016
1017 /* The first compatible and available node found is considered the main core */
1018 do {
1019 node = of_find_compatible_node(node, NULL, compatible);
1020 if (of_device_is_available(node))
1021 break;
1022 } while (node);
1023
1024 if (!node)
1025 return -EINVAL;
1026
1027 is_main_core = (vpu->dev->of_node == node);
1028
1029 of_node_put(node);
1030
1031 if (!is_main_core) {
1032 dev_info(vpu->dev, "missing multi-core support, ignoring this instance\n");
1033 return -ENODEV;
1034 }
1035
1036 return 0;
1037 }
1038
hantro_get_v4l2_m2m_dev(struct hantro_dev * vpu)1039 static struct v4l2_m2m_dev *hantro_get_v4l2_m2m_dev(struct hantro_dev *vpu)
1040 {
1041 struct device_node *node;
1042 struct hantro_dev *shared_vpu;
1043
1044 if (!vpu->variant || !vpu->variant->shared_devices)
1045 goto init_new_m2m_dev;
1046
1047 for_each_matching_node(node, vpu->variant->shared_devices) {
1048 struct platform_device *pdev;
1049 struct v4l2_m2m_dev *m2m_dev;
1050
1051 pdev = of_find_device_by_node(node);
1052 if (!pdev)
1053 continue;
1054
1055 shared_vpu = platform_get_drvdata(pdev);
1056 if (IS_ERR_OR_NULL(shared_vpu) || shared_vpu == vpu) {
1057 platform_device_put(pdev);
1058 continue;
1059 }
1060
1061 v4l2_m2m_get(shared_vpu->m2m_dev);
1062 m2m_dev = shared_vpu->m2m_dev;
1063 platform_device_put(pdev);
1064
1065 of_node_put(node);
1066
1067 return m2m_dev;
1068 }
1069
1070 init_new_m2m_dev:
1071 return v4l2_m2m_init(&vpu_m2m_ops);
1072 }
1073
hantro_probe(struct platform_device * pdev)1074 static int hantro_probe(struct platform_device *pdev)
1075 {
1076 const struct of_device_id *match;
1077 struct hantro_dev *vpu;
1078 int num_bases;
1079 int i, ret;
1080
1081 vpu = devm_kzalloc(&pdev->dev, sizeof(*vpu), GFP_KERNEL);
1082 if (!vpu)
1083 return -ENOMEM;
1084
1085 vpu->dev = &pdev->dev;
1086 vpu->pdev = pdev;
1087 mutex_init(&vpu->vpu_mutex);
1088 spin_lock_init(&vpu->irqlock);
1089
1090 match = of_match_node(of_hantro_match, pdev->dev.of_node);
1091 vpu->variant = match->data;
1092
1093 ret = hantro_disable_multicore(vpu);
1094 if (ret)
1095 return ret;
1096
1097 /*
1098 * Support for nxp,imx8mq-vpu is kept for backwards compatibility
1099 * but it's deprecated. Please update your DTS file to use
1100 * nxp,imx8mq-vpu-g1 or nxp,imx8mq-vpu-g2 instead.
1101 */
1102 if (of_device_is_compatible(pdev->dev.of_node, "nxp,imx8mq-vpu"))
1103 dev_warn(&pdev->dev, "%s compatible is deprecated\n",
1104 match->compatible);
1105
1106 INIT_DELAYED_WORK(&vpu->watchdog_work, hantro_watchdog);
1107
1108 vpu->clocks = devm_kcalloc(&pdev->dev, vpu->variant->num_clocks,
1109 sizeof(*vpu->clocks), GFP_KERNEL);
1110 if (!vpu->clocks)
1111 return -ENOMEM;
1112
1113 if (vpu->variant->num_clocks > 1) {
1114 for (i = 0; i < vpu->variant->num_clocks; i++)
1115 vpu->clocks[i].id = vpu->variant->clk_names[i];
1116
1117 ret = devm_clk_bulk_get(&pdev->dev, vpu->variant->num_clocks,
1118 vpu->clocks);
1119 if (ret)
1120 return ret;
1121 } else {
1122 /*
1123 * If the driver has a single clk, chances are there will be no
1124 * actual name in the DT bindings.
1125 */
1126 vpu->clocks[0].clk = devm_clk_get(&pdev->dev, NULL);
1127 if (IS_ERR(vpu->clocks[0].clk))
1128 return PTR_ERR(vpu->clocks[0].clk);
1129 }
1130
1131 vpu->resets = devm_reset_control_array_get_optional_exclusive(&pdev->dev);
1132 if (IS_ERR(vpu->resets))
1133 return PTR_ERR(vpu->resets);
1134
1135 num_bases = vpu->variant->num_regs ?: 1;
1136 vpu->reg_bases = devm_kcalloc(&pdev->dev, num_bases,
1137 sizeof(*vpu->reg_bases), GFP_KERNEL);
1138 if (!vpu->reg_bases)
1139 return -ENOMEM;
1140
1141 for (i = 0; i < num_bases; i++) {
1142 vpu->reg_bases[i] = vpu->variant->reg_names ?
1143 devm_platform_ioremap_resource_byname(pdev, vpu->variant->reg_names[i]) :
1144 devm_platform_ioremap_resource(pdev, 0);
1145 if (IS_ERR(vpu->reg_bases[i]))
1146 return PTR_ERR(vpu->reg_bases[i]);
1147 }
1148 vpu->enc_base = vpu->reg_bases[0] + vpu->variant->enc_offset;
1149 vpu->dec_base = vpu->reg_bases[0] + vpu->variant->dec_offset;
1150
1151 /**
1152 * TODO: Eventually allow taking advantage of full 64-bit address space.
1153 * Until then we assume the MSB portion of buffers' base addresses is
1154 * always 0 due to this masking operation.
1155 */
1156 ret = dma_set_coherent_mask(vpu->dev, DMA_BIT_MASK(32));
1157 if (ret) {
1158 dev_err(vpu->dev, "Could not set DMA coherent mask.\n");
1159 return ret;
1160 }
1161 vb2_dma_contig_set_max_seg_size(&pdev->dev, DMA_BIT_MASK(32));
1162
1163 for (i = 0; i < vpu->variant->num_irqs; i++) {
1164 const char *irq_name;
1165 int irq;
1166
1167 if (!vpu->variant->irqs[i].handler)
1168 continue;
1169
1170 if (vpu->variant->num_irqs > 1) {
1171 irq_name = vpu->variant->irqs[i].name;
1172 irq = platform_get_irq_byname(vpu->pdev, irq_name);
1173 } else {
1174 /*
1175 * If the driver has a single IRQ, chances are there
1176 * will be no actual name in the DT bindings.
1177 */
1178 irq_name = "default";
1179 irq = platform_get_irq(vpu->pdev, 0);
1180 }
1181 if (irq < 0)
1182 return irq;
1183
1184 ret = devm_request_irq(vpu->dev, irq,
1185 vpu->variant->irqs[i].handler, 0,
1186 dev_name(vpu->dev), vpu);
1187 if (ret) {
1188 dev_err(vpu->dev, "Could not request %s IRQ.\n",
1189 irq_name);
1190 return ret;
1191 }
1192 }
1193
1194 if (vpu->variant->init) {
1195 ret = vpu->variant->init(vpu);
1196 if (ret) {
1197 dev_err(&pdev->dev, "Failed to init VPU hardware\n");
1198 return ret;
1199 }
1200 }
1201
1202 pm_runtime_set_autosuspend_delay(vpu->dev, 100);
1203 pm_runtime_use_autosuspend(vpu->dev);
1204 pm_runtime_enable(vpu->dev);
1205
1206 ret = reset_control_deassert(vpu->resets);
1207 if (ret) {
1208 dev_err(&pdev->dev, "Failed to deassert resets\n");
1209 goto err_pm_disable;
1210 }
1211
1212 ret = clk_bulk_prepare(vpu->variant->num_clocks, vpu->clocks);
1213 if (ret) {
1214 dev_err(&pdev->dev, "Failed to prepare clocks\n");
1215 goto err_rst_assert;
1216 }
1217
1218 ret = v4l2_device_register(&pdev->dev, &vpu->v4l2_dev);
1219 if (ret) {
1220 dev_err(&pdev->dev, "Failed to register v4l2 device\n");
1221 goto err_clk_unprepare;
1222 }
1223 platform_set_drvdata(pdev, vpu);
1224
1225 vpu->m2m_dev = hantro_get_v4l2_m2m_dev(vpu);
1226 if (IS_ERR(vpu->m2m_dev)) {
1227 v4l2_err(&vpu->v4l2_dev, "Failed to init mem2mem device\n");
1228 ret = PTR_ERR(vpu->m2m_dev);
1229 goto err_v4l2_unreg;
1230 }
1231
1232 vpu->mdev.dev = vpu->dev;
1233 strscpy(vpu->mdev.model, DRIVER_NAME, sizeof(vpu->mdev.model));
1234 media_device_init(&vpu->mdev);
1235 vpu->mdev.ops = &hantro_m2m_media_ops;
1236 vpu->v4l2_dev.mdev = &vpu->mdev;
1237
1238 ret = hantro_add_enc_func(vpu);
1239 if (ret) {
1240 dev_err(&pdev->dev, "Failed to register encoder\n");
1241 goto err_m2m_rel;
1242 }
1243
1244 ret = hantro_add_dec_func(vpu);
1245 if (ret) {
1246 dev_err(&pdev->dev, "Failed to register decoder\n");
1247 goto err_rm_enc_func;
1248 }
1249
1250 ret = media_device_register(&vpu->mdev);
1251 if (ret) {
1252 v4l2_err(&vpu->v4l2_dev, "Failed to register mem2mem media device\n");
1253 goto err_rm_dec_func;
1254 }
1255
1256 return 0;
1257
1258 err_rm_dec_func:
1259 hantro_remove_dec_func(vpu);
1260 err_rm_enc_func:
1261 hantro_remove_enc_func(vpu);
1262 err_m2m_rel:
1263 media_device_cleanup(&vpu->mdev);
1264 v4l2_m2m_put(vpu->m2m_dev);
1265 err_v4l2_unreg:
1266 v4l2_device_unregister(&vpu->v4l2_dev);
1267 err_clk_unprepare:
1268 clk_bulk_unprepare(vpu->variant->num_clocks, vpu->clocks);
1269 err_rst_assert:
1270 reset_control_assert(vpu->resets);
1271 err_pm_disable:
1272 pm_runtime_dont_use_autosuspend(vpu->dev);
1273 pm_runtime_disable(vpu->dev);
1274 return ret;
1275 }
1276
hantro_remove(struct platform_device * pdev)1277 static void hantro_remove(struct platform_device *pdev)
1278 {
1279 struct hantro_dev *vpu = platform_get_drvdata(pdev);
1280
1281 v4l2_info(&vpu->v4l2_dev, "Removing %s\n", pdev->name);
1282
1283 media_device_unregister(&vpu->mdev);
1284 hantro_remove_dec_func(vpu);
1285 hantro_remove_enc_func(vpu);
1286 media_device_cleanup(&vpu->mdev);
1287 v4l2_m2m_put(vpu->m2m_dev);
1288 v4l2_device_unregister(&vpu->v4l2_dev);
1289 clk_bulk_unprepare(vpu->variant->num_clocks, vpu->clocks);
1290 reset_control_assert(vpu->resets);
1291 pm_runtime_dont_use_autosuspend(vpu->dev);
1292 pm_runtime_disable(vpu->dev);
1293 }
1294
1295 #ifdef CONFIG_PM
hantro_runtime_resume(struct device * dev)1296 static int hantro_runtime_resume(struct device *dev)
1297 {
1298 struct hantro_dev *vpu = dev_get_drvdata(dev);
1299
1300 if (vpu->variant->runtime_resume)
1301 return vpu->variant->runtime_resume(vpu);
1302
1303 return 0;
1304 }
1305 #endif
1306
1307 static const struct dev_pm_ops hantro_pm_ops = {
1308 SET_SYSTEM_SLEEP_PM_OPS(pm_runtime_force_suspend,
1309 pm_runtime_force_resume)
1310 SET_RUNTIME_PM_OPS(NULL, hantro_runtime_resume, NULL)
1311 };
1312
1313 static struct platform_driver hantro_driver = {
1314 .probe = hantro_probe,
1315 .remove = hantro_remove,
1316 .driver = {
1317 .name = DRIVER_NAME,
1318 .of_match_table = of_hantro_match,
1319 .pm = &hantro_pm_ops,
1320 },
1321 };
1322 module_platform_driver(hantro_driver);
1323
1324 MODULE_LICENSE("GPL v2");
1325 MODULE_AUTHOR("Alpha Lin <Alpha.Lin@Rock-Chips.com>");
1326 MODULE_AUTHOR("Tomasz Figa <tfiga@chromium.org>");
1327 MODULE_AUTHOR("Ezequiel Garcia <ezequiel@collabora.com>");
1328 MODULE_DESCRIPTION("Hantro VPU codec driver");
1329