xref: /linux/drivers/media/platform/qcom/camss/camss-vfe-gen1.c (revision a4eb44a6435d6d8f9e642407a4a06f65eb90ca04)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * camss-vfe-gen1.c
4  *
5  * Qualcomm MSM Camera Subsystem - VFE Common functionality for Gen 1 versions of hw (4.1, 4.7..)
6  *
7  * Copyright (C) 2020 Linaro Ltd.
8  */
9 
10 #include "camss.h"
11 #include "camss-vfe.h"
12 #include "camss-vfe-gen1.h"
13 
14 /* Max number of frame drop updates per frame */
15 #define VFE_FRAME_DROP_UPDATES 2
16 #define VFE_NEXT_SOF_MS 500
17 
18 int vfe_gen1_halt(struct vfe_device *vfe)
19 {
20 	unsigned long time;
21 
22 	reinit_completion(&vfe->halt_complete);
23 
24 	vfe->ops_gen1->halt_request(vfe);
25 
26 	time = wait_for_completion_timeout(&vfe->halt_complete,
27 					   msecs_to_jiffies(VFE_HALT_TIMEOUT_MS));
28 	if (!time) {
29 		dev_err(vfe->camss->dev, "VFE halt timeout\n");
30 		return -EIO;
31 	}
32 
33 	return 0;
34 }
35 
36 static int vfe_disable_output(struct vfe_line *line)
37 {
38 	struct vfe_device *vfe = to_vfe(line);
39 	struct vfe_output *output = &line->output;
40 	const struct vfe_hw_ops *ops = vfe->ops;
41 	unsigned long flags;
42 	unsigned long time;
43 	unsigned int i;
44 
45 	spin_lock_irqsave(&vfe->output_lock, flags);
46 
47 	output->gen1.wait_sof = 1;
48 	spin_unlock_irqrestore(&vfe->output_lock, flags);
49 
50 	time = wait_for_completion_timeout(&output->sof, msecs_to_jiffies(VFE_NEXT_SOF_MS));
51 	if (!time)
52 		dev_err(vfe->camss->dev, "VFE sof timeout\n");
53 
54 	spin_lock_irqsave(&vfe->output_lock, flags);
55 	for (i = 0; i < output->wm_num; i++)
56 		vfe->ops_gen1->wm_enable(vfe, output->wm_idx[i], 0);
57 
58 	ops->reg_update(vfe, line->id);
59 	output->wait_reg_update = 1;
60 	spin_unlock_irqrestore(&vfe->output_lock, flags);
61 
62 	time = wait_for_completion_timeout(&output->reg_update, msecs_to_jiffies(VFE_NEXT_SOF_MS));
63 	if (!time)
64 		dev_err(vfe->camss->dev, "VFE reg update timeout\n");
65 
66 	spin_lock_irqsave(&vfe->output_lock, flags);
67 
68 	if (line->id != VFE_LINE_PIX) {
69 		vfe->ops_gen1->wm_frame_based(vfe, output->wm_idx[0], 0);
70 		vfe->ops_gen1->bus_disconnect_wm_from_rdi(vfe, output->wm_idx[0], line->id);
71 		vfe->ops_gen1->enable_irq_wm_line(vfe, output->wm_idx[0], line->id, 0);
72 		vfe->ops_gen1->set_cgc_override(vfe, output->wm_idx[0], 0);
73 		spin_unlock_irqrestore(&vfe->output_lock, flags);
74 	} else {
75 		for (i = 0; i < output->wm_num; i++) {
76 			vfe->ops_gen1->wm_line_based(vfe, output->wm_idx[i], NULL, i, 0);
77 			vfe->ops_gen1->set_cgc_override(vfe, output->wm_idx[i], 0);
78 		}
79 
80 		vfe->ops_gen1->enable_irq_pix_line(vfe, 0, line->id, 0);
81 		vfe->ops_gen1->set_module_cfg(vfe, 0);
82 		vfe->ops_gen1->set_realign_cfg(vfe, line, 0);
83 		vfe->ops_gen1->set_xbar_cfg(vfe, output, 0);
84 		vfe->ops_gen1->set_camif_cmd(vfe, 0);
85 
86 		spin_unlock_irqrestore(&vfe->output_lock, flags);
87 
88 		vfe->ops_gen1->camif_wait_for_stop(vfe, vfe->camss->dev);
89 	}
90 
91 	return 0;
92 }
93 
94 /*
95  * vfe_gen1_disable - Disable streaming on VFE line
96  * @line: VFE line
97  *
98  * Return 0 on success or a negative error code otherwise
99  */
100 int vfe_gen1_disable(struct vfe_line *line)
101 {
102 	struct vfe_device *vfe = to_vfe(line);
103 
104 	vfe_disable_output(line);
105 
106 	vfe_put_output(line);
107 
108 	mutex_lock(&vfe->stream_lock);
109 
110 	if (vfe->stream_count == 1)
111 		vfe->ops_gen1->bus_enable_wr_if(vfe, 0);
112 
113 	vfe->stream_count--;
114 
115 	mutex_unlock(&vfe->stream_lock);
116 
117 	return 0;
118 }
119 
120 static void vfe_output_init_addrs(struct vfe_device *vfe,
121 				  struct vfe_output *output, u8 sync,
122 				  struct vfe_line *line)
123 {
124 	u32 ping_addr;
125 	u32 pong_addr;
126 	unsigned int i;
127 
128 	output->gen1.active_buf = 0;
129 
130 	for (i = 0; i < output->wm_num; i++) {
131 		if (output->buf[0])
132 			ping_addr = output->buf[0]->addr[i];
133 		else
134 			ping_addr = 0;
135 
136 		if (output->buf[1])
137 			pong_addr = output->buf[1]->addr[i];
138 		else
139 			pong_addr = ping_addr;
140 
141 		vfe->ops_gen1->wm_set_ping_addr(vfe, output->wm_idx[i], ping_addr);
142 		vfe->ops_gen1->wm_set_pong_addr(vfe, output->wm_idx[i], pong_addr);
143 		if (sync)
144 			vfe->ops_gen1->bus_reload_wm(vfe, output->wm_idx[i]);
145 	}
146 }
147 
148 static void vfe_output_frame_drop(struct vfe_device *vfe,
149 				  struct vfe_output *output,
150 				  u32 drop_pattern)
151 {
152 	u8 drop_period;
153 	unsigned int i;
154 
155 	/* We need to toggle update period to be valid on next frame */
156 	output->drop_update_idx++;
157 	output->drop_update_idx %= VFE_FRAME_DROP_UPDATES;
158 	drop_period = VFE_FRAME_DROP_VAL + output->drop_update_idx;
159 
160 	for (i = 0; i < output->wm_num; i++) {
161 		vfe->ops_gen1->wm_set_framedrop_period(vfe, output->wm_idx[i], drop_period);
162 		vfe->ops_gen1->wm_set_framedrop_pattern(vfe, output->wm_idx[i], drop_pattern);
163 	}
164 
165 	vfe->ops->reg_update(vfe, container_of(output, struct vfe_line, output)->id);
166 }
167 
168 static int vfe_enable_output(struct vfe_line *line)
169 {
170 	struct vfe_device *vfe = to_vfe(line);
171 	struct vfe_output *output = &line->output;
172 	const struct vfe_hw_ops *ops = vfe->ops;
173 	struct media_entity *sensor;
174 	unsigned long flags;
175 	unsigned int frame_skip = 0;
176 	unsigned int i;
177 	u16 ub_size;
178 
179 	ub_size = vfe->ops_gen1->get_ub_size(vfe->id);
180 	if (!ub_size)
181 		return -EINVAL;
182 
183 	sensor = camss_find_sensor(&line->subdev.entity);
184 	if (sensor) {
185 		struct v4l2_subdev *subdev = media_entity_to_v4l2_subdev(sensor);
186 
187 		v4l2_subdev_call(subdev, sensor, g_skip_frames, &frame_skip);
188 		/* Max frame skip is 29 frames */
189 		if (frame_skip > VFE_FRAME_DROP_VAL - 1)
190 			frame_skip = VFE_FRAME_DROP_VAL - 1;
191 	}
192 
193 	spin_lock_irqsave(&vfe->output_lock, flags);
194 
195 	ops->reg_update_clear(vfe, line->id);
196 
197 	if (output->state != VFE_OUTPUT_RESERVED) {
198 		dev_err(vfe->camss->dev, "Output is not in reserved state %d\n", output->state);
199 		spin_unlock_irqrestore(&vfe->output_lock, flags);
200 		return -EINVAL;
201 	}
202 	output->state = VFE_OUTPUT_IDLE;
203 
204 	output->buf[0] = vfe_buf_get_pending(output);
205 	output->buf[1] = vfe_buf_get_pending(output);
206 
207 	if (!output->buf[0] && output->buf[1]) {
208 		output->buf[0] = output->buf[1];
209 		output->buf[1] = NULL;
210 	}
211 
212 	if (output->buf[0])
213 		output->state = VFE_OUTPUT_SINGLE;
214 
215 	if (output->buf[1])
216 		output->state = VFE_OUTPUT_CONTINUOUS;
217 
218 	switch (output->state) {
219 	case VFE_OUTPUT_SINGLE:
220 		vfe_output_frame_drop(vfe, output, 1 << frame_skip);
221 		break;
222 	case VFE_OUTPUT_CONTINUOUS:
223 		vfe_output_frame_drop(vfe, output, 3 << frame_skip);
224 		break;
225 	default:
226 		vfe_output_frame_drop(vfe, output, 0);
227 		break;
228 	}
229 
230 	output->sequence = 0;
231 	output->gen1.wait_sof = 0;
232 	output->wait_reg_update = 0;
233 	reinit_completion(&output->sof);
234 	reinit_completion(&output->reg_update);
235 
236 	vfe_output_init_addrs(vfe, output, 0, line);
237 
238 	if (line->id != VFE_LINE_PIX) {
239 		vfe->ops_gen1->set_cgc_override(vfe, output->wm_idx[0], 1);
240 		vfe->ops_gen1->enable_irq_wm_line(vfe, output->wm_idx[0], line->id, 1);
241 		vfe->ops_gen1->bus_connect_wm_to_rdi(vfe, output->wm_idx[0], line->id);
242 		vfe->ops_gen1->wm_set_subsample(vfe, output->wm_idx[0]);
243 		vfe->ops_gen1->set_rdi_cid(vfe, line->id, 0);
244 		vfe->ops_gen1->wm_set_ub_cfg(vfe, output->wm_idx[0],
245 					    (ub_size + 1) * output->wm_idx[0], ub_size);
246 		vfe->ops_gen1->wm_frame_based(vfe, output->wm_idx[0], 1);
247 		vfe->ops_gen1->wm_enable(vfe, output->wm_idx[0], 1);
248 		vfe->ops_gen1->bus_reload_wm(vfe, output->wm_idx[0]);
249 	} else {
250 		ub_size /= output->wm_num;
251 		for (i = 0; i < output->wm_num; i++) {
252 			vfe->ops_gen1->set_cgc_override(vfe, output->wm_idx[i], 1);
253 			vfe->ops_gen1->wm_set_subsample(vfe, output->wm_idx[i]);
254 			vfe->ops_gen1->wm_set_ub_cfg(vfe, output->wm_idx[i],
255 						     (ub_size + 1) * output->wm_idx[i], ub_size);
256 			vfe->ops_gen1->wm_line_based(vfe, output->wm_idx[i],
257 						     &line->video_out.active_fmt.fmt.pix_mp, i, 1);
258 			vfe->ops_gen1->wm_enable(vfe, output->wm_idx[i], 1);
259 			vfe->ops_gen1->bus_reload_wm(vfe, output->wm_idx[i]);
260 		}
261 		vfe->ops_gen1->enable_irq_pix_line(vfe, 0, line->id, 1);
262 		vfe->ops_gen1->set_module_cfg(vfe, 1);
263 		vfe->ops_gen1->set_camif_cfg(vfe, line);
264 		vfe->ops_gen1->set_realign_cfg(vfe, line, 1);
265 		vfe->ops_gen1->set_xbar_cfg(vfe, output, 1);
266 		vfe->ops_gen1->set_demux_cfg(vfe, line);
267 		vfe->ops_gen1->set_scale_cfg(vfe, line);
268 		vfe->ops_gen1->set_crop_cfg(vfe, line);
269 		vfe->ops_gen1->set_clamp_cfg(vfe);
270 		vfe->ops_gen1->set_camif_cmd(vfe, 1);
271 	}
272 
273 	ops->reg_update(vfe, line->id);
274 
275 	spin_unlock_irqrestore(&vfe->output_lock, flags);
276 
277 	return 0;
278 }
279 
280 static int vfe_get_output(struct vfe_line *line)
281 {
282 	struct vfe_device *vfe = to_vfe(line);
283 	struct vfe_output *output;
284 	struct v4l2_format *f = &line->video_out.active_fmt;
285 	unsigned long flags;
286 	int i;
287 	int wm_idx;
288 
289 	spin_lock_irqsave(&vfe->output_lock, flags);
290 
291 	output = &line->output;
292 	if (output->state != VFE_OUTPUT_OFF) {
293 		dev_err(vfe->camss->dev, "Output is running\n");
294 		goto error;
295 	}
296 	output->state = VFE_OUTPUT_RESERVED;
297 
298 	output->gen1.active_buf = 0;
299 
300 	switch (f->fmt.pix_mp.pixelformat) {
301 	case V4L2_PIX_FMT_NV12:
302 	case V4L2_PIX_FMT_NV21:
303 	case V4L2_PIX_FMT_NV16:
304 	case V4L2_PIX_FMT_NV61:
305 		output->wm_num = 2;
306 		break;
307 	default:
308 		output->wm_num = 1;
309 		break;
310 	}
311 
312 	for (i = 0; i < output->wm_num; i++) {
313 		wm_idx = vfe_reserve_wm(vfe, line->id);
314 		if (wm_idx < 0) {
315 			dev_err(vfe->camss->dev, "Can not reserve wm\n");
316 			goto error_get_wm;
317 		}
318 		output->wm_idx[i] = wm_idx;
319 	}
320 
321 	output->drop_update_idx = 0;
322 
323 	spin_unlock_irqrestore(&vfe->output_lock, flags);
324 
325 	return 0;
326 
327 error_get_wm:
328 	for (i--; i >= 0; i--)
329 		vfe_release_wm(vfe, output->wm_idx[i]);
330 	output->state = VFE_OUTPUT_OFF;
331 error:
332 	spin_unlock_irqrestore(&vfe->output_lock, flags);
333 
334 	return -EINVAL;
335 }
336 
337 int vfe_gen1_enable(struct vfe_line *line)
338 {
339 	struct vfe_device *vfe = to_vfe(line);
340 	int ret;
341 
342 	mutex_lock(&vfe->stream_lock);
343 
344 	if (!vfe->stream_count) {
345 		vfe->ops_gen1->enable_irq_common(vfe);
346 		vfe->ops_gen1->bus_enable_wr_if(vfe, 1);
347 		vfe->ops_gen1->set_qos(vfe);
348 		vfe->ops_gen1->set_ds(vfe);
349 	}
350 
351 	vfe->stream_count++;
352 
353 	mutex_unlock(&vfe->stream_lock);
354 
355 	ret = vfe_get_output(line);
356 	if (ret < 0)
357 		goto error_get_output;
358 
359 	ret = vfe_enable_output(line);
360 	if (ret < 0)
361 		goto error_enable_output;
362 
363 	vfe->was_streaming = 1;
364 
365 	return 0;
366 
367 error_enable_output:
368 	vfe_put_output(line);
369 
370 error_get_output:
371 	mutex_lock(&vfe->stream_lock);
372 
373 	if (vfe->stream_count == 1)
374 		vfe->ops_gen1->bus_enable_wr_if(vfe, 0);
375 
376 	vfe->stream_count--;
377 
378 	mutex_unlock(&vfe->stream_lock);
379 
380 	return ret;
381 }
382 
383 static void vfe_output_update_ping_addr(struct vfe_device *vfe,
384 					struct vfe_output *output, u8 sync,
385 					struct vfe_line *line)
386 {
387 	u32 addr;
388 	unsigned int i;
389 
390 	for (i = 0; i < output->wm_num; i++) {
391 		if (output->buf[0])
392 			addr = output->buf[0]->addr[i];
393 		else
394 			addr = 0;
395 
396 		vfe->ops_gen1->wm_set_ping_addr(vfe, output->wm_idx[i], addr);
397 		if (sync)
398 			vfe->ops_gen1->bus_reload_wm(vfe, output->wm_idx[i]);
399 	}
400 }
401 
402 static void vfe_output_update_pong_addr(struct vfe_device *vfe,
403 					struct vfe_output *output, u8 sync,
404 					struct vfe_line *line)
405 {
406 	u32 addr;
407 	unsigned int i;
408 
409 	for (i = 0; i < output->wm_num; i++) {
410 		if (output->buf[1])
411 			addr = output->buf[1]->addr[i];
412 		else
413 			addr = 0;
414 
415 		vfe->ops_gen1->wm_set_pong_addr(vfe, output->wm_idx[i], addr);
416 		if (sync)
417 			vfe->ops_gen1->bus_reload_wm(vfe, output->wm_idx[i]);
418 	}
419 }
420 
421 static void vfe_buf_update_wm_on_next(struct vfe_device *vfe,
422 				      struct vfe_output *output)
423 {
424 	switch (output->state) {
425 	case VFE_OUTPUT_CONTINUOUS:
426 		vfe_output_frame_drop(vfe, output, 3);
427 		break;
428 	case VFE_OUTPUT_SINGLE:
429 	default:
430 		dev_err_ratelimited(vfe->camss->dev,
431 				    "Next buf in wrong state! %d\n",
432 				    output->state);
433 		break;
434 	}
435 }
436 
437 static void vfe_buf_update_wm_on_last(struct vfe_device *vfe,
438 				      struct vfe_output *output)
439 {
440 	switch (output->state) {
441 	case VFE_OUTPUT_CONTINUOUS:
442 		output->state = VFE_OUTPUT_SINGLE;
443 		vfe_output_frame_drop(vfe, output, 1);
444 		break;
445 	case VFE_OUTPUT_SINGLE:
446 		output->state = VFE_OUTPUT_STOPPING;
447 		vfe_output_frame_drop(vfe, output, 0);
448 		break;
449 	default:
450 		dev_err_ratelimited(vfe->camss->dev,
451 				    "Last buff in wrong state! %d\n",
452 				    output->state);
453 		break;
454 	}
455 }
456 
457 static void vfe_buf_update_wm_on_new(struct vfe_device *vfe,
458 				     struct vfe_output *output,
459 				     struct camss_buffer *new_buf,
460 				     struct vfe_line *line)
461 {
462 	int inactive_idx;
463 
464 	switch (output->state) {
465 	case VFE_OUTPUT_SINGLE:
466 		inactive_idx = !output->gen1.active_buf;
467 
468 		if (!output->buf[inactive_idx]) {
469 			output->buf[inactive_idx] = new_buf;
470 
471 			if (inactive_idx)
472 				vfe_output_update_pong_addr(vfe, output, 0, line);
473 			else
474 				vfe_output_update_ping_addr(vfe, output, 0, line);
475 
476 			vfe_output_frame_drop(vfe, output, 3);
477 			output->state = VFE_OUTPUT_CONTINUOUS;
478 		} else {
479 			vfe_buf_add_pending(output, new_buf);
480 			dev_err_ratelimited(vfe->camss->dev,
481 					    "Inactive buffer is busy\n");
482 		}
483 		break;
484 
485 	case VFE_OUTPUT_IDLE:
486 		if (!output->buf[0]) {
487 			output->buf[0] = new_buf;
488 
489 			vfe_output_init_addrs(vfe, output, 1, line);
490 			vfe_output_frame_drop(vfe, output, 1);
491 
492 			output->state = VFE_OUTPUT_SINGLE;
493 		} else {
494 			vfe_buf_add_pending(output, new_buf);
495 			dev_err_ratelimited(vfe->camss->dev,
496 					    "Output idle with buffer set!\n");
497 		}
498 		break;
499 
500 	case VFE_OUTPUT_CONTINUOUS:
501 	default:
502 		vfe_buf_add_pending(output, new_buf);
503 		break;
504 	}
505 }
506 
507 /*
508  * vfe_isr_halt_ack - Process halt ack
509  * @vfe: VFE Device
510  */
511 static void vfe_isr_halt_ack(struct vfe_device *vfe)
512 {
513 	complete(&vfe->halt_complete);
514 	vfe->ops_gen1->halt_clear(vfe);
515 }
516 
517 /*
518  * vfe_isr_sof - Process start of frame interrupt
519  * @vfe: VFE Device
520  * @line_id: VFE line
521  */
522 static void vfe_isr_sof(struct vfe_device *vfe, enum vfe_line_id line_id)
523 {
524 	struct vfe_output *output;
525 	unsigned long flags;
526 
527 	spin_lock_irqsave(&vfe->output_lock, flags);
528 	output = &vfe->line[line_id].output;
529 	if (output->gen1.wait_sof) {
530 		output->gen1.wait_sof = 0;
531 		complete(&output->sof);
532 	}
533 	spin_unlock_irqrestore(&vfe->output_lock, flags);
534 }
535 
536 /*
537  * vfe_isr_reg_update - Process reg update interrupt
538  * @vfe: VFE Device
539  * @line_id: VFE line
540  */
541 static void vfe_isr_reg_update(struct vfe_device *vfe, enum vfe_line_id line_id)
542 {
543 	struct vfe_output *output;
544 	struct vfe_line *line = &vfe->line[line_id];
545 	unsigned long flags;
546 
547 	spin_lock_irqsave(&vfe->output_lock, flags);
548 	vfe->ops->reg_update_clear(vfe, line_id);
549 
550 	output = &line->output;
551 
552 	if (output->wait_reg_update) {
553 		output->wait_reg_update = 0;
554 		complete(&output->reg_update);
555 		spin_unlock_irqrestore(&vfe->output_lock, flags);
556 		return;
557 	}
558 
559 	if (output->state == VFE_OUTPUT_STOPPING) {
560 		/* Release last buffer when hw is idle */
561 		if (output->last_buffer) {
562 			vb2_buffer_done(&output->last_buffer->vb.vb2_buf,
563 					VB2_BUF_STATE_DONE);
564 			output->last_buffer = NULL;
565 		}
566 		output->state = VFE_OUTPUT_IDLE;
567 
568 		/* Buffers received in stopping state are queued in */
569 		/* dma pending queue, start next capture here */
570 
571 		output->buf[0] = vfe_buf_get_pending(output);
572 		output->buf[1] = vfe_buf_get_pending(output);
573 
574 		if (!output->buf[0] && output->buf[1]) {
575 			output->buf[0] = output->buf[1];
576 			output->buf[1] = NULL;
577 		}
578 
579 		if (output->buf[0])
580 			output->state = VFE_OUTPUT_SINGLE;
581 
582 		if (output->buf[1])
583 			output->state = VFE_OUTPUT_CONTINUOUS;
584 
585 		switch (output->state) {
586 		case VFE_OUTPUT_SINGLE:
587 			vfe_output_frame_drop(vfe, output, 2);
588 			break;
589 		case VFE_OUTPUT_CONTINUOUS:
590 			vfe_output_frame_drop(vfe, output, 3);
591 			break;
592 		default:
593 			vfe_output_frame_drop(vfe, output, 0);
594 			break;
595 		}
596 
597 		vfe_output_init_addrs(vfe, output, 1, &vfe->line[line_id]);
598 	}
599 
600 	spin_unlock_irqrestore(&vfe->output_lock, flags);
601 }
602 
603 /*
604  * vfe_isr_wm_done - Process write master done interrupt
605  * @vfe: VFE Device
606  * @wm: Write master id
607  */
608 static void vfe_isr_wm_done(struct vfe_device *vfe, u8 wm)
609 {
610 	struct camss_buffer *ready_buf;
611 	struct vfe_output *output;
612 	dma_addr_t *new_addr;
613 	unsigned long flags;
614 	u32 active_index;
615 	u64 ts = ktime_get_ns();
616 	unsigned int i;
617 
618 	active_index = vfe->ops_gen1->wm_get_ping_pong_status(vfe, wm);
619 
620 	spin_lock_irqsave(&vfe->output_lock, flags);
621 
622 	if (vfe->wm_output_map[wm] == VFE_LINE_NONE) {
623 		dev_err_ratelimited(vfe->camss->dev,
624 				    "Received wm done for unmapped index\n");
625 		goto out_unlock;
626 	}
627 	output = &vfe->line[vfe->wm_output_map[wm]].output;
628 
629 	if (output->gen1.active_buf == active_index && 0) {
630 		dev_err_ratelimited(vfe->camss->dev,
631 				    "Active buffer mismatch!\n");
632 		goto out_unlock;
633 	}
634 	output->gen1.active_buf = active_index;
635 
636 	ready_buf = output->buf[!active_index];
637 	if (!ready_buf) {
638 		dev_err_ratelimited(vfe->camss->dev,
639 				    "Missing ready buf %d %d!\n",
640 				    !active_index, output->state);
641 		goto out_unlock;
642 	}
643 
644 	ready_buf->vb.vb2_buf.timestamp = ts;
645 	ready_buf->vb.sequence = output->sequence++;
646 
647 	/* Get next buffer */
648 	output->buf[!active_index] = vfe_buf_get_pending(output);
649 	if (!output->buf[!active_index]) {
650 		/* No next buffer - set same address */
651 		new_addr = ready_buf->addr;
652 		vfe_buf_update_wm_on_last(vfe, output);
653 	} else {
654 		new_addr = output->buf[!active_index]->addr;
655 		vfe_buf_update_wm_on_next(vfe, output);
656 	}
657 
658 	if (active_index)
659 		for (i = 0; i < output->wm_num; i++)
660 			vfe->ops_gen1->wm_set_ping_addr(vfe, output->wm_idx[i], new_addr[i]);
661 	else
662 		for (i = 0; i < output->wm_num; i++)
663 			vfe->ops_gen1->wm_set_pong_addr(vfe, output->wm_idx[i], new_addr[i]);
664 
665 	spin_unlock_irqrestore(&vfe->output_lock, flags);
666 
667 	if (output->state == VFE_OUTPUT_STOPPING)
668 		output->last_buffer = ready_buf;
669 	else
670 		vb2_buffer_done(&ready_buf->vb.vb2_buf, VB2_BUF_STATE_DONE);
671 
672 	return;
673 
674 out_unlock:
675 	spin_unlock_irqrestore(&vfe->output_lock, flags);
676 }
677 
678 /*
679  * vfe_queue_buffer - Add empty buffer
680  * @vid: Video device structure
681  * @buf: Buffer to be enqueued
682  *
683  * Add an empty buffer - depending on the current number of buffers it will be
684  * put in pending buffer queue or directly given to the hardware to be filled.
685  *
686  * Return 0 on success or a negative error code otherwise
687  */
688 static int vfe_queue_buffer(struct camss_video *vid, struct camss_buffer *buf)
689 {
690 	struct vfe_line *line = container_of(vid, struct vfe_line, video_out);
691 	struct vfe_device *vfe = to_vfe(line);
692 	struct vfe_output *output;
693 	unsigned long flags;
694 
695 	output = &line->output;
696 
697 	spin_lock_irqsave(&vfe->output_lock, flags);
698 
699 	vfe_buf_update_wm_on_new(vfe, output, buf, line);
700 
701 	spin_unlock_irqrestore(&vfe->output_lock, flags);
702 
703 	return 0;
704 }
705 
706 #define CALC_WORD(width, M, N) (((width) * (M) + (N) - 1) / (N))
707 
708 int vfe_word_per_line(u32 format, u32 width)
709 {
710 	int val = 0;
711 
712 	switch (format) {
713 	case V4L2_PIX_FMT_NV12:
714 	case V4L2_PIX_FMT_NV21:
715 	case V4L2_PIX_FMT_NV16:
716 	case V4L2_PIX_FMT_NV61:
717 		val = CALC_WORD(width, 1, 8);
718 		break;
719 	case V4L2_PIX_FMT_YUYV:
720 	case V4L2_PIX_FMT_YVYU:
721 	case V4L2_PIX_FMT_UYVY:
722 	case V4L2_PIX_FMT_VYUY:
723 		val = CALC_WORD(width, 2, 8);
724 		break;
725 	}
726 
727 	return val;
728 }
729 
730 const struct vfe_isr_ops vfe_isr_ops_gen1 = {
731 	.reset_ack = vfe_isr_reset_ack,
732 	.halt_ack = vfe_isr_halt_ack,
733 	.reg_update = vfe_isr_reg_update,
734 	.sof = vfe_isr_sof,
735 	.comp_done = vfe_isr_comp_done,
736 	.wm_done = vfe_isr_wm_done,
737 };
738 
739 const struct camss_video_ops vfe_video_ops_gen1 = {
740 	.queue_buffer = vfe_queue_buffer,
741 	.flush_buffers = vfe_flush_buffers,
742 };
743