xref: /linux/drivers/gpu/drm/tegra/hub.c (revision 1553a1c48281243359a9529a10ddb551f3b967ab)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Copyright (C) 2017 NVIDIA CORPORATION.  All rights reserved.
4  */
5 
6 #include <linux/clk.h>
7 #include <linux/delay.h>
8 #include <linux/dma-mapping.h>
9 #include <linux/host1x.h>
10 #include <linux/module.h>
11 #include <linux/of.h>
12 #include <linux/of_graph.h>
13 #include <linux/of_platform.h>
14 #include <linux/platform_device.h>
15 #include <linux/pm_runtime.h>
16 #include <linux/reset.h>
17 
18 #include <drm/drm_atomic.h>
19 #include <drm/drm_atomic_helper.h>
20 #include <drm/drm_blend.h>
21 #include <drm/drm_fourcc.h>
22 #include <drm/drm_framebuffer.h>
23 #include <drm/drm_probe_helper.h>
24 
25 #include "drm.h"
26 #include "dc.h"
27 #include "plane.h"
28 
29 #define NFB 24
30 
31 static const u32 tegra_shared_plane_formats[] = {
32 	DRM_FORMAT_ARGB1555,
33 	DRM_FORMAT_RGB565,
34 	DRM_FORMAT_RGBA5551,
35 	DRM_FORMAT_ARGB8888,
36 	DRM_FORMAT_ABGR8888,
37 	/* new on Tegra114 */
38 	DRM_FORMAT_ABGR4444,
39 	DRM_FORMAT_ABGR1555,
40 	DRM_FORMAT_BGRA5551,
41 	DRM_FORMAT_XRGB1555,
42 	DRM_FORMAT_RGBX5551,
43 	DRM_FORMAT_XBGR1555,
44 	DRM_FORMAT_BGRX5551,
45 	DRM_FORMAT_BGR565,
46 	DRM_FORMAT_XRGB8888,
47 	DRM_FORMAT_XBGR8888,
48 	/* planar formats */
49 	DRM_FORMAT_UYVY,
50 	DRM_FORMAT_YUYV,
51 	DRM_FORMAT_YUV420,
52 	DRM_FORMAT_YUV422,
53 };
54 
55 static const u64 tegra_shared_plane_modifiers[] = {
56 	DRM_FORMAT_MOD_LINEAR,
57 	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(0),
58 	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(1),
59 	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(2),
60 	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(3),
61 	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(4),
62 	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(5),
63 	/*
64 	 * The GPU sector layout is only supported on Tegra194, but these will
65 	 * be filtered out later on by ->format_mod_supported() on SoCs where
66 	 * it isn't supported.
67 	 */
68 	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(0) | DRM_FORMAT_MOD_NVIDIA_SECTOR_LAYOUT,
69 	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(1) | DRM_FORMAT_MOD_NVIDIA_SECTOR_LAYOUT,
70 	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(2) | DRM_FORMAT_MOD_NVIDIA_SECTOR_LAYOUT,
71 	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(3) | DRM_FORMAT_MOD_NVIDIA_SECTOR_LAYOUT,
72 	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(4) | DRM_FORMAT_MOD_NVIDIA_SECTOR_LAYOUT,
73 	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(5) | DRM_FORMAT_MOD_NVIDIA_SECTOR_LAYOUT,
74 	/* sentinel */
75 	DRM_FORMAT_MOD_INVALID
76 };
77 
78 static inline unsigned int tegra_plane_offset(struct tegra_plane *plane,
79 					      unsigned int offset)
80 {
81 	if (offset >= 0x500 && offset <= 0x581) {
82 		offset = 0x000 + (offset - 0x500);
83 		return plane->offset + offset;
84 	}
85 
86 	if (offset >= 0x700 && offset <= 0x73c) {
87 		offset = 0x180 + (offset - 0x700);
88 		return plane->offset + offset;
89 	}
90 
91 	if (offset >= 0x800 && offset <= 0x83e) {
92 		offset = 0x1c0 + (offset - 0x800);
93 		return plane->offset + offset;
94 	}
95 
96 	dev_WARN(plane->dc->dev, "invalid offset: %x\n", offset);
97 
98 	return plane->offset + offset;
99 }
100 
101 static inline u32 tegra_plane_readl(struct tegra_plane *plane,
102 				    unsigned int offset)
103 {
104 	return tegra_dc_readl(plane->dc, tegra_plane_offset(plane, offset));
105 }
106 
107 static inline void tegra_plane_writel(struct tegra_plane *plane, u32 value,
108 				      unsigned int offset)
109 {
110 	tegra_dc_writel(plane->dc, value, tegra_plane_offset(plane, offset));
111 }
112 
113 static int tegra_windowgroup_enable(struct tegra_windowgroup *wgrp)
114 {
115 	int err = 0;
116 
117 	mutex_lock(&wgrp->lock);
118 
119 	if (wgrp->usecount == 0) {
120 		err = host1x_client_resume(wgrp->parent);
121 		if (err < 0) {
122 			dev_err(wgrp->parent->dev, "failed to resume: %d\n", err);
123 			goto unlock;
124 		}
125 
126 		reset_control_deassert(wgrp->rst);
127 	}
128 
129 	wgrp->usecount++;
130 
131 unlock:
132 	mutex_unlock(&wgrp->lock);
133 	return err;
134 }
135 
136 static void tegra_windowgroup_disable(struct tegra_windowgroup *wgrp)
137 {
138 	int err;
139 
140 	mutex_lock(&wgrp->lock);
141 
142 	if (wgrp->usecount == 1) {
143 		err = reset_control_assert(wgrp->rst);
144 		if (err < 0) {
145 			pr_err("failed to assert reset for window group %u\n",
146 			       wgrp->index);
147 		}
148 
149 		host1x_client_suspend(wgrp->parent);
150 	}
151 
152 	wgrp->usecount--;
153 	mutex_unlock(&wgrp->lock);
154 }
155 
156 int tegra_display_hub_prepare(struct tegra_display_hub *hub)
157 {
158 	unsigned int i;
159 
160 	/*
161 	 * XXX Enabling/disabling windowgroups needs to happen when the owner
162 	 * display controller is disabled. There's currently no good point at
163 	 * which this could be executed, so unconditionally enable all window
164 	 * groups for now.
165 	 */
166 	for (i = 0; i < hub->soc->num_wgrps; i++) {
167 		struct tegra_windowgroup *wgrp = &hub->wgrps[i];
168 
169 		/* Skip orphaned window group whose parent DC is disabled */
170 		if (wgrp->parent)
171 			tegra_windowgroup_enable(wgrp);
172 	}
173 
174 	return 0;
175 }
176 
177 void tegra_display_hub_cleanup(struct tegra_display_hub *hub)
178 {
179 	unsigned int i;
180 
181 	/*
182 	 * XXX Remove this once window groups can be more fine-grainedly
183 	 * enabled and disabled.
184 	 */
185 	for (i = 0; i < hub->soc->num_wgrps; i++) {
186 		struct tegra_windowgroup *wgrp = &hub->wgrps[i];
187 
188 		/* Skip orphaned window group whose parent DC is disabled */
189 		if (wgrp->parent)
190 			tegra_windowgroup_disable(wgrp);
191 	}
192 }
193 
194 static void tegra_shared_plane_update(struct tegra_plane *plane)
195 {
196 	struct tegra_dc *dc = plane->dc;
197 	unsigned long timeout;
198 	u32 mask, value;
199 
200 	mask = COMMON_UPDATE | WIN_A_UPDATE << plane->base.index;
201 	tegra_dc_writel(dc, mask, DC_CMD_STATE_CONTROL);
202 
203 	timeout = jiffies + msecs_to_jiffies(1000);
204 
205 	while (time_before(jiffies, timeout)) {
206 		value = tegra_dc_readl(dc, DC_CMD_STATE_CONTROL);
207 		if ((value & mask) == 0)
208 			break;
209 
210 		usleep_range(100, 400);
211 	}
212 }
213 
214 static void tegra_shared_plane_activate(struct tegra_plane *plane)
215 {
216 	struct tegra_dc *dc = plane->dc;
217 	unsigned long timeout;
218 	u32 mask, value;
219 
220 	mask = COMMON_ACTREQ | WIN_A_ACT_REQ << plane->base.index;
221 	tegra_dc_writel(dc, mask, DC_CMD_STATE_CONTROL);
222 
223 	timeout = jiffies + msecs_to_jiffies(1000);
224 
225 	while (time_before(jiffies, timeout)) {
226 		value = tegra_dc_readl(dc, DC_CMD_STATE_CONTROL);
227 		if ((value & mask) == 0)
228 			break;
229 
230 		usleep_range(100, 400);
231 	}
232 }
233 
234 static unsigned int
235 tegra_shared_plane_get_owner(struct tegra_plane *plane, struct tegra_dc *dc)
236 {
237 	unsigned int offset =
238 		tegra_plane_offset(plane, DC_WIN_CORE_WINDOWGROUP_SET_CONTROL);
239 
240 	return tegra_dc_readl(dc, offset) & OWNER_MASK;
241 }
242 
243 static bool tegra_dc_owns_shared_plane(struct tegra_dc *dc,
244 				       struct tegra_plane *plane)
245 {
246 	struct device *dev = dc->dev;
247 
248 	if (tegra_shared_plane_get_owner(plane, dc) == dc->pipe) {
249 		if (plane->dc == dc)
250 			return true;
251 
252 		dev_WARN(dev, "head %u owns window %u but is not attached\n",
253 			 dc->pipe, plane->index);
254 	}
255 
256 	return false;
257 }
258 
259 static int tegra_shared_plane_set_owner(struct tegra_plane *plane,
260 					struct tegra_dc *new)
261 {
262 	unsigned int offset =
263 		tegra_plane_offset(plane, DC_WIN_CORE_WINDOWGROUP_SET_CONTROL);
264 	struct tegra_dc *old = plane->dc, *dc = new ? new : old;
265 	struct device *dev = new ? new->dev : old->dev;
266 	unsigned int owner, index = plane->index;
267 	u32 value;
268 
269 	value = tegra_dc_readl(dc, offset);
270 	owner = value & OWNER_MASK;
271 
272 	if (new && (owner != OWNER_MASK && owner != new->pipe)) {
273 		dev_WARN(dev, "window %u owned by head %u\n", index, owner);
274 		return -EBUSY;
275 	}
276 
277 	/*
278 	 * This seems to happen whenever the head has been disabled with one
279 	 * or more windows being active. This is harmless because we'll just
280 	 * reassign the window to the new head anyway.
281 	 */
282 	if (old && owner == OWNER_MASK)
283 		dev_dbg(dev, "window %u not owned by head %u but %u\n", index,
284 			old->pipe, owner);
285 
286 	value &= ~OWNER_MASK;
287 
288 	if (new)
289 		value |= OWNER(new->pipe);
290 	else
291 		value |= OWNER_MASK;
292 
293 	tegra_dc_writel(dc, value, offset);
294 
295 	plane->dc = new;
296 
297 	return 0;
298 }
299 
300 static void tegra_shared_plane_setup_scaler(struct tegra_plane *plane)
301 {
302 	static const unsigned int coeffs[192] = {
303 		0x00000000, 0x3c70e400, 0x3bb037e4, 0x0c51cc9c,
304 		0x00100001, 0x3bf0dbfa, 0x3d00f406, 0x3fe003ff,
305 		0x00300002, 0x3b80cbf5, 0x3da1040d, 0x3fb003fe,
306 		0x00400002, 0x3b20bff1, 0x3e511015, 0x3f9003fc,
307 		0x00500002, 0x3ad0b3ed, 0x3f21201d, 0x3f5003fb,
308 		0x00500003, 0x3aa0a3e9, 0x3ff13026, 0x3f2007f9,
309 		0x00500403, 0x3a7097e6, 0x00e1402f, 0x3ee007f7,
310 		0x00500403, 0x3a608be4, 0x01d14c38, 0x3ea00bf6,
311 		0x00500403, 0x3a507fe2, 0x02e15c42, 0x3e500ff4,
312 		0x00500402, 0x3a6073e1, 0x03f16c4d, 0x3e000ff2,
313 		0x00400402, 0x3a706be0, 0x05117858, 0x3db013f0,
314 		0x00300402, 0x3a905fe0, 0x06318863, 0x3d6017ee,
315 		0x00300402, 0x3ab057e0, 0x0771986e, 0x3d001beb,
316 		0x00200001, 0x3af04fe1, 0x08a1a47a, 0x3cb023e9,
317 		0x00100001, 0x3b2047e2, 0x09e1b485, 0x3c6027e7,
318 		0x00100000, 0x3b703fe2, 0x0b11c091, 0x3c002fe6,
319 		0x3f203800, 0x0391103f, 0x3ff0a014, 0x0811606c,
320 		0x3f2037ff, 0x0351083c, 0x03e11842, 0x3f203c00,
321 		0x3f302fff, 0x03010439, 0x04311c45, 0x3f104401,
322 		0x3f302fff, 0x02c0fc35, 0x04812448, 0x3f104802,
323 		0x3f4027ff, 0x0270f832, 0x04c1284b, 0x3f205003,
324 		0x3f4023ff, 0x0230f030, 0x0511304e, 0x3f205403,
325 		0x3f601fff, 0x01f0e82d, 0x05613451, 0x3f205c04,
326 		0x3f701bfe, 0x01b0e02a, 0x05a13c54, 0x3f306006,
327 		0x3f7017fe, 0x0170d827, 0x05f14057, 0x3f406807,
328 		0x3f8017ff, 0x0140d424, 0x0641445a, 0x3f406c08,
329 		0x3fa013ff, 0x0100cc22, 0x0681485d, 0x3f507409,
330 		0x3fa00fff, 0x00d0c41f, 0x06d14c60, 0x3f607c0b,
331 		0x3fc00fff, 0x0090bc1c, 0x07115063, 0x3f80840c,
332 		0x3fd00bff, 0x0070b41a, 0x07515465, 0x3f908c0e,
333 		0x3fe007ff, 0x0040b018, 0x07915868, 0x3fb0900f,
334 		0x3ff00400, 0x0010a816, 0x07d15c6a, 0x3fd09811,
335 		0x00a04c0e, 0x0460f442, 0x0240a827, 0x05c15859,
336 		0x0090440d, 0x0440f040, 0x0480fc43, 0x00b05010,
337 		0x0080400c, 0x0410ec3e, 0x04910044, 0x00d05411,
338 		0x0070380b, 0x03f0e83d, 0x04b10846, 0x00e05812,
339 		0x0060340a, 0x03d0e43b, 0x04d10c48, 0x00f06013,
340 		0x00503009, 0x03b0e039, 0x04e11449, 0x01106415,
341 		0x00402c08, 0x0390d838, 0x05011c4b, 0x01206c16,
342 		0x00302807, 0x0370d436, 0x0511204c, 0x01407018,
343 		0x00302406, 0x0340d034, 0x0531244e, 0x01507419,
344 		0x00202005, 0x0320cc32, 0x05412c50, 0x01707c1b,
345 		0x00101c04, 0x0300c431, 0x05613451, 0x0180801d,
346 		0x00101803, 0x02e0c02f, 0x05713853, 0x01a0881e,
347 		0x00101002, 0x02b0bc2d, 0x05814054, 0x01c08c20,
348 		0x00000c02, 0x02a0b82c, 0x05914455, 0x01e09421,
349 		0x00000801, 0x0280b02a, 0x05a14c57, 0x02009c23,
350 		0x00000400, 0x0260ac28, 0x05b15458, 0x0220a025,
351 	};
352 	unsigned int ratio, row, column;
353 
354 	for (ratio = 0; ratio <= 2; ratio++) {
355 		for (row = 0; row <= 15; row++) {
356 			for (column = 0; column <= 3; column++) {
357 				unsigned int index = (ratio << 6) + (row << 2) + column;
358 				u32 value;
359 
360 				value = COEFF_INDEX(index) | COEFF_DATA(coeffs[index]);
361 				tegra_plane_writel(plane, value,
362 						   DC_WIN_WINDOWGROUP_SET_INPUT_SCALER_COEFF);
363 			}
364 		}
365 	}
366 }
367 
368 static void tegra_dc_assign_shared_plane(struct tegra_dc *dc,
369 					 struct tegra_plane *plane)
370 {
371 	u32 value;
372 	int err;
373 
374 	if (!tegra_dc_owns_shared_plane(dc, plane)) {
375 		err = tegra_shared_plane_set_owner(plane, dc);
376 		if (err < 0)
377 			return;
378 	}
379 
380 	value = tegra_plane_readl(plane, DC_WIN_CORE_IHUB_LINEBUF_CONFIG);
381 	value |= MODE_FOUR_LINES;
382 	tegra_plane_writel(plane, value, DC_WIN_CORE_IHUB_LINEBUF_CONFIG);
383 
384 	value = tegra_plane_readl(plane, DC_WIN_CORE_IHUB_WGRP_FETCH_METER);
385 	value = SLOTS(1);
386 	tegra_plane_writel(plane, value, DC_WIN_CORE_IHUB_WGRP_FETCH_METER);
387 
388 	/* disable watermark */
389 	value = tegra_plane_readl(plane, DC_WIN_CORE_IHUB_WGRP_LATENCY_CTLA);
390 	value &= ~LATENCY_CTL_MODE_ENABLE;
391 	tegra_plane_writel(plane, value, DC_WIN_CORE_IHUB_WGRP_LATENCY_CTLA);
392 
393 	value = tegra_plane_readl(plane, DC_WIN_CORE_IHUB_WGRP_LATENCY_CTLB);
394 	value |= WATERMARK_MASK;
395 	tegra_plane_writel(plane, value, DC_WIN_CORE_IHUB_WGRP_LATENCY_CTLB);
396 
397 	/* pipe meter */
398 	value = tegra_plane_readl(plane, DC_WIN_CORE_PRECOMP_WGRP_PIPE_METER);
399 	value = PIPE_METER_INT(0) | PIPE_METER_FRAC(0);
400 	tegra_plane_writel(plane, value, DC_WIN_CORE_PRECOMP_WGRP_PIPE_METER);
401 
402 	/* mempool entries */
403 	value = tegra_plane_readl(plane, DC_WIN_CORE_IHUB_WGRP_POOL_CONFIG);
404 	value = MEMPOOL_ENTRIES(0x331);
405 	tegra_plane_writel(plane, value, DC_WIN_CORE_IHUB_WGRP_POOL_CONFIG);
406 
407 	value = tegra_plane_readl(plane, DC_WIN_CORE_IHUB_THREAD_GROUP);
408 	value &= ~THREAD_NUM_MASK;
409 	value |= THREAD_NUM(plane->base.index);
410 	value |= THREAD_GROUP_ENABLE;
411 	tegra_plane_writel(plane, value, DC_WIN_CORE_IHUB_THREAD_GROUP);
412 
413 	tegra_shared_plane_setup_scaler(plane);
414 
415 	tegra_shared_plane_update(plane);
416 	tegra_shared_plane_activate(plane);
417 }
418 
419 static void tegra_dc_remove_shared_plane(struct tegra_dc *dc,
420 					 struct tegra_plane *plane)
421 {
422 	tegra_shared_plane_set_owner(plane, NULL);
423 }
424 
425 static int tegra_shared_plane_atomic_check(struct drm_plane *plane,
426 					   struct drm_atomic_state *state)
427 {
428 	struct drm_plane_state *new_plane_state = drm_atomic_get_new_plane_state(state,
429 										 plane);
430 	struct tegra_plane_state *plane_state = to_tegra_plane_state(new_plane_state);
431 	struct tegra_shared_plane *tegra = to_tegra_shared_plane(plane);
432 	struct tegra_bo_tiling *tiling = &plane_state->tiling;
433 	struct tegra_dc *dc = to_tegra_dc(new_plane_state->crtc);
434 	int err;
435 
436 	/* no need for further checks if the plane is being disabled */
437 	if (!new_plane_state->crtc || !new_plane_state->fb)
438 		return 0;
439 
440 	err = tegra_plane_format(new_plane_state->fb->format->format,
441 				 &plane_state->format,
442 				 &plane_state->swap);
443 	if (err < 0)
444 		return err;
445 
446 	err = tegra_fb_get_tiling(new_plane_state->fb, tiling);
447 	if (err < 0)
448 		return err;
449 
450 	if (tiling->mode == TEGRA_BO_TILING_MODE_BLOCK &&
451 	    !dc->soc->supports_block_linear) {
452 		DRM_ERROR("hardware doesn't support block linear mode\n");
453 		return -EINVAL;
454 	}
455 
456 	if (tiling->sector_layout == TEGRA_BO_SECTOR_LAYOUT_GPU &&
457 	    !dc->soc->supports_sector_layout) {
458 		DRM_ERROR("hardware doesn't support GPU sector layout\n");
459 		return -EINVAL;
460 	}
461 
462 	/*
463 	 * Tegra doesn't support different strides for U and V planes so we
464 	 * error out if the user tries to display a framebuffer with such a
465 	 * configuration.
466 	 */
467 	if (new_plane_state->fb->format->num_planes > 2) {
468 		if (new_plane_state->fb->pitches[2] != new_plane_state->fb->pitches[1]) {
469 			DRM_ERROR("unsupported UV-plane configuration\n");
470 			return -EINVAL;
471 		}
472 	}
473 
474 	/* XXX scaling is not yet supported, add a check here */
475 
476 	err = tegra_plane_state_add(&tegra->base, new_plane_state);
477 	if (err < 0)
478 		return err;
479 
480 	return 0;
481 }
482 
483 static void tegra_shared_plane_atomic_disable(struct drm_plane *plane,
484 					      struct drm_atomic_state *state)
485 {
486 	struct drm_plane_state *old_state = drm_atomic_get_old_plane_state(state,
487 									   plane);
488 	struct tegra_plane *p = to_tegra_plane(plane);
489 	struct tegra_dc *dc;
490 	u32 value;
491 	int err;
492 
493 	/* rien ne va plus */
494 	if (!old_state || !old_state->crtc)
495 		return;
496 
497 	dc = to_tegra_dc(old_state->crtc);
498 
499 	err = host1x_client_resume(&dc->client);
500 	if (err < 0) {
501 		dev_err(dc->dev, "failed to resume: %d\n", err);
502 		return;
503 	}
504 
505 	/*
506 	 * XXX Legacy helpers seem to sometimes call ->atomic_disable() even
507 	 * on planes that are already disabled. Make sure we fallback to the
508 	 * head for this particular state instead of crashing.
509 	 */
510 	if (WARN_ON(p->dc == NULL))
511 		p->dc = dc;
512 
513 	value = tegra_plane_readl(p, DC_WIN_WIN_OPTIONS);
514 	value &= ~WIN_ENABLE;
515 	tegra_plane_writel(p, value, DC_WIN_WIN_OPTIONS);
516 
517 	tegra_dc_remove_shared_plane(dc, p);
518 
519 	host1x_client_suspend(&dc->client);
520 }
521 
522 static inline u32 compute_phase_incr(fixed20_12 in, unsigned int out)
523 {
524 	u64 tmp, tmp1, tmp2;
525 
526 	tmp = (u64)dfixed_trunc(in);
527 	tmp2 = (u64)out;
528 	tmp1 = (tmp << NFB) + (tmp2 >> 1);
529 	do_div(tmp1, tmp2);
530 
531 	return lower_32_bits(tmp1);
532 }
533 
534 static void tegra_shared_plane_atomic_update(struct drm_plane *plane,
535 					     struct drm_atomic_state *state)
536 {
537 	struct drm_plane_state *new_state = drm_atomic_get_new_plane_state(state,
538 									   plane);
539 	struct tegra_plane_state *tegra_plane_state = to_tegra_plane_state(new_state);
540 	struct tegra_dc *dc = to_tegra_dc(new_state->crtc);
541 	unsigned int zpos = new_state->normalized_zpos;
542 	struct drm_framebuffer *fb = new_state->fb;
543 	struct tegra_plane *p = to_tegra_plane(plane);
544 	u32 value, min_width, bypass = 0;
545 	dma_addr_t base, addr_flag = 0;
546 	unsigned int bpc, planes;
547 	bool yuv;
548 	int err;
549 
550 	/* rien ne va plus */
551 	if (!new_state->crtc || !new_state->fb)
552 		return;
553 
554 	if (!new_state->visible) {
555 		tegra_shared_plane_atomic_disable(plane, state);
556 		return;
557 	}
558 
559 	err = host1x_client_resume(&dc->client);
560 	if (err < 0) {
561 		dev_err(dc->dev, "failed to resume: %d\n", err);
562 		return;
563 	}
564 
565 	yuv = tegra_plane_format_is_yuv(tegra_plane_state->format, &planes, &bpc);
566 
567 	tegra_dc_assign_shared_plane(dc, p);
568 
569 	tegra_plane_writel(p, VCOUNTER, DC_WIN_CORE_ACT_CONTROL);
570 
571 	/* blending */
572 	value = BLEND_FACTOR_DST_ALPHA_ZERO | BLEND_FACTOR_SRC_ALPHA_K2 |
573 		BLEND_FACTOR_DST_COLOR_NEG_K1_TIMES_SRC |
574 		BLEND_FACTOR_SRC_COLOR_K1_TIMES_SRC;
575 	tegra_plane_writel(p, value, DC_WIN_BLEND_MATCH_SELECT);
576 
577 	value = BLEND_FACTOR_DST_ALPHA_ZERO | BLEND_FACTOR_SRC_ALPHA_K2 |
578 		BLEND_FACTOR_DST_COLOR_NEG_K1_TIMES_SRC |
579 		BLEND_FACTOR_SRC_COLOR_K1_TIMES_SRC;
580 	tegra_plane_writel(p, value, DC_WIN_BLEND_NOMATCH_SELECT);
581 
582 	value = K2(255) | K1(255) | WINDOW_LAYER_DEPTH(255 - zpos);
583 	tegra_plane_writel(p, value, DC_WIN_BLEND_LAYER_CONTROL);
584 
585 	/* scaling */
586 	min_width = min(new_state->src_w >> 16, new_state->crtc_w);
587 
588 	value = tegra_plane_readl(p, DC_WINC_PRECOMP_WGRP_PIPE_CAPC);
589 
590 	if (min_width < MAX_PIXELS_5TAP444(value)) {
591 		value = HORIZONTAL_TAPS_5 | VERTICAL_TAPS_5;
592 	} else {
593 		value = tegra_plane_readl(p, DC_WINC_PRECOMP_WGRP_PIPE_CAPE);
594 
595 		if (min_width < MAX_PIXELS_2TAP444(value))
596 			value = HORIZONTAL_TAPS_2 | VERTICAL_TAPS_2;
597 		else
598 			dev_err(dc->dev, "invalid minimum width: %u\n", min_width);
599 	}
600 
601 	value = HORIZONTAL_TAPS_5 | VERTICAL_TAPS_5;
602 	tegra_plane_writel(p, value, DC_WIN_WINDOWGROUP_SET_CONTROL_INPUT_SCALER);
603 
604 	if (new_state->src_w != new_state->crtc_w << 16) {
605 		fixed20_12 width = dfixed_init(new_state->src_w >> 16);
606 		u32 incr = compute_phase_incr(width, new_state->crtc_w) & ~0x1;
607 		u32 init = (1 << (NFB - 1)) + (incr >> 1);
608 
609 		tegra_plane_writel(p, incr, DC_WIN_SET_INPUT_SCALER_HPHASE_INCR);
610 		tegra_plane_writel(p, init, DC_WIN_SET_INPUT_SCALER_H_START_PHASE);
611 	} else {
612 		bypass |= INPUT_SCALER_HBYPASS;
613 	}
614 
615 	if (new_state->src_h != new_state->crtc_h << 16) {
616 		fixed20_12 height = dfixed_init(new_state->src_h >> 16);
617 		u32 incr = compute_phase_incr(height, new_state->crtc_h) & ~0x1;
618 		u32 init = (1 << (NFB - 1)) + (incr >> 1);
619 
620 		tegra_plane_writel(p, incr, DC_WIN_SET_INPUT_SCALER_VPHASE_INCR);
621 		tegra_plane_writel(p, init, DC_WIN_SET_INPUT_SCALER_V_START_PHASE);
622 	} else {
623 		bypass |= INPUT_SCALER_VBYPASS;
624 	}
625 
626 	tegra_plane_writel(p, bypass, DC_WIN_WINDOWGROUP_SET_INPUT_SCALER_USAGE);
627 
628 	/* disable compression */
629 	tegra_plane_writel(p, 0, DC_WINBUF_CDE_CONTROL);
630 
631 #ifdef CONFIG_ARCH_DMA_ADDR_T_64BIT
632 	/*
633 	 * Physical address bit 39 in Tegra194 is used as a switch for special
634 	 * logic that swizzles the memory using either the legacy Tegra or the
635 	 * dGPU sector layout.
636 	 */
637 	if (tegra_plane_state->tiling.sector_layout == TEGRA_BO_SECTOR_LAYOUT_GPU)
638 		addr_flag = BIT_ULL(39);
639 #endif
640 
641 	base = tegra_plane_state->iova[0] + fb->offsets[0];
642 	base |= addr_flag;
643 
644 	tegra_plane_writel(p, tegra_plane_state->format, DC_WIN_COLOR_DEPTH);
645 	tegra_plane_writel(p, 0, DC_WIN_PRECOMP_WGRP_PARAMS);
646 
647 	value = V_POSITION(new_state->crtc_y) |
648 		H_POSITION(new_state->crtc_x);
649 	tegra_plane_writel(p, value, DC_WIN_POSITION);
650 
651 	value = V_SIZE(new_state->crtc_h) | H_SIZE(new_state->crtc_w);
652 	tegra_plane_writel(p, value, DC_WIN_SIZE);
653 
654 	value = WIN_ENABLE | COLOR_EXPAND;
655 	tegra_plane_writel(p, value, DC_WIN_WIN_OPTIONS);
656 
657 	value = V_SIZE(new_state->src_h >> 16) | H_SIZE(new_state->src_w >> 16);
658 	tegra_plane_writel(p, value, DC_WIN_CROPPED_SIZE);
659 
660 	tegra_plane_writel(p, upper_32_bits(base), DC_WINBUF_START_ADDR_HI);
661 	tegra_plane_writel(p, lower_32_bits(base), DC_WINBUF_START_ADDR);
662 
663 	value = PITCH(fb->pitches[0]);
664 	tegra_plane_writel(p, value, DC_WIN_PLANAR_STORAGE);
665 
666 	if (yuv && planes > 1) {
667 		base = tegra_plane_state->iova[1] + fb->offsets[1];
668 		base |= addr_flag;
669 
670 		tegra_plane_writel(p, upper_32_bits(base), DC_WINBUF_START_ADDR_HI_U);
671 		tegra_plane_writel(p, lower_32_bits(base), DC_WINBUF_START_ADDR_U);
672 
673 		if (planes > 2) {
674 			base = tegra_plane_state->iova[2] + fb->offsets[2];
675 			base |= addr_flag;
676 
677 			tegra_plane_writel(p, upper_32_bits(base), DC_WINBUF_START_ADDR_HI_V);
678 			tegra_plane_writel(p, lower_32_bits(base), DC_WINBUF_START_ADDR_V);
679 		}
680 
681 		value = PITCH_U(fb->pitches[1]);
682 
683 		if (planes > 2)
684 			value |= PITCH_V(fb->pitches[2]);
685 
686 		tegra_plane_writel(p, value, DC_WIN_PLANAR_STORAGE_UV);
687 	} else {
688 		tegra_plane_writel(p, 0, DC_WINBUF_START_ADDR_U);
689 		tegra_plane_writel(p, 0, DC_WINBUF_START_ADDR_HI_U);
690 		tegra_plane_writel(p, 0, DC_WINBUF_START_ADDR_V);
691 		tegra_plane_writel(p, 0, DC_WINBUF_START_ADDR_HI_V);
692 		tegra_plane_writel(p, 0, DC_WIN_PLANAR_STORAGE_UV);
693 	}
694 
695 	value = CLAMP_BEFORE_BLEND | INPUT_RANGE_FULL;
696 
697 	if (yuv) {
698 		if (bpc < 12)
699 			value |= DEGAMMA_YUV8_10;
700 		else
701 			value |= DEGAMMA_YUV12;
702 
703 		/* XXX parameterize */
704 		value |= COLOR_SPACE_YUV_2020;
705 	} else {
706 		if (!tegra_plane_format_is_indexed(tegra_plane_state->format))
707 			value |= DEGAMMA_SRGB;
708 	}
709 
710 	tegra_plane_writel(p, value, DC_WIN_SET_PARAMS);
711 
712 	value = OFFSET_X(new_state->src_y >> 16) |
713 		OFFSET_Y(new_state->src_x >> 16);
714 	tegra_plane_writel(p, value, DC_WINBUF_CROPPED_POINT);
715 
716 	if (dc->soc->supports_block_linear) {
717 		unsigned long height = tegra_plane_state->tiling.value;
718 
719 		/* XXX */
720 		switch (tegra_plane_state->tiling.mode) {
721 		case TEGRA_BO_TILING_MODE_PITCH:
722 			value = DC_WINBUF_SURFACE_KIND_BLOCK_HEIGHT(0) |
723 				DC_WINBUF_SURFACE_KIND_PITCH;
724 			break;
725 
726 		/* XXX not supported on Tegra186 and later */
727 		case TEGRA_BO_TILING_MODE_TILED:
728 			value = DC_WINBUF_SURFACE_KIND_TILED;
729 			break;
730 
731 		case TEGRA_BO_TILING_MODE_BLOCK:
732 			value = DC_WINBUF_SURFACE_KIND_BLOCK_HEIGHT(height) |
733 				DC_WINBUF_SURFACE_KIND_BLOCK;
734 			break;
735 		}
736 
737 		tegra_plane_writel(p, value, DC_WINBUF_SURFACE_KIND);
738 	}
739 
740 	/* disable gamut CSC */
741 	value = tegra_plane_readl(p, DC_WIN_WINDOW_SET_CONTROL);
742 	value &= ~CONTROL_CSC_ENABLE;
743 	tegra_plane_writel(p, value, DC_WIN_WINDOW_SET_CONTROL);
744 
745 	host1x_client_suspend(&dc->client);
746 }
747 
748 static const struct drm_plane_helper_funcs tegra_shared_plane_helper_funcs = {
749 	.prepare_fb = tegra_plane_prepare_fb,
750 	.cleanup_fb = tegra_plane_cleanup_fb,
751 	.atomic_check = tegra_shared_plane_atomic_check,
752 	.atomic_update = tegra_shared_plane_atomic_update,
753 	.atomic_disable = tegra_shared_plane_atomic_disable,
754 };
755 
756 struct drm_plane *tegra_shared_plane_create(struct drm_device *drm,
757 					    struct tegra_dc *dc,
758 					    unsigned int wgrp,
759 					    unsigned int index)
760 {
761 	enum drm_plane_type type = DRM_PLANE_TYPE_OVERLAY;
762 	struct tegra_drm *tegra = drm->dev_private;
763 	struct tegra_display_hub *hub = tegra->hub;
764 	struct tegra_shared_plane *plane;
765 	unsigned int possible_crtcs;
766 	unsigned int num_formats;
767 	const u64 *modifiers;
768 	struct drm_plane *p;
769 	const u32 *formats;
770 	int err;
771 
772 	plane = kzalloc(sizeof(*plane), GFP_KERNEL);
773 	if (!plane)
774 		return ERR_PTR(-ENOMEM);
775 
776 	plane->base.offset = 0x0a00 + 0x0300 * index;
777 	plane->base.index = index;
778 
779 	plane->wgrp = &hub->wgrps[wgrp];
780 	plane->wgrp->parent = &dc->client;
781 
782 	p = &plane->base.base;
783 
784 	/* planes can be assigned to arbitrary CRTCs */
785 	possible_crtcs = BIT(tegra->num_crtcs) - 1;
786 
787 	num_formats = ARRAY_SIZE(tegra_shared_plane_formats);
788 	formats = tegra_shared_plane_formats;
789 	modifiers = tegra_shared_plane_modifiers;
790 
791 	err = drm_universal_plane_init(drm, p, possible_crtcs,
792 				       &tegra_plane_funcs, formats,
793 				       num_formats, modifiers, type, NULL);
794 	if (err < 0) {
795 		kfree(plane);
796 		return ERR_PTR(err);
797 	}
798 
799 	drm_plane_helper_add(p, &tegra_shared_plane_helper_funcs);
800 	drm_plane_create_zpos_property(p, 0, 0, 255);
801 
802 	return p;
803 }
804 
805 static struct drm_private_state *
806 tegra_display_hub_duplicate_state(struct drm_private_obj *obj)
807 {
808 	struct tegra_display_hub_state *state;
809 
810 	state = kmemdup(obj->state, sizeof(*state), GFP_KERNEL);
811 	if (!state)
812 		return NULL;
813 
814 	__drm_atomic_helper_private_obj_duplicate_state(obj, &state->base);
815 
816 	return &state->base;
817 }
818 
819 static void tegra_display_hub_destroy_state(struct drm_private_obj *obj,
820 					    struct drm_private_state *state)
821 {
822 	struct tegra_display_hub_state *hub_state =
823 		to_tegra_display_hub_state(state);
824 
825 	kfree(hub_state);
826 }
827 
828 static const struct drm_private_state_funcs tegra_display_hub_state_funcs = {
829 	.atomic_duplicate_state = tegra_display_hub_duplicate_state,
830 	.atomic_destroy_state = tegra_display_hub_destroy_state,
831 };
832 
833 static struct tegra_display_hub_state *
834 tegra_display_hub_get_state(struct tegra_display_hub *hub,
835 			    struct drm_atomic_state *state)
836 {
837 	struct drm_private_state *priv;
838 
839 	priv = drm_atomic_get_private_obj_state(state, &hub->base);
840 	if (IS_ERR(priv))
841 		return ERR_CAST(priv);
842 
843 	return to_tegra_display_hub_state(priv);
844 }
845 
846 int tegra_display_hub_atomic_check(struct drm_device *drm,
847 				   struct drm_atomic_state *state)
848 {
849 	struct tegra_drm *tegra = drm->dev_private;
850 	struct tegra_display_hub_state *hub_state;
851 	struct drm_crtc_state *old, *new;
852 	struct drm_crtc *crtc;
853 	unsigned int i;
854 
855 	if (!tegra->hub)
856 		return 0;
857 
858 	hub_state = tegra_display_hub_get_state(tegra->hub, state);
859 	if (IS_ERR(hub_state))
860 		return PTR_ERR(hub_state);
861 
862 	/*
863 	 * The display hub display clock needs to be fed by the display clock
864 	 * with the highest frequency to ensure proper functioning of all the
865 	 * displays.
866 	 *
867 	 * Note that this isn't used before Tegra186, but it doesn't hurt and
868 	 * conditionalizing it would make the code less clean.
869 	 */
870 	for_each_oldnew_crtc_in_state(state, crtc, old, new, i) {
871 		struct tegra_dc_state *dc = to_dc_state(new);
872 
873 		if (new->active) {
874 			if (!hub_state->clk || dc->pclk > hub_state->rate) {
875 				hub_state->dc = to_tegra_dc(dc->base.crtc);
876 				hub_state->clk = hub_state->dc->clk;
877 				hub_state->rate = dc->pclk;
878 			}
879 		}
880 	}
881 
882 	return 0;
883 }
884 
885 static void tegra_display_hub_update(struct tegra_dc *dc)
886 {
887 	u32 value;
888 	int err;
889 
890 	err = host1x_client_resume(&dc->client);
891 	if (err < 0) {
892 		dev_err(dc->dev, "failed to resume: %d\n", err);
893 		return;
894 	}
895 
896 	value = tegra_dc_readl(dc, DC_CMD_IHUB_COMMON_MISC_CTL);
897 	value &= ~LATENCY_EVENT;
898 	tegra_dc_writel(dc, value, DC_CMD_IHUB_COMMON_MISC_CTL);
899 
900 	value = tegra_dc_readl(dc, DC_DISP_IHUB_COMMON_DISPLAY_FETCH_METER);
901 	value = CURS_SLOTS(1) | WGRP_SLOTS(1);
902 	tegra_dc_writel(dc, value, DC_DISP_IHUB_COMMON_DISPLAY_FETCH_METER);
903 
904 	tegra_dc_writel(dc, COMMON_UPDATE, DC_CMD_STATE_CONTROL);
905 	tegra_dc_readl(dc, DC_CMD_STATE_CONTROL);
906 	tegra_dc_writel(dc, COMMON_ACTREQ, DC_CMD_STATE_CONTROL);
907 	tegra_dc_readl(dc, DC_CMD_STATE_CONTROL);
908 
909 	host1x_client_suspend(&dc->client);
910 }
911 
912 void tegra_display_hub_atomic_commit(struct drm_device *drm,
913 				     struct drm_atomic_state *state)
914 {
915 	struct tegra_drm *tegra = drm->dev_private;
916 	struct tegra_display_hub *hub = tegra->hub;
917 	struct tegra_display_hub_state *hub_state;
918 	struct device *dev = hub->client.dev;
919 	int err;
920 
921 	hub_state = to_tegra_display_hub_state(hub->base.state);
922 
923 	if (hub_state->clk) {
924 		err = clk_set_rate(hub_state->clk, hub_state->rate);
925 		if (err < 0)
926 			dev_err(dev, "failed to set rate of %pC to %lu Hz\n",
927 				hub_state->clk, hub_state->rate);
928 
929 		err = clk_set_parent(hub->clk_disp, hub_state->clk);
930 		if (err < 0)
931 			dev_err(dev, "failed to set parent of %pC to %pC: %d\n",
932 				hub->clk_disp, hub_state->clk, err);
933 	}
934 
935 	if (hub_state->dc)
936 		tegra_display_hub_update(hub_state->dc);
937 }
938 
939 static int tegra_display_hub_init(struct host1x_client *client)
940 {
941 	struct tegra_display_hub *hub = to_tegra_display_hub(client);
942 	struct drm_device *drm = dev_get_drvdata(client->host);
943 	struct tegra_drm *tegra = drm->dev_private;
944 	struct tegra_display_hub_state *state;
945 
946 	state = kzalloc(sizeof(*state), GFP_KERNEL);
947 	if (!state)
948 		return -ENOMEM;
949 
950 	drm_atomic_private_obj_init(drm, &hub->base, &state->base,
951 				    &tegra_display_hub_state_funcs);
952 
953 	tegra->hub = hub;
954 
955 	return 0;
956 }
957 
958 static int tegra_display_hub_exit(struct host1x_client *client)
959 {
960 	struct drm_device *drm = dev_get_drvdata(client->host);
961 	struct tegra_drm *tegra = drm->dev_private;
962 
963 	drm_atomic_private_obj_fini(&tegra->hub->base);
964 	tegra->hub = NULL;
965 
966 	return 0;
967 }
968 
969 static int tegra_display_hub_runtime_suspend(struct host1x_client *client)
970 {
971 	struct tegra_display_hub *hub = to_tegra_display_hub(client);
972 	struct device *dev = client->dev;
973 	unsigned int i = hub->num_heads;
974 	int err;
975 
976 	err = reset_control_assert(hub->rst);
977 	if (err < 0)
978 		return err;
979 
980 	while (i--)
981 		clk_disable_unprepare(hub->clk_heads[i]);
982 
983 	clk_disable_unprepare(hub->clk_hub);
984 	clk_disable_unprepare(hub->clk_dsc);
985 	clk_disable_unprepare(hub->clk_disp);
986 
987 	pm_runtime_put_sync(dev);
988 
989 	return 0;
990 }
991 
992 static int tegra_display_hub_runtime_resume(struct host1x_client *client)
993 {
994 	struct tegra_display_hub *hub = to_tegra_display_hub(client);
995 	struct device *dev = client->dev;
996 	unsigned int i;
997 	int err;
998 
999 	err = pm_runtime_resume_and_get(dev);
1000 	if (err < 0) {
1001 		dev_err(dev, "failed to get runtime PM: %d\n", err);
1002 		return err;
1003 	}
1004 
1005 	err = clk_prepare_enable(hub->clk_disp);
1006 	if (err < 0)
1007 		goto put_rpm;
1008 
1009 	err = clk_prepare_enable(hub->clk_dsc);
1010 	if (err < 0)
1011 		goto disable_disp;
1012 
1013 	err = clk_prepare_enable(hub->clk_hub);
1014 	if (err < 0)
1015 		goto disable_dsc;
1016 
1017 	for (i = 0; i < hub->num_heads; i++) {
1018 		err = clk_prepare_enable(hub->clk_heads[i]);
1019 		if (err < 0)
1020 			goto disable_heads;
1021 	}
1022 
1023 	err = reset_control_deassert(hub->rst);
1024 	if (err < 0)
1025 		goto disable_heads;
1026 
1027 	return 0;
1028 
1029 disable_heads:
1030 	while (i--)
1031 		clk_disable_unprepare(hub->clk_heads[i]);
1032 
1033 	clk_disable_unprepare(hub->clk_hub);
1034 disable_dsc:
1035 	clk_disable_unprepare(hub->clk_dsc);
1036 disable_disp:
1037 	clk_disable_unprepare(hub->clk_disp);
1038 put_rpm:
1039 	pm_runtime_put_sync(dev);
1040 	return err;
1041 }
1042 
1043 static const struct host1x_client_ops tegra_display_hub_ops = {
1044 	.init = tegra_display_hub_init,
1045 	.exit = tegra_display_hub_exit,
1046 	.suspend = tegra_display_hub_runtime_suspend,
1047 	.resume = tegra_display_hub_runtime_resume,
1048 };
1049 
1050 static int tegra_display_hub_probe(struct platform_device *pdev)
1051 {
1052 	u64 dma_mask = dma_get_mask(pdev->dev.parent);
1053 	struct device_node *child = NULL;
1054 	struct tegra_display_hub *hub;
1055 	struct clk *clk;
1056 	unsigned int i;
1057 	int err;
1058 
1059 	err = dma_coerce_mask_and_coherent(&pdev->dev, dma_mask);
1060 	if (err < 0) {
1061 		dev_err(&pdev->dev, "failed to set DMA mask: %d\n", err);
1062 		return err;
1063 	}
1064 
1065 	hub = devm_kzalloc(&pdev->dev, sizeof(*hub), GFP_KERNEL);
1066 	if (!hub)
1067 		return -ENOMEM;
1068 
1069 	hub->soc = of_device_get_match_data(&pdev->dev);
1070 
1071 	hub->clk_disp = devm_clk_get(&pdev->dev, "disp");
1072 	if (IS_ERR(hub->clk_disp)) {
1073 		err = PTR_ERR(hub->clk_disp);
1074 		return err;
1075 	}
1076 
1077 	if (hub->soc->supports_dsc) {
1078 		hub->clk_dsc = devm_clk_get(&pdev->dev, "dsc");
1079 		if (IS_ERR(hub->clk_dsc)) {
1080 			err = PTR_ERR(hub->clk_dsc);
1081 			return err;
1082 		}
1083 	}
1084 
1085 	hub->clk_hub = devm_clk_get(&pdev->dev, "hub");
1086 	if (IS_ERR(hub->clk_hub)) {
1087 		err = PTR_ERR(hub->clk_hub);
1088 		return err;
1089 	}
1090 
1091 	hub->rst = devm_reset_control_get(&pdev->dev, "misc");
1092 	if (IS_ERR(hub->rst)) {
1093 		err = PTR_ERR(hub->rst);
1094 		return err;
1095 	}
1096 
1097 	hub->wgrps = devm_kcalloc(&pdev->dev, hub->soc->num_wgrps,
1098 				  sizeof(*hub->wgrps), GFP_KERNEL);
1099 	if (!hub->wgrps)
1100 		return -ENOMEM;
1101 
1102 	for (i = 0; i < hub->soc->num_wgrps; i++) {
1103 		struct tegra_windowgroup *wgrp = &hub->wgrps[i];
1104 		char id[16];
1105 
1106 		snprintf(id, sizeof(id), "wgrp%u", i);
1107 		mutex_init(&wgrp->lock);
1108 		wgrp->usecount = 0;
1109 		wgrp->index = i;
1110 
1111 		wgrp->rst = devm_reset_control_get(&pdev->dev, id);
1112 		if (IS_ERR(wgrp->rst))
1113 			return PTR_ERR(wgrp->rst);
1114 
1115 		err = reset_control_assert(wgrp->rst);
1116 		if (err < 0)
1117 			return err;
1118 	}
1119 
1120 	hub->num_heads = of_get_child_count(pdev->dev.of_node);
1121 
1122 	hub->clk_heads = devm_kcalloc(&pdev->dev, hub->num_heads, sizeof(clk),
1123 				      GFP_KERNEL);
1124 	if (!hub->clk_heads)
1125 		return -ENOMEM;
1126 
1127 	for (i = 0; i < hub->num_heads; i++) {
1128 		child = of_get_next_child(pdev->dev.of_node, child);
1129 		if (!child) {
1130 			dev_err(&pdev->dev, "failed to find node for head %u\n",
1131 				i);
1132 			return -ENODEV;
1133 		}
1134 
1135 		clk = devm_get_clk_from_child(&pdev->dev, child, "dc");
1136 		if (IS_ERR(clk)) {
1137 			dev_err(&pdev->dev, "failed to get clock for head %u\n",
1138 				i);
1139 			of_node_put(child);
1140 			return PTR_ERR(clk);
1141 		}
1142 
1143 		hub->clk_heads[i] = clk;
1144 	}
1145 
1146 	of_node_put(child);
1147 
1148 	/* XXX: enable clock across reset? */
1149 	err = reset_control_assert(hub->rst);
1150 	if (err < 0)
1151 		return err;
1152 
1153 	platform_set_drvdata(pdev, hub);
1154 	pm_runtime_enable(&pdev->dev);
1155 
1156 	INIT_LIST_HEAD(&hub->client.list);
1157 	hub->client.ops = &tegra_display_hub_ops;
1158 	hub->client.dev = &pdev->dev;
1159 
1160 	err = host1x_client_register(&hub->client);
1161 	if (err < 0)
1162 		dev_err(&pdev->dev, "failed to register host1x client: %d\n",
1163 			err);
1164 
1165 	err = devm_of_platform_populate(&pdev->dev);
1166 	if (err < 0)
1167 		goto unregister;
1168 
1169 	return err;
1170 
1171 unregister:
1172 	host1x_client_unregister(&hub->client);
1173 	pm_runtime_disable(&pdev->dev);
1174 	return err;
1175 }
1176 
1177 static void tegra_display_hub_remove(struct platform_device *pdev)
1178 {
1179 	struct tegra_display_hub *hub = platform_get_drvdata(pdev);
1180 	unsigned int i;
1181 
1182 	host1x_client_unregister(&hub->client);
1183 
1184 	for (i = 0; i < hub->soc->num_wgrps; i++) {
1185 		struct tegra_windowgroup *wgrp = &hub->wgrps[i];
1186 
1187 		mutex_destroy(&wgrp->lock);
1188 	}
1189 
1190 	pm_runtime_disable(&pdev->dev);
1191 }
1192 
1193 static const struct tegra_display_hub_soc tegra186_display_hub = {
1194 	.num_wgrps = 6,
1195 	.supports_dsc = true,
1196 };
1197 
1198 static const struct tegra_display_hub_soc tegra194_display_hub = {
1199 	.num_wgrps = 6,
1200 	.supports_dsc = false,
1201 };
1202 
1203 static const struct of_device_id tegra_display_hub_of_match[] = {
1204 	{
1205 		.compatible = "nvidia,tegra194-display",
1206 		.data = &tegra194_display_hub
1207 	}, {
1208 		.compatible = "nvidia,tegra186-display",
1209 		.data = &tegra186_display_hub
1210 	}, {
1211 		/* sentinel */
1212 	}
1213 };
1214 MODULE_DEVICE_TABLE(of, tegra_display_hub_of_match);
1215 
1216 struct platform_driver tegra_display_hub_driver = {
1217 	.driver = {
1218 		.name = "tegra-display-hub",
1219 		.of_match_table = tegra_display_hub_of_match,
1220 	},
1221 	.probe = tegra_display_hub_probe,
1222 	.remove_new = tegra_display_hub_remove,
1223 };
1224