xref: /linux/drivers/gpu/drm/xe/display/xe_display.c (revision f7d7ccf92f2b9398781f791b4af1a74a9f65b5c3)
1 // SPDX-License-Identifier: MIT
2 /*
3  * Copyright © 2023 Intel Corporation
4  */
5 
6 #include "xe_display.h"
7 #include "regs/xe_regs.h"
8 
9 #include <linux/fb.h>
10 
11 #include <drm/drm_drv.h>
12 #include <drm/drm_managed.h>
13 #include <drm/xe_drm.h>
14 
15 #include "soc/intel_dram.h"
16 #include "i915_drv.h"		/* FIXME: HAS_DISPLAY() depends on this */
17 #include "intel_acpi.h"
18 #include "intel_audio.h"
19 #include "intel_bw.h"
20 #include "intel_display.h"
21 #include "intel_display_driver.h"
22 #include "intel_display_irq.h"
23 #include "intel_display_types.h"
24 #include "intel_dmc.h"
25 #include "intel_dp.h"
26 #include "intel_encoder.h"
27 #include "intel_fbdev.h"
28 #include "intel_hdcp.h"
29 #include "intel_hotplug.h"
30 #include "intel_opregion.h"
31 #include "xe_module.h"
32 
33 /* Xe device functions */
34 
35 static bool has_display(struct xe_device *xe)
36 {
37 	return HAS_DISPLAY(xe);
38 }
39 
40 /**
41  * xe_display_driver_probe_defer - Detect if we need to wait for other drivers
42  *				   early on
43  * @pdev: PCI device
44  *
45  * Returns: true if probe needs to be deferred, false otherwise
46  */
47 bool xe_display_driver_probe_defer(struct pci_dev *pdev)
48 {
49 	if (!xe_modparam.enable_display)
50 		return 0;
51 
52 	return intel_display_driver_probe_defer(pdev);
53 }
54 
55 /**
56  * xe_display_driver_set_hooks - Add driver flags and hooks for display
57  * @driver: DRM device driver
58  *
59  * Set features and function hooks in @driver that are needed for driving the
60  * display IP. This sets the driver's capability of driving display, regardless
61  * if the device has it enabled
62  */
63 void xe_display_driver_set_hooks(struct drm_driver *driver)
64 {
65 	if (!xe_modparam.enable_display)
66 		return;
67 
68 	driver->driver_features |= DRIVER_MODESET | DRIVER_ATOMIC;
69 }
70 
71 static void unset_display_features(struct xe_device *xe)
72 {
73 	xe->drm.driver_features &= ~(DRIVER_MODESET | DRIVER_ATOMIC);
74 }
75 
76 static void display_destroy(struct drm_device *dev, void *dummy)
77 {
78 	struct xe_device *xe = to_xe_device(dev);
79 
80 	destroy_workqueue(xe->display.hotplug.dp_wq);
81 }
82 
83 /**
84  * xe_display_create - create display struct
85  * @xe: XE device instance
86  *
87  * Initialize all fields used by the display part.
88  *
89  * TODO: once everything can be inside a single struct, make the struct opaque
90  * to the rest of xe and return it to be xe->display.
91  *
92  * Returns: 0 on success
93  */
94 int xe_display_create(struct xe_device *xe)
95 {
96 	spin_lock_init(&xe->display.fb_tracking.lock);
97 
98 	xe->display.hotplug.dp_wq = alloc_ordered_workqueue("xe-dp", 0);
99 
100 	return drmm_add_action_or_reset(&xe->drm, display_destroy, NULL);
101 }
102 
103 static void xe_display_fini_nommio(struct drm_device *dev, void *dummy)
104 {
105 	struct xe_device *xe = to_xe_device(dev);
106 
107 	if (!xe->info.enable_display)
108 		return;
109 
110 	intel_power_domains_cleanup(xe);
111 }
112 
113 int xe_display_init_nommio(struct xe_device *xe)
114 {
115 	if (!xe->info.enable_display)
116 		return 0;
117 
118 	/* Fake uncore lock */
119 	spin_lock_init(&xe->uncore.lock);
120 
121 	/* This must be called before any calls to HAS_PCH_* */
122 	intel_detect_pch(xe);
123 
124 	return drmm_add_action_or_reset(&xe->drm, xe_display_fini_nommio, xe);
125 }
126 
127 static void xe_display_fini_noirq(void *arg)
128 {
129 	struct xe_device *xe = arg;
130 
131 	if (!xe->info.enable_display)
132 		return;
133 
134 	intel_display_driver_remove_noirq(xe);
135 	intel_opregion_cleanup(xe);
136 }
137 
138 int xe_display_init_noirq(struct xe_device *xe)
139 {
140 	int err;
141 
142 	if (!xe->info.enable_display)
143 		return 0;
144 
145 	intel_display_driver_early_probe(xe);
146 
147 	/* Early display init.. */
148 	intel_opregion_setup(xe);
149 
150 	/*
151 	 * Fill the dram structure to get the system dram info. This will be
152 	 * used for memory latency calculation.
153 	 */
154 	intel_dram_detect(xe);
155 
156 	intel_bw_init_hw(xe);
157 
158 	intel_display_device_info_runtime_init(xe);
159 
160 	err = intel_display_driver_probe_noirq(xe);
161 	if (err) {
162 		intel_opregion_cleanup(xe);
163 		return err;
164 	}
165 
166 	return devm_add_action_or_reset(xe->drm.dev, xe_display_fini_noirq, xe);
167 }
168 
169 static void xe_display_fini_noaccel(void *arg)
170 {
171 	struct xe_device *xe = arg;
172 
173 	if (!xe->info.enable_display)
174 		return;
175 
176 	intel_display_driver_remove_nogem(xe);
177 }
178 
179 int xe_display_init_noaccel(struct xe_device *xe)
180 {
181 	int err;
182 
183 	if (!xe->info.enable_display)
184 		return 0;
185 
186 	err = intel_display_driver_probe_nogem(xe);
187 	if (err)
188 		return err;
189 
190 	return devm_add_action_or_reset(xe->drm.dev, xe_display_fini_noaccel, xe);
191 }
192 
193 int xe_display_init(struct xe_device *xe)
194 {
195 	if (!xe->info.enable_display)
196 		return 0;
197 
198 	return intel_display_driver_probe(xe);
199 }
200 
201 void xe_display_fini(struct xe_device *xe)
202 {
203 	if (!xe->info.enable_display)
204 		return;
205 
206 	intel_hpd_poll_fini(xe);
207 
208 	intel_hdcp_component_fini(xe);
209 	intel_audio_deinit(xe);
210 }
211 
212 void xe_display_register(struct xe_device *xe)
213 {
214 	if (!xe->info.enable_display)
215 		return;
216 
217 	intel_display_driver_register(xe);
218 	intel_register_dsm_handler();
219 	intel_power_domains_enable(xe);
220 }
221 
222 void xe_display_unregister(struct xe_device *xe)
223 {
224 	if (!xe->info.enable_display)
225 		return;
226 
227 	intel_unregister_dsm_handler();
228 	intel_power_domains_disable(xe);
229 	intel_display_driver_unregister(xe);
230 }
231 
232 void xe_display_driver_remove(struct xe_device *xe)
233 {
234 	if (!xe->info.enable_display)
235 		return;
236 
237 	intel_display_driver_remove(xe);
238 }
239 
240 /* IRQ-related functions */
241 
242 void xe_display_irq_handler(struct xe_device *xe, u32 master_ctl)
243 {
244 	if (!xe->info.enable_display)
245 		return;
246 
247 	if (master_ctl & DISPLAY_IRQ)
248 		gen11_display_irq_handler(xe);
249 }
250 
251 void xe_display_irq_enable(struct xe_device *xe, u32 gu_misc_iir)
252 {
253 	if (!xe->info.enable_display)
254 		return;
255 
256 	if (gu_misc_iir & GU_MISC_GSE)
257 		intel_opregion_asle_intr(xe);
258 }
259 
260 void xe_display_irq_reset(struct xe_device *xe)
261 {
262 	if (!xe->info.enable_display)
263 		return;
264 
265 	gen11_display_irq_reset(xe);
266 }
267 
268 void xe_display_irq_postinstall(struct xe_device *xe, struct xe_gt *gt)
269 {
270 	if (!xe->info.enable_display)
271 		return;
272 
273 	if (gt->info.id == XE_GT0)
274 		gen11_de_irq_postinstall(xe);
275 }
276 
277 static bool suspend_to_idle(void)
278 {
279 #if IS_ENABLED(CONFIG_ACPI_SLEEP)
280 	if (acpi_target_system_state() < ACPI_STATE_S3)
281 		return true;
282 #endif
283 	return false;
284 }
285 
286 static void xe_display_flush_cleanup_work(struct xe_device *xe)
287 {
288 	struct intel_crtc *crtc;
289 
290 	for_each_intel_crtc(&xe->drm, crtc) {
291 		struct drm_crtc_commit *commit;
292 
293 		spin_lock(&crtc->base.commit_lock);
294 		commit = list_first_entry_or_null(&crtc->base.commit_list,
295 						  struct drm_crtc_commit, commit_entry);
296 		if (commit)
297 			drm_crtc_commit_get(commit);
298 		spin_unlock(&crtc->base.commit_lock);
299 
300 		if (commit) {
301 			wait_for_completion(&commit->cleanup_done);
302 			drm_crtc_commit_put(commit);
303 		}
304 	}
305 }
306 
307 void xe_display_pm_suspend(struct xe_device *xe, bool runtime)
308 {
309 	bool s2idle = suspend_to_idle();
310 	if (!xe->info.enable_display)
311 		return;
312 
313 	/*
314 	 * We do a lot of poking in a lot of registers, make sure they work
315 	 * properly.
316 	 */
317 	intel_power_domains_disable(xe);
318 	if (has_display(xe))
319 		drm_kms_helper_poll_disable(&xe->drm);
320 
321 	if (!runtime)
322 		intel_display_driver_suspend(xe);
323 
324 	xe_display_flush_cleanup_work(xe);
325 
326 	intel_dp_mst_suspend(xe);
327 
328 	intel_hpd_cancel_work(xe);
329 
330 	intel_encoder_suspend_all(&xe->display);
331 
332 	intel_opregion_suspend(xe, s2idle ? PCI_D1 : PCI_D3cold);
333 
334 	intel_fbdev_set_suspend(&xe->drm, FBINFO_STATE_SUSPENDED, true);
335 
336 	intel_dmc_suspend(xe);
337 }
338 
339 void xe_display_pm_suspend_late(struct xe_device *xe)
340 {
341 	bool s2idle = suspend_to_idle();
342 	if (!xe->info.enable_display)
343 		return;
344 
345 	intel_power_domains_suspend(xe, s2idle);
346 
347 	intel_display_power_suspend_late(xe);
348 }
349 
350 void xe_display_pm_resume_early(struct xe_device *xe)
351 {
352 	if (!xe->info.enable_display)
353 		return;
354 
355 	intel_display_power_resume_early(xe);
356 
357 	intel_power_domains_resume(xe);
358 }
359 
360 void xe_display_pm_resume(struct xe_device *xe, bool runtime)
361 {
362 	if (!xe->info.enable_display)
363 		return;
364 
365 	intel_dmc_resume(xe);
366 
367 	if (has_display(xe))
368 		drm_mode_config_reset(&xe->drm);
369 
370 	intel_display_driver_init_hw(xe);
371 	intel_hpd_init(xe);
372 
373 	/* MST sideband requires HPD interrupts enabled */
374 	intel_dp_mst_resume(xe);
375 	if (!runtime)
376 		intel_display_driver_resume(xe);
377 
378 	intel_hpd_poll_disable(xe);
379 	if (has_display(xe))
380 		drm_kms_helper_poll_enable(&xe->drm);
381 
382 	intel_opregion_resume(xe);
383 
384 	intel_fbdev_set_suspend(&xe->drm, FBINFO_STATE_RUNNING, false);
385 
386 	intel_power_domains_enable(xe);
387 }
388 
389 static void display_device_remove(struct drm_device *dev, void *arg)
390 {
391 	struct xe_device *xe = arg;
392 
393 	intel_display_device_remove(xe);
394 }
395 
396 int xe_display_probe(struct xe_device *xe)
397 {
398 	int err;
399 
400 	if (!xe->info.enable_display)
401 		goto no_display;
402 
403 	intel_display_device_probe(xe);
404 
405 	err = drmm_add_action_or_reset(&xe->drm, display_device_remove, xe);
406 	if (err)
407 		return err;
408 
409 	if (has_display(xe))
410 		return 0;
411 
412 no_display:
413 	xe->info.enable_display = false;
414 	unset_display_features(xe);
415 	return 0;
416 }
417