1 // SPDX-License-Identifier: MIT
2 /*
3 * Copyright © 2023 Intel Corporation
4 */
5
6 #include "xe_display.h"
7 #include "regs/xe_regs.h"
8
9 #include <linux/fb.h>
10
11 #include <drm/drm_drv.h>
12 #include <drm/drm_managed.h>
13 #include <uapi/drm/xe_drm.h>
14
15 #include "soc/intel_dram.h"
16 #include "i915_drv.h" /* FIXME: HAS_DISPLAY() depends on this */
17 #include "intel_acpi.h"
18 #include "intel_audio.h"
19 #include "intel_bw.h"
20 #include "intel_display.h"
21 #include "intel_display_driver.h"
22 #include "intel_display_irq.h"
23 #include "intel_display_types.h"
24 #include "intel_dmc.h"
25 #include "intel_dp.h"
26 #include "intel_encoder.h"
27 #include "intel_fbdev.h"
28 #include "intel_hdcp.h"
29 #include "intel_hotplug.h"
30 #include "intel_opregion.h"
31 #include "xe_module.h"
32
33 /* Xe device functions */
34
has_display(struct xe_device * xe)35 static bool has_display(struct xe_device *xe)
36 {
37 return HAS_DISPLAY(xe);
38 }
39
40 /**
41 * xe_display_driver_probe_defer - Detect if we need to wait for other drivers
42 * early on
43 * @pdev: PCI device
44 *
45 * Returns: true if probe needs to be deferred, false otherwise
46 */
xe_display_driver_probe_defer(struct pci_dev * pdev)47 bool xe_display_driver_probe_defer(struct pci_dev *pdev)
48 {
49 if (!xe_modparam.probe_display)
50 return 0;
51
52 return intel_display_driver_probe_defer(pdev);
53 }
54
55 /**
56 * xe_display_driver_set_hooks - Add driver flags and hooks for display
57 * @driver: DRM device driver
58 *
59 * Set features and function hooks in @driver that are needed for driving the
60 * display IP. This sets the driver's capability of driving display, regardless
61 * if the device has it enabled
62 */
xe_display_driver_set_hooks(struct drm_driver * driver)63 void xe_display_driver_set_hooks(struct drm_driver *driver)
64 {
65 if (!xe_modparam.probe_display)
66 return;
67
68 driver->driver_features |= DRIVER_MODESET | DRIVER_ATOMIC;
69 }
70
unset_display_features(struct xe_device * xe)71 static void unset_display_features(struct xe_device *xe)
72 {
73 xe->drm.driver_features &= ~(DRIVER_MODESET | DRIVER_ATOMIC);
74 }
75
display_destroy(struct drm_device * dev,void * dummy)76 static void display_destroy(struct drm_device *dev, void *dummy)
77 {
78 struct xe_device *xe = to_xe_device(dev);
79
80 destroy_workqueue(xe->display.hotplug.dp_wq);
81 }
82
83 /**
84 * xe_display_create - create display struct
85 * @xe: XE device instance
86 *
87 * Initialize all fields used by the display part.
88 *
89 * TODO: once everything can be inside a single struct, make the struct opaque
90 * to the rest of xe and return it to be xe->display.
91 *
92 * Returns: 0 on success
93 */
xe_display_create(struct xe_device * xe)94 int xe_display_create(struct xe_device *xe)
95 {
96 spin_lock_init(&xe->display.fb_tracking.lock);
97
98 xe->display.hotplug.dp_wq = alloc_ordered_workqueue("xe-dp", 0);
99
100 return drmm_add_action_or_reset(&xe->drm, display_destroy, NULL);
101 }
102
xe_display_fini_nommio(struct drm_device * dev,void * dummy)103 static void xe_display_fini_nommio(struct drm_device *dev, void *dummy)
104 {
105 struct xe_device *xe = to_xe_device(dev);
106
107 if (!xe->info.probe_display)
108 return;
109
110 intel_power_domains_cleanup(xe);
111 }
112
xe_display_init_nommio(struct xe_device * xe)113 int xe_display_init_nommio(struct xe_device *xe)
114 {
115 if (!xe->info.probe_display)
116 return 0;
117
118 /* Fake uncore lock */
119 spin_lock_init(&xe->uncore.lock);
120
121 /* This must be called before any calls to HAS_PCH_* */
122 intel_detect_pch(xe);
123
124 return drmm_add_action_or_reset(&xe->drm, xe_display_fini_nommio, xe);
125 }
126
xe_display_fini_noirq(void * arg)127 static void xe_display_fini_noirq(void *arg)
128 {
129 struct xe_device *xe = arg;
130 struct intel_display *display = &xe->display;
131
132 if (!xe->info.probe_display)
133 return;
134
135 intel_display_driver_remove_noirq(xe);
136 intel_opregion_cleanup(display);
137 }
138
xe_display_init_noirq(struct xe_device * xe)139 int xe_display_init_noirq(struct xe_device *xe)
140 {
141 struct intel_display *display = &xe->display;
142 int err;
143
144 if (!xe->info.probe_display)
145 return 0;
146
147 intel_display_driver_early_probe(xe);
148
149 /* Early display init.. */
150 intel_opregion_setup(display);
151
152 /*
153 * Fill the dram structure to get the system dram info. This will be
154 * used for memory latency calculation.
155 */
156 intel_dram_detect(xe);
157
158 intel_bw_init_hw(xe);
159
160 intel_display_device_info_runtime_init(xe);
161
162 err = intel_display_driver_probe_noirq(xe);
163 if (err) {
164 intel_opregion_cleanup(display);
165 return err;
166 }
167
168 return devm_add_action_or_reset(xe->drm.dev, xe_display_fini_noirq, xe);
169 }
170
xe_display_fini_noaccel(void * arg)171 static void xe_display_fini_noaccel(void *arg)
172 {
173 struct xe_device *xe = arg;
174
175 if (!xe->info.probe_display)
176 return;
177
178 intel_display_driver_remove_nogem(xe);
179 }
180
xe_display_init_noaccel(struct xe_device * xe)181 int xe_display_init_noaccel(struct xe_device *xe)
182 {
183 int err;
184
185 if (!xe->info.probe_display)
186 return 0;
187
188 err = intel_display_driver_probe_nogem(xe);
189 if (err)
190 return err;
191
192 return devm_add_action_or_reset(xe->drm.dev, xe_display_fini_noaccel, xe);
193 }
194
xe_display_init(struct xe_device * xe)195 int xe_display_init(struct xe_device *xe)
196 {
197 if (!xe->info.probe_display)
198 return 0;
199
200 return intel_display_driver_probe(xe);
201 }
202
xe_display_fini(struct xe_device * xe)203 void xe_display_fini(struct xe_device *xe)
204 {
205 if (!xe->info.probe_display)
206 return;
207
208 intel_hpd_poll_fini(xe);
209
210 intel_hdcp_component_fini(xe);
211 intel_audio_deinit(xe);
212 }
213
xe_display_register(struct xe_device * xe)214 void xe_display_register(struct xe_device *xe)
215 {
216 if (!xe->info.probe_display)
217 return;
218
219 intel_display_driver_register(xe);
220 intel_register_dsm_handler();
221 intel_power_domains_enable(xe);
222 }
223
xe_display_unregister(struct xe_device * xe)224 void xe_display_unregister(struct xe_device *xe)
225 {
226 if (!xe->info.probe_display)
227 return;
228
229 intel_unregister_dsm_handler();
230 intel_power_domains_disable(xe);
231 intel_display_driver_unregister(xe);
232 }
233
xe_display_driver_remove(struct xe_device * xe)234 void xe_display_driver_remove(struct xe_device *xe)
235 {
236 if (!xe->info.probe_display)
237 return;
238
239 intel_display_driver_remove(xe);
240 }
241
242 /* IRQ-related functions */
243
xe_display_irq_handler(struct xe_device * xe,u32 master_ctl)244 void xe_display_irq_handler(struct xe_device *xe, u32 master_ctl)
245 {
246 if (!xe->info.probe_display)
247 return;
248
249 if (master_ctl & DISPLAY_IRQ)
250 gen11_display_irq_handler(xe);
251 }
252
xe_display_irq_enable(struct xe_device * xe,u32 gu_misc_iir)253 void xe_display_irq_enable(struct xe_device *xe, u32 gu_misc_iir)
254 {
255 struct intel_display *display = &xe->display;
256
257 if (!xe->info.probe_display)
258 return;
259
260 if (gu_misc_iir & GU_MISC_GSE)
261 intel_opregion_asle_intr(display);
262 }
263
xe_display_irq_reset(struct xe_device * xe)264 void xe_display_irq_reset(struct xe_device *xe)
265 {
266 if (!xe->info.probe_display)
267 return;
268
269 gen11_display_irq_reset(xe);
270 }
271
xe_display_irq_postinstall(struct xe_device * xe,struct xe_gt * gt)272 void xe_display_irq_postinstall(struct xe_device *xe, struct xe_gt *gt)
273 {
274 if (!xe->info.probe_display)
275 return;
276
277 if (gt->info.id == XE_GT0)
278 gen11_de_irq_postinstall(xe);
279 }
280
suspend_to_idle(void)281 static bool suspend_to_idle(void)
282 {
283 #if IS_ENABLED(CONFIG_ACPI_SLEEP)
284 if (acpi_target_system_state() < ACPI_STATE_S3)
285 return true;
286 #endif
287 return false;
288 }
289
xe_display_flush_cleanup_work(struct xe_device * xe)290 static void xe_display_flush_cleanup_work(struct xe_device *xe)
291 {
292 struct intel_crtc *crtc;
293
294 for_each_intel_crtc(&xe->drm, crtc) {
295 struct drm_crtc_commit *commit;
296
297 spin_lock(&crtc->base.commit_lock);
298 commit = list_first_entry_or_null(&crtc->base.commit_list,
299 struct drm_crtc_commit, commit_entry);
300 if (commit)
301 drm_crtc_commit_get(commit);
302 spin_unlock(&crtc->base.commit_lock);
303
304 if (commit) {
305 wait_for_completion(&commit->cleanup_done);
306 drm_crtc_commit_put(commit);
307 }
308 }
309 }
310
311 /* TODO: System and runtime suspend/resume sequences will be sanitized as a follow-up. */
xe_display_pm_runtime_suspend(struct xe_device * xe)312 void xe_display_pm_runtime_suspend(struct xe_device *xe)
313 {
314 if (!xe->info.probe_display)
315 return;
316
317 if (xe->d3cold.allowed)
318 xe_display_pm_suspend(xe, true);
319
320 intel_hpd_poll_enable(xe);
321 }
322
xe_display_pm_suspend(struct xe_device * xe,bool runtime)323 void xe_display_pm_suspend(struct xe_device *xe, bool runtime)
324 {
325 struct intel_display *display = &xe->display;
326 bool s2idle = suspend_to_idle();
327 if (!xe->info.probe_display)
328 return;
329
330 /*
331 * We do a lot of poking in a lot of registers, make sure they work
332 * properly.
333 */
334 intel_power_domains_disable(xe);
335 intel_fbdev_set_suspend(&xe->drm, FBINFO_STATE_SUSPENDED, true);
336 if (!runtime && has_display(xe)) {
337 drm_kms_helper_poll_disable(&xe->drm);
338 intel_display_driver_disable_user_access(xe);
339 intel_display_driver_suspend(xe);
340 }
341
342 xe_display_flush_cleanup_work(xe);
343
344 intel_dp_mst_suspend(xe);
345
346 intel_hpd_cancel_work(xe);
347
348 if (!runtime && has_display(xe)) {
349 intel_display_driver_suspend_access(xe);
350 intel_encoder_suspend_all(&xe->display);
351 }
352
353 intel_opregion_suspend(display, s2idle ? PCI_D1 : PCI_D3cold);
354
355 intel_dmc_suspend(xe);
356 }
357
xe_display_pm_suspend_late(struct xe_device * xe)358 void xe_display_pm_suspend_late(struct xe_device *xe)
359 {
360 bool s2idle = suspend_to_idle();
361 if (!xe->info.probe_display)
362 return;
363
364 intel_power_domains_suspend(xe, s2idle);
365
366 intel_display_power_suspend_late(xe);
367 }
368
xe_display_pm_runtime_resume(struct xe_device * xe)369 void xe_display_pm_runtime_resume(struct xe_device *xe)
370 {
371 if (!xe->info.probe_display)
372 return;
373
374 intel_hpd_poll_disable(xe);
375
376 if (xe->d3cold.allowed)
377 xe_display_pm_resume(xe, true);
378 }
379
xe_display_pm_resume_early(struct xe_device * xe)380 void xe_display_pm_resume_early(struct xe_device *xe)
381 {
382 if (!xe->info.probe_display)
383 return;
384
385 intel_display_power_resume_early(xe);
386
387 intel_power_domains_resume(xe);
388 }
389
xe_display_pm_resume(struct xe_device * xe,bool runtime)390 void xe_display_pm_resume(struct xe_device *xe, bool runtime)
391 {
392 struct intel_display *display = &xe->display;
393
394 if (!xe->info.probe_display)
395 return;
396
397 intel_dmc_resume(xe);
398
399 if (has_display(xe))
400 drm_mode_config_reset(&xe->drm);
401
402 intel_display_driver_init_hw(xe);
403 intel_hpd_init(xe);
404
405 if (!runtime && has_display(xe))
406 intel_display_driver_resume_access(xe);
407
408 /* MST sideband requires HPD interrupts enabled */
409 intel_dp_mst_resume(xe);
410 if (!runtime && has_display(xe)) {
411 intel_display_driver_resume(xe);
412 drm_kms_helper_poll_enable(&xe->drm);
413 intel_display_driver_enable_user_access(xe);
414 intel_hpd_poll_disable(xe);
415 }
416
417 intel_opregion_resume(display);
418
419 intel_fbdev_set_suspend(&xe->drm, FBINFO_STATE_RUNNING, false);
420
421 intel_power_domains_enable(xe);
422 }
423
display_device_remove(struct drm_device * dev,void * arg)424 static void display_device_remove(struct drm_device *dev, void *arg)
425 {
426 struct xe_device *xe = arg;
427
428 intel_display_device_remove(xe);
429 }
430
xe_display_probe(struct xe_device * xe)431 int xe_display_probe(struct xe_device *xe)
432 {
433 int err;
434
435 if (!xe->info.probe_display)
436 goto no_display;
437
438 intel_display_device_probe(xe);
439
440 err = drmm_add_action_or_reset(&xe->drm, display_device_remove, xe);
441 if (err)
442 return err;
443
444 if (has_display(xe))
445 return 0;
446
447 no_display:
448 xe->info.probe_display = false;
449 unset_display_features(xe);
450 return 0;
451 }
452