xref: /linux/drivers/gpu/drm/drm_gem_atomic_helper.c (revision 55d0969c451159cff86949b38c39171cab962069)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 
3 #include <linux/dma-resv.h>
4 #include <linux/dma-fence-chain.h>
5 
6 #include <drm/drm_atomic_state_helper.h>
7 #include <drm/drm_atomic_uapi.h>
8 #include <drm/drm_framebuffer.h>
9 #include <drm/drm_gem.h>
10 #include <drm/drm_gem_atomic_helper.h>
11 #include <drm/drm_gem_framebuffer_helper.h>
12 #include <drm/drm_simple_kms_helper.h>
13 
14 #include "drm_internal.h"
15 
16 /**
17  * DOC: overview
18  *
19  * The GEM atomic helpers library implements generic atomic-commit
20  * functions for drivers that use GEM objects. Currently, it provides
21  * synchronization helpers, and plane state and framebuffer BO mappings
22  * for planes with shadow buffers.
23  *
24  * Before scanout, a plane's framebuffer needs to be synchronized with
25  * possible writers that draw into the framebuffer. All drivers should
26  * call drm_gem_plane_helper_prepare_fb() from their implementation of
27  * struct &drm_plane_helper.prepare_fb . It sets the plane's fence from
28  * the framebuffer so that the DRM core can synchronize access automatically.
29  * drm_gem_plane_helper_prepare_fb() can also be used directly as
30  * implementation of prepare_fb.
31  *
32  * .. code-block:: c
33  *
34  *	#include <drm/drm_gem_atomic_helper.h>
35  *
36  *	struct drm_plane_helper_funcs driver_plane_helper_funcs = {
37  *		...,
38  *		. prepare_fb = drm_gem_plane_helper_prepare_fb,
39  *	};
40  *
41  * A driver using a shadow buffer copies the content of the shadow buffers
42  * into the HW's framebuffer memory during an atomic update. This requires
43  * a mapping of the shadow buffer into kernel address space. The mappings
44  * cannot be established by commit-tail functions, such as atomic_update,
45  * as this would violate locking rules around dma_buf_vmap().
46  *
47  * The helpers for shadow-buffered planes establish and release mappings,
48  * and provide struct drm_shadow_plane_state, which stores the plane's mapping
49  * for commit-tail functions.
50  *
51  * Shadow-buffered planes can easily be enabled by using the provided macros
52  * %DRM_GEM_SHADOW_PLANE_FUNCS and %DRM_GEM_SHADOW_PLANE_HELPER_FUNCS.
53  * These macros set up the plane and plane-helper callbacks to point to the
54  * shadow-buffer helpers.
55  *
56  * .. code-block:: c
57  *
58  *	#include <drm/drm_gem_atomic_helper.h>
59  *
60  *	struct drm_plane_funcs driver_plane_funcs = {
61  *		...,
62  *		DRM_GEM_SHADOW_PLANE_FUNCS,
63  *	};
64  *
65  *	struct drm_plane_helper_funcs driver_plane_helper_funcs = {
66  *		...,
67  *		DRM_GEM_SHADOW_PLANE_HELPER_FUNCS,
68  *	};
69  *
70  * In the driver's atomic-update function, shadow-buffer mappings are available
71  * from the plane state. Use to_drm_shadow_plane_state() to upcast from
72  * struct drm_plane_state.
73  *
74  * .. code-block:: c
75  *
76  *	void driver_plane_atomic_update(struct drm_plane *plane,
77  *					struct drm_plane_state *old_plane_state)
78  *	{
79  *		struct drm_plane_state *plane_state = plane->state;
80  *		struct drm_shadow_plane_state *shadow_plane_state =
81  *			to_drm_shadow_plane_state(plane_state);
82  *
83  *		// access shadow buffer via shadow_plane_state->map
84  *	}
85  *
86  * A mapping address for each of the framebuffer's buffer object is stored in
87  * struct &drm_shadow_plane_state.map. The mappings are valid while the state
88  * is being used.
89  *
90  * Drivers that use struct drm_simple_display_pipe can use
91  * %DRM_GEM_SIMPLE_DISPLAY_PIPE_SHADOW_PLANE_FUNCS to initialize the rsp
92  * callbacks. Access to shadow-buffer mappings is similar to regular
93  * atomic_update.
94  *
95  * .. code-block:: c
96  *
97  *	struct drm_simple_display_pipe_funcs driver_pipe_funcs = {
98  *		...,
99  *		DRM_GEM_SIMPLE_DISPLAY_PIPE_SHADOW_PLANE_FUNCS,
100  *	};
101  *
102  *	void driver_pipe_enable(struct drm_simple_display_pipe *pipe,
103  *				struct drm_crtc_state *crtc_state,
104  *				struct drm_plane_state *plane_state)
105  *	{
106  *		struct drm_shadow_plane_state *shadow_plane_state =
107  *			to_drm_shadow_plane_state(plane_state);
108  *
109  *		// access shadow buffer via shadow_plane_state->map
110  *	}
111  */
112 
113 /*
114  * Plane Helpers
115  */
116 
117 /**
118  * drm_gem_plane_helper_prepare_fb() - Prepare a GEM backed framebuffer
119  * @plane: Plane
120  * @state: Plane state the fence will be attached to
121  *
122  * This function extracts the exclusive fence from &drm_gem_object.resv and
123  * attaches it to plane state for the atomic helper to wait on. This is
124  * necessary to correctly implement implicit synchronization for any buffers
125  * shared as a struct &dma_buf. This function can be used as the
126  * &drm_plane_helper_funcs.prepare_fb callback.
127  *
128  * There is no need for &drm_plane_helper_funcs.cleanup_fb hook for simple
129  * GEM based framebuffer drivers which have their buffers always pinned in
130  * memory.
131  *
132  * This function is the default implementation for GEM drivers of
133  * &drm_plane_helper_funcs.prepare_fb if no callback is provided.
134  */
135 int drm_gem_plane_helper_prepare_fb(struct drm_plane *plane,
136 				    struct drm_plane_state *state)
137 {
138 	struct dma_fence *fence = dma_fence_get(state->fence);
139 	enum dma_resv_usage usage;
140 	size_t i;
141 	int ret;
142 
143 	if (!state->fb)
144 		return 0;
145 
146 	/*
147 	 * Only add the kernel fences here if there is already a fence set via
148 	 * explicit fencing interfaces on the atomic ioctl.
149 	 *
150 	 * This way explicit fencing can be used to overrule implicit fencing,
151 	 * which is important to make explicit fencing use-cases work: One
152 	 * example is using one buffer for 2 screens with different refresh
153 	 * rates. Implicit fencing will clamp rendering to the refresh rate of
154 	 * the slower screen, whereas explicit fence allows 2 independent
155 	 * render and display loops on a single buffer. If a driver allows
156 	 * obeys both implicit and explicit fences for plane updates, then it
157 	 * will break all the benefits of explicit fencing.
158 	 */
159 	usage = fence ? DMA_RESV_USAGE_KERNEL : DMA_RESV_USAGE_WRITE;
160 
161 	for (i = 0; i < state->fb->format->num_planes; ++i) {
162 		struct drm_gem_object *obj = drm_gem_fb_get_obj(state->fb, i);
163 		struct dma_fence *new;
164 
165 		if (!obj) {
166 			ret = -EINVAL;
167 			goto error;
168 		}
169 
170 		ret = dma_resv_get_singleton(obj->resv, usage, &new);
171 		if (ret)
172 			goto error;
173 
174 		if (new && fence) {
175 			struct dma_fence_chain *chain = dma_fence_chain_alloc();
176 
177 			if (!chain) {
178 				ret = -ENOMEM;
179 				goto error;
180 			}
181 
182 			dma_fence_chain_init(chain, fence, new, 1);
183 			fence = &chain->base;
184 
185 		} else if (new) {
186 			fence = new;
187 		}
188 	}
189 
190 	dma_fence_put(state->fence);
191 	state->fence = fence;
192 	return 0;
193 
194 error:
195 	dma_fence_put(fence);
196 	return ret;
197 }
198 EXPORT_SYMBOL_GPL(drm_gem_plane_helper_prepare_fb);
199 
200 /*
201  * Shadow-buffered Planes
202  */
203 
204 /**
205  * __drm_gem_duplicate_shadow_plane_state - duplicates shadow-buffered plane state
206  * @plane: the plane
207  * @new_shadow_plane_state: the new shadow-buffered plane state
208  *
209  * This function duplicates shadow-buffered plane state. This is helpful for drivers
210  * that subclass struct drm_shadow_plane_state.
211  *
212  * The function does not duplicate existing mappings of the shadow buffers.
213  * Mappings are maintained during the atomic commit by the plane's prepare_fb
214  * and cleanup_fb helpers. See drm_gem_prepare_shadow_fb() and drm_gem_cleanup_shadow_fb()
215  * for corresponding helpers.
216  */
217 void
218 __drm_gem_duplicate_shadow_plane_state(struct drm_plane *plane,
219 				       struct drm_shadow_plane_state *new_shadow_plane_state)
220 {
221 	struct drm_plane_state *plane_state = plane->state;
222 	struct drm_shadow_plane_state *shadow_plane_state =
223 		to_drm_shadow_plane_state(plane_state);
224 
225 	__drm_atomic_helper_plane_duplicate_state(plane, &new_shadow_plane_state->base);
226 
227 	drm_format_conv_state_copy(&new_shadow_plane_state->fmtcnv_state,
228 				   &shadow_plane_state->fmtcnv_state);
229 }
230 EXPORT_SYMBOL(__drm_gem_duplicate_shadow_plane_state);
231 
232 /**
233  * drm_gem_duplicate_shadow_plane_state - duplicates shadow-buffered plane state
234  * @plane: the plane
235  *
236  * This function implements struct &drm_plane_funcs.atomic_duplicate_state for
237  * shadow-buffered planes. It assumes the existing state to be of type
238  * struct drm_shadow_plane_state and it allocates the new state to be of this
239  * type.
240  *
241  * The function does not duplicate existing mappings of the shadow buffers.
242  * Mappings are maintained during the atomic commit by the plane's prepare_fb
243  * and cleanup_fb helpers. See drm_gem_prepare_shadow_fb() and drm_gem_cleanup_shadow_fb()
244  * for corresponding helpers.
245  *
246  * Returns:
247  * A pointer to a new plane state on success, or NULL otherwise.
248  */
249 struct drm_plane_state *
250 drm_gem_duplicate_shadow_plane_state(struct drm_plane *plane)
251 {
252 	struct drm_plane_state *plane_state = plane->state;
253 	struct drm_shadow_plane_state *new_shadow_plane_state;
254 
255 	if (!plane_state)
256 		return NULL;
257 
258 	new_shadow_plane_state = kzalloc(sizeof(*new_shadow_plane_state), GFP_KERNEL);
259 	if (!new_shadow_plane_state)
260 		return NULL;
261 	__drm_gem_duplicate_shadow_plane_state(plane, new_shadow_plane_state);
262 
263 	return &new_shadow_plane_state->base;
264 }
265 EXPORT_SYMBOL(drm_gem_duplicate_shadow_plane_state);
266 
267 /**
268  * __drm_gem_destroy_shadow_plane_state - cleans up shadow-buffered plane state
269  * @shadow_plane_state: the shadow-buffered plane state
270  *
271  * This function cleans up shadow-buffered plane state. Helpful for drivers that
272  * subclass struct drm_shadow_plane_state.
273  */
274 void __drm_gem_destroy_shadow_plane_state(struct drm_shadow_plane_state *shadow_plane_state)
275 {
276 	drm_format_conv_state_release(&shadow_plane_state->fmtcnv_state);
277 	__drm_atomic_helper_plane_destroy_state(&shadow_plane_state->base);
278 }
279 EXPORT_SYMBOL(__drm_gem_destroy_shadow_plane_state);
280 
281 /**
282  * drm_gem_destroy_shadow_plane_state - deletes shadow-buffered plane state
283  * @plane: the plane
284  * @plane_state: the plane state of type struct drm_shadow_plane_state
285  *
286  * This function implements struct &drm_plane_funcs.atomic_destroy_state
287  * for shadow-buffered planes. It expects that mappings of shadow buffers
288  * have been released already.
289  */
290 void drm_gem_destroy_shadow_plane_state(struct drm_plane *plane,
291 					struct drm_plane_state *plane_state)
292 {
293 	struct drm_shadow_plane_state *shadow_plane_state =
294 		to_drm_shadow_plane_state(plane_state);
295 
296 	__drm_gem_destroy_shadow_plane_state(shadow_plane_state);
297 	kfree(shadow_plane_state);
298 }
299 EXPORT_SYMBOL(drm_gem_destroy_shadow_plane_state);
300 
301 /**
302  * __drm_gem_reset_shadow_plane - resets a shadow-buffered plane
303  * @plane: the plane
304  * @shadow_plane_state: the shadow-buffered plane state
305  *
306  * This function resets state for shadow-buffered planes. Helpful
307  * for drivers that subclass struct drm_shadow_plane_state.
308  */
309 void __drm_gem_reset_shadow_plane(struct drm_plane *plane,
310 				  struct drm_shadow_plane_state *shadow_plane_state)
311 {
312 	__drm_atomic_helper_plane_reset(plane, &shadow_plane_state->base);
313 	drm_format_conv_state_init(&shadow_plane_state->fmtcnv_state);
314 }
315 EXPORT_SYMBOL(__drm_gem_reset_shadow_plane);
316 
317 /**
318  * drm_gem_reset_shadow_plane - resets a shadow-buffered plane
319  * @plane: the plane
320  *
321  * This function implements struct &drm_plane_funcs.reset_plane for
322  * shadow-buffered planes. It assumes the current plane state to be
323  * of type struct drm_shadow_plane and it allocates the new state of
324  * this type.
325  */
326 void drm_gem_reset_shadow_plane(struct drm_plane *plane)
327 {
328 	struct drm_shadow_plane_state *shadow_plane_state;
329 
330 	if (plane->state) {
331 		drm_gem_destroy_shadow_plane_state(plane, plane->state);
332 		plane->state = NULL; /* must be set to NULL here */
333 	}
334 
335 	shadow_plane_state = kzalloc(sizeof(*shadow_plane_state), GFP_KERNEL);
336 	if (!shadow_plane_state)
337 		return;
338 	__drm_gem_reset_shadow_plane(plane, shadow_plane_state);
339 }
340 EXPORT_SYMBOL(drm_gem_reset_shadow_plane);
341 
342 /**
343  * drm_gem_begin_shadow_fb_access - prepares shadow framebuffers for CPU access
344  * @plane: the plane
345  * @plane_state: the plane state of type struct drm_shadow_plane_state
346  *
347  * This function implements struct &drm_plane_helper_funcs.begin_fb_access. It
348  * maps all buffer objects of the plane's framebuffer into kernel address
349  * space and stores them in struct &drm_shadow_plane_state.map. The first data
350  * bytes are available in struct &drm_shadow_plane_state.data.
351  *
352  * See drm_gem_end_shadow_fb_access() for cleanup.
353  *
354  * Returns:
355  * 0 on success, or a negative errno code otherwise.
356  */
357 int drm_gem_begin_shadow_fb_access(struct drm_plane *plane, struct drm_plane_state *plane_state)
358 {
359 	struct drm_shadow_plane_state *shadow_plane_state = to_drm_shadow_plane_state(plane_state);
360 	struct drm_framebuffer *fb = plane_state->fb;
361 
362 	if (!fb)
363 		return 0;
364 
365 	return drm_gem_fb_vmap(fb, shadow_plane_state->map, shadow_plane_state->data);
366 }
367 EXPORT_SYMBOL(drm_gem_begin_shadow_fb_access);
368 
369 /**
370  * drm_gem_end_shadow_fb_access - releases shadow framebuffers from CPU access
371  * @plane: the plane
372  * @plane_state: the plane state of type struct drm_shadow_plane_state
373  *
374  * This function implements struct &drm_plane_helper_funcs.end_fb_access. It
375  * undoes all effects of drm_gem_begin_shadow_fb_access() in reverse order.
376  *
377  * See drm_gem_begin_shadow_fb_access() for more information.
378  */
379 void drm_gem_end_shadow_fb_access(struct drm_plane *plane, struct drm_plane_state *plane_state)
380 {
381 	struct drm_shadow_plane_state *shadow_plane_state = to_drm_shadow_plane_state(plane_state);
382 	struct drm_framebuffer *fb = plane_state->fb;
383 
384 	if (!fb)
385 		return;
386 
387 	drm_gem_fb_vunmap(fb, shadow_plane_state->map);
388 }
389 EXPORT_SYMBOL(drm_gem_end_shadow_fb_access);
390 
391 /**
392  * drm_gem_simple_kms_begin_shadow_fb_access - prepares shadow framebuffers for CPU access
393  * @pipe: the simple display pipe
394  * @plane_state: the plane state of type struct drm_shadow_plane_state
395  *
396  * This function implements struct drm_simple_display_funcs.begin_fb_access.
397  *
398  * See drm_gem_begin_shadow_fb_access() for details and
399  * drm_gem_simple_kms_cleanup_shadow_fb() for cleanup.
400  *
401  * Returns:
402  * 0 on success, or a negative errno code otherwise.
403  */
404 int drm_gem_simple_kms_begin_shadow_fb_access(struct drm_simple_display_pipe *pipe,
405 					      struct drm_plane_state *plane_state)
406 {
407 	return drm_gem_begin_shadow_fb_access(&pipe->plane, plane_state);
408 }
409 EXPORT_SYMBOL(drm_gem_simple_kms_begin_shadow_fb_access);
410 
411 /**
412  * drm_gem_simple_kms_end_shadow_fb_access - releases shadow framebuffers from CPU access
413  * @pipe: the simple display pipe
414  * @plane_state: the plane state of type struct drm_shadow_plane_state
415  *
416  * This function implements struct drm_simple_display_funcs.end_fb_access.
417  * It undoes all effects of drm_gem_simple_kms_begin_shadow_fb_access() in
418  * reverse order.
419  *
420  * See drm_gem_simple_kms_begin_shadow_fb_access().
421  */
422 void drm_gem_simple_kms_end_shadow_fb_access(struct drm_simple_display_pipe *pipe,
423 					     struct drm_plane_state *plane_state)
424 {
425 	drm_gem_end_shadow_fb_access(&pipe->plane, plane_state);
426 }
427 EXPORT_SYMBOL(drm_gem_simple_kms_end_shadow_fb_access);
428 
429 /**
430  * drm_gem_simple_kms_reset_shadow_plane - resets a shadow-buffered plane
431  * @pipe: the simple display pipe
432  *
433  * This function implements struct drm_simple_display_funcs.reset_plane
434  * for shadow-buffered planes.
435  */
436 void drm_gem_simple_kms_reset_shadow_plane(struct drm_simple_display_pipe *pipe)
437 {
438 	drm_gem_reset_shadow_plane(&pipe->plane);
439 }
440 EXPORT_SYMBOL(drm_gem_simple_kms_reset_shadow_plane);
441 
442 /**
443  * drm_gem_simple_kms_duplicate_shadow_plane_state - duplicates shadow-buffered plane state
444  * @pipe: the simple display pipe
445  *
446  * This function implements struct drm_simple_display_funcs.duplicate_plane_state
447  * for shadow-buffered planes. It does not duplicate existing mappings of the shadow
448  * buffers. Mappings are maintained during the atomic commit by the plane's prepare_fb
449  * and cleanup_fb helpers.
450  *
451  * Returns:
452  * A pointer to a new plane state on success, or NULL otherwise.
453  */
454 struct drm_plane_state *
455 drm_gem_simple_kms_duplicate_shadow_plane_state(struct drm_simple_display_pipe *pipe)
456 {
457 	return drm_gem_duplicate_shadow_plane_state(&pipe->plane);
458 }
459 EXPORT_SYMBOL(drm_gem_simple_kms_duplicate_shadow_plane_state);
460 
461 /**
462  * drm_gem_simple_kms_destroy_shadow_plane_state - resets shadow-buffered plane state
463  * @pipe: the simple display pipe
464  * @plane_state: the plane state of type struct drm_shadow_plane_state
465  *
466  * This function implements struct drm_simple_display_funcs.destroy_plane_state
467  * for shadow-buffered planes. It expects that mappings of shadow buffers
468  * have been released already.
469  */
470 void drm_gem_simple_kms_destroy_shadow_plane_state(struct drm_simple_display_pipe *pipe,
471 						   struct drm_plane_state *plane_state)
472 {
473 	drm_gem_destroy_shadow_plane_state(&pipe->plane, plane_state);
474 }
475 EXPORT_SYMBOL(drm_gem_simple_kms_destroy_shadow_plane_state);
476