xref: /linux/include/uapi/drm/pvr_drm.h (revision 0ea5c948cb64bab5bc7a5516774eb8536f05aa0d)
1 /* SPDX-License-Identifier: (GPL-2.0-only WITH Linux-syscall-note) OR MIT */
2 /* Copyright (c) 2023 Imagination Technologies Ltd. */
3 
4 #ifndef PVR_DRM_UAPI_H
5 #define PVR_DRM_UAPI_H
6 
7 #include "drm.h"
8 
9 #include <linux/const.h>
10 #include <linux/types.h>
11 
12 #if defined(__cplusplus)
13 extern "C" {
14 #endif
15 
16 /**
17  * DOC: PowerVR UAPI
18  *
19  * The PowerVR IOCTL argument structs have a few limitations in place, in
20  * addition to the standard kernel restrictions:
21  *
22  *  - All members must be type-aligned.
23  *  - The overall struct must be padded to 64-bit alignment.
24  *  - Explicit padding is almost always required. This takes the form of
25  *    ``_padding_[x]`` members of sufficient size to pad to the next power-of-two
26  *    alignment, where [x] is the offset into the struct in hexadecimal. Arrays
27  *    are never used for alignment. Padding fields must be zeroed; this is
28  *    always checked.
29  *  - Unions may only appear as the last member of a struct.
30  *  - Individual union members may grow in the future. The space between the
31  *    end of a union member and the end of its containing union is considered
32  *    "implicit padding" and must be zeroed. This is always checked.
33  *
34  * In addition to the IOCTL argument structs, the PowerVR UAPI makes use of
35  * DEV_QUERY argument structs. These are used to fetch information about the
36  * device and runtime. These structs are subject to the same rules set out
37  * above.
38  */
39 
40 /**
41  * struct drm_pvr_obj_array - Container used to pass arrays of objects
42  *
43  * It is not unusual to have to extend objects to pass new parameters, and the DRM
44  * ioctl infrastructure is supporting that by padding ioctl arguments with zeros
45  * when the data passed by userspace is smaller than the struct defined in the
46  * drm_ioctl_desc, thus keeping things backward compatible. This type is just
47  * applying the same concepts to indirect objects passed through arrays referenced
48  * from the main ioctl arguments structure: the stride basically defines the size
49  * of the object passed by userspace, which allows the kernel driver to pad with
50  * zeros when it's smaller than the size of the object it expects.
51  *
52  * Use ``DRM_PVR_OBJ_ARRAY()`` to fill object array fields, unless you
53  * have a very good reason not to.
54  */
55 struct drm_pvr_obj_array {
56 	/** @stride: Stride of object struct. Used for versioning. */
57 	__u32 stride;
58 
59 	/** @count: Number of objects in the array. */
60 	__u32 count;
61 
62 	/** @array: User pointer to an array of objects. */
63 	__u64 array;
64 };
65 
66 /**
67  * DRM_PVR_OBJ_ARRAY() - Helper macro for filling &struct drm_pvr_obj_array.
68  * @cnt: Number of elements pointed to py @ptr.
69  * @ptr: Pointer to start of a C array.
70  *
71  * Return: Literal of type &struct drm_pvr_obj_array.
72  */
73 #define DRM_PVR_OBJ_ARRAY(cnt, ptr) \
74 	{ .stride = sizeof((ptr)[0]), .count = (cnt), .array = (__u64)(uintptr_t)(ptr) }
75 
76 /**
77  * DOC: PowerVR IOCTL interface
78  */
79 
80 /**
81  * PVR_IOCTL() - Build a PowerVR IOCTL number
82  * @_ioctl: An incrementing id for this IOCTL. Added to %DRM_COMMAND_BASE.
83  * @_mode: Must be one of %DRM_IOR, %DRM_IOW or %DRM_IOWR.
84  * @_data: The type of the args struct passed by this IOCTL.
85  *
86  * The struct referred to by @_data must have a ``drm_pvr_ioctl_`` prefix and an
87  * ``_args suffix``. They are therefore omitted from @_data.
88  *
89  * This should only be used to build the constants described below; it should
90  * never be used to call an IOCTL directly.
91  *
92  * Return: An IOCTL number to be passed to ioctl() from userspace.
93  */
94 #define PVR_IOCTL(_ioctl, _mode, _data) \
95 	_mode(DRM_COMMAND_BASE + (_ioctl), struct drm_pvr_ioctl_##_data##_args)
96 
97 #define DRM_IOCTL_PVR_DEV_QUERY PVR_IOCTL(0x00, DRM_IOWR, dev_query)
98 #define DRM_IOCTL_PVR_CREATE_BO PVR_IOCTL(0x01, DRM_IOWR, create_bo)
99 #define DRM_IOCTL_PVR_GET_BO_MMAP_OFFSET PVR_IOCTL(0x02, DRM_IOWR, get_bo_mmap_offset)
100 #define DRM_IOCTL_PVR_CREATE_VM_CONTEXT PVR_IOCTL(0x03, DRM_IOWR, create_vm_context)
101 #define DRM_IOCTL_PVR_DESTROY_VM_CONTEXT PVR_IOCTL(0x04, DRM_IOW, destroy_vm_context)
102 #define DRM_IOCTL_PVR_VM_MAP PVR_IOCTL(0x05, DRM_IOW, vm_map)
103 #define DRM_IOCTL_PVR_VM_UNMAP PVR_IOCTL(0x06, DRM_IOW, vm_unmap)
104 #define DRM_IOCTL_PVR_CREATE_CONTEXT PVR_IOCTL(0x07, DRM_IOWR, create_context)
105 #define DRM_IOCTL_PVR_DESTROY_CONTEXT PVR_IOCTL(0x08, DRM_IOW, destroy_context)
106 #define DRM_IOCTL_PVR_CREATE_FREE_LIST PVR_IOCTL(0x09, DRM_IOWR, create_free_list)
107 #define DRM_IOCTL_PVR_DESTROY_FREE_LIST PVR_IOCTL(0x0a, DRM_IOW, destroy_free_list)
108 #define DRM_IOCTL_PVR_CREATE_HWRT_DATASET PVR_IOCTL(0x0b, DRM_IOWR, create_hwrt_dataset)
109 #define DRM_IOCTL_PVR_DESTROY_HWRT_DATASET PVR_IOCTL(0x0c, DRM_IOW, destroy_hwrt_dataset)
110 #define DRM_IOCTL_PVR_SUBMIT_JOBS PVR_IOCTL(0x0d, DRM_IOW, submit_jobs)
111 
112 /**
113  * DOC: PowerVR IOCTL DEV_QUERY interface
114  */
115 
116 /**
117  * struct drm_pvr_dev_query_gpu_info - Container used to fetch information about
118  * the graphics processor.
119  *
120  * When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set
121  * to %DRM_PVR_DEV_QUERY_GPU_INFO_GET.
122  */
123 struct drm_pvr_dev_query_gpu_info {
124 	/**
125 	 * @gpu_id: GPU identifier.
126 	 *
127 	 * For all currently supported GPUs this is the BVNC encoded as a 64-bit
128 	 * value as follows:
129 	 *
130 	 *    +--------+--------+--------+-------+
131 	 *    | 63..48 | 47..32 | 31..16 | 15..0 |
132 	 *    +========+========+========+=======+
133 	 *    | B      | V      | N      | C     |
134 	 *    +--------+--------+--------+-------+
135 	 */
136 	__u64 gpu_id;
137 
138 	/**
139 	 * @num_phantoms: Number of Phantoms present.
140 	 */
141 	__u32 num_phantoms;
142 
143 	/** @_padding_c: Reserved. This field must be zeroed. */
144 	__u32 _padding_c;
145 };
146 
147 /**
148  * struct drm_pvr_dev_query_runtime_info - Container used to fetch information
149  * about the graphics runtime.
150  *
151  * When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set
152  * to %DRM_PVR_DEV_QUERY_RUNTIME_INFO_GET.
153  */
154 struct drm_pvr_dev_query_runtime_info {
155 	/**
156 	 * @free_list_min_pages: Minimum allowed free list size,
157 	 * in PM physical pages.
158 	 */
159 	__u64 free_list_min_pages;
160 
161 	/**
162 	 * @free_list_max_pages: Maximum allowed free list size,
163 	 * in PM physical pages.
164 	 */
165 	__u64 free_list_max_pages;
166 
167 	/**
168 	 * @common_store_alloc_region_size: Size of the Allocation
169 	 * Region within the Common Store used for coefficient and shared
170 	 * registers, in dwords.
171 	 */
172 	__u32 common_store_alloc_region_size;
173 
174 	/**
175 	 * @common_store_partition_space_size: Size of the
176 	 * Partition Space within the Common Store for output buffers, in
177 	 * dwords.
178 	 */
179 	__u32 common_store_partition_space_size;
180 
181 	/**
182 	 * @max_coeffs: Maximum coefficients, in dwords.
183 	 */
184 	__u32 max_coeffs;
185 
186 	/**
187 	 * @cdm_max_local_mem_size_regs: Maximum amount of local
188 	 * memory available to a compute kernel, in dwords.
189 	 */
190 	__u32 cdm_max_local_mem_size_regs;
191 };
192 
193 /**
194  * struct drm_pvr_dev_query_quirks - Container used to fetch information about
195  * hardware fixes for which the device may require support in the user mode
196  * driver.
197  *
198  * When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set
199  * to %DRM_PVR_DEV_QUERY_QUIRKS_GET.
200  */
201 struct drm_pvr_dev_query_quirks {
202 	/**
203 	 * @quirks: A userspace address for the hardware quirks __u32 array.
204 	 *
205 	 * The first @musthave_count items in the list are quirks that the
206 	 * client must support for this device. If userspace does not support
207 	 * all these quirks then functionality is not guaranteed and client
208 	 * initialisation must fail.
209 	 * The remaining quirks in the list affect userspace and the kernel or
210 	 * firmware. They are disabled by default and require userspace to
211 	 * opt-in. The opt-in mechanism depends on the quirk.
212 	 */
213 	__u64 quirks;
214 
215 	/** @count: Length of @quirks (number of __u32). */
216 	__u16 count;
217 
218 	/**
219 	 * @musthave_count: The number of entries in @quirks that are
220 	 * mandatory, starting at index 0.
221 	 */
222 	__u16 musthave_count;
223 
224 	/** @_padding_c: Reserved. This field must be zeroed. */
225 	__u32 _padding_c;
226 };
227 
228 /**
229  * struct drm_pvr_dev_query_enhancements - Container used to fetch information
230  * about optional enhancements supported by the device that require support in
231  * the user mode driver.
232  *
233  * When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set
234  * to %DRM_PVR_DEV_ENHANCEMENTS_GET.
235  */
236 struct drm_pvr_dev_query_enhancements {
237 	/**
238 	 * @enhancements: A userspace address for the hardware enhancements
239 	 * __u32 array.
240 	 *
241 	 * These enhancements affect userspace and the kernel or firmware. They
242 	 * are disabled by default and require userspace to opt-in. The opt-in
243 	 * mechanism depends on the enhancement.
244 	 */
245 	__u64 enhancements;
246 
247 	/** @count: Length of @enhancements (number of __u32). */
248 	__u16 count;
249 
250 	/** @_padding_a: Reserved. This field must be zeroed. */
251 	__u16 _padding_a;
252 
253 	/** @_padding_c: Reserved. This field must be zeroed. */
254 	__u32 _padding_c;
255 };
256 
257 /**
258  * enum drm_pvr_heap_id - Array index for heap info data returned by
259  * %DRM_PVR_DEV_QUERY_HEAP_INFO_GET.
260  *
261  * For compatibility reasons all indices will be present in the returned array,
262  * however some heaps may not be present. These are indicated where
263  * &struct drm_pvr_heap.size is set to zero.
264  */
265 enum drm_pvr_heap_id {
266 	/** @DRM_PVR_HEAP_GENERAL: General purpose heap. */
267 	DRM_PVR_HEAP_GENERAL = 0,
268 	/** @DRM_PVR_HEAP_PDS_CODE_DATA: PDS code and data heap. */
269 	DRM_PVR_HEAP_PDS_CODE_DATA,
270 	/** @DRM_PVR_HEAP_USC_CODE: USC code heap. */
271 	DRM_PVR_HEAP_USC_CODE,
272 	/** @DRM_PVR_HEAP_RGNHDR: Region header heap. Only used if GPU has BRN63142. */
273 	DRM_PVR_HEAP_RGNHDR,
274 	/** @DRM_PVR_HEAP_VIS_TEST: Visibility test heap. */
275 	DRM_PVR_HEAP_VIS_TEST,
276 	/** @DRM_PVR_HEAP_TRANSFER_FRAG: Transfer fragment heap. */
277 	DRM_PVR_HEAP_TRANSFER_FRAG,
278 
279 	/**
280 	 * @DRM_PVR_HEAP_COUNT: The number of heaps returned by
281 	 * %DRM_PVR_DEV_QUERY_HEAP_INFO_GET.
282 	 *
283 	 * More heaps may be added, so this also serves as the copy limit when
284 	 * sent by the caller.
285 	 */
286 	DRM_PVR_HEAP_COUNT
287 	/* Please only add additional heaps above DRM_PVR_HEAP_COUNT! */
288 };
289 
290 /**
291  * struct drm_pvr_heap - Container holding information about a single heap.
292  *
293  * This will always be fetched as an array.
294  */
295 struct drm_pvr_heap {
296 	/** @base: Base address of heap. */
297 	__u64 base;
298 
299 	/** @size: Size of heap, in bytes. Will be 0 if the heap is not present. */
300 	__u64 size;
301 
302 	/** @flags: Flags for this heap. Currently always 0. */
303 	__u32 flags;
304 
305 	/** @page_size_log2: Log2 of page size. */
306 	__u32 page_size_log2;
307 };
308 
309 /**
310  * struct drm_pvr_dev_query_heap_info - Container used to fetch information
311  * about heaps supported by the device driver.
312  *
313  * Please note all driver-supported heaps will be returned up to &heaps.count.
314  * Some heaps will not be present in all devices, which will be indicated by
315  * &struct drm_pvr_heap.size being set to zero.
316  *
317  * When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set
318  * to %DRM_PVR_DEV_QUERY_HEAP_INFO_GET.
319  */
320 struct drm_pvr_dev_query_heap_info {
321 	/**
322 	 * @heaps: Array of &struct drm_pvr_heap. If pointer is NULL, the count
323 	 * and stride will be updated with those known to the driver version, to
324 	 * facilitate allocation by the caller.
325 	 */
326 	struct drm_pvr_obj_array heaps;
327 };
328 
329 /**
330  * enum drm_pvr_static_data_area_usage - Array index for static data area info
331  * returned by %DRM_PVR_DEV_QUERY_STATIC_DATA_AREAS_GET.
332  *
333  * For compatibility reasons all indices will be present in the returned array,
334  * however some areas may not be present. These are indicated where
335  * &struct drm_pvr_static_data_area.size is set to zero.
336  */
337 enum drm_pvr_static_data_area_usage {
338 	/**
339 	 * @DRM_PVR_STATIC_DATA_AREA_EOT: End of Tile PDS program code segment.
340 	 *
341 	 * The End of Tile PDS task runs at completion of a tile during a fragment job, and is
342 	 * responsible for emitting the tile to the Pixel Back End.
343 	 */
344 	DRM_PVR_STATIC_DATA_AREA_EOT = 0,
345 
346 	/**
347 	 * @DRM_PVR_STATIC_DATA_AREA_FENCE: MCU fence area, used during cache flush and
348 	 * invalidation.
349 	 *
350 	 * This must point to valid physical memory but the contents otherwise are not used.
351 	 */
352 	DRM_PVR_STATIC_DATA_AREA_FENCE,
353 
354 	/**
355 	 * @DRM_PVR_STATIC_DATA_AREA_VDM_SYNC: VDM sync program.
356 	 *
357 	 * The VDM sync program is used to synchronise multiple areas of the GPU hardware.
358 	 */
359 	DRM_PVR_STATIC_DATA_AREA_VDM_SYNC,
360 
361 	/**
362 	 * @DRM_PVR_STATIC_DATA_AREA_YUV_CSC: YUV coefficients.
363 	 *
364 	 * Area contains up to 16 slots with stride of 64 bytes. Each is a 3x4 matrix of u16 fixed
365 	 * point numbers, with 1 sign bit, 2 integer bits and 13 fractional bits.
366 	 *
367 	 * The slots are :
368 	 * 0 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_RGB_IDENTITY_KHR
369 	 * 1 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_IDENTITY_KHR (full range)
370 	 * 2 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_IDENTITY_KHR (conformant range)
371 	 * 3 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_709_KHR (full range)
372 	 * 4 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_709_KHR (conformant range)
373 	 * 5 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_601_KHR (full range)
374 	 * 6 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_601_KHR (conformant range)
375 	 * 7 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_2020_KHR (full range)
376 	 * 8 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_2020_KHR (conformant range)
377 	 * 9 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_601_KHR (conformant range, 10 bit)
378 	 * 10 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_709_KHR (conformant range, 10 bit)
379 	 * 11 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_2020_KHR (conformant range, 10 bit)
380 	 * 14 = Identity (biased)
381 	 * 15 = Identity
382 	 */
383 	DRM_PVR_STATIC_DATA_AREA_YUV_CSC,
384 };
385 
386 /**
387  * struct drm_pvr_static_data_area - Container holding information about a
388  * single static data area.
389  *
390  * This will always be fetched as an array.
391  */
392 struct drm_pvr_static_data_area {
393 	/**
394 	 * @area_usage: Usage of static data area.
395 	 * See &enum drm_pvr_static_data_area_usage.
396 	 */
397 	__u16 area_usage;
398 
399 	/**
400 	 * @location_heap_id: Array index of heap where this of static data
401 	 * area is located. This array is fetched using
402 	 * %DRM_PVR_DEV_QUERY_HEAP_INFO_GET.
403 	 */
404 	__u16 location_heap_id;
405 
406 	/** @size: Size of static data area. Not present if set to zero. */
407 	__u32 size;
408 
409 	/** @offset: Offset of static data area from start of heap. */
410 	__u64 offset;
411 };
412 
413 /**
414  * struct drm_pvr_dev_query_static_data_areas - Container used to fetch
415  * information about the static data areas in heaps supported by the device
416  * driver.
417  *
418  * Please note all driver-supported static data areas will be returned up to
419  * &static_data_areas.count. Some will not be present for all devices which,
420  * will be indicated by &struct drm_pvr_static_data_area.size being set to zero.
421  *
422  * Further, some heaps will not be present either. See &struct
423  * drm_pvr_dev_query_heap_info.
424  *
425  * When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set
426  * to %DRM_PVR_DEV_QUERY_STATIC_DATA_AREAS_GET.
427  */
428 struct drm_pvr_dev_query_static_data_areas {
429 	/**
430 	 * @static_data_areas: Array of &struct drm_pvr_static_data_area. If
431 	 * pointer is NULL, the count and stride will be updated with those
432 	 * known to the driver version, to facilitate allocation by the caller.
433 	 */
434 	struct drm_pvr_obj_array static_data_areas;
435 };
436 
437 /**
438  * enum drm_pvr_dev_query - For use with &drm_pvr_ioctl_dev_query_args.type to
439  * indicate the type of the receiving container.
440  *
441  * Append only. Do not reorder.
442  */
443 enum drm_pvr_dev_query {
444 	/**
445 	 * @DRM_PVR_DEV_QUERY_GPU_INFO_GET: The dev query args contain a pointer
446 	 * to &struct drm_pvr_dev_query_gpu_info.
447 	 */
448 	DRM_PVR_DEV_QUERY_GPU_INFO_GET = 0,
449 
450 	/**
451 	 * @DRM_PVR_DEV_QUERY_RUNTIME_INFO_GET: The dev query args contain a
452 	 * pointer to &struct drm_pvr_dev_query_runtime_info.
453 	 */
454 	DRM_PVR_DEV_QUERY_RUNTIME_INFO_GET,
455 
456 	/**
457 	 * @DRM_PVR_DEV_QUERY_QUIRKS_GET: The dev query args contain a pointer
458 	 * to &struct drm_pvr_dev_query_quirks.
459 	 */
460 	DRM_PVR_DEV_QUERY_QUIRKS_GET,
461 
462 	/**
463 	 * @DRM_PVR_DEV_QUERY_ENHANCEMENTS_GET: The dev query args contain a
464 	 * pointer to &struct drm_pvr_dev_query_enhancements.
465 	 */
466 	DRM_PVR_DEV_QUERY_ENHANCEMENTS_GET,
467 
468 	/**
469 	 * @DRM_PVR_DEV_QUERY_HEAP_INFO_GET: The dev query args contain a
470 	 * pointer to &struct drm_pvr_dev_query_heap_info.
471 	 */
472 	DRM_PVR_DEV_QUERY_HEAP_INFO_GET,
473 
474 	/**
475 	 * @DRM_PVR_DEV_QUERY_STATIC_DATA_AREAS_GET: The dev query args contain
476 	 * a pointer to &struct drm_pvr_dev_query_static_data_areas.
477 	 */
478 	DRM_PVR_DEV_QUERY_STATIC_DATA_AREAS_GET,
479 };
480 
481 /**
482  * struct drm_pvr_ioctl_dev_query_args - Arguments for %DRM_IOCTL_PVR_DEV_QUERY.
483  */
484 struct drm_pvr_ioctl_dev_query_args {
485 	/**
486 	 * @type: Type of query and output struct. See &enum drm_pvr_dev_query.
487 	 */
488 	__u32 type;
489 
490 	/**
491 	 * @size: Size of the receiving struct, see @type.
492 	 *
493 	 * After a successful call this will be updated to the written byte
494 	 * length.
495 	 * Can also be used to get the minimum byte length (see @pointer).
496 	 * This allows additional fields to be appended to the structs in
497 	 * future.
498 	 */
499 	__u32 size;
500 
501 	/**
502 	 * @pointer: Pointer to struct @type.
503 	 *
504 	 * Must be large enough to contain @size bytes.
505 	 * If pointer is NULL, the expected size will be returned in the @size
506 	 * field, but no other data will be written.
507 	 */
508 	__u64 pointer;
509 };
510 
511 /**
512  * DOC: PowerVR IOCTL CREATE_BO interface
513  */
514 
515 /**
516  * DOC: Flags for CREATE_BO
517  *
518  * We use "device" to refer to the GPU here because of the ambiguity between CPU and GPU in some
519  * fonts.
520  *
521  * Device mapping options
522  *    :DRM_PVR_BO_BYPASS_DEVICE_CACHE: Specify that device accesses to this memory will bypass the
523  *       cache. This is used for buffers that will either be regularly updated by the CPU (eg free
524  *       lists) or will be accessed only once and therefore isn't worth caching (eg partial render
525  *       buffers).
526  *       By default, the device flushes its memory caches after every job, so this is not normally
527  *       required for coherency.
528  *    :DRM_PVR_BO_PM_FW_PROTECT: Specify that only the Parameter Manager (PM) and/or firmware
529  *       processor should be allowed to access this memory when mapped to the device. It is not
530  *       valid to specify this flag with DRM_PVR_BO_ALLOW_CPU_USERSPACE_ACCESS.
531  *
532  * CPU mapping options
533  *    :DRM_PVR_BO_ALLOW_CPU_USERSPACE_ACCESS: Allow userspace to map and access the contents of this
534  *       memory. It is not valid to specify this flag with DRM_PVR_BO_PM_FW_PROTECT.
535  */
536 #define DRM_PVR_BO_BYPASS_DEVICE_CACHE _BITULL(0)
537 #define DRM_PVR_BO_PM_FW_PROTECT _BITULL(1)
538 #define DRM_PVR_BO_ALLOW_CPU_USERSPACE_ACCESS _BITULL(2)
539 /* Bits 3..63 are reserved. */
540 
541 #define DRM_PVR_BO_FLAGS_MASK (DRM_PVR_BO_BYPASS_DEVICE_CACHE | DRM_PVR_BO_PM_FW_PROTECT | \
542 			       DRM_PVR_BO_ALLOW_CPU_USERSPACE_ACCESS)
543 
544 /**
545  * struct drm_pvr_ioctl_create_bo_args - Arguments for %DRM_IOCTL_PVR_CREATE_BO
546  */
547 struct drm_pvr_ioctl_create_bo_args {
548 	/**
549 	 * @size: [IN] Size of buffer object to create. This must be page size
550 	 * aligned.
551 	 */
552 	__u64 size;
553 
554 	/**
555 	 * @handle: [OUT] GEM handle of the new buffer object for use in
556 	 * userspace.
557 	 */
558 	__u32 handle;
559 
560 	/** @_padding_c: Reserved. This field must be zeroed. */
561 	__u32 _padding_c;
562 
563 	/**
564 	 * @flags: [IN] Options which will affect the behaviour of this
565 	 * creation operation and future mapping operations on the created
566 	 * object. This field must be a valid combination of ``DRM_PVR_BO_*``
567 	 * values, with all bits marked as reserved set to zero.
568 	 */
569 	__u64 flags;
570 };
571 
572 /**
573  * DOC: PowerVR IOCTL GET_BO_MMAP_OFFSET interface
574  */
575 
576 /**
577  * struct drm_pvr_ioctl_get_bo_mmap_offset_args - Arguments for
578  * %DRM_IOCTL_PVR_GET_BO_MMAP_OFFSET
579  *
580  * Like other DRM drivers, the "mmap" IOCTL doesn't actually map any memory.
581  * Instead, it allocates a fake offset which refers to the specified buffer
582  * object. This offset can be used with a real mmap call on the DRM device
583  * itself.
584  */
585 struct drm_pvr_ioctl_get_bo_mmap_offset_args {
586 	/** @handle: [IN] GEM handle of the buffer object to be mapped. */
587 	__u32 handle;
588 
589 	/** @_padding_4: Reserved. This field must be zeroed. */
590 	__u32 _padding_4;
591 
592 	/** @offset: [OUT] Fake offset to use in the real mmap call. */
593 	__u64 offset;
594 };
595 
596 /**
597  * DOC: PowerVR IOCTL CREATE_VM_CONTEXT and DESTROY_VM_CONTEXT interfaces
598  */
599 
600 /**
601  * struct drm_pvr_ioctl_create_vm_context_args - Arguments for
602  * %DRM_IOCTL_PVR_CREATE_VM_CONTEXT
603  */
604 struct drm_pvr_ioctl_create_vm_context_args {
605 	/** @handle: [OUT] Handle for new VM context. */
606 	__u32 handle;
607 
608 	/** @_padding_4: Reserved. This field must be zeroed. */
609 	__u32 _padding_4;
610 };
611 
612 /**
613  * struct drm_pvr_ioctl_destroy_vm_context_args - Arguments for
614  * %DRM_IOCTL_PVR_DESTROY_VM_CONTEXT
615  */
616 struct drm_pvr_ioctl_destroy_vm_context_args {
617 	/**
618 	 * @handle: [IN] Handle for VM context to be destroyed.
619 	 */
620 	__u32 handle;
621 
622 	/** @_padding_4: Reserved. This field must be zeroed. */
623 	__u32 _padding_4;
624 };
625 
626 /**
627  * DOC: PowerVR IOCTL VM_MAP and VM_UNMAP interfaces
628  *
629  * The VM UAPI allows userspace to create buffer object mappings in GPU virtual address space.
630  *
631  * The client is responsible for managing GPU address space. It should allocate mappings within
632  * the heaps returned by %DRM_PVR_DEV_QUERY_HEAP_INFO_GET.
633  *
634  * %DRM_IOCTL_PVR_VM_MAP creates a new mapping. The client provides the target virtual address for
635  * the mapping. Size and offset within the mapped buffer object can be specified, so the client can
636  * partially map a buffer.
637  *
638  * %DRM_IOCTL_PVR_VM_UNMAP removes a mapping. The entire mapping will be removed from GPU address
639  * space only if the size of the mapping matches that known to the driver.
640  */
641 
642 /**
643  * struct drm_pvr_ioctl_vm_map_args - Arguments for %DRM_IOCTL_PVR_VM_MAP.
644  */
645 struct drm_pvr_ioctl_vm_map_args {
646 	/**
647 	 * @vm_context_handle: [IN] Handle for VM context for this mapping to
648 	 * exist in.
649 	 */
650 	__u32 vm_context_handle;
651 
652 	/** @flags: [IN] Flags which affect this mapping. Currently always 0. */
653 	__u32 flags;
654 
655 	/**
656 	 * @device_addr: [IN] Requested device-virtual address for the mapping.
657 	 * This must be non-zero and aligned to the device page size for the
658 	 * heap containing the requested address. It is an error to specify an
659 	 * address which is not contained within one of the heaps returned by
660 	 * %DRM_PVR_DEV_QUERY_HEAP_INFO_GET.
661 	 */
662 	__u64 device_addr;
663 
664 	/**
665 	 * @handle: [IN] Handle of the target buffer object. This must be a
666 	 * valid handle returned by %DRM_IOCTL_PVR_CREATE_BO.
667 	 */
668 	__u32 handle;
669 
670 	/** @_padding_14: Reserved. This field must be zeroed. */
671 	__u32 _padding_14;
672 
673 	/**
674 	 * @offset: [IN] Offset into the target bo from which to begin the
675 	 * mapping.
676 	 */
677 	__u64 offset;
678 
679 	/**
680 	 * @size: [IN] Size of the requested mapping. Must be aligned to
681 	 * the device page size for the heap containing the requested address,
682 	 * as well as the host page size. When added to @device_addr, the
683 	 * result must not overflow the heap which contains @device_addr (i.e.
684 	 * the range specified by @device_addr and @size must be completely
685 	 * contained within a single heap specified by
686 	 * %DRM_PVR_DEV_QUERY_HEAP_INFO_GET).
687 	 */
688 	__u64 size;
689 };
690 
691 /**
692  * struct drm_pvr_ioctl_vm_unmap_args - Arguments for %DRM_IOCTL_PVR_VM_UNMAP.
693  */
694 struct drm_pvr_ioctl_vm_unmap_args {
695 	/**
696 	 * @vm_context_handle: [IN] Handle for VM context that this mapping
697 	 * exists in.
698 	 */
699 	__u32 vm_context_handle;
700 
701 	/** @_padding_4: Reserved. This field must be zeroed. */
702 	__u32 _padding_4;
703 
704 	/**
705 	 * @device_addr: [IN] Device-virtual address at the start of the target
706 	 * mapping. This must be non-zero.
707 	 */
708 	__u64 device_addr;
709 
710 	/**
711 	 * @size: Size in bytes of the target mapping. This must be non-zero.
712 	 */
713 	__u64 size;
714 };
715 
716 /**
717  * DOC: PowerVR IOCTL CREATE_CONTEXT and DESTROY_CONTEXT interfaces
718  */
719 
720 /**
721  * enum drm_pvr_ctx_priority - Arguments for
722  * &drm_pvr_ioctl_create_context_args.priority
723  */
724 enum drm_pvr_ctx_priority {
725 	/** @DRM_PVR_CTX_PRIORITY_LOW: Priority below normal. */
726 	DRM_PVR_CTX_PRIORITY_LOW = -512,
727 
728 	/** @DRM_PVR_CTX_PRIORITY_NORMAL: Normal priority. */
729 	DRM_PVR_CTX_PRIORITY_NORMAL = 0,
730 
731 	/**
732 	 * @DRM_PVR_CTX_PRIORITY_HIGH: Priority above normal.
733 	 * Note this requires ``CAP_SYS_NICE`` or ``DRM_MASTER``.
734 	 */
735 	DRM_PVR_CTX_PRIORITY_HIGH = 512,
736 };
737 
738 /**
739  * enum drm_pvr_ctx_type - Arguments for
740  * &struct drm_pvr_ioctl_create_context_args.type
741  */
742 enum drm_pvr_ctx_type {
743 	/**
744 	 * @DRM_PVR_CTX_TYPE_RENDER: Render context.
745 	 */
746 	DRM_PVR_CTX_TYPE_RENDER = 0,
747 
748 	/**
749 	 * @DRM_PVR_CTX_TYPE_COMPUTE: Compute context.
750 	 */
751 	DRM_PVR_CTX_TYPE_COMPUTE,
752 
753 	/**
754 	 * @DRM_PVR_CTX_TYPE_TRANSFER_FRAG: Transfer context for fragment data
755 	 * master.
756 	 */
757 	DRM_PVR_CTX_TYPE_TRANSFER_FRAG,
758 };
759 
760 /**
761  * struct drm_pvr_ioctl_create_context_args - Arguments for
762  * %DRM_IOCTL_PVR_CREATE_CONTEXT
763  */
764 struct drm_pvr_ioctl_create_context_args {
765 	/**
766 	 * @type: [IN] Type of context to create.
767 	 *
768 	 * This must be one of the values defined by &enum drm_pvr_ctx_type.
769 	 */
770 	__u32 type;
771 
772 	/** @flags: [IN] Flags for context. */
773 	__u32 flags;
774 
775 	/**
776 	 * @priority: [IN] Priority of new context.
777 	 *
778 	 * This must be one of the values defined by &enum drm_pvr_ctx_priority.
779 	 */
780 	__s32 priority;
781 
782 	/** @handle: [OUT] Handle for new context. */
783 	__u32 handle;
784 
785 	/**
786 	 * @static_context_state: [IN] Pointer to static context state stream.
787 	 */
788 	__u64 static_context_state;
789 
790 	/**
791 	 * @static_context_state_len: [IN] Length of static context state, in bytes.
792 	 */
793 	__u32 static_context_state_len;
794 
795 	/**
796 	 * @vm_context_handle: [IN] Handle for VM context that this context is
797 	 * associated with.
798 	 */
799 	__u32 vm_context_handle;
800 
801 	/**
802 	 * @callstack_addr: [IN] Address for initial call stack pointer. Only valid
803 	 * if @type is %DRM_PVR_CTX_TYPE_RENDER, otherwise must be 0.
804 	 */
805 	__u64 callstack_addr;
806 };
807 
808 /**
809  * struct drm_pvr_ioctl_destroy_context_args - Arguments for
810  * %DRM_IOCTL_PVR_DESTROY_CONTEXT
811  */
812 struct drm_pvr_ioctl_destroy_context_args {
813 	/**
814 	 * @handle: [IN] Handle for context to be destroyed.
815 	 */
816 	__u32 handle;
817 
818 	/** @_padding_4: Reserved. This field must be zeroed. */
819 	__u32 _padding_4;
820 };
821 
822 /**
823  * DOC: PowerVR IOCTL CREATE_FREE_LIST and DESTROY_FREE_LIST interfaces
824  */
825 
826 /**
827  * struct drm_pvr_ioctl_create_free_list_args - Arguments for
828  * %DRM_IOCTL_PVR_CREATE_FREE_LIST
829  *
830  * Free list arguments have the following constraints :
831  *
832  * - @max_num_pages must be greater than zero.
833  * - @grow_threshold must be between 0 and 100.
834  * - @grow_num_pages must be less than or equal to &max_num_pages.
835  * - @initial_num_pages, @max_num_pages and @grow_num_pages must be multiples
836  *   of 4.
837  * - When &grow_num_pages is 0, @initial_num_pages must be equal to
838  *   @max_num_pages.
839  * - When &grow_num_pages is non-zero, @initial_num_pages must be less than
840  *   @max_num_pages.
841  */
842 struct drm_pvr_ioctl_create_free_list_args {
843 	/**
844 	 * @free_list_gpu_addr: [IN] Address of GPU mapping of buffer object
845 	 * containing memory to be used by free list.
846 	 *
847 	 * The mapped region of the buffer object must be at least
848 	 * @max_num_pages * ``sizeof(__u32)``.
849 	 *
850 	 * The buffer object must have been created with
851 	 * %DRM_PVR_BO_DEVICE_PM_FW_PROTECT set and
852 	 * %DRM_PVR_BO_CPU_ALLOW_USERSPACE_ACCESS not set.
853 	 */
854 	__u64 free_list_gpu_addr;
855 
856 	/** @initial_num_pages: [IN] Pages initially allocated to free list. */
857 	__u32 initial_num_pages;
858 
859 	/** @max_num_pages: [IN] Maximum number of pages in free list. */
860 	__u32 max_num_pages;
861 
862 	/** @grow_num_pages: [IN] Pages to grow free list by per request. */
863 	__u32 grow_num_pages;
864 
865 	/**
866 	 * @grow_threshold: [IN] Percentage of FL memory used that should
867 	 * trigger a new grow request.
868 	 */
869 	__u32 grow_threshold;
870 
871 	/**
872 	 * @vm_context_handle: [IN] Handle for VM context that the free list buffer
873 	 * object is mapped in.
874 	 */
875 	__u32 vm_context_handle;
876 
877 	/**
878 	 * @handle: [OUT] Handle for created free list.
879 	 */
880 	__u32 handle;
881 };
882 
883 /**
884  * struct drm_pvr_ioctl_destroy_free_list_args - Arguments for
885  * %DRM_IOCTL_PVR_DESTROY_FREE_LIST
886  */
887 struct drm_pvr_ioctl_destroy_free_list_args {
888 	/**
889 	 * @handle: [IN] Handle for free list to be destroyed.
890 	 */
891 	__u32 handle;
892 
893 	/** @_padding_4: Reserved. This field must be zeroed. */
894 	__u32 _padding_4;
895 };
896 
897 /**
898  * DOC: PowerVR IOCTL CREATE_HWRT_DATASET and DESTROY_HWRT_DATASET interfaces
899  */
900 
901 /**
902  * struct drm_pvr_create_hwrt_geom_data_args - Geometry data arguments used for
903  * &struct drm_pvr_ioctl_create_hwrt_dataset_args.geom_data_args.
904  */
905 struct drm_pvr_create_hwrt_geom_data_args {
906 	/** @tpc_dev_addr: [IN] Tail pointer cache GPU virtual address. */
907 	__u64 tpc_dev_addr;
908 
909 	/** @tpc_size: [IN] Size of TPC, in bytes. */
910 	__u32 tpc_size;
911 
912 	/** @tpc_stride: [IN] Stride between layers in TPC, in pages */
913 	__u32 tpc_stride;
914 
915 	/** @vheap_table_dev_addr: [IN] VHEAP table GPU virtual address. */
916 	__u64 vheap_table_dev_addr;
917 
918 	/** @rtc_dev_addr: [IN] Render Target Cache virtual address. */
919 	__u64 rtc_dev_addr;
920 };
921 
922 /**
923  * struct drm_pvr_create_hwrt_rt_data_args - Render target arguments used for
924  * &struct drm_pvr_ioctl_create_hwrt_dataset_args.rt_data_args.
925  */
926 struct drm_pvr_create_hwrt_rt_data_args {
927 	/** @pm_mlist_dev_addr: [IN] PM MLIST GPU virtual address. */
928 	__u64 pm_mlist_dev_addr;
929 
930 	/** @macrotile_array_dev_addr: [IN] Macrotile array GPU virtual address. */
931 	__u64 macrotile_array_dev_addr;
932 
933 	/** @region_header_dev_addr: [IN] Region header array GPU virtual address. */
934 	__u64 region_header_dev_addr;
935 };
936 
937 #define PVR_DRM_HWRT_FREE_LIST_LOCAL 0
938 #define PVR_DRM_HWRT_FREE_LIST_GLOBAL 1U
939 
940 /**
941  * struct drm_pvr_ioctl_create_hwrt_dataset_args - Arguments for
942  * %DRM_IOCTL_PVR_CREATE_HWRT_DATASET
943  */
944 struct drm_pvr_ioctl_create_hwrt_dataset_args {
945 	/** @geom_data_args: [IN] Geometry data arguments. */
946 	struct drm_pvr_create_hwrt_geom_data_args geom_data_args;
947 
948 	/**
949 	 * @rt_data_args: [IN] Array of render target arguments.
950 	 *
951 	 * Each entry in this array represents a render target in a double buffered
952 	 * setup.
953 	 */
954 	struct drm_pvr_create_hwrt_rt_data_args rt_data_args[2];
955 
956 	/**
957 	 * @free_list_handles: [IN] Array of free list handles.
958 	 *
959 	 * free_list_handles[PVR_DRM_HWRT_FREE_LIST_LOCAL] must have initial
960 	 * size of at least that reported by
961 	 * &drm_pvr_dev_query_runtime_info.free_list_min_pages.
962 	 */
963 	__u32 free_list_handles[2];
964 
965 	/** @width: [IN] Width in pixels. */
966 	__u32 width;
967 
968 	/** @height: [IN] Height in pixels. */
969 	__u32 height;
970 
971 	/** @samples: [IN] Number of samples. */
972 	__u32 samples;
973 
974 	/** @layers: [IN] Number of layers. */
975 	__u32 layers;
976 
977 	/** @isp_merge_lower_x: [IN] Lower X coefficient for triangle merging. */
978 	__u32 isp_merge_lower_x;
979 
980 	/** @isp_merge_lower_y: [IN] Lower Y coefficient for triangle merging. */
981 	__u32 isp_merge_lower_y;
982 
983 	/** @isp_merge_scale_x: [IN] Scale X coefficient for triangle merging. */
984 	__u32 isp_merge_scale_x;
985 
986 	/** @isp_merge_scale_y: [IN] Scale Y coefficient for triangle merging. */
987 	__u32 isp_merge_scale_y;
988 
989 	/** @isp_merge_upper_x: [IN] Upper X coefficient for triangle merging. */
990 	__u32 isp_merge_upper_x;
991 
992 	/** @isp_merge_upper_y: [IN] Upper Y coefficient for triangle merging. */
993 	__u32 isp_merge_upper_y;
994 
995 	/**
996 	 * @region_header_size: [IN] Size of region header array. This common field is used by
997 	 * both render targets in this data set.
998 	 *
999 	 * The units for this field differ depending on what version of the simple internal
1000 	 * parameter format the device uses. If format 2 is in use then this is interpreted as the
1001 	 * number of region headers. For other formats it is interpreted as the size in dwords.
1002 	 */
1003 	__u32 region_header_size;
1004 
1005 	/**
1006 	 * @handle: [OUT] Handle for created HWRT dataset.
1007 	 */
1008 	__u32 handle;
1009 };
1010 
1011 /**
1012  * struct drm_pvr_ioctl_destroy_hwrt_dataset_args - Arguments for
1013  * %DRM_IOCTL_PVR_DESTROY_HWRT_DATASET
1014  */
1015 struct drm_pvr_ioctl_destroy_hwrt_dataset_args {
1016 	/**
1017 	 * @handle: [IN] Handle for HWRT dataset to be destroyed.
1018 	 */
1019 	__u32 handle;
1020 
1021 	/** @_padding_4: Reserved. This field must be zeroed. */
1022 	__u32 _padding_4;
1023 };
1024 
1025 /**
1026  * DOC: PowerVR IOCTL SUBMIT_JOBS interface
1027  */
1028 
1029 /**
1030  * DOC: Flags for the drm_pvr_sync_op object.
1031  *
1032  * .. c:macro:: DRM_PVR_SYNC_OP_HANDLE_TYPE_MASK
1033  *
1034  *    Handle type mask for the drm_pvr_sync_op::flags field.
1035  *
1036  * .. c:macro:: DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_SYNCOBJ
1037  *
1038  *    Indicates the handle passed in drm_pvr_sync_op::handle is a syncobj handle.
1039  *    This is the default type.
1040  *
1041  * .. c:macro:: DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_TIMELINE_SYNCOBJ
1042  *
1043  *    Indicates the handle passed in drm_pvr_sync_op::handle is a timeline syncobj handle.
1044  *
1045  * .. c:macro:: DRM_PVR_SYNC_OP_FLAG_SIGNAL
1046  *
1047  *    Signal operation requested. The out-fence bound to the job will be attached to
1048  *    the syncobj whose handle is passed in drm_pvr_sync_op::handle.
1049  *
1050  * .. c:macro:: DRM_PVR_SYNC_OP_FLAG_WAIT
1051  *
1052  *    Wait operation requested. The job will wait for this particular syncobj or syncobj
1053  *    point to be signaled before being started.
1054  *    This is the default operation.
1055  */
1056 #define DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_MASK 0xf
1057 #define DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_SYNCOBJ 0
1058 #define DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_TIMELINE_SYNCOBJ 1
1059 #define DRM_PVR_SYNC_OP_FLAG_SIGNAL _BITULL(31)
1060 #define DRM_PVR_SYNC_OP_FLAG_WAIT 0
1061 
1062 #define DRM_PVR_SYNC_OP_FLAGS_MASK (DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_MASK | \
1063 				    DRM_PVR_SYNC_OP_FLAG_SIGNAL)
1064 
1065 /**
1066  * struct drm_pvr_sync_op - Object describing a sync operation
1067  */
1068 struct drm_pvr_sync_op {
1069 	/** @handle: Handle of sync object. */
1070 	__u32 handle;
1071 
1072 	/** @flags: Combination of ``DRM_PVR_SYNC_OP_FLAG_`` flags. */
1073 	__u32 flags;
1074 
1075 	/** @value: Timeline value for this drm_syncobj. MBZ for a binary syncobj. */
1076 	__u64 value;
1077 };
1078 
1079 /**
1080  * DOC: Flags for SUBMIT_JOB ioctl geometry command.
1081  *
1082  * .. c:macro:: DRM_PVR_SUBMIT_JOB_GEOM_CMD_FIRST
1083  *
1084  *    Indicates if this the first command to be issued for a render.
1085  *
1086  * .. c:macro:: DRM_PVR_SUBMIT_JOB_GEOM_CMD_LAST
1087  *
1088  *    Indicates if this the last command to be issued for a render.
1089  *
1090  * .. c:macro:: DRM_PVR_SUBMIT_JOB_GEOM_CMD_SINGLE_CORE
1091  *
1092  *    Forces to use single core in a multi core device.
1093  *
1094  * .. c:macro:: DRM_PVR_SUBMIT_JOB_GEOM_CMD_FLAGS_MASK
1095  *
1096  *    Logical OR of all the geometry cmd flags.
1097  */
1098 #define DRM_PVR_SUBMIT_JOB_GEOM_CMD_FIRST _BITULL(0)
1099 #define DRM_PVR_SUBMIT_JOB_GEOM_CMD_LAST _BITULL(1)
1100 #define DRM_PVR_SUBMIT_JOB_GEOM_CMD_SINGLE_CORE _BITULL(2)
1101 #define DRM_PVR_SUBMIT_JOB_GEOM_CMD_FLAGS_MASK                                 \
1102 	(DRM_PVR_SUBMIT_JOB_GEOM_CMD_FIRST |                                   \
1103 	 DRM_PVR_SUBMIT_JOB_GEOM_CMD_LAST |                                    \
1104 	 DRM_PVR_SUBMIT_JOB_GEOM_CMD_SINGLE_CORE)
1105 
1106 /**
1107  * DOC: Flags for SUBMIT_JOB ioctl fragment command.
1108  *
1109  * .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_SINGLE_CORE
1110  *
1111  *    Use single core in a multi core setup.
1112  *
1113  * .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_DEPTHBUFFER
1114  *
1115  *    Indicates whether a depth buffer is present.
1116  *
1117  * .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_STENCILBUFFER
1118  *
1119  *    Indicates whether a stencil buffer is present.
1120  *
1121  * .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_PREVENT_CDM_OVERLAP
1122  *
1123  *    Disallow compute overlapped with this render.
1124  *
1125  * .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_GET_VIS_RESULTS
1126  *
1127  *    Indicates whether this render produces visibility results.
1128  *
1129  * .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_SCRATCHBUFFER
1130  *
1131  *    Indicates whether partial renders write to a scratch buffer instead of
1132  *    the final surface. It also forces the full screen copy expected to be
1133  *    present on the last render after all partial renders have completed.
1134  *
1135  * .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_DISABLE_PIXELMERGE
1136  *
1137  *    Disable pixel merging for this render.
1138  *
1139  * .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_FLAGS_MASK
1140  *
1141  *    Logical OR of all the fragment cmd flags.
1142  */
1143 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_SINGLE_CORE _BITULL(0)
1144 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_DEPTHBUFFER _BITULL(1)
1145 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_STENCILBUFFER _BITULL(2)
1146 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_PREVENT_CDM_OVERLAP _BITULL(3)
1147 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_SCRATCHBUFFER _BITULL(4)
1148 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_GET_VIS_RESULTS _BITULL(5)
1149 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_PARTIAL_RENDER _BITULL(6)
1150 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_DISABLE_PIXELMERGE _BITULL(7)
1151 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_FLAGS_MASK                                 \
1152 	(DRM_PVR_SUBMIT_JOB_FRAG_CMD_SINGLE_CORE |                             \
1153 	 DRM_PVR_SUBMIT_JOB_FRAG_CMD_DEPTHBUFFER |                             \
1154 	 DRM_PVR_SUBMIT_JOB_FRAG_CMD_STENCILBUFFER |                           \
1155 	 DRM_PVR_SUBMIT_JOB_FRAG_CMD_PREVENT_CDM_OVERLAP |                     \
1156 	 DRM_PVR_SUBMIT_JOB_FRAG_CMD_SCRATCHBUFFER |                           \
1157 	 DRM_PVR_SUBMIT_JOB_FRAG_CMD_GET_VIS_RESULTS |                         \
1158 	 DRM_PVR_SUBMIT_JOB_FRAG_CMD_PARTIAL_RENDER |                          \
1159 	 DRM_PVR_SUBMIT_JOB_FRAG_CMD_DISABLE_PIXELMERGE)
1160 
1161 /**
1162  * DOC: Flags for SUBMIT_JOB ioctl compute command.
1163  *
1164  * .. c:macro:: DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_PREVENT_ALL_OVERLAP
1165  *
1166  *    Disallow other jobs overlapped with this compute.
1167  *
1168  * .. c:macro:: DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_SINGLE_CORE
1169  *
1170  *    Forces to use single core in a multi core device.
1171  *
1172  * .. c:macro:: DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_FLAGS_MASK
1173  *
1174  *    Logical OR of all the compute cmd flags.
1175  */
1176 #define DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_PREVENT_ALL_OVERLAP _BITULL(0)
1177 #define DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_SINGLE_CORE _BITULL(1)
1178 #define DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_FLAGS_MASK         \
1179 	(DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_PREVENT_ALL_OVERLAP | \
1180 	 DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_SINGLE_CORE)
1181 
1182 /**
1183  * DOC: Flags for SUBMIT_JOB ioctl transfer command.
1184  *
1185  * .. c:macro:: DRM_PVR_SUBMIT_JOB_TRANSFER_CMD_SINGLE_CORE
1186  *
1187  *    Forces job to use a single core in a multi core device.
1188  *
1189  * .. c:macro:: DRM_PVR_SUBMIT_JOB_TRANSFER_CMD_FLAGS_MASK
1190  *
1191  *    Logical OR of all the transfer cmd flags.
1192  */
1193 #define DRM_PVR_SUBMIT_JOB_TRANSFER_CMD_SINGLE_CORE _BITULL(0)
1194 
1195 #define DRM_PVR_SUBMIT_JOB_TRANSFER_CMD_FLAGS_MASK \
1196 	DRM_PVR_SUBMIT_JOB_TRANSFER_CMD_SINGLE_CORE
1197 
1198 /**
1199  * enum drm_pvr_job_type - Arguments for &struct drm_pvr_job.job_type
1200  */
1201 enum drm_pvr_job_type {
1202 	/** @DRM_PVR_JOB_TYPE_GEOMETRY: Job type is geometry. */
1203 	DRM_PVR_JOB_TYPE_GEOMETRY = 0,
1204 
1205 	/** @DRM_PVR_JOB_TYPE_FRAGMENT: Job type is fragment. */
1206 	DRM_PVR_JOB_TYPE_FRAGMENT,
1207 
1208 	/** @DRM_PVR_JOB_TYPE_COMPUTE: Job type is compute. */
1209 	DRM_PVR_JOB_TYPE_COMPUTE,
1210 
1211 	/** @DRM_PVR_JOB_TYPE_TRANSFER_FRAG: Job type is a fragment transfer. */
1212 	DRM_PVR_JOB_TYPE_TRANSFER_FRAG,
1213 };
1214 
1215 /**
1216  * struct drm_pvr_hwrt_data_ref - Reference HWRT data
1217  */
1218 struct drm_pvr_hwrt_data_ref {
1219 	/** @set_handle: HWRT data set handle. */
1220 	__u32 set_handle;
1221 
1222 	/** @data_index: Index of the HWRT data inside the data set. */
1223 	__u32 data_index;
1224 };
1225 
1226 /**
1227  * struct drm_pvr_job - Job arguments passed to the %DRM_IOCTL_PVR_SUBMIT_JOBS ioctl
1228  */
1229 struct drm_pvr_job {
1230 	/**
1231 	 * @type: [IN] Type of job being submitted
1232 	 *
1233 	 * This must be one of the values defined by &enum drm_pvr_job_type.
1234 	 */
1235 	__u32 type;
1236 
1237 	/**
1238 	 * @context_handle: [IN] Context handle.
1239 	 *
1240 	 * When @job_type is %DRM_PVR_JOB_TYPE_RENDER, %DRM_PVR_JOB_TYPE_COMPUTE or
1241 	 * %DRM_PVR_JOB_TYPE_TRANSFER_FRAG, this must be a valid handle returned by
1242 	 * %DRM_IOCTL_PVR_CREATE_CONTEXT. The type of context must be compatible
1243 	 * with the type of job being submitted.
1244 	 *
1245 	 * When @job_type is %DRM_PVR_JOB_TYPE_NULL, this must be zero.
1246 	 */
1247 	__u32 context_handle;
1248 
1249 	/**
1250 	 * @flags: [IN] Flags for command.
1251 	 *
1252 	 * Those are job-dependent. See all ``DRM_PVR_SUBMIT_JOB_*``.
1253 	 */
1254 	__u32 flags;
1255 
1256 	/**
1257 	 * @cmd_stream_len: [IN] Length of command stream, in bytes.
1258 	 */
1259 	__u32 cmd_stream_len;
1260 
1261 	/**
1262 	 * @cmd_stream: [IN] Pointer to command stream for command.
1263 	 *
1264 	 * The command stream must be u64-aligned.
1265 	 */
1266 	__u64 cmd_stream;
1267 
1268 	/** @sync_ops: [IN] Fragment sync operations. */
1269 	struct drm_pvr_obj_array sync_ops;
1270 
1271 	/**
1272 	 * @hwrt: [IN] HWRT data used by render jobs (geometry or fragment).
1273 	 *
1274 	 * Must be zero for non-render jobs.
1275 	 */
1276 	struct drm_pvr_hwrt_data_ref hwrt;
1277 };
1278 
1279 /**
1280  * struct drm_pvr_ioctl_submit_jobs_args - Arguments for %DRM_IOCTL_PVR_SUBMIT_JOB
1281  *
1282  * If the syscall returns an error it is important to check the value of
1283  * @jobs.count. This indicates the index into @jobs.array where the
1284  * error occurred.
1285  */
1286 struct drm_pvr_ioctl_submit_jobs_args {
1287 	/** @jobs: [IN] Array of jobs to submit. */
1288 	struct drm_pvr_obj_array jobs;
1289 };
1290 
1291 #if defined(__cplusplus)
1292 }
1293 #endif
1294 
1295 #endif /* PVR_DRM_UAPI_H */
1296