xref: /linux/include/uapi/drm/panthor_drm.h (revision 58809f614e0e3f4e12b489bddf680bfeb31c0a20)
1 /* SPDX-License-Identifier: MIT */
2 /* Copyright (C) 2023 Collabora ltd. */
3 #ifndef _PANTHOR_DRM_H_
4 #define _PANTHOR_DRM_H_
5 
6 #include "drm.h"
7 
8 #if defined(__cplusplus)
9 extern "C" {
10 #endif
11 
12 /**
13  * DOC: Introduction
14  *
15  * This documentation describes the Panthor IOCTLs.
16  *
17  * Just a few generic rules about the data passed to the Panthor IOCTLs:
18  *
19  * - Structures must be aligned on 64-bit/8-byte. If the object is not
20  *   naturally aligned, a padding field must be added.
21  * - Fields must be explicitly aligned to their natural type alignment with
22  *   pad[0..N] fields.
23  * - All padding fields will be checked by the driver to make sure they are
24  *   zeroed.
25  * - Flags can be added, but not removed/replaced.
26  * - New fields can be added to the main structures (the structures
27  *   directly passed to the ioctl). Those fields can be added at the end of
28  *   the structure, or replace existing padding fields. Any new field being
29  *   added must preserve the behavior that existed before those fields were
30  *   added when a value of zero is passed.
31  * - New fields can be added to indirect objects (objects pointed by the
32  *   main structure), iff those objects are passed a size to reflect the
33  *   size known by the userspace driver (see drm_panthor_obj_array::stride
34  *   or drm_panthor_dev_query::size).
35  * - If the kernel driver is too old to know some fields, those will be
36  *   ignored if zero, and otherwise rejected (and so will be zero on output).
37  * - If userspace is too old to know some fields, those will be zeroed
38  *   (input) before the structure is parsed by the kernel driver.
39  * - Each new flag/field addition must come with a driver version update so
40  *   the userspace driver doesn't have to trial and error to know which
41  *   flags are supported.
42  * - Structures should not contain unions, as this would defeat the
43  *   extensibility of such structures.
44  * - IOCTLs can't be removed or replaced. New IOCTL IDs should be placed
45  *   at the end of the drm_panthor_ioctl_id enum.
46  */
47 
48 /**
49  * DOC: MMIO regions exposed to userspace.
50  *
51  * .. c:macro:: DRM_PANTHOR_USER_MMIO_OFFSET
52  *
53  * File offset for all MMIO regions being exposed to userspace. Don't use
54  * this value directly, use DRM_PANTHOR_USER_<name>_OFFSET values instead.
55  * pgoffset passed to mmap2() is an unsigned long, which forces us to use a
56  * different offset on 32-bit and 64-bit systems.
57  *
58  * .. c:macro:: DRM_PANTHOR_USER_FLUSH_ID_MMIO_OFFSET
59  *
60  * File offset for the LATEST_FLUSH_ID register. The Userspace driver controls
61  * GPU cache flushing through CS instructions, but the flush reduction
62  * mechanism requires a flush_id. This flush_id could be queried with an
63  * ioctl, but Arm provides a well-isolated register page containing only this
64  * read-only register, so let's expose this page through a static mmap offset
65  * and allow direct mapping of this MMIO region so we can avoid the
66  * user <-> kernel round-trip.
67  */
68 #define DRM_PANTHOR_USER_MMIO_OFFSET_32BIT	(1ull << 43)
69 #define DRM_PANTHOR_USER_MMIO_OFFSET_64BIT	(1ull << 56)
70 #define DRM_PANTHOR_USER_MMIO_OFFSET		(sizeof(unsigned long) < 8 ? \
71 						 DRM_PANTHOR_USER_MMIO_OFFSET_32BIT : \
72 						 DRM_PANTHOR_USER_MMIO_OFFSET_64BIT)
73 #define DRM_PANTHOR_USER_FLUSH_ID_MMIO_OFFSET	(DRM_PANTHOR_USER_MMIO_OFFSET | 0)
74 
75 /**
76  * DOC: IOCTL IDs
77  *
78  * enum drm_panthor_ioctl_id - IOCTL IDs
79  *
80  * Place new ioctls at the end, don't re-order, don't replace or remove entries.
81  *
82  * These IDs are not meant to be used directly. Use the DRM_IOCTL_PANTHOR_xxx
83  * definitions instead.
84  */
85 enum drm_panthor_ioctl_id {
86 	/** @DRM_PANTHOR_DEV_QUERY: Query device information. */
87 	DRM_PANTHOR_DEV_QUERY = 0,
88 
89 	/** @DRM_PANTHOR_VM_CREATE: Create a VM. */
90 	DRM_PANTHOR_VM_CREATE,
91 
92 	/** @DRM_PANTHOR_VM_DESTROY: Destroy a VM. */
93 	DRM_PANTHOR_VM_DESTROY,
94 
95 	/** @DRM_PANTHOR_VM_BIND: Bind/unbind memory to a VM. */
96 	DRM_PANTHOR_VM_BIND,
97 
98 	/** @DRM_PANTHOR_VM_GET_STATE: Get VM state. */
99 	DRM_PANTHOR_VM_GET_STATE,
100 
101 	/** @DRM_PANTHOR_BO_CREATE: Create a buffer object. */
102 	DRM_PANTHOR_BO_CREATE,
103 
104 	/**
105 	 * @DRM_PANTHOR_BO_MMAP_OFFSET: Get the file offset to pass to
106 	 * mmap to map a GEM object.
107 	 */
108 	DRM_PANTHOR_BO_MMAP_OFFSET,
109 
110 	/** @DRM_PANTHOR_GROUP_CREATE: Create a scheduling group. */
111 	DRM_PANTHOR_GROUP_CREATE,
112 
113 	/** @DRM_PANTHOR_GROUP_DESTROY: Destroy a scheduling group. */
114 	DRM_PANTHOR_GROUP_DESTROY,
115 
116 	/**
117 	 * @DRM_PANTHOR_GROUP_SUBMIT: Submit jobs to queues belonging
118 	 * to a specific scheduling group.
119 	 */
120 	DRM_PANTHOR_GROUP_SUBMIT,
121 
122 	/** @DRM_PANTHOR_GROUP_GET_STATE: Get the state of a scheduling group. */
123 	DRM_PANTHOR_GROUP_GET_STATE,
124 
125 	/** @DRM_PANTHOR_TILER_HEAP_CREATE: Create a tiler heap. */
126 	DRM_PANTHOR_TILER_HEAP_CREATE,
127 
128 	/** @DRM_PANTHOR_TILER_HEAP_DESTROY: Destroy a tiler heap. */
129 	DRM_PANTHOR_TILER_HEAP_DESTROY,
130 
131 	/** @DRM_PANTHOR_BO_SET_LABEL: Label a BO. */
132 	DRM_PANTHOR_BO_SET_LABEL,
133 
134 	/**
135 	 * @DRM_PANTHOR_SET_USER_MMIO_OFFSET: Set the offset to use as the user MMIO offset.
136 	 *
137 	 * The default behavior is to pick the MMIO offset based on the size of the pgoff_t
138 	 * type seen by the process that manipulates the FD, such that a 32-bit process can
139 	 * always map the user MMIO ranges. But this approach doesn't work well for emulators
140 	 * like FEX, where the emulator is an 64-bit binary which might be executing 32-bit
141 	 * code. In that case, the kernel thinks it's the 64-bit process and assumes
142 	 * DRM_PANTHOR_USER_MMIO_OFFSET_64BIT is in use, but the UMD library expects
143 	 * DRM_PANTHOR_USER_MMIO_OFFSET_32BIT, because it can't mmap() anything above the
144 	 * pgoff_t size.
145 	 */
146 	DRM_PANTHOR_SET_USER_MMIO_OFFSET,
147 };
148 
149 /**
150  * DOC: IOCTL arguments
151  */
152 
153 /**
154  * struct drm_panthor_obj_array - Object array.
155  *
156  * This object is used to pass an array of objects whose size is subject to changes in
157  * future versions of the driver. In order to support this mutability, we pass a stride
158  * describing the size of the object as known by userspace.
159  *
160  * You shouldn't fill drm_panthor_obj_array fields directly. You should instead use
161  * the DRM_PANTHOR_OBJ_ARRAY() macro that takes care of initializing the stride to
162  * the object size.
163  */
164 struct drm_panthor_obj_array {
165 	/** @stride: Stride of object struct. Used for versioning. */
166 	__u32 stride;
167 
168 	/** @count: Number of objects in the array. */
169 	__u32 count;
170 
171 	/** @array: User pointer to an array of objects. */
172 	__u64 array;
173 };
174 
175 /**
176  * DRM_PANTHOR_OBJ_ARRAY() - Initialize a drm_panthor_obj_array field.
177  * @cnt: Number of elements in the array.
178  * @ptr: Pointer to the array to pass to the kernel.
179  *
180  * Macro initializing a drm_panthor_obj_array based on the object size as known
181  * by userspace.
182  */
183 #define DRM_PANTHOR_OBJ_ARRAY(cnt, ptr) \
184 	{ .stride = sizeof((ptr)[0]), .count = (cnt), .array = (__u64)(uintptr_t)(ptr) }
185 
186 /**
187  * enum drm_panthor_sync_op_flags - Synchronization operation flags.
188  */
189 enum drm_panthor_sync_op_flags {
190 	/** @DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_MASK: Synchronization handle type mask. */
191 	DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_MASK = 0xff,
192 
193 	/** @DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_SYNCOBJ: Synchronization object type. */
194 	DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_SYNCOBJ = 0,
195 
196 	/**
197 	 * @DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_TIMELINE_SYNCOBJ: Timeline synchronization
198 	 * object type.
199 	 */
200 	DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_TIMELINE_SYNCOBJ = 1,
201 
202 	/** @DRM_PANTHOR_SYNC_OP_WAIT: Wait operation. */
203 	DRM_PANTHOR_SYNC_OP_WAIT = 0 << 31,
204 
205 	/** @DRM_PANTHOR_SYNC_OP_SIGNAL: Signal operation. */
206 	DRM_PANTHOR_SYNC_OP_SIGNAL = (int)(1u << 31),
207 };
208 
209 /**
210  * struct drm_panthor_sync_op - Synchronization operation.
211  */
212 struct drm_panthor_sync_op {
213 	/** @flags: Synchronization operation flags. Combination of DRM_PANTHOR_SYNC_OP values. */
214 	__u32 flags;
215 
216 	/** @handle: Sync handle. */
217 	__u32 handle;
218 
219 	/**
220 	 * @timeline_value: MBZ if
221 	 * (flags & DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_MASK) !=
222 	 * DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_TIMELINE_SYNCOBJ.
223 	 */
224 	__u64 timeline_value;
225 };
226 
227 /**
228  * enum drm_panthor_dev_query_type - Query type
229  *
230  * Place new types at the end, don't re-order, don't remove or replace.
231  */
232 enum drm_panthor_dev_query_type {
233 	/** @DRM_PANTHOR_DEV_QUERY_GPU_INFO: Query GPU information. */
234 	DRM_PANTHOR_DEV_QUERY_GPU_INFO = 0,
235 
236 	/** @DRM_PANTHOR_DEV_QUERY_CSIF_INFO: Query command-stream interface information. */
237 	DRM_PANTHOR_DEV_QUERY_CSIF_INFO,
238 
239 	/** @DRM_PANTHOR_DEV_QUERY_TIMESTAMP_INFO: Query timestamp information. */
240 	DRM_PANTHOR_DEV_QUERY_TIMESTAMP_INFO,
241 
242 	/**
243 	 * @DRM_PANTHOR_DEV_QUERY_GROUP_PRIORITIES_INFO: Query allowed group priorities information.
244 	 */
245 	DRM_PANTHOR_DEV_QUERY_GROUP_PRIORITIES_INFO,
246 };
247 
248 /**
249  * struct drm_panthor_gpu_info - GPU information
250  *
251  * Structure grouping all queryable information relating to the GPU.
252  */
253 struct drm_panthor_gpu_info {
254 	/** @gpu_id : GPU ID. */
255 	__u32 gpu_id;
256 #define DRM_PANTHOR_ARCH_MAJOR(x)		((x) >> 28)
257 #define DRM_PANTHOR_ARCH_MINOR(x)		(((x) >> 24) & 0xf)
258 #define DRM_PANTHOR_ARCH_REV(x)			(((x) >> 20) & 0xf)
259 #define DRM_PANTHOR_PRODUCT_MAJOR(x)		(((x) >> 16) & 0xf)
260 #define DRM_PANTHOR_VERSION_MAJOR(x)		(((x) >> 12) & 0xf)
261 #define DRM_PANTHOR_VERSION_MINOR(x)		(((x) >> 4) & 0xff)
262 #define DRM_PANTHOR_VERSION_STATUS(x)		((x) & 0xf)
263 
264 	/** @gpu_rev: GPU revision. */
265 	__u32 gpu_rev;
266 
267 	/** @csf_id: Command stream frontend ID. */
268 	__u32 csf_id;
269 #define DRM_PANTHOR_CSHW_MAJOR(x)		(((x) >> 26) & 0x3f)
270 #define DRM_PANTHOR_CSHW_MINOR(x)		(((x) >> 20) & 0x3f)
271 #define DRM_PANTHOR_CSHW_REV(x)			(((x) >> 16) & 0xf)
272 #define DRM_PANTHOR_MCU_MAJOR(x)		(((x) >> 10) & 0x3f)
273 #define DRM_PANTHOR_MCU_MINOR(x)		(((x) >> 4) & 0x3f)
274 #define DRM_PANTHOR_MCU_REV(x)			((x) & 0xf)
275 
276 	/** @l2_features: L2-cache features. */
277 	__u32 l2_features;
278 
279 	/** @tiler_features: Tiler features. */
280 	__u32 tiler_features;
281 
282 	/** @mem_features: Memory features. */
283 	__u32 mem_features;
284 
285 	/** @mmu_features: MMU features. */
286 	__u32 mmu_features;
287 #define DRM_PANTHOR_MMU_VA_BITS(x)		((x) & 0xff)
288 
289 	/** @thread_features: Thread features. */
290 	__u32 thread_features;
291 
292 	/** @max_threads: Maximum number of threads. */
293 	__u32 max_threads;
294 
295 	/** @thread_max_workgroup_size: Maximum workgroup size. */
296 	__u32 thread_max_workgroup_size;
297 
298 	/**
299 	 * @thread_max_barrier_size: Maximum number of threads that can wait
300 	 * simultaneously on a barrier.
301 	 */
302 	__u32 thread_max_barrier_size;
303 
304 	/** @coherency_features: Coherency features. */
305 	__u32 coherency_features;
306 
307 	/** @texture_features: Texture features. */
308 	__u32 texture_features[4];
309 
310 	/** @as_present: Bitmask encoding the number of address-space exposed by the MMU. */
311 	__u32 as_present;
312 
313 	/** @pad0: MBZ. */
314 	__u32 pad0;
315 
316 	/** @shader_present: Bitmask encoding the shader cores exposed by the GPU. */
317 	__u64 shader_present;
318 
319 	/** @l2_present: Bitmask encoding the L2 caches exposed by the GPU. */
320 	__u64 l2_present;
321 
322 	/** @tiler_present: Bitmask encoding the tiler units exposed by the GPU. */
323 	__u64 tiler_present;
324 
325 	/** @core_features: Used to discriminate core variants when they exist. */
326 	__u32 core_features;
327 
328 	/** @pad: MBZ. */
329 	__u32 pad;
330 
331 	/** @gpu_features: Bitmask describing supported GPU-wide features */
332 	__u64 gpu_features;
333 };
334 
335 /**
336  * struct drm_panthor_csif_info - Command stream interface information
337  *
338  * Structure grouping all queryable information relating to the command stream interface.
339  */
340 struct drm_panthor_csif_info {
341 	/** @csg_slot_count: Number of command stream group slots exposed by the firmware. */
342 	__u32 csg_slot_count;
343 
344 	/** @cs_slot_count: Number of command stream slots per group. */
345 	__u32 cs_slot_count;
346 
347 	/** @cs_reg_count: Number of command stream registers. */
348 	__u32 cs_reg_count;
349 
350 	/** @scoreboard_slot_count: Number of scoreboard slots. */
351 	__u32 scoreboard_slot_count;
352 
353 	/**
354 	 * @unpreserved_cs_reg_count: Number of command stream registers reserved by
355 	 * the kernel driver to call a userspace command stream.
356 	 *
357 	 * All registers can be used by a userspace command stream, but the
358 	 * [cs_slot_count - unpreserved_cs_reg_count .. cs_slot_count] registers are
359 	 * used by the kernel when DRM_PANTHOR_IOCTL_GROUP_SUBMIT is called.
360 	 */
361 	__u32 unpreserved_cs_reg_count;
362 
363 	/**
364 	 * @pad: Padding field, set to zero.
365 	 */
366 	__u32 pad;
367 };
368 
369 /**
370  * struct drm_panthor_timestamp_info - Timestamp information
371  *
372  * Structure grouping all queryable information relating to the GPU timestamp.
373  */
374 struct drm_panthor_timestamp_info {
375 	/**
376 	 * @timestamp_frequency: The frequency of the timestamp timer or 0 if
377 	 * unknown.
378 	 */
379 	__u64 timestamp_frequency;
380 
381 	/** @current_timestamp: The current timestamp. */
382 	__u64 current_timestamp;
383 
384 	/** @timestamp_offset: The offset of the timestamp timer. */
385 	__u64 timestamp_offset;
386 };
387 
388 /**
389  * struct drm_panthor_group_priorities_info - Group priorities information
390  *
391  * Structure grouping all queryable information relating to the allowed group priorities.
392  */
393 struct drm_panthor_group_priorities_info {
394 	/**
395 	 * @allowed_mask: Bitmask of the allowed group priorities.
396 	 *
397 	 * Each bit represents a variant of the enum drm_panthor_group_priority.
398 	 */
399 	__u8 allowed_mask;
400 
401 	/** @pad: Padding fields, MBZ. */
402 	__u8 pad[3];
403 };
404 
405 /**
406  * struct drm_panthor_dev_query - Arguments passed to DRM_PANTHOR_IOCTL_DEV_QUERY
407  */
408 struct drm_panthor_dev_query {
409 	/** @type: the query type (see drm_panthor_dev_query_type). */
410 	__u32 type;
411 
412 	/**
413 	 * @size: size of the type being queried.
414 	 *
415 	 * If pointer is NULL, size is updated by the driver to provide the
416 	 * output structure size. If pointer is not NULL, the driver will
417 	 * only copy min(size, actual_structure_size) bytes to the pointer,
418 	 * and update the size accordingly. This allows us to extend query
419 	 * types without breaking userspace.
420 	 */
421 	__u32 size;
422 
423 	/**
424 	 * @pointer: user pointer to a query type struct.
425 	 *
426 	 * Pointer can be NULL, in which case, nothing is copied, but the
427 	 * actual structure size is returned. If not NULL, it must point to
428 	 * a location that's large enough to hold size bytes.
429 	 */
430 	__u64 pointer;
431 };
432 
433 /**
434  * struct drm_panthor_vm_create - Arguments passed to DRM_PANTHOR_IOCTL_VM_CREATE
435  */
436 struct drm_panthor_vm_create {
437 	/** @flags: VM flags, MBZ. */
438 	__u32 flags;
439 
440 	/** @id: Returned VM ID. */
441 	__u32 id;
442 
443 	/**
444 	 * @user_va_range: Size of the VA space reserved for user objects.
445 	 *
446 	 * The kernel will pick the remaining space to map kernel-only objects to the
447 	 * VM (heap chunks, heap context, ring buffers, kernel synchronization objects,
448 	 * ...). If the space left for kernel objects is too small, kernel object
449 	 * allocation will fail further down the road. One can use
450 	 * drm_panthor_gpu_info::mmu_features to extract the total virtual address
451 	 * range, and chose a user_va_range that leaves some space to the kernel.
452 	 *
453 	 * If user_va_range is zero, the kernel will pick a sensible value based on
454 	 * TASK_SIZE and the virtual range supported by the GPU MMU (the kernel/user
455 	 * split should leave enough VA space for userspace processes to support SVM,
456 	 * while still allowing the kernel to map some amount of kernel objects in
457 	 * the kernel VA range). The value chosen by the driver will be returned in
458 	 * @user_va_range.
459 	 *
460 	 * User VA space always starts at 0x0, kernel VA space is always placed after
461 	 * the user VA range.
462 	 */
463 	__u64 user_va_range;
464 };
465 
466 /**
467  * struct drm_panthor_vm_destroy - Arguments passed to DRM_PANTHOR_IOCTL_VM_DESTROY
468  */
469 struct drm_panthor_vm_destroy {
470 	/** @id: ID of the VM to destroy. */
471 	__u32 id;
472 
473 	/** @pad: MBZ. */
474 	__u32 pad;
475 };
476 
477 /**
478  * enum drm_panthor_vm_bind_op_flags - VM bind operation flags
479  */
480 enum drm_panthor_vm_bind_op_flags {
481 	/**
482 	 * @DRM_PANTHOR_VM_BIND_OP_MAP_READONLY: Map the memory read-only.
483 	 *
484 	 * Only valid with DRM_PANTHOR_VM_BIND_OP_TYPE_MAP.
485 	 */
486 	DRM_PANTHOR_VM_BIND_OP_MAP_READONLY = 1 << 0,
487 
488 	/**
489 	 * @DRM_PANTHOR_VM_BIND_OP_MAP_NOEXEC: Map the memory not-executable.
490 	 *
491 	 * Only valid with DRM_PANTHOR_VM_BIND_OP_TYPE_MAP.
492 	 */
493 	DRM_PANTHOR_VM_BIND_OP_MAP_NOEXEC = 1 << 1,
494 
495 	/**
496 	 * @DRM_PANTHOR_VM_BIND_OP_MAP_UNCACHED: Map the memory uncached.
497 	 *
498 	 * Only valid with DRM_PANTHOR_VM_BIND_OP_TYPE_MAP.
499 	 */
500 	DRM_PANTHOR_VM_BIND_OP_MAP_UNCACHED = 1 << 2,
501 
502 	/**
503 	 * @DRM_PANTHOR_VM_BIND_OP_TYPE_MASK: Mask used to determine the type of operation.
504 	 */
505 	DRM_PANTHOR_VM_BIND_OP_TYPE_MASK = (int)(0xfu << 28),
506 
507 	/** @DRM_PANTHOR_VM_BIND_OP_TYPE_MAP: Map operation. */
508 	DRM_PANTHOR_VM_BIND_OP_TYPE_MAP = 0 << 28,
509 
510 	/** @DRM_PANTHOR_VM_BIND_OP_TYPE_UNMAP: Unmap operation. */
511 	DRM_PANTHOR_VM_BIND_OP_TYPE_UNMAP = 1 << 28,
512 
513 	/**
514 	 * @DRM_PANTHOR_VM_BIND_OP_TYPE_SYNC_ONLY: No VM operation.
515 	 *
516 	 * Just serves as a synchronization point on a VM queue.
517 	 *
518 	 * Only valid if %DRM_PANTHOR_VM_BIND_ASYNC is set in drm_panthor_vm_bind::flags,
519 	 * and drm_panthor_vm_bind_op::syncs contains at least one element.
520 	 */
521 	DRM_PANTHOR_VM_BIND_OP_TYPE_SYNC_ONLY = 2 << 28,
522 };
523 
524 /**
525  * struct drm_panthor_vm_bind_op - VM bind operation
526  */
527 struct drm_panthor_vm_bind_op {
528 	/** @flags: Combination of drm_panthor_vm_bind_op_flags flags. */
529 	__u32 flags;
530 
531 	/**
532 	 * @bo_handle: Handle of the buffer object to map.
533 	 * MBZ for unmap or sync-only operations.
534 	 */
535 	__u32 bo_handle;
536 
537 	/**
538 	 * @bo_offset: Buffer object offset.
539 	 * MBZ for unmap or sync-only operations.
540 	 */
541 	__u64 bo_offset;
542 
543 	/**
544 	 * @va: Virtual address to map/unmap.
545 	 * MBZ for sync-only operations.
546 	 */
547 	__u64 va;
548 
549 	/**
550 	 * @size: Size to map/unmap.
551 	 * MBZ for sync-only operations.
552 	 */
553 	__u64 size;
554 
555 	/**
556 	 * @syncs: Array of struct drm_panthor_sync_op synchronization
557 	 * operations.
558 	 *
559 	 * This array must be empty if %DRM_PANTHOR_VM_BIND_ASYNC is not set on
560 	 * the drm_panthor_vm_bind object containing this VM bind operation.
561 	 *
562 	 * This array shall not be empty for sync-only operations.
563 	 */
564 	struct drm_panthor_obj_array syncs;
565 
566 };
567 
568 /**
569  * enum drm_panthor_vm_bind_flags - VM bind flags
570  */
571 enum drm_panthor_vm_bind_flags {
572 	/**
573 	 * @DRM_PANTHOR_VM_BIND_ASYNC: VM bind operations are queued to the VM
574 	 * queue instead of being executed synchronously.
575 	 */
576 	DRM_PANTHOR_VM_BIND_ASYNC = 1 << 0,
577 };
578 
579 /**
580  * struct drm_panthor_vm_bind - Arguments passed to DRM_IOCTL_PANTHOR_VM_BIND
581  */
582 struct drm_panthor_vm_bind {
583 	/** @vm_id: VM targeted by the bind request. */
584 	__u32 vm_id;
585 
586 	/** @flags: Combination of drm_panthor_vm_bind_flags flags. */
587 	__u32 flags;
588 
589 	/** @ops: Array of struct drm_panthor_vm_bind_op bind operations. */
590 	struct drm_panthor_obj_array ops;
591 };
592 
593 /**
594  * enum drm_panthor_vm_state - VM states.
595  */
596 enum drm_panthor_vm_state {
597 	/**
598 	 * @DRM_PANTHOR_VM_STATE_USABLE: VM is usable.
599 	 *
600 	 * New VM operations will be accepted on this VM.
601 	 */
602 	DRM_PANTHOR_VM_STATE_USABLE,
603 
604 	/**
605 	 * @DRM_PANTHOR_VM_STATE_UNUSABLE: VM is unusable.
606 	 *
607 	 * Something put the VM in an unusable state (like an asynchronous
608 	 * VM_BIND request failing for any reason).
609 	 *
610 	 * Once the VM is in this state, all new MAP operations will be
611 	 * rejected, and any GPU job targeting this VM will fail.
612 	 * UNMAP operations are still accepted.
613 	 *
614 	 * The only way to recover from an unusable VM is to create a new
615 	 * VM, and destroy the old one.
616 	 */
617 	DRM_PANTHOR_VM_STATE_UNUSABLE,
618 };
619 
620 /**
621  * struct drm_panthor_vm_get_state - Get VM state.
622  */
623 struct drm_panthor_vm_get_state {
624 	/** @vm_id: VM targeted by the get_state request. */
625 	__u32 vm_id;
626 
627 	/**
628 	 * @state: state returned by the driver.
629 	 *
630 	 * Must be one of the enum drm_panthor_vm_state values.
631 	 */
632 	__u32 state;
633 };
634 
635 /**
636  * enum drm_panthor_bo_flags - Buffer object flags, passed at creation time.
637  */
638 enum drm_panthor_bo_flags {
639 	/** @DRM_PANTHOR_BO_NO_MMAP: The buffer object will never be CPU-mapped in userspace. */
640 	DRM_PANTHOR_BO_NO_MMAP = (1 << 0),
641 };
642 
643 /**
644  * struct drm_panthor_bo_create - Arguments passed to DRM_IOCTL_PANTHOR_BO_CREATE.
645  */
646 struct drm_panthor_bo_create {
647 	/**
648 	 * @size: Requested size for the object
649 	 *
650 	 * The (page-aligned) allocated size for the object will be returned.
651 	 */
652 	__u64 size;
653 
654 	/**
655 	 * @flags: Flags. Must be a combination of drm_panthor_bo_flags flags.
656 	 */
657 	__u32 flags;
658 
659 	/**
660 	 * @exclusive_vm_id: Exclusive VM this buffer object will be mapped to.
661 	 *
662 	 * If not zero, the field must refer to a valid VM ID, and implies that:
663 	 *  - the buffer object will only ever be bound to that VM
664 	 *  - cannot be exported as a PRIME fd
665 	 */
666 	__u32 exclusive_vm_id;
667 
668 	/**
669 	 * @handle: Returned handle for the object.
670 	 *
671 	 * Object handles are nonzero.
672 	 */
673 	__u32 handle;
674 
675 	/** @pad: MBZ. */
676 	__u32 pad;
677 };
678 
679 /**
680  * struct drm_panthor_bo_mmap_offset - Arguments passed to DRM_IOCTL_PANTHOR_BO_MMAP_OFFSET.
681  */
682 struct drm_panthor_bo_mmap_offset {
683 	/** @handle: Handle of the object we want an mmap offset for. */
684 	__u32 handle;
685 
686 	/** @pad: MBZ. */
687 	__u32 pad;
688 
689 	/** @offset: The fake offset to use for subsequent mmap calls. */
690 	__u64 offset;
691 };
692 
693 /**
694  * struct drm_panthor_queue_create - Queue creation arguments.
695  */
696 struct drm_panthor_queue_create {
697 	/**
698 	 * @priority: Defines the priority of queues inside a group. Goes from 0 to 15,
699 	 * 15 being the highest priority.
700 	 */
701 	__u8 priority;
702 
703 	/** @pad: Padding fields, MBZ. */
704 	__u8 pad[3];
705 
706 	/** @ringbuf_size: Size of the ring buffer to allocate to this queue. */
707 	__u32 ringbuf_size;
708 };
709 
710 /**
711  * enum drm_panthor_group_priority - Scheduling group priority
712  */
713 enum drm_panthor_group_priority {
714 	/** @PANTHOR_GROUP_PRIORITY_LOW: Low priority group. */
715 	PANTHOR_GROUP_PRIORITY_LOW = 0,
716 
717 	/** @PANTHOR_GROUP_PRIORITY_MEDIUM: Medium priority group. */
718 	PANTHOR_GROUP_PRIORITY_MEDIUM,
719 
720 	/**
721 	 * @PANTHOR_GROUP_PRIORITY_HIGH: High priority group.
722 	 *
723 	 * Requires CAP_SYS_NICE or DRM_MASTER.
724 	 */
725 	PANTHOR_GROUP_PRIORITY_HIGH,
726 
727 	/**
728 	 * @PANTHOR_GROUP_PRIORITY_REALTIME: Realtime priority group.
729 	 *
730 	 * Requires CAP_SYS_NICE or DRM_MASTER.
731 	 */
732 	PANTHOR_GROUP_PRIORITY_REALTIME,
733 };
734 
735 /**
736  * struct drm_panthor_group_create - Arguments passed to DRM_IOCTL_PANTHOR_GROUP_CREATE
737  */
738 struct drm_panthor_group_create {
739 	/** @queues: Array of drm_panthor_queue_create elements. */
740 	struct drm_panthor_obj_array queues;
741 
742 	/**
743 	 * @max_compute_cores: Maximum number of cores that can be used by compute
744 	 * jobs across CS queues bound to this group.
745 	 *
746 	 * Must be less or equal to the number of bits set in @compute_core_mask.
747 	 */
748 	__u8 max_compute_cores;
749 
750 	/**
751 	 * @max_fragment_cores: Maximum number of cores that can be used by fragment
752 	 * jobs across CS queues bound to this group.
753 	 *
754 	 * Must be less or equal to the number of bits set in @fragment_core_mask.
755 	 */
756 	__u8 max_fragment_cores;
757 
758 	/**
759 	 * @max_tiler_cores: Maximum number of tilers that can be used by tiler jobs
760 	 * across CS queues bound to this group.
761 	 *
762 	 * Must be less or equal to the number of bits set in @tiler_core_mask.
763 	 */
764 	__u8 max_tiler_cores;
765 
766 	/** @priority: Group priority (see enum drm_panthor_group_priority). */
767 	__u8 priority;
768 
769 	/** @pad: Padding field, MBZ. */
770 	__u32 pad;
771 
772 	/**
773 	 * @compute_core_mask: Mask encoding cores that can be used for compute jobs.
774 	 *
775 	 * This field must have at least @max_compute_cores bits set.
776 	 *
777 	 * The bits set here should also be set in drm_panthor_gpu_info::shader_present.
778 	 */
779 	__u64 compute_core_mask;
780 
781 	/**
782 	 * @fragment_core_mask: Mask encoding cores that can be used for fragment jobs.
783 	 *
784 	 * This field must have at least @max_fragment_cores bits set.
785 	 *
786 	 * The bits set here should also be set in drm_panthor_gpu_info::shader_present.
787 	 */
788 	__u64 fragment_core_mask;
789 
790 	/**
791 	 * @tiler_core_mask: Mask encoding cores that can be used for tiler jobs.
792 	 *
793 	 * This field must have at least @max_tiler_cores bits set.
794 	 *
795 	 * The bits set here should also be set in drm_panthor_gpu_info::tiler_present.
796 	 */
797 	__u64 tiler_core_mask;
798 
799 	/**
800 	 * @vm_id: VM ID to bind this group to.
801 	 *
802 	 * All submission to queues bound to this group will use this VM.
803 	 */
804 	__u32 vm_id;
805 
806 	/**
807 	 * @group_handle: Returned group handle. Passed back when submitting jobs or
808 	 * destroying a group.
809 	 */
810 	__u32 group_handle;
811 };
812 
813 /**
814  * struct drm_panthor_group_destroy - Arguments passed to DRM_IOCTL_PANTHOR_GROUP_DESTROY
815  */
816 struct drm_panthor_group_destroy {
817 	/** @group_handle: Group to destroy */
818 	__u32 group_handle;
819 
820 	/** @pad: Padding field, MBZ. */
821 	__u32 pad;
822 };
823 
824 /**
825  * struct drm_panthor_queue_submit - Job submission arguments.
826  *
827  * This is describing the userspace command stream to call from the kernel
828  * command stream ring-buffer. Queue submission is always part of a group
829  * submission, taking one or more jobs to submit to the underlying queues.
830  */
831 struct drm_panthor_queue_submit {
832 	/** @queue_index: Index of the queue inside a group. */
833 	__u32 queue_index;
834 
835 	/**
836 	 * @stream_size: Size of the command stream to execute.
837 	 *
838 	 * Must be 64-bit/8-byte aligned (the size of a CS instruction)
839 	 *
840 	 * Can be zero if stream_addr is zero too.
841 	 *
842 	 * When the stream size is zero, the queue submit serves as a
843 	 * synchronization point.
844 	 */
845 	__u32 stream_size;
846 
847 	/**
848 	 * @stream_addr: GPU address of the command stream to execute.
849 	 *
850 	 * Must be aligned on 64-byte.
851 	 *
852 	 * Can be zero is stream_size is zero too.
853 	 */
854 	__u64 stream_addr;
855 
856 	/**
857 	 * @latest_flush: FLUSH_ID read at the time the stream was built.
858 	 *
859 	 * This allows cache flush elimination for the automatic
860 	 * flush+invalidate(all) done at submission time, which is needed to
861 	 * ensure the GPU doesn't get garbage when reading the indirect command
862 	 * stream buffers. If you want the cache flush to happen
863 	 * unconditionally, pass a zero here.
864 	 *
865 	 * Ignored when stream_size is zero.
866 	 */
867 	__u32 latest_flush;
868 
869 	/** @pad: MBZ. */
870 	__u32 pad;
871 
872 	/** @syncs: Array of struct drm_panthor_sync_op sync operations. */
873 	struct drm_panthor_obj_array syncs;
874 };
875 
876 /**
877  * struct drm_panthor_group_submit - Arguments passed to DRM_IOCTL_PANTHOR_GROUP_SUBMIT
878  */
879 struct drm_panthor_group_submit {
880 	/** @group_handle: Handle of the group to queue jobs to. */
881 	__u32 group_handle;
882 
883 	/** @pad: MBZ. */
884 	__u32 pad;
885 
886 	/** @queue_submits: Array of drm_panthor_queue_submit objects. */
887 	struct drm_panthor_obj_array queue_submits;
888 };
889 
890 /**
891  * enum drm_panthor_group_state_flags - Group state flags
892  */
893 enum drm_panthor_group_state_flags {
894 	/**
895 	 * @DRM_PANTHOR_GROUP_STATE_TIMEDOUT: Group had unfinished jobs.
896 	 *
897 	 * When a group ends up with this flag set, no jobs can be submitted to its queues.
898 	 */
899 	DRM_PANTHOR_GROUP_STATE_TIMEDOUT = 1 << 0,
900 
901 	/**
902 	 * @DRM_PANTHOR_GROUP_STATE_FATAL_FAULT: Group had fatal faults.
903 	 *
904 	 * When a group ends up with this flag set, no jobs can be submitted to its queues.
905 	 */
906 	DRM_PANTHOR_GROUP_STATE_FATAL_FAULT = 1 << 1,
907 
908 	/**
909 	 * @DRM_PANTHOR_GROUP_STATE_INNOCENT: Group was killed during a reset caused by other
910 	 * groups.
911 	 *
912 	 * This flag can only be set if DRM_PANTHOR_GROUP_STATE_TIMEDOUT is set and
913 	 * DRM_PANTHOR_GROUP_STATE_FATAL_FAULT is not.
914 	 */
915 	DRM_PANTHOR_GROUP_STATE_INNOCENT = 1 << 2,
916 };
917 
918 /**
919  * struct drm_panthor_group_get_state - Arguments passed to DRM_IOCTL_PANTHOR_GROUP_GET_STATE
920  *
921  * Used to query the state of a group and decide whether a new group should be created to
922  * replace it.
923  */
924 struct drm_panthor_group_get_state {
925 	/** @group_handle: Handle of the group to query state on */
926 	__u32 group_handle;
927 
928 	/**
929 	 * @state: Combination of DRM_PANTHOR_GROUP_STATE_* flags encoding the
930 	 * group state.
931 	 */
932 	__u32 state;
933 
934 	/** @fatal_queues: Bitmask of queues that faced fatal faults. */
935 	__u32 fatal_queues;
936 
937 	/** @pad: MBZ */
938 	__u32 pad;
939 };
940 
941 /**
942  * struct drm_panthor_tiler_heap_create - Arguments passed to DRM_IOCTL_PANTHOR_TILER_HEAP_CREATE
943  */
944 struct drm_panthor_tiler_heap_create {
945 	/** @vm_id: VM ID the tiler heap should be mapped to */
946 	__u32 vm_id;
947 
948 	/** @initial_chunk_count: Initial number of chunks to allocate. Must be at least one. */
949 	__u32 initial_chunk_count;
950 
951 	/**
952 	 * @chunk_size: Chunk size.
953 	 *
954 	 * Must be page-aligned and lie in the [128k:8M] range.
955 	 */
956 	__u32 chunk_size;
957 
958 	/**
959 	 * @max_chunks: Maximum number of chunks that can be allocated.
960 	 *
961 	 * Must be at least @initial_chunk_count.
962 	 */
963 	__u32 max_chunks;
964 
965 	/**
966 	 * @target_in_flight: Maximum number of in-flight render passes.
967 	 *
968 	 * If the heap has more than tiler jobs in-flight, the FW will wait for render
969 	 * passes to finish before queuing new tiler jobs.
970 	 */
971 	__u32 target_in_flight;
972 
973 	/** @handle: Returned heap handle. Passed back to DESTROY_TILER_HEAP. */
974 	__u32 handle;
975 
976 	/** @tiler_heap_ctx_gpu_va: Returned heap GPU virtual address returned */
977 	__u64 tiler_heap_ctx_gpu_va;
978 
979 	/**
980 	 * @first_heap_chunk_gpu_va: First heap chunk.
981 	 *
982 	 * The tiler heap is formed of heap chunks forming a single-link list. This
983 	 * is the first element in the list.
984 	 */
985 	__u64 first_heap_chunk_gpu_va;
986 };
987 
988 /**
989  * struct drm_panthor_tiler_heap_destroy - Arguments passed to DRM_IOCTL_PANTHOR_TILER_HEAP_DESTROY
990  */
991 struct drm_panthor_tiler_heap_destroy {
992 	/**
993 	 * @handle: Handle of the tiler heap to destroy.
994 	 *
995 	 * Must be a valid heap handle returned by DRM_IOCTL_PANTHOR_TILER_HEAP_CREATE.
996 	 */
997 	__u32 handle;
998 
999 	/** @pad: Padding field, MBZ. */
1000 	__u32 pad;
1001 };
1002 
1003 /**
1004  * struct drm_panthor_bo_set_label - Arguments passed to DRM_IOCTL_PANTHOR_BO_SET_LABEL
1005  */
1006 struct drm_panthor_bo_set_label {
1007 	/** @handle: Handle of the buffer object to label. */
1008 	__u32 handle;
1009 
1010 	/**  @pad: MBZ. */
1011 	__u32 pad;
1012 
1013 	/**
1014 	 * @label: User pointer to a NUL-terminated string
1015 	 *
1016 	 * Length cannot be greater than 4096
1017 	 */
1018 	__u64 label;
1019 };
1020 
1021 /**
1022  * struct drm_panthor_set_user_mmio_offset - Arguments passed to
1023  * DRM_IOCTL_PANTHOR_SET_USER_MMIO_OFFSET
1024  *
1025  * This ioctl is only really useful if you want to support userspace
1026  * CPU emulation environments where the size of an unsigned long differs
1027  * between the host and the guest architectures.
1028  */
1029 struct drm_panthor_set_user_mmio_offset {
1030 	/**
1031 	 * @offset: User MMIO offset to use.
1032 	 *
1033 	 * Must be either DRM_PANTHOR_USER_MMIO_OFFSET_32BIT or
1034 	 * DRM_PANTHOR_USER_MMIO_OFFSET_64BIT.
1035 	 *
1036 	 * Use DRM_PANTHOR_USER_MMIO_OFFSET (which selects OFFSET_32BIT or
1037 	 * OFFSET_64BIT based on the size of an unsigned long) unless you
1038 	 * have a very good reason to overrule this decision.
1039 	 */
1040 	__u64 offset;
1041 };
1042 
1043 /**
1044  * DRM_IOCTL_PANTHOR() - Build a Panthor IOCTL number
1045  * @__access: Access type. Must be R, W or RW.
1046  * @__id: One of the DRM_PANTHOR_xxx id.
1047  * @__type: Suffix of the type being passed to the IOCTL.
1048  *
1049  * Don't use this macro directly, use the DRM_IOCTL_PANTHOR_xxx
1050  * values instead.
1051  *
1052  * Return: An IOCTL number to be passed to ioctl() from userspace.
1053  */
1054 #define DRM_IOCTL_PANTHOR(__access, __id, __type) \
1055 	DRM_IO ## __access(DRM_COMMAND_BASE + DRM_PANTHOR_ ## __id, \
1056 			   struct drm_panthor_ ## __type)
1057 
1058 enum {
1059 	DRM_IOCTL_PANTHOR_DEV_QUERY =
1060 		DRM_IOCTL_PANTHOR(WR, DEV_QUERY, dev_query),
1061 	DRM_IOCTL_PANTHOR_VM_CREATE =
1062 		DRM_IOCTL_PANTHOR(WR, VM_CREATE, vm_create),
1063 	DRM_IOCTL_PANTHOR_VM_DESTROY =
1064 		DRM_IOCTL_PANTHOR(WR, VM_DESTROY, vm_destroy),
1065 	DRM_IOCTL_PANTHOR_VM_BIND =
1066 		DRM_IOCTL_PANTHOR(WR, VM_BIND, vm_bind),
1067 	DRM_IOCTL_PANTHOR_VM_GET_STATE =
1068 		DRM_IOCTL_PANTHOR(WR, VM_GET_STATE, vm_get_state),
1069 	DRM_IOCTL_PANTHOR_BO_CREATE =
1070 		DRM_IOCTL_PANTHOR(WR, BO_CREATE, bo_create),
1071 	DRM_IOCTL_PANTHOR_BO_MMAP_OFFSET =
1072 		DRM_IOCTL_PANTHOR(WR, BO_MMAP_OFFSET, bo_mmap_offset),
1073 	DRM_IOCTL_PANTHOR_GROUP_CREATE =
1074 		DRM_IOCTL_PANTHOR(WR, GROUP_CREATE, group_create),
1075 	DRM_IOCTL_PANTHOR_GROUP_DESTROY =
1076 		DRM_IOCTL_PANTHOR(WR, GROUP_DESTROY, group_destroy),
1077 	DRM_IOCTL_PANTHOR_GROUP_SUBMIT =
1078 		DRM_IOCTL_PANTHOR(WR, GROUP_SUBMIT, group_submit),
1079 	DRM_IOCTL_PANTHOR_GROUP_GET_STATE =
1080 		DRM_IOCTL_PANTHOR(WR, GROUP_GET_STATE, group_get_state),
1081 	DRM_IOCTL_PANTHOR_TILER_HEAP_CREATE =
1082 		DRM_IOCTL_PANTHOR(WR, TILER_HEAP_CREATE, tiler_heap_create),
1083 	DRM_IOCTL_PANTHOR_TILER_HEAP_DESTROY =
1084 		DRM_IOCTL_PANTHOR(WR, TILER_HEAP_DESTROY, tiler_heap_destroy),
1085 	DRM_IOCTL_PANTHOR_BO_SET_LABEL =
1086 		DRM_IOCTL_PANTHOR(WR, BO_SET_LABEL, bo_set_label),
1087 	DRM_IOCTL_PANTHOR_SET_USER_MMIO_OFFSET =
1088 		DRM_IOCTL_PANTHOR(WR, SET_USER_MMIO_OFFSET, set_user_mmio_offset),
1089 };
1090 
1091 #if defined(__cplusplus)
1092 }
1093 #endif
1094 
1095 #endif /* _PANTHOR_DRM_H_ */
1096