xref: /linux/include/uapi/drm/panthor_drm.h (revision 3fd6c59042dbba50391e30862beac979491145fe)
1 /* SPDX-License-Identifier: MIT */
2 /* Copyright (C) 2023 Collabora ltd. */
3 #ifndef _PANTHOR_DRM_H_
4 #define _PANTHOR_DRM_H_
5 
6 #include "drm.h"
7 
8 #if defined(__cplusplus)
9 extern "C" {
10 #endif
11 
12 /**
13  * DOC: Introduction
14  *
15  * This documentation describes the Panthor IOCTLs.
16  *
17  * Just a few generic rules about the data passed to the Panthor IOCTLs:
18  *
19  * - Structures must be aligned on 64-bit/8-byte. If the object is not
20  *   naturally aligned, a padding field must be added.
21  * - Fields must be explicitly aligned to their natural type alignment with
22  *   pad[0..N] fields.
23  * - All padding fields will be checked by the driver to make sure they are
24  *   zeroed.
25  * - Flags can be added, but not removed/replaced.
26  * - New fields can be added to the main structures (the structures
27  *   directly passed to the ioctl). Those fields can be added at the end of
28  *   the structure, or replace existing padding fields. Any new field being
29  *   added must preserve the behavior that existed before those fields were
30  *   added when a value of zero is passed.
31  * - New fields can be added to indirect objects (objects pointed by the
32  *   main structure), iff those objects are passed a size to reflect the
33  *   size known by the userspace driver (see drm_panthor_obj_array::stride
34  *   or drm_panthor_dev_query::size).
35  * - If the kernel driver is too old to know some fields, those will be
36  *   ignored if zero, and otherwise rejected (and so will be zero on output).
37  * - If userspace is too old to know some fields, those will be zeroed
38  *   (input) before the structure is parsed by the kernel driver.
39  * - Each new flag/field addition must come with a driver version update so
40  *   the userspace driver doesn't have to trial and error to know which
41  *   flags are supported.
42  * - Structures should not contain unions, as this would defeat the
43  *   extensibility of such structures.
44  * - IOCTLs can't be removed or replaced. New IOCTL IDs should be placed
45  *   at the end of the drm_panthor_ioctl_id enum.
46  */
47 
48 /**
49  * DOC: MMIO regions exposed to userspace.
50  *
51  * .. c:macro:: DRM_PANTHOR_USER_MMIO_OFFSET
52  *
53  * File offset for all MMIO regions being exposed to userspace. Don't use
54  * this value directly, use DRM_PANTHOR_USER_<name>_OFFSET values instead.
55  * pgoffset passed to mmap2() is an unsigned long, which forces us to use a
56  * different offset on 32-bit and 64-bit systems.
57  *
58  * .. c:macro:: DRM_PANTHOR_USER_FLUSH_ID_MMIO_OFFSET
59  *
60  * File offset for the LATEST_FLUSH_ID register. The Userspace driver controls
61  * GPU cache flushing through CS instructions, but the flush reduction
62  * mechanism requires a flush_id. This flush_id could be queried with an
63  * ioctl, but Arm provides a well-isolated register page containing only this
64  * read-only register, so let's expose this page through a static mmap offset
65  * and allow direct mapping of this MMIO region so we can avoid the
66  * user <-> kernel round-trip.
67  */
68 #define DRM_PANTHOR_USER_MMIO_OFFSET_32BIT	(1ull << 43)
69 #define DRM_PANTHOR_USER_MMIO_OFFSET_64BIT	(1ull << 56)
70 #define DRM_PANTHOR_USER_MMIO_OFFSET		(sizeof(unsigned long) < 8 ? \
71 						 DRM_PANTHOR_USER_MMIO_OFFSET_32BIT : \
72 						 DRM_PANTHOR_USER_MMIO_OFFSET_64BIT)
73 #define DRM_PANTHOR_USER_FLUSH_ID_MMIO_OFFSET	(DRM_PANTHOR_USER_MMIO_OFFSET | 0)
74 
75 /**
76  * DOC: IOCTL IDs
77  *
78  * enum drm_panthor_ioctl_id - IOCTL IDs
79  *
80  * Place new ioctls at the end, don't re-order, don't replace or remove entries.
81  *
82  * These IDs are not meant to be used directly. Use the DRM_IOCTL_PANTHOR_xxx
83  * definitions instead.
84  */
85 enum drm_panthor_ioctl_id {
86 	/** @DRM_PANTHOR_DEV_QUERY: Query device information. */
87 	DRM_PANTHOR_DEV_QUERY = 0,
88 
89 	/** @DRM_PANTHOR_VM_CREATE: Create a VM. */
90 	DRM_PANTHOR_VM_CREATE,
91 
92 	/** @DRM_PANTHOR_VM_DESTROY: Destroy a VM. */
93 	DRM_PANTHOR_VM_DESTROY,
94 
95 	/** @DRM_PANTHOR_VM_BIND: Bind/unbind memory to a VM. */
96 	DRM_PANTHOR_VM_BIND,
97 
98 	/** @DRM_PANTHOR_VM_GET_STATE: Get VM state. */
99 	DRM_PANTHOR_VM_GET_STATE,
100 
101 	/** @DRM_PANTHOR_BO_CREATE: Create a buffer object. */
102 	DRM_PANTHOR_BO_CREATE,
103 
104 	/**
105 	 * @DRM_PANTHOR_BO_MMAP_OFFSET: Get the file offset to pass to
106 	 * mmap to map a GEM object.
107 	 */
108 	DRM_PANTHOR_BO_MMAP_OFFSET,
109 
110 	/** @DRM_PANTHOR_GROUP_CREATE: Create a scheduling group. */
111 	DRM_PANTHOR_GROUP_CREATE,
112 
113 	/** @DRM_PANTHOR_GROUP_DESTROY: Destroy a scheduling group. */
114 	DRM_PANTHOR_GROUP_DESTROY,
115 
116 	/**
117 	 * @DRM_PANTHOR_GROUP_SUBMIT: Submit jobs to queues belonging
118 	 * to a specific scheduling group.
119 	 */
120 	DRM_PANTHOR_GROUP_SUBMIT,
121 
122 	/** @DRM_PANTHOR_GROUP_GET_STATE: Get the state of a scheduling group. */
123 	DRM_PANTHOR_GROUP_GET_STATE,
124 
125 	/** @DRM_PANTHOR_TILER_HEAP_CREATE: Create a tiler heap. */
126 	DRM_PANTHOR_TILER_HEAP_CREATE,
127 
128 	/** @DRM_PANTHOR_TILER_HEAP_DESTROY: Destroy a tiler heap. */
129 	DRM_PANTHOR_TILER_HEAP_DESTROY,
130 };
131 
132 /**
133  * DRM_IOCTL_PANTHOR() - Build a Panthor IOCTL number
134  * @__access: Access type. Must be R, W or RW.
135  * @__id: One of the DRM_PANTHOR_xxx id.
136  * @__type: Suffix of the type being passed to the IOCTL.
137  *
138  * Don't use this macro directly, use the DRM_IOCTL_PANTHOR_xxx
139  * values instead.
140  *
141  * Return: An IOCTL number to be passed to ioctl() from userspace.
142  */
143 #define DRM_IOCTL_PANTHOR(__access, __id, __type) \
144 	DRM_IO ## __access(DRM_COMMAND_BASE + DRM_PANTHOR_ ## __id, \
145 			   struct drm_panthor_ ## __type)
146 
147 #define DRM_IOCTL_PANTHOR_DEV_QUERY \
148 	DRM_IOCTL_PANTHOR(WR, DEV_QUERY, dev_query)
149 #define DRM_IOCTL_PANTHOR_VM_CREATE \
150 	DRM_IOCTL_PANTHOR(WR, VM_CREATE, vm_create)
151 #define DRM_IOCTL_PANTHOR_VM_DESTROY \
152 	DRM_IOCTL_PANTHOR(WR, VM_DESTROY, vm_destroy)
153 #define DRM_IOCTL_PANTHOR_VM_BIND \
154 	DRM_IOCTL_PANTHOR(WR, VM_BIND, vm_bind)
155 #define DRM_IOCTL_PANTHOR_VM_GET_STATE \
156 	DRM_IOCTL_PANTHOR(WR, VM_GET_STATE, vm_get_state)
157 #define DRM_IOCTL_PANTHOR_BO_CREATE \
158 	DRM_IOCTL_PANTHOR(WR, BO_CREATE, bo_create)
159 #define DRM_IOCTL_PANTHOR_BO_MMAP_OFFSET \
160 	DRM_IOCTL_PANTHOR(WR, BO_MMAP_OFFSET, bo_mmap_offset)
161 #define DRM_IOCTL_PANTHOR_GROUP_CREATE \
162 	DRM_IOCTL_PANTHOR(WR, GROUP_CREATE, group_create)
163 #define DRM_IOCTL_PANTHOR_GROUP_DESTROY \
164 	DRM_IOCTL_PANTHOR(WR, GROUP_DESTROY, group_destroy)
165 #define DRM_IOCTL_PANTHOR_GROUP_SUBMIT \
166 	DRM_IOCTL_PANTHOR(WR, GROUP_SUBMIT, group_submit)
167 #define DRM_IOCTL_PANTHOR_GROUP_GET_STATE \
168 	DRM_IOCTL_PANTHOR(WR, GROUP_GET_STATE, group_get_state)
169 #define DRM_IOCTL_PANTHOR_TILER_HEAP_CREATE \
170 	DRM_IOCTL_PANTHOR(WR, TILER_HEAP_CREATE, tiler_heap_create)
171 #define DRM_IOCTL_PANTHOR_TILER_HEAP_DESTROY \
172 	DRM_IOCTL_PANTHOR(WR, TILER_HEAP_DESTROY, tiler_heap_destroy)
173 
174 /**
175  * DOC: IOCTL arguments
176  */
177 
178 /**
179  * struct drm_panthor_obj_array - Object array.
180  *
181  * This object is used to pass an array of objects whose size is subject to changes in
182  * future versions of the driver. In order to support this mutability, we pass a stride
183  * describing the size of the object as known by userspace.
184  *
185  * You shouldn't fill drm_panthor_obj_array fields directly. You should instead use
186  * the DRM_PANTHOR_OBJ_ARRAY() macro that takes care of initializing the stride to
187  * the object size.
188  */
189 struct drm_panthor_obj_array {
190 	/** @stride: Stride of object struct. Used for versioning. */
191 	__u32 stride;
192 
193 	/** @count: Number of objects in the array. */
194 	__u32 count;
195 
196 	/** @array: User pointer to an array of objects. */
197 	__u64 array;
198 };
199 
200 /**
201  * DRM_PANTHOR_OBJ_ARRAY() - Initialize a drm_panthor_obj_array field.
202  * @cnt: Number of elements in the array.
203  * @ptr: Pointer to the array to pass to the kernel.
204  *
205  * Macro initializing a drm_panthor_obj_array based on the object size as known
206  * by userspace.
207  */
208 #define DRM_PANTHOR_OBJ_ARRAY(cnt, ptr) \
209 	{ .stride = sizeof((ptr)[0]), .count = (cnt), .array = (__u64)(uintptr_t)(ptr) }
210 
211 /**
212  * enum drm_panthor_sync_op_flags - Synchronization operation flags.
213  */
214 enum drm_panthor_sync_op_flags {
215 	/** @DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_MASK: Synchronization handle type mask. */
216 	DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_MASK = 0xff,
217 
218 	/** @DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_SYNCOBJ: Synchronization object type. */
219 	DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_SYNCOBJ = 0,
220 
221 	/**
222 	 * @DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_TIMELINE_SYNCOBJ: Timeline synchronization
223 	 * object type.
224 	 */
225 	DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_TIMELINE_SYNCOBJ = 1,
226 
227 	/** @DRM_PANTHOR_SYNC_OP_WAIT: Wait operation. */
228 	DRM_PANTHOR_SYNC_OP_WAIT = 0 << 31,
229 
230 	/** @DRM_PANTHOR_SYNC_OP_SIGNAL: Signal operation. */
231 	DRM_PANTHOR_SYNC_OP_SIGNAL = (int)(1u << 31),
232 };
233 
234 /**
235  * struct drm_panthor_sync_op - Synchronization operation.
236  */
237 struct drm_panthor_sync_op {
238 	/** @flags: Synchronization operation flags. Combination of DRM_PANTHOR_SYNC_OP values. */
239 	__u32 flags;
240 
241 	/** @handle: Sync handle. */
242 	__u32 handle;
243 
244 	/**
245 	 * @timeline_value: MBZ if
246 	 * (flags & DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_MASK) !=
247 	 * DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_TIMELINE_SYNCOBJ.
248 	 */
249 	__u64 timeline_value;
250 };
251 
252 /**
253  * enum drm_panthor_dev_query_type - Query type
254  *
255  * Place new types at the end, don't re-order, don't remove or replace.
256  */
257 enum drm_panthor_dev_query_type {
258 	/** @DRM_PANTHOR_DEV_QUERY_GPU_INFO: Query GPU information. */
259 	DRM_PANTHOR_DEV_QUERY_GPU_INFO = 0,
260 
261 	/** @DRM_PANTHOR_DEV_QUERY_CSIF_INFO: Query command-stream interface information. */
262 	DRM_PANTHOR_DEV_QUERY_CSIF_INFO,
263 
264 	/** @DRM_PANTHOR_DEV_QUERY_TIMESTAMP_INFO: Query timestamp information. */
265 	DRM_PANTHOR_DEV_QUERY_TIMESTAMP_INFO,
266 
267 	/**
268 	 * @DRM_PANTHOR_DEV_QUERY_GROUP_PRIORITIES_INFO: Query allowed group priorities information.
269 	 */
270 	DRM_PANTHOR_DEV_QUERY_GROUP_PRIORITIES_INFO,
271 };
272 
273 /**
274  * struct drm_panthor_gpu_info - GPU information
275  *
276  * Structure grouping all queryable information relating to the GPU.
277  */
278 struct drm_panthor_gpu_info {
279 	/** @gpu_id : GPU ID. */
280 	__u32 gpu_id;
281 #define DRM_PANTHOR_ARCH_MAJOR(x)		((x) >> 28)
282 #define DRM_PANTHOR_ARCH_MINOR(x)		(((x) >> 24) & 0xf)
283 #define DRM_PANTHOR_ARCH_REV(x)			(((x) >> 20) & 0xf)
284 #define DRM_PANTHOR_PRODUCT_MAJOR(x)		(((x) >> 16) & 0xf)
285 #define DRM_PANTHOR_VERSION_MAJOR(x)		(((x) >> 12) & 0xf)
286 #define DRM_PANTHOR_VERSION_MINOR(x)		(((x) >> 4) & 0xff)
287 #define DRM_PANTHOR_VERSION_STATUS(x)		((x) & 0xf)
288 
289 	/** @gpu_rev: GPU revision. */
290 	__u32 gpu_rev;
291 
292 	/** @csf_id: Command stream frontend ID. */
293 	__u32 csf_id;
294 #define DRM_PANTHOR_CSHW_MAJOR(x)		(((x) >> 26) & 0x3f)
295 #define DRM_PANTHOR_CSHW_MINOR(x)		(((x) >> 20) & 0x3f)
296 #define DRM_PANTHOR_CSHW_REV(x)			(((x) >> 16) & 0xf)
297 #define DRM_PANTHOR_MCU_MAJOR(x)		(((x) >> 10) & 0x3f)
298 #define DRM_PANTHOR_MCU_MINOR(x)		(((x) >> 4) & 0x3f)
299 #define DRM_PANTHOR_MCU_REV(x)			((x) & 0xf)
300 
301 	/** @l2_features: L2-cache features. */
302 	__u32 l2_features;
303 
304 	/** @tiler_features: Tiler features. */
305 	__u32 tiler_features;
306 
307 	/** @mem_features: Memory features. */
308 	__u32 mem_features;
309 
310 	/** @mmu_features: MMU features. */
311 	__u32 mmu_features;
312 #define DRM_PANTHOR_MMU_VA_BITS(x)		((x) & 0xff)
313 
314 	/** @thread_features: Thread features. */
315 	__u32 thread_features;
316 
317 	/** @max_threads: Maximum number of threads. */
318 	__u32 max_threads;
319 
320 	/** @thread_max_workgroup_size: Maximum workgroup size. */
321 	__u32 thread_max_workgroup_size;
322 
323 	/**
324 	 * @thread_max_barrier_size: Maximum number of threads that can wait
325 	 * simultaneously on a barrier.
326 	 */
327 	__u32 thread_max_barrier_size;
328 
329 	/** @coherency_features: Coherency features. */
330 	__u32 coherency_features;
331 
332 	/** @texture_features: Texture features. */
333 	__u32 texture_features[4];
334 
335 	/** @as_present: Bitmask encoding the number of address-space exposed by the MMU. */
336 	__u32 as_present;
337 
338 	/** @shader_present: Bitmask encoding the shader cores exposed by the GPU. */
339 	__u64 shader_present;
340 
341 	/** @l2_present: Bitmask encoding the L2 caches exposed by the GPU. */
342 	__u64 l2_present;
343 
344 	/** @tiler_present: Bitmask encoding the tiler units exposed by the GPU. */
345 	__u64 tiler_present;
346 
347 	/** @core_features: Used to discriminate core variants when they exist. */
348 	__u32 core_features;
349 
350 	/** @pad: MBZ. */
351 	__u32 pad;
352 };
353 
354 /**
355  * struct drm_panthor_csif_info - Command stream interface information
356  *
357  * Structure grouping all queryable information relating to the command stream interface.
358  */
359 struct drm_panthor_csif_info {
360 	/** @csg_slot_count: Number of command stream group slots exposed by the firmware. */
361 	__u32 csg_slot_count;
362 
363 	/** @cs_slot_count: Number of command stream slots per group. */
364 	__u32 cs_slot_count;
365 
366 	/** @cs_reg_count: Number of command stream registers. */
367 	__u32 cs_reg_count;
368 
369 	/** @scoreboard_slot_count: Number of scoreboard slots. */
370 	__u32 scoreboard_slot_count;
371 
372 	/**
373 	 * @unpreserved_cs_reg_count: Number of command stream registers reserved by
374 	 * the kernel driver to call a userspace command stream.
375 	 *
376 	 * All registers can be used by a userspace command stream, but the
377 	 * [cs_slot_count - unpreserved_cs_reg_count .. cs_slot_count] registers are
378 	 * used by the kernel when DRM_PANTHOR_IOCTL_GROUP_SUBMIT is called.
379 	 */
380 	__u32 unpreserved_cs_reg_count;
381 
382 	/**
383 	 * @pad: Padding field, set to zero.
384 	 */
385 	__u32 pad;
386 };
387 
388 /**
389  * struct drm_panthor_timestamp_info - Timestamp information
390  *
391  * Structure grouping all queryable information relating to the GPU timestamp.
392  */
393 struct drm_panthor_timestamp_info {
394 	/**
395 	 * @timestamp_frequency: The frequency of the timestamp timer or 0 if
396 	 * unknown.
397 	 */
398 	__u64 timestamp_frequency;
399 
400 	/** @current_timestamp: The current timestamp. */
401 	__u64 current_timestamp;
402 
403 	/** @timestamp_offset: The offset of the timestamp timer. */
404 	__u64 timestamp_offset;
405 };
406 
407 /**
408  * struct drm_panthor_group_priorities_info - Group priorities information
409  *
410  * Structure grouping all queryable information relating to the allowed group priorities.
411  */
412 struct drm_panthor_group_priorities_info {
413 	/**
414 	 * @allowed_mask: Bitmask of the allowed group priorities.
415 	 *
416 	 * Each bit represents a variant of the enum drm_panthor_group_priority.
417 	 */
418 	__u8 allowed_mask;
419 
420 	/** @pad: Padding fields, MBZ. */
421 	__u8 pad[3];
422 };
423 
424 /**
425  * struct drm_panthor_dev_query - Arguments passed to DRM_PANTHOR_IOCTL_DEV_QUERY
426  */
427 struct drm_panthor_dev_query {
428 	/** @type: the query type (see drm_panthor_dev_query_type). */
429 	__u32 type;
430 
431 	/**
432 	 * @size: size of the type being queried.
433 	 *
434 	 * If pointer is NULL, size is updated by the driver to provide the
435 	 * output structure size. If pointer is not NULL, the driver will
436 	 * only copy min(size, actual_structure_size) bytes to the pointer,
437 	 * and update the size accordingly. This allows us to extend query
438 	 * types without breaking userspace.
439 	 */
440 	__u32 size;
441 
442 	/**
443 	 * @pointer: user pointer to a query type struct.
444 	 *
445 	 * Pointer can be NULL, in which case, nothing is copied, but the
446 	 * actual structure size is returned. If not NULL, it must point to
447 	 * a location that's large enough to hold size bytes.
448 	 */
449 	__u64 pointer;
450 };
451 
452 /**
453  * struct drm_panthor_vm_create - Arguments passed to DRM_PANTHOR_IOCTL_VM_CREATE
454  */
455 struct drm_panthor_vm_create {
456 	/** @flags: VM flags, MBZ. */
457 	__u32 flags;
458 
459 	/** @id: Returned VM ID. */
460 	__u32 id;
461 
462 	/**
463 	 * @user_va_range: Size of the VA space reserved for user objects.
464 	 *
465 	 * The kernel will pick the remaining space to map kernel-only objects to the
466 	 * VM (heap chunks, heap context, ring buffers, kernel synchronization objects,
467 	 * ...). If the space left for kernel objects is too small, kernel object
468 	 * allocation will fail further down the road. One can use
469 	 * drm_panthor_gpu_info::mmu_features to extract the total virtual address
470 	 * range, and chose a user_va_range that leaves some space to the kernel.
471 	 *
472 	 * If user_va_range is zero, the kernel will pick a sensible value based on
473 	 * TASK_SIZE and the virtual range supported by the GPU MMU (the kernel/user
474 	 * split should leave enough VA space for userspace processes to support SVM,
475 	 * while still allowing the kernel to map some amount of kernel objects in
476 	 * the kernel VA range). The value chosen by the driver will be returned in
477 	 * @user_va_range.
478 	 *
479 	 * User VA space always starts at 0x0, kernel VA space is always placed after
480 	 * the user VA range.
481 	 */
482 	__u64 user_va_range;
483 };
484 
485 /**
486  * struct drm_panthor_vm_destroy - Arguments passed to DRM_PANTHOR_IOCTL_VM_DESTROY
487  */
488 struct drm_panthor_vm_destroy {
489 	/** @id: ID of the VM to destroy. */
490 	__u32 id;
491 
492 	/** @pad: MBZ. */
493 	__u32 pad;
494 };
495 
496 /**
497  * enum drm_panthor_vm_bind_op_flags - VM bind operation flags
498  */
499 enum drm_panthor_vm_bind_op_flags {
500 	/**
501 	 * @DRM_PANTHOR_VM_BIND_OP_MAP_READONLY: Map the memory read-only.
502 	 *
503 	 * Only valid with DRM_PANTHOR_VM_BIND_OP_TYPE_MAP.
504 	 */
505 	DRM_PANTHOR_VM_BIND_OP_MAP_READONLY = 1 << 0,
506 
507 	/**
508 	 * @DRM_PANTHOR_VM_BIND_OP_MAP_NOEXEC: Map the memory not-executable.
509 	 *
510 	 * Only valid with DRM_PANTHOR_VM_BIND_OP_TYPE_MAP.
511 	 */
512 	DRM_PANTHOR_VM_BIND_OP_MAP_NOEXEC = 1 << 1,
513 
514 	/**
515 	 * @DRM_PANTHOR_VM_BIND_OP_MAP_UNCACHED: Map the memory uncached.
516 	 *
517 	 * Only valid with DRM_PANTHOR_VM_BIND_OP_TYPE_MAP.
518 	 */
519 	DRM_PANTHOR_VM_BIND_OP_MAP_UNCACHED = 1 << 2,
520 
521 	/**
522 	 * @DRM_PANTHOR_VM_BIND_OP_TYPE_MASK: Mask used to determine the type of operation.
523 	 */
524 	DRM_PANTHOR_VM_BIND_OP_TYPE_MASK = (int)(0xfu << 28),
525 
526 	/** @DRM_PANTHOR_VM_BIND_OP_TYPE_MAP: Map operation. */
527 	DRM_PANTHOR_VM_BIND_OP_TYPE_MAP = 0 << 28,
528 
529 	/** @DRM_PANTHOR_VM_BIND_OP_TYPE_UNMAP: Unmap operation. */
530 	DRM_PANTHOR_VM_BIND_OP_TYPE_UNMAP = 1 << 28,
531 
532 	/**
533 	 * @DRM_PANTHOR_VM_BIND_OP_TYPE_SYNC_ONLY: No VM operation.
534 	 *
535 	 * Just serves as a synchronization point on a VM queue.
536 	 *
537 	 * Only valid if %DRM_PANTHOR_VM_BIND_ASYNC is set in drm_panthor_vm_bind::flags,
538 	 * and drm_panthor_vm_bind_op::syncs contains at least one element.
539 	 */
540 	DRM_PANTHOR_VM_BIND_OP_TYPE_SYNC_ONLY = 2 << 28,
541 };
542 
543 /**
544  * struct drm_panthor_vm_bind_op - VM bind operation
545  */
546 struct drm_panthor_vm_bind_op {
547 	/** @flags: Combination of drm_panthor_vm_bind_op_flags flags. */
548 	__u32 flags;
549 
550 	/**
551 	 * @bo_handle: Handle of the buffer object to map.
552 	 * MBZ for unmap or sync-only operations.
553 	 */
554 	__u32 bo_handle;
555 
556 	/**
557 	 * @bo_offset: Buffer object offset.
558 	 * MBZ for unmap or sync-only operations.
559 	 */
560 	__u64 bo_offset;
561 
562 	/**
563 	 * @va: Virtual address to map/unmap.
564 	 * MBZ for sync-only operations.
565 	 */
566 	__u64 va;
567 
568 	/**
569 	 * @size: Size to map/unmap.
570 	 * MBZ for sync-only operations.
571 	 */
572 	__u64 size;
573 
574 	/**
575 	 * @syncs: Array of struct drm_panthor_sync_op synchronization
576 	 * operations.
577 	 *
578 	 * This array must be empty if %DRM_PANTHOR_VM_BIND_ASYNC is not set on
579 	 * the drm_panthor_vm_bind object containing this VM bind operation.
580 	 *
581 	 * This array shall not be empty for sync-only operations.
582 	 */
583 	struct drm_panthor_obj_array syncs;
584 
585 };
586 
587 /**
588  * enum drm_panthor_vm_bind_flags - VM bind flags
589  */
590 enum drm_panthor_vm_bind_flags {
591 	/**
592 	 * @DRM_PANTHOR_VM_BIND_ASYNC: VM bind operations are queued to the VM
593 	 * queue instead of being executed synchronously.
594 	 */
595 	DRM_PANTHOR_VM_BIND_ASYNC = 1 << 0,
596 };
597 
598 /**
599  * struct drm_panthor_vm_bind - Arguments passed to DRM_IOCTL_PANTHOR_VM_BIND
600  */
601 struct drm_panthor_vm_bind {
602 	/** @vm_id: VM targeted by the bind request. */
603 	__u32 vm_id;
604 
605 	/** @flags: Combination of drm_panthor_vm_bind_flags flags. */
606 	__u32 flags;
607 
608 	/** @ops: Array of struct drm_panthor_vm_bind_op bind operations. */
609 	struct drm_panthor_obj_array ops;
610 };
611 
612 /**
613  * enum drm_panthor_vm_state - VM states.
614  */
615 enum drm_panthor_vm_state {
616 	/**
617 	 * @DRM_PANTHOR_VM_STATE_USABLE: VM is usable.
618 	 *
619 	 * New VM operations will be accepted on this VM.
620 	 */
621 	DRM_PANTHOR_VM_STATE_USABLE,
622 
623 	/**
624 	 * @DRM_PANTHOR_VM_STATE_UNUSABLE: VM is unusable.
625 	 *
626 	 * Something put the VM in an unusable state (like an asynchronous
627 	 * VM_BIND request failing for any reason).
628 	 *
629 	 * Once the VM is in this state, all new MAP operations will be
630 	 * rejected, and any GPU job targeting this VM will fail.
631 	 * UNMAP operations are still accepted.
632 	 *
633 	 * The only way to recover from an unusable VM is to create a new
634 	 * VM, and destroy the old one.
635 	 */
636 	DRM_PANTHOR_VM_STATE_UNUSABLE,
637 };
638 
639 /**
640  * struct drm_panthor_vm_get_state - Get VM state.
641  */
642 struct drm_panthor_vm_get_state {
643 	/** @vm_id: VM targeted by the get_state request. */
644 	__u32 vm_id;
645 
646 	/**
647 	 * @state: state returned by the driver.
648 	 *
649 	 * Must be one of the enum drm_panthor_vm_state values.
650 	 */
651 	__u32 state;
652 };
653 
654 /**
655  * enum drm_panthor_bo_flags - Buffer object flags, passed at creation time.
656  */
657 enum drm_panthor_bo_flags {
658 	/** @DRM_PANTHOR_BO_NO_MMAP: The buffer object will never be CPU-mapped in userspace. */
659 	DRM_PANTHOR_BO_NO_MMAP = (1 << 0),
660 };
661 
662 /**
663  * struct drm_panthor_bo_create - Arguments passed to DRM_IOCTL_PANTHOR_BO_CREATE.
664  */
665 struct drm_panthor_bo_create {
666 	/**
667 	 * @size: Requested size for the object
668 	 *
669 	 * The (page-aligned) allocated size for the object will be returned.
670 	 */
671 	__u64 size;
672 
673 	/**
674 	 * @flags: Flags. Must be a combination of drm_panthor_bo_flags flags.
675 	 */
676 	__u32 flags;
677 
678 	/**
679 	 * @exclusive_vm_id: Exclusive VM this buffer object will be mapped to.
680 	 *
681 	 * If not zero, the field must refer to a valid VM ID, and implies that:
682 	 *  - the buffer object will only ever be bound to that VM
683 	 *  - cannot be exported as a PRIME fd
684 	 */
685 	__u32 exclusive_vm_id;
686 
687 	/**
688 	 * @handle: Returned handle for the object.
689 	 *
690 	 * Object handles are nonzero.
691 	 */
692 	__u32 handle;
693 
694 	/** @pad: MBZ. */
695 	__u32 pad;
696 };
697 
698 /**
699  * struct drm_panthor_bo_mmap_offset - Arguments passed to DRM_IOCTL_PANTHOR_BO_MMAP_OFFSET.
700  */
701 struct drm_panthor_bo_mmap_offset {
702 	/** @handle: Handle of the object we want an mmap offset for. */
703 	__u32 handle;
704 
705 	/** @pad: MBZ. */
706 	__u32 pad;
707 
708 	/** @offset: The fake offset to use for subsequent mmap calls. */
709 	__u64 offset;
710 };
711 
712 /**
713  * struct drm_panthor_queue_create - Queue creation arguments.
714  */
715 struct drm_panthor_queue_create {
716 	/**
717 	 * @priority: Defines the priority of queues inside a group. Goes from 0 to 15,
718 	 * 15 being the highest priority.
719 	 */
720 	__u8 priority;
721 
722 	/** @pad: Padding fields, MBZ. */
723 	__u8 pad[3];
724 
725 	/** @ringbuf_size: Size of the ring buffer to allocate to this queue. */
726 	__u32 ringbuf_size;
727 };
728 
729 /**
730  * enum drm_panthor_group_priority - Scheduling group priority
731  */
732 enum drm_panthor_group_priority {
733 	/** @PANTHOR_GROUP_PRIORITY_LOW: Low priority group. */
734 	PANTHOR_GROUP_PRIORITY_LOW = 0,
735 
736 	/** @PANTHOR_GROUP_PRIORITY_MEDIUM: Medium priority group. */
737 	PANTHOR_GROUP_PRIORITY_MEDIUM,
738 
739 	/**
740 	 * @PANTHOR_GROUP_PRIORITY_HIGH: High priority group.
741 	 *
742 	 * Requires CAP_SYS_NICE or DRM_MASTER.
743 	 */
744 	PANTHOR_GROUP_PRIORITY_HIGH,
745 
746 	/**
747 	 * @PANTHOR_GROUP_PRIORITY_REALTIME: Realtime priority group.
748 	 *
749 	 * Requires CAP_SYS_NICE or DRM_MASTER.
750 	 */
751 	PANTHOR_GROUP_PRIORITY_REALTIME,
752 };
753 
754 /**
755  * struct drm_panthor_group_create - Arguments passed to DRM_IOCTL_PANTHOR_GROUP_CREATE
756  */
757 struct drm_panthor_group_create {
758 	/** @queues: Array of drm_panthor_queue_create elements. */
759 	struct drm_panthor_obj_array queues;
760 
761 	/**
762 	 * @max_compute_cores: Maximum number of cores that can be used by compute
763 	 * jobs across CS queues bound to this group.
764 	 *
765 	 * Must be less or equal to the number of bits set in @compute_core_mask.
766 	 */
767 	__u8 max_compute_cores;
768 
769 	/**
770 	 * @max_fragment_cores: Maximum number of cores that can be used by fragment
771 	 * jobs across CS queues bound to this group.
772 	 *
773 	 * Must be less or equal to the number of bits set in @fragment_core_mask.
774 	 */
775 	__u8 max_fragment_cores;
776 
777 	/**
778 	 * @max_tiler_cores: Maximum number of tilers that can be used by tiler jobs
779 	 * across CS queues bound to this group.
780 	 *
781 	 * Must be less or equal to the number of bits set in @tiler_core_mask.
782 	 */
783 	__u8 max_tiler_cores;
784 
785 	/** @priority: Group priority (see enum drm_panthor_group_priority). */
786 	__u8 priority;
787 
788 	/** @pad: Padding field, MBZ. */
789 	__u32 pad;
790 
791 	/**
792 	 * @compute_core_mask: Mask encoding cores that can be used for compute jobs.
793 	 *
794 	 * This field must have at least @max_compute_cores bits set.
795 	 *
796 	 * The bits set here should also be set in drm_panthor_gpu_info::shader_present.
797 	 */
798 	__u64 compute_core_mask;
799 
800 	/**
801 	 * @fragment_core_mask: Mask encoding cores that can be used for fragment jobs.
802 	 *
803 	 * This field must have at least @max_fragment_cores bits set.
804 	 *
805 	 * The bits set here should also be set in drm_panthor_gpu_info::shader_present.
806 	 */
807 	__u64 fragment_core_mask;
808 
809 	/**
810 	 * @tiler_core_mask: Mask encoding cores that can be used for tiler jobs.
811 	 *
812 	 * This field must have at least @max_tiler_cores bits set.
813 	 *
814 	 * The bits set here should also be set in drm_panthor_gpu_info::tiler_present.
815 	 */
816 	__u64 tiler_core_mask;
817 
818 	/**
819 	 * @vm_id: VM ID to bind this group to.
820 	 *
821 	 * All submission to queues bound to this group will use this VM.
822 	 */
823 	__u32 vm_id;
824 
825 	/**
826 	 * @group_handle: Returned group handle. Passed back when submitting jobs or
827 	 * destroying a group.
828 	 */
829 	__u32 group_handle;
830 };
831 
832 /**
833  * struct drm_panthor_group_destroy - Arguments passed to DRM_IOCTL_PANTHOR_GROUP_DESTROY
834  */
835 struct drm_panthor_group_destroy {
836 	/** @group_handle: Group to destroy */
837 	__u32 group_handle;
838 
839 	/** @pad: Padding field, MBZ. */
840 	__u32 pad;
841 };
842 
843 /**
844  * struct drm_panthor_queue_submit - Job submission arguments.
845  *
846  * This is describing the userspace command stream to call from the kernel
847  * command stream ring-buffer. Queue submission is always part of a group
848  * submission, taking one or more jobs to submit to the underlying queues.
849  */
850 struct drm_panthor_queue_submit {
851 	/** @queue_index: Index of the queue inside a group. */
852 	__u32 queue_index;
853 
854 	/**
855 	 * @stream_size: Size of the command stream to execute.
856 	 *
857 	 * Must be 64-bit/8-byte aligned (the size of a CS instruction)
858 	 *
859 	 * Can be zero if stream_addr is zero too.
860 	 *
861 	 * When the stream size is zero, the queue submit serves as a
862 	 * synchronization point.
863 	 */
864 	__u32 stream_size;
865 
866 	/**
867 	 * @stream_addr: GPU address of the command stream to execute.
868 	 *
869 	 * Must be aligned on 64-byte.
870 	 *
871 	 * Can be zero is stream_size is zero too.
872 	 */
873 	__u64 stream_addr;
874 
875 	/**
876 	 * @latest_flush: FLUSH_ID read at the time the stream was built.
877 	 *
878 	 * This allows cache flush elimination for the automatic
879 	 * flush+invalidate(all) done at submission time, which is needed to
880 	 * ensure the GPU doesn't get garbage when reading the indirect command
881 	 * stream buffers. If you want the cache flush to happen
882 	 * unconditionally, pass a zero here.
883 	 *
884 	 * Ignored when stream_size is zero.
885 	 */
886 	__u32 latest_flush;
887 
888 	/** @pad: MBZ. */
889 	__u32 pad;
890 
891 	/** @syncs: Array of struct drm_panthor_sync_op sync operations. */
892 	struct drm_panthor_obj_array syncs;
893 };
894 
895 /**
896  * struct drm_panthor_group_submit - Arguments passed to DRM_IOCTL_PANTHOR_GROUP_SUBMIT
897  */
898 struct drm_panthor_group_submit {
899 	/** @group_handle: Handle of the group to queue jobs to. */
900 	__u32 group_handle;
901 
902 	/** @pad: MBZ. */
903 	__u32 pad;
904 
905 	/** @queue_submits: Array of drm_panthor_queue_submit objects. */
906 	struct drm_panthor_obj_array queue_submits;
907 };
908 
909 /**
910  * enum drm_panthor_group_state_flags - Group state flags
911  */
912 enum drm_panthor_group_state_flags {
913 	/**
914 	 * @DRM_PANTHOR_GROUP_STATE_TIMEDOUT: Group had unfinished jobs.
915 	 *
916 	 * When a group ends up with this flag set, no jobs can be submitted to its queues.
917 	 */
918 	DRM_PANTHOR_GROUP_STATE_TIMEDOUT = 1 << 0,
919 
920 	/**
921 	 * @DRM_PANTHOR_GROUP_STATE_FATAL_FAULT: Group had fatal faults.
922 	 *
923 	 * When a group ends up with this flag set, no jobs can be submitted to its queues.
924 	 */
925 	DRM_PANTHOR_GROUP_STATE_FATAL_FAULT = 1 << 1,
926 };
927 
928 /**
929  * struct drm_panthor_group_get_state - Arguments passed to DRM_IOCTL_PANTHOR_GROUP_GET_STATE
930  *
931  * Used to query the state of a group and decide whether a new group should be created to
932  * replace it.
933  */
934 struct drm_panthor_group_get_state {
935 	/** @group_handle: Handle of the group to query state on */
936 	__u32 group_handle;
937 
938 	/**
939 	 * @state: Combination of DRM_PANTHOR_GROUP_STATE_* flags encoding the
940 	 * group state.
941 	 */
942 	__u32 state;
943 
944 	/** @fatal_queues: Bitmask of queues that faced fatal faults. */
945 	__u32 fatal_queues;
946 
947 	/** @pad: MBZ */
948 	__u32 pad;
949 };
950 
951 /**
952  * struct drm_panthor_tiler_heap_create - Arguments passed to DRM_IOCTL_PANTHOR_TILER_HEAP_CREATE
953  */
954 struct drm_panthor_tiler_heap_create {
955 	/** @vm_id: VM ID the tiler heap should be mapped to */
956 	__u32 vm_id;
957 
958 	/** @initial_chunk_count: Initial number of chunks to allocate. Must be at least one. */
959 	__u32 initial_chunk_count;
960 
961 	/**
962 	 * @chunk_size: Chunk size.
963 	 *
964 	 * Must be page-aligned and lie in the [128k:8M] range.
965 	 */
966 	__u32 chunk_size;
967 
968 	/**
969 	 * @max_chunks: Maximum number of chunks that can be allocated.
970 	 *
971 	 * Must be at least @initial_chunk_count.
972 	 */
973 	__u32 max_chunks;
974 
975 	/**
976 	 * @target_in_flight: Maximum number of in-flight render passes.
977 	 *
978 	 * If the heap has more than tiler jobs in-flight, the FW will wait for render
979 	 * passes to finish before queuing new tiler jobs.
980 	 */
981 	__u32 target_in_flight;
982 
983 	/** @handle: Returned heap handle. Passed back to DESTROY_TILER_HEAP. */
984 	__u32 handle;
985 
986 	/** @tiler_heap_ctx_gpu_va: Returned heap GPU virtual address returned */
987 	__u64 tiler_heap_ctx_gpu_va;
988 
989 	/**
990 	 * @first_heap_chunk_gpu_va: First heap chunk.
991 	 *
992 	 * The tiler heap is formed of heap chunks forming a single-link list. This
993 	 * is the first element in the list.
994 	 */
995 	__u64 first_heap_chunk_gpu_va;
996 };
997 
998 /**
999  * struct drm_panthor_tiler_heap_destroy - Arguments passed to DRM_IOCTL_PANTHOR_TILER_HEAP_DESTROY
1000  */
1001 struct drm_panthor_tiler_heap_destroy {
1002 	/**
1003 	 * @handle: Handle of the tiler heap to destroy.
1004 	 *
1005 	 * Must be a valid heap handle returned by DRM_IOCTL_PANTHOR_TILER_HEAP_CREATE.
1006 	 */
1007 	__u32 handle;
1008 
1009 	/** @pad: Padding field, MBZ. */
1010 	__u32 pad;
1011 };
1012 
1013 #if defined(__cplusplus)
1014 }
1015 #endif
1016 
1017 #endif /* _PANTHOR_DRM_H_ */
1018