1 /* SPDX-License-Identifier: (GPL-2.0-only WITH Linux-syscall-note) OR MIT */ 2 /* Copyright (c) 2023 Imagination Technologies Ltd. */ 3 4 #ifndef PVR_DRM_UAPI_H 5 #define PVR_DRM_UAPI_H 6 7 #include "drm.h" 8 9 #include <linux/const.h> 10 #include <linux/types.h> 11 12 #if defined(__cplusplus) 13 extern "C" { 14 #endif 15 16 /** 17 * DOC: PowerVR UAPI 18 * 19 * The PowerVR IOCTL argument structs have a few limitations in place, in 20 * addition to the standard kernel restrictions: 21 * 22 * - All members must be type-aligned. 23 * - The overall struct must be padded to 64-bit alignment. 24 * - Explicit padding is almost always required. This takes the form of 25 * ``_padding_[x]`` members of sufficient size to pad to the next power-of-two 26 * alignment, where [x] is the offset into the struct in hexadecimal. Arrays 27 * are never used for alignment. Padding fields must be zeroed; this is 28 * always checked. 29 * - Unions may only appear as the last member of a struct. 30 * - Individual union members may grow in the future. The space between the 31 * end of a union member and the end of its containing union is considered 32 * "implicit padding" and must be zeroed. This is always checked. 33 * 34 * In addition to the IOCTL argument structs, the PowerVR UAPI makes use of 35 * DEV_QUERY argument structs. These are used to fetch information about the 36 * device and runtime. These structs are subject to the same rules set out 37 * above. 38 */ 39 40 /** 41 * struct drm_pvr_obj_array - Container used to pass arrays of objects 42 * 43 * It is not unusual to have to extend objects to pass new parameters, and the DRM 44 * ioctl infrastructure is supporting that by padding ioctl arguments with zeros 45 * when the data passed by userspace is smaller than the struct defined in the 46 * drm_ioctl_desc, thus keeping things backward compatible. This type is just 47 * applying the same concepts to indirect objects passed through arrays referenced 48 * from the main ioctl arguments structure: the stride basically defines the size 49 * of the object passed by userspace, which allows the kernel driver to pad with 50 * zeros when it's smaller than the size of the object it expects. 51 * 52 * Use ``DRM_PVR_OBJ_ARRAY()`` to fill object array fields, unless you 53 * have a very good reason not to. 54 */ 55 struct drm_pvr_obj_array { 56 /** @stride: Stride of object struct. Used for versioning. */ 57 __u32 stride; 58 59 /** @count: Number of objects in the array. */ 60 __u32 count; 61 62 /** @array: User pointer to an array of objects. */ 63 __u64 array; 64 }; 65 66 /** 67 * DRM_PVR_OBJ_ARRAY() - Helper macro for filling &struct drm_pvr_obj_array. 68 * @cnt: Number of elements pointed to py @ptr. 69 * @ptr: Pointer to start of a C array. 70 * 71 * Return: Literal of type &struct drm_pvr_obj_array. 72 */ 73 #define DRM_PVR_OBJ_ARRAY(cnt, ptr) \ 74 { .stride = sizeof((ptr)[0]), .count = (cnt), .array = (__u64)(uintptr_t)(ptr) } 75 76 /** 77 * DOC: PowerVR IOCTL interface 78 */ 79 80 /** 81 * PVR_IOCTL() - Build a PowerVR IOCTL number 82 * @_ioctl: An incrementing id for this IOCTL. Added to %DRM_COMMAND_BASE. 83 * @_mode: Must be one of %DRM_IOR, %DRM_IOW or %DRM_IOWR. 84 * @_data: The type of the args struct passed by this IOCTL. 85 * 86 * The struct referred to by @_data must have a ``drm_pvr_ioctl_`` prefix and an 87 * ``_args suffix``. They are therefore omitted from @_data. 88 * 89 * This should only be used to build the constants described below; it should 90 * never be used to call an IOCTL directly. 91 * 92 * Return: An IOCTL number to be passed to ioctl() from userspace. 93 */ 94 #define PVR_IOCTL(_ioctl, _mode, _data) \ 95 _mode(DRM_COMMAND_BASE + (_ioctl), struct drm_pvr_ioctl_##_data##_args) 96 97 #define DRM_IOCTL_PVR_DEV_QUERY PVR_IOCTL(0x00, DRM_IOWR, dev_query) 98 #define DRM_IOCTL_PVR_CREATE_BO PVR_IOCTL(0x01, DRM_IOWR, create_bo) 99 #define DRM_IOCTL_PVR_GET_BO_MMAP_OFFSET PVR_IOCTL(0x02, DRM_IOWR, get_bo_mmap_offset) 100 #define DRM_IOCTL_PVR_CREATE_VM_CONTEXT PVR_IOCTL(0x03, DRM_IOWR, create_vm_context) 101 #define DRM_IOCTL_PVR_DESTROY_VM_CONTEXT PVR_IOCTL(0x04, DRM_IOW, destroy_vm_context) 102 #define DRM_IOCTL_PVR_VM_MAP PVR_IOCTL(0x05, DRM_IOW, vm_map) 103 #define DRM_IOCTL_PVR_VM_UNMAP PVR_IOCTL(0x06, DRM_IOW, vm_unmap) 104 #define DRM_IOCTL_PVR_CREATE_CONTEXT PVR_IOCTL(0x07, DRM_IOWR, create_context) 105 #define DRM_IOCTL_PVR_DESTROY_CONTEXT PVR_IOCTL(0x08, DRM_IOW, destroy_context) 106 #define DRM_IOCTL_PVR_CREATE_FREE_LIST PVR_IOCTL(0x09, DRM_IOWR, create_free_list) 107 #define DRM_IOCTL_PVR_DESTROY_FREE_LIST PVR_IOCTL(0x0a, DRM_IOW, destroy_free_list) 108 #define DRM_IOCTL_PVR_CREATE_HWRT_DATASET PVR_IOCTL(0x0b, DRM_IOWR, create_hwrt_dataset) 109 #define DRM_IOCTL_PVR_DESTROY_HWRT_DATASET PVR_IOCTL(0x0c, DRM_IOW, destroy_hwrt_dataset) 110 #define DRM_IOCTL_PVR_SUBMIT_JOBS PVR_IOCTL(0x0d, DRM_IOW, submit_jobs) 111 112 /** 113 * DOC: PowerVR IOCTL DEV_QUERY interface 114 */ 115 116 /** 117 * struct drm_pvr_dev_query_gpu_info - Container used to fetch information about 118 * the graphics processor. 119 * 120 * When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set 121 * to %DRM_PVR_DEV_QUERY_GPU_INFO_GET. 122 */ 123 struct drm_pvr_dev_query_gpu_info { 124 /** 125 * @gpu_id: GPU identifier. 126 * 127 * For all currently supported GPUs this is the BVNC encoded as a 64-bit 128 * value as follows: 129 * 130 * +--------+--------+--------+-------+ 131 * | 63..48 | 47..32 | 31..16 | 15..0 | 132 * +========+========+========+=======+ 133 * | B | V | N | C | 134 * +--------+--------+--------+-------+ 135 */ 136 __u64 gpu_id; 137 138 /** 139 * @num_phantoms: Number of Phantoms present. 140 */ 141 __u32 num_phantoms; 142 143 /** @_padding_c: Reserved. This field must be zeroed. */ 144 __u32 _padding_c; 145 }; 146 147 /** 148 * struct drm_pvr_dev_query_runtime_info - Container used to fetch information 149 * about the graphics runtime. 150 * 151 * When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set 152 * to %DRM_PVR_DEV_QUERY_RUNTIME_INFO_GET. 153 */ 154 struct drm_pvr_dev_query_runtime_info { 155 /** 156 * @free_list_min_pages: Minimum allowed free list size, 157 * in PM physical pages. 158 */ 159 __u64 free_list_min_pages; 160 161 /** 162 * @free_list_max_pages: Maximum allowed free list size, 163 * in PM physical pages. 164 */ 165 __u64 free_list_max_pages; 166 167 /** 168 * @common_store_alloc_region_size: Size of the Allocation 169 * Region within the Common Store used for coefficient and shared 170 * registers, in dwords. 171 */ 172 __u32 common_store_alloc_region_size; 173 174 /** 175 * @common_store_partition_space_size: Size of the 176 * Partition Space within the Common Store for output buffers, in 177 * dwords. 178 */ 179 __u32 common_store_partition_space_size; 180 181 /** 182 * @max_coeffs: Maximum coefficients, in dwords. 183 */ 184 __u32 max_coeffs; 185 186 /** 187 * @cdm_max_local_mem_size_regs: Maximum amount of local 188 * memory available to a compute kernel, in dwords. 189 */ 190 __u32 cdm_max_local_mem_size_regs; 191 }; 192 193 /** 194 * struct drm_pvr_dev_query_quirks - Container used to fetch information about 195 * hardware fixes for which the device may require support in the user mode 196 * driver. 197 * 198 * When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set 199 * to %DRM_PVR_DEV_QUERY_QUIRKS_GET. 200 */ 201 struct drm_pvr_dev_query_quirks { 202 /** 203 * @quirks: A userspace address for the hardware quirks __u32 array. 204 * 205 * The first @musthave_count items in the list are quirks that the 206 * client must support for this device. If userspace does not support 207 * all these quirks then functionality is not guaranteed and client 208 * initialisation must fail. 209 * The remaining quirks in the list affect userspace and the kernel or 210 * firmware. They are disabled by default and require userspace to 211 * opt-in. The opt-in mechanism depends on the quirk. 212 */ 213 __u64 quirks; 214 215 /** @count: Length of @quirks (number of __u32). */ 216 __u16 count; 217 218 /** 219 * @musthave_count: The number of entries in @quirks that are 220 * mandatory, starting at index 0. 221 */ 222 __u16 musthave_count; 223 224 /** @_padding_c: Reserved. This field must be zeroed. */ 225 __u32 _padding_c; 226 }; 227 228 /** 229 * struct drm_pvr_dev_query_enhancements - Container used to fetch information 230 * about optional enhancements supported by the device that require support in 231 * the user mode driver. 232 * 233 * When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set 234 * to %DRM_PVR_DEV_ENHANCEMENTS_GET. 235 */ 236 struct drm_pvr_dev_query_enhancements { 237 /** 238 * @enhancements: A userspace address for the hardware enhancements 239 * __u32 array. 240 * 241 * These enhancements affect userspace and the kernel or firmware. They 242 * are disabled by default and require userspace to opt-in. The opt-in 243 * mechanism depends on the enhancement. 244 */ 245 __u64 enhancements; 246 247 /** @count: Length of @enhancements (number of __u32). */ 248 __u16 count; 249 250 /** @_padding_a: Reserved. This field must be zeroed. */ 251 __u16 _padding_a; 252 253 /** @_padding_c: Reserved. This field must be zeroed. */ 254 __u32 _padding_c; 255 }; 256 257 /** 258 * enum drm_pvr_heap_id - Array index for heap info data returned by 259 * %DRM_PVR_DEV_QUERY_HEAP_INFO_GET. 260 * 261 * For compatibility reasons all indices will be present in the returned array, 262 * however some heaps may not be present. These are indicated where 263 * &struct drm_pvr_heap.size is set to zero. 264 */ 265 enum drm_pvr_heap_id { 266 /** @DRM_PVR_HEAP_GENERAL: General purpose heap. */ 267 DRM_PVR_HEAP_GENERAL = 0, 268 /** @DRM_PVR_HEAP_PDS_CODE_DATA: PDS code and data heap. */ 269 DRM_PVR_HEAP_PDS_CODE_DATA, 270 /** @DRM_PVR_HEAP_USC_CODE: USC code heap. */ 271 DRM_PVR_HEAP_USC_CODE, 272 /** @DRM_PVR_HEAP_RGNHDR: Region header heap. Only used if GPU has BRN63142. */ 273 DRM_PVR_HEAP_RGNHDR, 274 /** @DRM_PVR_HEAP_VIS_TEST: Visibility test heap. */ 275 DRM_PVR_HEAP_VIS_TEST, 276 /** @DRM_PVR_HEAP_TRANSFER_FRAG: Transfer fragment heap. */ 277 DRM_PVR_HEAP_TRANSFER_FRAG, 278 279 /** 280 * @DRM_PVR_HEAP_COUNT: The number of heaps returned by 281 * %DRM_PVR_DEV_QUERY_HEAP_INFO_GET. 282 * 283 * More heaps may be added, so this also serves as the copy limit when 284 * sent by the caller. 285 */ 286 DRM_PVR_HEAP_COUNT 287 /* Please only add additional heaps above DRM_PVR_HEAP_COUNT! */ 288 }; 289 290 /** 291 * struct drm_pvr_heap - Container holding information about a single heap. 292 * 293 * This will always be fetched as an array. 294 */ 295 struct drm_pvr_heap { 296 /** @base: Base address of heap. */ 297 __u64 base; 298 299 /** @size: Size of heap, in bytes. Will be 0 if the heap is not present. */ 300 __u64 size; 301 302 /** @flags: Flags for this heap. Currently always 0. */ 303 __u32 flags; 304 305 /** @page_size_log2: Log2 of page size. */ 306 __u32 page_size_log2; 307 }; 308 309 /** 310 * struct drm_pvr_dev_query_heap_info - Container used to fetch information 311 * about heaps supported by the device driver. 312 * 313 * Please note all driver-supported heaps will be returned up to &heaps.count. 314 * Some heaps will not be present in all devices, which will be indicated by 315 * &struct drm_pvr_heap.size being set to zero. 316 * 317 * When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set 318 * to %DRM_PVR_DEV_QUERY_HEAP_INFO_GET. 319 */ 320 struct drm_pvr_dev_query_heap_info { 321 /** 322 * @heaps: Array of &struct drm_pvr_heap. If pointer is NULL, the count 323 * and stride will be updated with those known to the driver version, to 324 * facilitate allocation by the caller. 325 */ 326 struct drm_pvr_obj_array heaps; 327 }; 328 329 /** 330 * enum drm_pvr_static_data_area_usage - Array index for static data area info 331 * returned by %DRM_PVR_DEV_QUERY_STATIC_DATA_AREAS_GET. 332 * 333 * For compatibility reasons all indices will be present in the returned array, 334 * however some areas may not be present. These are indicated where 335 * &struct drm_pvr_static_data_area.size is set to zero. 336 */ 337 enum drm_pvr_static_data_area_usage { 338 /** 339 * @DRM_PVR_STATIC_DATA_AREA_EOT: End of Tile PDS program code segment. 340 * 341 * The End of Tile PDS task runs at completion of a tile during a fragment job, and is 342 * responsible for emitting the tile to the Pixel Back End. 343 */ 344 DRM_PVR_STATIC_DATA_AREA_EOT = 0, 345 346 /** 347 * @DRM_PVR_STATIC_DATA_AREA_FENCE: MCU fence area, used during cache flush and 348 * invalidation. 349 * 350 * This must point to valid physical memory but the contents otherwise are not used. 351 */ 352 DRM_PVR_STATIC_DATA_AREA_FENCE, 353 354 /** 355 * @DRM_PVR_STATIC_DATA_AREA_VDM_SYNC: VDM sync program. 356 * 357 * The VDM sync program is used to synchronise multiple areas of the GPU hardware. 358 */ 359 DRM_PVR_STATIC_DATA_AREA_VDM_SYNC, 360 361 /** 362 * @DRM_PVR_STATIC_DATA_AREA_YUV_CSC: YUV coefficients. 363 * 364 * Area contains up to 16 slots with stride of 64 bytes. Each is a 3x4 matrix of u16 fixed 365 * point numbers, with 1 sign bit, 2 integer bits and 13 fractional bits. 366 * 367 * The slots are : 368 * 0 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_RGB_IDENTITY_KHR 369 * 1 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_IDENTITY_KHR (full range) 370 * 2 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_IDENTITY_KHR (conformant range) 371 * 3 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_709_KHR (full range) 372 * 4 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_709_KHR (conformant range) 373 * 5 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_601_KHR (full range) 374 * 6 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_601_KHR (conformant range) 375 * 7 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_2020_KHR (full range) 376 * 8 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_2020_KHR (conformant range) 377 * 9 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_601_KHR (conformant range, 10 bit) 378 * 10 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_709_KHR (conformant range, 10 bit) 379 * 11 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_2020_KHR (conformant range, 10 bit) 380 * 14 = Identity (biased) 381 * 15 = Identity 382 */ 383 DRM_PVR_STATIC_DATA_AREA_YUV_CSC, 384 }; 385 386 /** 387 * struct drm_pvr_static_data_area - Container holding information about a 388 * single static data area. 389 * 390 * This will always be fetched as an array. 391 */ 392 struct drm_pvr_static_data_area { 393 /** 394 * @area_usage: Usage of static data area. 395 * See &enum drm_pvr_static_data_area_usage. 396 */ 397 __u16 area_usage; 398 399 /** 400 * @location_heap_id: Array index of heap where this of static data 401 * area is located. This array is fetched using 402 * %DRM_PVR_DEV_QUERY_HEAP_INFO_GET. 403 */ 404 __u16 location_heap_id; 405 406 /** @size: Size of static data area. Not present if set to zero. */ 407 __u32 size; 408 409 /** @offset: Offset of static data area from start of heap. */ 410 __u64 offset; 411 }; 412 413 /** 414 * struct drm_pvr_dev_query_static_data_areas - Container used to fetch 415 * information about the static data areas in heaps supported by the device 416 * driver. 417 * 418 * Please note all driver-supported static data areas will be returned up to 419 * &static_data_areas.count. Some will not be present for all devices which, 420 * will be indicated by &struct drm_pvr_static_data_area.size being set to zero. 421 * 422 * Further, some heaps will not be present either. See &struct 423 * drm_pvr_dev_query_heap_info. 424 * 425 * When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set 426 * to %DRM_PVR_DEV_QUERY_STATIC_DATA_AREAS_GET. 427 */ 428 struct drm_pvr_dev_query_static_data_areas { 429 /** 430 * @static_data_areas: Array of &struct drm_pvr_static_data_area. If 431 * pointer is NULL, the count and stride will be updated with those 432 * known to the driver version, to facilitate allocation by the caller. 433 */ 434 struct drm_pvr_obj_array static_data_areas; 435 }; 436 437 /** 438 * enum drm_pvr_dev_query - For use with &drm_pvr_ioctl_dev_query_args.type to 439 * indicate the type of the receiving container. 440 * 441 * Append only. Do not reorder. 442 */ 443 enum drm_pvr_dev_query { 444 /** 445 * @DRM_PVR_DEV_QUERY_GPU_INFO_GET: The dev query args contain a pointer 446 * to &struct drm_pvr_dev_query_gpu_info. 447 */ 448 DRM_PVR_DEV_QUERY_GPU_INFO_GET = 0, 449 450 /** 451 * @DRM_PVR_DEV_QUERY_RUNTIME_INFO_GET: The dev query args contain a 452 * pointer to &struct drm_pvr_dev_query_runtime_info. 453 */ 454 DRM_PVR_DEV_QUERY_RUNTIME_INFO_GET, 455 456 /** 457 * @DRM_PVR_DEV_QUERY_QUIRKS_GET: The dev query args contain a pointer 458 * to &struct drm_pvr_dev_query_quirks. 459 */ 460 DRM_PVR_DEV_QUERY_QUIRKS_GET, 461 462 /** 463 * @DRM_PVR_DEV_QUERY_ENHANCEMENTS_GET: The dev query args contain a 464 * pointer to &struct drm_pvr_dev_query_enhancements. 465 */ 466 DRM_PVR_DEV_QUERY_ENHANCEMENTS_GET, 467 468 /** 469 * @DRM_PVR_DEV_QUERY_HEAP_INFO_GET: The dev query args contain a 470 * pointer to &struct drm_pvr_dev_query_heap_info. 471 */ 472 DRM_PVR_DEV_QUERY_HEAP_INFO_GET, 473 474 /** 475 * @DRM_PVR_DEV_QUERY_STATIC_DATA_AREAS_GET: The dev query args contain 476 * a pointer to &struct drm_pvr_dev_query_static_data_areas. 477 */ 478 DRM_PVR_DEV_QUERY_STATIC_DATA_AREAS_GET, 479 }; 480 481 /** 482 * struct drm_pvr_ioctl_dev_query_args - Arguments for %DRM_IOCTL_PVR_DEV_QUERY. 483 */ 484 struct drm_pvr_ioctl_dev_query_args { 485 /** 486 * @type: Type of query and output struct. See &enum drm_pvr_dev_query. 487 */ 488 __u32 type; 489 490 /** 491 * @size: Size of the receiving struct, see @type. 492 * 493 * After a successful call this will be updated to the written byte 494 * length. 495 * Can also be used to get the minimum byte length (see @pointer). 496 * This allows additional fields to be appended to the structs in 497 * future. 498 */ 499 __u32 size; 500 501 /** 502 * @pointer: Pointer to struct @type. 503 * 504 * Must be large enough to contain @size bytes. 505 * If pointer is NULL, the expected size will be returned in the @size 506 * field, but no other data will be written. 507 */ 508 __u64 pointer; 509 }; 510 511 /** 512 * DOC: PowerVR IOCTL CREATE_BO interface 513 */ 514 515 /** 516 * DOC: Flags for CREATE_BO 517 * 518 * We use "device" to refer to the GPU here because of the ambiguity between CPU and GPU in some 519 * fonts. 520 * 521 * Device mapping options 522 * :DRM_PVR_BO_BYPASS_DEVICE_CACHE: Specify that device accesses to this memory will bypass the 523 * cache. This is used for buffers that will either be regularly updated by the CPU (eg free 524 * lists) or will be accessed only once and therefore isn't worth caching (eg partial render 525 * buffers). 526 * By default, the device flushes its memory caches after every job, so this is not normally 527 * required for coherency. 528 * :DRM_PVR_BO_PM_FW_PROTECT: Specify that only the Parameter Manager (PM) and/or firmware 529 * processor should be allowed to access this memory when mapped to the device. It is not 530 * valid to specify this flag with DRM_PVR_BO_ALLOW_CPU_USERSPACE_ACCESS. 531 * 532 * CPU mapping options 533 * :DRM_PVR_BO_ALLOW_CPU_USERSPACE_ACCESS: Allow userspace to map and access the contents of this 534 * memory. It is not valid to specify this flag with DRM_PVR_BO_PM_FW_PROTECT. 535 */ 536 #define DRM_PVR_BO_BYPASS_DEVICE_CACHE _BITULL(0) 537 #define DRM_PVR_BO_PM_FW_PROTECT _BITULL(1) 538 #define DRM_PVR_BO_ALLOW_CPU_USERSPACE_ACCESS _BITULL(2) 539 /* Bits 3..63 are reserved. */ 540 541 #define DRM_PVR_BO_FLAGS_MASK (DRM_PVR_BO_BYPASS_DEVICE_CACHE | DRM_PVR_BO_PM_FW_PROTECT | \ 542 DRM_PVR_BO_ALLOW_CPU_USERSPACE_ACCESS) 543 544 /** 545 * struct drm_pvr_ioctl_create_bo_args - Arguments for %DRM_IOCTL_PVR_CREATE_BO 546 */ 547 struct drm_pvr_ioctl_create_bo_args { 548 /** 549 * @size: [IN] Size of buffer object to create. This must be page size 550 * aligned. 551 */ 552 __u64 size; 553 554 /** 555 * @handle: [OUT] GEM handle of the new buffer object for use in 556 * userspace. 557 */ 558 __u32 handle; 559 560 /** @_padding_c: Reserved. This field must be zeroed. */ 561 __u32 _padding_c; 562 563 /** 564 * @flags: [IN] Options which will affect the behaviour of this 565 * creation operation and future mapping operations on the created 566 * object. This field must be a valid combination of ``DRM_PVR_BO_*`` 567 * values, with all bits marked as reserved set to zero. 568 */ 569 __u64 flags; 570 }; 571 572 /** 573 * DOC: PowerVR IOCTL GET_BO_MMAP_OFFSET interface 574 */ 575 576 /** 577 * struct drm_pvr_ioctl_get_bo_mmap_offset_args - Arguments for 578 * %DRM_IOCTL_PVR_GET_BO_MMAP_OFFSET 579 * 580 * Like other DRM drivers, the "mmap" IOCTL doesn't actually map any memory. 581 * Instead, it allocates a fake offset which refers to the specified buffer 582 * object. This offset can be used with a real mmap call on the DRM device 583 * itself. 584 */ 585 struct drm_pvr_ioctl_get_bo_mmap_offset_args { 586 /** @handle: [IN] GEM handle of the buffer object to be mapped. */ 587 __u32 handle; 588 589 /** @_padding_4: Reserved. This field must be zeroed. */ 590 __u32 _padding_4; 591 592 /** @offset: [OUT] Fake offset to use in the real mmap call. */ 593 __u64 offset; 594 }; 595 596 /** 597 * DOC: PowerVR IOCTL CREATE_VM_CONTEXT and DESTROY_VM_CONTEXT interfaces 598 */ 599 600 /** 601 * struct drm_pvr_ioctl_create_vm_context_args - Arguments for 602 * %DRM_IOCTL_PVR_CREATE_VM_CONTEXT 603 */ 604 struct drm_pvr_ioctl_create_vm_context_args { 605 /** @handle: [OUT] Handle for new VM context. */ 606 __u32 handle; 607 608 /** @_padding_4: Reserved. This field must be zeroed. */ 609 __u32 _padding_4; 610 }; 611 612 /** 613 * struct drm_pvr_ioctl_destroy_vm_context_args - Arguments for 614 * %DRM_IOCTL_PVR_DESTROY_VM_CONTEXT 615 */ 616 struct drm_pvr_ioctl_destroy_vm_context_args { 617 /** 618 * @handle: [IN] Handle for VM context to be destroyed. 619 */ 620 __u32 handle; 621 622 /** @_padding_4: Reserved. This field must be zeroed. */ 623 __u32 _padding_4; 624 }; 625 626 /** 627 * DOC: PowerVR IOCTL VM_MAP and VM_UNMAP interfaces 628 * 629 * The VM UAPI allows userspace to create buffer object mappings in GPU virtual address space. 630 * 631 * The client is responsible for managing GPU address space. It should allocate mappings within 632 * the heaps returned by %DRM_PVR_DEV_QUERY_HEAP_INFO_GET. 633 * 634 * %DRM_IOCTL_PVR_VM_MAP creates a new mapping. The client provides the target virtual address for 635 * the mapping. Size and offset within the mapped buffer object can be specified, so the client can 636 * partially map a buffer. 637 * 638 * %DRM_IOCTL_PVR_VM_UNMAP removes a mapping. The entire mapping will be removed from GPU address 639 * space only if the size of the mapping matches that known to the driver. 640 */ 641 642 /** 643 * struct drm_pvr_ioctl_vm_map_args - Arguments for %DRM_IOCTL_PVR_VM_MAP. 644 */ 645 struct drm_pvr_ioctl_vm_map_args { 646 /** 647 * @vm_context_handle: [IN] Handle for VM context for this mapping to 648 * exist in. 649 */ 650 __u32 vm_context_handle; 651 652 /** @flags: [IN] Flags which affect this mapping. Currently always 0. */ 653 __u32 flags; 654 655 /** 656 * @device_addr: [IN] Requested device-virtual address for the mapping. 657 * This must be non-zero and aligned to the device page size for the 658 * heap containing the requested address. It is an error to specify an 659 * address which is not contained within one of the heaps returned by 660 * %DRM_PVR_DEV_QUERY_HEAP_INFO_GET. 661 */ 662 __u64 device_addr; 663 664 /** 665 * @handle: [IN] Handle of the target buffer object. This must be a 666 * valid handle returned by %DRM_IOCTL_PVR_CREATE_BO. 667 */ 668 __u32 handle; 669 670 /** @_padding_14: Reserved. This field must be zeroed. */ 671 __u32 _padding_14; 672 673 /** 674 * @offset: [IN] Offset into the target bo from which to begin the 675 * mapping. 676 */ 677 __u64 offset; 678 679 /** 680 * @size: [IN] Size of the requested mapping. Must be aligned to 681 * the device page size for the heap containing the requested address, 682 * as well as the host page size. When added to @device_addr, the 683 * result must not overflow the heap which contains @device_addr (i.e. 684 * the range specified by @device_addr and @size must be completely 685 * contained within a single heap specified by 686 * %DRM_PVR_DEV_QUERY_HEAP_INFO_GET). 687 */ 688 __u64 size; 689 }; 690 691 /** 692 * struct drm_pvr_ioctl_vm_unmap_args - Arguments for %DRM_IOCTL_PVR_VM_UNMAP. 693 */ 694 struct drm_pvr_ioctl_vm_unmap_args { 695 /** 696 * @vm_context_handle: [IN] Handle for VM context that this mapping 697 * exists in. 698 */ 699 __u32 vm_context_handle; 700 701 /** @_padding_4: Reserved. This field must be zeroed. */ 702 __u32 _padding_4; 703 704 /** 705 * @device_addr: [IN] Device-virtual address at the start of the target 706 * mapping. This must be non-zero. 707 */ 708 __u64 device_addr; 709 710 /** 711 * @size: Size in bytes of the target mapping. This must be non-zero. 712 */ 713 __u64 size; 714 }; 715 716 /** 717 * DOC: PowerVR IOCTL CREATE_CONTEXT and DESTROY_CONTEXT interfaces 718 */ 719 720 /** 721 * enum drm_pvr_ctx_priority - Arguments for 722 * &drm_pvr_ioctl_create_context_args.priority 723 */ 724 enum drm_pvr_ctx_priority { 725 /** @DRM_PVR_CTX_PRIORITY_LOW: Priority below normal. */ 726 DRM_PVR_CTX_PRIORITY_LOW = -512, 727 728 /** @DRM_PVR_CTX_PRIORITY_NORMAL: Normal priority. */ 729 DRM_PVR_CTX_PRIORITY_NORMAL = 0, 730 731 /** 732 * @DRM_PVR_CTX_PRIORITY_HIGH: Priority above normal. 733 * Note this requires ``CAP_SYS_NICE`` or ``DRM_MASTER``. 734 */ 735 DRM_PVR_CTX_PRIORITY_HIGH = 512, 736 }; 737 738 /** 739 * enum drm_pvr_ctx_type - Arguments for 740 * &struct drm_pvr_ioctl_create_context_args.type 741 */ 742 enum drm_pvr_ctx_type { 743 /** 744 * @DRM_PVR_CTX_TYPE_RENDER: Render context. 745 */ 746 DRM_PVR_CTX_TYPE_RENDER = 0, 747 748 /** 749 * @DRM_PVR_CTX_TYPE_COMPUTE: Compute context. 750 */ 751 DRM_PVR_CTX_TYPE_COMPUTE, 752 753 /** 754 * @DRM_PVR_CTX_TYPE_TRANSFER_FRAG: Transfer context for fragment data 755 * master. 756 */ 757 DRM_PVR_CTX_TYPE_TRANSFER_FRAG, 758 }; 759 760 /** 761 * struct drm_pvr_ioctl_create_context_args - Arguments for 762 * %DRM_IOCTL_PVR_CREATE_CONTEXT 763 */ 764 struct drm_pvr_ioctl_create_context_args { 765 /** 766 * @type: [IN] Type of context to create. 767 * 768 * This must be one of the values defined by &enum drm_pvr_ctx_type. 769 */ 770 __u32 type; 771 772 /** @flags: [IN] Flags for context. */ 773 __u32 flags; 774 775 /** 776 * @priority: [IN] Priority of new context. 777 * 778 * This must be one of the values defined by &enum drm_pvr_ctx_priority. 779 */ 780 __s32 priority; 781 782 /** @handle: [OUT] Handle for new context. */ 783 __u32 handle; 784 785 /** 786 * @static_context_state: [IN] Pointer to static context state stream. 787 */ 788 __u64 static_context_state; 789 790 /** 791 * @static_context_state_len: [IN] Length of static context state, in bytes. 792 */ 793 __u32 static_context_state_len; 794 795 /** 796 * @vm_context_handle: [IN] Handle for VM context that this context is 797 * associated with. 798 */ 799 __u32 vm_context_handle; 800 801 /** 802 * @callstack_addr: [IN] Address for initial call stack pointer. Only valid 803 * if @type is %DRM_PVR_CTX_TYPE_RENDER, otherwise must be 0. 804 */ 805 __u64 callstack_addr; 806 }; 807 808 /** 809 * struct drm_pvr_ioctl_destroy_context_args - Arguments for 810 * %DRM_IOCTL_PVR_DESTROY_CONTEXT 811 */ 812 struct drm_pvr_ioctl_destroy_context_args { 813 /** 814 * @handle: [IN] Handle for context to be destroyed. 815 */ 816 __u32 handle; 817 818 /** @_padding_4: Reserved. This field must be zeroed. */ 819 __u32 _padding_4; 820 }; 821 822 /** 823 * DOC: PowerVR IOCTL CREATE_FREE_LIST and DESTROY_FREE_LIST interfaces 824 */ 825 826 /** 827 * struct drm_pvr_ioctl_create_free_list_args - Arguments for 828 * %DRM_IOCTL_PVR_CREATE_FREE_LIST 829 * 830 * Free list arguments have the following constraints : 831 * 832 * - @max_num_pages must be greater than zero. 833 * - @grow_threshold must be between 0 and 100. 834 * - @grow_num_pages must be less than or equal to &max_num_pages. 835 * - @initial_num_pages, @max_num_pages and @grow_num_pages must be multiples 836 * of 4. 837 * - When &grow_num_pages is 0, @initial_num_pages must be equal to 838 * @max_num_pages. 839 * - When &grow_num_pages is non-zero, @initial_num_pages must be less than 840 * @max_num_pages. 841 */ 842 struct drm_pvr_ioctl_create_free_list_args { 843 /** 844 * @free_list_gpu_addr: [IN] Address of GPU mapping of buffer object 845 * containing memory to be used by free list. 846 * 847 * The mapped region of the buffer object must be at least 848 * @max_num_pages * ``sizeof(__u32)``. 849 * 850 * The buffer object must have been created with 851 * %DRM_PVR_BO_DEVICE_PM_FW_PROTECT set and 852 * %DRM_PVR_BO_CPU_ALLOW_USERSPACE_ACCESS not set. 853 */ 854 __u64 free_list_gpu_addr; 855 856 /** @initial_num_pages: [IN] Pages initially allocated to free list. */ 857 __u32 initial_num_pages; 858 859 /** @max_num_pages: [IN] Maximum number of pages in free list. */ 860 __u32 max_num_pages; 861 862 /** @grow_num_pages: [IN] Pages to grow free list by per request. */ 863 __u32 grow_num_pages; 864 865 /** 866 * @grow_threshold: [IN] Percentage of FL memory used that should 867 * trigger a new grow request. 868 */ 869 __u32 grow_threshold; 870 871 /** 872 * @vm_context_handle: [IN] Handle for VM context that the free list buffer 873 * object is mapped in. 874 */ 875 __u32 vm_context_handle; 876 877 /** 878 * @handle: [OUT] Handle for created free list. 879 */ 880 __u32 handle; 881 }; 882 883 /** 884 * struct drm_pvr_ioctl_destroy_free_list_args - Arguments for 885 * %DRM_IOCTL_PVR_DESTROY_FREE_LIST 886 */ 887 struct drm_pvr_ioctl_destroy_free_list_args { 888 /** 889 * @handle: [IN] Handle for free list to be destroyed. 890 */ 891 __u32 handle; 892 893 /** @_padding_4: Reserved. This field must be zeroed. */ 894 __u32 _padding_4; 895 }; 896 897 /** 898 * DOC: PowerVR IOCTL CREATE_HWRT_DATASET and DESTROY_HWRT_DATASET interfaces 899 */ 900 901 /** 902 * struct drm_pvr_create_hwrt_geom_data_args - Geometry data arguments used for 903 * &struct drm_pvr_ioctl_create_hwrt_dataset_args.geom_data_args. 904 */ 905 struct drm_pvr_create_hwrt_geom_data_args { 906 /** @tpc_dev_addr: [IN] Tail pointer cache GPU virtual address. */ 907 __u64 tpc_dev_addr; 908 909 /** @tpc_size: [IN] Size of TPC, in bytes. */ 910 __u32 tpc_size; 911 912 /** @tpc_stride: [IN] Stride between layers in TPC, in pages */ 913 __u32 tpc_stride; 914 915 /** @vheap_table_dev_addr: [IN] VHEAP table GPU virtual address. */ 916 __u64 vheap_table_dev_addr; 917 918 /** @rtc_dev_addr: [IN] Render Target Cache virtual address. */ 919 __u64 rtc_dev_addr; 920 }; 921 922 /** 923 * struct drm_pvr_create_hwrt_rt_data_args - Render target arguments used for 924 * &struct drm_pvr_ioctl_create_hwrt_dataset_args.rt_data_args. 925 */ 926 struct drm_pvr_create_hwrt_rt_data_args { 927 /** @pm_mlist_dev_addr: [IN] PM MLIST GPU virtual address. */ 928 __u64 pm_mlist_dev_addr; 929 930 /** @macrotile_array_dev_addr: [IN] Macrotile array GPU virtual address. */ 931 __u64 macrotile_array_dev_addr; 932 933 /** @region_header_dev_addr: [IN] Region header array GPU virtual address. */ 934 __u64 region_header_dev_addr; 935 }; 936 937 #define PVR_DRM_HWRT_FREE_LIST_LOCAL 0 938 #define PVR_DRM_HWRT_FREE_LIST_GLOBAL 1U 939 940 /** 941 * struct drm_pvr_ioctl_create_hwrt_dataset_args - Arguments for 942 * %DRM_IOCTL_PVR_CREATE_HWRT_DATASET 943 */ 944 struct drm_pvr_ioctl_create_hwrt_dataset_args { 945 /** @geom_data_args: [IN] Geometry data arguments. */ 946 struct drm_pvr_create_hwrt_geom_data_args geom_data_args; 947 948 /** 949 * @rt_data_args: [IN] Array of render target arguments. 950 * 951 * Each entry in this array represents a render target in a double buffered 952 * setup. 953 */ 954 struct drm_pvr_create_hwrt_rt_data_args rt_data_args[2]; 955 956 /** 957 * @free_list_handles: [IN] Array of free list handles. 958 * 959 * free_list_handles[PVR_DRM_HWRT_FREE_LIST_LOCAL] must have initial 960 * size of at least that reported by 961 * &drm_pvr_dev_query_runtime_info.free_list_min_pages. 962 */ 963 __u32 free_list_handles[2]; 964 965 /** @width: [IN] Width in pixels. */ 966 __u32 width; 967 968 /** @height: [IN] Height in pixels. */ 969 __u32 height; 970 971 /** @samples: [IN] Number of samples. */ 972 __u32 samples; 973 974 /** @layers: [IN] Number of layers. */ 975 __u32 layers; 976 977 /** @isp_merge_lower_x: [IN] Lower X coefficient for triangle merging. */ 978 __u32 isp_merge_lower_x; 979 980 /** @isp_merge_lower_y: [IN] Lower Y coefficient for triangle merging. */ 981 __u32 isp_merge_lower_y; 982 983 /** @isp_merge_scale_x: [IN] Scale X coefficient for triangle merging. */ 984 __u32 isp_merge_scale_x; 985 986 /** @isp_merge_scale_y: [IN] Scale Y coefficient for triangle merging. */ 987 __u32 isp_merge_scale_y; 988 989 /** @isp_merge_upper_x: [IN] Upper X coefficient for triangle merging. */ 990 __u32 isp_merge_upper_x; 991 992 /** @isp_merge_upper_y: [IN] Upper Y coefficient for triangle merging. */ 993 __u32 isp_merge_upper_y; 994 995 /** 996 * @region_header_size: [IN] Size of region header array. This common field is used by 997 * both render targets in this data set. 998 * 999 * The units for this field differ depending on what version of the simple internal 1000 * parameter format the device uses. If format 2 is in use then this is interpreted as the 1001 * number of region headers. For other formats it is interpreted as the size in dwords. 1002 */ 1003 __u32 region_header_size; 1004 1005 /** 1006 * @handle: [OUT] Handle for created HWRT dataset. 1007 */ 1008 __u32 handle; 1009 }; 1010 1011 /** 1012 * struct drm_pvr_ioctl_destroy_hwrt_dataset_args - Arguments for 1013 * %DRM_IOCTL_PVR_DESTROY_HWRT_DATASET 1014 */ 1015 struct drm_pvr_ioctl_destroy_hwrt_dataset_args { 1016 /** 1017 * @handle: [IN] Handle for HWRT dataset to be destroyed. 1018 */ 1019 __u32 handle; 1020 1021 /** @_padding_4: Reserved. This field must be zeroed. */ 1022 __u32 _padding_4; 1023 }; 1024 1025 /** 1026 * DOC: PowerVR IOCTL SUBMIT_JOBS interface 1027 */ 1028 1029 /** 1030 * DOC: Flags for the drm_pvr_sync_op object. 1031 * 1032 * .. c:macro:: DRM_PVR_SYNC_OP_HANDLE_TYPE_MASK 1033 * 1034 * Handle type mask for the drm_pvr_sync_op::flags field. 1035 * 1036 * .. c:macro:: DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_SYNCOBJ 1037 * 1038 * Indicates the handle passed in drm_pvr_sync_op::handle is a syncobj handle. 1039 * This is the default type. 1040 * 1041 * .. c:macro:: DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_TIMELINE_SYNCOBJ 1042 * 1043 * Indicates the handle passed in drm_pvr_sync_op::handle is a timeline syncobj handle. 1044 * 1045 * .. c:macro:: DRM_PVR_SYNC_OP_FLAG_SIGNAL 1046 * 1047 * Signal operation requested. The out-fence bound to the job will be attached to 1048 * the syncobj whose handle is passed in drm_pvr_sync_op::handle. 1049 * 1050 * .. c:macro:: DRM_PVR_SYNC_OP_FLAG_WAIT 1051 * 1052 * Wait operation requested. The job will wait for this particular syncobj or syncobj 1053 * point to be signaled before being started. 1054 * This is the default operation. 1055 */ 1056 #define DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_MASK 0xf 1057 #define DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_SYNCOBJ 0 1058 #define DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_TIMELINE_SYNCOBJ 1 1059 #define DRM_PVR_SYNC_OP_FLAG_SIGNAL _BITULL(31) 1060 #define DRM_PVR_SYNC_OP_FLAG_WAIT 0 1061 1062 #define DRM_PVR_SYNC_OP_FLAGS_MASK (DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_MASK | \ 1063 DRM_PVR_SYNC_OP_FLAG_SIGNAL) 1064 1065 /** 1066 * struct drm_pvr_sync_op - Object describing a sync operation 1067 */ 1068 struct drm_pvr_sync_op { 1069 /** @handle: Handle of sync object. */ 1070 __u32 handle; 1071 1072 /** @flags: Combination of ``DRM_PVR_SYNC_OP_FLAG_`` flags. */ 1073 __u32 flags; 1074 1075 /** @value: Timeline value for this drm_syncobj. MBZ for a binary syncobj. */ 1076 __u64 value; 1077 }; 1078 1079 /** 1080 * DOC: Flags for SUBMIT_JOB ioctl geometry command. 1081 * 1082 * .. c:macro:: DRM_PVR_SUBMIT_JOB_GEOM_CMD_FIRST 1083 * 1084 * Indicates if this the first command to be issued for a render. 1085 * 1086 * .. c:macro:: DRM_PVR_SUBMIT_JOB_GEOM_CMD_LAST 1087 * 1088 * Indicates if this the last command to be issued for a render. 1089 * 1090 * .. c:macro:: DRM_PVR_SUBMIT_JOB_GEOM_CMD_SINGLE_CORE 1091 * 1092 * Forces to use single core in a multi core device. 1093 * 1094 * .. c:macro:: DRM_PVR_SUBMIT_JOB_GEOM_CMD_FLAGS_MASK 1095 * 1096 * Logical OR of all the geometry cmd flags. 1097 */ 1098 #define DRM_PVR_SUBMIT_JOB_GEOM_CMD_FIRST _BITULL(0) 1099 #define DRM_PVR_SUBMIT_JOB_GEOM_CMD_LAST _BITULL(1) 1100 #define DRM_PVR_SUBMIT_JOB_GEOM_CMD_SINGLE_CORE _BITULL(2) 1101 #define DRM_PVR_SUBMIT_JOB_GEOM_CMD_FLAGS_MASK \ 1102 (DRM_PVR_SUBMIT_JOB_GEOM_CMD_FIRST | \ 1103 DRM_PVR_SUBMIT_JOB_GEOM_CMD_LAST | \ 1104 DRM_PVR_SUBMIT_JOB_GEOM_CMD_SINGLE_CORE) 1105 1106 /** 1107 * DOC: Flags for SUBMIT_JOB ioctl fragment command. 1108 * 1109 * .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_SINGLE_CORE 1110 * 1111 * Use single core in a multi core setup. 1112 * 1113 * .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_DEPTHBUFFER 1114 * 1115 * Indicates whether a depth buffer is present. 1116 * 1117 * .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_STENCILBUFFER 1118 * 1119 * Indicates whether a stencil buffer is present. 1120 * 1121 * .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_PREVENT_CDM_OVERLAP 1122 * 1123 * Disallow compute overlapped with this render. 1124 * 1125 * .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_GET_VIS_RESULTS 1126 * 1127 * Indicates whether this render produces visibility results. 1128 * 1129 * .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_SCRATCHBUFFER 1130 * 1131 * Indicates whether partial renders write to a scratch buffer instead of 1132 * the final surface. It also forces the full screen copy expected to be 1133 * present on the last render after all partial renders have completed. 1134 * 1135 * .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_DISABLE_PIXELMERGE 1136 * 1137 * Disable pixel merging for this render. 1138 * 1139 * .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_FLAGS_MASK 1140 * 1141 * Logical OR of all the fragment cmd flags. 1142 */ 1143 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_SINGLE_CORE _BITULL(0) 1144 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_DEPTHBUFFER _BITULL(1) 1145 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_STENCILBUFFER _BITULL(2) 1146 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_PREVENT_CDM_OVERLAP _BITULL(3) 1147 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_SCRATCHBUFFER _BITULL(4) 1148 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_GET_VIS_RESULTS _BITULL(5) 1149 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_PARTIAL_RENDER _BITULL(6) 1150 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_DISABLE_PIXELMERGE _BITULL(7) 1151 #define DRM_PVR_SUBMIT_JOB_FRAG_CMD_FLAGS_MASK \ 1152 (DRM_PVR_SUBMIT_JOB_FRAG_CMD_SINGLE_CORE | \ 1153 DRM_PVR_SUBMIT_JOB_FRAG_CMD_DEPTHBUFFER | \ 1154 DRM_PVR_SUBMIT_JOB_FRAG_CMD_STENCILBUFFER | \ 1155 DRM_PVR_SUBMIT_JOB_FRAG_CMD_PREVENT_CDM_OVERLAP | \ 1156 DRM_PVR_SUBMIT_JOB_FRAG_CMD_SCRATCHBUFFER | \ 1157 DRM_PVR_SUBMIT_JOB_FRAG_CMD_GET_VIS_RESULTS | \ 1158 DRM_PVR_SUBMIT_JOB_FRAG_CMD_PARTIAL_RENDER | \ 1159 DRM_PVR_SUBMIT_JOB_FRAG_CMD_DISABLE_PIXELMERGE) 1160 1161 /** 1162 * DOC: Flags for SUBMIT_JOB ioctl compute command. 1163 * 1164 * .. c:macro:: DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_PREVENT_ALL_OVERLAP 1165 * 1166 * Disallow other jobs overlapped with this compute. 1167 * 1168 * .. c:macro:: DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_SINGLE_CORE 1169 * 1170 * Forces to use single core in a multi core device. 1171 * 1172 * .. c:macro:: DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_FLAGS_MASK 1173 * 1174 * Logical OR of all the compute cmd flags. 1175 */ 1176 #define DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_PREVENT_ALL_OVERLAP _BITULL(0) 1177 #define DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_SINGLE_CORE _BITULL(1) 1178 #define DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_FLAGS_MASK \ 1179 (DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_PREVENT_ALL_OVERLAP | \ 1180 DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_SINGLE_CORE) 1181 1182 /** 1183 * DOC: Flags for SUBMIT_JOB ioctl transfer command. 1184 * 1185 * .. c:macro:: DRM_PVR_SUBMIT_JOB_TRANSFER_CMD_SINGLE_CORE 1186 * 1187 * Forces job to use a single core in a multi core device. 1188 * 1189 * .. c:macro:: DRM_PVR_SUBMIT_JOB_TRANSFER_CMD_FLAGS_MASK 1190 * 1191 * Logical OR of all the transfer cmd flags. 1192 */ 1193 #define DRM_PVR_SUBMIT_JOB_TRANSFER_CMD_SINGLE_CORE _BITULL(0) 1194 1195 #define DRM_PVR_SUBMIT_JOB_TRANSFER_CMD_FLAGS_MASK \ 1196 DRM_PVR_SUBMIT_JOB_TRANSFER_CMD_SINGLE_CORE 1197 1198 /** 1199 * enum drm_pvr_job_type - Arguments for &struct drm_pvr_job.job_type 1200 */ 1201 enum drm_pvr_job_type { 1202 /** @DRM_PVR_JOB_TYPE_GEOMETRY: Job type is geometry. */ 1203 DRM_PVR_JOB_TYPE_GEOMETRY = 0, 1204 1205 /** @DRM_PVR_JOB_TYPE_FRAGMENT: Job type is fragment. */ 1206 DRM_PVR_JOB_TYPE_FRAGMENT, 1207 1208 /** @DRM_PVR_JOB_TYPE_COMPUTE: Job type is compute. */ 1209 DRM_PVR_JOB_TYPE_COMPUTE, 1210 1211 /** @DRM_PVR_JOB_TYPE_TRANSFER_FRAG: Job type is a fragment transfer. */ 1212 DRM_PVR_JOB_TYPE_TRANSFER_FRAG, 1213 }; 1214 1215 /** 1216 * struct drm_pvr_hwrt_data_ref - Reference HWRT data 1217 */ 1218 struct drm_pvr_hwrt_data_ref { 1219 /** @set_handle: HWRT data set handle. */ 1220 __u32 set_handle; 1221 1222 /** @data_index: Index of the HWRT data inside the data set. */ 1223 __u32 data_index; 1224 }; 1225 1226 /** 1227 * struct drm_pvr_job - Job arguments passed to the %DRM_IOCTL_PVR_SUBMIT_JOBS ioctl 1228 */ 1229 struct drm_pvr_job { 1230 /** 1231 * @type: [IN] Type of job being submitted 1232 * 1233 * This must be one of the values defined by &enum drm_pvr_job_type. 1234 */ 1235 __u32 type; 1236 1237 /** 1238 * @context_handle: [IN] Context handle. 1239 * 1240 * When @job_type is %DRM_PVR_JOB_TYPE_RENDER, %DRM_PVR_JOB_TYPE_COMPUTE or 1241 * %DRM_PVR_JOB_TYPE_TRANSFER_FRAG, this must be a valid handle returned by 1242 * %DRM_IOCTL_PVR_CREATE_CONTEXT. The type of context must be compatible 1243 * with the type of job being submitted. 1244 * 1245 * When @job_type is %DRM_PVR_JOB_TYPE_NULL, this must be zero. 1246 */ 1247 __u32 context_handle; 1248 1249 /** 1250 * @flags: [IN] Flags for command. 1251 * 1252 * Those are job-dependent. See all ``DRM_PVR_SUBMIT_JOB_*``. 1253 */ 1254 __u32 flags; 1255 1256 /** 1257 * @cmd_stream_len: [IN] Length of command stream, in bytes. 1258 */ 1259 __u32 cmd_stream_len; 1260 1261 /** 1262 * @cmd_stream: [IN] Pointer to command stream for command. 1263 * 1264 * The command stream must be u64-aligned. 1265 */ 1266 __u64 cmd_stream; 1267 1268 /** @sync_ops: [IN] Fragment sync operations. */ 1269 struct drm_pvr_obj_array sync_ops; 1270 1271 /** 1272 * @hwrt: [IN] HWRT data used by render jobs (geometry or fragment). 1273 * 1274 * Must be zero for non-render jobs. 1275 */ 1276 struct drm_pvr_hwrt_data_ref hwrt; 1277 }; 1278 1279 /** 1280 * struct drm_pvr_ioctl_submit_jobs_args - Arguments for %DRM_IOCTL_PVR_SUBMIT_JOB 1281 * 1282 * If the syscall returns an error it is important to check the value of 1283 * @jobs.count. This indicates the index into @jobs.array where the 1284 * error occurred. 1285 */ 1286 struct drm_pvr_ioctl_submit_jobs_args { 1287 /** @jobs: [IN] Array of jobs to submit. */ 1288 struct drm_pvr_obj_array jobs; 1289 }; 1290 1291 #if defined(__cplusplus) 1292 } 1293 #endif 1294 1295 #endif /* PVR_DRM_UAPI_H */ 1296