1c8b75bcaSEric Anholt /* 2c8b75bcaSEric Anholt * Copyright (C) 2015 Broadcom 3c8b75bcaSEric Anholt * 4c8b75bcaSEric Anholt * This program is free software; you can redistribute it and/or modify 5c8b75bcaSEric Anholt * it under the terms of the GNU General Public License version 2 as 6c8b75bcaSEric Anholt * published by the Free Software Foundation. 7c8b75bcaSEric Anholt */ 8c8b75bcaSEric Anholt 9c8b75bcaSEric Anholt #include "drmP.h" 10c8b75bcaSEric Anholt #include "drm_gem_cma_helper.h" 11c8b75bcaSEric Anholt 12c8b75bcaSEric Anholt struct vc4_dev { 13c8b75bcaSEric Anholt struct drm_device *dev; 14c8b75bcaSEric Anholt 15c8b75bcaSEric Anholt struct vc4_hdmi *hdmi; 16c8b75bcaSEric Anholt struct vc4_hvs *hvs; 17c8b75bcaSEric Anholt struct vc4_crtc *crtc[3]; 18d3f5168aSEric Anholt struct vc4_v3d *v3d; 1948666d56SDerek Foreman 2048666d56SDerek Foreman struct drm_fbdev_cma *fbdev; 21c826a6e1SEric Anholt 2221461365SEric Anholt struct vc4_hang_state *hang_state; 2321461365SEric Anholt 24c826a6e1SEric Anholt /* The kernel-space BO cache. Tracks buffers that have been 25c826a6e1SEric Anholt * unreferenced by all other users (refcounts of 0!) but not 26c826a6e1SEric Anholt * yet freed, so we can do cheap allocations. 27c826a6e1SEric Anholt */ 28c826a6e1SEric Anholt struct vc4_bo_cache { 29c826a6e1SEric Anholt /* Array of list heads for entries in the BO cache, 30c826a6e1SEric Anholt * based on number of pages, so we can do O(1) lookups 31c826a6e1SEric Anholt * in the cache when allocating. 32c826a6e1SEric Anholt */ 33c826a6e1SEric Anholt struct list_head *size_list; 34c826a6e1SEric Anholt uint32_t size_list_size; 35c826a6e1SEric Anholt 36c826a6e1SEric Anholt /* List of all BOs in the cache, ordered by age, so we 37c826a6e1SEric Anholt * can do O(1) lookups when trying to free old 38c826a6e1SEric Anholt * buffers. 39c826a6e1SEric Anholt */ 40c826a6e1SEric Anholt struct list_head time_list; 41c826a6e1SEric Anholt struct work_struct time_work; 42c826a6e1SEric Anholt struct timer_list time_timer; 43c826a6e1SEric Anholt } bo_cache; 44c826a6e1SEric Anholt 45c826a6e1SEric Anholt struct vc4_bo_stats { 46c826a6e1SEric Anholt u32 num_allocated; 47c826a6e1SEric Anholt u32 size_allocated; 48c826a6e1SEric Anholt u32 num_cached; 49c826a6e1SEric Anholt u32 size_cached; 50c826a6e1SEric Anholt } bo_stats; 51c826a6e1SEric Anholt 52c826a6e1SEric Anholt /* Protects bo_cache and the BO stats. */ 53c826a6e1SEric Anholt struct mutex bo_lock; 54d5b1a78aSEric Anholt 55d5b1a78aSEric Anholt /* Sequence number for the last job queued in job_list. 56d5b1a78aSEric Anholt * Starts at 0 (no jobs emitted). 57d5b1a78aSEric Anholt */ 58d5b1a78aSEric Anholt uint64_t emit_seqno; 59d5b1a78aSEric Anholt 60d5b1a78aSEric Anholt /* Sequence number for the last completed job on the GPU. 61d5b1a78aSEric Anholt * Starts at 0 (no jobs completed). 62d5b1a78aSEric Anholt */ 63d5b1a78aSEric Anholt uint64_t finished_seqno; 64d5b1a78aSEric Anholt 65d5b1a78aSEric Anholt /* List of all struct vc4_exec_info for jobs to be executed. 66d5b1a78aSEric Anholt * The first job in the list is the one currently programmed 67d5b1a78aSEric Anholt * into ct0ca/ct1ca for execution. 68d5b1a78aSEric Anholt */ 69d5b1a78aSEric Anholt struct list_head job_list; 70d5b1a78aSEric Anholt /* List of the finished vc4_exec_infos waiting to be freed by 71d5b1a78aSEric Anholt * job_done_work. 72d5b1a78aSEric Anholt */ 73d5b1a78aSEric Anholt struct list_head job_done_list; 74d5b1a78aSEric Anholt /* Spinlock used to synchronize the job_list and seqno 75d5b1a78aSEric Anholt * accesses between the IRQ handler and GEM ioctls. 76d5b1a78aSEric Anholt */ 77d5b1a78aSEric Anholt spinlock_t job_lock; 78d5b1a78aSEric Anholt wait_queue_head_t job_wait_queue; 79d5b1a78aSEric Anholt struct work_struct job_done_work; 80d5b1a78aSEric Anholt 81b501baccSEric Anholt /* List of struct vc4_seqno_cb for callbacks to be made from a 82b501baccSEric Anholt * workqueue when the given seqno is passed. 83b501baccSEric Anholt */ 84b501baccSEric Anholt struct list_head seqno_cb_list; 85b501baccSEric Anholt 86d5b1a78aSEric Anholt /* The binner overflow memory that's currently set up in 87d5b1a78aSEric Anholt * BPOA/BPOS registers. When overflow occurs and a new one is 88d5b1a78aSEric Anholt * allocated, the previous one will be moved to 89d5b1a78aSEric Anholt * vc4->current_exec's free list. 90d5b1a78aSEric Anholt */ 91d5b1a78aSEric Anholt struct vc4_bo *overflow_mem; 92d5b1a78aSEric Anholt struct work_struct overflow_mem_work; 93d5b1a78aSEric Anholt 94d5b1a78aSEric Anholt struct { 95d5b1a78aSEric Anholt struct timer_list timer; 96d5b1a78aSEric Anholt struct work_struct reset_work; 97d5b1a78aSEric Anholt } hangcheck; 98d5b1a78aSEric Anholt 99d5b1a78aSEric Anholt struct semaphore async_modeset; 100c8b75bcaSEric Anholt }; 101c8b75bcaSEric Anholt 102c8b75bcaSEric Anholt static inline struct vc4_dev * 103c8b75bcaSEric Anholt to_vc4_dev(struct drm_device *dev) 104c8b75bcaSEric Anholt { 105c8b75bcaSEric Anholt return (struct vc4_dev *)dev->dev_private; 106c8b75bcaSEric Anholt } 107c8b75bcaSEric Anholt 108c8b75bcaSEric Anholt struct vc4_bo { 109c8b75bcaSEric Anholt struct drm_gem_cma_object base; 110c826a6e1SEric Anholt 111d5b1a78aSEric Anholt /* seqno of the last job to render to this BO. */ 112d5b1a78aSEric Anholt uint64_t seqno; 113d5b1a78aSEric Anholt 114c826a6e1SEric Anholt /* List entry for the BO's position in either 115c826a6e1SEric Anholt * vc4_exec_info->unref_list or vc4_dev->bo_cache.time_list 116c826a6e1SEric Anholt */ 117c826a6e1SEric Anholt struct list_head unref_head; 118c826a6e1SEric Anholt 119c826a6e1SEric Anholt /* Time in jiffies when the BO was put in vc4->bo_cache. */ 120c826a6e1SEric Anholt unsigned long free_time; 121c826a6e1SEric Anholt 122c826a6e1SEric Anholt /* List entry for the BO's position in vc4_dev->bo_cache.size_list */ 123c826a6e1SEric Anholt struct list_head size_head; 124463873d5SEric Anholt 125463873d5SEric Anholt /* Struct for shader validation state, if created by 126463873d5SEric Anholt * DRM_IOCTL_VC4_CREATE_SHADER_BO. 127463873d5SEric Anholt */ 128463873d5SEric Anholt struct vc4_validated_shader_info *validated_shader; 129c8b75bcaSEric Anholt }; 130c8b75bcaSEric Anholt 131c8b75bcaSEric Anholt static inline struct vc4_bo * 132c8b75bcaSEric Anholt to_vc4_bo(struct drm_gem_object *bo) 133c8b75bcaSEric Anholt { 134c8b75bcaSEric Anholt return (struct vc4_bo *)bo; 135c8b75bcaSEric Anholt } 136c8b75bcaSEric Anholt 137b501baccSEric Anholt struct vc4_seqno_cb { 138b501baccSEric Anholt struct work_struct work; 139b501baccSEric Anholt uint64_t seqno; 140b501baccSEric Anholt void (*func)(struct vc4_seqno_cb *cb); 141b501baccSEric Anholt }; 142b501baccSEric Anholt 143d3f5168aSEric Anholt struct vc4_v3d { 144d3f5168aSEric Anholt struct platform_device *pdev; 145d3f5168aSEric Anholt void __iomem *regs; 146d3f5168aSEric Anholt }; 147d3f5168aSEric Anholt 148c8b75bcaSEric Anholt struct vc4_hvs { 149c8b75bcaSEric Anholt struct platform_device *pdev; 150c8b75bcaSEric Anholt void __iomem *regs; 151c8b75bcaSEric Anholt void __iomem *dlist; 152c8b75bcaSEric Anholt }; 153c8b75bcaSEric Anholt 154c8b75bcaSEric Anholt struct vc4_plane { 155c8b75bcaSEric Anholt struct drm_plane base; 156c8b75bcaSEric Anholt }; 157c8b75bcaSEric Anholt 158c8b75bcaSEric Anholt static inline struct vc4_plane * 159c8b75bcaSEric Anholt to_vc4_plane(struct drm_plane *plane) 160c8b75bcaSEric Anholt { 161c8b75bcaSEric Anholt return (struct vc4_plane *)plane; 162c8b75bcaSEric Anholt } 163c8b75bcaSEric Anholt 164c8b75bcaSEric Anholt enum vc4_encoder_type { 165c8b75bcaSEric Anholt VC4_ENCODER_TYPE_HDMI, 166c8b75bcaSEric Anholt VC4_ENCODER_TYPE_VEC, 167c8b75bcaSEric Anholt VC4_ENCODER_TYPE_DSI0, 168c8b75bcaSEric Anholt VC4_ENCODER_TYPE_DSI1, 169c8b75bcaSEric Anholt VC4_ENCODER_TYPE_SMI, 170c8b75bcaSEric Anholt VC4_ENCODER_TYPE_DPI, 171c8b75bcaSEric Anholt }; 172c8b75bcaSEric Anholt 173c8b75bcaSEric Anholt struct vc4_encoder { 174c8b75bcaSEric Anholt struct drm_encoder base; 175c8b75bcaSEric Anholt enum vc4_encoder_type type; 176c8b75bcaSEric Anholt u32 clock_select; 177c8b75bcaSEric Anholt }; 178c8b75bcaSEric Anholt 179c8b75bcaSEric Anholt static inline struct vc4_encoder * 180c8b75bcaSEric Anholt to_vc4_encoder(struct drm_encoder *encoder) 181c8b75bcaSEric Anholt { 182c8b75bcaSEric Anholt return container_of(encoder, struct vc4_encoder, base); 183c8b75bcaSEric Anholt } 184c8b75bcaSEric Anholt 185d3f5168aSEric Anholt #define V3D_READ(offset) readl(vc4->v3d->regs + offset) 186d3f5168aSEric Anholt #define V3D_WRITE(offset, val) writel(val, vc4->v3d->regs + offset) 187c8b75bcaSEric Anholt #define HVS_READ(offset) readl(vc4->hvs->regs + offset) 188c8b75bcaSEric Anholt #define HVS_WRITE(offset, val) writel(val, vc4->hvs->regs + offset) 189c8b75bcaSEric Anholt 190d5b1a78aSEric Anholt struct vc4_exec_info { 191d5b1a78aSEric Anholt /* Sequence number for this bin/render job. */ 192d5b1a78aSEric Anholt uint64_t seqno; 193d5b1a78aSEric Anholt 194*c4ce60dcSEric Anholt /* Last current addresses the hardware was processing when the 195*c4ce60dcSEric Anholt * hangcheck timer checked on us. 196*c4ce60dcSEric Anholt */ 197*c4ce60dcSEric Anholt uint32_t last_ct0ca, last_ct1ca; 198*c4ce60dcSEric Anholt 199d5b1a78aSEric Anholt /* Kernel-space copy of the ioctl arguments */ 200d5b1a78aSEric Anholt struct drm_vc4_submit_cl *args; 201d5b1a78aSEric Anholt 202d5b1a78aSEric Anholt /* This is the array of BOs that were looked up at the start of exec. 203d5b1a78aSEric Anholt * Command validation will use indices into this array. 204d5b1a78aSEric Anholt */ 205d5b1a78aSEric Anholt struct drm_gem_cma_object **bo; 206d5b1a78aSEric Anholt uint32_t bo_count; 207d5b1a78aSEric Anholt 208d5b1a78aSEric Anholt /* Pointers for our position in vc4->job_list */ 209d5b1a78aSEric Anholt struct list_head head; 210d5b1a78aSEric Anholt 211d5b1a78aSEric Anholt /* List of other BOs used in the job that need to be released 212d5b1a78aSEric Anholt * once the job is complete. 213d5b1a78aSEric Anholt */ 214d5b1a78aSEric Anholt struct list_head unref_list; 215d5b1a78aSEric Anholt 216d5b1a78aSEric Anholt /* Current unvalidated indices into @bo loaded by the non-hardware 217d5b1a78aSEric Anholt * VC4_PACKET_GEM_HANDLES. 218d5b1a78aSEric Anholt */ 219d5b1a78aSEric Anholt uint32_t bo_index[2]; 220d5b1a78aSEric Anholt 221d5b1a78aSEric Anholt /* This is the BO where we store the validated command lists, shader 222d5b1a78aSEric Anholt * records, and uniforms. 223d5b1a78aSEric Anholt */ 224d5b1a78aSEric Anholt struct drm_gem_cma_object *exec_bo; 225d5b1a78aSEric Anholt 226d5b1a78aSEric Anholt /** 227d5b1a78aSEric Anholt * This tracks the per-shader-record state (packet 64) that 228d5b1a78aSEric Anholt * determines the length of the shader record and the offset 229d5b1a78aSEric Anholt * it's expected to be found at. It gets read in from the 230d5b1a78aSEric Anholt * command lists. 231d5b1a78aSEric Anholt */ 232d5b1a78aSEric Anholt struct vc4_shader_state { 233d5b1a78aSEric Anholt uint32_t addr; 234d5b1a78aSEric Anholt /* Maximum vertex index referenced by any primitive using this 235d5b1a78aSEric Anholt * shader state. 236d5b1a78aSEric Anholt */ 237d5b1a78aSEric Anholt uint32_t max_index; 238d5b1a78aSEric Anholt } *shader_state; 239d5b1a78aSEric Anholt 240d5b1a78aSEric Anholt /** How many shader states the user declared they were using. */ 241d5b1a78aSEric Anholt uint32_t shader_state_size; 242d5b1a78aSEric Anholt /** How many shader state records the validator has seen. */ 243d5b1a78aSEric Anholt uint32_t shader_state_count; 244d5b1a78aSEric Anholt 245d5b1a78aSEric Anholt bool found_tile_binning_mode_config_packet; 246d5b1a78aSEric Anholt bool found_start_tile_binning_packet; 247d5b1a78aSEric Anholt bool found_increment_semaphore_packet; 248d5b1a78aSEric Anholt bool found_flush; 249d5b1a78aSEric Anholt uint8_t bin_tiles_x, bin_tiles_y; 250d5b1a78aSEric Anholt struct drm_gem_cma_object *tile_bo; 251d5b1a78aSEric Anholt uint32_t tile_alloc_offset; 252d5b1a78aSEric Anholt 253d5b1a78aSEric Anholt /** 254d5b1a78aSEric Anholt * Computed addresses pointing into exec_bo where we start the 255d5b1a78aSEric Anholt * bin thread (ct0) and render thread (ct1). 256d5b1a78aSEric Anholt */ 257d5b1a78aSEric Anholt uint32_t ct0ca, ct0ea; 258d5b1a78aSEric Anholt uint32_t ct1ca, ct1ea; 259d5b1a78aSEric Anholt 260d5b1a78aSEric Anholt /* Pointer to the unvalidated bin CL (if present). */ 261d5b1a78aSEric Anholt void *bin_u; 262d5b1a78aSEric Anholt 263d5b1a78aSEric Anholt /* Pointers to the shader recs. These paddr gets incremented as CL 264d5b1a78aSEric Anholt * packets are relocated in validate_gl_shader_state, and the vaddrs 265d5b1a78aSEric Anholt * (u and v) get incremented and size decremented as the shader recs 266d5b1a78aSEric Anholt * themselves are validated. 267d5b1a78aSEric Anholt */ 268d5b1a78aSEric Anholt void *shader_rec_u; 269d5b1a78aSEric Anholt void *shader_rec_v; 270d5b1a78aSEric Anholt uint32_t shader_rec_p; 271d5b1a78aSEric Anholt uint32_t shader_rec_size; 272d5b1a78aSEric Anholt 273d5b1a78aSEric Anholt /* Pointers to the uniform data. These pointers are incremented, and 274d5b1a78aSEric Anholt * size decremented, as each batch of uniforms is uploaded. 275d5b1a78aSEric Anholt */ 276d5b1a78aSEric Anholt void *uniforms_u; 277d5b1a78aSEric Anholt void *uniforms_v; 278d5b1a78aSEric Anholt uint32_t uniforms_p; 279d5b1a78aSEric Anholt uint32_t uniforms_size; 280d5b1a78aSEric Anholt }; 281d5b1a78aSEric Anholt 282d5b1a78aSEric Anholt static inline struct vc4_exec_info * 283d5b1a78aSEric Anholt vc4_first_job(struct vc4_dev *vc4) 284d5b1a78aSEric Anholt { 285d5b1a78aSEric Anholt if (list_empty(&vc4->job_list)) 286d5b1a78aSEric Anholt return NULL; 287d5b1a78aSEric Anholt return list_first_entry(&vc4->job_list, struct vc4_exec_info, head); 288d5b1a78aSEric Anholt } 289d5b1a78aSEric Anholt 290c8b75bcaSEric Anholt /** 291463873d5SEric Anholt * struct vc4_texture_sample_info - saves the offsets into the UBO for texture 292463873d5SEric Anholt * setup parameters. 293463873d5SEric Anholt * 294463873d5SEric Anholt * This will be used at draw time to relocate the reference to the texture 295463873d5SEric Anholt * contents in p0, and validate that the offset combined with 296463873d5SEric Anholt * width/height/stride/etc. from p1 and p2/p3 doesn't sample outside the BO. 297463873d5SEric Anholt * Note that the hardware treats unprovided config parameters as 0, so not all 298463873d5SEric Anholt * of them need to be set up for every texure sample, and we'll store ~0 as 299463873d5SEric Anholt * the offset to mark the unused ones. 300463873d5SEric Anholt * 301463873d5SEric Anholt * See the VC4 3D architecture guide page 41 ("Texture and Memory Lookup Unit 302463873d5SEric Anholt * Setup") for definitions of the texture parameters. 303463873d5SEric Anholt */ 304463873d5SEric Anholt struct vc4_texture_sample_info { 305463873d5SEric Anholt bool is_direct; 306463873d5SEric Anholt uint32_t p_offset[4]; 307463873d5SEric Anholt }; 308463873d5SEric Anholt 309463873d5SEric Anholt /** 310463873d5SEric Anholt * struct vc4_validated_shader_info - information about validated shaders that 311463873d5SEric Anholt * needs to be used from command list validation. 312463873d5SEric Anholt * 313463873d5SEric Anholt * For a given shader, each time a shader state record references it, we need 314463873d5SEric Anholt * to verify that the shader doesn't read more uniforms than the shader state 315463873d5SEric Anholt * record's uniform BO pointer can provide, and we need to apply relocations 316463873d5SEric Anholt * and validate the shader state record's uniforms that define the texture 317463873d5SEric Anholt * samples. 318463873d5SEric Anholt */ 319463873d5SEric Anholt struct vc4_validated_shader_info { 320463873d5SEric Anholt uint32_t uniforms_size; 321463873d5SEric Anholt uint32_t uniforms_src_size; 322463873d5SEric Anholt uint32_t num_texture_samples; 323463873d5SEric Anholt struct vc4_texture_sample_info *texture_samples; 324463873d5SEric Anholt }; 325463873d5SEric Anholt 326463873d5SEric Anholt /** 327c8b75bcaSEric Anholt * _wait_for - magic (register) wait macro 328c8b75bcaSEric Anholt * 329c8b75bcaSEric Anholt * Does the right thing for modeset paths when run under kdgb or similar atomic 330c8b75bcaSEric Anholt * contexts. Note that it's important that we check the condition again after 331c8b75bcaSEric Anholt * having timed out, since the timeout could be due to preemption or similar and 332c8b75bcaSEric Anholt * we've never had a chance to check the condition before the timeout. 333c8b75bcaSEric Anholt */ 334c8b75bcaSEric Anholt #define _wait_for(COND, MS, W) ({ \ 335c8b75bcaSEric Anholt unsigned long timeout__ = jiffies + msecs_to_jiffies(MS) + 1; \ 336c8b75bcaSEric Anholt int ret__ = 0; \ 337c8b75bcaSEric Anholt while (!(COND)) { \ 338c8b75bcaSEric Anholt if (time_after(jiffies, timeout__)) { \ 339c8b75bcaSEric Anholt if (!(COND)) \ 340c8b75bcaSEric Anholt ret__ = -ETIMEDOUT; \ 341c8b75bcaSEric Anholt break; \ 342c8b75bcaSEric Anholt } \ 343c8b75bcaSEric Anholt if (W && drm_can_sleep()) { \ 344c8b75bcaSEric Anholt msleep(W); \ 345c8b75bcaSEric Anholt } else { \ 346c8b75bcaSEric Anholt cpu_relax(); \ 347c8b75bcaSEric Anholt } \ 348c8b75bcaSEric Anholt } \ 349c8b75bcaSEric Anholt ret__; \ 350c8b75bcaSEric Anholt }) 351c8b75bcaSEric Anholt 352c8b75bcaSEric Anholt #define wait_for(COND, MS) _wait_for(COND, MS, 1) 353c8b75bcaSEric Anholt 354c8b75bcaSEric Anholt /* vc4_bo.c */ 355c826a6e1SEric Anholt struct drm_gem_object *vc4_create_object(struct drm_device *dev, size_t size); 356c8b75bcaSEric Anholt void vc4_free_object(struct drm_gem_object *gem_obj); 357c826a6e1SEric Anholt struct vc4_bo *vc4_bo_create(struct drm_device *dev, size_t size, 358c826a6e1SEric Anholt bool from_cache); 359c8b75bcaSEric Anholt int vc4_dumb_create(struct drm_file *file_priv, 360c8b75bcaSEric Anholt struct drm_device *dev, 361c8b75bcaSEric Anholt struct drm_mode_create_dumb *args); 362c8b75bcaSEric Anholt struct dma_buf *vc4_prime_export(struct drm_device *dev, 363c8b75bcaSEric Anholt struct drm_gem_object *obj, int flags); 364d5bc60f6SEric Anholt int vc4_create_bo_ioctl(struct drm_device *dev, void *data, 365d5bc60f6SEric Anholt struct drm_file *file_priv); 366463873d5SEric Anholt int vc4_create_shader_bo_ioctl(struct drm_device *dev, void *data, 367463873d5SEric Anholt struct drm_file *file_priv); 368d5bc60f6SEric Anholt int vc4_mmap_bo_ioctl(struct drm_device *dev, void *data, 369d5bc60f6SEric Anholt struct drm_file *file_priv); 37021461365SEric Anholt int vc4_get_hang_state_ioctl(struct drm_device *dev, void *data, 37121461365SEric Anholt struct drm_file *file_priv); 372463873d5SEric Anholt int vc4_mmap(struct file *filp, struct vm_area_struct *vma); 373463873d5SEric Anholt int vc4_prime_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma); 374463873d5SEric Anholt void *vc4_prime_vmap(struct drm_gem_object *obj); 375c826a6e1SEric Anholt void vc4_bo_cache_init(struct drm_device *dev); 376c826a6e1SEric Anholt void vc4_bo_cache_destroy(struct drm_device *dev); 377c826a6e1SEric Anholt int vc4_bo_stats_debugfs(struct seq_file *m, void *arg); 378c8b75bcaSEric Anholt 379c8b75bcaSEric Anholt /* vc4_crtc.c */ 380c8b75bcaSEric Anholt extern struct platform_driver vc4_crtc_driver; 3811f43710aSDave Airlie int vc4_enable_vblank(struct drm_device *dev, unsigned int crtc_id); 3821f43710aSDave Airlie void vc4_disable_vblank(struct drm_device *dev, unsigned int crtc_id); 383c8b75bcaSEric Anholt void vc4_cancel_page_flip(struct drm_crtc *crtc, struct drm_file *file); 384c8b75bcaSEric Anholt int vc4_crtc_debugfs_regs(struct seq_file *m, void *arg); 385c8b75bcaSEric Anholt 386c8b75bcaSEric Anholt /* vc4_debugfs.c */ 387c8b75bcaSEric Anholt int vc4_debugfs_init(struct drm_minor *minor); 388c8b75bcaSEric Anholt void vc4_debugfs_cleanup(struct drm_minor *minor); 389c8b75bcaSEric Anholt 390c8b75bcaSEric Anholt /* vc4_drv.c */ 391c8b75bcaSEric Anholt void __iomem *vc4_ioremap_regs(struct platform_device *dev, int index); 392c8b75bcaSEric Anholt 393d5b1a78aSEric Anholt /* vc4_gem.c */ 394d5b1a78aSEric Anholt void vc4_gem_init(struct drm_device *dev); 395d5b1a78aSEric Anholt void vc4_gem_destroy(struct drm_device *dev); 396d5b1a78aSEric Anholt int vc4_submit_cl_ioctl(struct drm_device *dev, void *data, 397d5b1a78aSEric Anholt struct drm_file *file_priv); 398d5b1a78aSEric Anholt int vc4_wait_seqno_ioctl(struct drm_device *dev, void *data, 399d5b1a78aSEric Anholt struct drm_file *file_priv); 400d5b1a78aSEric Anholt int vc4_wait_bo_ioctl(struct drm_device *dev, void *data, 401d5b1a78aSEric Anholt struct drm_file *file_priv); 402d5b1a78aSEric Anholt void vc4_submit_next_job(struct drm_device *dev); 403d5b1a78aSEric Anholt int vc4_wait_for_seqno(struct drm_device *dev, uint64_t seqno, 404d5b1a78aSEric Anholt uint64_t timeout_ns, bool interruptible); 405d5b1a78aSEric Anholt void vc4_job_handle_completed(struct vc4_dev *vc4); 406b501baccSEric Anholt int vc4_queue_seqno_cb(struct drm_device *dev, 407b501baccSEric Anholt struct vc4_seqno_cb *cb, uint64_t seqno, 408b501baccSEric Anholt void (*func)(struct vc4_seqno_cb *cb)); 409d5b1a78aSEric Anholt 410c8b75bcaSEric Anholt /* vc4_hdmi.c */ 411c8b75bcaSEric Anholt extern struct platform_driver vc4_hdmi_driver; 412c8b75bcaSEric Anholt int vc4_hdmi_debugfs_regs(struct seq_file *m, void *unused); 413c8b75bcaSEric Anholt 414d5b1a78aSEric Anholt /* vc4_irq.c */ 415d5b1a78aSEric Anholt irqreturn_t vc4_irq(int irq, void *arg); 416d5b1a78aSEric Anholt void vc4_irq_preinstall(struct drm_device *dev); 417d5b1a78aSEric Anholt int vc4_irq_postinstall(struct drm_device *dev); 418d5b1a78aSEric Anholt void vc4_irq_uninstall(struct drm_device *dev); 419d5b1a78aSEric Anholt void vc4_irq_reset(struct drm_device *dev); 420d5b1a78aSEric Anholt 421c8b75bcaSEric Anholt /* vc4_hvs.c */ 422c8b75bcaSEric Anholt extern struct platform_driver vc4_hvs_driver; 423c8b75bcaSEric Anholt void vc4_hvs_dump_state(struct drm_device *dev); 424c8b75bcaSEric Anholt int vc4_hvs_debugfs_regs(struct seq_file *m, void *unused); 425c8b75bcaSEric Anholt 426c8b75bcaSEric Anholt /* vc4_kms.c */ 427c8b75bcaSEric Anholt int vc4_kms_load(struct drm_device *dev); 428c8b75bcaSEric Anholt 429c8b75bcaSEric Anholt /* vc4_plane.c */ 430c8b75bcaSEric Anholt struct drm_plane *vc4_plane_init(struct drm_device *dev, 431c8b75bcaSEric Anholt enum drm_plane_type type); 432c8b75bcaSEric Anholt u32 vc4_plane_write_dlist(struct drm_plane *plane, u32 __iomem *dlist); 433c8b75bcaSEric Anholt u32 vc4_plane_dlist_size(struct drm_plane_state *state); 434b501baccSEric Anholt void vc4_plane_async_set_fb(struct drm_plane *plane, 435b501baccSEric Anholt struct drm_framebuffer *fb); 436463873d5SEric Anholt 437d3f5168aSEric Anholt /* vc4_v3d.c */ 438d3f5168aSEric Anholt extern struct platform_driver vc4_v3d_driver; 439d3f5168aSEric Anholt int vc4_v3d_debugfs_ident(struct seq_file *m, void *unused); 440d3f5168aSEric Anholt int vc4_v3d_debugfs_regs(struct seq_file *m, void *unused); 441d5b1a78aSEric Anholt int vc4_v3d_set_power(struct vc4_dev *vc4, bool on); 442d5b1a78aSEric Anholt 443d5b1a78aSEric Anholt /* vc4_validate.c */ 444d5b1a78aSEric Anholt int 445d5b1a78aSEric Anholt vc4_validate_bin_cl(struct drm_device *dev, 446d5b1a78aSEric Anholt void *validated, 447d5b1a78aSEric Anholt void *unvalidated, 448d5b1a78aSEric Anholt struct vc4_exec_info *exec); 449d5b1a78aSEric Anholt 450d5b1a78aSEric Anholt int 451d5b1a78aSEric Anholt vc4_validate_shader_recs(struct drm_device *dev, struct vc4_exec_info *exec); 452d5b1a78aSEric Anholt 453d5b1a78aSEric Anholt struct drm_gem_cma_object *vc4_use_bo(struct vc4_exec_info *exec, 454d5b1a78aSEric Anholt uint32_t hindex); 455d5b1a78aSEric Anholt 456d5b1a78aSEric Anholt int vc4_get_rcl(struct drm_device *dev, struct vc4_exec_info *exec); 457d5b1a78aSEric Anholt 458d5b1a78aSEric Anholt bool vc4_check_tex_size(struct vc4_exec_info *exec, 459d5b1a78aSEric Anholt struct drm_gem_cma_object *fbo, 460d5b1a78aSEric Anholt uint32_t offset, uint8_t tiling_format, 461d5b1a78aSEric Anholt uint32_t width, uint32_t height, uint8_t cpp); 462d3f5168aSEric Anholt 463463873d5SEric Anholt /* vc4_validate_shader.c */ 464463873d5SEric Anholt struct vc4_validated_shader_info * 465463873d5SEric Anholt vc4_validate_shader(struct drm_gem_cma_object *shader_obj); 466