1c8b75bcaSEric Anholt /* 2c8b75bcaSEric Anholt * Copyright (C) 2015 Broadcom 3c8b75bcaSEric Anholt * 4c8b75bcaSEric Anholt * This program is free software; you can redistribute it and/or modify 5c8b75bcaSEric Anholt * it under the terms of the GNU General Public License version 2 as 6c8b75bcaSEric Anholt * published by the Free Software Foundation. 7c8b75bcaSEric Anholt */ 8c8b75bcaSEric Anholt 9c8b75bcaSEric Anholt #include "drmP.h" 10c8b75bcaSEric Anholt #include "drm_gem_cma_helper.h" 11c8b75bcaSEric Anholt 12c8b75bcaSEric Anholt struct vc4_dev { 13c8b75bcaSEric Anholt struct drm_device *dev; 14c8b75bcaSEric Anholt 15c8b75bcaSEric Anholt struct vc4_hdmi *hdmi; 16c8b75bcaSEric Anholt struct vc4_hvs *hvs; 17c8b75bcaSEric Anholt struct vc4_crtc *crtc[3]; 18d3f5168aSEric Anholt struct vc4_v3d *v3d; 1948666d56SDerek Foreman 2048666d56SDerek Foreman struct drm_fbdev_cma *fbdev; 21c826a6e1SEric Anholt 2221461365SEric Anholt struct vc4_hang_state *hang_state; 2321461365SEric Anholt 24c826a6e1SEric Anholt /* The kernel-space BO cache. Tracks buffers that have been 25c826a6e1SEric Anholt * unreferenced by all other users (refcounts of 0!) but not 26c826a6e1SEric Anholt * yet freed, so we can do cheap allocations. 27c826a6e1SEric Anholt */ 28c826a6e1SEric Anholt struct vc4_bo_cache { 29c826a6e1SEric Anholt /* Array of list heads for entries in the BO cache, 30c826a6e1SEric Anholt * based on number of pages, so we can do O(1) lookups 31c826a6e1SEric Anholt * in the cache when allocating. 32c826a6e1SEric Anholt */ 33c826a6e1SEric Anholt struct list_head *size_list; 34c826a6e1SEric Anholt uint32_t size_list_size; 35c826a6e1SEric Anholt 36c826a6e1SEric Anholt /* List of all BOs in the cache, ordered by age, so we 37c826a6e1SEric Anholt * can do O(1) lookups when trying to free old 38c826a6e1SEric Anholt * buffers. 39c826a6e1SEric Anholt */ 40c826a6e1SEric Anholt struct list_head time_list; 41c826a6e1SEric Anholt struct work_struct time_work; 42c826a6e1SEric Anholt struct timer_list time_timer; 43c826a6e1SEric Anholt } bo_cache; 44c826a6e1SEric Anholt 45c826a6e1SEric Anholt struct vc4_bo_stats { 46c826a6e1SEric Anholt u32 num_allocated; 47c826a6e1SEric Anholt u32 size_allocated; 48c826a6e1SEric Anholt u32 num_cached; 49c826a6e1SEric Anholt u32 size_cached; 50c826a6e1SEric Anholt } bo_stats; 51c826a6e1SEric Anholt 52c826a6e1SEric Anholt /* Protects bo_cache and the BO stats. */ 53c826a6e1SEric Anholt struct mutex bo_lock; 54d5b1a78aSEric Anholt 55d5b1a78aSEric Anholt /* Sequence number for the last job queued in job_list. 56d5b1a78aSEric Anholt * Starts at 0 (no jobs emitted). 57d5b1a78aSEric Anholt */ 58d5b1a78aSEric Anholt uint64_t emit_seqno; 59d5b1a78aSEric Anholt 60d5b1a78aSEric Anholt /* Sequence number for the last completed job on the GPU. 61d5b1a78aSEric Anholt * Starts at 0 (no jobs completed). 62d5b1a78aSEric Anholt */ 63d5b1a78aSEric Anholt uint64_t finished_seqno; 64d5b1a78aSEric Anholt 65d5b1a78aSEric Anholt /* List of all struct vc4_exec_info for jobs to be executed. 66d5b1a78aSEric Anholt * The first job in the list is the one currently programmed 67d5b1a78aSEric Anholt * into ct0ca/ct1ca for execution. 68d5b1a78aSEric Anholt */ 69d5b1a78aSEric Anholt struct list_head job_list; 70d5b1a78aSEric Anholt /* List of the finished vc4_exec_infos waiting to be freed by 71d5b1a78aSEric Anholt * job_done_work. 72d5b1a78aSEric Anholt */ 73d5b1a78aSEric Anholt struct list_head job_done_list; 74d5b1a78aSEric Anholt /* Spinlock used to synchronize the job_list and seqno 75d5b1a78aSEric Anholt * accesses between the IRQ handler and GEM ioctls. 76d5b1a78aSEric Anholt */ 77d5b1a78aSEric Anholt spinlock_t job_lock; 78d5b1a78aSEric Anholt wait_queue_head_t job_wait_queue; 79d5b1a78aSEric Anholt struct work_struct job_done_work; 80d5b1a78aSEric Anholt 81b501baccSEric Anholt /* List of struct vc4_seqno_cb for callbacks to be made from a 82b501baccSEric Anholt * workqueue when the given seqno is passed. 83b501baccSEric Anholt */ 84b501baccSEric Anholt struct list_head seqno_cb_list; 85b501baccSEric Anholt 86d5b1a78aSEric Anholt /* The binner overflow memory that's currently set up in 87d5b1a78aSEric Anholt * BPOA/BPOS registers. When overflow occurs and a new one is 88d5b1a78aSEric Anholt * allocated, the previous one will be moved to 89d5b1a78aSEric Anholt * vc4->current_exec's free list. 90d5b1a78aSEric Anholt */ 91d5b1a78aSEric Anholt struct vc4_bo *overflow_mem; 92d5b1a78aSEric Anholt struct work_struct overflow_mem_work; 93d5b1a78aSEric Anholt 94*36cb6253SEric Anholt int power_refcount; 95*36cb6253SEric Anholt 96*36cb6253SEric Anholt /* Mutex controlling the power refcount. */ 97*36cb6253SEric Anholt struct mutex power_lock; 98*36cb6253SEric Anholt 99d5b1a78aSEric Anholt struct { 100d5b1a78aSEric Anholt struct timer_list timer; 101d5b1a78aSEric Anholt struct work_struct reset_work; 102d5b1a78aSEric Anholt } hangcheck; 103d5b1a78aSEric Anholt 104d5b1a78aSEric Anholt struct semaphore async_modeset; 105c8b75bcaSEric Anholt }; 106c8b75bcaSEric Anholt 107c8b75bcaSEric Anholt static inline struct vc4_dev * 108c8b75bcaSEric Anholt to_vc4_dev(struct drm_device *dev) 109c8b75bcaSEric Anholt { 110c8b75bcaSEric Anholt return (struct vc4_dev *)dev->dev_private; 111c8b75bcaSEric Anholt } 112c8b75bcaSEric Anholt 113c8b75bcaSEric Anholt struct vc4_bo { 114c8b75bcaSEric Anholt struct drm_gem_cma_object base; 115c826a6e1SEric Anholt 116d5b1a78aSEric Anholt /* seqno of the last job to render to this BO. */ 117d5b1a78aSEric Anholt uint64_t seqno; 118d5b1a78aSEric Anholt 119c826a6e1SEric Anholt /* List entry for the BO's position in either 120c826a6e1SEric Anholt * vc4_exec_info->unref_list or vc4_dev->bo_cache.time_list 121c826a6e1SEric Anholt */ 122c826a6e1SEric Anholt struct list_head unref_head; 123c826a6e1SEric Anholt 124c826a6e1SEric Anholt /* Time in jiffies when the BO was put in vc4->bo_cache. */ 125c826a6e1SEric Anholt unsigned long free_time; 126c826a6e1SEric Anholt 127c826a6e1SEric Anholt /* List entry for the BO's position in vc4_dev->bo_cache.size_list */ 128c826a6e1SEric Anholt struct list_head size_head; 129463873d5SEric Anholt 130463873d5SEric Anholt /* Struct for shader validation state, if created by 131463873d5SEric Anholt * DRM_IOCTL_VC4_CREATE_SHADER_BO. 132463873d5SEric Anholt */ 133463873d5SEric Anholt struct vc4_validated_shader_info *validated_shader; 134c8b75bcaSEric Anholt }; 135c8b75bcaSEric Anholt 136c8b75bcaSEric Anholt static inline struct vc4_bo * 137c8b75bcaSEric Anholt to_vc4_bo(struct drm_gem_object *bo) 138c8b75bcaSEric Anholt { 139c8b75bcaSEric Anholt return (struct vc4_bo *)bo; 140c8b75bcaSEric Anholt } 141c8b75bcaSEric Anholt 142b501baccSEric Anholt struct vc4_seqno_cb { 143b501baccSEric Anholt struct work_struct work; 144b501baccSEric Anholt uint64_t seqno; 145b501baccSEric Anholt void (*func)(struct vc4_seqno_cb *cb); 146b501baccSEric Anholt }; 147b501baccSEric Anholt 148d3f5168aSEric Anholt struct vc4_v3d { 149001bdb55SEric Anholt struct vc4_dev *vc4; 150d3f5168aSEric Anholt struct platform_device *pdev; 151d3f5168aSEric Anholt void __iomem *regs; 152d3f5168aSEric Anholt }; 153d3f5168aSEric Anholt 154c8b75bcaSEric Anholt struct vc4_hvs { 155c8b75bcaSEric Anholt struct platform_device *pdev; 156c8b75bcaSEric Anholt void __iomem *regs; 157c8b75bcaSEric Anholt void __iomem *dlist; 158c8b75bcaSEric Anholt }; 159c8b75bcaSEric Anholt 160c8b75bcaSEric Anholt struct vc4_plane { 161c8b75bcaSEric Anholt struct drm_plane base; 162c8b75bcaSEric Anholt }; 163c8b75bcaSEric Anholt 164c8b75bcaSEric Anholt static inline struct vc4_plane * 165c8b75bcaSEric Anholt to_vc4_plane(struct drm_plane *plane) 166c8b75bcaSEric Anholt { 167c8b75bcaSEric Anholt return (struct vc4_plane *)plane; 168c8b75bcaSEric Anholt } 169c8b75bcaSEric Anholt 170c8b75bcaSEric Anholt enum vc4_encoder_type { 171c8b75bcaSEric Anholt VC4_ENCODER_TYPE_HDMI, 172c8b75bcaSEric Anholt VC4_ENCODER_TYPE_VEC, 173c8b75bcaSEric Anholt VC4_ENCODER_TYPE_DSI0, 174c8b75bcaSEric Anholt VC4_ENCODER_TYPE_DSI1, 175c8b75bcaSEric Anholt VC4_ENCODER_TYPE_SMI, 176c8b75bcaSEric Anholt VC4_ENCODER_TYPE_DPI, 177c8b75bcaSEric Anholt }; 178c8b75bcaSEric Anholt 179c8b75bcaSEric Anholt struct vc4_encoder { 180c8b75bcaSEric Anholt struct drm_encoder base; 181c8b75bcaSEric Anholt enum vc4_encoder_type type; 182c8b75bcaSEric Anholt u32 clock_select; 183c8b75bcaSEric Anholt }; 184c8b75bcaSEric Anholt 185c8b75bcaSEric Anholt static inline struct vc4_encoder * 186c8b75bcaSEric Anholt to_vc4_encoder(struct drm_encoder *encoder) 187c8b75bcaSEric Anholt { 188c8b75bcaSEric Anholt return container_of(encoder, struct vc4_encoder, base); 189c8b75bcaSEric Anholt } 190c8b75bcaSEric Anholt 191d3f5168aSEric Anholt #define V3D_READ(offset) readl(vc4->v3d->regs + offset) 192d3f5168aSEric Anholt #define V3D_WRITE(offset, val) writel(val, vc4->v3d->regs + offset) 193c8b75bcaSEric Anholt #define HVS_READ(offset) readl(vc4->hvs->regs + offset) 194c8b75bcaSEric Anholt #define HVS_WRITE(offset, val) writel(val, vc4->hvs->regs + offset) 195c8b75bcaSEric Anholt 196d5b1a78aSEric Anholt struct vc4_exec_info { 197d5b1a78aSEric Anholt /* Sequence number for this bin/render job. */ 198d5b1a78aSEric Anholt uint64_t seqno; 199d5b1a78aSEric Anholt 200c4ce60dcSEric Anholt /* Last current addresses the hardware was processing when the 201c4ce60dcSEric Anholt * hangcheck timer checked on us. 202c4ce60dcSEric Anholt */ 203c4ce60dcSEric Anholt uint32_t last_ct0ca, last_ct1ca; 204c4ce60dcSEric Anholt 205d5b1a78aSEric Anholt /* Kernel-space copy of the ioctl arguments */ 206d5b1a78aSEric Anholt struct drm_vc4_submit_cl *args; 207d5b1a78aSEric Anholt 208d5b1a78aSEric Anholt /* This is the array of BOs that were looked up at the start of exec. 209d5b1a78aSEric Anholt * Command validation will use indices into this array. 210d5b1a78aSEric Anholt */ 211d5b1a78aSEric Anholt struct drm_gem_cma_object **bo; 212d5b1a78aSEric Anholt uint32_t bo_count; 213d5b1a78aSEric Anholt 214d5b1a78aSEric Anholt /* Pointers for our position in vc4->job_list */ 215d5b1a78aSEric Anholt struct list_head head; 216d5b1a78aSEric Anholt 217d5b1a78aSEric Anholt /* List of other BOs used in the job that need to be released 218d5b1a78aSEric Anholt * once the job is complete. 219d5b1a78aSEric Anholt */ 220d5b1a78aSEric Anholt struct list_head unref_list; 221d5b1a78aSEric Anholt 222d5b1a78aSEric Anholt /* Current unvalidated indices into @bo loaded by the non-hardware 223d5b1a78aSEric Anholt * VC4_PACKET_GEM_HANDLES. 224d5b1a78aSEric Anholt */ 225d5b1a78aSEric Anholt uint32_t bo_index[2]; 226d5b1a78aSEric Anholt 227d5b1a78aSEric Anholt /* This is the BO where we store the validated command lists, shader 228d5b1a78aSEric Anholt * records, and uniforms. 229d5b1a78aSEric Anholt */ 230d5b1a78aSEric Anholt struct drm_gem_cma_object *exec_bo; 231d5b1a78aSEric Anholt 232d5b1a78aSEric Anholt /** 233d5b1a78aSEric Anholt * This tracks the per-shader-record state (packet 64) that 234d5b1a78aSEric Anholt * determines the length of the shader record and the offset 235d5b1a78aSEric Anholt * it's expected to be found at. It gets read in from the 236d5b1a78aSEric Anholt * command lists. 237d5b1a78aSEric Anholt */ 238d5b1a78aSEric Anholt struct vc4_shader_state { 239d5b1a78aSEric Anholt uint32_t addr; 240d5b1a78aSEric Anholt /* Maximum vertex index referenced by any primitive using this 241d5b1a78aSEric Anholt * shader state. 242d5b1a78aSEric Anholt */ 243d5b1a78aSEric Anholt uint32_t max_index; 244d5b1a78aSEric Anholt } *shader_state; 245d5b1a78aSEric Anholt 246d5b1a78aSEric Anholt /** How many shader states the user declared they were using. */ 247d5b1a78aSEric Anholt uint32_t shader_state_size; 248d5b1a78aSEric Anholt /** How many shader state records the validator has seen. */ 249d5b1a78aSEric Anholt uint32_t shader_state_count; 250d5b1a78aSEric Anholt 251d5b1a78aSEric Anholt bool found_tile_binning_mode_config_packet; 252d5b1a78aSEric Anholt bool found_start_tile_binning_packet; 253d5b1a78aSEric Anholt bool found_increment_semaphore_packet; 254d5b1a78aSEric Anholt bool found_flush; 255d5b1a78aSEric Anholt uint8_t bin_tiles_x, bin_tiles_y; 256d5b1a78aSEric Anholt struct drm_gem_cma_object *tile_bo; 257d5b1a78aSEric Anholt uint32_t tile_alloc_offset; 258d5b1a78aSEric Anholt 259d5b1a78aSEric Anholt /** 260d5b1a78aSEric Anholt * Computed addresses pointing into exec_bo where we start the 261d5b1a78aSEric Anholt * bin thread (ct0) and render thread (ct1). 262d5b1a78aSEric Anholt */ 263d5b1a78aSEric Anholt uint32_t ct0ca, ct0ea; 264d5b1a78aSEric Anholt uint32_t ct1ca, ct1ea; 265d5b1a78aSEric Anholt 266d5b1a78aSEric Anholt /* Pointer to the unvalidated bin CL (if present). */ 267d5b1a78aSEric Anholt void *bin_u; 268d5b1a78aSEric Anholt 269d5b1a78aSEric Anholt /* Pointers to the shader recs. These paddr gets incremented as CL 270d5b1a78aSEric Anholt * packets are relocated in validate_gl_shader_state, and the vaddrs 271d5b1a78aSEric Anholt * (u and v) get incremented and size decremented as the shader recs 272d5b1a78aSEric Anholt * themselves are validated. 273d5b1a78aSEric Anholt */ 274d5b1a78aSEric Anholt void *shader_rec_u; 275d5b1a78aSEric Anholt void *shader_rec_v; 276d5b1a78aSEric Anholt uint32_t shader_rec_p; 277d5b1a78aSEric Anholt uint32_t shader_rec_size; 278d5b1a78aSEric Anholt 279d5b1a78aSEric Anholt /* Pointers to the uniform data. These pointers are incremented, and 280d5b1a78aSEric Anholt * size decremented, as each batch of uniforms is uploaded. 281d5b1a78aSEric Anholt */ 282d5b1a78aSEric Anholt void *uniforms_u; 283d5b1a78aSEric Anholt void *uniforms_v; 284d5b1a78aSEric Anholt uint32_t uniforms_p; 285d5b1a78aSEric Anholt uint32_t uniforms_size; 286d5b1a78aSEric Anholt }; 287d5b1a78aSEric Anholt 288d5b1a78aSEric Anholt static inline struct vc4_exec_info * 289d5b1a78aSEric Anholt vc4_first_job(struct vc4_dev *vc4) 290d5b1a78aSEric Anholt { 291d5b1a78aSEric Anholt if (list_empty(&vc4->job_list)) 292d5b1a78aSEric Anholt return NULL; 293d5b1a78aSEric Anholt return list_first_entry(&vc4->job_list, struct vc4_exec_info, head); 294d5b1a78aSEric Anholt } 295d5b1a78aSEric Anholt 296c8b75bcaSEric Anholt /** 297463873d5SEric Anholt * struct vc4_texture_sample_info - saves the offsets into the UBO for texture 298463873d5SEric Anholt * setup parameters. 299463873d5SEric Anholt * 300463873d5SEric Anholt * This will be used at draw time to relocate the reference to the texture 301463873d5SEric Anholt * contents in p0, and validate that the offset combined with 302463873d5SEric Anholt * width/height/stride/etc. from p1 and p2/p3 doesn't sample outside the BO. 303463873d5SEric Anholt * Note that the hardware treats unprovided config parameters as 0, so not all 304463873d5SEric Anholt * of them need to be set up for every texure sample, and we'll store ~0 as 305463873d5SEric Anholt * the offset to mark the unused ones. 306463873d5SEric Anholt * 307463873d5SEric Anholt * See the VC4 3D architecture guide page 41 ("Texture and Memory Lookup Unit 308463873d5SEric Anholt * Setup") for definitions of the texture parameters. 309463873d5SEric Anholt */ 310463873d5SEric Anholt struct vc4_texture_sample_info { 311463873d5SEric Anholt bool is_direct; 312463873d5SEric Anholt uint32_t p_offset[4]; 313463873d5SEric Anholt }; 314463873d5SEric Anholt 315463873d5SEric Anholt /** 316463873d5SEric Anholt * struct vc4_validated_shader_info - information about validated shaders that 317463873d5SEric Anholt * needs to be used from command list validation. 318463873d5SEric Anholt * 319463873d5SEric Anholt * For a given shader, each time a shader state record references it, we need 320463873d5SEric Anholt * to verify that the shader doesn't read more uniforms than the shader state 321463873d5SEric Anholt * record's uniform BO pointer can provide, and we need to apply relocations 322463873d5SEric Anholt * and validate the shader state record's uniforms that define the texture 323463873d5SEric Anholt * samples. 324463873d5SEric Anholt */ 325463873d5SEric Anholt struct vc4_validated_shader_info { 326463873d5SEric Anholt uint32_t uniforms_size; 327463873d5SEric Anholt uint32_t uniforms_src_size; 328463873d5SEric Anholt uint32_t num_texture_samples; 329463873d5SEric Anholt struct vc4_texture_sample_info *texture_samples; 330463873d5SEric Anholt }; 331463873d5SEric Anholt 332463873d5SEric Anholt /** 333c8b75bcaSEric Anholt * _wait_for - magic (register) wait macro 334c8b75bcaSEric Anholt * 335c8b75bcaSEric Anholt * Does the right thing for modeset paths when run under kdgb or similar atomic 336c8b75bcaSEric Anholt * contexts. Note that it's important that we check the condition again after 337c8b75bcaSEric Anholt * having timed out, since the timeout could be due to preemption or similar and 338c8b75bcaSEric Anholt * we've never had a chance to check the condition before the timeout. 339c8b75bcaSEric Anholt */ 340c8b75bcaSEric Anholt #define _wait_for(COND, MS, W) ({ \ 341c8b75bcaSEric Anholt unsigned long timeout__ = jiffies + msecs_to_jiffies(MS) + 1; \ 342c8b75bcaSEric Anholt int ret__ = 0; \ 343c8b75bcaSEric Anholt while (!(COND)) { \ 344c8b75bcaSEric Anholt if (time_after(jiffies, timeout__)) { \ 345c8b75bcaSEric Anholt if (!(COND)) \ 346c8b75bcaSEric Anholt ret__ = -ETIMEDOUT; \ 347c8b75bcaSEric Anholt break; \ 348c8b75bcaSEric Anholt } \ 349c8b75bcaSEric Anholt if (W && drm_can_sleep()) { \ 350c8b75bcaSEric Anholt msleep(W); \ 351c8b75bcaSEric Anholt } else { \ 352c8b75bcaSEric Anholt cpu_relax(); \ 353c8b75bcaSEric Anholt } \ 354c8b75bcaSEric Anholt } \ 355c8b75bcaSEric Anholt ret__; \ 356c8b75bcaSEric Anholt }) 357c8b75bcaSEric Anholt 358c8b75bcaSEric Anholt #define wait_for(COND, MS) _wait_for(COND, MS, 1) 359c8b75bcaSEric Anholt 360c8b75bcaSEric Anholt /* vc4_bo.c */ 361c826a6e1SEric Anholt struct drm_gem_object *vc4_create_object(struct drm_device *dev, size_t size); 362c8b75bcaSEric Anholt void vc4_free_object(struct drm_gem_object *gem_obj); 363c826a6e1SEric Anholt struct vc4_bo *vc4_bo_create(struct drm_device *dev, size_t size, 364c826a6e1SEric Anholt bool from_cache); 365c8b75bcaSEric Anholt int vc4_dumb_create(struct drm_file *file_priv, 366c8b75bcaSEric Anholt struct drm_device *dev, 367c8b75bcaSEric Anholt struct drm_mode_create_dumb *args); 368c8b75bcaSEric Anholt struct dma_buf *vc4_prime_export(struct drm_device *dev, 369c8b75bcaSEric Anholt struct drm_gem_object *obj, int flags); 370d5bc60f6SEric Anholt int vc4_create_bo_ioctl(struct drm_device *dev, void *data, 371d5bc60f6SEric Anholt struct drm_file *file_priv); 372463873d5SEric Anholt int vc4_create_shader_bo_ioctl(struct drm_device *dev, void *data, 373463873d5SEric Anholt struct drm_file *file_priv); 374d5bc60f6SEric Anholt int vc4_mmap_bo_ioctl(struct drm_device *dev, void *data, 375d5bc60f6SEric Anholt struct drm_file *file_priv); 37621461365SEric Anholt int vc4_get_hang_state_ioctl(struct drm_device *dev, void *data, 37721461365SEric Anholt struct drm_file *file_priv); 378463873d5SEric Anholt int vc4_mmap(struct file *filp, struct vm_area_struct *vma); 379463873d5SEric Anholt int vc4_prime_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma); 380463873d5SEric Anholt void *vc4_prime_vmap(struct drm_gem_object *obj); 381c826a6e1SEric Anholt void vc4_bo_cache_init(struct drm_device *dev); 382c826a6e1SEric Anholt void vc4_bo_cache_destroy(struct drm_device *dev); 383c826a6e1SEric Anholt int vc4_bo_stats_debugfs(struct seq_file *m, void *arg); 384c8b75bcaSEric Anholt 385c8b75bcaSEric Anholt /* vc4_crtc.c */ 386c8b75bcaSEric Anholt extern struct platform_driver vc4_crtc_driver; 3871f43710aSDave Airlie int vc4_enable_vblank(struct drm_device *dev, unsigned int crtc_id); 3881f43710aSDave Airlie void vc4_disable_vblank(struct drm_device *dev, unsigned int crtc_id); 389c8b75bcaSEric Anholt void vc4_cancel_page_flip(struct drm_crtc *crtc, struct drm_file *file); 390c8b75bcaSEric Anholt int vc4_crtc_debugfs_regs(struct seq_file *m, void *arg); 391c8b75bcaSEric Anholt 392c8b75bcaSEric Anholt /* vc4_debugfs.c */ 393c8b75bcaSEric Anholt int vc4_debugfs_init(struct drm_minor *minor); 394c8b75bcaSEric Anholt void vc4_debugfs_cleanup(struct drm_minor *minor); 395c8b75bcaSEric Anholt 396c8b75bcaSEric Anholt /* vc4_drv.c */ 397c8b75bcaSEric Anholt void __iomem *vc4_ioremap_regs(struct platform_device *dev, int index); 398c8b75bcaSEric Anholt 399d5b1a78aSEric Anholt /* vc4_gem.c */ 400d5b1a78aSEric Anholt void vc4_gem_init(struct drm_device *dev); 401d5b1a78aSEric Anholt void vc4_gem_destroy(struct drm_device *dev); 402d5b1a78aSEric Anholt int vc4_submit_cl_ioctl(struct drm_device *dev, void *data, 403d5b1a78aSEric Anholt struct drm_file *file_priv); 404d5b1a78aSEric Anholt int vc4_wait_seqno_ioctl(struct drm_device *dev, void *data, 405d5b1a78aSEric Anholt struct drm_file *file_priv); 406d5b1a78aSEric Anholt int vc4_wait_bo_ioctl(struct drm_device *dev, void *data, 407d5b1a78aSEric Anholt struct drm_file *file_priv); 408d5b1a78aSEric Anholt void vc4_submit_next_job(struct drm_device *dev); 409d5b1a78aSEric Anholt int vc4_wait_for_seqno(struct drm_device *dev, uint64_t seqno, 410d5b1a78aSEric Anholt uint64_t timeout_ns, bool interruptible); 411d5b1a78aSEric Anholt void vc4_job_handle_completed(struct vc4_dev *vc4); 412b501baccSEric Anholt int vc4_queue_seqno_cb(struct drm_device *dev, 413b501baccSEric Anholt struct vc4_seqno_cb *cb, uint64_t seqno, 414b501baccSEric Anholt void (*func)(struct vc4_seqno_cb *cb)); 415d5b1a78aSEric Anholt 416c8b75bcaSEric Anholt /* vc4_hdmi.c */ 417c8b75bcaSEric Anholt extern struct platform_driver vc4_hdmi_driver; 418c8b75bcaSEric Anholt int vc4_hdmi_debugfs_regs(struct seq_file *m, void *unused); 419c8b75bcaSEric Anholt 420d5b1a78aSEric Anholt /* vc4_irq.c */ 421d5b1a78aSEric Anholt irqreturn_t vc4_irq(int irq, void *arg); 422d5b1a78aSEric Anholt void vc4_irq_preinstall(struct drm_device *dev); 423d5b1a78aSEric Anholt int vc4_irq_postinstall(struct drm_device *dev); 424d5b1a78aSEric Anholt void vc4_irq_uninstall(struct drm_device *dev); 425d5b1a78aSEric Anholt void vc4_irq_reset(struct drm_device *dev); 426d5b1a78aSEric Anholt 427c8b75bcaSEric Anholt /* vc4_hvs.c */ 428c8b75bcaSEric Anholt extern struct platform_driver vc4_hvs_driver; 429c8b75bcaSEric Anholt void vc4_hvs_dump_state(struct drm_device *dev); 430c8b75bcaSEric Anholt int vc4_hvs_debugfs_regs(struct seq_file *m, void *unused); 431c8b75bcaSEric Anholt 432c8b75bcaSEric Anholt /* vc4_kms.c */ 433c8b75bcaSEric Anholt int vc4_kms_load(struct drm_device *dev); 434c8b75bcaSEric Anholt 435c8b75bcaSEric Anholt /* vc4_plane.c */ 436c8b75bcaSEric Anholt struct drm_plane *vc4_plane_init(struct drm_device *dev, 437c8b75bcaSEric Anholt enum drm_plane_type type); 438c8b75bcaSEric Anholt u32 vc4_plane_write_dlist(struct drm_plane *plane, u32 __iomem *dlist); 439c8b75bcaSEric Anholt u32 vc4_plane_dlist_size(struct drm_plane_state *state); 440b501baccSEric Anholt void vc4_plane_async_set_fb(struct drm_plane *plane, 441b501baccSEric Anholt struct drm_framebuffer *fb); 442463873d5SEric Anholt 443d3f5168aSEric Anholt /* vc4_v3d.c */ 444d3f5168aSEric Anholt extern struct platform_driver vc4_v3d_driver; 445d3f5168aSEric Anholt int vc4_v3d_debugfs_ident(struct seq_file *m, void *unused); 446d3f5168aSEric Anholt int vc4_v3d_debugfs_regs(struct seq_file *m, void *unused); 447d5b1a78aSEric Anholt 448d5b1a78aSEric Anholt /* vc4_validate.c */ 449d5b1a78aSEric Anholt int 450d5b1a78aSEric Anholt vc4_validate_bin_cl(struct drm_device *dev, 451d5b1a78aSEric Anholt void *validated, 452d5b1a78aSEric Anholt void *unvalidated, 453d5b1a78aSEric Anholt struct vc4_exec_info *exec); 454d5b1a78aSEric Anholt 455d5b1a78aSEric Anholt int 456d5b1a78aSEric Anholt vc4_validate_shader_recs(struct drm_device *dev, struct vc4_exec_info *exec); 457d5b1a78aSEric Anholt 458d5b1a78aSEric Anholt struct drm_gem_cma_object *vc4_use_bo(struct vc4_exec_info *exec, 459d5b1a78aSEric Anholt uint32_t hindex); 460d5b1a78aSEric Anholt 461d5b1a78aSEric Anholt int vc4_get_rcl(struct drm_device *dev, struct vc4_exec_info *exec); 462d5b1a78aSEric Anholt 463d5b1a78aSEric Anholt bool vc4_check_tex_size(struct vc4_exec_info *exec, 464d5b1a78aSEric Anholt struct drm_gem_cma_object *fbo, 465d5b1a78aSEric Anholt uint32_t offset, uint8_t tiling_format, 466d5b1a78aSEric Anholt uint32_t width, uint32_t height, uint8_t cpp); 467d3f5168aSEric Anholt 468463873d5SEric Anholt /* vc4_validate_shader.c */ 469463873d5SEric Anholt struct vc4_validated_shader_info * 470463873d5SEric Anholt vc4_validate_shader(struct drm_gem_cma_object *shader_obj); 471