1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * (C) COPYRIGHT 2018 ARM Limited. All rights reserved. 4 * Author: James.Qian.Wang <james.qian.wang@arm.com> 5 * 6 */ 7 #include <linux/component.h> 8 #include <linux/interrupt.h> 9 10 #include <drm/drm_atomic.h> 11 #include <drm/drm_atomic_helper.h> 12 #include <drm/drm_drv.h> 13 #include <drm/drm_fb_helper.h> 14 #include <drm/drm_gem_cma_helper.h> 15 #include <drm/drm_gem_framebuffer_helper.h> 16 #include <drm/drm_irq.h> 17 #include <drm/drm_vblank.h> 18 19 #include "komeda_dev.h" 20 #include "komeda_framebuffer.h" 21 #include "komeda_kms.h" 22 23 DEFINE_DRM_GEM_CMA_FOPS(komeda_cma_fops); 24 25 static int komeda_gem_cma_dumb_create(struct drm_file *file, 26 struct drm_device *dev, 27 struct drm_mode_create_dumb *args) 28 { 29 struct komeda_dev *mdev = dev->dev_private; 30 u32 pitch = DIV_ROUND_UP(args->width * args->bpp, 8); 31 32 args->pitch = ALIGN(pitch, mdev->chip.bus_width); 33 34 return drm_gem_cma_dumb_create_internal(file, dev, args); 35 } 36 37 static irqreturn_t komeda_kms_irq_handler(int irq, void *data) 38 { 39 struct drm_device *drm = data; 40 struct komeda_dev *mdev = drm->dev_private; 41 struct komeda_kms_dev *kms = to_kdev(drm); 42 struct komeda_events evts; 43 irqreturn_t status; 44 u32 i; 45 46 /* Call into the CHIP to recognize events */ 47 memset(&evts, 0, sizeof(evts)); 48 status = mdev->funcs->irq_handler(mdev, &evts); 49 50 /* Notify the crtc to handle the events */ 51 for (i = 0; i < kms->n_crtcs; i++) 52 komeda_crtc_handle_event(&kms->crtcs[i], &evts); 53 54 return status; 55 } 56 57 static struct drm_driver komeda_kms_driver = { 58 .driver_features = DRIVER_GEM | DRIVER_MODESET | DRIVER_ATOMIC | 59 DRIVER_PRIME | DRIVER_HAVE_IRQ, 60 .lastclose = drm_fb_helper_lastclose, 61 .gem_free_object_unlocked = drm_gem_cma_free_object, 62 .gem_vm_ops = &drm_gem_cma_vm_ops, 63 .dumb_create = komeda_gem_cma_dumb_create, 64 .prime_handle_to_fd = drm_gem_prime_handle_to_fd, 65 .prime_fd_to_handle = drm_gem_prime_fd_to_handle, 66 .gem_prime_export = drm_gem_prime_export, 67 .gem_prime_import = drm_gem_prime_import, 68 .gem_prime_get_sg_table = drm_gem_cma_prime_get_sg_table, 69 .gem_prime_import_sg_table = drm_gem_cma_prime_import_sg_table, 70 .gem_prime_vmap = drm_gem_cma_prime_vmap, 71 .gem_prime_vunmap = drm_gem_cma_prime_vunmap, 72 .gem_prime_mmap = drm_gem_cma_prime_mmap, 73 .fops = &komeda_cma_fops, 74 .name = "komeda", 75 .desc = "Arm Komeda Display Processor driver", 76 .date = "20181101", 77 .major = 0, 78 .minor = 1, 79 }; 80 81 static void komeda_kms_commit_tail(struct drm_atomic_state *old_state) 82 { 83 struct drm_device *dev = old_state->dev; 84 85 drm_atomic_helper_commit_modeset_disables(dev, old_state); 86 87 drm_atomic_helper_commit_planes(dev, old_state, 0); 88 89 drm_atomic_helper_commit_modeset_enables(dev, old_state); 90 91 drm_atomic_helper_wait_for_flip_done(dev, old_state); 92 93 drm_atomic_helper_commit_hw_done(old_state); 94 95 drm_atomic_helper_cleanup_planes(dev, old_state); 96 } 97 98 static const struct drm_mode_config_helper_funcs komeda_mode_config_helpers = { 99 .atomic_commit_tail = komeda_kms_commit_tail, 100 }; 101 102 static int komeda_plane_state_list_add(struct drm_plane_state *plane_st, 103 struct list_head *zorder_list) 104 { 105 struct komeda_plane_state *new = to_kplane_st(plane_st); 106 struct komeda_plane_state *node, *last; 107 108 last = list_empty(zorder_list) ? 109 NULL : list_last_entry(zorder_list, typeof(*last), zlist_node); 110 111 /* Considering the list sequence is zpos increasing, so if list is empty 112 * or the zpos of new node bigger than the last node in list, no need 113 * loop and just insert the new one to the tail of the list. 114 */ 115 if (!last || (new->base.zpos > last->base.zpos)) { 116 list_add_tail(&new->zlist_node, zorder_list); 117 return 0; 118 } 119 120 /* Build the list by zpos increasing */ 121 list_for_each_entry(node, zorder_list, zlist_node) { 122 if (new->base.zpos < node->base.zpos) { 123 list_add_tail(&new->zlist_node, &node->zlist_node); 124 break; 125 } else if (node->base.zpos == new->base.zpos) { 126 struct drm_plane *a = node->base.plane; 127 struct drm_plane *b = new->base.plane; 128 129 /* Komeda doesn't support setting a same zpos for 130 * different planes. 131 */ 132 DRM_DEBUG_ATOMIC("PLANE: %s and PLANE: %s are configured same zpos: %d.\n", 133 a->name, b->name, node->base.zpos); 134 return -EINVAL; 135 } 136 } 137 138 return 0; 139 } 140 141 static int komeda_crtc_normalize_zpos(struct drm_crtc *crtc, 142 struct drm_crtc_state *crtc_st) 143 { 144 struct drm_atomic_state *state = crtc_st->state; 145 struct komeda_crtc *kcrtc = to_kcrtc(crtc); 146 struct komeda_crtc_state *kcrtc_st = to_kcrtc_st(crtc_st); 147 struct komeda_plane_state *kplane_st; 148 struct drm_plane_state *plane_st; 149 struct drm_framebuffer *fb; 150 struct drm_plane *plane; 151 struct list_head zorder_list; 152 int order = 0, err; 153 154 DRM_DEBUG_ATOMIC("[CRTC:%d:%s] calculating normalized zpos values\n", 155 crtc->base.id, crtc->name); 156 157 INIT_LIST_HEAD(&zorder_list); 158 159 /* This loop also added all effected planes into the new state */ 160 drm_for_each_plane_mask(plane, crtc->dev, crtc_st->plane_mask) { 161 plane_st = drm_atomic_get_plane_state(state, plane); 162 if (IS_ERR(plane_st)) 163 return PTR_ERR(plane_st); 164 165 /* Build a list by zpos increasing */ 166 err = komeda_plane_state_list_add(plane_st, &zorder_list); 167 if (err) 168 return err; 169 } 170 171 kcrtc_st->max_slave_zorder = 0; 172 173 list_for_each_entry(kplane_st, &zorder_list, zlist_node) { 174 plane_st = &kplane_st->base; 175 fb = plane_st->fb; 176 plane = plane_st->plane; 177 178 plane_st->normalized_zpos = order++; 179 /* When layer_split has been enabled, one plane will be handled 180 * by two separated komeda layers (left/right), which may needs 181 * two zorders. 182 * - zorder: for left_layer for left display part. 183 * - zorder + 1: will be reserved for right layer. 184 */ 185 if (to_kplane_st(plane_st)->layer_split) 186 order++; 187 188 DRM_DEBUG_ATOMIC("[PLANE:%d:%s] zpos:%d, normalized zpos: %d\n", 189 plane->base.id, plane->name, 190 plane_st->zpos, plane_st->normalized_zpos); 191 192 /* calculate max slave zorder */ 193 if (has_bit(drm_plane_index(plane), kcrtc->slave_planes)) 194 kcrtc_st->max_slave_zorder = 195 max(plane_st->normalized_zpos, 196 kcrtc_st->max_slave_zorder); 197 } 198 199 crtc_st->zpos_changed = true; 200 201 return 0; 202 } 203 204 static int komeda_kms_check(struct drm_device *dev, 205 struct drm_atomic_state *state) 206 { 207 struct drm_crtc *crtc; 208 struct drm_crtc_state *old_crtc_st, *new_crtc_st; 209 int i, err; 210 211 err = drm_atomic_helper_check_modeset(dev, state); 212 if (err) 213 return err; 214 215 /* Komeda need to re-calculate resource assumption in every commit 216 * so need to add all affected_planes (even unchanged) to 217 * drm_atomic_state. 218 */ 219 for_each_oldnew_crtc_in_state(state, crtc, old_crtc_st, new_crtc_st, i) { 220 err = drm_atomic_add_affected_planes(state, crtc); 221 if (err) 222 return err; 223 224 err = komeda_crtc_normalize_zpos(crtc, new_crtc_st); 225 if (err) 226 return err; 227 } 228 229 err = drm_atomic_helper_check_planes(dev, state); 230 if (err) 231 return err; 232 233 return 0; 234 } 235 236 static const struct drm_mode_config_funcs komeda_mode_config_funcs = { 237 .fb_create = komeda_fb_create, 238 .atomic_check = komeda_kms_check, 239 .atomic_commit = drm_atomic_helper_commit, 240 }; 241 242 static void komeda_kms_mode_config_init(struct komeda_kms_dev *kms, 243 struct komeda_dev *mdev) 244 { 245 struct drm_mode_config *config = &kms->base.mode_config; 246 247 drm_mode_config_init(&kms->base); 248 249 komeda_kms_setup_crtcs(kms, mdev); 250 251 /* Get value from dev */ 252 config->min_width = 0; 253 config->min_height = 0; 254 config->max_width = 4096; 255 config->max_height = 4096; 256 config->allow_fb_modifiers = true; 257 258 config->funcs = &komeda_mode_config_funcs; 259 config->helper_private = &komeda_mode_config_helpers; 260 } 261 262 struct komeda_kms_dev *komeda_kms_attach(struct komeda_dev *mdev) 263 { 264 struct komeda_kms_dev *kms = kzalloc(sizeof(*kms), GFP_KERNEL); 265 struct drm_device *drm; 266 int err; 267 268 if (!kms) 269 return ERR_PTR(-ENOMEM); 270 271 drm = &kms->base; 272 err = drm_dev_init(drm, &komeda_kms_driver, mdev->dev); 273 if (err) 274 goto free_kms; 275 276 drm->dev_private = mdev; 277 278 komeda_kms_mode_config_init(kms, mdev); 279 280 err = komeda_kms_add_private_objs(kms, mdev); 281 if (err) 282 goto cleanup_mode_config; 283 284 err = komeda_kms_add_planes(kms, mdev); 285 if (err) 286 goto cleanup_mode_config; 287 288 err = drm_vblank_init(drm, kms->n_crtcs); 289 if (err) 290 goto cleanup_mode_config; 291 292 err = komeda_kms_add_crtcs(kms, mdev); 293 if (err) 294 goto cleanup_mode_config; 295 296 err = komeda_kms_add_wb_connectors(kms, mdev); 297 if (err) 298 goto cleanup_mode_config; 299 300 err = component_bind_all(mdev->dev, kms); 301 if (err) 302 goto cleanup_mode_config; 303 304 drm_mode_config_reset(drm); 305 306 err = devm_request_irq(drm->dev, mdev->irq, 307 komeda_kms_irq_handler, IRQF_SHARED, 308 drm->driver->name, drm); 309 if (err) 310 goto cleanup_mode_config; 311 312 err = mdev->funcs->enable_irq(mdev); 313 if (err) 314 goto cleanup_mode_config; 315 316 drm->irq_enabled = true; 317 318 err = drm_dev_register(drm, 0); 319 if (err) 320 goto cleanup_mode_config; 321 322 return kms; 323 324 cleanup_mode_config: 325 drm->irq_enabled = false; 326 drm_mode_config_cleanup(drm); 327 komeda_kms_cleanup_private_objs(kms); 328 free_kms: 329 kfree(kms); 330 return ERR_PTR(err); 331 } 332 333 void komeda_kms_detach(struct komeda_kms_dev *kms) 334 { 335 struct drm_device *drm = &kms->base; 336 struct komeda_dev *mdev = drm->dev_private; 337 338 drm->irq_enabled = false; 339 mdev->funcs->disable_irq(mdev); 340 drm_dev_unregister(drm); 341 component_unbind_all(mdev->dev, drm); 342 komeda_kms_cleanup_private_objs(kms); 343 drm_mode_config_cleanup(drm); 344 drm->dev_private = NULL; 345 drm_dev_put(drm); 346 } 347