1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /* 3 * linux/drivers/video/mmp/hw/mmp_ctrl.c 4 * Marvell MMP series Display Controller support 5 * 6 * Copyright (C) 2012 Marvell Technology Group Ltd. 7 * Authors: Guoqing Li <ligq@marvell.com> 8 * Lisa Du <cldu@marvell.com> 9 * Zhou Zhu <zzhu3@marvell.com> 10 */ 11 #include <linux/module.h> 12 #include <linux/moduleparam.h> 13 #include <linux/kernel.h> 14 #include <linux/errno.h> 15 #include <linux/string.h> 16 #include <linux/interrupt.h> 17 #include <linux/slab.h> 18 #include <linux/delay.h> 19 #include <linux/platform_device.h> 20 #include <linux/dma-mapping.h> 21 #include <linux/clk.h> 22 #include <linux/err.h> 23 #include <linux/vmalloc.h> 24 #include <linux/uaccess.h> 25 #include <linux/kthread.h> 26 #include <linux/io.h> 27 28 #include "mmp_ctrl.h" 29 30 static irqreturn_t ctrl_handle_irq(int irq, void *dev_id) 31 { 32 struct mmphw_ctrl *ctrl = (struct mmphw_ctrl *)dev_id; 33 u32 isr, imask, tmp; 34 35 isr = readl_relaxed(ctrl->reg_base + SPU_IRQ_ISR); 36 imask = readl_relaxed(ctrl->reg_base + SPU_IRQ_ENA); 37 38 do { 39 /* clear clock only */ 40 tmp = readl_relaxed(ctrl->reg_base + SPU_IRQ_ISR); 41 if (tmp & isr) 42 writel_relaxed(~isr, ctrl->reg_base + SPU_IRQ_ISR); 43 } while ((isr = readl_relaxed(ctrl->reg_base + SPU_IRQ_ISR)) & imask); 44 45 return IRQ_HANDLED; 46 } 47 48 static u32 fmt_to_reg(struct mmp_overlay *overlay, int pix_fmt) 49 { 50 u32 rbswap = 0, uvswap = 0, yuvswap = 0, 51 csc_en = 0, val = 0, 52 vid = overlay_is_vid(overlay); 53 54 switch (pix_fmt) { 55 case PIXFMT_RGB565: 56 case PIXFMT_RGB1555: 57 case PIXFMT_RGB888PACK: 58 case PIXFMT_RGB888UNPACK: 59 case PIXFMT_RGBA888: 60 rbswap = 1; 61 break; 62 case PIXFMT_VYUY: 63 case PIXFMT_YVU422P: 64 case PIXFMT_YVU420P: 65 uvswap = 1; 66 break; 67 case PIXFMT_YUYV: 68 yuvswap = 1; 69 break; 70 default: 71 break; 72 } 73 74 switch (pix_fmt) { 75 case PIXFMT_RGB565: 76 case PIXFMT_BGR565: 77 break; 78 case PIXFMT_RGB1555: 79 case PIXFMT_BGR1555: 80 val = 0x1; 81 break; 82 case PIXFMT_RGB888PACK: 83 case PIXFMT_BGR888PACK: 84 val = 0x2; 85 break; 86 case PIXFMT_RGB888UNPACK: 87 case PIXFMT_BGR888UNPACK: 88 val = 0x3; 89 break; 90 case PIXFMT_RGBA888: 91 case PIXFMT_BGRA888: 92 val = 0x4; 93 break; 94 case PIXFMT_UYVY: 95 case PIXFMT_VYUY: 96 case PIXFMT_YUYV: 97 val = 0x5; 98 csc_en = 1; 99 break; 100 case PIXFMT_YUV422P: 101 case PIXFMT_YVU422P: 102 val = 0x6; 103 csc_en = 1; 104 break; 105 case PIXFMT_YUV420P: 106 case PIXFMT_YVU420P: 107 val = 0x7; 108 csc_en = 1; 109 break; 110 default: 111 break; 112 } 113 114 return (dma_palette(0) | dma_fmt(vid, val) | 115 dma_swaprb(vid, rbswap) | dma_swapuv(vid, uvswap) | 116 dma_swapyuv(vid, yuvswap) | dma_csc(vid, csc_en)); 117 } 118 119 static void dmafetch_set_fmt(struct mmp_overlay *overlay) 120 { 121 u32 tmp; 122 struct mmp_path *path = overlay->path; 123 tmp = readl_relaxed(ctrl_regs(path) + dma_ctrl(0, path->id)); 124 tmp &= ~dma_mask(overlay_is_vid(overlay)); 125 tmp |= fmt_to_reg(overlay, overlay->win.pix_fmt); 126 writel_relaxed(tmp, ctrl_regs(path) + dma_ctrl(0, path->id)); 127 } 128 129 static void overlay_set_win(struct mmp_overlay *overlay, struct mmp_win *win) 130 { 131 struct lcd_regs *regs = path_regs(overlay->path); 132 133 /* assert win supported */ 134 memcpy(&overlay->win, win, sizeof(struct mmp_win)); 135 136 mutex_lock(&overlay->access_ok); 137 138 if (overlay_is_vid(overlay)) { 139 writel_relaxed(win->pitch[0], 140 (void __iomem *)®s->v_pitch_yc); 141 writel_relaxed(win->pitch[2] << 16 | win->pitch[1], 142 (void __iomem *)®s->v_pitch_uv); 143 144 writel_relaxed((win->ysrc << 16) | win->xsrc, 145 (void __iomem *)®s->v_size); 146 writel_relaxed((win->ydst << 16) | win->xdst, 147 (void __iomem *)®s->v_size_z); 148 writel_relaxed(win->ypos << 16 | win->xpos, 149 (void __iomem *)®s->v_start); 150 } else { 151 writel_relaxed(win->pitch[0], (void __iomem *)®s->g_pitch); 152 153 writel_relaxed((win->ysrc << 16) | win->xsrc, 154 (void __iomem *)®s->g_size); 155 writel_relaxed((win->ydst << 16) | win->xdst, 156 (void __iomem *)®s->g_size_z); 157 writel_relaxed(win->ypos << 16 | win->xpos, 158 (void __iomem *)®s->g_start); 159 } 160 161 dmafetch_set_fmt(overlay); 162 mutex_unlock(&overlay->access_ok); 163 } 164 165 static void dmafetch_onoff(struct mmp_overlay *overlay, int on) 166 { 167 u32 mask = overlay_is_vid(overlay) ? CFG_DMA_ENA_MASK : 168 CFG_GRA_ENA_MASK; 169 u32 enable = overlay_is_vid(overlay) ? CFG_DMA_ENA(1) : CFG_GRA_ENA(1); 170 u32 tmp; 171 struct mmp_path *path = overlay->path; 172 173 mutex_lock(&overlay->access_ok); 174 tmp = readl_relaxed(ctrl_regs(path) + dma_ctrl(0, path->id)); 175 tmp &= ~mask; 176 tmp |= (on ? enable : 0); 177 writel(tmp, ctrl_regs(path) + dma_ctrl(0, path->id)); 178 mutex_unlock(&overlay->access_ok); 179 } 180 181 static void path_enabledisable(struct mmp_path *path, int on) 182 { 183 u32 tmp; 184 mutex_lock(&path->access_ok); 185 tmp = readl_relaxed(ctrl_regs(path) + LCD_SCLK(path)); 186 if (on) 187 tmp &= ~SCLK_DISABLE; 188 else 189 tmp |= SCLK_DISABLE; 190 writel_relaxed(tmp, ctrl_regs(path) + LCD_SCLK(path)); 191 mutex_unlock(&path->access_ok); 192 } 193 194 static void path_onoff(struct mmp_path *path, int on) 195 { 196 if (path->status == on) { 197 dev_info(path->dev, "path %s is already %s\n", 198 path->name, stat_name(path->status)); 199 return; 200 } 201 202 if (on) { 203 path_enabledisable(path, 1); 204 205 if (path->panel && path->panel->set_onoff) 206 path->panel->set_onoff(path->panel, 1); 207 } else { 208 if (path->panel && path->panel->set_onoff) 209 path->panel->set_onoff(path->panel, 0); 210 211 path_enabledisable(path, 0); 212 } 213 path->status = on; 214 } 215 216 static void overlay_set_onoff(struct mmp_overlay *overlay, int on) 217 { 218 if (overlay->status == on) { 219 dev_info(overlay_to_ctrl(overlay)->dev, "overlay %s is already %s\n", 220 overlay->path->name, stat_name(overlay->status)); 221 return; 222 } 223 overlay->status = on; 224 dmafetch_onoff(overlay, on); 225 if (overlay->path->ops.check_status(overlay->path) 226 != overlay->path->status) 227 path_onoff(overlay->path, on); 228 } 229 230 static void overlay_set_fetch(struct mmp_overlay *overlay, int fetch_id) 231 { 232 overlay->dmafetch_id = fetch_id; 233 } 234 235 static int overlay_set_addr(struct mmp_overlay *overlay, struct mmp_addr *addr) 236 { 237 struct lcd_regs *regs = path_regs(overlay->path); 238 239 /* FIXME: assert addr supported */ 240 memcpy(&overlay->addr, addr, sizeof(struct mmp_addr)); 241 242 if (overlay_is_vid(overlay)) { 243 writel_relaxed(addr->phys[0], (void __iomem *)®s->v_y0); 244 writel_relaxed(addr->phys[1], (void __iomem *)®s->v_u0); 245 writel_relaxed(addr->phys[2], (void __iomem *)®s->v_v0); 246 } else 247 writel_relaxed(addr->phys[0], (void __iomem *)®s->g_0); 248 249 return overlay->addr.phys[0]; 250 } 251 252 static void path_set_mode(struct mmp_path *path, struct mmp_mode *mode) 253 { 254 struct lcd_regs *regs = path_regs(path); 255 u32 total_x, total_y, vsync_ctrl, tmp, sclk_src, sclk_div, 256 link_config = path_to_path_plat(path)->link_config, 257 dsi_rbswap = path_to_path_plat(path)->link_config; 258 259 /* FIXME: assert videomode supported */ 260 memcpy(&path->mode, mode, sizeof(struct mmp_mode)); 261 262 mutex_lock(&path->access_ok); 263 264 /* polarity of timing signals */ 265 tmp = readl_relaxed(ctrl_regs(path) + intf_ctrl(path->id)) & 0x1; 266 tmp |= mode->vsync_invert ? 0 : 0x8; 267 tmp |= mode->hsync_invert ? 0 : 0x4; 268 tmp |= link_config & CFG_DUMBMODE_MASK; 269 tmp |= CFG_DUMB_ENA(1); 270 writel_relaxed(tmp, ctrl_regs(path) + intf_ctrl(path->id)); 271 272 /* interface rb_swap setting */ 273 tmp = readl_relaxed(ctrl_regs(path) + intf_rbswap_ctrl(path->id)) & 274 (~(CFG_INTFRBSWAP_MASK)); 275 tmp |= dsi_rbswap & CFG_INTFRBSWAP_MASK; 276 writel_relaxed(tmp, ctrl_regs(path) + intf_rbswap_ctrl(path->id)); 277 278 writel_relaxed((mode->yres << 16) | mode->xres, 279 (void __iomem *)®s->screen_active); 280 writel_relaxed((mode->left_margin << 16) | mode->right_margin, 281 (void __iomem *)®s->screen_h_porch); 282 writel_relaxed((mode->upper_margin << 16) | mode->lower_margin, 283 (void __iomem *)®s->screen_v_porch); 284 total_x = mode->xres + mode->left_margin + mode->right_margin + 285 mode->hsync_len; 286 total_y = mode->yres + mode->upper_margin + mode->lower_margin + 287 mode->vsync_len; 288 writel_relaxed((total_y << 16) | total_x, 289 (void __iomem *)®s->screen_size); 290 291 /* vsync ctrl */ 292 if (path->output_type == PATH_OUT_DSI) 293 vsync_ctrl = 0x01330133; 294 else 295 vsync_ctrl = ((mode->xres + mode->right_margin) << 16) 296 | (mode->xres + mode->right_margin); 297 writel_relaxed(vsync_ctrl, (void __iomem *)®s->vsync_ctrl); 298 299 /* set pixclock div */ 300 sclk_src = clk_get_rate(path_to_ctrl(path)->clk); 301 sclk_div = sclk_src / mode->pixclock_freq; 302 if (sclk_div * mode->pixclock_freq < sclk_src) 303 sclk_div++; 304 305 dev_info(path->dev, "%s sclk_src %d sclk_div 0x%x pclk %d\n", 306 __func__, sclk_src, sclk_div, mode->pixclock_freq); 307 308 tmp = readl_relaxed(ctrl_regs(path) + LCD_SCLK(path)); 309 tmp &= ~CLK_INT_DIV_MASK; 310 tmp |= sclk_div; 311 writel_relaxed(tmp, ctrl_regs(path) + LCD_SCLK(path)); 312 313 mutex_unlock(&path->access_ok); 314 } 315 316 static const struct mmp_overlay_ops mmphw_overlay_ops = { 317 .set_fetch = overlay_set_fetch, 318 .set_onoff = overlay_set_onoff, 319 .set_win = overlay_set_win, 320 .set_addr = overlay_set_addr, 321 }; 322 323 static void ctrl_set_default(struct mmphw_ctrl *ctrl) 324 { 325 u32 tmp, irq_mask; 326 327 /* 328 * LCD Global control(LCD_TOP_CTRL) should be configed before 329 * any other LCD registers read/write, or there maybe issues. 330 */ 331 tmp = readl_relaxed(ctrl->reg_base + LCD_TOP_CTRL); 332 tmp |= 0xfff0; 333 writel_relaxed(tmp, ctrl->reg_base + LCD_TOP_CTRL); 334 335 336 /* disable all interrupts */ 337 irq_mask = path_imasks(0) | err_imask(0) | 338 path_imasks(1) | err_imask(1); 339 tmp = readl_relaxed(ctrl->reg_base + SPU_IRQ_ENA); 340 tmp &= ~irq_mask; 341 tmp |= irq_mask; 342 writel_relaxed(tmp, ctrl->reg_base + SPU_IRQ_ENA); 343 } 344 345 static void path_set_default(struct mmp_path *path) 346 { 347 struct lcd_regs *regs = path_regs(path); 348 u32 dma_ctrl1, mask, tmp, path_config; 349 350 path_config = path_to_path_plat(path)->path_config; 351 352 /* Configure IOPAD: should be parallel only */ 353 if (PATH_OUT_PARALLEL == path->output_type) { 354 mask = CFG_IOPADMODE_MASK | CFG_BURST_MASK | CFG_BOUNDARY_MASK; 355 tmp = readl_relaxed(ctrl_regs(path) + SPU_IOPAD_CONTROL); 356 tmp &= ~mask; 357 tmp |= path_config; 358 writel_relaxed(tmp, ctrl_regs(path) + SPU_IOPAD_CONTROL); 359 } 360 361 /* Select path clock source */ 362 tmp = readl_relaxed(ctrl_regs(path) + LCD_SCLK(path)); 363 tmp &= ~SCLK_SRC_SEL_MASK; 364 tmp |= path_config; 365 writel_relaxed(tmp, ctrl_regs(path) + LCD_SCLK(path)); 366 367 /* 368 * Configure default bits: vsync triggers DMA, 369 * power save enable, configure alpha registers to 370 * display 100% graphics, and set pixel command. 371 */ 372 dma_ctrl1 = 0x2032ff81; 373 374 dma_ctrl1 |= CFG_VSYNC_INV_MASK; 375 writel_relaxed(dma_ctrl1, ctrl_regs(path) + dma_ctrl(1, path->id)); 376 377 /* Configure default register values */ 378 writel_relaxed(0x00000000, (void __iomem *)®s->blank_color); 379 writel_relaxed(0x00000000, (void __iomem *)®s->g_1); 380 writel_relaxed(0x00000000, (void __iomem *)®s->g_start); 381 382 /* 383 * 1.enable multiple burst request in DMA AXI 384 * bus arbiter for faster read if not tv path; 385 * 2.enable horizontal smooth filter; 386 */ 387 mask = CFG_GRA_HSMOOTH_MASK | CFG_DMA_HSMOOTH_MASK | CFG_ARBFAST_ENA(1); 388 tmp = readl_relaxed(ctrl_regs(path) + dma_ctrl(0, path->id)); 389 tmp |= mask; 390 if (PATH_TV == path->id) 391 tmp &= ~CFG_ARBFAST_ENA(1); 392 writel_relaxed(tmp, ctrl_regs(path) + dma_ctrl(0, path->id)); 393 } 394 395 static int path_init(struct mmphw_path_plat *path_plat, 396 struct mmp_mach_path_config *config) 397 { 398 struct mmphw_ctrl *ctrl = path_plat->ctrl; 399 struct mmp_path_info *path_info; 400 struct mmp_path *path = NULL; 401 402 dev_info(ctrl->dev, "%s: %s\n", __func__, config->name); 403 404 /* init driver data */ 405 path_info = kzalloc(sizeof(*path_info), GFP_KERNEL); 406 if (!path_info) 407 return 0; 408 409 path_info->name = config->name; 410 path_info->id = path_plat->id; 411 path_info->dev = ctrl->dev; 412 path_info->overlay_num = config->overlay_num; 413 path_info->overlay_ops = &mmphw_overlay_ops; 414 path_info->set_mode = path_set_mode; 415 path_info->plat_data = path_plat; 416 417 /* create/register platform device */ 418 path = mmp_register_path(path_info); 419 if (!path) { 420 kfree(path_info); 421 return 0; 422 } 423 path_plat->path = path; 424 path_plat->path_config = config->path_config; 425 path_plat->link_config = config->link_config; 426 path_plat->dsi_rbswap = config->dsi_rbswap; 427 path_set_default(path); 428 429 kfree(path_info); 430 return 1; 431 } 432 433 static void path_deinit(struct mmphw_path_plat *path_plat) 434 { 435 if (!path_plat) 436 return; 437 438 mmp_unregister_path(path_plat->path); 439 } 440 441 static int mmphw_probe(struct platform_device *pdev) 442 { 443 struct mmp_mach_plat_info *mi; 444 struct resource *res; 445 int ret, i, irq; 446 struct mmphw_path_plat *path_plat; 447 struct mmphw_ctrl *ctrl = NULL; 448 449 /* get resources from platform data */ 450 res = platform_get_resource(pdev, IORESOURCE_MEM, 0); 451 if (res == NULL) { 452 dev_err(&pdev->dev, "%s: no IO memory defined\n", __func__); 453 ret = -ENOENT; 454 goto failed; 455 } 456 457 irq = platform_get_irq(pdev, 0); 458 if (irq < 0) { 459 ret = -ENOENT; 460 goto failed; 461 } 462 463 /* get configs from platform data */ 464 mi = pdev->dev.platform_data; 465 if (mi == NULL || !mi->path_num || !mi->paths) { 466 dev_err(&pdev->dev, "%s: no platform data defined\n", __func__); 467 ret = -EINVAL; 468 goto failed; 469 } 470 471 /* allocate */ 472 ctrl = devm_kzalloc(&pdev->dev, 473 struct_size(ctrl, path_plats, mi->path_num), 474 GFP_KERNEL); 475 if (!ctrl) { 476 ret = -ENOMEM; 477 goto failed; 478 } 479 480 ctrl->name = mi->name; 481 ctrl->path_num = mi->path_num; 482 ctrl->dev = &pdev->dev; 483 ctrl->irq = irq; 484 platform_set_drvdata(pdev, ctrl); 485 mutex_init(&ctrl->access_ok); 486 487 /* map registers.*/ 488 if (!devm_request_mem_region(ctrl->dev, res->start, 489 resource_size(res), ctrl->name)) { 490 dev_err(ctrl->dev, 491 "can't request region for resource %pR\n", res); 492 ret = -EINVAL; 493 goto failed; 494 } 495 496 ctrl->reg_base = devm_ioremap(ctrl->dev, 497 res->start, resource_size(res)); 498 if (ctrl->reg_base == NULL) { 499 dev_err(ctrl->dev, "%s: res %pR map failed\n", __func__, res); 500 ret = -ENOMEM; 501 goto failed; 502 } 503 504 /* request irq */ 505 ret = devm_request_irq(ctrl->dev, ctrl->irq, ctrl_handle_irq, 506 IRQF_SHARED, "lcd_controller", ctrl); 507 if (ret < 0) { 508 dev_err(ctrl->dev, "%s unable to request IRQ %d\n", 509 __func__, ctrl->irq); 510 ret = -ENXIO; 511 goto failed; 512 } 513 514 /* get clock */ 515 ctrl->clk = devm_clk_get(ctrl->dev, mi->clk_name); 516 if (IS_ERR(ctrl->clk)) { 517 ret = PTR_ERR(ctrl->clk); 518 dev_err_probe(ctrl->dev, ret, 519 "unable to get clk %s\n", mi->clk_name); 520 goto failed; 521 } 522 ret = clk_prepare_enable(ctrl->clk); 523 if (ret) 524 goto failed; 525 526 /* init global regs */ 527 ctrl_set_default(ctrl); 528 529 /* init pathes from machine info and register them */ 530 for (i = 0; i < ctrl->path_num; i++) { 531 /* get from config and machine info */ 532 path_plat = &ctrl->path_plats[i]; 533 path_plat->id = i; 534 path_plat->ctrl = ctrl; 535 536 /* path init */ 537 if (!path_init(path_plat, &mi->paths[i])) { 538 ret = -EINVAL; 539 goto failed_path_init; 540 } 541 } 542 543 #ifdef CONFIG_MMP_DISP_SPI 544 ret = lcd_spi_register(ctrl); 545 if (ret < 0) 546 goto failed_path_init; 547 #endif 548 549 dev_info(ctrl->dev, "device init done\n"); 550 551 return 0; 552 553 failed_path_init: 554 for (i = 0; i < ctrl->path_num; i++) { 555 path_plat = &ctrl->path_plats[i]; 556 path_deinit(path_plat); 557 } 558 559 clk_disable_unprepare(ctrl->clk); 560 failed: 561 dev_err(&pdev->dev, "device init failed\n"); 562 563 return ret; 564 } 565 566 static struct platform_driver mmphw_driver = { 567 .driver = { 568 .name = "mmp-disp", 569 }, 570 .probe = mmphw_probe, 571 }; 572 573 static int mmphw_init(void) 574 { 575 return platform_driver_register(&mmphw_driver); 576 } 577 module_init(mmphw_init); 578 579 MODULE_AUTHOR("Li Guoqing<ligq@marvell.com>"); 580 MODULE_DESCRIPTION("Framebuffer driver for mmp"); 581 MODULE_LICENSE("GPL"); 582