1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Copyright (C) 2020 Unisoc Inc. 4 */ 5 6 #include <linux/component.h> 7 #include <linux/module.h> 8 #include <linux/of.h> 9 #include <linux/platform_device.h> 10 #include <video/mipi_display.h> 11 12 #include <drm/drm_atomic_helper.h> 13 #include <drm/drm_bridge.h> 14 #include <drm/drm_of.h> 15 #include <drm/drm_probe_helper.h> 16 17 #include "sprd_drm.h" 18 #include "sprd_dpu.h" 19 #include "sprd_dsi.h" 20 21 #define SOFT_RESET 0x04 22 #define MASK_PROTOCOL_INT 0x0C 23 #define MASK_INTERNAL_INT 0x14 24 #define DSI_MODE_CFG 0x18 25 26 #define VIRTUAL_CHANNEL_ID 0x1C 27 #define GEN_RX_VCID GENMASK(1, 0) 28 #define VIDEO_PKT_VCID GENMASK(3, 2) 29 30 #define DPI_VIDEO_FORMAT 0x20 31 #define DPI_VIDEO_MODE_FORMAT GENMASK(5, 0) 32 #define LOOSELY18_EN BIT(6) 33 34 #define VIDEO_PKT_CONFIG 0x24 35 #define VIDEO_PKT_SIZE GENMASK(15, 0) 36 #define VIDEO_LINE_CHUNK_NUM GENMASK(31, 16) 37 38 #define VIDEO_LINE_HBLK_TIME 0x28 39 #define VIDEO_LINE_HBP_TIME GENMASK(15, 0) 40 #define VIDEO_LINE_HSA_TIME GENMASK(31, 16) 41 42 #define VIDEO_LINE_TIME 0x2C 43 44 #define VIDEO_VBLK_LINES 0x30 45 #define VFP_LINES GENMASK(9, 0) 46 #define VBP_LINES GENMASK(19, 10) 47 #define VSA_LINES GENMASK(29, 20) 48 49 #define VIDEO_VACTIVE_LINES 0x34 50 51 #define VID_MODE_CFG 0x38 52 #define VID_MODE_TYPE GENMASK(1, 0) 53 #define LP_VSA_EN BIT(8) 54 #define LP_VBP_EN BIT(9) 55 #define LP_VFP_EN BIT(10) 56 #define LP_VACT_EN BIT(11) 57 #define LP_HBP_EN BIT(12) 58 #define LP_HFP_EN BIT(13) 59 #define FRAME_BTA_ACK_EN BIT(14) 60 61 #define TIMEOUT_CNT_CLK_CONFIG 0x40 62 #define HTX_TO_CONFIG 0x44 63 #define LRX_H_TO_CONFIG 0x48 64 65 #define TX_ESC_CLK_CONFIG 0x5C 66 67 #define CMD_MODE_CFG 0x68 68 #define TEAR_FX_EN BIT(0) 69 70 #define GEN_HDR 0x6C 71 #define GEN_DT GENMASK(5, 0) 72 #define GEN_VC GENMASK(7, 6) 73 74 #define GEN_PLD_DATA 0x70 75 76 #define PHY_CLK_LANE_LP_CTRL 0x74 77 #define PHY_CLKLANE_TX_REQ_HS BIT(0) 78 #define AUTO_CLKLANE_CTRL_EN BIT(1) 79 80 #define PHY_INTERFACE_CTRL 0x78 81 #define RF_PHY_SHUTDOWN BIT(0) 82 #define RF_PHY_RESET_N BIT(1) 83 #define RF_PHY_CLK_EN BIT(2) 84 85 #define CMD_MODE_STATUS 0x98 86 #define GEN_CMD_RDATA_FIFO_EMPTY BIT(1) 87 #define GEN_CMD_WDATA_FIFO_EMPTY BIT(3) 88 #define GEN_CMD_CMD_FIFO_EMPTY BIT(5) 89 #define GEN_CMD_RDCMD_DONE BIT(7) 90 91 #define PHY_STATUS 0x9C 92 #define PHY_LOCK BIT(1) 93 94 #define PHY_MIN_STOP_TIME 0xA0 95 #define PHY_LANE_NUM_CONFIG 0xA4 96 97 #define PHY_CLKLANE_TIME_CONFIG 0xA8 98 #define PHY_CLKLANE_LP_TO_HS_TIME GENMASK(15, 0) 99 #define PHY_CLKLANE_HS_TO_LP_TIME GENMASK(31, 16) 100 101 #define PHY_DATALANE_TIME_CONFIG 0xAC 102 #define PHY_DATALANE_LP_TO_HS_TIME GENMASK(15, 0) 103 #define PHY_DATALANE_HS_TO_LP_TIME GENMASK(31, 16) 104 105 #define MAX_READ_TIME 0xB0 106 107 #define RX_PKT_CHECK_CONFIG 0xB4 108 #define RX_PKT_ECC_EN BIT(0) 109 #define RX_PKT_CRC_EN BIT(1) 110 111 #define TA_EN 0xB8 112 113 #define EOTP_EN 0xBC 114 #define TX_EOTP_EN BIT(0) 115 #define RX_EOTP_EN BIT(1) 116 117 #define VIDEO_NULLPKT_SIZE 0xC0 118 #define DCS_WM_PKT_SIZE 0xC4 119 120 #define VIDEO_SIG_DELAY_CONFIG 0xD0 121 #define VIDEO_SIG_DELAY GENMASK(23, 0) 122 123 #define PHY_TST_CTRL0 0xF0 124 #define PHY_TESTCLR BIT(0) 125 #define PHY_TESTCLK BIT(1) 126 127 #define PHY_TST_CTRL1 0xF4 128 #define PHY_TESTDIN GENMASK(7, 0) 129 #define PHY_TESTDOUT GENMASK(15, 8) 130 #define PHY_TESTEN BIT(16) 131 132 #define host_to_dsi(host) \ 133 container_of(host, struct sprd_dsi, host) 134 135 static inline u32 136 dsi_reg_rd(struct dsi_context *ctx, u32 offset, u32 mask, 137 u32 shift) 138 { 139 return (readl(ctx->base + offset) & mask) >> shift; 140 } 141 142 static inline void 143 dsi_reg_wr(struct dsi_context *ctx, u32 offset, u32 mask, 144 u32 shift, u32 val) 145 { 146 u32 ret; 147 148 ret = readl(ctx->base + offset); 149 ret &= ~mask; 150 ret |= (val << shift) & mask; 151 writel(ret, ctx->base + offset); 152 } 153 154 static inline void 155 dsi_reg_up(struct dsi_context *ctx, u32 offset, u32 mask, 156 u32 val) 157 { 158 u32 ret = readl(ctx->base + offset); 159 160 writel((ret & ~mask) | (val & mask), ctx->base + offset); 161 } 162 163 static int regmap_tst_io_write(void *context, u32 reg, u32 val) 164 { 165 struct sprd_dsi *dsi = context; 166 struct dsi_context *ctx = &dsi->ctx; 167 168 if (val > 0xff || reg > 0xff) 169 return -EINVAL; 170 171 drm_dbg(dsi->drm, "reg = 0x%02x, val = 0x%02x\n", reg, val); 172 173 dsi_reg_up(ctx, PHY_TST_CTRL1, PHY_TESTEN, PHY_TESTEN); 174 dsi_reg_wr(ctx, PHY_TST_CTRL1, PHY_TESTDIN, 0, reg); 175 dsi_reg_up(ctx, PHY_TST_CTRL0, PHY_TESTCLK, PHY_TESTCLK); 176 dsi_reg_up(ctx, PHY_TST_CTRL0, PHY_TESTCLK, 0); 177 dsi_reg_up(ctx, PHY_TST_CTRL1, PHY_TESTEN, 0); 178 dsi_reg_wr(ctx, PHY_TST_CTRL1, PHY_TESTDIN, 0, val); 179 dsi_reg_up(ctx, PHY_TST_CTRL0, PHY_TESTCLK, PHY_TESTCLK); 180 dsi_reg_up(ctx, PHY_TST_CTRL0, PHY_TESTCLK, 0); 181 182 return 0; 183 } 184 185 static int regmap_tst_io_read(void *context, u32 reg, u32 *val) 186 { 187 struct sprd_dsi *dsi = context; 188 struct dsi_context *ctx = &dsi->ctx; 189 int ret; 190 191 if (reg > 0xff) 192 return -EINVAL; 193 194 dsi_reg_up(ctx, PHY_TST_CTRL1, PHY_TESTEN, PHY_TESTEN); 195 dsi_reg_wr(ctx, PHY_TST_CTRL1, PHY_TESTDIN, 0, reg); 196 dsi_reg_up(ctx, PHY_TST_CTRL0, PHY_TESTCLK, PHY_TESTCLK); 197 dsi_reg_up(ctx, PHY_TST_CTRL0, PHY_TESTCLK, 0); 198 dsi_reg_up(ctx, PHY_TST_CTRL1, PHY_TESTEN, 0); 199 200 udelay(1); 201 202 ret = dsi_reg_rd(ctx, PHY_TST_CTRL1, PHY_TESTDOUT, 8); 203 if (ret < 0) 204 return ret; 205 206 *val = ret; 207 208 drm_dbg(dsi->drm, "reg = 0x%02x, val = 0x%02x\n", reg, *val); 209 return 0; 210 } 211 212 static struct regmap_bus regmap_tst_io = { 213 .reg_write = regmap_tst_io_write, 214 .reg_read = regmap_tst_io_read, 215 }; 216 217 static const struct regmap_config byte_config = { 218 .reg_bits = 8, 219 .val_bits = 8, 220 }; 221 222 static int dphy_wait_pll_locked(struct dsi_context *ctx) 223 { 224 struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx); 225 int i; 226 227 for (i = 0; i < 50000; i++) { 228 if (dsi_reg_rd(ctx, PHY_STATUS, PHY_LOCK, 1)) 229 return 0; 230 udelay(3); 231 } 232 233 drm_err(dsi->drm, "dphy pll can not be locked\n"); 234 return -ETIMEDOUT; 235 } 236 237 static int dsi_wait_tx_payload_fifo_empty(struct dsi_context *ctx) 238 { 239 int i; 240 241 for (i = 0; i < 5000; i++) { 242 if (dsi_reg_rd(ctx, CMD_MODE_STATUS, GEN_CMD_WDATA_FIFO_EMPTY, 3)) 243 return 0; 244 udelay(1); 245 } 246 247 return -ETIMEDOUT; 248 } 249 250 static int dsi_wait_tx_cmd_fifo_empty(struct dsi_context *ctx) 251 { 252 int i; 253 254 for (i = 0; i < 5000; i++) { 255 if (dsi_reg_rd(ctx, CMD_MODE_STATUS, GEN_CMD_CMD_FIFO_EMPTY, 5)) 256 return 0; 257 udelay(1); 258 } 259 260 return -ETIMEDOUT; 261 } 262 263 static int dsi_wait_rd_resp_completed(struct dsi_context *ctx) 264 { 265 int i; 266 267 for (i = 0; i < 10000; i++) { 268 if (dsi_reg_rd(ctx, CMD_MODE_STATUS, GEN_CMD_RDCMD_DONE, 7)) 269 return 0; 270 udelay(10); 271 } 272 273 return -ETIMEDOUT; 274 } 275 276 static u16 calc_bytes_per_pixel_x100(int coding) 277 { 278 u16 bpp_x100; 279 280 switch (coding) { 281 case COLOR_CODE_16BIT_CONFIG1: 282 case COLOR_CODE_16BIT_CONFIG2: 283 case COLOR_CODE_16BIT_CONFIG3: 284 bpp_x100 = 200; 285 break; 286 case COLOR_CODE_18BIT_CONFIG1: 287 case COLOR_CODE_18BIT_CONFIG2: 288 bpp_x100 = 225; 289 break; 290 case COLOR_CODE_24BIT: 291 bpp_x100 = 300; 292 break; 293 case COLOR_CODE_COMPRESSTION: 294 bpp_x100 = 100; 295 break; 296 case COLOR_CODE_20BIT_YCC422_LOOSELY: 297 bpp_x100 = 250; 298 break; 299 case COLOR_CODE_24BIT_YCC422: 300 bpp_x100 = 300; 301 break; 302 case COLOR_CODE_16BIT_YCC422: 303 bpp_x100 = 200; 304 break; 305 case COLOR_CODE_30BIT: 306 bpp_x100 = 375; 307 break; 308 case COLOR_CODE_36BIT: 309 bpp_x100 = 450; 310 break; 311 case COLOR_CODE_12BIT_YCC420: 312 bpp_x100 = 150; 313 break; 314 default: 315 DRM_ERROR("invalid color coding"); 316 bpp_x100 = 0; 317 break; 318 } 319 320 return bpp_x100; 321 } 322 323 static u8 calc_video_size_step(int coding) 324 { 325 u8 video_size_step; 326 327 switch (coding) { 328 case COLOR_CODE_16BIT_CONFIG1: 329 case COLOR_CODE_16BIT_CONFIG2: 330 case COLOR_CODE_16BIT_CONFIG3: 331 case COLOR_CODE_18BIT_CONFIG1: 332 case COLOR_CODE_18BIT_CONFIG2: 333 case COLOR_CODE_24BIT: 334 case COLOR_CODE_COMPRESSTION: 335 return video_size_step = 1; 336 case COLOR_CODE_20BIT_YCC422_LOOSELY: 337 case COLOR_CODE_24BIT_YCC422: 338 case COLOR_CODE_16BIT_YCC422: 339 case COLOR_CODE_30BIT: 340 case COLOR_CODE_36BIT: 341 case COLOR_CODE_12BIT_YCC420: 342 return video_size_step = 2; 343 default: 344 DRM_ERROR("invalid color coding"); 345 return 0; 346 } 347 } 348 349 static u16 round_video_size(int coding, u16 video_size) 350 { 351 switch (coding) { 352 case COLOR_CODE_16BIT_YCC422: 353 case COLOR_CODE_24BIT_YCC422: 354 case COLOR_CODE_20BIT_YCC422_LOOSELY: 355 case COLOR_CODE_12BIT_YCC420: 356 /* round up active H pixels to a multiple of 2 */ 357 if ((video_size % 2) != 0) 358 video_size += 1; 359 break; 360 default: 361 break; 362 } 363 364 return video_size; 365 } 366 367 #define SPRD_MIPI_DSI_FMT_DSC 0xff 368 static u32 fmt_to_coding(u32 fmt) 369 { 370 switch (fmt) { 371 case MIPI_DSI_FMT_RGB565: 372 return COLOR_CODE_16BIT_CONFIG1; 373 case MIPI_DSI_FMT_RGB666: 374 case MIPI_DSI_FMT_RGB666_PACKED: 375 return COLOR_CODE_18BIT_CONFIG1; 376 case MIPI_DSI_FMT_RGB888: 377 return COLOR_CODE_24BIT; 378 case SPRD_MIPI_DSI_FMT_DSC: 379 return COLOR_CODE_COMPRESSTION; 380 default: 381 DRM_ERROR("Unsupported format (%d)\n", fmt); 382 return COLOR_CODE_24BIT; 383 } 384 } 385 386 #define ns_to_cycle(ns, byte_clk) \ 387 DIV_ROUND_UP((ns) * (byte_clk), 1000000) 388 389 static void sprd_dsi_init(struct dsi_context *ctx) 390 { 391 struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx); 392 u32 byte_clk = dsi->slave->hs_rate / 8; 393 u16 data_hs2lp, data_lp2hs, clk_hs2lp, clk_lp2hs; 394 u16 max_rd_time; 395 int div; 396 397 writel(0, ctx->base + SOFT_RESET); 398 writel(0xffffffff, ctx->base + MASK_PROTOCOL_INT); 399 writel(0xffffffff, ctx->base + MASK_INTERNAL_INT); 400 writel(1, ctx->base + DSI_MODE_CFG); 401 dsi_reg_up(ctx, EOTP_EN, RX_EOTP_EN, 0); 402 dsi_reg_up(ctx, EOTP_EN, TX_EOTP_EN, 0); 403 dsi_reg_up(ctx, RX_PKT_CHECK_CONFIG, RX_PKT_ECC_EN, RX_PKT_ECC_EN); 404 dsi_reg_up(ctx, RX_PKT_CHECK_CONFIG, RX_PKT_CRC_EN, RX_PKT_CRC_EN); 405 writel(1, ctx->base + TA_EN); 406 dsi_reg_up(ctx, VIRTUAL_CHANNEL_ID, VIDEO_PKT_VCID, 0); 407 dsi_reg_up(ctx, VIRTUAL_CHANNEL_ID, GEN_RX_VCID, 0); 408 409 div = DIV_ROUND_UP(byte_clk, dsi->slave->lp_rate); 410 writel(div, ctx->base + TX_ESC_CLK_CONFIG); 411 412 max_rd_time = ns_to_cycle(ctx->max_rd_time, byte_clk); 413 writel(max_rd_time, ctx->base + MAX_READ_TIME); 414 415 data_hs2lp = ns_to_cycle(ctx->data_hs2lp, byte_clk); 416 data_lp2hs = ns_to_cycle(ctx->data_lp2hs, byte_clk); 417 clk_hs2lp = ns_to_cycle(ctx->clk_hs2lp, byte_clk); 418 clk_lp2hs = ns_to_cycle(ctx->clk_lp2hs, byte_clk); 419 dsi_reg_wr(ctx, PHY_DATALANE_TIME_CONFIG, 420 PHY_DATALANE_HS_TO_LP_TIME, 16, data_hs2lp); 421 dsi_reg_wr(ctx, PHY_DATALANE_TIME_CONFIG, 422 PHY_DATALANE_LP_TO_HS_TIME, 0, data_lp2hs); 423 dsi_reg_wr(ctx, PHY_CLKLANE_TIME_CONFIG, 424 PHY_CLKLANE_HS_TO_LP_TIME, 16, clk_hs2lp); 425 dsi_reg_wr(ctx, PHY_CLKLANE_TIME_CONFIG, 426 PHY_CLKLANE_LP_TO_HS_TIME, 0, clk_lp2hs); 427 428 writel(1, ctx->base + SOFT_RESET); 429 } 430 431 /* 432 * Free up resources and shutdown host controller and PHY 433 */ 434 static void sprd_dsi_fini(struct dsi_context *ctx) 435 { 436 writel(0xffffffff, ctx->base + MASK_PROTOCOL_INT); 437 writel(0xffffffff, ctx->base + MASK_INTERNAL_INT); 438 writel(0, ctx->base + SOFT_RESET); 439 } 440 441 /* 442 * If not in burst mode, it will compute the video and null packet sizes 443 * according to necessity. 444 * Configure timers for data lanes and/or clock lane to return to LP when 445 * bandwidth is not filled by data. 446 */ 447 static int sprd_dsi_dpi_video(struct dsi_context *ctx) 448 { 449 struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx); 450 struct videomode *vm = &ctx->vm; 451 u32 byte_clk = dsi->slave->hs_rate / 8; 452 u16 bpp_x100; 453 u16 video_size; 454 u32 ratio_x1000; 455 u16 null_pkt_size = 0; 456 u8 video_size_step; 457 u32 hs_to; 458 u32 total_bytes; 459 u32 bytes_per_chunk; 460 u32 chunks = 0; 461 u32 bytes_left = 0; 462 u32 chunk_overhead; 463 const u8 pkt_header = 6; 464 u8 coding; 465 int div; 466 u16 hline; 467 u16 byte_cycle; 468 469 coding = fmt_to_coding(dsi->slave->format); 470 video_size = round_video_size(coding, vm->hactive); 471 bpp_x100 = calc_bytes_per_pixel_x100(coding); 472 video_size_step = calc_video_size_step(coding); 473 ratio_x1000 = byte_clk * 1000 / (vm->pixelclock / 1000); 474 hline = vm->hactive + vm->hsync_len + vm->hfront_porch + 475 vm->hback_porch; 476 477 writel(0, ctx->base + SOFT_RESET); 478 dsi_reg_wr(ctx, VID_MODE_CFG, FRAME_BTA_ACK_EN, 15, ctx->frame_ack_en); 479 dsi_reg_wr(ctx, DPI_VIDEO_FORMAT, DPI_VIDEO_MODE_FORMAT, 0, coding); 480 dsi_reg_wr(ctx, VID_MODE_CFG, VID_MODE_TYPE, 0, ctx->burst_mode); 481 byte_cycle = 95 * hline * ratio_x1000 / 100000; 482 dsi_reg_wr(ctx, VIDEO_SIG_DELAY_CONFIG, VIDEO_SIG_DELAY, 0, byte_cycle); 483 byte_cycle = hline * ratio_x1000 / 1000; 484 writel(byte_cycle, ctx->base + VIDEO_LINE_TIME); 485 byte_cycle = vm->hsync_len * ratio_x1000 / 1000; 486 dsi_reg_wr(ctx, VIDEO_LINE_HBLK_TIME, VIDEO_LINE_HSA_TIME, 16, byte_cycle); 487 byte_cycle = vm->hback_porch * ratio_x1000 / 1000; 488 dsi_reg_wr(ctx, VIDEO_LINE_HBLK_TIME, VIDEO_LINE_HBP_TIME, 0, byte_cycle); 489 writel(vm->vactive, ctx->base + VIDEO_VACTIVE_LINES); 490 dsi_reg_wr(ctx, VIDEO_VBLK_LINES, VFP_LINES, 0, vm->vfront_porch); 491 dsi_reg_wr(ctx, VIDEO_VBLK_LINES, VBP_LINES, 10, vm->vback_porch); 492 dsi_reg_wr(ctx, VIDEO_VBLK_LINES, VSA_LINES, 20, vm->vsync_len); 493 dsi_reg_up(ctx, VID_MODE_CFG, LP_HBP_EN | LP_HFP_EN | LP_VACT_EN | 494 LP_VFP_EN | LP_VBP_EN | LP_VSA_EN, LP_HBP_EN | LP_HFP_EN | 495 LP_VACT_EN | LP_VFP_EN | LP_VBP_EN | LP_VSA_EN); 496 497 hs_to = (hline * vm->vactive) + (2 * bpp_x100) / 100; 498 for (div = 0x80; (div < hs_to) && (div > 2); div--) { 499 if ((hs_to % div) == 0) { 500 writel(div, ctx->base + TIMEOUT_CNT_CLK_CONFIG); 501 writel(hs_to / div, ctx->base + LRX_H_TO_CONFIG); 502 writel(hs_to / div, ctx->base + HTX_TO_CONFIG); 503 break; 504 } 505 } 506 507 if (ctx->burst_mode == VIDEO_BURST_WITH_SYNC_PULSES) { 508 dsi_reg_wr(ctx, VIDEO_PKT_CONFIG, VIDEO_PKT_SIZE, 0, video_size); 509 writel(0, ctx->base + VIDEO_NULLPKT_SIZE); 510 dsi_reg_up(ctx, VIDEO_PKT_CONFIG, VIDEO_LINE_CHUNK_NUM, 0); 511 } else { 512 /* non burst transmission */ 513 null_pkt_size = 0; 514 515 /* bytes to be sent - first as one chunk */ 516 bytes_per_chunk = vm->hactive * bpp_x100 / 100 + pkt_header; 517 518 /* hline total bytes from the DPI interface */ 519 total_bytes = (vm->hactive + vm->hfront_porch) * 520 ratio_x1000 / dsi->slave->lanes / 1000; 521 522 /* check if the pixels actually fit on the DSI link */ 523 if (total_bytes < bytes_per_chunk) { 524 drm_err(dsi->drm, "current resolution can not be set\n"); 525 return -EINVAL; 526 } 527 528 chunk_overhead = total_bytes - bytes_per_chunk; 529 530 /* overhead higher than 1 -> enable multi packets */ 531 if (chunk_overhead > 1) { 532 /* multi packets */ 533 for (video_size = video_size_step; 534 video_size < vm->hactive; 535 video_size += video_size_step) { 536 if (vm->hactive * 1000 / video_size % 1000) 537 continue; 538 539 chunks = vm->hactive / video_size; 540 bytes_per_chunk = bpp_x100 * video_size / 100 541 + pkt_header; 542 if (total_bytes >= (bytes_per_chunk * chunks)) { 543 bytes_left = total_bytes - 544 bytes_per_chunk * chunks; 545 break; 546 } 547 } 548 549 /* prevent overflow (unsigned - unsigned) */ 550 if (bytes_left > (pkt_header * chunks)) { 551 null_pkt_size = (bytes_left - 552 pkt_header * chunks) / chunks; 553 /* avoid register overflow */ 554 if (null_pkt_size > 1023) 555 null_pkt_size = 1023; 556 } 557 558 } else { 559 /* single packet */ 560 chunks = 1; 561 562 /* must be a multiple of 4 except 18 loosely */ 563 for (video_size = vm->hactive; 564 (video_size % video_size_step) != 0; 565 video_size++) 566 ; 567 } 568 569 dsi_reg_wr(ctx, VIDEO_PKT_CONFIG, VIDEO_PKT_SIZE, 0, video_size); 570 writel(null_pkt_size, ctx->base + VIDEO_NULLPKT_SIZE); 571 dsi_reg_wr(ctx, VIDEO_PKT_CONFIG, VIDEO_LINE_CHUNK_NUM, 16, chunks); 572 } 573 574 writel(ctx->int0_mask, ctx->base + MASK_PROTOCOL_INT); 575 writel(ctx->int1_mask, ctx->base + MASK_INTERNAL_INT); 576 writel(1, ctx->base + SOFT_RESET); 577 578 return 0; 579 } 580 581 static void sprd_dsi_edpi_video(struct dsi_context *ctx) 582 { 583 struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx); 584 const u32 fifo_depth = 1096; 585 const u32 word_length = 4; 586 u32 hactive = ctx->vm.hactive; 587 u32 bpp_x100; 588 u32 max_fifo_len; 589 u8 coding; 590 591 coding = fmt_to_coding(dsi->slave->format); 592 bpp_x100 = calc_bytes_per_pixel_x100(coding); 593 max_fifo_len = word_length * fifo_depth * 100 / bpp_x100; 594 595 writel(0, ctx->base + SOFT_RESET); 596 dsi_reg_wr(ctx, DPI_VIDEO_FORMAT, DPI_VIDEO_MODE_FORMAT, 0, coding); 597 dsi_reg_wr(ctx, CMD_MODE_CFG, TEAR_FX_EN, 0, ctx->te_ack_en); 598 599 if (max_fifo_len > hactive) 600 writel(hactive, ctx->base + DCS_WM_PKT_SIZE); 601 else 602 writel(max_fifo_len, ctx->base + DCS_WM_PKT_SIZE); 603 604 writel(ctx->int0_mask, ctx->base + MASK_PROTOCOL_INT); 605 writel(ctx->int1_mask, ctx->base + MASK_INTERNAL_INT); 606 writel(1, ctx->base + SOFT_RESET); 607 } 608 609 /* 610 * Send a packet on the generic interface, 611 * this function has an active delay to wait for the buffer to clear. 612 * The delay is limited to: 613 * (param_length / 4) x DSIH_FIFO_ACTIVE_WAIT x register access time 614 * the controller restricts the sending of. 615 * 616 * This function will not be able to send Null and Blanking packets due to 617 * controller restriction 618 */ 619 static int sprd_dsi_wr_pkt(struct dsi_context *ctx, u8 vc, u8 type, 620 const u8 *param, u16 len) 621 { 622 struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx); 623 u8 wc_lsbyte, wc_msbyte; 624 u32 payload; 625 int i, j, ret; 626 627 if (vc > 3) 628 return -EINVAL; 629 630 /* 1st: for long packet, must config payload first */ 631 ret = dsi_wait_tx_payload_fifo_empty(ctx); 632 if (ret) { 633 drm_err(dsi->drm, "tx payload fifo is not empty\n"); 634 return ret; 635 } 636 637 if (len > 2) { 638 for (i = 0, j = 0; i < len; i += j) { 639 payload = 0; 640 for (j = 0; (j < 4) && ((j + i) < (len)); j++) 641 payload |= param[i + j] << (j * 8); 642 643 writel(payload, ctx->base + GEN_PLD_DATA); 644 } 645 wc_lsbyte = len & 0xff; 646 wc_msbyte = len >> 8; 647 } else { 648 wc_lsbyte = (len > 0) ? param[0] : 0; 649 wc_msbyte = (len > 1) ? param[1] : 0; 650 } 651 652 /* 2nd: then set packet header */ 653 ret = dsi_wait_tx_cmd_fifo_empty(ctx); 654 if (ret) { 655 drm_err(dsi->drm, "tx cmd fifo is not empty\n"); 656 return ret; 657 } 658 659 writel(type | (vc << 6) | (wc_lsbyte << 8) | (wc_msbyte << 16), 660 ctx->base + GEN_HDR); 661 662 return 0; 663 } 664 665 /* 666 * Send READ packet to peripheral using the generic interface, 667 * this will force command mode and stop video mode (because of BTA). 668 * 669 * This function has an active delay to wait for the buffer to clear, 670 * the delay is limited to 2 x DSIH_FIFO_ACTIVE_WAIT 671 * (waiting for command buffer, and waiting for receiving) 672 * @note this function will enable BTA 673 */ 674 static int sprd_dsi_rd_pkt(struct dsi_context *ctx, u8 vc, u8 type, 675 u8 msb_byte, u8 lsb_byte, 676 u8 *buffer, u8 bytes_to_read) 677 { 678 struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx); 679 int i, ret; 680 int count = 0; 681 u32 temp; 682 683 if (vc > 3) 684 return -EINVAL; 685 686 /* 1st: send read command to peripheral */ 687 ret = dsi_reg_rd(ctx, CMD_MODE_STATUS, GEN_CMD_CMD_FIFO_EMPTY, 5); 688 if (!ret) 689 return -EIO; 690 691 writel(type | (vc << 6) | (lsb_byte << 8) | (msb_byte << 16), 692 ctx->base + GEN_HDR); 693 694 /* 2nd: wait peripheral response completed */ 695 ret = dsi_wait_rd_resp_completed(ctx); 696 if (ret) { 697 drm_err(dsi->drm, "wait read response time out\n"); 698 return ret; 699 } 700 701 /* 3rd: get data from rx payload fifo */ 702 ret = dsi_reg_rd(ctx, CMD_MODE_STATUS, GEN_CMD_RDATA_FIFO_EMPTY, 1); 703 if (ret) { 704 drm_err(dsi->drm, "rx payload fifo empty\n"); 705 return -EIO; 706 } 707 708 for (i = 0; i < 100; i++) { 709 temp = readl(ctx->base + GEN_PLD_DATA); 710 711 if (count < bytes_to_read) 712 buffer[count++] = temp & 0xff; 713 if (count < bytes_to_read) 714 buffer[count++] = (temp >> 8) & 0xff; 715 if (count < bytes_to_read) 716 buffer[count++] = (temp >> 16) & 0xff; 717 if (count < bytes_to_read) 718 buffer[count++] = (temp >> 24) & 0xff; 719 720 ret = dsi_reg_rd(ctx, CMD_MODE_STATUS, GEN_CMD_RDATA_FIFO_EMPTY, 1); 721 if (ret) 722 return count; 723 } 724 725 return 0; 726 } 727 728 static void sprd_dsi_set_work_mode(struct dsi_context *ctx, u8 mode) 729 { 730 if (mode == DSI_MODE_CMD) 731 writel(1, ctx->base + DSI_MODE_CFG); 732 else 733 writel(0, ctx->base + DSI_MODE_CFG); 734 } 735 736 static void sprd_dsi_state_reset(struct dsi_context *ctx) 737 { 738 writel(0, ctx->base + SOFT_RESET); 739 udelay(100); 740 writel(1, ctx->base + SOFT_RESET); 741 } 742 743 static int sprd_dphy_init(struct dsi_context *ctx) 744 { 745 struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx); 746 int ret; 747 748 dsi_reg_up(ctx, PHY_INTERFACE_CTRL, RF_PHY_RESET_N, 0); 749 dsi_reg_up(ctx, PHY_INTERFACE_CTRL, RF_PHY_SHUTDOWN, 0); 750 dsi_reg_up(ctx, PHY_INTERFACE_CTRL, RF_PHY_CLK_EN, 0); 751 752 dsi_reg_up(ctx, PHY_TST_CTRL0, PHY_TESTCLR, 0); 753 dsi_reg_up(ctx, PHY_TST_CTRL0, PHY_TESTCLR, PHY_TESTCLR); 754 dsi_reg_up(ctx, PHY_TST_CTRL0, PHY_TESTCLR, 0); 755 756 dphy_pll_config(ctx); 757 dphy_timing_config(ctx); 758 759 dsi_reg_up(ctx, PHY_INTERFACE_CTRL, RF_PHY_SHUTDOWN, RF_PHY_SHUTDOWN); 760 dsi_reg_up(ctx, PHY_INTERFACE_CTRL, RF_PHY_RESET_N, RF_PHY_RESET_N); 761 writel(0x1C, ctx->base + PHY_MIN_STOP_TIME); 762 dsi_reg_up(ctx, PHY_INTERFACE_CTRL, RF_PHY_CLK_EN, RF_PHY_CLK_EN); 763 writel(dsi->slave->lanes - 1, ctx->base + PHY_LANE_NUM_CONFIG); 764 765 ret = dphy_wait_pll_locked(ctx); 766 if (ret) { 767 drm_err(dsi->drm, "dphy initial failed\n"); 768 return ret; 769 } 770 771 return 0; 772 } 773 774 static void sprd_dphy_fini(struct dsi_context *ctx) 775 { 776 dsi_reg_up(ctx, PHY_INTERFACE_CTRL, RF_PHY_RESET_N, 0); 777 dsi_reg_up(ctx, PHY_INTERFACE_CTRL, RF_PHY_SHUTDOWN, 0); 778 dsi_reg_up(ctx, PHY_INTERFACE_CTRL, RF_PHY_RESET_N, RF_PHY_RESET_N); 779 } 780 781 static void sprd_dsi_encoder_mode_set(struct drm_encoder *encoder, 782 struct drm_display_mode *mode, 783 struct drm_display_mode *adj_mode) 784 { 785 struct sprd_dsi *dsi = encoder_to_dsi(encoder); 786 787 drm_display_mode_to_videomode(adj_mode, &dsi->ctx.vm); 788 } 789 790 static void sprd_dsi_encoder_enable(struct drm_encoder *encoder) 791 { 792 struct sprd_dsi *dsi = encoder_to_dsi(encoder); 793 struct sprd_dpu *dpu = to_sprd_crtc(encoder->crtc); 794 struct dsi_context *ctx = &dsi->ctx; 795 796 if (ctx->enabled) { 797 drm_warn(dsi->drm, "dsi is initialized\n"); 798 return; 799 } 800 801 sprd_dsi_init(ctx); 802 if (ctx->work_mode == DSI_MODE_VIDEO) 803 sprd_dsi_dpi_video(ctx); 804 else 805 sprd_dsi_edpi_video(ctx); 806 807 sprd_dphy_init(ctx); 808 809 sprd_dsi_set_work_mode(ctx, ctx->work_mode); 810 sprd_dsi_state_reset(ctx); 811 812 if (dsi->slave->mode_flags & MIPI_DSI_CLOCK_NON_CONTINUOUS) { 813 dsi_reg_up(ctx, PHY_CLK_LANE_LP_CTRL, AUTO_CLKLANE_CTRL_EN, 814 AUTO_CLKLANE_CTRL_EN); 815 } else { 816 dsi_reg_up(ctx, PHY_CLK_LANE_LP_CTRL, RF_PHY_CLK_EN, RF_PHY_CLK_EN); 817 dsi_reg_up(ctx, PHY_CLK_LANE_LP_CTRL, PHY_CLKLANE_TX_REQ_HS, 818 PHY_CLKLANE_TX_REQ_HS); 819 dphy_wait_pll_locked(ctx); 820 } 821 822 sprd_dpu_run(dpu); 823 824 ctx->enabled = true; 825 } 826 827 static void sprd_dsi_encoder_disable(struct drm_encoder *encoder) 828 { 829 struct sprd_dsi *dsi = encoder_to_dsi(encoder); 830 struct sprd_dpu *dpu = to_sprd_crtc(encoder->crtc); 831 struct dsi_context *ctx = &dsi->ctx; 832 833 if (!ctx->enabled) { 834 drm_warn(dsi->drm, "dsi isn't initialized\n"); 835 return; 836 } 837 838 sprd_dpu_stop(dpu); 839 sprd_dphy_fini(ctx); 840 sprd_dsi_fini(ctx); 841 842 ctx->enabled = false; 843 } 844 845 static const struct drm_encoder_helper_funcs sprd_encoder_helper_funcs = { 846 .mode_set = sprd_dsi_encoder_mode_set, 847 .enable = sprd_dsi_encoder_enable, 848 .disable = sprd_dsi_encoder_disable 849 }; 850 851 static const struct drm_encoder_funcs sprd_encoder_funcs = { 852 .destroy = drm_encoder_cleanup, 853 }; 854 855 static int sprd_dsi_encoder_init(struct sprd_dsi *dsi, 856 struct device *dev) 857 { 858 struct drm_encoder *encoder = &dsi->encoder; 859 u32 crtc_mask; 860 int ret; 861 862 crtc_mask = drm_of_find_possible_crtcs(dsi->drm, dev->of_node); 863 if (!crtc_mask) { 864 drm_err(dsi->drm, "failed to find crtc mask\n"); 865 return -EINVAL; 866 } 867 868 drm_dbg(dsi->drm, "find possible crtcs: 0x%08x\n", crtc_mask); 869 870 encoder->possible_crtcs = crtc_mask; 871 ret = drm_encoder_init(dsi->drm, encoder, &sprd_encoder_funcs, 872 DRM_MODE_ENCODER_DSI, NULL); 873 if (ret) { 874 drm_err(dsi->drm, "failed to init dsi encoder\n"); 875 return ret; 876 } 877 878 drm_encoder_helper_add(encoder, &sprd_encoder_helper_funcs); 879 880 return 0; 881 } 882 883 static int sprd_dsi_bridge_init(struct sprd_dsi *dsi, 884 struct device *dev) 885 { 886 int ret; 887 888 dsi->panel_bridge = devm_drm_of_get_bridge(dev, dev->of_node, 1, 0); 889 if (IS_ERR(dsi->panel_bridge)) 890 return PTR_ERR(dsi->panel_bridge); 891 892 ret = drm_bridge_attach(&dsi->encoder, dsi->panel_bridge, NULL, 0); 893 if (ret) 894 return ret; 895 896 return 0; 897 } 898 899 static int sprd_dsi_context_init(struct sprd_dsi *dsi, 900 struct device *dev) 901 { 902 struct platform_device *pdev = to_platform_device(dev); 903 struct dsi_context *ctx = &dsi->ctx; 904 struct resource *res; 905 906 res = platform_get_resource(pdev, IORESOURCE_MEM, 0); 907 if (!res) { 908 dev_err(dev, "failed to get I/O resource\n"); 909 return -EINVAL; 910 } 911 912 ctx->base = devm_ioremap(dev, res->start, resource_size(res)); 913 if (!ctx->base) { 914 drm_err(dsi->drm, "failed to map dsi host registers\n"); 915 return -ENXIO; 916 } 917 918 ctx->regmap = devm_regmap_init(dev, ®map_tst_io, dsi, &byte_config); 919 if (IS_ERR(ctx->regmap)) { 920 drm_err(dsi->drm, "dphy regmap init failed\n"); 921 return PTR_ERR(ctx->regmap); 922 } 923 924 ctx->data_hs2lp = 120; 925 ctx->data_lp2hs = 500; 926 ctx->clk_hs2lp = 4; 927 ctx->clk_lp2hs = 15; 928 ctx->max_rd_time = 6000; 929 ctx->int0_mask = 0xffffffff; 930 ctx->int1_mask = 0xffffffff; 931 ctx->enabled = true; 932 933 return 0; 934 } 935 936 static int sprd_dsi_bind(struct device *dev, struct device *master, void *data) 937 { 938 struct drm_device *drm = data; 939 struct sprd_dsi *dsi = dev_get_drvdata(dev); 940 int ret; 941 942 dsi->drm = drm; 943 944 ret = sprd_dsi_encoder_init(dsi, dev); 945 if (ret) 946 return ret; 947 948 ret = sprd_dsi_bridge_init(dsi, dev); 949 if (ret) 950 return ret; 951 952 ret = sprd_dsi_context_init(dsi, dev); 953 if (ret) 954 return ret; 955 956 return 0; 957 } 958 959 static void sprd_dsi_unbind(struct device *dev, 960 struct device *master, void *data) 961 { 962 struct sprd_dsi *dsi = dev_get_drvdata(dev); 963 964 drm_of_panel_bridge_remove(dev->of_node, 1, 0); 965 966 drm_encoder_cleanup(&dsi->encoder); 967 } 968 969 static const struct component_ops dsi_component_ops = { 970 .bind = sprd_dsi_bind, 971 .unbind = sprd_dsi_unbind, 972 }; 973 974 static int sprd_dsi_host_attach(struct mipi_dsi_host *host, 975 struct mipi_dsi_device *slave) 976 { 977 struct sprd_dsi *dsi = host_to_dsi(host); 978 struct dsi_context *ctx = &dsi->ctx; 979 980 dsi->slave = slave; 981 982 if (slave->mode_flags & MIPI_DSI_MODE_VIDEO) 983 ctx->work_mode = DSI_MODE_VIDEO; 984 else 985 ctx->work_mode = DSI_MODE_CMD; 986 987 if (slave->mode_flags & MIPI_DSI_MODE_VIDEO_BURST) 988 ctx->burst_mode = VIDEO_BURST_WITH_SYNC_PULSES; 989 else if (slave->mode_flags & MIPI_DSI_MODE_VIDEO_SYNC_PULSE) 990 ctx->burst_mode = VIDEO_NON_BURST_WITH_SYNC_PULSES; 991 else 992 ctx->burst_mode = VIDEO_NON_BURST_WITH_SYNC_EVENTS; 993 994 return component_add(host->dev, &dsi_component_ops); 995 } 996 997 static int sprd_dsi_host_detach(struct mipi_dsi_host *host, 998 struct mipi_dsi_device *slave) 999 { 1000 component_del(host->dev, &dsi_component_ops); 1001 1002 return 0; 1003 } 1004 1005 static ssize_t sprd_dsi_host_transfer(struct mipi_dsi_host *host, 1006 const struct mipi_dsi_msg *msg) 1007 { 1008 struct sprd_dsi *dsi = host_to_dsi(host); 1009 const u8 *tx_buf = msg->tx_buf; 1010 1011 if (msg->rx_buf && msg->rx_len) { 1012 u8 lsb = (msg->tx_len > 0) ? tx_buf[0] : 0; 1013 u8 msb = (msg->tx_len > 1) ? tx_buf[1] : 0; 1014 1015 return sprd_dsi_rd_pkt(&dsi->ctx, msg->channel, msg->type, 1016 msb, lsb, msg->rx_buf, msg->rx_len); 1017 } 1018 1019 if (msg->tx_buf && msg->tx_len) 1020 return sprd_dsi_wr_pkt(&dsi->ctx, msg->channel, msg->type, 1021 tx_buf, msg->tx_len); 1022 1023 return 0; 1024 } 1025 1026 static const struct mipi_dsi_host_ops sprd_dsi_host_ops = { 1027 .attach = sprd_dsi_host_attach, 1028 .detach = sprd_dsi_host_detach, 1029 .transfer = sprd_dsi_host_transfer, 1030 }; 1031 1032 static const struct of_device_id dsi_match_table[] = { 1033 { .compatible = "sprd,sharkl3-dsi-host" }, 1034 { /* sentinel */ }, 1035 }; 1036 1037 static int sprd_dsi_probe(struct platform_device *pdev) 1038 { 1039 struct device *dev = &pdev->dev; 1040 struct sprd_dsi *dsi; 1041 1042 dsi = devm_kzalloc(dev, sizeof(*dsi), GFP_KERNEL); 1043 if (!dsi) 1044 return -ENOMEM; 1045 1046 dev_set_drvdata(dev, dsi); 1047 1048 dsi->host.ops = &sprd_dsi_host_ops; 1049 dsi->host.dev = dev; 1050 1051 return mipi_dsi_host_register(&dsi->host); 1052 } 1053 1054 static void sprd_dsi_remove(struct platform_device *pdev) 1055 { 1056 struct sprd_dsi *dsi = dev_get_drvdata(&pdev->dev); 1057 1058 mipi_dsi_host_unregister(&dsi->host); 1059 } 1060 1061 struct platform_driver sprd_dsi_driver = { 1062 .probe = sprd_dsi_probe, 1063 .remove_new = sprd_dsi_remove, 1064 .driver = { 1065 .name = "sprd-dsi-drv", 1066 .of_match_table = dsi_match_table, 1067 }, 1068 }; 1069 1070 MODULE_AUTHOR("Leon He <leon.he@unisoc.com>"); 1071 MODULE_AUTHOR("Kevin Tang <kevin.tang@unisoc.com>"); 1072 MODULE_DESCRIPTION("Unisoc MIPI DSI HOST Controller Driver"); 1073 MODULE_LICENSE("GPL v2"); 1074