1 /* 2 * Copyright 2018 Red Hat Inc. 3 * 4 * Permission is hereby granted, free of charge, to any person obtaining a 5 * copy of this software and associated documentation files (the "Software"), 6 * to deal in the Software without restriction, including without limitation 7 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 8 * and/or sell copies of the Software, and to permit persons to whom the 9 * Software is furnished to do so, subject to the following conditions: 10 * 11 * The above copyright notice and this permission notice shall be included in 12 * all copies or substantial portions of the Software. 13 * 14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 20 * OTHER DEALINGS IN THE SOFTWARE. 21 */ 22 #include "priv.h" 23 #include "chan.h" 24 #include "hdmi.h" 25 #include "head.h" 26 #include "ior.h" 27 #include "outp.h" 28 29 #include <core/client.h> 30 #include <core/gpuobj.h> 31 #include <core/ramht.h> 32 #include <subdev/timer.h> 33 34 #include <nvif/class.h> 35 #include <nvif/unpack.h> 36 37 static void 38 gv100_sor_hda_device_entry(struct nvkm_ior *ior, int head) 39 { 40 struct nvkm_device *device = ior->disp->engine.subdev.device; 41 const u32 hoff = 0x800 * head; 42 43 nvkm_mask(device, 0x616528 + hoff, 0x00000070, head << 4); 44 } 45 46 const struct nvkm_ior_func_hda 47 gv100_sor_hda = { 48 .hpd = gf119_sor_hda_hpd, 49 .eld = gf119_sor_hda_eld, 50 .device_entry = gv100_sor_hda_device_entry, 51 }; 52 53 void 54 gv100_sor_dp_watermark(struct nvkm_ior *sor, int head, u8 watermark) 55 { 56 struct nvkm_device *device = sor->disp->engine.subdev.device; 57 const u32 hoff = head * 0x800; 58 59 nvkm_mask(device, 0x616550 + hoff, 0x0c00003f, 0x08000000 | watermark); 60 } 61 62 void 63 gv100_sor_dp_audio_sym(struct nvkm_ior *sor, int head, u16 h, u32 v) 64 { 65 struct nvkm_device *device = sor->disp->engine.subdev.device; 66 const u32 hoff = head * 0x800; 67 68 nvkm_mask(device, 0x616568 + hoff, 0x0000ffff, h); 69 nvkm_mask(device, 0x61656c + hoff, 0x00ffffff, v); 70 } 71 72 void 73 gv100_sor_dp_audio(struct nvkm_ior *sor, int head, bool enable) 74 { 75 struct nvkm_device *device = sor->disp->engine.subdev.device; 76 const u32 hoff = 0x800 * head; 77 const u32 data = 0x80000000 | (0x00000001 * enable); 78 const u32 mask = 0x8000000d; 79 80 nvkm_mask(device, 0x616560 + hoff, mask, data); 81 nvkm_msec(device, 2000, 82 if (!(nvkm_rd32(device, 0x616560 + hoff) & 0x80000000)) 83 break; 84 ); 85 } 86 87 static const struct nvkm_ior_func_dp 88 gv100_sor_dp = { 89 .lanes = { 0, 1, 2, 3 }, 90 .links = gf119_sor_dp_links, 91 .power = g94_sor_dp_power, 92 .pattern = gm107_sor_dp_pattern, 93 .drive = gm200_sor_dp_drive, 94 .audio = gv100_sor_dp_audio, 95 .audio_sym = gv100_sor_dp_audio_sym, 96 .watermark = gv100_sor_dp_watermark, 97 }; 98 99 static void 100 gv100_sor_hdmi_infoframe_vsi(struct nvkm_ior *ior, int head, void *data, u32 size) 101 { 102 struct nvkm_device *device = ior->disp->engine.subdev.device; 103 struct packed_hdmi_infoframe vsi; 104 const u32 hoff = head * 0x400; 105 106 pack_hdmi_infoframe(&vsi, data, size); 107 108 nvkm_mask(device, 0x6f0100 + hoff, 0x00010001, 0x00000000); 109 if (!size) 110 return; 111 112 nvkm_wr32(device, 0x6f0108 + hoff, vsi.header); 113 nvkm_wr32(device, 0x6f010c + hoff, vsi.subpack0_low); 114 nvkm_wr32(device, 0x6f0110 + hoff, vsi.subpack0_high); 115 nvkm_wr32(device, 0x6f0114 + hoff, 0x00000000); 116 nvkm_wr32(device, 0x6f0118 + hoff, 0x00000000); 117 nvkm_wr32(device, 0x6f011c + hoff, 0x00000000); 118 nvkm_wr32(device, 0x6f0120 + hoff, 0x00000000); 119 nvkm_wr32(device, 0x6f0124 + hoff, 0x00000000); 120 nvkm_mask(device, 0x6f0100 + hoff, 0x00000001, 0x00000001); 121 } 122 123 static void 124 gv100_sor_hdmi_infoframe_avi(struct nvkm_ior *ior, int head, void *data, u32 size) 125 { 126 struct nvkm_device *device = ior->disp->engine.subdev.device; 127 struct packed_hdmi_infoframe avi; 128 const u32 hoff = head * 0x400; 129 130 pack_hdmi_infoframe(&avi, data, size); 131 132 nvkm_mask(device, 0x6f0000 + hoff, 0x00000001, 0x00000000); 133 if (!size) 134 return; 135 136 nvkm_wr32(device, 0x6f0008 + hoff, avi.header); 137 nvkm_wr32(device, 0x6f000c + hoff, avi.subpack0_low); 138 nvkm_wr32(device, 0x6f0010 + hoff, avi.subpack0_high); 139 nvkm_wr32(device, 0x6f0014 + hoff, avi.subpack1_low); 140 nvkm_wr32(device, 0x6f0018 + hoff, avi.subpack1_high); 141 142 nvkm_mask(device, 0x6f0000 + hoff, 0x00000001, 0x00000001); 143 } 144 145 static void 146 gv100_sor_hdmi_ctrl(struct nvkm_ior *ior, int head, bool enable, u8 max_ac_packet, u8 rekey) 147 { 148 struct nvkm_device *device = ior->disp->engine.subdev.device; 149 const u32 ctrl = 0x40000000 * enable | 150 max_ac_packet << 16 | 151 rekey; 152 const u32 hoff = head * 0x800; 153 const u32 hdmi = head * 0x400; 154 155 if (!(ctrl & 0x40000000)) { 156 nvkm_mask(device, 0x6165c0 + hoff, 0x40000000, 0x00000000); 157 nvkm_mask(device, 0x6f0100 + hdmi, 0x00000001, 0x00000000); 158 nvkm_mask(device, 0x6f00c0 + hdmi, 0x00000001, 0x00000000); 159 nvkm_mask(device, 0x6f0000 + hdmi, 0x00000001, 0x00000000); 160 return; 161 } 162 163 /* General Control (GCP). */ 164 nvkm_mask(device, 0x6f00c0 + hdmi, 0x00000001, 0x00000000); 165 nvkm_wr32(device, 0x6f00cc + hdmi, 0x00000010); 166 nvkm_mask(device, 0x6f00c0 + hdmi, 0x00000001, 0x00000001); 167 168 /* Audio Clock Regeneration (ACR). */ 169 nvkm_wr32(device, 0x6f0080 + hdmi, 0x82000000); 170 171 /* NV_PDISP_SF_HDMI_CTRL. */ 172 nvkm_mask(device, 0x6165c0 + hoff, 0x401f007f, ctrl); 173 } 174 175 const struct nvkm_ior_func_hdmi 176 gv100_sor_hdmi = { 177 .ctrl = gv100_sor_hdmi_ctrl, 178 .scdc = gm200_sor_hdmi_scdc, 179 .infoframe_avi = gv100_sor_hdmi_infoframe_avi, 180 .infoframe_vsi = gv100_sor_hdmi_infoframe_vsi, 181 }; 182 183 void 184 gv100_sor_state(struct nvkm_ior *sor, struct nvkm_ior_state *state) 185 { 186 struct nvkm_device *device = sor->disp->engine.subdev.device; 187 const u32 coff = (state == &sor->arm) * 0x8000 + sor->id * 0x20; 188 u32 ctrl = nvkm_rd32(device, 0x680300 + coff); 189 190 state->proto_evo = (ctrl & 0x00000f00) >> 8; 191 switch (state->proto_evo) { 192 case 0: state->proto = LVDS; state->link = 1; break; 193 case 1: state->proto = TMDS; state->link = 1; break; 194 case 2: state->proto = TMDS; state->link = 2; break; 195 case 5: state->proto = TMDS; state->link = 3; break; 196 case 8: state->proto = DP; state->link = 1; break; 197 case 9: state->proto = DP; state->link = 2; break; 198 default: 199 state->proto = UNKNOWN; 200 break; 201 } 202 203 state->head = ctrl & 0x000000ff; 204 } 205 206 static const struct nvkm_ior_func 207 gv100_sor = { 208 .route = { 209 .get = gm200_sor_route_get, 210 .set = gm200_sor_route_set, 211 }, 212 .state = gv100_sor_state, 213 .power = nv50_sor_power, 214 .clock = gf119_sor_clock, 215 .bl = >215_sor_bl, 216 .hdmi = &gv100_sor_hdmi, 217 .dp = &gv100_sor_dp, 218 .hda = &gv100_sor_hda, 219 }; 220 221 static int 222 gv100_sor_new(struct nvkm_disp *disp, int id) 223 { 224 struct nvkm_device *device = disp->engine.subdev.device; 225 u32 hda; 226 227 if (!((hda = nvkm_rd32(device, 0x08a15c)) & 0x40000000)) 228 hda = nvkm_rd32(device, 0x118fb0) >> 8; 229 230 return nvkm_ior_new_(&gv100_sor, disp, SOR, id, hda & BIT(id)); 231 } 232 233 int 234 gv100_sor_cnt(struct nvkm_disp *disp, unsigned long *pmask) 235 { 236 struct nvkm_device *device = disp->engine.subdev.device; 237 238 *pmask = (nvkm_rd32(device, 0x610060) & 0x0000ff00) >> 8; 239 return (nvkm_rd32(device, 0x610074) & 0x00000f00) >> 8; 240 } 241 242 static void 243 gv100_head_vblank_put(struct nvkm_head *head) 244 { 245 struct nvkm_device *device = head->disp->engine.subdev.device; 246 nvkm_mask(device, 0x611d80 + (head->id * 4), 0x00000004, 0x00000000); 247 } 248 249 static void 250 gv100_head_vblank_get(struct nvkm_head *head) 251 { 252 struct nvkm_device *device = head->disp->engine.subdev.device; 253 nvkm_mask(device, 0x611d80 + (head->id * 4), 0x00000004, 0x00000004); 254 } 255 256 static void 257 gv100_head_rgpos(struct nvkm_head *head, u16 *hline, u16 *vline) 258 { 259 struct nvkm_device *device = head->disp->engine.subdev.device; 260 const u32 hoff = head->id * 0x800; 261 /* vline read locks hline. */ 262 *vline = nvkm_rd32(device, 0x616330 + hoff) & 0x0000ffff; 263 *hline = nvkm_rd32(device, 0x616334 + hoff) & 0x0000ffff; 264 } 265 266 static void 267 gv100_head_state(struct nvkm_head *head, struct nvkm_head_state *state) 268 { 269 struct nvkm_device *device = head->disp->engine.subdev.device; 270 const u32 hoff = (state == &head->arm) * 0x8000 + head->id * 0x400; 271 u32 data; 272 273 data = nvkm_rd32(device, 0x682064 + hoff); 274 state->vtotal = (data & 0xffff0000) >> 16; 275 state->htotal = (data & 0x0000ffff); 276 data = nvkm_rd32(device, 0x682068 + hoff); 277 state->vsynce = (data & 0xffff0000) >> 16; 278 state->hsynce = (data & 0x0000ffff); 279 data = nvkm_rd32(device, 0x68206c + hoff); 280 state->vblanke = (data & 0xffff0000) >> 16; 281 state->hblanke = (data & 0x0000ffff); 282 data = nvkm_rd32(device, 0x682070 + hoff); 283 state->vblanks = (data & 0xffff0000) >> 16; 284 state->hblanks = (data & 0x0000ffff); 285 state->hz = nvkm_rd32(device, 0x68200c + hoff); 286 287 data = nvkm_rd32(device, 0x682004 + hoff); 288 switch ((data & 0x000000f0) >> 4) { 289 case 5: state->or.depth = 30; break; 290 case 4: state->or.depth = 24; break; 291 case 1: state->or.depth = 18; break; 292 default: 293 state->or.depth = 18; 294 WARN_ON(1); 295 break; 296 } 297 } 298 299 static const struct nvkm_head_func 300 gv100_head = { 301 .state = gv100_head_state, 302 .rgpos = gv100_head_rgpos, 303 .rgclk = gf119_head_rgclk, 304 .vblank_get = gv100_head_vblank_get, 305 .vblank_put = gv100_head_vblank_put, 306 }; 307 308 int 309 gv100_head_new(struct nvkm_disp *disp, int id) 310 { 311 struct nvkm_device *device = disp->engine.subdev.device; 312 313 if (!(nvkm_rd32(device, 0x610060) & (0x00000001 << id))) 314 return 0; 315 316 return nvkm_head_new_(&gv100_head, disp, id); 317 } 318 319 int 320 gv100_head_cnt(struct nvkm_disp *disp, unsigned long *pmask) 321 { 322 struct nvkm_device *device = disp->engine.subdev.device; 323 324 *pmask = nvkm_rd32(device, 0x610060) & 0x000000ff; 325 return nvkm_rd32(device, 0x610074) & 0x0000000f; 326 } 327 328 const struct nvkm_event_func 329 gv100_disp_chan_uevent = { 330 }; 331 332 u64 333 gv100_disp_chan_user(struct nvkm_disp_chan *chan, u64 *psize) 334 { 335 *psize = 0x1000; 336 return 0x690000 + ((chan->chid.user - 1) * 0x1000); 337 } 338 339 static int 340 gv100_disp_dmac_idle(struct nvkm_disp_chan *chan) 341 { 342 struct nvkm_device *device = chan->disp->engine.subdev.device; 343 const u32 soff = (chan->chid.ctrl - 1) * 0x04; 344 nvkm_msec(device, 2000, 345 u32 stat = nvkm_rd32(device, 0x610664 + soff); 346 if ((stat & 0x000f0000) == 0x00040000) 347 return 0; 348 ); 349 return -EBUSY; 350 } 351 352 int 353 gv100_disp_dmac_bind(struct nvkm_disp_chan *chan, 354 struct nvkm_object *object, u32 handle) 355 { 356 return nvkm_ramht_insert(chan->disp->ramht, object, chan->chid.user, -9, handle, 357 chan->chid.user << 25 | 0x00000040); 358 } 359 360 void 361 gv100_disp_dmac_fini(struct nvkm_disp_chan *chan) 362 { 363 struct nvkm_device *device = chan->disp->engine.subdev.device; 364 const u32 uoff = (chan->chid.ctrl - 1) * 0x1000; 365 const u32 coff = chan->chid.ctrl * 0x04; 366 nvkm_mask(device, 0x6104e0 + coff, 0x00000010, 0x00000000); 367 gv100_disp_dmac_idle(chan); 368 nvkm_mask(device, 0x6104e0 + coff, 0x00000002, 0x00000000); 369 chan->suspend_put = nvkm_rd32(device, 0x690000 + uoff); 370 } 371 372 int 373 gv100_disp_dmac_init(struct nvkm_disp_chan *chan) 374 { 375 struct nvkm_subdev *subdev = &chan->disp->engine.subdev; 376 struct nvkm_device *device = subdev->device; 377 const u32 uoff = (chan->chid.ctrl - 1) * 0x1000; 378 const u32 poff = chan->chid.ctrl * 0x10; 379 const u32 coff = chan->chid.ctrl * 0x04; 380 381 nvkm_wr32(device, 0x610b24 + poff, lower_32_bits(chan->push)); 382 nvkm_wr32(device, 0x610b20 + poff, upper_32_bits(chan->push)); 383 nvkm_wr32(device, 0x610b28 + poff, 0x00000001); 384 nvkm_wr32(device, 0x610b2c + poff, 0x00000040); 385 386 nvkm_mask(device, 0x6104e0 + coff, 0x00000010, 0x00000010); 387 nvkm_wr32(device, 0x690000 + uoff, chan->suspend_put); 388 nvkm_wr32(device, 0x6104e0 + coff, 0x00000013); 389 return gv100_disp_dmac_idle(chan); 390 } 391 392 static void 393 gv100_disp_wimm_intr(struct nvkm_disp_chan *chan, bool en) 394 { 395 struct nvkm_device *device = chan->disp->engine.subdev.device; 396 const u32 mask = 0x00000001 << chan->head; 397 const u32 data = en ? mask : 0; 398 nvkm_mask(device, 0x611da8, mask, data); 399 } 400 401 static const struct nvkm_disp_chan_func 402 gv100_disp_wimm_func = { 403 .push = nv50_disp_dmac_push, 404 .init = gv100_disp_dmac_init, 405 .fini = gv100_disp_dmac_fini, 406 .intr = gv100_disp_wimm_intr, 407 .user = gv100_disp_chan_user, 408 }; 409 410 const struct nvkm_disp_chan_user 411 gv100_disp_wimm = { 412 .func = &gv100_disp_wimm_func, 413 .ctrl = 33, 414 .user = 33, 415 }; 416 417 static const struct nvkm_disp_mthd_list 418 gv100_disp_wndw_mthd_base = { 419 .mthd = 0x0000, 420 .addr = 0x000000, 421 .data = { 422 { 0x0200, 0x690200 }, 423 { 0x020c, 0x69020c }, 424 { 0x0210, 0x690210 }, 425 { 0x0214, 0x690214 }, 426 { 0x0218, 0x690218 }, 427 { 0x021c, 0x69021c }, 428 { 0x0220, 0x690220 }, 429 { 0x0224, 0x690224 }, 430 { 0x0228, 0x690228 }, 431 { 0x022c, 0x69022c }, 432 { 0x0230, 0x690230 }, 433 { 0x0234, 0x690234 }, 434 { 0x0238, 0x690238 }, 435 { 0x0240, 0x690240 }, 436 { 0x0244, 0x690244 }, 437 { 0x0248, 0x690248 }, 438 { 0x024c, 0x69024c }, 439 { 0x0250, 0x690250 }, 440 { 0x0254, 0x690254 }, 441 { 0x0260, 0x690260 }, 442 { 0x0264, 0x690264 }, 443 { 0x0268, 0x690268 }, 444 { 0x026c, 0x69026c }, 445 { 0x0270, 0x690270 }, 446 { 0x0274, 0x690274 }, 447 { 0x0280, 0x690280 }, 448 { 0x0284, 0x690284 }, 449 { 0x0288, 0x690288 }, 450 { 0x028c, 0x69028c }, 451 { 0x0290, 0x690290 }, 452 { 0x0298, 0x690298 }, 453 { 0x029c, 0x69029c }, 454 { 0x02a0, 0x6902a0 }, 455 { 0x02a4, 0x6902a4 }, 456 { 0x02a8, 0x6902a8 }, 457 { 0x02ac, 0x6902ac }, 458 { 0x02b0, 0x6902b0 }, 459 { 0x02b4, 0x6902b4 }, 460 { 0x02b8, 0x6902b8 }, 461 { 0x02bc, 0x6902bc }, 462 { 0x02c0, 0x6902c0 }, 463 { 0x02c4, 0x6902c4 }, 464 { 0x02c8, 0x6902c8 }, 465 { 0x02cc, 0x6902cc }, 466 { 0x02d0, 0x6902d0 }, 467 { 0x02d4, 0x6902d4 }, 468 { 0x02d8, 0x6902d8 }, 469 { 0x02dc, 0x6902dc }, 470 { 0x02e0, 0x6902e0 }, 471 { 0x02e4, 0x6902e4 }, 472 { 0x02e8, 0x6902e8 }, 473 { 0x02ec, 0x6902ec }, 474 { 0x02f0, 0x6902f0 }, 475 { 0x02f4, 0x6902f4 }, 476 { 0x02f8, 0x6902f8 }, 477 { 0x02fc, 0x6902fc }, 478 { 0x0300, 0x690300 }, 479 { 0x0304, 0x690304 }, 480 { 0x0308, 0x690308 }, 481 { 0x0310, 0x690310 }, 482 { 0x0314, 0x690314 }, 483 { 0x0318, 0x690318 }, 484 { 0x031c, 0x69031c }, 485 { 0x0320, 0x690320 }, 486 { 0x0324, 0x690324 }, 487 { 0x0328, 0x690328 }, 488 { 0x032c, 0x69032c }, 489 { 0x033c, 0x69033c }, 490 { 0x0340, 0x690340 }, 491 { 0x0344, 0x690344 }, 492 { 0x0348, 0x690348 }, 493 { 0x034c, 0x69034c }, 494 { 0x0350, 0x690350 }, 495 { 0x0354, 0x690354 }, 496 { 0x0358, 0x690358 }, 497 { 0x0364, 0x690364 }, 498 { 0x0368, 0x690368 }, 499 { 0x036c, 0x69036c }, 500 { 0x0370, 0x690370 }, 501 { 0x0374, 0x690374 }, 502 { 0x0380, 0x690380 }, 503 {} 504 } 505 }; 506 507 static const struct nvkm_disp_chan_mthd 508 gv100_disp_wndw_mthd = { 509 .name = "Window", 510 .addr = 0x001000, 511 .prev = 0x000800, 512 .data = { 513 { "Global", 1, &gv100_disp_wndw_mthd_base }, 514 {} 515 } 516 }; 517 518 static void 519 gv100_disp_wndw_intr(struct nvkm_disp_chan *chan, bool en) 520 { 521 struct nvkm_device *device = chan->disp->engine.subdev.device; 522 const u32 mask = 0x00000001 << chan->head; 523 const u32 data = en ? mask : 0; 524 nvkm_mask(device, 0x611da4, mask, data); 525 } 526 527 static const struct nvkm_disp_chan_func 528 gv100_disp_wndw_func = { 529 .push = nv50_disp_dmac_push, 530 .init = gv100_disp_dmac_init, 531 .fini = gv100_disp_dmac_fini, 532 .intr = gv100_disp_wndw_intr, 533 .user = gv100_disp_chan_user, 534 .bind = gv100_disp_dmac_bind, 535 }; 536 537 const struct nvkm_disp_chan_user 538 gv100_disp_wndw = { 539 .func = &gv100_disp_wndw_func, 540 .ctrl = 1, 541 .user = 1, 542 .mthd = &gv100_disp_wndw_mthd, 543 }; 544 545 int 546 gv100_disp_wndw_cnt(struct nvkm_disp *disp, unsigned long *pmask) 547 { 548 struct nvkm_device *device = disp->engine.subdev.device; 549 550 *pmask = nvkm_rd32(device, 0x610064); 551 return (nvkm_rd32(device, 0x610074) & 0x03f00000) >> 20; 552 } 553 554 static int 555 gv100_disp_curs_idle(struct nvkm_disp_chan *chan) 556 { 557 struct nvkm_device *device = chan->disp->engine.subdev.device; 558 const u32 soff = (chan->chid.ctrl - 1) * 0x04; 559 nvkm_msec(device, 2000, 560 u32 stat = nvkm_rd32(device, 0x610664 + soff); 561 if ((stat & 0x00070000) == 0x00040000) 562 return 0; 563 ); 564 return -EBUSY; 565 } 566 567 static void 568 gv100_disp_curs_intr(struct nvkm_disp_chan *chan, bool en) 569 { 570 struct nvkm_device *device = chan->disp->engine.subdev.device; 571 const u32 mask = 0x00010000 << chan->head; 572 const u32 data = en ? mask : 0; 573 nvkm_mask(device, 0x611dac, mask, data); 574 } 575 576 static void 577 gv100_disp_curs_fini(struct nvkm_disp_chan *chan) 578 { 579 struct nvkm_device *device = chan->disp->engine.subdev.device; 580 const u32 hoff = chan->chid.ctrl * 4; 581 nvkm_mask(device, 0x6104e0 + hoff, 0x00000010, 0x00000010); 582 gv100_disp_curs_idle(chan); 583 nvkm_mask(device, 0x6104e0 + hoff, 0x00000001, 0x00000000); 584 } 585 586 static int 587 gv100_disp_curs_init(struct nvkm_disp_chan *chan) 588 { 589 struct nvkm_subdev *subdev = &chan->disp->engine.subdev; 590 struct nvkm_device *device = subdev->device; 591 nvkm_wr32(device, 0x6104e0 + chan->chid.ctrl * 4, 0x00000001); 592 return gv100_disp_curs_idle(chan); 593 } 594 595 static const struct nvkm_disp_chan_func 596 gv100_disp_curs_func = { 597 .init = gv100_disp_curs_init, 598 .fini = gv100_disp_curs_fini, 599 .intr = gv100_disp_curs_intr, 600 .user = gv100_disp_chan_user, 601 }; 602 603 const struct nvkm_disp_chan_user 604 gv100_disp_curs = { 605 .func = &gv100_disp_curs_func, 606 .ctrl = 73, 607 .user = 73, 608 }; 609 610 static const struct nvkm_disp_mthd_list 611 gv100_disp_core_mthd_base = { 612 .mthd = 0x0000, 613 .addr = 0x000000, 614 .data = { 615 { 0x0200, 0x680200 }, 616 { 0x0208, 0x680208 }, 617 { 0x020c, 0x68020c }, 618 { 0x0210, 0x680210 }, 619 { 0x0214, 0x680214 }, 620 { 0x0218, 0x680218 }, 621 { 0x021c, 0x68021c }, 622 {} 623 } 624 }; 625 626 static const struct nvkm_disp_mthd_list 627 gv100_disp_core_mthd_sor = { 628 .mthd = 0x0020, 629 .addr = 0x000020, 630 .data = { 631 { 0x0300, 0x680300 }, 632 { 0x0304, 0x680304 }, 633 { 0x0308, 0x680308 }, 634 { 0x030c, 0x68030c }, 635 {} 636 } 637 }; 638 639 static const struct nvkm_disp_mthd_list 640 gv100_disp_core_mthd_wndw = { 641 .mthd = 0x0080, 642 .addr = 0x000080, 643 .data = { 644 { 0x1000, 0x681000 }, 645 { 0x1004, 0x681004 }, 646 { 0x1008, 0x681008 }, 647 { 0x100c, 0x68100c }, 648 { 0x1010, 0x681010 }, 649 {} 650 } 651 }; 652 653 static const struct nvkm_disp_mthd_list 654 gv100_disp_core_mthd_head = { 655 .mthd = 0x0400, 656 .addr = 0x000400, 657 .data = { 658 { 0x2000, 0x682000 }, 659 { 0x2004, 0x682004 }, 660 { 0x2008, 0x682008 }, 661 { 0x200c, 0x68200c }, 662 { 0x2014, 0x682014 }, 663 { 0x2018, 0x682018 }, 664 { 0x201c, 0x68201c }, 665 { 0x2020, 0x682020 }, 666 { 0x2028, 0x682028 }, 667 { 0x202c, 0x68202c }, 668 { 0x2030, 0x682030 }, 669 { 0x2038, 0x682038 }, 670 { 0x203c, 0x68203c }, 671 { 0x2048, 0x682048 }, 672 { 0x204c, 0x68204c }, 673 { 0x2050, 0x682050 }, 674 { 0x2054, 0x682054 }, 675 { 0x2058, 0x682058 }, 676 { 0x205c, 0x68205c }, 677 { 0x2060, 0x682060 }, 678 { 0x2064, 0x682064 }, 679 { 0x2068, 0x682068 }, 680 { 0x206c, 0x68206c }, 681 { 0x2070, 0x682070 }, 682 { 0x2074, 0x682074 }, 683 { 0x2078, 0x682078 }, 684 { 0x207c, 0x68207c }, 685 { 0x2080, 0x682080 }, 686 { 0x2088, 0x682088 }, 687 { 0x2090, 0x682090 }, 688 { 0x209c, 0x68209c }, 689 { 0x20a0, 0x6820a0 }, 690 { 0x20a4, 0x6820a4 }, 691 { 0x20a8, 0x6820a8 }, 692 { 0x20ac, 0x6820ac }, 693 { 0x2180, 0x682180 }, 694 { 0x2184, 0x682184 }, 695 { 0x218c, 0x68218c }, 696 { 0x2194, 0x682194 }, 697 { 0x2198, 0x682198 }, 698 { 0x219c, 0x68219c }, 699 { 0x21a0, 0x6821a0 }, 700 { 0x21a4, 0x6821a4 }, 701 { 0x2214, 0x682214 }, 702 { 0x2218, 0x682218 }, 703 {} 704 } 705 }; 706 707 static const struct nvkm_disp_chan_mthd 708 gv100_disp_core_mthd = { 709 .name = "Core", 710 .addr = 0x000000, 711 .prev = 0x008000, 712 .data = { 713 { "Global", 1, &gv100_disp_core_mthd_base }, 714 { "SOR", 4, &gv100_disp_core_mthd_sor }, 715 { "WINDOW", 8, &gv100_disp_core_mthd_wndw }, 716 { "HEAD", 4, &gv100_disp_core_mthd_head }, 717 {} 718 } 719 }; 720 721 static int 722 gv100_disp_core_idle(struct nvkm_disp_chan *chan) 723 { 724 struct nvkm_device *device = chan->disp->engine.subdev.device; 725 nvkm_msec(device, 2000, 726 u32 stat = nvkm_rd32(device, 0x610630); 727 if ((stat & 0x001f0000) == 0x000b0000) 728 return 0; 729 ); 730 return -EBUSY; 731 } 732 733 static u64 734 gv100_disp_core_user(struct nvkm_disp_chan *chan, u64 *psize) 735 { 736 *psize = 0x10000; 737 return 0x680000; 738 } 739 740 static void 741 gv100_disp_core_intr(struct nvkm_disp_chan *chan, bool en) 742 { 743 struct nvkm_device *device = chan->disp->engine.subdev.device; 744 const u32 mask = 0x00000001; 745 const u32 data = en ? mask : 0; 746 nvkm_mask(device, 0x611dac, mask, data); 747 } 748 749 static void 750 gv100_disp_core_fini(struct nvkm_disp_chan *chan) 751 { 752 struct nvkm_device *device = chan->disp->engine.subdev.device; 753 nvkm_mask(device, 0x6104e0, 0x00000010, 0x00000000); 754 gv100_disp_core_idle(chan); 755 nvkm_mask(device, 0x6104e0, 0x00000002, 0x00000000); 756 chan->suspend_put = nvkm_rd32(device, 0x680000); 757 } 758 759 static int 760 gv100_disp_core_init(struct nvkm_disp_chan *chan) 761 { 762 struct nvkm_subdev *subdev = &chan->disp->engine.subdev; 763 struct nvkm_device *device = subdev->device; 764 765 nvkm_wr32(device, 0x610b24, lower_32_bits(chan->push)); 766 nvkm_wr32(device, 0x610b20, upper_32_bits(chan->push)); 767 nvkm_wr32(device, 0x610b28, 0x00000001); 768 nvkm_wr32(device, 0x610b2c, 0x00000040); 769 770 nvkm_mask(device, 0x6104e0, 0x00000010, 0x00000010); 771 nvkm_wr32(device, 0x680000, chan->suspend_put); 772 nvkm_wr32(device, 0x6104e0, 0x00000013); 773 return gv100_disp_core_idle(chan); 774 } 775 776 static const struct nvkm_disp_chan_func 777 gv100_disp_core_func = { 778 .push = nv50_disp_dmac_push, 779 .init = gv100_disp_core_init, 780 .fini = gv100_disp_core_fini, 781 .intr = gv100_disp_core_intr, 782 .user = gv100_disp_core_user, 783 .bind = gv100_disp_dmac_bind, 784 }; 785 786 const struct nvkm_disp_chan_user 787 gv100_disp_core = { 788 .func = &gv100_disp_core_func, 789 .ctrl = 0, 790 .user = 0, 791 .mthd = &gv100_disp_core_mthd, 792 }; 793 794 #define gv100_disp_caps(p) container_of((p), struct gv100_disp_caps, object) 795 796 struct gv100_disp_caps { 797 struct nvkm_object object; 798 struct nvkm_disp *disp; 799 }; 800 801 static int 802 gv100_disp_caps_map(struct nvkm_object *object, void *argv, u32 argc, 803 enum nvkm_object_map *type, u64 *addr, u64 *size) 804 { 805 struct gv100_disp_caps *caps = gv100_disp_caps(object); 806 struct nvkm_device *device = caps->disp->engine.subdev.device; 807 *type = NVKM_OBJECT_MAP_IO; 808 *addr = 0x640000 + device->func->resource_addr(device, 0); 809 *size = 0x1000; 810 return 0; 811 } 812 813 static const struct nvkm_object_func 814 gv100_disp_caps = { 815 .map = gv100_disp_caps_map, 816 }; 817 818 int 819 gv100_disp_caps_new(const struct nvkm_oclass *oclass, void *argv, u32 argc, 820 struct nvkm_object **pobject) 821 { 822 struct nvkm_disp *disp = nvkm_udisp(oclass->parent); 823 struct gv100_disp_caps *caps; 824 825 if (!(caps = kzalloc(sizeof(*caps), GFP_KERNEL))) 826 return -ENOMEM; 827 *pobject = &caps->object; 828 829 nvkm_object_ctor(&gv100_disp_caps, oclass, &caps->object); 830 caps->disp = disp; 831 return 0; 832 } 833 834 void 835 gv100_disp_super(struct work_struct *work) 836 { 837 struct nvkm_disp *disp = container_of(work, struct nvkm_disp, super.work); 838 struct nvkm_subdev *subdev = &disp->engine.subdev; 839 struct nvkm_device *device = subdev->device; 840 struct nvkm_head *head; 841 u32 stat, mask[4]; 842 843 mutex_lock(&disp->super.mutex); 844 stat = nvkm_rd32(device, 0x6107a8); 845 846 nvkm_debug(subdev, "supervisor %d: %08x\n", ffs(disp->super.pending), stat); 847 list_for_each_entry(head, &disp->heads, head) { 848 mask[head->id] = nvkm_rd32(device, 0x6107ac + (head->id * 4)); 849 HEAD_DBG(head, "%08x", mask[head->id]); 850 } 851 852 if (disp->super.pending & 0x00000001) { 853 nv50_disp_chan_mthd(disp->chan[0], NV_DBG_DEBUG); 854 nv50_disp_super_1(disp); 855 list_for_each_entry(head, &disp->heads, head) { 856 if (!(mask[head->id] & 0x00001000)) 857 continue; 858 nv50_disp_super_1_0(disp, head); 859 } 860 } else 861 if (disp->super.pending & 0x00000002) { 862 list_for_each_entry(head, &disp->heads, head) { 863 if (!(mask[head->id] & 0x00001000)) 864 continue; 865 nv50_disp_super_2_0(disp, head); 866 } 867 list_for_each_entry(head, &disp->heads, head) { 868 if (!(mask[head->id] & 0x00010000)) 869 continue; 870 nv50_disp_super_2_1(disp, head); 871 } 872 list_for_each_entry(head, &disp->heads, head) { 873 if (!(mask[head->id] & 0x00001000)) 874 continue; 875 nv50_disp_super_2_2(disp, head); 876 } 877 } else 878 if (disp->super.pending & 0x00000004) { 879 list_for_each_entry(head, &disp->heads, head) { 880 if (!(mask[head->id] & 0x00001000)) 881 continue; 882 nv50_disp_super_3_0(disp, head); 883 } 884 } 885 886 list_for_each_entry(head, &disp->heads, head) 887 nvkm_wr32(device, 0x6107ac + (head->id * 4), 0x00000000); 888 889 nvkm_wr32(device, 0x6107a8, 0x80000000); 890 mutex_unlock(&disp->super.mutex); 891 } 892 893 static void 894 gv100_disp_exception(struct nvkm_disp *disp, int chid) 895 { 896 struct nvkm_subdev *subdev = &disp->engine.subdev; 897 struct nvkm_device *device = subdev->device; 898 u32 stat = nvkm_rd32(device, 0x611020 + (chid * 12)); 899 u32 type = (stat & 0x00007000) >> 12; 900 u32 mthd = (stat & 0x00000fff) << 2; 901 const struct nvkm_enum *reason = 902 nvkm_enum_find(nv50_disp_intr_error_type, type); 903 904 /*TODO: Suspect 33->41 are for WRBK channel exceptions, but we 905 * don't support those currently. 906 * 907 * CORE+WIN CHIDs map directly to the FE_EXCEPT() slots. 908 */ 909 if (chid <= 32) { 910 u32 data = nvkm_rd32(device, 0x611024 + (chid * 12)); 911 u32 code = nvkm_rd32(device, 0x611028 + (chid * 12)); 912 nvkm_error(subdev, "chid %d stat %08x reason %d [%s] " 913 "mthd %04x data %08x code %08x\n", 914 chid, stat, type, reason ? reason->name : "", 915 mthd, data, code); 916 } else { 917 nvkm_error(subdev, "chid %d stat %08x reason %d [%s] " 918 "mthd %04x\n", 919 chid, stat, type, reason ? reason->name : "", mthd); 920 } 921 922 if (chid < ARRAY_SIZE(disp->chan) && disp->chan[chid]) { 923 switch (mthd) { 924 case 0x0200: 925 nv50_disp_chan_mthd(disp->chan[chid], NV_DBG_ERROR); 926 break; 927 default: 928 break; 929 } 930 } 931 932 nvkm_wr32(device, 0x611020 + (chid * 12), 0x90000000); 933 } 934 935 static void 936 gv100_disp_intr_ctrl_disp(struct nvkm_disp *disp) 937 { 938 struct nvkm_subdev *subdev = &disp->engine.subdev; 939 struct nvkm_device *device = subdev->device; 940 u32 stat = nvkm_rd32(device, 0x611c30); 941 942 if (stat & 0x00000007) { 943 disp->super.pending = (stat & 0x00000007); 944 queue_work(disp->super.wq, &disp->super.work); 945 nvkm_wr32(device, 0x611860, disp->super.pending); 946 stat &= ~0x00000007; 947 } 948 949 /*TODO: I would guess this is VBIOS_RELEASE, however, NFI how to 950 * ACK it, nor does RM appear to bother. 951 */ 952 if (stat & 0x00000008) 953 stat &= ~0x00000008; 954 955 if (stat & 0x00000080) { 956 u32 error = nvkm_mask(device, 0x611848, 0x00000000, 0x00000000); 957 nvkm_warn(subdev, "error %08x\n", error); 958 stat &= ~0x00000080; 959 } 960 961 if (stat & 0x00000100) { 962 unsigned long wndws = nvkm_rd32(device, 0x611858); 963 unsigned long other = nvkm_rd32(device, 0x61185c); 964 int wndw; 965 966 nvkm_wr32(device, 0x611858, wndws); 967 nvkm_wr32(device, 0x61185c, other); 968 969 /* AWAKEN_OTHER_CORE. */ 970 if (other & 0x00000001) 971 nv50_disp_chan_uevent_send(disp, 0); 972 973 /* AWAKEN_WIN_CH(n). */ 974 for_each_set_bit(wndw, &wndws, disp->wndw.nr) { 975 nv50_disp_chan_uevent_send(disp, 1 + wndw); 976 } 977 } 978 979 if (stat) 980 nvkm_warn(subdev, "ctrl %08x\n", stat); 981 } 982 983 static void 984 gv100_disp_intr_exc_other(struct nvkm_disp *disp) 985 { 986 struct nvkm_subdev *subdev = &disp->engine.subdev; 987 struct nvkm_device *device = subdev->device; 988 u32 stat = nvkm_rd32(device, 0x611854); 989 unsigned long mask; 990 int head; 991 992 if (stat & 0x00000001) { 993 nvkm_wr32(device, 0x611854, 0x00000001); 994 gv100_disp_exception(disp, 0); 995 stat &= ~0x00000001; 996 } 997 998 if ((mask = (stat & 0x00ff0000) >> 16)) { 999 for_each_set_bit(head, &mask, disp->wndw.nr) { 1000 nvkm_wr32(device, 0x611854, 0x00010000 << head); 1001 gv100_disp_exception(disp, 73 + head); 1002 stat &= ~(0x00010000 << head); 1003 } 1004 } 1005 1006 if (stat) { 1007 nvkm_warn(subdev, "exception %08x\n", stat); 1008 nvkm_wr32(device, 0x611854, stat); 1009 } 1010 } 1011 1012 static void 1013 gv100_disp_intr_exc_winim(struct nvkm_disp *disp) 1014 { 1015 struct nvkm_subdev *subdev = &disp->engine.subdev; 1016 struct nvkm_device *device = subdev->device; 1017 unsigned long stat = nvkm_rd32(device, 0x611850); 1018 int wndw; 1019 1020 for_each_set_bit(wndw, &stat, disp->wndw.nr) { 1021 nvkm_wr32(device, 0x611850, BIT(wndw)); 1022 gv100_disp_exception(disp, 33 + wndw); 1023 stat &= ~BIT(wndw); 1024 } 1025 1026 if (stat) { 1027 nvkm_warn(subdev, "wimm %08x\n", (u32)stat); 1028 nvkm_wr32(device, 0x611850, stat); 1029 } 1030 } 1031 1032 static void 1033 gv100_disp_intr_exc_win(struct nvkm_disp *disp) 1034 { 1035 struct nvkm_subdev *subdev = &disp->engine.subdev; 1036 struct nvkm_device *device = subdev->device; 1037 unsigned long stat = nvkm_rd32(device, 0x61184c); 1038 int wndw; 1039 1040 for_each_set_bit(wndw, &stat, disp->wndw.nr) { 1041 nvkm_wr32(device, 0x61184c, BIT(wndw)); 1042 gv100_disp_exception(disp, 1 + wndw); 1043 stat &= ~BIT(wndw); 1044 } 1045 1046 if (stat) { 1047 nvkm_warn(subdev, "wndw %08x\n", (u32)stat); 1048 nvkm_wr32(device, 0x61184c, stat); 1049 } 1050 } 1051 1052 static void 1053 gv100_disp_intr_head_timing(struct nvkm_disp *disp, int head) 1054 { 1055 struct nvkm_subdev *subdev = &disp->engine.subdev; 1056 struct nvkm_device *device = subdev->device; 1057 u32 stat = nvkm_rd32(device, 0x611800 + (head * 0x04)); 1058 1059 /* LAST_DATA, LOADV. */ 1060 if (stat & 0x00000003) { 1061 nvkm_wr32(device, 0x611800 + (head * 0x04), stat & 0x00000003); 1062 stat &= ~0x00000003; 1063 } 1064 1065 if (stat & 0x00000004) { 1066 nvkm_disp_vblank(disp, head); 1067 nvkm_wr32(device, 0x611800 + (head * 0x04), 0x00000004); 1068 stat &= ~0x00000004; 1069 } 1070 1071 if (stat) { 1072 nvkm_warn(subdev, "head %08x\n", stat); 1073 nvkm_wr32(device, 0x611800 + (head * 0x04), stat); 1074 } 1075 } 1076 1077 void 1078 gv100_disp_intr(struct nvkm_disp *disp) 1079 { 1080 struct nvkm_subdev *subdev = &disp->engine.subdev; 1081 struct nvkm_device *device = subdev->device; 1082 u32 stat = nvkm_rd32(device, 0x611ec0); 1083 unsigned long mask; 1084 int head; 1085 1086 if ((mask = (stat & 0x000000ff))) { 1087 for_each_set_bit(head, &mask, 8) { 1088 gv100_disp_intr_head_timing(disp, head); 1089 stat &= ~BIT(head); 1090 } 1091 } 1092 1093 if (stat & 0x00000200) { 1094 gv100_disp_intr_exc_win(disp); 1095 stat &= ~0x00000200; 1096 } 1097 1098 if (stat & 0x00000400) { 1099 gv100_disp_intr_exc_winim(disp); 1100 stat &= ~0x00000400; 1101 } 1102 1103 if (stat & 0x00000800) { 1104 gv100_disp_intr_exc_other(disp); 1105 stat &= ~0x00000800; 1106 } 1107 1108 if (stat & 0x00001000) { 1109 gv100_disp_intr_ctrl_disp(disp); 1110 stat &= ~0x00001000; 1111 } 1112 1113 if (stat) 1114 nvkm_warn(subdev, "intr %08x\n", stat); 1115 } 1116 1117 void 1118 gv100_disp_fini(struct nvkm_disp *disp, bool suspend) 1119 { 1120 struct nvkm_device *device = disp->engine.subdev.device; 1121 nvkm_wr32(device, 0x611db0, 0x00000000); 1122 } 1123 1124 static int 1125 gv100_disp_init(struct nvkm_disp *disp) 1126 { 1127 struct nvkm_device *device = disp->engine.subdev.device; 1128 struct nvkm_head *head; 1129 int i, j; 1130 u32 tmp; 1131 1132 /* Claim ownership of display. */ 1133 if (nvkm_rd32(device, 0x6254e8) & 0x00000002) { 1134 nvkm_mask(device, 0x6254e8, 0x00000001, 0x00000000); 1135 if (nvkm_msec(device, 2000, 1136 if (!(nvkm_rd32(device, 0x6254e8) & 0x00000002)) 1137 break; 1138 ) < 0) 1139 return -EBUSY; 1140 } 1141 1142 /* Lock pin capabilities. */ 1143 tmp = nvkm_rd32(device, 0x610068); 1144 nvkm_wr32(device, 0x640008, tmp); 1145 1146 /* SOR capabilities. */ 1147 for (i = 0; i < disp->sor.nr; i++) { 1148 tmp = nvkm_rd32(device, 0x61c000 + (i * 0x800)); 1149 nvkm_mask(device, 0x640000, 0x00000100 << i, 0x00000100 << i); 1150 nvkm_wr32(device, 0x640144 + (i * 0x08), tmp); 1151 } 1152 1153 /* Head capabilities. */ 1154 list_for_each_entry(head, &disp->heads, head) { 1155 const int id = head->id; 1156 1157 /* RG. */ 1158 tmp = nvkm_rd32(device, 0x616300 + (id * 0x800)); 1159 nvkm_wr32(device, 0x640048 + (id * 0x020), tmp); 1160 1161 /* POSTCOMP. */ 1162 for (j = 0; j < 6 * 4; j += 4) { 1163 tmp = nvkm_rd32(device, 0x616100 + (id * 0x800) + j); 1164 nvkm_wr32(device, 0x640030 + (id * 0x20) + j, tmp); 1165 } 1166 } 1167 1168 /* Window capabilities. */ 1169 for (i = 0; i < disp->wndw.nr; i++) { 1170 nvkm_mask(device, 0x640004, 1 << i, 1 << i); 1171 for (j = 0; j < 6 * 4; j += 4) { 1172 tmp = nvkm_rd32(device, 0x630050 + (i * 0x800) + j); 1173 nvkm_wr32(device, 0x6401e4 + (i * 0x20) + j, tmp); 1174 } 1175 } 1176 1177 /* IHUB capabilities. */ 1178 for (i = 0; i < 4; i++) { 1179 tmp = nvkm_rd32(device, 0x62e000 + (i * 0x04)); 1180 nvkm_wr32(device, 0x640010 + (i * 0x04), tmp); 1181 } 1182 1183 nvkm_mask(device, 0x610078, 0x00000001, 0x00000001); 1184 1185 /* Setup instance memory. */ 1186 switch (nvkm_memory_target(disp->inst->memory)) { 1187 case NVKM_MEM_TARGET_VRAM: tmp = 0x00000001; break; 1188 case NVKM_MEM_TARGET_NCOH: tmp = 0x00000002; break; 1189 case NVKM_MEM_TARGET_HOST: tmp = 0x00000003; break; 1190 default: 1191 break; 1192 } 1193 nvkm_wr32(device, 0x610010, 0x00000008 | tmp); 1194 nvkm_wr32(device, 0x610014, disp->inst->addr >> 16); 1195 1196 /* CTRL_DISP: AWAKEN, ERROR, SUPERVISOR[1-3]. */ 1197 nvkm_wr32(device, 0x611cf0, 0x00000187); /* MSK. */ 1198 nvkm_wr32(device, 0x611db0, 0x00000187); /* EN. */ 1199 1200 /* EXC_OTHER: CURSn, CORE. */ 1201 nvkm_wr32(device, 0x611cec, disp->head.mask << 16 | 1202 0x00000001); /* MSK. */ 1203 nvkm_wr32(device, 0x611dac, 0x00000000); /* EN. */ 1204 1205 /* EXC_WINIM. */ 1206 nvkm_wr32(device, 0x611ce8, disp->wndw.mask); /* MSK. */ 1207 nvkm_wr32(device, 0x611da8, 0x00000000); /* EN. */ 1208 1209 /* EXC_WIN. */ 1210 nvkm_wr32(device, 0x611ce4, disp->wndw.mask); /* MSK. */ 1211 nvkm_wr32(device, 0x611da4, 0x00000000); /* EN. */ 1212 1213 /* HEAD_TIMING(n): VBLANK. */ 1214 list_for_each_entry(head, &disp->heads, head) { 1215 const u32 hoff = head->id * 4; 1216 nvkm_wr32(device, 0x611cc0 + hoff, 0x00000004); /* MSK. */ 1217 nvkm_wr32(device, 0x611d80 + hoff, 0x00000000); /* EN. */ 1218 } 1219 1220 /* OR. */ 1221 nvkm_wr32(device, 0x611cf4, 0x00000000); /* MSK. */ 1222 nvkm_wr32(device, 0x611db4, 0x00000000); /* EN. */ 1223 return 0; 1224 } 1225 1226 static const struct nvkm_disp_func 1227 gv100_disp = { 1228 .oneinit = nv50_disp_oneinit, 1229 .init = gv100_disp_init, 1230 .fini = gv100_disp_fini, 1231 .intr = gv100_disp_intr, 1232 .super = gv100_disp_super, 1233 .uevent = &gv100_disp_chan_uevent, 1234 .wndw = { .cnt = gv100_disp_wndw_cnt }, 1235 .head = { .cnt = gv100_head_cnt, .new = gv100_head_new }, 1236 .sor = { .cnt = gv100_sor_cnt, .new = gv100_sor_new }, 1237 .ramht_size = 0x2000, 1238 .root = { 0, 0,GV100_DISP }, 1239 .user = { 1240 {{-1,-1,GV100_DISP_CAPS }, gv100_disp_caps_new }, 1241 {{ 0, 0,GV100_DISP_CURSOR }, nvkm_disp_chan_new, &gv100_disp_curs }, 1242 {{ 0, 0,GV100_DISP_WINDOW_IMM_CHANNEL_DMA}, nvkm_disp_wndw_new, &gv100_disp_wimm }, 1243 {{ 0, 0,GV100_DISP_CORE_CHANNEL_DMA }, nvkm_disp_core_new, &gv100_disp_core }, 1244 {{ 0, 0,GV100_DISP_WINDOW_CHANNEL_DMA }, nvkm_disp_wndw_new, &gv100_disp_wndw }, 1245 {} 1246 }, 1247 }; 1248 1249 int 1250 gv100_disp_new(struct nvkm_device *device, enum nvkm_subdev_type type, int inst, 1251 struct nvkm_disp **pdisp) 1252 { 1253 return nvkm_disp_new_(&gv100_disp, device, type, inst, pdisp); 1254 } 1255