xref: /linux/drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c (revision de80193308f43d3ae52cd3561e8ba77cd1437311)
1 /*
2  * Copyright 2007 Matthieu CASTET <castet.matthieu@free.fr>
3  * All Rights Reserved.
4  *
5  * Permission is hereby granted, free of charge, to any person obtaining a
6  * copy of this software and associated documentation files (the "Software"),
7  * to deal in the Software without restriction, including without limitation
8  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
9  * and/or sell copies of the Software, and to permit persons to whom the
10  * Software is furnished to do so, subject to the following conditions:
11  *
12  * The above copyright notice and this permission notice (including the next
13  * paragr) shall be included in all copies or substantial portions of the
14  * Software.
15  *
16  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
19  * PRECISION INSIGHT AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
20  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
21  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
22  * DEALINGS IN THE SOFTWARE.
23  */
24 #include "nv10.h"
25 #include "regs.h"
26 
27 #include <core/client.h>
28 #include <core/gpuobj.h>
29 #include <engine/fifo.h>
30 #include <engine/fifo/chan.h>
31 #include <subdev/fb.h>
32 
33 struct pipe_state {
34 	u32 pipe_0x0000[0x040/4];
35 	u32 pipe_0x0040[0x010/4];
36 	u32 pipe_0x0200[0x0c0/4];
37 	u32 pipe_0x4400[0x080/4];
38 	u32 pipe_0x6400[0x3b0/4];
39 	u32 pipe_0x6800[0x2f0/4];
40 	u32 pipe_0x6c00[0x030/4];
41 	u32 pipe_0x7000[0x130/4];
42 	u32 pipe_0x7400[0x0c0/4];
43 	u32 pipe_0x7800[0x0c0/4];
44 };
45 
46 static int nv10_gr_ctx_regs[] = {
47 	NV10_PGRAPH_CTX_SWITCH(0),
48 	NV10_PGRAPH_CTX_SWITCH(1),
49 	NV10_PGRAPH_CTX_SWITCH(2),
50 	NV10_PGRAPH_CTX_SWITCH(3),
51 	NV10_PGRAPH_CTX_SWITCH(4),
52 	NV10_PGRAPH_CTX_CACHE(0, 0),
53 	NV10_PGRAPH_CTX_CACHE(0, 1),
54 	NV10_PGRAPH_CTX_CACHE(0, 2),
55 	NV10_PGRAPH_CTX_CACHE(0, 3),
56 	NV10_PGRAPH_CTX_CACHE(0, 4),
57 	NV10_PGRAPH_CTX_CACHE(1, 0),
58 	NV10_PGRAPH_CTX_CACHE(1, 1),
59 	NV10_PGRAPH_CTX_CACHE(1, 2),
60 	NV10_PGRAPH_CTX_CACHE(1, 3),
61 	NV10_PGRAPH_CTX_CACHE(1, 4),
62 	NV10_PGRAPH_CTX_CACHE(2, 0),
63 	NV10_PGRAPH_CTX_CACHE(2, 1),
64 	NV10_PGRAPH_CTX_CACHE(2, 2),
65 	NV10_PGRAPH_CTX_CACHE(2, 3),
66 	NV10_PGRAPH_CTX_CACHE(2, 4),
67 	NV10_PGRAPH_CTX_CACHE(3, 0),
68 	NV10_PGRAPH_CTX_CACHE(3, 1),
69 	NV10_PGRAPH_CTX_CACHE(3, 2),
70 	NV10_PGRAPH_CTX_CACHE(3, 3),
71 	NV10_PGRAPH_CTX_CACHE(3, 4),
72 	NV10_PGRAPH_CTX_CACHE(4, 0),
73 	NV10_PGRAPH_CTX_CACHE(4, 1),
74 	NV10_PGRAPH_CTX_CACHE(4, 2),
75 	NV10_PGRAPH_CTX_CACHE(4, 3),
76 	NV10_PGRAPH_CTX_CACHE(4, 4),
77 	NV10_PGRAPH_CTX_CACHE(5, 0),
78 	NV10_PGRAPH_CTX_CACHE(5, 1),
79 	NV10_PGRAPH_CTX_CACHE(5, 2),
80 	NV10_PGRAPH_CTX_CACHE(5, 3),
81 	NV10_PGRAPH_CTX_CACHE(5, 4),
82 	NV10_PGRAPH_CTX_CACHE(6, 0),
83 	NV10_PGRAPH_CTX_CACHE(6, 1),
84 	NV10_PGRAPH_CTX_CACHE(6, 2),
85 	NV10_PGRAPH_CTX_CACHE(6, 3),
86 	NV10_PGRAPH_CTX_CACHE(6, 4),
87 	NV10_PGRAPH_CTX_CACHE(7, 0),
88 	NV10_PGRAPH_CTX_CACHE(7, 1),
89 	NV10_PGRAPH_CTX_CACHE(7, 2),
90 	NV10_PGRAPH_CTX_CACHE(7, 3),
91 	NV10_PGRAPH_CTX_CACHE(7, 4),
92 	NV10_PGRAPH_CTX_USER,
93 	NV04_PGRAPH_DMA_START_0,
94 	NV04_PGRAPH_DMA_START_1,
95 	NV04_PGRAPH_DMA_LENGTH,
96 	NV04_PGRAPH_DMA_MISC,
97 	NV10_PGRAPH_DMA_PITCH,
98 	NV04_PGRAPH_BOFFSET0,
99 	NV04_PGRAPH_BBASE0,
100 	NV04_PGRAPH_BLIMIT0,
101 	NV04_PGRAPH_BOFFSET1,
102 	NV04_PGRAPH_BBASE1,
103 	NV04_PGRAPH_BLIMIT1,
104 	NV04_PGRAPH_BOFFSET2,
105 	NV04_PGRAPH_BBASE2,
106 	NV04_PGRAPH_BLIMIT2,
107 	NV04_PGRAPH_BOFFSET3,
108 	NV04_PGRAPH_BBASE3,
109 	NV04_PGRAPH_BLIMIT3,
110 	NV04_PGRAPH_BOFFSET4,
111 	NV04_PGRAPH_BBASE4,
112 	NV04_PGRAPH_BLIMIT4,
113 	NV04_PGRAPH_BOFFSET5,
114 	NV04_PGRAPH_BBASE5,
115 	NV04_PGRAPH_BLIMIT5,
116 	NV04_PGRAPH_BPITCH0,
117 	NV04_PGRAPH_BPITCH1,
118 	NV04_PGRAPH_BPITCH2,
119 	NV04_PGRAPH_BPITCH3,
120 	NV04_PGRAPH_BPITCH4,
121 	NV10_PGRAPH_SURFACE,
122 	NV10_PGRAPH_STATE,
123 	NV04_PGRAPH_BSWIZZLE2,
124 	NV04_PGRAPH_BSWIZZLE5,
125 	NV04_PGRAPH_BPIXEL,
126 	NV10_PGRAPH_NOTIFY,
127 	NV04_PGRAPH_PATT_COLOR0,
128 	NV04_PGRAPH_PATT_COLOR1,
129 	NV04_PGRAPH_PATT_COLORRAM, /* 64 values from 0x400900 to 0x4009fc */
130 	0x00400904,
131 	0x00400908,
132 	0x0040090c,
133 	0x00400910,
134 	0x00400914,
135 	0x00400918,
136 	0x0040091c,
137 	0x00400920,
138 	0x00400924,
139 	0x00400928,
140 	0x0040092c,
141 	0x00400930,
142 	0x00400934,
143 	0x00400938,
144 	0x0040093c,
145 	0x00400940,
146 	0x00400944,
147 	0x00400948,
148 	0x0040094c,
149 	0x00400950,
150 	0x00400954,
151 	0x00400958,
152 	0x0040095c,
153 	0x00400960,
154 	0x00400964,
155 	0x00400968,
156 	0x0040096c,
157 	0x00400970,
158 	0x00400974,
159 	0x00400978,
160 	0x0040097c,
161 	0x00400980,
162 	0x00400984,
163 	0x00400988,
164 	0x0040098c,
165 	0x00400990,
166 	0x00400994,
167 	0x00400998,
168 	0x0040099c,
169 	0x004009a0,
170 	0x004009a4,
171 	0x004009a8,
172 	0x004009ac,
173 	0x004009b0,
174 	0x004009b4,
175 	0x004009b8,
176 	0x004009bc,
177 	0x004009c0,
178 	0x004009c4,
179 	0x004009c8,
180 	0x004009cc,
181 	0x004009d0,
182 	0x004009d4,
183 	0x004009d8,
184 	0x004009dc,
185 	0x004009e0,
186 	0x004009e4,
187 	0x004009e8,
188 	0x004009ec,
189 	0x004009f0,
190 	0x004009f4,
191 	0x004009f8,
192 	0x004009fc,
193 	NV04_PGRAPH_PATTERN,	/* 2 values from 0x400808 to 0x40080c */
194 	0x0040080c,
195 	NV04_PGRAPH_PATTERN_SHAPE,
196 	NV03_PGRAPH_MONO_COLOR0,
197 	NV04_PGRAPH_ROP3,
198 	NV04_PGRAPH_CHROMA,
199 	NV04_PGRAPH_BETA_AND,
200 	NV04_PGRAPH_BETA_PREMULT,
201 	0x00400e70,
202 	0x00400e74,
203 	0x00400e78,
204 	0x00400e7c,
205 	0x00400e80,
206 	0x00400e84,
207 	0x00400e88,
208 	0x00400e8c,
209 	0x00400ea0,
210 	0x00400ea4,
211 	0x00400ea8,
212 	0x00400e90,
213 	0x00400e94,
214 	0x00400e98,
215 	0x00400e9c,
216 	NV10_PGRAPH_WINDOWCLIP_HORIZONTAL, /* 8 values from 0x400f00-0x400f1c */
217 	NV10_PGRAPH_WINDOWCLIP_VERTICAL,   /* 8 values from 0x400f20-0x400f3c */
218 	0x00400f04,
219 	0x00400f24,
220 	0x00400f08,
221 	0x00400f28,
222 	0x00400f0c,
223 	0x00400f2c,
224 	0x00400f10,
225 	0x00400f30,
226 	0x00400f14,
227 	0x00400f34,
228 	0x00400f18,
229 	0x00400f38,
230 	0x00400f1c,
231 	0x00400f3c,
232 	NV10_PGRAPH_XFMODE0,
233 	NV10_PGRAPH_XFMODE1,
234 	NV10_PGRAPH_GLOBALSTATE0,
235 	NV10_PGRAPH_GLOBALSTATE1,
236 	NV04_PGRAPH_STORED_FMT,
237 	NV04_PGRAPH_SOURCE_COLOR,
238 	NV03_PGRAPH_ABS_X_RAM,	/* 32 values from 0x400400 to 0x40047c */
239 	NV03_PGRAPH_ABS_Y_RAM,	/* 32 values from 0x400480 to 0x4004fc */
240 	0x00400404,
241 	0x00400484,
242 	0x00400408,
243 	0x00400488,
244 	0x0040040c,
245 	0x0040048c,
246 	0x00400410,
247 	0x00400490,
248 	0x00400414,
249 	0x00400494,
250 	0x00400418,
251 	0x00400498,
252 	0x0040041c,
253 	0x0040049c,
254 	0x00400420,
255 	0x004004a0,
256 	0x00400424,
257 	0x004004a4,
258 	0x00400428,
259 	0x004004a8,
260 	0x0040042c,
261 	0x004004ac,
262 	0x00400430,
263 	0x004004b0,
264 	0x00400434,
265 	0x004004b4,
266 	0x00400438,
267 	0x004004b8,
268 	0x0040043c,
269 	0x004004bc,
270 	0x00400440,
271 	0x004004c0,
272 	0x00400444,
273 	0x004004c4,
274 	0x00400448,
275 	0x004004c8,
276 	0x0040044c,
277 	0x004004cc,
278 	0x00400450,
279 	0x004004d0,
280 	0x00400454,
281 	0x004004d4,
282 	0x00400458,
283 	0x004004d8,
284 	0x0040045c,
285 	0x004004dc,
286 	0x00400460,
287 	0x004004e0,
288 	0x00400464,
289 	0x004004e4,
290 	0x00400468,
291 	0x004004e8,
292 	0x0040046c,
293 	0x004004ec,
294 	0x00400470,
295 	0x004004f0,
296 	0x00400474,
297 	0x004004f4,
298 	0x00400478,
299 	0x004004f8,
300 	0x0040047c,
301 	0x004004fc,
302 	NV03_PGRAPH_ABS_UCLIP_XMIN,
303 	NV03_PGRAPH_ABS_UCLIP_XMAX,
304 	NV03_PGRAPH_ABS_UCLIP_YMIN,
305 	NV03_PGRAPH_ABS_UCLIP_YMAX,
306 	0x00400550,
307 	0x00400558,
308 	0x00400554,
309 	0x0040055c,
310 	NV03_PGRAPH_ABS_UCLIPA_XMIN,
311 	NV03_PGRAPH_ABS_UCLIPA_XMAX,
312 	NV03_PGRAPH_ABS_UCLIPA_YMIN,
313 	NV03_PGRAPH_ABS_UCLIPA_YMAX,
314 	NV03_PGRAPH_ABS_ICLIP_XMAX,
315 	NV03_PGRAPH_ABS_ICLIP_YMAX,
316 	NV03_PGRAPH_XY_LOGIC_MISC0,
317 	NV03_PGRAPH_XY_LOGIC_MISC1,
318 	NV03_PGRAPH_XY_LOGIC_MISC2,
319 	NV03_PGRAPH_XY_LOGIC_MISC3,
320 	NV03_PGRAPH_CLIPX_0,
321 	NV03_PGRAPH_CLIPX_1,
322 	NV03_PGRAPH_CLIPY_0,
323 	NV03_PGRAPH_CLIPY_1,
324 	NV10_PGRAPH_COMBINER0_IN_ALPHA,
325 	NV10_PGRAPH_COMBINER1_IN_ALPHA,
326 	NV10_PGRAPH_COMBINER0_IN_RGB,
327 	NV10_PGRAPH_COMBINER1_IN_RGB,
328 	NV10_PGRAPH_COMBINER_COLOR0,
329 	NV10_PGRAPH_COMBINER_COLOR1,
330 	NV10_PGRAPH_COMBINER0_OUT_ALPHA,
331 	NV10_PGRAPH_COMBINER1_OUT_ALPHA,
332 	NV10_PGRAPH_COMBINER0_OUT_RGB,
333 	NV10_PGRAPH_COMBINER1_OUT_RGB,
334 	NV10_PGRAPH_COMBINER_FINAL0,
335 	NV10_PGRAPH_COMBINER_FINAL1,
336 	0x00400e00,
337 	0x00400e04,
338 	0x00400e08,
339 	0x00400e0c,
340 	0x00400e10,
341 	0x00400e14,
342 	0x00400e18,
343 	0x00400e1c,
344 	0x00400e20,
345 	0x00400e24,
346 	0x00400e28,
347 	0x00400e2c,
348 	0x00400e30,
349 	0x00400e34,
350 	0x00400e38,
351 	0x00400e3c,
352 	NV04_PGRAPH_PASSTHRU_0,
353 	NV04_PGRAPH_PASSTHRU_1,
354 	NV04_PGRAPH_PASSTHRU_2,
355 	NV10_PGRAPH_DIMX_TEXTURE,
356 	NV10_PGRAPH_WDIMX_TEXTURE,
357 	NV10_PGRAPH_DVD_COLORFMT,
358 	NV10_PGRAPH_SCALED_FORMAT,
359 	NV04_PGRAPH_MISC24_0,
360 	NV04_PGRAPH_MISC24_1,
361 	NV04_PGRAPH_MISC24_2,
362 	NV03_PGRAPH_X_MISC,
363 	NV03_PGRAPH_Y_MISC,
364 	NV04_PGRAPH_VALID1,
365 	NV04_PGRAPH_VALID2,
366 };
367 
368 static int nv17_gr_ctx_regs[] = {
369 	NV10_PGRAPH_DEBUG_4,
370 	0x004006b0,
371 	0x00400eac,
372 	0x00400eb0,
373 	0x00400eb4,
374 	0x00400eb8,
375 	0x00400ebc,
376 	0x00400ec0,
377 	0x00400ec4,
378 	0x00400ec8,
379 	0x00400ecc,
380 	0x00400ed0,
381 	0x00400ed4,
382 	0x00400ed8,
383 	0x00400edc,
384 	0x00400ee0,
385 	0x00400a00,
386 	0x00400a04,
387 };
388 
389 #define nv10_gr(p) container_of((p), struct nv10_gr, base)
390 
391 struct nv10_gr {
392 	struct nvkm_gr base;
393 	struct nv10_gr_chan *chan[32];
394 	spinlock_t lock;
395 };
396 
397 #define nv10_gr_chan(p) container_of((p), struct nv10_gr_chan, object)
398 
399 struct nv10_gr_chan {
400 	struct nvkm_object object;
401 	struct nv10_gr *gr;
402 	int chid;
403 	int nv10[ARRAY_SIZE(nv10_gr_ctx_regs)];
404 	int nv17[ARRAY_SIZE(nv17_gr_ctx_regs)];
405 	struct pipe_state pipe_state;
406 	u32 lma_window[4];
407 };
408 
409 
410 /*******************************************************************************
411  * Graphics object classes
412  ******************************************************************************/
413 
414 #define PIPE_SAVE(gr, state, addr)					\
415 	do {								\
416 		int __i;						\
417 		nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, addr);		\
418 		for (__i = 0; __i < ARRAY_SIZE(state); __i++)		\
419 			state[__i] = nvkm_rd32(device, NV10_PGRAPH_PIPE_DATA); \
420 	} while (0)
421 
422 #define PIPE_RESTORE(gr, state, addr)					\
423 	do {								\
424 		int __i;						\
425 		nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, addr);		\
426 		for (__i = 0; __i < ARRAY_SIZE(state); __i++)		\
427 			nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, state[__i]); \
428 	} while (0)
429 
430 static void
431 nv17_gr_mthd_lma_window(struct nv10_gr_chan *chan, u32 mthd, u32 data)
432 {
433 	struct nvkm_device *device = chan->object.engine->subdev.device;
434 	struct nvkm_gr *gr = &chan->gr->base;
435 	struct pipe_state *pipe = &chan->pipe_state;
436 	u32 pipe_0x0040[1], pipe_0x64c0[8], pipe_0x6a80[3], pipe_0x6ab0[3];
437 	u32 xfmode0, xfmode1;
438 	int i;
439 
440 	chan->lma_window[(mthd - 0x1638) / 4] = data;
441 
442 	if (mthd != 0x1644)
443 		return;
444 
445 	nv04_gr_idle(gr);
446 
447 	PIPE_SAVE(device, pipe_0x0040, 0x0040);
448 	PIPE_SAVE(device, pipe->pipe_0x0200, 0x0200);
449 
450 	PIPE_RESTORE(device, chan->lma_window, 0x6790);
451 
452 	nv04_gr_idle(gr);
453 
454 	xfmode0 = nvkm_rd32(device, NV10_PGRAPH_XFMODE0);
455 	xfmode1 = nvkm_rd32(device, NV10_PGRAPH_XFMODE1);
456 
457 	PIPE_SAVE(device, pipe->pipe_0x4400, 0x4400);
458 	PIPE_SAVE(device, pipe_0x64c0, 0x64c0);
459 	PIPE_SAVE(device, pipe_0x6ab0, 0x6ab0);
460 	PIPE_SAVE(device, pipe_0x6a80, 0x6a80);
461 
462 	nv04_gr_idle(gr);
463 
464 	nvkm_wr32(device, NV10_PGRAPH_XFMODE0, 0x10000000);
465 	nvkm_wr32(device, NV10_PGRAPH_XFMODE1, 0x00000000);
466 	nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x000064c0);
467 	for (i = 0; i < 4; i++)
468 		nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x3f800000);
469 	for (i = 0; i < 4; i++)
470 		nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000);
471 
472 	nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00006ab0);
473 	for (i = 0; i < 3; i++)
474 		nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x3f800000);
475 
476 	nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00006a80);
477 	for (i = 0; i < 3; i++)
478 		nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000);
479 
480 	nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00000040);
481 	nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000008);
482 
483 	PIPE_RESTORE(device, pipe->pipe_0x0200, 0x0200);
484 
485 	nv04_gr_idle(gr);
486 
487 	PIPE_RESTORE(device, pipe_0x0040, 0x0040);
488 
489 	nvkm_wr32(device, NV10_PGRAPH_XFMODE0, xfmode0);
490 	nvkm_wr32(device, NV10_PGRAPH_XFMODE1, xfmode1);
491 
492 	PIPE_RESTORE(device, pipe_0x64c0, 0x64c0);
493 	PIPE_RESTORE(device, pipe_0x6ab0, 0x6ab0);
494 	PIPE_RESTORE(device, pipe_0x6a80, 0x6a80);
495 	PIPE_RESTORE(device, pipe->pipe_0x4400, 0x4400);
496 
497 	nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x000000c0);
498 	nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000);
499 
500 	nv04_gr_idle(gr);
501 }
502 
503 static void
504 nv17_gr_mthd_lma_enable(struct nv10_gr_chan *chan, u32 mthd, u32 data)
505 {
506 	struct nvkm_device *device = chan->object.engine->subdev.device;
507 	struct nvkm_gr *gr = &chan->gr->base;
508 
509 	nv04_gr_idle(gr);
510 
511 	nvkm_mask(device, NV10_PGRAPH_DEBUG_4, 0x00000100, 0x00000100);
512 	nvkm_mask(device, 0x4006b0, 0x08000000, 0x08000000);
513 }
514 
515 static bool
516 nv17_gr_mthd_celcius(struct nv10_gr_chan *chan, u32 mthd, u32 data)
517 {
518 	void (*func)(struct nv10_gr_chan *, u32, u32);
519 	switch (mthd) {
520 	case 0x1638 ... 0x1644:
521 		     func = nv17_gr_mthd_lma_window; break;
522 	case 0x1658: func = nv17_gr_mthd_lma_enable; break;
523 	default:
524 		return false;
525 	}
526 	func(chan, mthd, data);
527 	return true;
528 }
529 
530 static bool
531 nv10_gr_mthd(struct nv10_gr_chan *chan, u8 class, u32 mthd, u32 data)
532 {
533 	bool (*func)(struct nv10_gr_chan *, u32, u32);
534 	switch (class) {
535 	case 0x99: func = nv17_gr_mthd_celcius; break;
536 	default:
537 		return false;
538 	}
539 	return func(chan, mthd, data);
540 }
541 
542 /*******************************************************************************
543  * PGRAPH context
544  ******************************************************************************/
545 
546 static struct nv10_gr_chan *
547 nv10_gr_channel(struct nv10_gr *gr)
548 {
549 	struct nvkm_device *device = gr->base.engine.subdev.device;
550 	struct nv10_gr_chan *chan = NULL;
551 	if (nvkm_rd32(device, 0x400144) & 0x00010000) {
552 		int chid = nvkm_rd32(device, 0x400148) >> 24;
553 		if (chid < ARRAY_SIZE(gr->chan))
554 			chan = gr->chan[chid];
555 	}
556 	return chan;
557 }
558 
559 static void
560 nv10_gr_save_pipe(struct nv10_gr_chan *chan)
561 {
562 	struct nv10_gr *gr = chan->gr;
563 	struct pipe_state *pipe = &chan->pipe_state;
564 	struct nvkm_device *device = gr->base.engine.subdev.device;
565 
566 	PIPE_SAVE(gr, pipe->pipe_0x4400, 0x4400);
567 	PIPE_SAVE(gr, pipe->pipe_0x0200, 0x0200);
568 	PIPE_SAVE(gr, pipe->pipe_0x6400, 0x6400);
569 	PIPE_SAVE(gr, pipe->pipe_0x6800, 0x6800);
570 	PIPE_SAVE(gr, pipe->pipe_0x6c00, 0x6c00);
571 	PIPE_SAVE(gr, pipe->pipe_0x7000, 0x7000);
572 	PIPE_SAVE(gr, pipe->pipe_0x7400, 0x7400);
573 	PIPE_SAVE(gr, pipe->pipe_0x7800, 0x7800);
574 	PIPE_SAVE(gr, pipe->pipe_0x0040, 0x0040);
575 	PIPE_SAVE(gr, pipe->pipe_0x0000, 0x0000);
576 }
577 
578 static void
579 nv10_gr_load_pipe(struct nv10_gr_chan *chan)
580 {
581 	struct nv10_gr *gr = chan->gr;
582 	struct pipe_state *pipe = &chan->pipe_state;
583 	struct nvkm_device *device = gr->base.engine.subdev.device;
584 	u32 xfmode0, xfmode1;
585 	int i;
586 
587 	nv04_gr_idle(&gr->base);
588 	/* XXX check haiku comments */
589 	xfmode0 = nvkm_rd32(device, NV10_PGRAPH_XFMODE0);
590 	xfmode1 = nvkm_rd32(device, NV10_PGRAPH_XFMODE1);
591 	nvkm_wr32(device, NV10_PGRAPH_XFMODE0, 0x10000000);
592 	nvkm_wr32(device, NV10_PGRAPH_XFMODE1, 0x00000000);
593 	nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x000064c0);
594 	for (i = 0; i < 4; i++)
595 		nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x3f800000);
596 	for (i = 0; i < 4; i++)
597 		nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000);
598 
599 	nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00006ab0);
600 	for (i = 0; i < 3; i++)
601 		nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x3f800000);
602 
603 	nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00006a80);
604 	for (i = 0; i < 3; i++)
605 		nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000);
606 
607 	nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00000040);
608 	nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000008);
609 
610 
611 	PIPE_RESTORE(gr, pipe->pipe_0x0200, 0x0200);
612 	nv04_gr_idle(&gr->base);
613 
614 	/* restore XFMODE */
615 	nvkm_wr32(device, NV10_PGRAPH_XFMODE0, xfmode0);
616 	nvkm_wr32(device, NV10_PGRAPH_XFMODE1, xfmode1);
617 	PIPE_RESTORE(gr, pipe->pipe_0x6400, 0x6400);
618 	PIPE_RESTORE(gr, pipe->pipe_0x6800, 0x6800);
619 	PIPE_RESTORE(gr, pipe->pipe_0x6c00, 0x6c00);
620 	PIPE_RESTORE(gr, pipe->pipe_0x7000, 0x7000);
621 	PIPE_RESTORE(gr, pipe->pipe_0x7400, 0x7400);
622 	PIPE_RESTORE(gr, pipe->pipe_0x7800, 0x7800);
623 	PIPE_RESTORE(gr, pipe->pipe_0x4400, 0x4400);
624 	PIPE_RESTORE(gr, pipe->pipe_0x0000, 0x0000);
625 	PIPE_RESTORE(gr, pipe->pipe_0x0040, 0x0040);
626 	nv04_gr_idle(&gr->base);
627 }
628 
629 static void
630 nv10_gr_create_pipe(struct nv10_gr_chan *chan)
631 {
632 	struct nv10_gr *gr = chan->gr;
633 	struct nvkm_subdev *subdev = &gr->base.engine.subdev;
634 	struct pipe_state *pipe_state = &chan->pipe_state;
635 	u32 *pipe_state_addr;
636 	int i;
637 #define PIPE_INIT(addr) \
638 	do { \
639 		pipe_state_addr = pipe_state->pipe_##addr; \
640 	} while (0)
641 #define PIPE_INIT_END(addr) \
642 	do { \
643 		u32 *__end_addr = pipe_state->pipe_##addr + \
644 				ARRAY_SIZE(pipe_state->pipe_##addr); \
645 		if (pipe_state_addr != __end_addr) \
646 			nvkm_error(subdev, "incomplete pipe init for 0x%x :  %p/%p\n", \
647 				addr, pipe_state_addr, __end_addr); \
648 	} while (0)
649 #define NV_WRITE_PIPE_INIT(value) *(pipe_state_addr++) = value
650 
651 	PIPE_INIT(0x0200);
652 	for (i = 0; i < 48; i++)
653 		NV_WRITE_PIPE_INIT(0x00000000);
654 	PIPE_INIT_END(0x0200);
655 
656 	PIPE_INIT(0x6400);
657 	for (i = 0; i < 211; i++)
658 		NV_WRITE_PIPE_INIT(0x00000000);
659 	NV_WRITE_PIPE_INIT(0x3f800000);
660 	NV_WRITE_PIPE_INIT(0x40000000);
661 	NV_WRITE_PIPE_INIT(0x40000000);
662 	NV_WRITE_PIPE_INIT(0x40000000);
663 	NV_WRITE_PIPE_INIT(0x40000000);
664 	NV_WRITE_PIPE_INIT(0x00000000);
665 	NV_WRITE_PIPE_INIT(0x00000000);
666 	NV_WRITE_PIPE_INIT(0x3f800000);
667 	NV_WRITE_PIPE_INIT(0x00000000);
668 	NV_WRITE_PIPE_INIT(0x3f000000);
669 	NV_WRITE_PIPE_INIT(0x3f000000);
670 	NV_WRITE_PIPE_INIT(0x00000000);
671 	NV_WRITE_PIPE_INIT(0x00000000);
672 	NV_WRITE_PIPE_INIT(0x00000000);
673 	NV_WRITE_PIPE_INIT(0x00000000);
674 	NV_WRITE_PIPE_INIT(0x3f800000);
675 	NV_WRITE_PIPE_INIT(0x00000000);
676 	NV_WRITE_PIPE_INIT(0x00000000);
677 	NV_WRITE_PIPE_INIT(0x00000000);
678 	NV_WRITE_PIPE_INIT(0x00000000);
679 	NV_WRITE_PIPE_INIT(0x00000000);
680 	NV_WRITE_PIPE_INIT(0x3f800000);
681 	NV_WRITE_PIPE_INIT(0x3f800000);
682 	NV_WRITE_PIPE_INIT(0x3f800000);
683 	NV_WRITE_PIPE_INIT(0x3f800000);
684 	PIPE_INIT_END(0x6400);
685 
686 	PIPE_INIT(0x6800);
687 	for (i = 0; i < 162; i++)
688 		NV_WRITE_PIPE_INIT(0x00000000);
689 	NV_WRITE_PIPE_INIT(0x3f800000);
690 	for (i = 0; i < 25; i++)
691 		NV_WRITE_PIPE_INIT(0x00000000);
692 	PIPE_INIT_END(0x6800);
693 
694 	PIPE_INIT(0x6c00);
695 	NV_WRITE_PIPE_INIT(0x00000000);
696 	NV_WRITE_PIPE_INIT(0x00000000);
697 	NV_WRITE_PIPE_INIT(0x00000000);
698 	NV_WRITE_PIPE_INIT(0x00000000);
699 	NV_WRITE_PIPE_INIT(0xbf800000);
700 	NV_WRITE_PIPE_INIT(0x00000000);
701 	NV_WRITE_PIPE_INIT(0x00000000);
702 	NV_WRITE_PIPE_INIT(0x00000000);
703 	NV_WRITE_PIPE_INIT(0x00000000);
704 	NV_WRITE_PIPE_INIT(0x00000000);
705 	NV_WRITE_PIPE_INIT(0x00000000);
706 	NV_WRITE_PIPE_INIT(0x00000000);
707 	PIPE_INIT_END(0x6c00);
708 
709 	PIPE_INIT(0x7000);
710 	NV_WRITE_PIPE_INIT(0x00000000);
711 	NV_WRITE_PIPE_INIT(0x00000000);
712 	NV_WRITE_PIPE_INIT(0x00000000);
713 	NV_WRITE_PIPE_INIT(0x00000000);
714 	NV_WRITE_PIPE_INIT(0x00000000);
715 	NV_WRITE_PIPE_INIT(0x00000000);
716 	NV_WRITE_PIPE_INIT(0x00000000);
717 	NV_WRITE_PIPE_INIT(0x00000000);
718 	NV_WRITE_PIPE_INIT(0x00000000);
719 	NV_WRITE_PIPE_INIT(0x00000000);
720 	NV_WRITE_PIPE_INIT(0x00000000);
721 	NV_WRITE_PIPE_INIT(0x00000000);
722 	NV_WRITE_PIPE_INIT(0x7149f2ca);
723 	NV_WRITE_PIPE_INIT(0x00000000);
724 	NV_WRITE_PIPE_INIT(0x00000000);
725 	NV_WRITE_PIPE_INIT(0x00000000);
726 	NV_WRITE_PIPE_INIT(0x7149f2ca);
727 	NV_WRITE_PIPE_INIT(0x00000000);
728 	NV_WRITE_PIPE_INIT(0x00000000);
729 	NV_WRITE_PIPE_INIT(0x00000000);
730 	NV_WRITE_PIPE_INIT(0x7149f2ca);
731 	NV_WRITE_PIPE_INIT(0x00000000);
732 	NV_WRITE_PIPE_INIT(0x00000000);
733 	NV_WRITE_PIPE_INIT(0x00000000);
734 	NV_WRITE_PIPE_INIT(0x7149f2ca);
735 	NV_WRITE_PIPE_INIT(0x00000000);
736 	NV_WRITE_PIPE_INIT(0x00000000);
737 	NV_WRITE_PIPE_INIT(0x00000000);
738 	NV_WRITE_PIPE_INIT(0x7149f2ca);
739 	NV_WRITE_PIPE_INIT(0x00000000);
740 	NV_WRITE_PIPE_INIT(0x00000000);
741 	NV_WRITE_PIPE_INIT(0x00000000);
742 	NV_WRITE_PIPE_INIT(0x7149f2ca);
743 	NV_WRITE_PIPE_INIT(0x00000000);
744 	NV_WRITE_PIPE_INIT(0x00000000);
745 	NV_WRITE_PIPE_INIT(0x00000000);
746 	NV_WRITE_PIPE_INIT(0x7149f2ca);
747 	NV_WRITE_PIPE_INIT(0x00000000);
748 	NV_WRITE_PIPE_INIT(0x00000000);
749 	NV_WRITE_PIPE_INIT(0x00000000);
750 	NV_WRITE_PIPE_INIT(0x7149f2ca);
751 	for (i = 0; i < 35; i++)
752 		NV_WRITE_PIPE_INIT(0x00000000);
753 	PIPE_INIT_END(0x7000);
754 
755 	PIPE_INIT(0x7400);
756 	for (i = 0; i < 48; i++)
757 		NV_WRITE_PIPE_INIT(0x00000000);
758 	PIPE_INIT_END(0x7400);
759 
760 	PIPE_INIT(0x7800);
761 	for (i = 0; i < 48; i++)
762 		NV_WRITE_PIPE_INIT(0x00000000);
763 	PIPE_INIT_END(0x7800);
764 
765 	PIPE_INIT(0x4400);
766 	for (i = 0; i < 32; i++)
767 		NV_WRITE_PIPE_INIT(0x00000000);
768 	PIPE_INIT_END(0x4400);
769 
770 	PIPE_INIT(0x0000);
771 	for (i = 0; i < 16; i++)
772 		NV_WRITE_PIPE_INIT(0x00000000);
773 	PIPE_INIT_END(0x0000);
774 
775 	PIPE_INIT(0x0040);
776 	for (i = 0; i < 4; i++)
777 		NV_WRITE_PIPE_INIT(0x00000000);
778 	PIPE_INIT_END(0x0040);
779 
780 #undef PIPE_INIT
781 #undef PIPE_INIT_END
782 #undef NV_WRITE_PIPE_INIT
783 }
784 
785 static int
786 nv10_gr_ctx_regs_find_offset(struct nv10_gr *gr, int reg)
787 {
788 	struct nvkm_subdev *subdev = &gr->base.engine.subdev;
789 	int i;
790 	for (i = 0; i < ARRAY_SIZE(nv10_gr_ctx_regs); i++) {
791 		if (nv10_gr_ctx_regs[i] == reg)
792 			return i;
793 	}
794 	nvkm_error(subdev, "unknown offset nv10_ctx_regs %d\n", reg);
795 	return -1;
796 }
797 
798 static int
799 nv17_gr_ctx_regs_find_offset(struct nv10_gr *gr, int reg)
800 {
801 	struct nvkm_subdev *subdev = &gr->base.engine.subdev;
802 	int i;
803 	for (i = 0; i < ARRAY_SIZE(nv17_gr_ctx_regs); i++) {
804 		if (nv17_gr_ctx_regs[i] == reg)
805 			return i;
806 	}
807 	nvkm_error(subdev, "unknown offset nv17_ctx_regs %d\n", reg);
808 	return -1;
809 }
810 
811 static void
812 nv10_gr_load_dma_vtxbuf(struct nv10_gr_chan *chan, int chid, u32 inst)
813 {
814 	struct nv10_gr *gr = chan->gr;
815 	struct nvkm_device *device = gr->base.engine.subdev.device;
816 	u32 st2, st2_dl, st2_dh, fifo_ptr, fifo[0x60/4];
817 	u32 ctx_user, ctx_switch[5];
818 	int i, subchan = -1;
819 
820 	/* NV10TCL_DMA_VTXBUF (method 0x18c) modifies hidden state
821 	 * that cannot be restored via MMIO. Do it through the FIFO
822 	 * instead.
823 	 */
824 
825 	/* Look for a celsius object */
826 	for (i = 0; i < 8; i++) {
827 		int class = nvkm_rd32(device, NV10_PGRAPH_CTX_CACHE(i, 0)) & 0xfff;
828 
829 		if (class == 0x56 || class == 0x96 || class == 0x99) {
830 			subchan = i;
831 			break;
832 		}
833 	}
834 
835 	if (subchan < 0 || !inst)
836 		return;
837 
838 	/* Save the current ctx object */
839 	ctx_user = nvkm_rd32(device, NV10_PGRAPH_CTX_USER);
840 	for (i = 0; i < 5; i++)
841 		ctx_switch[i] = nvkm_rd32(device, NV10_PGRAPH_CTX_SWITCH(i));
842 
843 	/* Save the FIFO state */
844 	st2 = nvkm_rd32(device, NV10_PGRAPH_FFINTFC_ST2);
845 	st2_dl = nvkm_rd32(device, NV10_PGRAPH_FFINTFC_ST2_DL);
846 	st2_dh = nvkm_rd32(device, NV10_PGRAPH_FFINTFC_ST2_DH);
847 	fifo_ptr = nvkm_rd32(device, NV10_PGRAPH_FFINTFC_FIFO_PTR);
848 
849 	for (i = 0; i < ARRAY_SIZE(fifo); i++)
850 		fifo[i] = nvkm_rd32(device, 0x4007a0 + 4 * i);
851 
852 	/* Switch to the celsius subchannel */
853 	for (i = 0; i < 5; i++)
854 		nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(i),
855 			nvkm_rd32(device, NV10_PGRAPH_CTX_CACHE(subchan, i)));
856 	nvkm_mask(device, NV10_PGRAPH_CTX_USER, 0xe000, subchan << 13);
857 
858 	/* Inject NV10TCL_DMA_VTXBUF */
859 	nvkm_wr32(device, NV10_PGRAPH_FFINTFC_FIFO_PTR, 0);
860 	nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2,
861 		0x2c000000 | chid << 20 | subchan << 16 | 0x18c);
862 	nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2_DL, inst);
863 	nvkm_mask(device, NV10_PGRAPH_CTX_CONTROL, 0, 0x10000);
864 	nvkm_mask(device, NV04_PGRAPH_FIFO, 0x00000001, 0x00000001);
865 	nvkm_mask(device, NV04_PGRAPH_FIFO, 0x00000001, 0x00000000);
866 
867 	/* Restore the FIFO state */
868 	for (i = 0; i < ARRAY_SIZE(fifo); i++)
869 		nvkm_wr32(device, 0x4007a0 + 4 * i, fifo[i]);
870 
871 	nvkm_wr32(device, NV10_PGRAPH_FFINTFC_FIFO_PTR, fifo_ptr);
872 	nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2, st2);
873 	nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2_DL, st2_dl);
874 	nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2_DH, st2_dh);
875 
876 	/* Restore the current ctx object */
877 	for (i = 0; i < 5; i++)
878 		nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(i), ctx_switch[i]);
879 	nvkm_wr32(device, NV10_PGRAPH_CTX_USER, ctx_user);
880 }
881 
882 static int
883 nv10_gr_load_context(struct nv10_gr_chan *chan, int chid)
884 {
885 	struct nv10_gr *gr = chan->gr;
886 	struct nvkm_device *device = gr->base.engine.subdev.device;
887 	u32 inst;
888 	int i;
889 
890 	for (i = 0; i < ARRAY_SIZE(nv10_gr_ctx_regs); i++)
891 		nvkm_wr32(device, nv10_gr_ctx_regs[i], chan->nv10[i]);
892 
893 	if (device->card_type >= NV_11 && device->chipset >= 0x17) {
894 		for (i = 0; i < ARRAY_SIZE(nv17_gr_ctx_regs); i++)
895 			nvkm_wr32(device, nv17_gr_ctx_regs[i], chan->nv17[i]);
896 	}
897 
898 	nv10_gr_load_pipe(chan);
899 
900 	inst = nvkm_rd32(device, NV10_PGRAPH_GLOBALSTATE1) & 0xffff;
901 	nv10_gr_load_dma_vtxbuf(chan, chid, inst);
902 
903 	nvkm_wr32(device, NV10_PGRAPH_CTX_CONTROL, 0x10010100);
904 	nvkm_mask(device, NV10_PGRAPH_CTX_USER, 0xff000000, chid << 24);
905 	nvkm_mask(device, NV10_PGRAPH_FFINTFC_ST2, 0x30000000, 0x00000000);
906 	return 0;
907 }
908 
909 static int
910 nv10_gr_unload_context(struct nv10_gr_chan *chan)
911 {
912 	struct nv10_gr *gr = chan->gr;
913 	struct nvkm_device *device = gr->base.engine.subdev.device;
914 	int i;
915 
916 	for (i = 0; i < ARRAY_SIZE(nv10_gr_ctx_regs); i++)
917 		chan->nv10[i] = nvkm_rd32(device, nv10_gr_ctx_regs[i]);
918 
919 	if (device->card_type >= NV_11 && device->chipset >= 0x17) {
920 		for (i = 0; i < ARRAY_SIZE(nv17_gr_ctx_regs); i++)
921 			chan->nv17[i] = nvkm_rd32(device, nv17_gr_ctx_regs[i]);
922 	}
923 
924 	nv10_gr_save_pipe(chan);
925 
926 	nvkm_wr32(device, NV10_PGRAPH_CTX_CONTROL, 0x10000000);
927 	nvkm_mask(device, NV10_PGRAPH_CTX_USER, 0xff000000, 0x1f000000);
928 	return 0;
929 }
930 
931 static void
932 nv10_gr_context_switch(struct nv10_gr *gr)
933 {
934 	struct nvkm_device *device = gr->base.engine.subdev.device;
935 	struct nv10_gr_chan *prev = NULL;
936 	struct nv10_gr_chan *next = NULL;
937 	int chid;
938 
939 	nv04_gr_idle(&gr->base);
940 
941 	/* If previous context is valid, we need to save it */
942 	prev = nv10_gr_channel(gr);
943 	if (prev)
944 		nv10_gr_unload_context(prev);
945 
946 	/* load context for next channel */
947 	chid = (nvkm_rd32(device, NV04_PGRAPH_TRAPPED_ADDR) >> 20) & 0x1f;
948 	next = gr->chan[chid];
949 	if (next)
950 		nv10_gr_load_context(next, chid);
951 }
952 
953 static int
954 nv10_gr_chan_fini(struct nvkm_object *object, bool suspend)
955 {
956 	struct nv10_gr_chan *chan = nv10_gr_chan(object);
957 	struct nv10_gr *gr = chan->gr;
958 	struct nvkm_device *device = gr->base.engine.subdev.device;
959 	unsigned long flags;
960 
961 	spin_lock_irqsave(&gr->lock, flags);
962 	nvkm_mask(device, NV04_PGRAPH_FIFO, 0x00000001, 0x00000000);
963 	if (nv10_gr_channel(gr) == chan)
964 		nv10_gr_unload_context(chan);
965 	nvkm_mask(device, NV04_PGRAPH_FIFO, 0x00000001, 0x00000001);
966 	spin_unlock_irqrestore(&gr->lock, flags);
967 	return 0;
968 }
969 
970 static void *
971 nv10_gr_chan_dtor(struct nvkm_object *object)
972 {
973 	struct nv10_gr_chan *chan = nv10_gr_chan(object);
974 	struct nv10_gr *gr = chan->gr;
975 	unsigned long flags;
976 
977 	spin_lock_irqsave(&gr->lock, flags);
978 	gr->chan[chan->chid] = NULL;
979 	spin_unlock_irqrestore(&gr->lock, flags);
980 	return chan;
981 }
982 
983 static const struct nvkm_object_func
984 nv10_gr_chan = {
985 	.dtor = nv10_gr_chan_dtor,
986 	.fini = nv10_gr_chan_fini,
987 };
988 
989 #define NV_WRITE_CTX(reg, val) do { \
990 	int offset = nv10_gr_ctx_regs_find_offset(gr, reg); \
991 	if (offset > 0) \
992 		chan->nv10[offset] = val; \
993 	} while (0)
994 
995 #define NV17_WRITE_CTX(reg, val) do { \
996 	int offset = nv17_gr_ctx_regs_find_offset(gr, reg); \
997 	if (offset > 0) \
998 		chan->nv17[offset] = val; \
999 	} while (0)
1000 
1001 int
1002 nv10_gr_chan_new(struct nvkm_gr *base, struct nvkm_chan *fifoch,
1003 		 const struct nvkm_oclass *oclass, struct nvkm_object **pobject)
1004 {
1005 	struct nv10_gr *gr = nv10_gr(base);
1006 	struct nv10_gr_chan *chan;
1007 	struct nvkm_device *device = gr->base.engine.subdev.device;
1008 	unsigned long flags;
1009 
1010 	if (!(chan = kzalloc(sizeof(*chan), GFP_KERNEL)))
1011 		return -ENOMEM;
1012 	nvkm_object_ctor(&nv10_gr_chan, oclass, &chan->object);
1013 	chan->gr = gr;
1014 	chan->chid = fifoch->id;
1015 	*pobject = &chan->object;
1016 
1017 	NV_WRITE_CTX(0x00400e88, 0x08000000);
1018 	NV_WRITE_CTX(0x00400e9c, 0x4b7fffff);
1019 	NV_WRITE_CTX(NV03_PGRAPH_XY_LOGIC_MISC0, 0x0001ffff);
1020 	NV_WRITE_CTX(0x00400e10, 0x00001000);
1021 	NV_WRITE_CTX(0x00400e14, 0x00001000);
1022 	NV_WRITE_CTX(0x00400e30, 0x00080008);
1023 	NV_WRITE_CTX(0x00400e34, 0x00080008);
1024 	if (device->card_type >= NV_11 && device->chipset >= 0x17) {
1025 		/* is it really needed ??? */
1026 		NV17_WRITE_CTX(NV10_PGRAPH_DEBUG_4,
1027 			       nvkm_rd32(device, NV10_PGRAPH_DEBUG_4));
1028 		NV17_WRITE_CTX(0x004006b0, nvkm_rd32(device, 0x004006b0));
1029 		NV17_WRITE_CTX(0x00400eac, 0x0fff0000);
1030 		NV17_WRITE_CTX(0x00400eb0, 0x0fff0000);
1031 		NV17_WRITE_CTX(0x00400ec0, 0x00000080);
1032 		NV17_WRITE_CTX(0x00400ed0, 0x00000080);
1033 	}
1034 	NV_WRITE_CTX(NV10_PGRAPH_CTX_USER, chan->chid << 24);
1035 
1036 	nv10_gr_create_pipe(chan);
1037 
1038 	spin_lock_irqsave(&gr->lock, flags);
1039 	gr->chan[chan->chid] = chan;
1040 	spin_unlock_irqrestore(&gr->lock, flags);
1041 	return 0;
1042 }
1043 
1044 /*******************************************************************************
1045  * PGRAPH engine/subdev functions
1046  ******************************************************************************/
1047 
1048 void
1049 nv10_gr_tile(struct nvkm_gr *base, int i, struct nvkm_fb_tile *tile)
1050 {
1051 	struct nv10_gr *gr = nv10_gr(base);
1052 	struct nvkm_device *device = gr->base.engine.subdev.device;
1053 	struct nvkm_fifo *fifo = device->fifo;
1054 	unsigned long flags;
1055 
1056 	nvkm_fifo_pause(fifo, &flags);
1057 	nv04_gr_idle(&gr->base);
1058 
1059 	nvkm_wr32(device, NV10_PGRAPH_TLIMIT(i), tile->limit);
1060 	nvkm_wr32(device, NV10_PGRAPH_TSIZE(i), tile->pitch);
1061 	nvkm_wr32(device, NV10_PGRAPH_TILE(i), tile->addr);
1062 
1063 	nvkm_fifo_start(fifo, &flags);
1064 }
1065 
1066 const struct nvkm_bitfield nv10_gr_intr_name[] = {
1067 	{ NV_PGRAPH_INTR_NOTIFY, "NOTIFY" },
1068 	{ NV_PGRAPH_INTR_ERROR,  "ERROR"  },
1069 	{}
1070 };
1071 
1072 const struct nvkm_bitfield nv10_gr_nstatus[] = {
1073 	{ NV10_PGRAPH_NSTATUS_STATE_IN_USE,       "STATE_IN_USE" },
1074 	{ NV10_PGRAPH_NSTATUS_INVALID_STATE,      "INVALID_STATE" },
1075 	{ NV10_PGRAPH_NSTATUS_BAD_ARGUMENT,       "BAD_ARGUMENT" },
1076 	{ NV10_PGRAPH_NSTATUS_PROTECTION_FAULT,   "PROTECTION_FAULT" },
1077 	{}
1078 };
1079 
1080 void
1081 nv10_gr_intr(struct nvkm_gr *base)
1082 {
1083 	struct nv10_gr *gr = nv10_gr(base);
1084 	struct nvkm_subdev *subdev = &gr->base.engine.subdev;
1085 	struct nvkm_device *device = subdev->device;
1086 	u32 stat = nvkm_rd32(device, NV03_PGRAPH_INTR);
1087 	u32 nsource = nvkm_rd32(device, NV03_PGRAPH_NSOURCE);
1088 	u32 nstatus = nvkm_rd32(device, NV03_PGRAPH_NSTATUS);
1089 	u32 addr = nvkm_rd32(device, NV04_PGRAPH_TRAPPED_ADDR);
1090 	u32 chid = (addr & 0x01f00000) >> 20;
1091 	u32 subc = (addr & 0x00070000) >> 16;
1092 	u32 mthd = (addr & 0x00001ffc);
1093 	u32 data = nvkm_rd32(device, NV04_PGRAPH_TRAPPED_DATA);
1094 	u32 class = nvkm_rd32(device, 0x400160 + subc * 4) & 0xfff;
1095 	u32 show = stat;
1096 	char msg[128], src[128], sta[128];
1097 	struct nv10_gr_chan *chan;
1098 	unsigned long flags;
1099 
1100 	spin_lock_irqsave(&gr->lock, flags);
1101 	chan = gr->chan[chid];
1102 
1103 	if (stat & NV_PGRAPH_INTR_ERROR) {
1104 		if (chan && (nsource & NV03_PGRAPH_NSOURCE_ILLEGAL_MTHD)) {
1105 			if (!nv10_gr_mthd(chan, class, mthd, data))
1106 				show &= ~NV_PGRAPH_INTR_ERROR;
1107 		}
1108 	}
1109 
1110 	if (stat & NV_PGRAPH_INTR_CONTEXT_SWITCH) {
1111 		nvkm_wr32(device, NV03_PGRAPH_INTR, NV_PGRAPH_INTR_CONTEXT_SWITCH);
1112 		stat &= ~NV_PGRAPH_INTR_CONTEXT_SWITCH;
1113 		show &= ~NV_PGRAPH_INTR_CONTEXT_SWITCH;
1114 		nv10_gr_context_switch(gr);
1115 	}
1116 
1117 	nvkm_wr32(device, NV03_PGRAPH_INTR, stat);
1118 	nvkm_wr32(device, NV04_PGRAPH_FIFO, 0x00000001);
1119 
1120 	if (show) {
1121 		nvkm_snprintbf(msg, sizeof(msg), nv10_gr_intr_name, show);
1122 		nvkm_snprintbf(src, sizeof(src), nv04_gr_nsource, nsource);
1123 		nvkm_snprintbf(sta, sizeof(sta), nv10_gr_nstatus, nstatus);
1124 		nvkm_error(subdev, "intr %08x [%s] nsource %08x [%s] "
1125 				   "nstatus %08x [%s] ch %d [%s] subc %d "
1126 				   "class %04x mthd %04x data %08x\n",
1127 			   show, msg, nsource, src, nstatus, sta, chid,
1128 			   chan ? chan->object.client->name : "unknown",
1129 			   subc, class, mthd, data);
1130 	}
1131 
1132 	spin_unlock_irqrestore(&gr->lock, flags);
1133 }
1134 
1135 int
1136 nv10_gr_init(struct nvkm_gr *base)
1137 {
1138 	struct nv10_gr *gr = nv10_gr(base);
1139 	struct nvkm_device *device = gr->base.engine.subdev.device;
1140 
1141 	nvkm_wr32(device, NV03_PGRAPH_INTR   , 0xFFFFFFFF);
1142 	nvkm_wr32(device, NV03_PGRAPH_INTR_EN, 0xFFFFFFFF);
1143 
1144 	nvkm_wr32(device, NV04_PGRAPH_DEBUG_0, 0xFFFFFFFF);
1145 	nvkm_wr32(device, NV04_PGRAPH_DEBUG_0, 0x00000000);
1146 	nvkm_wr32(device, NV04_PGRAPH_DEBUG_1, 0x00118700);
1147 	/* nvkm_wr32(device, NV04_PGRAPH_DEBUG_2, 0x24E00810); */ /* 0x25f92ad9 */
1148 	nvkm_wr32(device, NV04_PGRAPH_DEBUG_2, 0x25f92ad9);
1149 	nvkm_wr32(device, NV04_PGRAPH_DEBUG_3, 0x55DE0830 | (1 << 29) | (1 << 31));
1150 
1151 	if (device->card_type >= NV_11 && device->chipset >= 0x17) {
1152 		nvkm_wr32(device, NV10_PGRAPH_DEBUG_4, 0x1f000000);
1153 		nvkm_wr32(device, 0x400a10, 0x03ff3fb6);
1154 		nvkm_wr32(device, 0x400838, 0x002f8684);
1155 		nvkm_wr32(device, 0x40083c, 0x00115f3f);
1156 		nvkm_wr32(device, 0x4006b0, 0x40000020);
1157 	} else {
1158 		nvkm_wr32(device, NV10_PGRAPH_DEBUG_4, 0x00000000);
1159 	}
1160 
1161 	nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(0), 0x00000000);
1162 	nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(1), 0x00000000);
1163 	nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(2), 0x00000000);
1164 	nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(3), 0x00000000);
1165 	nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(4), 0x00000000);
1166 	nvkm_wr32(device, NV10_PGRAPH_STATE, 0xFFFFFFFF);
1167 
1168 	nvkm_mask(device, NV10_PGRAPH_CTX_USER, 0xff000000, 0x1f000000);
1169 	nvkm_wr32(device, NV10_PGRAPH_CTX_CONTROL, 0x10000100);
1170 	nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2, 0x08000000);
1171 	return 0;
1172 }
1173 
1174 int
1175 nv10_gr_new_(const struct nvkm_gr_func *func, struct nvkm_device *device,
1176 	     enum nvkm_subdev_type type, int inst, struct nvkm_gr **pgr)
1177 {
1178 	struct nv10_gr *gr;
1179 
1180 	if (!(gr = kzalloc(sizeof(*gr), GFP_KERNEL)))
1181 		return -ENOMEM;
1182 	spin_lock_init(&gr->lock);
1183 	*pgr = &gr->base;
1184 
1185 	return nvkm_gr_ctor(func, device, type, inst, true, &gr->base);
1186 }
1187 
1188 static const struct nvkm_gr_func
1189 nv10_gr = {
1190 	.init = nv10_gr_init,
1191 	.intr = nv10_gr_intr,
1192 	.tile = nv10_gr_tile,
1193 	.chan_new = nv10_gr_chan_new,
1194 	.sclass = {
1195 		{ -1, -1, 0x0012, &nv04_gr_object }, /* beta1 */
1196 		{ -1, -1, 0x0019, &nv04_gr_object }, /* clip */
1197 		{ -1, -1, 0x0030, &nv04_gr_object }, /* null */
1198 		{ -1, -1, 0x0039, &nv04_gr_object }, /* m2mf */
1199 		{ -1, -1, 0x0043, &nv04_gr_object }, /* rop */
1200 		{ -1, -1, 0x0044, &nv04_gr_object }, /* pattern */
1201 		{ -1, -1, 0x004a, &nv04_gr_object }, /* gdi */
1202 		{ -1, -1, 0x0052, &nv04_gr_object }, /* swzsurf */
1203 		{ -1, -1, 0x005f, &nv04_gr_object }, /* blit */
1204 		{ -1, -1, 0x0062, &nv04_gr_object }, /* surf2d */
1205 		{ -1, -1, 0x0072, &nv04_gr_object }, /* beta4 */
1206 		{ -1, -1, 0x0089, &nv04_gr_object }, /* sifm */
1207 		{ -1, -1, 0x008a, &nv04_gr_object }, /* ifc */
1208 		{ -1, -1, 0x009f, &nv04_gr_object }, /* blit */
1209 		{ -1, -1, 0x0093, &nv04_gr_object }, /* surf3d */
1210 		{ -1, -1, 0x0094, &nv04_gr_object }, /* ttri */
1211 		{ -1, -1, 0x0095, &nv04_gr_object }, /* mtri */
1212 		{ -1, -1, 0x0056, &nv04_gr_object }, /* celcius */
1213 		{}
1214 	}
1215 };
1216 
1217 int
1218 nv10_gr_new(struct nvkm_device *device, enum nvkm_subdev_type type, int inst, struct nvkm_gr **pgr)
1219 {
1220 	return nv10_gr_new_(&nv10_gr, device, type, inst, pgr);
1221 }
1222