/linux/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
H A D | r535.c | 154 args->userdMem.base = nvkm_memory_addr(chan->userd.mem) + chan->userd.base; in r535_chan_ramfc_write() 155 args->userdMem.size = fifo->func->chan.func->userd->size; in r535_chan_ramfc_write() 225 } *userd; variable 232 struct r535_chan_userd *userd; in r535_chan_id_put() local 234 mutex_lock(&fifo->userd.mutex); in r535_chan_id_put() 235 list_for_each_entry(userd, &fifo->userd.list, head) { in r535_chan_id_put() 236 if (userd->map == chan->userd.mem) { in r535_chan_id_put() 237 u32 chid = chan->userd.base / chan->func->userd->size; in r535_chan_id_put() 239 userd->used &= ~BIT(chid); in r535_chan_id_put() 240 if (!userd->used) { in r535_chan_id_put() [all …]
|
H A D | chan.c | 286 nvkm_memory_unref(&chan->userd.mem); in nvkm_chan_del() 352 struct nvkm_memory *userd, u64 ouserd, struct nvkm_chan **pchan) in nvkm_chan_new_() argument 362 ((func->userd->bar < 0) == !userd) || in nvkm_chan_new_() 369 func->userd->bar < 0, userd, func->ramfc->ctxdma, dmaobj, in nvkm_chan_new_() 447 if (func->userd->bar < 0) { in nvkm_chan_new_() 448 if (ouserd + chan->func->userd->size >= in nvkm_chan_new_() 449 nvkm_memory_size(userd)) { in nvkm_chan_new_() 454 ret = nvkm_memory_kmap(userd, &chan->userd.mem); in nvkm_chan_new_() 460 chan->userd.base = ouserd; in nvkm_chan_new_() 462 chan->userd.mem = nvkm_memory_ref(fifo->userd.mem); in nvkm_chan_new_() [all …]
|
H A D | gf100.c | 87 const u64 userd = nvkm_memory_addr(chan->userd.mem) + chan->userd.base; in gf100_chan_ramfc_write() local 91 nvkm_wo32(chan->inst, 0x08, lower_32_bits(userd)); in gf100_chan_ramfc_write() 92 nvkm_wo32(chan->inst, 0x0c, upper_32_bits(userd)); in gf100_chan_ramfc_write() 120 nvkm_kmap(chan->userd.mem); in gf100_chan_userd_clear() 121 nvkm_wo32(chan->userd.mem, chan->userd.base + 0x040, 0x00000000); in gf100_chan_userd_clear() 122 nvkm_wo32(chan->userd.mem, chan->userd.base + 0x044, 0x00000000); in gf100_chan_userd_clear() 123 nvkm_wo32(chan->userd.mem, chan->userd.base + 0x048, 0x00000000); in gf100_chan_userd_clear() 124 nvkm_wo32(chan->userd.mem, chan->userd.base + 0x04c, 0x00000000); in gf100_chan_userd_clear() 125 nvkm_wo32(chan->userd.mem, chan->userd.base + 0x050, 0x00000000); in gf100_chan_userd_clear() 126 nvkm_wo32(chan->userd.mem, chan->userd.base + 0x058, 0x00000000); in gf100_chan_userd_clear() [all …]
|
H A D | base.c | 306 if (fifo->func->chan.func->userd->bar == 1) { in nvkm_fifo_oneinit() 310 fifo->func->chan.func->userd->size, 0, true, in nvkm_fifo_oneinit() 311 &fifo->userd.mem); in nvkm_fifo_oneinit() 315 ret = nvkm_vmm_get(bar1, 12, nvkm_memory_size(fifo->userd.mem), &fifo->userd.bar1); in nvkm_fifo_oneinit() 319 ret = nvkm_memory_map(fifo->userd.mem, 0, bar1, fifo->userd.bar1, NULL, 0); in nvkm_fifo_oneinit() 340 if (fifo->userd.bar1) in nvkm_fifo_dtor() 341 nvkm_vmm_put(nvkm_bar_bar1_vmm(engine->subdev.device), &fifo->userd.bar1); in nvkm_fifo_dtor() 342 nvkm_memory_unref(&fifo->userd.mem); in nvkm_fifo_dtor() 352 mutex_destroy(&fifo->userd.mutex); in nvkm_fifo_dtor() 394 INIT_LIST_HEAD(&fifo->userd.list); in nvkm_fifo_new_() [all …]
|
H A D | uchan.c | 261 if (chan->func->userd->bar < 0) in nvkm_uchan_map() 265 *addr = device->func->resource_addr(device, chan->func->userd->bar) + in nvkm_uchan_map() 266 chan->func->userd->base + chan->userd.base; in nvkm_uchan_map() 267 *size = chan->func->userd->size; in nvkm_uchan_map() 337 struct nvkm_memory *userd = NULL; in nvkm_uchan_new() local 369 userd = nvkm_umem_search(oclass->client, args->v0.huserd); in nvkm_uchan_new() 370 if (IS_ERR(userd)) { in nvkm_uchan_new() 371 ret = PTR_ERR(userd); in nvkm_uchan_new() 372 userd = NULL; in nvkm_uchan_new() 388 args->v0.length, userd, args->v0.ouserd, &uchan->chan); in nvkm_uchan_new() [all …]
|
H A D | gv100.c | 43 const u64 userd = nvkm_memory_addr(chan->userd.mem) + chan->userd.base; in gv100_chan_ramfc_write() local 47 nvkm_wo32(chan->inst, 0x008, lower_32_bits(userd)); in gv100_chan_ramfc_write() 48 nvkm_wo32(chan->inst, 0x00c, upper_32_bits(userd)); in gv100_chan_ramfc_write() 81 .userd = &gv100_chan_userd, 185 const u64 user = nvkm_memory_addr(chan->userd.mem) + chan->userd.base; in gv100_runl_insert_chan()
|
H A D | chan.h | 20 int (*id_get)(struct nvkm_chan *, struct nvkm_memory *userd, u64 ouserd); 34 } *userd; member 61 u64 offset, u64 length, struct nvkm_memory *userd, u64 userd_bar1,
|
H A D | gk104.c | 84 const u64 userd = nvkm_memory_addr(chan->userd.mem) + chan->userd.base; in gk104_chan_ramfc_write() local 88 nvkm_wo32(chan->inst, 0x08, lower_32_bits(userd)); in gk104_chan_ramfc_write() 89 nvkm_wo32(chan->inst, 0x0c, upper_32_bits(userd)); in gk104_chan_ramfc_write() 124 .userd = &gk104_chan_userd, 748 if (fifo->func->chan.func->userd->bar == 1) in gk104_fifo_init() 749 nvkm_wr32(device, 0x002254, 0x10000000 | fifo->userd.bar1->addr >> 12); in gk104_fifo_init()
|
H A D | nv10.c | 80 .userd = &nv04_chan_userd,
|
H A D | gk110.c | 51 .userd = &gk104_chan_userd,
|
H A D | nv17.c | 86 .userd = &nv04_chan_userd,
|
H A D | gm107.c | 36 .userd = &gk104_chan_userd,
|
H A D | nv40.c | 104 .userd = &nv40_chan_userd,
|
H A D | tu102.c | 52 .userd = &gv100_chan_userd,
|
H A D | g84.c | 97 .userd = &nv50_chan_userd,
|
H A D | nv50.c | 141 .userd = &nv50_chan_userd,
|
H A D | nv04.c | 170 .userd = &nv04_chan_userd,
|
H A D | ga100.c | 100 .userd = &gv100_chan_userd,
|
/linux/drivers/gpu/drm/nouveau/ |
H A D | nouveau_dma.c | 45 val = nvif_rd32(chan->userd, chan->user_get); in READ_GET() 47 val |= (uint64_t)nvif_rd32(chan->userd, chan->user_get_hi) << 32; in READ_GET() 92 nvif_wr32(chan->userd, 0x8c, chan->dma.ib_put); in nv50_dma_push() 104 uint32_t get = nvif_rd32(chan->userd, 0x88); in nv50_dma_push_wait()
|
H A D | nouveau_chan.h | 18 struct nvif_object *userd; member
|
H A D | nouveau_chan.c | 336 chan->userd = &chan->mem_userd.object; in nouveau_channel_ctor() 338 chan->userd = &chan->user; in nouveau_channel_ctor() 367 ret = nvif_object_map(chan->userd, NULL, 0); in nouveau_channel_init()
|
/linux/drivers/gpu/drm/nouveau/include/nvkm/engine/ |
H A D | fifo.h | 26 } userd; member 84 } userd; member
|