Home
last modified time | relevance | path

Searched refs:userd (Results 1 – 22 of 22) sorted by relevance

/linux/drivers/gpu/drm/nouveau/nvkm/engine/fifo/
H A Dr535.c154 args->userdMem.base = nvkm_memory_addr(chan->userd.mem) + chan->userd.base; in r535_chan_ramfc_write()
155 args->userdMem.size = fifo->func->chan.func->userd->size; in r535_chan_ramfc_write()
225 } *userd; variable
232 struct r535_chan_userd *userd; in r535_chan_id_put() local
234 mutex_lock(&fifo->userd.mutex); in r535_chan_id_put()
235 list_for_each_entry(userd, &fifo->userd.list, head) { in r535_chan_id_put()
236 if (userd->map == chan->userd.mem) { in r535_chan_id_put()
237 u32 chid = chan->userd.base / chan->func->userd->size; in r535_chan_id_put()
239 userd->used &= ~BIT(chid); in r535_chan_id_put()
240 if (!userd->used) { in r535_chan_id_put()
[all …]
H A Dchan.c286 nvkm_memory_unref(&chan->userd.mem); in nvkm_chan_del()
352 struct nvkm_memory *userd, u64 ouserd, struct nvkm_chan **pchan) in nvkm_chan_new_() argument
362 ((func->userd->bar < 0) == !userd) || in nvkm_chan_new_()
369 func->userd->bar < 0, userd, func->ramfc->ctxdma, dmaobj, in nvkm_chan_new_()
447 if (func->userd->bar < 0) { in nvkm_chan_new_()
448 if (ouserd + chan->func->userd->size >= in nvkm_chan_new_()
449 nvkm_memory_size(userd)) { in nvkm_chan_new_()
454 ret = nvkm_memory_kmap(userd, &chan->userd.mem); in nvkm_chan_new_()
460 chan->userd.base = ouserd; in nvkm_chan_new_()
462 chan->userd.mem = nvkm_memory_ref(fifo->userd.mem); in nvkm_chan_new_()
[all …]
H A Dgf100.c87 const u64 userd = nvkm_memory_addr(chan->userd.mem) + chan->userd.base; in gf100_chan_ramfc_write() local
91 nvkm_wo32(chan->inst, 0x08, lower_32_bits(userd)); in gf100_chan_ramfc_write()
92 nvkm_wo32(chan->inst, 0x0c, upper_32_bits(userd)); in gf100_chan_ramfc_write()
120 nvkm_kmap(chan->userd.mem); in gf100_chan_userd_clear()
121 nvkm_wo32(chan->userd.mem, chan->userd.base + 0x040, 0x00000000); in gf100_chan_userd_clear()
122 nvkm_wo32(chan->userd.mem, chan->userd.base + 0x044, 0x00000000); in gf100_chan_userd_clear()
123 nvkm_wo32(chan->userd.mem, chan->userd.base + 0x048, 0x00000000); in gf100_chan_userd_clear()
124 nvkm_wo32(chan->userd.mem, chan->userd.base + 0x04c, 0x00000000); in gf100_chan_userd_clear()
125 nvkm_wo32(chan->userd.mem, chan->userd.base + 0x050, 0x00000000); in gf100_chan_userd_clear()
126 nvkm_wo32(chan->userd.mem, chan->userd.base + 0x058, 0x00000000); in gf100_chan_userd_clear()
[all …]
H A Dbase.c306 if (fifo->func->chan.func->userd->bar == 1) { in nvkm_fifo_oneinit()
310 fifo->func->chan.func->userd->size, 0, true, in nvkm_fifo_oneinit()
311 &fifo->userd.mem); in nvkm_fifo_oneinit()
315 ret = nvkm_vmm_get(bar1, 12, nvkm_memory_size(fifo->userd.mem), &fifo->userd.bar1); in nvkm_fifo_oneinit()
319 ret = nvkm_memory_map(fifo->userd.mem, 0, bar1, fifo->userd.bar1, NULL, 0); in nvkm_fifo_oneinit()
340 if (fifo->userd.bar1) in nvkm_fifo_dtor()
341 nvkm_vmm_put(nvkm_bar_bar1_vmm(engine->subdev.device), &fifo->userd.bar1); in nvkm_fifo_dtor()
342 nvkm_memory_unref(&fifo->userd.mem); in nvkm_fifo_dtor()
352 mutex_destroy(&fifo->userd.mutex); in nvkm_fifo_dtor()
394 INIT_LIST_HEAD(&fifo->userd.list); in nvkm_fifo_new_()
[all …]
H A Duchan.c261 if (chan->func->userd->bar < 0) in nvkm_uchan_map()
265 *addr = device->func->resource_addr(device, chan->func->userd->bar) + in nvkm_uchan_map()
266 chan->func->userd->base + chan->userd.base; in nvkm_uchan_map()
267 *size = chan->func->userd->size; in nvkm_uchan_map()
337 struct nvkm_memory *userd = NULL; in nvkm_uchan_new() local
369 userd = nvkm_umem_search(oclass->client, args->v0.huserd); in nvkm_uchan_new()
370 if (IS_ERR(userd)) { in nvkm_uchan_new()
371 ret = PTR_ERR(userd); in nvkm_uchan_new()
372 userd = NULL; in nvkm_uchan_new()
388 args->v0.length, userd, args->v0.ouserd, &uchan->chan); in nvkm_uchan_new()
[all …]
H A Dgv100.c43 const u64 userd = nvkm_memory_addr(chan->userd.mem) + chan->userd.base; in gv100_chan_ramfc_write() local
47 nvkm_wo32(chan->inst, 0x008, lower_32_bits(userd)); in gv100_chan_ramfc_write()
48 nvkm_wo32(chan->inst, 0x00c, upper_32_bits(userd)); in gv100_chan_ramfc_write()
81 .userd = &gv100_chan_userd,
185 const u64 user = nvkm_memory_addr(chan->userd.mem) + chan->userd.base; in gv100_runl_insert_chan()
H A Dchan.h20 int (*id_get)(struct nvkm_chan *, struct nvkm_memory *userd, u64 ouserd);
34 } *userd; member
61 u64 offset, u64 length, struct nvkm_memory *userd, u64 userd_bar1,
H A Dgk104.c84 const u64 userd = nvkm_memory_addr(chan->userd.mem) + chan->userd.base; in gk104_chan_ramfc_write() local
88 nvkm_wo32(chan->inst, 0x08, lower_32_bits(userd)); in gk104_chan_ramfc_write()
89 nvkm_wo32(chan->inst, 0x0c, upper_32_bits(userd)); in gk104_chan_ramfc_write()
124 .userd = &gk104_chan_userd,
748 if (fifo->func->chan.func->userd->bar == 1) in gk104_fifo_init()
749 nvkm_wr32(device, 0x002254, 0x10000000 | fifo->userd.bar1->addr >> 12); in gk104_fifo_init()
H A Dnv10.c80 .userd = &nv04_chan_userd,
H A Dgk110.c51 .userd = &gk104_chan_userd,
H A Dnv17.c86 .userd = &nv04_chan_userd,
H A Dgm107.c36 .userd = &gk104_chan_userd,
H A Dnv40.c104 .userd = &nv40_chan_userd,
H A Dtu102.c52 .userd = &gv100_chan_userd,
H A Dg84.c97 .userd = &nv50_chan_userd,
H A Dnv50.c141 .userd = &nv50_chan_userd,
H A Dnv04.c170 .userd = &nv04_chan_userd,
H A Dga100.c100 .userd = &gv100_chan_userd,
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_dma.c45 val = nvif_rd32(chan->userd, chan->user_get); in READ_GET()
47 val |= (uint64_t)nvif_rd32(chan->userd, chan->user_get_hi) << 32; in READ_GET()
92 nvif_wr32(chan->userd, 0x8c, chan->dma.ib_put); in nv50_dma_push()
104 uint32_t get = nvif_rd32(chan->userd, 0x88); in nv50_dma_push_wait()
H A Dnouveau_chan.h18 struct nvif_object *userd; member
H A Dnouveau_chan.c336 chan->userd = &chan->mem_userd.object; in nouveau_channel_ctor()
338 chan->userd = &chan->user; in nouveau_channel_ctor()
367 ret = nvif_object_map(chan->userd, NULL, 0); in nouveau_channel_init()
/linux/drivers/gpu/drm/nouveau/include/nvkm/engine/
H A Dfifo.h26 } userd; member
84 } userd; member