Home
last modified time | relevance | path

Searched refs:length_dw (Results 1 – 25 of 33) sorted by relevance

12

/linux/drivers/gpu/drm/radeon/
H A Dsi_dma.c78 ib->ptr[ib->length_dw++] = DMA_PACKET(DMA_PACKET_COPY, in si_dma_vm_copy_pages()
80 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in si_dma_vm_copy_pages()
81 ib->ptr[ib->length_dw++] = lower_32_bits(src); in si_dma_vm_copy_pages()
82 ib->ptr[ib->length_dw++] = upper_32_bits(pe) & 0xff; in si_dma_vm_copy_pages()
83 ib->ptr[ib->length_dw++] = upper_32_bits(src) & 0xff; in si_dma_vm_copy_pages()
119 ib->ptr[ib->length_dw++] = DMA_PACKET(DMA_PACKET_WRITE, 0, 0, 0, ndw); in si_dma_vm_write_pages()
120 ib->ptr[ib->length_dw++] = pe; in si_dma_vm_write_pages()
121 ib->ptr[ib->length_dw++] = upper_32_bits(pe) & 0xff; in si_dma_vm_write_pages()
132 ib->ptr[ib->length_dw++] = value; in si_dma_vm_write_pages()
133 ib->ptr[ib->length_dw++] = upper_32_bits(value); in si_dma_vm_write_pages()
[all …]
H A Dni_dma.c145 radeon_ring_write(ring, (ib->length_dw << 12) | (upper_32_bits(ib->gpu_addr) & 0xFF)); in cayman_dma_ring_ib_execute()
326 ib->ptr[ib->length_dw++] = DMA_PACKET(DMA_PACKET_COPY, in cayman_dma_vm_copy_pages()
328 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in cayman_dma_vm_copy_pages()
329 ib->ptr[ib->length_dw++] = lower_32_bits(src); in cayman_dma_vm_copy_pages()
330 ib->ptr[ib->length_dw++] = upper_32_bits(pe) & 0xff; in cayman_dma_vm_copy_pages()
331 ib->ptr[ib->length_dw++] = upper_32_bits(src) & 0xff; in cayman_dma_vm_copy_pages()
367 ib->ptr[ib->length_dw++] = DMA_PACKET(DMA_PACKET_WRITE, in cayman_dma_vm_write_pages()
369 ib->ptr[ib->length_dw++] = pe; in cayman_dma_vm_write_pages()
370 ib->ptr[ib->length_dw++] = upper_32_bits(pe) & 0xff; in cayman_dma_vm_write_pages()
381 ib->ptr[ib->length_dw++] = value; in cayman_dma_vm_write_pages()
[all …]
H A Dcik_sdma.c156 radeon_ring_write(ring, ib->length_dw); in cik_sdma_ring_ib_execute()
731 ib.length_dw = 5; in cik_sdma_ib_test()
812 ib->ptr[ib->length_dw++] = SDMA_PACKET(SDMA_OPCODE_COPY, in cik_sdma_vm_copy_pages()
814 ib->ptr[ib->length_dw++] = bytes; in cik_sdma_vm_copy_pages()
815 ib->ptr[ib->length_dw++] = 0; /* src/dst endian swap */ in cik_sdma_vm_copy_pages()
816 ib->ptr[ib->length_dw++] = lower_32_bits(src); in cik_sdma_vm_copy_pages()
817 ib->ptr[ib->length_dw++] = upper_32_bits(src); in cik_sdma_vm_copy_pages()
818 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in cik_sdma_vm_copy_pages()
819 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in cik_sdma_vm_copy_pages()
855 ib->ptr[ib->length_dw++] = SDMA_PACKET(SDMA_OPCODE_WRITE, in cik_sdma_vm_write_pages()
[all …]
H A Dradeon_vm.c410 ib.length_dw = 0; in radeon_vm_clear_bo()
414 WARN_ON(ib.length_dw > 64); in radeon_vm_clear_bo()
662 ib.length_dw = 0; in radeon_vm_update_page_directory()
699 if (ib.length_dw != 0) { in radeon_vm_update_page_directory()
703 WARN_ON(ib.length_dw > ndw); in radeon_vm_update_page_directory()
1000 ib.length_dw = 0; in radeon_vm_bo_update()
1018 WARN_ON(ib.length_dw > ndw); in radeon_vm_bo_update()
H A Dr600_dma.c362 ib.length_dw = 4; in r600_dma_ib_test()
426 radeon_ring_write(ring, (ib->length_dw << 16) | (upper_32_bits(ib->gpu_addr) & 0xFF)); in r600_dma_ring_ib_execute()
H A Devergreen_dma.c89 radeon_ring_write(ring, (ib->length_dw << 12) | (upper_32_bits(ib->gpu_addr) & 0xFF)); in evergreen_dma_ring_ib_execute()
H A Dradeon_ib.c133 if (!ib->length_dw || !ring->ready) { in radeon_ib_schedule()
H A Dradeon_trace.h41 __entry->dw = p->chunk_ib->length_dw;
H A Duvd_v1_0.c489 radeon_ring_write(ring, ib->length_dw); in uvd_v1_0_ib_execute()
/linux/drivers/gpu/drm/amd/amdgpu/
H A Dsdma_v6_0.c287 amdgpu_ring_write(ring, ib->length_dw); in sdma_v6_0_ring_emit_ib()
1017 ib.length_dw = 8; in sdma_v6_0_ring_test_ib()
1065 ib->ptr[ib->length_dw++] = SDMA_PKT_COPY_LINEAR_HEADER_OP(SDMA_OP_COPY) | in sdma_v6_0_vm_copy_pte()
1067 ib->ptr[ib->length_dw++] = bytes - 1; in sdma_v6_0_vm_copy_pte()
1068 ib->ptr[ib->length_dw++] = 0; /* src/dst endian swap */ in sdma_v6_0_vm_copy_pte()
1069 ib->ptr[ib->length_dw++] = lower_32_bits(src); in sdma_v6_0_vm_copy_pte()
1070 ib->ptr[ib->length_dw++] = upper_32_bits(src); in sdma_v6_0_vm_copy_pte()
1071 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in sdma_v6_0_vm_copy_pte()
1072 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in sdma_v6_0_vm_copy_pte()
1093 ib->ptr[ib->length_dw++] = SDMA_PKT_COPY_LINEAR_HEADER_OP(SDMA_OP_WRITE) | in sdma_v6_0_vm_write_pte()
[all …]
H A Damdgpu_cs.c214 p->chunks[i].length_dw = user_chunk.length_dw; in amdgpu_cs_pass1()
216 size = p->chunks[i].length_dw; in amdgpu_cs_pass1()
398 ib->length_dw = chunk_ib->ib_bytes / 4; in amdgpu_cs_p2_ib()
411 num_deps = chunk->length_dw * 4 / in amdgpu_cs_p2_dependencies()
482 num_deps = chunk->length_dw * 4 / in amdgpu_cs_p2_syncobj_in()
500 num_deps = chunk->length_dw * 4 / in amdgpu_cs_p2_syncobj_timeline_wait()
520 num_deps = chunk->length_dw * 4 / in amdgpu_cs_p2_syncobj_out()
554 num_deps = chunk->length_dw * 4 / in amdgpu_cs_p2_syncobj_timeline_signal()
1045 if ((va_start + ib->length_dw * 4) > in amdgpu_cs_patch_ibs()
1059 memcpy(ib->ptr, kptr, ib->length_dw * 4); in amdgpu_cs_patch_ibs()
H A Damdgpu_cs.h39 uint32_t length_dw; member
H A Dvcn_sw_ring.c53 amdgpu_ring_write(ring, ib->length_dw); in vcn_dec_sw_ring_emit_ib()
H A Damdgpu_ring.c164 u32 count = ib->length_dw & align_mask; in amdgpu_ring_generic_pad_ib()
169 memset32(&ib->ptr[ib->length_dw], ring->funcs->nop, count); in amdgpu_ring_generic_pad_ib()
171 ib->length_dw += count; in amdgpu_ring_generic_pad_ib()
H A Damdgpu_jpeg.c212 ib->length_dw = 16; in amdgpu_jpeg_dec_set_reg()
571 for (i = 0; i < ib->length_dw ; i += 2) { in amdgpu_jpeg_dec_parse_cs()
H A Damdgpu_vpe.c538 amdgpu_ring_write(ring, ib->length_dw); in vpe_ring_emit_ib()
847 ib.length_dw = 8; in vpe_ring_test_ib()
H A Damdgpu_ttm.c243 WARN_ON(job->ibs[0].length_dw > num_dw); in amdgpu_ttm_map_buffer()
1501 WARN_ON(job->ibs[0].length_dw > num_dw); in amdgpu_ttm_access_memory_sdma()
2282 WARN_ON(job->ibs[0].length_dw > num_dw); in amdgpu_copy_buffer()
2331 WARN_ON(job->ibs[0].length_dw > num_dw); in amdgpu_ttm_fill_mem()
H A Damdgpu_ring.h102 uint32_t length_dw; member
H A Damdgpu_amdkfd.c678 ib->length_dw = ib_len; in amdgpu_amdkfd_submit_ib()
H A Dgfx_v7_0.c2223 control |= ib->length_dw | (vmid << 24); in gfx_v7_0_ring_emit_ib_gfx()
2241 u32 control = INDIRECT_BUFFER_VALID | ib->length_dw | (vmid << 24); in gfx_v7_0_ring_emit_ib_compute()
2316 ib.length_dw = 3; in gfx_v7_0_ring_test_ib()
H A Damdgpu_gmc.c702 job->ibs->ptr[job->ibs->length_dw++] = ring->funcs->nop; in amdgpu_gmc_flush_gpu_tlb()
H A Dgfx_v6_0.c1872 control |= ib->length_dw | (vmid << 24); in gfx_v6_0_ring_emit_ib()
1911 ib.length_dw = 3; in gfx_v6_0_ring_test_ib()
/linux/include/uapi/drm/
H A Dradeon_drm.h969 __u32 length_dw; member
H A Damdgpu_drm.h965 __u32 length_dw; member
/linux/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_migrate.c83 WARN_ON(job->ibs[0].length_dw > num_dw); in svm_migrate_gart_map()

12