| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_gem.c | 584 struct amdgpu_bo *robj; in amdgpu_mode_dumb_mmap() local 590 robj = gem_to_amdgpu_bo(gobj); in amdgpu_mode_dumb_mmap() 591 if (amdgpu_ttm_tt_get_usermm(robj->tbo.ttm) || in amdgpu_mode_dumb_mmap() 592 (robj->flags & AMDGPU_GEM_CREATE_NO_CPU_ACCESS)) { in amdgpu_mode_dumb_mmap() 596 *offset_p = amdgpu_bo_mmap_offset(robj); in amdgpu_mode_dumb_mmap() 644 struct amdgpu_bo *robj; in amdgpu_gem_wait_idle_ioctl() local 654 robj = gem_to_amdgpu_bo(gobj); in amdgpu_gem_wait_idle_ioctl() 655 ret = dma_resv_wait_timeout(robj->tbo.base.resv, DMA_RESV_USAGE_READ, in amdgpu_gem_wait_idle_ioctl() 677 struct amdgpu_bo *robj; in amdgpu_gem_metadata_ioctl() local 684 robj = gem_to_amdgpu_bo(gobj); in amdgpu_gem_metadata_ioctl() [all …]
|
| H A D | amdgpu_display.c | 1742 amdgpu_display_robj_is_fb(struct amdgpu_device *adev, struct amdgpu_bo *robj) in amdgpu_display_robj_is_fb() argument 1750 if (gem_to_amdgpu_bo(fb_helper->buffer->gem) != robj) in amdgpu_display_robj_is_fb() 1778 struct amdgpu_bo *robj; in amdgpu_display_suspend_helper() local 1793 robj = gem_to_amdgpu_bo(fb->obj[0]); in amdgpu_display_suspend_helper() 1794 if (!amdgpu_display_robj_is_fb(adev, robj)) { in amdgpu_display_suspend_helper() 1795 r = amdgpu_bo_reserve(robj, true); in amdgpu_display_suspend_helper() 1797 amdgpu_bo_unpin(robj); in amdgpu_display_suspend_helper() 1798 amdgpu_bo_unreserve(robj); in amdgpu_display_suspend_helper()
|
| H A D | amdgpu.h | 669 struct amdgpu_bo *robj; member
|
| H A D | amdgpu_device.c | 1617 &adev->mem_scratch.robj, in amdgpu_device_mem_scratch_init() 1631 amdgpu_bo_free_kernel(&adev->mem_scratch.robj, NULL, NULL); in amdgpu_device_mem_scratch_fini()
|
| /linux/drivers/gpu/drm/radeon/ |
| H A D | evergreen_cs.c | 1223 track->db_z_read_bo = reloc->robj; in evergreen_cs_handle_reg() 1235 track->db_z_write_bo = reloc->robj; in evergreen_cs_handle_reg() 1247 track->db_s_read_bo = reloc->robj; in evergreen_cs_handle_reg() 1259 track->db_s_write_bo = reloc->robj; in evergreen_cs_handle_reg() 1283 track->vgt_strmout_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1507 track->cb_color_fmask_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1524 track->cb_color_cmask_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1565 track->cb_color_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1581 track->cb_color_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1593 track->htile_bo = reloc->robj; in evergreen_cs_handle_reg() [all …]
|
| H A D | r100_track.h | 16 struct radeon_bo *robj; member 23 struct radeon_bo *robj; member 28 struct radeon_bo *robj; member 39 struct radeon_bo *robj; member
|
| H A D | radeon_cursor.c | 287 struct radeon_bo *robj; in radeon_crtc_cursor_set2() local 309 robj = gem_to_radeon_bo(obj); in radeon_crtc_cursor_set2() 310 ret = radeon_bo_reserve(robj, false); in radeon_crtc_cursor_set2() 316 ret = radeon_bo_pin_restricted(robj, RADEON_GEM_DOMAIN_VRAM, in radeon_crtc_cursor_set2() 319 radeon_bo_unreserve(robj); in radeon_crtc_cursor_set2() 351 struct radeon_bo *robj = gem_to_radeon_bo(radeon_crtc->cursor_bo); in radeon_crtc_cursor_set2() local 352 ret = radeon_bo_reserve(robj, false); in radeon_crtc_cursor_set2() 354 radeon_bo_unpin(robj); in radeon_crtc_cursor_set2() 355 radeon_bo_unreserve(robj); in radeon_crtc_cursor_set2()
|
| H A D | r600_cs.c | 1088 track->vgt_strmout_bo[tmp] = reloc->robj; in r600_cs_check_reg() 1215 track->cb_color_frag_bo[tmp] = reloc->robj; in r600_cs_check_reg() 1246 track->cb_color_tile_bo[tmp] = reloc->robj; in r600_cs_check_reg() 1286 track->cb_color_bo[tmp] = reloc->robj; in r600_cs_check_reg() 1299 track->db_bo = reloc->robj; in r600_cs_check_reg() 1312 track->htile_bo = reloc->robj; in r600_cs_check_reg() 1813 if ((tmp + size) > radeon_bo_size(reloc->robj)) { in r600_packet3_check() 1815 tmp + size, radeon_bo_size(reloc->robj)); in r600_packet3_check() 1843 if ((tmp + size) > radeon_bo_size(reloc->robj)) { in r600_packet3_check() 1845 tmp + size, radeon_bo_size(reloc->robj)); in r600_packet3_check() [all …]
|
| H A D | r100.c | 1363 track->arrays[i + 0].robj = reloc->robj; in r100_packet3_load_vbpntr() 1373 track->arrays[i + 1].robj = reloc->robj; in r100_packet3_load_vbpntr() 1387 track->arrays[i + 0].robj = reloc->robj; in r100_packet3_load_vbpntr() 1624 track->zb.robj = reloc->robj; in r100_packet0_check() 1637 track->cb[0].robj = reloc->robj; in r100_packet0_check() 1664 track->textures[i].robj = reloc->robj; in r100_packet0_check() 1682 track->textures[0].cube_info[i].robj = reloc->robj; in r100_packet0_check() 1700 track->textures[1].cube_info[i].robj = reloc->robj; in r100_packet0_check() 1718 track->textures[2].cube_info[i].robj = reloc->robj; in r100_packet0_check() 1923 struct radeon_bo *robj) in r100_cs_track_check_pkt3_indx_buffer() argument [all …]
|
| H A D | radeon_cs.c | 115 p->relocs[i].robj = gem_to_radeon_bo(gobj); in radeon_cs_parser_relocs() 161 if (radeon_ttm_tt_has_userptr(p->rdev, p->relocs[i].robj->tbo.ttm)) { in radeon_cs_parser_relocs() 175 if (p->relocs[i].robj->prime_shared_count) { in radeon_cs_parser_relocs() 255 resv = reloc->robj->tbo.base.resv; in radeon_cs_sync_rings() 400 if (la->robj->tbo.base.size > lb->robj->tbo.base.size) in cmp_size_smaller_first() 402 if (la->robj->tbo.base.size < lb->robj->tbo.base.size) in cmp_size_smaller_first() 434 dma_resv_add_fence(reloc->robj->tbo.base.resv, in radeon_cs_parser_fini() 446 struct radeon_bo *bo = parser->relocs[i].robj; in radeon_cs_parser_fini() 529 bo = p->relocs[i].robj; in radeon_bo_vm_update_pte()
|
| H A D | radeon_device.c | 1577 struct radeon_bo *robj; in radeon_suspend_kms() local 1580 struct radeon_bo *robj = gem_to_radeon_bo(radeon_crtc->cursor_bo); in radeon_suspend_kms() local 1581 r = radeon_bo_reserve(robj, false); in radeon_suspend_kms() 1583 radeon_bo_unpin(robj); in radeon_suspend_kms() 1584 radeon_bo_unreserve(robj); in radeon_suspend_kms() 1591 robj = gem_to_radeon_bo(fb->obj[0]); in radeon_suspend_kms() 1593 if (!radeon_fbdev_robj_is_fb(rdev, robj)) { in radeon_suspend_kms() 1594 r = radeon_bo_reserve(robj, false); in radeon_suspend_kms() 1596 radeon_bo_unpin(robj); in radeon_suspend_kms() 1597 radeon_bo_unreserve(robj); in radeon_suspend_kms() [all …]
|
| H A D | r300.c | 134 if (rdev->gart.robj) { in rv370_pcie_gart_init() 157 if (rdev->gart.robj == NULL) { in rv370_pcie_gart_enable() 672 track->cb[i].robj = reloc->robj; in r300_packet0_check() 685 track->zb.robj = reloc->robj; in r300_packet0_check() 730 track->textures[i].robj = reloc->robj; in r300_packet0_check() 1129 track->aa.robj = reloc->robj; in r300_packet0_check() 1199 r = r100_cs_track_check_pkt3_indx_buffer(p, pkt, reloc->robj); in r300_packet3_check()
|
| H A D | radeon_fbdev.c | 278 bool radeon_fbdev_robj_is_fb(struct radeon_device *rdev, struct radeon_bo *robj) in radeon_fbdev_robj_is_fb() argument 289 if (gobj != &robj->tbo.base) in radeon_fbdev_robj_is_fb()
|
| H A D | r600.c | 1116 if (rdev->gart.robj) { in r600_pcie_gart_init() 1133 if (rdev->gart.robj == NULL) { in r600_pcie_gart_enable() 1511 if (rdev->vram_scratch.robj == NULL) { in r600_vram_scratch_init() 1514 0, NULL, NULL, &rdev->vram_scratch.robj); in r600_vram_scratch_init() 1520 r = radeon_bo_reserve(rdev->vram_scratch.robj, false); in r600_vram_scratch_init() 1523 r = radeon_bo_pin(rdev->vram_scratch.robj, in r600_vram_scratch_init() 1526 radeon_bo_unreserve(rdev->vram_scratch.robj); in r600_vram_scratch_init() 1529 r = radeon_bo_kmap(rdev->vram_scratch.robj, in r600_vram_scratch_init() 1532 radeon_bo_unpin(rdev->vram_scratch.robj); in r600_vram_scratch_init() 1533 radeon_bo_unreserve(rdev->vram_scratch.robj); in r600_vram_scratch_init() [all …]
|
| H A D | radeon_object.c | 478 r = drm_exec_prepare_obj(exec, &lobj->robj->tbo.base, in radeon_bo_list_validate() 487 struct radeon_bo *bo = lobj->robj; in radeon_bo_list_validate()
|
| H A D | radeon_vm.c | 141 list[0].robj = vm->page_directory; in radeon_vm_get_bos() 152 list[idx].robj = vm->page_tables[i].bo; in radeon_vm_get_bos()
|
| H A D | rs600.c | 549 if (rdev->gart.robj) { in rs600_gart_init() 567 if (rdev->gart.robj == NULL) { in rs600_gart_enable()
|
| H A D | rv770.c | 899 if (rdev->gart.robj == NULL) { in rv770_pcie_gart_enable()
|
| H A D | ni.c | 1252 if (rdev->gart.robj == NULL) { in cayman_pcie_gart_enable()
|
| /linux/drivers/gpu/drm/etnaviv/ |
| H A D | etnaviv_gem.c | 449 struct dma_resv *robj = obj->resv; in etnaviv_gem_describe() local 458 r = dma_resv_lock(robj, NULL); in etnaviv_gem_describe() 462 dma_resv_describe(robj, m); in etnaviv_gem_describe() 463 dma_resv_unlock(robj); in etnaviv_gem_describe()
|
| H A D | etnaviv_gem_submit.c | 179 struct dma_resv *robj = bo->obj->base.resv; in submit_fence_sync() local 181 ret = dma_resv_reserve_fences(robj, 1); in submit_fence_sync()
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_bo.c | 346 struct sg_table *sg, struct dma_resv *robj) in nouveau_bo_init() argument 353 .resv = robj, in nouveau_bo_init() 361 sg, robj, nouveau_bo_del_ttm); in nouveau_bo_init() 367 if (!robj) in nouveau_bo_init() 376 struct sg_table *sg, struct dma_resv *robj, in nouveau_bo_new() argument 396 ret = nouveau_bo_init(nvbo, size, align, domain, sg, robj); in nouveau_bo_new()
|
| /linux/drivers/gpu/drm/msm/ |
| H A D | msm_gem.c | 934 struct dma_resv *robj = obj->resv; in msm_gem_describe() local 1010 dma_resv_describe(robj, m); in msm_gem_describe()
|