Home
last modified time | relevance | path

Searched refs:tbo (Results 1 – 25 of 72) sorted by relevance

123

/linux/drivers/gpu/drm/loongson/
H A Dlsdc_ttm.c54 if (lbo->tbo.base.size <= PAGE_SIZE) in lsdc_bo_set_placement()
94 lsdc_ttm_tt_create(struct ttm_buffer_object *tbo, uint32_t page_flags) in lsdc_ttm_tt_create() argument
103 ret = ttm_sg_tt_init(tt, tbo, page_flags, ttm_cached); in lsdc_ttm_tt_create()
140 static void lsdc_bo_evict_flags(struct ttm_buffer_object *tbo, in lsdc_bo_evict_flags() argument
143 struct ttm_resource *resource = tbo->resource; in lsdc_bo_evict_flags()
144 struct lsdc_bo *lbo = to_lsdc_bo(tbo); in lsdc_bo_evict_flags()
159 static int lsdc_bo_move(struct ttm_buffer_object *tbo, in lsdc_bo_move() argument
165 struct drm_device *ddev = tbo->base.dev; in lsdc_bo_move()
166 struct ttm_resource *old_mem = tbo->resource; in lsdc_bo_move()
167 struct lsdc_bo *lbo = to_lsdc_bo(tbo); in lsdc_bo_move()
[all …]
H A Dlsdc_ttm.h24 struct ttm_buffer_object tbo; member
52 static inline struct lsdc_bo *to_lsdc_bo(struct ttm_buffer_object *tbo) in to_lsdc_bo() argument
54 return container_of(tbo, struct lsdc_bo, tbo); in to_lsdc_bo()
59 return container_of(gem, struct lsdc_bo, tbo.base); in gem_to_lsdc_bo()
/linux/drivers/gpu/drm/qxl/
H A Dqxl_object.c32 static void qxl_ttm_bo_destroy(struct ttm_buffer_object *tbo) in qxl_ttm_bo_destroy() argument
37 bo = to_qxl_bo(tbo); in qxl_ttm_bo_destroy()
38 qdev = to_qxl(bo->tbo.base.dev); in qxl_ttm_bo_destroy()
45 drm_gem_object_release(&bo->tbo.base); in qxl_ttm_bo_destroy()
62 if (qbo->tbo.base.size <= PAGE_SIZE) in qxl_ttm_placement_from_domain()
123 r = drm_gem_object_init(&qdev->ddev, &bo->tbo.base, size); in qxl_bo_create()
128 bo->tbo.base.funcs = &qxl_object_funcs; in qxl_bo_create()
138 bo->tbo.priority = priority; in qxl_bo_create()
139 r = ttm_bo_init_reserved(&qdev->mman.bdev, &bo->tbo, type, in qxl_bo_create()
150 ttm_bo_pin(&bo->tbo); in qxl_bo_create()
[all …]
H A Dqxl_object.h34 r = ttm_bo_reserve(&bo->tbo, true, false, NULL); in qxl_bo_reserve()
37 struct drm_device *ddev = bo->tbo.base.dev; in qxl_bo_reserve()
48 ttm_bo_unreserve(&bo->tbo); in qxl_bo_unreserve()
53 return bo->tbo.base.size; in qxl_bo_size()
H A Dqxl_release.c195 if (!bo->tbo.pin_count) { in qxl_release_validate_bo()
197 ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); in qxl_release_validate_bo()
202 ret = dma_resv_reserve_fences(bo->tbo.base.resv, 1); in qxl_release_validate_bo()
207 ret = qxl_bo_check_id(to_qxl(bo->tbo.base.dev), bo); in qxl_release_validate_bo()
228 &entry->bo->tbo.base, in qxl_release_reserve_list()
424 bdev = bo->tbo.bdev; in qxl_release_fence_buffer_objects()
438 dma_resv_add_fence(bo->tbo.base.resv, &release->base, in qxl_release_fence_buffer_objects()
440 ttm_bo_move_to_lru_tail_unlocked(&bo->tbo); in qxl_release_fence_buffer_objects()
H A Dqxl_drv.h73 struct ttm_buffer_object tbo; member
94 #define gem_to_qxl_bo(gobj) container_of((gobj), struct qxl_bo, tbo.base)
95 #define to_qxl_bo(tobj) container_of((tobj), struct qxl_bo, tbo)
284 (bo->tbo.resource->mem_type == TTM_PL_VRAM) in qxl_bo_physical_address()
289 return slot->high_bits | ((bo->tbo.resource->start << PAGE_SHIFT) + offset); in qxl_bo_physical_address()
/linux/drivers/gpu/drm/radeon/
H A Dradeon_object.c52 static void radeon_ttm_bo_destroy(struct ttm_buffer_object *tbo) in radeon_ttm_bo_destroy() argument
56 bo = container_of(tbo, struct radeon_bo, tbo); in radeon_ttm_bo_destroy()
63 if (bo->tbo.base.import_attach) in radeon_ttm_bo_destroy()
64 drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg); in radeon_ttm_bo_destroy()
65 drm_gem_object_release(&bo->tbo.base); in radeon_ttm_bo_destroy()
153 drm_gem_private_object_init(rdev_to_drm(rdev), &bo->tbo.base, size); in radeon_bo_create()
154 bo->tbo.base.funcs = &radeon_gem_object_funcs; in radeon_bo_create()
204 r = ttm_bo_init_validate(&rdev->mman.bdev, &bo->tbo, type, in radeon_bo_create()
223 r = dma_resv_wait_timeout(bo->tbo.base.resv, DMA_RESV_USAGE_KERNEL, in radeon_bo_kmap()
234 r = ttm_bo_kmap(&bo->tbo, 0, PFN_UP(bo->tbo.base.size), &bo->kmap); in radeon_bo_kmap()
[all …]
H A Dradeon_object.h68 r = ttm_bo_reserve(&bo->tbo, !no_intr, false, NULL); in radeon_bo_reserve()
79 ttm_bo_unreserve(&bo->tbo); in radeon_bo_unreserve()
96 rdev = radeon_get_rdev(bo->tbo.bdev); in radeon_bo_gpu_offset()
98 switch (bo->tbo.resource->mem_type) { in radeon_bo_gpu_offset()
107 return (bo->tbo.resource->start << PAGE_SHIFT) + start; in radeon_bo_gpu_offset()
112 return bo->tbo.base.size; in radeon_bo_size()
117 return bo->tbo.base.size / RADEON_GPU_PAGE_SIZE; in radeon_bo_ngpu_pages()
122 return (bo->tbo.page_alignment << PAGE_SHIFT) / RADEON_GPU_PAGE_SIZE; in radeon_bo_gpu_page_alignment()
133 return drm_vma_node_offset_addr(&bo->tbo.base.vma_node); in radeon_bo_mmap_offset()
H A Dradeon_mn.c57 if (!bo->tbo.ttm || !radeon_ttm_tt_is_bound(bo->tbo.bdev, bo->tbo.ttm)) in radeon_mn_invalidate()
69 r = dma_resv_wait_timeout(bo->tbo.base.resv, DMA_RESV_USAGE_BOOKKEEP, in radeon_mn_invalidate()
75 r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); in radeon_mn_invalidate()
H A Dradeon_prime.c41 return drm_prime_pages_to_sg(obj->dev, bo->tbo.ttm->pages, in radeon_gem_prime_get_sg_table()
42 bo->tbo.ttm->num_pages); in radeon_gem_prime_get_sg_table()
61 bo->tbo.base.funcs = &radeon_gem_object_funcs; in radeon_gem_prime_import_sg_table()
68 return &bo->tbo.base; in radeon_gem_prime_import_sg_table()
98 if (radeon_ttm_tt_has_userptr(bo->rdev, bo->tbo.ttm)) in radeon_gem_prime_export()
H A Dradeon_cs.c161 if (radeon_ttm_tt_has_userptr(p->rdev, p->relocs[i].robj->tbo.ttm)) { in radeon_cs_parser_relocs()
255 resv = reloc->robj->tbo.base.resv; in radeon_cs_sync_rings()
400 if (la->robj->tbo.base.size > lb->robj->tbo.base.size) in cmp_size_smaller_first()
402 if (la->robj->tbo.base.size < lb->robj->tbo.base.size) in cmp_size_smaller_first()
434 dma_resv_add_fence(reloc->robj->tbo.base.resv, in radeon_cs_parser_fini()
450 drm_gem_object_put(&bo->tbo.base); in radeon_cs_parser_fini()
522 rdev->ring_tmp_bo.bo->tbo.resource); in radeon_bo_vm_update_pte()
536 r = radeon_vm_bo_update(rdev, bo_va, bo->tbo.resource); in radeon_bo_vm_update_pte()
542 r = dma_resv_reserve_fences(bo->tbo.base.resv, 1); in radeon_bo_vm_update_pte()
/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_dma_buf.c64 return amdgpu_ttm_adev(bo->tbo.bdev); in dma_buf_attach_adev()
84 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_dma_buf_attach()
103 r = dma_resv_lock(bo->tbo.base.resv, NULL); in amdgpu_dma_buf_attach()
109 dma_resv_unlock(bo->tbo.base.resv); in amdgpu_dma_buf_attach()
188 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_dma_buf_map()
192 if (!bo->tbo.pin_count) { in amdgpu_dma_buf_map()
203 r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); in amdgpu_dma_buf_map()
208 switch (bo->tbo.resource->mem_type) { in amdgpu_dma_buf_map()
211 bo->tbo.ttm->pages, in amdgpu_dma_buf_map()
212 bo->tbo.ttm->num_pages); in amdgpu_dma_buf_map()
[all …]
H A Damdgpu_amdkfd_gpuvm.c311 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_amdkfd_release_notify()
346 ret = amdgpu_gem_object_create(adev, mem->bo->tbo.base.size, 1, in create_dmamap_sg_bo()
348 ttm_bo_type_sg, mem->bo->tbo.base.resv, &gem_obj, 0); in create_dmamap_sg_bo()
383 dma_resv_replace_fences(bo->tbo.base.resv, ef->base.context, in amdgpu_amdkfd_remove_eviction_fence()
400 struct dma_resv *resv = &bo->tbo.base._resv; in amdgpu_amdkfd_remove_all_eviction_fences()
423 if (WARN(amdgpu_ttm_tt_get_usermm(bo->tbo.ttm), in amdgpu_amdkfd_bo_validate()
428 if (bo->tbo.pin_count) in amdgpu_amdkfd_bo_validate()
433 ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); in amdgpu_amdkfd_bo_validate()
456 ret = dma_resv_reserve_fences(bo->tbo.base.resv, 1); in amdgpu_amdkfd_bo_validate_and_fence()
460 dma_resv_add_fence(bo->tbo.base.resv, fence, in amdgpu_amdkfd_bo_validate_and_fence()
[all …]
H A Damdgpu_vm.c138 dma_resv_assert_held(vm->root.bo->tbo.base.resv); in amdgpu_vm_assert_locked()
157 if (bo->tbo.type == ttm_bo_type_kernel) in amdgpu_vm_bo_evicted()
285 if (!bo || bo->tbo.type != ttm_bo_type_kernel) in amdgpu_vm_bo_reset_state_machine()
309 dma_resv_assert_held(bo->tbo.base.resv); in amdgpu_vm_update_shared()
311 shared = drm_gem_object_is_shared_for_memory_stats(&bo->tbo.base); in amdgpu_vm_update_shared()
423 base->shared = drm_gem_object_is_shared_for_memory_stats(&bo->tbo.base); in amdgpu_vm_bo_base_init()
424 amdgpu_vm_update_stats_locked(base, bo->tbo.resource, +1); in amdgpu_vm_bo_base_init()
430 dma_resv_assert_held(vm->root.bo->tbo.base.resv); in amdgpu_vm_bo_base_init()
432 ttm_bo_set_bulk_move(&bo->tbo, &vm->lru_bulk_move); in amdgpu_vm_bo_base_init()
433 if (bo->tbo.type == ttm_bo_type_kernel && bo->parent) in amdgpu_vm_bo_base_init()
[all …]
H A Damdgpu_gem.c160 ttm_bo_fini(&aobj->tbo); in amdgpu_gem_object_free()
193 *obj = &bo->tbo.base; in amdgpu_gem_object_create()
230 struct amdgpu_device *adev = amdgpu_ttm_adev(abo->tbo.bdev); in amdgpu_gem_object_open()
238 mm = amdgpu_ttm_tt_get_usermm(abo->tbo.ttm); in amdgpu_gem_object_open()
248 r = drm_exec_prepare_obj(&exec, &abo->tbo.base, 1); in amdgpu_gem_object_open()
317 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_gem_object_close()
328 r = drm_exec_prepare_obj(&exec, &bo->tbo.base, 1); in amdgpu_gem_object_close()
371 if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) in amdgpu_gem_object_mmap()
450 resv = vm->root.bo->tbo.base.resv; in amdgpu_gem_create_ioctl()
534 r = amdgpu_ttm_tt_set_userptr(&bo->tbo, args->addr, args->flags); in amdgpu_gem_userptr_ioctl()
[all …]
H A Damdgpu_ttm.c341 (abo_src->tbo.resource->mem_type == TTM_PL_VRAM)) in amdgpu_ttm_copy_mem_to_mem()
720 struct ttm_tt *ttm = bo->tbo.ttm; in amdgpu_ttm_tt_get_user_pages()
879 struct ttm_buffer_object *tbo, in amdgpu_ttm_gart_bind() argument
882 struct amdgpu_bo *abo = ttm_to_amdgpu_bo(tbo); in amdgpu_ttm_gart_bind()
883 struct ttm_tt *ttm = tbo->ttm; in amdgpu_ttm_gart_bind()
1021 void amdgpu_ttm_recover_gart(struct ttm_buffer_object *tbo) in amdgpu_ttm_recover_gart() argument
1023 struct amdgpu_device *adev = amdgpu_ttm_adev(tbo->bdev); in amdgpu_ttm_recover_gart()
1026 if (!tbo->ttm) in amdgpu_ttm_recover_gart()
1029 flags = amdgpu_ttm_tt_pte_flags(adev, tbo->ttm, tbo->resource); in amdgpu_ttm_recover_gart()
1030 amdgpu_ttm_gart_bind(adev, tbo, flags); in amdgpu_ttm_recover_gart()
[all …]
H A Damdgpu_cs.c144 if (amdgpu_ttm_tt_get_usermm(p->uf_bo->tbo.ttm)) in amdgpu_cs_p1_user_fence()
799 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_cs_bo_validate()
804 .resv = bo->tbo.base.resv in amdgpu_cs_bo_validate()
809 if (bo->tbo.pin_count) in amdgpu_cs_bo_validate()
816 (!bo->tbo.base.dma_buf || in amdgpu_cs_bo_validate()
817 list_empty(&bo->tbo.base.dma_buf->attachments))) { in amdgpu_cs_bo_validate()
837 r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); in amdgpu_cs_bo_validate()
841 amdgpu_res_cpu_visible(adev, bo->tbo.resource)) in amdgpu_cs_bo_validate()
901 for (i = 0; i < bo->tbo.ttm->num_pages; i++) { in amdgpu_cs_parser_bos()
902 if (bo->tbo.ttm->pages[i] != in amdgpu_cs_parser_bos()
[all …]
H A Damdgpu_vram_mgr.c279 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_vram_mgr_bo_visible_size()
280 struct ttm_resource *res = bo->tbo.resource; in amdgpu_vram_mgr_bo_visible_size()
442 struct ttm_buffer_object *tbo, in amdgpu_vram_mgr_new() argument
448 struct amdgpu_bo *bo = ttm_to_amdgpu_bo(tbo); in amdgpu_vram_mgr_new()
465 if (tbo->type != ttm_bo_type_kernel) in amdgpu_vram_mgr_new()
478 tbo->page_alignment); in amdgpu_vram_mgr_new()
485 ttm_resource_init(tbo, place, &vres->base); in amdgpu_vram_mgr_new()
524 if (tbo->page_alignment) in amdgpu_vram_mgr_new()
525 min_block_size = (u64)tbo->page_alignment << PAGE_SHIFT; in amdgpu_vram_mgr_new()
550 tbo->page_alignment); in amdgpu_vram_mgr_new()
H A Damdgpu_gtt_mgr.c116 struct ttm_buffer_object *tbo, in amdgpu_gtt_mgr_new() argument
121 uint32_t num_pages = PFN_UP(tbo->base.size); in amdgpu_gtt_mgr_new()
129 ttm_resource_init(tbo, place, &node->base); in amdgpu_gtt_mgr_new()
139 num_pages, tbo->page_alignment, in amdgpu_gtt_mgr_new()
/linux/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_bo.c45 WARN_ON(kref_read(&vbo->tbo.base.refcount) != 0); in vmw_bo_free()
72 drm_gem_object_release(&vbo->tbo.base); in vmw_bo_free()
93 struct ttm_buffer_object *bo = &buf->tbo; in vmw_bo_pin_in_placement()
129 struct ttm_buffer_object *bo = &buf->tbo; in vmw_bo_pin_in_vram_or_gmr()
198 struct ttm_buffer_object *bo = &buf->tbo; in vmw_bo_pin_in_start_of_vram()
214 buf->tbo.pin_count == 0) { in vmw_bo_pin_in_start_of_vram()
255 struct ttm_buffer_object *bo = &buf->tbo; in vmw_bo_unpin()
302 struct ttm_buffer_object *bo = &vbo->tbo; in vmw_bo_pin_reserved()
347 return vmw_bo_map_and_cache_size(vbo, vbo->tbo.base.size); in vmw_bo_map_and_cache()
352 struct ttm_buffer_object *bo = &vbo->tbo; in vmw_bo_map_and_cache_size()
[all …]
H A Dvmwgfx_cursor_plane.c158 ttm_bo_unpin(&(*vbo)->tbo); in vmw_cursor_mob_destroy()
179 ret = ttm_bo_reserve(&vbo->tbo, true, false, NULL); in vmw_cursor_mob_unmap()
182 ttm_bo_unreserve(&vbo->tbo); in vmw_cursor_mob_unmap()
209 if (vcp->cursor_mobs[i]->tbo.base.size < in vmw_cursor_mob_put()
210 vps->cursor.mob->tbo.base.size) { in vmw_cursor_mob_put()
245 if (vps->cursor.mob->tbo.base.size >= size) in vmw_cursor_mob_get()
253 vcp->cursor_mobs[i]->tbo.base.size >= size) { in vmw_cursor_mob_get()
267 ret = ttm_bo_reserve(&vps->cursor.mob->tbo, false, false, NULL); in vmw_cursor_mob_get()
273 ttm_bo_unreserve(&vps->cursor.mob->tbo); in vmw_cursor_mob_get()
280 ttm_bo_unreserve(&vps->cursor.mob->tbo); in vmw_cursor_mob_get()
[all …]
H A Dvmwgfx_page_dirty.c71 pgoff_t offset = drm_vma_node_start(&vbo->tbo.base.vma_node); in vmw_bo_dirty_scan_pagetable()
72 struct address_space *mapping = vbo->tbo.bdev->dev_mapping; in vmw_bo_dirty_scan_pagetable()
109 unsigned long offset = drm_vma_node_start(&vbo->tbo.base.vma_node); in vmw_bo_dirty_scan_mkwrite()
110 struct address_space *mapping = vbo->tbo.bdev->dev_mapping; in vmw_bo_dirty_scan_mkwrite()
116 num_marked = wp_shared_mapping_range(vbo->tbo.bdev->dev_mapping, in vmw_bo_dirty_scan_mkwrite()
174 unsigned long offset = drm_vma_node_start(&vbo->tbo.base.vma_node); in vmw_bo_dirty_pre_unmap()
175 struct address_space *mapping = vbo->tbo.bdev->dev_mapping; in vmw_bo_dirty_pre_unmap()
198 unsigned long offset = drm_vma_node_start(&vbo->tbo.base.vma_node); in vmw_bo_dirty_unmap()
199 struct address_space *mapping = vbo->tbo.bdev->dev_mapping; in vmw_bo_dirty_unmap()
219 pgoff_t num_pages = PFN_UP(vbo->tbo.resource->size); in vmw_bo_dirty_add()
[all …]
H A Dvmwgfx_blit.c428 container_of(bo->tbo.bdev, struct vmw_private, bdev); in map_external()
432 if (drm_gem_is_imported(&bo->tbo.base)) { in map_external()
433 ret = dma_buf_vmap(bo->tbo.base.dma_buf, map); in map_external()
450 if (drm_gem_is_imported(&bo->tbo.base)) in unmap_external()
451 dma_buf_vunmap(bo->tbo.base.dma_buf, map); in unmap_external()
463 container_of(dst->tbo.bdev, struct vmw_private, bdev); in vmw_external_bo_copy()
464 size_t dst_size = dst->tbo.resource->size; in vmw_external_bo_copy()
465 size_t src_size = src->tbo.resource->size; in vmw_external_bo_copy()
548 struct ttm_buffer_object *src = &vmw_src->tbo; in vmw_bo_cpu_blit()
549 struct ttm_buffer_object *dst = &vmw_dst->tbo; in vmw_bo_cpu_blit()
H A Dvmwgfx_mob.c207 bo = &otable->page_table->pt_bo->tbo; in vmw_takedown_otable_base()
267 &batch->otable_bo->tbo, in vmw_otable_batch_setup()
284 vmw_bo_unpin_unlocked(&batch->otable_bo->tbo); in vmw_otable_batch_setup()
335 struct ttm_buffer_object *bo = &batch->otable_bo->tbo; in vmw_otable_batch_takedown()
501 struct ttm_buffer_object *bo = &mob->pt_bo->tbo; in vmw_mob_pt_setup()
538 vmw_bo_unpin_unlocked(&mob->pt_bo->tbo); in vmw_mob_destroy()
558 struct ttm_buffer_object *bo = &mob->pt_bo->tbo; in vmw_mob_unbind()
650 vmw_bo_unpin_unlocked(&mob->pt_bo->tbo); in vmw_mob_bind()
/linux/drivers/gpu/drm/xe/
H A Dxe_ttm_sys_mgr.c18 struct ttm_buffer_object *tbo; member
29 struct ttm_buffer_object *tbo, in xe_ttm_sys_mgr_new() argument
40 node->tbo = tbo; in xe_ttm_sys_mgr_new()
41 ttm_resource_init(tbo, place, &node->base.base); in xe_ttm_sys_mgr_new()

123