/linux/drivers/gpu/drm/radeon/ |
H A D | radeon_gem.c | 41 struct dma_buf *radeon_gem_prime_export(struct drm_gem_object *gobj, 83 static void radeon_gem_object_free(struct drm_gem_object *gobj) in radeon_gem_object_free() argument 85 struct radeon_bo *robj = gem_to_radeon_bo(gobj); in radeon_gem_object_free() 142 static int radeon_gem_set_domain(struct drm_gem_object *gobj, in radeon_gem_set_domain() argument 150 robj = gem_to_radeon_bo(gobj); in radeon_gem_set_domain() 317 struct drm_gem_object *gobj; in radeon_gem_create_ioctl() local 326 false, &gobj); in radeon_gem_create_ioctl() 332 r = drm_gem_handle_create(filp, gobj, &handle); in radeon_gem_create_ioctl() 334 drm_gem_object_put(gobj); in radeon_gem_create_ioctl() 351 struct drm_gem_object *gobj; in radeon_gem_userptr_ioctl() local [all …]
|
H A D | radeon_fbdev.c | 41 static void radeon_fbdev_destroy_pinned_object(struct drm_gem_object *gobj) in radeon_fbdev_destroy_pinned_object() argument 43 struct radeon_bo *rbo = gem_to_radeon_bo(gobj); in radeon_fbdev_destroy_pinned_object() 52 drm_gem_object_put(gobj); in radeon_fbdev_destroy_pinned_object() 61 struct drm_gem_object *gobj = NULL; in radeon_fbdev_create_pinned_object() local 83 0, true, &gobj); in radeon_fbdev_create_pinned_object() 88 rbo = gem_to_radeon_bo(gobj); in radeon_fbdev_create_pinned_object() 132 *gobj_p = gobj; in radeon_fbdev_create_pinned_object() 136 radeon_fbdev_destroy_pinned_object(gobj); in radeon_fbdev_create_pinned_object() 178 struct drm_gem_object *gobj = drm_gem_fb_get_obj(fb, 0); in radeon_fbdev_fb_destroy() local 185 radeon_fbdev_destroy_pinned_object(gobj); in radeon_fbdev_fb_destroy() [all …]
|
H A D | radeon_prime.c | 94 struct dma_buf *radeon_gem_prime_export(struct drm_gem_object *gobj, in radeon_gem_prime_export() argument 97 struct radeon_bo *bo = gem_to_radeon_bo(gobj); in radeon_gem_prime_export() 100 return drm_gem_prime_export(gobj, flags); in radeon_gem_prime_export()
|
H A D | radeon_prime.h | 29 struct dma_buf *radeon_gem_prime_export(struct drm_gem_object *gobj,
|
/linux/drivers/gpu/drm/tests/ |
H A D | drm_exec_test.c | 57 struct drm_gem_object gobj = { }; in test_lock() local 61 drm_gem_private_object_init(priv->drm, &gobj, PAGE_SIZE); in test_lock() 65 ret = drm_exec_lock_obj(&exec, &gobj); in test_lock() 77 struct drm_gem_object gobj = { }; in test_lock_unlock() local 81 drm_gem_private_object_init(priv->drm, &gobj, PAGE_SIZE); in test_lock_unlock() 85 ret = drm_exec_lock_obj(&exec, &gobj); in test_lock_unlock() 91 drm_exec_unlock_obj(&exec, &gobj); in test_lock_unlock() 92 ret = drm_exec_lock_obj(&exec, &gobj); in test_lock_unlock() 104 struct drm_gem_object gobj = { }; in test_duplicates() local 108 drm_gem_private_object_init(priv->drm, &gobj, PAGE_SIZE); in test_duplicates() [all …]
|
/linux/drivers/gpu/drm/loongson/ |
H A D | lsdc_gem.c | 148 struct drm_gem_object *gobj; in lsdc_gem_object_create() local 163 gobj = &lbo->tbo.base; in lsdc_gem_object_create() 164 gobj->funcs = &lsdc_gem_object_funcs; in lsdc_gem_object_create() 171 return gobj; in lsdc_gem_object_create() 181 struct drm_gem_object *gobj; in lsdc_prime_import_sg_table() local 185 gobj = lsdc_gem_object_create(ddev, LSDC_GEM_DOMAIN_GTT, size, false, in lsdc_prime_import_sg_table() 189 if (IS_ERR(gobj)) { in lsdc_prime_import_sg_table() 191 return gobj; in lsdc_prime_import_sg_table() 194 lbo = gem_to_lsdc_bo(gobj); in lsdc_prime_import_sg_table() 197 return gobj; in lsdc_prime_import_sg_table() [all …]
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_gem.c | 87 static void amdgpu_gem_object_free(struct drm_gem_object *gobj) in amdgpu_gem_object_free() argument 89 struct amdgpu_bo *aobj = gem_to_amdgpu_bo(gobj); in amdgpu_gem_object_free() 138 struct drm_gem_object *gobj; in amdgpu_gem_force_release() local 143 idr_for_each_entry(&file->object_idr, gobj, handle) { in amdgpu_gem_force_release() 145 drm_gem_object_put(gobj); in amdgpu_gem_force_release() 320 struct drm_gem_object *gobj; in amdgpu_gem_create_ioctl() local 377 flags, ttm_bo_type_device, resv, &gobj, fpriv->xcp_id + 1); in amdgpu_gem_create_ioctl() 394 struct amdgpu_bo *abo = gem_to_amdgpu_bo(gobj); in amdgpu_gem_create_ioctl() 403 r = drm_gem_handle_create(filp, gobj, &handle); in amdgpu_gem_create_ioctl() 405 drm_gem_object_put(gobj); in amdgpu_gem_create_ioctl() [all …]
|
H A D | amdgpu_dma_buf.c | 260 struct dma_buf *amdgpu_gem_prime_export(struct drm_gem_object *gobj, in amdgpu_gem_prime_export() argument 263 struct amdgpu_bo *bo = gem_to_amdgpu_bo(gobj); in amdgpu_gem_prime_export() 270 buf = drm_gem_prime_export(gobj, flags); in amdgpu_gem_prime_export() 294 struct drm_gem_object *gobj; in amdgpu_dma_buf_create_obj() local 312 ttm_bo_type_sg, resv, &gobj, 0); in amdgpu_dma_buf_create_obj() 316 bo = gem_to_amdgpu_bo(gobj); in amdgpu_dma_buf_create_obj() 321 return gobj; in amdgpu_dma_buf_create_obj() 460 struct drm_gem_object *gobj; in amdgpu_dmabuf_is_xgmi_accessible() local 469 gobj = dma_buf->priv; in amdgpu_dmabuf_is_xgmi_accessible() 470 bo = gem_to_amdgpu_bo(gobj); in amdgpu_dmabuf_is_xgmi_accessible()
|
H A D | amdgpu_bo_list.c | 92 struct drm_gem_object *gobj; in amdgpu_bo_list_create() local 96 gobj = drm_gem_object_lookup(filp, info[i].bo_handle); in amdgpu_bo_list_create() 97 if (!gobj) { in amdgpu_bo_list_create() 102 bo = amdgpu_bo_ref(gem_to_amdgpu_bo(gobj)); in amdgpu_bo_list_create() 103 drm_gem_object_put(gobj); in amdgpu_bo_list_create()
|
H A D | amdgpu_gem.h | 34 #define gem_to_amdgpu_bo(gobj) container_of((gobj), struct amdgpu_bo, tbo.base) argument
|
H A D | amdgpu_amdkfd_gpuvm.c | 838 struct drm_gem_object *gobj; in kfd_mem_attach_dmabuf() local 845 gobj = amdgpu_gem_prime_import(adev_to_drm(adev), mem->dmabuf); in kfd_mem_attach_dmabuf() 846 if (IS_ERR(gobj)) in kfd_mem_attach_dmabuf() 847 return PTR_ERR(gobj); in kfd_mem_attach_dmabuf() 849 *bo = gem_to_amdgpu_bo(gobj); in kfd_mem_attach_dmabuf() 1726 struct drm_gem_object *gobj = NULL; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() local 1819 bo_type, NULL, &gobj, xcp_id + 1); in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1825 ret = drm_vma_node_allow(&gobj->vma_node, drm_priv); in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1830 ret = drm_gem_handle_create(adev->kfd.client.file, gobj, &(*mem)->gem_handle); in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1833 bo = gem_to_amdgpu_bo(gobj); in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() [all …]
|
H A D | amdgpu_dma_buf.h | 28 struct dma_buf *amdgpu_gem_prime_export(struct drm_gem_object *gobj,
|
H A D | amdgpu_cs.c | 129 struct drm_gem_object *gobj; in amdgpu_cs_p1_user_fence() local 132 gobj = drm_gem_object_lookup(p->filp, data->handle); in amdgpu_cs_p1_user_fence() 133 if (gobj == NULL) in amdgpu_cs_p1_user_fence() 136 p->uf_bo = amdgpu_bo_ref(gem_to_amdgpu_bo(gobj)); in amdgpu_cs_p1_user_fence() 137 drm_gem_object_put(gobj); in amdgpu_cs_p1_user_fence() 1282 struct drm_gem_object *gobj; in amdgpu_cs_submit() local 1333 drm_exec_for_each_locked_object(&p->exec, index, gobj) { in amdgpu_cs_submit() 1335 ttm_bo_move_to_lru_tail_unlocked(&gem_to_amdgpu_bo(gobj)->tbo); in amdgpu_cs_submit() 1342 dma_resv_add_fence(gobj->resv, in amdgpu_cs_submit() 1348 dma_resv_add_fence(gobj->resv, p->fence, DMA_RESV_USAGE_WRITE); in amdgpu_cs_submit()
|
/linux/drivers/accel/amdxdna/ |
H A D | amdxdna_gem.c | 58 static void amdxdna_gem_obj_free(struct drm_gem_object *gobj) in amdxdna_gem_obj_free() argument 60 struct amdxdna_dev *xdna = to_xdna_dev(gobj->dev); in amdxdna_gem_obj_free() 61 struct amdxdna_gem_obj *abo = to_xdna_obj(gobj); in amdxdna_gem_obj_free() 75 drm_gem_object_release(gobj); in amdxdna_gem_obj_free() 84 drm_gem_vunmap_unlocked(gobj, &map); in amdxdna_gem_obj_free() 161 static int amdxdna_gem_obj_mmap(struct drm_gem_object *gobj, in amdxdna_gem_obj_mmap() argument 164 struct amdxdna_gem_obj *abo = to_xdna_obj(gobj); in amdxdna_gem_obj_mmap() 168 ret = amdxdna_hmm_register(abo, vma->vm_start, gobj->size); in amdxdna_gem_obj_mmap() 176 num_pages = gobj->size >> PAGE_SHIFT; in amdxdna_gem_obj_mmap() 202 struct drm_gem_object *gobj = vma->vm_private_data; in amdxdna_gem_vm_close() local [all …]
|
H A D | amdxdna_ctx.c | 357 struct drm_gem_object *gobj; in amdxdna_arg_bos_lookup() local 364 gobj = drm_gem_object_lookup(client->filp, bo_hdls[i]); in amdxdna_arg_bos_lookup() 365 if (!gobj) { in amdxdna_arg_bos_lookup() 369 abo = to_xdna_obj(gobj); in amdxdna_arg_bos_lookup() 374 job->bos[i] = gobj; in amdxdna_arg_bos_lookup() 381 drm_gem_object_put(gobj); in amdxdna_arg_bos_lookup() 387 job->bos[i] = gobj; in amdxdna_arg_bos_lookup()
|
H A D | amdxdna_gem.h | 38 static inline struct amdxdna_gem_obj *to_xdna_obj(struct drm_gem_object *gobj) in to_xdna_obj() argument 40 return container_of(gobj, struct amdxdna_gem_obj, base.base); in to_xdna_obj()
|
H A D | aie2_message.c | 383 struct drm_gem_object *gobj; in aie2_config_cu() local 401 gobj = drm_gem_object_lookup(hwctx->client->filp, cu->cu_bo); in aie2_config_cu() 402 if (!gobj) { in aie2_config_cu() 406 abo = to_xdna_obj(gobj); in aie2_config_cu() 409 drm_gem_object_put(gobj); in aie2_config_cu() 419 drm_gem_object_put(gobj); in aie2_config_cu()
|
/linux/drivers/gpu/drm/qxl/ |
H A D | qxl_gem.c | 31 void qxl_gem_object_free(struct drm_gem_object *gobj) in qxl_gem_object_free() argument 33 struct qxl_bo *qobj = gem_to_qxl_bo(gobj); in qxl_gem_object_free() 37 qdev = to_qxl(gobj->dev); in qxl_gem_object_free() 86 struct drm_gem_object **gobj, in qxl_gem_object_create_with_handle() argument 104 if (gobj) in qxl_gem_object_create_with_handle() 105 *gobj = local_gobj; in qxl_gem_object_create_with_handle()
|
H A D | qxl_dumb.c | 37 struct drm_gem_object *gobj; in qxl_mode_dumb_create() local 66 args->size, &surf, &gobj, in qxl_mode_dumb_create() 70 qobj = gem_to_qxl_bo(gobj); in qxl_mode_dumb_create() 72 drm_gem_object_put(gobj); in qxl_mode_dumb_create()
|
H A D | qxl_ioctl.c | 114 struct drm_gem_object *gobj; in qxlhw_handle_to_bo() local 118 gobj = drm_gem_object_lookup(file_priv, handle); in qxlhw_handle_to_bo() 119 if (!gobj) in qxlhw_handle_to_bo() 122 qobj = gem_to_qxl_bo(gobj); in qxlhw_handle_to_bo() 125 drm_gem_object_put(gobj); in qxlhw_handle_to_bo() 303 struct drm_gem_object *gobj = NULL; in qxl_update_area_ioctl() local 311 gobj = drm_gem_object_lookup(file, update_area->handle); in qxl_update_area_ioctl() 312 if (gobj == NULL) in qxl_update_area_ioctl() 315 qobj = gem_to_qxl_bo(gobj); in qxl_update_area_ioctl() 339 drm_gem_object_put(gobj); in qxl_update_area_ioctl()
|
H A D | qxl_drv.h | 94 #define gem_to_qxl_bo(gobj) container_of((gobj), struct qxl_bo, tbo.base) argument 310 struct drm_gem_object **gobj, 312 void qxl_gem_object_free(struct drm_gem_object *gobj);
|
/linux/drivers/gpu/drm/nouveau/ |
H A D | nouveau_prime.c | 107 struct dma_buf *nouveau_gem_prime_export(struct drm_gem_object *gobj, in nouveau_gem_prime_export() argument 110 struct nouveau_bo *nvbo = nouveau_gem_object(gobj); in nouveau_gem_prime_export() 115 return drm_gem_prime_export(gobj, flags); in nouveau_gem_prime_export()
|
/linux/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_gem.c | 36 static void vmw_gem_object_free(struct drm_gem_object *gobj) in vmw_gem_object_free() argument 38 struct ttm_buffer_object *bo = drm_gem_ttm_of_gem(gobj); in vmw_gem_object_free() 322 struct drm_gem_object *gobj; in vmw_debugfs_gem_info_show() local 340 idr_for_each_entry(&file->object_idr, gobj, id) { in vmw_debugfs_gem_info_show() 341 struct vmw_bo *bo = to_vmw_bo(gobj); in vmw_debugfs_gem_info_show()
|
H A D | vmwgfx_bo.h | 231 static inline struct vmw_bo *to_vmw_bo(struct drm_gem_object *gobj) in to_vmw_bo() argument 233 return container_of((gobj), struct vmw_bo, tbo.base); in to_vmw_bo()
|
/linux/drivers/gpu/drm/virtio/ |
H A D | virtgpu_ioctl.c | 205 struct drm_gem_object *gobj = NULL; in virtio_gpu_resource_info_ioctl() local 208 gobj = drm_gem_object_lookup(file, ri->bo_handle); in virtio_gpu_resource_info_ioctl() 209 if (gobj == NULL) in virtio_gpu_resource_info_ioctl() 212 qobj = gem_to_virtio_gpu_obj(gobj); in virtio_gpu_resource_info_ioctl() 219 drm_gem_object_put(gobj); in virtio_gpu_resource_info_ioctl()
|