| /linux/drivers/gpu/drm/radeon/ |
| H A D | radeon_fbdev.c | 41 static void radeon_fbdev_destroy_pinned_object(struct drm_gem_object *gobj) in radeon_fbdev_destroy_pinned_object() argument 43 struct radeon_bo *rbo = gem_to_radeon_bo(gobj); in radeon_fbdev_destroy_pinned_object() 52 drm_gem_object_put(gobj); in radeon_fbdev_destroy_pinned_object() 61 struct drm_gem_object *gobj = NULL; in radeon_fbdev_create_pinned_object() local 82 0, true, &gobj); in radeon_fbdev_create_pinned_object() 87 rbo = gem_to_radeon_bo(gobj); in radeon_fbdev_create_pinned_object() 131 *gobj_p = gobj; in radeon_fbdev_create_pinned_object() 135 radeon_fbdev_destroy_pinned_object(gobj); in radeon_fbdev_create_pinned_object() 175 struct drm_gem_object *gobj = drm_gem_fb_get_obj(fb, 0); in radeon_fbdev_fb_destroy() local 182 radeon_fbdev_destroy_pinned_object(gobj); in radeon_fbdev_fb_destroy() [all …]
|
| H A D | radeon_prime.c | 94 struct dma_buf *radeon_gem_prime_export(struct drm_gem_object *gobj, in radeon_gem_prime_export() argument 97 struct radeon_bo *bo = gem_to_radeon_bo(gobj); in radeon_gem_prime_export() 100 return drm_gem_prime_export(gobj, flags); in radeon_gem_prime_export()
|
| H A D | radeon_prime.h | 29 struct dma_buf *radeon_gem_prime_export(struct drm_gem_object *gobj,
|
| H A D | radeon_cs.c | 105 struct drm_gem_object *gobj; in radeon_cs_parser_relocs() local 109 gobj = drm_gem_object_lookup(p->filp, r->handle); in radeon_cs_parser_relocs() 110 if (gobj == NULL) { in radeon_cs_parser_relocs() 115 p->relocs[i].robj = gem_to_radeon_bo(gobj); in radeon_cs_parser_relocs()
|
| /linux/drivers/accel/amdxdna/ |
| H A D | amdxdna_gem.c | 256 static void amdxdna_gem_dev_obj_free(struct drm_gem_object *gobj) in amdxdna_gem_dev_obj_free() argument 258 struct amdxdna_dev *xdna = to_xdna_dev(gobj->dev); in amdxdna_gem_dev_obj_free() 259 struct amdxdna_gem_obj *abo = to_xdna_obj(gobj); in amdxdna_gem_dev_obj_free() 266 drm_gem_object_release(gobj); in amdxdna_gem_dev_obj_free() 326 static int amdxdna_gem_obj_mmap(struct drm_gem_object *gobj, in amdxdna_gem_obj_mmap() argument 329 struct amdxdna_dev *xdna = to_xdna_dev(gobj->dev); in amdxdna_gem_obj_mmap() 330 struct amdxdna_gem_obj *abo = to_xdna_obj(gobj); in amdxdna_gem_obj_mmap() 344 drm_vma_node_offset_addr(&gobj->vma_node), abo->type, in amdxdna_gem_obj_mmap() 345 vma->vm_start, gobj->size); in amdxdna_gem_obj_mmap() 355 struct drm_gem_object *gobj = dma_buf->priv; in amdxdna_gem_dmabuf_mmap() local [all …]
|
| H A D | amdxdna_ctx.c | 336 struct drm_gem_object *gobj; in amdxdna_hwctx_sync_debug_bo() local 342 gobj = drm_gem_object_lookup(client->filp, debug_bo_hdl); in amdxdna_hwctx_sync_debug_bo() 343 if (!gobj) in amdxdna_hwctx_sync_debug_bo() 346 abo = to_xdna_obj(gobj); in amdxdna_hwctx_sync_debug_bo() 359 drm_gem_object_put(gobj); in amdxdna_hwctx_sync_debug_bo() 380 struct drm_gem_object *gobj; in amdxdna_arg_bos_lookup() local 387 gobj = drm_gem_object_lookup(client->filp, bo_hdls[i]); in amdxdna_arg_bos_lookup() 388 if (!gobj) { in amdxdna_arg_bos_lookup() 392 abo = to_xdna_obj(gobj); in amdxdna_arg_bos_lookup() 397 job->bos[i] = gobj; in amdxdna_arg_bos_lookup() [all …]
|
| H A D | aie2_ctx.c | 1068 struct drm_gem_object *gobj = to_gobj(abo); in aie2_hmm_invalidate() local 1071 ret = dma_resv_wait_timeout(gobj->resv, DMA_RESV_USAGE_BOOKKEEP, in aie2_hmm_invalidate()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_gem.c | 155 static void amdgpu_gem_object_free(struct drm_gem_object *gobj) in amdgpu_gem_object_free() argument 157 struct amdgpu_bo *aobj = gem_to_amdgpu_bo(gobj); in amdgpu_gem_object_free() 206 struct drm_gem_object *gobj; in amdgpu_gem_force_release() local 211 idr_for_each_entry(&file->object_idr, gobj, handle) { in amdgpu_gem_force_release() 213 drm_gem_object_put(gobj); in amdgpu_gem_force_release() 412 struct drm_gem_object *gobj; in amdgpu_gem_create_ioctl() local 457 flags, ttm_bo_type_device, resv, &gobj, fpriv->xcp_id + 1); in amdgpu_gem_create_ioctl() 474 struct amdgpu_bo *abo = gem_to_amdgpu_bo(gobj); in amdgpu_gem_create_ioctl() 483 r = drm_gem_handle_create(filp, gobj, &handle); in amdgpu_gem_create_ioctl() 485 drm_gem_object_put(gobj); in amdgpu_gem_create_ioctl() [all …]
|
| H A D | amdgpu_dma_buf.c | 378 struct dma_buf *amdgpu_gem_prime_export(struct drm_gem_object *gobj, in amdgpu_gem_prime_export() argument 381 struct amdgpu_bo *bo = gem_to_amdgpu_bo(gobj); in amdgpu_gem_prime_export() 400 buf = drm_gem_prime_export(gobj, flags); in amdgpu_gem_prime_export() 424 struct drm_gem_object *gobj; in amdgpu_dma_buf_create_obj() local 442 ttm_bo_type_sg, resv, &gobj, 0); in amdgpu_dma_buf_create_obj() 446 bo = gem_to_amdgpu_bo(gobj); in amdgpu_dma_buf_create_obj() 451 return gobj; in amdgpu_dma_buf_create_obj() 597 struct drm_gem_object *gobj; in amdgpu_dmabuf_is_xgmi_accessible() local 609 gobj = dma_buf->priv; in amdgpu_dmabuf_is_xgmi_accessible() 610 bo = gem_to_amdgpu_bo(gobj); in amdgpu_dmabuf_is_xgmi_accessible()
|
| H A D | amdgpu_bo_list.c | 90 struct drm_gem_object *gobj; in amdgpu_bo_list_create() local 94 gobj = drm_gem_object_lookup(filp, info[i].bo_handle); in amdgpu_bo_list_create() 95 if (!gobj) { in amdgpu_bo_list_create() 100 bo = amdgpu_bo_ref(gem_to_amdgpu_bo(gobj)); in amdgpu_bo_list_create() 101 drm_gem_object_put(gobj); in amdgpu_bo_list_create()
|
| H A D | amdgpu_amdkfd_gpuvm.c | 841 struct drm_gem_object *gobj; in kfd_mem_attach_dmabuf() local 848 gobj = amdgpu_gem_prime_import(adev_to_drm(adev), mem->dmabuf); in kfd_mem_attach_dmabuf() 849 if (IS_ERR(gobj)) in kfd_mem_attach_dmabuf() 850 return PTR_ERR(gobj); in kfd_mem_attach_dmabuf() 852 *bo = gem_to_amdgpu_bo(gobj); in kfd_mem_attach_dmabuf() 1716 struct drm_gem_object *gobj = NULL; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() local 1809 bo_type, NULL, &gobj, xcp_id + 1); in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1815 ret = drm_vma_node_allow(&gobj->vma_node, drm_priv); in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1820 ret = drm_gem_handle_create(adev->kfd.client.file, gobj, &(*mem)->gem_handle); in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1823 bo = gem_to_amdgpu_bo(gobj); in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() [all …]
|
| H A D | amdgpu_dma_buf.h | 28 struct dma_buf *amdgpu_gem_prime_export(struct drm_gem_object *gobj,
|
| H A D | amdgpu_cs.c | 130 struct drm_gem_object *gobj; in amdgpu_cs_p1_user_fence() local 133 gobj = drm_gem_object_lookup(p->filp, data->handle); in amdgpu_cs_p1_user_fence() 134 if (gobj == NULL) in amdgpu_cs_p1_user_fence() 137 p->uf_bo = amdgpu_bo_ref(gem_to_amdgpu_bo(gobj)); in amdgpu_cs_p1_user_fence() 138 drm_gem_object_put(gobj); in amdgpu_cs_p1_user_fence() 1289 struct drm_gem_object *gobj; in amdgpu_cs_submit() local 1340 drm_exec_for_each_locked_object(&p->exec, index, gobj) { in amdgpu_cs_submit() 1342 ttm_bo_move_to_lru_tail_unlocked(&gem_to_amdgpu_bo(gobj)->tbo); in amdgpu_cs_submit() 1349 dma_resv_add_fence(gobj->resv, in amdgpu_cs_submit() 1355 dma_resv_add_fence(gobj->resv, p->fence, DMA_RESV_USAGE_WRITE); in amdgpu_cs_submit()
|
| H A D | amdgpu_userq.c | 572 struct drm_gem_object *gobj; in amdgpu_userq_get_doorbell_index() local 576 gobj = drm_gem_object_lookup(filp, db_info->doorbell_handle); in amdgpu_userq_get_doorbell_index() 577 if (gobj == NULL) { in amdgpu_userq_get_doorbell_index() 582 db_obj->obj = amdgpu_bo_ref(gem_to_amdgpu_bo(gobj)); in amdgpu_userq_get_doorbell_index() 583 drm_gem_object_put(gobj); in amdgpu_userq_get_doorbell_index()
|
| H A D | amdgpu_ttm.c | 697 struct drm_gem_object *gobj; member 930 attach = gtt->gobj->import_attach; in amdgpu_ttm_backend_bind() 1048 } else if (ttm->sg && drm_gem_is_imported(gtt->gobj)) { in amdgpu_ttm_backend_unbind() 1051 attach = gtt->gobj->import_attach; in amdgpu_ttm_backend_unbind() 1179 gtt->gobj = &bo->base; in amdgpu_ttm_tt_create()
|
| /linux/drivers/gpu/drm/qxl/ |
| H A D | qxl_dumb.c | 37 struct drm_gem_object *gobj; in qxl_mode_dumb_create() local 66 args->size, &surf, &gobj, in qxl_mode_dumb_create() 70 qobj = gem_to_qxl_bo(gobj); in qxl_mode_dumb_create() 72 drm_gem_object_put(gobj); in qxl_mode_dumb_create()
|
| H A D | qxl_ioctl.c | 116 struct drm_gem_object *gobj; in qxlhw_handle_to_bo() local 120 gobj = drm_gem_object_lookup(file_priv, handle); in qxlhw_handle_to_bo() 121 if (!gobj) in qxlhw_handle_to_bo() 124 qobj = gem_to_qxl_bo(gobj); in qxlhw_handle_to_bo() 127 drm_gem_object_put(gobj); in qxlhw_handle_to_bo() 304 struct drm_gem_object *gobj = NULL; in qxl_update_area_ioctl() local 312 gobj = drm_gem_object_lookup(file, update_area->handle); in qxl_update_area_ioctl() 313 if (gobj == NULL) in qxl_update_area_ioctl() 316 qobj = gem_to_qxl_bo(gobj); in qxl_update_area_ioctl() 340 drm_gem_object_put(gobj); in qxl_update_area_ioctl()
|
| H A D | qxl_drv.h | 94 #define gem_to_qxl_bo(gobj) container_of((gobj), struct qxl_bo, tbo.base) argument 310 struct drm_gem_object **gobj, 312 void qxl_gem_object_free(struct drm_gem_object *gobj);
|
| H A D | qxl_display.c | 1220 struct drm_gem_object *gobj; in qxl_create_monitors_object() local 1227 false, false, NULL, &gobj); in qxl_create_monitors_object() 1232 qdev->monitors_config_bo = gem_to_qxl_bo(gobj); in qxl_create_monitors_object()
|
| /linux/drivers/media/mc/ |
| H A D | mc-entity.c | 93 static void dev_dbg_obj(const char *event_name, struct media_gobj *gobj) in dev_dbg_obj() argument 96 switch (media_type(gobj)) { in dev_dbg_obj() 98 dev_dbg(gobj->mdev->dev, in dev_dbg_obj() 100 event_name, media_id(gobj), in dev_dbg_obj() 101 gobj_to_entity(gobj)->name); in dev_dbg_obj() 105 struct media_link *link = gobj_to_link(gobj); in dev_dbg_obj() 107 dev_dbg(gobj->mdev->dev, in dev_dbg_obj() 109 event_name, media_id(gobj), link_type_name(link), in dev_dbg_obj() 116 struct media_pad *pad = gobj_to_pad(gobj); in dev_dbg_obj() 118 dev_dbg(gobj->mdev->dev, in dev_dbg_obj() [all …]
|
| /linux/drivers/gpu/drm/virtio/ |
| H A D | virtgpu_ioctl.c | 205 struct drm_gem_object *gobj = NULL; in virtio_gpu_resource_info_ioctl() local 208 gobj = drm_gem_object_lookup(file, ri->bo_handle); in virtio_gpu_resource_info_ioctl() 209 if (gobj == NULL) in virtio_gpu_resource_info_ioctl() 212 qobj = gem_to_virtio_gpu_obj(gobj); in virtio_gpu_resource_info_ioctl() 219 drm_gem_object_put(gobj); in virtio_gpu_resource_info_ioctl()
|
| H A D | virtgpu_gem.c | 65 struct drm_gem_object *gobj; in virtio_gpu_mode_dumb_create() local 90 ret = virtio_gpu_gem_create(file_priv, dev, ¶ms, &gobj, in virtio_gpu_mode_dumb_create()
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_gem.h | 40 struct dma_buf *nouveau_gem_prime_export(struct drm_gem_object *gobj,
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_bo.c | 645 struct drm_gem_object *gobj; in vmw_user_bo_lookup() local 647 gobj = drm_gem_object_lookup(filp, handle); in vmw_user_bo_lookup() 648 if (!gobj) { in vmw_user_bo_lookup() 654 *out = to_vmw_bo(gobj); in vmw_user_bo_lookup()
|
| /linux/drivers/gpu/drm/armada/ |
| H A D | armada_gem.c | 23 struct drm_gem_object *gobj = vmf->vma->vm_private_data; in armada_gem_vm_fault() local 24 struct armada_gem_object *obj = drm_to_armada_gem(gobj); in armada_gem_vm_fault()
|