| /linux/drivers/accel/rocket/ |
| H A D | rocket_gem.c | 66 struct drm_gem_object *gem_obj; in rocket_ioctl_create_bo() local 74 gem_obj = &shmem_obj->base; in rocket_ioctl_create_bo() 75 rkt_obj = to_rocket_bo(gem_obj); in rocket_ioctl_create_bo() 82 ret = drm_gem_handle_create(file, gem_obj, &args->handle); in rocket_ioctl_create_bo() 83 drm_gem_object_put(gem_obj); in rocket_ioctl_create_bo() 112 args->offset = drm_vma_node_offset_addr(&gem_obj->vma_node); in rocket_ioctl_create_bo() 123 drm_gem_shmem_object_free(gem_obj); in rocket_ioctl_create_bo() 132 struct drm_gem_object *gem_obj; in rocket_ioctl_prep_bo() local 141 gem_obj = drm_gem_object_lookup(file, args->handle); in rocket_ioctl_prep_bo() 142 if (!gem_obj) in rocket_ioctl_prep_bo() [all …]
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_gem_dma_helper.c | 76 struct drm_gem_object *gem_obj; in __drm_gem_dma_create() local 80 gem_obj = drm->driver->gem_create_object(drm, size); in __drm_gem_dma_create() 81 if (IS_ERR(gem_obj)) in __drm_gem_dma_create() 82 return ERR_CAST(gem_obj); in __drm_gem_dma_create() 83 dma_obj = to_drm_gem_dma_obj(gem_obj); in __drm_gem_dma_create() 88 gem_obj = &dma_obj->base; in __drm_gem_dma_create() 91 if (!gem_obj->funcs) in __drm_gem_dma_create() 92 gem_obj->funcs = &drm_gem_dma_default_funcs; in __drm_gem_dma_create() 95 drm_gem_private_object_init(drm, gem_obj, size); in __drm_gem_dma_create() 100 ret = drm_gem_object_init(drm, gem_obj, size); in __drm_gem_dma_create() [all …]
|
| /linux/drivers/gpu/drm/xen/ |
| H A D | xen_drm_front_gem.c | 41 to_xen_gem_obj(struct drm_gem_object *gem_obj) in to_xen_gem_obj() argument 43 return container_of(gem_obj, struct xen_gem_object, base); in to_xen_gem_obj() 60 static int xen_drm_front_gem_object_mmap(struct drm_gem_object *gem_obj, in xen_drm_front_gem_object_mmap() argument 63 struct xen_gem_object *xen_obj = to_xen_gem_obj(gem_obj); in xen_drm_front_gem_object_mmap() 66 vma->vm_ops = gem_obj->funcs->vm_ops; in xen_drm_front_gem_object_mmap() 202 void xen_drm_front_gem_free_object_unlocked(struct drm_gem_object *gem_obj) in xen_drm_front_gem_free_object_unlocked() argument 204 struct xen_gem_object *xen_obj = to_xen_gem_obj(gem_obj); in xen_drm_front_gem_free_object_unlocked() 221 drm_gem_object_release(gem_obj); in xen_drm_front_gem_free_object_unlocked() 225 struct page **xen_drm_front_gem_get_pages(struct drm_gem_object *gem_obj) in xen_drm_front_gem_get_pages() argument 227 struct xen_gem_object *xen_obj = to_xen_gem_obj(gem_obj); in xen_drm_front_gem_get_pages() [all …]
|
| H A D | xen_drm_front_gem.h | 28 struct sg_table *xen_drm_front_gem_get_sg_table(struct drm_gem_object *gem_obj); 32 void xen_drm_front_gem_free_object_unlocked(struct drm_gem_object *gem_obj); 34 int xen_drm_front_gem_prime_vmap(struct drm_gem_object *gem_obj, 37 void xen_drm_front_gem_prime_vunmap(struct drm_gem_object *gem_obj,
|
| H A D | xen_drm_front.h | 128 static inline u64 xen_drm_front_dbuf_to_cookie(struct drm_gem_object *gem_obj) in xen_drm_front_dbuf_to_cookie() argument 130 return (uintptr_t)gem_obj; in xen_drm_front_dbuf_to_cookie()
|
| /linux/drivers/gpu/drm/imagination/ |
| H A D | pvr_gem.c | 41 static int pvr_gem_mmap(struct drm_gem_object *gem_obj, struct vm_area_struct *vma) in pvr_gem_mmap() argument 43 struct pvr_gem_object *pvr_obj = gem_to_pvr_gem(gem_obj); in pvr_gem_mmap() 135 struct drm_gem_object *gem_obj = gem_from_pvr_gem(pvr_obj); in pvr_gem_object_into_handle() local 141 err = drm_gem_handle_create(file, gem_obj, &new_handle); in pvr_gem_object_into_handle() 180 struct drm_gem_object *gem_obj; in pvr_gem_object_from_handle() local 182 gem_obj = drm_gem_object_lookup(file, handle); in pvr_gem_object_from_handle() 183 if (!gem_obj) in pvr_gem_object_from_handle() 186 return gem_to_pvr_gem(gem_obj); in pvr_gem_object_from_handle() 314 struct drm_gem_object *gem_obj; in pvr_gem_create_object() local 321 gem_obj = gem_from_pvr_gem(pvr_obj); in pvr_gem_create_object() [all …]
|
| H A D | pvr_drv.c | 171 struct drm_gem_object *gem_obj; in pvr_ioctl_get_bo_mmap_offset() local 196 gem_obj = gem_from_pvr_gem(pvr_obj); in pvr_ioctl_get_bo_mmap_offset() 203 ret = drm_gem_create_mmap_offset(gem_obj); in pvr_ioctl_get_bo_mmap_offset() 206 drm_gem_object_put(gem_obj); in pvr_ioctl_get_bo_mmap_offset() 214 args->offset = drm_vma_node_offset_addr(&gem_obj->vma_node); in pvr_ioctl_get_bo_mmap_offset()
|
| H A D | pvr_fw.c | 1177 struct drm_gem_object *gem_obj = gem_from_pvr_gem(pvr_obj); in pvr_fw_object_fw_map() local 1195 gem_obj->size, 0, 0, in pvr_fw_object_fw_map() 1203 fw_obj->fw_mm_node.size = gem_obj->size; in pvr_fw_object_fw_map() 1242 struct drm_gem_object *gem_obj = gem_from_pvr_gem(pvr_obj); in pvr_fw_object_fw_unmap() local 1243 struct pvr_device *pvr_dev = to_pvr_device(gem_obj->dev); in pvr_fw_object_fw_unmap() 1429 struct drm_gem_object *gem_obj = gem_from_pvr_gem(pvr_obj); in pvr_fw_object_destroy() local 1430 struct pvr_device *pvr_dev = to_pvr_device(gem_obj->dev); in pvr_fw_object_destroy()
|
| /linux/drivers/gpu/drm/v3d/ |
| H A D | v3d_bo.c | 236 struct drm_gem_object *gem_obj; in v3d_mmap_bo_ioctl() local 243 gem_obj = drm_gem_object_lookup(file_priv, args->handle); in v3d_mmap_bo_ioctl() 244 if (!gem_obj) { in v3d_mmap_bo_ioctl() 249 args->offset = drm_vma_node_offset_addr(&gem_obj->vma_node); in v3d_mmap_bo_ioctl() 250 drm_gem_object_put(gem_obj); in v3d_mmap_bo_ioctl() 259 struct drm_gem_object *gem_obj; in v3d_get_bo_offset_ioctl() local 262 gem_obj = drm_gem_object_lookup(file_priv, args->handle); in v3d_get_bo_offset_ioctl() 263 if (!gem_obj) { in v3d_get_bo_offset_ioctl() 267 bo = to_v3d_bo(gem_obj); in v3d_get_bo_offset_ioctl() 271 drm_gem_object_put(gem_obj); in v3d_get_bo_offset_ioctl()
|
| /linux/drivers/gpu/drm/panfrost/ |
| H A D | panfrost_drv.c | 379 struct drm_gem_object *gem_obj; in panfrost_ioctl_wait_bo() local 385 gem_obj = drm_gem_object_lookup(file_priv, args->handle); in panfrost_ioctl_wait_bo() 386 if (!gem_obj) in panfrost_ioctl_wait_bo() 389 ret = dma_resv_wait_timeout(gem_obj->resv, DMA_RESV_USAGE_READ, in panfrost_ioctl_wait_bo() 394 drm_gem_object_put(gem_obj); in panfrost_ioctl_wait_bo() 403 struct drm_gem_object *gem_obj; in panfrost_ioctl_mmap_bo() local 411 gem_obj = drm_gem_object_lookup(file_priv, args->handle); in panfrost_ioctl_mmap_bo() 412 if (!gem_obj) { in panfrost_ioctl_mmap_bo() 418 if (to_panfrost_bo(gem_obj)->is_heap) { in panfrost_ioctl_mmap_bo() 423 ret = drm_gem_create_mmap_offset(gem_obj); in panfrost_ioctl_mmap_bo() [all …]
|
| /linux/drivers/gpu/drm/vc4/ |
| H A D | vc4_bo.c | 133 static void vc4_bo_set_label(struct drm_gem_object *gem_obj, int label) in vc4_bo_set_label() argument 135 struct vc4_bo *bo = to_vc4_bo(gem_obj); in vc4_bo_set_label() 136 struct vc4_dev *vc4 = to_vc4_dev(gem_obj->dev); in vc4_bo_set_label() 142 vc4->bo_labels[label].size_allocated += gem_obj->size; in vc4_bo_set_label() 146 vc4->bo_labels[bo->label].size_allocated -= gem_obj->size; in vc4_bo_set_label() 814 struct drm_gem_object *gem_obj; in vc4_mmap_bo_ioctl() local 819 gem_obj = drm_gem_object_lookup(file_priv, args->handle); in vc4_mmap_bo_ioctl() 820 if (!gem_obj) { in vc4_mmap_bo_ioctl() 826 args->offset = drm_vma_node_offset_addr(&gem_obj->vma_node); in vc4_mmap_bo_ioctl() 828 drm_gem_object_put(gem_obj); in vc4_mmap_bo_ioctl() [all …]
|
| H A D | vc4_gem.c | 1217 struct drm_gem_object *gem_obj; in vc4_gem_madvise_ioctl() local 1235 gem_obj = drm_gem_object_lookup(file_priv, args->handle); in vc4_gem_madvise_ioctl() 1236 if (!gem_obj) { in vc4_gem_madvise_ioctl() 1241 bo = to_vc4_bo(gem_obj); in vc4_gem_madvise_ioctl() 1253 if (gem_obj->import_attach) { in vc4_gem_madvise_ioctl() 1289 drm_gem_object_put(gem_obj); in vc4_gem_madvise_ioctl()
|
| H A D | vc4_kms.c | 547 struct drm_gem_object *gem_obj; in vc4_fb_create() local 550 gem_obj = drm_gem_object_lookup(file_priv, in vc4_fb_create() 552 if (!gem_obj) { in vc4_fb_create() 557 bo = to_vc4_bo(gem_obj); in vc4_fb_create() 568 drm_gem_object_put(gem_obj); in vc4_fb_create()
|
| /linux/drivers/gpu/drm/tests/ |
| H A D | drm_gem_shmem_test.c | 72 struct drm_gem_object *gem_obj; in drm_gem_shmem_test_obj_create_private() local 112 gem_obj = drm_gem_shmem_prime_import_sg_table(drm_dev, &attach_mock, sgt); in drm_gem_shmem_test_obj_create_private() 113 KUNIT_ASSERT_NOT_ERR_OR_NULL(test, gem_obj); in drm_gem_shmem_test_obj_create_private() 114 KUNIT_EXPECT_EQ(test, gem_obj->size, TEST_SIZE); in drm_gem_shmem_test_obj_create_private() 115 KUNIT_EXPECT_NULL(test, gem_obj->filp); in drm_gem_shmem_test_obj_create_private() 116 KUNIT_EXPECT_NOT_NULL(test, gem_obj->funcs); in drm_gem_shmem_test_obj_create_private() 122 shmem = to_drm_gem_shmem_obj(gem_obj); in drm_gem_shmem_test_obj_create_private()
|
| /linux/drivers/gpu/drm/i915/gem/ |
| H A D | i915_gem_dmabuf.c | 216 struct dma_buf *i915_gem_prime_export(struct drm_gem_object *gem_obj, int flags) in i915_gem_prime_export() argument 218 struct drm_i915_gem_object *obj = to_intel_bo(gem_obj); in i915_gem_prime_export() 222 exp_info.size = gem_obj->size; in i915_gem_prime_export() 224 exp_info.priv = gem_obj; in i915_gem_prime_export() 233 return drm_gem_dmabuf_export(gem_obj->dev, &exp_info); in i915_gem_prime_export()
|
| H A D | i915_gem_dmabuf.h | 16 struct dma_buf *i915_gem_prime_export(struct drm_gem_object *gem_obj, int flags);
|
| /linux/include/drm/ |
| H A D | drm_gem_ttm_helper.h | 14 #define drm_gem_ttm_of_gem(gem_obj) \ argument 15 container_of(gem_obj, struct ttm_buffer_object, base)
|
| H A D | drm_gem_dma_helper.h | 32 #define to_drm_gem_dma_obj(gem_obj) \ argument 33 container_of(gem_obj, struct drm_gem_dma_object, base)
|
| /linux/drivers/gpu/drm/renesas/rcar-du/ |
| H A D | rcar_du_kms.c | 377 struct drm_gem_object *gem_obj; in rcar_du_gem_prime_import_sg_table() local 388 gem_obj = &dma_obj->base; in rcar_du_gem_prime_import_sg_table() 389 gem_obj->funcs = &rcar_du_gem_funcs; in rcar_du_gem_prime_import_sg_table() 391 drm_gem_private_object_init(dev, gem_obj, attach->dmabuf->size); in rcar_du_gem_prime_import_sg_table() 394 ret = drm_gem_create_mmap_offset(gem_obj); in rcar_du_gem_prime_import_sg_table() 396 drm_gem_object_release(gem_obj); in rcar_du_gem_prime_import_sg_table() 404 return gem_obj; in rcar_du_gem_prime_import_sg_table()
|
| /linux/drivers/gpu/drm/i915/gt/ |
| H A D | intel_gsc.c | 63 intf->gem_obj = obj; in gsc_ext_om_alloc() 74 struct drm_i915_gem_object *obj = fetch_and_zero(&intf->gem_obj); in gsc_ext_om_destroy() 220 adev->ext_op_mem.start = i915_gem_object_get_dma_address(intf->gem_obj, 0); in gsc_init_one()
|
| H A D | intel_gsc.h | 33 struct drm_i915_gem_object *gem_obj; member
|
| /linux/drivers/gpu/drm/i915/gvt/ |
| H A D | dmabuf.c | 172 static void vgpu_gem_release(struct drm_i915_gem_object *gem_obj) in vgpu_gem_release() argument 175 struct intel_vgpu_fb_info *fb_info = gem_obj->gvt_info; in vgpu_gem_release() 181 gem_obj->base.dma_buf = NULL; in vgpu_gem_release() 186 gem_obj->base.dma_buf = NULL; in vgpu_gem_release()
|
| /linux/drivers/gpu/drm/ttm/tests/ |
| H A D | ttm_kunit_helpers.c | 174 struct drm_gem_object gem_obj = { }; in ttm_bo_kunit_init() local 181 bo->base = gem_obj; in ttm_bo_kunit_init()
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_bo.c | 3296 struct drm_gem_object *gem_obj; in xe_gem_mmap_offset_ioctl() local 3322 gem_obj = drm_gem_object_lookup(file, args->handle); in xe_gem_mmap_offset_ioctl() 3323 if (XE_IOCTL_DBG(xe, !gem_obj)) in xe_gem_mmap_offset_ioctl() 3327 args->offset = drm_vma_node_offset_addr(&gem_obj->vma_node); in xe_gem_mmap_offset_ioctl() 3329 xe_bo_put(gem_to_xe_bo(gem_obj)); in xe_gem_mmap_offset_ioctl()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_amdkfd_gpuvm.c | 334 struct drm_gem_object *gem_obj; in create_dmamap_sg_bo() local 348 ttm_bo_type_sg, mem->bo->tbo.base.resv, &gem_obj, 0); in create_dmamap_sg_bo() 357 *bo_out = gem_to_amdgpu_bo(gem_obj); in create_dmamap_sg_bo()
|