| /linux/drivers/gpu/drm/ |
| H A D | drm_gem_dma_helper.c | 75 struct drm_gem_dma_object *dma_obj; in __drm_gem_dma_create() local 83 dma_obj = to_drm_gem_dma_obj(gem_obj); in __drm_gem_dma_create() 85 dma_obj = kzalloc_obj(*dma_obj); in __drm_gem_dma_create() 86 if (!dma_obj) in __drm_gem_dma_create() 88 gem_obj = &dma_obj->base; in __drm_gem_dma_create() 98 dma_obj->map_noncoherent = false; in __drm_gem_dma_create() 111 return dma_obj; in __drm_gem_dma_create() 114 kfree(dma_obj); in __drm_gem_dma_create() 139 struct drm_gem_dma_object *dma_obj; in drm_gem_dma_create() local 144 dma_obj = __drm_gem_dma_create(drm, size, false); in drm_gem_dma_create() [all …]
|
| H A D | drm_fbdev_dma.c | 204 struct drm_gem_dma_object *dma_obj = to_drm_gem_dma_obj(buffer->gem); in drm_fbdev_dma_driver_fbdev_probe_tail() local 213 if (dma_obj->map_noncoherent) in drm_fbdev_dma_driver_fbdev_probe_tail()
|
| /linux/drivers/gpu/drm/mediatek/ |
| H A D | mtk_gem.c | 25 struct drm_gem_dma_object *dma_obj = to_drm_gem_dma_obj(obj); in mtk_gem_free_object() local 28 if (dma_obj->sgt) in mtk_gem_free_object() 29 drm_prime_gem_destroy(obj, dma_obj->sgt); in mtk_gem_free_object() 31 dma_free_wc(priv->dma_dev, dma_obj->base.size, in mtk_gem_free_object() 32 dma_obj->vaddr, dma_obj->dma_addr); in mtk_gem_free_object() 37 kfree(dma_obj); in mtk_gem_free_object() 48 struct drm_gem_dma_object *dma_obj = to_drm_gem_dma_obj(obj); in mtk_gem_prime_get_sg_table() local 57 ret = dma_get_sgtable(priv->dma_dev, sgt, dma_obj->vaddr, in mtk_gem_prime_get_sg_table() 58 dma_obj->dma_addr, obj->size); in mtk_gem_prime_get_sg_table() 80 struct drm_gem_dma_object *dma_obj; in mtk_gem_init() local [all …]
|
| H A D | mtk_plane.c | 117 struct drm_gem_dma_object *dma_obj; in mtk_plane_update_new_state() local 127 dma_obj = to_drm_gem_dma_obj(gem); in mtk_plane_update_new_state() 128 addr = dma_obj->dma_addr; in mtk_plane_update_new_state()
|
| /linux/include/drm/ |
| H A D | drm_gem_dma_helper.h | 37 void drm_gem_dma_free(struct drm_gem_dma_object *dma_obj); 38 void drm_gem_dma_print_info(const struct drm_gem_dma_object *dma_obj, 40 struct sg_table *drm_gem_dma_get_sg_table(struct drm_gem_dma_object *dma_obj); 41 int drm_gem_dma_vmap(struct drm_gem_dma_object *dma_obj, 43 int drm_gem_dma_mmap(struct drm_gem_dma_object *dma_obj, struct vm_area_struct *vma); 60 struct drm_gem_dma_object *dma_obj = to_drm_gem_dma_obj(obj); in drm_gem_dma_object_free() local 62 drm_gem_dma_free(dma_obj); in drm_gem_dma_object_free() 77 const struct drm_gem_dma_object *dma_obj = to_drm_gem_dma_obj(obj); in drm_gem_dma_object_print_info() local 79 drm_gem_dma_print_info(dma_obj, p, indent); in drm_gem_dma_object_print_info() 94 struct drm_gem_dma_object *dma_obj = to_drm_gem_dma_obj(obj); in drm_gem_dma_object_get_sg_table() local [all …]
|
| /linux/drivers/gpu/drm/imx/ipuv3/ |
| H A D | ipuv3-plane.c | 129 struct drm_gem_dma_object *dma_obj; in drm_plane_state_to_eba() local 133 dma_obj = drm_fb_dma_get_gem_obj(fb, plane); in drm_plane_state_to_eba() 134 BUG_ON(!dma_obj); in drm_plane_state_to_eba() 136 return dma_obj->dma_addr + fb->offsets[plane] + fb->pitches[plane] * y + in drm_plane_state_to_eba() 144 struct drm_gem_dma_object *dma_obj; in drm_plane_state_to_ubo() local 149 dma_obj = drm_fb_dma_get_gem_obj(fb, 1); in drm_plane_state_to_ubo() 150 BUG_ON(!dma_obj); in drm_plane_state_to_ubo() 155 return dma_obj->dma_addr + fb->offsets[1] + fb->pitches[1] * y + in drm_plane_state_to_ubo() 163 struct drm_gem_dma_object *dma_obj; in drm_plane_state_to_vbo() local 168 dma_obj = drm_fb_dma_get_gem_obj(fb, 2); in drm_plane_state_to_vbo() [all …]
|
| /linux/drivers/gpu/drm/imx/dcss/ |
| H A D | dcss-plane.c | 151 struct drm_gem_dma_object *dma_obj; in dcss_plane_atomic_check() local 160 dma_obj = drm_fb_dma_get_gem_obj(fb, 0); in dcss_plane_atomic_check() 161 WARN_ON(!dma_obj); in dcss_plane_atomic_check() 222 struct drm_gem_dma_object *dma_obj = drm_fb_dma_get_gem_obj(fb, 0); in dcss_plane_atomic_set_base() local 228 p1_ba = dma_obj->dma_addr + fb->offsets[0] + in dcss_plane_atomic_set_base() 235 p1_ba = dma_obj->dma_addr + fb->offsets[0] + in dcss_plane_atomic_set_base() 241 p2_ba = dma_obj->dma_addr + fb->offsets[1] + in dcss_plane_atomic_set_base()
|
| /linux/drivers/gpu/drm/renesas/rcar-du/ |
| H A D | rcar_du_kms.c | 376 struct drm_gem_dma_object *dma_obj; in rcar_du_gem_prime_import_sg_table() local 384 dma_obj = kzalloc_obj(*dma_obj); in rcar_du_gem_prime_import_sg_table() 385 if (!dma_obj) in rcar_du_gem_prime_import_sg_table() 388 gem_obj = &dma_obj->base; in rcar_du_gem_prime_import_sg_table() 392 dma_obj->map_noncoherent = false; in rcar_du_gem_prime_import_sg_table() 397 kfree(dma_obj); in rcar_du_gem_prime_import_sg_table() 401 dma_obj->dma_addr = 0; in rcar_du_gem_prime_import_sg_table() 402 dma_obj->sgt = sgt; in rcar_du_gem_prime_import_sg_table()
|
| /linux/drivers/gpu/drm/vc4/ |
| H A D | vc4_bo.c | 427 struct drm_gem_dma_object *dma_obj; in vc4_bo_create() local 444 dma_obj = drm_gem_dma_create(dev, size); in vc4_bo_create() 445 if (IS_ERR(dma_obj)) { in vc4_bo_create() 451 dma_obj = drm_gem_dma_create(dev, size); in vc4_bo_create() 454 if (IS_ERR(dma_obj)) { in vc4_bo_create() 467 dma_obj = drm_gem_dma_create(dev, size); in vc4_bo_create() 470 if (IS_ERR(dma_obj)) { in vc4_bo_create() 476 bo = to_vc4_bo(&dma_obj->base); in vc4_bo_create() 485 vc4_bo_set_label(&dma_obj->base, type); in vc4_bo_create()
|
| /linux/drivers/gpu/drm/arm/ |
| H A D | malidp_planes.c | 327 struct drm_gem_dma_object *dma_obj; in malidp_check_pages_threshold() local 332 dma_obj = to_drm_gem_dma_obj(obj); in malidp_check_pages_threshold() 334 if (dma_obj->sgt) in malidp_check_pages_threshold() 335 sgt = dma_obj->sgt; in malidp_check_pages_threshold() 346 if (!dma_obj->sgt) in malidp_check_pages_threshold() 353 if (!dma_obj->sgt) in malidp_check_pages_threshold()
|