Home
last modified time | relevance | path

Searched refs:vgdev (Results 1 – 11 of 11) sorted by relevance

/linux/drivers/gpu/drm/virtio/
H A Dvirtgpu_kms.c38 struct virtio_gpu_device *vgdev = in virtio_gpu_config_changed_work_func() local
44 virtio_cread_le(vgdev->vdev, struct virtio_gpu_config, in virtio_gpu_config_changed_work_func()
47 if (vgdev->num_scanouts) { in virtio_gpu_config_changed_work_func()
48 if (vgdev->has_edid) in virtio_gpu_config_changed_work_func()
49 virtio_gpu_cmd_get_edids(vgdev); in virtio_gpu_config_changed_work_func()
50 virtio_gpu_cmd_get_display_info(vgdev); in virtio_gpu_config_changed_work_func()
51 virtio_gpu_notify(vgdev); in virtio_gpu_config_changed_work_func()
52 drm_helper_hpd_irq_event(vgdev->ddev); in virtio_gpu_config_changed_work_func()
56 virtio_cwrite_le(vgdev->vdev, struct virtio_gpu_config, in virtio_gpu_config_changed_work_func()
68 static void virtio_gpu_get_capsets(struct virtio_gpu_device *vgdev, in virtio_gpu_get_capsets() argument
[all …]
H A Dvirtgpu_vq.c60 struct virtio_gpu_device *vgdev = dev->dev_private; in virtio_gpu_ctrl_ack() local
62 schedule_work(&vgdev->ctrlq.dequeue_work); in virtio_gpu_ctrl_ack()
68 struct virtio_gpu_device *vgdev = dev->dev_private; in virtio_gpu_cursor_ack() local
70 schedule_work(&vgdev->cursorq.dequeue_work); in virtio_gpu_cursor_ack()
73 int virtio_gpu_alloc_vbufs(struct virtio_gpu_device *vgdev) in virtio_gpu_alloc_vbufs() argument
75 vgdev->vbufs = kmem_cache_create("virtio-gpu-vbufs", in virtio_gpu_alloc_vbufs()
79 if (!vgdev->vbufs) in virtio_gpu_alloc_vbufs()
84 void virtio_gpu_free_vbufs(struct virtio_gpu_device *vgdev) in virtio_gpu_free_vbufs() argument
86 kmem_cache_destroy(vgdev->vbufs); in virtio_gpu_free_vbufs()
87 vgdev->vbufs = NULL; in virtio_gpu_free_vbufs()
[all …]
H A Dvirtgpu_ioctl.c42 static void virtio_gpu_create_context_locked(struct virtio_gpu_device *vgdev, in virtio_gpu_create_context_locked() argument
46 virtio_gpu_cmd_context_create(vgdev, vfpriv->ctx_id, in virtio_gpu_create_context_locked()
54 virtio_gpu_cmd_context_create(vgdev, vfpriv->ctx_id, in virtio_gpu_create_context_locked()
64 struct virtio_gpu_device *vgdev = dev->dev_private; in virtio_gpu_create_context() local
71 virtio_gpu_create_context_locked(vgdev, vfpriv); in virtio_gpu_create_context()
80 struct virtio_gpu_device *vgdev = dev->dev_private; in virtio_gpu_map_ioctl() local
83 return drm_gem_dumb_map_offset(file, vgdev->ddev, in virtio_gpu_map_ioctl()
91 struct virtio_gpu_device *vgdev = dev->dev_private; in virtio_gpu_getparam_ioctl() local
97 value = vgdev->has_virgl_3d ? 1 : 0; in virtio_gpu_getparam_ioctl()
103 value = vgdev->has_resource_blob ? 1 : 0; in virtio_gpu_getparam_ioctl()
[all …]
H A Dvirtgpu_object.c36 int virtio_gpu_resource_id_get(struct virtio_gpu_device *vgdev, uint32_t *resid) in virtio_gpu_resource_id_get() argument
51 int handle = ida_alloc(&vgdev->resource_ida, GFP_KERNEL); in virtio_gpu_resource_id_get()
60 static void virtio_gpu_resource_id_put(struct virtio_gpu_device *vgdev, uint32_t id) in virtio_gpu_resource_id_put() argument
63 ida_free(&vgdev->resource_ida, id - 1); in virtio_gpu_resource_id_put()
68 struct virtio_gpu_device *vgdev = bo->base.base.dev->dev_private; in virtio_gpu_cleanup_object() local
70 virtio_gpu_resource_id_put(vgdev, bo->hw_res_handle); in virtio_gpu_cleanup_object()
76 spin_lock(&vgdev->host_visible_lock); in virtio_gpu_cleanup_object()
80 spin_unlock(&vgdev->host_visible_lock); in virtio_gpu_cleanup_object()
94 struct virtio_gpu_device *vgdev = bo->base.base.dev->dev_private; in virtio_gpu_free_object() local
97 virtio_gpu_cmd_unref_resource(vgdev, bo); in virtio_gpu_free_object()
[all …]
H A Dvirtgpu_gem.c37 struct virtio_gpu_device *vgdev = dev->dev_private; in virtio_gpu_gem_create() local
42 ret = virtio_gpu_object_create(vgdev, params, &obj, NULL); in virtio_gpu_gem_create()
67 struct virtio_gpu_device *vgdev = dev->dev_private; in virtio_gpu_mode_dumb_create() local
84 if (vgdev->has_resource_blob && !vgdev->has_virgl_3d) { in virtio_gpu_mode_dumb_create()
105 struct virtio_gpu_device *vgdev = obj->dev->dev_private; in virtio_gpu_gem_object_open() local
109 if (!vgdev->has_virgl_3d) in virtio_gpu_gem_object_open()
115 if (!vgdev->has_context_init) in virtio_gpu_gem_object_open()
124 virtio_gpu_cmd_context_attach_resource(vgdev, vfpriv->ctx_id, objs); in virtio_gpu_gem_object_open()
128 virtio_gpu_notify(vgdev); in virtio_gpu_gem_object_open()
135 struct virtio_gpu_device *vgdev = obj->dev->dev_private; in virtio_gpu_gem_object_close() local
[all …]
H A Dvirtgpu_display.c94 struct virtio_gpu_device *vgdev = dev->dev_private; in virtio_gpu_crtc_mode_set_nofb() local
97 virtio_gpu_cmd_set_scanout(vgdev, output->index, 0, in virtio_gpu_crtc_mode_set_nofb()
100 virtio_gpu_notify(vgdev); in virtio_gpu_crtc_mode_set_nofb()
113 struct virtio_gpu_device *vgdev = dev->dev_private; in virtio_gpu_crtc_atomic_disable() local
118 virtio_gpu_cmd_set_scanout(vgdev, output->index, 0, 0, 0, 0, 0); in virtio_gpu_crtc_atomic_disable()
119 virtio_gpu_notify(vgdev); in virtio_gpu_crtc_atomic_disable()
274 static int vgdev_output_init(struct virtio_gpu_device *vgdev, int index) in vgdev_output_init() argument
276 struct drm_device *dev = vgdev->ddev; in vgdev_output_init()
277 struct virtio_gpu_output *output = vgdev->outputs + index; in vgdev_output_init()
291 primary = virtio_gpu_plane_init(vgdev, DRM_PLANE_TYPE_PRIMARY, index); in vgdev_output_init()
[all …]
H A Dvirtgpu_vram.c9 struct virtio_gpu_device *vgdev = obj->dev->dev_private; in virtio_gpu_vram_free() local
14 spin_lock(&vgdev->host_visible_lock); in virtio_gpu_vram_free()
16 spin_unlock(&vgdev->host_visible_lock); in virtio_gpu_vram_free()
19 virtio_gpu_cmd_unmap(vgdev, bo); in virtio_gpu_vram_free()
21 virtio_gpu_cmd_unref_resource(vgdev, bo); in virtio_gpu_vram_free()
22 virtio_gpu_notify(vgdev); in virtio_gpu_vram_free()
36 struct virtio_gpu_device *vgdev = obj->dev->dev_private; in virtio_gpu_vram_mmap() local
45 wait_event(vgdev->resp_wq, vram->map_state != STATE_INITIALIZING); in virtio_gpu_vram_mmap()
76 struct virtio_gpu_device *vgdev = bo->base.base.dev->dev_private; in virtio_gpu_vram_map_dma_buf() local
89 if (!is_virtio_device(dev) || !vgdev->has_resource_assign_uuid) { in virtio_gpu_vram_map_dma_buf()
[all …]
H A Dvirtgpu_plane.c134 static int virtio_gpu_panic_update_dumb_bo(struct virtio_gpu_device *vgdev, in virtio_gpu_panic_update_dumb_bo() argument
153 return virtio_gpu_panic_cmd_transfer_to_host_2d(vgdev, off, w, h, x, y, in virtio_gpu_panic_update_dumb_bo()
157 static void virtio_gpu_update_dumb_bo(struct virtio_gpu_device *vgdev, in virtio_gpu_update_dumb_bo() argument
176 virtio_gpu_cmd_transfer_to_host_2d(vgdev, off, w, h, x, y, in virtio_gpu_update_dumb_bo()
186 struct virtio_gpu_device *vgdev = dev->dev_private; in virtio_gpu_panic_resource_flush() local
193 virtio_gpu_panic_cmd_resource_flush(vgdev, bo->hw_res_handle, x, y, in virtio_gpu_panic_resource_flush()
195 virtio_gpu_panic_notify(vgdev); in virtio_gpu_panic_resource_flush()
203 struct virtio_gpu_device *vgdev = dev->dev_private; in virtio_gpu_resource_flush() local
219 virtio_gpu_cmd_resource_flush(vgdev, bo->hw_res_handle, x, y, in virtio_gpu_resource_flush()
222 virtio_gpu_notify(vgdev); in virtio_gpu_resource_flush()
[all …]
H A Dvirtgpu_prime.c37 struct virtio_gpu_device *vgdev = obj->dev->dev_private; in virtgpu_virtio_get_uuid() local
39 wait_event(vgdev->resp_wq, bo->uuid_state != STATE_INITIALIZING); in virtgpu_virtio_get_uuid()
91 int virtio_gpu_resource_assign_uuid(struct virtio_gpu_device *vgdev, in virtio_gpu_resource_assign_uuid() argument
102 return virtio_gpu_cmd_resource_assign_uuid(vgdev, objs); in virtio_gpu_resource_assign_uuid()
110 struct virtio_gpu_device *vgdev = dev->dev_private; in virtgpu_gem_prime_export() local
117 if (vgdev->has_resource_assign_uuid) { in virtgpu_gem_prime_export()
118 ret = virtio_gpu_resource_assign_uuid(vgdev, bo); in virtgpu_gem_prime_export()
122 virtio_gpu_notify(vgdev); in virtgpu_gem_prime_export()
204 struct virtio_gpu_device *vgdev = obj->dev->dev_private; in virtgpu_dma_buf_free_obj() local
219 virtio_gpu_cmd_unref_resource(vgdev, bo); in virtgpu_dma_buf_free_obj()
[all …]
H A Dvirtgpu_fence.c58 struct virtio_gpu_fence *virtio_gpu_fence_alloc(struct virtio_gpu_device *vgdev, in virtio_gpu_fence_alloc() argument
63 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_alloc()
84 void virtio_gpu_fence_emit(struct virtio_gpu_device *vgdev, in virtio_gpu_fence_emit() argument
88 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_emit()
110 void virtio_gpu_fence_event_process(struct virtio_gpu_device *vgdev, in virtio_gpu_fence_event_process() argument
113 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_event_process()
118 atomic64_set(&vgdev->fence_drv.last_fence_id, fence_id); in virtio_gpu_fence_event_process()
139 drm_send_event(vgdev->ddev, &curr->e->base); in virtio_gpu_fence_event_process()
149 drm_send_event(vgdev->ddev, &signaled->e->base); in virtio_gpu_fence_event_process()
H A Dvirtgpu_submit.c39 struct virtio_gpu_device *vgdev; member
353 virtio_gpu_cmd_submit(submit->vgdev, submit->buf, submit->exbuf->size, in virtio_gpu_submit()
356 virtio_gpu_notify(submit->vgdev); in virtio_gpu_submit()
374 struct virtio_gpu_device *vgdev = dev->dev_private; in virtio_gpu_init_submit() local
391 out_fence = virtio_gpu_fence_alloc(vgdev, fence_ctx, ring_idx); in virtio_gpu_init_submit()
408 submit->vgdev = vgdev; in virtio_gpu_init_submit()
476 struct virtio_gpu_device *vgdev = dev->dev_private; in virtio_gpu_execbuffer_ioctl() local
478 u64 fence_ctx = vgdev->fence_drv.context; in virtio_gpu_execbuffer_ioctl()
484 if (!vgdev->has_virgl_3d) in virtio_gpu_execbuffer_ioctl()