xref: /linux/drivers/gpu/drm/virtio/virtgpu_prime.c (revision 4b132aacb0768ac1e652cf517097ea6f237214b9)
1 /*
2  * Copyright 2014 Canonical
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8  * and/or sell copies of the Software, and to permit persons to whom the
9  * Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice shall be included in
12  * all copies or substantial portions of the Software.
13  *
14  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17  * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20  * OTHER DEALINGS IN THE SOFTWARE.
21  *
22  * Authors: Andreas Pokorny
23  */
24 
25 #include <drm/drm_prime.h>
26 #include <linux/virtio_dma_buf.h>
27 
28 #include "virtgpu_drv.h"
29 
30 static int virtgpu_virtio_get_uuid(struct dma_buf *buf,
31 				   uuid_t *uuid)
32 {
33 	struct drm_gem_object *obj = buf->priv;
34 	struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(obj);
35 	struct virtio_gpu_device *vgdev = obj->dev->dev_private;
36 
37 	wait_event(vgdev->resp_wq, bo->uuid_state != STATE_INITIALIZING);
38 	if (bo->uuid_state != STATE_OK)
39 		return -ENODEV;
40 
41 	uuid_copy(uuid, &bo->uuid);
42 
43 	return 0;
44 }
45 
46 static struct sg_table *
47 virtgpu_gem_map_dma_buf(struct dma_buf_attachment *attach,
48 			enum dma_data_direction dir)
49 {
50 	struct drm_gem_object *obj = attach->dmabuf->priv;
51 	struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(obj);
52 
53 	if (virtio_gpu_is_vram(bo))
54 		return virtio_gpu_vram_map_dma_buf(bo, attach->dev, dir);
55 
56 	return drm_gem_map_dma_buf(attach, dir);
57 }
58 
59 static void virtgpu_gem_unmap_dma_buf(struct dma_buf_attachment *attach,
60 				      struct sg_table *sgt,
61 				      enum dma_data_direction dir)
62 {
63 	struct drm_gem_object *obj = attach->dmabuf->priv;
64 	struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(obj);
65 
66 	if (virtio_gpu_is_vram(bo)) {
67 		virtio_gpu_vram_unmap_dma_buf(attach->dev, sgt, dir);
68 		return;
69 	}
70 
71 	drm_gem_unmap_dma_buf(attach, sgt, dir);
72 }
73 
74 static const struct virtio_dma_buf_ops virtgpu_dmabuf_ops =  {
75 	.ops = {
76 		.cache_sgt_mapping = true,
77 		.attach = virtio_dma_buf_attach,
78 		.detach = drm_gem_map_detach,
79 		.map_dma_buf = virtgpu_gem_map_dma_buf,
80 		.unmap_dma_buf = virtgpu_gem_unmap_dma_buf,
81 		.release = drm_gem_dmabuf_release,
82 		.mmap = drm_gem_dmabuf_mmap,
83 		.vmap = drm_gem_dmabuf_vmap,
84 		.vunmap = drm_gem_dmabuf_vunmap,
85 	},
86 	.device_attach = drm_gem_map_attach,
87 	.get_uuid = virtgpu_virtio_get_uuid,
88 };
89 
90 int virtio_gpu_resource_assign_uuid(struct virtio_gpu_device *vgdev,
91 				    struct virtio_gpu_object *bo)
92 {
93 	struct virtio_gpu_object_array *objs;
94 
95 	objs = virtio_gpu_array_alloc(1);
96 	if (!objs)
97 		return -ENOMEM;
98 
99 	virtio_gpu_array_add_obj(objs, &bo->base.base);
100 
101 	return virtio_gpu_cmd_resource_assign_uuid(vgdev, objs);
102 }
103 
104 struct dma_buf *virtgpu_gem_prime_export(struct drm_gem_object *obj,
105 					 int flags)
106 {
107 	struct dma_buf *buf;
108 	struct drm_device *dev = obj->dev;
109 	struct virtio_gpu_device *vgdev = dev->dev_private;
110 	struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(obj);
111 	int ret = 0;
112 	bool blob = bo->host3d_blob || bo->guest_blob;
113 	DEFINE_DMA_BUF_EXPORT_INFO(exp_info);
114 
115 	if (!blob) {
116 		if (vgdev->has_resource_assign_uuid) {
117 			ret = virtio_gpu_resource_assign_uuid(vgdev, bo);
118 			if (ret)
119 				return ERR_PTR(ret);
120 
121 			virtio_gpu_notify(vgdev);
122 		} else {
123 			bo->uuid_state = STATE_ERR;
124 		}
125 	} else if (!(bo->blob_flags & VIRTGPU_BLOB_FLAG_USE_CROSS_DEVICE)) {
126 		bo->uuid_state = STATE_ERR;
127 	}
128 
129 	exp_info.ops = &virtgpu_dmabuf_ops.ops;
130 	exp_info.size = obj->size;
131 	exp_info.flags = flags;
132 	exp_info.priv = obj;
133 	exp_info.resv = obj->resv;
134 
135 	buf = virtio_dma_buf_export(&exp_info);
136 	if (IS_ERR(buf))
137 		return buf;
138 
139 	drm_dev_get(dev);
140 	drm_gem_object_get(obj);
141 
142 	return buf;
143 }
144 
145 struct drm_gem_object *virtgpu_gem_prime_import(struct drm_device *dev,
146 						struct dma_buf *buf)
147 {
148 	struct drm_gem_object *obj;
149 
150 	if (buf->ops == &virtgpu_dmabuf_ops.ops) {
151 		obj = buf->priv;
152 		if (obj->dev == dev) {
153 			/*
154 			 * Importing dmabuf exported from our own gem increases
155 			 * refcount on gem itself instead of f_count of dmabuf.
156 			 */
157 			drm_gem_object_get(obj);
158 			return obj;
159 		}
160 	}
161 
162 	return drm_gem_prime_import(dev, buf);
163 }
164 
165 struct drm_gem_object *virtgpu_gem_prime_import_sg_table(
166 	struct drm_device *dev, struct dma_buf_attachment *attach,
167 	struct sg_table *table)
168 {
169 	return ERR_PTR(-ENODEV);
170 }
171