xref: /linux/drivers/gpu/drm/qxl/qxl_ioctl.c (revision b82779648dfd3814df4e381f086326ec70fd791f)
1 /*
2  * Copyright 2013 Red Hat Inc.
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8  * and/or sell copies of the Software, and to permit persons to whom the
9  * Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice shall be included in
12  * all copies or substantial portions of the Software.
13  *
14  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17  * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20  * OTHER DEALINGS IN THE SOFTWARE.
21  *
22  * Authors: Dave Airlie
23  *          Alon Levy
24  */
25 
26 #include <linux/pci.h>
27 #include <linux/uaccess.h>
28 
29 #include "qxl_drv.h"
30 #include "qxl_object.h"
31 
32 /*
33  * TODO: allocating a new gem(in qxl_bo) for each request.
34  * This is wasteful since bo's are page aligned.
35  */
qxl_alloc_ioctl(struct drm_device * dev,void * data,struct drm_file * file_priv)36 int qxl_alloc_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv)
37 {
38 	struct qxl_device *qdev = to_qxl(dev);
39 	struct drm_qxl_alloc *qxl_alloc = data;
40 	int ret;
41 	uint32_t handle;
42 	u32 domain = QXL_GEM_DOMAIN_VRAM;
43 
44 	if (qxl_alloc->size == 0) {
45 		DRM_ERROR("invalid size %d\n", qxl_alloc->size);
46 		return -EINVAL;
47 	}
48 	ret = qxl_gem_object_create_with_handle(qdev, file_priv,
49 						domain,
50 						qxl_alloc->size,
51 						NULL,
52 						NULL, &handle);
53 	if (ret) {
54 		DRM_ERROR("%s: failed to create gem ret=%d\n",
55 			  __func__, ret);
56 		return -ENOMEM;
57 	}
58 	qxl_alloc->handle = handle;
59 	return 0;
60 }
61 
qxl_map_ioctl(struct drm_device * dev,void * data,struct drm_file * file_priv)62 int qxl_map_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv)
63 {
64 	struct qxl_device *qdev = to_qxl(dev);
65 	struct drm_qxl_map *qxl_map = data;
66 
67 	return drm_gem_ttm_dumb_map_offset(file_priv, &qdev->ddev, qxl_map->handle,
68 					   &qxl_map->offset);
69 }
70 
71 struct qxl_reloc_info {
72 	int type;
73 	struct qxl_bo *dst_bo;
74 	uint32_t dst_offset;
75 	struct qxl_bo *src_bo;
76 	int src_offset;
77 };
78 
79 /*
80  * dst must be validated, i.e. whole bo on vram/surfacesram (right now all bo's
81  * are on vram).
82  * *(dst + dst_off) = qxl_bo_physical_address(src, src_off)
83  */
84 static void
apply_reloc(struct qxl_device * qdev,struct qxl_reloc_info * info)85 apply_reloc(struct qxl_device *qdev, struct qxl_reloc_info *info)
86 {
87 	void *reloc_page;
88 
89 	reloc_page = qxl_bo_kmap_atomic_page(qdev, info->dst_bo, info->dst_offset & PAGE_MASK);
90 	*(uint64_t *)(reloc_page + (info->dst_offset & ~PAGE_MASK)) = qxl_bo_physical_address(qdev,
91 											      info->src_bo,
92 											      info->src_offset);
93 	qxl_bo_kunmap_atomic_page(qdev, info->dst_bo, reloc_page);
94 }
95 
96 static void
apply_surf_reloc(struct qxl_device * qdev,struct qxl_reloc_info * info)97 apply_surf_reloc(struct qxl_device *qdev, struct qxl_reloc_info *info)
98 {
99 	uint32_t id = 0;
100 	void *reloc_page;
101 
102 	if (info->src_bo && !info->src_bo->is_primary)
103 		id = info->src_bo->surface_id;
104 
105 	reloc_page = qxl_bo_kmap_atomic_page(qdev, info->dst_bo, info->dst_offset & PAGE_MASK);
106 	*(uint32_t *)(reloc_page + (info->dst_offset & ~PAGE_MASK)) = id;
107 	qxl_bo_kunmap_atomic_page(qdev, info->dst_bo, reloc_page);
108 }
109 
110 /* return holding the reference to this object */
qxlhw_handle_to_bo(struct drm_file * file_priv,uint64_t handle,struct qxl_release * release,struct qxl_bo ** qbo_p)111 static int qxlhw_handle_to_bo(struct drm_file *file_priv, uint64_t handle,
112 			      struct qxl_release *release, struct qxl_bo **qbo_p)
113 {
114 	struct drm_gem_object *gobj;
115 	struct qxl_bo *qobj;
116 	int ret;
117 
118 	gobj = drm_gem_object_lookup(file_priv, handle);
119 	if (!gobj)
120 		return -EINVAL;
121 
122 	qobj = gem_to_qxl_bo(gobj);
123 
124 	ret = qxl_release_list_add(release, qobj);
125 	drm_gem_object_put(gobj);
126 	if (ret)
127 		return ret;
128 
129 	*qbo_p = qobj;
130 	return 0;
131 }
132 
133 /*
134  * Usage of execbuffer:
135  * Relocations need to take into account the full QXLDrawable size.
136  * However, the command as passed from user space must *not* contain the initial
137  * QXLReleaseInfo struct (first XXX bytes)
138  */
qxl_process_single_command(struct qxl_device * qdev,struct drm_qxl_command * cmd,struct drm_file * file_priv)139 static int qxl_process_single_command(struct qxl_device *qdev,
140 				      struct drm_qxl_command *cmd,
141 				      struct drm_file *file_priv)
142 {
143 	struct qxl_reloc_info *reloc_info;
144 	int release_type;
145 	struct qxl_release *release;
146 	struct qxl_bo *cmd_bo;
147 	void *fb_cmd;
148 	int i, ret;
149 	int unwritten;
150 
151 	switch (cmd->type) {
152 	case QXL_CMD_DRAW:
153 		release_type = QXL_RELEASE_DRAWABLE;
154 		break;
155 	case QXL_CMD_SURFACE:
156 	case QXL_CMD_CURSOR:
157 	default:
158 		DRM_DEBUG("Only draw commands in execbuffers\n");
159 		return -EINVAL;
160 	}
161 
162 	if (cmd->command_size > PAGE_SIZE - sizeof(union qxl_release_info))
163 		return -EINVAL;
164 
165 	if (!access_ok(u64_to_user_ptr(cmd->command),
166 		       cmd->command_size))
167 		return -EFAULT;
168 
169 	reloc_info = kmalloc_array(cmd->relocs_num,
170 				   sizeof(struct qxl_reloc_info), GFP_KERNEL);
171 	if (!reloc_info)
172 		return -ENOMEM;
173 
174 	ret = qxl_alloc_release_reserved(qdev,
175 					 sizeof(union qxl_release_info) +
176 					 cmd->command_size,
177 					 release_type,
178 					 &release,
179 					 &cmd_bo);
180 	if (ret)
181 		goto out_free_reloc;
182 
183 	/* TODO copy slow path code from i915 */
184 	fb_cmd = qxl_bo_kmap_atomic_page(qdev, cmd_bo, (release->release_offset & PAGE_MASK));
185 	unwritten = __copy_from_user_inatomic_nocache
186 		(fb_cmd + sizeof(union qxl_release_info) + (release->release_offset & ~PAGE_MASK),
187 		 u64_to_user_ptr(cmd->command), cmd->command_size);
188 
189 	{
190 		struct qxl_drawable *draw = fb_cmd;
191 
192 		draw->mm_time = qdev->rom->mm_clock;
193 	}
194 
195 	qxl_bo_kunmap_atomic_page(qdev, cmd_bo, fb_cmd);
196 	if (unwritten) {
197 		DRM_ERROR("got unwritten %d\n", unwritten);
198 		ret = -EFAULT;
199 		goto out_free_release;
200 	}
201 
202 	/* fill out reloc info structs */
203 	for (i = 0; i < cmd->relocs_num; ++i) {
204 		struct drm_qxl_reloc reloc;
205 		struct drm_qxl_reloc __user *u = u64_to_user_ptr(cmd->relocs);
206 
207 		if (copy_from_user(&reloc, u + i, sizeof(reloc))) {
208 			ret = -EFAULT;
209 			goto out_free_bos;
210 		}
211 
212 		/* add the bos to the list of bos to validate -
213 		   need to validate first then process relocs? */
214 		if (reloc.reloc_type != QXL_RELOC_TYPE_BO && reloc.reloc_type != QXL_RELOC_TYPE_SURF) {
215 			DRM_DEBUG("unknown reloc type %d\n", reloc.reloc_type);
216 
217 			ret = -EINVAL;
218 			goto out_free_bos;
219 		}
220 		reloc_info[i].type = reloc.reloc_type;
221 
222 		if (reloc.dst_handle) {
223 			ret = qxlhw_handle_to_bo(file_priv, reloc.dst_handle, release,
224 						 &reloc_info[i].dst_bo);
225 			if (ret)
226 				goto out_free_bos;
227 			reloc_info[i].dst_offset = reloc.dst_offset;
228 		} else {
229 			reloc_info[i].dst_bo = cmd_bo;
230 			reloc_info[i].dst_offset = reloc.dst_offset + release->release_offset;
231 		}
232 
233 		/* reserve and validate the reloc dst bo */
234 		if (reloc.reloc_type == QXL_RELOC_TYPE_BO || reloc.src_handle) {
235 			ret = qxlhw_handle_to_bo(file_priv, reloc.src_handle, release,
236 						 &reloc_info[i].src_bo);
237 			if (ret)
238 				goto out_free_bos;
239 			reloc_info[i].src_offset = reloc.src_offset;
240 		} else {
241 			reloc_info[i].src_bo = NULL;
242 			reloc_info[i].src_offset = 0;
243 		}
244 	}
245 
246 	/* validate all buffers */
247 	ret = qxl_release_reserve_list(release, false);
248 	if (ret)
249 		goto out_free_bos;
250 
251 	for (i = 0; i < cmd->relocs_num; ++i) {
252 		if (reloc_info[i].type == QXL_RELOC_TYPE_BO)
253 			apply_reloc(qdev, &reloc_info[i]);
254 		else if (reloc_info[i].type == QXL_RELOC_TYPE_SURF)
255 			apply_surf_reloc(qdev, &reloc_info[i]);
256 	}
257 
258 	qxl_release_fence_buffer_objects(release);
259 	ret = qxl_push_command_ring_release(qdev, release, cmd->type, true);
260 
261 out_free_bos:
262 out_free_release:
263 	if (ret)
264 		qxl_release_free(qdev, release);
265 out_free_reloc:
266 	kfree(reloc_info);
267 	return ret;
268 }
269 
qxl_execbuffer_ioctl(struct drm_device * dev,void * data,struct drm_file * file_priv)270 int qxl_execbuffer_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv)
271 {
272 	struct qxl_device *qdev = to_qxl(dev);
273 	struct drm_qxl_execbuffer *execbuffer = data;
274 	struct drm_qxl_command user_cmd;
275 	int cmd_num;
276 	int ret;
277 
278 	for (cmd_num = 0; cmd_num < execbuffer->commands_num; ++cmd_num) {
279 
280 		struct drm_qxl_command __user *commands =
281 			u64_to_user_ptr(execbuffer->commands);
282 
283 		if (copy_from_user(&user_cmd, commands + cmd_num,
284 				       sizeof(user_cmd)))
285 			return -EFAULT;
286 
287 		ret = qxl_process_single_command(qdev, &user_cmd, file_priv);
288 		if (ret)
289 			return ret;
290 	}
291 	return 0;
292 }
293 
qxl_update_area_ioctl(struct drm_device * dev,void * data,struct drm_file * file)294 int qxl_update_area_ioctl(struct drm_device *dev, void *data, struct drm_file *file)
295 {
296 	struct qxl_device *qdev = to_qxl(dev);
297 	struct drm_qxl_update_area *update_area = data;
298 	struct qxl_rect area = {.left = update_area->left,
299 				.top = update_area->top,
300 				.right = update_area->right,
301 				.bottom = update_area->bottom};
302 	int ret;
303 	struct drm_gem_object *gobj = NULL;
304 	struct qxl_bo *qobj = NULL;
305 	struct ttm_operation_ctx ctx = { true, false };
306 
307 	if (update_area->left >= update_area->right ||
308 	    update_area->top >= update_area->bottom)
309 		return -EINVAL;
310 
311 	gobj = drm_gem_object_lookup(file, update_area->handle);
312 	if (gobj == NULL)
313 		return -ENOENT;
314 
315 	qobj = gem_to_qxl_bo(gobj);
316 
317 	ret = qxl_bo_reserve(qobj);
318 	if (ret)
319 		goto out;
320 
321 	if (!qobj->tbo.pin_count) {
322 		qxl_ttm_placement_from_domain(qobj, qobj->type);
323 		ret = ttm_bo_validate(&qobj->tbo, &qobj->placement, &ctx);
324 		if (unlikely(ret))
325 			goto out;
326 	}
327 
328 	ret = qxl_bo_check_id(qdev, qobj);
329 	if (ret)
330 		goto out2;
331 	if (!qobj->surface_id)
332 		DRM_ERROR("got update area for surface with no id %d\n", update_area->handle);
333 	ret = qxl_io_update_area(qdev, qobj, &area);
334 
335 out2:
336 	qxl_bo_unreserve(qobj);
337 
338 out:
339 	drm_gem_object_put(gobj);
340 	return ret;
341 }
342 
qxl_getparam_ioctl(struct drm_device * dev,void * data,struct drm_file * file_priv)343 int qxl_getparam_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv)
344 {
345 	struct qxl_device *qdev = to_qxl(dev);
346 	struct drm_qxl_getparam *param = data;
347 
348 	switch (param->param) {
349 	case QXL_PARAM_NUM_SURFACES:
350 		param->value = qdev->rom->n_surfaces;
351 		break;
352 	case QXL_PARAM_MAX_RELOCS:
353 		param->value = QXL_MAX_RES;
354 		break;
355 	default:
356 		return -EINVAL;
357 	}
358 	return 0;
359 }
360 
qxl_clientcap_ioctl(struct drm_device * dev,void * data,struct drm_file * file_priv)361 int qxl_clientcap_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv)
362 {
363 	struct qxl_device *qdev = to_qxl(dev);
364 	struct pci_dev *pdev = to_pci_dev(dev->dev);
365 	struct drm_qxl_clientcap *param = data;
366 	int byte, idx;
367 
368 	byte = param->index / 8;
369 	idx = param->index % 8;
370 
371 	if (pdev->revision < 4)
372 		return -ENOSYS;
373 
374 	if (byte >= 58)
375 		return -ENOSYS;
376 
377 	if (qdev->rom->client_capabilities[byte] & (1 << idx))
378 		return 0;
379 	return -ENOSYS;
380 }
381 
qxl_alloc_surf_ioctl(struct drm_device * dev,void * data,struct drm_file * file)382 int qxl_alloc_surf_ioctl(struct drm_device *dev, void *data, struct drm_file *file)
383 {
384 	struct qxl_device *qdev = to_qxl(dev);
385 	struct drm_qxl_alloc_surf *param = data;
386 	int handle;
387 	int ret;
388 	int size, actual_stride;
389 	struct qxl_surface surf;
390 
391 	/* work out size allocate bo with handle */
392 	actual_stride = param->stride < 0 ? -param->stride : param->stride;
393 	size = actual_stride * param->height + actual_stride;
394 
395 	surf.format = param->format;
396 	surf.width = param->width;
397 	surf.height = param->height;
398 	surf.stride = param->stride;
399 	surf.data = 0;
400 
401 	ret = qxl_gem_object_create_with_handle(qdev, file,
402 						QXL_GEM_DOMAIN_SURFACE,
403 						size,
404 						&surf,
405 						NULL, &handle);
406 	if (ret) {
407 		DRM_ERROR("%s: failed to create gem ret=%d\n",
408 			  __func__, ret);
409 		return -ENOMEM;
410 	} else
411 		param->handle = handle;
412 	return ret;
413 }
414