1 /************************************************************************** 2 * 3 * Copyright © 2009-2015 VMware, Inc., Palo Alto, CA., USA 4 * All Rights Reserved. 5 * 6 * Permission is hereby granted, free of charge, to any person obtaining a 7 * copy of this software and associated documentation files (the 8 * "Software"), to deal in the Software without restriction, including 9 * without limitation the rights to use, copy, modify, merge, publish, 10 * distribute, sub license, and/or sell copies of the Software, and to 11 * permit persons to whom the Software is furnished to do so, subject to 12 * the following conditions: 13 * 14 * The above copyright notice and this permission notice (including the 15 * next paragraph) shall be included in all copies or substantial portions 16 * of the Software. 17 * 18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 20 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL 21 * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, 22 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR 23 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE 24 * USE OR OTHER DEALINGS IN THE SOFTWARE. 25 * 26 **************************************************************************/ 27 28 #include "vmwgfx_drv.h" 29 #include <drm/vmwgfx_drm.h> 30 #include "vmwgfx_kms.h" 31 #include "device_include/svga3d_caps.h" 32 33 struct svga_3d_compat_cap { 34 SVGA3dCapsRecordHeader header; 35 SVGA3dCapPair pairs[SVGA3D_DEVCAP_MAX]; 36 }; 37 38 int vmw_getparam_ioctl(struct drm_device *dev, void *data, 39 struct drm_file *file_priv) 40 { 41 struct vmw_private *dev_priv = vmw_priv(dev); 42 struct drm_vmw_getparam_arg *param = 43 (struct drm_vmw_getparam_arg *)data; 44 struct vmw_fpriv *vmw_fp = vmw_fpriv(file_priv); 45 46 switch (param->param) { 47 case DRM_VMW_PARAM_NUM_STREAMS: 48 param->value = vmw_overlay_num_overlays(dev_priv); 49 break; 50 case DRM_VMW_PARAM_NUM_FREE_STREAMS: 51 param->value = vmw_overlay_num_free_overlays(dev_priv); 52 break; 53 case DRM_VMW_PARAM_3D: 54 param->value = vmw_fifo_have_3d(dev_priv) ? 1 : 0; 55 break; 56 case DRM_VMW_PARAM_HW_CAPS: 57 param->value = dev_priv->capabilities; 58 break; 59 case DRM_VMW_PARAM_FIFO_CAPS: 60 param->value = dev_priv->fifo.capabilities; 61 break; 62 case DRM_VMW_PARAM_MAX_FB_SIZE: 63 param->value = dev_priv->prim_bb_mem; 64 break; 65 case DRM_VMW_PARAM_FIFO_HW_VERSION: 66 { 67 u32 *fifo_mem = dev_priv->mmio_virt; 68 const struct vmw_fifo_state *fifo = &dev_priv->fifo; 69 70 if ((dev_priv->capabilities & SVGA_CAP_GBOBJECTS)) { 71 param->value = SVGA3D_HWVERSION_WS8_B1; 72 break; 73 } 74 75 param->value = 76 vmw_mmio_read(fifo_mem + 77 ((fifo->capabilities & 78 SVGA_FIFO_CAP_3D_HWVERSION_REVISED) ? 79 SVGA_FIFO_3D_HWVERSION_REVISED : 80 SVGA_FIFO_3D_HWVERSION)); 81 break; 82 } 83 case DRM_VMW_PARAM_MAX_SURF_MEMORY: 84 if ((dev_priv->capabilities & SVGA_CAP_GBOBJECTS) && 85 !vmw_fp->gb_aware) 86 param->value = dev_priv->max_mob_pages * PAGE_SIZE / 2; 87 else 88 param->value = dev_priv->memory_size; 89 break; 90 case DRM_VMW_PARAM_3D_CAPS_SIZE: 91 if ((dev_priv->capabilities & SVGA_CAP_GBOBJECTS) && 92 vmw_fp->gb_aware) 93 param->value = SVGA3D_DEVCAP_MAX * sizeof(uint32_t); 94 else if (dev_priv->capabilities & SVGA_CAP_GBOBJECTS) 95 param->value = sizeof(struct svga_3d_compat_cap) + 96 sizeof(uint32_t); 97 else 98 param->value = (SVGA_FIFO_3D_CAPS_LAST - 99 SVGA_FIFO_3D_CAPS + 1) * 100 sizeof(uint32_t); 101 break; 102 case DRM_VMW_PARAM_MAX_MOB_MEMORY: 103 vmw_fp->gb_aware = true; 104 param->value = dev_priv->max_mob_pages * PAGE_SIZE; 105 break; 106 case DRM_VMW_PARAM_MAX_MOB_SIZE: 107 param->value = dev_priv->max_mob_size; 108 break; 109 case DRM_VMW_PARAM_SCREEN_TARGET: 110 param->value = 111 (dev_priv->active_display_unit == vmw_du_screen_target); 112 break; 113 case DRM_VMW_PARAM_DX: 114 param->value = dev_priv->has_dx; 115 break; 116 default: 117 DRM_ERROR("Illegal vmwgfx get param request: %d\n", 118 param->param); 119 return -EINVAL; 120 } 121 122 return 0; 123 } 124 125 static u32 vmw_mask_multisample(unsigned int cap, u32 fmt_value) 126 { 127 /* If the header is updated, update the format test as well! */ 128 BUILD_BUG_ON(SVGA3D_DEVCAP_DXFMT_BC5_UNORM + 1 != SVGA3D_DEVCAP_MAX); 129 130 if (cap >= SVGA3D_DEVCAP_DXFMT_X8R8G8B8 && 131 cap <= SVGA3D_DEVCAP_DXFMT_BC5_UNORM) 132 fmt_value &= ~(SVGADX_DXFMT_MULTISAMPLE_2 | 133 SVGADX_DXFMT_MULTISAMPLE_4 | 134 SVGADX_DXFMT_MULTISAMPLE_8); 135 else if (cap == SVGA3D_DEVCAP_MULTISAMPLE_MASKABLESAMPLES) 136 return 0; 137 138 return fmt_value; 139 } 140 141 static int vmw_fill_compat_cap(struct vmw_private *dev_priv, void *bounce, 142 size_t size) 143 { 144 struct svga_3d_compat_cap *compat_cap = 145 (struct svga_3d_compat_cap *) bounce; 146 unsigned int i; 147 size_t pair_offset = offsetof(struct svga_3d_compat_cap, pairs); 148 unsigned int max_size; 149 150 if (size < pair_offset) 151 return -EINVAL; 152 153 max_size = (size - pair_offset) / sizeof(SVGA3dCapPair); 154 155 if (max_size > SVGA3D_DEVCAP_MAX) 156 max_size = SVGA3D_DEVCAP_MAX; 157 158 compat_cap->header.length = 159 (pair_offset + max_size * sizeof(SVGA3dCapPair)) / sizeof(u32); 160 compat_cap->header.type = SVGA3DCAPS_RECORD_DEVCAPS; 161 162 spin_lock(&dev_priv->cap_lock); 163 for (i = 0; i < max_size; ++i) { 164 vmw_write(dev_priv, SVGA_REG_DEV_CAP, i); 165 compat_cap->pairs[i][0] = i; 166 compat_cap->pairs[i][1] = vmw_mask_multisample 167 (i, vmw_read(dev_priv, SVGA_REG_DEV_CAP)); 168 } 169 spin_unlock(&dev_priv->cap_lock); 170 171 return 0; 172 } 173 174 175 int vmw_get_cap_3d_ioctl(struct drm_device *dev, void *data, 176 struct drm_file *file_priv) 177 { 178 struct drm_vmw_get_3d_cap_arg *arg = 179 (struct drm_vmw_get_3d_cap_arg *) data; 180 struct vmw_private *dev_priv = vmw_priv(dev); 181 uint32_t size; 182 u32 *fifo_mem; 183 void __user *buffer = (void __user *)((unsigned long)(arg->buffer)); 184 void *bounce; 185 int ret; 186 bool gb_objects = !!(dev_priv->capabilities & SVGA_CAP_GBOBJECTS); 187 struct vmw_fpriv *vmw_fp = vmw_fpriv(file_priv); 188 189 if (unlikely(arg->pad64 != 0)) { 190 DRM_ERROR("Illegal GET_3D_CAP argument.\n"); 191 return -EINVAL; 192 } 193 194 if (gb_objects && vmw_fp->gb_aware) 195 size = SVGA3D_DEVCAP_MAX * sizeof(uint32_t); 196 else if (gb_objects) 197 size = sizeof(struct svga_3d_compat_cap) + sizeof(uint32_t); 198 else 199 size = (SVGA_FIFO_3D_CAPS_LAST - SVGA_FIFO_3D_CAPS + 1) * 200 sizeof(uint32_t); 201 202 if (arg->max_size < size) 203 size = arg->max_size; 204 205 bounce = vzalloc(size); 206 if (unlikely(bounce == NULL)) { 207 DRM_ERROR("Failed to allocate bounce buffer for 3D caps.\n"); 208 return -ENOMEM; 209 } 210 211 if (gb_objects && vmw_fp->gb_aware) { 212 int i, num; 213 uint32_t *bounce32 = (uint32_t *) bounce; 214 215 num = size / sizeof(uint32_t); 216 if (num > SVGA3D_DEVCAP_MAX) 217 num = SVGA3D_DEVCAP_MAX; 218 219 spin_lock(&dev_priv->cap_lock); 220 for (i = 0; i < num; ++i) { 221 vmw_write(dev_priv, SVGA_REG_DEV_CAP, i); 222 *bounce32++ = vmw_mask_multisample 223 (i, vmw_read(dev_priv, SVGA_REG_DEV_CAP)); 224 } 225 spin_unlock(&dev_priv->cap_lock); 226 } else if (gb_objects) { 227 ret = vmw_fill_compat_cap(dev_priv, bounce, size); 228 if (unlikely(ret != 0)) 229 goto out_err; 230 } else { 231 fifo_mem = dev_priv->mmio_virt; 232 memcpy(bounce, &fifo_mem[SVGA_FIFO_3D_CAPS], size); 233 } 234 235 ret = copy_to_user(buffer, bounce, size); 236 if (ret) 237 ret = -EFAULT; 238 out_err: 239 vfree(bounce); 240 241 if (unlikely(ret != 0)) 242 DRM_ERROR("Failed to report 3D caps info.\n"); 243 244 return ret; 245 } 246 247 int vmw_present_ioctl(struct drm_device *dev, void *data, 248 struct drm_file *file_priv) 249 { 250 struct ttm_object_file *tfile = vmw_fpriv(file_priv)->tfile; 251 struct vmw_private *dev_priv = vmw_priv(dev); 252 struct drm_vmw_present_arg *arg = 253 (struct drm_vmw_present_arg *)data; 254 struct vmw_surface *surface; 255 struct drm_vmw_rect __user *clips_ptr; 256 struct drm_vmw_rect *clips = NULL; 257 struct drm_framebuffer *fb; 258 struct vmw_framebuffer *vfb; 259 struct vmw_resource *res; 260 uint32_t num_clips; 261 int ret; 262 263 num_clips = arg->num_clips; 264 clips_ptr = (struct drm_vmw_rect __user *)(unsigned long)arg->clips_ptr; 265 266 if (unlikely(num_clips == 0)) 267 return 0; 268 269 if (clips_ptr == NULL) { 270 DRM_ERROR("Variable clips_ptr must be specified.\n"); 271 ret = -EINVAL; 272 goto out_clips; 273 } 274 275 clips = kcalloc(num_clips, sizeof(*clips), GFP_KERNEL); 276 if (clips == NULL) { 277 DRM_ERROR("Failed to allocate clip rect list.\n"); 278 ret = -ENOMEM; 279 goto out_clips; 280 } 281 282 ret = copy_from_user(clips, clips_ptr, num_clips * sizeof(*clips)); 283 if (ret) { 284 DRM_ERROR("Failed to copy clip rects from userspace.\n"); 285 ret = -EFAULT; 286 goto out_no_copy; 287 } 288 289 drm_modeset_lock_all(dev); 290 291 fb = drm_framebuffer_lookup(dev, arg->fb_id); 292 if (!fb) { 293 DRM_ERROR("Invalid framebuffer id.\n"); 294 ret = -ENOENT; 295 goto out_no_fb; 296 } 297 vfb = vmw_framebuffer_to_vfb(fb); 298 299 ret = ttm_read_lock(&dev_priv->reservation_sem, true); 300 if (unlikely(ret != 0)) 301 goto out_no_ttm_lock; 302 303 ret = vmw_user_resource_lookup_handle(dev_priv, tfile, arg->sid, 304 user_surface_converter, 305 &res); 306 if (ret) 307 goto out_no_surface; 308 309 surface = vmw_res_to_srf(res); 310 ret = vmw_kms_present(dev_priv, file_priv, 311 vfb, surface, arg->sid, 312 arg->dest_x, arg->dest_y, 313 clips, num_clips); 314 315 /* vmw_user_surface_lookup takes one ref so does new_fb */ 316 vmw_surface_unreference(&surface); 317 318 out_no_surface: 319 ttm_read_unlock(&dev_priv->reservation_sem); 320 out_no_ttm_lock: 321 drm_framebuffer_unreference(fb); 322 out_no_fb: 323 drm_modeset_unlock_all(dev); 324 out_no_copy: 325 kfree(clips); 326 out_clips: 327 return ret; 328 } 329 330 int vmw_present_readback_ioctl(struct drm_device *dev, void *data, 331 struct drm_file *file_priv) 332 { 333 struct vmw_private *dev_priv = vmw_priv(dev); 334 struct drm_vmw_present_readback_arg *arg = 335 (struct drm_vmw_present_readback_arg *)data; 336 struct drm_vmw_fence_rep __user *user_fence_rep = 337 (struct drm_vmw_fence_rep __user *) 338 (unsigned long)arg->fence_rep; 339 struct drm_vmw_rect __user *clips_ptr; 340 struct drm_vmw_rect *clips = NULL; 341 struct drm_framebuffer *fb; 342 struct vmw_framebuffer *vfb; 343 uint32_t num_clips; 344 int ret; 345 346 num_clips = arg->num_clips; 347 clips_ptr = (struct drm_vmw_rect __user *)(unsigned long)arg->clips_ptr; 348 349 if (unlikely(num_clips == 0)) 350 return 0; 351 352 if (clips_ptr == NULL) { 353 DRM_ERROR("Argument clips_ptr must be specified.\n"); 354 ret = -EINVAL; 355 goto out_clips; 356 } 357 358 clips = kcalloc(num_clips, sizeof(*clips), GFP_KERNEL); 359 if (clips == NULL) { 360 DRM_ERROR("Failed to allocate clip rect list.\n"); 361 ret = -ENOMEM; 362 goto out_clips; 363 } 364 365 ret = copy_from_user(clips, clips_ptr, num_clips * sizeof(*clips)); 366 if (ret) { 367 DRM_ERROR("Failed to copy clip rects from userspace.\n"); 368 ret = -EFAULT; 369 goto out_no_copy; 370 } 371 372 drm_modeset_lock_all(dev); 373 374 fb = drm_framebuffer_lookup(dev, arg->fb_id); 375 if (!fb) { 376 DRM_ERROR("Invalid framebuffer id.\n"); 377 ret = -ENOENT; 378 goto out_no_fb; 379 } 380 381 vfb = vmw_framebuffer_to_vfb(fb); 382 if (!vfb->dmabuf) { 383 DRM_ERROR("Framebuffer not dmabuf backed.\n"); 384 ret = -EINVAL; 385 goto out_no_ttm_lock; 386 } 387 388 ret = ttm_read_lock(&dev_priv->reservation_sem, true); 389 if (unlikely(ret != 0)) 390 goto out_no_ttm_lock; 391 392 ret = vmw_kms_readback(dev_priv, file_priv, 393 vfb, user_fence_rep, 394 clips, num_clips); 395 396 ttm_read_unlock(&dev_priv->reservation_sem); 397 out_no_ttm_lock: 398 drm_framebuffer_unreference(fb); 399 out_no_fb: 400 drm_modeset_unlock_all(dev); 401 out_no_copy: 402 kfree(clips); 403 out_clips: 404 return ret; 405 } 406 407 408 /** 409 * vmw_fops_poll - wrapper around the drm_poll function 410 * 411 * @filp: See the linux fops poll documentation. 412 * @wait: See the linux fops poll documentation. 413 * 414 * Wrapper around the drm_poll function that makes sure the device is 415 * processing the fifo if drm_poll decides to wait. 416 */ 417 unsigned int vmw_fops_poll(struct file *filp, struct poll_table_struct *wait) 418 { 419 struct drm_file *file_priv = filp->private_data; 420 struct vmw_private *dev_priv = 421 vmw_priv(file_priv->minor->dev); 422 423 vmw_fifo_ping_host(dev_priv, SVGA_SYNC_GENERIC); 424 return drm_poll(filp, wait); 425 } 426 427 428 /** 429 * vmw_fops_read - wrapper around the drm_read function 430 * 431 * @filp: See the linux fops read documentation. 432 * @buffer: See the linux fops read documentation. 433 * @count: See the linux fops read documentation. 434 * offset: See the linux fops read documentation. 435 * 436 * Wrapper around the drm_read function that makes sure the device is 437 * processing the fifo if drm_read decides to wait. 438 */ 439 ssize_t vmw_fops_read(struct file *filp, char __user *buffer, 440 size_t count, loff_t *offset) 441 { 442 struct drm_file *file_priv = filp->private_data; 443 struct vmw_private *dev_priv = 444 vmw_priv(file_priv->minor->dev); 445 446 vmw_fifo_ping_host(dev_priv, SVGA_SYNC_GENERIC); 447 return drm_read(filp, buffer, count, offset); 448 } 449