1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * Copyright (C) 2013 Red Hat 4 * Author: Rob Clark <robdclark@gmail.com> 5 */ 6 7 #include <linux/dma-map-ops.h> 8 #include <linux/vmalloc.h> 9 #include <linux/spinlock.h> 10 #include <linux/shmem_fs.h> 11 #include <linux/dma-buf.h> 12 13 #include <drm/drm_dumb_buffers.h> 14 #include <drm/drm_prime.h> 15 #include <drm/drm_file.h> 16 #include <drm/drm_fourcc.h> 17 18 #include <trace/events/gpu_mem.h> 19 20 #include "msm_drv.h" 21 #include "msm_gem.h" 22 #include "msm_gpu.h" 23 #include "msm_kms.h" 24 25 static void update_device_mem(struct msm_drm_private *priv, ssize_t size) 26 { 27 uint64_t total_mem = atomic64_add_return(size, &priv->total_mem); 28 trace_gpu_mem_total(0, 0, total_mem); 29 } 30 31 static void update_ctx_mem(struct drm_file *file, ssize_t size) 32 { 33 struct msm_context *ctx = file->driver_priv; 34 uint64_t ctx_mem = atomic64_add_return(size, &ctx->ctx_mem); 35 36 rcu_read_lock(); /* Locks file->pid! */ 37 trace_gpu_mem_total(0, pid_nr(rcu_dereference(file->pid)), ctx_mem); 38 rcu_read_unlock(); 39 40 } 41 42 static int msm_gem_open(struct drm_gem_object *obj, struct drm_file *file) 43 { 44 msm_gem_vma_get(obj); 45 update_ctx_mem(file, obj->size); 46 return 0; 47 } 48 49 static void put_iova_spaces(struct drm_gem_object *obj, struct drm_gpuvm *vm, 50 bool close, const char *reason); 51 52 static void msm_gem_close(struct drm_gem_object *obj, struct drm_file *file) 53 { 54 struct msm_context *ctx = file->driver_priv; 55 struct drm_exec exec; 56 57 update_ctx_mem(file, -obj->size); 58 msm_gem_vma_put(obj); 59 60 /* 61 * If VM isn't created yet, nothing to cleanup. And in fact calling 62 * put_iova_spaces() with vm=NULL would be bad, in that it will tear- 63 * down the mappings of shared buffers in other contexts. 64 */ 65 if (!ctx->vm) 66 return; 67 68 /* 69 * VM_BIND does not depend on implicit teardown of VMAs on handle 70 * close, but instead on implicit teardown of the VM when the device 71 * is closed (see msm_gem_vm_close()) 72 */ 73 if (msm_context_is_vmbind(ctx)) 74 return; 75 76 /* 77 * TODO we might need to kick this to a queue to avoid blocking 78 * in CLOSE ioctl 79 */ 80 dma_resv_wait_timeout(obj->resv, DMA_RESV_USAGE_BOOKKEEP, false, 81 MAX_SCHEDULE_TIMEOUT); 82 83 msm_gem_lock_vm_and_obj(&exec, obj, ctx->vm); 84 put_iova_spaces(obj, ctx->vm, true, "close"); 85 drm_exec_fini(&exec); /* drop locks */ 86 } 87 88 /* 89 * Get/put for kms->vm VMA 90 */ 91 92 void msm_gem_vma_get(struct drm_gem_object *obj) 93 { 94 atomic_inc(&to_msm_bo(obj)->vma_ref); 95 } 96 97 void msm_gem_vma_put(struct drm_gem_object *obj) 98 { 99 struct msm_drm_private *priv = obj->dev->dev_private; 100 101 if (atomic_dec_return(&to_msm_bo(obj)->vma_ref)) 102 return; 103 104 if (!priv->kms) 105 return; 106 107 #ifdef CONFIG_DRM_MSM_KMS 108 struct drm_exec exec; 109 110 msm_gem_lock_vm_and_obj(&exec, obj, priv->kms->vm); 111 put_iova_spaces(obj, priv->kms->vm, true, "vma_put"); 112 drm_exec_fini(&exec); /* drop locks */ 113 #endif 114 } 115 116 /* 117 * Cache sync.. this is a bit over-complicated, to fit dma-mapping 118 * API. Really GPU cache is out of scope here (handled on cmdstream) 119 * and all we need to do is invalidate newly allocated pages before 120 * mapping to CPU as uncached/writecombine. 121 * 122 * On top of this, we have the added headache, that depending on 123 * display generation, the display's iommu may be wired up to either 124 * the toplevel drm device (mdss), or to the mdp sub-node, meaning 125 * that here we either have dma-direct or iommu ops. 126 * 127 * Let this be a cautionary tail of abstraction gone wrong. 128 */ 129 130 static void sync_for_device(struct msm_gem_object *msm_obj) 131 { 132 struct device *dev = msm_obj->base.dev->dev; 133 134 dma_map_sgtable(dev, msm_obj->sgt, DMA_BIDIRECTIONAL, 0); 135 } 136 137 static void sync_for_cpu(struct msm_gem_object *msm_obj) 138 { 139 struct device *dev = msm_obj->base.dev->dev; 140 141 dma_unmap_sgtable(dev, msm_obj->sgt, DMA_BIDIRECTIONAL, 0); 142 } 143 144 static void update_lru_active(struct drm_gem_object *obj) 145 { 146 struct msm_drm_private *priv = obj->dev->dev_private; 147 struct msm_gem_object *msm_obj = to_msm_bo(obj); 148 149 GEM_WARN_ON(!msm_obj->pages); 150 151 if (msm_obj->pin_count) { 152 drm_gem_lru_move_tail_locked(&priv->lru.pinned, obj); 153 } else if (msm_obj->madv == MSM_MADV_WILLNEED) { 154 drm_gem_lru_move_tail_locked(&priv->lru.willneed, obj); 155 } else { 156 GEM_WARN_ON(msm_obj->madv != MSM_MADV_DONTNEED); 157 158 drm_gem_lru_move_tail_locked(&priv->lru.dontneed, obj); 159 } 160 } 161 162 static void update_lru_locked(struct drm_gem_object *obj) 163 { 164 struct msm_drm_private *priv = obj->dev->dev_private; 165 struct msm_gem_object *msm_obj = to_msm_bo(obj); 166 167 msm_gem_assert_locked(&msm_obj->base); 168 169 if (!msm_obj->pages) { 170 GEM_WARN_ON(msm_obj->pin_count); 171 172 drm_gem_lru_move_tail_locked(&priv->lru.unbacked, obj); 173 } else { 174 update_lru_active(obj); 175 } 176 } 177 178 static void update_lru(struct drm_gem_object *obj) 179 { 180 struct msm_drm_private *priv = obj->dev->dev_private; 181 182 mutex_lock(&priv->lru.lock); 183 update_lru_locked(obj); 184 mutex_unlock(&priv->lru.lock); 185 } 186 187 static struct page **get_pages(struct drm_gem_object *obj) 188 { 189 struct msm_gem_object *msm_obj = to_msm_bo(obj); 190 191 msm_gem_assert_locked(obj); 192 193 if (!msm_obj->pages) { 194 struct drm_device *dev = obj->dev; 195 struct page **p; 196 size_t npages = obj->size >> PAGE_SHIFT; 197 198 p = drm_gem_get_pages(obj); 199 200 if (IS_ERR(p)) { 201 DRM_DEV_ERROR(dev->dev, "could not get pages: %ld\n", 202 PTR_ERR(p)); 203 return p; 204 } 205 206 update_device_mem(dev->dev_private, obj->size); 207 208 msm_obj->pages = p; 209 210 msm_obj->sgt = drm_prime_pages_to_sg(obj->dev, p, npages); 211 if (IS_ERR(msm_obj->sgt)) { 212 void *ptr = ERR_CAST(msm_obj->sgt); 213 214 DRM_DEV_ERROR(dev->dev, "failed to allocate sgt\n"); 215 msm_obj->sgt = NULL; 216 return ptr; 217 } 218 219 /* For non-cached buffers, ensure the new pages are clean 220 * because display controller, GPU, etc. are not coherent: 221 */ 222 if (msm_obj->flags & MSM_BO_WC) 223 sync_for_device(msm_obj); 224 225 update_lru(obj); 226 } 227 228 return msm_obj->pages; 229 } 230 231 static void put_pages(struct drm_gem_object *obj) 232 { 233 struct msm_gem_object *msm_obj = to_msm_bo(obj); 234 235 /* 236 * Skip gpuvm in the object free path to avoid a WARN_ON() splat. 237 * See explaination in msm_gem_assert_locked() 238 */ 239 if (kref_read(&obj->refcount)) 240 drm_gpuvm_bo_gem_evict(obj, true); 241 242 if (msm_obj->pages) { 243 if (msm_obj->sgt) { 244 /* For non-cached buffers, ensure the new 245 * pages are clean because display controller, 246 * GPU, etc. are not coherent: 247 */ 248 if (msm_obj->flags & MSM_BO_WC) 249 sync_for_cpu(msm_obj); 250 251 sg_free_table(msm_obj->sgt); 252 kfree(msm_obj->sgt); 253 msm_obj->sgt = NULL; 254 } 255 256 update_device_mem(obj->dev->dev_private, -obj->size); 257 258 drm_gem_put_pages(obj, msm_obj->pages, true, false); 259 260 msm_obj->pages = NULL; 261 update_lru(obj); 262 } 263 } 264 265 struct page **msm_gem_get_pages_locked(struct drm_gem_object *obj, unsigned madv) 266 { 267 struct msm_gem_object *msm_obj = to_msm_bo(obj); 268 269 msm_gem_assert_locked(obj); 270 271 if (msm_obj->madv > madv) { 272 DRM_DEV_DEBUG_DRIVER(obj->dev->dev, "Invalid madv state: %u vs %u\n", 273 msm_obj->madv, madv); 274 return ERR_PTR(-EBUSY); 275 } 276 277 return get_pages(obj); 278 } 279 280 /* 281 * Update the pin count of the object, call under lru.lock 282 */ 283 void msm_gem_pin_obj_locked(struct drm_gem_object *obj) 284 { 285 struct msm_drm_private *priv = obj->dev->dev_private; 286 287 msm_gem_assert_locked(obj); 288 289 to_msm_bo(obj)->pin_count++; 290 drm_gem_lru_move_tail_locked(&priv->lru.pinned, obj); 291 } 292 293 static void pin_obj_locked(struct drm_gem_object *obj) 294 { 295 struct msm_drm_private *priv = obj->dev->dev_private; 296 297 mutex_lock(&priv->lru.lock); 298 msm_gem_pin_obj_locked(obj); 299 mutex_unlock(&priv->lru.lock); 300 } 301 302 struct page **msm_gem_pin_pages_locked(struct drm_gem_object *obj) 303 { 304 struct page **p; 305 306 msm_gem_assert_locked(obj); 307 308 p = msm_gem_get_pages_locked(obj, MSM_MADV_WILLNEED); 309 if (!IS_ERR(p)) 310 pin_obj_locked(obj); 311 312 return p; 313 } 314 315 void msm_gem_unpin_pages_locked(struct drm_gem_object *obj) 316 { 317 msm_gem_assert_locked(obj); 318 319 msm_gem_unpin_locked(obj); 320 } 321 322 static pgprot_t msm_gem_pgprot(struct msm_gem_object *msm_obj, pgprot_t prot) 323 { 324 if (msm_obj->flags & MSM_BO_WC) 325 return pgprot_writecombine(prot); 326 return prot; 327 } 328 329 static vm_fault_t msm_gem_fault(struct vm_fault *vmf) 330 { 331 struct vm_area_struct *vma = vmf->vma; 332 struct drm_gem_object *obj = vma->vm_private_data; 333 struct msm_gem_object *msm_obj = to_msm_bo(obj); 334 struct page **pages; 335 unsigned long pfn; 336 pgoff_t pgoff; 337 int err; 338 vm_fault_t ret; 339 340 /* 341 * vm_ops.open/drm_gem_mmap_obj and close get and put 342 * a reference on obj. So, we dont need to hold one here. 343 */ 344 err = msm_gem_lock_interruptible(obj); 345 if (err) { 346 ret = VM_FAULT_NOPAGE; 347 goto out; 348 } 349 350 if (GEM_WARN_ON(msm_obj->madv != MSM_MADV_WILLNEED)) { 351 msm_gem_unlock(obj); 352 return VM_FAULT_SIGBUS; 353 } 354 355 /* make sure we have pages attached now */ 356 pages = get_pages(obj); 357 if (IS_ERR(pages)) { 358 ret = vmf_error(PTR_ERR(pages)); 359 goto out_unlock; 360 } 361 362 /* We don't use vmf->pgoff since that has the fake offset: */ 363 pgoff = (vmf->address - vma->vm_start) >> PAGE_SHIFT; 364 365 pfn = page_to_pfn(pages[pgoff]); 366 367 VERB("Inserting %p pfn %lx, pa %lx", (void *)vmf->address, 368 pfn, pfn << PAGE_SHIFT); 369 370 ret = vmf_insert_pfn(vma, vmf->address, pfn); 371 372 out_unlock: 373 msm_gem_unlock(obj); 374 out: 375 return ret; 376 } 377 378 379 static struct drm_gpuva *lookup_vma(struct drm_gem_object *obj, 380 struct drm_gpuvm *vm) 381 { 382 struct drm_gpuvm_bo *vm_bo; 383 384 msm_gem_assert_locked(obj); 385 386 drm_gem_for_each_gpuvm_bo (vm_bo, obj) { 387 struct drm_gpuva *vma; 388 389 drm_gpuvm_bo_for_each_va (vma, vm_bo) { 390 if (vma->vm == vm) { 391 /* lookup_vma() should only be used in paths 392 * with at most one vma per vm 393 */ 394 GEM_WARN_ON(!list_is_singular(&vm_bo->list.gpuva)); 395 396 return vma; 397 } 398 } 399 } 400 401 return NULL; 402 } 403 404 /* 405 * If close is true, this also closes the VMA (releasing the allocated 406 * iova range) in addition to removing the iommu mapping. In the eviction 407 * case (!close), we keep the iova allocated, but only remove the iommu 408 * mapping. 409 */ 410 static void 411 put_iova_spaces(struct drm_gem_object *obj, struct drm_gpuvm *vm, 412 bool close, const char *reason) 413 { 414 struct drm_gpuvm_bo *vm_bo, *tmp; 415 416 msm_gem_assert_locked(obj); 417 418 drm_gem_for_each_gpuvm_bo_safe (vm_bo, tmp, obj) { 419 struct drm_gpuva *vma, *vmatmp; 420 421 if (vm && vm_bo->vm != vm) 422 continue; 423 424 drm_gpuvm_bo_get(vm_bo); 425 426 drm_gpuvm_bo_for_each_va_safe (vma, vmatmp, vm_bo) { 427 msm_gem_vma_unmap(vma, reason); 428 if (close) 429 msm_gem_vma_close(vma); 430 } 431 432 drm_gpuvm_bo_put(vm_bo); 433 } 434 } 435 436 static struct drm_gpuva *get_vma_locked(struct drm_gem_object *obj, 437 struct drm_gpuvm *vm, u64 range_start, 438 u64 range_end) 439 { 440 struct drm_gpuva *vma; 441 442 msm_gem_assert_locked(obj); 443 444 vma = lookup_vma(obj, vm); 445 446 if (!vma) { 447 vma = msm_gem_vma_new(vm, obj, 0, range_start, range_end); 448 } else { 449 GEM_WARN_ON(vma->va.addr < range_start); 450 GEM_WARN_ON((vma->va.addr + obj->size) > range_end); 451 } 452 453 return vma; 454 } 455 456 int msm_gem_prot(struct drm_gem_object *obj) 457 { 458 struct msm_gem_object *msm_obj = to_msm_bo(obj); 459 int prot = IOMMU_READ; 460 461 if (!(msm_obj->flags & MSM_BO_GPU_READONLY)) 462 prot |= IOMMU_WRITE; 463 464 if (msm_obj->flags & MSM_BO_MAP_PRIV) 465 prot |= IOMMU_PRIV; 466 467 if (msm_obj->flags & MSM_BO_CACHED_COHERENT) 468 prot |= IOMMU_CACHE; 469 470 return prot; 471 } 472 473 int msm_gem_pin_vma_locked(struct drm_gem_object *obj, struct drm_gpuva *vma) 474 { 475 struct msm_gem_object *msm_obj = to_msm_bo(obj); 476 struct page **pages; 477 int prot = msm_gem_prot(obj); 478 479 msm_gem_assert_locked(obj); 480 481 pages = msm_gem_get_pages_locked(obj, MSM_MADV_WILLNEED); 482 if (IS_ERR(pages)) 483 return PTR_ERR(pages); 484 485 return msm_gem_vma_map(vma, prot, msm_obj->sgt); 486 } 487 488 void msm_gem_unpin_locked(struct drm_gem_object *obj) 489 { 490 struct msm_drm_private *priv = obj->dev->dev_private; 491 struct msm_gem_object *msm_obj = to_msm_bo(obj); 492 493 msm_gem_assert_locked(obj); 494 495 mutex_lock(&priv->lru.lock); 496 msm_obj->pin_count--; 497 GEM_WARN_ON(msm_obj->pin_count < 0); 498 update_lru_locked(obj); 499 mutex_unlock(&priv->lru.lock); 500 } 501 502 /* Special unpin path for use in fence-signaling path, avoiding the need 503 * to hold the obj lock by only depending on things that a protected by 504 * the LRU lock. In particular we know that that we already have backing 505 * and and that the object's dma_resv has the fence for the current 506 * submit/job which will prevent us racing against page eviction. 507 */ 508 void msm_gem_unpin_active(struct drm_gem_object *obj) 509 { 510 struct msm_gem_object *msm_obj = to_msm_bo(obj); 511 512 msm_obj->pin_count--; 513 GEM_WARN_ON(msm_obj->pin_count < 0); 514 update_lru_active(obj); 515 } 516 517 struct drm_gpuva *msm_gem_get_vma_locked(struct drm_gem_object *obj, 518 struct drm_gpuvm *vm) 519 { 520 return get_vma_locked(obj, vm, 0, U64_MAX); 521 } 522 523 static int get_and_pin_iova_range_locked(struct drm_gem_object *obj, 524 struct drm_gpuvm *vm, uint64_t *iova, 525 u64 range_start, u64 range_end) 526 { 527 struct drm_gpuva *vma; 528 int ret; 529 530 msm_gem_assert_locked(obj); 531 532 if (to_msm_bo(obj)->flags & MSM_BO_NO_SHARE) 533 return -EINVAL; 534 535 vma = get_vma_locked(obj, vm, range_start, range_end); 536 if (IS_ERR(vma)) 537 return PTR_ERR(vma); 538 539 ret = msm_gem_pin_vma_locked(obj, vma); 540 if (!ret) { 541 *iova = vma->va.addr; 542 pin_obj_locked(obj); 543 } 544 545 return ret; 546 } 547 548 /* 549 * get iova and pin it. Should have a matching put 550 * limits iova to specified range (in pages) 551 */ 552 int msm_gem_get_and_pin_iova_range(struct drm_gem_object *obj, 553 struct drm_gpuvm *vm, uint64_t *iova, 554 u64 range_start, u64 range_end) 555 { 556 struct drm_exec exec; 557 int ret; 558 559 msm_gem_lock_vm_and_obj(&exec, obj, vm); 560 ret = get_and_pin_iova_range_locked(obj, vm, iova, range_start, range_end); 561 drm_exec_fini(&exec); /* drop locks */ 562 563 return ret; 564 } 565 566 /* get iova and pin it. Should have a matching put */ 567 int msm_gem_get_and_pin_iova(struct drm_gem_object *obj, struct drm_gpuvm *vm, 568 uint64_t *iova) 569 { 570 return msm_gem_get_and_pin_iova_range(obj, vm, iova, 0, U64_MAX); 571 } 572 573 /* 574 * Get an iova but don't pin it. Doesn't need a put because iovas are currently 575 * valid for the life of the object 576 */ 577 int msm_gem_get_iova(struct drm_gem_object *obj, struct drm_gpuvm *vm, 578 uint64_t *iova) 579 { 580 struct drm_gpuva *vma; 581 struct drm_exec exec; 582 int ret = 0; 583 584 msm_gem_lock_vm_and_obj(&exec, obj, vm); 585 vma = get_vma_locked(obj, vm, 0, U64_MAX); 586 if (IS_ERR(vma)) { 587 ret = PTR_ERR(vma); 588 } else { 589 *iova = vma->va.addr; 590 } 591 drm_exec_fini(&exec); /* drop locks */ 592 593 return ret; 594 } 595 596 static int clear_iova(struct drm_gem_object *obj, 597 struct drm_gpuvm *vm) 598 { 599 struct drm_gpuva *vma = lookup_vma(obj, vm); 600 601 if (!vma) 602 return 0; 603 604 msm_gem_vma_unmap(vma, NULL); 605 msm_gem_vma_close(vma); 606 607 return 0; 608 } 609 610 /* 611 * Get the requested iova but don't pin it. Fails if the requested iova is 612 * not available. Doesn't need a put because iovas are currently valid for 613 * the life of the object. 614 * 615 * Setting an iova of zero will clear the vma. 616 */ 617 int msm_gem_set_iova(struct drm_gem_object *obj, 618 struct drm_gpuvm *vm, uint64_t iova) 619 { 620 struct drm_exec exec; 621 int ret = 0; 622 623 msm_gem_lock_vm_and_obj(&exec, obj, vm); 624 if (!iova) { 625 ret = clear_iova(obj, vm); 626 } else { 627 struct drm_gpuva *vma; 628 vma = get_vma_locked(obj, vm, iova, iova + obj->size); 629 if (IS_ERR(vma)) { 630 ret = PTR_ERR(vma); 631 } else if (GEM_WARN_ON(vma->va.addr != iova)) { 632 clear_iova(obj, vm); 633 ret = -EBUSY; 634 } 635 } 636 drm_exec_fini(&exec); /* drop locks */ 637 638 return ret; 639 } 640 641 static bool is_kms_vm(struct drm_gpuvm *vm) 642 { 643 #ifdef CONFIG_DRM_MSM_KMS 644 struct msm_drm_private *priv = vm->drm->dev_private; 645 646 return priv->kms && (priv->kms->vm == vm); 647 #else 648 return false; 649 #endif 650 } 651 652 /* 653 * Unpin a iova by updating the reference counts. The memory isn't actually 654 * purged until something else (shrinker, mm_notifier, destroy, etc) decides 655 * to get rid of it 656 */ 657 void msm_gem_unpin_iova(struct drm_gem_object *obj, struct drm_gpuvm *vm) 658 { 659 struct drm_gpuva *vma; 660 struct drm_exec exec; 661 662 msm_gem_lock_vm_and_obj(&exec, obj, vm); 663 vma = lookup_vma(obj, vm); 664 if (vma) { 665 msm_gem_unpin_locked(obj); 666 } 667 if (!is_kms_vm(vm)) 668 put_iova_spaces(obj, vm, true, "close"); 669 drm_exec_fini(&exec); /* drop locks */ 670 } 671 672 int msm_gem_dumb_create(struct drm_file *file, struct drm_device *dev, 673 struct drm_mode_create_dumb *args) 674 { 675 u32 fourcc; 676 u64 pitch_align; 677 int ret; 678 679 /* 680 * Adreno needs pitch aligned to 32 pixels. Compute the number 681 * of bytes for a block of 32 pixels at the given color format. 682 * Use the result as pitch alignment. 683 */ 684 fourcc = drm_driver_color_mode_format(dev, args->bpp); 685 if (fourcc != DRM_FORMAT_INVALID) { 686 const struct drm_format_info *info; 687 688 info = drm_format_info(fourcc); 689 if (!info) 690 return -EINVAL; 691 pitch_align = drm_format_info_min_pitch(info, 0, 32); 692 } else { 693 pitch_align = round_up(args->width, 32) * DIV_ROUND_UP(args->bpp, SZ_8); 694 } 695 if (!pitch_align || pitch_align > U32_MAX) 696 return -EINVAL; 697 ret = drm_mode_size_dumb(dev, args, pitch_align, 0); 698 if (ret) 699 return ret; 700 701 return msm_gem_new_handle(dev, file, args->size, 702 MSM_BO_SCANOUT | MSM_BO_WC, &args->handle, "dumb"); 703 } 704 705 static void *get_vaddr(struct drm_gem_object *obj, unsigned madv) 706 { 707 struct msm_gem_object *msm_obj = to_msm_bo(obj); 708 struct page **pages; 709 int ret = 0; 710 711 msm_gem_assert_locked(obj); 712 713 if (drm_gem_is_imported(obj)) 714 return ERR_PTR(-ENODEV); 715 716 pages = msm_gem_get_pages_locked(obj, madv); 717 if (IS_ERR(pages)) 718 return ERR_CAST(pages); 719 720 pin_obj_locked(obj); 721 722 /* increment vmap_count *before* vmap() call, so shrinker can 723 * check vmap_count (is_vunmapable()) outside of msm_obj lock. 724 * This guarantees that we won't try to msm_gem_vunmap() this 725 * same object from within the vmap() call (while we already 726 * hold msm_obj lock) 727 */ 728 msm_obj->vmap_count++; 729 730 if (!msm_obj->vaddr) { 731 msm_obj->vaddr = vmap(pages, obj->size >> PAGE_SHIFT, 732 VM_MAP, msm_gem_pgprot(msm_obj, PAGE_KERNEL)); 733 if (msm_obj->vaddr == NULL) { 734 ret = -ENOMEM; 735 goto fail; 736 } 737 } 738 739 return msm_obj->vaddr; 740 741 fail: 742 msm_obj->vmap_count--; 743 msm_gem_unpin_locked(obj); 744 return ERR_PTR(ret); 745 } 746 747 void *msm_gem_get_vaddr_locked(struct drm_gem_object *obj) 748 { 749 return get_vaddr(obj, MSM_MADV_WILLNEED); 750 } 751 752 void *msm_gem_get_vaddr(struct drm_gem_object *obj) 753 { 754 void *ret; 755 756 msm_gem_lock(obj); 757 ret = msm_gem_get_vaddr_locked(obj); 758 msm_gem_unlock(obj); 759 760 return ret; 761 } 762 763 /* 764 * Don't use this! It is for the very special case of dumping 765 * submits from GPU hangs or faults, were the bo may already 766 * be MSM_MADV_DONTNEED, but we know the buffer is still on the 767 * active list. 768 */ 769 void *msm_gem_get_vaddr_active(struct drm_gem_object *obj) 770 { 771 return get_vaddr(obj, __MSM_MADV_PURGED); 772 } 773 774 void msm_gem_put_vaddr_locked(struct drm_gem_object *obj) 775 { 776 struct msm_gem_object *msm_obj = to_msm_bo(obj); 777 778 msm_gem_assert_locked(obj); 779 GEM_WARN_ON(msm_obj->vmap_count < 1); 780 781 msm_obj->vmap_count--; 782 msm_gem_unpin_locked(obj); 783 } 784 785 void msm_gem_put_vaddr(struct drm_gem_object *obj) 786 { 787 msm_gem_lock(obj); 788 msm_gem_put_vaddr_locked(obj); 789 msm_gem_unlock(obj); 790 } 791 792 /* Update madvise status, returns true if not purged, else 793 * false or -errno. 794 */ 795 int msm_gem_madvise(struct drm_gem_object *obj, unsigned madv) 796 { 797 struct msm_drm_private *priv = obj->dev->dev_private; 798 struct msm_gem_object *msm_obj = to_msm_bo(obj); 799 800 msm_gem_lock(obj); 801 802 mutex_lock(&priv->lru.lock); 803 804 if (msm_obj->madv != __MSM_MADV_PURGED) 805 msm_obj->madv = madv; 806 807 madv = msm_obj->madv; 808 809 /* If the obj is inactive, we might need to move it 810 * between inactive lists 811 */ 812 update_lru_locked(obj); 813 814 mutex_unlock(&priv->lru.lock); 815 816 msm_gem_unlock(obj); 817 818 return (madv != __MSM_MADV_PURGED); 819 } 820 821 void msm_gem_purge(struct drm_gem_object *obj) 822 { 823 struct drm_device *dev = obj->dev; 824 struct msm_drm_private *priv = obj->dev->dev_private; 825 struct msm_gem_object *msm_obj = to_msm_bo(obj); 826 827 msm_gem_assert_locked(obj); 828 GEM_WARN_ON(!is_purgeable(msm_obj)); 829 830 /* Get rid of any iommu mapping(s): */ 831 put_iova_spaces(obj, NULL, false, "purge"); 832 833 msm_gem_vunmap(obj); 834 835 drm_vma_node_unmap(&obj->vma_node, dev->anon_inode->i_mapping); 836 837 put_pages(obj); 838 839 mutex_lock(&priv->lru.lock); 840 /* A one-way transition: */ 841 msm_obj->madv = __MSM_MADV_PURGED; 842 mutex_unlock(&priv->lru.lock); 843 844 drm_gem_free_mmap_offset(obj); 845 846 /* Our goal here is to return as much of the memory as 847 * is possible back to the system as we are called from OOM. 848 * To do this we must instruct the shmfs to drop all of its 849 * backing pages, *now*. 850 */ 851 shmem_truncate_range(file_inode(obj->filp), 0, (loff_t)-1); 852 853 invalidate_mapping_pages(file_inode(obj->filp)->i_mapping, 854 0, (loff_t)-1); 855 } 856 857 /* 858 * Unpin the backing pages and make them available to be swapped out. 859 */ 860 void msm_gem_evict(struct drm_gem_object *obj) 861 { 862 struct drm_device *dev = obj->dev; 863 struct msm_gem_object *msm_obj = to_msm_bo(obj); 864 865 msm_gem_assert_locked(obj); 866 GEM_WARN_ON(is_unevictable(msm_obj)); 867 868 /* Get rid of any iommu mapping(s): */ 869 put_iova_spaces(obj, NULL, false, "evict"); 870 871 drm_vma_node_unmap(&obj->vma_node, dev->anon_inode->i_mapping); 872 873 put_pages(obj); 874 } 875 876 void msm_gem_vunmap(struct drm_gem_object *obj) 877 { 878 struct msm_gem_object *msm_obj = to_msm_bo(obj); 879 880 msm_gem_assert_locked(obj); 881 882 if (!msm_obj->vaddr || GEM_WARN_ON(!is_vunmapable(msm_obj))) 883 return; 884 885 vunmap(msm_obj->vaddr); 886 msm_obj->vaddr = NULL; 887 } 888 889 bool msm_gem_active(struct drm_gem_object *obj) 890 { 891 msm_gem_assert_locked(obj); 892 893 if (to_msm_bo(obj)->pin_count) 894 return true; 895 896 return !dma_resv_test_signaled(obj->resv, DMA_RESV_USAGE_BOOKKEEP); 897 } 898 899 int msm_gem_cpu_prep(struct drm_gem_object *obj, uint32_t op, ktime_t *timeout) 900 { 901 bool write = !!(op & MSM_PREP_WRITE); 902 unsigned long remain = 903 op & MSM_PREP_NOSYNC ? 0 : timeout_to_jiffies(timeout); 904 long ret; 905 906 if (op & MSM_PREP_BOOST) { 907 dma_resv_set_deadline(obj->resv, dma_resv_usage_rw(write), 908 ktime_get()); 909 } 910 911 ret = dma_resv_wait_timeout(obj->resv, dma_resv_usage_rw(write), 912 true, remain); 913 if (ret == 0) 914 return remain == 0 ? -EBUSY : -ETIMEDOUT; 915 else if (ret < 0) 916 return ret; 917 918 /* TODO cache maintenance */ 919 920 return 0; 921 } 922 923 int msm_gem_cpu_fini(struct drm_gem_object *obj) 924 { 925 /* TODO cache maintenance */ 926 return 0; 927 } 928 929 #ifdef CONFIG_DEBUG_FS 930 void msm_gem_describe(struct drm_gem_object *obj, struct seq_file *m, 931 struct msm_gem_stats *stats) 932 { 933 struct msm_gem_object *msm_obj = to_msm_bo(obj); 934 struct dma_resv *robj = obj->resv; 935 uint64_t off = drm_vma_node_start(&obj->vma_node); 936 const char *madv; 937 938 if (!msm_gem_trylock(obj)) 939 return; 940 941 stats->all.count++; 942 stats->all.size += obj->size; 943 944 if (msm_gem_active(obj)) { 945 stats->active.count++; 946 stats->active.size += obj->size; 947 } 948 949 if (msm_obj->pages) { 950 stats->resident.count++; 951 stats->resident.size += obj->size; 952 } 953 954 switch (msm_obj->madv) { 955 case __MSM_MADV_PURGED: 956 stats->purged.count++; 957 stats->purged.size += obj->size; 958 madv = " purged"; 959 break; 960 case MSM_MADV_DONTNEED: 961 stats->purgeable.count++; 962 stats->purgeable.size += obj->size; 963 madv = " purgeable"; 964 break; 965 case MSM_MADV_WILLNEED: 966 default: 967 madv = ""; 968 break; 969 } 970 971 seq_printf(m, "%08x: %c %2d (%2d) %08llx %p", 972 msm_obj->flags, msm_gem_active(obj) ? 'A' : 'I', 973 obj->name, kref_read(&obj->refcount), 974 off, msm_obj->vaddr); 975 976 seq_printf(m, " %08zu %9s %-32s\n", obj->size, madv, msm_obj->name); 977 978 if (!list_empty(&obj->gpuva.list)) { 979 struct drm_gpuvm_bo *vm_bo; 980 981 seq_puts(m, " vmas:"); 982 983 drm_gem_for_each_gpuvm_bo (vm_bo, obj) { 984 struct drm_gpuva *vma; 985 986 drm_gpuvm_bo_for_each_va (vma, vm_bo) { 987 const char *name, *comm; 988 struct msm_gem_vm *vm = to_msm_vm(vma->vm); 989 struct task_struct *task = 990 get_pid_task(vm->pid, PIDTYPE_PID); 991 if (task) { 992 comm = kstrdup(task->comm, GFP_KERNEL); 993 put_task_struct(task); 994 } else { 995 comm = NULL; 996 } 997 name = vm->base.name; 998 999 seq_printf(m, " [%s%s%s: vm=%p, %08llx, %smapped]", 1000 name, comm ? ":" : "", comm ? comm : "", 1001 vma->vm, vma->va.addr, 1002 to_msm_vma(vma)->mapped ? "" : "un"); 1003 kfree(comm); 1004 } 1005 } 1006 1007 seq_puts(m, "\n"); 1008 } 1009 1010 dma_resv_describe(robj, m); 1011 msm_gem_unlock(obj); 1012 } 1013 1014 void msm_gem_describe_objects(struct list_head *list, struct seq_file *m) 1015 { 1016 struct msm_gem_stats stats = {}; 1017 struct msm_gem_object *msm_obj; 1018 1019 seq_puts(m, " flags id ref offset kaddr size madv name\n"); 1020 list_for_each_entry(msm_obj, list, node) { 1021 struct drm_gem_object *obj = &msm_obj->base; 1022 seq_puts(m, " "); 1023 msm_gem_describe(obj, m, &stats); 1024 } 1025 1026 seq_printf(m, "Total: %4d objects, %9zu bytes\n", 1027 stats.all.count, stats.all.size); 1028 seq_printf(m, "Active: %4d objects, %9zu bytes\n", 1029 stats.active.count, stats.active.size); 1030 seq_printf(m, "Resident: %4d objects, %9zu bytes\n", 1031 stats.resident.count, stats.resident.size); 1032 seq_printf(m, "Purgeable: %4d objects, %9zu bytes\n", 1033 stats.purgeable.count, stats.purgeable.size); 1034 seq_printf(m, "Purged: %4d objects, %9zu bytes\n", 1035 stats.purged.count, stats.purged.size); 1036 } 1037 #endif 1038 1039 /* don't call directly! Use drm_gem_object_put() */ 1040 static void msm_gem_free_object(struct drm_gem_object *obj) 1041 { 1042 struct msm_gem_object *msm_obj = to_msm_bo(obj); 1043 struct drm_device *dev = obj->dev; 1044 struct msm_drm_private *priv = dev->dev_private; 1045 struct drm_exec exec; 1046 1047 mutex_lock(&priv->obj_lock); 1048 list_del(&msm_obj->node); 1049 mutex_unlock(&priv->obj_lock); 1050 1051 /* 1052 * We need to lock any VMs the object is still attached to, but not 1053 * the object itself (see explaination in msm_gem_assert_locked()), 1054 * so just open-code this special case. 1055 * 1056 * Note that we skip the dance if we aren't attached to any VM. This 1057 * is load bearing. The driver needs to support two usage models: 1058 * 1059 * 1. Legacy kernel managed VM: Userspace expects the VMA's to be 1060 * implicitly torn down when the object is freed, the VMA's do 1061 * not hold a hard reference to the BO. 1062 * 1063 * 2. VM_BIND, userspace managed VM: The VMA holds a reference to the 1064 * BO. This can be dropped when the VM is closed and it's associated 1065 * VMAs are torn down. (See msm_gem_vm_close()). 1066 * 1067 * In the latter case the last reference to a BO can be dropped while 1068 * we already have the VM locked. It would have already been removed 1069 * from the gpuva list, but lockdep doesn't know that. Or understand 1070 * the differences between the two usage models. 1071 */ 1072 if (!list_empty(&obj->gpuva.list)) { 1073 drm_exec_init(&exec, 0, 0); 1074 drm_exec_until_all_locked (&exec) { 1075 struct drm_gpuvm_bo *vm_bo; 1076 drm_gem_for_each_gpuvm_bo (vm_bo, obj) { 1077 drm_exec_lock_obj(&exec, 1078 drm_gpuvm_resv_obj(vm_bo->vm)); 1079 drm_exec_retry_on_contention(&exec); 1080 } 1081 } 1082 put_iova_spaces(obj, NULL, true, "free"); 1083 drm_exec_fini(&exec); /* drop locks */ 1084 } 1085 1086 if (drm_gem_is_imported(obj)) { 1087 GEM_WARN_ON(msm_obj->vaddr); 1088 1089 /* Don't drop the pages for imported dmabuf, as they are not 1090 * ours, just free the array we allocated: 1091 */ 1092 kvfree(msm_obj->pages); 1093 1094 drm_prime_gem_destroy(obj, msm_obj->sgt); 1095 } else { 1096 msm_gem_vunmap(obj); 1097 put_pages(obj); 1098 } 1099 1100 /* 1101 * In error paths, we could end up here before msm_gem_new_handle() 1102 * has changed obj->resv to point to the shared resv. In this case, 1103 * we don't want to drop a ref to the shared r_obj that we haven't 1104 * taken yet. 1105 */ 1106 if ((msm_obj->flags & MSM_BO_NO_SHARE) && (obj->resv != &obj->_resv)) { 1107 struct drm_gem_object *r_obj = 1108 container_of(obj->resv, struct drm_gem_object, _resv); 1109 1110 /* Drop reference we hold to shared resv obj: */ 1111 drm_gem_object_put(r_obj); 1112 } 1113 1114 drm_gem_object_release(obj); 1115 1116 kfree(msm_obj->metadata); 1117 kfree(msm_obj); 1118 } 1119 1120 static int msm_gem_object_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma) 1121 { 1122 struct msm_gem_object *msm_obj = to_msm_bo(obj); 1123 1124 vm_flags_set(vma, VM_PFNMAP | VM_DONTEXPAND | VM_DONTDUMP); 1125 vma->vm_page_prot = msm_gem_pgprot(msm_obj, vm_get_page_prot(vma->vm_flags)); 1126 1127 return 0; 1128 } 1129 1130 /* convenience method to construct a GEM buffer object, and userspace handle */ 1131 int msm_gem_new_handle(struct drm_device *dev, struct drm_file *file, 1132 size_t size, uint32_t flags, uint32_t *handle, 1133 char *name) 1134 { 1135 struct drm_gem_object *obj; 1136 int ret; 1137 1138 obj = msm_gem_new(dev, size, flags); 1139 1140 if (IS_ERR(obj)) 1141 return PTR_ERR(obj); 1142 1143 if (name) 1144 msm_gem_object_set_name(obj, "%s", name); 1145 1146 if (flags & MSM_BO_NO_SHARE) { 1147 struct msm_context *ctx = file->driver_priv; 1148 struct drm_gem_object *r_obj = drm_gpuvm_resv_obj(ctx->vm); 1149 1150 drm_gem_object_get(r_obj); 1151 1152 obj->resv = r_obj->resv; 1153 } 1154 1155 ret = drm_gem_handle_create(file, obj, handle); 1156 1157 /* drop reference from allocate - handle holds it now */ 1158 drm_gem_object_put(obj); 1159 1160 return ret; 1161 } 1162 1163 static enum drm_gem_object_status msm_gem_status(struct drm_gem_object *obj) 1164 { 1165 struct msm_gem_object *msm_obj = to_msm_bo(obj); 1166 enum drm_gem_object_status status = 0; 1167 1168 if (msm_obj->pages) 1169 status |= DRM_GEM_OBJECT_RESIDENT; 1170 1171 if (msm_obj->madv == MSM_MADV_DONTNEED) 1172 status |= DRM_GEM_OBJECT_PURGEABLE; 1173 1174 return status; 1175 } 1176 1177 static const struct vm_operations_struct vm_ops = { 1178 .fault = msm_gem_fault, 1179 .open = drm_gem_vm_open, 1180 .close = drm_gem_vm_close, 1181 }; 1182 1183 static const struct drm_gem_object_funcs msm_gem_object_funcs = { 1184 .free = msm_gem_free_object, 1185 .open = msm_gem_open, 1186 .close = msm_gem_close, 1187 .export = msm_gem_prime_export, 1188 .pin = msm_gem_prime_pin, 1189 .unpin = msm_gem_prime_unpin, 1190 .get_sg_table = msm_gem_prime_get_sg_table, 1191 .vmap = msm_gem_prime_vmap, 1192 .vunmap = msm_gem_prime_vunmap, 1193 .mmap = msm_gem_object_mmap, 1194 .status = msm_gem_status, 1195 .vm_ops = &vm_ops, 1196 }; 1197 1198 static int msm_gem_new_impl(struct drm_device *dev, uint32_t flags, 1199 struct drm_gem_object **obj) 1200 { 1201 struct msm_drm_private *priv = dev->dev_private; 1202 struct msm_gem_object *msm_obj; 1203 1204 switch (flags & MSM_BO_CACHE_MASK) { 1205 case MSM_BO_CACHED: 1206 case MSM_BO_WC: 1207 break; 1208 case MSM_BO_CACHED_COHERENT: 1209 if (priv->has_cached_coherent) 1210 break; 1211 fallthrough; 1212 default: 1213 DRM_DEV_DEBUG(dev->dev, "invalid cache flag: %x\n", 1214 (flags & MSM_BO_CACHE_MASK)); 1215 return -EINVAL; 1216 } 1217 1218 msm_obj = kzalloc(sizeof(*msm_obj), GFP_KERNEL); 1219 if (!msm_obj) 1220 return -ENOMEM; 1221 1222 msm_obj->flags = flags; 1223 msm_obj->madv = MSM_MADV_WILLNEED; 1224 1225 INIT_LIST_HEAD(&msm_obj->node); 1226 1227 *obj = &msm_obj->base; 1228 (*obj)->funcs = &msm_gem_object_funcs; 1229 1230 return 0; 1231 } 1232 1233 struct drm_gem_object *msm_gem_new(struct drm_device *dev, size_t size, uint32_t flags) 1234 { 1235 struct msm_drm_private *priv = dev->dev_private; 1236 struct msm_gem_object *msm_obj; 1237 struct drm_gem_object *obj = NULL; 1238 int ret; 1239 1240 size = PAGE_ALIGN(size); 1241 1242 /* Disallow zero sized objects as they make the underlying 1243 * infrastructure grumpy 1244 */ 1245 if (size == 0) 1246 return ERR_PTR(-EINVAL); 1247 1248 ret = msm_gem_new_impl(dev, flags, &obj); 1249 if (ret) 1250 return ERR_PTR(ret); 1251 1252 msm_obj = to_msm_bo(obj); 1253 1254 ret = drm_gem_object_init(dev, obj, size); 1255 if (ret) 1256 goto fail; 1257 /* 1258 * Our buffers are kept pinned, so allocating them from the 1259 * MOVABLE zone is a really bad idea, and conflicts with CMA. 1260 * See comments above new_inode() why this is required _and_ 1261 * expected if you're going to pin these pages. 1262 */ 1263 mapping_set_gfp_mask(obj->filp->f_mapping, GFP_HIGHUSER); 1264 1265 drm_gem_lru_move_tail(&priv->lru.unbacked, obj); 1266 1267 mutex_lock(&priv->obj_lock); 1268 list_add_tail(&msm_obj->node, &priv->objects); 1269 mutex_unlock(&priv->obj_lock); 1270 1271 ret = drm_gem_create_mmap_offset(obj); 1272 if (ret) 1273 goto fail; 1274 1275 return obj; 1276 1277 fail: 1278 drm_gem_object_put(obj); 1279 return ERR_PTR(ret); 1280 } 1281 1282 struct drm_gem_object *msm_gem_import(struct drm_device *dev, 1283 struct dma_buf *dmabuf, struct sg_table *sgt) 1284 { 1285 struct msm_drm_private *priv = dev->dev_private; 1286 struct msm_gem_object *msm_obj; 1287 struct drm_gem_object *obj; 1288 size_t size, npages; 1289 int ret; 1290 1291 size = PAGE_ALIGN(dmabuf->size); 1292 1293 ret = msm_gem_new_impl(dev, MSM_BO_WC, &obj); 1294 if (ret) 1295 return ERR_PTR(ret); 1296 1297 drm_gem_private_object_init(dev, obj, size); 1298 1299 npages = size / PAGE_SIZE; 1300 1301 msm_obj = to_msm_bo(obj); 1302 msm_gem_lock(obj); 1303 msm_obj->sgt = sgt; 1304 msm_obj->pages = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL); 1305 if (!msm_obj->pages) { 1306 msm_gem_unlock(obj); 1307 ret = -ENOMEM; 1308 goto fail; 1309 } 1310 1311 ret = drm_prime_sg_to_page_array(sgt, msm_obj->pages, npages); 1312 if (ret) { 1313 msm_gem_unlock(obj); 1314 goto fail; 1315 } 1316 1317 msm_gem_unlock(obj); 1318 1319 drm_gem_lru_move_tail(&priv->lru.pinned, obj); 1320 1321 mutex_lock(&priv->obj_lock); 1322 list_add_tail(&msm_obj->node, &priv->objects); 1323 mutex_unlock(&priv->obj_lock); 1324 1325 ret = drm_gem_create_mmap_offset(obj); 1326 if (ret) 1327 goto fail; 1328 1329 return obj; 1330 1331 fail: 1332 drm_gem_object_put(obj); 1333 return ERR_PTR(ret); 1334 } 1335 1336 void *msm_gem_kernel_new(struct drm_device *dev, size_t size, uint32_t flags, 1337 struct drm_gpuvm *vm, struct drm_gem_object **bo, 1338 uint64_t *iova) 1339 { 1340 void *vaddr; 1341 struct drm_gem_object *obj = msm_gem_new(dev, size, flags); 1342 int ret; 1343 1344 if (IS_ERR(obj)) 1345 return ERR_CAST(obj); 1346 1347 if (iova) { 1348 ret = msm_gem_get_and_pin_iova(obj, vm, iova); 1349 if (ret) 1350 goto err; 1351 } 1352 1353 vaddr = msm_gem_get_vaddr(obj); 1354 if (IS_ERR(vaddr)) { 1355 msm_gem_unpin_iova(obj, vm); 1356 ret = PTR_ERR(vaddr); 1357 goto err; 1358 } 1359 1360 if (bo) 1361 *bo = obj; 1362 1363 return vaddr; 1364 err: 1365 drm_gem_object_put(obj); 1366 1367 return ERR_PTR(ret); 1368 1369 } 1370 1371 void msm_gem_kernel_put(struct drm_gem_object *bo, struct drm_gpuvm *vm) 1372 { 1373 if (IS_ERR_OR_NULL(bo)) 1374 return; 1375 1376 msm_gem_put_vaddr(bo); 1377 msm_gem_unpin_iova(bo, vm); 1378 drm_gem_object_put(bo); 1379 } 1380 1381 void msm_gem_object_set_name(struct drm_gem_object *bo, const char *fmt, ...) 1382 { 1383 struct msm_gem_object *msm_obj = to_msm_bo(bo); 1384 va_list ap; 1385 1386 if (!fmt) 1387 return; 1388 1389 va_start(ap, fmt); 1390 vsnprintf(msm_obj->name, sizeof(msm_obj->name), fmt, ap); 1391 va_end(ap); 1392 } 1393