Lines Matching +full:fifo +full:- +full:width

1 // SPDX-License-Identifier: GPL-2.0 OR MIT
4 * Copyright 2009-2023 VMware, Inc., Palo Alto, CA., USA
20 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
55 #include <linux/dma-mapping.h>
288 MODULE_PARM_DESC(assume_16bpp, "Assume 16-bpp when filtering modes");
305 { SVGA_CAP_EXTENDED_FIFO, "extended fifo" },
354 ARRAY_SIZE(buf) - offset, in vmw_print_bitmap()
377 drm_info(&dev_priv->drm, "Available shader model: %s.\n", in vmw_print_sm_type()
378 names[dev_priv->sm_type]); in vmw_print_sm_type()
382 * vmw_dummy_query_bo_create - create a bo to hold a dummy query result
419 ret = ttm_bo_kmap(&vbo->tbo, 0, 1, &map); in vmw_dummy_query_bo_create()
422 result->totalSize = sizeof(*result); in vmw_dummy_query_bo_create()
423 result->state = SVGA3D_QUERYSTATE_PENDING; in vmw_dummy_query_bo_create()
424 result->result32 = 0xff; in vmw_dummy_query_bo_create()
428 ttm_bo_unreserve(&vbo->tbo); in vmw_dummy_query_bo_create()
434 dev_priv->dummy_query_bo = vbo; in vmw_dummy_query_bo_create()
443 dev_priv->enable_state = vmw_read(dev_priv, SVGA_REG_ENABLE); in vmw_device_init()
444 dev_priv->config_done_state = vmw_read(dev_priv, SVGA_REG_CONFIG_DONE); in vmw_device_init()
445 dev_priv->traces_state = vmw_read(dev_priv, SVGA_REG_TRACES); in vmw_device_init()
451 (dev_priv->capabilities & SVGA_CAP_TRACES) != 0; in vmw_device_init()
454 dev_priv->fifo = vmw_fifo_create(dev_priv); in vmw_device_init()
455 if (IS_ERR(dev_priv->fifo)) { in vmw_device_init()
456 int err = PTR_ERR(dev_priv->fifo); in vmw_device_init()
457 dev_priv->fifo = NULL; in vmw_device_init()
459 } else if (!dev_priv->fifo) { in vmw_device_init()
463 dev_priv->last_read_seqno = vmw_fence_read(dev_priv); in vmw_device_init()
464 atomic_set(&dev_priv->marker_seq, dev_priv->last_read_seqno); in vmw_device_init()
477 vmw->last_read_seqno = vmw_fence_read(vmw); in vmw_device_fini()
480 vmw->config_done_state); in vmw_device_fini()
482 vmw->enable_state); in vmw_device_fini()
484 vmw->traces_state); in vmw_device_fini()
490 * vmw_request_device_late - Perform late device setup
503 if (dev_priv->has_mob) { in vmw_request_device_late()
512 if (dev_priv->cman) { in vmw_request_device_late()
513 ret = vmw_cmdbuf_set_pool_size(dev_priv->cman, 256*4096); in vmw_request_device_late()
515 struct vmw_cmdbuf_man *man = dev_priv->cman; in vmw_request_device_late()
517 dev_priv->cman = NULL; in vmw_request_device_late()
534 vmw_fence_fifo_up(dev_priv->fman); in vmw_request_device()
535 dev_priv->cman = vmw_cmdbuf_man_create(dev_priv); in vmw_request_device()
536 if (IS_ERR(dev_priv->cman)) { in vmw_request_device()
537 dev_priv->cman = NULL; in vmw_request_device()
538 dev_priv->sm_type = VMW_SM_LEGACY; in vmw_request_device()
552 if (dev_priv->cman) in vmw_request_device()
553 vmw_cmdbuf_remove_pool(dev_priv->cman); in vmw_request_device()
554 if (dev_priv->has_mob) { in vmw_request_device()
557 man = ttm_manager_type(&dev_priv->bdev, VMW_PL_MOB); in vmw_request_device()
558 ttm_resource_manager_evict_all(&dev_priv->bdev, man); in vmw_request_device()
561 if (dev_priv->cman) in vmw_request_device()
562 vmw_cmdbuf_man_destroy(dev_priv->cman); in vmw_request_device()
564 vmw_fence_fifo_down(dev_priv->fman); in vmw_request_device()
570 * vmw_release_device_early - Early part of fifo takedown.
584 BUG_ON(dev_priv->pinned_bo != NULL); in vmw_release_device_early()
586 vmw_bo_unreference(&dev_priv->dummy_query_bo); in vmw_release_device_early()
587 if (dev_priv->cman) in vmw_release_device_early()
588 vmw_cmdbuf_remove_pool(dev_priv->cman); in vmw_release_device_early()
590 if (dev_priv->has_mob) { in vmw_release_device_early()
593 man = ttm_manager_type(&dev_priv->bdev, VMW_PL_MOB); in vmw_release_device_early()
594 ttm_resource_manager_evict_all(&dev_priv->bdev, man); in vmw_release_device_early()
600 * vmw_release_device_late - Late part of fifo takedown.
609 vmw_fence_fifo_down(dev_priv->fman); in vmw_release_device_late()
610 if (dev_priv->cman) in vmw_release_device_late()
611 vmw_cmdbuf_man_destroy(dev_priv->cman); in vmw_release_device_late()
617 * Sets the initial_[width|height] fields on the given vmw_private.
619 * It does so by reading SVGA_REG_[WIDTH|HEIGHT] regs and then
620 * clamping the value to fb_max_[width|height] fields and the
621 * VMW_MIN_INITIAL_[WIDTH|HEIGHT].
623 * VMW_MIN_INITIAL_[WIDTH|HEIGHT].
627 uint32_t width; in vmw_get_initial_size() local
630 width = vmw_read(dev_priv, SVGA_REG_WIDTH); in vmw_get_initial_size()
633 width = max_t(uint32_t, width, VMWGFX_MIN_INITIAL_WIDTH); in vmw_get_initial_size()
636 if (width > dev_priv->fb_max_width || in vmw_get_initial_size()
637 height > dev_priv->fb_max_height) { in vmw_get_initial_size()
643 width = VMWGFX_MIN_INITIAL_WIDTH; in vmw_get_initial_size()
647 dev_priv->initial_width = width; in vmw_get_initial_size()
648 dev_priv->initial_height = height; in vmw_get_initial_size()
652 * vmw_dma_select_mode - Determine how DMA mappings should be set up for this
659 * If this function decides that DMA is not possible, it returns -EINVAL.
676 dev_priv->map_mode = vmw_dma_alloc_coherent; in vmw_dma_select_mode()
678 dev_priv->map_mode = vmw_dma_map_bind; in vmw_dma_select_mode()
680 dev_priv->map_mode = vmw_dma_map_populate; in vmw_dma_select_mode()
682 drm_info(&dev_priv->drm, in vmw_dma_select_mode()
683 "DMA map mode: %s\n", names[dev_priv->map_mode]); in vmw_dma_select_mode()
688 * vmw_dma_masks - set required page- and dma masks
690 * @dev_priv: Pointer to struct drm-device
692 * With 32-bit we can only handle 32 bit PFNs. Optionally set that
693 * restriction also for 64-bit systems.
697 struct drm_device *dev = &dev_priv->drm; in vmw_dma_masks()
700 ret = dma_set_mask_and_coherent(dev->dev, DMA_BIT_MASK(64)); in vmw_dma_masks()
702 drm_info(&dev_priv->drm, in vmw_dma_masks()
704 return dma_set_mask_and_coherent(dev->dev, DMA_BIT_MASK(44)); in vmw_dma_masks()
713 ret = ttm_range_man_init(&dev_priv->bdev, TTM_PL_VRAM, false, in vmw_vram_manager_init()
714 dev_priv->vram_size >> PAGE_SHIFT); in vmw_vram_manager_init()
715 ttm_resource_manager_set_used(ttm_manager_type(&dev_priv->bdev, TTM_PL_VRAM), false); in vmw_vram_manager_init()
721 ttm_range_man_fini(&dev_priv->bdev, TTM_PL_VRAM); in vmw_vram_manager_fini()
732 struct pci_dev *pdev = to_pci_dev(dev->drm.dev); in vmw_setup_pci_resources()
740 dev->pci_id = pci_id; in vmw_setup_pci_resources()
744 dev->vram_start = pci_resource_start(pdev, 2); in vmw_setup_pci_resources()
745 dev->vram_size = pci_resource_len(pdev, 2); in vmw_setup_pci_resources()
747 drm_info(&dev->drm, in vmw_setup_pci_resources()
750 dev->rmmio = devm_ioremap(dev->drm.dev, in vmw_setup_pci_resources()
753 if (!dev->rmmio) { in vmw_setup_pci_resources()
754 drm_err(&dev->drm, in vmw_setup_pci_resources()
757 return -ENOMEM; in vmw_setup_pci_resources()
760 dev->io_start = pci_resource_start(pdev, 0); in vmw_setup_pci_resources()
761 dev->vram_start = pci_resource_start(pdev, 1); in vmw_setup_pci_resources()
762 dev->vram_size = pci_resource_len(pdev, 1); in vmw_setup_pci_resources()
766 drm_info(&dev->drm, in vmw_setup_pci_resources()
767 "FIFO at %pa size is %llu KiB\n", in vmw_setup_pci_resources()
769 dev->fifo_mem = devm_memremap(dev->drm.dev, in vmw_setup_pci_resources()
774 if (IS_ERR(dev->fifo_mem)) { in vmw_setup_pci_resources()
775 drm_err(&dev->drm, in vmw_setup_pci_resources()
776 "Failed mapping FIFO memory.\n"); in vmw_setup_pci_resources()
778 return PTR_ERR(dev->fifo_mem); in vmw_setup_pci_resources()
782 return -EINVAL; in vmw_setup_pci_resources()
791 drm_info(&dev->drm, in vmw_setup_pci_resources()
793 &dev->vram_start, (uint64_t)dev->vram_size / 1024); in vmw_setup_pci_resources()
806 drm_err(&dev->drm, in vmw_detect_version()
808 svga_id, dev->pci_id); in vmw_detect_version()
809 return -ENOSYS; in vmw_detect_version()
812 drm_info(&dev->drm, in vmw_detect_version()
819 if ((dev->capabilities2 & SVGA_CAP2_DX2) != 0) { in vmw_write_driver_id()
840 struct vmw_sw_context *sw_context = &dev_priv->ctx; in vmw_sw_context_init()
842 hash_init(sw_context->res_ht); in vmw_sw_context_init()
847 struct vmw_sw_context *sw_context = &dev_priv->ctx; in vmw_sw_context_fini()
849 vfree(sw_context->cmd_bounce); in vmw_sw_context_fini()
850 if (sw_context->staged_bindings) in vmw_sw_context_fini()
851 vmw_binding_state_free(sw_context->staged_bindings); in vmw_sw_context_fini()
859 struct pci_dev *pdev = to_pci_dev(dev_priv->drm.dev); in vmw_driver_load()
863 mutex_init(&dev_priv->cmdbuf_mutex); in vmw_driver_load()
864 mutex_init(&dev_priv->binding_mutex); in vmw_driver_load()
865 spin_lock_init(&dev_priv->resource_lock); in vmw_driver_load()
866 spin_lock_init(&dev_priv->hw_lock); in vmw_driver_load()
867 spin_lock_init(&dev_priv->waiter_lock); in vmw_driver_load()
868 spin_lock_init(&dev_priv->cursor_lock); in vmw_driver_load()
879 idr_init_base(&dev_priv->res_idr[i], 1); in vmw_driver_load()
880 INIT_LIST_HEAD(&dev_priv->res_lru[i]); in vmw_driver_load()
883 init_waitqueue_head(&dev_priv->fence_queue); in vmw_driver_load()
884 init_waitqueue_head(&dev_priv->fifo_queue); in vmw_driver_load()
885 dev_priv->fence_queue_waiters = 0; in vmw_driver_load()
886 dev_priv->fifo_queue_waiters = 0; in vmw_driver_load()
888 dev_priv->used_memory_size = 0; in vmw_driver_load()
890 dev_priv->assume_16bpp = !!vmw_assume_16bpp; in vmw_driver_load()
892 dev_priv->capabilities = vmw_read(dev_priv, SVGA_REG_CAPABILITIES); in vmw_driver_load()
893 vmw_print_bitmap(&dev_priv->drm, "Capabilities", in vmw_driver_load()
894 dev_priv->capabilities, in vmw_driver_load()
896 if (dev_priv->capabilities & SVGA_CAP_CAP2_REGISTER) { in vmw_driver_load()
897 dev_priv->capabilities2 = vmw_read(dev_priv, SVGA_REG_CAP2); in vmw_driver_load()
898 vmw_print_bitmap(&dev_priv->drm, "Capabilities2", in vmw_driver_load()
899 dev_priv->capabilities2, in vmw_driver_load()
905 drm_err_once(&dev_priv->drm, in vmw_driver_load()
907 drm_err_once(&dev_priv->drm, in vmw_driver_load()
909 drm_err_once(&dev_priv->drm, in vmw_driver_load()
917 drm_info(&dev_priv->drm, in vmw_driver_load()
920 if (dev_priv->capabilities & SVGA_CAP_GBOBJECTS) in vmw_driver_load()
921 drm_info(&dev_priv->drm, in vmw_driver_load()
925 dev_priv->vram_size = vmw_read(dev_priv, SVGA_REG_VRAM_SIZE); in vmw_driver_load()
926 dev_priv->fifo_mem_size = vmw_read(dev_priv, SVGA_REG_MEM_SIZE); in vmw_driver_load()
927 dev_priv->fb_max_width = vmw_read(dev_priv, SVGA_REG_MAX_WIDTH); in vmw_driver_load()
928 dev_priv->fb_max_height = vmw_read(dev_priv, SVGA_REG_MAX_HEIGHT); in vmw_driver_load()
932 if (dev_priv->capabilities & SVGA_CAP_GMR2) { in vmw_driver_load()
933 dev_priv->max_gmr_ids = in vmw_driver_load()
935 dev_priv->max_gmr_pages = in vmw_driver_load()
937 dev_priv->memory_size = in vmw_driver_load()
939 dev_priv->memory_size -= dev_priv->vram_size; in vmw_driver_load()
945 dev_priv->memory_size = 512*1024*1024; in vmw_driver_load()
947 dev_priv->max_mob_pages = 0; in vmw_driver_load()
948 dev_priv->max_mob_size = 0; in vmw_driver_load()
949 if (dev_priv->capabilities & SVGA_CAP_GBOBJECTS) { in vmw_driver_load()
952 if (dev_priv->capabilities2 & SVGA_CAP2_GB_MEMSIZE_2) in vmw_driver_load()
960 dev_priv->max_mob_pages = mem_size * 1024 / PAGE_SIZE; in vmw_driver_load()
961 dev_priv->max_primary_mem = in vmw_driver_load()
963 dev_priv->max_mob_size = in vmw_driver_load()
965 dev_priv->stdu_max_width = in vmw_driver_load()
967 dev_priv->stdu_max_height = in vmw_driver_load()
972 dev_priv->texture_max_width = vmw_read(dev_priv, in vmw_driver_load()
976 dev_priv->texture_max_height = vmw_read(dev_priv, in vmw_driver_load()
979 dev_priv->texture_max_width = 8192; in vmw_driver_load()
980 dev_priv->texture_max_height = 8192; in vmw_driver_load()
981 dev_priv->max_primary_mem = dev_priv->vram_size; in vmw_driver_load()
983 drm_info(&dev_priv->drm, in vmw_driver_load()
984 "Legacy memory limits: VRAM = %llu KiB, FIFO = %llu KiB, surface = %u KiB\n", in vmw_driver_load()
985 (u64)dev_priv->vram_size / 1024, in vmw_driver_load()
986 (u64)dev_priv->fifo_mem_size / 1024, in vmw_driver_load()
987 dev_priv->memory_size / 1024); in vmw_driver_load()
989 drm_info(&dev_priv->drm, in vmw_driver_load()
991 dev_priv->max_mob_size / 1024, dev_priv->max_mob_pages); in vmw_driver_load()
997 dma_set_max_seg_size(dev_priv->drm.dev, U32_MAX); in vmw_driver_load()
999 if (dev_priv->capabilities & SVGA_CAP_GMR2) { in vmw_driver_load()
1000 drm_info(&dev_priv->drm, in vmw_driver_load()
1002 (unsigned)dev_priv->max_gmr_ids); in vmw_driver_load()
1003 drm_info(&dev_priv->drm, in vmw_driver_load()
1005 (unsigned)dev_priv->max_gmr_pages); in vmw_driver_load()
1007 drm_info(&dev_priv->drm, in vmw_driver_load()
1009 (uint64_t)dev_priv->max_primary_mem / 1024); in vmw_driver_load()
1011 /* Need mmio memory to check for fifo pitchlock cap. */ in vmw_driver_load()
1012 if (!(dev_priv->capabilities & SVGA_CAP_DISPLAY_TOPOLOGY) && in vmw_driver_load()
1013 !(dev_priv->capabilities & SVGA_CAP_PITCHLOCK) && in vmw_driver_load()
1015 ret = -ENOSYS; in vmw_driver_load()
1020 dev_priv->tdev = ttm_object_device_init(&vmw_prime_dmabuf_ops); in vmw_driver_load()
1022 if (unlikely(dev_priv->tdev == NULL)) { in vmw_driver_load()
1023 drm_err(&dev_priv->drm, in vmw_driver_load()
1025 ret = -ENOMEM; in vmw_driver_load()
1029 if (dev_priv->capabilities & SVGA_CAP_IRQMASK) { in vmw_driver_load()
1032 drm_err(&dev_priv->drm, in vmw_driver_load()
1038 dev_priv->fman = vmw_fence_manager_init(dev_priv); in vmw_driver_load()
1039 if (unlikely(dev_priv->fman == NULL)) { in vmw_driver_load()
1040 ret = -ENOMEM; in vmw_driver_load()
1044 ret = ttm_device_init(&dev_priv->bdev, &vmw_bo_driver, in vmw_driver_load()
1045 dev_priv->drm.dev, in vmw_driver_load()
1046 dev_priv->drm.anon_inode->i_mapping, in vmw_driver_load()
1047 dev_priv->drm.vma_offset_manager, in vmw_driver_load()
1048 dev_priv->map_mode == vmw_dma_alloc_coherent, in vmw_driver_load()
1051 drm_err(&dev_priv->drm, in vmw_driver_load()
1063 drm_err(&dev_priv->drm, in vmw_driver_load()
1070 drm_err(&dev_priv->drm, in vmw_driver_load()
1080 dev_priv->has_gmr = true; in vmw_driver_load()
1082 if (((dev_priv->capabilities & (SVGA_CAP_GMR | SVGA_CAP_GMR2)) == 0) || in vmw_driver_load()
1085 drm_info(&dev_priv->drm, in vmw_driver_load()
1088 dev_priv->has_gmr = false; in vmw_driver_load()
1091 if (dev_priv->capabilities & SVGA_CAP_GBOBJECTS && !refuse_dma) { in vmw_driver_load()
1092 dev_priv->has_mob = true; in vmw_driver_load()
1095 drm_info(&dev_priv->drm, in vmw_driver_load()
1098 dev_priv->has_mob = false; in vmw_driver_load()
1101 drm_info(&dev_priv->drm, in vmw_driver_load()
1104 dev_priv->has_mob = false; in vmw_driver_load()
1108 if (dev_priv->has_mob && (dev_priv->capabilities & SVGA_CAP_DX)) { in vmw_driver_load()
1110 dev_priv->sm_type = VMW_SM_4; in vmw_driver_load()
1115 (dev_priv->capabilities2 & SVGA_CAP2_DX2)) { in vmw_driver_load()
1117 dev_priv->sm_type = VMW_SM_4_1; in vmw_driver_load()
1119 (dev_priv->capabilities2 & SVGA_CAP2_DX3)) { in vmw_driver_load()
1121 dev_priv->sm_type = VMW_SM_5; in vmw_driver_load()
1123 dev_priv->sm_type = VMW_SM_5_1X; in vmw_driver_load()
1143 dev_priv->pm_nb.notifier_call = vmwgfx_pm_notifier; in vmw_driver_load()
1144 register_pm_notifier(&dev_priv->pm_nb); in vmw_driver_load()
1152 if (dev_priv->has_mob) { in vmw_driver_load()
1156 if (dev_priv->has_gmr) in vmw_driver_load()
1161 ttm_device_fini(&dev_priv->bdev); in vmw_driver_load()
1163 vmw_fence_manager_takedown(dev_priv->fman); in vmw_driver_load()
1165 if (dev_priv->capabilities & SVGA_CAP_IRQMASK) in vmw_driver_load()
1166 vmw_irq_uninstall(&dev_priv->drm); in vmw_driver_load()
1168 ttm_object_device_release(&dev_priv->tdev); in vmw_driver_load()
1171 idr_destroy(&dev_priv->res_idr[i]); in vmw_driver_load()
1173 if (dev_priv->ctx.staged_bindings) in vmw_driver_load()
1174 vmw_binding_state_free(dev_priv->ctx.staged_bindings); in vmw_driver_load()
1183 struct pci_dev *pdev = to_pci_dev(dev->dev); in vmw_driver_unload()
1186 unregister_pm_notifier(&dev_priv->pm_nb); in vmw_driver_unload()
1197 if (dev_priv->has_gmr) in vmw_driver_unload()
1201 if (dev_priv->has_mob) { in vmw_driver_unload()
1207 ttm_device_fini(&dev_priv->bdev); in vmw_driver_unload()
1209 vmw_fence_manager_takedown(dev_priv->fman); in vmw_driver_unload()
1210 if (dev_priv->capabilities & SVGA_CAP_IRQMASK) in vmw_driver_unload()
1211 vmw_irq_uninstall(&dev_priv->drm); in vmw_driver_unload()
1213 ttm_object_device_release(&dev_priv->tdev); in vmw_driver_unload()
1216 idr_destroy(&dev_priv->res_idr[i]); in vmw_driver_unload()
1228 ttm_object_file_release(&vmw_fp->tfile); in vmw_postclose()
1236 int ret = -ENOMEM; in vmw_driver_open()
1242 vmw_fp->tfile = ttm_object_file_init(dev_priv->tdev); in vmw_driver_open()
1243 if (unlikely(vmw_fp->tfile == NULL)) in vmw_driver_open()
1246 file_priv->driver_priv = vmw_fp; in vmw_driver_open()
1260 struct drm_file *file_priv = filp->private_data; in vmw_generic_ioctl()
1261 struct drm_device *dev = file_priv->minor->dev; in vmw_generic_ioctl()
1270 && (nr < DRM_COMMAND_BASE + dev->driver->num_ioctls)) { in vmw_generic_ioctl()
1272 &vmw_ioctls[nr - DRM_COMMAND_BASE]; in vmw_generic_ioctl()
1279 return -EACCES; in vmw_generic_ioctl()
1282 if (unlikely(ioctl->cmd != cmd)) in vmw_generic_ioctl()
1285 flags = ioctl->flags; in vmw_generic_ioctl()
1287 return -EINVAL; in vmw_generic_ioctl()
1293 nr - DRM_COMMAND_BASE); in vmw_generic_ioctl()
1295 return -EINVAL; in vmw_generic_ioctl()
1340 return vmw->pci_id == VMWGFX_PCI_ID_SVGA3; in vmwgfx_supported()
1342 drm_warn_once(&vmw->drm, in vmwgfx_supported()
1349 * __vmw_svga_enable - Enable SVGA mode, FIFO and use of VRAM.
1352 * Needs the reservation sem to be held in non-exclusive mode.
1356 struct ttm_resource_manager *man = ttm_manager_type(&dev_priv->bdev, TTM_PL_VRAM); in __vmw_svga_enable()
1365 * vmw_svga_enable - Enable SVGA mode, FIFO and use of VRAM.
1375 * __vmw_svga_disable - Disable SVGA mode and use of VRAM.
1383 struct ttm_resource_manager *man = ttm_manager_type(&dev_priv->bdev, TTM_PL_VRAM); in __vmw_svga_disable()
1394 * vmw_svga_disable - Disable SVGA_MODE, and use of VRAM. Keep the fifo
1402 struct ttm_resource_manager *man = ttm_manager_type(&dev_priv->bdev, TTM_PL_VRAM); in vmw_svga_disable()
1415 vmw_kms_lost_device(&dev_priv->drm); in vmw_svga_disable()
1417 if (ttm_resource_manager_evict_all(&dev_priv->bdev, man)) in vmw_svga_disable()
1436 struct drm_minor *minor = vmw->drm.primary; in vmw_debugfs_resource_managers_init()
1437 struct dentry *root = minor->debugfs_root; in vmw_debugfs_resource_managers_init()
1439 ttm_resource_manager_create_debugfs(ttm_manager_type(&vmw->bdev, TTM_PL_SYSTEM), in vmw_debugfs_resource_managers_init()
1441 ttm_resource_manager_create_debugfs(ttm_manager_type(&vmw->bdev, TTM_PL_VRAM), in vmw_debugfs_resource_managers_init()
1443 if (vmw->has_gmr) in vmw_debugfs_resource_managers_init()
1444 ttm_resource_manager_create_debugfs(ttm_manager_type(&vmw->bdev, VMW_PL_GMR), in vmw_debugfs_resource_managers_init()
1446 if (vmw->has_mob) { in vmw_debugfs_resource_managers_init()
1447 ttm_resource_manager_create_debugfs(ttm_manager_type(&vmw->bdev, VMW_PL_MOB), in vmw_debugfs_resource_managers_init()
1449 ttm_resource_manager_create_debugfs(ttm_manager_type(&vmw->bdev, VMW_PL_SYSTEM), in vmw_debugfs_resource_managers_init()
1467 * Once user-space processes have been frozen, we can release in vmwgfx_pm_notifier()
1470 dev_priv->suspend_locked = true; in vmwgfx_pm_notifier()
1474 if (READ_ONCE(dev_priv->suspend_locked)) { in vmwgfx_pm_notifier()
1475 dev_priv->suspend_locked = false; in vmwgfx_pm_notifier()
1489 if (dev_priv->refuse_hibernation) in vmw_pci_suspend()
1490 return -EBUSY; in vmw_pci_suspend()
1534 * No user-space processes should be running now. in vmw_pm_freeze()
1536 ret = vmw_kms_suspend(&dev_priv->drm); in vmw_pm_freeze()
1545 while (ttm_device_swapout(&dev_priv->bdev, &ctx, GFP_KERNEL) > 0); in vmw_pm_freeze()
1547 if (atomic_read(&dev_priv->num_fifo_resources) != 0) { in vmw_pm_freeze()
1551 dev_priv->suspend_locked = false; in vmw_pm_freeze()
1552 if (dev_priv->suspend_state) in vmw_pm_freeze()
1554 return -EBUSY; in vmw_pm_freeze()
1557 vmw_fence_fifo_down(dev_priv->fman); in vmw_pm_freeze()
1581 vmw_fence_fifo_up(dev_priv->fman); in vmw_pm_restore()
1582 dev_priv->suspend_locked = false; in vmw_pm_restore()
1583 if (dev_priv->suspend_state) in vmw_pm_restore()
1584 vmw_kms_resume(&dev_priv->drm); in vmw_pm_restore()
1662 vmw = devm_drm_dev_alloc(&pdev->dev, &driver, in vmw_probe()
1669 pci_set_drvdata(pdev, &vmw->drm); in vmw_probe()
1671 ret = vmw_driver_load(vmw, ent->device); in vmw_probe()
1675 ret = drm_dev_register(&vmw->drm, 0); in vmw_probe()
1681 drm_client_setup(&vmw->drm, NULL); in vmw_probe()
1688 vmw_driver_unload(&vmw->drm); in vmw_probe()