Lines Matching +full:reg +full:- +full:names

1 // SPDX-License-Identifier: GPL-2.0 OR MIT
4 * Copyright 2009-2023 VMware, Inc., Palo Alto, CA., USA
20 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
52 #include <linux/dma-mapping.h>
284 MODULE_PARM_DESC(assume_16bpp, "Assume 16-bpp when filtering modes");
325 { SVGA_CAP2_SCREENDMA_REG, "screendma reg" },
350 ARRAY_SIZE(buf) - offset, in vmw_print_bitmap()
364 static const char *names[] = { in vmw_print_sm_type() local
372 BUILD_BUG_ON(ARRAY_SIZE(names) != (VMW_SM_MAX + 1)); in vmw_print_sm_type()
373 drm_info(&dev_priv->drm, "Available shader model: %s.\n", in vmw_print_sm_type()
374 names[dev_priv->sm_type]); in vmw_print_sm_type()
378 * vmw_dummy_query_bo_create - create a bo to hold a dummy query result
414 ret = ttm_bo_reserve(&vbo->tbo, false, true, NULL); in vmw_dummy_query_bo_create()
418 ret = ttm_bo_kmap(&vbo->tbo, 0, 1, &map); in vmw_dummy_query_bo_create()
421 result->totalSize = sizeof(*result); in vmw_dummy_query_bo_create()
422 result->state = SVGA3D_QUERYSTATE_PENDING; in vmw_dummy_query_bo_create()
423 result->result32 = 0xff; in vmw_dummy_query_bo_create()
427 ttm_bo_unreserve(&vbo->tbo); in vmw_dummy_query_bo_create()
433 dev_priv->dummy_query_bo = vbo; in vmw_dummy_query_bo_create()
442 dev_priv->enable_state = vmw_read(dev_priv, SVGA_REG_ENABLE); in vmw_device_init()
443 dev_priv->config_done_state = vmw_read(dev_priv, SVGA_REG_CONFIG_DONE); in vmw_device_init()
444 dev_priv->traces_state = vmw_read(dev_priv, SVGA_REG_TRACES); in vmw_device_init()
450 (dev_priv->capabilities & SVGA_CAP_TRACES) != 0; in vmw_device_init()
453 dev_priv->fifo = vmw_fifo_create(dev_priv); in vmw_device_init()
454 if (IS_ERR(dev_priv->fifo)) { in vmw_device_init()
455 int err = PTR_ERR(dev_priv->fifo); in vmw_device_init()
456 dev_priv->fifo = NULL; in vmw_device_init()
458 } else if (!dev_priv->fifo) { in vmw_device_init()
462 dev_priv->last_read_seqno = vmw_fence_read(dev_priv); in vmw_device_init()
463 atomic_set(&dev_priv->marker_seq, dev_priv->last_read_seqno); in vmw_device_init()
476 vmw->last_read_seqno = vmw_fence_read(vmw); in vmw_device_fini()
479 vmw->config_done_state); in vmw_device_fini()
481 vmw->enable_state); in vmw_device_fini()
483 vmw->traces_state); in vmw_device_fini()
489 * vmw_request_device_late - Perform late device setup
502 if (dev_priv->has_mob) { in vmw_request_device_late()
511 if (dev_priv->cman) { in vmw_request_device_late()
512 ret = vmw_cmdbuf_set_pool_size(dev_priv->cman, 256*4096); in vmw_request_device_late()
514 struct vmw_cmdbuf_man *man = dev_priv->cman; in vmw_request_device_late()
516 dev_priv->cman = NULL; in vmw_request_device_late()
533 vmw_fence_fifo_up(dev_priv->fman); in vmw_request_device()
534 dev_priv->cman = vmw_cmdbuf_man_create(dev_priv); in vmw_request_device()
535 if (IS_ERR(dev_priv->cman)) { in vmw_request_device()
536 dev_priv->cman = NULL; in vmw_request_device()
537 dev_priv->sm_type = VMW_SM_LEGACY; in vmw_request_device()
551 if (dev_priv->cman) in vmw_request_device()
552 vmw_cmdbuf_remove_pool(dev_priv->cman); in vmw_request_device()
553 if (dev_priv->has_mob) { in vmw_request_device()
556 man = ttm_manager_type(&dev_priv->bdev, VMW_PL_MOB); in vmw_request_device()
557 ttm_resource_manager_evict_all(&dev_priv->bdev, man); in vmw_request_device()
560 if (dev_priv->cman) in vmw_request_device()
561 vmw_cmdbuf_man_destroy(dev_priv->cman); in vmw_request_device()
563 vmw_fence_fifo_down(dev_priv->fman); in vmw_request_device()
569 * vmw_release_device_early - Early part of fifo takedown.
583 BUG_ON(dev_priv->pinned_bo != NULL); in vmw_release_device_early()
585 vmw_bo_unreference(&dev_priv->dummy_query_bo); in vmw_release_device_early()
586 if (dev_priv->cman) in vmw_release_device_early()
587 vmw_cmdbuf_remove_pool(dev_priv->cman); in vmw_release_device_early()
589 if (dev_priv->has_mob) { in vmw_release_device_early()
592 man = ttm_manager_type(&dev_priv->bdev, VMW_PL_MOB); in vmw_release_device_early()
593 ttm_resource_manager_evict_all(&dev_priv->bdev, man); in vmw_release_device_early()
599 * vmw_release_device_late - Late part of fifo takedown.
608 vmw_fence_fifo_down(dev_priv->fman); in vmw_release_device_late()
609 if (dev_priv->cman) in vmw_release_device_late()
610 vmw_cmdbuf_man_destroy(dev_priv->cman); in vmw_release_device_late()
635 if (width > dev_priv->fb_max_width || in vmw_get_initial_size()
636 height > dev_priv->fb_max_height) { in vmw_get_initial_size()
646 dev_priv->initial_width = width; in vmw_get_initial_size()
647 dev_priv->initial_height = height; in vmw_get_initial_size()
651 * vmw_dma_select_mode - Determine how DMA mappings should be set up for this
658 * If this function decides that DMA is not possible, it returns -EINVAL.
664 static const char *names[vmw_dma_map_max] = { in vmw_dma_select_mode() local
671 return -EINVAL; in vmw_dma_select_mode()
674 dev_priv->map_mode = vmw_dma_alloc_coherent; in vmw_dma_select_mode()
676 dev_priv->map_mode = vmw_dma_map_bind; in vmw_dma_select_mode()
678 dev_priv->map_mode = vmw_dma_map_populate; in vmw_dma_select_mode()
680 drm_info(&dev_priv->drm, in vmw_dma_select_mode()
681 "DMA map mode: %s\n", names[dev_priv->map_mode]); in vmw_dma_select_mode()
686 * vmw_dma_masks - set required page- and dma masks
688 * @dev_priv: Pointer to struct drm-device
690 * With 32-bit we can only handle 32 bit PFNs. Optionally set that
691 * restriction also for 64-bit systems.
695 struct drm_device *dev = &dev_priv->drm; in vmw_dma_masks()
698 ret = dma_set_mask_and_coherent(dev->dev, DMA_BIT_MASK(64)); in vmw_dma_masks()
700 drm_info(&dev_priv->drm, in vmw_dma_masks()
702 return dma_set_mask_and_coherent(dev->dev, DMA_BIT_MASK(44)); in vmw_dma_masks()
711 ret = ttm_range_man_init(&dev_priv->bdev, TTM_PL_VRAM, false, in vmw_vram_manager_init()
712 dev_priv->vram_size >> PAGE_SHIFT); in vmw_vram_manager_init()
713 ttm_resource_manager_set_used(ttm_manager_type(&dev_priv->bdev, TTM_PL_VRAM), false); in vmw_vram_manager_init()
719 ttm_range_man_fini(&dev_priv->bdev, TTM_PL_VRAM); in vmw_vram_manager_fini()
730 struct pci_dev *pdev = to_pci_dev(dev->drm.dev); in vmw_setup_pci_resources()
738 dev->pci_id = pci_id; in vmw_setup_pci_resources()
742 dev->vram_start = pci_resource_start(pdev, 2); in vmw_setup_pci_resources()
743 dev->vram_size = pci_resource_len(pdev, 2); in vmw_setup_pci_resources()
745 drm_info(&dev->drm, in vmw_setup_pci_resources()
748 dev->rmmio = devm_ioremap(dev->drm.dev, in vmw_setup_pci_resources()
751 if (!dev->rmmio) { in vmw_setup_pci_resources()
752 drm_err(&dev->drm, in vmw_setup_pci_resources()
755 return -ENOMEM; in vmw_setup_pci_resources()
758 dev->io_start = pci_resource_start(pdev, 0); in vmw_setup_pci_resources()
759 dev->vram_start = pci_resource_start(pdev, 1); in vmw_setup_pci_resources()
760 dev->vram_size = pci_resource_len(pdev, 1); in vmw_setup_pci_resources()
764 drm_info(&dev->drm, in vmw_setup_pci_resources()
767 dev->fifo_mem = devm_memremap(dev->drm.dev, in vmw_setup_pci_resources()
772 if (IS_ERR(dev->fifo_mem)) { in vmw_setup_pci_resources()
773 drm_err(&dev->drm, in vmw_setup_pci_resources()
776 return PTR_ERR(dev->fifo_mem); in vmw_setup_pci_resources()
780 return -EINVAL; in vmw_setup_pci_resources()
789 drm_info(&dev->drm, in vmw_setup_pci_resources()
791 &dev->vram_start, (uint64_t)dev->vram_size / 1024); in vmw_setup_pci_resources()
804 drm_err(&dev->drm, in vmw_detect_version()
806 svga_id, dev->pci_id); in vmw_detect_version()
807 return -ENOSYS; in vmw_detect_version()
810 drm_info(&dev->drm, in vmw_detect_version()
817 if ((dev->capabilities2 & SVGA_CAP2_DX2) != 0) { in vmw_write_driver_id()
838 struct vmw_sw_context *sw_context = &dev_priv->ctx; in vmw_sw_context_init()
840 hash_init(sw_context->res_ht); in vmw_sw_context_init()
845 struct vmw_sw_context *sw_context = &dev_priv->ctx; in vmw_sw_context_fini()
847 vfree(sw_context->cmd_bounce); in vmw_sw_context_fini()
848 if (sw_context->staged_bindings) in vmw_sw_context_fini()
849 vmw_binding_state_free(sw_context->staged_bindings); in vmw_sw_context_fini()
857 struct pci_dev *pdev = to_pci_dev(dev_priv->drm.dev); in vmw_driver_load()
859 dev_priv->drm.dev_private = dev_priv; in vmw_driver_load()
863 mutex_init(&dev_priv->cmdbuf_mutex); in vmw_driver_load()
864 mutex_init(&dev_priv->binding_mutex); in vmw_driver_load()
865 spin_lock_init(&dev_priv->resource_lock); in vmw_driver_load()
866 spin_lock_init(&dev_priv->hw_lock); in vmw_driver_load()
867 spin_lock_init(&dev_priv->waiter_lock); in vmw_driver_load()
868 spin_lock_init(&dev_priv->cursor_lock); in vmw_driver_load()
879 idr_init_base(&dev_priv->res_idr[i], 1); in vmw_driver_load()
880 INIT_LIST_HEAD(&dev_priv->res_lru[i]); in vmw_driver_load()
883 init_waitqueue_head(&dev_priv->fence_queue); in vmw_driver_load()
884 init_waitqueue_head(&dev_priv->fifo_queue); in vmw_driver_load()
885 dev_priv->fence_queue_waiters = 0; in vmw_driver_load()
886 dev_priv->fifo_queue_waiters = 0; in vmw_driver_load()
888 dev_priv->used_memory_size = 0; in vmw_driver_load()
890 dev_priv->assume_16bpp = !!vmw_assume_16bpp; in vmw_driver_load()
892 dev_priv->capabilities = vmw_read(dev_priv, SVGA_REG_CAPABILITIES); in vmw_driver_load()
893 vmw_print_bitmap(&dev_priv->drm, "Capabilities", in vmw_driver_load()
894 dev_priv->capabilities, in vmw_driver_load()
896 if (dev_priv->capabilities & SVGA_CAP_CAP2_REGISTER) { in vmw_driver_load()
897 dev_priv->capabilities2 = vmw_read(dev_priv, SVGA_REG_CAP2); in vmw_driver_load()
898 vmw_print_bitmap(&dev_priv->drm, "Capabilities2", in vmw_driver_load()
899 dev_priv->capabilities2, in vmw_driver_load()
905 drm_err_once(&dev_priv->drm, in vmw_driver_load()
907 drm_err_once(&dev_priv->drm, in vmw_driver_load()
909 drm_err_once(&dev_priv->drm, in vmw_driver_load()
915 drm_info(&dev_priv->drm, in vmw_driver_load()
918 if (dev_priv->capabilities & SVGA_CAP_GBOBJECTS) in vmw_driver_load()
919 drm_info(&dev_priv->drm, in vmw_driver_load()
923 dev_priv->vram_size = vmw_read(dev_priv, SVGA_REG_VRAM_SIZE); in vmw_driver_load()
924 dev_priv->fifo_mem_size = vmw_read(dev_priv, SVGA_REG_MEM_SIZE); in vmw_driver_load()
925 dev_priv->fb_max_width = vmw_read(dev_priv, SVGA_REG_MAX_WIDTH); in vmw_driver_load()
926 dev_priv->fb_max_height = vmw_read(dev_priv, SVGA_REG_MAX_HEIGHT); in vmw_driver_load()
930 if (dev_priv->capabilities & SVGA_CAP_GMR2) { in vmw_driver_load()
931 dev_priv->max_gmr_ids = in vmw_driver_load()
933 dev_priv->max_gmr_pages = in vmw_driver_load()
935 dev_priv->memory_size = in vmw_driver_load()
937 dev_priv->memory_size -= dev_priv->vram_size; in vmw_driver_load()
943 dev_priv->memory_size = 512*1024*1024; in vmw_driver_load()
945 dev_priv->max_mob_pages = 0; in vmw_driver_load()
946 dev_priv->max_mob_size = 0; in vmw_driver_load()
947 if (dev_priv->capabilities & SVGA_CAP_GBOBJECTS) { in vmw_driver_load()
950 if (dev_priv->capabilities2 & SVGA_CAP2_GB_MEMSIZE_2) in vmw_driver_load()
962 if (!(dev_priv->capabilities & SVGA_CAP_3D)) in vmw_driver_load()
965 dev_priv->max_mob_pages = mem_size * 1024 / PAGE_SIZE; in vmw_driver_load()
966 dev_priv->max_primary_mem = in vmw_driver_load()
968 dev_priv->max_mob_size = in vmw_driver_load()
970 dev_priv->stdu_max_width = in vmw_driver_load()
972 dev_priv->stdu_max_height = in vmw_driver_load()
977 dev_priv->texture_max_width = vmw_read(dev_priv, in vmw_driver_load()
981 dev_priv->texture_max_height = vmw_read(dev_priv, in vmw_driver_load()
984 dev_priv->texture_max_width = 8192; in vmw_driver_load()
985 dev_priv->texture_max_height = 8192; in vmw_driver_load()
986 dev_priv->max_primary_mem = dev_priv->vram_size; in vmw_driver_load()
988 drm_info(&dev_priv->drm, in vmw_driver_load()
990 (u64)dev_priv->vram_size / 1024, in vmw_driver_load()
991 (u64)dev_priv->fifo_mem_size / 1024, in vmw_driver_load()
992 dev_priv->memory_size / 1024); in vmw_driver_load()
994 drm_info(&dev_priv->drm, in vmw_driver_load()
996 dev_priv->max_mob_size / 1024, dev_priv->max_mob_pages); in vmw_driver_load()
1002 dma_set_max_seg_size(dev_priv->drm.dev, U32_MAX); in vmw_driver_load()
1004 if (dev_priv->capabilities & SVGA_CAP_GMR2) { in vmw_driver_load()
1005 drm_info(&dev_priv->drm, in vmw_driver_load()
1007 (unsigned)dev_priv->max_gmr_ids); in vmw_driver_load()
1008 drm_info(&dev_priv->drm, in vmw_driver_load()
1010 (unsigned)dev_priv->max_gmr_pages); in vmw_driver_load()
1012 drm_info(&dev_priv->drm, in vmw_driver_load()
1014 (uint64_t)dev_priv->max_primary_mem / 1024); in vmw_driver_load()
1017 if (!(dev_priv->capabilities & SVGA_CAP_DISPLAY_TOPOLOGY) && in vmw_driver_load()
1018 !(dev_priv->capabilities & SVGA_CAP_PITCHLOCK) && in vmw_driver_load()
1020 ret = -ENOSYS; in vmw_driver_load()
1025 dev_priv->tdev = ttm_object_device_init(&vmw_prime_dmabuf_ops); in vmw_driver_load()
1027 if (unlikely(dev_priv->tdev == NULL)) { in vmw_driver_load()
1028 drm_err(&dev_priv->drm, in vmw_driver_load()
1030 ret = -ENOMEM; in vmw_driver_load()
1034 if (dev_priv->capabilities & SVGA_CAP_IRQMASK) { in vmw_driver_load()
1037 drm_err(&dev_priv->drm, in vmw_driver_load()
1043 dev_priv->fman = vmw_fence_manager_init(dev_priv); in vmw_driver_load()
1044 if (unlikely(dev_priv->fman == NULL)) { in vmw_driver_load()
1045 ret = -ENOMEM; in vmw_driver_load()
1049 ret = ttm_device_init(&dev_priv->bdev, &vmw_bo_driver, in vmw_driver_load()
1050 dev_priv->drm.dev, in vmw_driver_load()
1051 dev_priv->drm.anon_inode->i_mapping, in vmw_driver_load()
1052 dev_priv->drm.vma_offset_manager, in vmw_driver_load()
1053 dev_priv->map_mode == vmw_dma_alloc_coherent, in vmw_driver_load()
1056 drm_err(&dev_priv->drm, in vmw_driver_load()
1068 drm_err(&dev_priv->drm, in vmw_driver_load()
1075 drm_err(&dev_priv->drm, in vmw_driver_load()
1085 dev_priv->has_gmr = true; in vmw_driver_load()
1087 if (((dev_priv->capabilities & (SVGA_CAP_GMR | SVGA_CAP_GMR2)) == 0) || in vmw_driver_load()
1090 drm_info(&dev_priv->drm, in vmw_driver_load()
1093 dev_priv->has_gmr = false; in vmw_driver_load()
1096 if (dev_priv->capabilities & SVGA_CAP_GBOBJECTS && !refuse_dma) { in vmw_driver_load()
1097 dev_priv->has_mob = true; in vmw_driver_load()
1100 drm_info(&dev_priv->drm, in vmw_driver_load()
1103 dev_priv->has_mob = false; in vmw_driver_load()
1106 drm_info(&dev_priv->drm, in vmw_driver_load()
1109 dev_priv->has_mob = false; in vmw_driver_load()
1113 if (dev_priv->has_mob && (dev_priv->capabilities & SVGA_CAP_DX)) { in vmw_driver_load()
1115 dev_priv->sm_type = VMW_SM_4; in vmw_driver_load()
1120 (dev_priv->capabilities2 & SVGA_CAP2_DX2)) { in vmw_driver_load()
1122 dev_priv->sm_type = VMW_SM_4_1; in vmw_driver_load()
1124 (dev_priv->capabilities2 & SVGA_CAP2_DX3)) { in vmw_driver_load()
1126 dev_priv->sm_type = VMW_SM_5; in vmw_driver_load()
1128 dev_priv->sm_type = VMW_SM_5_1X; in vmw_driver_load()
1148 dev_priv->pm_nb.notifier_call = vmwgfx_pm_notifier; in vmw_driver_load()
1149 register_pm_notifier(&dev_priv->pm_nb); in vmw_driver_load()
1157 if (dev_priv->has_mob) { in vmw_driver_load()
1161 if (dev_priv->has_gmr) in vmw_driver_load()
1166 ttm_device_fini(&dev_priv->bdev); in vmw_driver_load()
1168 vmw_fence_manager_takedown(dev_priv->fman); in vmw_driver_load()
1170 if (dev_priv->capabilities & SVGA_CAP_IRQMASK) in vmw_driver_load()
1171 vmw_irq_uninstall(&dev_priv->drm); in vmw_driver_load()
1173 ttm_object_device_release(&dev_priv->tdev); in vmw_driver_load()
1176 idr_destroy(&dev_priv->res_idr[i]); in vmw_driver_load()
1178 if (dev_priv->ctx.staged_bindings) in vmw_driver_load()
1179 vmw_binding_state_free(dev_priv->ctx.staged_bindings); in vmw_driver_load()
1188 struct pci_dev *pdev = to_pci_dev(dev->dev); in vmw_driver_unload()
1191 unregister_pm_notifier(&dev_priv->pm_nb); in vmw_driver_unload()
1201 if (dev_priv->has_gmr) in vmw_driver_unload()
1205 if (dev_priv->has_mob) { in vmw_driver_unload()
1211 ttm_device_fini(&dev_priv->bdev); in vmw_driver_unload()
1213 vmw_fence_manager_takedown(dev_priv->fman); in vmw_driver_unload()
1214 if (dev_priv->capabilities & SVGA_CAP_IRQMASK) in vmw_driver_unload()
1215 vmw_irq_uninstall(&dev_priv->drm); in vmw_driver_unload()
1217 ttm_object_device_release(&dev_priv->tdev); in vmw_driver_unload()
1220 idr_destroy(&dev_priv->res_idr[i]); in vmw_driver_unload()
1232 ttm_object_file_release(&vmw_fp->tfile); in vmw_postclose()
1240 int ret = -ENOMEM; in vmw_driver_open()
1246 vmw_fp->tfile = ttm_object_file_init(dev_priv->tdev); in vmw_driver_open()
1247 if (unlikely(vmw_fp->tfile == NULL)) in vmw_driver_open()
1250 file_priv->driver_priv = vmw_fp; in vmw_driver_open()
1264 struct drm_file *file_priv = filp->private_data; in vmw_generic_ioctl()
1265 struct drm_device *dev = file_priv->minor->dev; in vmw_generic_ioctl()
1274 && (nr < DRM_COMMAND_BASE + dev->driver->num_ioctls)) { in vmw_generic_ioctl()
1276 &vmw_ioctls[nr - DRM_COMMAND_BASE]; in vmw_generic_ioctl()
1283 return -EACCES; in vmw_generic_ioctl()
1286 if (unlikely(ioctl->cmd != cmd)) in vmw_generic_ioctl()
1289 flags = ioctl->flags; in vmw_generic_ioctl()
1291 return -EINVAL; in vmw_generic_ioctl()
1297 nr - DRM_COMMAND_BASE); in vmw_generic_ioctl()
1299 return -EINVAL; in vmw_generic_ioctl()
1344 return vmw->pci_id == VMWGFX_PCI_ID_SVGA3; in vmwgfx_supported()
1346 drm_warn_once(&vmw->drm, in vmwgfx_supported()
1353 * __vmw_svga_enable - Enable SVGA mode, FIFO and use of VRAM.
1356 * Needs the reservation sem to be held in non-exclusive mode.
1360 struct ttm_resource_manager *man = ttm_manager_type(&dev_priv->bdev, TTM_PL_VRAM); in __vmw_svga_enable()
1369 * vmw_svga_enable - Enable SVGA mode, FIFO and use of VRAM.
1379 * __vmw_svga_disable - Disable SVGA mode and use of VRAM.
1387 struct ttm_resource_manager *man = ttm_manager_type(&dev_priv->bdev, TTM_PL_VRAM); in __vmw_svga_disable()
1398 * vmw_svga_disable - Disable SVGA_MODE, and use of VRAM. Keep the fifo
1406 struct ttm_resource_manager *man = ttm_manager_type(&dev_priv->bdev, TTM_PL_VRAM); in vmw_svga_disable()
1419 vmw_kms_lost_device(&dev_priv->drm); in vmw_svga_disable()
1421 if (ttm_resource_manager_evict_all(&dev_priv->bdev, man)) in vmw_svga_disable()
1440 struct drm_minor *minor = vmw->drm.primary; in vmw_debugfs_resource_managers_init()
1441 struct dentry *root = minor->debugfs_root; in vmw_debugfs_resource_managers_init()
1443 ttm_resource_manager_create_debugfs(ttm_manager_type(&vmw->bdev, TTM_PL_SYSTEM), in vmw_debugfs_resource_managers_init()
1445 ttm_resource_manager_create_debugfs(ttm_manager_type(&vmw->bdev, TTM_PL_VRAM), in vmw_debugfs_resource_managers_init()
1447 ttm_resource_manager_create_debugfs(ttm_manager_type(&vmw->bdev, VMW_PL_GMR), in vmw_debugfs_resource_managers_init()
1449 ttm_resource_manager_create_debugfs(ttm_manager_type(&vmw->bdev, VMW_PL_MOB), in vmw_debugfs_resource_managers_init()
1451 ttm_resource_manager_create_debugfs(ttm_manager_type(&vmw->bdev, VMW_PL_SYSTEM), in vmw_debugfs_resource_managers_init()
1468 * Once user-space processes have been frozen, we can release in vmwgfx_pm_notifier()
1471 dev_priv->suspend_locked = true; in vmwgfx_pm_notifier()
1475 if (READ_ONCE(dev_priv->suspend_locked)) { in vmwgfx_pm_notifier()
1476 dev_priv->suspend_locked = false; in vmwgfx_pm_notifier()
1490 if (dev_priv->refuse_hibernation) in vmw_pci_suspend()
1491 return -EBUSY; in vmw_pci_suspend()
1535 * No user-space processes should be running now. in vmw_pm_freeze()
1537 ret = vmw_kms_suspend(&dev_priv->drm); in vmw_pm_freeze()
1546 while (ttm_device_swapout(&dev_priv->bdev, &ctx, GFP_KERNEL) > 0); in vmw_pm_freeze()
1548 if (atomic_read(&dev_priv->num_fifo_resources) != 0) { in vmw_pm_freeze()
1552 dev_priv->suspend_locked = false; in vmw_pm_freeze()
1553 if (dev_priv->suspend_state) in vmw_pm_freeze()
1555 return -EBUSY; in vmw_pm_freeze()
1558 vmw_fence_fifo_down(dev_priv->fman); in vmw_pm_freeze()
1582 vmw_fence_fifo_up(dev_priv->fman); in vmw_pm_restore()
1583 dev_priv->suspend_locked = false; in vmw_pm_restore()
1584 if (dev_priv->suspend_state) in vmw_pm_restore()
1585 vmw_kms_resume(&dev_priv->drm); in vmw_pm_restore()
1660 vmw = devm_drm_dev_alloc(&pdev->dev, &driver, in vmw_probe()
1667 pci_set_drvdata(pdev, &vmw->drm); in vmw_probe()
1669 ret = vmw_driver_load(vmw, ent->device); in vmw_probe()
1673 ret = drm_dev_register(&vmw->drm, 0); in vmw_probe()
1679 drm_fbdev_generic_setup(&vmw->drm, 0); in vmw_probe()
1686 vmw_driver_unload(&vmw->drm); in vmw_probe()