| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_pm.c | 132 int xe_pm_block_on_suspend(struct xe_device *xe) in xe_pm_block_on_suspend() argument 136 return wait_for_completion_interruptible(&xe->pm_block); in xe_pm_block_on_suspend() 146 bool xe_rpm_reclaim_safe(const struct xe_device *xe) in xe_rpm_reclaim_safe() argument 148 return !xe->d3cold.capable; in xe_rpm_reclaim_safe() 151 static void xe_rpm_lockmap_acquire(const struct xe_device *xe) in xe_rpm_lockmap_acquire() argument 153 lock_map_acquire(xe_rpm_reclaim_safe(xe) ? in xe_rpm_lockmap_acquire() 158 static void xe_rpm_lockmap_release(const struct xe_device *xe) in xe_rpm_lockmap_release() argument 160 lock_map_release(xe_rpm_reclaim_safe(xe) ? in xe_rpm_lockmap_release() 171 int xe_pm_suspend(struct xe_device *xe) in xe_pm_suspend() argument 177 drm_dbg(&xe->drm, "Suspending device\n"); in xe_pm_suspend() [all …]
|
| H A D | xe_device.c | 83 struct xe_device *xe = to_xe_device(dev); in xe_file_open() local 101 xef->xe = xe; in xe_file_open() 164 struct xe_device *xe = to_xe_device(dev); in xe_file_close() local 170 guard(xe_pm_runtime)(xe); in xe_file_close() 218 struct xe_device *xe = to_xe_device(file_priv->minor->dev); in xe_drm_ioctl() local 221 if (xe_device_wedged(xe)) in xe_drm_ioctl() 224 ACQUIRE(xe_pm_runtime_ioctl, pm)(xe); in xe_drm_ioctl() 236 struct xe_device *xe = to_xe_device(file_priv->minor->dev); in xe_drm_compat_ioctl() local 239 if (xe_device_wedged(xe)) in xe_drm_compat_ioctl() 242 ACQUIRE(xe_pm_runtime_ioctl, pm)(xe); in xe_drm_compat_ioctl() [all …]
|
| H A D | xe_bo_evict.c | 15 static int xe_bo_apply_to_pinned(struct xe_device *xe, in xe_bo_apply_to_pinned() argument 24 spin_lock(&xe->pinned.lock); in xe_bo_apply_to_pinned() 32 spin_unlock(&xe->pinned.lock); in xe_bo_apply_to_pinned() 36 spin_lock(&xe->pinned.lock); in xe_bo_apply_to_pinned() 44 spin_unlock(&xe->pinned.lock); in xe_bo_apply_to_pinned() 47 spin_lock(&xe->pinned.lock); in xe_bo_apply_to_pinned() 50 spin_unlock(&xe->pinned.lock); in xe_bo_apply_to_pinned() 64 int xe_bo_notifier_prepare_all_pinned(struct xe_device *xe) in xe_bo_notifier_prepare_all_pinned() argument 68 ret = xe_bo_apply_to_pinned(xe, &xe->pinned.early.kernel_bo_present, in xe_bo_notifier_prepare_all_pinned() 69 &xe->pinned.early.kernel_bo_present, in xe_bo_notifier_prepare_all_pinned() [all …]
|
| H A D | xe_sriov_pf_migration.c | 19 static struct xe_sriov_migration_state *pf_pick_migration(struct xe_device *xe, unsigned int vfid) in pf_pick_migration() argument 21 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_pick_migration() 22 xe_assert(xe, vfid <= xe_sriov_pf_get_totalvfs(xe)); in pf_pick_migration() 24 return &xe->sriov.pf.vfs[vfid].migration; in pf_pick_migration() 34 wait_queue_head_t *xe_sriov_pf_migration_waitqueue(struct xe_device *xe, unsigned int vfid) in xe_sriov_pf_migration_waitqueue() argument 36 return &pf_pick_migration(xe, vfid)->wq; in xe_sriov_pf_migration_waitqueue() 45 bool xe_sriov_pf_migration_supported(struct xe_device *xe) in xe_sriov_pf_migration_supported() argument 47 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_migration_supported() 49 return IS_ENABLED(CONFIG_DRM_XE_DEBUG) || !xe->sriov.pf.migration.disabled; in xe_sriov_pf_migration_supported() 57 void xe_sriov_pf_migration_disable(struct xe_device *xe, const char *fmt, ...) in xe_sriov_pf_migration_disable() argument [all …]
|
| H A D | xe_debugfs.c | 42 static void read_residency_counter(struct xe_device *xe, struct xe_mmio *mmio, in read_residency_counter() argument 48 ret = xe_pmt_telem_read(to_pci_dev(xe->drm.dev), in read_residency_counter() 52 drm_warn(&xe->drm, "%s counter failed to read, ret %d\n", name, ret); in read_residency_counter() 66 struct xe_device *xe = node_to_xe(m->private); in info() local 71 guard(xe_pm_runtime)(xe); in info() 73 drm_printf(&p, "graphics_verx100 %d\n", xe->info.graphics_verx100); in info() 74 drm_printf(&p, "media_verx100 %d\n", xe->info.media_verx100); in info() 76 xe_step_name(xe->info.step.graphics), in info() 77 xe_step_name(xe->info.step.media), in info() 78 xe_step_name(xe->info.step.basedie)); in info() [all …]
|
| H A D | xe_pci.c | 516 find_subplatform(const struct xe_device *xe, const struct xe_device_desc *desc) in find_subplatform() argument 523 if (*id == xe->info.devid) in find_subplatform() 534 static int read_gmdid(struct xe_device *xe, enum xe_gmdid_type type, u32 *ver, u32 *revid) in read_gmdid() argument 536 struct xe_mmio *mmio = xe_root_tile_mmio(xe); in read_gmdid() 540 KUNIT_STATIC_STUB_REDIRECT(read_gmdid, xe, type, ver, revid); in read_gmdid() 542 if (IS_SRIOV_VF(xe)) { in read_gmdid() 557 if (type == GMDID_MEDIA && !xe_configfs_media_gt_allowed(to_pci_dev(xe->drm.dev))) { in read_gmdid() 566 gt->tile = &xe->tiles[0]; in read_gmdid() 628 static int handle_gmdid(struct xe_device *xe, in handle_gmdid() argument 640 ret = read_gmdid(xe, GMDID_GRAPHICS, &ver, graphics_revid); in handle_gmdid() [all …]
|
| H A D | xe_sriov_pf_sysfs.c | 70 struct xe_device *xe; member 77 ssize_t (*show)(struct xe_device *xe, char *buf); 78 ssize_t (*store)(struct xe_device *xe, const char *buf, size_t count); 96 ssize_t (*show)(struct xe_device *xe, unsigned int vfid, char *buf); 97 ssize_t (*store)(struct xe_device *xe, unsigned int vfid, const char *buf, size_t count); 117 static ssize_t xe_sriov_dev_attr_##NAME##_store(struct xe_device *xe, \ 127 err = xe_sriov_pf_provision_bulk_apply_##ITEM(xe, value); \ 154 static bool sched_priority_bulk_high_allowed(struct xe_device *xe) in sched_priority_bulk_high_allowed() argument 160 static ssize_t xe_sriov_dev_attr_sched_priority_store(struct xe_device *xe, in xe_sriov_dev_attr_sched_priority_store() argument 167 if (!sched_priority_bulk_high_allowed(xe)) in xe_sriov_dev_attr_sched_priority_store() [all …]
|
| H A D | xe_sriov_pf_service.c | 24 void xe_sriov_pf_service_init(struct xe_device *xe) in xe_sriov_pf_service_init() argument 29 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_service_init() 32 xe->sriov.pf.service.version.base.major = GUC_RELAY_VERSION_BASE_MAJOR; in xe_sriov_pf_service_init() 33 xe->sriov.pf.service.version.base.minor = GUC_RELAY_VERSION_BASE_MINOR; in xe_sriov_pf_service_init() 36 xe->sriov.pf.service.version.latest.major = GUC_RELAY_VERSION_LATEST_MAJOR; in xe_sriov_pf_service_init() 37 xe->sriov.pf.service.version.latest.minor = GUC_RELAY_VERSION_LATEST_MINOR; in xe_sriov_pf_service_init() 41 static int pf_negotiate_version(struct xe_device *xe, in pf_negotiate_version() argument 45 struct xe_sriov_pf_service_version base = xe->sriov.pf.service.version.base; in pf_negotiate_version() 46 struct xe_sriov_pf_service_version latest = xe->sriov.pf.service.version.latest; in pf_negotiate_version() 48 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_negotiate_version() [all …]
|
| H A D | xe_sriov_packet.c | 15 static struct mutex *pf_migration_mutex(struct xe_device *xe, unsigned int vfid) in pf_migration_mutex() argument 17 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_migration_mutex() 18 xe_assert(xe, vfid <= xe_sriov_pf_get_totalvfs(xe)); in pf_migration_mutex() 20 return &xe->sriov.pf.vfs[vfid].migration.lock; in pf_migration_mutex() 23 static struct xe_sriov_packet **pf_pick_pending(struct xe_device *xe, unsigned int vfid) in pf_pick_pending() argument 25 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_pick_pending() 26 xe_assert(xe, vfid <= xe_sriov_pf_get_totalvfs(xe)); in pf_pick_pending() 27 lockdep_assert_held(pf_migration_mutex(xe, vfid)); in pf_pick_pending() 29 return &xe->sriov.pf.vfs[vfid].migration.pending; in pf_pick_pending() 33 pf_pick_descriptor(struct xe_device *xe, unsigned int vfid) in pf_pick_descriptor() argument [all …]
|
| H A D | xe_sriov_vf_ccs.c | 121 static u64 get_ccs_bb_pool_size(struct xe_device *xe) in get_ccs_bb_pool_size() argument 128 ccs_mem_size = div64_u64(sys_mem_size, NUM_BYTES_PER_CCS_BYTE(xe)); in get_ccs_bb_pool_size() 144 struct xe_device *xe = tile_to_xe(tile); in alloc_bb_pool() local 149 bb_pool_size = get_ccs_bb_pool_size(xe); in alloc_bb_pool() 150 xe_sriov_info(xe, "Allocating %s CCS BB pool size = %lldMB\n", in alloc_bb_pool() 157 xe_sriov_err(xe, "Suballocator init failed with error: %pe\n", in alloc_bb_pool() 164 xe_map_memset(xe, &sa_manager->bo->vmap, offset, MI_NOOP, in alloc_bb_pool() 166 xe_map_memset(xe, &sa_manager->shadow->vmap, offset, MI_NOOP, in alloc_bb_pool() 170 xe_map_wr(xe, &sa_manager->bo->vmap, offset, u32, MI_BATCH_BUFFER_END); in alloc_bb_pool() 171 xe_map_wr(xe, &sa_manager->shadow->vmap, offset, u32, MI_BATCH_BUFFER_END); in alloc_bb_pool() [all …]
|
| H A D | xe_pm.h | 16 int xe_pm_suspend(struct xe_device *xe); 17 int xe_pm_resume(struct xe_device *xe); 19 int xe_pm_init_early(struct xe_device *xe); 20 int xe_pm_init(struct xe_device *xe); 21 void xe_pm_fini(struct xe_device *xe); 22 bool xe_pm_runtime_suspended(struct xe_device *xe); 23 int xe_pm_runtime_suspend(struct xe_device *xe); 24 int xe_pm_runtime_resume(struct xe_device *xe); 25 void xe_pm_runtime_get(struct xe_device *xe); 26 int xe_pm_runtime_get_ioctl(struct xe_device *xe); [all …]
|
| H A D | xe_bo.c | 95 static bool resource_is_stolen_vram(struct xe_device *xe, struct ttm_resource *res) in resource_is_stolen_vram() argument 97 return res->mem_type == XE_PL_STOLEN && IS_DGFX(xe); in resource_is_stolen_vram() 167 mem_type_to_migrate(struct xe_device *xe, u32 mem_type) in mem_type_to_migrate() argument 171 xe_assert(xe, mem_type == XE_PL_STOLEN || mem_type_is_vram(mem_type)); in mem_type_to_migrate() 172 tile = &xe->tiles[mem_type == XE_PL_STOLEN ? 0 : (mem_type - XE_PL_VRAM0)]; in mem_type_to_migrate() 178 struct xe_device *xe = ttm_to_xe_device(res->bo->bdev); in res_to_mem_region() local 182 xe_assert(xe, resource_is_vram(res)); in res_to_mem_region() 183 mgr = ttm_manager_type(&xe->ttm, res->mem_type); in res_to_mem_region() 189 static void try_add_system(struct xe_device *xe, struct xe_bo *bo, in try_add_system() argument 193 xe_assert(xe, *c < ARRAY_SIZE(bo->placements)); in try_add_system() [all …]
|
| H A D | xe_heci_gsc.c | 108 static int heci_gsc_irq_setup(struct xe_device *xe) in heci_gsc_irq_setup() argument 110 struct xe_heci_gsc *heci_gsc = &xe->heci_gsc; in heci_gsc_irq_setup() 115 drm_err(&xe->drm, "gsc irq error %d\n", heci_gsc->irq); in heci_gsc_irq_setup() 121 drm_err(&xe->drm, "gsc irq init failed %d\n", ret); in heci_gsc_irq_setup() 126 static int heci_gsc_add_device(struct xe_device *xe, const struct heci_gsc_def *def) in heci_gsc_add_device() argument 128 struct xe_heci_gsc *heci_gsc = &xe->heci_gsc; in heci_gsc_add_device() 129 struct pci_dev *pdev = to_pci_dev(xe->drm.dev); in heci_gsc_add_device() 154 drm_err(&xe->drm, "gsc aux init failed %d\n", ret); in heci_gsc_add_device() 162 drm_err(&xe->drm, "gsc aux add failed %d\n", ret); in heci_gsc_add_device() 171 int xe_heci_gsc_init(struct xe_device *xe) in xe_heci_gsc_init() argument [all …]
|
| H A D | xe_nvm.c | 44 static bool xe_nvm_non_posted_erase(struct xe_device *xe) in xe_nvm_non_posted_erase() argument 46 struct xe_mmio *mmio = xe_root_tile_mmio(xe); in xe_nvm_non_posted_erase() 48 switch (xe->info.platform) { in xe_nvm_non_posted_erase() 58 static bool xe_nvm_writable_override(struct xe_device *xe) in xe_nvm_writable_override() argument 60 struct xe_mmio *mmio = xe_root_tile_mmio(xe); in xe_nvm_writable_override() 65 switch (xe->info.platform) { in xe_nvm_writable_override() 87 drm_err(&xe->drm, "Unknown platform\n"); in xe_nvm_writable_override() 93 drm_info(&xe->drm, "NVM access overridden by jumper\n"); in xe_nvm_writable_override() 99 struct xe_device *xe = arg; in xe_nvm_fini() local 100 struct intel_dg_nvm_dev *nvm = xe->nvm; in xe_nvm_fini() [all …]
|
| H A D | xe_exec_queue.c | 110 static int exec_queue_user_extensions(struct xe_device *xe, struct xe_exec_queue *q, 165 static int alloc_dep_schedulers(struct xe_device *xe, struct xe_exec_queue *q) in alloc_dep_schedulers() argument 186 dep_scheduler = xe_dep_scheduler_create(xe, wq, q->name, in alloc_dep_schedulers() 198 static struct xe_exec_queue *__xe_exec_queue_alloc(struct xe_device *xe, in __xe_exec_queue_alloc() argument 244 err = alloc_dep_schedulers(xe, q); in __xe_exec_queue_alloc() 259 err = exec_queue_user_extensions(xe, q, extensions); in __xe_exec_queue_alloc() 338 struct xe_exec_queue *xe_exec_queue_create(struct xe_device *xe, struct xe_vm *vm, in xe_exec_queue_create() argument 347 …xe_assert(xe, !vm || (!!(vm->flags & XE_VM_FLAG_GSC) == !!(hwe->engine_id == XE_HW_ENGINE_GSCCS0))… in xe_exec_queue_create() 349 q = __xe_exec_queue_alloc(xe, vm, logical_mask, width, hwe, flags, in xe_exec_queue_create() 366 err = xe_pxp_exec_queue_add(xe->pxp, q); in xe_exec_queue_create() [all …]
|
| H A D | xe_late_bind_fw.c | 58 struct xe_device *xe = late_bind_fw_to_xe(lb_fw); in parse_cpd_header() local 67 xe_assert(xe, manifest_entry); in parse_cpd_header() 73 drm_err(&xe->drm, "%s late binding fw: Invalid CPD header length %u!\n", in parse_cpd_header() 80 drm_err(&xe->drm, "%s late binding fw: too small! %zu < %zu\n", in parse_cpd_header() 92 drm_err(&xe->drm, "%s late binding fw: Failed to find manifest_entry\n", in parse_cpd_header() 99 drm_err(&xe->drm, "%s late binding fw: too small! %zu < %zu\n", in parse_cpd_header() 115 struct xe_device *xe = late_bind_fw_to_xe(lb_fw); in parse_lb_layout() local 123 xe_assert(xe, fpt_entry); in parse_lb_layout() 129 drm_err(&xe->drm, "%s late binding fw: Invalid FPT header length %u!\n", in parse_lb_layout() 136 drm_err(&xe->drm, "%s late binding fw: too small! %zu < %zu\n", in parse_lb_layout() [all …]
|
| H A D | xe_pmu.c | 93 struct xe_device *xe = container_of(event->pmu, typeof(*xe), pmu.base); in event_to_gt() local 96 return xe_device_get_gt(xe, gt); in event_to_gt() 101 struct xe_device *xe = container_of(event->pmu, typeof(*xe), pmu.base); in event_to_hwe() local 110 hwe = xe_hw_engine_lookup(xe, eci); in event_to_hwe() 135 struct xe_device *xe = container_of(event->pmu, typeof(*xe), pmu.base); in event_gt_forcewake() local 143 gt = xe_device_get_gt(xe, config_to_gt_id(config)); in event_gt_forcewake() 163 struct xe_device *xe = container_of(pmu, typeof(*xe), pmu); in event_supported() local 164 struct xe_gt *gt = xe_device_get_gt(xe, gt_id); in event_supported() 175 struct xe_device *xe = container_of(event->pmu, typeof(*xe), pmu.base); in event_param_valid() local 180 gt = xe_device_get_gt(xe, config_to_gt_id(config)); in event_param_valid() [all …]
|
| H A D | xe_sriov_vfio.c | 21 bool xe_sriov_vfio_migration_supported(struct xe_device *xe) in xe_sriov_vfio_migration_supported() argument 23 if (!IS_SRIOV_PF(xe)) in xe_sriov_vfio_migration_supported() 26 return xe_sriov_pf_migration_supported(xe); in xe_sriov_vfio_migration_supported() 31 _type xe_sriov_vfio_##_func(struct xe_device *xe, unsigned int vfid) \ 33 if (!IS_SRIOV_PF(xe)) \ 35 if (vfid == PFID || vfid > xe_sriov_pf_num_vfs(xe)) \ 38 guard(xe_pm_runtime_noresume)(xe); \ 40 return xe_sriov_pf_##_impl(xe, vfid); \ 54 ssize_t xe_sriov_vfio_data_read(struct xe_device *xe, unsigned int vfid, in xe_sriov_vfio_data_read() argument 57 if (!IS_SRIOV_PF(xe)) in xe_sriov_vfio_data_read() [all …]
|
| H A D | xe_mmio.c | 27 struct xe_device *xe = arg; in tiles_fini() local 31 for_each_remote_tile(tile, xe, id) in tiles_fini() 54 static void mmio_multi_tile_setup(struct xe_device *xe, size_t tile_mmio_size) in mmio_multi_tile_setup() argument 63 if (xe->info.tile_count == 1) in mmio_multi_tile_setup() 66 for_each_remote_tile(tile, xe, id) in mmio_multi_tile_setup() 67 xe_mmio_init(&tile->mmio, tile, xe->mmio.regs + id * tile_mmio_size, SZ_4M); in mmio_multi_tile_setup() 70 int xe_mmio_probe_tiles(struct xe_device *xe) in xe_mmio_probe_tiles() argument 74 mmio_multi_tile_setup(xe, tile_mmio_size); in xe_mmio_probe_tiles() 76 return devm_add_action_or_reset(xe->drm.dev, tiles_fini, xe); in xe_mmio_probe_tiles() 81 struct xe_device *xe = arg; in mmio_fini() local [all …]
|
| H A D | xe_vm_madvise.c | 54 vm_dbg(&vm->xe->drm, "VMA's in range: start=0x%016llx, end=0x%016llx", addr, addr + range); in get_vmas() 83 vm_dbg(&vm->xe->drm, "madvise_range-num_vmas = %d\n", madvise_range->num_vmas); in get_vmas() 88 static void madvise_preferred_mem_loc(struct xe_device *xe, struct xe_vm *vm, in madvise_preferred_mem_loc() argument 95 xe_assert(vm->xe, op->type == DRM_XE_MEM_RANGE_ATTR_PREFERRED_LOC); in madvise_preferred_mem_loc() 121 static void madvise_atomic(struct xe_device *xe, struct xe_vm *vm, in madvise_atomic() argument 129 xe_assert(vm->xe, op->type == DRM_XE_MEM_RANGE_ATTR_ATOMIC); in madvise_atomic() 130 xe_assert(vm->xe, op->atomic.val <= DRM_XE_ATOMIC_CPU); in madvise_atomic() 135 xe->info.has_device_atomics_on_smem)) { in madvise_atomic() 163 static void madvise_pat_index(struct xe_device *xe, struct xe_vm *vm, in madvise_pat_index() argument 170 xe_assert(vm->xe, op->type == DRM_XE_MEM_RANGE_ATTR_PAT); in madvise_pat_index() [all …]
|
| H A D | xe_query.c | 50 static size_t calc_hw_engine_info_size(struct xe_device *xe) in calc_hw_engine_info_size() argument 58 for_each_gt(gt, xe, gt_id) in calc_hw_engine_info_size() 115 query_engine_cycles(struct xe_device *xe, in query_engine_cycles() argument 126 if (IS_SRIOV_VF(xe)) in query_engine_cycles() 132 } else if (XE_IOCTL_DBG(xe, query->size != size)) { in query_engine_cycles() 145 if (eci->gt_id >= xe->info.max_gt_per_tile) in query_engine_cycles() 148 gt = xe_device_get_gt(xe, eci->gt_id); in query_engine_cycles() 168 if (GRAPHICS_VER(xe) >= 20) in query_engine_cycles() 183 static int query_engines(struct xe_device *xe, in query_engines() argument 186 size_t size = calc_hw_engine_info_size(xe); in query_engines() [all …]
|
| H A D | xe_shrinker.c | 27 struct xe_device *xe; member 57 static s64 __xe_shrinker_walk(struct xe_device *xe, in __xe_shrinker_walk() argument 66 struct ttm_resource_manager *man = ttm_manager_type(&xe->ttm, mem_type); in __xe_shrinker_walk() 90 xe_assert(xe, !IS_ERR(ttm_bo)); in __xe_shrinker_walk() 102 static s64 xe_shrinker_walk(struct xe_device *xe, in xe_shrinker_walk() argument 113 lret = __xe_shrinker_walk(xe, ctx, save_flags, to_scan, scanned); in xe_shrinker_walk() 120 lret = __xe_shrinker_walk(xe, ctx, save_flags, to_scan, scanned); in xe_shrinker_walk() 129 lret = __xe_shrinker_walk(xe, ctx, flags, to_scan, scanned); in xe_shrinker_walk() 169 struct xe_device *xe = shrinker->xe; in xe_shrinker_runtime_pm_get() local 171 if (IS_DGFX(xe) || !xe_device_has_flat_ccs(xe) || in xe_shrinker_runtime_pm_get() [all …]
|
| /linux/drivers/gpu/drm/xe/tests/ |
| H A D | xe_sriov_pf_service_kunit.c | 19 struct xe_device *xe; in pf_service_test_init() local 24 xe = test->priv; in pf_service_test_init() 25 KUNIT_ASSERT_EQ(test, xe_sriov_init(xe), 0); in pf_service_test_init() 27 xe_sriov_pf_service_init(xe); in pf_service_test_init() 33 KUNIT_ASSERT_NE(test, 0, xe->sriov.pf.service.version.base.major); in pf_service_test_init() 34 KUNIT_ASSERT_NE(test, 0, xe->sriov.pf.service.version.latest.major); in pf_service_test_init() 35 KUNIT_ASSERT_LE(test, xe->sriov.pf.service.version.base.major, in pf_service_test_init() 36 xe->sriov.pf.service.version.latest.major); in pf_service_test_init() 37 if (xe->sriov.pf.service.version.base.major == xe->sriov.pf.service.version.latest.major) in pf_service_test_init() 38 KUNIT_ASSERT_LE(test, xe->sriov.pf.service.version.base.minor, in pf_service_test_init() [all …]
|
| H A D | xe_kunit_helpers.c | 36 struct xe_device *xe; in xe_kunit_helper_alloc_xe_device() local 38 xe = drm_kunit_helper_alloc_drm_device(test, dev, in xe_kunit_helper_alloc_xe_device() 41 KUNIT_ASSERT_NOT_ERR_OR_NULL(test, xe); in xe_kunit_helper_alloc_xe_device() 42 return xe; in xe_kunit_helper_alloc_xe_device() 73 struct xe_device *xe; in xe_kunit_helper_xe_device_test_init() local 80 xe = xe_kunit_helper_alloc_xe_device(test, dev); in xe_kunit_helper_xe_device_test_init() 81 KUNIT_ASSERT_NOT_ERR_OR_NULL(test, xe); in xe_kunit_helper_xe_device_test_init() 83 err = xe_pci_fake_device_init(xe); in xe_kunit_helper_xe_device_test_init() 89 test->priv = xe; in xe_kunit_helper_xe_device_test_init() 117 struct xe_device *xe = xe_device_const_cast(test->param_value); in xe_kunit_helper_xe_device_live_test_init() local [all …]
|
| /linux/drivers/gpu/drm/xe/display/ |
| H A D | xe_hdcp_gsc.c | 27 struct xe_device *xe; member 37 struct xe_device *xe = to_xe_device(drm); in intel_hdcp_gsc_check_status() local 38 struct xe_tile *tile = xe_device_get_root_tile(xe); in intel_hdcp_gsc_check_status() 43 drm_dbg_kms(&xe->drm, in intel_hdcp_gsc_check_status() 48 guard(xe_pm_runtime)(xe); in intel_hdcp_gsc_check_status() 51 drm_dbg_kms(&xe->drm, in intel_hdcp_gsc_check_status() 60 static int intel_hdcp_gsc_initialize_message(struct xe_device *xe, in intel_hdcp_gsc_initialize_message() argument 68 bo = xe_bo_create_pin_map_novm(xe, xe_device_get_root_tile(xe), PAGE_SIZE * 2, in intel_hdcp_gsc_initialize_message() 74 drm_err(&xe->drm, "Failed to allocate bo for HDCP streaming command!\n"); in intel_hdcp_gsc_initialize_message() 81 xe_map_memset(xe, &bo->vmap, 0, 0, xe_bo_size(bo)); in intel_hdcp_gsc_initialize_message() [all …]
|