Home
last modified time | relevance | path

Searched refs:xe (Results 1 – 25 of 239) sorted by relevance

12345678910

/linux/drivers/gpu/drm/xe/
H A Dxe_pm.c132 int xe_pm_block_on_suspend(struct xe_device *xe) in xe_pm_block_on_suspend() argument
136 return wait_for_completion_interruptible(&xe->pm_block); in xe_pm_block_on_suspend()
146 bool xe_rpm_reclaim_safe(const struct xe_device *xe) in xe_rpm_reclaim_safe() argument
148 return !xe->d3cold.capable; in xe_rpm_reclaim_safe()
151 static void xe_rpm_lockmap_acquire(const struct xe_device *xe) in xe_rpm_lockmap_acquire() argument
153 lock_map_acquire(xe_rpm_reclaim_safe(xe) ? in xe_rpm_lockmap_acquire()
158 static void xe_rpm_lockmap_release(const struct xe_device *xe) in xe_rpm_lockmap_release() argument
160 lock_map_release(xe_rpm_reclaim_safe(xe) ? in xe_rpm_lockmap_release()
171 int xe_pm_suspend(struct xe_device *xe) in xe_pm_suspend() argument
177 drm_dbg(&xe->drm, "Suspending device\n"); in xe_pm_suspend()
[all …]
H A Dxe_device.c83 struct xe_device *xe = to_xe_device(dev); in xe_file_open() local
101 xef->xe = xe; in xe_file_open()
164 struct xe_device *xe = to_xe_device(dev); in xe_file_close() local
170 guard(xe_pm_runtime)(xe); in xe_file_close()
218 struct xe_device *xe = to_xe_device(file_priv->minor->dev); in xe_drm_ioctl() local
221 if (xe_device_wedged(xe)) in xe_drm_ioctl()
224 ACQUIRE(xe_pm_runtime_ioctl, pm)(xe); in xe_drm_ioctl()
236 struct xe_device *xe = to_xe_device(file_priv->minor->dev); in xe_drm_compat_ioctl() local
239 if (xe_device_wedged(xe)) in xe_drm_compat_ioctl()
242 ACQUIRE(xe_pm_runtime_ioctl, pm)(xe); in xe_drm_compat_ioctl()
[all …]
H A Dxe_bo_evict.c15 static int xe_bo_apply_to_pinned(struct xe_device *xe, in xe_bo_apply_to_pinned() argument
24 spin_lock(&xe->pinned.lock); in xe_bo_apply_to_pinned()
32 spin_unlock(&xe->pinned.lock); in xe_bo_apply_to_pinned()
36 spin_lock(&xe->pinned.lock); in xe_bo_apply_to_pinned()
44 spin_unlock(&xe->pinned.lock); in xe_bo_apply_to_pinned()
47 spin_lock(&xe->pinned.lock); in xe_bo_apply_to_pinned()
50 spin_unlock(&xe->pinned.lock); in xe_bo_apply_to_pinned()
64 int xe_bo_notifier_prepare_all_pinned(struct xe_device *xe) in xe_bo_notifier_prepare_all_pinned() argument
68 ret = xe_bo_apply_to_pinned(xe, &xe->pinned.early.kernel_bo_present, in xe_bo_notifier_prepare_all_pinned()
69 &xe->pinned.early.kernel_bo_present, in xe_bo_notifier_prepare_all_pinned()
[all …]
H A Dxe_sriov_pf_migration.c19 static struct xe_sriov_migration_state *pf_pick_migration(struct xe_device *xe, unsigned int vfid) in pf_pick_migration() argument
21 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_pick_migration()
22 xe_assert(xe, vfid <= xe_sriov_pf_get_totalvfs(xe)); in pf_pick_migration()
24 return &xe->sriov.pf.vfs[vfid].migration; in pf_pick_migration()
34 wait_queue_head_t *xe_sriov_pf_migration_waitqueue(struct xe_device *xe, unsigned int vfid) in xe_sriov_pf_migration_waitqueue() argument
36 return &pf_pick_migration(xe, vfid)->wq; in xe_sriov_pf_migration_waitqueue()
45 bool xe_sriov_pf_migration_supported(struct xe_device *xe) in xe_sriov_pf_migration_supported() argument
47 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_migration_supported()
49 return IS_ENABLED(CONFIG_DRM_XE_DEBUG) || !xe->sriov.pf.migration.disabled; in xe_sriov_pf_migration_supported()
57 void xe_sriov_pf_migration_disable(struct xe_device *xe, const char *fmt, ...) in xe_sriov_pf_migration_disable() argument
[all …]
H A Dxe_debugfs.c42 static void read_residency_counter(struct xe_device *xe, struct xe_mmio *mmio, in read_residency_counter() argument
48 ret = xe_pmt_telem_read(to_pci_dev(xe->drm.dev), in read_residency_counter()
52 drm_warn(&xe->drm, "%s counter failed to read, ret %d\n", name, ret); in read_residency_counter()
66 struct xe_device *xe = node_to_xe(m->private); in info() local
71 guard(xe_pm_runtime)(xe); in info()
73 drm_printf(&p, "graphics_verx100 %d\n", xe->info.graphics_verx100); in info()
74 drm_printf(&p, "media_verx100 %d\n", xe->info.media_verx100); in info()
76 xe_step_name(xe->info.step.graphics), in info()
77 xe_step_name(xe->info.step.media), in info()
78 xe_step_name(xe->info.step.basedie)); in info()
[all …]
H A Dxe_pci.c516 find_subplatform(const struct xe_device *xe, const struct xe_device_desc *desc) in find_subplatform() argument
523 if (*id == xe->info.devid) in find_subplatform()
534 static int read_gmdid(struct xe_device *xe, enum xe_gmdid_type type, u32 *ver, u32 *revid) in read_gmdid() argument
536 struct xe_mmio *mmio = xe_root_tile_mmio(xe); in read_gmdid()
540 KUNIT_STATIC_STUB_REDIRECT(read_gmdid, xe, type, ver, revid); in read_gmdid()
542 if (IS_SRIOV_VF(xe)) { in read_gmdid()
557 if (type == GMDID_MEDIA && !xe_configfs_media_gt_allowed(to_pci_dev(xe->drm.dev))) { in read_gmdid()
566 gt->tile = &xe->tiles[0]; in read_gmdid()
628 static int handle_gmdid(struct xe_device *xe, in handle_gmdid() argument
640 ret = read_gmdid(xe, GMDID_GRAPHICS, &ver, graphics_revid); in handle_gmdid()
[all …]
H A Dxe_sriov_pf_sysfs.c70 struct xe_device *xe; member
77 ssize_t (*show)(struct xe_device *xe, char *buf);
78 ssize_t (*store)(struct xe_device *xe, const char *buf, size_t count);
96 ssize_t (*show)(struct xe_device *xe, unsigned int vfid, char *buf);
97 ssize_t (*store)(struct xe_device *xe, unsigned int vfid, const char *buf, size_t count);
117 static ssize_t xe_sriov_dev_attr_##NAME##_store(struct xe_device *xe, \
127 err = xe_sriov_pf_provision_bulk_apply_##ITEM(xe, value); \
154 static bool sched_priority_bulk_high_allowed(struct xe_device *xe) in sched_priority_bulk_high_allowed() argument
160 static ssize_t xe_sriov_dev_attr_sched_priority_store(struct xe_device *xe, in xe_sriov_dev_attr_sched_priority_store() argument
167 if (!sched_priority_bulk_high_allowed(xe)) in xe_sriov_dev_attr_sched_priority_store()
[all …]
H A Dxe_sriov_pf_service.c24 void xe_sriov_pf_service_init(struct xe_device *xe) in xe_sriov_pf_service_init() argument
29 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_service_init()
32 xe->sriov.pf.service.version.base.major = GUC_RELAY_VERSION_BASE_MAJOR; in xe_sriov_pf_service_init()
33 xe->sriov.pf.service.version.base.minor = GUC_RELAY_VERSION_BASE_MINOR; in xe_sriov_pf_service_init()
36 xe->sriov.pf.service.version.latest.major = GUC_RELAY_VERSION_LATEST_MAJOR; in xe_sriov_pf_service_init()
37 xe->sriov.pf.service.version.latest.minor = GUC_RELAY_VERSION_LATEST_MINOR; in xe_sriov_pf_service_init()
41 static int pf_negotiate_version(struct xe_device *xe, in pf_negotiate_version() argument
45 struct xe_sriov_pf_service_version base = xe->sriov.pf.service.version.base; in pf_negotiate_version()
46 struct xe_sriov_pf_service_version latest = xe->sriov.pf.service.version.latest; in pf_negotiate_version()
48 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_negotiate_version()
[all …]
H A Dxe_sriov_packet.c15 static struct mutex *pf_migration_mutex(struct xe_device *xe, unsigned int vfid) in pf_migration_mutex() argument
17 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_migration_mutex()
18 xe_assert(xe, vfid <= xe_sriov_pf_get_totalvfs(xe)); in pf_migration_mutex()
20 return &xe->sriov.pf.vfs[vfid].migration.lock; in pf_migration_mutex()
23 static struct xe_sriov_packet **pf_pick_pending(struct xe_device *xe, unsigned int vfid) in pf_pick_pending() argument
25 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_pick_pending()
26 xe_assert(xe, vfid <= xe_sriov_pf_get_totalvfs(xe)); in pf_pick_pending()
27 lockdep_assert_held(pf_migration_mutex(xe, vfid)); in pf_pick_pending()
29 return &xe->sriov.pf.vfs[vfid].migration.pending; in pf_pick_pending()
33 pf_pick_descriptor(struct xe_device *xe, unsigned int vfid) in pf_pick_descriptor() argument
[all …]
H A Dxe_sriov_vf_ccs.c121 static u64 get_ccs_bb_pool_size(struct xe_device *xe) in get_ccs_bb_pool_size() argument
128 ccs_mem_size = div64_u64(sys_mem_size, NUM_BYTES_PER_CCS_BYTE(xe)); in get_ccs_bb_pool_size()
144 struct xe_device *xe = tile_to_xe(tile); in alloc_bb_pool() local
149 bb_pool_size = get_ccs_bb_pool_size(xe); in alloc_bb_pool()
150 xe_sriov_info(xe, "Allocating %s CCS BB pool size = %lldMB\n", in alloc_bb_pool()
157 xe_sriov_err(xe, "Suballocator init failed with error: %pe\n", in alloc_bb_pool()
164 xe_map_memset(xe, &sa_manager->bo->vmap, offset, MI_NOOP, in alloc_bb_pool()
166 xe_map_memset(xe, &sa_manager->shadow->vmap, offset, MI_NOOP, in alloc_bb_pool()
170 xe_map_wr(xe, &sa_manager->bo->vmap, offset, u32, MI_BATCH_BUFFER_END); in alloc_bb_pool()
171 xe_map_wr(xe, &sa_manager->shadow->vmap, offset, u32, MI_BATCH_BUFFER_END); in alloc_bb_pool()
[all …]
H A Dxe_pm.h16 int xe_pm_suspend(struct xe_device *xe);
17 int xe_pm_resume(struct xe_device *xe);
19 int xe_pm_init_early(struct xe_device *xe);
20 int xe_pm_init(struct xe_device *xe);
21 void xe_pm_fini(struct xe_device *xe);
22 bool xe_pm_runtime_suspended(struct xe_device *xe);
23 int xe_pm_runtime_suspend(struct xe_device *xe);
24 int xe_pm_runtime_resume(struct xe_device *xe);
25 void xe_pm_runtime_get(struct xe_device *xe);
26 int xe_pm_runtime_get_ioctl(struct xe_device *xe);
[all …]
H A Dxe_bo.c95 static bool resource_is_stolen_vram(struct xe_device *xe, struct ttm_resource *res) in resource_is_stolen_vram() argument
97 return res->mem_type == XE_PL_STOLEN && IS_DGFX(xe); in resource_is_stolen_vram()
167 mem_type_to_migrate(struct xe_device *xe, u32 mem_type) in mem_type_to_migrate() argument
171 xe_assert(xe, mem_type == XE_PL_STOLEN || mem_type_is_vram(mem_type)); in mem_type_to_migrate()
172 tile = &xe->tiles[mem_type == XE_PL_STOLEN ? 0 : (mem_type - XE_PL_VRAM0)]; in mem_type_to_migrate()
178 struct xe_device *xe = ttm_to_xe_device(res->bo->bdev); in res_to_mem_region() local
182 xe_assert(xe, resource_is_vram(res)); in res_to_mem_region()
183 mgr = ttm_manager_type(&xe->ttm, res->mem_type); in res_to_mem_region()
189 static void try_add_system(struct xe_device *xe, struct xe_bo *bo, in try_add_system() argument
193 xe_assert(xe, *c < ARRAY_SIZE(bo->placements)); in try_add_system()
[all …]
H A Dxe_heci_gsc.c108 static int heci_gsc_irq_setup(struct xe_device *xe) in heci_gsc_irq_setup() argument
110 struct xe_heci_gsc *heci_gsc = &xe->heci_gsc; in heci_gsc_irq_setup()
115 drm_err(&xe->drm, "gsc irq error %d\n", heci_gsc->irq); in heci_gsc_irq_setup()
121 drm_err(&xe->drm, "gsc irq init failed %d\n", ret); in heci_gsc_irq_setup()
126 static int heci_gsc_add_device(struct xe_device *xe, const struct heci_gsc_def *def) in heci_gsc_add_device() argument
128 struct xe_heci_gsc *heci_gsc = &xe->heci_gsc; in heci_gsc_add_device()
129 struct pci_dev *pdev = to_pci_dev(xe->drm.dev); in heci_gsc_add_device()
154 drm_err(&xe->drm, "gsc aux init failed %d\n", ret); in heci_gsc_add_device()
162 drm_err(&xe->drm, "gsc aux add failed %d\n", ret); in heci_gsc_add_device()
171 int xe_heci_gsc_init(struct xe_device *xe) in xe_heci_gsc_init() argument
[all …]
H A Dxe_nvm.c44 static bool xe_nvm_non_posted_erase(struct xe_device *xe) in xe_nvm_non_posted_erase() argument
46 struct xe_mmio *mmio = xe_root_tile_mmio(xe); in xe_nvm_non_posted_erase()
48 switch (xe->info.platform) { in xe_nvm_non_posted_erase()
58 static bool xe_nvm_writable_override(struct xe_device *xe) in xe_nvm_writable_override() argument
60 struct xe_mmio *mmio = xe_root_tile_mmio(xe); in xe_nvm_writable_override()
65 switch (xe->info.platform) { in xe_nvm_writable_override()
87 drm_err(&xe->drm, "Unknown platform\n"); in xe_nvm_writable_override()
93 drm_info(&xe->drm, "NVM access overridden by jumper\n"); in xe_nvm_writable_override()
99 struct xe_device *xe = arg; in xe_nvm_fini() local
100 struct intel_dg_nvm_dev *nvm = xe->nvm; in xe_nvm_fini()
[all …]
H A Dxe_exec_queue.c110 static int exec_queue_user_extensions(struct xe_device *xe, struct xe_exec_queue *q,
165 static int alloc_dep_schedulers(struct xe_device *xe, struct xe_exec_queue *q) in alloc_dep_schedulers() argument
186 dep_scheduler = xe_dep_scheduler_create(xe, wq, q->name, in alloc_dep_schedulers()
198 static struct xe_exec_queue *__xe_exec_queue_alloc(struct xe_device *xe, in __xe_exec_queue_alloc() argument
244 err = alloc_dep_schedulers(xe, q); in __xe_exec_queue_alloc()
259 err = exec_queue_user_extensions(xe, q, extensions); in __xe_exec_queue_alloc()
338 struct xe_exec_queue *xe_exec_queue_create(struct xe_device *xe, struct xe_vm *vm, in xe_exec_queue_create() argument
347 …xe_assert(xe, !vm || (!!(vm->flags & XE_VM_FLAG_GSC) == !!(hwe->engine_id == XE_HW_ENGINE_GSCCS0))… in xe_exec_queue_create()
349 q = __xe_exec_queue_alloc(xe, vm, logical_mask, width, hwe, flags, in xe_exec_queue_create()
366 err = xe_pxp_exec_queue_add(xe->pxp, q); in xe_exec_queue_create()
[all …]
H A Dxe_late_bind_fw.c58 struct xe_device *xe = late_bind_fw_to_xe(lb_fw); in parse_cpd_header() local
67 xe_assert(xe, manifest_entry); in parse_cpd_header()
73 drm_err(&xe->drm, "%s late binding fw: Invalid CPD header length %u!\n", in parse_cpd_header()
80 drm_err(&xe->drm, "%s late binding fw: too small! %zu < %zu\n", in parse_cpd_header()
92 drm_err(&xe->drm, "%s late binding fw: Failed to find manifest_entry\n", in parse_cpd_header()
99 drm_err(&xe->drm, "%s late binding fw: too small! %zu < %zu\n", in parse_cpd_header()
115 struct xe_device *xe = late_bind_fw_to_xe(lb_fw); in parse_lb_layout() local
123 xe_assert(xe, fpt_entry); in parse_lb_layout()
129 drm_err(&xe->drm, "%s late binding fw: Invalid FPT header length %u!\n", in parse_lb_layout()
136 drm_err(&xe->drm, "%s late binding fw: too small! %zu < %zu\n", in parse_lb_layout()
[all …]
H A Dxe_pmu.c93 struct xe_device *xe = container_of(event->pmu, typeof(*xe), pmu.base); in event_to_gt() local
96 return xe_device_get_gt(xe, gt); in event_to_gt()
101 struct xe_device *xe = container_of(event->pmu, typeof(*xe), pmu.base); in event_to_hwe() local
110 hwe = xe_hw_engine_lookup(xe, eci); in event_to_hwe()
135 struct xe_device *xe = container_of(event->pmu, typeof(*xe), pmu.base); in event_gt_forcewake() local
143 gt = xe_device_get_gt(xe, config_to_gt_id(config)); in event_gt_forcewake()
163 struct xe_device *xe = container_of(pmu, typeof(*xe), pmu); in event_supported() local
164 struct xe_gt *gt = xe_device_get_gt(xe, gt_id); in event_supported()
175 struct xe_device *xe = container_of(event->pmu, typeof(*xe), pmu.base); in event_param_valid() local
180 gt = xe_device_get_gt(xe, config_to_gt_id(config)); in event_param_valid()
[all …]
H A Dxe_sriov_vfio.c21 bool xe_sriov_vfio_migration_supported(struct xe_device *xe) in xe_sriov_vfio_migration_supported() argument
23 if (!IS_SRIOV_PF(xe)) in xe_sriov_vfio_migration_supported()
26 return xe_sriov_pf_migration_supported(xe); in xe_sriov_vfio_migration_supported()
31 _type xe_sriov_vfio_##_func(struct xe_device *xe, unsigned int vfid) \
33 if (!IS_SRIOV_PF(xe)) \
35 if (vfid == PFID || vfid > xe_sriov_pf_num_vfs(xe)) \
38 guard(xe_pm_runtime_noresume)(xe); \
40 return xe_sriov_pf_##_impl(xe, vfid); \
54 ssize_t xe_sriov_vfio_data_read(struct xe_device *xe, unsigned int vfid, in xe_sriov_vfio_data_read() argument
57 if (!IS_SRIOV_PF(xe)) in xe_sriov_vfio_data_read()
[all …]
H A Dxe_mmio.c27 struct xe_device *xe = arg; in tiles_fini() local
31 for_each_remote_tile(tile, xe, id) in tiles_fini()
54 static void mmio_multi_tile_setup(struct xe_device *xe, size_t tile_mmio_size) in mmio_multi_tile_setup() argument
63 if (xe->info.tile_count == 1) in mmio_multi_tile_setup()
66 for_each_remote_tile(tile, xe, id) in mmio_multi_tile_setup()
67 xe_mmio_init(&tile->mmio, tile, xe->mmio.regs + id * tile_mmio_size, SZ_4M); in mmio_multi_tile_setup()
70 int xe_mmio_probe_tiles(struct xe_device *xe) in xe_mmio_probe_tiles() argument
74 mmio_multi_tile_setup(xe, tile_mmio_size); in xe_mmio_probe_tiles()
76 return devm_add_action_or_reset(xe->drm.dev, tiles_fini, xe); in xe_mmio_probe_tiles()
81 struct xe_device *xe = arg; in mmio_fini() local
[all …]
H A Dxe_vm_madvise.c54 vm_dbg(&vm->xe->drm, "VMA's in range: start=0x%016llx, end=0x%016llx", addr, addr + range); in get_vmas()
83 vm_dbg(&vm->xe->drm, "madvise_range-num_vmas = %d\n", madvise_range->num_vmas); in get_vmas()
88 static void madvise_preferred_mem_loc(struct xe_device *xe, struct xe_vm *vm, in madvise_preferred_mem_loc() argument
95 xe_assert(vm->xe, op->type == DRM_XE_MEM_RANGE_ATTR_PREFERRED_LOC); in madvise_preferred_mem_loc()
121 static void madvise_atomic(struct xe_device *xe, struct xe_vm *vm, in madvise_atomic() argument
129 xe_assert(vm->xe, op->type == DRM_XE_MEM_RANGE_ATTR_ATOMIC); in madvise_atomic()
130 xe_assert(vm->xe, op->atomic.val <= DRM_XE_ATOMIC_CPU); in madvise_atomic()
135 xe->info.has_device_atomics_on_smem)) { in madvise_atomic()
163 static void madvise_pat_index(struct xe_device *xe, struct xe_vm *vm, in madvise_pat_index() argument
170 xe_assert(vm->xe, op->type == DRM_XE_MEM_RANGE_ATTR_PAT); in madvise_pat_index()
[all …]
H A Dxe_query.c50 static size_t calc_hw_engine_info_size(struct xe_device *xe) in calc_hw_engine_info_size() argument
58 for_each_gt(gt, xe, gt_id) in calc_hw_engine_info_size()
115 query_engine_cycles(struct xe_device *xe, in query_engine_cycles() argument
126 if (IS_SRIOV_VF(xe)) in query_engine_cycles()
132 } else if (XE_IOCTL_DBG(xe, query->size != size)) { in query_engine_cycles()
145 if (eci->gt_id >= xe->info.max_gt_per_tile) in query_engine_cycles()
148 gt = xe_device_get_gt(xe, eci->gt_id); in query_engine_cycles()
168 if (GRAPHICS_VER(xe) >= 20) in query_engine_cycles()
183 static int query_engines(struct xe_device *xe, in query_engines() argument
186 size_t size = calc_hw_engine_info_size(xe); in query_engines()
[all …]
H A Dxe_shrinker.c27 struct xe_device *xe; member
57 static s64 __xe_shrinker_walk(struct xe_device *xe, in __xe_shrinker_walk() argument
66 struct ttm_resource_manager *man = ttm_manager_type(&xe->ttm, mem_type); in __xe_shrinker_walk()
90 xe_assert(xe, !IS_ERR(ttm_bo)); in __xe_shrinker_walk()
102 static s64 xe_shrinker_walk(struct xe_device *xe, in xe_shrinker_walk() argument
113 lret = __xe_shrinker_walk(xe, ctx, save_flags, to_scan, scanned); in xe_shrinker_walk()
120 lret = __xe_shrinker_walk(xe, ctx, save_flags, to_scan, scanned); in xe_shrinker_walk()
129 lret = __xe_shrinker_walk(xe, ctx, flags, to_scan, scanned); in xe_shrinker_walk()
169 struct xe_device *xe = shrinker->xe; in xe_shrinker_runtime_pm_get() local
171 if (IS_DGFX(xe) || !xe_device_has_flat_ccs(xe) || in xe_shrinker_runtime_pm_get()
[all …]
/linux/drivers/gpu/drm/xe/tests/
H A Dxe_sriov_pf_service_kunit.c19 struct xe_device *xe; in pf_service_test_init() local
24 xe = test->priv; in pf_service_test_init()
25 KUNIT_ASSERT_EQ(test, xe_sriov_init(xe), 0); in pf_service_test_init()
27 xe_sriov_pf_service_init(xe); in pf_service_test_init()
33 KUNIT_ASSERT_NE(test, 0, xe->sriov.pf.service.version.base.major); in pf_service_test_init()
34 KUNIT_ASSERT_NE(test, 0, xe->sriov.pf.service.version.latest.major); in pf_service_test_init()
35 KUNIT_ASSERT_LE(test, xe->sriov.pf.service.version.base.major, in pf_service_test_init()
36 xe->sriov.pf.service.version.latest.major); in pf_service_test_init()
37 if (xe->sriov.pf.service.version.base.major == xe->sriov.pf.service.version.latest.major) in pf_service_test_init()
38 KUNIT_ASSERT_LE(test, xe->sriov.pf.service.version.base.minor, in pf_service_test_init()
[all …]
H A Dxe_kunit_helpers.c36 struct xe_device *xe; in xe_kunit_helper_alloc_xe_device() local
38 xe = drm_kunit_helper_alloc_drm_device(test, dev, in xe_kunit_helper_alloc_xe_device()
41 KUNIT_ASSERT_NOT_ERR_OR_NULL(test, xe); in xe_kunit_helper_alloc_xe_device()
42 return xe; in xe_kunit_helper_alloc_xe_device()
73 struct xe_device *xe; in xe_kunit_helper_xe_device_test_init() local
80 xe = xe_kunit_helper_alloc_xe_device(test, dev); in xe_kunit_helper_xe_device_test_init()
81 KUNIT_ASSERT_NOT_ERR_OR_NULL(test, xe); in xe_kunit_helper_xe_device_test_init()
83 err = xe_pci_fake_device_init(xe); in xe_kunit_helper_xe_device_test_init()
89 test->priv = xe; in xe_kunit_helper_xe_device_test_init()
117 struct xe_device *xe = xe_device_const_cast(test->param_value); in xe_kunit_helper_xe_device_live_test_init() local
[all …]
/linux/drivers/gpu/drm/xe/display/
H A Dxe_hdcp_gsc.c27 struct xe_device *xe; member
37 struct xe_device *xe = to_xe_device(drm); in intel_hdcp_gsc_check_status() local
38 struct xe_tile *tile = xe_device_get_root_tile(xe); in intel_hdcp_gsc_check_status()
43 drm_dbg_kms(&xe->drm, in intel_hdcp_gsc_check_status()
48 guard(xe_pm_runtime)(xe); in intel_hdcp_gsc_check_status()
51 drm_dbg_kms(&xe->drm, in intel_hdcp_gsc_check_status()
60 static int intel_hdcp_gsc_initialize_message(struct xe_device *xe, in intel_hdcp_gsc_initialize_message() argument
68 bo = xe_bo_create_pin_map_novm(xe, xe_device_get_root_tile(xe), PAGE_SIZE * 2, in intel_hdcp_gsc_initialize_message()
74 drm_err(&xe->drm, "Failed to allocate bo for HDCP streaming command!\n"); in intel_hdcp_gsc_initialize_message()
81 xe_map_memset(xe, &bo->vmap, 0, 0, xe_bo_size(bo)); in intel_hdcp_gsc_initialize_message()
[all …]

12345678910