| /linux/drivers/gpu/drm/i915/gt/uc/ |
| H A D | intel_guc_ads.c | 84 static u32 guc_ads_regset_size(struct intel_guc *guc) in guc_ads_regset_size() argument 86 GEM_BUG_ON(!guc->ads_regset_size); in guc_ads_regset_size() 87 return guc->ads_regset_size; in guc_ads_regset_size() 90 static u32 guc_ads_golden_ctxt_size(struct intel_guc *guc) in guc_ads_golden_ctxt_size() argument 92 return PAGE_ALIGN(guc->ads_golden_ctxt_size); in guc_ads_golden_ctxt_size() 95 static u32 guc_ads_waklv_size(struct intel_guc *guc) in guc_ads_waklv_size() argument 97 return PAGE_ALIGN(guc->ads_waklv_size); in guc_ads_waklv_size() 100 static u32 guc_ads_capture_size(struct intel_guc *guc) in guc_ads_capture_size() argument 102 return PAGE_ALIGN(guc->ads_capture_size); in guc_ads_capture_size() 105 static u32 guc_ads_private_data_size(struct intel_guc *guc) in guc_ads_private_data_size() argument [all …]
|
| H A D | intel_guc_rc.c | 13 static bool __guc_rc_supported(struct intel_guc *guc) in __guc_rc_supported() argument 16 return guc->submission_supported && in __guc_rc_supported() 17 GRAPHICS_VER(guc_to_i915(guc)) >= 12; in __guc_rc_supported() 20 static bool __guc_rc_selected(struct intel_guc *guc) in __guc_rc_selected() argument 22 if (!intel_guc_rc_is_supported(guc)) in __guc_rc_selected() 25 return guc->submission_selected; in __guc_rc_selected() 28 void intel_guc_rc_init_early(struct intel_guc *guc) in intel_guc_rc_init_early() argument 30 guc->rc_supported = __guc_rc_supported(guc); in intel_guc_rc_init_early() 31 guc->rc_selected = __guc_rc_selected(guc); in intel_guc_rc_init_early() 34 static int guc_action_control_gucrc(struct intel_guc *guc, bool enable) in guc_action_control_gucrc() argument [all …]
|
| H A D | intel_guc_capture.c | 292 guc_capture_alloc_steered_lists(struct intel_guc *guc, in guc_capture_alloc_steered_lists() argument 295 struct intel_gt *gt = guc_to_gt(guc); in guc_capture_alloc_steered_lists() 307 if (!list || guc->capture->extlists) in guc_capture_alloc_steered_lists() 347 guc_dbg(guc, "capture found %d ext-regs.\n", num_tot_regs); in guc_capture_alloc_steered_lists() 348 guc->capture->extlists = extlists; in guc_capture_alloc_steered_lists() 352 guc_capture_get_device_reglist(struct intel_guc *guc) in guc_capture_get_device_reglist() argument 354 struct drm_i915_private *i915 = guc_to_i915(guc); in guc_capture_get_device_reglist() 368 guc_capture_alloc_steered_lists(guc, lists); in guc_capture_get_device_reglist() 412 guc_capture_list_init(struct intel_guc *guc, u32 owner, u32 type, u32 classid, in guc_capture_list_init() argument 416 const struct __guc_mmio_reg_descr_group *reglists = guc->capture->reglists; in guc_capture_list_init() [all …]
|
| H A D | intel_guc_submission.h | 16 void intel_guc_submission_init_early(struct intel_guc *guc); 17 int intel_guc_submission_init(struct intel_guc *guc); 18 int intel_guc_submission_enable(struct intel_guc *guc); 19 void intel_guc_submission_disable(struct intel_guc *guc); 20 void intel_guc_submission_fini(struct intel_guc *guc); 21 int intel_guc_preempt_work_create(struct intel_guc *guc); 22 void intel_guc_preempt_work_destroy(struct intel_guc *guc); 24 void intel_guc_submission_print_info(struct intel_guc *guc, 26 void intel_guc_submission_print_context_info(struct intel_guc *guc, 36 int intel_guc_wait_for_pending_msg(struct intel_guc *guc, [all …]
|
| H A D | intel_guc_rc.h | 11 void intel_guc_rc_init_early(struct intel_guc *guc); 13 static inline bool intel_guc_rc_is_supported(struct intel_guc *guc) in intel_guc_rc_is_supported() argument 15 return guc->rc_supported; in intel_guc_rc_is_supported() 18 static inline bool intel_guc_rc_is_wanted(struct intel_guc *guc) in intel_guc_rc_is_wanted() argument 20 return guc->submission_selected && intel_guc_rc_is_supported(guc); in intel_guc_rc_is_wanted() 23 static inline bool intel_guc_rc_is_used(struct intel_guc *guc) in intel_guc_rc_is_used() argument 25 return intel_guc_submission_is_used(guc) && intel_guc_rc_is_wanted(guc); in intel_guc_rc_is_used() 28 int intel_guc_rc_enable(struct intel_guc *guc); 29 int intel_guc_rc_disable(struct intel_guc *guc);
|
| H A D | intel_guc_ads.h | 16 int intel_guc_ads_create(struct intel_guc *guc); 17 void intel_guc_ads_destroy(struct intel_guc *guc); 18 void intel_guc_ads_init_late(struct intel_guc *guc); 19 void intel_guc_ads_reset(struct intel_guc *guc); 20 void intel_guc_ads_print_policy_info(struct intel_guc *guc, 23 u32 intel_guc_engine_usage_offset(struct intel_guc *guc);
|
| H A D | intel_guc_capture.h | 26 void intel_guc_capture_process(struct intel_guc *guc); 27 int intel_guc_capture_getlist(struct intel_guc *guc, u32 owner, u32 type, u32 classid, 29 int intel_guc_capture_getlistsize(struct intel_guc *guc, u32 owner, u32 type, u32 classid, 31 int intel_guc_capture_getnullheader(struct intel_guc *guc, void **outptr, size_t *size); 32 void intel_guc_capture_destroy(struct intel_guc *guc); 33 int intel_guc_capture_init(struct intel_guc *guc);
|
| H A D | intel_uc.h | 33 struct intel_guc guc; member 89 uc_state_checkers(guc, guc); 91 uc_state_checkers(guc, guc_submission); 92 uc_state_checkers(guc, guc_slpc); 93 uc_state_checkers(guc, guc_rc); 101 return intel_guc_wait_for_idle(&uc->guc, timeout); in intel_uc_wait_for_idle()
|
| H A D | intel_guc_fw.h | 11 int intel_guc_fw_upload(struct intel_guc *guc);
|
| H A D | intel_guc_debugfs.h | 12 void intel_guc_debugfs_register(struct intel_guc *guc, struct dentry *root);
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_guc.c | 52 static u32 guc_bo_ggtt_addr(struct xe_guc *guc, in guc_bo_ggtt_addr() argument 55 struct xe_device *xe = guc_to_xe(guc); in guc_bo_ggtt_addr() 64 addr = __xe_bo_ggtt_addr(bo, gt_to_tile(guc_to_gt(guc))->id); in guc_bo_ggtt_addr() 67 xe_assert(xe, addr >= xe_wopcm_size(guc_to_xe(guc))); in guc_bo_ggtt_addr() 74 static u32 guc_ctl_debug_flags(struct xe_guc *guc) in guc_ctl_debug_flags() argument 76 u32 level = xe_guc_log_get_level(&guc->log); in guc_ctl_debug_flags() 87 static u32 guc_ctl_feature_flags(struct xe_guc *guc) in guc_ctl_feature_flags() argument 89 struct xe_device *xe = guc_to_xe(guc); in guc_ctl_feature_flags() 98 if (xe_guc_using_main_gamctrl_queues(guc)) in guc_ctl_feature_flags() 101 if (GRAPHICS_VER(xe) >= 35 && !IS_DGFX(xe) && xe_gt_is_media_type(guc_to_gt(guc))) in guc_ctl_feature_flags() [all …]
|
| H A D | xe_guc_submit.c | 50 static int guc_submit_reset_prepare(struct xe_guc *guc); 55 return &q->gt->uc.guc; in exec_queue_to_guc() 78 return atomic_read(&q->guc->state) & EXEC_QUEUE_STATE_REGISTERED; in exec_queue_registered() 83 atomic_or(EXEC_QUEUE_STATE_REGISTERED, &q->guc->state); in set_exec_queue_registered() 88 atomic_and(~EXEC_QUEUE_STATE_REGISTERED, &q->guc->state); in clear_exec_queue_registered() 93 return atomic_read(&q->guc->state) & EXEC_QUEUE_STATE_ENABLED; in exec_queue_enabled() 98 atomic_or(EXEC_QUEUE_STATE_ENABLED, &q->guc->state); in set_exec_queue_enabled() 103 atomic_and(~EXEC_QUEUE_STATE_ENABLED, &q->guc->state); in clear_exec_queue_enabled() 108 return atomic_read(&q->guc->state) & EXEC_QUEUE_STATE_PENDING_ENABLE; in exec_queue_pending_enable() 113 atomic_or(EXEC_QUEUE_STATE_PENDING_ENABLE, &q->guc->state); in set_exec_queue_pending_enable() [all …]
|
| H A D | xe_guc.h | 25 #define GUC_SUBMIT_VER(guc) \ argument 26 MAKE_GUC_VER_STRUCT((guc)->fw.versions.found[XE_UC_FW_VER_COMPATIBILITY]) 27 #define GUC_FIRMWARE_VER(guc) \ argument 28 MAKE_GUC_VER_STRUCT((guc)->fw.versions.found[XE_UC_FW_VER_RELEASE]) 29 #define GUC_FIRMWARE_VER_AT_LEAST(guc, ver...) \ argument 30 xe_guc_fw_version_at_least((guc), MAKE_GUC_VER_ARGS(ver)) 34 void xe_guc_comm_init_early(struct xe_guc *guc); 35 int xe_guc_init_noalloc(struct xe_guc *guc); 36 int xe_guc_init(struct xe_guc *guc); 37 int xe_guc_init_post_hwconfig(struct xe_guc *guc); [all …]
|
| H A D | xe_guc_hwconfig.h | 14 int xe_guc_hwconfig_init(struct xe_guc *guc); 15 u32 xe_guc_hwconfig_size(struct xe_guc *guc); 16 void xe_guc_hwconfig_copy(struct xe_guc *guc, void *dst); 17 void xe_guc_hwconfig_dump(struct xe_guc *guc, struct drm_printer *p); 18 int xe_guc_hwconfig_lookup_u32(struct xe_guc *guc, u32 attribute, u32 *val);
|
| H A D | xe_gt_sriov_pf_monitor.c | 32 gt->sriov.pf.vfs[vfid].monitor.guc.events[e] = 0; in xe_gt_sriov_pf_monitor_flr() 41 gt->sriov.pf.vfs[vfid].monitor.guc.events[e]++; in pf_update_event_counter() 130 if (data->guc.events[e]) in xe_gt_sriov_pf_monitor_print_events() 139 #define __value(TAG, NAME, ...) , #NAME, data->guc.events[MAKE_XE_GUC_KLV_THRESHOLD_INDEX(TAG)] in xe_gt_sriov_pf_monitor_print_events()
|
| H A D | xe_guc_ct.c | 54 return container_of(ct, struct xe_gt, uc.guc.ct); in ct_to_gt() 465 struct xe_guc *guc = ct_to_guc(ct); in guc_ct_ctb_h2g_register() local 473 err = xe_guc_self_cfg64(guc, in guc_ct_ctb_h2g_register() 479 err = xe_guc_self_cfg64(guc, in guc_ct_ctb_h2g_register() 485 return xe_guc_self_cfg32(guc, in guc_ct_ctb_h2g_register() 492 struct xe_guc *guc = ct_to_guc(ct); in guc_ct_ctb_g2h_register() local 500 err = xe_guc_self_cfg64(guc, in guc_ct_ctb_g2h_register() 506 err = xe_guc_self_cfg64(guc, in guc_ct_ctb_g2h_register() 512 return xe_guc_self_cfg32(guc, in guc_ct_ctb_g2h_register() 1588 struct xe_guc *guc = ct_to_guc(ct); in process_g2h_msg() local [all …]
|
| H A D | xe_guc_ads.c | 45 return container_of(ads, struct xe_gt, uc.guc.ads); in ads_to_gt() 348 if (GUC_FIRMWARE_VER_AT_LEAST(>->uc.guc, 70, 44) && XE_GT_WA(gt, 16026508708)) in guc_waklv_init() 351 if (GUC_FIRMWARE_VER_AT_LEAST(>->uc.guc, 70, 47) && XE_GT_WA(gt, 16026007364)) { in guc_waklv_init() 587 static inline bool get_capture_list(struct xe_guc_ads *ads, struct xe_guc *guc, struct xe_gt *gt, in get_capture_list() argument 593 if (!xe_guc_capture_getlistsize(guc, owner, type, class, size)) { in get_capture_list() 597 else if (!xe_guc_capture_getlist(guc, owner, type, class, pptr)) in get_capture_list() 606 struct xe_guc *guc = ads_to_guc(ads); in guc_capture_prep_lists() local 627 if (!xe_guc_capture_getnullheader(guc, &ptr, &size)) in guc_capture_prep_lists() 650 write_empty_list = get_capture_list(ads, guc, gt, i, in guc_capture_prep_lists() 665 write_empty_list = get_capture_list(ads, guc, gt, i, in guc_capture_prep_lists() [all …]
|
| H A D | xe_gt_sriov_pf_migration.c | 154 static int guc_action_vf_save_restore(struct xe_guc *guc, u32 vfid, u32 opcode, in guc_action_vf_save_restore() argument 168 return xe_guc_ct_send_block(&guc->ct, request, ARRAY_SIZE(request)); in guc_action_vf_save_restore() 176 ret = guc_action_vf_save_restore(>->uc.guc, vfid, GUC_PF_OPCODE_VF_SAVE, 0, 0); in pf_send_guc_query_vf_mig_data_size() 185 struct xe_guc *guc = >->uc.guc; in pf_send_guc_save_vf_mig_data() local 186 CLASS(xe_guc_buf, buf)(&guc->buf, ndwords); in pf_send_guc_save_vf_mig_data() 198 ret = guc_action_vf_save_restore(guc, vfid, GUC_PF_OPCODE_VF_SAVE, in pf_send_guc_save_vf_mig_data() 215 struct xe_guc *guc = >->uc.guc; in pf_send_guc_restore_vf_mig_data() local 216 CLASS(xe_guc_buf_from_data, buf)(&guc->buf, src, size); in pf_send_guc_restore_vf_mig_data() 225 ret = guc_action_vf_save_restore(guc, vfid, GUC_PF_OPCODE_VF_RESTORE, in pf_send_guc_restore_vf_mig_data() 1029 if (!GUC_FIRMWARE_VER_AT_LEAST(>->uc.guc, 70, 54)) in pf_gt_migration_check_support()
|
| H A D | xe_pmu.c | 287 val = xe_guc_engine_activity_active_ticks(>->uc.guc, hwe, function_id); in read_engine_events() 289 val = xe_guc_engine_activity_total_ticks(>->uc.guc, hwe, function_id); in read_engine_events() 308 return xe_guc_pc_get_act_freq(>->uc.guc.pc); in __xe_pmu_event_read() 310 return xe_guc_pc_get_cur_freq_fw(>->uc.guc.pc); in __xe_pmu_event_read() 517 if (xe_guc_engine_activity_supported(>->uc.guc)) { in set_supported_events()
|
| H A D | xe_guc_log.c | 104 return container_of(log, struct xe_gt, uc.guc.log); in log_to_gt() 192 struct xe_guc *guc = log_to_guc(log); in xe_guc_log_snapshot_capture() local 221 snapshot->ver_found = guc->fw.versions.found[XE_UC_FW_VER_RELEASE]; in xe_guc_log_snapshot_capture() 222 snapshot->ver_want = guc->fw.versions.wanted; in xe_guc_log_snapshot_capture() 223 snapshot->path = guc->fw.path; in xe_guc_log_snapshot_capture()
|
| H A D | xe_gt_sriov_pf_monitor_types.h | 19 } guc; member
|
| H A D | xe_guc_debugfs.h | 12 void xe_guc_debugfs_register(struct xe_guc *guc, struct dentry *parent);
|
| H A D | xe_wopcm_types.h | 23 } guc; member
|
| /linux/drivers/gpu/drm/i915/gt/ |
| H A D | intel_gt.h | 102 static inline struct intel_gt *guc_to_gt(struct intel_guc *guc) in guc_to_gt() argument 104 return container_of(guc, struct intel_gt, uc.guc); in guc_to_gt() 122 static inline struct drm_i915_private *guc_to_i915(struct intel_guc *guc) in guc_to_i915() argument 124 return guc_to_gt(guc)->i915; in guc_to_i915() 129 return >->uc.guc; in gt_to_guc()
|
| /linux/drivers/gpu/drm/i915/selftests/ |
| H A D | intel_scheduler_helpers.c | 70 err = intel_guc_global_policies_update(&engine->gt->uc.guc); in intel_selftest_modify_policy() 89 return intel_guc_global_policies_update(&engine->gt->uc.guc); in intel_selftest_restore_policy()
|