| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_hw_engine.c | 277 struct xe_hw_engine *hwe = arg; in hw_engine_fini() local 279 if (hwe->exl_port) in hw_engine_fini() 280 xe_execlist_port_destroy(hwe->exl_port); in hw_engine_fini() 282 hwe->gt = NULL; in hw_engine_fini() 295 void xe_hw_engine_mmio_write32(struct xe_hw_engine *hwe, in xe_hw_engine_mmio_write32() argument 298 xe_gt_assert(hwe->gt, !(reg.addr & hwe->mmio_base)); in xe_hw_engine_mmio_write32() 299 xe_force_wake_assert_held(gt_to_fw(hwe->gt), hwe->domain); in xe_hw_engine_mmio_write32() 301 reg.addr += hwe->mmio_base; in xe_hw_engine_mmio_write32() 303 xe_mmio_write32(&hwe->gt->mmio, reg, val); in xe_hw_engine_mmio_write32() 316 u32 xe_hw_engine_mmio_read32(struct xe_hw_engine *hwe, struct xe_reg reg) in xe_hw_engine_mmio_read32() argument [all …]
|
| H A D | xe_hw_engine.h | 54 void xe_hw_engine_handle_irq(struct xe_hw_engine *hwe, u16 intr_vec); 55 void xe_hw_engine_enable_ring(struct xe_hw_engine *hwe); 59 xe_hw_engine_snapshot_capture(struct xe_hw_engine *hwe, struct xe_exec_queue *q); 61 void xe_hw_engine_print(struct xe_hw_engine *hwe, struct drm_printer *p); 62 void xe_hw_engine_setup_default_lrc_state(struct xe_hw_engine *hwe); 64 bool xe_hw_engine_is_reserved(struct xe_hw_engine *hwe); 70 static inline bool xe_hw_engine_is_valid(struct xe_hw_engine *hwe) in xe_hw_engine_is_valid() argument 72 return hwe->name; in xe_hw_engine_is_valid() 76 u64 xe_hw_engine_read_timestamp(struct xe_hw_engine *hwe); 77 enum xe_force_wake_domains xe_hw_engine_to_fw_domain(struct xe_hw_engine *hwe); [all …]
|
| H A D | xe_gt.c | 220 struct xe_hw_engine *hwe = q->hwe; in emit_wa_job() local 221 struct xe_reg_sr *sr = &hwe->reg_lrc; in emit_wa_job() 263 if (hwe->class == XE_ENGINE_CLASS_RENDER) in emit_wa_job() 268 bb_len += xe_gt_lrc_size(gt, hwe->class) / sizeof(u32); in emit_wa_job() 270 xe_gt_dbg(gt, "LRC %s WA job: %zu dwords\n", hwe->name, bb_len); in emit_wa_job() 316 *cs++ = CS_MMIO_GROUP_INSTANCE_SELECT(hwe->mmio_base).addr; in emit_wa_job() 365 *cs++ = CS_MMIO_GROUP_INSTANCE_SELECT(q->hwe->mmio_base).addr; in emit_wa_job() 385 struct xe_hw_engine *hwe; in xe_gt_record_default_lrcs() local 389 for_each_hw_engine(hwe, gt, id) { in xe_gt_record_default_lrcs() 393 if (gt->default_lrc[hwe->class]) in xe_gt_record_default_lrcs() [all …]
|
| H A D | xe_reg_whitelist.c | 24 const struct xe_hw_engine *hwe) in match_not_render() argument 26 return hwe->class != XE_ENGINE_CLASS_RENDER; in match_not_render() 31 const struct xe_hw_engine *hwe) in match_has_mert() argument 145 static void whitelist_apply_to_hwe(struct xe_hw_engine *hwe) in whitelist_apply_to_hwe() argument 147 struct xe_reg_sr *sr = &hwe->reg_whitelist; in whitelist_apply_to_hwe() 153 xe_gt_dbg(hwe->gt, "Add %s whitelist to engine\n", sr->name); in whitelist_apply_to_hwe() 154 p = xe_gt_dbg_printer(hwe->gt); in whitelist_apply_to_hwe() 159 .reg = RING_FORCE_TO_NONPRIV(hwe->mmio_base, slot), in whitelist_apply_to_hwe() 166 xe_gt_err(hwe->gt, in whitelist_apply_to_hwe() 168 hwe->name, RING_MAX_NONPRIV_SLOTS); in whitelist_apply_to_hwe() [all …]
|
| H A D | xe_lrc.c | 186 const struct xe_hw_engine *hwe) in set_offsets() argument 196 const u32 base = hwe->mmio_base; in set_offsets() 217 xe_gt_assert(hwe->gt, count); in set_offsets() 647 static void set_context_control(u32 *regs, struct xe_hw_engine *hwe) in set_context_control() argument 652 if (xe_gt_has_indirect_ring_state(hwe->gt)) in set_context_control() 657 static void set_memory_based_intr(u32 *regs, struct xe_hw_engine *hwe) in set_memory_based_intr() argument 659 struct xe_memirq *memirq = >_to_tile(hwe->gt)->memirq; in set_memory_based_intr() 660 struct xe_device *xe = gt_to_xe(hwe->gt); in set_memory_based_intr() 675 regs[CTX_INT_STATUS_REPORT_PTR] = xe_memirq_status_ptr(memirq, hwe); in set_memory_based_intr() 677 regs[CTX_INT_SRC_REPORT_PTR] = xe_memirq_source_ptr(memirq, hwe); in set_memory_based_intr() [all …]
|
| H A D | xe_drm_client.c | 279 struct xe_hw_engine *hwe = xe_gt_any_hw_engine(gt); in any_engine() local 281 if (hwe) in any_engine() 282 return hwe; in any_engine() 298 struct xe_hw_engine *hwe; in force_wake_get_any_engine() local 302 hwe = any_engine(xe); in force_wake_get_any_engine() 303 if (!hwe) in force_wake_get_any_engine() 306 domain = xe_hw_engine_to_fw_domain(hwe); in force_wake_get_any_engine() 308 fw_ref = xe_force_wake_constructor(gt_to_fw(hwe->gt), domain); in force_wake_get_any_engine() 310 *phwe = hwe; /* valid forcewake */ in force_wake_get_any_engine() 321 struct xe_hw_engine *hwe; in show_run_ticks() local [all …]
|
| H A D | xe_exec_queue.c | 203 u16 width, struct xe_hw_engine *hwe, in __xe_exec_queue_alloc() argument 207 struct xe_gt *gt = hwe->gt; in __xe_exec_queue_alloc() 219 q->hwe = hwe; in __xe_exec_queue_alloc() 221 q->class = hwe->class; in __xe_exec_queue_alloc() 225 q->fence_irq = >->fence_irq[hwe->class]; in __xe_exec_queue_alloc() 226 q->ring_ops = gt->ring_ops[hwe->class]; in __xe_exec_queue_alloc() 237 q->sched_props.timeslice_us = hwe->eclass->sched_props.timeslice_us; in __xe_exec_queue_alloc() 239 hwe->eclass->sched_props.preempt_timeout_us; in __xe_exec_queue_alloc() 241 hwe->eclass->sched_props.job_timeout_ms; in __xe_exec_queue_alloc() 382 lrc = xe_lrc_create(q->hwe, q->vm, q->replay_state, in __xe_exec_queue_init() [all …]
|
| H A D | xe_guc_ads.c | 238 struct xe_hw_engine *hwe; in calculate_regset_size() local 242 for_each_hw_engine(hwe, gt, id) in calculate_regset_size() 243 xa_for_each(&hwe->reg_sr.xa, sr_idx, sr_entry) in calculate_regset_size() 256 struct xe_hw_engine *hwe; in engine_enable_mask() local 260 for_each_hw_engine(hwe, gt, id) in engine_enable_mask() 261 if (hwe->class == class) in engine_enable_mask() 262 mask |= BIT(hwe->instance); in engine_enable_mask() 542 struct xe_hw_engine *hwe; in guc_mapping_table_init() local 547 for_each_hw_engine(hwe, gt, id) { in guc_mapping_table_init() 550 guc_class = xe_engine_class_to_guc_class(hwe->class); in guc_mapping_table_init() [all …]
|
| H A D | xe_oa.c | 96 struct xe_hw_engine *hwe; member 453 if (stream->hwe->class != XE_ENGINE_CLASS_COMPUTE) in __oa_ccs_select() 456 val = REG_FIELD_PREP(OAG_OACONTROL_OA_CCS_SELECT_MASK, stream->hwe->instance); in __oa_ccs_select() 458 REG_FIELD_GET(OAG_OACONTROL_OA_CCS_SELECT_MASK, val) == stream->hwe->instance); in __oa_ccs_select() 751 OACTXCONTROL(stream->hwe->mmio_base), in xe_oa_configure_oar_context() 759 RING_CONTEXT_CONTROL(stream->hwe->mmio_base), in xe_oa_configure_oar_context() 776 OACTXCONTROL(stream->hwe->mmio_base), in xe_oa_configure_oac_context() 784 RING_CONTEXT_CONTROL(stream->hwe->mmio_base), in xe_oa_configure_oac_context() 800 switch (stream->hwe->class) { in xe_oa_configure_oa_context() 860 struct xe_gt *gt = stream->hwe->gt; in xe_oa_stream_destroy() [all …]
|
| H A D | xe_tuning.c | 214 void xe_tuning_process_engine(struct xe_hw_engine *hwe) in xe_tuning_process_engine() argument 216 struct xe_rtp_process_ctx ctx = XE_RTP_PROCESS_CTX_INITIALIZER(hwe); in xe_tuning_process_engine() 219 hwe->gt->tuning_active.engine, in xe_tuning_process_engine() 222 &hwe->reg_sr, false); in xe_tuning_process_engine() 234 void xe_tuning_process_lrc(struct xe_hw_engine *hwe) in xe_tuning_process_lrc() argument 236 struct xe_rtp_process_ctx ctx = XE_RTP_PROCESS_CTX_INITIALIZER(hwe); in xe_tuning_process_lrc() 239 hwe->gt->tuning_active.lrc, in xe_tuning_process_lrc() 242 &hwe->reg_lrc, true); in xe_tuning_process_lrc()
|
| H A D | xe_gt_ccs_mode.c | 51 struct xe_hw_engine *hwe; in __xe_gt_apply_ccs_mode() local 54 for_each_hw_engine(hwe, gt, id) { in __xe_gt_apply_ccs_mode() 55 if (hwe->class != XE_ENGINE_CLASS_COMPUTE) in __xe_gt_apply_ccs_mode() 58 if (hwe->logical_instance >= num_engines) in __xe_gt_apply_ccs_mode() 61 config |= BIT(hwe->instance) << XE_HW_ENGINE_CCS0; in __xe_gt_apply_ccs_mode() 68 mode |= CCS_MODE_CSLICE(cslice, hwe->instance); in __xe_gt_apply_ccs_mode()
|
| H A D | xe_pmu.c | 104 struct xe_hw_engine *hwe; in event_to_hwe() local 110 hwe = xe_hw_engine_lookup(xe, eci); in event_to_hwe() 111 if (!hwe || xe_hw_engine_is_reserved(hwe)) in event_to_hwe() 114 return hwe; in event_to_hwe() 278 struct xe_hw_engine *hwe; in read_engine_events() local 285 hwe = event_to_hwe(event); in read_engine_events() 287 val = xe_guc_engine_activity_active_ticks(>->uc.guc, hwe, function_id); in read_engine_events() 289 val = xe_guc_engine_activity_total_ticks(>->uc.guc, hwe, function_id); in read_engine_events()
|
| H A D | xe_lrc.h | 54 struct xe_lrc *xe_lrc_create(struct xe_hw_engine *hwe, struct xe_vm *vm, 111 void xe_default_lrc_update_memirq_regs_with_address(struct xe_hw_engine *hwe); 112 void xe_lrc_update_memirq_regs_with_address(struct xe_lrc *lrc, struct xe_hw_engine *hwe, 156 int xe_lrc_setup_wa_bb_with_scratch(struct xe_lrc *lrc, struct xe_hw_engine *hwe,
|
| H A D | xe_hw_engine_class_sysfs.c | 626 struct xe_hw_engine *hwe; in xe_hw_engine_class_sysfs_init() local 642 for_each_hw_engine(hwe, gt, id) { in xe_hw_engine_class_sysfs_init() 646 if (hwe->class == XE_ENGINE_CLASS_OTHER || in xe_hw_engine_class_sysfs_init() 647 hwe->class == XE_ENGINE_CLASS_MAX) in xe_hw_engine_class_sysfs_init() 650 if ((class_mask >> hwe->class) & 1) in xe_hw_engine_class_sysfs_init() 653 class_mask |= 1 << hwe->class; in xe_hw_engine_class_sysfs_init() 654 name = xe_hw_engine_class_to_str(hwe->class); in xe_hw_engine_class_sysfs_init() 666 keclass->eclass = hwe->eclass; in xe_hw_engine_class_sysfs_init()
|
| H A D | xe_wa.c | 863 void xe_wa_process_engine(struct xe_hw_engine *hwe) in xe_wa_process_engine() argument 865 struct xe_rtp_process_ctx ctx = XE_RTP_PROCESS_CTX_INITIALIZER(hwe); in xe_wa_process_engine() 867 xe_rtp_process_ctx_enable_active_tracking(&ctx, hwe->gt->wa_active.engine, in xe_wa_process_engine() 870 &hwe->reg_sr, false); in xe_wa_process_engine() 881 void xe_wa_process_lrc(struct xe_hw_engine *hwe) in xe_wa_process_lrc() argument 883 struct xe_rtp_process_ctx ctx = XE_RTP_PROCESS_CTX_INITIALIZER(hwe); in xe_wa_process_lrc() 885 xe_rtp_process_ctx_enable_active_tracking(&ctx, hwe->gt->wa_active.lrc, in xe_wa_process_lrc() 888 &hwe->reg_lrc, true); in xe_wa_process_lrc()
|
| H A D | xe_gsc.c | 383 void xe_gsc_hwe_irq_handler(struct xe_hw_engine *hwe, u16 intr_vec) in xe_gsc_hwe_irq_handler() argument 385 struct xe_gt *gt = hwe->gt; in xe_gsc_hwe_irq_handler() 461 struct xe_hw_engine *hwe = xe_gt_hw_engine(gt, XE_ENGINE_CLASS_OTHER, 0, true); in xe_gsc_init_post_hwconfig() local 470 if (!hwe) in xe_gsc_init_post_hwconfig() 480 BIT(hwe->logical_instance), 1, hwe, in xe_gsc_init_post_hwconfig()
|
| H A D | xe_hw_engine_types.h | 144 void (*irq_handler)(struct xe_hw_engine *hwe, u16 intr_vec); 169 struct xe_hw_engine *hwe; member
|
| H A D | xe_reg_sr.c | 259 struct xe_hw_engine *hwe, in xe_reg_sr_lrc_check() argument 267 int ret = xe_lrc_lookup_default_reg_value(gt, hwe->class, offset, &val); in xe_reg_sr_lrc_check() 271 drm_printf(p, "%#8lx :: not found in LRC for %s\n", offset, hwe->name); in xe_reg_sr_lrc_check()
|
| H A D | xe_execlist.h | 18 struct xe_hw_engine *hwe);
|
| H A D | xe_reg_whitelist.h | 16 void xe_reg_whitelist_process_engine(struct xe_hw_engine *hwe);
|
| H A D | xe_ring_ops.c | 83 static int emit_aux_table_inv(struct xe_hw_engine *hwe, u32 *dw, int i) in emit_aux_table_inv() argument 85 struct xe_gt *gt = hwe->gt; in emit_aux_table_inv() 87 gt->ring_ops[hwe->class]->emit_aux_table_inv; in emit_aux_table_inv() 396 i = emit_aux_table_inv(job->q->hwe, dw, i); in __emit_job_gen12_video() 465 i = emit_aux_table_inv(job->q->hwe, dw, i); in __emit_job_gen12_render_compute()
|
| H A D | xe_gsc.h | 21 void xe_gsc_hwe_irq_handler(struct xe_hw_engine *hwe, u16 intr_vec);
|
| H A D | xe_execlist_types.h | 19 struct xe_hw_engine *hwe; member
|
| H A D | xe_gt_sriov_pf_debugfs.c | 376 struct xe_hw_engine *hwe; in sched_group_engines_read() local 383 for_each_hw_engine(hwe, gt, id) { in sched_group_engines_read() 384 u8 guc_class = xe_engine_class_to_guc_class(hwe->class); in sched_group_engines_read() 387 if (mask & BIT(hwe->logical_instance)) { in sched_group_engines_read() 388 strlcat(engines, hwe->name, sizeof(engines)); in sched_group_engines_read()
|
| /linux/drivers/net/ethernet/airoha/ |
| H A D | airoha_ppe.c | 301 struct airoha_foe_entry *hwe, in airoha_ppe_foe_entry_prepare() argument 311 memset(hwe, 0, sizeof(*hwe)); in airoha_ppe_foe_entry_prepare() 320 hwe->ib1 = val; in airoha_ppe_foe_entry_prepare() 370 hwe->ipv4.orig_tuple.ports = ports_pad; in airoha_ppe_foe_entry_prepare() 372 hwe->ipv6.ports = ports_pad; in airoha_ppe_foe_entry_prepare() 375 airoha_ppe_foe_set_bridge_addrs(&hwe->bridge, &data->eth); in airoha_ppe_foe_entry_prepare() 376 hwe->bridge.data = qdata; in airoha_ppe_foe_entry_prepare() 377 hwe->bridge.ib2 = val; in airoha_ppe_foe_entry_prepare() 378 l2 = &hwe->bridge.l2.common; in airoha_ppe_foe_entry_prepare() 380 hwe->ipv6.data = qdata; in airoha_ppe_foe_entry_prepare() [all …]
|