Home
last modified time | relevance | path

Searched refs:lrc (Results 1 – 19 of 19) sorted by relevance

/linux/drivers/gpu/drm/xe/
H A Dxe_lrc.c77 lrc_to_xe(struct xe_lrc *lrc) in lrc_to_xe() argument
79 return gt_to_xe(lrc->fence_ctx.gt); in lrc_to_xe()
695 static inline bool xe_lrc_has_indirect_ring_state(struct xe_lrc *lrc) in xe_lrc_has_indirect_ring_state() argument
697 return lrc->flags & XE_LRC_FLAG_INDIRECT_RING_STATE; in xe_lrc_has_indirect_ring_state()
700 static inline u32 __xe_lrc_ring_offset(struct xe_lrc *lrc) in __xe_lrc_ring_offset() argument
705 u32 xe_lrc_pphwsp_offset(struct xe_lrc *lrc) in xe_lrc_pphwsp_offset() argument
707 return lrc->ring.size; in xe_lrc_pphwsp_offset()
720 u32 xe_lrc_regs_offset(struct xe_lrc *lrc) in xe_lrc_regs_offset() argument
722 return xe_lrc_pphwsp_offset(lrc) + LRC_PPHWSP_SIZE; in xe_lrc_regs_offset()
744 static inline u32 __xe_lrc_seqno_offset(struct xe_lrc *lrc) in __xe_lrc_seqno_offset() argument
[all …]
H A Dxe_trace_lrc.h19 #define __dev_name_lrc(lrc) dev_name(gt_to_xe((lrc)->fence_ctx.gt)->drm.dev) argument
22 TP_PROTO(struct xe_lrc *lrc, uint64_t old),
23 TP_ARGS(lrc, old),
25 __field(struct xe_lrc *, lrc)
28 __string(name, lrc->fence_ctx.name)
29 __string(device_id, __dev_name_lrc(lrc))
33 __entry->lrc = lrc;
35 __entry->new = lrc->ctx_timestamp;
40 __entry->lrc, __get_str(name),
H A Dxe_execlist.c43 static void __start_lrc(struct xe_hw_engine *hwe, struct xe_lrc *lrc, in __start_lrc() argument
52 lrc_desc = xe_lrc_descriptor(lrc); in __start_lrc()
66 xe_lrc_write_ctx_reg(lrc, CTX_RING_TAIL, lrc->ring.tail); in __start_lrc()
67 lrc->ring.old_tail = lrc->ring.tail; in __start_lrc()
116 __start_lrc(port->hwe, exl->q->lrc[0], port->last_ctx_id); in __xe_execlist_port_start()
130 xe_lrc_write_ring(port->lrc, noop, sizeof(noop)); in __xe_execlist_port_idle()
131 __start_lrc(port->hwe, port->lrc, 0); in __xe_execlist_port_idle()
137 struct xe_lrc *lrc = exl->q->lrc[0]; in xe_execlist_is_idle() local
139 return lrc->ring.tail == lrc->ring.old_tail; in xe_execlist_is_idle()
271 port->lrc = xe_lrc_create(hwe, NULL, NULL, SZ_16K, XE_IRQ_DEFAULT_MSIX, 0); in xe_execlist_port_create()
[all …]
H A Dxe_sriov_vf_ccs.c181 struct xe_lrc *lrc = xe_exec_queue_lrc(ctx->mig_q); in ccs_rw_update_ring() local
190 lrc->ring.tail = 0; in ccs_rw_update_ring()
191 xe_lrc_set_ring_head(lrc, 0); in ccs_rw_update_ring()
200 xe_lrc_write_ring(lrc, dw, i * sizeof(u32)); in ccs_rw_update_ring()
201 xe_lrc_set_ring_tail(lrc, lrc->ring.tail); in ccs_rw_update_ring()
311 struct xe_lrc *lrc = xe_exec_queue_lrc(ctx->mig_q); in xe_sriov_vf_ccs_fini() local
317 xe_lrc_set_ring_tail(lrc, xe_lrc_ring_head(lrc)); in xe_sriov_vf_ccs_fini()
392 struct xe_lrc *lrc = xe_exec_queue_lrc(ctx->mig_q); in xe_sriov_vf_ccs_rw_update_bb_addr() local
396 xe_map_wr(xe, &lrc->bo->vmap, XE_SRIOV_VF_CCS_RW_BB_ADDR_OFFSET, u32, addr); in xe_sriov_vf_ccs_rw_update_bb_addr()
H A Dxe_exec_queue.c116 struct xe_lrc *lrc; in xe_exec_queue_group_cleanup() local
132 xa_for_each(&group->xa, idx, lrc) in xe_exec_queue_group_cleanup()
133 xe_lrc_put(lrc); in xe_exec_queue_group_cleanup()
211 q = kzalloc_flex(*q, lrc, width); in __xe_exec_queue_alloc()
306 struct xe_lrc *lrc; in __xe_exec_queue_init() local
309 lrc = xe_lrc_create(q->hwe, q->vm, q->replay_state, in __xe_exec_queue_init()
311 if (IS_ERR(lrc)) { in __xe_exec_queue_init()
312 err = PTR_ERR(lrc); in __xe_exec_queue_init()
317 WRITE_ONCE(q->lrc[i], lrc); in __xe_exec_queue_init()
324 xe_lrc_put(q->lrc[i]); in __xe_exec_queue_init()
[all …]
H A Dxe_guc_submit.c718 u32, lower_32_bits(xe_lrc_descriptor(q->lrc[0]))); in xe_guc_exec_queue_group_cgp_update()
764 xe_lrc_set_multi_queue_priority(q->lrc[0], q->multi_queue.priority); in xe_guc_exec_queue_group_cgp_sync()
848 struct xe_lrc *lrc = q->lrc[i]; in __register_mlrc_exec_queue() local
850 action[len++] = lower_32_bits(xe_lrc_descriptor(lrc)); in __register_mlrc_exec_queue()
851 action[len++] = upper_32_bits(xe_lrc_descriptor(lrc)); in __register_mlrc_exec_queue()
902 struct xe_lrc *lrc = q->lrc[0]; in register_exec_queue() local
912 info.hwlrca_lo = lower_32_bits(xe_lrc_descriptor(lrc)); in register_exec_queue()
913 info.hwlrca_hi = upper_32_bits(xe_lrc_descriptor(lrc)); in register_exec_queue()
925 u64 ggtt_addr = xe_lrc_parallel_ggtt_addr(lrc); in register_exec_queue()
926 struct iosys_map map = xe_lrc_parallel_map(lrc); in register_exec_queue()
[all …]
H A Dxe_execlist_types.h31 struct xe_lrc *lrc; member
H A Dxe_exec_queue_types.h251 struct xe_lrc *lrc[] __counted_by(width);
H A Dxe_wa.c979 xe_rtp_process_ctx_enable_active_tracking(&ctx, hwe->gt->wa_active.lrc, in xe_wa_process_lrc()
1032 gt->wa_active.lrc = p; in xe_wa_gt_init()
1072 for_each_set_bit(idx, gt->wa_active.lrc, ARRAY_SIZE(lrc_was)) in xe_wa_gt_dump()
H A Dxe_gt.c397 &q->lrc[0]->bo->vmap, in xe_gt_record_default_lrcs()
398 xe_lrc_pphwsp_offset(q->lrc[0]), in xe_gt_record_default_lrcs()
H A Dxe_migrate.c1052 return migrate->q->lrc[0]; in xe_migrate_lrc()
/linux/drivers/nfc/fdp/
H A Di2c.c66 u8 lrc = 0; in fdp_nci_i2c_add_len_lrc() local
76 lrc ^= skb->data[i]; in fdp_nci_i2c_add_len_lrc()
78 skb_put_u8(skb, lrc); in fdp_nci_i2c_add_len_lrc()
132 u8 tmp[FDP_NCI_I2C_MAX_PAYLOAD], lrc, k; in fdp_nci_i2c_read() local
151 for (lrc = i = 0; i < r; i++) in fdp_nci_i2c_read()
152 lrc ^= tmp[i]; in fdp_nci_i2c_read()
159 if (lrc) { in fdp_nci_i2c_read()
/linux/drivers/gpu/drm/i915/gt/
H A Dintel_execlists_submission.c406 rq->context->lrc.desc |= CTX_DESC_FORCE_RESTORE; in __unwind_incomplete_requests()
463 ce->lrc.lrca = lrc_update_regs(ce, engine, head); in reset_active()
492 ce->lrc.ccid = ce->tag; in __execlists_schedule_in()
499 ce->lrc.ccid = tag << (XEHP_SW_CTX_ID_SHIFT - 32); in __execlists_schedule_in()
509 ce->lrc.ccid = (1 + tag) << (GEN11_SW_CTX_ID_SHIFT - 32); in __execlists_schedule_in()
514 ce->lrc.ccid |= engine->execlists.ccid; in __execlists_schedule_in()
522 CE_TRACE(ce, "schedule-in, ccid:%x\n", ce->lrc.ccid); in __execlists_schedule_in()
598 CE_TRACE(ce, "schedule-out, ccid:%x\n", ce->lrc.ccid); in __execlists_schedule_out()
612 ccid = ce->lrc.ccid; in __execlists_schedule_out()
678 desc = ce->lrc.desc; in execlists_update_context()
[all …]
H A Dintel_engine_cs.c2182 rq->context->lrc.ccid, in intel_engine_print_registers()
2196 rq->context->lrc.ccid, in intel_engine_print_registers()
/linux/drivers/gpu/drm/i915/gt/uc/
H A Dintel_guc_capture.c1555 (n->lrca & CTX_GTT_ADDRESS_MASK) == (ce->lrc.lrca & CTX_GTT_ADDRESS_MASK)) in intel_guc_capture_is_matching_engine()
1587 (n->lrca & CTX_GTT_ADDRESS_MASK) == (ce->lrc.lrca & CTX_GTT_ADDRESS_MASK)) { in intel_guc_capture_get_matching_node()
1597 ce->guc_id.id, ce->lrc.lrca); in intel_guc_capture_get_matching_node()
H A Dintel_guc_submission.c881 *wqi++ = ce->lrc.lrca; in __guc_wq_item_append()
2515 action[len++] = lower_32_bits(child->lrc.lrca); in __guc_action_register_multi_lrc_v70()
2516 action[len++] = upper_32_bits(child->lrc.lrca); in __guc_action_register_multi_lrc_v70()
2816 desc->hw_context_desc = ce->lrc.lrca; in prepare_context_registration_info_v69()
2853 desc->hw_context_desc = child->lrc.lrca; in prepare_context_registration_info_v69()
2887 info->hwlrca_lo = lower_32_bits(ce->lrc.lrca); in prepare_context_registration_info_v70()
2888 info->hwlrca_hi = upper_32_bits(ce->lrc.lrca); in prepare_context_registration_info_v70()
3006 (ce->lrc.lrca & CTX_GTT_ADDRESS_MASK)) in __guc_context_pin()
5543 drm_printf(p, "\tHW Context Desc: 0x%08x\n", ce->lrc.lrca); in guc_log_context()
/linux/drivers/scsi/ibmvscsi_tgt/
H A Dibmvscsi_tgt.c3923 long lrc; in ibmvscsis_enable_tpg() local
3928 lrc = ibmvscsis_enable_change_state(vscsi); in ibmvscsis_enable_tpg()
3929 if (lrc) in ibmvscsis_enable_tpg()
3931 lrc, vscsi->state); in ibmvscsis_enable_tpg()
/linux/drivers/scsi/
H A DFlashPoint.c7482 unsigned char lrc; in FPT_CalcLrc() local
7483 lrc = 0; in FPT_CalcLrc()
7485 lrc ^= buffer[i]; in FPT_CalcLrc()
7486 return lrc; in FPT_CalcLrc()
/linux/drivers/gpu/drm/i915/
H A Di915_perf.c1575 stream->specific_ctx_id = ce->lrc.lrca >> 12; in oa_get_render_ctx_id()