Searched refs:q_ctx (Results 1 – 5 of 5) sorted by relevance
| /linux/drivers/media/v4l2-core/ |
| H A D | v4l2-mem2mem.c | 133 struct vb2_v4l2_buffer *v4l2_m2m_next_buf(struct v4l2_m2m_queue_ctx *q_ctx) in v4l2_m2m_next_buf() argument 138 spin_lock_irqsave(&q_ctx->rdy_spinlock, flags); in v4l2_m2m_next_buf() 140 if (list_empty(&q_ctx->rdy_queue)) { in v4l2_m2m_next_buf() 141 spin_unlock_irqrestore(&q_ctx->rdy_spinlock, flags); in v4l2_m2m_next_buf() 145 b = list_first_entry(&q_ctx->rdy_queue, struct v4l2_m2m_buffer, list); in v4l2_m2m_next_buf() 146 spin_unlock_irqrestore(&q_ctx->rdy_spinlock, flags); in v4l2_m2m_next_buf() 151 struct vb2_v4l2_buffer *v4l2_m2m_last_buf(struct v4l2_m2m_queue_ctx *q_ctx) in v4l2_m2m_last_buf() argument 156 spin_lock_irqsave(&q_ctx->rdy_spinlock, flags); in v4l2_m2m_last_buf() 158 if (list_empty(&q_ctx->rdy_queue)) { in v4l2_m2m_last_buf() 159 spin_unlock_irqrestore(&q_ctx->rdy_spinlock, flags); in v4l2_m2m_last_buf() [all …]
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_sched.c | 578 struct ice_q_ctx *q_ctx; in ice_alloc_lan_q_ctx() local 586 q_ctx = devm_kcalloc(ice_hw_to_dev(hw), new_numqs, in ice_alloc_lan_q_ctx() 587 sizeof(*q_ctx), GFP_KERNEL); in ice_alloc_lan_q_ctx() 588 if (!q_ctx) in ice_alloc_lan_q_ctx() 592 q_ctx[idx].q_handle = ICE_INVAL_Q_HANDLE; in ice_alloc_lan_q_ctx() 593 q_ctx[idx].q_teid = ICE_INVAL_TEID; in ice_alloc_lan_q_ctx() 596 vsi_ctx->lan_q_ctx[tc] = q_ctx; in ice_alloc_lan_q_ctx() 604 q_ctx = devm_kcalloc(ice_hw_to_dev(hw), new_numqs, in ice_alloc_lan_q_ctx() 605 sizeof(*q_ctx), GFP_KERNEL); in ice_alloc_lan_q_ctx() 606 if (!q_ctx) in ice_alloc_lan_q_ctx() [all …]
|
| H A D | ice_lag.c | 520 struct ice_q_ctx *q_ctx; in ice_lag_qbuf_recfg() local 525 q_ctx = ice_get_lan_q_ctx(hw, vsi_num, tc, i); in ice_lag_qbuf_recfg() 526 if (!q_ctx) { in ice_lag_qbuf_recfg() 531 if (q_ctx->q_teid == ICE_INVAL_TEID) { in ice_lag_qbuf_recfg() 536 if (q_ctx->q_handle == ICE_INVAL_Q_HANDLE) { in ice_lag_qbuf_recfg() 542 qid = pf->vsi[vsi_num]->txq_map[q_ctx->q_handle]; in ice_lag_qbuf_recfg() 545 qbuf->queue_info[count].q_teid = cpu_to_le32(q_ctx->q_teid); in ice_lag_qbuf_recfg() 945 struct ice_q_ctx *q_ctx; in ice_lag_aa_qbuf_recfg() local 948 q_ctx = ice_get_lan_q_ctx(hw, vsi_num, 0, qnum); in ice_lag_aa_qbuf_recfg() 949 if (!q_ctx) { in ice_lag_aa_qbuf_recfg() [all …]
|
| H A D | ice_common.c | 4915 struct ice_q_ctx *q_ctx; in ice_get_lan_q_ctx() local 4924 q_ctx = vsi->lan_q_ctx[tc]; in ice_get_lan_q_ctx() 4925 return &q_ctx[q_handle]; in ice_get_lan_q_ctx() 4948 struct ice_q_ctx *q_ctx; in ice_ena_vsi_txq() local 4965 q_ctx = ice_get_lan_q_ctx(hw, vsi_handle, tc, q_handle); in ice_ena_vsi_txq() 4966 if (!q_ctx) { in ice_ena_vsi_txq() 5018 q_ctx->q_handle = q_handle; in ice_ena_vsi_txq() 5019 q_ctx->q_teid = le32_to_cpu(node.node_teid); in ice_ena_vsi_txq() 5024 status = ice_sched_replay_q_bw(pi, q_ctx); in ice_ena_vsi_txq() 5054 struct ice_q_ctx *q_ctx; in ice_dis_vsi_txq() local [all …]
|
| H A D | ice_sched.h | 178 int ice_sched_replay_q_bw(struct ice_port_info *pi, struct ice_q_ctx *q_ctx);
|