Home
last modified time | relevance | path

Searched refs:seq_id (Results 1 – 25 of 49) sorted by relevance

12

/linux/drivers/net/ethernet/huawei/hinic3/
H A Dhinic3_mgmt.c81 u8 seq_id, msg_id, seg_len, is_last; in hinic3_recv_msg_add_seg() local
87 seq_id = MBOX_MSG_HEADER_GET(msg_header, SEQID); in hinic3_recv_msg_add_seg()
97 if (seq_id == 0) { in hinic3_recv_msg_add_seg()
98 recv_msg->seq_id = seq_id; in hinic3_recv_msg_add_seg()
100 } else if (seq_id != recv_msg->seq_id + 1 || in hinic3_recv_msg_add_seg()
105 offset = seq_id * MGMT_SEG_LEN_MAX; in hinic3_recv_msg_add_seg()
112 recv_msg->seq_id = seq_id; in hinic3_recv_msg_add_seg()
174 recv_msg->seq_id = MGMT_BOGUS_SEQ_ID; in hinic3_recv_mgmt_msg_handler()
185 recv_msg->seq_id = MGMT_BOGUS_SEQ_ID; in hinic3_recv_mgmt_msg_handler()
195 recv_msg->seq_id = MGMT_BOGUS_SEQ_ID; in alloc_recv_msg()
H A Dhinic3_mbox.c117 u8 seq_id, seg_len, msg_id, mod; in mbox_segment_valid() local
120 seq_id = MBOX_MSG_HEADER_GET(mbox_header, SEQID); in mbox_segment_valid()
128 if (seq_id > MBOX_SEQ_ID_MAX_VAL || seg_len > MBOX_SEG_LEN || in mbox_segment_valid()
129 (seq_id == MBOX_SEQ_ID_MAX_VAL && seg_len > MBOX_LAST_SEG_MAX_LEN)) in mbox_segment_valid()
132 if (seq_id == 0) { in mbox_segment_valid()
133 msg_desc->seq_id = seq_id; in mbox_segment_valid()
138 if (seq_id != msg_desc->seq_id + 1 || in mbox_segment_valid()
143 msg_desc->seq_id = seq_id; in mbox_segment_valid()
151 src_func_idx, msg_desc->seq_id, msg_desc->msg_info.msg_id, in mbox_segment_valid()
155 seg_len, seq_id, msg_id, mod, cmd); in mbox_segment_valid()
[all …]
/linux/drivers/net/ethernet/marvell/octeontx2/af/
H A Drvu_npc_hash.c502 static void rvu_npc_exact_free_id(struct rvu *rvu, u32 seq_id) in rvu_npc_exact_free_id() argument
508 clear_bit(seq_id, table->id_bmap); in rvu_npc_exact_free_id()
510 dev_dbg(rvu->dev, "%s: freed id %d\n", __func__, seq_id); in rvu_npc_exact_free_id()
519 static bool rvu_npc_exact_alloc_id(struct rvu *rvu, u32 *seq_id) in rvu_npc_exact_alloc_id() argument
540 *seq_id = idx; in rvu_npc_exact_alloc_id()
541 dev_dbg(rvu->dev, "%s: Allocated id (%d)\n", __func__, *seq_id); in rvu_npc_exact_alloc_id()
748 __rvu_npc_exact_find_entry_by_seq_id(struct rvu *rvu, u32 seq_id) in __rvu_npc_exact_find_entry_by_seq_id() argument
758 if (entry->seq_id != seq_id) in __rvu_npc_exact_find_entry_by_seq_id()
786 u8 ctype, u32 *seq_id, bool cmd, u32 mcam_idx, u16 pcifunc) in rvu_npc_exact_add_to_list() argument
794 if (!rvu_npc_exact_alloc_id(rvu, seq_id)) { in rvu_npc_exact_add_to_list()
[all …]
H A Drvu_npc_hash.h152 u32 seq_id; /* Sequence number of entry */ member
/linux/drivers/net/ethernet/broadcom/bnxt/
H A Dbnxt_hwrm.c383 token->seq_id = bp->hwrm_cmd_seq++; in __hwrm_acquire_token()
386 token->seq_id = bp->hwrm_cmd_kong_seq++; in __hwrm_acquire_token()
405 hwrm_update_token(struct bnxt *bp, u16 seq_id, enum bnxt_hwrm_wait_state state) in hwrm_update_token() argument
411 if (token->seq_id == seq_id) { in hwrm_update_token()
418 netdev_err(bp->dev, "Invalid hwrm seq id %d\n", seq_id); in hwrm_update_token()
426 u32 seq = le16_to_cpu(req->seq_id); in hwrm_req_dbg()
511 ctx->req->seq_id = cpu_to_le16(token->seq_id); in __hwrm_send()
592 __le16 seen_out_of_seq = ctx->req->seq_id; /* will never see */ in __hwrm_send()
611 __le16 resp_seq = READ_ONCE(ctx->resp->seq_id); in __hwrm_send()
613 if (resp_seq == ctx->req->seq_id) in __hwrm_send()
[all …]
H A Dbnxt_hwrm.h54 u16 seq_id; member
/linux/drivers/net/ethernet/broadcom/bnge/
H A Dbnge_hwrm.c241 token->seq_id = bd->hwrm_cmd_seq++; in bnge_hwrm_create_token()
244 token->seq_id = bd->hwrm_cmd_kong_seq++; in bnge_hwrm_create_token()
267 u32 seq = le16_to_cpu(req->seq_id); in bnge_hwrm_req_dbg()
321 ctx->req->seq_id = cpu_to_le16(token->seq_id); in __hwrm_send_ctx()
375 __le16 seen_out_of_seq = ctx->req->seq_id; /* will never see */ in __hwrm_send_ctx()
388 __le16 resp_seq = READ_ONCE(ctx->resp->seq_id); in __hwrm_send_ctx()
390 if (resp_seq == ctx->req->seq_id) in __hwrm_send_ctx()
394 le16_to_cpu(resp_seq), req_type, le16_to_cpu(ctx->req->seq_id)); in __hwrm_send_ctx()
413 le16_to_cpu(ctx->req->seq_id), len); in __hwrm_send_ctx()
436 le16_to_cpu(ctx->req->seq_id), len, *valid); in __hwrm_send_ctx()
[all …]
/linux/drivers/gpu/drm/i915/display/
H A Dintel_dsi_vbt.c610 static const char *sequence_name(enum mipi_seq seq_id) in sequence_name() argument
612 if (seq_id < ARRAY_SIZE(seq_name)) in sequence_name()
613 return seq_name[seq_id]; in sequence_name()
619 enum mipi_seq seq_id) in intel_dsi_vbt_exec() argument
627 seq_id >= ARRAY_SIZE(connector->panel.vbt.dsi.sequence))) in intel_dsi_vbt_exec()
630 data = connector->panel.vbt.dsi.sequence[seq_id]; in intel_dsi_vbt_exec()
634 drm_WARN_ON(display->drm, *data != seq_id); in intel_dsi_vbt_exec()
637 seq_id, sequence_name(seq_id)); in intel_dsi_vbt_exec()
687 enum mipi_seq seq_id) in intel_dsi_vbt_exec_sequence() argument
689 if (seq_id == MIPI_SEQ_POWER_ON && intel_dsi->gpio_panel) in intel_dsi_vbt_exec_sequence()
[all …]
H A Dintel_dsi_vbt.h17 enum mipi_seq seq_id);
/linux/drivers/net/ethernet/huawei/hinic/
H A Dhinic_hw_mbox.c373 u8 seq_id, u8 seg_len) in check_mbox_seq_id_and_seg_len() argument
375 if (seq_id > SEQ_ID_MAX_VAL || seg_len > MBOX_SEG_LEN) in check_mbox_seq_id_and_seg_len()
378 if (seq_id == 0) { in check_mbox_seq_id_and_seg_len()
379 recv_mbox->seq_id = seq_id; in check_mbox_seq_id_and_seg_len()
381 if (seq_id != recv_mbox->seq_id + 1) in check_mbox_seq_id_and_seg_len()
384 recv_mbox->seq_id = seq_id; in check_mbox_seq_id_and_seg_len()
433 u8 seq_id, seg_len; in recv_mbox_handler() local
437 seq_id = HINIC_MBOX_HEADER_GET(mbox_header, SEQID); in recv_mbox_handler()
441 if (!check_mbox_seq_id_and_seg_len(recv_mbox, seq_id, seg_len)) { in recv_mbox_handler()
444 src_func_idx, recv_mbox->seq_id, seq_id, seg_len); in recv_mbox_handler()
[all …]
H A Dhinic_hw_mgmt.c492 int seq_id, seg_len; in recv_mgmt_msg_handler() local
495 seq_id = HINIC_MSG_HEADER_GET(*header, SEQID); in recv_mgmt_msg_handler()
498 if (seq_id >= (MAX_MSG_LEN / SEGMENT_LEN)) { in recv_mgmt_msg_handler()
504 memcpy(recv_msg->msg + seq_id * SEGMENT_LEN, msg_body, seg_len); in recv_mgmt_msg_handler()
H A Dhinic_hw_mbox.h52 u8 seq_id; member
/linux/net/bridge/
H A Dbr_private_mrp.h46 u32 seq_id; member
117 __be16 seq_id; member
/linux/drivers/net/ethernet/qlogic/qlcnic/
H A Dqlcnic_sriov.h32 u16 seq_id; member
41 u16 seq_id;
/linux/drivers/input/misc/
H A Dda7280.c221 u8 seq_id; member
632 haptics->gpi_ctl[num].seq_id = in da7280_haptics_upload_effect()
637 haptics->gpi_ctl[num].seq_id); in da7280_haptics_upload_effect()
850 haptics->gpi_ctl[i].seq_id = DA7280_GPI_SEQ_ID_DFT + i; in da7280_parse_properties()
853 haptics->gpi_ctl[i].seq_id = val; in da7280_parse_properties()
1102 haptics->gpi_ctl[i].seq_id) | in da7280_init()
/linux/drivers/scsi/bfa/
H A Dbfi_ms.h601 u8 seq_id; /* seq id of the msg */ member
607 u8 seq_id; /* seq id of the msg */ member
615 u8 seq_id; /* seq id of the msg */ member
/linux/drivers/iommu/intel/
H A Ddmar.c1053 iommu->seq_id = ida_alloc_range(&dmar_seq_ids, 0, in alloc_iommu()
1055 if (iommu->seq_id < 0) { in alloc_iommu()
1057 err = iommu->seq_id; in alloc_iommu()
1060 snprintf(iommu->name, sizeof(iommu->name), "dmar%d", iommu->seq_id); in alloc_iommu()
1079 iommu->seq_id); in alloc_iommu()
1087 iommu->seq_id); in alloc_iommu()
1121 pr_debug("Cannot alloc PMU for iommu (seq_id = %d)\n", iommu->seq_id); in alloc_iommu()
1162 ida_free(&dmar_seq_ids, iommu->seq_id); in alloc_iommu()
1199 ida_free(&dmar_seq_ids, iommu->seq_id); in free_iommu()
2025 irq = dmar_alloc_hwirq(iommu->seq_id, iommu->node, iommu); in dmar_set_interrupt()
H A Dirq_remapping.c131 pr_warn("IR%d: can't allocate an IRTE\n", iommu->seq_id); in alloc_irte()
544 pr_err("IR%d: failed to allocate 1M of pages\n", iommu->seq_id); in intel_setup_irq_remapping()
550 pr_err("IR%d: failed to allocate bitmap\n", iommu->seq_id); in intel_setup_irq_remapping()
554 info.fwnode = irq_domain_alloc_named_id_fwnode("INTEL-IR", iommu->seq_id); in intel_setup_irq_remapping()
560 pr_err("IR%d: failed to allocate irqdomain\n", iommu->seq_id); in intel_setup_irq_remapping()
913 scope->enumeration_id, drhd->address, iommu->seq_id); in ir_parse_one_ioapic_scope()
H A Dperfmon.c731 irq = dmar_alloc_hwirq(IOMMU_IRQ_ID_OFFSET_PERF + iommu->seq_id, iommu->node, iommu); in iommu_pmu_set_interrupt()
735 snprintf(iommu_pmu->irq_name, sizeof(iommu_pmu->irq_name), "dmar%d-perf", iommu->seq_id); in iommu_pmu_set_interrupt()
777 pr_err("Failed to register PMU for iommu (seq_id = %d)\n", iommu->seq_id); in iommu_pmu_register()
/linux/drivers/net/phy/
H A Dmicrochip_rds_ptp.c382 if (skb_sig != rx_ts->seq_id) in mchp_rds_ptp_match_skb()
433 if (skb_sig != rx_ts->seq_id) in mchp_rds_ptp_match_rx_skb()
933 u32 seconds, u32 nsec, u16 seq_id) in mchp_rds_ptp_match_tx_skb() argument
946 if (skb_sig != seq_id) in mchp_rds_ptp_match_tx_skb()
1008 rx_ts->seq_id = rc; in mchp_rds_ptp_get_rx_ts()
H A Dmicrel.c422 u16 seq_id; member
3096 u32 *seconds, u32 *nano_seconds, u16 *seq_id) in lan8814_ptp_rx_ts_get() argument
3110 *seq_id = lanphy_read_page_reg(phydev, LAN8814_PAGE_PORT_REGS, in lan8814_ptp_rx_ts_get()
3115 u32 *seconds, u32 *nano_seconds, u16 *seq_id) in lan8814_ptp_tx_ts_get() argument
3129 *seq_id = lanphy_read_page_reg(phydev, LAN8814_PAGE_PORT_REGS, in lan8814_ptp_tx_ts_get()
3338 if (memcmp(&skb_sig, &rx_ts->seq_id, sizeof(rx_ts->seq_id))) in lan8814_match_rx_skb()
4004 u32 seconds, u32 nsec, u16 seq_id) in lan8814_match_tx_skb() argument
4017 if (memcmp(&skb_sig, &seq_id, sizeof(seq_id))) in lan8814_match_tx_skb()
4037 u16 seq_id; in lan8814_dequeue_tx_skb() local
4039 lan8814_ptp_tx_ts_get(phydev, &seconds, &nsec, &seq_id); in lan8814_dequeue_tx_skb()
[all …]
/linux/drivers/scsi/qla2xxx/
H A Dqla_nvme.h132 uint8_t seq_id; member
H A Dqla_target.h145 __le16 seq_id; member
261 uint8_t seq_id; member
278 uint8_t seq_id; member
/linux/include/linux/qed/
H A Dfcoe_common.h85 __le16 seq_id; member
329 u8 seq_id; member
/linux/drivers/scsi/bnx2fc/
H A D57xx_hsi_bnx2fc.h178 u8 seq_id; member
617 u8 seq_id; member

12