Lines Matching refs:rxcmp1
2001 struct rx_cmp_ext *rxcmp1, u32 *cmpl_ts) in bnxt_rx_ts_valid() argument
2003 u32 ts = le32_to_cpu(rxcmp1->rx_cmp_timestamp); in bnxt_rx_ts_valid()
2017 struct rx_cmp_ext *rxcmp1) in bnxt_rx_vlan() argument
2023 __le32 flags2 = rxcmp1->rx_cmp_flags2; in bnxt_rx_vlan()
2029 meta_data = le32_to_cpu(rxcmp1->rx_cmp_meta_data); in bnxt_rx_vlan()
2046 vtag = RX_CMP_METADATA0_TCI(rxcmp1); in bnxt_rx_vlan()
2088 struct rx_cmp_ext *rxcmp1; in bnxt_rx_pkt() local
2116 rxcmp1 = (struct rx_cmp_ext *) in bnxt_rx_pkt()
2119 if (!RX_CMP_VALID(rxcmp1, tmp_raw_cons)) in bnxt_rx_pkt()
2132 (struct rx_tpa_start_cmp_ext *)rxcmp1); in bnxt_rx_pkt()
2140 (struct rx_tpa_end_cmp_ext *)rxcmp1, event); in bnxt_rx_pkt()
2185 if (rxcmp1->rx_cmp_cfa_code_errors_v2 & RX_CMP_L2_ERRORS) { in bnxt_rx_pkt()
2186 u32 rx_err = le32_to_cpu(rxcmp1->rx_cmp_cfa_code_errors_v2); in bnxt_rx_pkt()
2301 dev = bnxt_get_pkt_dev(bp, RX_CMP_CFA_CODE(rxcmp1)); in bnxt_rx_pkt()
2305 skb = bnxt_rx_vlan(skb, cmp_type, rxcmp, rxcmp1); in bnxt_rx_pkt()
2311 if (RX_CMP_L4_CS_OK(rxcmp1)) { in bnxt_rx_pkt()
2314 skb->csum_level = RX_CMP_ENCAP(rxcmp1); in bnxt_rx_pkt()
2317 if (rxcmp1->rx_cmp_cfa_code_errors_v2 & RX_CMP_L4_CS_ERR_BITS) { in bnxt_rx_pkt()
2323 if (bnxt_rx_ts_valid(bp, flags, rxcmp1, &cmpl_ts)) { in bnxt_rx_pkt()
2367 struct rx_cmp_ext *rxcmp1; in bnxt_force_rx_discard() local
2379 rxcmp1 = (struct rx_cmp_ext *) in bnxt_force_rx_discard()
2382 if (!RX_CMP_VALID(rxcmp1, tmp_raw_cons)) in bnxt_force_rx_discard()
2392 rxcmp1->rx_cmp_cfa_code_errors_v2 |= in bnxt_force_rx_discard()
2397 tpa_end1 = (struct rx_tpa_end_cmp_ext *)rxcmp1; in bnxt_force_rx_discard()
3143 struct rx_cmp_ext *rxcmp1; in bnxt_poll_nitroa0() local
3166 rxcmp1 = (struct rx_cmp_ext *) in bnxt_poll_nitroa0()
3169 if (!RX_CMP_VALID(rxcmp1, tmp_raw_cons)) in bnxt_poll_nitroa0()
3173 rxcmp1->rx_cmp_cfa_code_errors_v2 |= in bnxt_poll_nitroa0()