Lines Matching refs:cqe

1024         struct oce_nic_tx_cqe *cqe;  in oce_drain_wq_cq()  local
1031 cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_nic_tx_cqe); in oce_drain_wq_cq()
1032 if (cqe->u0.dw[3] == 0) in oce_drain_wq_cq()
1034 cqe->u0.dw[3] = 0; in oce_drain_wq_cq()
1067 struct oce_nic_rx_cqe *cqe; in oce_drain_rq_cq() local
1074 cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_nic_rx_cqe); in oce_drain_rq_cq()
1076 while (RQ_CQE_VALID(cqe)) { in oce_drain_rq_cq()
1077 RQ_CQE_INVALIDATE(cqe); in oce_drain_rq_cq()
1079 cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, in oce_drain_rq_cq()
1113 struct nic_hwlro_singleton_cqe *cqe; in oce_rx_cq_clean_hwlro() local
1121 cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, struct nic_hwlro_singleton_cqe); in oce_rx_cq_clean_hwlro()
1122 if(cqe->valid) { in oce_rx_cq_clean_hwlro()
1123 if(cqe->cqe_type == 0) { /* singleton cqe */ in oce_rx_cq_clean_hwlro()
1129 num_frags = cqe->pkt_size / rq->cfg.frag_size; in oce_rx_cq_clean_hwlro()
1130 if(cqe->pkt_size % rq->cfg.frag_size) in oce_rx_cq_clean_hwlro()
1134 if(!cqe->pkt_size) in oce_rx_cq_clean_hwlro()
1136 cqe->valid = 0; in oce_rx_cq_clean_hwlro()
1138 }else if(cqe->cqe_type == 0x1) { /* first part */ in oce_rx_cq_clean_hwlro()
1144 rq->cqe_firstpart = (struct nic_hwlro_cqe_part1 *)cqe; in oce_rx_cq_clean_hwlro()
1146 }else if(cqe->cqe_type == 0x2) { /* second part */ in oce_rx_cq_clean_hwlro()
1147 cqe2 = (struct nic_hwlro_cqe_part2 *)cqe; in oce_rx_cq_clean_hwlro()
1188 struct oce_nic_rx_cqe *cqe; in oce_rx_cq_clean() local
1199 cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_nic_rx_cqe); in oce_rx_cq_clean()
1200 if(RQ_CQE_VALID(cqe)) { in oce_rx_cq_clean()
1201 DW_SWAP((uint32_t *) cqe, sizeof(oce_rq_cqe)); in oce_rx_cq_clean()
1202 oce_discard_rx_comp(rq, cqe->u0.s.num_fragments); in oce_rx_cq_clean()
1204 … if((cqe->u0.s.num_fragments==0)&&(cqe->u0.s.pkt_size == 0)&&(cqe->u0.s.error == 0)) in oce_rx_cq_clean()
1207 RQ_CQE_INVALIDATE(cqe); in oce_rx_cq_clean()