Lines Matching refs:gpd_current
430 struct qmu_gpd *gpd_current = NULL; in qmu_tx_zlp_error_handler() local
441 gpd_current = gpd_dma_to_virt(ring, cur_gpd_dma); in qmu_tx_zlp_error_handler()
443 if (GPD_DATA_LEN(mtu, le32_to_cpu(gpd_current->dw3_info)) != 0) { in qmu_tx_zlp_error_handler()
449 trace_mtu3_zlp_exp_gpd(mep, gpd_current); in qmu_tx_zlp_error_handler()
463 gpd_current->dw0_info |= cpu_to_le32(GPD_FLAGS_BPS | GPD_FLAGS_HWO); in qmu_tx_zlp_error_handler()
478 struct qmu_gpd *gpd_current = NULL; in qmu_error_rx() local
483 gpd_current = gpd_dma_to_virt(ring, cur_gpd_dma); in qmu_error_rx()
486 if (!mreq || mreq->gpd != gpd_current) { in qmu_error_rx()
494 gpd_current->dw0_info |= cpu_to_le32(GPD_FLAGS_BPS | GPD_FLAGS_HWO); in qmu_error_rx()
498 __func__, epnum, gpd_current, mreq); in qmu_error_rx()
514 struct qmu_gpd *gpd_current = NULL; in qmu_done_tx() local
521 gpd_current = gpd_dma_to_virt(ring, cur_gpd_dma); in qmu_done_tx()
524 __func__, epnum, gpd, gpd_current, ring->enqueue); in qmu_done_tx()
526 while (gpd && gpd != gpd_current && !GET_GPD_HWO(gpd)) { in qmu_done_tx()
554 struct qmu_gpd *gpd_current = NULL; in qmu_done_rx() local
560 gpd_current = gpd_dma_to_virt(ring, cur_gpd_dma); in qmu_done_rx()
563 __func__, epnum, gpd, gpd_current, ring->enqueue); in qmu_done_rx()
565 while (gpd && gpd != gpd_current && !GET_GPD_HWO(gpd)) { in qmu_done_rx()