// SPDX-License-Identifier: (GPL-2.0 OR MIT) // Copyright (c) 2017 Synopsys, Inc. and/or its affiliates. // stmmac Support for 5.xx Ethernet QoS cores #include #include #include "common.h" #include "dwmac4.h" #include "dwmac5.h" #include "stmmac.h" #include "stmmac_ptp.h" struct dwmac5_error_desc { bool valid; const char *desc; const char *detailed_desc; }; #define STAT_OFF(field) offsetof(struct stmmac_safety_stats, field) static void dwmac5_log_error(struct net_device *ndev, u32 value, bool corr, const char *module_name, const struct dwmac5_error_desc *desc, unsigned long field_offset, struct stmmac_safety_stats *stats) { unsigned long loc, mask; u8 *bptr = (u8 *)stats; unsigned long *ptr; ptr = (unsigned long *)(bptr + field_offset); mask = value; for_each_set_bit(loc, &mask, 32) { netdev_err(ndev, "Found %s error in %s: '%s: %s'\n", corr ? "correctable" : "uncorrectable", module_name, desc[loc].desc, desc[loc].detailed_desc); /* Update counters */ ptr[loc]++; } } static const struct dwmac5_error_desc dwmac5_mac_errors[32]= { { true, "ATPES", "Application Transmit Interface Parity Check Error" }, { true, "TPES", "TSO Data Path Parity Check Error" }, { true, "RDPES", "Read Descriptor Parity Check Error" }, { true, "MPES", "MTL Data Path Parity Check Error" }, { true, "MTSPES", "MTL TX Status Data Path Parity Check Error" }, { true, "ARPES", "Application Receive Interface Data Path Parity Check Error" }, { true, "CWPES", "CSR Write Data Path Parity Check Error" }, { true, "ASRPES", "AXI Slave Read Data Path Parity Check Error" }, { true, "TTES", "TX FSM Timeout Error" }, { true, "RTES", "RX FSM Timeout Error" }, { true, "CTES", "CSR FSM Timeout Error" }, { true, "ATES", "APP FSM Timeout Error" }, { true, "PTES", "PTP FSM Timeout Error" }, { true, "T125ES", "TX125 FSM Timeout Error" }, { true, "R125ES", "RX125 FSM Timeout Error" }, { true, "RVCTES", "REV MDC FSM Timeout Error" }, { true, "MSTTES", "Master Read/Write Timeout Error" }, { true, "SLVTES", "Slave Read/Write Timeout Error" }, { true, "ATITES", "Application Timeout on ATI Interface Error" }, { true, "ARITES", "Application Timeout on ARI Interface Error" }, { false, "UNKNOWN", "Unknown Error" }, /* 20 */ { false, "UNKNOWN", "Unknown Error" }, /* 21 */ { false, "UNKNOWN", "Unknown Error" }, /* 22 */ { false, "UNKNOWN", "Unknown Error" }, /* 23 */ { true, "FSMPES", "FSM State Parity Error" }, { false, "UNKNOWN", "Unknown Error" }, /* 25 */ { false, "UNKNOWN", "Unknown Error" }, /* 26 */ { false, "UNKNOWN", "Unknown Error" }, /* 27 */ { false, "UNKNOWN", "Unknown Error" }, /* 28 */ { false, "UNKNOWN", "Unknown Error" }, /* 29 */ { false, "UNKNOWN", "Unknown Error" }, /* 30 */ { false, "UNKNOWN", "Unknown Error" }, /* 31 */ }; static void dwmac5_handle_mac_err(struct net_device *ndev, void __iomem *ioaddr, bool correctable, struct stmmac_safety_stats *stats) { u32 value; value = readl(ioaddr + MAC_DPP_FSM_INT_STATUS); writel(value, ioaddr + MAC_DPP_FSM_INT_STATUS); dwmac5_log_error(ndev, value, correctable, "MAC", dwmac5_mac_errors, STAT_OFF(mac_errors), stats); } static const struct dwmac5_error_desc dwmac5_mtl_errors[32]= { { true, "TXCES", "MTL TX Memory Error" }, { true, "TXAMS", "MTL TX Memory Address Mismatch Error" }, { true, "TXUES", "MTL TX Memory Error" }, { false, "UNKNOWN", "Unknown Error" }, /* 3 */ { true, "RXCES", "MTL RX Memory Error" }, { true, "RXAMS", "MTL RX Memory Address Mismatch Error" }, { true, "RXUES", "MTL RX Memory Error" }, { false, "UNKNOWN", "Unknown Error" }, /* 7 */ { true, "ECES", "MTL EST Memory Error" }, { true, "EAMS", "MTL EST Memory Address Mismatch Error" }, { true, "EUES", "MTL EST Memory Error" }, { false, "UNKNOWN", "Unknown Error" }, /* 11 */ { true, "RPCES", "MTL RX Parser Memory Error" }, { true, "RPAMS", "MTL RX Parser Memory Address Mismatch Error" }, { true, "RPUES", "MTL RX Parser Memory Error" }, { false, "UNKNOWN", "Unknown Error" }, /* 15 */ { false, "UNKNOWN", "Unknown Error" }, /* 16 */ { false, "UNKNOWN", "Unknown Error" }, /* 17 */ { false, "UNKNOWN", "Unknown Error" }, /* 18 */ { false, "UNKNOWN", "Unknown Error" }, /* 19 */ { false, "UNKNOWN", "Unknown Error" }, /* 20 */ { false, "UNKNOWN", "Unknown Error" }, /* 21 */ { false, "UNKNOWN", "Unknown Error" }, /* 22 */ { false, "UNKNOWN", "Unknown Error" }, /* 23 */ { false, "UNKNOWN", "Unknown Error" }, /* 24 */ { false, "UNKNOWN", "Unknown Error" }, /* 25 */ { false, "UNKNOWN", "Unknown Error" }, /* 26 */ { false, "UNKNOWN", "Unknown Error" }, /* 27 */ { false, "UNKNOWN", "Unknown Error" }, /* 28 */ { false, "UNKNOWN", "Unknown Error" }, /* 29 */ { false, "UNKNOWN", "Unknown Error" }, /* 30 */ { false, "UNKNOWN", "Unknown Error" }, /* 31 */ }; static void dwmac5_handle_mtl_err(struct net_device *ndev, void __iomem *ioaddr, bool correctable, struct stmmac_safety_stats *stats) { u32 value; value = readl(ioaddr + MTL_ECC_INT_STATUS); writel(value, ioaddr + MTL_ECC_INT_STATUS); dwmac5_log_error(ndev, value, correctable, "MTL", dwmac5_mtl_errors, STAT_OFF(mtl_errors), stats); } static const struct dwmac5_error_desc dwmac5_dma_errors[32]= { { true, "TCES", "DMA TSO Memory Error" }, { true, "TAMS", "DMA TSO Memory Address Mismatch Error" }, { true, "TUES", "DMA TSO Memory Error" }, { false, "UNKNOWN", "Unknown Error" }, /* 3 */ { false, "UNKNOWN", "Unknown Error" }, /* 4 */ { false, "UNKNOWN", "Unknown Error" }, /* 5 */ { false, "UNKNOWN", "Unknown Error" }, /* 6 */ { false, "UNKNOWN", "Unknown Error" }, /* 7 */ { false, "UNKNOWN", "Unknown Error" }, /* 8 */ { false, "UNKNOWN", "Unknown Error" }, /* 9 */ { false, "UNKNOWN", "Unknown Error" }, /* 10 */ { false, "UNKNOWN", "Unknown Error" }, /* 11 */ { false, "UNKNOWN", "Unknown Error" }, /* 12 */ { false, "UNKNOWN", "Unknown Error" }, /* 13 */ { false, "UNKNOWN", "Unknown Error" }, /* 14 */ { false, "UNKNOWN", "Unknown Error" }, /* 15 */ { false, "UNKNOWN", "Unknown Error" }, /* 16 */ { false, "UNKNOWN", "Unknown Error" }, /* 17 */ { false, "UNKNOWN", "Unknown Error" }, /* 18 */ { false, "UNKNOWN", "Unknown Error" }, /* 19 */ { false, "UNKNOWN", "Unknown Error" }, /* 20 */ { false, "UNKNOWN", "Unknown Error" }, /* 21 */ { false, "UNKNOWN", "Unknown Error" }, /* 22 */ { false, "UNKNOWN", "Unknown Error" }, /* 23 */ { false, "UNKNOWN", "Unknown Error" }, /* 24 */ { false, "UNKNOWN", "Unknown Error" }, /* 25 */ { false, "UNKNOWN", "Unknown Error" }, /* 26 */ { false, "UNKNOWN", "Unknown Error" }, /* 27 */ { false, "UNKNOWN", "Unknown Error" }, /* 28 */ { false, "UNKNOWN", "Unknown Error" }, /* 29 */ { false, "UNKNOWN", "Unknown Error" }, /* 30 */ { false, "UNKNOWN", "Unknown Error" }, /* 31 */ }; static void dwmac5_handle_dma_err(struct net_device *ndev, void __iomem *ioaddr, bool correctable, struct stmmac_safety_stats *stats) { u32 value; value = readl(ioaddr + DMA_ECC_INT_STATUS); writel(value, ioaddr + DMA_ECC_INT_STATUS); dwmac5_log_error(ndev, value, correctable, "DMA", dwmac5_dma_errors, STAT_OFF(dma_errors), stats); } int dwmac5_safety_feat_config(void __iomem *ioaddr, unsigned int asp, struct stmmac_safety_feature_cfg *safety_feat_cfg) { struct stmmac_safety_feature_cfg all_safety_feats = { .tsoee = 1, .mrxpee = 1, .mestee = 1, .mrxee = 1, .mtxee = 1, .epsi = 1, .edpp = 1, .prtyen = 1, .tmouten = 1, }; u32 value; if (!asp) return -EINVAL; if (!safety_feat_cfg) safety_feat_cfg = &all_safety_feats; /* 1. Enable Safety Features */ value = readl(ioaddr + MTL_ECC_CONTROL); value |= MEEAO; /* MTL ECC Error Addr Status Override */ if (safety_feat_cfg->tsoee) value |= TSOEE; /* TSO ECC */ if (safety_feat_cfg->mrxpee) value |= MRXPEE; /* MTL RX Parser ECC */ if (safety_feat_cfg->mestee) value |= MESTEE; /* MTL EST ECC */ if (safety_feat_cfg->mrxee) value |= MRXEE; /* MTL RX FIFO ECC */ if (safety_feat_cfg->mtxee) value |= MTXEE; /* MTL TX FIFO ECC */ writel(value, ioaddr + MTL_ECC_CONTROL); /* 2. Enable MTL Safety Interrupts */ value = readl(ioaddr + MTL_ECC_INT_ENABLE); value |= RPCEIE; /* RX Parser Memory Correctable Error */ value |= ECEIE; /* EST Memory Correctable Error */ value |= RXCEIE; /* RX Memory Correctable Error */ value |= TXCEIE; /* TX Memory Correctable Error */ writel(value, ioaddr + MTL_ECC_INT_ENABLE); /* 3. Enable DMA Safety Interrupts */ value = readl(ioaddr + DMA_ECC_INT_ENABLE); value |= TCEIE; /* TSO Memory Correctable Error */ writel(value, ioaddr + DMA_ECC_INT_ENABLE); /* Only ECC Protection for External Memory feature is selected */ if (asp <= 0x1) return 0; /* 5. Enable Parity and Timeout for FSM */ value = readl(ioaddr + MAC_FSM_CONTROL); if (safety_feat_cfg->prtyen) value |= PRTYEN; /* FSM Parity Feature */ if (safety_feat_cfg->tmouten) value |= TMOUTEN; /* FSM Timeout Feature */ writel(value, ioaddr + MAC_FSM_CONTROL); /* 4. Enable Data Parity Protection */ value = readl(ioaddr + MTL_DPP_CONTROL); if (safety_feat_cfg->edpp) value |= EDPP; writel(value, ioaddr + MTL_DPP_CONTROL); /* * All the Automotive Safety features are selected without the "Parity * Port Enable for external interface" feature. */ if (asp <= 0x2) return 0; if (safety_feat_cfg->epsi) value |= EPSI; writel(value, ioaddr + MTL_DPP_CONTROL); return 0; } int dwmac5_safety_feat_irq_status(struct net_device *ndev, void __iomem *ioaddr, unsigned int asp, struct stmmac_safety_stats *stats) { bool err, corr; u32 mtl, dma; int ret = 0; if (!asp) return -EINVAL; mtl = readl(ioaddr + MTL_SAFETY_INT_STATUS); dma = readl(ioaddr + DMA_SAFETY_INT_STATUS); err = (mtl & MCSIS) || (dma & MCSIS); corr = false; if (err) { dwmac5_handle_mac_err(ndev, ioaddr, corr, stats); ret |= !corr; } err = (mtl & (MEUIS | MECIS)) || (dma & (MSUIS | MSCIS)); corr = (mtl & MECIS) || (dma & MSCIS); if (err) { dwmac5_handle_mtl_err(ndev, ioaddr, corr, stats); ret |= !corr; } err = dma & (DEUIS | DECIS); corr = dma & DECIS; if (err) { dwmac5_handle_dma_err(ndev, ioaddr, corr, stats); ret |= !corr; } return ret; } static const struct dwmac5_error { const struct dwmac5_error_desc *desc; } dwmac5_all_errors[] = { { dwmac5_mac_errors }, { dwmac5_mtl_errors }, { dwmac5_dma_errors }, }; int dwmac5_safety_feat_dump(struct stmmac_safety_stats *stats, int index, unsigned long *count, const char **desc) { int module = index / 32, offset = index % 32; unsigned long *ptr = (unsigned long *)stats; if (module >= ARRAY_SIZE(dwmac5_all_errors)) return -EINVAL; if (!dwmac5_all_errors[module].desc[offset].valid) return -EINVAL; if (count) *count = *(ptr + index); if (desc) *desc = dwmac5_all_errors[module].desc[offset].desc; return 0; } static int dwmac5_rxp_disable(void __iomem *ioaddr) { u32 val; val = readl(ioaddr + MTL_OPERATION_MODE); val &= ~MTL_FRPE; writel(val, ioaddr + MTL_OPERATION_MODE); return readl_poll_timeout(ioaddr + MTL_RXP_CONTROL_STATUS, val, val & RXPI, 1, 10000); } static void dwmac5_rxp_enable(void __iomem *ioaddr) { u32 val; val = readl(ioaddr + MTL_OPERATION_MODE); val |= MTL_FRPE; writel(val, ioaddr + MTL_OPERATION_MODE); } static int dwmac5_rxp_update_single_entry(void __iomem *ioaddr, struct stmmac_tc_entry *entry, int pos) { int ret, i; for (i = 0; i < (sizeof(entry->val) / sizeof(u32)); i++) { int real_pos = pos * (sizeof(entry->val) / sizeof(u32)) + i; u32 val; /* Wait for ready */ ret = readl_poll_timeout(ioaddr + MTL_RXP_IACC_CTRL_STATUS, val, !(val & STARTBUSY), 1, 10000); if (ret) return ret; /* Write data */ val = *((u32 *)&entry->val + i); writel(val, ioaddr + MTL_RXP_IACC_DATA); /* Write pos */ val = real_pos & ADDR; writel(val, ioaddr + MTL_RXP_IACC_CTRL_STATUS); /* Write OP */ val |= WRRDN; writel(val, ioaddr + MTL_RXP_IACC_CTRL_STATUS); /* Start Write */ val |= STARTBUSY; writel(val, ioaddr + MTL_RXP_IACC_CTRL_STATUS); /* Wait for done */ ret = readl_poll_timeout(ioaddr + MTL_RXP_IACC_CTRL_STATUS, val, !(val & STARTBUSY), 1, 10000); if (ret) return ret; } return 0; } static struct stmmac_tc_entry * dwmac5_rxp_get_next_entry(struct stmmac_tc_entry *entries, unsigned int count, u32 curr_prio) { struct stmmac_tc_entry *entry; u32 min_prio = ~0x0; int i, min_prio_idx; bool found = false; for (i = count - 1; i >= 0; i--) { entry = &entries[i]; /* Do not update unused entries */ if (!entry->in_use) continue; /* Do not update already updated entries (i.e. fragments) */ if (entry->in_hw) continue; /* Let last entry be updated last */ if (entry->is_last) continue; /* Do not return fragments */ if (entry->is_frag) continue; /* Check if we already checked this prio */ if (entry->prio < curr_prio) continue; /* Check if this is the minimum prio */ if (entry->prio < min_prio) { min_prio = entry->prio; min_prio_idx = i; found = true; } } if (found) return &entries[min_prio_idx]; return NULL; } int dwmac5_rxp_config(void __iomem *ioaddr, struct stmmac_tc_entry *entries, unsigned int count) { struct stmmac_tc_entry *entry, *frag; int i, ret, nve = 0; u32 curr_prio = 0; u32 old_val, val; /* Force disable RX */ old_val = readl(ioaddr + GMAC_CONFIG); val = old_val & ~GMAC_CONFIG_RE; writel(val, ioaddr + GMAC_CONFIG); /* Disable RX Parser */ ret = dwmac5_rxp_disable(ioaddr); if (ret) goto re_enable; /* Set all entries as NOT in HW */ for (i = 0; i < count; i++) { entry = &entries[i]; entry->in_hw = false; } /* Update entries by reverse order */ while (1) { entry = dwmac5_rxp_get_next_entry(entries, count, curr_prio); if (!entry) break; curr_prio = entry->prio; frag = entry->frag_ptr; /* Set special fragment requirements */ if (frag) { entry->val.af = 0; entry->val.rf = 0; entry->val.nc = 1; entry->val.ok_index = nve + 2; } ret = dwmac5_rxp_update_single_entry(ioaddr, entry, nve); if (ret) goto re_enable; entry->table_pos = nve++; entry->in_hw = true; if (frag && !frag->in_hw) { ret = dwmac5_rxp_update_single_entry(ioaddr, frag, nve); if (ret) goto re_enable; frag->table_pos = nve++; frag->in_hw = true; } } if (!nve) goto re_enable; /* Update all pass entry */ for (i = 0; i < count; i++) { entry = &entries[i]; if (!entry->is_last) continue; ret = dwmac5_rxp_update_single_entry(ioaddr, entry, nve); if (ret) goto re_enable; entry->table_pos = nve++; } /* Assume n. of parsable entries == n. of valid entries */ val = (nve << 16) & NPE; val |= nve & NVE; writel(val, ioaddr + MTL_RXP_CONTROL_STATUS); /* Enable RX Parser */ dwmac5_rxp_enable(ioaddr); re_enable: /* Re-enable RX */ writel(old_val, ioaddr + GMAC_CONFIG); return ret; } int dwmac5_flex_pps_config(void __iomem *ioaddr, int index, struct stmmac_pps_cfg *cfg, bool enable, u32 sub_second_inc, u32 systime_flags) { u32 tnsec = readl(ioaddr + MAC_PPSx_TARGET_TIME_NSEC(index)); u32 val = readl(ioaddr + MAC_PPS_CONTROL); u64 period; if (!cfg->available) return -EINVAL; if (tnsec & TRGTBUSY0) return -EBUSY; if (!sub_second_inc || !systime_flags) return -EINVAL; val &= ~PPSx_MASK(index); if (!enable) { val |= PPSCMDx(index, 0x5); val |= PPSEN0; writel(val, ioaddr + MAC_PPS_CONTROL); return 0; } val |= TRGTMODSELx(index, 0x2); val |= PPSEN0; writel(val, ioaddr + MAC_PPS_CONTROL); writel(cfg->start.tv_sec, ioaddr + MAC_PPSx_TARGET_TIME_SEC(index)); if (!(systime_flags & PTP_TCR_TSCTRLSSR)) cfg->start.tv_nsec = (cfg->start.tv_nsec * 1000) / 465; writel(cfg->start.tv_nsec, ioaddr + MAC_PPSx_TARGET_TIME_NSEC(index)); period = cfg->period.tv_sec * 1000000000; period += cfg->period.tv_nsec; do_div(period, sub_second_inc); if (period <= 1) return -EINVAL; writel(period - 1, ioaddr + MAC_PPSx_INTERVAL(index)); period >>= 1; if (period <= 1) return -EINVAL; writel(period - 1, ioaddr + MAC_PPSx_WIDTH(index)); /* Finally, activate it */ val |= PPSCMDx(index, 0x2); writel(val, ioaddr + MAC_PPS_CONTROL); return 0; } void dwmac5_fpe_configure(void __iomem *ioaddr, struct stmmac_fpe_cfg *cfg, u32 num_txq, u32 num_rxq, bool tx_enable, bool pmac_enable) { u32 value; if (tx_enable) { cfg->fpe_csr = EFPE; value = readl(ioaddr + GMAC_RXQ_CTRL1); value &= ~GMAC_RXQCTRL_FPRQ; value |= (num_rxq - 1) << GMAC_RXQCTRL_FPRQ_SHIFT; writel(value, ioaddr + GMAC_RXQ_CTRL1); } else { cfg->fpe_csr = 0; } writel(cfg->fpe_csr, ioaddr + MAC_FPE_CTRL_STS); value = readl(ioaddr + GMAC_INT_EN); if (pmac_enable) { if (!(value & GMAC_INT_FPE_EN)) { /* Dummy read to clear any pending masked interrupts */ readl(ioaddr + MAC_FPE_CTRL_STS); value |= GMAC_INT_FPE_EN; } } else { value &= ~GMAC_INT_FPE_EN; } writel(value, ioaddr + GMAC_INT_EN); } int dwmac5_fpe_irq_status(void __iomem *ioaddr, struct net_device *dev) { u32 value; int status; status = FPE_EVENT_UNKNOWN; /* Reads from the MAC_FPE_CTRL_STS register should only be performed * here, since the status flags of MAC_FPE_CTRL_STS are "clear on read" */ value = readl(ioaddr + MAC_FPE_CTRL_STS); if (value & TRSP) { status |= FPE_EVENT_TRSP; netdev_dbg(dev, "FPE: Respond mPacket is transmitted\n"); } if (value & TVER) { status |= FPE_EVENT_TVER; netdev_dbg(dev, "FPE: Verify mPacket is transmitted\n"); } if (value & RRSP) { status |= FPE_EVENT_RRSP; netdev_dbg(dev, "FPE: Respond mPacket is received\n"); } if (value & RVER) { status |= FPE_EVENT_RVER; netdev_dbg(dev, "FPE: Verify mPacket is received\n"); } return status; } void dwmac5_fpe_send_mpacket(void __iomem *ioaddr, struct stmmac_fpe_cfg *cfg, enum stmmac_mpacket_type type) { u32 value = cfg->fpe_csr; if (type == MPACKET_VERIFY) value |= SVER; else if (type == MPACKET_RESPONSE) value |= SRSP; writel(value, ioaddr + MAC_FPE_CTRL_STS); } int dwmac5_fpe_get_add_frag_size(const void __iomem *ioaddr) { return FIELD_GET(DWMAC5_ADD_FRAG_SZ, readl(ioaddr + MTL_FPE_CTRL_STS)); } void dwmac5_fpe_set_add_frag_size(void __iomem *ioaddr, u32 add_frag_size) { u32 value; value = readl(ioaddr + MTL_FPE_CTRL_STS); writel(u32_replace_bits(value, add_frag_size, DWMAC5_ADD_FRAG_SZ), ioaddr + MTL_FPE_CTRL_STS); } #define ALG_ERR_MSG "TX algorithm SP is not suitable for one-to-many mapping" #define WEIGHT_ERR_MSG "TXQ weight %u differs across other TXQs in TC: [%u]" int dwmac5_fpe_map_preemption_class(struct net_device *ndev, struct netlink_ext_ack *extack, u32 pclass) { u32 val, offset, count, queue_weight, preemptible_txqs = 0; struct stmmac_priv *priv = netdev_priv(ndev); u32 num_tc = ndev->num_tc; if (!pclass) goto update_mapping; /* DWMAC CORE4+ can not program TC:TXQ mapping to hardware. * * Synopsys Databook: * "The number of Tx DMA channels is equal to the number of Tx queues, * and is direct one-to-one mapping." */ for (u32 tc = 0; tc < num_tc; tc++) { count = ndev->tc_to_txq[tc].count; offset = ndev->tc_to_txq[tc].offset; if (pclass & BIT(tc)) preemptible_txqs |= GENMASK(offset + count - 1, offset); /* This is 1:1 mapping, go to next TC */ if (count == 1) continue; if (priv->plat->tx_sched_algorithm == MTL_TX_ALGORITHM_SP) { NL_SET_ERR_MSG_MOD(extack, ALG_ERR_MSG); return -EINVAL; } queue_weight = priv->plat->tx_queues_cfg[offset].weight; for (u32 i = 1; i < count; i++) { if (priv->plat->tx_queues_cfg[offset + i].weight != queue_weight) { NL_SET_ERR_MSG_FMT_MOD(extack, WEIGHT_ERR_MSG, queue_weight, tc); return -EINVAL; } } } update_mapping: val = readl(priv->ioaddr + MTL_FPE_CTRL_STS); writel(u32_replace_bits(val, preemptible_txqs, DWMAC5_PREEMPTION_CLASS), priv->ioaddr + MTL_FPE_CTRL_STS); return 0; }