| /linux/drivers/hsi/ |
| H A D | hsi_core.c | 63 kfree(cl->rx_cfg.channels); in hsi_client_release() 86 cl->rx_cfg = info->rx_cfg; in hsi_new_client() 87 if (cl->rx_cfg.channels) { in hsi_new_client() 88 size = cl->rx_cfg.num_channels * sizeof(*cl->rx_cfg.channels); in hsi_new_client() 89 cl->rx_cfg.channels = kmemdup(info->rx_cfg.channels, size, in hsi_new_client() 91 if (!cl->rx_cfg.channels) in hsi_new_client() 217 &cl->rx_cfg.mode); in hsi_add_client_from_dt() 226 cl->rx_cfg.mode = mode; in hsi_add_client_from_dt() 234 cl->rx_cfg.speed = cl->tx_cfg.speed; in hsi_add_client_from_dt() 237 &cl->rx_cfg.flow); in hsi_add_client_from_dt() [all …]
|
| /linux/drivers/net/ethernet/google/gve/ |
| H A D | gve_main.c | 118 for (ring = 0; ring < priv->rx_cfg.num_queues; ring++) { in gve_get_stats() 288 priv->rx_cfg.max_queues; in gve_alloc_stats_report() 511 priv->rx_cfg.max_queues = min_t(int, priv->rx_cfg.max_queues, in gve_alloc_notify_blocks() 516 priv->rx_cfg.max_queues); in gve_alloc_notify_blocks() 519 if (priv->rx_cfg.num_queues > priv->rx_cfg.max_queues) in gve_alloc_notify_blocks() 520 priv->rx_cfg.num_queues = priv->rx_cfg.max_queues; in gve_alloc_notify_blocks() 690 err = gve_init_rss_config(priv, priv->rx_cfg.num_queues); in gve_setup_device_resources() 844 num_rx_qpls = gve_num_rx_qpls(&priv->rx_cfg, gve_is_qpl(priv)); in gve_register_qpls() 868 num_rx_qpls = gve_num_rx_qpls(&priv->rx_cfg, gve_is_qpl(priv)); in gve_unregister_qpls() 904 err = gve_adminq_create_rx_queues(priv, priv->rx_cfg.num_queues); in gve_create_rings() [all …]
|
| H A D | gve_ethtool.c | 105 for (i = 0; i < priv->rx_cfg.num_queues; i++) in gve_get_strings() 139 (priv->rx_cfg.num_queues * NUM_GVE_RX_CNTS) + in gve_get_sset_count() 179 rx_qid_to_stats_idx = kmalloc_objs(int, priv->rx_cfg.num_queues); in gve_get_ethtool_stats() 182 for (ring = 0; ring < priv->rx_cfg.num_queues; ring++) { in gve_get_ethtool_stats() 203 ring < priv->rx_cfg.num_queues; ring++) { in gve_get_ethtool_stats() 279 nic_stats_len = (NIC_RX_STATS_REPORT_NUM * priv->rx_cfg.num_queues + in gve_get_ethtool_stats() 297 (priv->rx_cfg.num_queues - num_stopped_rxqs) + in gve_get_ethtool_stats() 315 if (queue_id < 0 || queue_id >= priv->rx_cfg.num_queues) { in gve_get_ethtool_stats() 323 for (ring = 0; ring < priv->rx_cfg.num_queues; ring++) { in gve_get_ethtool_stats() 382 i += priv->rx_cfg.num_queues * NUM_GVE_RX_CNTS; in gve_get_ethtool_stats() [all …]
|
| H A D | gve.h | 827 struct gve_rx_queue_config rx_cfg; member 1119 static inline u32 gve_num_rx_qpls(const struct gve_rx_queue_config *rx_cfg, in gve_num_rx_qpls() argument 1124 return rx_cfg->num_queues; in gve_num_rx_qpls()
|
| H A D | gve_flow_rule.c | 126 if (fsp->ring_cookie >= priv->rx_cfg.num_queues) in gve_generate_flow_rule()
|
| /linux/drivers/net/ethernet/apple/ |
| H A D | bmac.c | 840 unsigned short rx_cfg; in bmac_rx_off() local 842 rx_cfg = bmread(dev, RXCFG); in bmac_rx_off() 843 rx_cfg &= ~RxMACEnable; in bmac_rx_off() 844 bmwrite(dev, RXCFG, rx_cfg); in bmac_rx_off() 846 rx_cfg = bmread(dev, RXCFG); in bmac_rx_off() 847 } while (rx_cfg & RxMACEnable); in bmac_rx_off() 853 unsigned short rx_cfg; in bmac_rx_on() local 855 rx_cfg = bmread(dev, RXCFG); in bmac_rx_on() 856 rx_cfg |= RxMACEnable; in bmac_rx_on() 857 if (hash_enable) rx_cfg |= RxHashFilterEnable; in bmac_rx_on() [all …]
|
| /linux/drivers/net/wireless/intel/iwlwifi/pcie/ |
| H A D | ctxt-info.c | 168 struct iwl_context_info_rbd_cfg *rx_cfg; in iwl_pcie_ctxt_info_init() local 215 rx_cfg = &ctxt_info->rbd_cfg; in iwl_pcie_ctxt_info_init() 216 rx_cfg->free_rbd_addr = cpu_to_le64(trans_pcie->rxq->bd_dma); in iwl_pcie_ctxt_info_init() 217 rx_cfg->used_rbd_addr = cpu_to_le64(trans_pcie->rxq->used_bd_dma); in iwl_pcie_ctxt_info_init() 218 rx_cfg->status_wr_ptr = cpu_to_le64(trans_pcie->rxq->rb_stts_dma); in iwl_pcie_ctxt_info_init()
|
| /linux/drivers/hsi/clients/ |
| H A D | hsi_char.c | 355 tmp = cl->rx_cfg; in hsc_rx_set() 356 cl->rx_cfg.mode = rxc->mode; in hsc_rx_set() 357 cl->rx_cfg.num_hw_channels = rxc->channels; in hsc_rx_set() 358 cl->rx_cfg.flow = rxc->flow; in hsc_rx_set() 361 cl->rx_cfg = tmp; in hsc_rx_set() 372 rxc->mode = cl->rx_cfg.mode; in hsc_rx_get() 373 rxc->channels = cl->rx_cfg.num_hw_channels; in hsc_rx_get() 374 rxc->flow = cl->rx_cfg.flow; in hsc_rx_get() 425 if (channel->ch >= channel->cl->rx_cfg.num_hw_channels) in hsc_read()
|
| H A D | nokia-modem.c | 177 ssip.rx_cfg = cl->rx_cfg; in nokia_modem_probe() 200 cmtspeech.rx_cfg = cl->rx_cfg; in nokia_modem_probe()
|
| /linux/drivers/net/ethernet/brocade/bna/ |
| H A D | bna_tx_rx.c | 1625 cfg_req->rx_cfg.frame_size = bna_enet_mtu_get(&rx->bna->enet); in bna_bfi_rx_enet_start() 1651 cfg_req->rx_cfg.multi_buffer = in bna_bfi_rx_enet_start() 1690 cfg_req->rx_cfg.rxq_type = BFI_ENET_RXQ_LARGE_SMALL; in bna_bfi_rx_enet_start() 1694 cfg_req->rx_cfg.rxq_type = BFI_ENET_RXQ_HDS; in bna_bfi_rx_enet_start() 1695 cfg_req->rx_cfg.hds.type = rx->hds_cfg.hdr_type; in bna_bfi_rx_enet_start() 1696 cfg_req->rx_cfg.hds.force_offset = rx->hds_cfg.forced_offset; in bna_bfi_rx_enet_start() 1697 cfg_req->rx_cfg.hds.max_header_size = rx->hds_cfg.forced_offset; in bna_bfi_rx_enet_start() 1701 cfg_req->rx_cfg.rxq_type = BFI_ENET_RXQ_SINGLE; in bna_bfi_rx_enet_start() 1707 cfg_req->rx_cfg.strip_vlan = rx->rxf.vlan_strip_status; in bna_bfi_rx_enet_start() 1741 bna_rx_res_check(struct bna_rx_mod *rx_mod, struct bna_rx_config *rx_cfg) in bna_rx_res_check() argument [all …]
|
| H A D | bna.h | 346 struct bna_rx_config *rx_cfg,
|
| H A D | bfi_enet.h | 501 struct bfi_enet_rx_cfg rx_cfg; member
|
| /linux/include/linux/hsi/ |
| H A D | hsi.h | 104 struct hsi_config rx_cfg; member 129 struct hsi_config rx_cfg; member 224 struct hsi_config rx_cfg; member
|
| /linux/drivers/net/ethernet/micrel/ |
| H A D | ksz884x.c | 1208 u32 rx_cfg; member 2838 u32 rx_cfg; in set_flow_ctrl() local 2841 rx_cfg = hw->rx_cfg; in set_flow_ctrl() 2844 hw->rx_cfg |= DMA_RX_FLOW_ENABLE; in set_flow_ctrl() 2846 hw->rx_cfg &= ~DMA_RX_FLOW_ENABLE; in set_flow_ctrl() 2852 if (rx_cfg != hw->rx_cfg) in set_flow_ctrl() 2853 writel(hw->rx_cfg, hw->io + KS_DMA_RX_CTRL); in set_flow_ctrl() 3406 hw->rx_cfg = (DMA_RX_BROADCAST | DMA_RX_UNICAST | in hw_setup() 3408 hw->rx_cfg |= KS884X_DMA_RX_MULTICAST; in hw_setup() 3411 hw->rx_cfg |= (DMA_RX_CSUM_TCP | DMA_RX_CSUM_IP); in hw_setup() [all …]
|
| /linux/drivers/net/ethernet/ti/icssg/ |
| H A D | icssg_common.c | 496 struct k3_udma_glue_rx_channel_cfg rx_cfg; in prueth_init_rx_chns() local 513 memset(&rx_cfg, 0, sizeof(rx_cfg)); in prueth_init_rx_chns() 514 rx_cfg.swdata_size = PRUETH_NAV_SW_DATA_SIZE; in prueth_init_rx_chns() 515 rx_cfg.flow_id_num = max_rflows; in prueth_init_rx_chns() 516 rx_cfg.flow_id_base = -1; /* udmax will auto select flow id base */ in prueth_init_rx_chns() 523 &rx_cfg); in prueth_init_rx_chns() 561 for (i = 0; i < rx_cfg.flow_id_num; i++) { in prueth_init_rx_chns() 572 .rx_cfg = rxring_cfg, in prueth_init_rx_chns() 580 rx_flow_cfg.rx_cfg in prueth_init_rx_chns() [all...] |
| /linux/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
| H A D | hclge_debugfs.c | 2498 u8 rx_cfg; in hclge_get_vlan_rx_offload_cfg() local 2516 rx_cfg = req->vport_vlan_cfg; in hclge_get_vlan_rx_offload_cfg() 2517 vlan_cfg->strip_tag1 = hnae3_get_bit(rx_cfg, HCLGE_REM_TAG1_EN_B); in hclge_get_vlan_rx_offload_cfg() 2518 vlan_cfg->strip_tag2 = hnae3_get_bit(rx_cfg, HCLGE_REM_TAG2_EN_B); in hclge_get_vlan_rx_offload_cfg() 2519 vlan_cfg->drop_tag1 = hnae3_get_bit(rx_cfg, HCLGE_DISCARD_TAG1_EN_B); in hclge_get_vlan_rx_offload_cfg() 2520 vlan_cfg->drop_tag2 = hnae3_get_bit(rx_cfg, HCLGE_DISCARD_TAG2_EN_B); in hclge_get_vlan_rx_offload_cfg() 2521 vlan_cfg->pri_only1 = hnae3_get_bit(rx_cfg, HCLGE_SHOW_TAG1_EN_B); in hclge_get_vlan_rx_offload_cfg() 2522 vlan_cfg->pri_only2 = hnae3_get_bit(rx_cfg, HCLGE_SHOW_TAG2_EN_B); in hclge_get_vlan_rx_offload_cfg()
|
| /linux/drivers/net/ethernet/chelsio/cxgb3/ |
| H A D | t3_hw.c | 1133 static void t3_gate_rx_traffic(struct cmac *mac, u32 *rx_cfg, in t3_gate_rx_traffic() argument 1140 *rx_cfg = t3_read_reg(mac->adapter, A_XGM_RX_CFG); in t3_gate_rx_traffic() 1155 static void t3_open_rx_traffic(struct cmac *mac, u32 rx_cfg, in t3_open_rx_traffic() argument 1161 rx_cfg); in t3_open_rx_traffic() 1186 u32 rx_cfg, rx_hash_high, rx_hash_low; in t3_link_changed() local 1190 t3_gate_rx_traffic(mac, &rx_cfg, &rx_hash_high, &rx_hash_low); in t3_link_changed() 1199 t3_open_rx_traffic(mac, rx_cfg, rx_hash_high, rx_hash_low); in t3_link_changed() 1239 u32 rx_cfg, rx_hash_high, rx_hash_low; in t3_link_fault() local 1241 t3_gate_rx_traffic(mac, &rx_cfg, &rx_hash_high, &rx_hash_low); in t3_link_fault() 1249 t3_open_rx_traffic(mac, rx_cfg, rx_hash_high, rx_hash_low); in t3_link_fault()
|
| /linux/drivers/net/ethernet/ti/ |
| H A D | am65-cpsw-nuss.c | 2378 struct k3_udma_glue_rx_channel_cfg rx_cfg = { 0 }; in am65_cpsw_nuss_init_rx_chns() local 2389 rx_cfg.swdata_size = AM65_CPSW_NAV_SW_DATA_SIZE; in am65_cpsw_nuss_init_rx_chns() 2390 rx_cfg.flow_id_num = common->rx_ch_num_flows; in am65_cpsw_nuss_init_rx_chns() 2391 rx_cfg.flow_id_base = common->rx_flow_id_base; in am65_cpsw_nuss_init_rx_chns() 2395 rx_chn->descs_num = max_desc_num * rx_cfg.flow_id_num; in am65_cpsw_nuss_init_rx_chns() 2402 rx_chn->rx_chn = k3_udma_glue_request_rx_chn(dev, "rx", &rx_cfg); in am65_cpsw_nuss_init_rx_chns() 2428 for (i = 0; i < rx_cfg.flow_id_num; i++) { in am65_cpsw_nuss_init_rx_chns() 2439 .rx_cfg = rxring_cfg, in am65_cpsw_nuss_init_rx_chns() 2452 rx_flow_cfg.rx_cfg.size = max_desc_num; in am65_cpsw_nuss_init_rx_chns() 2454 rx_flow_cfg.rxfdq_cfg.size = max_desc_num * rx_cfg.flow_id_num; in am65_cpsw_nuss_init_rx_chns()
|
| /linux/drivers/usb/gadget/udc/ |
| H A D | bcm63xx_udc.c | 467 const struct iudma_ch_cfg *rx_cfg = &iudma_defaults[i]; in bcm63xx_fifo_setup() local 473 ((rx_fifo_slot + rx_cfg->n_fifo_slots - 1) << in bcm63xx_fifo_setup() 475 rx_fifo_slot += rx_cfg->n_fifo_slots; in bcm63xx_fifo_setup() 478 is_hs ? rx_cfg->max_pkt_hs : rx_cfg->max_pkt_fs, in bcm63xx_fifo_setup()
|
| /linux/drivers/net/ethernet/sun/ |
| H A D | niu.c | 412 u32 tx_cfg, rx_cfg; in serdes_init_niu_10g_fiber() local 416 rx_cfg = (PLL_RX_CFG_ENRX | PLL_RX_CFG_TERM_0P8VDDT | in serdes_init_niu_10g_fiber() 427 rx_cfg |= PLL_RX_CFG_ENTEST; in serdes_init_niu_10g_fiber() 438 int err = esr2_set_rx_cfg(np, i, rx_cfg); in serdes_init_niu_10g_fiber() 452 u32 tx_cfg, rx_cfg; in serdes_init_niu_1g_serdes() local 458 rx_cfg = (PLL_RX_CFG_ENRX | PLL_RX_CFG_TERM_0P8VDDT | in serdes_init_niu_1g_serdes() 463 rx_cfg |= PLL_RX_CFG_EQ_LP_ADAPTIVE; in serdes_init_niu_1g_serdes() 472 rx_cfg |= PLL_RX_CFG_ENTEST; in serdes_init_niu_1g_serdes() 506 err = esr2_set_rx_cfg(np, i, rx_cfg); in serdes_init_niu_1g_serdes() 546 u32 tx_cfg, rx_cfg, pll_cfg, pll_sts; in serdes_init_niu_10g_serdes() local [all …]
|
| /linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
| H A D | otx2_common.c | 1247 nixlf->rx_cfg = BIT_ULL(33) | BIT_ULL(35) | BIT_ULL(37); in otx2_config_nix()
|
| /linux/drivers/net/ethernet/marvell/octeontx2/af/ |
| H A D | rvu_nix.c | 1671 rvu_write64(rvu, blkaddr, NIX_AF_LFX_RX_CFG(nixlf), req->rx_cfg); in rvu_mbox_handler_nix_lf_alloc()
|