| /linux/drivers/net/ethernet/sis/ |
| H A D | sis190.c | 270 u32 rx_buf_sz; member 461 static inline void sis190_give_to_asic(struct RxDesc *desc, u32 rx_buf_sz) in sis190_give_to_asic() argument 466 desc->size = cpu_to_le32((rx_buf_sz & RX_BUF_MASK) | eor); in sis190_give_to_asic() 472 u32 rx_buf_sz) in sis190_map_to_asic() argument 475 sis190_give_to_asic(desc, rx_buf_sz); in sis190_map_to_asic() 490 u32 rx_buf_sz = tp->rx_buf_sz; in sis190_alloc_rx_skb() local 494 skb = netdev_alloc_skb(tp->dev, rx_buf_sz); in sis190_alloc_rx_skb() 497 mapping = dma_map_single(&tp->pci_dev->dev, skb->data, tp->rx_buf_sz, in sis190_alloc_rx_skb() 501 sis190_map_to_asic(desc, mapping, rx_buf_sz); in sis190_alloc_rx_skb() 545 dma_sync_single_for_cpu(&tp->pci_dev->dev, addr, tp->rx_buf_sz, in sis190_try_rx_copy() [all …]
|
| /linux/drivers/net/ethernet/packetengines/ |
| H A D | hamachi.c | 496 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member 1127 hmp->rx_buf_sz, DMA_FROM_DEVICE); in hamachi_tx_timeout() 1136 skb = netdev_alloc_skb_ip_align(dev, hmp->rx_buf_sz); in hamachi_tx_timeout() 1143 hmp->rx_buf_sz, in hamachi_tx_timeout() 1146 DescEndPacket | DescIntr | (hmp->rx_buf_sz - 2)); in hamachi_tx_timeout() 1180 hmp->rx_buf_sz = (dev->mtu <= 1492 ? PKT_BUF_SZ : in hamachi_init_ring() 1190 struct sk_buff *skb = netdev_alloc_skb(dev, hmp->rx_buf_sz + 2); in hamachi_init_ring() 1197 hmp->rx_buf_sz, in hamachi_init_ring() 1201 DescEndPacket | DescIntr | (hmp->rx_buf_sz -2)); in hamachi_init_ring() 1431 hmp->rx_buf_sz, DMA_FROM_DEVICE); in hamachi_rx() [all …]
|
| H A D | yellowfin.c | 324 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member 735 yp->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in yellowfin_init_ring() 739 cpu_to_le32(CMD_RX_BUF | INTR_ALWAYS | yp->rx_buf_sz); in yellowfin_init_ring() 745 struct sk_buff *skb = netdev_alloc_skb(dev, yp->rx_buf_sz + 2); in yellowfin_init_ring() 752 yp->rx_buf_sz, in yellowfin_init_ring() 1072 yp->rx_buf_sz, DMA_FROM_DEVICE); in yellowfin_rx() 1139 yp->rx_buf_sz, in yellowfin_rx() 1151 yp->rx_buf_sz, in yellowfin_rx() 1166 struct sk_buff *skb = netdev_alloc_skb(dev, yp->rx_buf_sz + 2); in yellowfin_rx() 1173 yp->rx_buf_sz, in yellowfin_rx() [all …]
|
| /linux/drivers/net/ethernet/dlink/ |
| H A D | dl2k.c | 497 np->rx_buf_sz = (dev->mtu <= 1500 ? PACKET_SIZE : dev->mtu + 32); in alloc_list() 513 skb = netdev_alloc_skb_ip_align(dev, np->rx_buf_sz); in alloc_list() 519 np->rx_buf_sz, DMA_FROM_DEVICE); in alloc_list() 528 np->rx_ring[i].fraginfo |= cpu_to_le64((u64)np->rx_buf_sz << 48); in alloc_list() 690 np->rx_buf_sz); in rio_timer() 701 np->rx_buf_sz, DMA_FROM_DEVICE)); in rio_timer() 704 cpu_to_le64((u64)np->rx_buf_sz << 48); in rio_timer() 983 np->rx_buf_sz, in receive_packet() 990 np->rx_buf_sz, in receive_packet() 998 np->rx_buf_sz, in receive_packet() [all …]
|
| H A D | sundance.c | 389 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member 1023 np->rx_buf_sz = (dev->mtu <= 1520 ? PKT_BUF_SZ : dev->mtu + 16); in init_ring() 1039 netdev_alloc_skb(dev, np->rx_buf_sz + 2); in init_ring() 1045 np->rx_buf_sz, DMA_FROM_DEVICE); in init_ring() 1052 np->rx_ring[i].frag.length = cpu_to_le32(np->rx_buf_sz | LastFrag); in init_ring() 1379 np->rx_buf_sz, DMA_FROM_DEVICE); in rx_poll() 1383 np->rx_buf_sz, DMA_FROM_DEVICE); in rx_poll() 1388 np->rx_buf_sz, DMA_FROM_DEVICE); in rx_poll() 1429 skb = netdev_alloc_skb(dev, np->rx_buf_sz + 2); in refill_rx() 1435 np->rx_buf_sz, DMA_FROM_DEVICE); in refill_rx() [all …]
|
| /linux/drivers/net/ethernet/smsc/ |
| H A D | epic100.c | 271 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member 907 ep->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in epic_init_ring() 912 ep->rx_ring[i].buflength = ep->rx_buf_sz; in epic_init_ring() 922 struct sk_buff *skb = netdev_alloc_skb(dev, ep->rx_buf_sz + 2); in epic_init_ring() 929 ep->rx_buf_sz, in epic_init_ring() 1193 ep->rx_buf_sz, in epic_rx() 1199 ep->rx_buf_sz, in epic_rx() 1204 ep->rx_buf_sz, in epic_rx() 1223 skb = ep->rx_skbuff[entry] = netdev_alloc_skb(dev, ep->rx_buf_sz + 2); in epic_rx() 1229 ep->rx_buf_sz, in epic_rx() [all …]
|
| /linux/drivers/net/ethernet/realtek/ |
| H A D | 8139cp.c | 347 unsigned rx_buf_sz; member 413 cp->rx_buf_sz = mtu + ETH_HLEN + 8; in cp_set_rxbufsize() 415 cp->rx_buf_sz = PKT_BUF_SZ; in cp_set_rxbufsize() 477 const unsigned buflen = cp->rx_buf_sz; in cp_rx_poll() 546 cp->rx_buf_sz); in cp_rx_poll() 548 desc->opts1 = cpu_to_le32(DescOwn | cp->rx_buf_sz); in cp_rx_poll() 1067 skb = netdev_alloc_skb_ip_align(dev, cp->rx_buf_sz); in cp_refill_rx() 1072 cp->rx_buf_sz, DMA_FROM_DEVICE); in cp_refill_rx() 1083 cpu_to_le32(DescOwn | RingEnd | cp->rx_buf_sz); in cp_refill_rx() 1086 cpu_to_le32(DescOwn | cp->rx_buf_sz); in cp_refill_rx() [all …]
|
| /linux/drivers/net/ethernet/dec/tulip/ |
| H A D | de2104x.c | 304 unsigned rx_buf_sz; member 436 buflen = copying_skb ? (len + RX_OFFSET) : de->rx_buf_sz; in de_rx() 479 cpu_to_le32(RingEnd | de->rx_buf_sz); in de_rx() 481 de->rx_ring[rx_tail].opts2 = cpu_to_le32(de->rx_buf_sz); in de_rx() 1283 skb = netdev_alloc_skb(de->dev, de->rx_buf_sz); in de_refill_rx() 1289 de->rx_buf_sz, in de_refill_rx() 1296 cpu_to_le32(RingEnd | de->rx_buf_sz); in de_refill_rx() 1298 de->rx_ring[i].opts2 = cpu_to_le32(de->rx_buf_sz); in de_refill_rx() 1345 de->rx_skb[i].mapping, de->rx_buf_sz, in de_clean_rings() 1390 de->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in de_open()
|
| H A D | winbond-840.c | 303 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member 791 np->rx_ring[i].length = np->rx_buf_sz; in init_rxtx_rings() 800 struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz); in init_rxtx_rings() 805 np->rx_buf_sz, in init_rxtx_rings() 962 np->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in alloc_ringdesc() 1256 skb = netdev_alloc_skb(dev, np->rx_buf_sz); in netdev_rx() 1262 np->rx_buf_sz, in netdev_rx()
|
| /linux/drivers/net/ethernet/realtek/rtase/ |
| H A D | rtase_main.c | 245 static void rtase_mark_to_asic(union rtase_rx_desc *desc, u32 rx_buf_sz) in rtase_mark_to_asic() argument 253 cpu_to_le32(RTASE_DESC_OWN | eor | rx_buf_sz)); in rtase_mark_to_asic() 355 u32 rx_buf_sz) in rtase_map_to_asic() argument 359 rtase_mark_to_asic(desc, rx_buf_sz); in rtase_map_to_asic() 387 rtase_map_to_asic(desc, mapping, tp->rx_buf_sz); in rtase_alloc_rx_data_buf() 527 rtase_mark_to_asic(desc, tp->rx_buf_sz); in rx_handler() 542 rtase_mark_to_asic(desc, tp->rx_buf_sz); in rx_handler() 548 tp->rx_buf_sz, DMA_FROM_DEVICE); in rx_handler() 553 rtase_mark_to_asic(desc, tp->rx_buf_sz); in rx_handler() 983 rtase_w16(tp, RTASE_RMS, tp->rx_buf_sz); in rtase_hw_config() [all …]
|
| H A D | rtase.h | 327 u32 rx_buf_sz; member
|
| /linux/drivers/net/ethernet/adaptec/ |
| H A D | starfire.c | 545 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member 905 writel((np->rx_buf_sz << RxBufferLenShift) | in netdev_open() 1127 np->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in init_ring() 1131 struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz); in init_ring() 1137 np->rx_buf_sz, in init_ring() 1472 np->rx_buf_sz, DMA_FROM_DEVICE); in __netdev_rx() 1585 skb = netdev_alloc_skb(dev, np->rx_buf_sz); in refill_rx_ring() 1591 np->rx_buf_sz, DMA_FROM_DEVICE); in refill_rx_ring() 1967 np->rx_buf_sz, DMA_FROM_DEVICE); in netdev_close()
|
| /linux/drivers/net/ethernet/sun/ |
| H A D | sungem.h | 939 #define RX_BUF_ALLOC_SIZE(gp) ((gp)->rx_buf_sz + 28 + RX_OFFSET + 64) 994 int rx_buf_sz; member
|
| H A D | sungem.c | 816 skb_put(new_skb, (gp->rx_buf_sz + RX_OFFSET)); in gem_rx() 1616 gp->rx_buf_sz = max(dev->mtu + ETH_HLEN + VLAN_HLEN, in gem_init_rings() 1631 skb_put(skb, (gp->rx_buf_sz + RX_OFFSET)); in gem_init_rings() 1813 writel(0x20000000 | (gp->rx_buf_sz + 4), gp->regs + MAC_MAXFSZ); in gem_init_mac() 1894 int max_frame = (gp->rx_buf_sz + 4 + 64) & ~63; in gem_init_pause_thresholds()
|
| /linux/drivers/net/ethernet/via/ |
| H A D | via-rhine.c | 470 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member 1205 const int size = rp->rx_buf_sz; in rhine_skb_dma_init() 1249 rp->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in alloc_rbufs() 1255 rp->rx_ring[i].desc_length = cpu_to_le32(rp->rx_buf_sz); in alloc_rbufs() 1294 rp->rx_buf_sz, DMA_FROM_DEVICE); in free_rbufs() 2098 rp->rx_buf_sz, in rhine_rx() 2107 rp->rx_buf_sz, in rhine_rx() 2119 rp->rx_buf_sz, in rhine_rx()
|
| /linux/drivers/net/wireless/ath/ath11k/ |
| H A D | debugfs.c | 956 int i, ret, rx_buf_sz = 0; in ath11k_write_pktlog_filter() local 999 rx_buf_sz, &tlv_filter); in ath11k_write_pktlog_filter() 1020 rx_buf_sz = DP_RX_BUFFER_SIZE; in ath11k_write_pktlog_filter() 1030 rx_buf_sz = DP_RX_BUFFER_SIZE_LITE; in ath11k_write_pktlog_filter() 1032 rx_buf_sz = DP_RX_BUFFER_SIZE; in ath11k_write_pktlog_filter() 1059 rx_buf_sz, &tlv_filter); in ath11k_write_pktlog_filter()
|
| /linux/drivers/net/ethernet/natsemi/ |
| H A D | natsemi.c | 561 unsigned int rx_buf_sz; member 1745 if (np->rx_buf_sz > NATSEMI_LONGPKT) in init_registers() 1934 unsigned int buflen = np->rx_buf_sz+NATSEMI_PADDING; in refill_rx() 1949 np->rx_ring[entry].cmd_status = cpu_to_le32(np->rx_buf_sz); in refill_rx() 1962 np->rx_buf_sz = ETH_DATA_LEN + NATSEMI_HEADERS; in set_bufsize() 1964 np->rx_buf_sz = dev->mtu + NATSEMI_HEADERS; in set_bufsize() 2025 unsigned int buflen = np->rx_buf_sz; in drain_rx() 2283 unsigned int buflen = np->rx_buf_sz; in netdev_rx() 2343 } else if (pkt_len > np->rx_buf_sz) { in netdev_rx()
|
| /linux/drivers/atm/ |
| H A D | iphase.c | 768 iadev->rx_buf_sz = IA_RX_BUF_SZ; 779 iadev->rx_buf_sz = IA_RX_BUF_SZ; 791 iadev->rx_buf_sz = IA_RX_BUF_SZ; 796 iadev->rx_buf_sz, iadev->rx_pkt_ram);) 1173 if (len > iadev->rx_buf_sz) { 1174 printk("Over %d bytes sdu received, dropped!!!\n", iadev->rx_buf_sz); 1334 if ((length > iadev->rx_buf_sz) || (length > 1481 writew(iadev->rx_buf_sz, iadev->reass_reg+BUF_SIZE); 1495 rx_pkt_start += iadev->rx_buf_sz; 3145 iadev->num_rx_desc, iadev->rx_buf_sz,
|
| H A D | iphase.h | 1016 u16 num_rx_desc, rx_buf_sz, rxing; member
|
| /linux/drivers/net/ethernet/renesas/ |
| H A D | sh_eth.h | 547 u32 rx_buf_sz; /* Based on MTU+slack. */ member
|
| /linux/drivers/net/ethernet/nvidia/ |
| H A D | forcedeth.c | 805 unsigned int rx_buf_sz; member 1818 struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz + NV_RX_ALLOC_PAD); in nv_alloc_rx() 1833 np->put_rx.orig->flaglen = cpu_to_le32(np->rx_buf_sz | NV_RX_AVAIL); in nv_alloc_rx() 1859 struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz + NV_RX_ALLOC_PAD); in nv_alloc_rx_optimized() 1875 np->put_rx.ex->flaglen = cpu_to_le32(np->rx_buf_sz | NV_RX2_AVAIL); in nv_alloc_rx_optimized() 3072 np->rx_buf_sz = ETH_DATA_LEN + NV_RX_HEADERS; in set_bufsize() 3074 np->rx_buf_sz = dev->mtu + NV_RX_HEADERS; in set_bufsize() 3119 writel(np->rx_buf_sz, base + NvRegOffloadConfig); in nv_change_mtu() 4201 writel(np->rx_buf_sz, base + NvRegOffloadConfig); in nv_do_nic_poll() 4760 writel(np->rx_buf_sz, base + NvRegOffloadConfig); in nv_set_ringparam() [all …]
|
| /linux/drivers/net/ethernet/huawei/hinic/ |
| H A D | hinic_hw_dev.h | 264 u16 rx_buf_sz; member
|
| H A D | hinic_hw_dev.c | 248 fw_ctxt.rx_buf_sz = HINIC_RX_BUF_SZ; in init_fw_ctxt()
|
| /linux/drivers/iio/adc/ |
| H A D | stm32-dfsdm-adc.c | 859 unsigned int rx_buf_sz = DFSDM_DMA_BUFFER_SIZE; in stm32_dfsdm_set_watermark() local 868 adc->buf_sz = min(rx_buf_sz, watermark * 2 * adc->nconv); in stm32_dfsdm_set_watermark()
|
| /linux/drivers/net/usb/ |
| H A D | r8152.c | 944 u32 rx_buf_sz; member 2049 unsigned int order = get_order(tp->rx_buf_sz); in alloc_rx_agg() 2806 agg->buffer, tp->rx_buf_sz, in r8152_submit_rx() 3255 u32 ocp_data = tp->rx_buf_sz - rx_reserved_size(tp->netdev->mtu); in r8153_set_rx_early_size() 9468 tp->rx_buf_sz = 16 * 1024; in rtl_ops_init() 9490 tp->rx_buf_sz = 16 * 1024; in rtl_ops_init() 9492 tp->rx_buf_sz = 32 * 1024; in rtl_ops_init() 9511 tp->rx_buf_sz = 32 * 1024; in rtl_ops_init() 9533 tp->rx_buf_sz = 48 * 1024; in rtl_ops_init() 9556 tp->rx_buf_sz = 48 * 1024; in rtl_ops_init() [all …]
|