/* * CDDL HEADER START * * The contents of this file are subject to the terms of the * Common Development and Distribution License (the "License"). * You may not use this file except in compliance with the License. * * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE * or http://www.opensolaris.org/os/licensing. * See the License for the specific language governing permissions * and limitations under the License. * * When distributing Covered Code, include this CDDL HEADER in each * file and include the License file at usr/src/OPENSOLARIS.LICENSE. * If applicable, add the following below this CDDL HEADER, with the * fields enclosed by brackets "[]" replaced with your own identifying * information: Portions Copyright [yyyy] [name of copyright owner] * * CDDL HEADER END */ /* * Copyright(c) 2007-2010 Intel Corporation. All rights reserved. */ /* * Copyright (c) 2008, 2010, Oracle and/or its affiliates. All rights reserved. * Copyright 2017 Joyent, Inc. */ #include "ixgbe_sw.h" /* function prototypes */ static mblk_t *ixgbe_rx_bind(ixgbe_rx_data_t *, uint32_t, uint32_t); static mblk_t *ixgbe_rx_copy(ixgbe_rx_data_t *, uint32_t, uint32_t); static void ixgbe_rx_assoc_hcksum(mblk_t *, uint32_t); static mblk_t *ixgbe_lro_bind(ixgbe_rx_data_t *, uint32_t, uint32_t, uint32_t); static mblk_t *ixgbe_lro_copy(ixgbe_rx_data_t *, uint32_t, uint32_t, uint32_t); static int ixgbe_lro_get_start(ixgbe_rx_data_t *, uint32_t); static uint32_t ixgbe_lro_get_first(ixgbe_rx_data_t *, uint32_t); #ifndef IXGBE_DEBUG #pragma inline(ixgbe_rx_assoc_hcksum) #pragma inline(ixgbe_lro_get_start) #pragma inline(ixgbe_lro_get_first) #endif /* * ixgbe_rx_recycle - The call-back function to reclaim rx buffer. * * This function is called when an mp is freed by the user thru * freeb call (Only for mp constructed through desballoc call). * It returns back the freed buffer to the free list. */ void ixgbe_rx_recycle(caddr_t arg) { ixgbe_t *ixgbe; ixgbe_rx_ring_t *rx_ring; ixgbe_rx_data_t *rx_data; rx_control_block_t *recycle_rcb; uint32_t free_index; uint32_t ref_cnt; recycle_rcb = (rx_control_block_t *)(uintptr_t)arg; rx_data = recycle_rcb->rx_data; rx_ring = rx_data->rx_ring; ixgbe = rx_ring->ixgbe; if (recycle_rcb->ref_cnt == 0) { /* * This case only happens when rx buffers are being freed * in ixgbe_stop() and freemsg() is called. */ return; } ASSERT(recycle_rcb->mp == NULL); /* * Using the recycled data buffer to generate a new mblk */ recycle_rcb->mp = desballoc((unsigned char *) recycle_rcb->rx_buf.address, recycle_rcb->rx_buf.size, 0, &recycle_rcb->free_rtn); /* * Put the recycled rx control block into free list */ mutex_enter(&rx_data->recycle_lock); free_index = rx_data->rcb_tail; ASSERT(rx_data->free_list[free_index] == NULL); rx_data->free_list[free_index] = recycle_rcb; rx_data->rcb_tail = NEXT_INDEX(free_index, 1, rx_data->free_list_size); mutex_exit(&rx_data->recycle_lock); /* * The atomic operation on the number of the available rx control * blocks in the free list is used to make the recycling mutual * exclusive with the receiving. */ atomic_inc_32(&rx_data->rcb_free); ASSERT(rx_data->rcb_free <= rx_data->free_list_size); /* * Considering the case that the interface is unplumbed * and there are still some buffers held by the upper layer. * When the buffer is returned back, we need to free it. */ ref_cnt = atomic_dec_32_nv(&recycle_rcb->ref_cnt); if (ref_cnt == 0) { if (recycle_rcb->mp != NULL) { freemsg(recycle_rcb->mp); recycle_rcb->mp = NULL; } ixgbe_free_dma_buffer(&recycle_rcb->rx_buf); mutex_enter(&ixgbe->rx_pending_lock); atomic_dec_32(&rx_data->rcb_pending); atomic_dec_32(&ixgbe->rcb_pending); /* * When there is not any buffer belonging to this rx_data * held by the upper layer, the rx_data can be freed. */ if ((rx_data->flag & IXGBE_RX_STOPPED) && (rx_data->rcb_pending == 0)) ixgbe_free_rx_ring_data(rx_data); mutex_exit(&ixgbe->rx_pending_lock); } } /* * ixgbe_rx_copy - Use copy to process the received packet. * * This function will use bcopy to process the packet * and send the copied packet upstream. */ static mblk_t * ixgbe_rx_copy(ixgbe_rx_data_t *rx_data, uint32_t index, uint32_t pkt_len) { ixgbe_t *ixgbe; rx_control_block_t *current_rcb; mblk_t *mp; ixgbe = rx_data->rx_ring->ixgbe; current_rcb = rx_data->work_list[index]; DMA_SYNC(¤t_rcb->rx_buf, DDI_DMA_SYNC_FORKERNEL); if (ixgbe_check_dma_handle(current_rcb->rx_buf.dma_handle) != DDI_FM_OK) { ddi_fm_service_impact(ixgbe->dip, DDI_SERVICE_DEGRADED); atomic_or_32(&ixgbe->ixgbe_state, IXGBE_ERROR); return (NULL); } /* * Allocate buffer to receive this packet */ mp = allocb(pkt_len + IPHDR_ALIGN_ROOM, 0); if (mp == NULL) { ixgbe_log(ixgbe, "ixgbe_rx_copy: allocate buffer failed"); return (NULL); } /* * Copy the data received into the new cluster */ mp->b_rptr += IPHDR_ALIGN_ROOM; bcopy(current_rcb->rx_buf.address, mp->b_rptr, pkt_len); mp->b_wptr = mp->b_rptr + pkt_len; return (mp); } /* * ixgbe_rx_bind - Use existing DMA buffer to build mblk for receiving. * * This function will use pre-bound DMA buffer to receive the packet * and build mblk that will be sent upstream. */ static mblk_t * ixgbe_rx_bind(ixgbe_rx_data_t *rx_data, uint32_t index, uint32_t pkt_len) { rx_control_block_t *current_rcb; rx_control_block_t *free_rcb; uint32_t free_index; mblk_t *mp; ixgbe_t *ixgbe = rx_data->rx_ring->ixgbe; /* * If the free list is empty, we cannot proceed to send * the current DMA buffer upstream. We'll have to return * and use bcopy to process the packet. */ if (ixgbe_atomic_reserve(&rx_data->rcb_free, 1) < 0) return (NULL); current_rcb = rx_data->work_list[index]; /* * If the mp of the rx control block is NULL, try to do * desballoc again. */ if (current_rcb->mp == NULL) { current_rcb->mp = desballoc((unsigned char *) current_rcb->rx_buf.address, current_rcb->rx_buf.size, 0, ¤t_rcb->free_rtn); /* * If it is failed to built a mblk using the current * DMA buffer, we have to return and use bcopy to * process the packet. */ if (current_rcb->mp == NULL) { atomic_inc_32(&rx_data->rcb_free); return (NULL); } } /* * Sync up the data received */ DMA_SYNC(¤t_rcb->rx_buf, DDI_DMA_SYNC_FORKERNEL); if (ixgbe_check_dma_handle(current_rcb->rx_buf.dma_handle) != DDI_FM_OK) { ddi_fm_service_impact(ixgbe->dip, DDI_SERVICE_DEGRADED); atomic_inc_32(&rx_data->rcb_free); atomic_or_32(&ixgbe->ixgbe_state, IXGBE_ERROR); return (NULL); } mp = current_rcb->mp; current_rcb->mp = NULL; atomic_inc_32(¤t_rcb->ref_cnt); mp->b_wptr = mp->b_rptr + pkt_len; mp->b_next = mp->b_cont = NULL; /* * Strip off one free rx control block from the free list */ free_index = rx_data->rcb_head; free_rcb = rx_data->free_list[free_index]; ASSERT(free_rcb != NULL); rx_data->free_list[free_index] = NULL; rx_data->rcb_head = NEXT_INDEX(free_index, 1, rx_data->free_list_size); /* * Put the rx control block to the work list */ rx_data->work_list[index] = free_rcb; return (mp); } /* * ixgbe_lro_bind - Use existing DMA buffer to build LRO mblk for receiving. * * This function will use pre-bound DMA buffers to receive the packet * and build LRO mblk that will be sent upstream. */ static mblk_t * ixgbe_lro_bind(ixgbe_rx_data_t *rx_data, uint32_t lro_start, uint32_t lro_num, uint32_t pkt_len) { rx_control_block_t *current_rcb; union ixgbe_adv_rx_desc *current_rbd; rx_control_block_t *free_rcb; uint32_t free_index; int lro_next; uint32_t last_pkt_len; uint32_t i; mblk_t *mp; mblk_t *mblk_head; mblk_t **mblk_tail; ixgbe_t *ixgbe = rx_data->rx_ring->ixgbe; /* * If the free list is empty, we cannot proceed to send * the current DMA buffer upstream. We'll have to return * and use bcopy to process the packet. */ if (ixgbe_atomic_reserve(&rx_data->rcb_free, lro_num) < 0) return (NULL); current_rcb = rx_data->work_list[lro_start]; /* * If any one of the rx data blocks can not support * lro bind operation, We'll have to return and use * bcopy to process the lro packet. */ for (i = lro_num; i > 0; i--) { /* * Sync up the data received */ DMA_SYNC(¤t_rcb->rx_buf, DDI_DMA_SYNC_FORKERNEL); if (ixgbe_check_dma_handle(current_rcb->rx_buf.dma_handle) != DDI_FM_OK) { ddi_fm_service_impact(ixgbe->dip, DDI_SERVICE_DEGRADED); atomic_add_32(&rx_data->rcb_free, lro_num); atomic_or_32(&ixgbe->ixgbe_state, IXGBE_ERROR); return (NULL); } /* * If the mp of the rx control block is NULL, try to do * desballoc again. */ if (current_rcb->mp == NULL) { current_rcb->mp = desballoc((unsigned char *) current_rcb->rx_buf.address, current_rcb->rx_buf.size, 0, ¤t_rcb->free_rtn); /* * If it is failed to built a mblk using the current * DMA buffer, we have to return and use bcopy to * process the packet. */ if (current_rcb->mp == NULL) { atomic_add_32(&rx_data->rcb_free, lro_num); return (NULL); } } if (current_rcb->lro_next != -1) lro_next = current_rcb->lro_next; current_rcb = rx_data->work_list[lro_next]; } mblk_head = NULL; mblk_tail = &mblk_head; lro_next = lro_start; last_pkt_len = pkt_len - ixgbe->rx_buf_size * (lro_num - 1); current_rcb = rx_data->work_list[lro_next]; current_rbd = &rx_data->rbd_ring[lro_next]; while (lro_num --) { mp = current_rcb->mp; current_rcb->mp = NULL; atomic_inc_32(¤t_rcb->ref_cnt); if (lro_num != 0) mp->b_wptr = mp->b_rptr + ixgbe->rx_buf_size; else mp->b_wptr = mp->b_rptr + last_pkt_len; mp->b_next = mp->b_cont = NULL; *mblk_tail = mp; mblk_tail = &mp->b_cont; /* * Strip off one free rx control block from the free list */ free_index = rx_data->rcb_head; free_rcb = rx_data->free_list[free_index]; ASSERT(free_rcb != NULL); rx_data->free_list[free_index] = NULL; rx_data->rcb_head = NEXT_INDEX(free_index, 1, rx_data->free_list_size); /* * Put the rx control block to the work list */ rx_data->work_list[lro_next] = free_rcb; lro_next = current_rcb->lro_next; current_rcb->lro_next = -1; current_rcb->lro_prev = -1; current_rcb->lro_pkt = B_FALSE; current_rbd->read.pkt_addr = free_rcb->rx_buf.dma_address; current_rbd->read.hdr_addr = 0; if (lro_next == -1) break; current_rcb = rx_data->work_list[lro_next]; current_rbd = &rx_data->rbd_ring[lro_next]; } return (mblk_head); } /* * ixgbe_lro_copy - Use copy to process the received LRO packet. * * This function will use bcopy to process the LRO packet * and send the copied packet upstream. */ static mblk_t * ixgbe_lro_copy(ixgbe_rx_data_t *rx_data, uint32_t lro_start, uint32_t lro_num, uint32_t pkt_len) { ixgbe_t *ixgbe; rx_control_block_t *current_rcb; union ixgbe_adv_rx_desc *current_rbd; mblk_t *mp; uint32_t last_pkt_len; int lro_next; uint32_t i; ixgbe = rx_data->rx_ring->ixgbe; /* * Allocate buffer to receive this LRO packet */ mp = allocb(pkt_len + IPHDR_ALIGN_ROOM, 0); if (mp == NULL) { ixgbe_log(ixgbe, "LRO copy MP alloc failed"); return (NULL); } current_rcb = rx_data->work_list[lro_start]; /* * Sync up the LRO packet data received */ for (i = lro_num; i > 0; i--) { DMA_SYNC(¤t_rcb->rx_buf, DDI_DMA_SYNC_FORKERNEL); if (ixgbe_check_dma_handle(current_rcb->rx_buf.dma_handle) != DDI_FM_OK) { ddi_fm_service_impact(ixgbe->dip, DDI_SERVICE_DEGRADED); atomic_or_32(&ixgbe->ixgbe_state, IXGBE_ERROR); return (NULL); } if (current_rcb->lro_next != -1) lro_next = current_rcb->lro_next; current_rcb = rx_data->work_list[lro_next]; } lro_next = lro_start; current_rcb = rx_data->work_list[lro_next]; current_rbd = &rx_data->rbd_ring[lro_next]; last_pkt_len = pkt_len - ixgbe->rx_buf_size * (lro_num - 1); /* * Copy the data received into the new cluster */ mp->b_rptr += IPHDR_ALIGN_ROOM; mp->b_wptr += IPHDR_ALIGN_ROOM; while (lro_num --) { if (lro_num != 0) { bcopy(current_rcb->rx_buf.address, mp->b_wptr, ixgbe->rx_buf_size); mp->b_wptr += ixgbe->rx_buf_size; } else { bcopy(current_rcb->rx_buf.address, mp->b_wptr, last_pkt_len); mp->b_wptr += last_pkt_len; } lro_next = current_rcb->lro_next; current_rcb->lro_next = -1; current_rcb->lro_prev = -1; current_rcb->lro_pkt = B_FALSE; current_rbd->read.pkt_addr = current_rcb->rx_buf.dma_address; current_rbd->read.hdr_addr = 0; if (lro_next == -1) break; current_rcb = rx_data->work_list[lro_next]; current_rbd = &rx_data->rbd_ring[lro_next]; } return (mp); } /* * ixgbe_lro_get_start - get the start rcb index in one LRO packet */ static int ixgbe_lro_get_start(ixgbe_rx_data_t *rx_data, uint32_t rx_next) { int lro_prev; int lro_start; uint32_t lro_num = 1; rx_control_block_t *prev_rcb; rx_control_block_t *current_rcb = rx_data->work_list[rx_next]; lro_prev = current_rcb->lro_prev; while (lro_prev != -1) { lro_num ++; prev_rcb = rx_data->work_list[lro_prev]; lro_start = lro_prev; lro_prev = prev_rcb->lro_prev; } rx_data->lro_num = lro_num; return (lro_start); } /* * ixgbe_lro_get_first - get the first LRO rcb index */ static uint32_t ixgbe_lro_get_first(ixgbe_rx_data_t *rx_data, uint32_t rx_next) { rx_control_block_t *current_rcb; uint32_t lro_first; lro_first = rx_data->lro_first; current_rcb = rx_data->work_list[lro_first]; while ((!current_rcb->lro_pkt) && (lro_first != rx_next)) { lro_first = NEXT_INDEX(lro_first, 1, rx_data->ring_size); current_rcb = rx_data->work_list[lro_first]; } rx_data->lro_first = lro_first; return (lro_first); } /* * ixgbe_rx_assoc_hcksum - Check the rx hardware checksum status and associate * the hcksum flags. */ static void ixgbe_rx_assoc_hcksum(mblk_t *mp, uint32_t status_error) { uint32_t hcksum_flags = 0; /* * Check TCP/UDP checksum */ if ((status_error & IXGBE_RXD_STAT_L4CS) && !(status_error & IXGBE_RXDADV_ERR_TCPE)) hcksum_flags |= HCK_FULLCKSUM_OK; /* * Check IP Checksum */ if ((status_error & IXGBE_RXD_STAT_IPCS) && !(status_error & IXGBE_RXDADV_ERR_IPE)) hcksum_flags |= HCK_IPV4_HDRCKSUM_OK; if (hcksum_flags != 0) { mac_hcksum_set(mp, 0, 0, 0, 0, hcksum_flags); } } /* * ixgbe_ring_rx - Receive the data of one ring. * * This function goes throught h/w descriptor in one specified rx ring, * receives the data if the descriptor status shows the data is ready. * It returns a chain of mblks containing the received data, to be * passed up to mac_rx(). */ mblk_t * ixgbe_ring_rx(ixgbe_rx_ring_t *rx_ring, int poll_bytes) { union ixgbe_adv_rx_desc *current_rbd; rx_control_block_t *current_rcb; mblk_t *mp; mblk_t *mblk_head; mblk_t **mblk_tail; uint32_t rx_next; uint32_t rx_tail; uint32_t pkt_len; uint32_t status_error; uint32_t pkt_num; uint32_t rsc_cnt; uint32_t lro_first; uint32_t lro_start; uint32_t lro_next; boolean_t lro_eop; uint32_t received_bytes; ixgbe_t *ixgbe = rx_ring->ixgbe; ixgbe_rx_data_t *rx_data; if ((ixgbe->ixgbe_state & IXGBE_SUSPENDED) || (ixgbe->ixgbe_state & IXGBE_ERROR) || (ixgbe->ixgbe_state & IXGBE_OVERTEMP) || !(ixgbe->ixgbe_state & IXGBE_STARTED)) return (NULL); rx_data = rx_ring->rx_data; lro_eop = B_FALSE; mblk_head = NULL; mblk_tail = &mblk_head; /* * Sync the receive descriptors before accepting the packets */ DMA_SYNC(&rx_data->rbd_area, DDI_DMA_SYNC_FORKERNEL); if (ixgbe_check_dma_handle(rx_data->rbd_area.dma_handle) != DDI_FM_OK) { ddi_fm_service_impact(ixgbe->dip, DDI_SERVICE_DEGRADED); atomic_or_32(&ixgbe->ixgbe_state, IXGBE_ERROR); return (NULL); } /* * Get the start point of rx bd ring which should be examined * during this cycle. */ rx_next = rx_data->rbd_next; current_rbd = &rx_data->rbd_ring[rx_next]; received_bytes = 0; pkt_num = 0; status_error = current_rbd->wb.upper.status_error; while (status_error & IXGBE_RXD_STAT_DD) { /* * If adapter has found errors, but the error * is hardware checksum error, this does not discard the * packet: let upper layer compute the checksum; * Otherwise discard the packet. */ if ((status_error & IXGBE_RXDADV_ERR_FRAME_ERR_MASK) || ((!ixgbe->lro_enable) && (!(status_error & IXGBE_RXD_STAT_EOP)))) { rx_ring->stat_frame_error++; goto rx_discard; } if ((status_error & IXGBE_RXDADV_ERR_TCPE) || (status_error & IXGBE_RXDADV_ERR_IPE)) rx_ring->stat_cksum_error++; if (ixgbe->lro_enable) { rsc_cnt = (current_rbd->wb.lower.lo_dword.data & IXGBE_RXDADV_RSCCNT_MASK) >> IXGBE_RXDADV_RSCCNT_SHIFT; if (rsc_cnt != 0) { if (status_error & IXGBE_RXD_STAT_EOP) { pkt_len = current_rbd->wb.upper.length; if (rx_data->work_list[rx_next]-> lro_prev != -1) { lro_start = ixgbe_lro_get_start(rx_data, rx_next); ixgbe->lro_pkt_count++; pkt_len += (rx_data->lro_num - 1) * ixgbe->rx_buf_size; lro_eop = B_TRUE; } } else { lro_next = (status_error & IXGBE_RXDADV_NEXTP_MASK) >> IXGBE_RXDADV_NEXTP_SHIFT; rx_data->work_list[lro_next]->lro_prev = rx_next; rx_data->work_list[rx_next]->lro_next = lro_next; rx_data->work_list[rx_next]->lro_pkt = B_TRUE; goto rx_discard; } } else { pkt_len = current_rbd->wb.upper.length; } } else { pkt_len = current_rbd->wb.upper.length; } if ((poll_bytes != IXGBE_POLL_NULL) && ((received_bytes + pkt_len) > poll_bytes)) break; received_bytes += pkt_len; mp = NULL; /* * For packets with length more than the copy threshold, * we'll first try to use the existing DMA buffer to build * an mblk and send the mblk upstream. * * If the first method fails, or the packet length is less * than the copy threshold, we'll allocate a new mblk and * copy the packet data to the new mblk. */ if (lro_eop) { mp = ixgbe_lro_bind(rx_data, lro_start, rx_data->lro_num, pkt_len); if (mp == NULL) mp = ixgbe_lro_copy(rx_data, lro_start, rx_data->lro_num, pkt_len); lro_eop = B_FALSE; rx_data->lro_num = 0; } else { if (pkt_len > ixgbe->rx_copy_thresh) mp = ixgbe_rx_bind(rx_data, rx_next, pkt_len); if (mp == NULL) mp = ixgbe_rx_copy(rx_data, rx_next, pkt_len); } if (mp != NULL) { /* * Check h/w checksum offload status */ if (ixgbe->rx_hcksum_enable) ixgbe_rx_assoc_hcksum(mp, status_error); *mblk_tail = mp; mblk_tail = &mp->b_next; } rx_discard: /* * Reset rx descriptor read bits */ current_rcb = rx_data->work_list[rx_next]; if (ixgbe->lro_enable) { if (!current_rcb->lro_pkt) { current_rbd->read.pkt_addr = current_rcb->rx_buf.dma_address; current_rbd->read.hdr_addr = 0; } } else { current_rbd->read.pkt_addr = current_rcb->rx_buf.dma_address; current_rbd->read.hdr_addr = 0; } rx_next = NEXT_INDEX(rx_next, 1, rx_data->ring_size); /* * The receive function is in interrupt context, so here * rx_limit_per_intr is used to avoid doing receiving too long * per interrupt. */ if (++pkt_num > ixgbe->rx_limit_per_intr) { rx_ring->stat_exceed_pkt++; break; } current_rbd = &rx_data->rbd_ring[rx_next]; status_error = current_rbd->wb.upper.status_error; } rx_ring->stat_rbytes += received_bytes; rx_ring->stat_ipackets += pkt_num; DMA_SYNC(&rx_data->rbd_area, DDI_DMA_SYNC_FORDEV); rx_data->rbd_next = rx_next; /* * Update the h/w tail accordingly */ if (ixgbe->lro_enable) { lro_first = ixgbe_lro_get_first(rx_data, rx_next); rx_tail = PREV_INDEX(lro_first, 1, rx_data->ring_size); } else rx_tail = PREV_INDEX(rx_next, 1, rx_data->ring_size); IXGBE_WRITE_REG(&ixgbe->hw, IXGBE_RDT(rx_ring->hw_index), rx_tail); if (ixgbe_check_acc_handle(ixgbe->osdep.reg_handle) != DDI_FM_OK) { ddi_fm_service_impact(ixgbe->dip, DDI_SERVICE_DEGRADED); atomic_or_32(&ixgbe->ixgbe_state, IXGBE_ERROR); } return (mblk_head); } mblk_t * ixgbe_ring_rx_poll(void *arg, int n_bytes) { ixgbe_rx_ring_t *rx_ring = (ixgbe_rx_ring_t *)arg; mblk_t *mp = NULL; ASSERT(n_bytes >= 0); if (n_bytes == 0) return (NULL); mutex_enter(&rx_ring->rx_lock); mp = ixgbe_ring_rx(rx_ring, n_bytes); mutex_exit(&rx_ring->rx_lock); return (mp); }