1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2008 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 */ 25 26 #ifndef _SYS_NXGE_NXGE_TXDMA_H 27 #define _SYS_NXGE_NXGE_TXDMA_H 28 29 #pragma ident "%Z%%M% %I% %E% SMI" 30 31 #ifdef __cplusplus 32 extern "C" { 33 #endif 34 35 #include <sys/nxge/nxge_txdma_hw.h> 36 #include <npi_txdma.h> 37 #include <sys/nxge/nxge_serialize.h> 38 39 #define TXDMA_PORT_BITMAP(nxgep) (nxgep->pt_config.tx_dma_map) 40 41 #define TXDMA_RECLAIM_PENDING_DEFAULT 64 42 #define TX_FULL_MARK 3 43 44 /* 45 * Transmit load balancing definitions. 46 */ 47 #define NXGE_TX_LB_TCPUDP 0 /* default policy */ 48 #define NXGE_TX_LB_HASH 1 /* from the hint data */ 49 #define NXGE_TX_LB_DEST_MAC 2 /* Dest. MAC */ 50 51 /* 52 * Descriptor ring empty: 53 * (1) head index is equal to tail index. 54 * (2) wrapped around bits are the same. 55 * Descriptor ring full: 56 * (1) head index is equal to tail index. 57 * (2) wrapped around bits are different. 58 * 59 */ 60 #define TXDMA_RING_EMPTY(head, head_wrap, tail, tail_wrap) \ 61 ((head == tail && head_wrap == tail_wrap) ? B_TRUE : B_FALSE) 62 63 #define TXDMA_RING_FULL(head, head_wrap, tail, tail_wrap) \ 64 ((head == tail && head_wrap != tail_wrap) ? B_TRUE : B_FALSE) 65 66 #define TXDMA_DESC_NEXT_INDEX(index, entries, wrap_mask) \ 67 ((index + entries) & wrap_mask) 68 69 #define TXDMA_DRR_WEIGHT_DEFAULT 0x001f 70 71 typedef enum { 72 NXGE_USE_SERIAL = 0, 73 NXGE_USE_START, 74 } nxge_tx_mode_t; 75 76 typedef struct _tx_msg_t { 77 nxge_os_block_mv_t flags; /* DMA, BCOPY, DVMA (?) */ 78 nxge_os_dma_common_t buf_dma; /* premapped buffer blocks */ 79 nxge_os_dma_handle_t buf_dma_handle; /* premapped buffer handle */ 80 nxge_os_dma_handle_t dma_handle; /* DMA handle for normal send */ 81 nxge_os_dma_handle_t dvma_handle; /* Fast DVMA handle */ 82 83 p_mblk_t tx_message; 84 uint32_t tx_msg_size; 85 size_t bytes_used; 86 int head; 87 int tail; 88 } tx_msg_t, *p_tx_msg_t; 89 90 /* 91 * TX Statistics. 92 */ 93 typedef struct _nxge_tx_ring_stats_t { 94 uint64_t opackets; 95 uint64_t obytes; 96 uint64_t oerrors; 97 98 uint32_t tx_inits; 99 uint32_t tx_no_buf; 100 101 uint32_t mbox_err; 102 uint32_t pkt_size_err; 103 uint32_t tx_ring_oflow; 104 uint32_t pre_buf_par_err; 105 uint32_t nack_pref; 106 uint32_t nack_pkt_rd; 107 uint32_t conf_part_err; 108 uint32_t pkt_part_err; 109 uint32_t tx_starts; 110 uint32_t tx_nocanput; 111 uint32_t tx_msgdup_fail; 112 uint32_t tx_allocb_fail; 113 uint32_t tx_no_desc; 114 uint32_t tx_dma_bind_fail; 115 uint32_t tx_uflo; 116 117 uint32_t tx_hdr_pkts; 118 uint32_t tx_ddi_pkts; 119 uint32_t tx_dvma_pkts; 120 121 uint32_t tx_max_pend; 122 uint32_t tx_jumbo_pkts; 123 124 txdma_ring_errlog_t errlog; 125 } nxge_tx_ring_stats_t, *p_nxge_tx_ring_stats_t; 126 127 typedef struct _tx_ring_t { 128 nxge_os_dma_common_t tdc_desc; 129 struct _nxge_t *nxgep; 130 p_tx_msg_t tx_msg_ring; 131 uint32_t tnblocks; 132 tx_rng_cfig_t tx_ring_cfig; 133 tx_ring_hdl_t tx_ring_hdl; 134 tx_ring_kick_t tx_ring_kick; 135 tx_cs_t tx_cs; 136 tx_dma_ent_msk_t tx_evmask; 137 txdma_mbh_t tx_mbox_mbh; 138 txdma_mbl_t tx_mbox_mbl; 139 log_page_vld_t page_valid; 140 log_page_mask_t page_mask_1; 141 log_page_mask_t page_mask_2; 142 log_page_value_t page_value_1; 143 log_page_value_t page_value_2; 144 log_page_relo_t page_reloc_1; 145 log_page_relo_t page_reloc_2; 146 log_page_hdl_t page_hdl; 147 txc_dma_max_burst_t max_burst; 148 boolean_t cfg_set; 149 #define NXGE_TX_RING_ONLINE 0x00 150 #define NXGE_TX_RING_OFFLINING 0x01 151 #define NXGE_TX_RING_OFFLINED 0x02 152 uint32_t tx_ring_offline; 153 154 nxge_os_mutex_t lock; 155 uint16_t index; 156 uint16_t tdc; 157 struct nxge_tdc_cfg *tdc_p; 158 uint_t tx_ring_size; 159 uint32_t num_chunks; 160 161 uint_t tx_wrap_mask; 162 uint_t rd_index; 163 uint_t wr_index; 164 boolean_t wr_index_wrap; 165 uint_t head_index; 166 boolean_t head_wrap; 167 tx_ring_hdl_t ring_head; 168 tx_ring_kick_t ring_kick_tail; 169 txdma_mailbox_t tx_mbox; 170 171 uint_t descs_pending; 172 boolean_t queueing; 173 174 nxge_os_mutex_t sq_lock; 175 nxge_serialize_t *serial; 176 p_mblk_t head; 177 p_mblk_t tail; 178 179 uint16_t ldg_group_id; 180 p_nxge_tx_ring_stats_t tdc_stats; 181 182 nxge_os_mutex_t dvma_lock; 183 uint_t dvma_wr_index; 184 uint_t dvma_rd_index; 185 uint_t dvma_pending; 186 uint_t dvma_available; 187 uint_t dvma_wrap_mask; 188 189 nxge_os_dma_handle_t *dvma_ring; 190 191 #if defined(sun4v) && defined(NIU_LP_WORKAROUND) 192 uint64_t hv_tx_buf_base_ioaddr_pp; 193 uint64_t hv_tx_buf_ioaddr_size; 194 uint64_t hv_tx_cntl_base_ioaddr_pp; 195 uint64_t hv_tx_cntl_ioaddr_size; 196 boolean_t hv_set; 197 #endif 198 } tx_ring_t, *p_tx_ring_t; 199 200 201 /* Transmit Mailbox */ 202 typedef struct _tx_mbox_t { 203 nxge_os_mutex_t lock; 204 uint16_t index; 205 struct _nxge_t *nxgep; 206 uint16_t tdc; 207 nxge_os_dma_common_t tx_mbox; 208 txdma_mbl_t tx_mbox_l; 209 txdma_mbh_t tx_mbox_h; 210 } tx_mbox_t, *p_tx_mbox_t; 211 212 typedef struct _tx_rings_t { 213 p_tx_ring_t *rings; 214 boolean_t txdesc_allocated; 215 uint32_t ndmas; 216 nxge_os_dma_common_t tdc_dma; 217 nxge_os_dma_common_t tdc_mbox; 218 } tx_rings_t, *p_tx_rings_t; 219 220 221 #if defined(_KERNEL) || (defined(COSIM) && !defined(IODIAG)) 222 223 typedef struct _tx_buf_rings_t { 224 struct _tx_buf_ring_t *txbuf_rings; 225 boolean_t txbuf_allocated; 226 } tx_buf_rings_t, *p_tx_buf_rings_t; 227 228 #endif 229 230 typedef struct _tx_mbox_areas_t { 231 p_tx_mbox_t *txmbox_areas_p; 232 boolean_t txmbox_allocated; 233 } tx_mbox_areas_t, *p_tx_mbox_areas_t; 234 235 typedef struct _tx_param_t { 236 nxge_logical_page_t tx_logical_pages[NXGE_MAX_LOGICAL_PAGES]; 237 } tx_param_t, *p_tx_param_t; 238 239 typedef struct _tx_params { 240 struct _tx_param_t *tx_param_p; 241 } tx_params_t, *p_tx_params_t; 242 243 /* 244 * Global register definitions per chip and they are initialized 245 * using the function zero control registers. 246 * . 247 */ 248 typedef struct _txdma_globals { 249 boolean_t mode32; 250 } txdma_globals_t, *p_txdma_globals; 251 252 253 #if defined(SOLARIS) && (defined(_KERNEL) || \ 254 (defined(COSIM) && !defined(IODIAG))) 255 256 /* 257 * Transmit prototypes. 258 */ 259 nxge_status_t nxge_init_txdma_channels(p_nxge_t); 260 void nxge_uninit_txdma_channels(p_nxge_t); 261 262 nxge_status_t nxge_init_txdma_channel(p_nxge_t, int); 263 void nxge_uninit_txdma_channel(p_nxge_t, int); 264 265 void nxge_setup_dma_common(p_nxge_dma_common_t, p_nxge_dma_common_t, 266 uint32_t, uint32_t); 267 nxge_status_t nxge_reset_txdma_channel(p_nxge_t, uint16_t, 268 uint64_t); 269 nxge_status_t nxge_init_txdma_channel_event_mask(p_nxge_t, 270 uint16_t, p_tx_dma_ent_msk_t); 271 nxge_status_t nxge_init_txdma_channel_cntl_stat(p_nxge_t, 272 uint16_t, uint64_t); 273 nxge_status_t nxge_enable_txdma_channel(p_nxge_t, uint16_t, 274 p_tx_ring_t, p_tx_mbox_t); 275 276 p_mblk_t nxge_tx_pkt_header_reserve(p_mblk_t, uint8_t *); 277 int nxge_tx_pkt_nmblocks(p_mblk_t, int *); 278 boolean_t nxge_txdma_reclaim(p_nxge_t, p_tx_ring_t, int); 279 280 void nxge_fill_tx_hdr(p_mblk_t, boolean_t, boolean_t, 281 int, uint8_t, p_tx_pkt_hdr_all_t, t_uscalar_t, t_uscalar_t); 282 283 nxge_status_t nxge_txdma_hw_mode(p_nxge_t, boolean_t); 284 void nxge_hw_start_tx(p_nxge_t); 285 void nxge_txdma_stop(p_nxge_t); 286 void nxge_txdma_stop_start(p_nxge_t); 287 void nxge_fixup_txdma_rings(p_nxge_t); 288 void nxge_txdma_hw_kick(p_nxge_t); 289 void nxge_txdma_fix_channel(p_nxge_t, uint16_t); 290 void nxge_txdma_fixup_channel(p_nxge_t, p_tx_ring_t, 291 uint16_t); 292 void nxge_txdma_hw_kick_channel(p_nxge_t, p_tx_ring_t, 293 uint16_t); 294 295 void nxge_txdma_regs_dump(p_nxge_t, int); 296 void nxge_txdma_regs_dump_channels(p_nxge_t); 297 298 void nxge_check_tx_hang(p_nxge_t); 299 void nxge_fixup_hung_txdma_rings(p_nxge_t); 300 301 void nxge_reclaim_rings(p_nxge_t); 302 int nxge_txdma_channel_hung(p_nxge_t, 303 p_tx_ring_t tx_ring_p, uint16_t); 304 int nxge_txdma_hung(p_nxge_t); 305 int nxge_txdma_stop_inj_err(p_nxge_t, int); 306 void nxge_txdma_inject_err(p_nxge_t, uint32_t, uint8_t); 307 308 extern nxge_status_t nxge_alloc_tx_mem_pool(p_nxge_t); 309 extern nxge_status_t nxge_alloc_txb(p_nxge_t nxgep, int channel); 310 extern void nxge_free_txb(p_nxge_t nxgep, int channel); 311 312 #endif 313 314 #ifdef __cplusplus 315 } 316 #endif 317 318 #endif /* _SYS_NXGE_NXGE_TXDMA_H */ 319