1 /* 2 * This file is provided under a CDDLv1 license. When using or 3 * redistributing this file, you may do so under this license. 4 * In redistributing this file this license must be included 5 * and no other modification of this header file is permitted. 6 * 7 * CDDL LICENSE SUMMARY 8 * 9 * Copyright(c) 1999 - 2008 Intel Corporation. All rights reserved. 10 * 11 * The contents of this file are subject to the terms of Version 12 * 1.0 of the Common Development and Distribution License (the "License"). 13 * 14 * You should have received a copy of the License with this software. 15 * You can obtain a copy of the License at 16 * http://www.opensolaris.org/os/licensing. 17 * See the License for the specific language governing permissions 18 * and limitations under the License. 19 */ 20 21 /* 22 * Copyright 2008 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms of the CDDLv1. 24 */ 25 26 #pragma ident "%Z%%M% %I% %E% SMI" 27 28 /* 29 * ********************************************************************** 30 * * 31 * Module Name: * 32 * e1000g_tx.c * 33 * * 34 * Abstract: * 35 * This file contains some routines that take care of Transmit, * 36 * make the hardware to send the data pointed by the packet out * 37 * on to the physical medium. * 38 * * 39 * ********************************************************************** 40 */ 41 42 #include "e1000g_sw.h" 43 #include "e1000g_debug.h" 44 45 static boolean_t e1000g_send(struct e1000g *, mblk_t *); 46 static int e1000g_tx_copy(e1000g_tx_ring_t *, 47 p_tx_sw_packet_t, mblk_t *, uint32_t); 48 static int e1000g_tx_bind(e1000g_tx_ring_t *, 49 p_tx_sw_packet_t, mblk_t *); 50 static boolean_t check_cksum_context(e1000g_tx_ring_t *, cksum_data_t *); 51 static int e1000g_fill_tx_ring(e1000g_tx_ring_t *, LIST_DESCRIBER *, 52 cksum_data_t *); 53 static void e1000g_fill_context_descriptor(cksum_data_t *, 54 struct e1000_context_desc *); 55 static int e1000g_fill_tx_desc(e1000g_tx_ring_t *, 56 p_tx_sw_packet_t, uint64_t, size_t); 57 static uint32_t e1000g_fill_82544_desc(uint64_t Address, size_t Length, 58 p_desc_array_t desc_array); 59 static int e1000g_tx_workaround_PCIX_82544(p_tx_sw_packet_t, uint64_t, size_t); 60 static int e1000g_tx_workaround_jumbo_82544(p_tx_sw_packet_t, uint64_t, size_t); 61 static void e1000g_82547_timeout(void *); 62 static void e1000g_82547_tx_move_tail(e1000g_tx_ring_t *); 63 static void e1000g_82547_tx_move_tail_work(e1000g_tx_ring_t *); 64 65 #ifndef E1000G_DEBUG 66 #pragma inline(e1000g_tx_copy) 67 #pragma inline(e1000g_tx_bind) 68 #pragma inline(check_cksum_context) 69 #pragma inline(e1000g_fill_tx_ring) 70 #pragma inline(e1000g_fill_context_descriptor) 71 #pragma inline(e1000g_fill_tx_desc) 72 #pragma inline(e1000g_fill_82544_desc) 73 #pragma inline(e1000g_tx_workaround_PCIX_82544) 74 #pragma inline(e1000g_tx_workaround_jumbo_82544) 75 #pragma inline(e1000g_free_tx_swpkt) 76 #endif 77 78 /* 79 * e1000g_free_tx_swpkt - free up the tx sw packet 80 * 81 * Unbind the previously bound DMA handle for a given 82 * transmit sw packet. And reset the sw packet data. 83 */ 84 void 85 e1000g_free_tx_swpkt(register p_tx_sw_packet_t packet) 86 { 87 switch (packet->data_transfer_type) { 88 case USE_BCOPY: 89 packet->tx_buf->len = 0; 90 break; 91 #ifdef __sparc 92 case USE_DVMA: 93 dvma_unload(packet->tx_dma_handle, 0, -1); 94 break; 95 #endif 96 case USE_DMA: 97 ddi_dma_unbind_handle(packet->tx_dma_handle); 98 break; 99 default: 100 break; 101 } 102 103 /* 104 * The mblk has been stripped off the sw packet 105 * and will be freed in a triggered soft intr. 106 */ 107 ASSERT(packet->mp == NULL); 108 109 packet->data_transfer_type = USE_NONE; 110 packet->num_mblk_frag = 0; 111 packet->num_desc = 0; 112 } 113 114 mblk_t * 115 e1000g_m_tx(void *arg, mblk_t *mp) 116 { 117 struct e1000g *Adapter = (struct e1000g *)arg; 118 mblk_t *next; 119 120 rw_enter(&Adapter->chip_lock, RW_READER); 121 122 if ((Adapter->chip_state != E1000G_START) || 123 (Adapter->link_state != LINK_STATE_UP)) { 124 freemsgchain(mp); 125 mp = NULL; 126 } 127 128 while (mp != NULL) { 129 next = mp->b_next; 130 mp->b_next = NULL; 131 132 if (!e1000g_send(Adapter, mp)) { 133 mp->b_next = next; 134 break; 135 } 136 137 mp = next; 138 } 139 140 rw_exit(&Adapter->chip_lock); 141 return (mp); 142 } 143 144 /* 145 * e1000g_send - send packets onto the wire 146 * 147 * Called from e1000g_m_tx with an mblk ready to send. this 148 * routine sets up the transmit descriptors and sends data to 149 * the wire. It also pushes the just transmitted packet to 150 * the used tx sw packet list. 151 */ 152 static boolean_t 153 e1000g_send(struct e1000g *Adapter, mblk_t *mp) 154 { 155 struct e1000_hw *hw; 156 p_tx_sw_packet_t packet; 157 LIST_DESCRIBER pending_list; 158 size_t len; 159 size_t msg_size; 160 uint32_t frag_count; 161 int desc_count; 162 uint32_t desc_total; 163 uint32_t force_bcopy; 164 mblk_t *nmp; 165 mblk_t *tmp; 166 e1000g_tx_ring_t *tx_ring; 167 cksum_data_t cksum; 168 169 hw = &Adapter->shared; 170 tx_ring = Adapter->tx_ring; 171 172 /* Get the total size and frags number of the message */ 173 force_bcopy = 0; 174 frag_count = 0; 175 msg_size = 0; 176 for (nmp = mp; nmp; nmp = nmp->b_cont) { 177 frag_count++; 178 msg_size += MBLKL(nmp); 179 } 180 181 /* Make sure packet is less than the max frame size */ 182 if (msg_size > hw->mac.max_frame_size - ETHERFCSL) { 183 /* 184 * For the over size packet, we'll just drop it. 185 * So we return B_TRUE here. 186 */ 187 E1000G_DEBUGLOG_1(Adapter, E1000G_WARN_LEVEL, 188 "Tx packet out of bound. length = %d \n", msg_size); 189 E1000G_STAT(tx_ring->stat_over_size); 190 freemsg(mp); 191 return (B_TRUE); 192 } 193 194 /* 195 * Check and reclaim tx descriptors. 196 * This low water mark check should be done all the time as 197 * Transmit interrupt delay can produce Transmit interrupts little 198 * late and that may cause few problems related to reaping Tx 199 * Descriptors... As you may run short of them before getting any 200 * transmit interrupt... 201 */ 202 if (tx_ring->resched_needed || 203 (tx_ring->tbd_avail < Adapter->tx_recycle_thresh)) { 204 (void) e1000g_recycle(tx_ring); 205 E1000G_DEBUG_STAT(tx_ring->stat_recycle); 206 207 if (tx_ring->tbd_avail < DEFAULT_TX_NO_RESOURCE) { 208 E1000G_DEBUG_STAT(tx_ring->stat_lack_desc); 209 goto tx_no_resource; 210 } 211 } 212 213 /* 214 * If there are many frags of the message, then bcopy them 215 * into one tx descriptor buffer will get better performance. 216 */ 217 if ((frag_count >= tx_ring->frags_limit) && 218 (msg_size <= Adapter->tx_buffer_size)) { 219 E1000G_DEBUG_STAT(tx_ring->stat_exceed_frags); 220 force_bcopy |= FORCE_BCOPY_EXCEED_FRAGS; 221 } 222 223 /* 224 * If the message size is less than the minimum ethernet packet size, 225 * we'll use bcopy to send it, and padd it to 60 bytes later. 226 */ 227 if (msg_size < ETHERMIN) { 228 E1000G_DEBUG_STAT(tx_ring->stat_under_size); 229 force_bcopy |= FORCE_BCOPY_UNDER_SIZE; 230 } 231 232 /* Initialize variables */ 233 desc_count = 1; /* The initial value should be greater than 0 */ 234 desc_total = 0; 235 QUEUE_INIT_LIST(&pending_list); 236 237 /* Retrieve checksum info */ 238 hcksum_retrieve(mp, NULL, NULL, &cksum.cksum_start, &cksum.cksum_stuff, 239 NULL, NULL, &cksum.cksum_flags); 240 241 if (((struct ether_vlan_header *)mp->b_rptr)->ether_tpid == 242 htons(ETHERTYPE_VLAN)) 243 cksum.ether_header_size = sizeof (struct ether_vlan_header); 244 else 245 cksum.ether_header_size = sizeof (struct ether_header); 246 247 /* Process each mblk fragment and fill tx descriptors */ 248 packet = NULL; 249 nmp = mp; 250 while (nmp) { 251 tmp = nmp->b_cont; 252 253 len = MBLKL(nmp); 254 /* Check zero length mblks */ 255 if (len == 0) { 256 E1000G_DEBUG_STAT(tx_ring->stat_empty_frags); 257 /* 258 * If there're no packet buffers have been used, 259 * or we just completed processing a buffer, then 260 * skip the empty mblk fragment. 261 * Otherwise, there's still a pending buffer that 262 * needs to be processed (tx_copy). 263 */ 264 if (desc_count > 0) { 265 nmp = tmp; 266 continue; 267 } 268 } 269 270 /* 271 * Get a new TxSwPacket to process mblk buffers. 272 */ 273 if (desc_count > 0) { 274 mutex_enter(&tx_ring->freelist_lock); 275 packet = (p_tx_sw_packet_t) 276 QUEUE_POP_HEAD(&tx_ring->free_list); 277 mutex_exit(&tx_ring->freelist_lock); 278 279 if (packet == NULL) { 280 E1000G_DEBUGLOG_0(Adapter, E1000G_INFO_LEVEL, 281 "No Tx SwPacket available\n"); 282 E1000G_STAT(tx_ring->stat_no_swpkt); 283 goto tx_send_failed; 284 } 285 QUEUE_PUSH_TAIL(&pending_list, &packet->Link); 286 } 287 288 ASSERT(packet); 289 /* 290 * If the size of the fragment is less than the tx_bcopy_thresh 291 * we'll use bcopy; Otherwise, we'll use DMA binding. 292 */ 293 if ((len <= Adapter->tx_bcopy_thresh) || force_bcopy) { 294 desc_count = 295 e1000g_tx_copy(tx_ring, packet, nmp, force_bcopy); 296 E1000G_DEBUG_STAT(tx_ring->stat_copy); 297 } else { 298 desc_count = 299 e1000g_tx_bind(tx_ring, packet, nmp); 300 E1000G_DEBUG_STAT(tx_ring->stat_bind); 301 } 302 303 if (desc_count > 0) 304 desc_total += desc_count; 305 else if (desc_count < 0) 306 goto tx_send_failed; 307 308 nmp = tmp; 309 } 310 311 /* Assign the message to the last sw packet */ 312 ASSERT(packet); 313 ASSERT(packet->mp == NULL); 314 packet->mp = mp; 315 316 /* Try to recycle the tx descriptors again */ 317 if (tx_ring->tbd_avail < (desc_total + 2)) { 318 E1000G_DEBUG_STAT(tx_ring->stat_recycle_retry); 319 (void) e1000g_recycle(tx_ring); 320 } 321 322 mutex_enter(&tx_ring->tx_lock); 323 324 /* 325 * If the number of available tx descriptors is not enough for transmit 326 * (one redundant descriptor and one hw checksum context descriptor are 327 * included), then return failure. 328 */ 329 if (tx_ring->tbd_avail < (desc_total + 2)) { 330 E1000G_DEBUGLOG_0(Adapter, E1000G_INFO_LEVEL, 331 "No Enough Tx descriptors\n"); 332 E1000G_STAT(tx_ring->stat_no_desc); 333 mutex_exit(&tx_ring->tx_lock); 334 goto tx_send_failed; 335 } 336 337 desc_count = e1000g_fill_tx_ring(tx_ring, &pending_list, &cksum); 338 339 mutex_exit(&tx_ring->tx_lock); 340 341 ASSERT(desc_count > 0); 342 343 /* Send successful */ 344 return (B_TRUE); 345 346 tx_send_failed: 347 /* 348 * Enable Transmit interrupts, so that the interrupt routine can 349 * call mac_tx_update() when transmit descriptors become available. 350 */ 351 tx_ring->resched_needed = B_TRUE; 352 if (!Adapter->tx_intr_enable) 353 e1000g_mask_tx_interrupt(Adapter); 354 355 /* Free pending TxSwPackets */ 356 packet = (p_tx_sw_packet_t)QUEUE_GET_HEAD(&pending_list); 357 while (packet) { 358 packet->mp = NULL; 359 e1000g_free_tx_swpkt(packet); 360 packet = (p_tx_sw_packet_t) 361 QUEUE_GET_NEXT(&pending_list, &packet->Link); 362 } 363 364 /* Return pending TxSwPackets to the "Free" list */ 365 mutex_enter(&tx_ring->freelist_lock); 366 QUEUE_APPEND(&tx_ring->free_list, &pending_list); 367 mutex_exit(&tx_ring->freelist_lock); 368 369 E1000G_STAT(tx_ring->stat_send_fail); 370 371 /* Message will be scheduled for re-transmit */ 372 return (B_FALSE); 373 374 tx_no_resource: 375 /* 376 * Enable Transmit interrupts, so that the interrupt routine can 377 * call mac_tx_update() when transmit descriptors become available. 378 */ 379 tx_ring->resched_needed = B_TRUE; 380 if (!Adapter->tx_intr_enable) 381 e1000g_mask_tx_interrupt(Adapter); 382 383 /* Message will be scheduled for re-transmit */ 384 return (B_FALSE); 385 } 386 387 static boolean_t 388 check_cksum_context(e1000g_tx_ring_t *tx_ring, cksum_data_t *cksum) 389 { 390 boolean_t cksum_load; 391 cksum_data_t *last; 392 393 cksum_load = B_FALSE; 394 last = &tx_ring->cksum_data; 395 396 if (cksum->cksum_flags != 0) { 397 if ((cksum->ether_header_size != last->ether_header_size) || 398 (cksum->cksum_flags != last->cksum_flags) || 399 (cksum->cksum_stuff != last->cksum_stuff) || 400 (cksum->cksum_start != last->cksum_start)) { 401 402 cksum_load = B_TRUE; 403 } 404 } 405 406 return (cksum_load); 407 } 408 409 static int 410 e1000g_fill_tx_ring(e1000g_tx_ring_t *tx_ring, LIST_DESCRIBER *pending_list, 411 cksum_data_t *cksum) 412 { 413 struct e1000g *Adapter; 414 struct e1000_hw *hw; 415 p_tx_sw_packet_t first_packet; 416 p_tx_sw_packet_t packet; 417 boolean_t cksum_load; 418 struct e1000_tx_desc *first_data_desc; 419 struct e1000_tx_desc *next_desc; 420 struct e1000_tx_desc *descriptor; 421 int desc_count; 422 int i; 423 424 Adapter = tx_ring->adapter; 425 hw = &Adapter->shared; 426 427 desc_count = 0; 428 first_packet = NULL; 429 first_data_desc = NULL; 430 descriptor = NULL; 431 432 next_desc = tx_ring->tbd_next; 433 434 /* IP Head/TCP/UDP checksum offload */ 435 cksum_load = check_cksum_context(tx_ring, cksum); 436 437 if (cksum_load) { 438 first_packet = (p_tx_sw_packet_t)QUEUE_GET_HEAD(pending_list); 439 440 descriptor = next_desc; 441 442 e1000g_fill_context_descriptor(cksum, 443 (struct e1000_context_desc *)descriptor); 444 445 /* Check the wrap-around case */ 446 if (descriptor == tx_ring->tbd_last) 447 next_desc = tx_ring->tbd_first; 448 else 449 next_desc++; 450 451 desc_count++; 452 } 453 454 first_data_desc = next_desc; 455 456 packet = (p_tx_sw_packet_t)QUEUE_GET_HEAD(pending_list); 457 while (packet) { 458 ASSERT(packet->num_desc); 459 460 for (i = 0; i < packet->num_desc; i++) { 461 ASSERT(tx_ring->tbd_avail > 0); 462 463 descriptor = next_desc; 464 descriptor->buffer_addr = 465 packet->desc[i].address; 466 descriptor->lower.data = 467 packet->desc[i].length; 468 469 /* Zero out status */ 470 descriptor->upper.data = 0; 471 472 descriptor->lower.data |= 473 E1000_TXD_CMD_DEXT | E1000_TXD_DTYP_D; 474 /* must set RS on every outgoing descriptor */ 475 descriptor->lower.data |= 476 E1000_TXD_CMD_RS; 477 478 /* Check the wrap-around case */ 479 if (descriptor == tx_ring->tbd_last) 480 next_desc = tx_ring->tbd_first; 481 else 482 next_desc++; 483 484 desc_count++; 485 } 486 487 if (first_packet != NULL) { 488 /* 489 * Count the checksum context descriptor for 490 * the first SwPacket. 491 */ 492 first_packet->num_desc++; 493 first_packet = NULL; 494 } 495 496 packet = (p_tx_sw_packet_t) 497 QUEUE_GET_NEXT(pending_list, &packet->Link); 498 } 499 500 ASSERT(descriptor); 501 502 if (cksum->cksum_flags) { 503 if (cksum->cksum_flags & HCK_IPV4_HDRCKSUM) 504 ((struct e1000_data_desc *)first_data_desc)-> 505 upper.fields.popts |= E1000_TXD_POPTS_IXSM; 506 if (cksum->cksum_flags & HCK_PARTIALCKSUM) 507 ((struct e1000_data_desc *)first_data_desc)-> 508 upper.fields.popts |= E1000_TXD_POPTS_TXSM; 509 } 510 511 /* 512 * Last Descriptor of Packet needs End Of Packet (EOP), Report 513 * Status (RS) and append Ethernet CRC (IFCS) bits set. 514 */ 515 if (Adapter->tx_intr_delay) { 516 descriptor->lower.data |= E1000_TXD_CMD_IDE | 517 E1000_TXD_CMD_EOP | E1000_TXD_CMD_IFCS; 518 } else { 519 descriptor->lower.data |= 520 E1000_TXD_CMD_EOP | E1000_TXD_CMD_IFCS; 521 } 522 523 /* 524 * Sync the Tx descriptors DMA buffer 525 */ 526 (void) ddi_dma_sync(tx_ring->tbd_dma_handle, 527 0, 0, DDI_DMA_SYNC_FORDEV); 528 529 tx_ring->tbd_next = next_desc; 530 531 /* 532 * Advance the Transmit Descriptor Tail (Tdt), this tells the 533 * FX1000 that this frame is available to transmit. 534 */ 535 if (hw->mac.type == e1000_82547) 536 e1000g_82547_tx_move_tail(tx_ring); 537 else 538 E1000_WRITE_REG(hw, E1000_TDT, 539 (uint32_t)(next_desc - tx_ring->tbd_first)); 540 541 if (e1000g_check_acc_handle(Adapter->osdep.reg_handle) != DDI_FM_OK) { 542 ddi_fm_service_impact(Adapter->dip, DDI_SERVICE_DEGRADED); 543 Adapter->chip_state = E1000G_ERROR; 544 } 545 546 /* Put the pending SwPackets to the "Used" list */ 547 mutex_enter(&tx_ring->usedlist_lock); 548 QUEUE_APPEND(&tx_ring->used_list, pending_list); 549 tx_ring->tbd_avail -= desc_count; 550 mutex_exit(&tx_ring->usedlist_lock); 551 552 /* Store the cksum data */ 553 if (cksum_load) 554 tx_ring->cksum_data = *cksum; 555 556 return (desc_count); 557 } 558 559 560 /* 561 * e1000g_tx_setup - setup tx data structures 562 * 563 * This routine initializes all of the transmit related 564 * structures. This includes the Transmit descriptors, 565 * and the tx_sw_packet structures. 566 */ 567 void 568 e1000g_tx_setup(struct e1000g *Adapter) 569 { 570 struct e1000_hw *hw; 571 p_tx_sw_packet_t packet; 572 UINT i; 573 uint32_t buf_high; 574 uint32_t buf_low; 575 uint32_t reg_tipg; 576 uint32_t reg_tctl; 577 uint32_t reg_tarc; 578 uint16_t speed, duplex; 579 int size; 580 e1000g_tx_ring_t *tx_ring; 581 582 hw = &Adapter->shared; 583 tx_ring = Adapter->tx_ring; 584 585 /* init the lists */ 586 /* 587 * Here we don't need to protect the lists using the 588 * usedlist_lock and freelist_lock, for they have 589 * been protected by the chip_lock. 590 */ 591 QUEUE_INIT_LIST(&tx_ring->used_list); 592 QUEUE_INIT_LIST(&tx_ring->free_list); 593 594 /* Go through and set up each SW_Packet */ 595 packet = tx_ring->packet_area; 596 for (i = 0; i < Adapter->tx_freelist_num; i++, packet++) { 597 /* Initialize this tx_sw_apcket area */ 598 e1000g_free_tx_swpkt(packet); 599 /* Add this tx_sw_packet to the free list */ 600 QUEUE_PUSH_TAIL(&tx_ring->free_list, 601 &packet->Link); 602 } 603 604 /* Setup TX descriptor pointers */ 605 tx_ring->tbd_next = tx_ring->tbd_first; 606 tx_ring->tbd_oldest = tx_ring->tbd_first; 607 608 /* 609 * Setup Hardware TX Registers 610 */ 611 /* Setup the Transmit Control Register (TCTL). */ 612 reg_tctl = E1000_TCTL_PSP | E1000_TCTL_EN | 613 (E1000_COLLISION_THRESHOLD << E1000_CT_SHIFT) | 614 (E1000_COLLISION_DISTANCE << E1000_COLD_SHIFT) | 615 E1000_TCTL_RTLC; 616 617 /* Enable the MULR bit */ 618 if (hw->bus.type == e1000_bus_type_pci_express) 619 reg_tctl |= E1000_TCTL_MULR; 620 621 E1000_WRITE_REG(hw, E1000_TCTL, reg_tctl); 622 623 if ((hw->mac.type == e1000_82571) || (hw->mac.type == e1000_82572)) { 624 e1000_get_speed_and_duplex(hw, &speed, &duplex); 625 626 reg_tarc = E1000_READ_REG(hw, E1000_TARC0); 627 reg_tarc |= (1 << 25); 628 if (speed == SPEED_1000) 629 reg_tarc |= (1 << 21); 630 E1000_WRITE_REG(hw, E1000_TARC0, reg_tarc); 631 632 reg_tarc = E1000_READ_REG(hw, E1000_TARC1); 633 reg_tarc |= (1 << 25); 634 if (reg_tctl & E1000_TCTL_MULR) 635 reg_tarc &= ~(1 << 28); 636 else 637 reg_tarc |= (1 << 28); 638 E1000_WRITE_REG(hw, E1000_TARC1, reg_tarc); 639 640 } else if (hw->mac.type == e1000_80003es2lan) { 641 reg_tarc = E1000_READ_REG(hw, E1000_TARC0); 642 reg_tarc |= 1; 643 if (hw->media_type == e1000_media_type_internal_serdes) 644 reg_tarc |= (1 << 20); 645 E1000_WRITE_REG(hw, E1000_TARC0, reg_tarc); 646 647 reg_tarc = E1000_READ_REG(hw, E1000_TARC1); 648 reg_tarc |= 1; 649 E1000_WRITE_REG(hw, E1000_TARC1, reg_tarc); 650 } 651 652 /* Setup HW Base and Length of Tx descriptor area */ 653 size = (Adapter->tx_desc_num * sizeof (struct e1000_tx_desc)); 654 E1000_WRITE_REG(hw, E1000_TDLEN, size); 655 size = E1000_READ_REG(hw, E1000_TDLEN); 656 657 buf_low = (uint32_t)tx_ring->tbd_dma_addr; 658 buf_high = (uint32_t)(tx_ring->tbd_dma_addr >> 32); 659 660 E1000_WRITE_REG(hw, E1000_TDBAL, buf_low); 661 E1000_WRITE_REG(hw, E1000_TDBAH, buf_high); 662 663 /* Setup our HW Tx Head & Tail descriptor pointers */ 664 E1000_WRITE_REG(hw, E1000_TDH, 0); 665 E1000_WRITE_REG(hw, E1000_TDT, 0); 666 667 /* Set the default values for the Tx Inter Packet Gap timer */ 668 if ((hw->mac.type == e1000_82542) && 669 ((hw->revision_id == E1000_REVISION_2) || 670 (hw->revision_id == E1000_REVISION_3))) { 671 reg_tipg = DEFAULT_82542_TIPG_IPGT; 672 reg_tipg |= 673 DEFAULT_82542_TIPG_IPGR1 << E1000_TIPG_IPGR1_SHIFT; 674 reg_tipg |= 675 DEFAULT_82542_TIPG_IPGR2 << E1000_TIPG_IPGR2_SHIFT; 676 } else { 677 if (hw->media_type == e1000_media_type_fiber) 678 reg_tipg = DEFAULT_82543_TIPG_IPGT_FIBER; 679 else 680 reg_tipg = DEFAULT_82543_TIPG_IPGT_COPPER; 681 reg_tipg |= 682 DEFAULT_82543_TIPG_IPGR1 << E1000_TIPG_IPGR1_SHIFT; 683 reg_tipg |= 684 DEFAULT_82543_TIPG_IPGR2 << E1000_TIPG_IPGR2_SHIFT; 685 } 686 E1000_WRITE_REG(hw, E1000_TIPG, reg_tipg); 687 688 /* Setup Transmit Interrupt Delay Value */ 689 E1000_WRITE_REG(hw, E1000_TIDV, Adapter->tx_intr_delay); 690 E1000G_DEBUGLOG_1(Adapter, E1000G_INFO_LEVEL, 691 "E1000_TIDV: 0x%x\n", Adapter->tx_intr_delay); 692 693 if (hw->mac.type >= e1000_82540) { 694 E1000_WRITE_REG(&Adapter->shared, E1000_TADV, 695 Adapter->tx_intr_abs_delay); 696 E1000G_DEBUGLOG_1(Adapter, E1000G_INFO_LEVEL, 697 "E1000_TADV: 0x%x\n", Adapter->tx_intr_abs_delay); 698 } 699 700 tx_ring->tbd_avail = Adapter->tx_desc_num; 701 702 /* For TCP/UDP checksum offload */ 703 tx_ring->cksum_data.cksum_stuff = 0; 704 tx_ring->cksum_data.cksum_start = 0; 705 tx_ring->cksum_data.cksum_flags = 0; 706 tx_ring->cksum_data.ether_header_size = 0; 707 } 708 709 /* 710 * e1000g_recycle - recycle the tx descriptors and tx sw packets 711 */ 712 int 713 e1000g_recycle(e1000g_tx_ring_t *tx_ring) 714 { 715 struct e1000g *Adapter; 716 LIST_DESCRIBER pending_list; 717 p_tx_sw_packet_t packet; 718 mblk_t *mp; 719 mblk_t *nmp; 720 struct e1000_tx_desc *descriptor; 721 int desc_count; 722 int is_intr; 723 724 /* 725 * This function will examine each TxSwPacket in the 'used' queue 726 * if the e1000g is done with it then the associated resources (Tx 727 * Descriptors) will be "freed" and the TxSwPacket will be 728 * returned to the 'free' queue. 729 */ 730 Adapter = tx_ring->adapter; 731 732 packet = (p_tx_sw_packet_t)QUEUE_GET_HEAD(&tx_ring->used_list); 733 if (packet == NULL) { 734 tx_ring->recycle_fail = 0; 735 tx_ring->stall_watchdog = 0; 736 return (0); 737 } 738 739 is_intr = servicing_interrupt(); 740 741 if (is_intr) 742 mutex_enter(&tx_ring->usedlist_lock); 743 else if (mutex_tryenter(&tx_ring->usedlist_lock) == 0) 744 return (0); 745 746 desc_count = 0; 747 QUEUE_INIT_LIST(&pending_list); 748 749 /* Sync the Tx descriptor DMA buffer */ 750 (void) ddi_dma_sync(tx_ring->tbd_dma_handle, 751 0, 0, DDI_DMA_SYNC_FORKERNEL); 752 if (e1000g_check_dma_handle( 753 tx_ring->tbd_dma_handle) != DDI_FM_OK) { 754 mutex_exit(&tx_ring->usedlist_lock); 755 ddi_fm_service_impact(Adapter->dip, DDI_SERVICE_DEGRADED); 756 Adapter->chip_state = E1000G_ERROR; 757 return (0); 758 } 759 760 /* 761 * While there are still TxSwPackets in the used queue check them 762 */ 763 while (packet = 764 (p_tx_sw_packet_t)QUEUE_GET_HEAD(&tx_ring->used_list)) { 765 766 /* 767 * Get hold of the next descriptor that the e1000g will 768 * report status back to (this will be the last descriptor 769 * of a given sw packet). We only want to free the 770 * sw packet (and it resources) if the e1000g is done 771 * with ALL of the descriptors. If the e1000g is done 772 * with the last one then it is done with all of them. 773 */ 774 ASSERT(packet->num_desc); 775 descriptor = tx_ring->tbd_oldest + (packet->num_desc - 1); 776 777 /* Check for wrap case */ 778 if (descriptor > tx_ring->tbd_last) 779 descriptor -= Adapter->tx_desc_num; 780 781 /* 782 * If the descriptor done bit is set free TxSwPacket and 783 * associated resources 784 */ 785 if (descriptor->upper.fields.status & E1000_TXD_STAT_DD) { 786 QUEUE_POP_HEAD(&tx_ring->used_list); 787 QUEUE_PUSH_TAIL(&pending_list, &packet->Link); 788 789 if (descriptor == tx_ring->tbd_last) 790 tx_ring->tbd_oldest = 791 tx_ring->tbd_first; 792 else 793 tx_ring->tbd_oldest = 794 descriptor + 1; 795 796 desc_count += packet->num_desc; 797 798 if (is_intr && (desc_count >= Adapter->tx_recycle_num)) 799 break; 800 } else { 801 /* 802 * Found a sw packet that the e1000g is not done 803 * with then there is no reason to check the rest 804 * of the queue. 805 */ 806 break; 807 } 808 } 809 810 tx_ring->tbd_avail += desc_count; 811 Adapter->tx_pkt_cnt += desc_count; 812 813 mutex_exit(&tx_ring->usedlist_lock); 814 815 if (desc_count == 0) { 816 tx_ring->recycle_fail++; 817 E1000G_DEBUG_STAT(tx_ring->stat_recycle_none); 818 return (0); 819 } 820 821 tx_ring->recycle_fail = 0; 822 tx_ring->stall_watchdog = 0; 823 824 mp = NULL; 825 nmp = NULL; 826 packet = (p_tx_sw_packet_t)QUEUE_GET_HEAD(&pending_list); 827 ASSERT(packet != NULL); 828 while (packet != NULL) { 829 if (packet->mp != NULL) { 830 ASSERT(packet->mp->b_next == NULL); 831 /* Assemble the message chain */ 832 if (mp == NULL) { 833 mp = packet->mp; 834 nmp = packet->mp; 835 } else { 836 nmp->b_next = packet->mp; 837 nmp = packet->mp; 838 } 839 /* Disconnect the message from the sw packet */ 840 packet->mp = NULL; 841 } 842 843 /* Free the TxSwPackets */ 844 e1000g_free_tx_swpkt(packet); 845 846 packet = (p_tx_sw_packet_t) 847 QUEUE_GET_NEXT(&pending_list, &packet->Link); 848 } 849 850 /* Return the TxSwPackets back to the FreeList */ 851 mutex_enter(&tx_ring->freelist_lock); 852 QUEUE_APPEND(&tx_ring->free_list, &pending_list); 853 mutex_exit(&tx_ring->freelist_lock); 854 855 if (mp != NULL) 856 freemsgchain(mp); 857 858 return (desc_count); 859 } 860 /* 861 * 82544 Coexistence issue workaround: 862 * There are 2 issues. 863 * 1. If a 32 bit split completion happens from P64H2 and another 864 * agent drives a 64 bit request/split completion after ONLY 865 * 1 idle clock (BRCM/Emulex/Adaptec fiber channel cards) then 866 * 82544 has a problem where in to clock all the data in, it 867 * looks at REQ64# signal and since it has changed so fast (i.e. 1 868 * idle clock turn around), it will fail to clock all the data in. 869 * Data coming from certain ending addresses has exposure to this issue. 870 * 871 * To detect this issue, following equation can be used... 872 * SIZE[3:0] + ADDR[2:0] = SUM[3:0]. 873 * If SUM[3:0] is in between 1 to 4, we will have this issue. 874 * 875 * ROOT CAUSE: 876 * The erratum involves the 82544 PCIX elasticity FIFO implementations as 877 * 64-bit FIFO's and flushing of the final partial-bytes corresponding 878 * to the end of a requested read burst. Under a specific burst condition 879 * of ending-data alignment and 32-byte split-completions, the final 880 * byte(s) of split-completion data require an extra clock cycle to flush 881 * into 64-bit FIFO orientation. An incorrect logic dependency on the 882 * REQ64# signal occurring during during this clock cycle may cause the 883 * residual byte(s) to be lost, thereby rendering the internal DMA client 884 * forever awaiting the final byte(s) for an outbound data-fetch. The 885 * erratum is confirmed to *only* occur if certain subsequent external 886 * 64-bit PCIX bus transactions occur immediately (minimum possible bus 887 * turn- around) following the odd-aligned 32-bit split-completion 888 * containing the final byte(s). Intel has confirmed that this has been 889 * seen only with chipset/bridges which have the capability to provide 890 * 32-bit split-completion data, and in the presence of newer PCIX bus 891 * agents which fully-optimize the inter-transaction turn-around (zero 892 * additional initiator latency when pre-granted bus ownership). 893 * 894 * This issue does not exist in PCI bus mode, when any agent is operating 895 * in 32 bit only mode or on chipsets that do not do 32 bit split 896 * completions for 64 bit read requests (Serverworks chipsets). P64H2 does 897 * 32 bit split completions for any read request that has bit 2 set to 1 898 * for the requested address and read request size is more than 8 bytes. 899 * 900 * 2. Another issue is related to 82544 driving DACs under the similar 901 * scenario (32 bit split completion followed by 64 bit transaction with 902 * only 1 cycle turnaround). This issue is still being root caused. We 903 * think that both of these issues can be avoided if following workaround 904 * is implemented. It seems DAC issues is related to ending addresses being 905 * 0x9, 0xA, 0xB, 0xC and hence ending up at odd boundaries in elasticity 906 * FIFO which does not get flushed due to REQ64# dependency. We will only 907 * know the full story after it has been simulated successfully by HW team. 908 * 909 * WORKAROUND: 910 * Make sure we do not have ending address as 1,2,3,4(Hang) or 9,a,b,c(DAC) 911 */ 912 static uint32_t 913 e1000g_fill_82544_desc(uint64_t address, 914 size_t length, p_desc_array_t desc_array) 915 { 916 /* 917 * Since issue is sensitive to length and address. 918 * Let us first check the address... 919 */ 920 uint32_t safe_terminator; 921 922 if (length <= 4) { 923 desc_array->descriptor[0].address = address; 924 desc_array->descriptor[0].length = length; 925 desc_array->elements = 1; 926 return (desc_array->elements); 927 } 928 safe_terminator = 929 (uint32_t)((((uint32_t)address & 0x7) + 930 (length & 0xF)) & 0xF); 931 /* 932 * if it does not fall between 0x1 to 0x4 and 0x9 to 0xC then 933 * return 934 */ 935 if (safe_terminator == 0 || 936 (safe_terminator > 4 && safe_terminator < 9) || 937 (safe_terminator > 0xC && safe_terminator <= 0xF)) { 938 desc_array->descriptor[0].address = address; 939 desc_array->descriptor[0].length = length; 940 desc_array->elements = 1; 941 return (desc_array->elements); 942 } 943 944 desc_array->descriptor[0].address = address; 945 desc_array->descriptor[0].length = length - 4; 946 desc_array->descriptor[1].address = address + (length - 4); 947 desc_array->descriptor[1].length = 4; 948 desc_array->elements = 2; 949 return (desc_array->elements); 950 } 951 952 static int 953 e1000g_tx_copy(e1000g_tx_ring_t *tx_ring, p_tx_sw_packet_t packet, 954 mblk_t *mp, uint32_t force_bcopy) 955 { 956 size_t len; 957 size_t len1; 958 dma_buffer_t *tx_buf; 959 mblk_t *nmp; 960 boolean_t finished; 961 int desc_count; 962 963 desc_count = 0; 964 tx_buf = packet->tx_buf; 965 len = MBLKL(mp); 966 967 ASSERT((tx_buf->len + len) <= tx_buf->size); 968 969 if (len > 0) { 970 bcopy(mp->b_rptr, 971 tx_buf->address + tx_buf->len, 972 len); 973 tx_buf->len += len; 974 975 packet->num_mblk_frag++; 976 } 977 978 nmp = mp->b_cont; 979 if (nmp == NULL) { 980 finished = B_TRUE; 981 } else { 982 len1 = MBLKL(nmp); 983 if ((tx_buf->len + len1) > tx_buf->size) 984 finished = B_TRUE; 985 else if (force_bcopy) 986 finished = B_FALSE; 987 else if (len1 > tx_ring->adapter->tx_bcopy_thresh) 988 finished = B_TRUE; 989 else 990 finished = B_FALSE; 991 } 992 993 if (finished) { 994 E1000G_DEBUG_STAT_COND(tx_ring->stat_multi_copy, 995 (tx_buf->len > len)); 996 997 /* 998 * If the packet is smaller than 64 bytes, which is the 999 * minimum ethernet packet size, pad the packet to make 1000 * it at least 60 bytes. The hardware will add 4 bytes 1001 * for CRC. 1002 */ 1003 if (force_bcopy & FORCE_BCOPY_UNDER_SIZE) { 1004 ASSERT(tx_buf->len < ETHERMIN); 1005 1006 bzero(tx_buf->address + tx_buf->len, 1007 ETHERMIN - tx_buf->len); 1008 tx_buf->len = ETHERMIN; 1009 } 1010 1011 #ifdef __sparc 1012 if (packet->dma_type == USE_DVMA) 1013 dvma_sync(tx_buf->dma_handle, 0, DDI_DMA_SYNC_FORDEV); 1014 else 1015 (void) ddi_dma_sync(tx_buf->dma_handle, 0, 1016 tx_buf->len, DDI_DMA_SYNC_FORDEV); 1017 #else 1018 (void) ddi_dma_sync(tx_buf->dma_handle, 0, 1019 tx_buf->len, DDI_DMA_SYNC_FORDEV); 1020 #endif 1021 1022 packet->data_transfer_type = USE_BCOPY; 1023 1024 desc_count = e1000g_fill_tx_desc(tx_ring, 1025 packet, 1026 tx_buf->dma_address, 1027 tx_buf->len); 1028 1029 if (desc_count <= 0) 1030 return (-1); 1031 } 1032 1033 return (desc_count); 1034 } 1035 1036 static int 1037 e1000g_tx_bind(e1000g_tx_ring_t *tx_ring, p_tx_sw_packet_t packet, mblk_t *mp) 1038 { 1039 int j; 1040 int mystat; 1041 size_t len; 1042 ddi_dma_cookie_t dma_cookie; 1043 uint_t ncookies; 1044 int desc_count; 1045 uint32_t desc_total; 1046 1047 desc_total = 0; 1048 len = MBLKL(mp); 1049 1050 /* 1051 * ddi_dma_addr_bind_handle() allocates DMA resources for a 1052 * memory object such that a device can perform DMA to or from 1053 * the object. DMA resources are allocated considering the 1054 * device's DMA attributes as expressed by ddi_dma_attr(9S) 1055 * (see ddi_dma_alloc_handle(9F)). 1056 * 1057 * ddi_dma_addr_bind_handle() fills in the first DMA cookie 1058 * pointed to by cookiep with the appropriate address, length, 1059 * and bus type. *ccountp is set to the number of DMA cookies 1060 * representing this DMA object. Subsequent DMA cookies must be 1061 * retrieved by calling ddi_dma_nextcookie(9F) the number of 1062 * times specified by *countp - 1. 1063 */ 1064 switch (packet->dma_type) { 1065 #ifdef __sparc 1066 case USE_DVMA: 1067 dvma_kaddr_load(packet->tx_dma_handle, 1068 (caddr_t)mp->b_rptr, len, 0, &dma_cookie); 1069 1070 dvma_sync(packet->tx_dma_handle, 0, 1071 DDI_DMA_SYNC_FORDEV); 1072 1073 ncookies = 1; 1074 packet->data_transfer_type = USE_DVMA; 1075 break; 1076 #endif 1077 case USE_DMA: 1078 if ((mystat = ddi_dma_addr_bind_handle( 1079 packet->tx_dma_handle, NULL, 1080 (caddr_t)mp->b_rptr, len, 1081 DDI_DMA_WRITE | DDI_DMA_STREAMING, 1082 DDI_DMA_DONTWAIT, 0, &dma_cookie, 1083 &ncookies)) != DDI_DMA_MAPPED) { 1084 1085 e1000g_log(tx_ring->adapter, CE_WARN, 1086 "Couldn't bind mblk buffer to Tx DMA handle: " 1087 "return: %X, Pkt: %X\n", 1088 mystat, packet); 1089 return (-1); 1090 } 1091 1092 /* 1093 * An implicit ddi_dma_sync() is done when the 1094 * ddi_dma_addr_bind_handle() is called. So we 1095 * don't need to explicitly call ddi_dma_sync() 1096 * here any more. 1097 */ 1098 ASSERT(ncookies); 1099 E1000G_DEBUG_STAT_COND(tx_ring->stat_multi_cookie, 1100 (ncookies > 1)); 1101 1102 /* 1103 * The data_transfer_type value must be set after the handle 1104 * has been bound, for it will be used in e1000g_free_tx_swpkt() 1105 * to decide whether we need to unbind the handle. 1106 */ 1107 packet->data_transfer_type = USE_DMA; 1108 break; 1109 default: 1110 ASSERT(B_FALSE); 1111 break; 1112 } 1113 1114 packet->num_mblk_frag++; 1115 1116 /* 1117 * Each address could span thru multpile cookie.. 1118 * Each cookie will have one descriptor 1119 */ 1120 for (j = ncookies; j != 0; j--) { 1121 1122 desc_count = e1000g_fill_tx_desc(tx_ring, 1123 packet, 1124 dma_cookie.dmac_laddress, 1125 dma_cookie.dmac_size); 1126 1127 if (desc_count <= 0) 1128 return (-1); 1129 1130 desc_total += desc_count; 1131 1132 /* 1133 * ddi_dma_nextcookie() retrieves subsequent DMA 1134 * cookies for a DMA object. 1135 * ddi_dma_nextcookie() fills in the 1136 * ddi_dma_cookie(9S) structure pointed to by 1137 * cookiep. The ddi_dma_cookie(9S) structure 1138 * must be allocated prior to calling 1139 * ddi_dma_nextcookie(). The DMA cookie count 1140 * returned by ddi_dma_buf_bind_handle(9F), 1141 * ddi_dma_addr_bind_handle(9F), or 1142 * ddi_dma_getwin(9F) indicates the number of DMA 1143 * cookies a DMA object consists of. If the 1144 * resulting cookie count, N, is larger than 1, 1145 * ddi_dma_nextcookie() must be called N-1 times 1146 * to retrieve all DMA cookies. 1147 */ 1148 if (j > 1) { 1149 ddi_dma_nextcookie(packet->tx_dma_handle, 1150 &dma_cookie); 1151 } 1152 } 1153 1154 return (desc_total); 1155 } 1156 1157 static void 1158 e1000g_fill_context_descriptor(cksum_data_t *cksum, 1159 struct e1000_context_desc *cksum_desc) 1160 { 1161 if (cksum->cksum_flags & HCK_IPV4_HDRCKSUM) { 1162 cksum_desc->lower_setup.ip_fields.ipcss = 1163 cksum->ether_header_size; 1164 cksum_desc->lower_setup.ip_fields.ipcso = 1165 cksum->ether_header_size + 1166 offsetof(struct ip, ip_sum); 1167 cksum_desc->lower_setup.ip_fields.ipcse = 1168 cksum->ether_header_size + 1169 sizeof (struct ip) - 1; 1170 } else 1171 cksum_desc->lower_setup.ip_config = 0; 1172 1173 if (cksum->cksum_flags & HCK_PARTIALCKSUM) { 1174 /* 1175 * The packet with same protocol has the following 1176 * stuff and start offset: 1177 * | Protocol | Stuff | Start | Checksum 1178 * | | Offset | Offset | Enable 1179 * | IPv4 + TCP | 0x24 | 0x14 | Yes 1180 * | IPv4 + UDP | 0x1A | 0x14 | Yes 1181 * | IPv6 + TCP | 0x20 | 0x10 | No 1182 * | IPv6 + UDP | 0x14 | 0x10 | No 1183 */ 1184 cksum_desc->upper_setup.tcp_fields.tucss = 1185 cksum->cksum_start + cksum->ether_header_size; 1186 cksum_desc->upper_setup.tcp_fields.tucso = 1187 cksum->cksum_stuff + cksum->ether_header_size; 1188 cksum_desc->upper_setup.tcp_fields.tucse = 0; 1189 } else 1190 cksum_desc->upper_setup.tcp_config = 0; 1191 1192 cksum_desc->cmd_and_length = E1000_TXD_CMD_DEXT; 1193 1194 /* 1195 * Zero out the options for TCP Segmentation Offload, 1196 * since we don't support it in this version 1197 */ 1198 cksum_desc->tcp_seg_setup.data = 0; 1199 } 1200 1201 static int 1202 e1000g_fill_tx_desc(e1000g_tx_ring_t *tx_ring, 1203 p_tx_sw_packet_t packet, uint64_t address, size_t size) 1204 { 1205 struct e1000_hw *hw = &tx_ring->adapter->shared; 1206 p_sw_desc_t desc; 1207 1208 if (hw->mac.type == e1000_82544) { 1209 if (hw->bus.type == e1000_bus_type_pcix) 1210 return (e1000g_tx_workaround_PCIX_82544(packet, 1211 address, size)); 1212 1213 if (size > JUMBO_FRAG_LENGTH) 1214 return (e1000g_tx_workaround_jumbo_82544(packet, 1215 address, size)); 1216 } 1217 1218 ASSERT(packet->num_desc < MAX_TX_DESC_PER_PACKET); 1219 1220 desc = &packet->desc[packet->num_desc]; 1221 desc->address = address; 1222 desc->length = size; 1223 1224 packet->num_desc++; 1225 1226 return (1); 1227 } 1228 1229 static int 1230 e1000g_tx_workaround_PCIX_82544(p_tx_sw_packet_t packet, 1231 uint64_t address, size_t size) 1232 { 1233 p_sw_desc_t desc; 1234 int desc_count; 1235 long size_left; 1236 size_t len; 1237 uint32_t counter; 1238 uint32_t array_elements; 1239 desc_array_t desc_array; 1240 1241 /* 1242 * Coexist Workaround for cordova: RP: 07/04/03 1243 * 1244 * RP: ERRATA: Workaround ISSUE: 1245 * 8kb_buffer_Lockup CONTROLLER: Cordova Breakup 1246 * Eachbuffer in to 8kb pieces until the 1247 * remainder is < 8kb 1248 */ 1249 size_left = size; 1250 desc_count = 0; 1251 1252 while (size_left > 0) { 1253 if (size_left > MAX_TX_BUF_SIZE) 1254 len = MAX_TX_BUF_SIZE; 1255 else 1256 len = size_left; 1257 1258 array_elements = e1000g_fill_82544_desc(address, 1259 len, &desc_array); 1260 1261 for (counter = 0; counter < array_elements; counter++) { 1262 ASSERT(packet->num_desc < MAX_TX_DESC_PER_PACKET); 1263 /* 1264 * Put in the buffer address 1265 */ 1266 desc = &packet->desc[packet->num_desc]; 1267 1268 desc->address = 1269 desc_array.descriptor[counter].address; 1270 desc->length = 1271 desc_array.descriptor[counter].length; 1272 1273 packet->num_desc++; 1274 desc_count++; 1275 } /* for */ 1276 1277 /* 1278 * Update the buffer address and length 1279 */ 1280 address += MAX_TX_BUF_SIZE; 1281 size_left -= MAX_TX_BUF_SIZE; 1282 } /* while */ 1283 1284 return (desc_count); 1285 } 1286 1287 static int 1288 e1000g_tx_workaround_jumbo_82544(p_tx_sw_packet_t packet, 1289 uint64_t address, size_t size) 1290 { 1291 p_sw_desc_t desc; 1292 int desc_count; 1293 long size_left; 1294 uint32_t offset; 1295 1296 /* 1297 * Workaround for Jumbo Frames on Cordova 1298 * PSD 06/01/2001 1299 */ 1300 size_left = size; 1301 desc_count = 0; 1302 offset = 0; 1303 while (size_left > 0) { 1304 ASSERT(packet->num_desc < MAX_TX_DESC_PER_PACKET); 1305 1306 desc = &packet->desc[packet->num_desc]; 1307 1308 desc->address = address + offset; 1309 1310 if (size_left > JUMBO_FRAG_LENGTH) 1311 desc->length = JUMBO_FRAG_LENGTH; 1312 else 1313 desc->length = size_left; 1314 1315 packet->num_desc++; 1316 desc_count++; 1317 1318 offset += desc->length; 1319 size_left -= JUMBO_FRAG_LENGTH; 1320 } 1321 1322 return (desc_count); 1323 } 1324 1325 #pragma inline(e1000g_82547_tx_move_tail_work) 1326 1327 static void 1328 e1000g_82547_tx_move_tail_work(e1000g_tx_ring_t *tx_ring) 1329 { 1330 struct e1000_hw *hw; 1331 uint16_t hw_tdt; 1332 uint16_t sw_tdt; 1333 struct e1000_tx_desc *tx_desc; 1334 uint16_t length = 0; 1335 boolean_t eop = B_FALSE; 1336 struct e1000g *Adapter; 1337 1338 Adapter = tx_ring->adapter; 1339 hw = &Adapter->shared; 1340 1341 hw_tdt = E1000_READ_REG(hw, E1000_TDT); 1342 sw_tdt = tx_ring->tbd_next - tx_ring->tbd_first; 1343 1344 while (hw_tdt != sw_tdt) { 1345 tx_desc = &(tx_ring->tbd_first[hw_tdt]); 1346 length += tx_desc->lower.flags.length; 1347 eop = tx_desc->lower.data & E1000_TXD_CMD_EOP; 1348 if (++hw_tdt == Adapter->tx_desc_num) 1349 hw_tdt = 0; 1350 1351 if (eop) { 1352 if ((Adapter->link_duplex == HALF_DUPLEX) && 1353 (e1000_fifo_workaround_82547(hw, length) 1354 != E1000_SUCCESS)) { 1355 if (tx_ring->timer_enable_82547) { 1356 ASSERT(tx_ring->timer_id_82547 == 0); 1357 tx_ring->timer_id_82547 = 1358 timeout(e1000g_82547_timeout, 1359 (void *)tx_ring, 1360 drv_usectohz(10000)); 1361 } 1362 return; 1363 1364 } else { 1365 E1000_WRITE_REG(hw, E1000_TDT, hw_tdt); 1366 e1000_update_tx_fifo_head_82547(hw, length); 1367 length = 0; 1368 } 1369 } 1370 } 1371 } 1372 1373 static void 1374 e1000g_82547_timeout(void *arg) 1375 { 1376 e1000g_tx_ring_t *tx_ring; 1377 1378 tx_ring = (e1000g_tx_ring_t *)arg; 1379 1380 mutex_enter(&tx_ring->tx_lock); 1381 1382 tx_ring->timer_id_82547 = 0; 1383 e1000g_82547_tx_move_tail_work(tx_ring); 1384 1385 mutex_exit(&tx_ring->tx_lock); 1386 } 1387 1388 static void 1389 e1000g_82547_tx_move_tail(e1000g_tx_ring_t *tx_ring) 1390 { 1391 timeout_id_t tid; 1392 1393 ASSERT(MUTEX_HELD(&tx_ring->tx_lock)); 1394 1395 tid = tx_ring->timer_id_82547; 1396 tx_ring->timer_id_82547 = 0; 1397 if (tid != 0) { 1398 tx_ring->timer_enable_82547 = B_FALSE; 1399 mutex_exit(&tx_ring->tx_lock); 1400 1401 (void) untimeout(tid); 1402 1403 mutex_enter(&tx_ring->tx_lock); 1404 } 1405 tx_ring->timer_enable_82547 = B_TRUE; 1406 e1000g_82547_tx_move_tail_work(tx_ring); 1407 } 1408