xref: /titanic_50/usr/src/uts/common/io/ixgbe/ixgbe_buf.c (revision 91d7f85e02991954d1e1bd44673df567ad8dcc87)
1 /*
2  * CDDL HEADER START
3  *
4  * Copyright(c) 2007-2009 Intel Corporation. All rights reserved.
5  * The contents of this file are subject to the terms of the
6  * Common Development and Distribution License (the "License").
7  * You may not use this file except in compliance with the License.
8  *
9  * You can obtain a copy of the license at:
10  *      http://www.opensolaris.org/os/licensing.
11  * See the License for the specific language governing permissions
12  * and limitations under the License.
13  *
14  * When using or redistributing this file, you may do so under the
15  * License only. No other modification of this header is permitted.
16  *
17  * If applicable, add the following below this CDDL HEADER, with the
18  * fields enclosed by brackets "[]" replaced with your own identifying
19  * information: Portions Copyright [yyyy] [name of copyright owner]
20  *
21  * CDDL HEADER END
22  */
23 
24 /*
25  * Copyright 2010 Sun Microsystems, Inc.  All rights reserved.
26  * Use is subject to license terms.
27  */
28 
29 #include "ixgbe_sw.h"
30 
31 static int ixgbe_alloc_tbd_ring(ixgbe_tx_ring_t *);
32 static void ixgbe_free_tbd_ring(ixgbe_tx_ring_t *);
33 static int ixgbe_alloc_rbd_ring(ixgbe_rx_data_t *);
34 static void ixgbe_free_rbd_ring(ixgbe_rx_data_t *);
35 static int ixgbe_alloc_dma_buffer(ixgbe_t *, dma_buffer_t *, size_t);
36 static int ixgbe_alloc_tcb_lists(ixgbe_tx_ring_t *);
37 static void ixgbe_free_tcb_lists(ixgbe_tx_ring_t *);
38 static int ixgbe_alloc_rcb_lists(ixgbe_rx_data_t *);
39 static void ixgbe_free_rcb_lists(ixgbe_rx_data_t *);
40 
41 #ifdef __sparc
42 #define	IXGBE_DMA_ALIGNMENT	0x0000000000002000ull
43 #else
44 #define	IXGBE_DMA_ALIGNMENT	0x0000000000001000ull
45 #endif
46 
47 /*
48  * DMA attributes for tx/rx descriptors.
49  */
50 static ddi_dma_attr_t ixgbe_desc_dma_attr = {
51 	DMA_ATTR_V0,			/* version number */
52 	0x0000000000000000ull,		/* low address */
53 	0xFFFFFFFFFFFFFFFFull,		/* high address */
54 	0x00000000FFFFFFFFull,		/* dma counter max */
55 	IXGBE_DMA_ALIGNMENT,		/* alignment */
56 	0x00000FFF,			/* burst sizes */
57 	0x00000001,			/* minimum transfer size */
58 	0x00000000FFFFFFFFull,		/* maximum transfer size */
59 	0xFFFFFFFFFFFFFFFFull,		/* maximum segment size */
60 	1,				/* scatter/gather list length */
61 	0x00000001,			/* granularity */
62 	DDI_DMA_FLAGERR			/* DMA flags */
63 };
64 
65 /*
66  * DMA attributes for tx/rx buffers.
67  */
68 static ddi_dma_attr_t ixgbe_buf_dma_attr = {
69 	DMA_ATTR_V0,			/* version number */
70 	0x0000000000000000ull,		/* low address */
71 	0xFFFFFFFFFFFFFFFFull,		/* high address */
72 	0x00000000FFFFFFFFull,		/* dma counter max */
73 	IXGBE_DMA_ALIGNMENT,		/* alignment */
74 	0x00000FFF,			/* burst sizes */
75 	0x00000001,			/* minimum transfer size */
76 	0x00000000FFFFFFFFull,		/* maximum transfer size */
77 	0xFFFFFFFFFFFFFFFFull,		/* maximum segment size	 */
78 	1,				/* scatter/gather list length */
79 	0x00000001,			/* granularity */
80 	DDI_DMA_FLAGERR			/* DMA flags */
81 };
82 
83 /*
84  * DMA attributes for transmit.
85  */
86 static ddi_dma_attr_t ixgbe_tx_dma_attr = {
87 	DMA_ATTR_V0,			/* version number */
88 	0x0000000000000000ull,		/* low address */
89 	0xFFFFFFFFFFFFFFFFull,		/* high address */
90 	0x00000000FFFFFFFFull,		/* dma counter max */
91 	1,				/* alignment */
92 	0x00000FFF,			/* burst sizes */
93 	0x00000001,			/* minimum transfer size */
94 	0x00000000FFFFFFFFull,		/* maximum transfer size */
95 	0xFFFFFFFFFFFFFFFFull,		/* maximum segment size	 */
96 	MAX_COOKIE,			/* scatter/gather list length */
97 	0x00000001,			/* granularity */
98 	DDI_DMA_FLAGERR			/* DMA flags */
99 };
100 
101 /*
102  * DMA access attributes for descriptors.
103  */
104 static ddi_device_acc_attr_t ixgbe_desc_acc_attr = {
105 	DDI_DEVICE_ATTR_V0,
106 	DDI_STRUCTURE_LE_ACC,
107 	DDI_STRICTORDER_ACC
108 };
109 
110 /*
111  * DMA access attributes for buffers.
112  */
113 static ddi_device_acc_attr_t ixgbe_buf_acc_attr = {
114 	DDI_DEVICE_ATTR_V0,
115 	DDI_NEVERSWAP_ACC,
116 	DDI_STRICTORDER_ACC
117 };
118 
119 /*
120  * ixgbe_alloc_dma - Allocate DMA resources for all rx/tx rings.
121  */
122 int
123 ixgbe_alloc_dma(ixgbe_t *ixgbe)
124 {
125 	ixgbe_rx_ring_t	*rx_ring;
126 	ixgbe_rx_data_t *rx_data;
127 	ixgbe_tx_ring_t *tx_ring;
128 	int i;
129 
130 	for (i = 0; i < ixgbe->num_rx_rings; i++) {
131 		/*
132 		 * Allocate receive desciptor ring and control block lists
133 		 */
134 		rx_ring = &ixgbe->rx_rings[i];
135 		rx_data = rx_ring->rx_data;
136 
137 		if (ixgbe_alloc_rbd_ring(rx_data) != IXGBE_SUCCESS)
138 			goto alloc_dma_failure;
139 
140 		if (ixgbe_alloc_rcb_lists(rx_data) != IXGBE_SUCCESS)
141 			goto alloc_dma_failure;
142 	}
143 
144 	for (i = 0; i < ixgbe->num_tx_rings; i++) {
145 		/*
146 		 * Allocate transmit desciptor ring and control block lists
147 		 */
148 		tx_ring = &ixgbe->tx_rings[i];
149 
150 		if (ixgbe_alloc_tbd_ring(tx_ring) != IXGBE_SUCCESS)
151 			goto alloc_dma_failure;
152 
153 		if (ixgbe_alloc_tcb_lists(tx_ring) != IXGBE_SUCCESS)
154 			goto alloc_dma_failure;
155 	}
156 
157 	return (IXGBE_SUCCESS);
158 
159 alloc_dma_failure:
160 	ixgbe_free_dma(ixgbe);
161 
162 	return (IXGBE_FAILURE);
163 }
164 
165 /*
166  * ixgbe_free_dma - Free all the DMA resources of all rx/tx rings.
167  */
168 void
169 ixgbe_free_dma(ixgbe_t *ixgbe)
170 {
171 	ixgbe_rx_ring_t *rx_ring;
172 	ixgbe_rx_data_t *rx_data;
173 	ixgbe_tx_ring_t *tx_ring;
174 	int i;
175 
176 	/*
177 	 * Free DMA resources of rx rings
178 	 */
179 	for (i = 0; i < ixgbe->num_rx_rings; i++) {
180 		rx_ring = &ixgbe->rx_rings[i];
181 		rx_data = rx_ring->rx_data;
182 
183 		ixgbe_free_rbd_ring(rx_data);
184 		ixgbe_free_rcb_lists(rx_data);
185 	}
186 
187 	/*
188 	 * Free DMA resources of tx rings
189 	 */
190 	for (i = 0; i < ixgbe->num_tx_rings; i++) {
191 		tx_ring = &ixgbe->tx_rings[i];
192 		ixgbe_free_tbd_ring(tx_ring);
193 		ixgbe_free_tcb_lists(tx_ring);
194 	}
195 }
196 
197 int
198 ixgbe_alloc_rx_ring_data(ixgbe_rx_ring_t *rx_ring)
199 {
200 	ixgbe_rx_data_t	*rx_data;
201 	ixgbe_t *ixgbe = rx_ring->ixgbe;
202 	uint32_t rcb_count;
203 
204 	/*
205 	 * Allocate memory for software receive rings
206 	 */
207 	rx_data = kmem_zalloc(sizeof (ixgbe_rx_data_t), KM_NOSLEEP);
208 
209 	if (rx_data == NULL) {
210 		ixgbe_error(ixgbe, "Allocate software receive rings failed");
211 		return (IXGBE_FAILURE);
212 	}
213 
214 	rx_data->rx_ring = rx_ring;
215 	mutex_init(&rx_data->recycle_lock, NULL,
216 	    MUTEX_DRIVER, DDI_INTR_PRI(ixgbe->intr_pri));
217 
218 	rx_data->ring_size = ixgbe->rx_ring_size;
219 	rx_data->free_list_size = ixgbe->rx_ring_size;
220 
221 	rx_data->rcb_head = 0;
222 	rx_data->rcb_tail = 0;
223 	rx_data->rcb_free = rx_data->free_list_size;
224 
225 	/*
226 	 * Allocate memory for the work list.
227 	 */
228 	rx_data->work_list = kmem_zalloc(sizeof (rx_control_block_t *) *
229 	    rx_data->ring_size, KM_NOSLEEP);
230 
231 	if (rx_data->work_list == NULL) {
232 		ixgbe_error(ixgbe,
233 		    "Could not allocate memory for rx work list");
234 		goto alloc_rx_data_failure;
235 	}
236 
237 	/*
238 	 * Allocate memory for the free list.
239 	 */
240 	rx_data->free_list = kmem_zalloc(sizeof (rx_control_block_t *) *
241 	    rx_data->free_list_size, KM_NOSLEEP);
242 
243 	if (rx_data->free_list == NULL) {
244 		ixgbe_error(ixgbe,
245 		    "Cound not allocate memory for rx free list");
246 		goto alloc_rx_data_failure;
247 	}
248 
249 	/*
250 	 * Allocate memory for the rx control blocks for work list and
251 	 * free list.
252 	 */
253 	rcb_count = rx_data->ring_size + rx_data->free_list_size;
254 	rx_data->rcb_area =
255 	    kmem_zalloc(sizeof (rx_control_block_t) * rcb_count,
256 	    KM_NOSLEEP);
257 
258 	if (rx_data->rcb_area == NULL) {
259 		ixgbe_error(ixgbe,
260 		    "Cound not allocate memory for rx control blocks");
261 		goto alloc_rx_data_failure;
262 	}
263 
264 	rx_ring->rx_data = rx_data;
265 	return (IXGBE_SUCCESS);
266 
267 alloc_rx_data_failure:
268 	ixgbe_free_rx_ring_data(rx_data);
269 	return (IXGBE_FAILURE);
270 }
271 
272 void
273 ixgbe_free_rx_ring_data(ixgbe_rx_data_t *rx_data)
274 {
275 	uint32_t rcb_count;
276 
277 	if (rx_data == NULL)
278 		return;
279 
280 	ASSERT(rx_data->rcb_pending == 0);
281 
282 	rcb_count = rx_data->ring_size + rx_data->free_list_size;
283 	if (rx_data->rcb_area != NULL) {
284 		kmem_free(rx_data->rcb_area,
285 		    sizeof (rx_control_block_t) * rcb_count);
286 		rx_data->rcb_area = NULL;
287 	}
288 
289 	if (rx_data->work_list != NULL) {
290 		kmem_free(rx_data->work_list,
291 		    sizeof (rx_control_block_t *) * rx_data->ring_size);
292 		rx_data->work_list = NULL;
293 	}
294 
295 	if (rx_data->free_list != NULL) {
296 		kmem_free(rx_data->free_list,
297 		    sizeof (rx_control_block_t *) * rx_data->free_list_size);
298 		rx_data->free_list = NULL;
299 	}
300 
301 	mutex_destroy(&rx_data->recycle_lock);
302 	kmem_free(rx_data, sizeof (ixgbe_rx_data_t));
303 }
304 
305 /*
306  * ixgbe_alloc_tbd_ring - Memory allocation for the tx descriptors of one ring.
307  */
308 static int
309 ixgbe_alloc_tbd_ring(ixgbe_tx_ring_t *tx_ring)
310 {
311 	int ret;
312 	size_t size;
313 	size_t len;
314 	uint_t cookie_num;
315 	dev_info_t *devinfo;
316 	ddi_dma_cookie_t cookie;
317 	ixgbe_t *ixgbe = tx_ring->ixgbe;
318 
319 	devinfo = ixgbe->dip;
320 	size = sizeof (union ixgbe_adv_tx_desc) * tx_ring->ring_size;
321 
322 	/*
323 	 * If tx head write-back is enabled, an extra tbd is allocated
324 	 * to save the head write-back value
325 	 */
326 	if (ixgbe->tx_head_wb_enable) {
327 		size += sizeof (union ixgbe_adv_tx_desc);
328 	}
329 
330 	/*
331 	 * Allocate a DMA handle for the transmit descriptor
332 	 * memory area.
333 	 */
334 	ret = ddi_dma_alloc_handle(devinfo, &ixgbe_desc_dma_attr,
335 	    DDI_DMA_DONTWAIT, NULL,
336 	    &tx_ring->tbd_area.dma_handle);
337 
338 	if (ret != DDI_SUCCESS) {
339 		ixgbe_error(ixgbe,
340 		    "Could not allocate tbd dma handle: %x", ret);
341 		tx_ring->tbd_area.dma_handle = NULL;
342 
343 		return (IXGBE_FAILURE);
344 	}
345 
346 	/*
347 	 * Allocate memory to DMA data to and from the transmit
348 	 * descriptors.
349 	 */
350 	ret = ddi_dma_mem_alloc(tx_ring->tbd_area.dma_handle,
351 	    size, &ixgbe_desc_acc_attr, DDI_DMA_CONSISTENT,
352 	    DDI_DMA_DONTWAIT, NULL,
353 	    (caddr_t *)&tx_ring->tbd_area.address,
354 	    &len, &tx_ring->tbd_area.acc_handle);
355 
356 	if (ret != DDI_SUCCESS) {
357 		ixgbe_error(ixgbe,
358 		    "Could not allocate tbd dma memory: %x", ret);
359 		tx_ring->tbd_area.acc_handle = NULL;
360 		tx_ring->tbd_area.address = NULL;
361 		if (tx_ring->tbd_area.dma_handle != NULL) {
362 			ddi_dma_free_handle(&tx_ring->tbd_area.dma_handle);
363 			tx_ring->tbd_area.dma_handle = NULL;
364 		}
365 		return (IXGBE_FAILURE);
366 	}
367 
368 	/*
369 	 * Initialize the entire transmit buffer descriptor area to zero
370 	 */
371 	bzero(tx_ring->tbd_area.address, len);
372 
373 	/*
374 	 * Allocates DMA resources for the memory that was allocated by
375 	 * the ddi_dma_mem_alloc call. The DMA resources then get bound to the
376 	 * the memory address
377 	 */
378 	ret = ddi_dma_addr_bind_handle(tx_ring->tbd_area.dma_handle,
379 	    NULL, (caddr_t)tx_ring->tbd_area.address,
380 	    len, DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
381 	    DDI_DMA_DONTWAIT, NULL, &cookie, &cookie_num);
382 
383 	if (ret != DDI_DMA_MAPPED) {
384 		ixgbe_error(ixgbe,
385 		    "Could not bind tbd dma resource: %x", ret);
386 		tx_ring->tbd_area.dma_address = NULL;
387 		if (tx_ring->tbd_area.acc_handle != NULL) {
388 			ddi_dma_mem_free(&tx_ring->tbd_area.acc_handle);
389 			tx_ring->tbd_area.acc_handle = NULL;
390 			tx_ring->tbd_area.address = NULL;
391 		}
392 		if (tx_ring->tbd_area.dma_handle != NULL) {
393 			ddi_dma_free_handle(&tx_ring->tbd_area.dma_handle);
394 			tx_ring->tbd_area.dma_handle = NULL;
395 		}
396 		return (IXGBE_FAILURE);
397 	}
398 
399 	ASSERT(cookie_num == 1);
400 
401 	tx_ring->tbd_area.dma_address = cookie.dmac_laddress;
402 	tx_ring->tbd_area.size = len;
403 
404 	tx_ring->tbd_ring = (union ixgbe_adv_tx_desc *)(uintptr_t)
405 	    tx_ring->tbd_area.address;
406 
407 	return (IXGBE_SUCCESS);
408 }
409 
410 /*
411  * ixgbe_free_tbd_ring - Free the tx descriptors of one ring.
412  */
413 static void
414 ixgbe_free_tbd_ring(ixgbe_tx_ring_t *tx_ring)
415 {
416 	if (tx_ring->tbd_area.dma_handle != NULL) {
417 		(void) ddi_dma_unbind_handle(tx_ring->tbd_area.dma_handle);
418 	}
419 	if (tx_ring->tbd_area.acc_handle != NULL) {
420 		ddi_dma_mem_free(&tx_ring->tbd_area.acc_handle);
421 		tx_ring->tbd_area.acc_handle = NULL;
422 	}
423 	if (tx_ring->tbd_area.dma_handle != NULL) {
424 		ddi_dma_free_handle(&tx_ring->tbd_area.dma_handle);
425 		tx_ring->tbd_area.dma_handle = NULL;
426 	}
427 	tx_ring->tbd_area.address = NULL;
428 	tx_ring->tbd_area.dma_address = NULL;
429 	tx_ring->tbd_area.size = 0;
430 
431 	tx_ring->tbd_ring = NULL;
432 }
433 
434 /*
435  * ixgbe_alloc_rbd_ring - Memory allocation for the rx descriptors of one ring.
436  */
437 static int
438 ixgbe_alloc_rbd_ring(ixgbe_rx_data_t *rx_data)
439 {
440 	int ret;
441 	size_t size;
442 	size_t len;
443 	uint_t cookie_num;
444 	dev_info_t *devinfo;
445 	ddi_dma_cookie_t cookie;
446 	ixgbe_t *ixgbe = rx_data->rx_ring->ixgbe;
447 
448 	devinfo = ixgbe->dip;
449 	size = sizeof (union ixgbe_adv_rx_desc) * rx_data->ring_size;
450 
451 	/*
452 	 * Allocate a new DMA handle for the receive descriptor
453 	 * memory area.
454 	 */
455 	ret = ddi_dma_alloc_handle(devinfo, &ixgbe_desc_dma_attr,
456 	    DDI_DMA_DONTWAIT, NULL,
457 	    &rx_data->rbd_area.dma_handle);
458 
459 	if (ret != DDI_SUCCESS) {
460 		ixgbe_error(ixgbe,
461 		    "Could not allocate rbd dma handle: %x", ret);
462 		rx_data->rbd_area.dma_handle = NULL;
463 		return (IXGBE_FAILURE);
464 	}
465 
466 	/*
467 	 * Allocate memory to DMA data to and from the receive
468 	 * descriptors.
469 	 */
470 	ret = ddi_dma_mem_alloc(rx_data->rbd_area.dma_handle,
471 	    size, &ixgbe_desc_acc_attr, DDI_DMA_CONSISTENT,
472 	    DDI_DMA_DONTWAIT, NULL,
473 	    (caddr_t *)&rx_data->rbd_area.address,
474 	    &len, &rx_data->rbd_area.acc_handle);
475 
476 	if (ret != DDI_SUCCESS) {
477 		ixgbe_error(ixgbe,
478 		    "Could not allocate rbd dma memory: %x", ret);
479 		rx_data->rbd_area.acc_handle = NULL;
480 		rx_data->rbd_area.address = NULL;
481 		if (rx_data->rbd_area.dma_handle != NULL) {
482 			ddi_dma_free_handle(&rx_data->rbd_area.dma_handle);
483 			rx_data->rbd_area.dma_handle = NULL;
484 		}
485 		return (IXGBE_FAILURE);
486 	}
487 
488 	/*
489 	 * Initialize the entire transmit buffer descriptor area to zero
490 	 */
491 	bzero(rx_data->rbd_area.address, len);
492 
493 	/*
494 	 * Allocates DMA resources for the memory that was allocated by
495 	 * the ddi_dma_mem_alloc call.
496 	 */
497 	ret = ddi_dma_addr_bind_handle(rx_data->rbd_area.dma_handle,
498 	    NULL, (caddr_t)rx_data->rbd_area.address,
499 	    len, DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
500 	    DDI_DMA_DONTWAIT, NULL, &cookie, &cookie_num);
501 
502 	if (ret != DDI_DMA_MAPPED) {
503 		ixgbe_error(ixgbe,
504 		    "Could not bind rbd dma resource: %x", ret);
505 		rx_data->rbd_area.dma_address = NULL;
506 		if (rx_data->rbd_area.acc_handle != NULL) {
507 			ddi_dma_mem_free(&rx_data->rbd_area.acc_handle);
508 			rx_data->rbd_area.acc_handle = NULL;
509 			rx_data->rbd_area.address = NULL;
510 		}
511 		if (rx_data->rbd_area.dma_handle != NULL) {
512 			ddi_dma_free_handle(&rx_data->rbd_area.dma_handle);
513 			rx_data->rbd_area.dma_handle = NULL;
514 		}
515 		return (IXGBE_FAILURE);
516 	}
517 
518 	ASSERT(cookie_num == 1);
519 
520 	rx_data->rbd_area.dma_address = cookie.dmac_laddress;
521 	rx_data->rbd_area.size = len;
522 
523 	rx_data->rbd_ring = (union ixgbe_adv_rx_desc *)(uintptr_t)
524 	    rx_data->rbd_area.address;
525 
526 	return (IXGBE_SUCCESS);
527 }
528 
529 /*
530  * ixgbe_free_rbd_ring - Free the rx descriptors of one ring.
531  */
532 static void
533 ixgbe_free_rbd_ring(ixgbe_rx_data_t *rx_data)
534 {
535 	if (rx_data->rbd_area.dma_handle != NULL) {
536 		(void) ddi_dma_unbind_handle(rx_data->rbd_area.dma_handle);
537 	}
538 	if (rx_data->rbd_area.acc_handle != NULL) {
539 		ddi_dma_mem_free(&rx_data->rbd_area.acc_handle);
540 		rx_data->rbd_area.acc_handle = NULL;
541 	}
542 	if (rx_data->rbd_area.dma_handle != NULL) {
543 		ddi_dma_free_handle(&rx_data->rbd_area.dma_handle);
544 		rx_data->rbd_area.dma_handle = NULL;
545 	}
546 	rx_data->rbd_area.address = NULL;
547 	rx_data->rbd_area.dma_address = NULL;
548 	rx_data->rbd_area.size = 0;
549 
550 	rx_data->rbd_ring = NULL;
551 }
552 
553 /*
554  * ixgbe_alloc_dma_buffer - Allocate DMA resources for a DMA buffer.
555  */
556 static int
557 ixgbe_alloc_dma_buffer(ixgbe_t *ixgbe, dma_buffer_t *buf, size_t size)
558 {
559 	int ret;
560 	dev_info_t *devinfo = ixgbe->dip;
561 	ddi_dma_cookie_t cookie;
562 	size_t len;
563 	uint_t cookie_num;
564 
565 	ret = ddi_dma_alloc_handle(devinfo,
566 	    &ixgbe_buf_dma_attr, DDI_DMA_DONTWAIT,
567 	    NULL, &buf->dma_handle);
568 
569 	if (ret != DDI_SUCCESS) {
570 		buf->dma_handle = NULL;
571 		ixgbe_error(ixgbe,
572 		    "Could not allocate dma buffer handle: %x", ret);
573 		return (IXGBE_FAILURE);
574 	}
575 
576 	ret = ddi_dma_mem_alloc(buf->dma_handle,
577 	    size, &ixgbe_buf_acc_attr, DDI_DMA_STREAMING,
578 	    DDI_DMA_DONTWAIT, NULL, &buf->address,
579 	    &len, &buf->acc_handle);
580 
581 	if (ret != DDI_SUCCESS) {
582 		buf->acc_handle = NULL;
583 		buf->address = NULL;
584 		if (buf->dma_handle != NULL) {
585 			ddi_dma_free_handle(&buf->dma_handle);
586 			buf->dma_handle = NULL;
587 		}
588 		ixgbe_error(ixgbe,
589 		    "Could not allocate dma buffer memory: %x", ret);
590 		return (IXGBE_FAILURE);
591 	}
592 
593 	ret = ddi_dma_addr_bind_handle(buf->dma_handle, NULL,
594 	    buf->address,
595 	    len, DDI_DMA_RDWR | DDI_DMA_STREAMING,
596 	    DDI_DMA_DONTWAIT, NULL, &cookie, &cookie_num);
597 
598 	if (ret != DDI_DMA_MAPPED) {
599 		buf->dma_address = NULL;
600 		if (buf->acc_handle != NULL) {
601 			ddi_dma_mem_free(&buf->acc_handle);
602 			buf->acc_handle = NULL;
603 			buf->address = NULL;
604 		}
605 		if (buf->dma_handle != NULL) {
606 			ddi_dma_free_handle(&buf->dma_handle);
607 			buf->dma_handle = NULL;
608 		}
609 		ixgbe_error(ixgbe,
610 		    "Could not bind dma buffer handle: %x", ret);
611 		return (IXGBE_FAILURE);
612 	}
613 
614 	ASSERT(cookie_num == 1);
615 
616 	buf->dma_address = cookie.dmac_laddress;
617 	buf->size = len;
618 	buf->len = 0;
619 
620 	return (IXGBE_SUCCESS);
621 }
622 
623 /*
624  * ixgbe_free_dma_buffer - Free one allocated area of dma memory and handle.
625  */
626 void
627 ixgbe_free_dma_buffer(dma_buffer_t *buf)
628 {
629 	if (buf->dma_handle != NULL) {
630 		(void) ddi_dma_unbind_handle(buf->dma_handle);
631 		buf->dma_address = NULL;
632 	} else {
633 		return;
634 	}
635 
636 	if (buf->acc_handle != NULL) {
637 		ddi_dma_mem_free(&buf->acc_handle);
638 		buf->acc_handle = NULL;
639 		buf->address = NULL;
640 	}
641 
642 	if (buf->dma_handle != NULL) {
643 		ddi_dma_free_handle(&buf->dma_handle);
644 		buf->dma_handle = NULL;
645 	}
646 
647 	buf->size = 0;
648 	buf->len = 0;
649 }
650 
651 /*
652  * ixgbe_alloc_tcb_lists - Memory allocation for the transmit control bolcks
653  * of one ring.
654  */
655 static int
656 ixgbe_alloc_tcb_lists(ixgbe_tx_ring_t *tx_ring)
657 {
658 	int i;
659 	int ret;
660 	tx_control_block_t *tcb;
661 	dma_buffer_t *tx_buf;
662 	ixgbe_t *ixgbe = tx_ring->ixgbe;
663 	dev_info_t *devinfo = ixgbe->dip;
664 
665 	/*
666 	 * Allocate memory for the work list.
667 	 */
668 	tx_ring->work_list = kmem_zalloc(sizeof (tx_control_block_t *) *
669 	    tx_ring->ring_size, KM_NOSLEEP);
670 
671 	if (tx_ring->work_list == NULL) {
672 		ixgbe_error(ixgbe,
673 		    "Cound not allocate memory for tx work list");
674 		return (IXGBE_FAILURE);
675 	}
676 
677 	/*
678 	 * Allocate memory for the free list.
679 	 */
680 	tx_ring->free_list = kmem_zalloc(sizeof (tx_control_block_t *) *
681 	    tx_ring->free_list_size, KM_NOSLEEP);
682 
683 	if (tx_ring->free_list == NULL) {
684 		kmem_free(tx_ring->work_list,
685 		    sizeof (tx_control_block_t *) * tx_ring->ring_size);
686 		tx_ring->work_list = NULL;
687 
688 		ixgbe_error(ixgbe,
689 		    "Cound not allocate memory for tx free list");
690 		return (IXGBE_FAILURE);
691 	}
692 
693 	/*
694 	 * Allocate memory for the tx control blocks of free list.
695 	 */
696 	tx_ring->tcb_area =
697 	    kmem_zalloc(sizeof (tx_control_block_t) *
698 	    tx_ring->free_list_size, KM_NOSLEEP);
699 
700 	if (tx_ring->tcb_area == NULL) {
701 		kmem_free(tx_ring->work_list,
702 		    sizeof (tx_control_block_t *) * tx_ring->ring_size);
703 		tx_ring->work_list = NULL;
704 
705 		kmem_free(tx_ring->free_list,
706 		    sizeof (tx_control_block_t *) * tx_ring->free_list_size);
707 		tx_ring->free_list = NULL;
708 
709 		ixgbe_error(ixgbe,
710 		    "Cound not allocate memory for tx control blocks");
711 		return (IXGBE_FAILURE);
712 	}
713 
714 	/*
715 	 * Allocate dma memory for the tx control block of free list.
716 	 */
717 	tcb = tx_ring->tcb_area;
718 	for (i = 0; i < tx_ring->free_list_size; i++, tcb++) {
719 		ASSERT(tcb != NULL);
720 
721 		tx_ring->free_list[i] = tcb;
722 
723 		/*
724 		 * Pre-allocate dma handles for transmit. These dma handles
725 		 * will be dynamically bound to the data buffers passed down
726 		 * from the upper layers at the time of transmitting.
727 		 */
728 		ret = ddi_dma_alloc_handle(devinfo,
729 		    &ixgbe_tx_dma_attr,
730 		    DDI_DMA_DONTWAIT, NULL,
731 		    &tcb->tx_dma_handle);
732 		if (ret != DDI_SUCCESS) {
733 			tcb->tx_dma_handle = NULL;
734 			ixgbe_error(ixgbe,
735 			    "Could not allocate tx dma handle: %x", ret);
736 			goto alloc_tcb_lists_fail;
737 		}
738 
739 		/*
740 		 * Pre-allocate transmit buffers for packets that the
741 		 * size is less than bcopy_thresh.
742 		 */
743 		tx_buf = &tcb->tx_buf;
744 
745 		ret = ixgbe_alloc_dma_buffer(ixgbe,
746 		    tx_buf, ixgbe->tx_buf_size);
747 
748 		if (ret != IXGBE_SUCCESS) {
749 			ASSERT(tcb->tx_dma_handle != NULL);
750 			ddi_dma_free_handle(&tcb->tx_dma_handle);
751 			tcb->tx_dma_handle = NULL;
752 			ixgbe_error(ixgbe, "Allocate tx dma buffer failed");
753 			goto alloc_tcb_lists_fail;
754 		}
755 
756 		tcb->last_index = MAX_TX_RING_SIZE;
757 	}
758 
759 	return (IXGBE_SUCCESS);
760 
761 alloc_tcb_lists_fail:
762 	ixgbe_free_tcb_lists(tx_ring);
763 
764 	return (IXGBE_FAILURE);
765 }
766 
767 /*
768  * ixgbe_free_tcb_lists - Release the memory allocated for
769  * the transmit control bolcks of one ring.
770  */
771 static void
772 ixgbe_free_tcb_lists(ixgbe_tx_ring_t *tx_ring)
773 {
774 	int i;
775 	tx_control_block_t *tcb;
776 
777 	tcb = tx_ring->tcb_area;
778 	if (tcb == NULL)
779 		return;
780 
781 	for (i = 0; i < tx_ring->free_list_size; i++, tcb++) {
782 		ASSERT(tcb != NULL);
783 
784 		/* Free the tx dma handle for dynamical binding */
785 		if (tcb->tx_dma_handle != NULL) {
786 			ddi_dma_free_handle(&tcb->tx_dma_handle);
787 			tcb->tx_dma_handle = NULL;
788 		} else {
789 			/*
790 			 * If the dma handle is NULL, then we don't
791 			 * have to check the remaining.
792 			 */
793 			break;
794 		}
795 
796 		ixgbe_free_dma_buffer(&tcb->tx_buf);
797 	}
798 
799 	if (tx_ring->tcb_area != NULL) {
800 		kmem_free(tx_ring->tcb_area,
801 		    sizeof (tx_control_block_t) * tx_ring->free_list_size);
802 		tx_ring->tcb_area = NULL;
803 	}
804 
805 	if (tx_ring->work_list != NULL) {
806 		kmem_free(tx_ring->work_list,
807 		    sizeof (tx_control_block_t *) * tx_ring->ring_size);
808 		tx_ring->work_list = NULL;
809 	}
810 
811 	if (tx_ring->free_list != NULL) {
812 		kmem_free(tx_ring->free_list,
813 		    sizeof (tx_control_block_t *) * tx_ring->free_list_size);
814 		tx_ring->free_list = NULL;
815 	}
816 }
817 
818 /*
819  * ixgbe_alloc_rcb_lists - Memory allocation for the receive control blocks
820  * of one ring.
821  */
822 static int
823 ixgbe_alloc_rcb_lists(ixgbe_rx_data_t *rx_data)
824 {
825 	int i;
826 	int ret;
827 	rx_control_block_t *rcb;
828 	ixgbe_t *ixgbe = rx_data->rx_ring->ixgbe;
829 	dma_buffer_t *rx_buf;
830 	uint32_t rcb_count;
831 
832 	/*
833 	 * Allocate memory for the rx control blocks for work list and
834 	 * free list.
835 	 */
836 	rcb_count = rx_data->ring_size + rx_data->free_list_size;
837 	rcb = rx_data->rcb_area;
838 
839 	for (i = 0; i < rcb_count; i++, rcb++) {
840 		ASSERT(rcb != NULL);
841 
842 		if (i < rx_data->ring_size) {
843 			/* Attach the rx control block to the work list */
844 			rx_data->work_list[i] = rcb;
845 		} else {
846 			/* Attach the rx control block to the free list */
847 			rx_data->free_list[i - rx_data->ring_size] = rcb;
848 		}
849 
850 		rx_buf = &rcb->rx_buf;
851 		ret = ixgbe_alloc_dma_buffer(ixgbe,
852 		    rx_buf, ixgbe->rx_buf_size);
853 
854 		if (ret != IXGBE_SUCCESS) {
855 			ixgbe_error(ixgbe, "Allocate rx dma buffer failed");
856 			goto alloc_rcb_lists_fail;
857 		}
858 
859 		rx_buf->size -= IPHDR_ALIGN_ROOM;
860 		rx_buf->address += IPHDR_ALIGN_ROOM;
861 		rx_buf->dma_address += IPHDR_ALIGN_ROOM;
862 
863 		rcb->ref_cnt = 1;
864 		rcb->rx_data = (ixgbe_rx_data_t *)rx_data;
865 		rcb->free_rtn.free_func = ixgbe_rx_recycle;
866 		rcb->free_rtn.free_arg = (char *)rcb;
867 		rcb->lro_prev = -1;
868 		rcb->lro_next = -1;
869 		rcb->lro_pkt = B_FALSE;
870 		rcb->mp = desballoc((unsigned char *)
871 		    rx_buf->address,
872 		    rx_buf->size,
873 		    0, &rcb->free_rtn);
874 	}
875 
876 	return (IXGBE_SUCCESS);
877 
878 alloc_rcb_lists_fail:
879 	ixgbe_free_rcb_lists(rx_data);
880 
881 	return (IXGBE_FAILURE);
882 }
883 
884 /*
885  * ixgbe_free_rcb_lists - Free the receive control blocks of one ring.
886  */
887 static void
888 ixgbe_free_rcb_lists(ixgbe_rx_data_t *rx_data)
889 {
890 	ixgbe_t *ixgbe;
891 	rx_control_block_t *rcb;
892 	uint32_t rcb_count;
893 	uint32_t ref_cnt;
894 	int i;
895 
896 	ixgbe = rx_data->rx_ring->ixgbe;
897 
898 	mutex_enter(&ixgbe->rx_pending_lock);
899 
900 	rcb = rx_data->rcb_area;
901 	rcb_count = rx_data->ring_size + rx_data->free_list_size;
902 
903 	for (i = 0; i < rcb_count; i++, rcb++) {
904 		ASSERT(rcb != NULL);
905 
906 		ref_cnt = atomic_dec_32_nv(&rcb->ref_cnt);
907 		if (ref_cnt == 0) {
908 			if (rcb->mp != NULL) {
909 				freemsg(rcb->mp);
910 				rcb->mp = NULL;
911 			}
912 			ixgbe_free_dma_buffer(&rcb->rx_buf);
913 		} else {
914 			atomic_inc_32(&rx_data->rcb_pending);
915 			atomic_inc_32(&ixgbe->rcb_pending);
916 		}
917 	}
918 
919 	mutex_exit(&ixgbe->rx_pending_lock);
920 }
921 
922 /*
923  * ixgbe_set_fma_flags - Set the attribute for fma support.
924  */
925 void
926 ixgbe_set_fma_flags(int dma_flag)
927 {
928 	if (dma_flag) {
929 		ixgbe_tx_dma_attr.dma_attr_flags = DDI_DMA_FLAGERR;
930 		ixgbe_buf_dma_attr.dma_attr_flags = DDI_DMA_FLAGERR;
931 		ixgbe_desc_dma_attr.dma_attr_flags = DDI_DMA_FLAGERR;
932 	} else {
933 		ixgbe_tx_dma_attr.dma_attr_flags = 0;
934 		ixgbe_buf_dma_attr.dma_attr_flags = 0;
935 		ixgbe_desc_dma_attr.dma_attr_flags = 0;
936 	}
937 }
938