xref: /freebsd/sys/contrib/vchiq/interface/vchiq_arm/vchiq_2835_arm.c (revision 39ee7a7a6bdd1557b1c3532abf60d139798ac88b)
1 /**
2  * Copyright (c) 2010-2012 Broadcom. All rights reserved.
3  *
4  * Redistribution and use in source and binary forms, with or without
5  * modification, are permitted provided that the following conditions
6  * are met:
7  * 1. Redistributions of source code must retain the above copyright
8  *    notice, this list of conditions, and the following disclaimer,
9  *    without modification.
10  * 2. Redistributions in binary form must reproduce the above copyright
11  *    notice, this list of conditions and the following disclaimer in the
12  *    documentation and/or other materials provided with the distribution.
13  * 3. The names of the above-listed copyright holders may not be used
14  *    to endorse or promote products derived from this software without
15  *    specific prior written permission.
16  *
17  * ALTERNATIVELY, this software may be distributed under the terms of the
18  * GNU General Public License ("GPL") version 2, as published by the Free
19  * Software Foundation.
20  *
21  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS
22  * IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
23  * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
24  * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
25  * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
26  * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
27  * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
28  * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
29  * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
30  * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
31  * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
32  */
33 
34 #include <interface/compat/vchi_bsd.h>
35 
36 #include <sys/malloc.h>
37 #include <sys/rwlock.h>
38 
39 #include <vm/vm.h>
40 #include <vm/pmap.h>
41 #include <vm/vm_extern.h>
42 #include <vm/vm_kern.h>
43 #include <vm/vm_map.h>
44 #include <vm/vm_object.h>
45 #include <vm/vm_page.h>
46 #include <vm/vm_pager.h>
47 #include <vm/vm_param.h>
48 #include <vm/vm_phys.h>
49 
50 #include <machine/bus.h>
51 #include <arm/broadcom/bcm2835/bcm2835_mbox.h>
52 #include <arm/broadcom/bcm2835/bcm2835_vcbus.h>
53 
54 MALLOC_DEFINE(M_VCPAGELIST, "vcpagelist", "VideoCore pagelist memory");
55 
56 #define TOTAL_SLOTS (VCHIQ_SLOT_ZERO_SLOTS + 2 * 32)
57 
58 #define VCHIQ_DOORBELL_IRQ IRQ_ARM_DOORBELL_0
59 #define VCHIQ_ARM_ADDRESS(x) ((void *)PHYS_TO_VCBUS(pmap_kextract((vm_offset_t)(x))))
60 
61 #include "vchiq_arm.h"
62 #include "vchiq_2835.h"
63 #include "vchiq_connected.h"
64 
65 #define MAX_FRAGMENTS (VCHIQ_NUM_CURRENT_BULKS * 2)
66 
67 typedef struct vchiq_2835_state_struct {
68    int inited;
69    VCHIQ_ARM_STATE_T arm_state;
70 } VCHIQ_2835_ARM_STATE_T;
71 
72 static char *g_slot_mem;
73 static int g_slot_mem_size;
74 vm_paddr_t g_slot_phys;
75 /* BSD DMA */
76 bus_dma_tag_t bcm_slots_dma_tag;
77 bus_dmamap_t bcm_slots_dma_map;
78 
79 static FRAGMENTS_T *g_fragments_base;
80 static FRAGMENTS_T *g_free_fragments;
81 struct semaphore g_free_fragments_sema;
82 
83 static DEFINE_SEMAPHORE(g_free_fragments_mutex);
84 
85 typedef struct bulkinfo_struct {
86 	PAGELIST_T	*pagelist;
87 	bus_dma_tag_t	pagelist_dma_tag;
88 	bus_dmamap_t	pagelist_dma_map;
89 	void		*buf;
90 	size_t		size;
91 } BULKINFO_T;
92 
93 static int
94 create_pagelist(char __user *buf, size_t count, unsigned short type,
95                 struct proc *p, BULKINFO_T *bi);
96 
97 static void
98 free_pagelist(BULKINFO_T *bi, int actual);
99 
100 static void
101 vchiq_dmamap_cb(void *arg, bus_dma_segment_t *segs, int nseg, int err)
102 {
103 	bus_addr_t *addr;
104 
105 	if (err)
106 		return;
107 
108 	addr = (bus_addr_t*)arg;
109 	*addr = PHYS_TO_VCBUS(segs[0].ds_addr);
110 }
111 
112 int __init
113 vchiq_platform_init(VCHIQ_STATE_T *state)
114 {
115 	VCHIQ_SLOT_ZERO_T *vchiq_slot_zero;
116 	int frag_mem_size;
117 	int err;
118 	int i;
119 
120 	/* Allocate space for the channels in coherent memory */
121 	g_slot_mem_size = PAGE_ALIGN(TOTAL_SLOTS * VCHIQ_SLOT_SIZE);
122 	frag_mem_size = PAGE_ALIGN(sizeof(FRAGMENTS_T) * MAX_FRAGMENTS);
123 
124 	err = bus_dma_tag_create(
125 	    NULL,
126 	    PAGE_SIZE, 0,	       /* alignment, boundary */
127 	    BUS_SPACE_MAXADDR_32BIT,    /* lowaddr */
128 	    BUS_SPACE_MAXADDR,	  /* highaddr */
129 	    NULL, NULL,		 /* filter, filterarg */
130 	    g_slot_mem_size + frag_mem_size, 1,		/* maxsize, nsegments */
131 	    g_slot_mem_size + frag_mem_size, 0,		/* maxsegsize, flags */
132 	    NULL, NULL,		 /* lockfunc, lockarg */
133 	    &bcm_slots_dma_tag);
134 
135 	err = bus_dmamem_alloc(bcm_slots_dma_tag, (void **)&g_slot_mem,
136 	    BUS_DMA_COHERENT | BUS_DMA_WAITOK, &bcm_slots_dma_map);
137 	if (err) {
138 		vchiq_log_error(vchiq_core_log_level, "Unable to allocate channel memory");
139 		err = -ENOMEM;
140 		goto failed_alloc;
141 	}
142 
143 	err = bus_dmamap_load(bcm_slots_dma_tag, bcm_slots_dma_map, g_slot_mem,
144 	    g_slot_mem_size + frag_mem_size, vchiq_dmamap_cb,
145 	    &g_slot_phys, 0);
146 
147 	if (err) {
148 		vchiq_log_error(vchiq_core_log_level, "cannot load DMA map");
149 		err = -ENOMEM;
150 		goto failed_load;
151 	}
152 
153 	WARN_ON(((int)g_slot_mem & (PAGE_SIZE - 1)) != 0);
154 
155 	vchiq_slot_zero = vchiq_init_slots(g_slot_mem, g_slot_mem_size);
156 	if (!vchiq_slot_zero) {
157 		err = -EINVAL;
158 		goto failed_init_slots;
159 	}
160 
161 	vchiq_slot_zero->platform_data[VCHIQ_PLATFORM_FRAGMENTS_OFFSET_IDX] =
162 		(int)g_slot_phys + g_slot_mem_size;
163 	vchiq_slot_zero->platform_data[VCHIQ_PLATFORM_FRAGMENTS_COUNT_IDX] =
164 		MAX_FRAGMENTS;
165 
166 	g_fragments_base = (FRAGMENTS_T *)(g_slot_mem + g_slot_mem_size);
167 	g_slot_mem_size += frag_mem_size;
168 
169 	g_free_fragments = g_fragments_base;
170 	for (i = 0; i < (MAX_FRAGMENTS - 1); i++) {
171 		*(FRAGMENTS_T **)&g_fragments_base[i] =
172 			&g_fragments_base[i + 1];
173 	}
174 	*(FRAGMENTS_T **)&g_fragments_base[i] = NULL;
175 	_sema_init(&g_free_fragments_sema, MAX_FRAGMENTS);
176 
177 	if (vchiq_init_state(state, vchiq_slot_zero, 0/*slave*/) !=
178 		VCHIQ_SUCCESS) {
179 		err = -EINVAL;
180 		goto failed_vchiq_init;
181 	}
182 
183 	bcm_mbox_write(BCM2835_MBOX_CHAN_VCHIQ, (unsigned int)g_slot_phys);
184 
185 	vchiq_log_info(vchiq_arm_log_level,
186 		"vchiq_init - done (slots %x, phys %x)",
187 		(unsigned int)vchiq_slot_zero, g_slot_phys);
188 
189    vchiq_call_connected_callbacks();
190 
191    return 0;
192 
193 failed_vchiq_init:
194 failed_init_slots:
195 failed_load:
196 	bus_dmamap_unload(bcm_slots_dma_tag, bcm_slots_dma_map);
197 failed_alloc:
198 	bus_dmamap_destroy(bcm_slots_dma_tag, bcm_slots_dma_map);
199 	bus_dma_tag_destroy(bcm_slots_dma_tag);
200 
201    return err;
202 }
203 
204 void __exit
205 vchiq_platform_exit(VCHIQ_STATE_T *state)
206 {
207 
208 	bus_dmamap_unload(bcm_slots_dma_tag, bcm_slots_dma_map);
209 	bus_dmamap_destroy(bcm_slots_dma_tag, bcm_slots_dma_map);
210 	bus_dma_tag_destroy(bcm_slots_dma_tag);
211 }
212 
213 VCHIQ_STATUS_T
214 vchiq_platform_init_state(VCHIQ_STATE_T *state)
215 {
216    VCHIQ_STATUS_T status = VCHIQ_SUCCESS;
217    state->platform_state = kzalloc(sizeof(VCHIQ_2835_ARM_STATE_T), GFP_KERNEL);
218    ((VCHIQ_2835_ARM_STATE_T*)state->platform_state)->inited = 1;
219    status = vchiq_arm_init_state(state, &((VCHIQ_2835_ARM_STATE_T*)state->platform_state)->arm_state);
220    if(status != VCHIQ_SUCCESS)
221    {
222       ((VCHIQ_2835_ARM_STATE_T*)state->platform_state)->inited = 0;
223    }
224    return status;
225 }
226 
227 VCHIQ_ARM_STATE_T*
228 vchiq_platform_get_arm_state(VCHIQ_STATE_T *state)
229 {
230    if(!((VCHIQ_2835_ARM_STATE_T*)state->platform_state)->inited)
231    {
232       BUG();
233    }
234    return &((VCHIQ_2835_ARM_STATE_T*)state->platform_state)->arm_state;
235 }
236 
237 int
238 vchiq_copy_from_user(void *dst, const void *src, int size)
239 {
240 
241 	if (((vm_offset_t)(src)) < VM_MIN_KERNEL_ADDRESS) {
242 		int error = copyin(src, dst, size);
243 		return error ? VCHIQ_ERROR : VCHIQ_SUCCESS;
244 	}
245 	else
246 		bcopy(src, dst, size);
247 
248 	return 0;
249 }
250 
251 VCHIQ_STATUS_T
252 vchiq_prepare_bulk_data(VCHIQ_BULK_T *bulk, VCHI_MEM_HANDLE_T memhandle,
253 	void *offset, int size, int dir)
254 {
255 	BULKINFO_T *bi;
256 	int ret;
257 
258 	WARN_ON(memhandle != VCHI_MEM_HANDLE_INVALID);
259 	bi = malloc(sizeof(*bi), M_VCPAGELIST, M_WAITOK | M_ZERO);
260 	if (bi == NULL)
261 		return VCHIQ_ERROR;
262 
263 	ret = create_pagelist((char __user *)offset, size,
264 			(dir == VCHIQ_BULK_RECEIVE)
265 			? PAGELIST_READ
266 			: PAGELIST_WRITE,
267 			current,
268 			bi);
269 	if (ret != 0)
270 		return VCHIQ_ERROR;
271 
272 	bulk->handle = memhandle;
273 	bulk->data = VCHIQ_ARM_ADDRESS(bi->pagelist);
274 
275 	/* Store the pagelist address in remote_data, which isn't used by the
276 	   slave. */
277 	bulk->remote_data = bi;
278 
279 	return VCHIQ_SUCCESS;
280 }
281 
282 void
283 vchiq_complete_bulk(VCHIQ_BULK_T *bulk)
284 {
285 	if (bulk && bulk->remote_data && bulk->actual)
286 		free_pagelist((BULKINFO_T *)bulk->remote_data, bulk->actual);
287 }
288 
289 void
290 vchiq_transfer_bulk(VCHIQ_BULK_T *bulk)
291 {
292 	/*
293 	 * This should only be called on the master (VideoCore) side, but
294 	 * provide an implementation to avoid the need for ifdefery.
295 	 */
296 	BUG();
297 }
298 
299 void
300 vchiq_dump_platform_state(void *dump_context)
301 {
302 	char buf[80];
303 	int len;
304 	len = snprintf(buf, sizeof(buf),
305 		"  Platform: 2835 (VC master)");
306 	vchiq_dump(dump_context, buf, len + 1);
307 }
308 
309 VCHIQ_STATUS_T
310 vchiq_platform_suspend(VCHIQ_STATE_T *state)
311 {
312    return VCHIQ_ERROR;
313 }
314 
315 VCHIQ_STATUS_T
316 vchiq_platform_resume(VCHIQ_STATE_T *state)
317 {
318    return VCHIQ_SUCCESS;
319 }
320 
321 void
322 vchiq_platform_paused(VCHIQ_STATE_T *state)
323 {
324 }
325 
326 void
327 vchiq_platform_resumed(VCHIQ_STATE_T *state)
328 {
329 }
330 
331 int
332 vchiq_platform_videocore_wanted(VCHIQ_STATE_T* state)
333 {
334    return 1; // autosuspend not supported - videocore always wanted
335 }
336 
337 int
338 vchiq_platform_use_suspend_timer(void)
339 {
340    return 0;
341 }
342 void
343 vchiq_dump_platform_use_state(VCHIQ_STATE_T *state)
344 {
345 	vchiq_log_info(vchiq_arm_log_level, "Suspend timer not in use");
346 }
347 void
348 vchiq_platform_handle_timeout(VCHIQ_STATE_T *state)
349 {
350 	(void)state;
351 }
352 /*
353  * Local functions
354  */
355 
356 static void
357 pagelist_page_free(vm_page_t pp)
358 {
359 	vm_page_lock(pp);
360 	vm_page_unwire(pp, PQ_INACTIVE);
361 	if (pp->wire_count == 0 && pp->object == NULL)
362 		vm_page_free(pp);
363 	vm_page_unlock(pp);
364 }
365 
366 /* There is a potential problem with partial cache lines (pages?)
367 ** at the ends of the block when reading. If the CPU accessed anything in
368 ** the same line (page?) then it may have pulled old data into the cache,
369 ** obscuring the new data underneath. We can solve this by transferring the
370 ** partial cache lines separately, and allowing the ARM to copy into the
371 ** cached area.
372 
373 ** N.B. This implementation plays slightly fast and loose with the Linux
374 ** driver programming rules, e.g. its use of __virt_to_bus instead of
375 ** dma_map_single, but it isn't a multi-platform driver and it benefits
376 ** from increased speed as a result.
377 */
378 
379 static int
380 create_pagelist(char __user *buf, size_t count, unsigned short type,
381 	struct proc *p, BULKINFO_T *bi)
382 {
383 	PAGELIST_T *pagelist;
384 	vm_page_t* pages;
385 	unsigned long *addrs;
386 	unsigned int num_pages, i;
387 	vm_offset_t offset;
388 	int pagelist_size;
389 	char *addr, *base_addr, *next_addr;
390 	int run, addridx, actual_pages;
391 	int err;
392 	vm_paddr_t pagelist_phys;
393 
394 	offset = (vm_offset_t)buf & (PAGE_SIZE - 1);
395 	num_pages = (count + offset + PAGE_SIZE - 1) / PAGE_SIZE;
396 
397 	bi->pagelist = NULL;
398 	bi->buf = buf;
399 	bi->size = count;
400 
401 	/* Allocate enough storage to hold the page pointers and the page
402 	** list
403 	*/
404 	pagelist_size = sizeof(PAGELIST_T) +
405 		(num_pages * sizeof(unsigned long)) +
406 		(num_pages * sizeof(pages[0]));
407 
408 	err = bus_dma_tag_create(
409 	    NULL,
410 	    PAGE_SIZE, 0,	       /* alignment, boundary */
411 	    BUS_SPACE_MAXADDR_32BIT,    /* lowaddr */
412 	    BUS_SPACE_MAXADDR,	  /* highaddr */
413 	    NULL, NULL,		 /* filter, filterarg */
414 	    pagelist_size, 1,		/* maxsize, nsegments */
415 	    pagelist_size, 0,		/* maxsegsize, flags */
416 	    NULL, NULL,		 /* lockfunc, lockarg */
417 	    &bi->pagelist_dma_tag);
418 
419 	err = bus_dmamem_alloc(bi->pagelist_dma_tag, (void **)&pagelist,
420 	    BUS_DMA_COHERENT | BUS_DMA_WAITOK, &bi->pagelist_dma_map);
421 	if (err) {
422 		vchiq_log_error(vchiq_core_log_level, "Unable to allocate pagelist memory");
423 		err = -ENOMEM;
424 		goto failed_alloc;
425 	}
426 
427 	err = bus_dmamap_load(bi->pagelist_dma_tag, bi->pagelist_dma_map, pagelist,
428 	    pagelist_size, vchiq_dmamap_cb,
429 	    &pagelist_phys, 0);
430 
431 	if (err) {
432 		vchiq_log_error(vchiq_core_log_level, "cannot load DMA map for pagelist memory");
433 		err = -ENOMEM;
434 		goto failed_load;
435 	}
436 
437 	vchiq_log_trace(vchiq_arm_log_level,
438 		"create_pagelist - %x", (unsigned int)pagelist);
439 	if (!pagelist)
440 		return -ENOMEM;
441 
442 	addrs = pagelist->addrs;
443 	pages = (vm_page_t*)(addrs + num_pages);
444 
445 	actual_pages = vm_fault_quick_hold_pages(&p->p_vmspace->vm_map,
446 	    (vm_offset_t)buf, count,
447 	    (type == PAGELIST_READ ? VM_PROT_WRITE : 0 ) | VM_PROT_READ, pages, num_pages);
448 
449 	if (actual_pages != num_pages) {
450 		vm_page_unhold_pages(pages, actual_pages);
451 		free(pagelist, M_VCPAGELIST);
452 		return (-ENOMEM);
453 	}
454 
455 	for (i = 0; i < actual_pages; i++) {
456 		vm_page_lock(pages[i]);
457 		vm_page_wire(pages[i]);
458 		vm_page_unhold(pages[i]);
459 		vm_page_unlock(pages[i]);
460 	}
461 
462 	pagelist->length = count;
463 	pagelist->type = type;
464 	pagelist->offset = offset;
465 
466 	/* Group the pages into runs of contiguous pages */
467 
468 	base_addr = (void *)PHYS_TO_VCBUS(VM_PAGE_TO_PHYS(pages[0]));
469 	next_addr = base_addr + PAGE_SIZE;
470 	addridx = 0;
471 	run = 0;
472 
473 	for (i = 1; i < num_pages; i++) {
474 		addr = (void *)PHYS_TO_VCBUS(VM_PAGE_TO_PHYS(pages[i]));
475 		if ((addr == next_addr) && (run < (PAGE_SIZE - 1))) {
476 			next_addr += PAGE_SIZE;
477 			run++;
478 		} else {
479 			addrs[addridx] = (unsigned long)base_addr + run;
480 			addridx++;
481 			base_addr = addr;
482 			next_addr = addr + PAGE_SIZE;
483 			run = 0;
484 		}
485 	}
486 
487 	addrs[addridx] = (unsigned long)base_addr + run;
488 	addridx++;
489 
490 	/* Partial cache lines (fragments) require special measures */
491 	if ((type == PAGELIST_READ) &&
492 		((pagelist->offset & (CACHE_LINE_SIZE - 1)) ||
493 		((pagelist->offset + pagelist->length) &
494 		(CACHE_LINE_SIZE - 1)))) {
495 		FRAGMENTS_T *fragments;
496 
497 		if (down_interruptible(&g_free_fragments_sema) != 0) {
498       			free(pagelist, M_VCPAGELIST);
499 			return -EINTR;
500 		}
501 
502 		WARN_ON(g_free_fragments == NULL);
503 
504 		down(&g_free_fragments_mutex);
505 		fragments = (FRAGMENTS_T *) g_free_fragments;
506 		WARN_ON(fragments == NULL);
507 		g_free_fragments = *(FRAGMENTS_T **) g_free_fragments;
508 		up(&g_free_fragments_mutex);
509 		pagelist->type =
510 			 PAGELIST_READ_WITH_FRAGMENTS + (fragments -
511 							 g_fragments_base);
512 	}
513 
514 	cpu_dcache_wbinv_range((vm_offset_t)buf, count);
515 
516 	bus_dmamap_sync(bi->pagelist_dma_tag, bi->pagelist_dma_map, BUS_DMASYNC_PREWRITE);
517 
518 	bi->pagelist = pagelist;
519 
520 	return 0;
521 
522 failed_load:
523 	bus_dmamap_unload(bi->pagelist_dma_tag, bi->pagelist_dma_map);
524 failed_alloc:
525 	bus_dmamap_destroy(bi->pagelist_dma_tag, bi->pagelist_dma_map);
526 	bus_dma_tag_destroy(bi->pagelist_dma_tag);
527 
528 	return err;
529 }
530 
531 static void
532 free_pagelist(BULKINFO_T *bi, int actual)
533 {
534 	vm_page_t*pages;
535 	unsigned int num_pages, i;
536 	PAGELIST_T *pagelist;
537 
538 	pagelist = bi->pagelist;
539 
540 	vchiq_log_trace(vchiq_arm_log_level,
541 		"free_pagelist - %x, %d", (unsigned int)pagelist, actual);
542 
543 	num_pages =
544 		(pagelist->length + pagelist->offset + PAGE_SIZE - 1) /
545 		PAGE_SIZE;
546 
547 	pages = (vm_page_t*)(pagelist->addrs + num_pages);
548 
549 	/* Deal with any partial cache lines (fragments) */
550 	if (pagelist->type >= PAGELIST_READ_WITH_FRAGMENTS) {
551 		FRAGMENTS_T *fragments = g_fragments_base +
552 			(pagelist->type - PAGELIST_READ_WITH_FRAGMENTS);
553 		int head_bytes, tail_bytes;
554 		head_bytes = (CACHE_LINE_SIZE - pagelist->offset) &
555 			(CACHE_LINE_SIZE - 1);
556 		tail_bytes = (pagelist->offset + actual) &
557 			(CACHE_LINE_SIZE - 1);
558 
559 		if ((actual >= 0) && (head_bytes != 0)) {
560 			if (head_bytes > actual)
561 				head_bytes = actual;
562 
563 			memcpy((char *)bi->buf,
564 				fragments->headbuf,
565 				head_bytes);
566 		}
567 
568 		if ((actual >= 0) && (head_bytes < actual) &&
569 			(tail_bytes != 0)) {
570 			memcpy((char *)bi->buf + actual - tail_bytes,
571 					 fragments->tailbuf, tail_bytes);
572 		}
573 
574 		down(&g_free_fragments_mutex);
575 		*(FRAGMENTS_T **) fragments = g_free_fragments;
576 		g_free_fragments = fragments;
577 		up(&g_free_fragments_mutex);
578 		up(&g_free_fragments_sema);
579 	}
580 
581 	for (i = 0; i < num_pages; i++) {
582 		if (pagelist->type != PAGELIST_WRITE) {
583 			vm_page_dirty(pages[i]);
584 			pagelist_page_free(pages[i]);
585 		}
586 	}
587 
588 	bus_dmamap_unload(bi->pagelist_dma_tag, bi->pagelist_dma_map);
589 	bus_dmamem_free(bi->pagelist_dma_tag, bi->pagelist, bi->pagelist_dma_map);
590 	bus_dmamap_destroy(bi->pagelist_dma_tag, bi->pagelist_dma_map);
591 	bus_dma_tag_destroy(bi->pagelist_dma_tag);
592 
593 	free(bi, M_VCPAGELIST);
594 }
595