xref: /freebsd/sys/dev/hpt27xx/hpt27xx_osm_bsd.c (revision 43a5ec4eb41567cc92586503212743d89686d78f)
1 /*-
2  * SPDX-License-Identifier: BSD-2-Clause-FreeBSD
3  *
4  * Copyright (c) 2011 HighPoint Technologies, Inc.
5  * All rights reserved.
6  *
7  * Redistribution and use in source and binary forms, with or without
8  * modification, are permitted provided that the following conditions
9  * are met:
10  * 1. Redistributions of source code must retain the above copyright
11  *    notice, this list of conditions and the following disclaimer.
12  * 2. Redistributions in binary form must reproduce the above copyright
13  *    notice, this list of conditions and the following disclaimer in the
14  *    documentation and/or other materials provided with the distribution.
15  *
16  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
17  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
18  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
19  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
20  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
21  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
22  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
23  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
24  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
25  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
26  * SUCH DAMAGE.
27  *
28  * $FreeBSD$
29  */
30 
31 #include <dev/hpt27xx/hpt27xx_config.h>
32 
33 #include <dev/hpt27xx/os_bsd.h>
34 #include <dev/hpt27xx/hptintf.h>
35 
36 static HIM *hpt_match(device_t dev, int scan)
37 {
38 	PCI_ID pci_id;
39 	HIM *him;
40 	int i;
41 
42 	for (him = him_list; him; him = him->next) {
43 		for (i=0; him->get_supported_device_id(i, &pci_id); i++) {
44 			if (scan && him->get_controller_count)
45 				him->get_controller_count(&pci_id,0,0);
46 			if ((pci_get_vendor(dev) == pci_id.vid) &&
47 				(pci_get_device(dev) == pci_id.did)){
48 				return (him);
49 			}
50 		}
51 	}
52 	return (NULL);
53 }
54 
55 static int hpt_probe(device_t dev)
56 {
57 	HIM *him;
58 
59 	him = hpt_match(dev, 0);
60 	if (him != NULL) {
61 		KdPrint(("hpt_probe: adapter at PCI %d:%d:%d, IRQ %d",
62 			pci_get_bus(dev), pci_get_slot(dev), pci_get_function(dev), pci_get_irq(dev)
63 			));
64 		device_set_desc(dev, him->name);
65 		return (BUS_PROBE_DEFAULT);
66 	}
67 
68 	return (ENXIO);
69 }
70 
71 static int hpt_attach(device_t dev)
72 {
73 	PHBA hba = (PHBA)device_get_softc(dev);
74 	HIM *him;
75 	PCI_ID pci_id;
76 	HPT_UINT size;
77 	PVBUS vbus;
78 	PVBUS_EXT vbus_ext;
79 
80 	KdPrint(("hpt_attach(%d/%d/%d)", pci_get_bus(dev), pci_get_slot(dev), pci_get_function(dev)));
81 
82 	him = hpt_match(dev, 1);
83 	hba->ext_type = EXT_TYPE_HBA;
84 	hba->ldm_adapter.him = him;
85 	pci_enable_busmaster(dev);
86 
87 	pci_id.vid = pci_get_vendor(dev);
88 	pci_id.did = pci_get_device(dev);
89 	pci_id.rev = pci_get_revid(dev);
90 	pci_id.subsys = (HPT_U32)(pci_get_subdevice(dev)) << 16 | pci_get_subvendor(dev);
91 
92 	size = him->get_adapter_size(&pci_id);
93 	hba->ldm_adapter.him_handle = malloc(size, M_DEVBUF, M_WAITOK);
94 	if (!hba->ldm_adapter.him_handle)
95 		return ENXIO;
96 
97 	hba->pcidev = dev;
98 	hba->pciaddr.tree = 0;
99 	hba->pciaddr.bus = pci_get_bus(dev);
100 	hba->pciaddr.device = pci_get_slot(dev);
101 	hba->pciaddr.function = pci_get_function(dev);
102 
103 	if (!him->create_adapter(&pci_id, hba->pciaddr, hba->ldm_adapter.him_handle, hba)) {
104 		free(hba->ldm_adapter.him_handle, M_DEVBUF);
105 		return ENXIO;
106 	}
107 
108 	os_printk("adapter at PCI %d:%d:%d, IRQ %d",
109 		hba->pciaddr.bus, hba->pciaddr.device, hba->pciaddr.function, pci_get_irq(dev));
110 
111 	if (!ldm_register_adapter(&hba->ldm_adapter)) {
112 		size = ldm_get_vbus_size();
113 		vbus_ext = malloc(sizeof(VBUS_EXT) + size, M_DEVBUF, M_WAITOK);
114 		if (!vbus_ext) {
115 			free(hba->ldm_adapter.him_handle, M_DEVBUF);
116 			return ENXIO;
117 		}
118 		memset(vbus_ext, 0, sizeof(VBUS_EXT));
119 		vbus_ext->ext_type = EXT_TYPE_VBUS;
120 		ldm_create_vbus((PVBUS)vbus_ext->vbus, vbus_ext);
121 		ldm_register_adapter(&hba->ldm_adapter);
122 	}
123 
124 	ldm_for_each_vbus(vbus, vbus_ext) {
125 		if (hba->ldm_adapter.vbus==vbus) {
126 			hba->vbus_ext = vbus_ext;
127 			hba->next = vbus_ext->hba_list;
128 			vbus_ext->hba_list = hba;
129 			break;
130 		}
131 	}
132 	return 0;
133 }
134 
135 /*
136  * Maybe we'd better to use the bus_dmamem_alloc to alloc DMA memory,
137  * but there are some problems currently (alignment, etc).
138  */
139 static __inline void *__get_free_pages(int order)
140 {
141 	/* don't use low memory - other devices may get starved */
142 	return contigmalloc(PAGE_SIZE<<order,
143 			M_DEVBUF, M_WAITOK, BUS_SPACE_MAXADDR_24BIT, BUS_SPACE_MAXADDR, PAGE_SIZE, 0);
144 }
145 
146 static __inline void free_pages(void *p, int order)
147 {
148 	contigfree(p, PAGE_SIZE<<order, M_DEVBUF);
149 }
150 
151 static int hpt_alloc_mem(PVBUS_EXT vbus_ext)
152 {
153 	PHBA hba;
154 	struct freelist *f;
155 	HPT_UINT i;
156 	void **p;
157 
158 	for (hba = vbus_ext->hba_list; hba; hba = hba->next)
159 		hba->ldm_adapter.him->get_meminfo(hba->ldm_adapter.him_handle);
160 
161 	ldm_get_mem_info((PVBUS)vbus_ext->vbus, 0);
162 
163 	for (f=vbus_ext->freelist_head; f; f=f->next) {
164 		KdPrint(("%s: %d*%d=%d bytes",
165 			f->tag, f->count, f->size, f->count*f->size));
166 		for (i=0; i<f->count; i++) {
167 			p = (void **)malloc(f->size, M_DEVBUF, M_WAITOK);
168 			if (!p)	return (ENXIO);
169 			*p = f->head;
170 			f->head = p;
171 		}
172 	}
173 
174 	for (f=vbus_ext->freelist_dma_head; f; f=f->next) {
175 		int order, size, j;
176 
177 		HPT_ASSERT((f->size & (f->alignment-1))==0);
178 
179 		for (order=0, size=PAGE_SIZE; size<f->size; order++, size<<=1)
180 			;
181 
182 		KdPrint(("%s: %d*%d=%d bytes, order %d",
183 			f->tag, f->count, f->size, f->count*f->size, order));
184 		HPT_ASSERT(f->alignment<=PAGE_SIZE);
185 
186 		for (i=0; i<f->count;) {
187 			p = (void **)__get_free_pages(order);
188 			if (!p) return -1;
189 			for (j = size/f->size; j && i<f->count; i++,j--) {
190 				*p = f->head;
191 				*(BUS_ADDRESS *)(p+1) = (BUS_ADDRESS)vtophys(p);
192 				f->head = p;
193 				p = (void **)((unsigned long)p + f->size);
194 			}
195 		}
196 	}
197 
198 	HPT_ASSERT(PAGE_SIZE==DMAPOOL_PAGE_SIZE);
199 
200 	for (i=0; i<os_max_cache_pages; i++) {
201 		p = (void **)__get_free_pages(0);
202 		if (!p) return -1;
203 		HPT_ASSERT(((HPT_UPTR)p & (DMAPOOL_PAGE_SIZE-1))==0);
204 		dmapool_put_page((PVBUS)vbus_ext->vbus, p, (BUS_ADDRESS)vtophys(p));
205 	}
206 
207 	return 0;
208 }
209 
210 static void hpt_free_mem(PVBUS_EXT vbus_ext)
211 {
212 	struct freelist *f;
213 	void *p;
214 	int i;
215 	BUS_ADDRESS bus;
216 
217 	for (f=vbus_ext->freelist_head; f; f=f->next) {
218 #if DBG
219 		if (f->count!=f->reserved_count) {
220 			KdPrint(("memory leak for freelist %s (%d/%d)", f->tag, f->count, f->reserved_count));
221 		}
222 #endif
223 		while ((p=freelist_get(f)))
224 			free(p, M_DEVBUF);
225 	}
226 
227 	for (i=0; i<os_max_cache_pages; i++) {
228 		p = dmapool_get_page((PVBUS)vbus_ext->vbus, &bus);
229 		HPT_ASSERT(p);
230 		free_pages(p, 0);
231 	}
232 
233 	for (f=vbus_ext->freelist_dma_head; f; f=f->next) {
234 		int order, size;
235 #if DBG
236 		if (f->count!=f->reserved_count) {
237 			KdPrint(("memory leak for dma freelist %s (%d/%d)", f->tag, f->count, f->reserved_count));
238 		}
239 #endif
240 		for (order=0, size=PAGE_SIZE; size<f->size; order++, size<<=1) ;
241 
242 		while ((p=freelist_get_dma(f, &bus))) {
243 			if (order)
244 				free_pages(p, order);
245 			else {
246 			/* can't free immediately since other blocks in this page may still be in the list */
247 				if (((HPT_UPTR)p & (PAGE_SIZE-1))==0)
248 					dmapool_put_page((PVBUS)vbus_ext->vbus, p, bus);
249 			}
250 		}
251 	}
252 
253 	while ((p = dmapool_get_page((PVBUS)vbus_ext->vbus, &bus)))
254 		free_pages(p, 0);
255 }
256 
257 static int hpt_init_vbus(PVBUS_EXT vbus_ext)
258 {
259 	PHBA hba;
260 
261 	for (hba = vbus_ext->hba_list; hba; hba = hba->next)
262 		if (!hba->ldm_adapter.him->initialize(hba->ldm_adapter.him_handle)) {
263 			KdPrint(("fail to initialize %p", hba));
264 			return -1;
265 		}
266 
267 	ldm_initialize_vbus((PVBUS)vbus_ext->vbus, &vbus_ext->hba_list->ldm_adapter);
268 	return 0;
269 }
270 
271 static void hpt_flush_done(PCOMMAND pCmd)
272 {
273 	PVDEV vd = pCmd->target;
274 
275 	if (mIsArray(vd->type) && vd->u.array.transform && vd!=vd->u.array.transform->target) {
276 		vd = vd->u.array.transform->target;
277 		HPT_ASSERT(vd);
278 		pCmd->target = vd;
279 		pCmd->Result = RETURN_PENDING;
280 		vdev_queue_cmd(pCmd);
281 		return;
282 	}
283 
284 	*(int *)pCmd->priv = 1;
285 	wakeup(pCmd);
286 }
287 
288 /*
289  * flush a vdev (without retry).
290  */
291 static int hpt_flush_vdev(PVBUS_EXT vbus_ext, PVDEV vd)
292 {
293 	PCOMMAND pCmd;
294 	int result = 0, done;
295 	HPT_UINT count;
296 
297 	KdPrint(("flusing dev %p", vd));
298 
299 	hpt_lock_vbus(vbus_ext);
300 
301 	if (mIsArray(vd->type) && vd->u.array.transform)
302 		count = max(vd->u.array.transform->source->cmds_per_request,
303 					vd->u.array.transform->target->cmds_per_request);
304 	else
305 		count = vd->cmds_per_request;
306 
307 	pCmd = ldm_alloc_cmds(vd->vbus, count);
308 
309 	if (!pCmd) {
310 		hpt_unlock_vbus(vbus_ext);
311 		return -1;
312 	}
313 
314 	pCmd->type = CMD_TYPE_FLUSH;
315 	pCmd->flags.hard_flush = 1;
316 	pCmd->target = vd;
317 	pCmd->done = hpt_flush_done;
318 	done = 0;
319 	pCmd->priv = &done;
320 
321 	ldm_queue_cmd(pCmd);
322 
323 	if (!done) {
324 		while (hpt_sleep(vbus_ext, pCmd, PPAUSE, "hptfls", HPT_OSM_TIMEOUT)) {
325 			ldm_reset_vbus(vd->vbus);
326 		}
327 	}
328 
329 	KdPrint(("flush result %d", pCmd->Result));
330 
331 	if (pCmd->Result!=RETURN_SUCCESS)
332 		result = -1;
333 
334 	ldm_free_cmds(pCmd);
335 
336 	hpt_unlock_vbus(vbus_ext);
337 
338 	return result;
339 }
340 
341 static void hpt_stop_tasks(PVBUS_EXT vbus_ext);
342 static void hpt_shutdown_vbus(PVBUS_EXT vbus_ext, int howto)
343 {
344 	PVBUS     vbus = (PVBUS)vbus_ext->vbus;
345 	PHBA hba;
346 	int i;
347 
348 	KdPrint(("hpt_shutdown_vbus"));
349 
350 	/* stop all ctl tasks and disable the worker taskqueue */
351 	hpt_stop_tasks(vbus_ext);
352 	vbus_ext->worker.ta_context = 0;
353 
354 	/* flush devices */
355 	for (i=0; i<osm_max_targets; i++) {
356 		PVDEV vd = ldm_find_target(vbus, i);
357 		if (vd) {
358 			/* retry once */
359 			if (hpt_flush_vdev(vbus_ext, vd))
360 				hpt_flush_vdev(vbus_ext, vd);
361 		}
362 	}
363 
364 	hpt_lock_vbus(vbus_ext);
365 	ldm_shutdown(vbus);
366 	hpt_unlock_vbus(vbus_ext);
367 
368 	ldm_release_vbus(vbus);
369 
370 	for (hba=vbus_ext->hba_list; hba; hba=hba->next)
371 		bus_teardown_intr(hba->pcidev, hba->irq_res, hba->irq_handle);
372 
373 	hpt_free_mem(vbus_ext);
374 
375 	while ((hba=vbus_ext->hba_list)) {
376 		vbus_ext->hba_list = hba->next;
377 		free(hba->ldm_adapter.him_handle, M_DEVBUF);
378 	}
379 	callout_drain(&vbus_ext->timer);
380 	mtx_destroy(&vbus_ext->lock);
381 	free(vbus_ext, M_DEVBUF);
382 	KdPrint(("hpt_shutdown_vbus done"));
383 }
384 
385 static void __hpt_do_tasks(PVBUS_EXT vbus_ext)
386 {
387 	OSM_TASK *tasks;
388 
389 	tasks = vbus_ext->tasks;
390 	vbus_ext->tasks = 0;
391 
392 	while (tasks) {
393 		OSM_TASK *t = tasks;
394 		tasks = t->next;
395 		t->next = 0;
396 		t->func(vbus_ext->vbus, t->data);
397 	}
398 }
399 
400 static void hpt_do_tasks(PVBUS_EXT vbus_ext, int pending)
401 {
402 	if(vbus_ext){
403 		hpt_lock_vbus(vbus_ext);
404 		__hpt_do_tasks(vbus_ext);
405 		hpt_unlock_vbus(vbus_ext);
406 	}
407 }
408 
409 static void hpt_action(struct cam_sim *sim, union ccb *ccb);
410 static void hpt_poll(struct cam_sim *sim);
411 static void hpt_async(void * callback_arg, u_int32_t code, struct cam_path * path, void * arg);
412 static void hpt_pci_intr(void *arg);
413 
414 static __inline POS_CMDEXT cmdext_get(PVBUS_EXT vbus_ext)
415 {
416 	POS_CMDEXT p = vbus_ext->cmdext_list;
417 	if (p)
418 		vbus_ext->cmdext_list = p->next;
419 	return p;
420 }
421 
422 static __inline void cmdext_put(POS_CMDEXT p)
423 {
424 	p->next = p->vbus_ext->cmdext_list;
425 	p->vbus_ext->cmdext_list = p;
426 }
427 
428 static void hpt_timeout(void *arg)
429 {
430 	PCOMMAND pCmd = (PCOMMAND)arg;
431 	POS_CMDEXT ext = (POS_CMDEXT)pCmd->priv;
432 
433 	KdPrint(("pCmd %p timeout", pCmd));
434 
435 	ldm_reset_vbus((PVBUS)ext->vbus_ext->vbus);
436 }
437 
438 static void os_cmddone(PCOMMAND pCmd)
439 {
440 	POS_CMDEXT ext = (POS_CMDEXT)pCmd->priv;
441 	union ccb *ccb = ext->ccb;
442 
443 	KdPrint(("<8>os_cmddone(%p, %d)", pCmd, pCmd->Result));
444 	callout_stop(&ext->timeout);
445 	switch(pCmd->Result) {
446 	case RETURN_SUCCESS:
447 		ccb->ccb_h.status = CAM_REQ_CMP;
448 		break;
449 	case RETURN_BAD_DEVICE:
450 		ccb->ccb_h.status = CAM_DEV_NOT_THERE;
451 		break;
452 	case RETURN_DEVICE_BUSY:
453 		ccb->ccb_h.status = CAM_BUSY;
454 		break;
455 	case RETURN_INVALID_REQUEST:
456 		ccb->ccb_h.status = CAM_REQ_INVALID;
457 		break;
458 	case RETURN_SELECTION_TIMEOUT:
459 		ccb->ccb_h.status = CAM_SEL_TIMEOUT;
460 		break;
461 	case RETURN_RETRY:
462 		ccb->ccb_h.status = CAM_BUSY;
463 		break;
464 	default:
465 		ccb->ccb_h.status = CAM_SCSI_STATUS_ERROR;
466 		break;
467 	}
468 
469 	if (pCmd->flags.data_in) {
470 		bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map, BUS_DMASYNC_POSTREAD);
471 	}
472 	else if (pCmd->flags.data_out) {
473 		bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map, BUS_DMASYNC_POSTWRITE);
474 	}
475 
476 	bus_dmamap_unload(ext->vbus_ext->io_dmat, ext->dma_map);
477 
478 	cmdext_put(ext);
479 	ldm_free_cmds(pCmd);
480 	xpt_done(ccb);
481 }
482 
483 static int os_buildsgl(PCOMMAND pCmd, PSG pSg, int logical)
484 {
485 	POS_CMDEXT ext = (POS_CMDEXT)pCmd->priv;
486 	union ccb *ccb = ext->ccb;
487 
488 	if(logical)	{
489 		os_set_sgptr(pSg, (HPT_U8 *)ccb->csio.data_ptr);
490 		pSg->size = ccb->csio.dxfer_len;
491 		pSg->eot = 1;
492 		return TRUE;
493 	}
494 	/* since we have provided physical sg, nobody will ask us to build physical sg */
495 	HPT_ASSERT(0);
496 	return FALSE;
497 }
498 
499 static void hpt_io_dmamap_callback(void *arg, bus_dma_segment_t *segs, int nsegs, int error)
500 {
501 	PCOMMAND pCmd = (PCOMMAND)arg;
502 	POS_CMDEXT ext = (POS_CMDEXT)pCmd->priv;
503 	PSG psg = pCmd->psg;
504 	int idx;
505 
506 	HPT_ASSERT(pCmd->flags.physical_sg);
507 
508 	if (error)
509 		panic("busdma error");
510 
511 	HPT_ASSERT(nsegs<=os_max_sg_descriptors);
512 
513 	if (nsegs != 0) {
514 		for (idx = 0; idx < nsegs; idx++, psg++) {
515 			psg->addr.bus = segs[idx].ds_addr;
516 			psg->size = segs[idx].ds_len;
517 			psg->eot = 0;
518 		}
519 		psg[-1].eot = 1;
520 
521 		if (pCmd->flags.data_in) {
522 			bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map,
523 			    BUS_DMASYNC_PREREAD);
524 		}
525 		else if (pCmd->flags.data_out) {
526 			bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map,
527 			    BUS_DMASYNC_PREWRITE);
528 		}
529 	}
530 	callout_reset(&ext->timeout, HPT_OSM_TIMEOUT, hpt_timeout, pCmd);
531 	ldm_queue_cmd(pCmd);
532 }
533 
534 static void hpt_scsi_io(PVBUS_EXT vbus_ext, union ccb *ccb)
535 {
536 	PVBUS vbus = (PVBUS)vbus_ext->vbus;
537 	PVDEV vd;
538 	PCOMMAND pCmd;
539 	POS_CMDEXT ext;
540 	HPT_U8 *cdb;
541 	int error;
542 
543 	if (ccb->ccb_h.flags & CAM_CDB_POINTER)
544 		cdb = ccb->csio.cdb_io.cdb_ptr;
545 	else
546 		cdb = ccb->csio.cdb_io.cdb_bytes;
547 
548 	KdPrint(("<8>hpt_scsi_io: ccb %x id %d lun %d cdb %x-%x-%x",
549 		ccb,
550 		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
551 		*(HPT_U32 *)&cdb[0], *(HPT_U32 *)&cdb[4], *(HPT_U32 *)&cdb[8]
552 	));
553 
554 	/* ccb->ccb_h.path_id is not our bus id - don't check it */
555 	if (ccb->ccb_h.target_lun != 0 ||
556 		ccb->ccb_h.target_id >= osm_max_targets ||
557 		(ccb->ccb_h.flags & CAM_CDB_PHYS))
558 	{
559 		ccb->ccb_h.status = CAM_TID_INVALID;
560 		xpt_done(ccb);
561 		return;
562 	}
563 
564 	vd = ldm_find_target(vbus, ccb->ccb_h.target_id);
565 
566 	if (!vd) {
567 		ccb->ccb_h.status = CAM_SEL_TIMEOUT;
568 		xpt_done(ccb);
569 		return;
570 	}
571 
572 	switch (cdb[0]) {
573 	case TEST_UNIT_READY:
574 	case START_STOP_UNIT:
575 	case SYNCHRONIZE_CACHE:
576 		ccb->ccb_h.status = CAM_REQ_CMP;
577 		break;
578 
579 	case INQUIRY:
580 		{
581 			PINQUIRYDATA inquiryData;
582 			memset(ccb->csio.data_ptr, 0, ccb->csio.dxfer_len);
583 			inquiryData = (PINQUIRYDATA)ccb->csio.data_ptr;
584 
585 			inquiryData->AdditionalLength = 31;
586 			inquiryData->CommandQueue = 1;
587 			memcpy(&inquiryData->VendorId, "HPT     ", 8);
588 			memcpy(&inquiryData->ProductId, "DISK 0_0        ", 16);
589 
590 			if (vd->target_id / 10) {
591 				inquiryData->ProductId[7] = (vd->target_id % 100) / 10 + '0';
592 				inquiryData->ProductId[8] = (vd->target_id % 100) % 10 + '0';
593 			}
594 			else
595 				inquiryData->ProductId[7] = (vd->target_id % 100) % 10 + '0';
596 
597 			memcpy(&inquiryData->ProductRevisionLevel, "4.00", 4);
598 
599 			ccb->ccb_h.status = CAM_REQ_CMP;
600 		}
601 		break;
602 
603 	case READ_CAPACITY:
604 	{
605 		HPT_U8 *rbuf = ccb->csio.data_ptr;
606 		HPT_U32 cap;
607 		HPT_U8 sector_size_shift = 0;
608 		HPT_U64 new_cap;
609 		HPT_U32 sector_size = 0;
610 
611 		if (mIsArray(vd->type))
612 			sector_size_shift = vd->u.array.sector_size_shift;
613 		else{
614 			if(vd->type == VD_RAW){
615 				sector_size = vd->u.raw.logical_sector_size;
616 			}
617 
618 			switch (sector_size) {
619 				case 0x1000:
620 					KdPrint(("set 4k setctor size in READ_CAPACITY"));
621 					sector_size_shift = 3;
622 					break;
623 				default:
624 					break;
625 			}
626 		}
627 		new_cap = vd->capacity >> sector_size_shift;
628 
629 		if (new_cap > 0xfffffffful)
630 			cap = 0xffffffff;
631 		else
632 			cap = new_cap - 1;
633 
634 		rbuf[0] = (HPT_U8)(cap>>24);
635 		rbuf[1] = (HPT_U8)(cap>>16);
636 		rbuf[2] = (HPT_U8)(cap>>8);
637 		rbuf[3] = (HPT_U8)cap;
638 		rbuf[4] = 0;
639 		rbuf[5] = 0;
640 		rbuf[6] = 2 << sector_size_shift;
641 		rbuf[7] = 0;
642 
643 		ccb->ccb_h.status = CAM_REQ_CMP;
644 		break;
645 	}
646 	case REPORT_LUNS:
647 	{
648 		HPT_U8 *rbuf = ccb->csio.data_ptr;
649 		memset(rbuf, 0, 16);
650 		rbuf[3] = 8;
651 		ccb->ccb_h.status = CAM_REQ_CMP;
652 		break;
653 	}
654 	case SERVICE_ACTION_IN:
655 	{
656 		HPT_U8 *rbuf = ccb->csio.data_ptr;
657 		HPT_U64	cap = 0;
658 		HPT_U8 sector_size_shift = 0;
659 		HPT_U32 sector_size = 0;
660 
661 		if(mIsArray(vd->type))
662 			sector_size_shift = vd->u.array.sector_size_shift;
663 		else{
664 			if(vd->type == VD_RAW){
665 				sector_size = vd->u.raw.logical_sector_size;
666 			}
667 
668 			switch (sector_size) {
669 				case 0x1000:
670 					KdPrint(("set 4k setctor size in SERVICE_ACTION_IN"));
671 					sector_size_shift = 3;
672 					break;
673 				default:
674 					break;
675 			}
676 		}
677 		cap = (vd->capacity >> sector_size_shift) - 1;
678 
679 		rbuf[0] = (HPT_U8)(cap>>56);
680 		rbuf[1] = (HPT_U8)(cap>>48);
681 		rbuf[2] = (HPT_U8)(cap>>40);
682 		rbuf[3] = (HPT_U8)(cap>>32);
683 		rbuf[4] = (HPT_U8)(cap>>24);
684 		rbuf[5] = (HPT_U8)(cap>>16);
685 		rbuf[6] = (HPT_U8)(cap>>8);
686 		rbuf[7] = (HPT_U8)cap;
687 		rbuf[8] = 0;
688 		rbuf[9] = 0;
689 		rbuf[10] = 2 << sector_size_shift;
690 		rbuf[11] = 0;
691 
692 		ccb->ccb_h.status = CAM_REQ_CMP;
693 		break;
694 	}
695 
696 	case READ_6:
697 	case READ_10:
698 	case READ_16:
699 	case WRITE_6:
700 	case WRITE_10:
701 	case WRITE_16:
702 	case 0x13:
703 	case 0x2f:
704 	case 0x8f: /* VERIFY_16 */
705 	{
706 		HPT_U8 sector_size_shift = 0;
707 		HPT_U32 sector_size = 0;
708 		pCmd = ldm_alloc_cmds(vbus, vd->cmds_per_request);
709 		if(!pCmd){
710 			KdPrint(("Failed to allocate command!"));
711 			ccb->ccb_h.status = CAM_BUSY;
712 			break;
713 		}
714 
715 		switch (cdb[0])	{
716 		case READ_6:
717 		case WRITE_6:
718 		case 0x13:
719 			pCmd->uCmd.Ide.Lba =  ((HPT_U32)cdb[1] << 16) | ((HPT_U32)cdb[2] << 8) | (HPT_U32)cdb[3];
720 			pCmd->uCmd.Ide.nSectors = (HPT_U16) cdb[4];
721 			break;
722 		case READ_16:
723 		case WRITE_16:
724 		case 0x8f: /* VERIFY_16 */
725 		{
726 			HPT_U64 block =
727 				((HPT_U64)cdb[2]<<56) |
728 				((HPT_U64)cdb[3]<<48) |
729 				((HPT_U64)cdb[4]<<40) |
730 				((HPT_U64)cdb[5]<<32) |
731 				((HPT_U64)cdb[6]<<24) |
732 				((HPT_U64)cdb[7]<<16) |
733 				((HPT_U64)cdb[8]<<8) |
734 				((HPT_U64)cdb[9]);
735 			pCmd->uCmd.Ide.Lba = block;
736 			pCmd->uCmd.Ide.nSectors = (HPT_U16)cdb[13] | ((HPT_U16)cdb[12]<<8);
737 			break;
738 		}
739 
740 		default:
741 			pCmd->uCmd.Ide.Lba = (HPT_U32)cdb[5] | ((HPT_U32)cdb[4] << 8) | ((HPT_U32)cdb[3] << 16) | ((HPT_U32)cdb[2] << 24);
742 			pCmd->uCmd.Ide.nSectors = (HPT_U16) cdb[8] | ((HPT_U16)cdb[7]<<8);
743 			break;
744 		}
745 
746 		if(mIsArray(vd->type)) {
747 			sector_size_shift = vd->u.array.sector_size_shift;
748 		}
749 		else{
750 			if(vd->type == VD_RAW){
751 				sector_size = vd->u.raw.logical_sector_size;
752 			}
753 
754 			switch (sector_size) {
755 				case 0x1000:
756 					KdPrint(("<8>resize sector size from 4k to 512"));
757 					sector_size_shift = 3;
758 					break;
759 				default:
760 					break;
761 	 		}
762 		}
763 		pCmd->uCmd.Ide.Lba <<= sector_size_shift;
764 		pCmd->uCmd.Ide.nSectors <<= sector_size_shift;
765 
766 
767 		switch (cdb[0]) {
768 		case READ_6:
769 		case READ_10:
770 		case READ_16:
771 			pCmd->flags.data_in = 1;
772 			break;
773 		case WRITE_6:
774 		case WRITE_10:
775 		case WRITE_16:
776 			pCmd->flags.data_out = 1;
777 			break;
778 		}
779 		pCmd->priv = ext = cmdext_get(vbus_ext);
780 		HPT_ASSERT(ext);
781 		ext->ccb = ccb;
782 		pCmd->target = vd;
783 		pCmd->done = os_cmddone;
784 		pCmd->buildsgl = os_buildsgl;
785 
786 		pCmd->psg = ext->psg;
787 		pCmd->flags.physical_sg = 1;
788 		error = bus_dmamap_load_ccb(vbus_ext->io_dmat,
789 					ext->dma_map, ccb,
790 					hpt_io_dmamap_callback, pCmd,
791 					BUS_DMA_WAITOK
792 				);
793 		KdPrint(("<8>bus_dmamap_load return %d", error));
794 		if (error && error!=EINPROGRESS) {
795 			os_printk("bus_dmamap_load error %d", error);
796 			cmdext_put(ext);
797 			ldm_free_cmds(pCmd);
798 			ccb->ccb_h.status = CAM_REQ_CMP_ERR;
799 			xpt_done(ccb);
800 		}
801 		return;
802 	}
803 
804 	default:
805 		ccb->ccb_h.status = CAM_REQ_INVALID;
806 		break;
807 	}
808 
809 	xpt_done(ccb);
810 	return;
811 }
812 
813 static void hpt_action(struct cam_sim *sim, union ccb *ccb)
814 {
815 	PVBUS_EXT vbus_ext = (PVBUS_EXT)cam_sim_softc(sim);
816 
817 	KdPrint(("<8>hpt_action(fn=%d, id=%d)", ccb->ccb_h.func_code, ccb->ccb_h.target_id));
818 
819 	hpt_assert_vbus_locked(vbus_ext);
820 	switch (ccb->ccb_h.func_code) {
821 
822 	case XPT_SCSI_IO:
823 		hpt_scsi_io(vbus_ext, ccb);
824 		return;
825 
826 	case XPT_RESET_BUS:
827 		ldm_reset_vbus((PVBUS)vbus_ext->vbus);
828 		break;
829 
830 	case XPT_GET_TRAN_SETTINGS:
831 	case XPT_SET_TRAN_SETTINGS:
832 		ccb->ccb_h.status = CAM_FUNC_NOTAVAIL;
833 		break;
834 
835 	case XPT_CALC_GEOMETRY:
836 		ccb->ccg.heads = 255;
837 		ccb->ccg.secs_per_track = 63;
838 		ccb->ccg.cylinders = ccb->ccg.volume_size / (ccb->ccg.heads * ccb->ccg.secs_per_track);
839 		ccb->ccb_h.status = CAM_REQ_CMP;
840 		break;
841 
842 	case XPT_PATH_INQ:
843 	{
844 		struct ccb_pathinq *cpi = &ccb->cpi;
845 
846 		cpi->version_num = 1;
847 		cpi->hba_inquiry = PI_SDTR_ABLE;
848 		cpi->target_sprt = 0;
849 		cpi->hba_misc = PIM_NOBUSRESET;
850 		cpi->hba_eng_cnt = 0;
851 		cpi->max_target = osm_max_targets;
852 		cpi->max_lun = 0;
853 		cpi->unit_number = cam_sim_unit(sim);
854 		cpi->bus_id = cam_sim_bus(sim);
855 		cpi->initiator_id = osm_max_targets;
856 		cpi->base_transfer_speed = 3300;
857 
858 		strlcpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN);
859 		strlcpy(cpi->hba_vid, "HPT   ", HBA_IDLEN);
860 		strlcpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN);
861 		cpi->transport = XPORT_SPI;
862 		cpi->transport_version = 2;
863 		cpi->protocol = PROTO_SCSI;
864 		cpi->protocol_version = SCSI_REV_2;
865 		cpi->ccb_h.status = CAM_REQ_CMP;
866 		break;
867 	}
868 
869 	default:
870 		ccb->ccb_h.status = CAM_REQ_INVALID;
871 		break;
872 	}
873 
874 	xpt_done(ccb);
875 	return;
876 }
877 
878 static void hpt_pci_intr(void *arg)
879 {
880 	PVBUS_EXT vbus_ext = (PVBUS_EXT)arg;
881 	hpt_lock_vbus(vbus_ext);
882 	ldm_intr((PVBUS)vbus_ext->vbus);
883 	hpt_unlock_vbus(vbus_ext);
884 }
885 
886 static void hpt_poll(struct cam_sim *sim)
887 {
888 	PVBUS_EXT vbus_ext = (PVBUS_EXT)cam_sim_softc(sim);
889 
890 	hpt_assert_vbus_locked(vbus_ext);
891 	ldm_intr((PVBUS)vbus_ext->vbus);
892 }
893 
894 static void hpt_async(void * callback_arg, u_int32_t code, struct cam_path * path, void * arg)
895 {
896 	KdPrint(("<8>hpt_async"));
897 }
898 
899 static int hpt_shutdown(device_t dev)
900 {
901 	KdPrint(("hpt_shutdown(dev=%p)", dev));
902 	return 0;
903 }
904 
905 static int hpt_detach(device_t dev)
906 {
907 	/* we don't allow the driver to be unloaded. */
908 	return EBUSY;
909 }
910 
911 static void hpt_ioctl_done(struct _IOCTL_ARG *arg)
912 {
913 	arg->ioctl_cmnd = 0;
914 	wakeup(arg);
915 }
916 
917 static void __hpt_do_ioctl(PVBUS_EXT vbus_ext, IOCTL_ARG *ioctl_args)
918 {
919 	ioctl_args->result = -1;
920 	ioctl_args->done = hpt_ioctl_done;
921 	ioctl_args->ioctl_cmnd = (void *)1;
922 
923 	hpt_lock_vbus(vbus_ext);
924 	ldm_ioctl((PVBUS)vbus_ext->vbus, ioctl_args);
925 
926 	while (ioctl_args->ioctl_cmnd) {
927 		if (hpt_sleep(vbus_ext, ioctl_args, PPAUSE, "hptctl", HPT_OSM_TIMEOUT)==0)
928 			break;
929 		ldm_reset_vbus((PVBUS)vbus_ext->vbus);
930 		__hpt_do_tasks(vbus_ext);
931 	}
932 
933 	/* KdPrint(("ioctl %x result %d", ioctl_args->dwIoControlCode, ioctl_args->result)); */
934 
935 	hpt_unlock_vbus(vbus_ext);
936 }
937 
938 static void hpt_do_ioctl(IOCTL_ARG *ioctl_args)
939 {
940 	PVBUS vbus;
941 	PVBUS_EXT vbus_ext;
942 
943 	ldm_for_each_vbus(vbus, vbus_ext) {
944 		__hpt_do_ioctl(vbus_ext, ioctl_args);
945 		if (ioctl_args->result!=HPT_IOCTL_RESULT_WRONG_VBUS)
946 			return;
947 	}
948 }
949 
950 #define HPT_DO_IOCTL(code, inbuf, insize, outbuf, outsize) ({\
951 	IOCTL_ARG arg;\
952 	arg.dwIoControlCode = code;\
953 	arg.lpInBuffer = inbuf;\
954 	arg.lpOutBuffer = outbuf;\
955 	arg.nInBufferSize = insize;\
956 	arg.nOutBufferSize = outsize;\
957 	arg.lpBytesReturned = 0;\
958 	hpt_do_ioctl(&arg);\
959 	arg.result;\
960 })
961 
962 #define DEVICEID_VALID(id) ((id) && ((HPT_U32)(id)!=0xffffffff))
963 
964 static int hpt_get_logical_devices(DEVICEID * pIds, int nMaxCount)
965 {
966 	int i;
967 	HPT_U32 count = nMaxCount-1;
968 
969 	if (HPT_DO_IOCTL(HPT_IOCTL_GET_LOGICAL_DEVICES,
970 			&count, sizeof(HPT_U32), pIds, sizeof(DEVICEID)*nMaxCount))
971 		return -1;
972 
973 	nMaxCount = (int)pIds[0];
974 	for (i=0; i<nMaxCount; i++) pIds[i] = pIds[i+1];
975 	return nMaxCount;
976 }
977 
978 static int hpt_get_device_info_v3(DEVICEID id, PLOGICAL_DEVICE_INFO_V3 pInfo)
979 {
980 	return HPT_DO_IOCTL(HPT_IOCTL_GET_DEVICE_INFO_V3,
981 				&id, sizeof(DEVICEID), pInfo, sizeof(LOGICAL_DEVICE_INFO_V3));
982 }
983 
984 /* not belong to this file logically, but we want to use ioctl interface */
985 static int __hpt_stop_tasks(PVBUS_EXT vbus_ext, DEVICEID id)
986 {
987 	LOGICAL_DEVICE_INFO_V3 devinfo;
988 	int i, result;
989 	DEVICEID param[2] = { id, 0 };
990 
991 	if (hpt_get_device_info_v3(id, &devinfo))
992 		return -1;
993 
994 	if (devinfo.Type!=LDT_ARRAY)
995 		return -1;
996 
997 	if (devinfo.u.array.Flags & ARRAY_FLAG_REBUILDING)
998 		param[1] = AS_REBUILD_ABORT;
999 	else if (devinfo.u.array.Flags & ARRAY_FLAG_VERIFYING)
1000 		param[1] = AS_VERIFY_ABORT;
1001 	else if (devinfo.u.array.Flags & ARRAY_FLAG_INITIALIZING)
1002 		param[1] = AS_INITIALIZE_ABORT;
1003 	else if (devinfo.u.array.Flags & ARRAY_FLAG_TRANSFORMING)
1004 		param[1] = AS_TRANSFORM_ABORT;
1005 	else
1006 		return -1;
1007 
1008 	KdPrint(("SET_ARRAY_STATE(%x, %d)", param[0], param[1]));
1009 	result = HPT_DO_IOCTL(HPT_IOCTL_SET_ARRAY_STATE,
1010 				param, sizeof(param), 0, 0);
1011 
1012 	for (i=0; i<devinfo.u.array.nDisk; i++)
1013 		if (DEVICEID_VALID(devinfo.u.array.Members[i]))
1014 			__hpt_stop_tasks(vbus_ext, devinfo.u.array.Members[i]);
1015 
1016 	return result;
1017 }
1018 
1019 static void hpt_stop_tasks(PVBUS_EXT vbus_ext)
1020 {
1021 	DEVICEID ids[32];
1022 	int i, count;
1023 
1024 	count = hpt_get_logical_devices((DEVICEID *)&ids, sizeof(ids)/sizeof(ids[0]));
1025 
1026 	for (i=0; i<count; i++)
1027 		__hpt_stop_tasks(vbus_ext, ids[i]);
1028 }
1029 
1030 static	d_open_t	hpt_open;
1031 static	d_close_t	hpt_close;
1032 static	d_ioctl_t	hpt_ioctl;
1033 static  int 		hpt_rescan_bus(void);
1034 
1035 static struct cdevsw hpt_cdevsw = {
1036 	.d_open =	hpt_open,
1037 	.d_close =	hpt_close,
1038 	.d_ioctl =	hpt_ioctl,
1039 	.d_name =	driver_name,
1040 	.d_version =	D_VERSION,
1041 };
1042 
1043 static struct intr_config_hook hpt_ich;
1044 
1045 /*
1046  * hpt_final_init will be called after all hpt_attach.
1047  */
1048 static void hpt_final_init(void *dummy)
1049 {
1050 	int       i,unit_number=0;
1051 	PVBUS_EXT vbus_ext;
1052 	PVBUS vbus;
1053 	PHBA hba;
1054 
1055 	/* Clear the config hook */
1056 	config_intrhook_disestablish(&hpt_ich);
1057 
1058 	/* allocate memory */
1059 	i = 0;
1060 	ldm_for_each_vbus(vbus, vbus_ext) {
1061 		if (hpt_alloc_mem(vbus_ext)) {
1062 			os_printk("out of memory");
1063 			return;
1064 		}
1065 		i++;
1066 	}
1067 
1068 	if (!i) {
1069 		if (bootverbose)
1070 			os_printk("no controller detected.");
1071 		return;
1072 	}
1073 
1074 	/* initializing hardware */
1075 	ldm_for_each_vbus(vbus, vbus_ext) {
1076 		/* make timer available here */
1077 		mtx_init(&vbus_ext->lock, "hptsleeplock", NULL, MTX_DEF);
1078 		callout_init_mtx(&vbus_ext->timer, &vbus_ext->lock, 0);
1079 		if (hpt_init_vbus(vbus_ext)) {
1080 			os_printk("fail to initialize hardware");
1081 			break; /* FIXME */
1082 		}
1083 	}
1084 
1085 	/* register CAM interface */
1086 	ldm_for_each_vbus(vbus, vbus_ext) {
1087 		struct cam_devq *devq;
1088 		struct ccb_setasync	ccb;
1089 
1090 		if (bus_dma_tag_create(NULL,/* parent */
1091 				4,	/* alignment */
1092 				BUS_SPACE_MAXADDR_32BIT+1, /* boundary */
1093 				BUS_SPACE_MAXADDR,	/* lowaddr */
1094 				BUS_SPACE_MAXADDR, 	/* highaddr */
1095 				NULL, NULL, 		/* filter, filterarg */
1096 				PAGE_SIZE * (os_max_sg_descriptors-1),	/* maxsize */
1097 				os_max_sg_descriptors,	/* nsegments */
1098 				0x10000,	/* maxsegsize */
1099 				BUS_DMA_WAITOK,		/* flags */
1100 				busdma_lock_mutex,	/* lockfunc */
1101 				&vbus_ext->lock,		/* lockfuncarg */
1102 				&vbus_ext->io_dmat	/* tag */))
1103 		{
1104 			return ;
1105 		}
1106 
1107 		for (i=0; i<os_max_queue_comm; i++) {
1108 			POS_CMDEXT ext = (POS_CMDEXT)malloc(sizeof(OS_CMDEXT), M_DEVBUF, M_WAITOK);
1109 			if (!ext) {
1110 				os_printk("Can't alloc cmdext(%d)", i);
1111 				return ;
1112 			}
1113 			ext->vbus_ext = vbus_ext;
1114 			ext->next = vbus_ext->cmdext_list;
1115 			vbus_ext->cmdext_list = ext;
1116 
1117 			if (bus_dmamap_create(vbus_ext->io_dmat, 0, &ext->dma_map)) {
1118 				os_printk("Can't create dma map(%d)", i);
1119 				return ;
1120 			}
1121 			callout_init_mtx(&ext->timeout, &vbus_ext->lock, 0);
1122 		}
1123 
1124 		if ((devq = cam_simq_alloc(os_max_queue_comm)) == NULL) {
1125 			os_printk("cam_simq_alloc failed");
1126 			return ;
1127 		}
1128 		vbus_ext->sim = cam_sim_alloc(hpt_action, hpt_poll, driver_name,
1129 				vbus_ext, unit_number, &vbus_ext->lock, os_max_queue_comm, /*tagged*/8,  devq);
1130 		unit_number++;
1131 		if (!vbus_ext->sim) {
1132 			os_printk("cam_sim_alloc failed");
1133 			cam_simq_free(devq);
1134 			return ;
1135 		}
1136 
1137 		hpt_lock_vbus(vbus_ext);
1138 		if (xpt_bus_register(vbus_ext->sim, NULL, 0) != CAM_SUCCESS) {
1139 			hpt_unlock_vbus(vbus_ext);
1140 			os_printk("xpt_bus_register failed");
1141 			cam_sim_free(vbus_ext->sim, /*free devq*/ TRUE);
1142 			vbus_ext->sim = NULL;
1143 			return ;
1144 		}
1145 
1146 		if (xpt_create_path(&vbus_ext->path, /*periph */ NULL,
1147 				cam_sim_path(vbus_ext->sim), CAM_TARGET_WILDCARD,
1148 				CAM_LUN_WILDCARD) != CAM_REQ_CMP)
1149 		{
1150 			hpt_unlock_vbus(vbus_ext);
1151 			os_printk("xpt_create_path failed");
1152 			xpt_bus_deregister(cam_sim_path(vbus_ext->sim));
1153 			cam_sim_free(vbus_ext->sim, /*free_devq*/TRUE);
1154 			vbus_ext->sim = NULL;
1155 			return ;
1156 		}
1157 
1158 		memset(&ccb, 0, sizeof(ccb));
1159 		xpt_setup_ccb(&ccb.ccb_h, vbus_ext->path, /*priority*/5);
1160 		ccb.ccb_h.func_code = XPT_SASYNC_CB;
1161 		ccb.event_enable = AC_LOST_DEVICE;
1162 		ccb.callback = hpt_async;
1163 		ccb.callback_arg = vbus_ext;
1164 		xpt_action((union ccb *)&ccb);
1165 		hpt_unlock_vbus(vbus_ext);
1166 
1167 		for (hba = vbus_ext->hba_list; hba; hba = hba->next) {
1168 			int rid = 0;
1169 			if ((hba->irq_res = bus_alloc_resource_any(hba->pcidev,
1170 				SYS_RES_IRQ, &rid, RF_SHAREABLE | RF_ACTIVE)) == NULL)
1171 			{
1172 				os_printk("can't allocate interrupt");
1173 				return ;
1174 			}
1175 			if (bus_setup_intr(hba->pcidev, hba->irq_res, INTR_TYPE_CAM | INTR_MPSAFE,
1176 				NULL, hpt_pci_intr, vbus_ext, &hba->irq_handle))
1177 			{
1178 				os_printk("can't set up interrupt");
1179 				return ;
1180 			}
1181 			hba->ldm_adapter.him->intr_control(hba->ldm_adapter.him_handle, HPT_TRUE);
1182 
1183 		}
1184 
1185 		vbus_ext->shutdown_eh = EVENTHANDLER_REGISTER(shutdown_final,
1186 									hpt_shutdown_vbus, vbus_ext, SHUTDOWN_PRI_DEFAULT);
1187 		if (!vbus_ext->shutdown_eh)
1188 			os_printk("Shutdown event registration failed");
1189 	}
1190 
1191 	ldm_for_each_vbus(vbus, vbus_ext) {
1192 		TASK_INIT(&vbus_ext->worker, 0, (task_fn_t *)hpt_do_tasks, vbus_ext);
1193 		if (vbus_ext->tasks)
1194 			TASK_ENQUEUE(&vbus_ext->worker);
1195 	}
1196 
1197 	make_dev(&hpt_cdevsw, DRIVER_MINOR, UID_ROOT, GID_OPERATOR,
1198 	    S_IRUSR | S_IWUSR, "%s", driver_name);
1199 }
1200 
1201 #if defined(KLD_MODULE)
1202 
1203 typedef struct driverlink *driverlink_t;
1204 struct driverlink {
1205 	kobj_class_t	driver;
1206 	TAILQ_ENTRY(driverlink) link;	/* list of drivers in devclass */
1207 };
1208 
1209 typedef TAILQ_HEAD(driver_list, driverlink) driver_list_t;
1210 
1211 struct devclass {
1212 	TAILQ_ENTRY(devclass) link;
1213 	devclass_t	parent;		/* parent in devclass hierarchy */
1214 	driver_list_t	drivers;     /* bus devclasses store drivers for bus */
1215 	char		*name;
1216 	device_t	*devices;	/* array of devices indexed by unit */
1217 	int		maxunit;	/* size of devices array */
1218 };
1219 
1220 static void override_kernel_driver(void)
1221 {
1222 	driverlink_t dl, dlfirst;
1223 	driver_t *tmpdriver;
1224 	devclass_t dc = devclass_find("pci");
1225 
1226 	if (dc){
1227 		dlfirst = TAILQ_FIRST(&dc->drivers);
1228 		for (dl = dlfirst; dl; dl = TAILQ_NEXT(dl, link)) {
1229 			if(strcmp(dl->driver->name, driver_name) == 0) {
1230 				tmpdriver=dl->driver;
1231 				dl->driver=dlfirst->driver;
1232 				dlfirst->driver=tmpdriver;
1233 				break;
1234 			}
1235 		}
1236 	}
1237 }
1238 
1239 #else
1240 #define override_kernel_driver()
1241 #endif
1242 
1243 static void hpt_init(void *dummy)
1244 {
1245 	if (bootverbose)
1246 		os_printk("%s %s", driver_name_long, driver_ver);
1247 
1248 	override_kernel_driver();
1249 	init_config();
1250 
1251 	hpt_ich.ich_func = hpt_final_init;
1252 	hpt_ich.ich_arg = NULL;
1253 	if (config_intrhook_establish(&hpt_ich) != 0) {
1254 		printf("%s: cannot establish configuration hook\n",
1255 		    driver_name_long);
1256 	}
1257 
1258 }
1259 SYSINIT(hptinit, SI_SUB_CONFIGURE, SI_ORDER_FIRST, hpt_init, NULL);
1260 
1261 /*
1262  * CAM driver interface
1263  */
1264 static device_method_t driver_methods[] = {
1265 	/* Device interface */
1266 	DEVMETHOD(device_probe,		hpt_probe),
1267 	DEVMETHOD(device_attach,	hpt_attach),
1268 	DEVMETHOD(device_detach,	hpt_detach),
1269 	DEVMETHOD(device_shutdown,	hpt_shutdown),
1270 	{ 0, 0 }
1271 };
1272 
1273 static driver_t hpt_pci_driver = {
1274 	driver_name,
1275 	driver_methods,
1276 	sizeof(HBA)
1277 };
1278 
1279 static devclass_t	hpt_devclass;
1280 
1281 #ifndef TARGETNAME
1282 #error "no TARGETNAME found"
1283 #endif
1284 
1285 /* use this to make TARGETNAME be expanded */
1286 #define __DRIVER_MODULE(p1, p2, p3, p4, p5, p6) DRIVER_MODULE(p1, p2, p3, p4, p5, p6)
1287 #define __MODULE_VERSION(p1, p2) MODULE_VERSION(p1, p2)
1288 #define __MODULE_DEPEND(p1, p2, p3, p4, p5) MODULE_DEPEND(p1, p2, p3, p4, p5)
1289 __DRIVER_MODULE(TARGETNAME, pci, hpt_pci_driver, hpt_devclass, 0, 0);
1290 __MODULE_VERSION(TARGETNAME, 1);
1291 __MODULE_DEPEND(TARGETNAME, cam, 1, 1, 1);
1292 
1293 static int hpt_open(struct cdev *dev, int flags, int devtype, struct thread *td)
1294 {
1295 	return 0;
1296 }
1297 
1298 static int hpt_close(struct cdev *dev, int flags, int devtype, struct thread *td)
1299 {
1300 	return 0;
1301 }
1302 
1303 static int hpt_ioctl(struct cdev *dev, u_long cmd, caddr_t data, int fflag, struct thread *td)
1304 {
1305 	PHPT_IOCTL_PARAM piop=(PHPT_IOCTL_PARAM)data;
1306 	IOCTL_ARG ioctl_args;
1307 	HPT_U32 bytesReturned = 0;
1308 
1309 	switch (cmd){
1310 	case HPT_DO_IOCONTROL:
1311 	{
1312 		if (piop->Magic == HPT_IOCTL_MAGIC || piop->Magic == HPT_IOCTL_MAGIC32) {
1313 			KdPrint(("<8>ioctl=%x in=%p len=%d out=%p len=%d\n",
1314 				piop->dwIoControlCode,
1315 				piop->lpInBuffer,
1316 				piop->nInBufferSize,
1317 				piop->lpOutBuffer,
1318 				piop->nOutBufferSize));
1319 
1320 		memset(&ioctl_args, 0, sizeof(ioctl_args));
1321 
1322 		ioctl_args.dwIoControlCode = piop->dwIoControlCode;
1323 		ioctl_args.nInBufferSize = piop->nInBufferSize;
1324 		ioctl_args.nOutBufferSize = piop->nOutBufferSize;
1325 		ioctl_args.lpBytesReturned = &bytesReturned;
1326 
1327 		if (ioctl_args.nInBufferSize) {
1328 			ioctl_args.lpInBuffer = malloc(ioctl_args.nInBufferSize, M_DEVBUF, M_WAITOK);
1329 			if (!ioctl_args.lpInBuffer)
1330 				goto invalid;
1331 			if (copyin((void*)piop->lpInBuffer,
1332 					ioctl_args.lpInBuffer, piop->nInBufferSize))
1333 				goto invalid;
1334 		}
1335 
1336 		if (ioctl_args.nOutBufferSize) {
1337 			ioctl_args.lpOutBuffer = malloc(ioctl_args.nOutBufferSize, M_DEVBUF, M_WAITOK | M_ZERO);
1338 			if (!ioctl_args.lpOutBuffer)
1339 				goto invalid;
1340 		}
1341 
1342 		hpt_do_ioctl(&ioctl_args);
1343 
1344 		if (ioctl_args.result==HPT_IOCTL_RESULT_OK) {
1345 			if (piop->nOutBufferSize) {
1346 				if (copyout(ioctl_args.lpOutBuffer,
1347 					(void*)piop->lpOutBuffer, piop->nOutBufferSize))
1348 					goto invalid;
1349 			}
1350 			if (piop->lpBytesReturned) {
1351 				if (copyout(&bytesReturned,
1352 					(void*)piop->lpBytesReturned, sizeof(HPT_U32)))
1353 					goto invalid;
1354 			}
1355 			if (ioctl_args.lpInBuffer) free(ioctl_args.lpInBuffer, M_DEVBUF);
1356 			if (ioctl_args.lpOutBuffer) free(ioctl_args.lpOutBuffer, M_DEVBUF);
1357 			return 0;
1358 		}
1359 invalid:
1360 		if (ioctl_args.lpInBuffer) free(ioctl_args.lpInBuffer, M_DEVBUF);
1361 		if (ioctl_args.lpOutBuffer) free(ioctl_args.lpOutBuffer, M_DEVBUF);
1362 		return EFAULT;
1363 	}
1364 	return EFAULT;
1365 	}
1366 
1367 	case HPT_SCAN_BUS:
1368 	{
1369 		return hpt_rescan_bus();
1370 	}
1371 	default:
1372 		KdPrint(("invalid command!"));
1373 		return EFAULT;
1374 	}
1375 
1376 }
1377 
1378 static int	hpt_rescan_bus(void)
1379 {
1380 	union ccb			*ccb;
1381 	PVBUS 				vbus;
1382 	PVBUS_EXT			vbus_ext;
1383 
1384 	ldm_for_each_vbus(vbus, vbus_ext) {
1385 		if ((ccb = xpt_alloc_ccb()) == NULL)
1386 		{
1387 			return(ENOMEM);
1388 		}
1389 		if (xpt_create_path(&ccb->ccb_h.path, NULL, cam_sim_path(vbus_ext->sim),
1390 			CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP)
1391 		{
1392 			xpt_free_ccb(ccb);
1393 			return(EIO);
1394 		}
1395 		xpt_rescan(ccb);
1396 	}
1397 	return(0);
1398 }
1399 
1400