xref: /freebsd/sys/dev/hptnr/hptnr_osm_bsd.c (revision 2a63c3be158216222d89a073dcbd6a72ee4aab5a)
1 /* $Id: osm_bsd.c,v 1.36 2010/05/11 03:12:11 lcn Exp $ */
2 /*-
3  * SPDX-License-Identifier: BSD-2-Clause
4  *
5  * HighPoint RAID Driver for FreeBSD
6  * Copyright (C) 2005-2011 HighPoint Technologies, Inc.
7  * All rights reserved.
8  *
9  * Redistribution and use in source and binary forms, with or without
10  * modification, are permitted provided that the following conditions
11  * are met:
12  * 1. Redistributions of source code must retain the above copyright
13  *    notice, this list of conditions and the following disclaimer.
14  * 2. Redistributions in binary form must reproduce the above copyright
15  *    notice, this list of conditions and the following disclaimer in the
16  *    documentation and/or other materials provided with the distribution.
17  *
18  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
19  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
20  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
21  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
22  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
23  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
24  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
25  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
26  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
27  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
28  * SUCH DAMAGE.
29  */
30 #include <dev/hptnr/hptnr_config.h>
31 #include <dev/hptnr/os_bsd.h>
32 #include <dev/hptnr/hptintf.h>
33 int msi = 0;
34 int debug_flag = 0;
35 static HIM *hpt_match(device_t dev, int scan)
36 {
37 	PCI_ID pci_id;
38 	HIM *him;
39 	int i;
40 
41 	for (him = him_list; him; him = him->next) {
42 		for (i=0; him->get_supported_device_id(i, &pci_id); i++) {
43 			if (scan && him->get_controller_count)
44 				him->get_controller_count(&pci_id,0,0);
45 			if ((pci_get_vendor(dev) == pci_id.vid) &&
46 				(pci_get_device(dev) == pci_id.did)){
47 				return (him);
48 			}
49 		}
50 	}
51 
52 	return (NULL);
53 }
54 
55 static int hpt_probe(device_t dev)
56 {
57 	HIM *him;
58 
59 	him = hpt_match(dev, 0);
60 	if (him != NULL) {
61 		KdPrint(("hpt_probe: adapter at PCI %d:%d:%d, IRQ %d",
62 			pci_get_bus(dev), pci_get_slot(dev), pci_get_function(dev), pci_get_irq(dev)
63 			));
64 		device_set_desc(dev, him->name);
65 		return (BUS_PROBE_DEFAULT);
66 	}
67 
68 	return (ENXIO);
69 }
70 
71 static int hpt_attach(device_t dev)
72 {
73 	PHBA hba = (PHBA)device_get_softc(dev);
74 	HIM *him;
75 	PCI_ID pci_id;
76 	HPT_UINT size;
77 	PVBUS vbus;
78 	PVBUS_EXT vbus_ext;
79 
80 	KdPrint(("hpt_attach(%d/%d/%d)", pci_get_bus(dev), pci_get_slot(dev), pci_get_function(dev)));
81 
82 	him = hpt_match(dev, 1);
83 	hba->ext_type = EXT_TYPE_HBA;
84 	hba->ldm_adapter.him = him;
85 
86 	pci_enable_busmaster(dev);
87 
88 	pci_id.vid = pci_get_vendor(dev);
89 	pci_id.did = pci_get_device(dev);
90 	pci_id.rev = pci_get_revid(dev);
91 	pci_id.subsys = (HPT_U32)(pci_get_subdevice(dev)) << 16 | pci_get_subvendor(dev);
92 
93 	size = him->get_adapter_size(&pci_id);
94 	hba->ldm_adapter.him_handle = malloc(size, M_DEVBUF, M_WAITOK);
95 
96 	hba->pcidev = dev;
97 	hba->pciaddr.tree = 0;
98 	hba->pciaddr.bus = pci_get_bus(dev);
99 	hba->pciaddr.device = pci_get_slot(dev);
100 	hba->pciaddr.function = pci_get_function(dev);
101 
102 	if (!him->create_adapter(&pci_id, hba->pciaddr, hba->ldm_adapter.him_handle, hba)) {
103 		free(hba->ldm_adapter.him_handle, M_DEVBUF);
104 		return ENXIO;
105 	}
106 
107 	os_printk("adapter at PCI %d:%d:%d, IRQ %d",
108 		hba->pciaddr.bus, hba->pciaddr.device, hba->pciaddr.function, pci_get_irq(dev));
109 
110 	if (!ldm_register_adapter(&hba->ldm_adapter)) {
111 		size = ldm_get_vbus_size();
112 		vbus_ext = malloc(sizeof(VBUS_EXT) + size, M_DEVBUF, M_WAITOK |
113 			M_ZERO);
114 		vbus_ext->ext_type = EXT_TYPE_VBUS;
115 		ldm_create_vbus((PVBUS)vbus_ext->vbus, vbus_ext);
116 		ldm_register_adapter(&hba->ldm_adapter);
117 	}
118 
119 	ldm_for_each_vbus(vbus, vbus_ext) {
120 		if (hba->ldm_adapter.vbus==vbus) {
121 			hba->vbus_ext = vbus_ext;
122 			hba->next = vbus_ext->hba_list;
123 			vbus_ext->hba_list = hba;
124 			break;
125 		}
126 	}
127 	return 0;
128 }
129 
130 /*
131  * Maybe we'd better to use the bus_dmamem_alloc to alloc DMA memory,
132  * but there are some problems currently (alignment, etc).
133  */
134 static __inline void *__get_free_pages(int order)
135 {
136 	/* don't use low memory - other devices may get starved */
137 	return contigmalloc(PAGE_SIZE<<order,
138 			M_DEVBUF, M_WAITOK, BUS_SPACE_MAXADDR_24BIT, BUS_SPACE_MAXADDR, PAGE_SIZE, 0);
139 }
140 
141 static __inline void free_pages(void *p, int order)
142 {
143 	contigfree(p, PAGE_SIZE<<order, M_DEVBUF);
144 }
145 
146 static int hpt_alloc_mem(PVBUS_EXT vbus_ext)
147 {
148 	PHBA hba;
149 	struct freelist *f;
150 	HPT_UINT i;
151 	void **p;
152 
153 	for (hba = vbus_ext->hba_list; hba; hba = hba->next)
154 		hba->ldm_adapter.him->get_meminfo(hba->ldm_adapter.him_handle);
155 
156 	ldm_get_mem_info((PVBUS)vbus_ext->vbus, 0);
157 
158 	for (f=vbus_ext->freelist_head; f; f=f->next) {
159 		KdPrint(("%s: %d*%d=%d bytes",
160 			f->tag, f->count, f->size, f->count*f->size));
161 		for (i=0; i<f->count; i++) {
162 			p = (void **)malloc(f->size, M_DEVBUF, M_WAITOK);
163 			if (!p)	return (ENXIO);
164 			*p = f->head;
165 			f->head = p;
166 		}
167 	}
168 
169 	for (f=vbus_ext->freelist_dma_head; f; f=f->next) {
170 		int order, size, j;
171 
172 		HPT_ASSERT((f->size & (f->alignment-1))==0);
173 
174 		for (order=0, size=PAGE_SIZE; size<f->size; order++, size<<=1)
175 			;
176 
177 		KdPrint(("%s: %d*%d=%d bytes, order %d",
178 			f->tag, f->count, f->size, f->count*f->size, order));
179 		HPT_ASSERT(f->alignment<=PAGE_SIZE);
180 
181 		for (i=0; i<f->count;) {
182 			p = (void **)__get_free_pages(order);
183 			if (!p) return -1;
184 			for (j = size/f->size; j && i<f->count; i++,j--) {
185 				*p = f->head;
186 				*(BUS_ADDRESS *)(p+1) = (BUS_ADDRESS)vtophys(p);
187 				f->head = p;
188 				p = (void **)((unsigned long)p + f->size);
189 			}
190 		}
191 	}
192 
193 	HPT_ASSERT(PAGE_SIZE==DMAPOOL_PAGE_SIZE);
194 
195 	for (i=0; i<os_max_cache_pages; i++) {
196 		p = (void **)__get_free_pages(0);
197 		if (!p) return -1;
198 		HPT_ASSERT(((HPT_UPTR)p & (DMAPOOL_PAGE_SIZE-1))==0);
199 		dmapool_put_page((PVBUS)vbus_ext->vbus, p, (BUS_ADDRESS)vtophys(p));
200 	}
201 
202 	return 0;
203 }
204 
205 static void hpt_free_mem(PVBUS_EXT vbus_ext)
206 {
207 	struct freelist *f;
208 	void *p;
209 	int i;
210 	BUS_ADDRESS bus;
211 
212 	for (f=vbus_ext->freelist_head; f; f=f->next) {
213 #if DBG
214 		if (f->count!=f->reserved_count) {
215 			KdPrint(("memory leak for freelist %s (%d/%d)", f->tag, f->count, f->reserved_count));
216 		}
217 #endif
218 		while ((p=freelist_get(f)))
219 			free(p, M_DEVBUF);
220 	}
221 
222 	for (i=0; i<os_max_cache_pages; i++) {
223 		p = dmapool_get_page((PVBUS)vbus_ext->vbus, &bus);
224 		HPT_ASSERT(p);
225 		free_pages(p, 0);
226 	}
227 
228 	for (f=vbus_ext->freelist_dma_head; f; f=f->next) {
229 		int order, size;
230 #if DBG
231 		if (f->count!=f->reserved_count) {
232 			KdPrint(("memory leak for dma freelist %s (%d/%d)", f->tag, f->count, f->reserved_count));
233 		}
234 #endif
235 		for (order=0, size=PAGE_SIZE; size<f->size; order++, size<<=1) ;
236 
237 		while ((p=freelist_get_dma(f, &bus))) {
238 			if (order)
239 				free_pages(p, order);
240 			else {
241 			/* can't free immediately since other blocks in this page may still be in the list */
242 				if (((HPT_UPTR)p & (PAGE_SIZE-1))==0)
243 					dmapool_put_page((PVBUS)vbus_ext->vbus, p, bus);
244 			}
245 		}
246 	}
247 
248 	while ((p = dmapool_get_page((PVBUS)vbus_ext->vbus, &bus)))
249 		free_pages(p, 0);
250 }
251 
252 static int hpt_init_vbus(PVBUS_EXT vbus_ext)
253 {
254 	PHBA hba;
255 
256 	for (hba = vbus_ext->hba_list; hba; hba = hba->next)
257 		if (!hba->ldm_adapter.him->initialize(hba->ldm_adapter.him_handle)) {
258 			KdPrint(("fail to initialize %p", hba));
259 			return -1;
260 		}
261 
262 	ldm_initialize_vbus((PVBUS)vbus_ext->vbus, &vbus_ext->hba_list->ldm_adapter);
263 	return 0;
264 }
265 
266 static void hpt_flush_done(PCOMMAND pCmd)
267 {
268 	PVDEV vd = pCmd->target;
269 
270 	if (mIsArray(vd->type) && vd->u.array.transform && vd!=vd->u.array.transform->target) {
271 		vd = vd->u.array.transform->target;
272 		HPT_ASSERT(vd);
273 		pCmd->target = vd;
274 		pCmd->Result = RETURN_PENDING;
275 		vdev_queue_cmd(pCmd);
276 		return;
277 	}
278 
279 	*(int *)pCmd->priv = 1;
280 	wakeup(pCmd);
281 }
282 
283 /*
284  * flush a vdev (without retry).
285  */
286 static int hpt_flush_vdev(PVBUS_EXT vbus_ext, PVDEV vd)
287 {
288 	PCOMMAND pCmd;
289 	int result = 0, done;
290 	HPT_UINT count;
291 
292 	KdPrint(("flusing dev %p", vd));
293 
294 	hpt_assert_vbus_locked(vbus_ext);
295 
296 	if (mIsArray(vd->type) && vd->u.array.transform)
297 		count = max(vd->u.array.transform->source->cmds_per_request,
298 					vd->u.array.transform->target->cmds_per_request);
299 	else
300 		count = vd->cmds_per_request;
301 
302 	pCmd = ldm_alloc_cmds(vd->vbus, count);
303 
304 	if (!pCmd) {
305 		return -1;
306 	}
307 
308 	pCmd->type = CMD_TYPE_FLUSH;
309 	pCmd->flags.hard_flush = 1;
310 	pCmd->target = vd;
311 	pCmd->done = hpt_flush_done;
312 	done = 0;
313 	pCmd->priv = &done;
314 
315 	ldm_queue_cmd(pCmd);
316 
317 	if (!done) {
318 		while (hpt_sleep(vbus_ext, pCmd, PPAUSE, "hptfls", HPT_OSM_TIMEOUT)) {
319 			ldm_reset_vbus(vd->vbus);
320 		}
321 	}
322 
323 	KdPrint(("flush result %d", pCmd->Result));
324 
325 	if (pCmd->Result!=RETURN_SUCCESS)
326 		result = -1;
327 
328 	ldm_free_cmds(pCmd);
329 
330 	return result;
331 }
332 
333 static void hpt_stop_tasks(PVBUS_EXT vbus_ext);
334 static void hpt_shutdown_vbus(PVBUS_EXT vbus_ext, int howto)
335 {
336 	PVBUS     vbus = (PVBUS)vbus_ext->vbus;
337 	PHBA hba;
338 	int i;
339 
340 	KdPrint(("hpt_shutdown_vbus"));
341 
342 	/* stop all ctl tasks and disable the worker taskqueue */
343 	hpt_stop_tasks(vbus_ext);
344 	hpt_lock_vbus(vbus_ext);
345 	vbus_ext->worker.ta_context = 0;
346 
347 	/* flush devices */
348 	for (i=0; i<osm_max_targets; i++) {
349 		PVDEV vd = ldm_find_target(vbus, i);
350 		if (vd) {
351 			/* retry once */
352 			if (hpt_flush_vdev(vbus_ext, vd))
353 				hpt_flush_vdev(vbus_ext, vd);
354 		}
355 	}
356 
357 	ldm_shutdown(vbus);
358 	hpt_unlock_vbus(vbus_ext);
359 
360 	ldm_release_vbus(vbus);
361 
362 	for (hba=vbus_ext->hba_list; hba; hba=hba->next)
363 		bus_teardown_intr(hba->pcidev, hba->irq_res, hba->irq_handle);
364 
365 	hpt_free_mem(vbus_ext);
366 
367 	while ((hba=vbus_ext->hba_list)) {
368 		vbus_ext->hba_list = hba->next;
369 		free(hba->ldm_adapter.him_handle, M_DEVBUF);
370 	}
371 
372 	callout_drain(&vbus_ext->timer);
373 	mtx_destroy(&vbus_ext->lock);
374 	free(vbus_ext, M_DEVBUF);
375 	KdPrint(("hpt_shutdown_vbus done"));
376 }
377 
378 static void __hpt_do_tasks(PVBUS_EXT vbus_ext)
379 {
380 	OSM_TASK *tasks;
381 
382 	tasks = vbus_ext->tasks;
383 	vbus_ext->tasks = 0;
384 
385 	while (tasks) {
386 		OSM_TASK *t = tasks;
387 		tasks = t->next;
388 		t->next = 0;
389 		t->func(vbus_ext->vbus, t->data);
390 	}
391 }
392 
393 static void hpt_do_tasks(PVBUS_EXT vbus_ext, int pending)
394 {
395 	if(vbus_ext){
396 		hpt_lock_vbus(vbus_ext);
397 		__hpt_do_tasks(vbus_ext);
398 		hpt_unlock_vbus(vbus_ext);
399 	}
400 }
401 
402 static void hpt_action(struct cam_sim *sim, union ccb *ccb);
403 static void hpt_poll(struct cam_sim *sim);
404 static void hpt_async(void * callback_arg, u_int32_t code, struct cam_path * path, void * arg);
405 static void hpt_pci_intr(void *arg);
406 
407 static __inline POS_CMDEXT cmdext_get(PVBUS_EXT vbus_ext)
408 {
409 	POS_CMDEXT p = vbus_ext->cmdext_list;
410 	if (p)
411 		vbus_ext->cmdext_list = p->next;
412 	return p;
413 }
414 
415 static __inline void cmdext_put(POS_CMDEXT p)
416 {
417 	p->next = p->vbus_ext->cmdext_list;
418 	p->vbus_ext->cmdext_list = p;
419 }
420 
421 static void hpt_timeout(void *arg)
422 {
423 	PCOMMAND pCmd = (PCOMMAND)arg;
424 	POS_CMDEXT ext = (POS_CMDEXT)pCmd->priv;
425 
426 	KdPrint(("pCmd %p timeout", pCmd));
427 
428 	ldm_reset_vbus((PVBUS)ext->vbus_ext->vbus);
429 }
430 
431 static void os_cmddone(PCOMMAND pCmd)
432 {
433 	POS_CMDEXT ext = (POS_CMDEXT)pCmd->priv;
434 	union ccb *ccb = ext->ccb;
435 	HPT_U8 *cdb;
436 
437 	if (ccb->ccb_h.flags & CAM_CDB_POINTER)
438 		cdb = ccb->csio.cdb_io.cdb_ptr;
439 	else
440 		cdb = ccb->csio.cdb_io.cdb_bytes;
441 
442 	KdPrint(("os_cmddone(%p, %d)", pCmd, pCmd->Result));
443 
444 	callout_stop(&ext->timeout);
445 	switch(cdb[0]) {
446 		case 0x85: /*ATA_16*/
447 		case 0xA1: /*ATA_12*/
448 		{
449 			PassthroughCmd *passthru = &pCmd->uCmd.Passthrough;
450 			HPT_U8 *sense_buffer = (HPT_U8 *)&ccb->csio.sense_data;
451 			memset(&ccb->csio.sense_data, 0,sizeof(ccb->csio.sense_data));
452 
453 			sense_buffer[0] = 0x72; /* Response Code */
454 			sense_buffer[7] = 14; /* Additional Sense Length */
455 
456 			sense_buffer[8] = 0x9; /* ATA Return Descriptor */
457 			sense_buffer[9] = 0xc; /* Additional Descriptor Length */
458 			sense_buffer[11] = (HPT_U8)passthru->bFeaturesReg; /* Error */
459 			sense_buffer[13] = (HPT_U8)passthru->bSectorCountReg;  /* Sector Count (7:0) */
460 			sense_buffer[15] = (HPT_U8)passthru->bLbaLowReg; /* LBA Low (7:0) */
461 			sense_buffer[17] = (HPT_U8)passthru->bLbaMidReg; /* LBA Mid (7:0) */
462 			sense_buffer[19] = (HPT_U8)passthru->bLbaHighReg; /* LBA High (7:0) */
463 
464 			if ((cdb[0] == 0x85) && (cdb[1] & 0x1))
465 			{
466 				sense_buffer[10] = 1;
467 				sense_buffer[12] = (HPT_U8)(passthru->bSectorCountReg >> 8); /* Sector Count (15:8) */
468 				sense_buffer[14] = (HPT_U8)(passthru->bLbaLowReg >> 8);	/* LBA Low (15:8) */
469 				sense_buffer[16] = (HPT_U8)(passthru->bLbaMidReg >> 8); /* LBA Mid (15:8) */
470 				sense_buffer[18] = (HPT_U8)(passthru->bLbaHighReg >> 8); /* LBA High (15:8) */
471 			}
472 
473 			sense_buffer[20] = (HPT_U8)passthru->bDriveHeadReg; /* Device */
474 			sense_buffer[21] = (HPT_U8)passthru->bCommandReg; /* Status */
475 			KdPrint(("sts 0x%x err 0x%x low 0x%x mid 0x%x hig 0x%x dh 0x%x sc 0x%x",
476 					 passthru->bCommandReg,
477 					 passthru->bFeaturesReg,
478 					 passthru->bLbaLowReg,
479 					 passthru->bLbaMidReg,
480 					 passthru->bLbaHighReg,
481 					 passthru->bDriveHeadReg,
482 					 passthru->bSectorCountReg));
483 			KdPrint(("result:0x%x,bFeaturesReg:0x%04x,bSectorCountReg:0x%04x,LBA:0x%04x%04x%04x ",
484 				pCmd->Result,passthru->bFeaturesReg,passthru->bSectorCountReg,
485 				passthru->bLbaHighReg,passthru->bLbaMidReg,passthru->bLbaLowReg));
486 		}
487 		default:
488 			break;
489 	}
490 
491 	switch(pCmd->Result) {
492 	case RETURN_SUCCESS:
493 		ccb->ccb_h.status = CAM_REQ_CMP;
494 		break;
495 	case RETURN_BAD_DEVICE:
496 		ccb->ccb_h.status = CAM_DEV_NOT_THERE;
497 		break;
498 	case RETURN_DEVICE_BUSY:
499 		ccb->ccb_h.status = CAM_BUSY;
500 		break;
501 	case RETURN_INVALID_REQUEST:
502 		ccb->ccb_h.status = CAM_REQ_INVALID;
503 		break;
504 	case RETURN_SELECTION_TIMEOUT:
505 		ccb->ccb_h.status = CAM_SEL_TIMEOUT;
506 		break;
507 	case RETURN_RETRY:
508 		ccb->ccb_h.status = CAM_BUSY;
509 		break;
510 	default:
511 		ccb->ccb_h.status = CAM_SCSI_STATUS_ERROR;
512 		break;
513 	}
514 
515 	if (pCmd->flags.data_in) {
516 		bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map, BUS_DMASYNC_POSTREAD);
517 	}
518 	else if (pCmd->flags.data_out) {
519 		bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map, BUS_DMASYNC_POSTWRITE);
520 	}
521 
522 	bus_dmamap_unload(ext->vbus_ext->io_dmat, ext->dma_map);
523 
524 	cmdext_put(ext);
525 	ldm_free_cmds(pCmd);
526 	xpt_done(ccb);
527 }
528 
529 static int os_buildsgl(PCOMMAND pCmd, PSG pSg, int logical)
530 {
531 	/* since we have provided physical sg, nobody will ask us to build physical sg */
532 	HPT_ASSERT(0);
533 	return FALSE;
534 }
535 
536 static void hpt_io_dmamap_callback(void *arg, bus_dma_segment_t *segs, int nsegs, int error)
537 {
538 	PCOMMAND pCmd = (PCOMMAND)arg;
539 	POS_CMDEXT ext = (POS_CMDEXT)pCmd->priv;
540 	PSG psg = pCmd->psg;
541 	int idx;
542 
543 	HPT_ASSERT(pCmd->flags.physical_sg);
544 
545 	if (error)
546 		panic("busdma error");
547 
548 	HPT_ASSERT(nsegs<=os_max_sg_descriptors);
549 
550 	if (nsegs != 0) {
551 		for (idx = 0; idx < nsegs; idx++, psg++) {
552 			psg->addr.bus = segs[idx].ds_addr;
553 			psg->size = segs[idx].ds_len;
554 			psg->eot = 0;
555 		}
556 		psg[-1].eot = 1;
557 
558 		if (pCmd->flags.data_in) {
559 			bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map,
560 			    BUS_DMASYNC_PREREAD);
561 		}
562 		else if (pCmd->flags.data_out) {
563 			bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map,
564 			    BUS_DMASYNC_PREWRITE);
565 		}
566 	}
567 
568 	callout_reset(&ext->timeout, HPT_OSM_TIMEOUT, hpt_timeout, pCmd);
569 	ldm_queue_cmd(pCmd);
570 }
571 
572 static void hpt_scsi_io(PVBUS_EXT vbus_ext, union ccb *ccb)
573 {
574 	PVBUS vbus = (PVBUS)vbus_ext->vbus;
575 	PVDEV vd;
576 	PCOMMAND pCmd;
577 	POS_CMDEXT ext;
578 	HPT_U8 *cdb;
579 
580 	if (ccb->ccb_h.flags & CAM_CDB_POINTER)
581 		cdb = ccb->csio.cdb_io.cdb_ptr;
582 	else
583 		cdb = ccb->csio.cdb_io.cdb_bytes;
584 
585 	KdPrint(("hpt_scsi_io: ccb %x id %d lun %d cdb %x-%x-%x",
586 		ccb,
587 		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
588 		*(HPT_U32 *)&cdb[0], *(HPT_U32 *)&cdb[4], *(HPT_U32 *)&cdb[8]
589 	));
590 
591 	/* ccb->ccb_h.path_id is not our bus id - don't check it */
592 	if (ccb->ccb_h.target_lun != 0 ||
593 		ccb->ccb_h.target_id >= osm_max_targets ||
594 		(ccb->ccb_h.flags & CAM_CDB_PHYS))
595 	{
596 		ccb->ccb_h.status = CAM_TID_INVALID;
597 		xpt_done(ccb);
598 		return;
599 	}
600 
601 	vd = ldm_find_target(vbus, ccb->ccb_h.target_id);
602 
603 	if (!vd) {
604 		ccb->ccb_h.status = CAM_SEL_TIMEOUT;
605 		xpt_done(ccb);
606 		return;
607 	}
608 
609 	switch (cdb[0]) {
610 	case TEST_UNIT_READY:
611 	case START_STOP_UNIT:
612 	case SYNCHRONIZE_CACHE:
613 		ccb->ccb_h.status = CAM_REQ_CMP;
614 		break;
615 
616 	case 0x85: /*ATA_16*/
617 	case 0xA1: /*ATA_12*/
618 	{
619 		int error;
620 		HPT_U8 prot;
621 		PassthroughCmd *passthru;
622 
623 		if (mIsArray(vd->type)) {
624 			ccb->ccb_h.status = CAM_REQ_INVALID;
625 			break;
626 		}
627 
628 		HPT_ASSERT(vd->type == VD_RAW && vd->u.raw.legacy_disk);
629 
630 		prot = (cdb[1] & 0x1e) >> 1;
631 
632 
633 		if (prot < 3 || prot > 5)
634 		{
635 			ccb->ccb_h.status = CAM_REQ_INVALID;
636 			break;
637 		}
638 
639 		pCmd = ldm_alloc_cmds(vbus, vd->cmds_per_request);
640 		if (!pCmd) {
641 			HPT_ASSERT(0);
642 			ccb->ccb_h.status = CAM_BUSY;
643 			break;
644 		}
645 
646 		passthru = &pCmd->uCmd.Passthrough;
647 		if (cdb[0] == 0x85/*ATA_16*/) {
648 			if (cdb[1] & 0x1) {
649 				passthru->bFeaturesReg =
650 					((HPT_U16)cdb[3] << 8)
651 						| cdb[4];
652 				passthru->bSectorCountReg =
653 					((HPT_U16)cdb[5] << 8) |
654 						cdb[6];
655 				passthru->bLbaLowReg =
656 					((HPT_U16)cdb[7] << 8) |
657 						cdb[8];
658 				passthru->bLbaMidReg =
659 					((HPT_U16)cdb[9] << 8) |
660 						cdb[10];
661 				passthru->bLbaHighReg =
662 					((HPT_U16)cdb[11] << 8) |
663 						cdb[12];
664 			} else {
665 				passthru->bFeaturesReg = cdb[4];
666 				passthru->bSectorCountReg = cdb[6];
667 				passthru->bLbaLowReg = cdb[8];
668 				passthru->bLbaMidReg = cdb[10];
669 				passthru->bLbaHighReg = cdb[12];
670 			}
671 			passthru->bDriveHeadReg = cdb[13];
672 			passthru->bCommandReg = cdb[14];
673 
674 		} else { /*ATA_12*/
675 
676 			passthru->bFeaturesReg = cdb[3];
677 			passthru->bSectorCountReg = cdb[4];
678 			passthru->bLbaLowReg = cdb[5];
679 			passthru->bLbaMidReg = cdb[6];
680 			passthru->bLbaHighReg = cdb[7];
681 			passthru->bDriveHeadReg = cdb[8];
682 			passthru->bCommandReg = cdb[9];
683 		}
684 
685 		if (cdb[1] & 0xe0) {
686 
687 
688 			if (!(passthru->bCommandReg == ATA_CMD_READ_MULTI ||
689 				passthru->bCommandReg == ATA_CMD_READ_MULTI_EXT ||
690 				passthru->bCommandReg == ATA_CMD_WRITE_MULTI ||
691 				passthru->bCommandReg == ATA_CMD_WRITE_MULTI_EXT ||
692 				passthru->bCommandReg == ATA_CMD_WRITE_MULTI_FUA_EXT)
693 				) {
694 				goto error;
695 			}
696 		}
697 
698 
699 		if (passthru->bFeaturesReg == ATA_SET_FEATURES_XFER &&
700 			passthru->bCommandReg == ATA_CMD_SET_FEATURES) {
701 			goto error;
702 		}
703 
704 
705 		passthru->nSectors = ccb->csio.dxfer_len/ATA_SECTOR_SIZE;
706 		switch (prot) {
707 			default: /*None data*/
708 				break;
709 			case 4: /*PIO data in, T_DIR=1 match check*/
710 				if ((cdb[2] & 3) &&
711 					(cdb[2] & 0x8) == 0)
712 				{
713 					OsPrint(("PIO data in, T_DIR=1 match check"));
714 					goto error;
715 				}
716 				pCmd->flags.data_in = 1;
717 						break;
718 			case 5: /*PIO data out, T_DIR=0 match check*/
719 				if ((cdb[2] & 3) &&
720 					(cdb[2] & 0x8))
721 				{
722 					OsPrint(("PIO data out, T_DIR=0 match check"));
723 					goto error;
724 				}
725 
726 				pCmd->flags.data_out = 1;
727 				break;
728 		}
729 		pCmd->type = CMD_TYPE_PASSTHROUGH;
730 		pCmd->priv = ext = cmdext_get(vbus_ext);
731 		HPT_ASSERT(ext);
732 		ext->ccb = ccb;
733 		pCmd->target = vd;
734 		pCmd->done = os_cmddone;
735 		pCmd->buildsgl = os_buildsgl;
736 		pCmd->psg = ext->psg;
737 
738 		if(!ccb->csio.dxfer_len)
739 		{
740 			ldm_queue_cmd(pCmd);
741 			return;
742 		}
743 		pCmd->flags.physical_sg = 1;
744 		error = bus_dmamap_load_ccb(vbus_ext->io_dmat,
745 					ext->dma_map, ccb,
746 					hpt_io_dmamap_callback, pCmd,
747 				    	BUS_DMA_WAITOK
748 					);
749 		KdPrint(("bus_dmamap_load return %d", error));
750 		if (error && error!=EINPROGRESS) {
751 			os_printk("bus_dmamap_load error %d", error);
752 			cmdext_put(ext);
753 			ldm_free_cmds(pCmd);
754 			ccb->ccb_h.status = CAM_REQ_CMP_ERR;
755 			xpt_done(ccb);
756 		}
757 		return;
758 error:
759 		ldm_free_cmds(pCmd);
760 		ccb->ccb_h.status = CAM_REQ_INVALID;
761 		break;
762 	}
763 
764 	case INQUIRY:
765 	{
766 		PINQUIRYDATA inquiryData;
767 		HIM_DEVICE_CONFIG devconf;
768 		HPT_U8 *rbuf;
769 
770 		memset(ccb->csio.data_ptr, 0, ccb->csio.dxfer_len);
771 		inquiryData = (PINQUIRYDATA)ccb->csio.data_ptr;
772 
773 		if (cdb[1] & 1) {
774 			rbuf = (HPT_U8 *)inquiryData;
775 			switch(cdb[2]) {
776 			case 0:
777 				rbuf[0] = 0;
778 				rbuf[1] = 0;
779 				rbuf[2] = 0;
780 				rbuf[3] = 3;
781 				rbuf[4] = 0;
782 				rbuf[5] = 0x80;
783 				rbuf[6] = 0x83;
784 				ccb->ccb_h.status = CAM_REQ_CMP;
785 				break;
786 			case 0x80: {
787 				rbuf[0] = 0;
788 				rbuf[1] = 0x80;
789 				rbuf[2] = 0;
790 				if (vd->type == VD_RAW) {
791 					rbuf[3] = 20;
792 					vd->u.raw.him->get_device_config(vd->u.raw.phy_dev,&devconf);
793 					memcpy(&rbuf[4], devconf.pIdentifyData->SerialNumber, 20);
794 					ldm_ide_fixstring(&rbuf[4], 20);
795 				} else {
796 					rbuf[3] = 1;
797 					rbuf[4] = 0x20;
798 				}
799 				ccb->ccb_h.status = CAM_REQ_CMP;
800 				break;
801 			}
802 			case 0x83:
803 				rbuf[0] = 0;
804 				rbuf[1] = 0x83;
805 				rbuf[2] = 0;
806 				rbuf[3] = 12;
807 				rbuf[4] = 1;
808 				rbuf[5] = 2;
809 				rbuf[6] = 0;
810 				rbuf[7] = 8;
811 				rbuf[8] = 0;
812 				rbuf[9] = 0x19;
813 				rbuf[10] = 0x3C;
814 				rbuf[11] = 0;
815 				rbuf[12] = 0;
816 				rbuf[13] = 0;
817 				rbuf[14] = 0;
818 				rbuf[15] = 0;
819 				ccb->ccb_h.status = CAM_REQ_CMP;
820 				break;
821 			default:
822 				ccb->ccb_h.status = CAM_REQ_INVALID;
823 				break;
824 			}
825 
826 			break;
827 		}
828 		else if (cdb[2]) {
829 			ccb->ccb_h.status = CAM_REQ_INVALID;
830 			break;
831 		}
832 
833 		inquiryData->DeviceType = 0; /*DIRECT_ACCESS_DEVICE*/
834 		inquiryData->Versions = 5; /*SPC-3*/
835 		inquiryData->ResponseDataFormat = 2;
836 		inquiryData->AdditionalLength = 0x5b;
837 		inquiryData->CommandQueue = 1;
838 
839 		if (ccb->csio.dxfer_len > 63) {
840 			rbuf = (HPT_U8 *)inquiryData;
841 			rbuf[58] = 0x60;
842 			rbuf[59] = 0x3;
843 
844 			rbuf[64] = 0x3;
845 			rbuf[66] = 0x3;
846 			rbuf[67] = 0x20;
847 
848 		}
849 
850 		if (vd->type == VD_RAW) {
851 			vd->u.raw.him->get_device_config(vd->u.raw.phy_dev,&devconf);
852 
853 			if ((devconf.pIdentifyData->GeneralConfiguration & 0x80))
854 				inquiryData->RemovableMedia = 1;
855 
856 
857 			memcpy(&inquiryData->VendorId, "ATA     ", 8);
858 			memcpy(&inquiryData->ProductId, devconf.pIdentifyData->ModelNumber, 16);
859 			ldm_ide_fixstring((HPT_U8 *)&inquiryData->ProductId, 16);
860 			memcpy(&inquiryData->ProductRevisionLevel, devconf.pIdentifyData->FirmwareRevision, 4);
861 			ldm_ide_fixstring((HPT_U8 *)&inquiryData->ProductRevisionLevel, 4);
862 			if (inquiryData->ProductRevisionLevel[0] == 0 || inquiryData->ProductRevisionLevel[0] == ' ')
863 				memcpy(&inquiryData->ProductRevisionLevel, "n/a ", 4);
864 		} else {
865 			memcpy(&inquiryData->VendorId, "HPT     ", 8);
866 			snprintf((char *)&inquiryData->ProductId, 16, "DISK_%d_%d        ",
867 				os_get_vbus_seq(vbus_ext), vd->target_id);
868 			inquiryData->ProductId[15] = ' ';
869 			memcpy(&inquiryData->ProductRevisionLevel, "4.00", 4);
870 		}
871 
872 		ccb->ccb_h.status = CAM_REQ_CMP;
873 		break;
874 	}
875 	case READ_CAPACITY:
876 	{
877 		HPT_U8 *rbuf = ccb->csio.data_ptr;
878 		HPT_U32 cap;
879 		HPT_U8 sector_size_shift = 0;
880 		HPT_U64 new_cap;
881 		HPT_U32 sector_size = 0;
882 
883 		if (mIsArray(vd->type))
884 			sector_size_shift = vd->u.array.sector_size_shift;
885 		else{
886 			if(vd->type == VD_RAW){
887 				sector_size = vd->u.raw.logical_sector_size;
888 			}
889 
890 			switch (sector_size) {
891 				case 0x1000:
892 					KdPrint(("set 4k setctor size in READ_CAPACITY"));
893 					sector_size_shift = 3;
894 					break;
895 				default:
896 					break;
897 			}
898 		}
899 		new_cap = vd->capacity >> sector_size_shift;
900 
901 		if (new_cap > 0xfffffffful)
902 			cap = 0xffffffff;
903 		else
904 			cap = new_cap - 1;
905 
906 		rbuf[0] = (HPT_U8)(cap>>24);
907 		rbuf[1] = (HPT_U8)(cap>>16);
908 		rbuf[2] = (HPT_U8)(cap>>8);
909 		rbuf[3] = (HPT_U8)cap;
910 		rbuf[4] = 0;
911 		rbuf[5] = 0;
912 		rbuf[6] = 2 << sector_size_shift;
913 		rbuf[7] = 0;
914 
915 		ccb->ccb_h.status = CAM_REQ_CMP;
916 		break;
917 	}
918 
919 	case REPORT_LUNS:
920 	{
921 		HPT_U8 *rbuf = ccb->csio.data_ptr;
922 		memset(rbuf, 0, 16);
923 		rbuf[3] = 8;
924 		ccb->ccb_h.status = CAM_REQ_CMP;
925 		break;
926 	}
927 	case SERVICE_ACTION_IN:
928 	{
929 		HPT_U8 *rbuf = ccb->csio.data_ptr;
930 		HPT_U64	cap = 0;
931 		HPT_U8 sector_size_shift = 0;
932 		HPT_U32 sector_size = 0;
933 
934 		if(mIsArray(vd->type))
935 			sector_size_shift = vd->u.array.sector_size_shift;
936 		else{
937 			if(vd->type == VD_RAW){
938 				sector_size = vd->u.raw.logical_sector_size;
939 			}
940 
941 			switch (sector_size) {
942 				case 0x1000:
943 					KdPrint(("set 4k setctor size in SERVICE_ACTION_IN"));
944 					sector_size_shift = 3;
945 					break;
946 				default:
947 					break;
948 			}
949 		}
950 		cap = (vd->capacity >> sector_size_shift) - 1;
951 
952 		rbuf[0] = (HPT_U8)(cap>>56);
953 		rbuf[1] = (HPT_U8)(cap>>48);
954 		rbuf[2] = (HPT_U8)(cap>>40);
955 		rbuf[3] = (HPT_U8)(cap>>32);
956 		rbuf[4] = (HPT_U8)(cap>>24);
957 		rbuf[5] = (HPT_U8)(cap>>16);
958 		rbuf[6] = (HPT_U8)(cap>>8);
959 		rbuf[7] = (HPT_U8)cap;
960 		rbuf[8] = 0;
961 		rbuf[9] = 0;
962 		rbuf[10] = 2 << sector_size_shift;
963 		rbuf[11] = 0;
964 
965 		if(!mIsArray(vd->type)){
966 			rbuf[13] = vd->u.raw.logicalsectors_per_physicalsector;
967 			rbuf[14] = (HPT_U8)((vd->u.raw.lowest_aligned >> 8) & 0x3f);
968 			rbuf[15] = (HPT_U8)(vd->u.raw.lowest_aligned);
969 		}
970 
971 		ccb->ccb_h.status = CAM_REQ_CMP;
972 		break;
973 	}
974 
975 	case READ_6:
976 	case READ_10:
977 	case READ_16:
978 	case WRITE_6:
979 	case WRITE_10:
980 	case WRITE_16:
981 	case 0x13:
982 	case 0x2f:
983 	case 0x8f: /* VERIFY_16 */
984 	{
985 		int error;
986 		HPT_U8 sector_size_shift = 0;
987 		HPT_U32 sector_size = 0;
988 		pCmd = ldm_alloc_cmds(vbus, vd->cmds_per_request);
989 		if(!pCmd){
990 			KdPrint(("Failed to allocate command!"));
991 			ccb->ccb_h.status = CAM_BUSY;
992 			break;
993 		}
994 
995 		switch (cdb[0])	{
996 		case READ_6:
997 		case WRITE_6:
998 		case 0x13:
999 			pCmd->uCmd.Ide.Lba =  ((HPT_U32)cdb[1] << 16) | ((HPT_U32)cdb[2] << 8) | (HPT_U32)cdb[3];
1000 			pCmd->uCmd.Ide.nSectors = (HPT_U16) cdb[4];
1001 			break;
1002 		case READ_16:
1003 		case WRITE_16:
1004 		case 0x8f: /* VERIFY_16 */
1005 		{
1006 			HPT_U64 block =
1007 				((HPT_U64)cdb[2]<<56) |
1008 				((HPT_U64)cdb[3]<<48) |
1009 				((HPT_U64)cdb[4]<<40) |
1010 				((HPT_U64)cdb[5]<<32) |
1011 				((HPT_U64)cdb[6]<<24) |
1012 				((HPT_U64)cdb[7]<<16) |
1013 				((HPT_U64)cdb[8]<<8) |
1014 				((HPT_U64)cdb[9]);
1015 			pCmd->uCmd.Ide.Lba = block;
1016 			pCmd->uCmd.Ide.nSectors = (HPT_U16)cdb[13] | ((HPT_U16)cdb[12]<<8);
1017 			break;
1018 		}
1019 
1020 		default:
1021 			pCmd->uCmd.Ide.Lba = (HPT_U32)cdb[5] | ((HPT_U32)cdb[4] << 8) | ((HPT_U32)cdb[3] << 16) | ((HPT_U32)cdb[2] << 24);
1022 			pCmd->uCmd.Ide.nSectors = (HPT_U16) cdb[8] | ((HPT_U16)cdb[7]<<8);
1023 			break;
1024 		}
1025 
1026 		if(mIsArray(vd->type)) {
1027 			sector_size_shift = vd->u.array.sector_size_shift;
1028 		}
1029 		else{
1030 			if(vd->type == VD_RAW){
1031 				sector_size = vd->u.raw.logical_sector_size;
1032 			}
1033 
1034 			switch (sector_size) {
1035 				case 0x1000:
1036 					KdPrint(("<8>resize sector size from 4k to 512"));
1037 					sector_size_shift = 3;
1038 					break;
1039 				default:
1040 					break;
1041 	 		}
1042 		}
1043 		pCmd->uCmd.Ide.Lba <<= sector_size_shift;
1044 		pCmd->uCmd.Ide.nSectors <<= sector_size_shift;
1045 
1046 
1047 		switch (cdb[0]) {
1048 		case READ_6:
1049 		case READ_10:
1050 		case READ_16:
1051 			pCmd->flags.data_in = 1;
1052 			break;
1053 		case WRITE_6:
1054 		case WRITE_10:
1055 		case WRITE_16:
1056 			pCmd->flags.data_out = 1;
1057 			break;
1058 		}
1059 		pCmd->priv = ext = cmdext_get(vbus_ext);
1060 		HPT_ASSERT(ext);
1061 		ext->ccb = ccb;
1062 		pCmd->target = vd;
1063 		pCmd->done = os_cmddone;
1064 		pCmd->buildsgl = os_buildsgl;
1065 		pCmd->psg = ext->psg;
1066 		pCmd->flags.physical_sg = 1;
1067 		error = bus_dmamap_load_ccb(vbus_ext->io_dmat,
1068 					ext->dma_map, ccb,
1069 					hpt_io_dmamap_callback, pCmd,
1070 				    	BUS_DMA_WAITOK
1071 					);
1072 		KdPrint(("bus_dmamap_load return %d", error));
1073 		if (error && error!=EINPROGRESS) {
1074 			os_printk("bus_dmamap_load error %d", error);
1075 			cmdext_put(ext);
1076 			ldm_free_cmds(pCmd);
1077 			ccb->ccb_h.status = CAM_REQ_CMP_ERR;
1078 			xpt_done(ccb);
1079 		}
1080 		return;
1081 	}
1082 
1083 	default:
1084 		ccb->ccb_h.status = CAM_REQ_INVALID;
1085 		break;
1086 	}
1087 
1088 	xpt_done(ccb);
1089 	return;
1090 }
1091 
1092 static void hpt_action(struct cam_sim *sim, union ccb *ccb)
1093 {
1094 	PVBUS_EXT vbus_ext = (PVBUS_EXT)cam_sim_softc(sim);
1095 
1096 	KdPrint(("hpt_action(fn=%d, id=%d)", ccb->ccb_h.func_code, ccb->ccb_h.target_id));
1097 
1098 	hpt_assert_vbus_locked(vbus_ext);
1099 	switch (ccb->ccb_h.func_code) {
1100 
1101 	case XPT_SCSI_IO:
1102 		hpt_scsi_io(vbus_ext, ccb);
1103 		return;
1104 
1105 	case XPT_RESET_BUS:
1106 		ldm_reset_vbus((PVBUS)vbus_ext->vbus);
1107 		break;
1108 
1109 	case XPT_GET_TRAN_SETTINGS:
1110 	case XPT_SET_TRAN_SETTINGS:
1111 		ccb->ccb_h.status = CAM_FUNC_NOTAVAIL;
1112 		break;
1113 
1114 	case XPT_CALC_GEOMETRY:
1115 		ccb->ccg.heads = 255;
1116 		ccb->ccg.secs_per_track = 63;
1117 		ccb->ccg.cylinders = ccb->ccg.volume_size / (ccb->ccg.heads * ccb->ccg.secs_per_track);
1118 		ccb->ccb_h.status = CAM_REQ_CMP;
1119 		break;
1120 
1121 	case XPT_PATH_INQ:
1122 	{
1123 		struct ccb_pathinq *cpi = &ccb->cpi;
1124 
1125 		cpi->version_num = 1;
1126 		cpi->hba_inquiry = PI_SDTR_ABLE;
1127 		cpi->target_sprt = 0;
1128 		cpi->hba_misc = PIM_NOBUSRESET;
1129 		cpi->hba_eng_cnt = 0;
1130 		cpi->max_target = osm_max_targets;
1131 		cpi->max_lun = 0;
1132 		cpi->unit_number = cam_sim_unit(sim);
1133 		cpi->bus_id = cam_sim_bus(sim);
1134 		cpi->initiator_id = osm_max_targets;
1135 		cpi->base_transfer_speed = 3300;
1136 
1137 		strlcpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN);
1138 		strlcpy(cpi->hba_vid, "HPT   ", HBA_IDLEN);
1139 		strlcpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN);
1140 		cpi->transport = XPORT_SPI;
1141 		cpi->transport_version = 2;
1142 		cpi->protocol = PROTO_SCSI;
1143 		cpi->protocol_version = SCSI_REV_2;
1144 		cpi->ccb_h.status = CAM_REQ_CMP;
1145 		break;
1146 	}
1147 
1148 	default:
1149 		ccb->ccb_h.status = CAM_REQ_INVALID;
1150 		break;
1151 	}
1152 
1153 	xpt_done(ccb);
1154 	return;
1155 }
1156 
1157 static void hpt_pci_intr(void *arg)
1158 {
1159 	PVBUS_EXT vbus_ext = (PVBUS_EXT)arg;
1160 	hpt_lock_vbus(vbus_ext);
1161 	ldm_intr((PVBUS)vbus_ext->vbus);
1162 	hpt_unlock_vbus(vbus_ext);
1163 }
1164 
1165 static void hpt_poll(struct cam_sim *sim)
1166 {
1167 	PVBUS_EXT vbus_ext = cam_sim_softc(sim);
1168 	hpt_assert_vbus_locked(vbus_ext);
1169 	ldm_intr((PVBUS)vbus_ext->vbus);
1170 }
1171 
1172 static void hpt_async(void * callback_arg, u_int32_t code, struct cam_path * path, void * arg)
1173 {
1174 	KdPrint(("hpt_async"));
1175 }
1176 
1177 static int hpt_shutdown(device_t dev)
1178 {
1179 	KdPrint(("hpt_shutdown(dev=%p)", dev));
1180 	return 0;
1181 }
1182 
1183 static int hpt_detach(device_t dev)
1184 {
1185 	/* we don't allow the driver to be unloaded. */
1186 	return EBUSY;
1187 }
1188 
1189 static void hpt_ioctl_done(struct _IOCTL_ARG *arg)
1190 {
1191 	arg->ioctl_cmnd = 0;
1192 	wakeup(arg);
1193 }
1194 
1195 static void __hpt_do_ioctl(PVBUS_EXT vbus_ext, IOCTL_ARG *ioctl_args)
1196 {
1197 	ioctl_args->result = -1;
1198 	ioctl_args->done = hpt_ioctl_done;
1199 	ioctl_args->ioctl_cmnd = (void *)1;
1200 
1201 	hpt_lock_vbus(vbus_ext);
1202 	ldm_ioctl((PVBUS)vbus_ext->vbus, ioctl_args);
1203 
1204 	while (ioctl_args->ioctl_cmnd) {
1205 		if (hpt_sleep(vbus_ext, ioctl_args, PPAUSE, "hptctl", HPT_OSM_TIMEOUT)==0)
1206 			break;
1207 		ldm_reset_vbus((PVBUS)vbus_ext->vbus);
1208 		__hpt_do_tasks(vbus_ext);
1209 	}
1210 
1211 	/* KdPrint(("ioctl %x result %d", ioctl_args->dwIoControlCode, ioctl_args->result)); */
1212 
1213 	hpt_unlock_vbus(vbus_ext);
1214 }
1215 
1216 static void hpt_do_ioctl(IOCTL_ARG *ioctl_args)
1217 {
1218 	PVBUS vbus;
1219 	PVBUS_EXT vbus_ext;
1220 
1221 	ldm_for_each_vbus(vbus, vbus_ext) {
1222 		__hpt_do_ioctl(vbus_ext, ioctl_args);
1223 		if (ioctl_args->result!=HPT_IOCTL_RESULT_WRONG_VBUS)
1224 			return;
1225 	}
1226 }
1227 
1228 #define HPT_DO_IOCTL(code, inbuf, insize, outbuf, outsize) ({\
1229 	IOCTL_ARG arg;\
1230 	arg.dwIoControlCode = code;\
1231 	arg.lpInBuffer = inbuf;\
1232 	arg.lpOutBuffer = outbuf;\
1233 	arg.nInBufferSize = insize;\
1234 	arg.nOutBufferSize = outsize;\
1235 	arg.lpBytesReturned = 0;\
1236 	hpt_do_ioctl(&arg);\
1237 	arg.result;\
1238 })
1239 
1240 #define DEVICEID_VALID(id) ((id) && ((HPT_U32)(id)!=0xffffffff))
1241 
1242 static int hpt_get_logical_devices(DEVICEID * pIds, int nMaxCount)
1243 {
1244 	int i;
1245 	HPT_U32 count = nMaxCount-1;
1246 
1247 	if (HPT_DO_IOCTL(HPT_IOCTL_GET_LOGICAL_DEVICES,
1248 			&count, sizeof(HPT_U32), pIds, sizeof(DEVICEID)*nMaxCount))
1249 		return -1;
1250 
1251 	nMaxCount = (int)pIds[0];
1252 	for (i=0; i<nMaxCount; i++) pIds[i] = pIds[i+1];
1253 	return nMaxCount;
1254 }
1255 
1256 static int hpt_get_device_info_v3(DEVICEID id, PLOGICAL_DEVICE_INFO_V3 pInfo)
1257 {
1258 	return HPT_DO_IOCTL(HPT_IOCTL_GET_DEVICE_INFO_V3,
1259 				&id, sizeof(DEVICEID), pInfo, sizeof(LOGICAL_DEVICE_INFO_V3));
1260 }
1261 
1262 /* not belong to this file logically, but we want to use ioctl interface */
1263 static int __hpt_stop_tasks(PVBUS_EXT vbus_ext, DEVICEID id)
1264 {
1265 	LOGICAL_DEVICE_INFO_V3 devinfo;
1266 	int i, result;
1267 	DEVICEID param[2] = { id, 0 };
1268 
1269 	if (hpt_get_device_info_v3(id, &devinfo))
1270 		return -1;
1271 
1272 	if (devinfo.Type!=LDT_ARRAY)
1273 		return -1;
1274 
1275 	if (devinfo.u.array.Flags & ARRAY_FLAG_REBUILDING)
1276 		param[1] = AS_REBUILD_ABORT;
1277 	else if (devinfo.u.array.Flags & ARRAY_FLAG_VERIFYING)
1278 		param[1] = AS_VERIFY_ABORT;
1279 	else if (devinfo.u.array.Flags & ARRAY_FLAG_INITIALIZING)
1280 		param[1] = AS_INITIALIZE_ABORT;
1281 	else if (devinfo.u.array.Flags & ARRAY_FLAG_TRANSFORMING)
1282 		param[1] = AS_TRANSFORM_ABORT;
1283 	else
1284 		return -1;
1285 
1286 	KdPrint(("SET_ARRAY_STATE(%x, %d)", param[0], param[1]));
1287 	result = HPT_DO_IOCTL(HPT_IOCTL_SET_ARRAY_STATE,
1288 				param, sizeof(param), 0, 0);
1289 
1290 	for (i=0; i<devinfo.u.array.nDisk; i++)
1291 		if (DEVICEID_VALID(devinfo.u.array.Members[i]))
1292 			__hpt_stop_tasks(vbus_ext, devinfo.u.array.Members[i]);
1293 
1294 	return result;
1295 }
1296 
1297 static void hpt_stop_tasks(PVBUS_EXT vbus_ext)
1298 {
1299 	DEVICEID ids[32];
1300 	int i, count;
1301 
1302 	count = hpt_get_logical_devices((DEVICEID *)&ids, sizeof(ids)/sizeof(ids[0]));
1303 
1304 	for (i=0; i<count; i++)
1305 		__hpt_stop_tasks(vbus_ext, ids[i]);
1306 }
1307 
1308 static	d_open_t	hpt_open;
1309 static	d_close_t	hpt_close;
1310 static	d_ioctl_t	hpt_ioctl;
1311 static  int 		hpt_rescan_bus(void);
1312 
1313 static struct cdevsw hpt_cdevsw = {
1314 	.d_open =	hpt_open,
1315 	.d_close =	hpt_close,
1316 	.d_ioctl =	hpt_ioctl,
1317 	.d_name =	driver_name,
1318 	.d_version =	D_VERSION,
1319 };
1320 
1321 static struct intr_config_hook hpt_ich;
1322 
1323 /*
1324  * hpt_final_init will be called after all hpt_attach.
1325  */
1326 static void hpt_final_init(void *dummy)
1327 {
1328 	int       i,unit_number=0;
1329 	PVBUS_EXT vbus_ext;
1330 	PVBUS vbus;
1331 	PHBA hba;
1332 
1333 	/* Clear the config hook */
1334 	config_intrhook_disestablish(&hpt_ich);
1335 
1336 	/* allocate memory */
1337 	i = 0;
1338 	ldm_for_each_vbus(vbus, vbus_ext) {
1339 		if (hpt_alloc_mem(vbus_ext)) {
1340 			os_printk("out of memory");
1341 			return;
1342 		}
1343 		i++;
1344 	}
1345 
1346 	if (!i) {
1347 		if (bootverbose)
1348 			os_printk("no controller detected.");
1349 		return;
1350 	}
1351 
1352 	/* initializing hardware */
1353 	ldm_for_each_vbus(vbus, vbus_ext) {
1354 		/* make timer available here */
1355 		mtx_init(&vbus_ext->lock, "hptsleeplock", NULL, MTX_DEF);
1356 		callout_init_mtx(&vbus_ext->timer, &vbus_ext->lock, 0);
1357 		if (hpt_init_vbus(vbus_ext)) {
1358 			os_printk("fail to initialize hardware");
1359 			break; /* FIXME */
1360 		}
1361 	}
1362 
1363 	/* register CAM interface */
1364 	ldm_for_each_vbus(vbus, vbus_ext) {
1365 		struct cam_devq *devq;
1366 		struct ccb_setasync	ccb;
1367 
1368 		if (bus_dma_tag_create(NULL,/* parent */
1369 				4,	/* alignment */
1370 				BUS_SPACE_MAXADDR_32BIT+1, /* boundary */
1371 				BUS_SPACE_MAXADDR,	/* lowaddr */
1372 				BUS_SPACE_MAXADDR, 	/* highaddr */
1373 				NULL, NULL, 		/* filter, filterarg */
1374 				PAGE_SIZE * (os_max_sg_descriptors-1),	/* maxsize */
1375 				os_max_sg_descriptors,	/* nsegments */
1376 				0x10000,	/* maxsegsize */
1377 				BUS_DMA_WAITOK,		/* flags */
1378 				busdma_lock_mutex,	/* lockfunc */
1379 				&vbus_ext->lock,		/* lockfuncarg */
1380 				&vbus_ext->io_dmat	/* tag */))
1381 		{
1382 			return ;
1383 		}
1384 
1385 		for (i=0; i<os_max_queue_comm; i++) {
1386 			POS_CMDEXT ext = (POS_CMDEXT)malloc(sizeof(OS_CMDEXT), M_DEVBUF, M_WAITOK);
1387 			if (!ext) {
1388 				os_printk("Can't alloc cmdext(%d)", i);
1389 				return ;
1390 			}
1391 			ext->vbus_ext = vbus_ext;
1392 			ext->next = vbus_ext->cmdext_list;
1393 			vbus_ext->cmdext_list = ext;
1394 
1395 			if (bus_dmamap_create(vbus_ext->io_dmat, 0, &ext->dma_map)) {
1396 				os_printk("Can't create dma map(%d)", i);
1397 				return ;
1398 			}
1399 			callout_init_mtx(&ext->timeout, &vbus_ext->lock, 0);
1400 		}
1401 
1402 		if ((devq = cam_simq_alloc(os_max_queue_comm)) == NULL) {
1403 			os_printk("cam_simq_alloc failed");
1404 			return ;
1405 		}
1406 
1407 		hpt_lock_vbus(vbus_ext);
1408 		vbus_ext->sim = cam_sim_alloc(hpt_action, hpt_poll, driver_name,
1409 				vbus_ext, unit_number, &vbus_ext->lock,
1410 				os_max_queue_comm, /*tagged*/8,  devq);
1411 		unit_number++;
1412 		if (!vbus_ext->sim) {
1413 			os_printk("cam_sim_alloc failed");
1414 			cam_simq_free(devq);
1415 			hpt_unlock_vbus(vbus_ext);
1416 			return ;
1417 		}
1418 
1419 		if (xpt_bus_register(vbus_ext->sim, NULL, 0) != CAM_SUCCESS) {
1420 			os_printk("xpt_bus_register failed");
1421 			cam_sim_free(vbus_ext->sim, /*free devq*/ TRUE);
1422 			vbus_ext->sim = NULL;
1423 			return ;
1424 		}
1425 
1426 		if (xpt_create_path(&vbus_ext->path, /*periph */ NULL,
1427 				cam_sim_path(vbus_ext->sim), CAM_TARGET_WILDCARD,
1428 				CAM_LUN_WILDCARD) != CAM_REQ_CMP)
1429 		{
1430 			os_printk("xpt_create_path failed");
1431 			xpt_bus_deregister(cam_sim_path(vbus_ext->sim));
1432 			cam_sim_free(vbus_ext->sim, /*free_devq*/TRUE);
1433 			hpt_unlock_vbus(vbus_ext);
1434 			vbus_ext->sim = NULL;
1435 			return ;
1436 		}
1437 		hpt_unlock_vbus(vbus_ext);
1438 
1439 		memset(&ccb, 0, sizeof(ccb));
1440 		xpt_setup_ccb(&ccb.ccb_h, vbus_ext->path, /*priority*/5);
1441 		ccb.ccb_h.func_code = XPT_SASYNC_CB;
1442 		ccb.event_enable = AC_LOST_DEVICE;
1443 		ccb.callback = hpt_async;
1444 		ccb.callback_arg = vbus_ext;
1445 		xpt_action((union ccb *)&ccb);
1446 
1447 		for (hba = vbus_ext->hba_list; hba; hba = hba->next) {
1448 			int rid = 0;
1449 			if ((hba->irq_res = bus_alloc_resource_any(hba->pcidev,
1450 				SYS_RES_IRQ, &rid, RF_SHAREABLE | RF_ACTIVE)) == NULL)
1451 			{
1452 				os_printk("can't allocate interrupt");
1453 				return ;
1454 			}
1455 
1456 			if (bus_setup_intr(hba->pcidev, hba->irq_res, INTR_TYPE_CAM | INTR_MPSAFE,
1457 				NULL, hpt_pci_intr, vbus_ext, &hba->irq_handle))
1458 			{
1459 				os_printk("can't set up interrupt");
1460 				return ;
1461 			}
1462 			hba->ldm_adapter.him->intr_control(hba->ldm_adapter.him_handle, HPT_TRUE);
1463 
1464 		}
1465 
1466 		vbus_ext->shutdown_eh = EVENTHANDLER_REGISTER(shutdown_final,
1467 									hpt_shutdown_vbus, vbus_ext, SHUTDOWN_PRI_DEFAULT);
1468 		if (!vbus_ext->shutdown_eh)
1469 			os_printk("Shutdown event registration failed");
1470 	}
1471 
1472 	ldm_for_each_vbus(vbus, vbus_ext) {
1473 		TASK_INIT(&vbus_ext->worker, 0, (task_fn_t *)hpt_do_tasks, vbus_ext);
1474 		if (vbus_ext->tasks)
1475 			TASK_ENQUEUE(&vbus_ext->worker);
1476 	}
1477 
1478 	make_dev(&hpt_cdevsw, DRIVER_MINOR, UID_ROOT, GID_OPERATOR,
1479 	    S_IRUSR | S_IWUSR, "%s", driver_name);
1480 }
1481 
1482 #if defined(KLD_MODULE)
1483 
1484 typedef struct driverlink *driverlink_t;
1485 struct driverlink {
1486 	kobj_class_t	driver;
1487 	TAILQ_ENTRY(driverlink) link;	/* list of drivers in devclass */
1488 };
1489 
1490 typedef TAILQ_HEAD(driver_list, driverlink) driver_list_t;
1491 
1492 struct devclass {
1493 	TAILQ_ENTRY(devclass) link;
1494 	devclass_t	parent;		/* parent in devclass hierarchy */
1495 	driver_list_t	drivers;     /* bus devclasses store drivers for bus */
1496 	char		*name;
1497 	device_t	*devices;	/* array of devices indexed by unit */
1498 	int		maxunit;	/* size of devices array */
1499 };
1500 
1501 static void override_kernel_driver(void)
1502 {
1503 	driverlink_t dl, dlfirst;
1504 	driver_t *tmpdriver;
1505 	devclass_t dc = devclass_find("pci");
1506 
1507 	if (dc){
1508 		dlfirst = TAILQ_FIRST(&dc->drivers);
1509 		for (dl = dlfirst; dl; dl = TAILQ_NEXT(dl, link)) {
1510 			if(strcmp(dl->driver->name, driver_name) == 0) {
1511 				tmpdriver=dl->driver;
1512 				dl->driver=dlfirst->driver;
1513 				dlfirst->driver=tmpdriver;
1514 				break;
1515 			}
1516 		}
1517 	}
1518 }
1519 
1520 #else
1521 #define override_kernel_driver()
1522 #endif
1523 
1524 static void hpt_init(void *dummy)
1525 {
1526 	if (bootverbose)
1527 		os_printk("%s %s", driver_name_long, driver_ver);
1528 
1529 	override_kernel_driver();
1530 	init_config();
1531 
1532 	hpt_ich.ich_func = hpt_final_init;
1533 	hpt_ich.ich_arg = NULL;
1534 	if (config_intrhook_establish(&hpt_ich) != 0) {
1535 		printf("%s: cannot establish configuration hook\n",
1536 		    driver_name_long);
1537 	}
1538 
1539 }
1540 SYSINIT(hptinit, SI_SUB_CONFIGURE, SI_ORDER_FIRST, hpt_init, NULL);
1541 
1542 /*
1543  * CAM driver interface
1544  */
1545 static device_method_t driver_methods[] = {
1546 	/* Device interface */
1547 	DEVMETHOD(device_probe,		hpt_probe),
1548 	DEVMETHOD(device_attach,	hpt_attach),
1549 	DEVMETHOD(device_detach,	hpt_detach),
1550 	DEVMETHOD(device_shutdown,	hpt_shutdown),
1551 	{ 0, 0 }
1552 };
1553 
1554 static driver_t hpt_pci_driver = {
1555 	driver_name,
1556 	driver_methods,
1557 	sizeof(HBA)
1558 };
1559 
1560 #ifndef TARGETNAME
1561 #error "no TARGETNAME found"
1562 #endif
1563 
1564 /* use this to make TARGETNAME be expanded */
1565 #define __DRIVER_MODULE(p1, p2, p3, p4, p5) DRIVER_MODULE(p1, p2, p3, p4, p5)
1566 #define __MODULE_VERSION(p1, p2) MODULE_VERSION(p1, p2)
1567 #define __MODULE_DEPEND(p1, p2, p3, p4, p5) MODULE_DEPEND(p1, p2, p3, p4, p5)
1568 __DRIVER_MODULE(TARGETNAME, pci, hpt_pci_driver, 0, 0);
1569 __MODULE_VERSION(TARGETNAME, 1);
1570 __MODULE_DEPEND(TARGETNAME, cam, 1, 1, 1);
1571 
1572 static int hpt_open(struct cdev *dev, int flags, int devtype, struct thread *td)
1573 {
1574 	return 0;
1575 }
1576 
1577 static int hpt_close(struct cdev *dev, int flags, int devtype, struct thread *td)
1578 {
1579 	return 0;
1580 }
1581 
1582 static int hpt_ioctl(struct cdev *dev, u_long cmd, caddr_t data, int fflag, struct thread *td)
1583 {
1584 	PHPT_IOCTL_PARAM piop=(PHPT_IOCTL_PARAM)data;
1585 	IOCTL_ARG ioctl_args;
1586 	HPT_U32 bytesReturned = 0;
1587 
1588 	switch (cmd){
1589 	case HPT_DO_IOCONTROL:
1590 	{
1591 		if (piop->Magic == HPT_IOCTL_MAGIC || piop->Magic == HPT_IOCTL_MAGIC32) {
1592 			KdPrint(("ioctl=%x in=%p len=%d out=%p len=%d\n",
1593 				piop->dwIoControlCode,
1594 				piop->lpInBuffer,
1595 				piop->nInBufferSize,
1596 				piop->lpOutBuffer,
1597 				piop->nOutBufferSize));
1598 
1599 		memset(&ioctl_args, 0, sizeof(ioctl_args));
1600 
1601 		ioctl_args.dwIoControlCode = piop->dwIoControlCode;
1602 		ioctl_args.nInBufferSize = piop->nInBufferSize;
1603 		ioctl_args.nOutBufferSize = piop->nOutBufferSize;
1604 		ioctl_args.lpBytesReturned = &bytesReturned;
1605 
1606 		if (ioctl_args.nInBufferSize) {
1607 			ioctl_args.lpInBuffer = malloc(ioctl_args.nInBufferSize, M_DEVBUF, M_WAITOK);
1608 			if (!ioctl_args.lpInBuffer)
1609 				goto invalid;
1610 			if (copyin((void*)piop->lpInBuffer,
1611 					ioctl_args.lpInBuffer, piop->nInBufferSize))
1612 				goto invalid;
1613 		}
1614 
1615 		if (ioctl_args.nOutBufferSize) {
1616 			ioctl_args.lpOutBuffer = malloc(ioctl_args.nOutBufferSize, M_DEVBUF, M_WAITOK | M_ZERO);
1617 			if (!ioctl_args.lpOutBuffer)
1618 				goto invalid;
1619 		}
1620 
1621 		hpt_do_ioctl(&ioctl_args);
1622 
1623 		if (ioctl_args.result==HPT_IOCTL_RESULT_OK) {
1624 			if (piop->nOutBufferSize) {
1625 				if (copyout(ioctl_args.lpOutBuffer,
1626 					(void*)piop->lpOutBuffer, piop->nOutBufferSize))
1627 					goto invalid;
1628 			}
1629 			if (piop->lpBytesReturned) {
1630 				if (copyout(&bytesReturned,
1631 					(void*)piop->lpBytesReturned, sizeof(HPT_U32)))
1632 					goto invalid;
1633 			}
1634 			if (ioctl_args.lpInBuffer) free(ioctl_args.lpInBuffer, M_DEVBUF);
1635 			if (ioctl_args.lpOutBuffer) free(ioctl_args.lpOutBuffer, M_DEVBUF);
1636 			return 0;
1637 		}
1638 invalid:
1639 		if (ioctl_args.lpInBuffer) free(ioctl_args.lpInBuffer, M_DEVBUF);
1640 		if (ioctl_args.lpOutBuffer) free(ioctl_args.lpOutBuffer, M_DEVBUF);
1641 		return EFAULT;
1642 	}
1643 	return EFAULT;
1644 	}
1645 
1646 	case HPT_SCAN_BUS:
1647 	{
1648 		return hpt_rescan_bus();
1649 	}
1650 	default:
1651 		KdPrint(("invalid command!"));
1652 		return EFAULT;
1653 	}
1654 
1655 }
1656 
1657 static int	hpt_rescan_bus(void)
1658 {
1659 	union ccb			*ccb;
1660 	PVBUS 				vbus;
1661 	PVBUS_EXT			vbus_ext;
1662 
1663 	ldm_for_each_vbus(vbus, vbus_ext) {
1664 		if ((ccb = xpt_alloc_ccb()) == NULL)
1665 		{
1666 			return(ENOMEM);
1667 		}
1668 		if (xpt_create_path(&ccb->ccb_h.path, NULL, cam_sim_path(vbus_ext->sim),
1669 			CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP)
1670 		{
1671 			xpt_free_ccb(ccb);
1672 			return(EIO);
1673 		}
1674 		xpt_rescan(ccb);
1675 	}
1676 	return(0);
1677 }
1678