xref: /freebsd/sys/dev/hptnr/hptnr_osm_bsd.c (revision d4eeb02986980bf33dd56c41ceb9fc5f180c0d47)
1 /* $Id: osm_bsd.c,v 1.36 2010/05/11 03:12:11 lcn Exp $ */
2 /*-
3  * SPDX-License-Identifier: BSD-2-Clause-FreeBSD
4  *
5  * HighPoint RAID Driver for FreeBSD
6  * Copyright (C) 2005-2011 HighPoint Technologies, Inc.
7  * All rights reserved.
8  *
9  * Redistribution and use in source and binary forms, with or without
10  * modification, are permitted provided that the following conditions
11  * are met:
12  * 1. Redistributions of source code must retain the above copyright
13  *    notice, this list of conditions and the following disclaimer.
14  * 2. Redistributions in binary form must reproduce the above copyright
15  *    notice, this list of conditions and the following disclaimer in the
16  *    documentation and/or other materials provided with the distribution.
17  *
18  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
19  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
20  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
21  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
22  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
23  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
24  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
25  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
26  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
27  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
28  * SUCH DAMAGE.
29  *
30  * $FreeBSD$
31  */
32 #include <dev/hptnr/hptnr_config.h>
33 #include <dev/hptnr/os_bsd.h>
34 #include <dev/hptnr/hptintf.h>
35 int msi = 0;
36 int debug_flag = 0;
37 static HIM *hpt_match(device_t dev, int scan)
38 {
39 	PCI_ID pci_id;
40 	HIM *him;
41 	int i;
42 
43 	for (him = him_list; him; him = him->next) {
44 		for (i=0; him->get_supported_device_id(i, &pci_id); i++) {
45 			if (scan && him->get_controller_count)
46 				him->get_controller_count(&pci_id,0,0);
47 			if ((pci_get_vendor(dev) == pci_id.vid) &&
48 				(pci_get_device(dev) == pci_id.did)){
49 				return (him);
50 			}
51 		}
52 	}
53 
54 	return (NULL);
55 }
56 
57 static int hpt_probe(device_t dev)
58 {
59 	HIM *him;
60 
61 	him = hpt_match(dev, 0);
62 	if (him != NULL) {
63 		KdPrint(("hpt_probe: adapter at PCI %d:%d:%d, IRQ %d",
64 			pci_get_bus(dev), pci_get_slot(dev), pci_get_function(dev), pci_get_irq(dev)
65 			));
66 		device_set_desc(dev, him->name);
67 		return (BUS_PROBE_DEFAULT);
68 	}
69 
70 	return (ENXIO);
71 }
72 
73 static int hpt_attach(device_t dev)
74 {
75 	PHBA hba = (PHBA)device_get_softc(dev);
76 	HIM *him;
77 	PCI_ID pci_id;
78 	HPT_UINT size;
79 	PVBUS vbus;
80 	PVBUS_EXT vbus_ext;
81 
82 	KdPrint(("hpt_attach(%d/%d/%d)", pci_get_bus(dev), pci_get_slot(dev), pci_get_function(dev)));
83 
84 	him = hpt_match(dev, 1);
85 	hba->ext_type = EXT_TYPE_HBA;
86 	hba->ldm_adapter.him = him;
87 
88 	pci_enable_busmaster(dev);
89 
90 	pci_id.vid = pci_get_vendor(dev);
91 	pci_id.did = pci_get_device(dev);
92 	pci_id.rev = pci_get_revid(dev);
93 	pci_id.subsys = (HPT_U32)(pci_get_subdevice(dev)) << 16 | pci_get_subvendor(dev);
94 
95 	size = him->get_adapter_size(&pci_id);
96 	hba->ldm_adapter.him_handle = malloc(size, M_DEVBUF, M_WAITOK);
97 
98 	hba->pcidev = dev;
99 	hba->pciaddr.tree = 0;
100 	hba->pciaddr.bus = pci_get_bus(dev);
101 	hba->pciaddr.device = pci_get_slot(dev);
102 	hba->pciaddr.function = pci_get_function(dev);
103 
104 	if (!him->create_adapter(&pci_id, hba->pciaddr, hba->ldm_adapter.him_handle, hba)) {
105 		free(hba->ldm_adapter.him_handle, M_DEVBUF);
106 		return ENXIO;
107 	}
108 
109 	os_printk("adapter at PCI %d:%d:%d, IRQ %d",
110 		hba->pciaddr.bus, hba->pciaddr.device, hba->pciaddr.function, pci_get_irq(dev));
111 
112 	if (!ldm_register_adapter(&hba->ldm_adapter)) {
113 		size = ldm_get_vbus_size();
114 		vbus_ext = malloc(sizeof(VBUS_EXT) + size, M_DEVBUF, M_WAITOK |
115 			M_ZERO);
116 		vbus_ext->ext_type = EXT_TYPE_VBUS;
117 		ldm_create_vbus((PVBUS)vbus_ext->vbus, vbus_ext);
118 		ldm_register_adapter(&hba->ldm_adapter);
119 	}
120 
121 	ldm_for_each_vbus(vbus, vbus_ext) {
122 		if (hba->ldm_adapter.vbus==vbus) {
123 			hba->vbus_ext = vbus_ext;
124 			hba->next = vbus_ext->hba_list;
125 			vbus_ext->hba_list = hba;
126 			break;
127 		}
128 	}
129 	return 0;
130 }
131 
132 /*
133  * Maybe we'd better to use the bus_dmamem_alloc to alloc DMA memory,
134  * but there are some problems currently (alignment, etc).
135  */
136 static __inline void *__get_free_pages(int order)
137 {
138 	/* don't use low memory - other devices may get starved */
139 	return contigmalloc(PAGE_SIZE<<order,
140 			M_DEVBUF, M_WAITOK, BUS_SPACE_MAXADDR_24BIT, BUS_SPACE_MAXADDR, PAGE_SIZE, 0);
141 }
142 
143 static __inline void free_pages(void *p, int order)
144 {
145 	contigfree(p, PAGE_SIZE<<order, M_DEVBUF);
146 }
147 
148 static int hpt_alloc_mem(PVBUS_EXT vbus_ext)
149 {
150 	PHBA hba;
151 	struct freelist *f;
152 	HPT_UINT i;
153 	void **p;
154 
155 	for (hba = vbus_ext->hba_list; hba; hba = hba->next)
156 		hba->ldm_adapter.him->get_meminfo(hba->ldm_adapter.him_handle);
157 
158 	ldm_get_mem_info((PVBUS)vbus_ext->vbus, 0);
159 
160 	for (f=vbus_ext->freelist_head; f; f=f->next) {
161 		KdPrint(("%s: %d*%d=%d bytes",
162 			f->tag, f->count, f->size, f->count*f->size));
163 		for (i=0; i<f->count; i++) {
164 			p = (void **)malloc(f->size, M_DEVBUF, M_WAITOK);
165 			if (!p)	return (ENXIO);
166 			*p = f->head;
167 			f->head = p;
168 		}
169 	}
170 
171 	for (f=vbus_ext->freelist_dma_head; f; f=f->next) {
172 		int order, size, j;
173 
174 		HPT_ASSERT((f->size & (f->alignment-1))==0);
175 
176 		for (order=0, size=PAGE_SIZE; size<f->size; order++, size<<=1)
177 			;
178 
179 		KdPrint(("%s: %d*%d=%d bytes, order %d",
180 			f->tag, f->count, f->size, f->count*f->size, order));
181 		HPT_ASSERT(f->alignment<=PAGE_SIZE);
182 
183 		for (i=0; i<f->count;) {
184 			p = (void **)__get_free_pages(order);
185 			if (!p) return -1;
186 			for (j = size/f->size; j && i<f->count; i++,j--) {
187 				*p = f->head;
188 				*(BUS_ADDRESS *)(p+1) = (BUS_ADDRESS)vtophys(p);
189 				f->head = p;
190 				p = (void **)((unsigned long)p + f->size);
191 			}
192 		}
193 	}
194 
195 	HPT_ASSERT(PAGE_SIZE==DMAPOOL_PAGE_SIZE);
196 
197 	for (i=0; i<os_max_cache_pages; i++) {
198 		p = (void **)__get_free_pages(0);
199 		if (!p) return -1;
200 		HPT_ASSERT(((HPT_UPTR)p & (DMAPOOL_PAGE_SIZE-1))==0);
201 		dmapool_put_page((PVBUS)vbus_ext->vbus, p, (BUS_ADDRESS)vtophys(p));
202 	}
203 
204 	return 0;
205 }
206 
207 static void hpt_free_mem(PVBUS_EXT vbus_ext)
208 {
209 	struct freelist *f;
210 	void *p;
211 	int i;
212 	BUS_ADDRESS bus;
213 
214 	for (f=vbus_ext->freelist_head; f; f=f->next) {
215 #if DBG
216 		if (f->count!=f->reserved_count) {
217 			KdPrint(("memory leak for freelist %s (%d/%d)", f->tag, f->count, f->reserved_count));
218 		}
219 #endif
220 		while ((p=freelist_get(f)))
221 			free(p, M_DEVBUF);
222 	}
223 
224 	for (i=0; i<os_max_cache_pages; i++) {
225 		p = dmapool_get_page((PVBUS)vbus_ext->vbus, &bus);
226 		HPT_ASSERT(p);
227 		free_pages(p, 0);
228 	}
229 
230 	for (f=vbus_ext->freelist_dma_head; f; f=f->next) {
231 		int order, size;
232 #if DBG
233 		if (f->count!=f->reserved_count) {
234 			KdPrint(("memory leak for dma freelist %s (%d/%d)", f->tag, f->count, f->reserved_count));
235 		}
236 #endif
237 		for (order=0, size=PAGE_SIZE; size<f->size; order++, size<<=1) ;
238 
239 		while ((p=freelist_get_dma(f, &bus))) {
240 			if (order)
241 				free_pages(p, order);
242 			else {
243 			/* can't free immediately since other blocks in this page may still be in the list */
244 				if (((HPT_UPTR)p & (PAGE_SIZE-1))==0)
245 					dmapool_put_page((PVBUS)vbus_ext->vbus, p, bus);
246 			}
247 		}
248 	}
249 
250 	while ((p = dmapool_get_page((PVBUS)vbus_ext->vbus, &bus)))
251 		free_pages(p, 0);
252 }
253 
254 static int hpt_init_vbus(PVBUS_EXT vbus_ext)
255 {
256 	PHBA hba;
257 
258 	for (hba = vbus_ext->hba_list; hba; hba = hba->next)
259 		if (!hba->ldm_adapter.him->initialize(hba->ldm_adapter.him_handle)) {
260 			KdPrint(("fail to initialize %p", hba));
261 			return -1;
262 		}
263 
264 	ldm_initialize_vbus((PVBUS)vbus_ext->vbus, &vbus_ext->hba_list->ldm_adapter);
265 	return 0;
266 }
267 
268 static void hpt_flush_done(PCOMMAND pCmd)
269 {
270 	PVDEV vd = pCmd->target;
271 
272 	if (mIsArray(vd->type) && vd->u.array.transform && vd!=vd->u.array.transform->target) {
273 		vd = vd->u.array.transform->target;
274 		HPT_ASSERT(vd);
275 		pCmd->target = vd;
276 		pCmd->Result = RETURN_PENDING;
277 		vdev_queue_cmd(pCmd);
278 		return;
279 	}
280 
281 	*(int *)pCmd->priv = 1;
282 	wakeup(pCmd);
283 }
284 
285 /*
286  * flush a vdev (without retry).
287  */
288 static int hpt_flush_vdev(PVBUS_EXT vbus_ext, PVDEV vd)
289 {
290 	PCOMMAND pCmd;
291 	int result = 0, done;
292 	HPT_UINT count;
293 
294 	KdPrint(("flusing dev %p", vd));
295 
296 	hpt_assert_vbus_locked(vbus_ext);
297 
298 	if (mIsArray(vd->type) && vd->u.array.transform)
299 		count = max(vd->u.array.transform->source->cmds_per_request,
300 					vd->u.array.transform->target->cmds_per_request);
301 	else
302 		count = vd->cmds_per_request;
303 
304 	pCmd = ldm_alloc_cmds(vd->vbus, count);
305 
306 	if (!pCmd) {
307 		return -1;
308 	}
309 
310 	pCmd->type = CMD_TYPE_FLUSH;
311 	pCmd->flags.hard_flush = 1;
312 	pCmd->target = vd;
313 	pCmd->done = hpt_flush_done;
314 	done = 0;
315 	pCmd->priv = &done;
316 
317 	ldm_queue_cmd(pCmd);
318 
319 	if (!done) {
320 		while (hpt_sleep(vbus_ext, pCmd, PPAUSE, "hptfls", HPT_OSM_TIMEOUT)) {
321 			ldm_reset_vbus(vd->vbus);
322 		}
323 	}
324 
325 	KdPrint(("flush result %d", pCmd->Result));
326 
327 	if (pCmd->Result!=RETURN_SUCCESS)
328 		result = -1;
329 
330 	ldm_free_cmds(pCmd);
331 
332 	return result;
333 }
334 
335 static void hpt_stop_tasks(PVBUS_EXT vbus_ext);
336 static void hpt_shutdown_vbus(PVBUS_EXT vbus_ext, int howto)
337 {
338 	PVBUS     vbus = (PVBUS)vbus_ext->vbus;
339 	PHBA hba;
340 	int i;
341 
342 	KdPrint(("hpt_shutdown_vbus"));
343 
344 	/* stop all ctl tasks and disable the worker taskqueue */
345 	hpt_stop_tasks(vbus_ext);
346 	hpt_lock_vbus(vbus_ext);
347 	vbus_ext->worker.ta_context = 0;
348 
349 	/* flush devices */
350 	for (i=0; i<osm_max_targets; i++) {
351 		PVDEV vd = ldm_find_target(vbus, i);
352 		if (vd) {
353 			/* retry once */
354 			if (hpt_flush_vdev(vbus_ext, vd))
355 				hpt_flush_vdev(vbus_ext, vd);
356 		}
357 	}
358 
359 	ldm_shutdown(vbus);
360 	hpt_unlock_vbus(vbus_ext);
361 
362 	ldm_release_vbus(vbus);
363 
364 	for (hba=vbus_ext->hba_list; hba; hba=hba->next)
365 		bus_teardown_intr(hba->pcidev, hba->irq_res, hba->irq_handle);
366 
367 	hpt_free_mem(vbus_ext);
368 
369 	while ((hba=vbus_ext->hba_list)) {
370 		vbus_ext->hba_list = hba->next;
371 		free(hba->ldm_adapter.him_handle, M_DEVBUF);
372 	}
373 
374 	callout_drain(&vbus_ext->timer);
375 	mtx_destroy(&vbus_ext->lock);
376 	free(vbus_ext, M_DEVBUF);
377 	KdPrint(("hpt_shutdown_vbus done"));
378 }
379 
380 static void __hpt_do_tasks(PVBUS_EXT vbus_ext)
381 {
382 	OSM_TASK *tasks;
383 
384 	tasks = vbus_ext->tasks;
385 	vbus_ext->tasks = 0;
386 
387 	while (tasks) {
388 		OSM_TASK *t = tasks;
389 		tasks = t->next;
390 		t->next = 0;
391 		t->func(vbus_ext->vbus, t->data);
392 	}
393 }
394 
395 static void hpt_do_tasks(PVBUS_EXT vbus_ext, int pending)
396 {
397 	if(vbus_ext){
398 		hpt_lock_vbus(vbus_ext);
399 		__hpt_do_tasks(vbus_ext);
400 		hpt_unlock_vbus(vbus_ext);
401 	}
402 }
403 
404 static void hpt_action(struct cam_sim *sim, union ccb *ccb);
405 static void hpt_poll(struct cam_sim *sim);
406 static void hpt_async(void * callback_arg, u_int32_t code, struct cam_path * path, void * arg);
407 static void hpt_pci_intr(void *arg);
408 
409 static __inline POS_CMDEXT cmdext_get(PVBUS_EXT vbus_ext)
410 {
411 	POS_CMDEXT p = vbus_ext->cmdext_list;
412 	if (p)
413 		vbus_ext->cmdext_list = p->next;
414 	return p;
415 }
416 
417 static __inline void cmdext_put(POS_CMDEXT p)
418 {
419 	p->next = p->vbus_ext->cmdext_list;
420 	p->vbus_ext->cmdext_list = p;
421 }
422 
423 static void hpt_timeout(void *arg)
424 {
425 	PCOMMAND pCmd = (PCOMMAND)arg;
426 	POS_CMDEXT ext = (POS_CMDEXT)pCmd->priv;
427 
428 	KdPrint(("pCmd %p timeout", pCmd));
429 
430 	ldm_reset_vbus((PVBUS)ext->vbus_ext->vbus);
431 }
432 
433 static void os_cmddone(PCOMMAND pCmd)
434 {
435 	POS_CMDEXT ext = (POS_CMDEXT)pCmd->priv;
436 	union ccb *ccb = ext->ccb;
437 	HPT_U8 *cdb;
438 
439 	if (ccb->ccb_h.flags & CAM_CDB_POINTER)
440 		cdb = ccb->csio.cdb_io.cdb_ptr;
441 	else
442 		cdb = ccb->csio.cdb_io.cdb_bytes;
443 
444 	KdPrint(("os_cmddone(%p, %d)", pCmd, pCmd->Result));
445 
446 	callout_stop(&ext->timeout);
447 	switch(cdb[0]) {
448 		case 0x85: /*ATA_16*/
449 		case 0xA1: /*ATA_12*/
450 		{
451 			PassthroughCmd *passthru = &pCmd->uCmd.Passthrough;
452 			HPT_U8 *sense_buffer = (HPT_U8 *)&ccb->csio.sense_data;
453 			memset(&ccb->csio.sense_data, 0,sizeof(ccb->csio.sense_data));
454 
455 			sense_buffer[0] = 0x72; /* Response Code */
456 			sense_buffer[7] = 14; /* Additional Sense Length */
457 
458 			sense_buffer[8] = 0x9; /* ATA Return Descriptor */
459 			sense_buffer[9] = 0xc; /* Additional Descriptor Length */
460 			sense_buffer[11] = (HPT_U8)passthru->bFeaturesReg; /* Error */
461 			sense_buffer[13] = (HPT_U8)passthru->bSectorCountReg;  /* Sector Count (7:0) */
462 			sense_buffer[15] = (HPT_U8)passthru->bLbaLowReg; /* LBA Low (7:0) */
463 			sense_buffer[17] = (HPT_U8)passthru->bLbaMidReg; /* LBA Mid (7:0) */
464 			sense_buffer[19] = (HPT_U8)passthru->bLbaHighReg; /* LBA High (7:0) */
465 
466 			if ((cdb[0] == 0x85) && (cdb[1] & 0x1))
467 			{
468 				sense_buffer[10] = 1;
469 				sense_buffer[12] = (HPT_U8)(passthru->bSectorCountReg >> 8); /* Sector Count (15:8) */
470 				sense_buffer[14] = (HPT_U8)(passthru->bLbaLowReg >> 8);	/* LBA Low (15:8) */
471 				sense_buffer[16] = (HPT_U8)(passthru->bLbaMidReg >> 8); /* LBA Mid (15:8) */
472 				sense_buffer[18] = (HPT_U8)(passthru->bLbaHighReg >> 8); /* LBA High (15:8) */
473 			}
474 
475 			sense_buffer[20] = (HPT_U8)passthru->bDriveHeadReg; /* Device */
476 			sense_buffer[21] = (HPT_U8)passthru->bCommandReg; /* Status */
477 			KdPrint(("sts 0x%x err 0x%x low 0x%x mid 0x%x hig 0x%x dh 0x%x sc 0x%x",
478 					 passthru->bCommandReg,
479 					 passthru->bFeaturesReg,
480 					 passthru->bLbaLowReg,
481 					 passthru->bLbaMidReg,
482 					 passthru->bLbaHighReg,
483 					 passthru->bDriveHeadReg,
484 					 passthru->bSectorCountReg));
485 			KdPrint(("result:0x%x,bFeaturesReg:0x%04x,bSectorCountReg:0x%04x,LBA:0x%04x%04x%04x ",
486 				pCmd->Result,passthru->bFeaturesReg,passthru->bSectorCountReg,
487 				passthru->bLbaHighReg,passthru->bLbaMidReg,passthru->bLbaLowReg));
488 		}
489 		default:
490 			break;
491 	}
492 
493 	switch(pCmd->Result) {
494 	case RETURN_SUCCESS:
495 		ccb->ccb_h.status = CAM_REQ_CMP;
496 		break;
497 	case RETURN_BAD_DEVICE:
498 		ccb->ccb_h.status = CAM_DEV_NOT_THERE;
499 		break;
500 	case RETURN_DEVICE_BUSY:
501 		ccb->ccb_h.status = CAM_BUSY;
502 		break;
503 	case RETURN_INVALID_REQUEST:
504 		ccb->ccb_h.status = CAM_REQ_INVALID;
505 		break;
506 	case RETURN_SELECTION_TIMEOUT:
507 		ccb->ccb_h.status = CAM_SEL_TIMEOUT;
508 		break;
509 	case RETURN_RETRY:
510 		ccb->ccb_h.status = CAM_BUSY;
511 		break;
512 	default:
513 		ccb->ccb_h.status = CAM_SCSI_STATUS_ERROR;
514 		break;
515 	}
516 
517 	if (pCmd->flags.data_in) {
518 		bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map, BUS_DMASYNC_POSTREAD);
519 	}
520 	else if (pCmd->flags.data_out) {
521 		bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map, BUS_DMASYNC_POSTWRITE);
522 	}
523 
524 	bus_dmamap_unload(ext->vbus_ext->io_dmat, ext->dma_map);
525 
526 	cmdext_put(ext);
527 	ldm_free_cmds(pCmd);
528 	xpt_done(ccb);
529 }
530 
531 static int os_buildsgl(PCOMMAND pCmd, PSG pSg, int logical)
532 {
533 	/* since we have provided physical sg, nobody will ask us to build physical sg */
534 	HPT_ASSERT(0);
535 	return FALSE;
536 }
537 
538 static void hpt_io_dmamap_callback(void *arg, bus_dma_segment_t *segs, int nsegs, int error)
539 {
540 	PCOMMAND pCmd = (PCOMMAND)arg;
541 	POS_CMDEXT ext = (POS_CMDEXT)pCmd->priv;
542 	PSG psg = pCmd->psg;
543 	int idx;
544 
545 	HPT_ASSERT(pCmd->flags.physical_sg);
546 
547 	if (error)
548 		panic("busdma error");
549 
550 	HPT_ASSERT(nsegs<=os_max_sg_descriptors);
551 
552 	if (nsegs != 0) {
553 		for (idx = 0; idx < nsegs; idx++, psg++) {
554 			psg->addr.bus = segs[idx].ds_addr;
555 			psg->size = segs[idx].ds_len;
556 			psg->eot = 0;
557 		}
558 		psg[-1].eot = 1;
559 
560 		if (pCmd->flags.data_in) {
561 			bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map,
562 			    BUS_DMASYNC_PREREAD);
563 		}
564 		else if (pCmd->flags.data_out) {
565 			bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map,
566 			    BUS_DMASYNC_PREWRITE);
567 		}
568 	}
569 
570 	callout_reset(&ext->timeout, HPT_OSM_TIMEOUT, hpt_timeout, pCmd);
571 	ldm_queue_cmd(pCmd);
572 }
573 
574 static void hpt_scsi_io(PVBUS_EXT vbus_ext, union ccb *ccb)
575 {
576 	PVBUS vbus = (PVBUS)vbus_ext->vbus;
577 	PVDEV vd;
578 	PCOMMAND pCmd;
579 	POS_CMDEXT ext;
580 	HPT_U8 *cdb;
581 
582 	if (ccb->ccb_h.flags & CAM_CDB_POINTER)
583 		cdb = ccb->csio.cdb_io.cdb_ptr;
584 	else
585 		cdb = ccb->csio.cdb_io.cdb_bytes;
586 
587 	KdPrint(("hpt_scsi_io: ccb %x id %d lun %d cdb %x-%x-%x",
588 		ccb,
589 		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
590 		*(HPT_U32 *)&cdb[0], *(HPT_U32 *)&cdb[4], *(HPT_U32 *)&cdb[8]
591 	));
592 
593 	/* ccb->ccb_h.path_id is not our bus id - don't check it */
594 	if (ccb->ccb_h.target_lun != 0 ||
595 		ccb->ccb_h.target_id >= osm_max_targets ||
596 		(ccb->ccb_h.flags & CAM_CDB_PHYS))
597 	{
598 		ccb->ccb_h.status = CAM_TID_INVALID;
599 		xpt_done(ccb);
600 		return;
601 	}
602 
603 	vd = ldm_find_target(vbus, ccb->ccb_h.target_id);
604 
605 	if (!vd) {
606 		ccb->ccb_h.status = CAM_SEL_TIMEOUT;
607 		xpt_done(ccb);
608 		return;
609 	}
610 
611 	switch (cdb[0]) {
612 	case TEST_UNIT_READY:
613 	case START_STOP_UNIT:
614 	case SYNCHRONIZE_CACHE:
615 		ccb->ccb_h.status = CAM_REQ_CMP;
616 		break;
617 
618 	case 0x85: /*ATA_16*/
619 	case 0xA1: /*ATA_12*/
620 	{
621 		int error;
622 		HPT_U8 prot;
623 		PassthroughCmd *passthru;
624 
625 		if (mIsArray(vd->type)) {
626 			ccb->ccb_h.status = CAM_REQ_INVALID;
627 			break;
628 		}
629 
630 		HPT_ASSERT(vd->type == VD_RAW && vd->u.raw.legacy_disk);
631 
632 		prot = (cdb[1] & 0x1e) >> 1;
633 
634 
635 		if (prot < 3 || prot > 5)
636 		{
637 			ccb->ccb_h.status = CAM_REQ_INVALID;
638 			break;
639 		}
640 
641 		pCmd = ldm_alloc_cmds(vbus, vd->cmds_per_request);
642 		if (!pCmd) {
643 			HPT_ASSERT(0);
644 			ccb->ccb_h.status = CAM_BUSY;
645 			break;
646 		}
647 
648 		passthru = &pCmd->uCmd.Passthrough;
649 		if (cdb[0] == 0x85/*ATA_16*/) {
650 			if (cdb[1] & 0x1) {
651 				passthru->bFeaturesReg =
652 					((HPT_U16)cdb[3] << 8)
653 						| cdb[4];
654 				passthru->bSectorCountReg =
655 					((HPT_U16)cdb[5] << 8) |
656 						cdb[6];
657 				passthru->bLbaLowReg =
658 					((HPT_U16)cdb[7] << 8) |
659 						cdb[8];
660 				passthru->bLbaMidReg =
661 					((HPT_U16)cdb[9] << 8) |
662 						cdb[10];
663 				passthru->bLbaHighReg =
664 					((HPT_U16)cdb[11] << 8) |
665 						cdb[12];
666 			} else {
667 				passthru->bFeaturesReg = cdb[4];
668 				passthru->bSectorCountReg = cdb[6];
669 				passthru->bLbaLowReg = cdb[8];
670 				passthru->bLbaMidReg = cdb[10];
671 				passthru->bLbaHighReg = cdb[12];
672 			}
673 			passthru->bDriveHeadReg = cdb[13];
674 			passthru->bCommandReg = cdb[14];
675 
676 		} else { /*ATA_12*/
677 
678 			passthru->bFeaturesReg = cdb[3];
679 			passthru->bSectorCountReg = cdb[4];
680 			passthru->bLbaLowReg = cdb[5];
681 			passthru->bLbaMidReg = cdb[6];
682 			passthru->bLbaHighReg = cdb[7];
683 			passthru->bDriveHeadReg = cdb[8];
684 			passthru->bCommandReg = cdb[9];
685 		}
686 
687 		if (cdb[1] & 0xe0) {
688 
689 
690 			if (!(passthru->bCommandReg == ATA_CMD_READ_MULTI ||
691 				passthru->bCommandReg == ATA_CMD_READ_MULTI_EXT ||
692 				passthru->bCommandReg == ATA_CMD_WRITE_MULTI ||
693 				passthru->bCommandReg == ATA_CMD_WRITE_MULTI_EXT ||
694 				passthru->bCommandReg == ATA_CMD_WRITE_MULTI_FUA_EXT)
695 				) {
696 				goto error;
697 			}
698 		}
699 
700 
701 		if (passthru->bFeaturesReg == ATA_SET_FEATURES_XFER &&
702 			passthru->bCommandReg == ATA_CMD_SET_FEATURES) {
703 			goto error;
704 		}
705 
706 
707 		passthru->nSectors = ccb->csio.dxfer_len/ATA_SECTOR_SIZE;
708 		switch (prot) {
709 			default: /*None data*/
710 				break;
711 			case 4: /*PIO data in, T_DIR=1 match check*/
712 				if ((cdb[2] & 3) &&
713 					(cdb[2] & 0x8) == 0)
714 				{
715 					OsPrint(("PIO data in, T_DIR=1 match check"));
716 					goto error;
717 				}
718 				pCmd->flags.data_in = 1;
719 						break;
720 			case 5: /*PIO data out, T_DIR=0 match check*/
721 				if ((cdb[2] & 3) &&
722 					(cdb[2] & 0x8))
723 				{
724 					OsPrint(("PIO data out, T_DIR=0 match check"));
725 					goto error;
726 				}
727 
728 				pCmd->flags.data_out = 1;
729 				break;
730 		}
731 		pCmd->type = CMD_TYPE_PASSTHROUGH;
732 		pCmd->priv = ext = cmdext_get(vbus_ext);
733 		HPT_ASSERT(ext);
734 		ext->ccb = ccb;
735 		pCmd->target = vd;
736 		pCmd->done = os_cmddone;
737 		pCmd->buildsgl = os_buildsgl;
738 		pCmd->psg = ext->psg;
739 
740 		if(!ccb->csio.dxfer_len)
741 		{
742 			ldm_queue_cmd(pCmd);
743 			return;
744 		}
745 		pCmd->flags.physical_sg = 1;
746 		error = bus_dmamap_load_ccb(vbus_ext->io_dmat,
747 					ext->dma_map, ccb,
748 					hpt_io_dmamap_callback, pCmd,
749 				    	BUS_DMA_WAITOK
750 					);
751 		KdPrint(("bus_dmamap_load return %d", error));
752 		if (error && error!=EINPROGRESS) {
753 			os_printk("bus_dmamap_load error %d", error);
754 			cmdext_put(ext);
755 			ldm_free_cmds(pCmd);
756 			ccb->ccb_h.status = CAM_REQ_CMP_ERR;
757 			xpt_done(ccb);
758 		}
759 		return;
760 error:
761 		ldm_free_cmds(pCmd);
762 		ccb->ccb_h.status = CAM_REQ_INVALID;
763 		break;
764 	}
765 
766 	case INQUIRY:
767 	{
768 		PINQUIRYDATA inquiryData;
769 		HIM_DEVICE_CONFIG devconf;
770 		HPT_U8 *rbuf;
771 
772 		memset(ccb->csio.data_ptr, 0, ccb->csio.dxfer_len);
773 		inquiryData = (PINQUIRYDATA)ccb->csio.data_ptr;
774 
775 		if (cdb[1] & 1) {
776 			rbuf = (HPT_U8 *)inquiryData;
777 			switch(cdb[2]) {
778 			case 0:
779 				rbuf[0] = 0;
780 				rbuf[1] = 0;
781 				rbuf[2] = 0;
782 				rbuf[3] = 3;
783 				rbuf[4] = 0;
784 				rbuf[5] = 0x80;
785 				rbuf[6] = 0x83;
786 				ccb->ccb_h.status = CAM_REQ_CMP;
787 				break;
788 			case 0x80: {
789 				rbuf[0] = 0;
790 				rbuf[1] = 0x80;
791 				rbuf[2] = 0;
792 				if (vd->type == VD_RAW) {
793 					rbuf[3] = 20;
794 					vd->u.raw.him->get_device_config(vd->u.raw.phy_dev,&devconf);
795 					memcpy(&rbuf[4], devconf.pIdentifyData->SerialNumber, 20);
796 					ldm_ide_fixstring(&rbuf[4], 20);
797 				} else {
798 					rbuf[3] = 1;
799 					rbuf[4] = 0x20;
800 				}
801 				ccb->ccb_h.status = CAM_REQ_CMP;
802 				break;
803 			}
804 			case 0x83:
805 				rbuf[0] = 0;
806 				rbuf[1] = 0x83;
807 				rbuf[2] = 0;
808 				rbuf[3] = 12;
809 				rbuf[4] = 1;
810 				rbuf[5] = 2;
811 				rbuf[6] = 0;
812 				rbuf[7] = 8;
813 				rbuf[8] = 0;
814 				rbuf[9] = 0x19;
815 				rbuf[10] = 0x3C;
816 				rbuf[11] = 0;
817 				rbuf[12] = 0;
818 				rbuf[13] = 0;
819 				rbuf[14] = 0;
820 				rbuf[15] = 0;
821 				ccb->ccb_h.status = CAM_REQ_CMP;
822 				break;
823 			default:
824 				ccb->ccb_h.status = CAM_REQ_INVALID;
825 				break;
826 			}
827 
828 			break;
829 		}
830 		else if (cdb[2]) {
831 			ccb->ccb_h.status = CAM_REQ_INVALID;
832 			break;
833 		}
834 
835 		inquiryData->DeviceType = 0; /*DIRECT_ACCESS_DEVICE*/
836 		inquiryData->Versions = 5; /*SPC-3*/
837 		inquiryData->ResponseDataFormat = 2;
838 		inquiryData->AdditionalLength = 0x5b;
839 		inquiryData->CommandQueue = 1;
840 
841 		if (ccb->csio.dxfer_len > 63) {
842 			rbuf = (HPT_U8 *)inquiryData;
843 			rbuf[58] = 0x60;
844 			rbuf[59] = 0x3;
845 
846 			rbuf[64] = 0x3;
847 			rbuf[66] = 0x3;
848 			rbuf[67] = 0x20;
849 
850 		}
851 
852 		if (vd->type == VD_RAW) {
853 			vd->u.raw.him->get_device_config(vd->u.raw.phy_dev,&devconf);
854 
855 			if ((devconf.pIdentifyData->GeneralConfiguration & 0x80))
856 				inquiryData->RemovableMedia = 1;
857 
858 
859 			memcpy(&inquiryData->VendorId, "ATA     ", 8);
860 			memcpy(&inquiryData->ProductId, devconf.pIdentifyData->ModelNumber, 16);
861 			ldm_ide_fixstring((HPT_U8 *)&inquiryData->ProductId, 16);
862 			memcpy(&inquiryData->ProductRevisionLevel, devconf.pIdentifyData->FirmwareRevision, 4);
863 			ldm_ide_fixstring((HPT_U8 *)&inquiryData->ProductRevisionLevel, 4);
864 			if (inquiryData->ProductRevisionLevel[0] == 0 || inquiryData->ProductRevisionLevel[0] == ' ')
865 				memcpy(&inquiryData->ProductRevisionLevel, "n/a ", 4);
866 		} else {
867 			memcpy(&inquiryData->VendorId, "HPT     ", 8);
868 			snprintf((char *)&inquiryData->ProductId, 16, "DISK_%d_%d        ",
869 				os_get_vbus_seq(vbus_ext), vd->target_id);
870 			inquiryData->ProductId[15] = ' ';
871 			memcpy(&inquiryData->ProductRevisionLevel, "4.00", 4);
872 		}
873 
874 		ccb->ccb_h.status = CAM_REQ_CMP;
875 		break;
876 	}
877 	case READ_CAPACITY:
878 	{
879 		HPT_U8 *rbuf = ccb->csio.data_ptr;
880 		HPT_U32 cap;
881 		HPT_U8 sector_size_shift = 0;
882 		HPT_U64 new_cap;
883 		HPT_U32 sector_size = 0;
884 
885 		if (mIsArray(vd->type))
886 			sector_size_shift = vd->u.array.sector_size_shift;
887 		else{
888 			if(vd->type == VD_RAW){
889 				sector_size = vd->u.raw.logical_sector_size;
890 			}
891 
892 			switch (sector_size) {
893 				case 0x1000:
894 					KdPrint(("set 4k setctor size in READ_CAPACITY"));
895 					sector_size_shift = 3;
896 					break;
897 				default:
898 					break;
899 			}
900 		}
901 		new_cap = vd->capacity >> sector_size_shift;
902 
903 		if (new_cap > 0xfffffffful)
904 			cap = 0xffffffff;
905 		else
906 			cap = new_cap - 1;
907 
908 		rbuf[0] = (HPT_U8)(cap>>24);
909 		rbuf[1] = (HPT_U8)(cap>>16);
910 		rbuf[2] = (HPT_U8)(cap>>8);
911 		rbuf[3] = (HPT_U8)cap;
912 		rbuf[4] = 0;
913 		rbuf[5] = 0;
914 		rbuf[6] = 2 << sector_size_shift;
915 		rbuf[7] = 0;
916 
917 		ccb->ccb_h.status = CAM_REQ_CMP;
918 		break;
919 	}
920 
921 	case REPORT_LUNS:
922 	{
923 		HPT_U8 *rbuf = ccb->csio.data_ptr;
924 		memset(rbuf, 0, 16);
925 		rbuf[3] = 8;
926 		ccb->ccb_h.status = CAM_REQ_CMP;
927 		break;
928 	}
929 	case SERVICE_ACTION_IN:
930 	{
931 		HPT_U8 *rbuf = ccb->csio.data_ptr;
932 		HPT_U64	cap = 0;
933 		HPT_U8 sector_size_shift = 0;
934 		HPT_U32 sector_size = 0;
935 
936 		if(mIsArray(vd->type))
937 			sector_size_shift = vd->u.array.sector_size_shift;
938 		else{
939 			if(vd->type == VD_RAW){
940 				sector_size = vd->u.raw.logical_sector_size;
941 			}
942 
943 			switch (sector_size) {
944 				case 0x1000:
945 					KdPrint(("set 4k setctor size in SERVICE_ACTION_IN"));
946 					sector_size_shift = 3;
947 					break;
948 				default:
949 					break;
950 			}
951 		}
952 		cap = (vd->capacity >> sector_size_shift) - 1;
953 
954 		rbuf[0] = (HPT_U8)(cap>>56);
955 		rbuf[1] = (HPT_U8)(cap>>48);
956 		rbuf[2] = (HPT_U8)(cap>>40);
957 		rbuf[3] = (HPT_U8)(cap>>32);
958 		rbuf[4] = (HPT_U8)(cap>>24);
959 		rbuf[5] = (HPT_U8)(cap>>16);
960 		rbuf[6] = (HPT_U8)(cap>>8);
961 		rbuf[7] = (HPT_U8)cap;
962 		rbuf[8] = 0;
963 		rbuf[9] = 0;
964 		rbuf[10] = 2 << sector_size_shift;
965 		rbuf[11] = 0;
966 
967 		if(!mIsArray(vd->type)){
968 			rbuf[13] = vd->u.raw.logicalsectors_per_physicalsector;
969 			rbuf[14] = (HPT_U8)((vd->u.raw.lowest_aligned >> 8) & 0x3f);
970 			rbuf[15] = (HPT_U8)(vd->u.raw.lowest_aligned);
971 		}
972 
973 		ccb->ccb_h.status = CAM_REQ_CMP;
974 		break;
975 	}
976 
977 	case READ_6:
978 	case READ_10:
979 	case READ_16:
980 	case WRITE_6:
981 	case WRITE_10:
982 	case WRITE_16:
983 	case 0x13:
984 	case 0x2f:
985 	case 0x8f: /* VERIFY_16 */
986 	{
987 		int error;
988 		HPT_U8 sector_size_shift = 0;
989 		HPT_U32 sector_size = 0;
990 		pCmd = ldm_alloc_cmds(vbus, vd->cmds_per_request);
991 		if(!pCmd){
992 			KdPrint(("Failed to allocate command!"));
993 			ccb->ccb_h.status = CAM_BUSY;
994 			break;
995 		}
996 
997 		switch (cdb[0])	{
998 		case READ_6:
999 		case WRITE_6:
1000 		case 0x13:
1001 			pCmd->uCmd.Ide.Lba =  ((HPT_U32)cdb[1] << 16) | ((HPT_U32)cdb[2] << 8) | (HPT_U32)cdb[3];
1002 			pCmd->uCmd.Ide.nSectors = (HPT_U16) cdb[4];
1003 			break;
1004 		case READ_16:
1005 		case WRITE_16:
1006 		case 0x8f: /* VERIFY_16 */
1007 		{
1008 			HPT_U64 block =
1009 				((HPT_U64)cdb[2]<<56) |
1010 				((HPT_U64)cdb[3]<<48) |
1011 				((HPT_U64)cdb[4]<<40) |
1012 				((HPT_U64)cdb[5]<<32) |
1013 				((HPT_U64)cdb[6]<<24) |
1014 				((HPT_U64)cdb[7]<<16) |
1015 				((HPT_U64)cdb[8]<<8) |
1016 				((HPT_U64)cdb[9]);
1017 			pCmd->uCmd.Ide.Lba = block;
1018 			pCmd->uCmd.Ide.nSectors = (HPT_U16)cdb[13] | ((HPT_U16)cdb[12]<<8);
1019 			break;
1020 		}
1021 
1022 		default:
1023 			pCmd->uCmd.Ide.Lba = (HPT_U32)cdb[5] | ((HPT_U32)cdb[4] << 8) | ((HPT_U32)cdb[3] << 16) | ((HPT_U32)cdb[2] << 24);
1024 			pCmd->uCmd.Ide.nSectors = (HPT_U16) cdb[8] | ((HPT_U16)cdb[7]<<8);
1025 			break;
1026 		}
1027 
1028 		if(mIsArray(vd->type)) {
1029 			sector_size_shift = vd->u.array.sector_size_shift;
1030 		}
1031 		else{
1032 			if(vd->type == VD_RAW){
1033 				sector_size = vd->u.raw.logical_sector_size;
1034 			}
1035 
1036 			switch (sector_size) {
1037 				case 0x1000:
1038 					KdPrint(("<8>resize sector size from 4k to 512"));
1039 					sector_size_shift = 3;
1040 					break;
1041 				default:
1042 					break;
1043 	 		}
1044 		}
1045 		pCmd->uCmd.Ide.Lba <<= sector_size_shift;
1046 		pCmd->uCmd.Ide.nSectors <<= sector_size_shift;
1047 
1048 
1049 		switch (cdb[0]) {
1050 		case READ_6:
1051 		case READ_10:
1052 		case READ_16:
1053 			pCmd->flags.data_in = 1;
1054 			break;
1055 		case WRITE_6:
1056 		case WRITE_10:
1057 		case WRITE_16:
1058 			pCmd->flags.data_out = 1;
1059 			break;
1060 		}
1061 		pCmd->priv = ext = cmdext_get(vbus_ext);
1062 		HPT_ASSERT(ext);
1063 		ext->ccb = ccb;
1064 		pCmd->target = vd;
1065 		pCmd->done = os_cmddone;
1066 		pCmd->buildsgl = os_buildsgl;
1067 		pCmd->psg = ext->psg;
1068 		pCmd->flags.physical_sg = 1;
1069 		error = bus_dmamap_load_ccb(vbus_ext->io_dmat,
1070 					ext->dma_map, ccb,
1071 					hpt_io_dmamap_callback, pCmd,
1072 				    	BUS_DMA_WAITOK
1073 					);
1074 		KdPrint(("bus_dmamap_load return %d", error));
1075 		if (error && error!=EINPROGRESS) {
1076 			os_printk("bus_dmamap_load error %d", error);
1077 			cmdext_put(ext);
1078 			ldm_free_cmds(pCmd);
1079 			ccb->ccb_h.status = CAM_REQ_CMP_ERR;
1080 			xpt_done(ccb);
1081 		}
1082 		return;
1083 	}
1084 
1085 	default:
1086 		ccb->ccb_h.status = CAM_REQ_INVALID;
1087 		break;
1088 	}
1089 
1090 	xpt_done(ccb);
1091 	return;
1092 }
1093 
1094 static void hpt_action(struct cam_sim *sim, union ccb *ccb)
1095 {
1096 	PVBUS_EXT vbus_ext = (PVBUS_EXT)cam_sim_softc(sim);
1097 
1098 	KdPrint(("hpt_action(fn=%d, id=%d)", ccb->ccb_h.func_code, ccb->ccb_h.target_id));
1099 
1100 	hpt_assert_vbus_locked(vbus_ext);
1101 	switch (ccb->ccb_h.func_code) {
1102 
1103 	case XPT_SCSI_IO:
1104 		hpt_scsi_io(vbus_ext, ccb);
1105 		return;
1106 
1107 	case XPT_RESET_BUS:
1108 		ldm_reset_vbus((PVBUS)vbus_ext->vbus);
1109 		break;
1110 
1111 	case XPT_GET_TRAN_SETTINGS:
1112 	case XPT_SET_TRAN_SETTINGS:
1113 		ccb->ccb_h.status = CAM_FUNC_NOTAVAIL;
1114 		break;
1115 
1116 	case XPT_CALC_GEOMETRY:
1117 		ccb->ccg.heads = 255;
1118 		ccb->ccg.secs_per_track = 63;
1119 		ccb->ccg.cylinders = ccb->ccg.volume_size / (ccb->ccg.heads * ccb->ccg.secs_per_track);
1120 		ccb->ccb_h.status = CAM_REQ_CMP;
1121 		break;
1122 
1123 	case XPT_PATH_INQ:
1124 	{
1125 		struct ccb_pathinq *cpi = &ccb->cpi;
1126 
1127 		cpi->version_num = 1;
1128 		cpi->hba_inquiry = PI_SDTR_ABLE;
1129 		cpi->target_sprt = 0;
1130 		cpi->hba_misc = PIM_NOBUSRESET;
1131 		cpi->hba_eng_cnt = 0;
1132 		cpi->max_target = osm_max_targets;
1133 		cpi->max_lun = 0;
1134 		cpi->unit_number = cam_sim_unit(sim);
1135 		cpi->bus_id = cam_sim_bus(sim);
1136 		cpi->initiator_id = osm_max_targets;
1137 		cpi->base_transfer_speed = 3300;
1138 
1139 		strlcpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN);
1140 		strlcpy(cpi->hba_vid, "HPT   ", HBA_IDLEN);
1141 		strlcpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN);
1142 		cpi->transport = XPORT_SPI;
1143 		cpi->transport_version = 2;
1144 		cpi->protocol = PROTO_SCSI;
1145 		cpi->protocol_version = SCSI_REV_2;
1146 		cpi->ccb_h.status = CAM_REQ_CMP;
1147 		break;
1148 	}
1149 
1150 	default:
1151 		ccb->ccb_h.status = CAM_REQ_INVALID;
1152 		break;
1153 	}
1154 
1155 	xpt_done(ccb);
1156 	return;
1157 }
1158 
1159 static void hpt_pci_intr(void *arg)
1160 {
1161 	PVBUS_EXT vbus_ext = (PVBUS_EXT)arg;
1162 	hpt_lock_vbus(vbus_ext);
1163 	ldm_intr((PVBUS)vbus_ext->vbus);
1164 	hpt_unlock_vbus(vbus_ext);
1165 }
1166 
1167 static void hpt_poll(struct cam_sim *sim)
1168 {
1169 	PVBUS_EXT vbus_ext = cam_sim_softc(sim);
1170 	hpt_assert_vbus_locked(vbus_ext);
1171 	ldm_intr((PVBUS)vbus_ext->vbus);
1172 }
1173 
1174 static void hpt_async(void * callback_arg, u_int32_t code, struct cam_path * path, void * arg)
1175 {
1176 	KdPrint(("hpt_async"));
1177 }
1178 
1179 static int hpt_shutdown(device_t dev)
1180 {
1181 	KdPrint(("hpt_shutdown(dev=%p)", dev));
1182 	return 0;
1183 }
1184 
1185 static int hpt_detach(device_t dev)
1186 {
1187 	/* we don't allow the driver to be unloaded. */
1188 	return EBUSY;
1189 }
1190 
1191 static void hpt_ioctl_done(struct _IOCTL_ARG *arg)
1192 {
1193 	arg->ioctl_cmnd = 0;
1194 	wakeup(arg);
1195 }
1196 
1197 static void __hpt_do_ioctl(PVBUS_EXT vbus_ext, IOCTL_ARG *ioctl_args)
1198 {
1199 	ioctl_args->result = -1;
1200 	ioctl_args->done = hpt_ioctl_done;
1201 	ioctl_args->ioctl_cmnd = (void *)1;
1202 
1203 	hpt_lock_vbus(vbus_ext);
1204 	ldm_ioctl((PVBUS)vbus_ext->vbus, ioctl_args);
1205 
1206 	while (ioctl_args->ioctl_cmnd) {
1207 		if (hpt_sleep(vbus_ext, ioctl_args, PPAUSE, "hptctl", HPT_OSM_TIMEOUT)==0)
1208 			break;
1209 		ldm_reset_vbus((PVBUS)vbus_ext->vbus);
1210 		__hpt_do_tasks(vbus_ext);
1211 	}
1212 
1213 	/* KdPrint(("ioctl %x result %d", ioctl_args->dwIoControlCode, ioctl_args->result)); */
1214 
1215 	hpt_unlock_vbus(vbus_ext);
1216 }
1217 
1218 static void hpt_do_ioctl(IOCTL_ARG *ioctl_args)
1219 {
1220 	PVBUS vbus;
1221 	PVBUS_EXT vbus_ext;
1222 
1223 	ldm_for_each_vbus(vbus, vbus_ext) {
1224 		__hpt_do_ioctl(vbus_ext, ioctl_args);
1225 		if (ioctl_args->result!=HPT_IOCTL_RESULT_WRONG_VBUS)
1226 			return;
1227 	}
1228 }
1229 
1230 #define HPT_DO_IOCTL(code, inbuf, insize, outbuf, outsize) ({\
1231 	IOCTL_ARG arg;\
1232 	arg.dwIoControlCode = code;\
1233 	arg.lpInBuffer = inbuf;\
1234 	arg.lpOutBuffer = outbuf;\
1235 	arg.nInBufferSize = insize;\
1236 	arg.nOutBufferSize = outsize;\
1237 	arg.lpBytesReturned = 0;\
1238 	hpt_do_ioctl(&arg);\
1239 	arg.result;\
1240 })
1241 
1242 #define DEVICEID_VALID(id) ((id) && ((HPT_U32)(id)!=0xffffffff))
1243 
1244 static int hpt_get_logical_devices(DEVICEID * pIds, int nMaxCount)
1245 {
1246 	int i;
1247 	HPT_U32 count = nMaxCount-1;
1248 
1249 	if (HPT_DO_IOCTL(HPT_IOCTL_GET_LOGICAL_DEVICES,
1250 			&count, sizeof(HPT_U32), pIds, sizeof(DEVICEID)*nMaxCount))
1251 		return -1;
1252 
1253 	nMaxCount = (int)pIds[0];
1254 	for (i=0; i<nMaxCount; i++) pIds[i] = pIds[i+1];
1255 	return nMaxCount;
1256 }
1257 
1258 static int hpt_get_device_info_v3(DEVICEID id, PLOGICAL_DEVICE_INFO_V3 pInfo)
1259 {
1260 	return HPT_DO_IOCTL(HPT_IOCTL_GET_DEVICE_INFO_V3,
1261 				&id, sizeof(DEVICEID), pInfo, sizeof(LOGICAL_DEVICE_INFO_V3));
1262 }
1263 
1264 /* not belong to this file logically, but we want to use ioctl interface */
1265 static int __hpt_stop_tasks(PVBUS_EXT vbus_ext, DEVICEID id)
1266 {
1267 	LOGICAL_DEVICE_INFO_V3 devinfo;
1268 	int i, result;
1269 	DEVICEID param[2] = { id, 0 };
1270 
1271 	if (hpt_get_device_info_v3(id, &devinfo))
1272 		return -1;
1273 
1274 	if (devinfo.Type!=LDT_ARRAY)
1275 		return -1;
1276 
1277 	if (devinfo.u.array.Flags & ARRAY_FLAG_REBUILDING)
1278 		param[1] = AS_REBUILD_ABORT;
1279 	else if (devinfo.u.array.Flags & ARRAY_FLAG_VERIFYING)
1280 		param[1] = AS_VERIFY_ABORT;
1281 	else if (devinfo.u.array.Flags & ARRAY_FLAG_INITIALIZING)
1282 		param[1] = AS_INITIALIZE_ABORT;
1283 	else if (devinfo.u.array.Flags & ARRAY_FLAG_TRANSFORMING)
1284 		param[1] = AS_TRANSFORM_ABORT;
1285 	else
1286 		return -1;
1287 
1288 	KdPrint(("SET_ARRAY_STATE(%x, %d)", param[0], param[1]));
1289 	result = HPT_DO_IOCTL(HPT_IOCTL_SET_ARRAY_STATE,
1290 				param, sizeof(param), 0, 0);
1291 
1292 	for (i=0; i<devinfo.u.array.nDisk; i++)
1293 		if (DEVICEID_VALID(devinfo.u.array.Members[i]))
1294 			__hpt_stop_tasks(vbus_ext, devinfo.u.array.Members[i]);
1295 
1296 	return result;
1297 }
1298 
1299 static void hpt_stop_tasks(PVBUS_EXT vbus_ext)
1300 {
1301 	DEVICEID ids[32];
1302 	int i, count;
1303 
1304 	count = hpt_get_logical_devices((DEVICEID *)&ids, sizeof(ids)/sizeof(ids[0]));
1305 
1306 	for (i=0; i<count; i++)
1307 		__hpt_stop_tasks(vbus_ext, ids[i]);
1308 }
1309 
1310 static	d_open_t	hpt_open;
1311 static	d_close_t	hpt_close;
1312 static	d_ioctl_t	hpt_ioctl;
1313 static  int 		hpt_rescan_bus(void);
1314 
1315 static struct cdevsw hpt_cdevsw = {
1316 	.d_open =	hpt_open,
1317 	.d_close =	hpt_close,
1318 	.d_ioctl =	hpt_ioctl,
1319 	.d_name =	driver_name,
1320 	.d_version =	D_VERSION,
1321 };
1322 
1323 static struct intr_config_hook hpt_ich;
1324 
1325 /*
1326  * hpt_final_init will be called after all hpt_attach.
1327  */
1328 static void hpt_final_init(void *dummy)
1329 {
1330 	int       i,unit_number=0;
1331 	PVBUS_EXT vbus_ext;
1332 	PVBUS vbus;
1333 	PHBA hba;
1334 
1335 	/* Clear the config hook */
1336 	config_intrhook_disestablish(&hpt_ich);
1337 
1338 	/* allocate memory */
1339 	i = 0;
1340 	ldm_for_each_vbus(vbus, vbus_ext) {
1341 		if (hpt_alloc_mem(vbus_ext)) {
1342 			os_printk("out of memory");
1343 			return;
1344 		}
1345 		i++;
1346 	}
1347 
1348 	if (!i) {
1349 		if (bootverbose)
1350 			os_printk("no controller detected.");
1351 		return;
1352 	}
1353 
1354 	/* initializing hardware */
1355 	ldm_for_each_vbus(vbus, vbus_ext) {
1356 		/* make timer available here */
1357 		mtx_init(&vbus_ext->lock, "hptsleeplock", NULL, MTX_DEF);
1358 		callout_init_mtx(&vbus_ext->timer, &vbus_ext->lock, 0);
1359 		if (hpt_init_vbus(vbus_ext)) {
1360 			os_printk("fail to initialize hardware");
1361 			break; /* FIXME */
1362 		}
1363 	}
1364 
1365 	/* register CAM interface */
1366 	ldm_for_each_vbus(vbus, vbus_ext) {
1367 		struct cam_devq *devq;
1368 		struct ccb_setasync	ccb;
1369 
1370 		if (bus_dma_tag_create(NULL,/* parent */
1371 				4,	/* alignment */
1372 				BUS_SPACE_MAXADDR_32BIT+1, /* boundary */
1373 				BUS_SPACE_MAXADDR,	/* lowaddr */
1374 				BUS_SPACE_MAXADDR, 	/* highaddr */
1375 				NULL, NULL, 		/* filter, filterarg */
1376 				PAGE_SIZE * (os_max_sg_descriptors-1),	/* maxsize */
1377 				os_max_sg_descriptors,	/* nsegments */
1378 				0x10000,	/* maxsegsize */
1379 				BUS_DMA_WAITOK,		/* flags */
1380 				busdma_lock_mutex,	/* lockfunc */
1381 				&vbus_ext->lock,		/* lockfuncarg */
1382 				&vbus_ext->io_dmat	/* tag */))
1383 		{
1384 			return ;
1385 		}
1386 
1387 		for (i=0; i<os_max_queue_comm; i++) {
1388 			POS_CMDEXT ext = (POS_CMDEXT)malloc(sizeof(OS_CMDEXT), M_DEVBUF, M_WAITOK);
1389 			if (!ext) {
1390 				os_printk("Can't alloc cmdext(%d)", i);
1391 				return ;
1392 			}
1393 			ext->vbus_ext = vbus_ext;
1394 			ext->next = vbus_ext->cmdext_list;
1395 			vbus_ext->cmdext_list = ext;
1396 
1397 			if (bus_dmamap_create(vbus_ext->io_dmat, 0, &ext->dma_map)) {
1398 				os_printk("Can't create dma map(%d)", i);
1399 				return ;
1400 			}
1401 			callout_init_mtx(&ext->timeout, &vbus_ext->lock, 0);
1402 		}
1403 
1404 		if ((devq = cam_simq_alloc(os_max_queue_comm)) == NULL) {
1405 			os_printk("cam_simq_alloc failed");
1406 			return ;
1407 		}
1408 
1409 		hpt_lock_vbus(vbus_ext);
1410 		vbus_ext->sim = cam_sim_alloc(hpt_action, hpt_poll, driver_name,
1411 				vbus_ext, unit_number, &vbus_ext->lock,
1412 				os_max_queue_comm, /*tagged*/8,  devq);
1413 		unit_number++;
1414 		if (!vbus_ext->sim) {
1415 			os_printk("cam_sim_alloc failed");
1416 			cam_simq_free(devq);
1417 			hpt_unlock_vbus(vbus_ext);
1418 			return ;
1419 		}
1420 
1421 		if (xpt_bus_register(vbus_ext->sim, NULL, 0) != CAM_SUCCESS) {
1422 			os_printk("xpt_bus_register failed");
1423 			cam_sim_free(vbus_ext->sim, /*free devq*/ TRUE);
1424 			vbus_ext->sim = NULL;
1425 			return ;
1426 		}
1427 
1428 		if (xpt_create_path(&vbus_ext->path, /*periph */ NULL,
1429 				cam_sim_path(vbus_ext->sim), CAM_TARGET_WILDCARD,
1430 				CAM_LUN_WILDCARD) != CAM_REQ_CMP)
1431 		{
1432 			os_printk("xpt_create_path failed");
1433 			xpt_bus_deregister(cam_sim_path(vbus_ext->sim));
1434 			cam_sim_free(vbus_ext->sim, /*free_devq*/TRUE);
1435 			hpt_unlock_vbus(vbus_ext);
1436 			vbus_ext->sim = NULL;
1437 			return ;
1438 		}
1439 		hpt_unlock_vbus(vbus_ext);
1440 
1441 		memset(&ccb, 0, sizeof(ccb));
1442 		xpt_setup_ccb(&ccb.ccb_h, vbus_ext->path, /*priority*/5);
1443 		ccb.ccb_h.func_code = XPT_SASYNC_CB;
1444 		ccb.event_enable = AC_LOST_DEVICE;
1445 		ccb.callback = hpt_async;
1446 		ccb.callback_arg = vbus_ext;
1447 		xpt_action((union ccb *)&ccb);
1448 
1449 		for (hba = vbus_ext->hba_list; hba; hba = hba->next) {
1450 			int rid = 0;
1451 			if ((hba->irq_res = bus_alloc_resource_any(hba->pcidev,
1452 				SYS_RES_IRQ, &rid, RF_SHAREABLE | RF_ACTIVE)) == NULL)
1453 			{
1454 				os_printk("can't allocate interrupt");
1455 				return ;
1456 			}
1457 
1458 			if (bus_setup_intr(hba->pcidev, hba->irq_res, INTR_TYPE_CAM | INTR_MPSAFE,
1459 				NULL, hpt_pci_intr, vbus_ext, &hba->irq_handle))
1460 			{
1461 				os_printk("can't set up interrupt");
1462 				return ;
1463 			}
1464 			hba->ldm_adapter.him->intr_control(hba->ldm_adapter.him_handle, HPT_TRUE);
1465 
1466 		}
1467 
1468 		vbus_ext->shutdown_eh = EVENTHANDLER_REGISTER(shutdown_final,
1469 									hpt_shutdown_vbus, vbus_ext, SHUTDOWN_PRI_DEFAULT);
1470 		if (!vbus_ext->shutdown_eh)
1471 			os_printk("Shutdown event registration failed");
1472 	}
1473 
1474 	ldm_for_each_vbus(vbus, vbus_ext) {
1475 		TASK_INIT(&vbus_ext->worker, 0, (task_fn_t *)hpt_do_tasks, vbus_ext);
1476 		if (vbus_ext->tasks)
1477 			TASK_ENQUEUE(&vbus_ext->worker);
1478 	}
1479 
1480 	make_dev(&hpt_cdevsw, DRIVER_MINOR, UID_ROOT, GID_OPERATOR,
1481 	    S_IRUSR | S_IWUSR, "%s", driver_name);
1482 }
1483 
1484 #if defined(KLD_MODULE)
1485 
1486 typedef struct driverlink *driverlink_t;
1487 struct driverlink {
1488 	kobj_class_t	driver;
1489 	TAILQ_ENTRY(driverlink) link;	/* list of drivers in devclass */
1490 };
1491 
1492 typedef TAILQ_HEAD(driver_list, driverlink) driver_list_t;
1493 
1494 struct devclass {
1495 	TAILQ_ENTRY(devclass) link;
1496 	devclass_t	parent;		/* parent in devclass hierarchy */
1497 	driver_list_t	drivers;     /* bus devclasses store drivers for bus */
1498 	char		*name;
1499 	device_t	*devices;	/* array of devices indexed by unit */
1500 	int		maxunit;	/* size of devices array */
1501 };
1502 
1503 static void override_kernel_driver(void)
1504 {
1505 	driverlink_t dl, dlfirst;
1506 	driver_t *tmpdriver;
1507 	devclass_t dc = devclass_find("pci");
1508 
1509 	if (dc){
1510 		dlfirst = TAILQ_FIRST(&dc->drivers);
1511 		for (dl = dlfirst; dl; dl = TAILQ_NEXT(dl, link)) {
1512 			if(strcmp(dl->driver->name, driver_name) == 0) {
1513 				tmpdriver=dl->driver;
1514 				dl->driver=dlfirst->driver;
1515 				dlfirst->driver=tmpdriver;
1516 				break;
1517 			}
1518 		}
1519 	}
1520 }
1521 
1522 #else
1523 #define override_kernel_driver()
1524 #endif
1525 
1526 static void hpt_init(void *dummy)
1527 {
1528 	if (bootverbose)
1529 		os_printk("%s %s", driver_name_long, driver_ver);
1530 
1531 	override_kernel_driver();
1532 	init_config();
1533 
1534 	hpt_ich.ich_func = hpt_final_init;
1535 	hpt_ich.ich_arg = NULL;
1536 	if (config_intrhook_establish(&hpt_ich) != 0) {
1537 		printf("%s: cannot establish configuration hook\n",
1538 		    driver_name_long);
1539 	}
1540 
1541 }
1542 SYSINIT(hptinit, SI_SUB_CONFIGURE, SI_ORDER_FIRST, hpt_init, NULL);
1543 
1544 /*
1545  * CAM driver interface
1546  */
1547 static device_method_t driver_methods[] = {
1548 	/* Device interface */
1549 	DEVMETHOD(device_probe,		hpt_probe),
1550 	DEVMETHOD(device_attach,	hpt_attach),
1551 	DEVMETHOD(device_detach,	hpt_detach),
1552 	DEVMETHOD(device_shutdown,	hpt_shutdown),
1553 	{ 0, 0 }
1554 };
1555 
1556 static driver_t hpt_pci_driver = {
1557 	driver_name,
1558 	driver_methods,
1559 	sizeof(HBA)
1560 };
1561 
1562 #ifndef TARGETNAME
1563 #error "no TARGETNAME found"
1564 #endif
1565 
1566 /* use this to make TARGETNAME be expanded */
1567 #define __DRIVER_MODULE(p1, p2, p3, p4, p5) DRIVER_MODULE(p1, p2, p3, p4, p5)
1568 #define __MODULE_VERSION(p1, p2) MODULE_VERSION(p1, p2)
1569 #define __MODULE_DEPEND(p1, p2, p3, p4, p5) MODULE_DEPEND(p1, p2, p3, p4, p5)
1570 __DRIVER_MODULE(TARGETNAME, pci, hpt_pci_driver, 0, 0);
1571 __MODULE_VERSION(TARGETNAME, 1);
1572 __MODULE_DEPEND(TARGETNAME, cam, 1, 1, 1);
1573 
1574 static int hpt_open(struct cdev *dev, int flags, int devtype, struct thread *td)
1575 {
1576 	return 0;
1577 }
1578 
1579 static int hpt_close(struct cdev *dev, int flags, int devtype, struct thread *td)
1580 {
1581 	return 0;
1582 }
1583 
1584 static int hpt_ioctl(struct cdev *dev, u_long cmd, caddr_t data, int fflag, struct thread *td)
1585 {
1586 	PHPT_IOCTL_PARAM piop=(PHPT_IOCTL_PARAM)data;
1587 	IOCTL_ARG ioctl_args;
1588 	HPT_U32 bytesReturned = 0;
1589 
1590 	switch (cmd){
1591 	case HPT_DO_IOCONTROL:
1592 	{
1593 		if (piop->Magic == HPT_IOCTL_MAGIC || piop->Magic == HPT_IOCTL_MAGIC32) {
1594 			KdPrint(("ioctl=%x in=%p len=%d out=%p len=%d\n",
1595 				piop->dwIoControlCode,
1596 				piop->lpInBuffer,
1597 				piop->nInBufferSize,
1598 				piop->lpOutBuffer,
1599 				piop->nOutBufferSize));
1600 
1601 		memset(&ioctl_args, 0, sizeof(ioctl_args));
1602 
1603 		ioctl_args.dwIoControlCode = piop->dwIoControlCode;
1604 		ioctl_args.nInBufferSize = piop->nInBufferSize;
1605 		ioctl_args.nOutBufferSize = piop->nOutBufferSize;
1606 		ioctl_args.lpBytesReturned = &bytesReturned;
1607 
1608 		if (ioctl_args.nInBufferSize) {
1609 			ioctl_args.lpInBuffer = malloc(ioctl_args.nInBufferSize, M_DEVBUF, M_WAITOK);
1610 			if (!ioctl_args.lpInBuffer)
1611 				goto invalid;
1612 			if (copyin((void*)piop->lpInBuffer,
1613 					ioctl_args.lpInBuffer, piop->nInBufferSize))
1614 				goto invalid;
1615 		}
1616 
1617 		if (ioctl_args.nOutBufferSize) {
1618 			ioctl_args.lpOutBuffer = malloc(ioctl_args.nOutBufferSize, M_DEVBUF, M_WAITOK | M_ZERO);
1619 			if (!ioctl_args.lpOutBuffer)
1620 				goto invalid;
1621 		}
1622 
1623 		hpt_do_ioctl(&ioctl_args);
1624 
1625 		if (ioctl_args.result==HPT_IOCTL_RESULT_OK) {
1626 			if (piop->nOutBufferSize) {
1627 				if (copyout(ioctl_args.lpOutBuffer,
1628 					(void*)piop->lpOutBuffer, piop->nOutBufferSize))
1629 					goto invalid;
1630 			}
1631 			if (piop->lpBytesReturned) {
1632 				if (copyout(&bytesReturned,
1633 					(void*)piop->lpBytesReturned, sizeof(HPT_U32)))
1634 					goto invalid;
1635 			}
1636 			if (ioctl_args.lpInBuffer) free(ioctl_args.lpInBuffer, M_DEVBUF);
1637 			if (ioctl_args.lpOutBuffer) free(ioctl_args.lpOutBuffer, M_DEVBUF);
1638 			return 0;
1639 		}
1640 invalid:
1641 		if (ioctl_args.lpInBuffer) free(ioctl_args.lpInBuffer, M_DEVBUF);
1642 		if (ioctl_args.lpOutBuffer) free(ioctl_args.lpOutBuffer, M_DEVBUF);
1643 		return EFAULT;
1644 	}
1645 	return EFAULT;
1646 	}
1647 
1648 	case HPT_SCAN_BUS:
1649 	{
1650 		return hpt_rescan_bus();
1651 	}
1652 	default:
1653 		KdPrint(("invalid command!"));
1654 		return EFAULT;
1655 	}
1656 
1657 }
1658 
1659 static int	hpt_rescan_bus(void)
1660 {
1661 	union ccb			*ccb;
1662 	PVBUS 				vbus;
1663 	PVBUS_EXT			vbus_ext;
1664 
1665 	ldm_for_each_vbus(vbus, vbus_ext) {
1666 		if ((ccb = xpt_alloc_ccb()) == NULL)
1667 		{
1668 			return(ENOMEM);
1669 		}
1670 		if (xpt_create_path(&ccb->ccb_h.path, NULL, cam_sim_path(vbus_ext->sim),
1671 			CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP)
1672 		{
1673 			xpt_free_ccb(ccb);
1674 			return(EIO);
1675 		}
1676 		xpt_rescan(ccb);
1677 	}
1678 	return(0);
1679 }
1680