xref: /freebsd/sys/dev/tws/tws.c (revision b633e08c705fe43180567eae26923d6f6f98c8d9)
1 /*-
2  * SPDX-License-Identifier: BSD-3-Clause
3  *
4  * Copyright (c) 2010, LSI Corp.
5  * All rights reserved.
6  * Author : Manjunath Ranganathaiah
7  * Support: freebsdraid@lsi.com
8  *
9  * Redistribution and use in source and binary forms, with or without
10  * modification, are permitted provided that the following conditions
11  * are met:
12  *
13  * 1. Redistributions of source code must retain the above copyright
14  *    notice, this list of conditions and the following disclaimer.
15  * 2. Redistributions in binary form must reproduce the above copyright
16  *    notice, this list of conditions and the following disclaimer in
17  *    the documentation and/or other materials provided with the
18  *    distribution.
19  * 3. Neither the name of the <ORGANIZATION> nor the names of its
20  *    contributors may be used to endorse or promote products derived
21  *    from this software without specific prior written permission.
22  *
23  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
24  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
25  * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
26  * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
27  * COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
28  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
29  * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
30  * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
31  * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
32  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
33  * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
34  * POSSIBILITY OF SUCH DAMAGE.
35  */
36 
37 #include <sys/cdefs.h>
38 __FBSDID("$FreeBSD$");
39 
40 #include <dev/tws/tws.h>
41 #include <dev/tws/tws_services.h>
42 #include <dev/tws/tws_hdm.h>
43 
44 #include <cam/cam.h>
45 #include <cam/cam_ccb.h>
46 #include <cam/cam_xpt.h>
47 
48 MALLOC_DEFINE(M_TWS, "twsbuf", "buffers used by tws driver");
49 int tws_queue_depth = TWS_MAX_REQS;
50 int tws_enable_msi = 0;
51 int tws_enable_msix = 0;
52 
53 /* externs */
54 extern int tws_cam_attach(struct tws_softc *sc);
55 extern void tws_cam_detach(struct tws_softc *sc);
56 extern int tws_init_ctlr(struct tws_softc *sc);
57 extern boolean tws_ctlr_ready(struct tws_softc *sc);
58 extern void tws_turn_off_interrupts(struct tws_softc *sc);
59 extern void tws_q_insert_tail(struct tws_softc *sc, struct tws_request *req,
60                                 u_int8_t q_type );
61 extern struct tws_request *tws_q_remove_request(struct tws_softc *sc,
62                                    struct tws_request *req, u_int8_t q_type );
63 extern struct tws_request *tws_q_remove_head(struct tws_softc *sc,
64                                                        u_int8_t q_type );
65 extern boolean tws_get_response(struct tws_softc *sc, u_int16_t *req_id);
66 extern boolean tws_ctlr_reset(struct tws_softc *sc);
67 extern void tws_intr(void *arg);
68 extern int tws_use_32bit_sgls;
69 
70 struct tws_request *tws_get_request(struct tws_softc *sc, u_int16_t type);
71 int tws_init_connect(struct tws_softc *sc, u_int16_t mc);
72 void tws_send_event(struct tws_softc *sc, u_int8_t event);
73 uint8_t tws_get_state(struct tws_softc *sc);
74 void tws_release_request(struct tws_request *req);
75 
76 /* Function prototypes */
77 static d_open_t     tws_open;
78 static d_close_t    tws_close;
79 static d_read_t     tws_read;
80 static d_write_t    tws_write;
81 extern d_ioctl_t    tws_ioctl;
82 
83 static int tws_init(struct tws_softc *sc);
84 static void tws_dmamap_cmds_load_cbfn(void *arg, bus_dma_segment_t *segs,
85                            int nseg, int error);
86 
87 static int tws_init_reqs(struct tws_softc *sc, u_int32_t dma_mem_size);
88 static int tws_init_aen_q(struct tws_softc *sc);
89 static int tws_init_trace_q(struct tws_softc *sc);
90 static int tws_setup_irq(struct tws_softc *sc);
91 int tws_setup_intr(struct tws_softc *sc, int irqs);
92 int tws_teardown_intr(struct tws_softc *sc);
93 
94 /* Character device entry points */
95 
96 static struct cdevsw tws_cdevsw = {
97     .d_version =    D_VERSION,
98     .d_open =   tws_open,
99     .d_close =  tws_close,
100     .d_read =   tws_read,
101     .d_write =  tws_write,
102     .d_ioctl =  tws_ioctl,
103     .d_name =   "tws",
104 };
105 
106 /*
107  * In the cdevsw routines, we find our softc by using the si_drv1 member
108  * of struct cdev.  We set this variable to point to our softc in our
109  * attach routine when we create the /dev entry.
110  */
111 
112 int
113 tws_open(struct cdev *dev, int oflags, int devtype, struct thread *td)
114 {
115     struct tws_softc *sc = dev->si_drv1;
116 
117     if ( sc )
118         TWS_TRACE_DEBUG(sc, "entry", dev, oflags);
119     return (0);
120 }
121 
122 int
123 tws_close(struct cdev *dev, int fflag, int devtype, struct thread *td)
124 {
125     struct tws_softc *sc = dev->si_drv1;
126 
127     if ( sc )
128         TWS_TRACE_DEBUG(sc, "entry", dev, fflag);
129     return (0);
130 }
131 
132 int
133 tws_read(struct cdev *dev, struct uio *uio, int ioflag)
134 {
135     struct tws_softc *sc = dev->si_drv1;
136 
137     if ( sc )
138         TWS_TRACE_DEBUG(sc, "entry", dev, ioflag);
139     return (0);
140 }
141 
142 int
143 tws_write(struct cdev *dev, struct uio *uio, int ioflag)
144 {
145     struct tws_softc *sc = dev->si_drv1;
146 
147     if ( sc )
148         TWS_TRACE_DEBUG(sc, "entry", dev, ioflag);
149     return (0);
150 }
151 
152 /* PCI Support Functions */
153 
154 /*
155  * Compare the device ID of this device against the IDs that this driver
156  * supports.  If there is a match, set the description and return success.
157  */
158 static int
159 tws_probe(device_t dev)
160 {
161     static u_int8_t first_ctlr = 1;
162 
163     if ((pci_get_vendor(dev) == TWS_VENDOR_ID) &&
164         (pci_get_device(dev) == TWS_DEVICE_ID)) {
165         device_set_desc(dev, "LSI 3ware SAS/SATA Storage Controller");
166         if (first_ctlr) {
167             printf("LSI 3ware device driver for SAS/SATA storage "
168                     "controllers, version: %s\n", TWS_DRIVER_VERSION_STRING);
169             first_ctlr = 0;
170         }
171 
172         return(BUS_PROBE_DEFAULT);
173     }
174     return (ENXIO);
175 }
176 
177 /* Attach function is only called if the probe is successful. */
178 
179 static int
180 tws_attach(device_t dev)
181 {
182     struct tws_softc *sc = device_get_softc(dev);
183     u_int32_t bar;
184     int error=0,i;
185 
186     /* no tracing yet */
187     /* Look up our softc and initialize its fields. */
188     sc->tws_dev = dev;
189     sc->device_id = pci_get_device(dev);
190     sc->subvendor_id = pci_get_subvendor(dev);
191     sc->subdevice_id = pci_get_subdevice(dev);
192 
193     /* Intialize mutexes */
194     mtx_init( &sc->q_lock, "tws_q_lock", NULL, MTX_DEF);
195     mtx_init( &sc->sim_lock,  "tws_sim_lock", NULL, MTX_DEF);
196     mtx_init( &sc->gen_lock,  "tws_gen_lock", NULL, MTX_DEF);
197     mtx_init( &sc->io_lock,  "tws_io_lock", NULL, MTX_DEF | MTX_RECURSE);
198     callout_init(&sc->stats_timer, 1);
199 
200     if ( tws_init_trace_q(sc) == FAILURE )
201         printf("trace init failure\n");
202     /* send init event */
203     mtx_lock(&sc->gen_lock);
204     tws_send_event(sc, TWS_INIT_START);
205     mtx_unlock(&sc->gen_lock);
206 
207 #if _BYTE_ORDER == _BIG_ENDIAN
208     TWS_TRACE(sc, "BIG endian", 0, 0);
209 #endif
210     /* sysctl context setup */
211     sysctl_ctx_init(&sc->tws_clist);
212     sc->tws_oidp = SYSCTL_ADD_NODE(&sc->tws_clist,
213         SYSCTL_STATIC_CHILDREN(_hw), OID_AUTO,
214 	device_get_nameunit(dev), CTLFLAG_RD | CTLFLAG_MPSAFE, 0, "");
215     if ( sc->tws_oidp == NULL ) {
216         tws_log(sc, SYSCTL_TREE_NODE_ADD);
217         goto attach_fail_1;
218     }
219     SYSCTL_ADD_STRING(&sc->tws_clist, SYSCTL_CHILDREN(sc->tws_oidp),
220                       OID_AUTO, "driver_version", CTLFLAG_RD,
221                       TWS_DRIVER_VERSION_STRING, 0, "TWS driver version");
222 
223     pci_enable_busmaster(dev);
224 
225     bar = pci_read_config(dev, TWS_PCI_BAR0, 4);
226     TWS_TRACE_DEBUG(sc, "bar0 ", bar, 0);
227     bar = pci_read_config(dev, TWS_PCI_BAR1, 4);
228     bar = bar & ~TWS_BIT2;
229     TWS_TRACE_DEBUG(sc, "bar1 ", bar, 0);
230 
231     /* MFA base address is BAR2 register used for
232      * push mode. Firmware will evatualy move to
233      * pull mode during witch this needs to change
234      */
235 #ifndef TWS_PULL_MODE_ENABLE
236     sc->mfa_base = (u_int64_t)pci_read_config(dev, TWS_PCI_BAR2, 4);
237     sc->mfa_base = sc->mfa_base & ~TWS_BIT2;
238     TWS_TRACE_DEBUG(sc, "bar2 ", sc->mfa_base, 0);
239 #endif
240 
241     /* allocate MMIO register space */
242     sc->reg_res_id = TWS_PCI_BAR1; /* BAR1 offset */
243     if ((sc->reg_res = bus_alloc_resource_any(dev, SYS_RES_MEMORY,
244                                 &(sc->reg_res_id), RF_ACTIVE))
245                                 == NULL) {
246         tws_log(sc, ALLOC_MEMORY_RES);
247         goto attach_fail_1;
248     }
249     sc->bus_tag = rman_get_bustag(sc->reg_res);
250     sc->bus_handle = rman_get_bushandle(sc->reg_res);
251 
252 #ifndef TWS_PULL_MODE_ENABLE
253     /* Allocate bus space for inbound mfa */
254     sc->mfa_res_id = TWS_PCI_BAR2; /* BAR2 offset */
255     if ((sc->mfa_res = bus_alloc_resource_any(dev, SYS_RES_MEMORY,
256                           &(sc->mfa_res_id), RF_ACTIVE))
257                                 == NULL) {
258         tws_log(sc, ALLOC_MEMORY_RES);
259         goto attach_fail_2;
260     }
261     sc->bus_mfa_tag = rman_get_bustag(sc->mfa_res);
262     sc->bus_mfa_handle = rman_get_bushandle(sc->mfa_res);
263 #endif
264 
265     /* Allocate and register our interrupt. */
266     sc->intr_type = TWS_INTx; /* default */
267 
268     if ( tws_enable_msi )
269         sc->intr_type = TWS_MSI;
270     if ( tws_setup_irq(sc) == FAILURE ) {
271         tws_log(sc, ALLOC_MEMORY_RES);
272         goto attach_fail_3;
273     }
274 
275     /*
276      * Create a /dev entry for this device.  The kernel will assign us
277      * a major number automatically.  We use the unit number of this
278      * device as the minor number and name the character device
279      * "tws<unit>".
280      */
281     sc->tws_cdev = make_dev(&tws_cdevsw, device_get_unit(dev),
282         UID_ROOT, GID_OPERATOR, S_IRUSR | S_IWUSR, "tws%u",
283         device_get_unit(dev));
284     sc->tws_cdev->si_drv1 = sc;
285 
286     if ( tws_init(sc) == FAILURE ) {
287         tws_log(sc, TWS_INIT_FAILURE);
288         goto attach_fail_4;
289     }
290     if ( tws_init_ctlr(sc) == FAILURE ) {
291         tws_log(sc, TWS_CTLR_INIT_FAILURE);
292         goto attach_fail_4;
293     }
294     if ((error = tws_cam_attach(sc))) {
295         tws_log(sc, TWS_CAM_ATTACH);
296         goto attach_fail_4;
297     }
298     /* send init complete event */
299     mtx_lock(&sc->gen_lock);
300     tws_send_event(sc, TWS_INIT_COMPLETE);
301     mtx_unlock(&sc->gen_lock);
302 
303     TWS_TRACE_DEBUG(sc, "attached successfully", 0, sc->device_id);
304     return(0);
305 
306 attach_fail_4:
307     tws_teardown_intr(sc);
308     destroy_dev(sc->tws_cdev);
309     if (sc->dma_mem_phys)
310 	    bus_dmamap_unload(sc->cmd_tag, sc->cmd_map);
311     if (sc->dma_mem)
312 	    bus_dmamem_free(sc->cmd_tag, sc->dma_mem, sc->cmd_map);
313     if (sc->cmd_tag)
314 	    bus_dma_tag_destroy(sc->cmd_tag);
315 attach_fail_3:
316     for(i=0;i<sc->irqs;i++) {
317         if ( sc->irq_res[i] ){
318             if (bus_release_resource(sc->tws_dev,
319                  SYS_RES_IRQ, sc->irq_res_id[i], sc->irq_res[i]))
320                 TWS_TRACE(sc, "bus irq res", 0, 0);
321         }
322     }
323 #ifndef TWS_PULL_MODE_ENABLE
324 attach_fail_2:
325 #endif
326     if ( sc->mfa_res ){
327         if (bus_release_resource(sc->tws_dev,
328                  SYS_RES_MEMORY, sc->mfa_res_id, sc->mfa_res))
329             TWS_TRACE(sc, "bus release ", 0, sc->mfa_res_id);
330     }
331     if ( sc->reg_res ){
332         if (bus_release_resource(sc->tws_dev,
333                  SYS_RES_MEMORY, sc->reg_res_id, sc->reg_res))
334             TWS_TRACE(sc, "bus release2 ", 0, sc->reg_res_id);
335     }
336 attach_fail_1:
337     mtx_destroy(&sc->q_lock);
338     mtx_destroy(&sc->sim_lock);
339     mtx_destroy(&sc->gen_lock);
340     mtx_destroy(&sc->io_lock);
341     sysctl_ctx_free(&sc->tws_clist);
342     return (ENXIO);
343 }
344 
345 /* Detach device. */
346 
347 static int
348 tws_detach(device_t dev)
349 {
350     struct tws_softc *sc = device_get_softc(dev);
351     int i;
352     u_int32_t reg;
353 
354     TWS_TRACE_DEBUG(sc, "entry", 0, 0);
355 
356     mtx_lock(&sc->gen_lock);
357     tws_send_event(sc, TWS_UNINIT_START);
358     mtx_unlock(&sc->gen_lock);
359 
360     /* needs to disable interrupt before detaching from cam */
361     tws_turn_off_interrupts(sc);
362     /* clear door bell */
363     tws_write_reg(sc, TWS_I2O0_HOBDBC, ~0, 4);
364     reg = tws_read_reg(sc, TWS_I2O0_HIMASK, 4);
365     TWS_TRACE_DEBUG(sc, "turn-off-intr", reg, 0);
366     sc->obfl_q_overrun = false;
367     tws_init_connect(sc, 1);
368 
369     /* Teardown the state in our softc created in our attach routine. */
370     /* Disconnect the interrupt handler. */
371     tws_teardown_intr(sc);
372 
373     /* Release irq resource */
374     for(i=0;i<sc->irqs;i++) {
375         if ( sc->irq_res[i] ){
376             if (bus_release_resource(sc->tws_dev,
377                      SYS_RES_IRQ, sc->irq_res_id[i], sc->irq_res[i]))
378                 TWS_TRACE(sc, "bus release irq resource",
379                                        i, sc->irq_res_id[i]);
380         }
381     }
382     if ( sc->intr_type == TWS_MSI ) {
383         pci_release_msi(sc->tws_dev);
384     }
385 
386     tws_cam_detach(sc);
387 
388     if (sc->dma_mem_phys)
389 	    bus_dmamap_unload(sc->cmd_tag, sc->cmd_map);
390     if (sc->dma_mem)
391 	    bus_dmamem_free(sc->cmd_tag, sc->dma_mem, sc->cmd_map);
392     if (sc->cmd_tag)
393 	    bus_dma_tag_destroy(sc->cmd_tag);
394 
395     /* Release memory resource */
396     if ( sc->mfa_res ){
397         if (bus_release_resource(sc->tws_dev,
398                  SYS_RES_MEMORY, sc->mfa_res_id, sc->mfa_res))
399             TWS_TRACE(sc, "bus release mem resource", 0, sc->mfa_res_id);
400     }
401     if ( sc->reg_res ){
402         if (bus_release_resource(sc->tws_dev,
403                  SYS_RES_MEMORY, sc->reg_res_id, sc->reg_res))
404             TWS_TRACE(sc, "bus release mem resource", 0, sc->reg_res_id);
405     }
406 
407     for ( i=0; i< tws_queue_depth; i++) {
408 	    if (sc->reqs[i].dma_map)
409 		    bus_dmamap_destroy(sc->data_tag, sc->reqs[i].dma_map);
410 	    callout_drain(&sc->reqs[i].timeout);
411     }
412 
413     callout_drain(&sc->stats_timer);
414     free(sc->reqs, M_TWS);
415     free(sc->sense_bufs, M_TWS);
416     xpt_free_ccb(sc->scan_ccb);
417     if (sc->ioctl_data_mem)
418             bus_dmamem_free(sc->data_tag, sc->ioctl_data_mem, sc->ioctl_data_map);
419     if (sc->data_tag)
420 	    bus_dma_tag_destroy(sc->data_tag);
421     free(sc->aen_q.q, M_TWS);
422     free(sc->trace_q.q, M_TWS);
423     mtx_destroy(&sc->q_lock);
424     mtx_destroy(&sc->sim_lock);
425     mtx_destroy(&sc->gen_lock);
426     mtx_destroy(&sc->io_lock);
427     destroy_dev(sc->tws_cdev);
428     sysctl_ctx_free(&sc->tws_clist);
429     return (0);
430 }
431 
432 int
433 tws_setup_intr(struct tws_softc *sc, int irqs)
434 {
435     int i, error;
436 
437     for(i=0;i<irqs;i++) {
438         if (!(sc->intr_handle[i])) {
439             if ((error = bus_setup_intr(sc->tws_dev, sc->irq_res[i],
440                                     INTR_TYPE_CAM | INTR_MPSAFE,
441                                     NULL,
442                                     tws_intr, sc, &sc->intr_handle[i]))) {
443                 tws_log(sc, SETUP_INTR_RES);
444                 return(FAILURE);
445             }
446         }
447     }
448     return(SUCCESS);
449 
450 }
451 
452 int
453 tws_teardown_intr(struct tws_softc *sc)
454 {
455     int i, error;
456 
457     for(i=0;i<sc->irqs;i++) {
458         if (sc->intr_handle[i]) {
459             error = bus_teardown_intr(sc->tws_dev,
460                                       sc->irq_res[i], sc->intr_handle[i]);
461             sc->intr_handle[i] = NULL;
462         }
463     }
464     return(SUCCESS);
465 }
466 
467 static int
468 tws_setup_irq(struct tws_softc *sc)
469 {
470     int messages;
471 
472     switch(sc->intr_type) {
473         case TWS_INTx :
474             sc->irqs = 1;
475             sc->irq_res_id[0] = 0;
476             sc->irq_res[0] = bus_alloc_resource_any(sc->tws_dev, SYS_RES_IRQ,
477                             &sc->irq_res_id[0], RF_SHAREABLE | RF_ACTIVE);
478             if ( ! sc->irq_res[0] )
479                 return(FAILURE);
480             if ( tws_setup_intr(sc, sc->irqs) == FAILURE )
481                 return(FAILURE);
482             device_printf(sc->tws_dev, "Using legacy INTx\n");
483             break;
484         case TWS_MSI :
485             sc->irqs = 1;
486             sc->irq_res_id[0] = 1;
487             messages = 1;
488             if (pci_alloc_msi(sc->tws_dev, &messages) != 0 ) {
489                 TWS_TRACE(sc, "pci alloc msi fail", 0, messages);
490                 return(FAILURE);
491             }
492             sc->irq_res[0] = bus_alloc_resource_any(sc->tws_dev, SYS_RES_IRQ,
493                               &sc->irq_res_id[0], RF_SHAREABLE | RF_ACTIVE);
494 
495             if ( !sc->irq_res[0]  )
496                 return(FAILURE);
497             if ( tws_setup_intr(sc, sc->irqs) == FAILURE )
498                 return(FAILURE);
499             device_printf(sc->tws_dev, "Using MSI\n");
500             break;
501     }
502 
503     return(SUCCESS);
504 }
505 
506 static int
507 tws_init(struct tws_softc *sc)
508 {
509 
510     u_int32_t max_sg_elements;
511     u_int32_t dma_mem_size;
512     int error;
513     u_int32_t reg;
514 
515     sc->seq_id = 0;
516     if ( tws_queue_depth > TWS_MAX_REQS )
517         tws_queue_depth = TWS_MAX_REQS;
518     if (tws_queue_depth < TWS_RESERVED_REQS+1)
519         tws_queue_depth = TWS_RESERVED_REQS+1;
520     sc->is64bit = (sizeof(bus_addr_t) == 8) ? true : false;
521     max_sg_elements = (sc->is64bit && !tws_use_32bit_sgls) ?
522                                  TWS_MAX_64BIT_SG_ELEMENTS :
523                                  TWS_MAX_32BIT_SG_ELEMENTS;
524     dma_mem_size = (sizeof(struct tws_command_packet) * tws_queue_depth) +
525                              (TWS_SECTOR_SIZE) ;
526     if ( bus_dma_tag_create(bus_get_dma_tag(sc->tws_dev), /* PCI parent */
527                             TWS_ALIGNMENT,           /* alignment */
528                             0,                       /* boundary */
529                             BUS_SPACE_MAXADDR_32BIT, /* lowaddr */
530                             BUS_SPACE_MAXADDR,       /* highaddr */
531                             NULL, NULL,              /* filter, filterarg */
532                             BUS_SPACE_MAXSIZE,       /* maxsize */
533                             max_sg_elements,         /* numsegs */
534                             BUS_SPACE_MAXSIZE,       /* maxsegsize */
535                             0,                       /* flags */
536                             NULL, NULL,              /* lockfunc, lockfuncarg */
537                             &sc->parent_tag          /* tag */
538                            )) {
539         TWS_TRACE_DEBUG(sc, "DMA parent tag Create fail", max_sg_elements,
540                                                     sc->is64bit);
541         return(ENOMEM);
542     }
543     /* In bound message frame requires 16byte alignment.
544      * Outbound MF's can live with 4byte alignment - for now just
545      * use 16 for both.
546      */
547     if ( bus_dma_tag_create(sc->parent_tag,       /* parent */
548                             TWS_IN_MF_ALIGNMENT,  /* alignment */
549                             0,                    /* boundary */
550                             BUS_SPACE_MAXADDR_32BIT, /* lowaddr */
551                             BUS_SPACE_MAXADDR,    /* highaddr */
552                             NULL, NULL,           /* filter, filterarg */
553                             dma_mem_size,         /* maxsize */
554                             1,                    /* numsegs */
555                             BUS_SPACE_MAXSIZE,    /* maxsegsize */
556                             0,                    /* flags */
557                             NULL, NULL,           /* lockfunc, lockfuncarg */
558                             &sc->cmd_tag          /* tag */
559                            )) {
560         TWS_TRACE_DEBUG(sc, "DMA cmd tag Create fail", max_sg_elements, sc->is64bit);
561         return(ENOMEM);
562     }
563 
564     if (bus_dmamem_alloc(sc->cmd_tag, &sc->dma_mem,
565                     BUS_DMA_NOWAIT, &sc->cmd_map)) {
566         TWS_TRACE_DEBUG(sc, "DMA mem alloc fail", max_sg_elements, sc->is64bit);
567         return(ENOMEM);
568     }
569 
570     /* if bus_dmamem_alloc succeeds then bus_dmamap_load will succeed */
571     sc->dma_mem_phys=0;
572     error = bus_dmamap_load(sc->cmd_tag, sc->cmd_map, sc->dma_mem,
573                     dma_mem_size, tws_dmamap_cmds_load_cbfn,
574                     &sc->dma_mem_phys, 0);
575 
576    /*
577     * Create a dma tag for data buffers; size will be the maximum
578     * possible I/O size (128kB).
579     */
580     if (bus_dma_tag_create(sc->parent_tag,         /* parent */
581                            TWS_ALIGNMENT,          /* alignment */
582                            0,                      /* boundary */
583                            BUS_SPACE_MAXADDR_32BIT,/* lowaddr */
584                            BUS_SPACE_MAXADDR,      /* highaddr */
585                            NULL, NULL,             /* filter, filterarg */
586                            TWS_MAX_IO_SIZE,        /* maxsize */
587                            max_sg_elements,        /* nsegments */
588                            TWS_MAX_IO_SIZE,        /* maxsegsize */
589                            BUS_DMA_ALLOCNOW,       /* flags */
590                            busdma_lock_mutex,      /* lockfunc */
591                            &sc->io_lock,           /* lockfuncarg */
592                            &sc->data_tag           /* tag */)) {
593         TWS_TRACE_DEBUG(sc, "DMA cmd tag Create fail", max_sg_elements, sc->is64bit);
594         return(ENOMEM);
595     }
596 
597     sc->reqs = malloc(sizeof(struct tws_request) * tws_queue_depth, M_TWS,
598                       M_WAITOK | M_ZERO);
599     sc->sense_bufs = malloc(sizeof(struct tws_sense) * tws_queue_depth, M_TWS,
600                       M_WAITOK | M_ZERO);
601     sc->scan_ccb = xpt_alloc_ccb();
602     if (bus_dmamem_alloc(sc->data_tag, (void **)&sc->ioctl_data_mem,
603             (BUS_DMA_NOWAIT | BUS_DMA_ZERO), &sc->ioctl_data_map)) {
604         device_printf(sc->tws_dev, "Cannot allocate ioctl data mem\n");
605         return(ENOMEM);
606     }
607 
608     if ( !tws_ctlr_ready(sc) )
609         if( !tws_ctlr_reset(sc) )
610             return(FAILURE);
611 
612     bzero(&sc->stats, sizeof(struct tws_stats));
613     tws_init_qs(sc);
614     tws_turn_off_interrupts(sc);
615 
616     /*
617      * enable pull mode by setting bit1 .
618      * setting bit0 to 1 will enable interrupt coalesing
619      * will revisit.
620      */
621 
622 #ifdef TWS_PULL_MODE_ENABLE
623 
624     reg = tws_read_reg(sc, TWS_I2O0_CTL, 4);
625     TWS_TRACE_DEBUG(sc, "i20 ctl", reg, TWS_I2O0_CTL);
626     tws_write_reg(sc, TWS_I2O0_CTL, reg | TWS_BIT1, 4);
627 
628 #endif
629 
630     TWS_TRACE_DEBUG(sc, "dma_mem_phys", sc->dma_mem_phys, TWS_I2O0_CTL);
631     if ( tws_init_reqs(sc, dma_mem_size) == FAILURE )
632         return(FAILURE);
633     if ( tws_init_aen_q(sc) == FAILURE )
634         return(FAILURE);
635 
636     return(SUCCESS);
637 
638 }
639 
640 static int
641 tws_init_aen_q(struct tws_softc *sc)
642 {
643     sc->aen_q.head=0;
644     sc->aen_q.tail=0;
645     sc->aen_q.depth=256;
646     sc->aen_q.overflow=0;
647     sc->aen_q.q = malloc(sizeof(struct tws_event_packet)*sc->aen_q.depth,
648                               M_TWS, M_WAITOK | M_ZERO);
649     return(SUCCESS);
650 }
651 
652 static int
653 tws_init_trace_q(struct tws_softc *sc)
654 {
655     sc->trace_q.head=0;
656     sc->trace_q.tail=0;
657     sc->trace_q.depth=256;
658     sc->trace_q.overflow=0;
659     sc->trace_q.q = malloc(sizeof(struct tws_trace_rec)*sc->trace_q.depth,
660                               M_TWS, M_WAITOK | M_ZERO);
661     return(SUCCESS);
662 }
663 
664 static int
665 tws_init_reqs(struct tws_softc *sc, u_int32_t dma_mem_size)
666 {
667 
668     struct tws_command_packet *cmd_buf;
669     cmd_buf = (struct tws_command_packet *)sc->dma_mem;
670     int i;
671 
672     bzero(cmd_buf, dma_mem_size);
673     TWS_TRACE_DEBUG(sc, "phy cmd", sc->dma_mem_phys, 0);
674     mtx_lock(&sc->q_lock);
675     for ( i=0; i< tws_queue_depth; i++)
676     {
677         if (bus_dmamap_create(sc->data_tag, 0, &sc->reqs[i].dma_map)) {
678             /* log a ENOMEM failure msg here */
679             mtx_unlock(&sc->q_lock);
680             return(FAILURE);
681         }
682         sc->reqs[i].cmd_pkt =  &cmd_buf[i];
683 
684         sc->sense_bufs[i].hdr = &cmd_buf[i].hdr ;
685         sc->sense_bufs[i].hdr_pkt_phy = sc->dma_mem_phys +
686                               (i * sizeof(struct tws_command_packet));
687 
688         sc->reqs[i].cmd_pkt_phy = sc->dma_mem_phys +
689                               sizeof(struct tws_command_header) +
690                               (i * sizeof(struct tws_command_packet));
691         sc->reqs[i].request_id = i;
692         sc->reqs[i].sc = sc;
693 
694         sc->reqs[i].cmd_pkt->hdr.header_desc.size_header = 128;
695 
696 	callout_init(&sc->reqs[i].timeout, 1);
697         sc->reqs[i].state = TWS_REQ_STATE_FREE;
698         if ( i >= TWS_RESERVED_REQS )
699             tws_q_insert_tail(sc, &sc->reqs[i], TWS_FREE_Q);
700     }
701     mtx_unlock(&sc->q_lock);
702     return(SUCCESS);
703 }
704 
705 static void
706 tws_dmamap_cmds_load_cbfn(void *arg, bus_dma_segment_t *segs,
707                            int nseg, int error)
708 {
709 
710     /* printf("command load done \n"); */
711 
712     *((bus_addr_t *)arg) = segs[0].ds_addr;
713 }
714 
715 void
716 tws_send_event(struct tws_softc *sc, u_int8_t event)
717 {
718     mtx_assert(&sc->gen_lock, MA_OWNED);
719     TWS_TRACE_DEBUG(sc, "received event ", 0, event);
720     switch (event) {
721         case TWS_INIT_START:
722             sc->tws_state = TWS_INIT;
723             break;
724 
725         case TWS_INIT_COMPLETE:
726             if (sc->tws_state != TWS_INIT) {
727                 device_printf(sc->tws_dev, "invalid state transition %d => TWS_ONLINE\n", sc->tws_state);
728             } else {
729                 sc->tws_state = TWS_ONLINE;
730             }
731             break;
732 
733         case TWS_RESET_START:
734             /* We can transition to reset state from any state except reset*/
735             if (sc->tws_state != TWS_RESET) {
736                 sc->tws_prev_state = sc->tws_state;
737                 sc->tws_state = TWS_RESET;
738             }
739             break;
740 
741         case TWS_RESET_COMPLETE:
742             if (sc->tws_state != TWS_RESET) {
743                 device_printf(sc->tws_dev, "invalid state transition %d => %d (previous state)\n", sc->tws_state, sc->tws_prev_state);
744             } else {
745                 sc->tws_state = sc->tws_prev_state;
746             }
747             break;
748 
749         case TWS_SCAN_FAILURE:
750             if (sc->tws_state != TWS_ONLINE) {
751                 device_printf(sc->tws_dev, "invalid state transition %d => TWS_OFFLINE\n", sc->tws_state);
752             } else {
753                 sc->tws_state = TWS_OFFLINE;
754             }
755             break;
756 
757         case TWS_UNINIT_START:
758             if ((sc->tws_state != TWS_ONLINE) && (sc->tws_state != TWS_OFFLINE)) {
759                 device_printf(sc->tws_dev, "invalid state transition %d => TWS_UNINIT\n", sc->tws_state);
760             } else {
761                 sc->tws_state = TWS_UNINIT;
762             }
763             break;
764     }
765 
766 }
767 
768 uint8_t
769 tws_get_state(struct tws_softc *sc)
770 {
771 
772     return((u_int8_t)sc->tws_state);
773 
774 }
775 
776 /* Called during system shutdown after sync. */
777 
778 static int
779 tws_shutdown(device_t dev)
780 {
781 
782     struct tws_softc *sc = device_get_softc(dev);
783 
784     TWS_TRACE_DEBUG(sc, "entry", 0, 0);
785 
786     tws_turn_off_interrupts(sc);
787     tws_init_connect(sc, 1);
788 
789     return (0);
790 }
791 
792 /*
793  * Device suspend routine.
794  */
795 static int
796 tws_suspend(device_t dev)
797 {
798     struct tws_softc *sc = device_get_softc(dev);
799 
800     if ( sc )
801         TWS_TRACE_DEBUG(sc, "entry", 0, 0);
802     return (0);
803 }
804 
805 /*
806  * Device resume routine.
807  */
808 static int
809 tws_resume(device_t dev)
810 {
811 
812     struct tws_softc *sc = device_get_softc(dev);
813 
814     if ( sc )
815         TWS_TRACE_DEBUG(sc, "entry", 0, 0);
816     return (0);
817 }
818 
819 struct tws_request *
820 tws_get_request(struct tws_softc *sc, u_int16_t type)
821 {
822     struct mtx *my_mutex = ((type == TWS_REQ_TYPE_SCSI_IO) ? &sc->q_lock : &sc->gen_lock);
823     struct tws_request *r = NULL;
824 
825     mtx_lock(my_mutex);
826 
827     if (type == TWS_REQ_TYPE_SCSI_IO) {
828         r = tws_q_remove_head(sc, TWS_FREE_Q);
829     } else {
830         if ( sc->reqs[type].state == TWS_REQ_STATE_FREE ) {
831             r = &sc->reqs[type];
832         }
833     }
834 
835     if ( r ) {
836         bzero(&r->cmd_pkt->cmd, sizeof(struct tws_command_apache));
837         r->data = NULL;
838         r->length = 0;
839         r->type = type;
840         r->flags = TWS_DIR_UNKNOWN;
841         r->error_code = TWS_REQ_RET_INVALID;
842         r->cb = NULL;
843         r->ccb_ptr = NULL;
844 	callout_stop(&r->timeout);
845         r->next = r->prev = NULL;
846 
847         r->state = ((type == TWS_REQ_TYPE_SCSI_IO) ? TWS_REQ_STATE_TRAN : TWS_REQ_STATE_BUSY);
848     }
849 
850     mtx_unlock(my_mutex);
851 
852     return(r);
853 }
854 
855 void
856 tws_release_request(struct tws_request *req)
857 {
858 
859     struct tws_softc *sc = req->sc;
860 
861     TWS_TRACE_DEBUG(sc, "entry", sc, 0);
862     mtx_lock(&sc->q_lock);
863     tws_q_insert_tail(sc, req, TWS_FREE_Q);
864     mtx_unlock(&sc->q_lock);
865 }
866 
867 static device_method_t tws_methods[] = {
868     /* Device interface */
869     DEVMETHOD(device_probe,     tws_probe),
870     DEVMETHOD(device_attach,    tws_attach),
871     DEVMETHOD(device_detach,    tws_detach),
872     DEVMETHOD(device_shutdown,  tws_shutdown),
873     DEVMETHOD(device_suspend,   tws_suspend),
874     DEVMETHOD(device_resume,    tws_resume),
875 
876     DEVMETHOD_END
877 };
878 
879 static driver_t tws_driver = {
880         "tws",
881         tws_methods,
882         sizeof(struct tws_softc)
883 };
884 
885 static devclass_t tws_devclass;
886 
887 /* DEFINE_CLASS_0(tws, tws_driver, tws_methods, sizeof(struct tws_softc)); */
888 DRIVER_MODULE(tws, pci, tws_driver, tws_devclass, 0, 0);
889 MODULE_DEPEND(tws, cam, 1, 1, 1);
890 MODULE_DEPEND(tws, pci, 1, 1, 1);
891 
892 TUNABLE_INT("hw.tws.queue_depth", &tws_queue_depth);
893 TUNABLE_INT("hw.tws.enable_msi", &tws_enable_msi);
894