/* * CDDL HEADER START * * The contents of this file are subject to the terms of the * Common Development and Distribution License (the "License"). * You may not use this file except in compliance with the License. * * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE * or http://www.opensolaris.org/os/licensing. * See the License for the specific language governing permissions * and limitations under the License. * * When distributing Covered Code, include this CDDL HEADER in each * file and include the License file at usr/src/OPENSOLARIS.LICENSE. * If applicable, add the following below this CDDL HEADER, with the * fields enclosed by brackets "[]" replaced with your own identifying * information: Portions Copyright [yyyy] [name of copyright owner] * * CDDL HEADER END */ /* * Copyright 2008 Sun Microsystems, Inc. All rights reserved. * Use is subject to license terms. */ /* * Universal Host Controller Driver (UHCI) * * The UHCI driver is a driver which interfaces to the Universal * Serial Bus Driver (USBA) and the Host Controller (HC). The interface to * the Host Controller is defined by the UHCI. * This file contains misc functions. */ #include #include #include #include /* Globals */ extern uint_t uhci_td_pool_size; /* Num TDs */ extern uint_t uhci_qh_pool_size; /* Num QHs */ extern ushort_t uhci_tree_bottom_nodes[]; extern void *uhci_statep; /* function prototypes */ static void uhci_build_interrupt_lattice(uhci_state_t *uhcip); static int uhci_init_frame_lst_table(dev_info_t *dip, uhci_state_t *uhcip); static uint_t uhci_lattice_height(uint_t bandwidth); static uint_t uhci_lattice_parent(uint_t node); static uint_t uhci_leftmost_leaf(uint_t node, uint_t height); static uint_t uhci_compute_total_bandwidth(usb_ep_descr_t *endpoint, usb_port_status_t port_status); static int uhci_bandwidth_adjust(uhci_state_t *uhcip, usb_ep_descr_t *endpoint, usb_port_status_t port_status); static uhci_td_t *uhci_allocate_td_from_pool(uhci_state_t *uhcip); static void uhci_fill_in_td(uhci_state_t *uhcip, uhci_td_t *td, uhci_td_t *current_dummy, uint32_t buffer_offset, size_t length, uhci_pipe_private_t *pp, uchar_t PID, usb_req_attrs_t attrs, uhci_trans_wrapper_t *tw); static uint32_t uhci_get_tw_paddr_by_offs(uhci_state_t *uhcip, uint32_t buffer_offset, size_t length, uhci_trans_wrapper_t *tw); static uhci_trans_wrapper_t *uhci_create_transfer_wrapper( uhci_state_t *uhcip, uhci_pipe_private_t *pp, size_t length, usb_flags_t usb_flags); static uhci_trans_wrapper_t *uhci_create_isoc_transfer_wrapper( uhci_state_t *uhcip, uhci_pipe_private_t *pp, usb_isoc_req_t *req, size_t length, usb_flags_t usb_flags); static int uhci_create_setup_pkt(uhci_state_t *uhcip, uhci_pipe_private_t *pp, uhci_trans_wrapper_t *tw); static void uhci_insert_ctrl_qh(uhci_state_t *uhcip, uhci_pipe_private_t *pp); static void uhci_remove_ctrl_qh(uhci_state_t *uhcip, uhci_pipe_private_t *pp); static void uhci_insert_intr_qh(uhci_state_t *uhcip, uhci_pipe_private_t *pp); static void uhci_remove_intr_qh(uhci_state_t *uhcip, uhci_pipe_private_t *pp); static void uhci_remove_bulk_qh(uhci_state_t *uhcip, uhci_pipe_private_t *pp); static void uhci_insert_bulk_qh(uhci_state_t *uhcip, uhci_pipe_private_t *pp); static void uhci_handle_bulk_td_errors(uhci_state_t *uhcip, uhci_td_t *td); static int uhci_alloc_memory_for_tds(uhci_state_t *uhcip, uint_t num_tds, uhci_bulk_isoc_xfer_t *info); static int uhci_alloc_bulk_isoc_tds(uhci_state_t *uhcip, uint_t num_tds, uhci_bulk_isoc_xfer_t *info); static void uhci_get_isoc_td_by_index(uhci_state_t *uhcip, uhci_bulk_isoc_xfer_t *info, uint_t index, uhci_td_t **tdpp, uhci_bulk_isoc_td_pool_t **td_pool_pp); static void uhci_get_bulk_td_by_paddr(uhci_state_t *uhcip, uhci_bulk_isoc_xfer_t *info, uint32_t paddr, uhci_bulk_isoc_td_pool_t **td_pool_pp); static int uhci_handle_isoc_receive(uhci_state_t *uhcip, uhci_pipe_private_t *pp, uhci_trans_wrapper_t *tw); static void uhci_delete_isoc_td(uhci_state_t *uhcip, uhci_td_t *td); #ifdef DEBUG static void uhci_print_td(uhci_state_t *uhcip, uhci_td_t *td); static void uhci_print_qh(uhci_state_t *uhcip, queue_head_t *qh); #endif /* * uhci_build_interrupt_lattice: * * Construct the interrupt lattice tree using static Queue Head pointers. * This interrupt lattice tree will have total of 63 queue heads and the * Host Controller (HC) processes queue heads every frame. */ static void uhci_build_interrupt_lattice(uhci_state_t *uhcip) { int half_list = NUM_INTR_QH_LISTS / 2; uint16_t i, j, k; uhci_td_t *sof_td, *isoc_td; uintptr_t addr; queue_head_t *list_array = uhcip->uhci_qh_pool_addr; queue_head_t *tmp_qh; frame_lst_table_t *frame_lst_tablep = uhcip->uhci_frame_lst_tablep; USB_DPRINTF_L4(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_build_interrupt_lattice:"); /* * Reserve the first 63 queue head structures in the pool as static * queue heads & these are required for constructing interrupt * lattice tree. */ for (i = 0; i < NUM_INTR_QH_LISTS; i++) { SetQH32(uhcip, list_array[i].link_ptr, HC_END_OF_LIST); SetQH32(uhcip, list_array[i].element_ptr, HC_END_OF_LIST); list_array[i].qh_flag = QUEUE_HEAD_FLAG_STATIC; list_array[i].node = i; } /* Build the interrupt lattice tree */ for (i = 0; i < half_list - 1; i++) { /* * The next pointer in the host controller queue head * descriptor must contain an iommu address. Calculate * the offset into the cpu address and add this to the * starting iommu address. */ addr = QH_PADDR(&list_array[i]) | HC_QUEUE_HEAD; SetQH32(uhcip, list_array[2*i + 1].link_ptr, addr); SetQH32(uhcip, list_array[2*i + 2].link_ptr, addr); } /* * Initialize the interrupt list in the Frame list Table * so that it points to the bottom of the tree. */ for (i = 0, j = 0; i < pow_2(TREE_HEIGHT); i++) { addr = QH_PADDR(&list_array[half_list + i - 1]); for (k = 0; k < pow_2(VIRTUAL_TREE_HEIGHT); k++) { SetFL32(uhcip, frame_lst_tablep[uhci_tree_bottom_nodes[j++]], addr | HC_QUEUE_HEAD); } } /* * Create a controller and bulk Queue heads */ uhcip->uhci_ctrl_xfers_q_head = uhci_alloc_queue_head(uhcip); tmp_qh = uhcip->uhci_ctrl_xfers_q_tail = uhcip->uhci_ctrl_xfers_q_head; SetQH32(uhcip, list_array[0].link_ptr, (QH_PADDR(tmp_qh) | HC_QUEUE_HEAD)); uhcip->uhci_bulk_xfers_q_head = uhci_alloc_queue_head(uhcip); uhcip->uhci_bulk_xfers_q_tail = uhcip->uhci_bulk_xfers_q_head; SetQH32(uhcip, tmp_qh->link_ptr, (QH_PADDR(uhcip->uhci_bulk_xfers_q_head)|HC_QUEUE_HEAD)); SetQH32(uhcip, uhcip->uhci_bulk_xfers_q_head->link_ptr, HC_END_OF_LIST); /* * Add a dummy TD to the static queue head 0. THis is used * to generate an at the end of frame. */ sof_td = uhci_allocate_td_from_pool(uhcip); SetQH32(uhcip, list_array[0].element_ptr, TD_PADDR(sof_td) | HC_TD_HEAD); SetTD32(uhcip, sof_td->link_ptr, HC_END_OF_LIST); uhcip->uhci_sof_td = sof_td; /* * Add a dummy td that is used to generate an interrupt for * every 1024 frames. */ isoc_td = uhci_allocate_td_from_pool(uhcip); SetTD32(uhcip, isoc_td->link_ptr, HC_END_OF_LIST); uhcip->uhci_isoc_td = isoc_td; uhcip->uhci_isoc_qh = uhci_alloc_queue_head(uhcip); SetQH32(uhcip, uhcip->uhci_isoc_qh->link_ptr, GetFL32(uhcip, uhcip->uhci_frame_lst_tablep[MAX_FRAME_NUM])); SetQH32(uhcip, uhcip->uhci_isoc_qh->element_ptr, TD_PADDR(isoc_td)); SetFL32(uhcip, uhcip->uhci_frame_lst_tablep[MAX_FRAME_NUM], QH_PADDR(uhcip->uhci_isoc_qh) | HC_QUEUE_HEAD); } /* * uhci_allocate_pools: * Allocate the system memory for the Queue Heads Descriptor and * for the Transfer Descriptor (TD) pools. Both QH and TD structures * must be aligned to a 16 byte boundary. */ int uhci_allocate_pools(uhci_state_t *uhcip) { dev_info_t *dip = uhcip->uhci_dip; size_t real_length; int i, result; uint_t ccount; ddi_device_acc_attr_t dev_attr; USB_DPRINTF_L4(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_allocate_pools:"); /* The host controller will be little endian */ dev_attr.devacc_attr_version = DDI_DEVICE_ATTR_V0; dev_attr.devacc_attr_endian_flags = DDI_STRUCTURE_LE_ACC; dev_attr.devacc_attr_dataorder = DDI_STRICTORDER_ACC; /* Allocate the TD pool DMA handle */ if (ddi_dma_alloc_handle(dip, &uhcip->uhci_dma_attr, DDI_DMA_SLEEP, 0, &uhcip->uhci_td_pool_dma_handle) != DDI_SUCCESS) { return (USB_FAILURE); } /* Allocate the memory for the TD pool */ if (ddi_dma_mem_alloc(uhcip->uhci_td_pool_dma_handle, uhci_td_pool_size * sizeof (uhci_td_t), &dev_attr, DDI_DMA_CONSISTENT, DDI_DMA_SLEEP, 0, (caddr_t *)&uhcip->uhci_td_pool_addr, &real_length, &uhcip->uhci_td_pool_mem_handle)) { return (USB_FAILURE); } /* Map the TD pool into the I/O address space */ result = ddi_dma_addr_bind_handle(uhcip->uhci_td_pool_dma_handle, NULL, (caddr_t)uhcip->uhci_td_pool_addr, real_length, DDI_DMA_RDWR | DDI_DMA_CONSISTENT, DDI_DMA_SLEEP, NULL, &uhcip->uhci_td_pool_cookie, &ccount); bzero((void *)uhcip->uhci_td_pool_addr, uhci_td_pool_size * sizeof (uhci_td_t)); /* Process the result */ if (result == DDI_DMA_MAPPED) { /* The cookie count should be 1 */ if (ccount != 1) { USB_DPRINTF_L2(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_allocate_pools: More than 1 cookie"); return (USB_FAILURE); } } else { USB_DPRINTF_L4(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_allocate_pools: Result = %d", result); uhci_decode_ddi_dma_addr_bind_handle_result(uhcip, result); return (USB_FAILURE); } uhcip->uhci_dma_addr_bind_flag |= UHCI_TD_POOL_BOUND; /* Initialize the TD pool */ for (i = 0; i < uhci_td_pool_size; i++) { uhcip->uhci_td_pool_addr[i].flag = TD_FLAG_FREE; } /* Allocate the TD pool DMA handle */ if (ddi_dma_alloc_handle(dip, &uhcip->uhci_dma_attr, DDI_DMA_SLEEP, 0, &uhcip->uhci_qh_pool_dma_handle) != DDI_SUCCESS) { return (USB_FAILURE); } /* Allocate the memory for the QH pool */ if (ddi_dma_mem_alloc(uhcip->uhci_qh_pool_dma_handle, uhci_qh_pool_size * sizeof (queue_head_t), &dev_attr, DDI_DMA_CONSISTENT, DDI_DMA_SLEEP, 0, (caddr_t *)&uhcip->uhci_qh_pool_addr, &real_length, &uhcip->uhci_qh_pool_mem_handle) != DDI_SUCCESS) { return (USB_FAILURE); } result = ddi_dma_addr_bind_handle(uhcip->uhci_qh_pool_dma_handle, NULL, (caddr_t)uhcip->uhci_qh_pool_addr, real_length, DDI_DMA_RDWR | DDI_DMA_CONSISTENT, DDI_DMA_SLEEP, NULL, &uhcip->uhci_qh_pool_cookie, &ccount); /* Process the result */ if (result == DDI_DMA_MAPPED) { /* The cookie count should be 1 */ if (ccount != 1) { USB_DPRINTF_L2(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_allocate_pools: More than 1 cookie"); return (USB_FAILURE); } } else { uhci_decode_ddi_dma_addr_bind_handle_result(uhcip, result); return (USB_FAILURE); } uhcip->uhci_dma_addr_bind_flag |= UHCI_QH_POOL_BOUND; bzero((void *)uhcip->uhci_qh_pool_addr, uhci_qh_pool_size * sizeof (queue_head_t)); /* Initialize the QH pool */ for (i = 0; i < uhci_qh_pool_size; i ++) { uhcip->uhci_qh_pool_addr[i].qh_flag = QUEUE_HEAD_FLAG_FREE; } USB_DPRINTF_L4(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_allocate_pools: Completed"); return (USB_SUCCESS); } /* * uhci_free_pools: * Cleanup on attach failure or detach */ void uhci_free_pools(uhci_state_t *uhcip) { int i, flag, rval; uhci_td_t *td; uhci_trans_wrapper_t *tw; USB_DPRINTF_L4(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_free_pools:"); if (uhcip->uhci_td_pool_addr && uhcip->uhci_td_pool_mem_handle) { for (i = 0; i < uhci_td_pool_size; i ++) { td = &uhcip->uhci_td_pool_addr[i]; flag = uhcip->uhci_td_pool_addr[i].flag; if ((flag != TD_FLAG_FREE) && (flag != TD_FLAG_DUMMY) && (td->tw != NULL)) { tw = td->tw; uhci_free_tw(uhcip, tw); } } if (uhcip->uhci_dma_addr_bind_flag & UHCI_TD_POOL_BOUND) { rval = ddi_dma_unbind_handle( uhcip->uhci_td_pool_dma_handle); ASSERT(rval == DDI_SUCCESS); } ddi_dma_mem_free(&uhcip->uhci_td_pool_mem_handle); } /* Free the TD pool */ if (uhcip->uhci_td_pool_dma_handle) { ddi_dma_free_handle(&uhcip->uhci_td_pool_dma_handle); } if (uhcip->uhci_qh_pool_addr && uhcip->uhci_qh_pool_mem_handle) { if (uhcip->uhci_dma_addr_bind_flag & UHCI_QH_POOL_BOUND) { rval = ddi_dma_unbind_handle( uhcip->uhci_qh_pool_dma_handle); ASSERT(rval == DDI_SUCCESS); } ddi_dma_mem_free(&uhcip->uhci_qh_pool_mem_handle); } /* Free the QH pool */ if (uhcip->uhci_qh_pool_dma_handle) { ddi_dma_free_handle(&uhcip->uhci_qh_pool_dma_handle); } /* Free the Frame list Table area */ if (uhcip->uhci_frame_lst_tablep && uhcip->uhci_flt_mem_handle) { if (uhcip->uhci_dma_addr_bind_flag & UHCI_FLA_POOL_BOUND) { rval = ddi_dma_unbind_handle( uhcip->uhci_flt_dma_handle); ASSERT(rval == DDI_SUCCESS); } ddi_dma_mem_free(&uhcip->uhci_flt_mem_handle); } if (uhcip->uhci_flt_dma_handle) { ddi_dma_free_handle(&uhcip->uhci_flt_dma_handle); } } /* * uhci_decode_ddi_dma_addr_bind_handle_result: * Process the return values of ddi_dma_addr_bind_handle() */ void uhci_decode_ddi_dma_addr_bind_handle_result(uhci_state_t *uhcip, int result) { char *msg; USB_DPRINTF_L4(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "uhci_decode_ddi_dma_addr_bind_handle_result:"); switch (result) { case DDI_DMA_PARTIAL_MAP: msg = "Partial transfers not allowed"; break; case DDI_DMA_INUSE: msg = "Handle is in use"; break; case DDI_DMA_NORESOURCES: msg = "No resources"; break; case DDI_DMA_NOMAPPING: msg = "No mapping"; break; case DDI_DMA_TOOBIG: msg = "Object is too big"; break; default: msg = "Unknown dma error"; } USB_DPRINTF_L4(PRINT_MASK_ALL, uhcip->uhci_log_hdl, "%s", msg); } /* * uhci_init_ctlr: * Initialize the Host Controller (HC). */ int uhci_init_ctlr(uhci_state_t *uhcip) { dev_info_t *dip = uhcip->uhci_dip; uint_t cmd_reg; uint_t frame_base_addr; mutex_enter(&uhcip->uhci_int_mutex); USB_DPRINTF_L4(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_init_ctlr:"); /* * When USB legacy mode is enabled, the BIOS manages the USB keyboard * attached to the UHCI controller. It has been observed that some * times the BIOS does not clear the interrupts in the legacy mode * register in the PCI configuration space. So, disable the SMI intrs * and route the intrs to PIRQD here. */ pci_config_put16(uhcip->uhci_config_handle, LEGACYMODE_REG_OFFSET, LEGACYMODE_REG_INIT_VALUE); /* * Disable all the interrupts. */ Set_OpReg16(USBINTR, DISABLE_ALL_INTRS); cmd_reg = Get_OpReg16(USBCMD); cmd_reg &= (~USBCMD_REG_HC_RUN); /* Stop the controller */ Set_OpReg16(USBCMD, cmd_reg); /* Reset the host controller */ Set_OpReg16(USBCMD, USBCMD_REG_GBL_RESET); /* Wait 10ms for reset to complete */ mutex_exit(&uhcip->uhci_int_mutex); delay(drv_usectohz(UHCI_RESET_DELAY)); mutex_enter(&uhcip->uhci_int_mutex); Set_OpReg16(USBCMD, 0); /* Set the frame number to zero */ Set_OpReg16(FRNUM, 0); if (uhcip->uhci_hc_soft_state == UHCI_CTLR_INIT_STATE) { /* Initialize the Frame list base address area */ if (uhci_init_frame_lst_table(dip, uhcip) != USB_SUCCESS) { mutex_exit(&uhcip->uhci_int_mutex); return (USB_FAILURE); } } /* Save the contents of the Frame Interval Registers */ uhcip->uhci_frame_interval = Get_OpReg8(SOFMOD); frame_base_addr = uhcip->uhci_flt_cookie.dmac_address; /* Set the Frame list base address */ Set_OpReg32(FRBASEADD, frame_base_addr); /* * Begin sending SOFs * Set the Host Controller Functional State to Operational */ cmd_reg = Get_OpReg16(USBCMD); cmd_reg |= (USBCMD_REG_HC_RUN | USBCMD_REG_MAXPKT_64 | USBCMD_REG_CONFIG_FLAG); Set_OpReg16(USBCMD, cmd_reg); /* * Verify the Command and interrupt enable registers, * a sanity check whether actually initialized or not */ cmd_reg = Get_OpReg16(USBCMD); if (!(cmd_reg & (USBCMD_REG_HC_RUN | USBCMD_REG_MAXPKT_64 | USBCMD_REG_CONFIG_FLAG))) { USB_DPRINTF_L2(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_init_ctlr: Controller initialization failed"); mutex_exit(&uhcip->uhci_int_mutex); return (USB_FAILURE); } /* * Set the ioc bit of the isoc intr td. This enables * the generation of an interrupt for every 1024 frames. */ SetTD_ioc(uhcip, uhcip->uhci_isoc_td, 1); /* Set host controller soft state to operational */ uhcip->uhci_hc_soft_state = UHCI_CTLR_OPERATIONAL_STATE; mutex_exit(&uhcip->uhci_int_mutex); USB_DPRINTF_L4(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_init_ctlr: Completed"); return (USB_SUCCESS); } /* * uhci_uninit_ctlr: * uninitialize the Host Controller (HC). */ void uhci_uninit_ctlr(uhci_state_t *uhcip) { if (uhcip->uhci_regs_handle) { /* Disable all the interrupts. */ Set_OpReg16(USBINTR, DISABLE_ALL_INTRS); /* Complete the current transaction and then halt. */ Set_OpReg16(USBCMD, 0); /* Wait for sometime */ mutex_exit(&uhcip->uhci_int_mutex); delay(drv_usectohz(UHCI_TIMEWAIT)); mutex_enter(&uhcip->uhci_int_mutex); } } /* * uhci_map_regs: * The Host Controller (HC) contains a set of on-chip operational * registers and which should be mapped into a non-cacheable * portion of the system addressable space. */ int uhci_map_regs(uhci_state_t *uhcip) { dev_info_t *dip = uhcip->uhci_dip; int index; uint32_t regs_prop_len; int32_t *regs_list; uint16_t command_reg; ddi_device_acc_attr_t attr; USB_DPRINTF_L4(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_map_regs:"); /* The host controller will be little endian */ attr.devacc_attr_version = DDI_DEVICE_ATTR_V0; attr.devacc_attr_endian_flags = DDI_STRUCTURE_LE_ACC; attr.devacc_attr_dataorder = DDI_STRICTORDER_ACC; if (ddi_prop_lookup_int_array(DDI_DEV_T_ANY, uhcip->uhci_dip, DDI_PROP_DONTPASS, "reg", ®s_list, ®s_prop_len) != DDI_PROP_SUCCESS) { return (USB_FAILURE); } for (index = 0; index * 5 < regs_prop_len; index++) { if (regs_list[index * 5] & UHCI_PROP_MASK) { break; } } /* * Deallocate the memory allocated by the ddi_prop_lookup_int_array */ ddi_prop_free(regs_list); if (index * 5 >= regs_prop_len) { return (USB_FAILURE); } /* Map in operational registers */ if (ddi_regs_map_setup(dip, index, (caddr_t *)&uhcip->uhci_regsp, 0, sizeof (hc_regs_t), &attr, &uhcip->uhci_regs_handle) != DDI_SUCCESS) { USB_DPRINTF_L2(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "ddi_regs_map_setup: failed"); return (USB_FAILURE); } if (pci_config_setup(dip, &uhcip->uhci_config_handle) != DDI_SUCCESS) { USB_DPRINTF_L2(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_map_regs: Config error"); return (USB_FAILURE); } /* Make sure Memory Access Enable and Master Enable are set */ command_reg = pci_config_get16(uhcip->uhci_config_handle, PCI_CONF_COMM); if (!(command_reg & (PCI_COMM_MAE | PCI_COMM_ME))) { USB_DPRINTF_L3(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_map_regs: No MAE/ME"); } command_reg |= PCI_COMM_MAE | PCI_COMM_ME; pci_config_put16(uhcip->uhci_config_handle, PCI_CONF_COMM, command_reg); /* * Check whether I/O base address is configured and enabled. */ if (!(command_reg & PCI_COMM_IO)) { USB_DPRINTF_L2(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "I/O Base address access disabled"); return (USB_FAILURE); } /* * Get the IO base address of the controller */ uhcip->uhci_iobase = (pci_config_get16(uhcip->uhci_config_handle, PCI_CONF_IOBASE) & PCI_CONF_IOBASE_MASK); USB_DPRINTF_L4(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_map_regs: Completed"); return (USB_SUCCESS); } void uhci_unmap_regs(uhci_state_t *uhcip) { /* Unmap the UHCI registers */ if (uhcip->uhci_regs_handle) { /* Reset the host controller */ Set_OpReg16(USBCMD, USBCMD_REG_GBL_RESET); ddi_regs_map_free(&uhcip->uhci_regs_handle); } if (uhcip->uhci_config_handle) { pci_config_teardown(&uhcip->uhci_config_handle); } } /* * uhci_set_dma_attributes: * Set the limits in the DMA attributes structure. Most of the values used * in the DMA limit structres are the default values as specified by the * Writing PCI device drivers document. */ void uhci_set_dma_attributes(uhci_state_t *uhcip) { USB_DPRINTF_L4(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_set_dma_attributes:"); /* Initialize the DMA attributes */ uhcip->uhci_dma_attr.dma_attr_version = DMA_ATTR_V0; uhcip->uhci_dma_attr.dma_attr_addr_lo = 0x00000000ull; uhcip->uhci_dma_attr.dma_attr_addr_hi = 0xfffffff0ull; /* 32 bit addressing */ uhcip->uhci_dma_attr.dma_attr_count_max = 0xffffffull; /* * Setting the dam_att_align to 512, some times fails the * binding handle. I dont know why ? But setting to 16 will * be right for our case (16 byte alignment required per * UHCI spec for TD descriptors). */ /* 16 byte alignment */ uhcip->uhci_dma_attr.dma_attr_align = 0x10; /* * Since PCI specification is byte alignment, the * burstsize field should be set to 1 for PCI devices. */ uhcip->uhci_dma_attr.dma_attr_burstsizes = 0x1; uhcip->uhci_dma_attr.dma_attr_minxfer = 0x1; uhcip->uhci_dma_attr.dma_attr_maxxfer = 0xffffffull; uhcip->uhci_dma_attr.dma_attr_seg = 0xffffffffull; uhcip->uhci_dma_attr.dma_attr_sgllen = 1; uhcip->uhci_dma_attr.dma_attr_granular = 1; uhcip->uhci_dma_attr.dma_attr_flags = 0; } uint_t pow_2(uint_t x) { return ((x == 0) ? 1 : (1 << x)); } uint_t log_2(uint_t x) { int ret_val = 0; while (x != 1) { ret_val++; x = x >> 1; } return (ret_val); } /* * uhci_obtain_state: */ uhci_state_t * uhci_obtain_state(dev_info_t *dip) { int instance = ddi_get_instance(dip); uhci_state_t *state = ddi_get_soft_state(uhci_statep, instance); ASSERT(state != NULL); return (state); } /* * uhci_alloc_hcdi_ops: * The HCDI interfaces or entry points are the software interfaces used by * the Universal Serial Bus Driver (USBA) to access the services of the * Host Controller Driver (HCD). During HCD initialization, inform USBA * about all available HCDI interfaces or entry points. */ usba_hcdi_ops_t * uhci_alloc_hcdi_ops(uhci_state_t *uhcip) { usba_hcdi_ops_t *hcdi_ops; USB_DPRINTF_L4(PRINT_MASK_HCDI, uhcip->uhci_log_hdl, "uhci_alloc_hcdi_ops:"); hcdi_ops = usba_alloc_hcdi_ops(); hcdi_ops->usba_hcdi_ops_version = HCDI_OPS_VERSION_1; hcdi_ops->usba_hcdi_pipe_open = uhci_hcdi_pipe_open; hcdi_ops->usba_hcdi_pipe_close = uhci_hcdi_pipe_close; hcdi_ops->usba_hcdi_pipe_reset = uhci_hcdi_pipe_reset; hcdi_ops->usba_hcdi_pipe_ctrl_xfer = uhci_hcdi_pipe_ctrl_xfer; hcdi_ops->usba_hcdi_pipe_bulk_xfer = uhci_hcdi_pipe_bulk_xfer; hcdi_ops->usba_hcdi_pipe_intr_xfer = uhci_hcdi_pipe_intr_xfer; hcdi_ops->usba_hcdi_pipe_isoc_xfer = uhci_hcdi_pipe_isoc_xfer; hcdi_ops->usba_hcdi_bulk_transfer_size = uhci_hcdi_bulk_transfer_size; hcdi_ops->usba_hcdi_pipe_stop_intr_polling = uhci_hcdi_pipe_stop_intr_polling; hcdi_ops->usba_hcdi_pipe_stop_isoc_polling = uhci_hcdi_pipe_stop_isoc_polling; hcdi_ops->usba_hcdi_get_current_frame_number = uhci_hcdi_get_current_frame_number; hcdi_ops->usba_hcdi_get_max_isoc_pkts = uhci_hcdi_get_max_isoc_pkts; hcdi_ops->usba_hcdi_console_input_init = uhci_hcdi_polled_input_init; hcdi_ops->usba_hcdi_console_input_enter = uhci_hcdi_polled_input_enter; hcdi_ops->usba_hcdi_console_read = uhci_hcdi_polled_read; hcdi_ops->usba_hcdi_console_input_exit = uhci_hcdi_polled_input_exit; hcdi_ops->usba_hcdi_console_input_fini = uhci_hcdi_polled_input_fini; hcdi_ops->usba_hcdi_console_output_init = uhci_hcdi_polled_output_init; hcdi_ops->usba_hcdi_console_output_enter = uhci_hcdi_polled_output_enter; hcdi_ops->usba_hcdi_console_write = uhci_hcdi_polled_write; hcdi_ops->usba_hcdi_console_output_exit = uhci_hcdi_polled_output_exit; hcdi_ops->usba_hcdi_console_output_fini = uhci_hcdi_polled_output_fini; return (hcdi_ops); } /* * uhci_init_frame_lst_table : * Allocate the system memory and initialize Host Controller * Frame list table area The starting of the Frame list Table * area must be 4096 byte aligned. */ static int uhci_init_frame_lst_table(dev_info_t *dip, uhci_state_t *uhcip) { int result; uint_t ccount; size_t real_length; ddi_device_acc_attr_t dev_attr; ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); USB_DPRINTF_L4(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_init_frame_lst_table:"); /* The host controller will be little endian */ dev_attr.devacc_attr_version = DDI_DEVICE_ATTR_V0; dev_attr.devacc_attr_endian_flags = DDI_STRUCTURE_LE_ACC; dev_attr.devacc_attr_dataorder = DDI_STRICTORDER_ACC; /* 4K alignment required */ uhcip->uhci_dma_attr.dma_attr_align = 0x1000; /* Create space for the HCCA block */ if (ddi_dma_alloc_handle(dip, &uhcip->uhci_dma_attr, DDI_DMA_SLEEP, 0, &uhcip->uhci_flt_dma_handle) != DDI_SUCCESS) { return (USB_FAILURE); } /* Reset to default 16 bytes */ uhcip->uhci_dma_attr.dma_attr_align = 0x10; if (ddi_dma_mem_alloc(uhcip->uhci_flt_dma_handle, SIZE_OF_FRAME_LST_TABLE, &dev_attr, DDI_DMA_CONSISTENT, DDI_DMA_SLEEP, 0, (caddr_t *)&uhcip->uhci_frame_lst_tablep, &real_length, &uhcip->uhci_flt_mem_handle)) { return (USB_FAILURE); } /* Map the whole Frame list base area into the I/O address space */ result = ddi_dma_addr_bind_handle(uhcip->uhci_flt_dma_handle, NULL, (caddr_t)uhcip->uhci_frame_lst_tablep, real_length, DDI_DMA_RDWR | DDI_DMA_CONSISTENT, DDI_DMA_SLEEP, NULL, &uhcip->uhci_flt_cookie, &ccount); if (result == DDI_DMA_MAPPED) { /* The cookie count should be 1 */ if (ccount != 1) { USB_DPRINTF_L2(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_init_frame_list_table: More than 1 cookie"); return (USB_FAILURE); } } else { uhci_decode_ddi_dma_addr_bind_handle_result(uhcip, result); return (USB_FAILURE); } uhcip->uhci_dma_addr_bind_flag |= UHCI_FLA_POOL_BOUND; bzero((void *)uhcip->uhci_frame_lst_tablep, real_length); /* Initialize the interrupt lists */ uhci_build_interrupt_lattice(uhcip); return (USB_SUCCESS); } /* * uhci_alloc_queue_head: * Allocate a queue head */ queue_head_t * uhci_alloc_queue_head(uhci_state_t *uhcip) { int index; uhci_td_t *dummy_td; queue_head_t *queue_head; USB_DPRINTF_L4(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "uhci_alloc_queue_head"); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); /* Allocate a dummy td first. */ if ((dummy_td = uhci_allocate_td_from_pool(uhcip)) == NULL) { USB_DPRINTF_L2(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "uhci_alloc_queue_head: allocate td from pool failed"); return (NULL); } /* * The first 63 queue heads in the Queue Head (QH) * buffer pool are reserved for building interrupt lattice * tree. Search for a blank Queue head in the QH buffer pool. */ for (index = NUM_STATIC_NODES; index < uhci_qh_pool_size; index++) { if (uhcip->uhci_qh_pool_addr[index].qh_flag == QUEUE_HEAD_FLAG_FREE) { break; } } USB_DPRINTF_L3(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "uhci_alloc_queue_head: Allocated %d", index); if (index == uhci_qh_pool_size) { USB_DPRINTF_L2(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "uhci_alloc_queue_head: All QH exhausted"); /* Free the dummy td allocated for this qh. */ dummy_td->flag = TD_FLAG_FREE; return (NULL); } queue_head = &uhcip->uhci_qh_pool_addr[index]; USB_DPRINTF_L4(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "uhci_alloc_queue_head: Allocated address 0x%p", (void *)queue_head); bzero((void *)queue_head, sizeof (queue_head_t)); SetQH32(uhcip, queue_head->link_ptr, HC_END_OF_LIST); SetQH32(uhcip, queue_head->element_ptr, HC_END_OF_LIST); queue_head->prev_qh = NULL; queue_head->qh_flag = QUEUE_HEAD_FLAG_BUSY; bzero((char *)dummy_td, sizeof (uhci_td_t)); queue_head->td_tailp = dummy_td; SetQH32(uhcip, queue_head->element_ptr, TD_PADDR(dummy_td)); return (queue_head); } /* * uhci_allocate_bandwidth: * Figure out whether or not this interval may be supported. Return * the index into the lattice if it can be supported. Return * allocation failure if it can not be supported. */ int uhci_allocate_bandwidth( uhci_state_t *uhcip, usba_pipe_handle_data_t *pipe_handle, uint_t *node) { int bandwidth; /* Requested bandwidth */ uint_t min, min_index; uint_t i; uint_t height; /* Bandwidth's height in the tree */ uint_t leftmost; uint_t length; uint32_t paddr; queue_head_t *tmp_qh; usb_ep_descr_t *endpoint = &pipe_handle->p_ep; ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); /* * Calculate the length in bytes of a transaction on this * periodic endpoint. */ mutex_enter(&pipe_handle->p_usba_device->usb_mutex); length = uhci_compute_total_bandwidth(endpoint, pipe_handle->p_usba_device->usb_port_status); mutex_exit(&pipe_handle->p_usba_device->usb_mutex); /* * If the length in bytes plus the allocated bandwidth exceeds * the maximum, return bandwidth allocation failure. */ if ((length + uhcip->uhci_bandwidth_intr_min + uhcip->uhci_bandwidth_isoch_sum) > (MAX_PERIODIC_BANDWIDTH)) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_allocate_bandwidth: " "Reached maximum bandwidth value and cannot allocate " "bandwidth for a given Interrupt/Isoch endpoint"); return (USB_NO_BANDWIDTH); } /* * ISOC xfers are not supported at this point type */ if (UHCI_XFER_TYPE(endpoint) == USB_EP_ATTR_ISOCH) { uhcip->uhci_bandwidth_isoch_sum += length; return (USB_SUCCESS); } /* * This is an interrupt endpoint. * Adjust bandwidth to be a power of 2 */ mutex_enter(&pipe_handle->p_usba_device->usb_mutex); bandwidth = uhci_bandwidth_adjust(uhcip, endpoint, pipe_handle->p_usba_device->usb_port_status); mutex_exit(&pipe_handle->p_usba_device->usb_mutex); /* * If this bandwidth can't be supported, * return allocation failure. */ if (bandwidth == USB_FAILURE) { return (USB_FAILURE); } USB_DPRINTF_L3(PRINT_MASK_BW, uhcip->uhci_log_hdl, "The new bandwidth is %d", bandwidth); /* Find the leaf with the smallest allocated bandwidth */ min_index = 0; min = uhcip->uhci_bandwidth[0]; for (i = 1; i < NUM_FRAME_LST_ENTRIES; i++) { if (uhcip->uhci_bandwidth[i] < min) { min_index = i; min = uhcip->uhci_bandwidth[i]; } } USB_DPRINTF_L3(PRINT_MASK_BW, uhcip->uhci_log_hdl, "The leaf with minimal bandwidth %d, " "The smallest bandwidth %d", min_index, min); /* * Find the index into the lattice given the * leaf with the smallest allocated bandwidth. */ height = uhci_lattice_height(bandwidth); USB_DPRINTF_L3(PRINT_MASK_BW, uhcip->uhci_log_hdl, "The height is %d", height); *node = uhci_tree_bottom_nodes[min_index]; /* check if there are isocs TDs scheduled for this frame */ if (uhcip->uhci_isoc_q_tailp[*node]) { paddr = (uhcip->uhci_isoc_q_tailp[*node]->link_ptr & FRAME_LST_PTR_MASK); } else { paddr = (uhcip->uhci_frame_lst_tablep[*node] & FRAME_LST_PTR_MASK); } tmp_qh = QH_VADDR(paddr); *node = tmp_qh->node; for (i = 0; i < height; i++) { *node = uhci_lattice_parent(*node); } USB_DPRINTF_L3(PRINT_MASK_BW, uhcip->uhci_log_hdl, "The real node is %d", *node); /* * Find the leftmost leaf in the subtree specified by the node. */ leftmost = uhci_leftmost_leaf(*node, height); USB_DPRINTF_L3(PRINT_MASK_BW, uhcip->uhci_log_hdl, "Leftmost %d", leftmost); for (i = leftmost; i < leftmost + (NUM_FRAME_LST_ENTRIES/bandwidth); i ++) { if ((length + uhcip->uhci_bandwidth_isoch_sum + uhcip->uhci_bandwidth[i]) > MAX_PERIODIC_BANDWIDTH) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_allocate_bandwidth: " "Reached maximum bandwidth value and cannot " "allocate bandwidth for Interrupt endpoint"); return (USB_NO_BANDWIDTH); } } /* * All the leaves for this node must be updated with the bandwidth. */ for (i = leftmost; i < leftmost + (NUM_FRAME_LST_ENTRIES/bandwidth); i ++) { uhcip->uhci_bandwidth[i] += length; } /* Find the leaf with the smallest allocated bandwidth */ min_index = 0; min = uhcip->uhci_bandwidth[0]; for (i = 1; i < NUM_FRAME_LST_ENTRIES; i++) { if (uhcip->uhci_bandwidth[i] < min) { min_index = i; min = uhcip->uhci_bandwidth[i]; } } /* Save the minimum for later use */ uhcip->uhci_bandwidth_intr_min = min; return (USB_SUCCESS); } /* * uhci_deallocate_bandwidth: * Deallocate bandwidth for the given node in the lattice * and the length of transfer. */ void uhci_deallocate_bandwidth(uhci_state_t *uhcip, usba_pipe_handle_data_t *pipe_handle) { uint_t bandwidth; uint_t height; uint_t leftmost; uint_t i; uint_t min; usb_ep_descr_t *endpoint = &pipe_handle->p_ep; uint_t node, length; uhci_pipe_private_t *pp = (uhci_pipe_private_t *)pipe_handle->p_hcd_private; /* This routine is protected by the uhci_int_mutex */ ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); /* Obtain the length */ mutex_enter(&pipe_handle->p_usba_device->usb_mutex); length = uhci_compute_total_bandwidth(endpoint, pipe_handle->p_usba_device->usb_port_status); mutex_exit(&pipe_handle->p_usba_device->usb_mutex); /* * If this is an isochronous endpoint, just delete endpoint's * bandwidth from the total allocated isochronous bandwidth. */ if (UHCI_XFER_TYPE(endpoint) == USB_EP_ATTR_ISOCH) { uhcip->uhci_bandwidth_isoch_sum -= length; return; } /* Obtain the node */ node = pp->pp_node; /* Adjust bandwidth to be a power of 2 */ mutex_enter(&pipe_handle->p_usba_device->usb_mutex); bandwidth = uhci_bandwidth_adjust(uhcip, endpoint, pipe_handle->p_usba_device->usb_port_status); mutex_exit(&pipe_handle->p_usba_device->usb_mutex); /* Find the height in the tree */ height = uhci_lattice_height(bandwidth); /* * Find the leftmost leaf in the subtree specified by the node */ leftmost = uhci_leftmost_leaf(node, height); /* Delete the bandwith from the appropriate lists */ for (i = leftmost; i < leftmost + (NUM_FRAME_LST_ENTRIES/bandwidth); i ++) { uhcip->uhci_bandwidth[i] -= length; } min = uhcip->uhci_bandwidth[0]; /* Recompute the minimum */ for (i = 1; i < NUM_FRAME_LST_ENTRIES; i++) { if (uhcip->uhci_bandwidth[i] < min) { min = uhcip->uhci_bandwidth[i]; } } /* Save the minimum for later use */ uhcip->uhci_bandwidth_intr_min = min; } /* * uhci_compute_total_bandwidth: * * Given a periodic endpoint (interrupt or isochronous) determine the total * bandwidth for one transaction. The UHCI host controller traverses the * endpoint descriptor lists on a first-come-first-serve basis. When the HC * services an endpoint, only a single transaction attempt is made. The HC * moves to the next Endpoint Descriptor after the first transaction attempt * rather than finishing the entire Transfer Descriptor. Therefore, when a * Transfer Descriptor is inserted into the lattice, we will only count the * number of bytes for one transaction. * * The following are the formulas used for calculating bandwidth in terms * bytes and it is for the single USB full speed and low speed transaction * respectively. The protocol overheads will be different for each of type * of USB transfer and all these formulas & protocol overheads are derived * from the 5.9.3 section of USB Specification & with the help of Bandwidth * Analysis white paper which is posted on the USB developer forum. * * Full-Speed: * Protocol overhead + ((MaxPacketSize * 7)/6 ) + Host_Delay * * Low-Speed: * Protocol overhead + Hub LS overhead + * (Low-Speed clock * ((MaxPacketSize * 7)/6 )) + Host_Delay */ static uint_t uhci_compute_total_bandwidth(usb_ep_descr_t *endpoint, usb_port_status_t port_status) { uint_t bandwidth; ushort_t MaxPacketSize = endpoint->wMaxPacketSize; /* Add Host Controller specific delay to required bandwidth */ bandwidth = HOST_CONTROLLER_DELAY; /* Add bit-stuffing overhead */ MaxPacketSize = (ushort_t)((MaxPacketSize * 7) / 6); /* Low Speed interrupt transaction */ if (port_status == USBA_LOW_SPEED_DEV) { /* Low Speed interrupt transaction */ bandwidth += (LOW_SPEED_PROTO_OVERHEAD + HUB_LOW_SPEED_PROTO_OVERHEAD + (LOW_SPEED_CLOCK * MaxPacketSize)); } else { /* Full Speed transaction */ bandwidth += MaxPacketSize; if (UHCI_XFER_TYPE(endpoint) == USB_EP_ATTR_INTR) { /* Full Speed interrupt transaction */ bandwidth += FS_NON_ISOC_PROTO_OVERHEAD; } else { /* Isochronus and input transaction */ if (UHCI_XFER_DIR(endpoint) == USB_EP_DIR_IN) { bandwidth += FS_ISOC_INPUT_PROTO_OVERHEAD; } else { /* Isochronus and output transaction */ bandwidth += FS_ISOC_OUTPUT_PROTO_OVERHEAD; } } } return (bandwidth); } /* * uhci_bandwidth_adjust: */ static int uhci_bandwidth_adjust( uhci_state_t *uhcip, usb_ep_descr_t *endpoint, usb_port_status_t port_status) { int i = 0; uint_t interval; /* * Get the polling interval from the endpoint descriptor */ interval = endpoint->bInterval; /* * The bInterval value in the endpoint descriptor can range * from 1 to 255ms. The interrupt lattice has 32 leaf nodes, * and the host controller cycles through these nodes every * 32ms. The longest polling interval that the controller * supports is 32ms. */ /* * Return an error if the polling interval is less than 1ms * and greater than 255ms */ if ((interval < MIN_POLL_INTERVAL) || (interval > MAX_POLL_INTERVAL)) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_bandwidth_adjust: Endpoint's poll interval must be " "between %d and %d ms", MIN_POLL_INTERVAL, MAX_POLL_INTERVAL); return (USB_FAILURE); } /* * According USB Specifications, a full-speed endpoint can * specify a desired polling interval 1ms to 255ms and a low * speed endpoints are limited to specifying only 10ms to * 255ms. But some old keyboards & mice uses polling interval * of 8ms. For compatibility purpose, we are using polling * interval between 8ms & 255ms for low speed endpoints. */ if ((port_status == USBA_LOW_SPEED_DEV) && (interval < MIN_LOW_SPEED_POLL_INTERVAL)) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_bandwidth_adjust: Low speed endpoint's poll interval " "must be >= %d ms, adjusted", MIN_LOW_SPEED_POLL_INTERVAL); interval = MIN_LOW_SPEED_POLL_INTERVAL; } /* * If polling interval is greater than 32ms, * adjust polling interval equal to 32ms. */ if (interval > 32) { interval = 32; } /* * Find the nearest power of 2 that's less * than interval. */ while ((pow_2(i)) <= interval) { i++; } return (pow_2((i - 1))); } /* * uhci_lattice_height: * Given the requested bandwidth, find the height in the tree at * which the nodes for this bandwidth fall. The height is measured * as the number of nodes from the leaf to the level specified by * bandwidth The root of the tree is at height TREE_HEIGHT. */ static uint_t uhci_lattice_height(uint_t bandwidth) { return (TREE_HEIGHT - (log_2(bandwidth))); } static uint_t uhci_lattice_parent(uint_t node) { return (((node % 2) == 0) ? ((node/2) - 1) : (node/2)); } /* * uhci_leftmost_leaf: * Find the leftmost leaf in the subtree specified by the node. * Height refers to number of nodes from the bottom of the tree * to the node, including the node. */ static uint_t uhci_leftmost_leaf(uint_t node, uint_t height) { node = pow_2(height + VIRTUAL_TREE_HEIGHT) * (node+1) - NUM_FRAME_LST_ENTRIES; return (node); } /* * uhci_insert_qh: * Add the Queue Head (QH) into the Host Controller's (HC) * appropriate queue head list. */ void uhci_insert_qh(uhci_state_t *uhcip, usba_pipe_handle_data_t *ph) { uhci_pipe_private_t *pp = (uhci_pipe_private_t *)ph->p_hcd_private; USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_insert_qh:"); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); switch (UHCI_XFER_TYPE(&ph->p_ep)) { case USB_EP_ATTR_CONTROL: uhci_insert_ctrl_qh(uhcip, pp); break; case USB_EP_ATTR_BULK: uhci_insert_bulk_qh(uhcip, pp); break; case USB_EP_ATTR_INTR: uhci_insert_intr_qh(uhcip, pp); break; case USB_EP_ATTR_ISOCH: USB_DPRINTF_L2(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_insert_qh: Illegal request"); break; } } /* * uhci_insert_ctrl_qh: * Insert a control QH into the Host Controller's (HC) control QH list. */ static void uhci_insert_ctrl_qh(uhci_state_t *uhcip, uhci_pipe_private_t *pp) { queue_head_t *qh = pp->pp_qh; USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_insert_ctrl_qh:"); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); if (uhcip->uhci_ctrl_xfers_q_head == uhcip->uhci_ctrl_xfers_q_tail) { uhcip->uhci_ctrl_xfers_q_head->prev_qh = UHCI_INVALID_PTR; } SetQH32(uhcip, qh->link_ptr, GetQH32(uhcip, uhcip->uhci_ctrl_xfers_q_tail->link_ptr)); qh->prev_qh = uhcip->uhci_ctrl_xfers_q_tail; SetQH32(uhcip, uhcip->uhci_ctrl_xfers_q_tail->link_ptr, QH_PADDR(qh) | HC_QUEUE_HEAD); uhcip->uhci_ctrl_xfers_q_tail = qh; } /* * uhci_insert_bulk_qh: * Insert a bulk QH into the Host Controller's (HC) bulk QH list. */ static void uhci_insert_bulk_qh(uhci_state_t *uhcip, uhci_pipe_private_t *pp) { queue_head_t *qh = pp->pp_qh; USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_insert_bulk_qh:"); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); if (uhcip->uhci_bulk_xfers_q_head == uhcip->uhci_bulk_xfers_q_tail) { uhcip->uhci_bulk_xfers_q_head->prev_qh = UHCI_INVALID_PTR; } else if (uhcip->uhci_bulk_xfers_q_head->link_ptr == uhcip->uhci_bulk_xfers_q_tail->link_ptr) { /* If there is already a loop, we should keep the loop. */ qh->link_ptr = uhcip->uhci_bulk_xfers_q_tail->link_ptr; } qh->prev_qh = uhcip->uhci_bulk_xfers_q_tail; SetQH32(uhcip, uhcip->uhci_bulk_xfers_q_tail->link_ptr, QH_PADDR(qh) | HC_QUEUE_HEAD); uhcip->uhci_bulk_xfers_q_tail = qh; } /* * uhci_insert_intr_qh: * Insert a periodic Queue head i.e Interrupt queue head into the * Host Controller's (HC) interrupt lattice tree. */ static void uhci_insert_intr_qh(uhci_state_t *uhcip, uhci_pipe_private_t *pp) { uint_t node = pp->pp_node; /* The appropriate node was */ /* found during the opening */ /* of the pipe. */ queue_head_t *qh = pp->pp_qh; queue_head_t *next_lattice_qh, *lattice_qh; USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_insert_intr_qh:"); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); /* Find the lattice queue head */ lattice_qh = &uhcip->uhci_qh_pool_addr[node]; next_lattice_qh = QH_VADDR(GetQH32(uhcip, lattice_qh->link_ptr) & QH_LINK_PTR_MASK); next_lattice_qh->prev_qh = qh; qh->link_ptr = lattice_qh->link_ptr; qh->prev_qh = lattice_qh; SetQH32(uhcip, lattice_qh->link_ptr, QH_PADDR(qh) | HC_QUEUE_HEAD); pp->pp_data_toggle = 0; } /* * uhci_insert_intr_td: * Create a TD and a data buffer for an interrupt endpoint. */ int uhci_insert_intr_td( uhci_state_t *uhcip, usba_pipe_handle_data_t *ph, usb_intr_req_t *req, usb_flags_t flags) { int error, pipe_dir; uint_t length, mps; uint32_t buf_offs; uhci_td_t *tmp_td; usb_intr_req_t *intr_reqp; uhci_pipe_private_t *pp = (uhci_pipe_private_t *)ph->p_hcd_private; uhci_trans_wrapper_t *tw; USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_insert_intr_td: req: 0x%p", (void *)req); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); /* Get the interrupt pipe direction */ pipe_dir = UHCI_XFER_DIR(&ph->p_ep); /* Get the current interrupt request pointer */ if (req) { length = req->intr_len; } else { ASSERT(pipe_dir == USB_EP_DIR_IN); length = (pp->pp_client_periodic_in_reqp) ? (((usb_intr_req_t *)pp-> pp_client_periodic_in_reqp)->intr_len) : ph->p_ep.wMaxPacketSize; } /* Check the size of interrupt request */ if (length > UHCI_MAX_TD_XFER_SIZE) { /* the length shouldn't exceed 8K */ USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_insert_intr_td: Intr request size 0x%x is " "more than 0x%x", length, UHCI_MAX_TD_XFER_SIZE); return (USB_INVALID_REQUEST); } USB_DPRINTF_L3(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_insert_intr_td: length: 0x%x", length); /* Allocate a transaction wrapper */ if ((tw = uhci_create_transfer_wrapper(uhcip, pp, length, flags)) == NULL) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_insert_intr_td: TW allocation failed"); return (USB_NO_RESOURCES); } /* * Initialize the callback and any callback * data for when the td completes. */ tw->tw_handle_td = uhci_handle_intr_td; tw->tw_handle_callback_value = NULL; tw->tw_direction = (pipe_dir == USB_EP_DIR_OUT) ? PID_OUT : PID_IN; tw->tw_curr_xfer_reqp = (usb_opaque_t)req; /* * If it is an Interrupt IN request and interrupt request is NULL, * allocate the usb interrupt request structure for the current * interrupt polling request. */ if (tw->tw_direction == PID_IN) { if ((error = uhci_allocate_periodic_in_resource(uhcip, pp, tw, flags)) != USB_SUCCESS) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_insert_intr_td: Interrupt request structure " "allocation failed"); /* free the transfer wrapper */ uhci_deallocate_tw(uhcip, pp, tw); return (error); } } intr_reqp = (usb_intr_req_t *)tw->tw_curr_xfer_reqp; ASSERT(tw->tw_curr_xfer_reqp != NULL); tw->tw_timeout_cnt = (intr_reqp->intr_attributes & USB_ATTRS_ONE_XFER) ? intr_reqp->intr_timeout : 0; /* DATA IN */ if (tw->tw_direction == PID_IN) { /* Insert the td onto the queue head */ error = uhci_insert_hc_td(uhcip, 0, length, pp, tw, PID_IN, intr_reqp->intr_attributes); if (error != USB_SUCCESS) { uhci_deallocate_periodic_in_resource(uhcip, pp, tw); /* free the transfer wrapper */ uhci_deallocate_tw(uhcip, pp, tw); return (USB_NO_RESOURCES); } tw->tw_bytes_xfered = 0; return (USB_SUCCESS); } if (req->intr_len) { /* DATA OUT */ ASSERT(req->intr_data != NULL); /* Copy the data into the message */ ddi_rep_put8(tw->tw_accesshandle, req->intr_data->b_rptr, (uint8_t *)tw->tw_buf, req->intr_len, DDI_DEV_AUTOINCR); } /* set tw->tw_claim flag, so that nobody else works on this tw. */ tw->tw_claim = UHCI_INTR_HDLR_CLAIMED; mps = ph->p_ep.wMaxPacketSize; buf_offs = 0; /* Insert tds onto the queue head */ while (length > 0) { error = uhci_insert_hc_td(uhcip, buf_offs, (length > mps) ? mps : length, pp, tw, PID_OUT, intr_reqp->intr_attributes); if (error != USB_SUCCESS) { /* no resource. */ break; } if (length <= mps) { /* inserted all data. */ length = 0; } else { buf_offs += mps; length -= mps; } } if (error != USB_SUCCESS) { USB_DPRINTF_L2(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "uhci_insert_intr_td: allocate td failed, free resource"); /* remove all the tds */ while (tw->tw_hctd_head != NULL) { uhci_delete_td(uhcip, tw->tw_hctd_head); } tw->tw_claim = UHCI_NOT_CLAIMED; uhci_deallocate_tw(uhcip, pp, tw); return (error); } /* allow HC to xfer the tds of this tw */ tmp_td = tw->tw_hctd_head; while (tmp_td != NULL) { SetTD_status(uhcip, tmp_td, UHCI_TD_ACTIVE); tmp_td = tmp_td->tw_td_next; } tw->tw_bytes_xfered = 0; tw->tw_claim = UHCI_NOT_CLAIMED; return (error); } /* * uhci_create_transfer_wrapper: * Create a Transaction Wrapper (TW) for non-isoc transfer types. * This involves the allocating of DMA resources. * * For non-isoc transfers, one DMA handle and one DMA buffer are * allocated per transfer. The DMA buffer may contain multiple * DMA cookies and the cookies should meet certain alignment * requirement to be able to fit in the multiple TDs. The alignment * needs to ensure: * 1. the size of a cookie be larger than max TD length (0x500) * 2. the size of a cookie be a multiple of wMaxPacketSize of the * ctrl/bulk pipes * * wMaxPacketSize for ctrl and bulk pipes may be 8, 16, 32 or 64 bytes. * So the alignment should be a multiple of 64. wMaxPacketSize for intr * pipes is a little different since it only specifies the max to be * 64 bytes, but as long as an intr transfer is limited to max TD length, * any alignment can work if the cookie size is larger than max TD length. * * Considering the above conditions, 2K alignment is used. 4K alignment * should also be fine. */ static uhci_trans_wrapper_t * uhci_create_transfer_wrapper( uhci_state_t *uhcip, uhci_pipe_private_t *pp, size_t length, usb_flags_t usb_flags) { size_t real_length; uhci_trans_wrapper_t *tw; ddi_device_acc_attr_t dev_attr; ddi_dma_attr_t dma_attr; int kmem_flag; int (*dmamem_wait)(caddr_t); usba_pipe_handle_data_t *ph = pp->pp_pipe_handle; USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_create_transfer_wrapper: length = 0x%lx flags = 0x%x", length, usb_flags); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); /* isochronous pipe should not call into this function */ if (UHCI_XFER_TYPE(&ph->p_ep) == USB_EP_ATTR_ISOCH) { return (NULL); } /* SLEEP flag should not be used in interrupt context */ if (servicing_interrupt()) { kmem_flag = KM_NOSLEEP; dmamem_wait = DDI_DMA_DONTWAIT; } else { kmem_flag = KM_SLEEP; dmamem_wait = DDI_DMA_SLEEP; } /* Allocate space for the transfer wrapper */ if ((tw = kmem_zalloc(sizeof (uhci_trans_wrapper_t), kmem_flag)) == NULL) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_create_transfer_wrapper: kmem_alloc failed"); return (NULL); } /* zero-length packet doesn't need to allocate dma memory */ if (length == 0) { goto dmadone; } /* allow sg lists for transfer wrapper dma memory */ bcopy(&uhcip->uhci_dma_attr, &dma_attr, sizeof (ddi_dma_attr_t)); dma_attr.dma_attr_sgllen = UHCI_DMA_ATTR_SGLLEN; dma_attr.dma_attr_align = UHCI_DMA_ATTR_ALIGN; /* Store the transfer length */ tw->tw_length = length; /* Allocate the DMA handle */ if (ddi_dma_alloc_handle(uhcip->uhci_dip, &dma_attr, dmamem_wait, 0, &tw->tw_dmahandle) != DDI_SUCCESS) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_create_transfer_wrapper: Alloc handle failed"); kmem_free(tw, sizeof (uhci_trans_wrapper_t)); return (NULL); } dev_attr.devacc_attr_version = DDI_DEVICE_ATTR_V0; dev_attr.devacc_attr_endian_flags = DDI_STRUCTURE_LE_ACC; dev_attr.devacc_attr_dataorder = DDI_STRICTORDER_ACC; /* Allocate the memory */ if (ddi_dma_mem_alloc(tw->tw_dmahandle, tw->tw_length, &dev_attr, DDI_DMA_CONSISTENT, dmamem_wait, NULL, (caddr_t *)&tw->tw_buf, &real_length, &tw->tw_accesshandle) != DDI_SUCCESS) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_create_transfer_wrapper: dma_mem_alloc fail"); ddi_dma_free_handle(&tw->tw_dmahandle); kmem_free(tw, sizeof (uhci_trans_wrapper_t)); return (NULL); } ASSERT(real_length >= length); /* Bind the handle */ if (ddi_dma_addr_bind_handle(tw->tw_dmahandle, NULL, (caddr_t)tw->tw_buf, real_length, DDI_DMA_RDWR|DDI_DMA_CONSISTENT, dmamem_wait, NULL, &tw->tw_cookie, &tw->tw_ncookies) != DDI_DMA_MAPPED) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_create_transfer_wrapper: Bind handle failed"); ddi_dma_mem_free(&tw->tw_accesshandle); ddi_dma_free_handle(&tw->tw_dmahandle); kmem_free(tw, sizeof (uhci_trans_wrapper_t)); return (NULL); } tw->tw_cookie_idx = 0; tw->tw_dma_offs = 0; dmadone: /* * Only allow one wrapper to be added at a time. Insert the * new transaction wrapper into the list for this pipe. */ if (pp->pp_tw_head == NULL) { pp->pp_tw_head = tw; pp->pp_tw_tail = tw; } else { pp->pp_tw_tail->tw_next = tw; pp->pp_tw_tail = tw; ASSERT(tw->tw_next == NULL); } /* Store a back pointer to the pipe private structure */ tw->tw_pipe_private = pp; /* Store the transfer type - synchronous or asynchronous */ tw->tw_flags = usb_flags; USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_create_transfer_wrapper: tw = 0x%p, ncookies = %u", (void *)tw, tw->tw_ncookies); return (tw); } /* * uhci_insert_hc_td: * Insert a Transfer Descriptor (TD) on an QH. */ int uhci_insert_hc_td( uhci_state_t *uhcip, uint32_t buffer_offset, size_t hcgtd_length, uhci_pipe_private_t *pp, uhci_trans_wrapper_t *tw, uchar_t PID, usb_req_attrs_t attrs) { uhci_td_t *td, *current_dummy; queue_head_t *qh = pp->pp_qh; ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); if ((td = uhci_allocate_td_from_pool(uhcip)) == NULL) { return (USB_NO_RESOURCES); } current_dummy = qh->td_tailp; USB_DPRINTF_L4(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "uhci_insert_hc_td: td %p, attrs = 0x%x", (void *)td, attrs); /* * Fill in the current dummy td and * add the new dummy to the end. */ uhci_fill_in_td(uhcip, td, current_dummy, buffer_offset, hcgtd_length, pp, PID, attrs, tw); /* * Allow HC hardware xfer the td, except interrupt out td. */ if ((tw->tw_handle_td != uhci_handle_intr_td) || (PID != PID_OUT)) { SetTD_status(uhcip, current_dummy, UHCI_TD_ACTIVE); } /* Insert this td onto the tw */ if (tw->tw_hctd_head == NULL) { ASSERT(tw->tw_hctd_tail == NULL); tw->tw_hctd_head = current_dummy; tw->tw_hctd_tail = current_dummy; } else { /* Add the td to the end of the list */ tw->tw_hctd_tail->tw_td_next = current_dummy; tw->tw_hctd_tail = current_dummy; } /* * Insert the TD on to the QH. When this occurs, * the Host Controller will see the newly filled in TD */ current_dummy->outst_td_next = NULL; current_dummy->outst_td_prev = uhcip->uhci_outst_tds_tail; if (uhcip->uhci_outst_tds_head == NULL) { uhcip->uhci_outst_tds_head = current_dummy; } else { uhcip->uhci_outst_tds_tail->outst_td_next = current_dummy; } uhcip->uhci_outst_tds_tail = current_dummy; current_dummy->tw = tw; return (USB_SUCCESS); } /* * uhci_fill_in_td: * Fill in the fields of a Transfer Descriptor (TD). */ static void uhci_fill_in_td( uhci_state_t *uhcip, uhci_td_t *td, uhci_td_t *current_dummy, uint32_t buffer_offset, size_t length, uhci_pipe_private_t *pp, uchar_t PID, usb_req_attrs_t attrs, uhci_trans_wrapper_t *tw) { usba_pipe_handle_data_t *ph = pp->pp_pipe_handle; uint32_t buf_addr; USB_DPRINTF_L4(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "uhci_fill_in_td: td 0x%p buf_offs 0x%x len 0x%lx " "attrs 0x%x", (void *)td, buffer_offset, length, attrs); /* * If this is an isochronous TD, just return */ if (UHCI_XFER_TYPE(&ph->p_ep) == USB_EP_ATTR_ISOCH) { return; } /* The maximum transfer length of UHCI cannot exceed 0x500 bytes */ ASSERT(length <= UHCI_MAX_TD_XFER_SIZE); bzero((char *)td, sizeof (uhci_td_t)); /* Clear the TD */ SetTD32(uhcip, current_dummy->link_ptr, TD_PADDR(td)); if (attrs & USB_ATTRS_SHORT_XFER_OK) { SetTD_spd(uhcip, current_dummy, 1); } mutex_enter(&ph->p_usba_device->usb_mutex); if (ph->p_usba_device->usb_port_status == USBA_LOW_SPEED_DEV) { SetTD_ls(uhcip, current_dummy, LOW_SPEED_DEVICE); } SetTD_c_err(uhcip, current_dummy, UHCI_MAX_ERR_COUNT); SetTD_mlen(uhcip, current_dummy, (length == 0) ? ZERO_LENGTH : (length - 1)); SetTD_dtogg(uhcip, current_dummy, pp->pp_data_toggle); /* Adjust the data toggle bit */ ADJ_DATA_TOGGLE(pp); SetTD_devaddr(uhcip, current_dummy, ph->p_usba_device->usb_addr); SetTD_endpt(uhcip, current_dummy, ph->p_ep.bEndpointAddress & END_POINT_ADDRESS_MASK); SetTD_PID(uhcip, current_dummy, PID); SetTD_ioc(uhcip, current_dummy, INTERRUPT_ON_COMPLETION); buf_addr = uhci_get_tw_paddr_by_offs(uhcip, buffer_offset, length, tw); SetTD32(uhcip, current_dummy->buffer_address, buf_addr); td->qh_td_prev = current_dummy; current_dummy->qh_td_prev = NULL; pp->pp_qh->td_tailp = td; mutex_exit(&ph->p_usba_device->usb_mutex); } /* * uhci_get_tw_paddr_by_offs: * Walk through the DMA cookies of a TW buffer to retrieve * the device address used for a TD. * * buffer_offset - the starting offset into the TW buffer, where the * TD should transfer from. When a TW has more than * one TD, the TDs must be filled in increasing order. */ static uint32_t uhci_get_tw_paddr_by_offs( uhci_state_t *uhcip, uint32_t buffer_offset, size_t length, uhci_trans_wrapper_t *tw) { uint32_t buf_addr; int rem_len; USB_DPRINTF_L4(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "uhci_get_tw_paddr_by_offs: buf_offs 0x%x len 0x%lx", buffer_offset, length); /* * TDs must be filled in increasing DMA offset order. * tw_dma_offs is initialized to be 0 at TW creation and * is only increased in this function. */ ASSERT(length == 0 || buffer_offset >= tw->tw_dma_offs); if (length == 0) { buf_addr = 0; return (buf_addr); } /* * Advance to the next DMA cookie until finding the cookie * that buffer_offset falls in. * It is very likely this loop will never repeat more than * once. It is here just to accommodate the case buffer_offset * is increased by multiple cookies during two consecutive * calls into this function. In that case, the interim DMA * buffer is allowed to be skipped. */ while ((tw->tw_dma_offs + tw->tw_cookie.dmac_size) <= buffer_offset) { /* * tw_dma_offs always points to the starting offset * of a cookie */ tw->tw_dma_offs += tw->tw_cookie.dmac_size; ddi_dma_nextcookie(tw->tw_dmahandle, &tw->tw_cookie); tw->tw_cookie_idx++; ASSERT(tw->tw_cookie_idx < tw->tw_ncookies); } /* * Counting the remained buffer length to be filled in * the TDs for current DMA cookie */ rem_len = (tw->tw_dma_offs + tw->tw_cookie.dmac_size) - buffer_offset; /* Calculate the beginning address of the buffer */ ASSERT(length <= rem_len); buf_addr = (buffer_offset - tw->tw_dma_offs) + tw->tw_cookie.dmac_address; USB_DPRINTF_L3(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "uhci_get_tw_paddr_by_offs: dmac_addr 0x%x dmac_size " "0x%lx idx %d", buf_addr, tw->tw_cookie.dmac_size, tw->tw_cookie_idx); return (buf_addr); } /* * uhci_modify_td_active_bits: * Sets active bit in all the tds of QH to INACTIVE so that * the HC stops processing the TD's related to the QH. */ void uhci_modify_td_active_bits( uhci_state_t *uhcip, uhci_pipe_private_t *pp) { uhci_td_t *td_head; usb_ep_descr_t *ept = &pp->pp_pipe_handle->p_ep; uhci_trans_wrapper_t *tw_head = pp->pp_tw_head; USB_DPRINTF_L4(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "uhci_modify_td_active_bits: tw head %p", (void *)tw_head); while (tw_head != NULL) { tw_head->tw_claim = UHCI_MODIFY_TD_BITS_CLAIMED; td_head = tw_head->tw_hctd_head; while (td_head) { if (UHCI_XFER_TYPE(ept) == USB_EP_ATTR_ISOCH) { SetTD_status(uhcip, td_head, GetTD_status(uhcip, td_head) & TD_INACTIVE); } else { SetTD32(uhcip, td_head->link_ptr, GetTD32(uhcip, td_head->link_ptr) | HC_END_OF_LIST); } td_head = td_head->tw_td_next; } tw_head = tw_head->tw_next; } } /* * uhci_insert_ctrl_td: * Create a TD and a data buffer for a control Queue Head. */ int uhci_insert_ctrl_td( uhci_state_t *uhcip, usba_pipe_handle_data_t *ph, usb_ctrl_req_t *ctrl_reqp, usb_flags_t flags) { uhci_pipe_private_t *pp = (uhci_pipe_private_t *)ph->p_hcd_private; uhci_trans_wrapper_t *tw; size_t ctrl_buf_size; USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_insert_ctrl_td: timeout: 0x%x", ctrl_reqp->ctrl_timeout); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); /* * If we have a control data phase, make the data buffer start * on the next 64-byte boundary so as to ensure the DMA cookie * can fit in the multiple TDs. The buffer in the range of * [SETUP_SIZE, UHCI_CTRL_EPT_MAX_SIZE) is just for padding * and not to be transferred. */ if (ctrl_reqp->ctrl_wLength) { ctrl_buf_size = UHCI_CTRL_EPT_MAX_SIZE + ctrl_reqp->ctrl_wLength; } else { ctrl_buf_size = SETUP_SIZE; } /* Allocate a transaction wrapper */ if ((tw = uhci_create_transfer_wrapper(uhcip, pp, ctrl_buf_size, flags)) == NULL) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_insert_ctrl_td: TW allocation failed"); return (USB_NO_RESOURCES); } pp->pp_data_toggle = 0; tw->tw_curr_xfer_reqp = (usb_opaque_t)ctrl_reqp; tw->tw_bytes_xfered = 0; tw->tw_bytes_pending = ctrl_reqp->ctrl_wLength; tw->tw_timeout_cnt = max(UHCI_CTRL_TIMEOUT, ctrl_reqp->ctrl_timeout); /* * Initialize the callback and any callback * data for when the td completes. */ tw->tw_handle_td = uhci_handle_ctrl_td; tw->tw_handle_callback_value = NULL; if ((uhci_create_setup_pkt(uhcip, pp, tw)) != USB_SUCCESS) { tw->tw_ctrl_state = 0; /* free the transfer wrapper */ uhci_deallocate_tw(uhcip, pp, tw); return (USB_NO_RESOURCES); } tw->tw_ctrl_state = SETUP; return (USB_SUCCESS); } /* * uhci_create_setup_pkt: * create a setup packet to initiate a control transfer. * * OHCI driver has seen the case where devices fail if there is * more than one control transfer to the device within a frame. * So, the UHCI ensures that only one TD will be put on the control * pipe to one device (to be consistent with OHCI driver). */ static int uhci_create_setup_pkt( uhci_state_t *uhcip, uhci_pipe_private_t *pp, uhci_trans_wrapper_t *tw) { int sdata; usb_ctrl_req_t *req = (usb_ctrl_req_t *)tw->tw_curr_xfer_reqp; USB_DPRINTF_L4(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "uhci_create_setup_pkt: 0x%x 0x%x 0x%x 0x%x 0x%x 0x%p", req->ctrl_bmRequestType, req->ctrl_bRequest, req->ctrl_wValue, req->ctrl_wIndex, req->ctrl_wLength, (void *)req->ctrl_data); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); ASSERT(tw != NULL); /* Create the first four bytes of the setup packet */ sdata = (req->ctrl_bmRequestType | (req->ctrl_bRequest << 8) | (req->ctrl_wValue << 16)); ddi_put32(tw->tw_accesshandle, (uint_t *)tw->tw_buf, sdata); /* Create the second four bytes */ sdata = (uint32_t)(req->ctrl_wIndex | (req->ctrl_wLength << 16)); ddi_put32(tw->tw_accesshandle, (uint_t *)(tw->tw_buf + sizeof (uint_t)), sdata); /* * The TD's are placed on the QH one at a time. * Once this TD is placed on the done list, the * data or status phase TD will be enqueued. */ if ((uhci_insert_hc_td(uhcip, 0, SETUP_SIZE, pp, tw, PID_SETUP, req->ctrl_attributes)) != USB_SUCCESS) { return (USB_NO_RESOURCES); } USB_DPRINTF_L3(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "Create_setup: pp = 0x%p, attrs = 0x%x", (void *)pp, req->ctrl_attributes); /* * If this control transfer has a data phase, record the * direction. If the data phase is an OUT transaction , * copy the data into the buffer of the transfer wrapper. */ if (req->ctrl_wLength != 0) { /* There is a data stage. Find the direction */ if (req->ctrl_bmRequestType & USB_DEV_REQ_DEV_TO_HOST) { tw->tw_direction = PID_IN; } else { tw->tw_direction = PID_OUT; /* Copy the data into the buffer */ ddi_rep_put8(tw->tw_accesshandle, req->ctrl_data->b_rptr, (uint8_t *)(tw->tw_buf + UHCI_CTRL_EPT_MAX_SIZE), req->ctrl_wLength, DDI_DEV_AUTOINCR); } } return (USB_SUCCESS); } /* * uhci_create_stats: * Allocate and initialize the uhci kstat structures */ void uhci_create_stats(uhci_state_t *uhcip) { int i; char kstatname[KSTAT_STRLEN]; char *usbtypes[USB_N_COUNT_KSTATS] = {"ctrl", "isoch", "bulk", "intr"}; uint_t instance = uhcip->uhci_instance; const char *dname = ddi_driver_name(uhcip->uhci_dip); uhci_intrs_stats_t *isp; if (UHCI_INTRS_STATS(uhcip) == NULL) { (void) snprintf(kstatname, KSTAT_STRLEN, "%s%d,intrs", dname, instance); UHCI_INTRS_STATS(uhcip) = kstat_create("usba", instance, kstatname, "usb_interrupts", KSTAT_TYPE_NAMED, sizeof (uhci_intrs_stats_t) / sizeof (kstat_named_t), KSTAT_FLAG_PERSISTENT); if (UHCI_INTRS_STATS(uhcip) != NULL) { isp = UHCI_INTRS_STATS_DATA(uhcip); kstat_named_init(&isp->uhci_intrs_hc_halted, "HC Halted", KSTAT_DATA_UINT64); kstat_named_init(&isp->uhci_intrs_hc_process_err, "HC Process Errors", KSTAT_DATA_UINT64); kstat_named_init(&isp->uhci_intrs_host_sys_err, "Host Sys Errors", KSTAT_DATA_UINT64); kstat_named_init(&isp->uhci_intrs_resume_detected, "Resume Detected", KSTAT_DATA_UINT64); kstat_named_init(&isp->uhci_intrs_usb_err_intr, "USB Error", KSTAT_DATA_UINT64); kstat_named_init(&isp->uhci_intrs_usb_intr, "USB Interrupts", KSTAT_DATA_UINT64); kstat_named_init(&isp->uhci_intrs_total, "Total Interrupts", KSTAT_DATA_UINT64); kstat_named_init(&isp->uhci_intrs_not_claimed, "Not Claimed", KSTAT_DATA_UINT64); UHCI_INTRS_STATS(uhcip)->ks_private = uhcip; UHCI_INTRS_STATS(uhcip)->ks_update = nulldev; kstat_install(UHCI_INTRS_STATS(uhcip)); } } if (UHCI_TOTAL_STATS(uhcip) == NULL) { (void) snprintf(kstatname, KSTAT_STRLEN, "%s%d,total", dname, instance); UHCI_TOTAL_STATS(uhcip) = kstat_create("usba", instance, kstatname, "usb_byte_count", KSTAT_TYPE_IO, 1, KSTAT_FLAG_PERSISTENT); if (UHCI_TOTAL_STATS(uhcip) != NULL) { kstat_install(UHCI_TOTAL_STATS(uhcip)); } } for (i = 0; i < USB_N_COUNT_KSTATS; i++) { if (uhcip->uhci_count_stats[i] == NULL) { (void) snprintf(kstatname, KSTAT_STRLEN, "%s%d,%s", dname, instance, usbtypes[i]); uhcip->uhci_count_stats[i] = kstat_create("usba", instance, kstatname, "usb_byte_count", KSTAT_TYPE_IO, 1, KSTAT_FLAG_PERSISTENT); if (uhcip->uhci_count_stats[i] != NULL) { kstat_install(uhcip->uhci_count_stats[i]); } } } } /* * uhci_destroy_stats: * Clean up uhci kstat structures */ void uhci_destroy_stats(uhci_state_t *uhcip) { int i; if (UHCI_INTRS_STATS(uhcip)) { kstat_delete(UHCI_INTRS_STATS(uhcip)); UHCI_INTRS_STATS(uhcip) = NULL; } if (UHCI_TOTAL_STATS(uhcip)) { kstat_delete(UHCI_TOTAL_STATS(uhcip)); UHCI_TOTAL_STATS(uhcip) = NULL; } for (i = 0; i < USB_N_COUNT_KSTATS; i++) { if (uhcip->uhci_count_stats[i]) { kstat_delete(uhcip->uhci_count_stats[i]); uhcip->uhci_count_stats[i] = NULL; } } } void uhci_do_intrs_stats(uhci_state_t *uhcip, int val) { if (UHCI_INTRS_STATS(uhcip) == NULL) { return; } UHCI_INTRS_STATS_DATA(uhcip)->uhci_intrs_total.value.ui64++; switch (val) { case USBSTS_REG_HC_HALTED: UHCI_INTRS_STATS_DATA(uhcip)->uhci_intrs_hc_halted.value.ui64++; break; case USBSTS_REG_HC_PROCESS_ERR: UHCI_INTRS_STATS_DATA(uhcip)-> uhci_intrs_hc_process_err.value.ui64++; break; case USBSTS_REG_HOST_SYS_ERR: UHCI_INTRS_STATS_DATA(uhcip)-> uhci_intrs_host_sys_err.value.ui64++; break; case USBSTS_REG_RESUME_DETECT: UHCI_INTRS_STATS_DATA(uhcip)-> uhci_intrs_resume_detected.value.ui64++; break; case USBSTS_REG_USB_ERR_INTR: UHCI_INTRS_STATS_DATA(uhcip)-> uhci_intrs_usb_err_intr.value.ui64++; break; case USBSTS_REG_USB_INTR: UHCI_INTRS_STATS_DATA(uhcip)->uhci_intrs_usb_intr.value.ui64++; break; default: UHCI_INTRS_STATS_DATA(uhcip)-> uhci_intrs_not_claimed.value.ui64++; break; } } void uhci_do_byte_stats(uhci_state_t *uhcip, size_t len, uint8_t attr, uint8_t addr) { uint8_t type = attr & USB_EP_ATTR_MASK; uint8_t dir = addr & USB_EP_DIR_MASK; switch (dir) { case USB_EP_DIR_IN: UHCI_TOTAL_STATS_DATA(uhcip)->reads++; UHCI_TOTAL_STATS_DATA(uhcip)->nread += len; switch (type) { case USB_EP_ATTR_CONTROL: UHCI_CTRL_STATS(uhcip)->reads++; UHCI_CTRL_STATS(uhcip)->nread += len; break; case USB_EP_ATTR_BULK: UHCI_BULK_STATS(uhcip)->reads++; UHCI_BULK_STATS(uhcip)->nread += len; break; case USB_EP_ATTR_INTR: UHCI_INTR_STATS(uhcip)->reads++; UHCI_INTR_STATS(uhcip)->nread += len; break; case USB_EP_ATTR_ISOCH: UHCI_ISOC_STATS(uhcip)->reads++; UHCI_ISOC_STATS(uhcip)->nread += len; break; } break; case USB_EP_DIR_OUT: UHCI_TOTAL_STATS_DATA(uhcip)->writes++; UHCI_TOTAL_STATS_DATA(uhcip)->nwritten += len; switch (type) { case USB_EP_ATTR_CONTROL: UHCI_CTRL_STATS(uhcip)->writes++; UHCI_CTRL_STATS(uhcip)->nwritten += len; break; case USB_EP_ATTR_BULK: UHCI_BULK_STATS(uhcip)->writes++; UHCI_BULK_STATS(uhcip)->nwritten += len; break; case USB_EP_ATTR_INTR: UHCI_INTR_STATS(uhcip)->writes++; UHCI_INTR_STATS(uhcip)->nwritten += len; break; case USB_EP_ATTR_ISOCH: UHCI_ISOC_STATS(uhcip)->writes++; UHCI_ISOC_STATS(uhcip)->nwritten += len; break; } break; } } /* * uhci_free_tw: * Free the Transfer Wrapper (TW). */ void uhci_free_tw(uhci_state_t *uhcip, uhci_trans_wrapper_t *tw) { int rval, i; USB_DPRINTF_L4(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "uhci_free_tw:"); ASSERT(tw != NULL); if (tw->tw_isoc_strtlen > 0) { ASSERT(tw->tw_isoc_bufs != NULL); for (i = 0; i < tw->tw_ncookies; i++) { rval = ddi_dma_unbind_handle( tw->tw_isoc_bufs[i].dma_handle); ASSERT(rval == USB_SUCCESS); ddi_dma_mem_free(&tw->tw_isoc_bufs[i].mem_handle); ddi_dma_free_handle(&tw->tw_isoc_bufs[i].dma_handle); } kmem_free(tw->tw_isoc_bufs, tw->tw_isoc_strtlen); } else if (tw->tw_dmahandle != NULL) { rval = ddi_dma_unbind_handle(tw->tw_dmahandle); ASSERT(rval == DDI_SUCCESS); ddi_dma_mem_free(&tw->tw_accesshandle); ddi_dma_free_handle(&tw->tw_dmahandle); } kmem_free(tw, sizeof (uhci_trans_wrapper_t)); } /* * uhci_deallocate_tw: * Deallocate of a Transaction Wrapper (TW) and this involves * the freeing of DMA resources. */ void uhci_deallocate_tw(uhci_state_t *uhcip, uhci_pipe_private_t *pp, uhci_trans_wrapper_t *tw) { uhci_trans_wrapper_t *head; USB_DPRINTF_L4(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "uhci_deallocate_tw:"); /* * If the transfer wrapper has no Host Controller (HC) * Transfer Descriptors (TD) associated with it, then * remove the transfer wrapper. The transfers are done * in FIFO order, so this should be the first transfer * wrapper on the list. */ if (tw->tw_hctd_head != NULL) { ASSERT(tw->tw_hctd_tail != NULL); return; } ASSERT(tw->tw_hctd_tail == NULL); ASSERT(pp->pp_tw_head != NULL); /* * If pp->pp_tw_head is NULL, set the tail also to NULL. */ head = pp->pp_tw_head; if (head == tw) { pp->pp_tw_head = head->tw_next; if (pp->pp_tw_head == NULL) { pp->pp_tw_tail = NULL; } } else { while (head->tw_next != tw) head = head->tw_next; head->tw_next = tw->tw_next; if (tw->tw_next == NULL) { pp->pp_tw_tail = head; } } uhci_free_tw(uhcip, tw); } void uhci_delete_td(uhci_state_t *uhcip, uhci_td_t *td) { uhci_td_t *tmp_td; uhci_trans_wrapper_t *tw = td->tw; if ((td->outst_td_next == NULL) && (td->outst_td_prev == NULL)) { uhcip->uhci_outst_tds_head = NULL; uhcip->uhci_outst_tds_tail = NULL; } else if (td->outst_td_next == NULL) { td->outst_td_prev->outst_td_next = NULL; uhcip->uhci_outst_tds_tail = td->outst_td_prev; } else if (td->outst_td_prev == NULL) { td->outst_td_next->outst_td_prev = NULL; uhcip->uhci_outst_tds_head = td->outst_td_next; } else { td->outst_td_prev->outst_td_next = td->outst_td_next; td->outst_td_next->outst_td_prev = td->outst_td_prev; } tmp_td = tw->tw_hctd_head; if (tmp_td != td) { while (tmp_td->tw_td_next != td) { tmp_td = tmp_td->tw_td_next; } ASSERT(tmp_td); tmp_td->tw_td_next = td->tw_td_next; if (td->tw_td_next == NULL) { tw->tw_hctd_tail = tmp_td; } } else { tw->tw_hctd_head = tw->tw_hctd_head->tw_td_next; if (tw->tw_hctd_head == NULL) { tw->tw_hctd_tail = NULL; } } td->flag = TD_FLAG_FREE; } void uhci_remove_tds_tws( uhci_state_t *uhcip, usba_pipe_handle_data_t *ph) { usb_opaque_t curr_reqp; uhci_pipe_private_t *pp = (uhci_pipe_private_t *)ph->p_hcd_private; usb_ep_descr_t *ept = &pp->pp_pipe_handle->p_ep; uhci_trans_wrapper_t *tw_tmp; uhci_trans_wrapper_t *tw_head = pp->pp_tw_head; while (tw_head != NULL) { tw_tmp = tw_head; tw_head = tw_head->tw_next; curr_reqp = tw_tmp->tw_curr_xfer_reqp; if (curr_reqp) { /* do this for control/bulk/intr */ if ((tw_tmp->tw_direction == PID_IN) && (UHCI_XFER_TYPE(ept) == USB_EP_ATTR_INTR)) { uhci_deallocate_periodic_in_resource(uhcip, pp, tw_tmp); } else { uhci_hcdi_callback(uhcip, pp, pp->pp_pipe_handle, tw_tmp, USB_CR_FLUSHED); } } /* end of curr_reqp */ if (tw_tmp->tw_claim != UHCI_MODIFY_TD_BITS_CLAIMED) { continue; } while (tw_tmp->tw_hctd_head != NULL) { uhci_delete_td(uhcip, tw_tmp->tw_hctd_head); } uhci_deallocate_tw(uhcip, pp, tw_tmp); } } /* * uhci_remove_qh: * Remove the Queue Head from the Host Controller's * appropriate QH list. */ void uhci_remove_qh(uhci_state_t *uhcip, uhci_pipe_private_t *pp) { uhci_td_t *dummy_td; ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_remove_qh:"); dummy_td = pp->pp_qh->td_tailp; dummy_td->flag = TD_FLAG_FREE; switch (UHCI_XFER_TYPE(&pp->pp_pipe_handle->p_ep)) { case USB_EP_ATTR_CONTROL: uhci_remove_ctrl_qh(uhcip, pp); break; case USB_EP_ATTR_BULK: uhci_remove_bulk_qh(uhcip, pp); break; case USB_EP_ATTR_INTR: uhci_remove_intr_qh(uhcip, pp); break; } } static void uhci_remove_intr_qh(uhci_state_t *uhcip, uhci_pipe_private_t *pp) { queue_head_t *qh = pp->pp_qh; queue_head_t *next_lattice_qh = QH_VADDR(GetQH32(uhcip, qh->link_ptr) & QH_LINK_PTR_MASK); qh->prev_qh->link_ptr = qh->link_ptr; next_lattice_qh->prev_qh = qh->prev_qh; qh->qh_flag = QUEUE_HEAD_FLAG_FREE; } /* * uhci_remove_bulk_qh: * Remove a bulk QH from the Host Controller's QH list. There may be a * loop for bulk QHs, we must care about this while removing a bulk QH. */ static void uhci_remove_bulk_qh(uhci_state_t *uhcip, uhci_pipe_private_t *pp) { queue_head_t *qh = pp->pp_qh; queue_head_t *next_lattice_qh; uint32_t paddr; paddr = (GetQH32(uhcip, qh->link_ptr) & QH_LINK_PTR_MASK); next_lattice_qh = (qh == uhcip->uhci_bulk_xfers_q_tail) ? 0 : QH_VADDR(paddr); if ((qh == uhcip->uhci_bulk_xfers_q_tail) && (qh->prev_qh == uhcip->uhci_bulk_xfers_q_head)) { SetQH32(uhcip, qh->prev_qh->link_ptr, HC_END_OF_LIST); } else { qh->prev_qh->link_ptr = qh->link_ptr; } if (next_lattice_qh == NULL) { uhcip->uhci_bulk_xfers_q_tail = qh->prev_qh; } else { next_lattice_qh->prev_qh = qh->prev_qh; } qh->qh_flag = QUEUE_HEAD_FLAG_FREE; } static void uhci_remove_ctrl_qh(uhci_state_t *uhcip, uhci_pipe_private_t *pp) { queue_head_t *qh = pp->pp_qh; queue_head_t *next_lattice_qh = QH_VADDR(GetQH32(uhcip, qh->link_ptr) & QH_LINK_PTR_MASK); qh->prev_qh->link_ptr = qh->link_ptr; if (next_lattice_qh->prev_qh != NULL) { next_lattice_qh->prev_qh = qh->prev_qh; } else { uhcip->uhci_ctrl_xfers_q_tail = qh->prev_qh; } qh->qh_flag = QUEUE_HEAD_FLAG_FREE; } /* * uhci_allocate_td_from_pool: * Allocate a Transfer Descriptor (TD) from the TD buffer pool. */ static uhci_td_t * uhci_allocate_td_from_pool(uhci_state_t *uhcip) { int index; uhci_td_t *td; ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); /* * Search for a blank Transfer Descriptor (TD) * in the TD buffer pool. */ for (index = 0; index < uhci_td_pool_size; index ++) { if (uhcip->uhci_td_pool_addr[index].flag == TD_FLAG_FREE) { break; } } if (index == uhci_td_pool_size) { USB_DPRINTF_L2(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "uhci_allocate_td_from_pool: TD exhausted"); return (NULL); } USB_DPRINTF_L4(PRINT_MASK_ALLOC, uhcip->uhci_log_hdl, "uhci_allocate_td_from_pool: Allocated %d", index); /* Create a new dummy for the end of the TD list */ td = &uhcip->uhci_td_pool_addr[index]; /* Mark the newly allocated TD as a dummy */ td->flag = TD_FLAG_DUMMY; td->qh_td_prev = NULL; return (td); } /* * uhci_insert_bulk_td: */ int uhci_insert_bulk_td( uhci_state_t *uhcip, usba_pipe_handle_data_t *ph, usb_bulk_req_t *req, usb_flags_t flags) { size_t length; uint_t mps; /* MaxPacketSize */ uint_t num_bulk_tds, i, j; uint32_t buf_offs; uhci_td_t *bulk_td_ptr; uhci_td_t *current_dummy, *tmp_td; uhci_pipe_private_t *pp = (uhci_pipe_private_t *)ph->p_hcd_private; uhci_trans_wrapper_t *tw; uhci_bulk_isoc_xfer_t *bulk_xfer_info; uhci_bulk_isoc_td_pool_t *td_pool_ptr; USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_insert_bulk_td: req: 0x%p, flags = 0x%x", (void *)req, flags); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); /* * Create transfer wrapper */ if ((tw = uhci_create_transfer_wrapper(uhcip, pp, req->bulk_len, flags)) == NULL) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_insert_bulk_td: TW allocation failed"); return (USB_NO_RESOURCES); } tw->tw_bytes_xfered = 0; tw->tw_bytes_pending = req->bulk_len; tw->tw_handle_td = uhci_handle_bulk_td; tw->tw_handle_callback_value = (usb_opaque_t)req->bulk_data; tw->tw_timeout_cnt = req->bulk_timeout; tw->tw_data = req->bulk_data; tw->tw_curr_xfer_reqp = (usb_opaque_t)req; /* Get the bulk pipe direction */ tw->tw_direction = (UHCI_XFER_DIR(&ph->p_ep) == USB_EP_DIR_OUT) ? PID_OUT : PID_IN; USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_insert_bulk_td: direction: 0x%x", tw->tw_direction); /* If the DATA OUT, copy the data into transfer buffer. */ if (tw->tw_direction == PID_OUT) { if (req->bulk_len) { ASSERT(req->bulk_data != NULL); /* Copy the data into the message */ ddi_rep_put8(tw->tw_accesshandle, req->bulk_data->b_rptr, (uint8_t *)tw->tw_buf, req->bulk_len, DDI_DEV_AUTOINCR); } } /* Get the max packet size. */ length = mps = pp->pp_pipe_handle->p_ep.wMaxPacketSize; /* * Calculate number of TD's to insert in the current frame interval. * Max number TD's allowed (driver implementation) is 128 * in one frame interval. Once all the TD's are completed * then the remaining TD's will be inserted into the lattice * in the uhci_handle_bulk_td(). */ if ((tw->tw_bytes_pending / mps) >= MAX_NUM_BULK_TDS_PER_XFER) { num_bulk_tds = MAX_NUM_BULK_TDS_PER_XFER; } else { num_bulk_tds = (tw->tw_bytes_pending / mps); if (tw->tw_bytes_pending % mps || tw->tw_bytes_pending == 0) { num_bulk_tds++; length = (tw->tw_bytes_pending % mps); } } /* * Allocate memory for the bulk xfer information structure */ if ((bulk_xfer_info = kmem_zalloc( sizeof (uhci_bulk_isoc_xfer_t), KM_NOSLEEP)) == NULL) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_insert_bulk_td: kmem_zalloc failed"); /* Free the transfer wrapper */ uhci_deallocate_tw(uhcip, pp, tw); return (USB_FAILURE); } /* Allocate memory for the bulk TD's */ if (uhci_alloc_bulk_isoc_tds(uhcip, num_bulk_tds, bulk_xfer_info) != USB_SUCCESS) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_insert_bulk_td: alloc_bulk_isoc_tds failed"); kmem_free(bulk_xfer_info, sizeof (uhci_bulk_isoc_xfer_t)); /* Free the transfer wrapper */ uhci_deallocate_tw(uhcip, pp, tw); return (USB_FAILURE); } td_pool_ptr = &bulk_xfer_info->td_pools[0]; bulk_td_ptr = (uhci_td_t *)td_pool_ptr->pool_addr; bulk_td_ptr[0].qh_td_prev = NULL; current_dummy = pp->pp_qh->td_tailp; buf_offs = 0; pp->pp_qh->bulk_xfer_info = bulk_xfer_info; /* Fill up all the bulk TD's */ for (i = 0; i < bulk_xfer_info->num_pools; i++) { for (j = 0; j < (td_pool_ptr->num_tds - 1); j++) { uhci_fill_in_bulk_isoc_td(uhcip, &bulk_td_ptr[j], &bulk_td_ptr[j+1], BULKTD_PADDR(td_pool_ptr, &bulk_td_ptr[j+1]), ph, buf_offs, mps, tw); buf_offs += mps; } /* fill in the last TD */ if (i == (bulk_xfer_info->num_pools - 1)) { uhci_fill_in_bulk_isoc_td(uhcip, &bulk_td_ptr[j], current_dummy, TD_PADDR(current_dummy), ph, buf_offs, length, tw); } else { /* fill in the TD at the tail of a pool */ tmp_td = &bulk_td_ptr[j]; td_pool_ptr = &bulk_xfer_info->td_pools[i + 1]; bulk_td_ptr = (uhci_td_t *)td_pool_ptr->pool_addr; uhci_fill_in_bulk_isoc_td(uhcip, tmp_td, &bulk_td_ptr[0], BULKTD_PADDR(td_pool_ptr, &bulk_td_ptr[0]), ph, buf_offs, mps, tw); buf_offs += mps; } } bulk_xfer_info->num_tds = (ushort_t)num_bulk_tds; /* * Point the end of the lattice tree to the start of the bulk xfers * queue head. This allows the HC to execute the same Queue Head/TD * in the same frame. There are some bulk devices, which NAKs after * completing each TD. As a result, the performance on such devices * is very bad. This loop will provide a chance to execute NAk'ed * bulk TDs again in the same frame. */ if (uhcip->uhci_pending_bulk_cmds++ == 0) { uhcip->uhci_bulk_xfers_q_tail->link_ptr = uhcip->uhci_bulk_xfers_q_head->link_ptr; USB_DPRINTF_L3(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_insert_bulk_td: count = %d no tds %d", uhcip->uhci_pending_bulk_cmds, num_bulk_tds); } /* Insert on the bulk queue head for the execution by HC */ SetQH32(uhcip, pp->pp_qh->element_ptr, bulk_xfer_info->td_pools[0].cookie.dmac_address); return (USB_SUCCESS); } /* * uhci_fill_in_bulk_isoc_td * Fills the bulk/isoc TD * * offset - different meanings for bulk and isoc TDs: * starting offset into the TW buffer for a bulk TD * and the index into the isoc packet list for an isoc TD */ void uhci_fill_in_bulk_isoc_td(uhci_state_t *uhcip, uhci_td_t *current_td, uhci_td_t *next_td, uint32_t next_td_paddr, usba_pipe_handle_data_t *ph, uint_t offset, uint_t length, uhci_trans_wrapper_t *tw) { uhci_pipe_private_t *pp = (uhci_pipe_private_t *)ph->p_hcd_private; usb_ep_descr_t *ept = &pp->pp_pipe_handle->p_ep; uint32_t buf_addr; USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_fill_in_bulk_isoc_td: tw 0x%p offs 0x%x length 0x%x", (void *)tw, offset, length); bzero((char *)current_td, sizeof (uhci_td_t)); SetTD32(uhcip, current_td->link_ptr, next_td_paddr | HC_DEPTH_FIRST); switch (UHCI_XFER_TYPE(ept)) { case USB_EP_ATTR_ISOCH: if (((usb_isoc_req_t *)tw->tw_curr_xfer_reqp)->isoc_attributes & USB_ATTRS_SHORT_XFER_OK) { SetTD_spd(uhcip, current_td, 1); } break; case USB_EP_ATTR_BULK: if (((usb_bulk_req_t *)tw->tw_curr_xfer_reqp)->bulk_attributes & USB_ATTRS_SHORT_XFER_OK) { SetTD_spd(uhcip, current_td, 1); } break; } mutex_enter(&ph->p_usba_device->usb_mutex); SetTD_c_err(uhcip, current_td, UHCI_MAX_ERR_COUNT); SetTD_status(uhcip, current_td, UHCI_TD_ACTIVE); SetTD_ioc(uhcip, current_td, INTERRUPT_ON_COMPLETION); SetTD_mlen(uhcip, current_td, (length == 0) ? ZERO_LENGTH : (length - 1)); SetTD_dtogg(uhcip, current_td, pp->pp_data_toggle); SetTD_devaddr(uhcip, current_td, ph->p_usba_device->usb_addr); SetTD_endpt(uhcip, current_td, ph->p_ep.bEndpointAddress & END_POINT_ADDRESS_MASK); SetTD_PID(uhcip, current_td, tw->tw_direction); /* Get the right buffer address for the current TD */ switch (UHCI_XFER_TYPE(ept)) { case USB_EP_ATTR_ISOCH: buf_addr = tw->tw_isoc_bufs[offset].cookie.dmac_address; break; case USB_EP_ATTR_BULK: buf_addr = uhci_get_tw_paddr_by_offs(uhcip, offset, length, tw); break; } SetTD32(uhcip, current_td->buffer_address, buf_addr); /* * Adjust the data toggle. * The data toggle bit must always be 0 for isoc transfers. * And set the "iso" bit in the TD for isoc transfers. */ if (UHCI_XFER_TYPE(ept) == USB_EP_ATTR_ISOCH) { pp->pp_data_toggle = 0; SetTD_iso(uhcip, current_td, 1); } else { ADJ_DATA_TOGGLE(pp); next_td->qh_td_prev = current_td; pp->pp_qh->td_tailp = next_td; } current_td->outst_td_next = NULL; current_td->outst_td_prev = uhcip->uhci_outst_tds_tail; if (uhcip->uhci_outst_tds_head == NULL) { uhcip->uhci_outst_tds_head = current_td; } else { uhcip->uhci_outst_tds_tail->outst_td_next = current_td; } uhcip->uhci_outst_tds_tail = current_td; current_td->tw = tw; if (tw->tw_hctd_head == NULL) { ASSERT(tw->tw_hctd_tail == NULL); tw->tw_hctd_head = current_td; tw->tw_hctd_tail = current_td; } else { /* Add the td to the end of the list */ tw->tw_hctd_tail->tw_td_next = current_td; tw->tw_hctd_tail = current_td; } mutex_exit(&ph->p_usba_device->usb_mutex); } /* * uhci_alloc_bulk_isoc_tds: * - Allocates the isoc/bulk TD pools. It will allocate one whole * pool to store all the TDs if the system allows. Only when the * first allocation fails, it tries to allocate several small * pools with each pool limited in physical page size. */ static int uhci_alloc_bulk_isoc_tds( uhci_state_t *uhcip, uint_t num_tds, uhci_bulk_isoc_xfer_t *info) { USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_alloc_bulk_isoc_tds: num_tds: 0x%x info: 0x%p", num_tds, (void *)info); info->num_pools = 1; /* allocate as a whole pool at the first time */ if (uhci_alloc_memory_for_tds(uhcip, num_tds, info) != USB_SUCCESS) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "alloc_memory_for_tds failed: num_tds %d num_pools %d", num_tds, info->num_pools); /* reduce the td number per pool and alloc again */ info->num_pools = num_tds / UHCI_MAX_TD_NUM_PER_POOL; if (num_tds % UHCI_MAX_TD_NUM_PER_POOL) { info->num_pools++; } if (uhci_alloc_memory_for_tds(uhcip, num_tds, info) != USB_SUCCESS) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "alloc_memory_for_tds failed: num_tds %d " "num_pools %d", num_tds, info->num_pools); return (USB_NO_RESOURCES); } } return (USB_SUCCESS); } /* * uhci_alloc_memory_for_tds: * - Allocates memory for the isoc/bulk td pools. */ static int uhci_alloc_memory_for_tds( uhci_state_t *uhcip, uint_t num_tds, uhci_bulk_isoc_xfer_t *info) { int result, i, j, err; size_t real_length; uint_t ccount, num; ddi_device_acc_attr_t dev_attr; uhci_bulk_isoc_td_pool_t *td_pool_ptr1, *td_pool_ptr2; USB_DPRINTF_L4(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_alloc_memory_for_tds: num_tds: 0x%x info: 0x%p " "num_pools: %u", num_tds, (void *)info, info->num_pools); /* The host controller will be little endian */ dev_attr.devacc_attr_version = DDI_DEVICE_ATTR_V0; dev_attr.devacc_attr_endian_flags = DDI_STRUCTURE_LE_ACC; dev_attr.devacc_attr_dataorder = DDI_STRICTORDER_ACC; /* Allocate the TD pool structures */ if ((info->td_pools = kmem_zalloc( (sizeof (uhci_bulk_isoc_td_pool_t) * info->num_pools), KM_SLEEP)) == NULL) { USB_DPRINTF_L2(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_alloc_memory_for_tds: alloc td_pools failed"); return (USB_FAILURE); } for (i = 0; i < info->num_pools; i++) { if (info->num_pools == 1) { num = num_tds; } else if (i < (info->num_pools - 1)) { num = UHCI_MAX_TD_NUM_PER_POOL; } else { num = (num_tds % UHCI_MAX_TD_NUM_PER_POOL); } td_pool_ptr1 = &info->td_pools[i]; /* Allocate the bulk TD pool DMA handle */ if (ddi_dma_alloc_handle(uhcip->uhci_dip, &uhcip->uhci_dma_attr, DDI_DMA_SLEEP, 0, &td_pool_ptr1->dma_handle) != DDI_SUCCESS) { for (j = 0; j < i; j++) { td_pool_ptr2 = &info->td_pools[j]; result = ddi_dma_unbind_handle( td_pool_ptr2->dma_handle); ASSERT(result == DDI_SUCCESS); ddi_dma_mem_free(&td_pool_ptr2->mem_handle); ddi_dma_free_handle(&td_pool_ptr2->dma_handle); } kmem_free(info->td_pools, (sizeof (uhci_bulk_isoc_td_pool_t) * info->num_pools)); return (USB_FAILURE); } /* Allocate the memory for the bulk TD pool */ if (ddi_dma_mem_alloc(td_pool_ptr1->dma_handle, num * sizeof (uhci_td_t), &dev_attr, DDI_DMA_CONSISTENT, DDI_DMA_SLEEP, 0, &td_pool_ptr1->pool_addr, &real_length, &td_pool_ptr1->mem_handle) != DDI_SUCCESS) { ddi_dma_free_handle(&td_pool_ptr1->dma_handle); for (j = 0; j < i; j++) { td_pool_ptr2 = &info->td_pools[j]; result = ddi_dma_unbind_handle( td_pool_ptr2->dma_handle); ASSERT(result == DDI_SUCCESS); ddi_dma_mem_free(&td_pool_ptr2->mem_handle); ddi_dma_free_handle(&td_pool_ptr2->dma_handle); } kmem_free(info->td_pools, (sizeof (uhci_bulk_isoc_td_pool_t) * info->num_pools)); return (USB_FAILURE); } /* Map the bulk TD pool into the I/O address space */ result = ddi_dma_addr_bind_handle(td_pool_ptr1->dma_handle, NULL, (caddr_t)td_pool_ptr1->pool_addr, real_length, DDI_DMA_RDWR | DDI_DMA_CONSISTENT, DDI_DMA_SLEEP, NULL, &td_pool_ptr1->cookie, &ccount); /* Process the result */ err = USB_SUCCESS; if (result != DDI_DMA_MAPPED) { USB_DPRINTF_L2(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_allocate_memory_for_tds: Result = %d", result); uhci_decode_ddi_dma_addr_bind_handle_result(uhcip, result); err = USB_FAILURE; } if ((result == DDI_DMA_MAPPED) && (ccount != 1)) { /* The cookie count should be 1 */ USB_DPRINTF_L2(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_allocate_memory_for_tds: " "More than 1 cookie"); result = ddi_dma_unbind_handle( td_pool_ptr1->dma_handle); ASSERT(result == DDI_SUCCESS); err = USB_FAILURE; } if (err == USB_FAILURE) { ddi_dma_mem_free(&td_pool_ptr1->mem_handle); ddi_dma_free_handle(&td_pool_ptr1->dma_handle); for (j = 0; j < i; j++) { td_pool_ptr2 = &info->td_pools[j]; result = ddi_dma_unbind_handle( td_pool_ptr2->dma_handle); ASSERT(result == DDI_SUCCESS); ddi_dma_mem_free(&td_pool_ptr2->mem_handle); ddi_dma_free_handle(&td_pool_ptr2->dma_handle); } kmem_free(info->td_pools, (sizeof (uhci_bulk_isoc_td_pool_t) * info->num_pools)); return (USB_FAILURE); } bzero((void *)td_pool_ptr1->pool_addr, num * sizeof (uhci_td_t)); td_pool_ptr1->num_tds = (ushort_t)num; } return (USB_SUCCESS); } /* * uhci_handle_bulk_td: * * Handles the completed bulk transfer descriptors */ void uhci_handle_bulk_td(uhci_state_t *uhcip, uhci_td_t *td) { uint_t num_bulk_tds, index, td_count, j; usb_cr_t error; uint_t length, bytes_xfered; ushort_t MaxPacketSize; uint32_t buf_offs, paddr; uhci_td_t *bulk_td_ptr, *current_dummy, *td_head; uhci_td_t *tmp_td; queue_head_t *qh, *next_qh; uhci_trans_wrapper_t *tw = td->tw; uhci_pipe_private_t *pp = tw->tw_pipe_private; uhci_bulk_isoc_xfer_t *bulk_xfer_info; uhci_bulk_isoc_td_pool_t *td_pool_ptr; usba_pipe_handle_data_t *ph; USB_DPRINTF_L4(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_handle_bulk_td: td = 0x%p tw = 0x%p", (void *)td, (void *)tw); /* * Update the tw_bytes_pending, and tw_bytes_xfered */ bytes_xfered = ZERO_LENGTH; /* * Check whether there are any errors occurred in the xfer. * If so, update the data_toggle for the queue head and * return error to the upper layer. */ if (GetTD_status(uhcip, td) & TD_STATUS_MASK) { uhci_handle_bulk_td_errors(uhcip, td); USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_handle_bulk_td: error; data toggle: 0x%x", pp->pp_data_toggle); return; } /* * Update the tw_bytes_pending, and tw_bytes_xfered */ bytes_xfered = GetTD_alen(uhcip, td); if (bytes_xfered != ZERO_LENGTH) { tw->tw_bytes_pending -= (bytes_xfered + 1); tw->tw_bytes_xfered += (bytes_xfered + 1); } /* * Get Bulk pipe information and pipe handle */ bulk_xfer_info = pp->pp_qh->bulk_xfer_info; ph = tw->tw_pipe_private->pp_pipe_handle; /* * Check whether data underrun occurred. * If so, complete the transfer * Update the data toggle bit */ if (bytes_xfered != GetTD_mlen(uhcip, td)) { bulk_xfer_info->num_tds = 1; USB_DPRINTF_L2(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_handle_bulk_td: Data underrun occured"); pp->pp_data_toggle = GetTD_dtogg(uhcip, td) == 0 ? 1 : 0; } /* * If the TD's in the current frame are completed, then check * whether we have any more bytes to xfer. If so, insert TD's. * If no more bytes needs to be transferred, then do callback to the * upper layer. * If the TD's in the current frame are not completed, then * just delete the TD from the linked lists. */ USB_DPRINTF_L3(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_handle_bulk_td: completed TD data toggle: 0x%x", GetTD_dtogg(uhcip, td)); if (--bulk_xfer_info->num_tds == 0) { uhci_delete_td(uhcip, td); if ((tw->tw_bytes_pending) && (GetTD_mlen(uhcip, td) - GetTD_alen(uhcip, td) == 0)) { MaxPacketSize = pp->pp_pipe_handle->p_ep.wMaxPacketSize; length = MaxPacketSize; qh = pp->pp_qh; paddr = GetQH32(uhcip, qh->link_ptr) & QH_LINK_PTR_MASK; if (GetQH32(uhcip, qh->link_ptr) != GetQH32(uhcip, uhcip->uhci_bulk_xfers_q_head->link_ptr)) { next_qh = QH_VADDR(paddr); SetQH32(uhcip, qh->prev_qh->link_ptr, paddr|(0x2)); next_qh->prev_qh = qh->prev_qh; SetQH32(uhcip, qh->link_ptr, GetQH32(uhcip, uhcip->uhci_bulk_xfers_q_head->link_ptr)); qh->prev_qh = uhcip->uhci_bulk_xfers_q_tail; SetQH32(uhcip, uhcip->uhci_bulk_xfers_q_tail->link_ptr, QH_PADDR(qh) | 0x2); uhcip->uhci_bulk_xfers_q_tail = qh; } if ((tw->tw_bytes_pending / MaxPacketSize) >= MAX_NUM_BULK_TDS_PER_XFER) { num_bulk_tds = MAX_NUM_BULK_TDS_PER_XFER; } else { num_bulk_tds = (tw->tw_bytes_pending / MaxPacketSize); if (tw->tw_bytes_pending % MaxPacketSize) { num_bulk_tds++; length = (tw->tw_bytes_pending % MaxPacketSize); } } current_dummy = pp->pp_qh->td_tailp; td_pool_ptr = &bulk_xfer_info->td_pools[0]; bulk_td_ptr = (uhci_td_t *)td_pool_ptr->pool_addr; buf_offs = tw->tw_bytes_xfered; td_count = num_bulk_tds; index = 0; /* reuse the TDs to transfer more data */ while (td_count > 0) { for (j = 0; (j < (td_pool_ptr->num_tds - 1)) && (td_count > 1); j++, td_count--) { uhci_fill_in_bulk_isoc_td(uhcip, &bulk_td_ptr[j], &bulk_td_ptr[j+1], BULKTD_PADDR(td_pool_ptr, &bulk_td_ptr[j+1]), ph, buf_offs, MaxPacketSize, tw); buf_offs += MaxPacketSize; } if (td_count == 1) { uhci_fill_in_bulk_isoc_td(uhcip, &bulk_td_ptr[j], current_dummy, TD_PADDR(current_dummy), ph, buf_offs, length, tw); break; } else { tmp_td = &bulk_td_ptr[j]; ASSERT(index < (bulk_xfer_info->num_pools - 1)); td_pool_ptr = &bulk_xfer_info-> td_pools[index + 1]; bulk_td_ptr = (uhci_td_t *) td_pool_ptr->pool_addr; uhci_fill_in_bulk_isoc_td(uhcip, tmp_td, &bulk_td_ptr[0], BULKTD_PADDR(td_pool_ptr, &bulk_td_ptr[0]), ph, buf_offs, MaxPacketSize, tw); buf_offs += MaxPacketSize; td_count--; index++; } } pp->pp_qh->bulk_xfer_info = bulk_xfer_info; bulk_xfer_info->num_tds = (ushort_t)num_bulk_tds; SetQH32(uhcip, pp->pp_qh->element_ptr, bulk_xfer_info->td_pools[0].cookie.dmac_address); } else { usba_pipe_handle_data_t *usb_pp = pp->pp_pipe_handle; pp->pp_qh->bulk_xfer_info = NULL; if (tw->tw_bytes_pending) { /* Update the element pointer */ SetQH32(uhcip, pp->pp_qh->element_ptr, TD_PADDR(pp->pp_qh->td_tailp)); /* Remove all the tds */ td_head = tw->tw_hctd_head; while (td_head != NULL) { uhci_delete_td(uhcip, td_head); td_head = tw->tw_hctd_head; } } if (tw->tw_direction == PID_IN) { usb_req_attrs_t attrs = ((usb_bulk_req_t *) tw->tw_curr_xfer_reqp)->bulk_attributes; error = USB_CR_OK; /* Data run occurred */ if (tw->tw_bytes_pending && (!(attrs & USB_ATTRS_SHORT_XFER_OK))) { error = USB_CR_DATA_UNDERRUN; } uhci_sendup_td_message(uhcip, error, tw); } else { uhci_do_byte_stats(uhcip, tw->tw_length, usb_pp->p_ep.bmAttributes, usb_pp->p_ep.bEndpointAddress); /* Data underrun occurred */ if (tw->tw_bytes_pending) { tw->tw_data->b_rptr += tw->tw_bytes_xfered; USB_DPRINTF_L2(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_handle_bulk_td: " "data underrun occurred"); uhci_hcdi_callback(uhcip, pp, tw->tw_pipe_private->pp_pipe_handle, tw, USB_CR_DATA_UNDERRUN); } else { uhci_hcdi_callback(uhcip, pp, tw->tw_pipe_private->pp_pipe_handle, tw, USB_CR_OK); } } /* direction */ /* Deallocate DMA memory */ uhci_deallocate_tw(uhcip, pp, tw); for (j = 0; j < bulk_xfer_info->num_pools; j++) { td_pool_ptr = &bulk_xfer_info->td_pools[j]; (void) ddi_dma_unbind_handle( td_pool_ptr->dma_handle); ddi_dma_mem_free(&td_pool_ptr->mem_handle); ddi_dma_free_handle(&td_pool_ptr->dma_handle); } kmem_free(bulk_xfer_info->td_pools, (sizeof (uhci_bulk_isoc_td_pool_t) * bulk_xfer_info->num_pools)); kmem_free(bulk_xfer_info, sizeof (uhci_bulk_isoc_xfer_t)); /* * When there are no pending bulk commands, point the * end of the lattice tree to NULL. This will make sure * that the HC control does not loop anymore and PCI * bus is not affected. */ if (--uhcip->uhci_pending_bulk_cmds == 0) { uhcip->uhci_bulk_xfers_q_tail->link_ptr = HC_END_OF_LIST; USB_DPRINTF_L3(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_handle_bulk_td: count = %d", uhcip->uhci_pending_bulk_cmds); } } } else { uhci_delete_td(uhcip, td); } } void uhci_handle_bulk_td_errors(uhci_state_t *uhcip, uhci_td_t *td) { usb_cr_t usb_err; uint32_t paddr_tail, element_ptr, paddr; uhci_td_t *next_td; uhci_pipe_private_t *pp; uhci_trans_wrapper_t *tw = td->tw; usba_pipe_handle_data_t *ph; uhci_bulk_isoc_td_pool_t *td_pool_ptr = NULL; USB_DPRINTF_L2(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_handle_bulk_td_errors: td = %p", (void *)td); #ifdef DEBUG uhci_print_td(uhcip, td); #endif tw = td->tw; ph = tw->tw_pipe_private->pp_pipe_handle; pp = (uhci_pipe_private_t *)ph->p_hcd_private; /* * Find the type of error occurred and return the error * to the upper layer. And adjust the data toggle. */ element_ptr = GetQH32(uhcip, pp->pp_qh->element_ptr) & QH_ELEMENT_PTR_MASK; paddr_tail = TD_PADDR(pp->pp_qh->td_tailp); /* * If a timeout occurs before a transfer has completed, * the timeout handler sets the CRC/Timeout bit and clears the Active * bit in the link_ptr for each td in the transfer. * It then waits (at least) 1 ms so that any tds the controller might * have been executing will have completed. * So at this point element_ptr will point to either: * 1) the next td for the transfer (which has not been executed, * and has the CRC/Timeout status bit set and Active bit cleared), * 2) the dummy td for this qh. * So if the element_ptr does not point to the dummy td, we know * it points to the next td that would have been executed. * That td has the data toggle we want to save. * All outstanding tds have been marked as CRC/Timeout, * so it doesn't matter which td we pass to uhci_parse_td_error * for the error status. */ if (element_ptr != paddr_tail) { paddr = (element_ptr & QH_ELEMENT_PTR_MASK); uhci_get_bulk_td_by_paddr(uhcip, pp->pp_qh->bulk_xfer_info, paddr, &td_pool_ptr); next_td = BULKTD_VADDR(td_pool_ptr, paddr); USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_handle_bulk_td_errors: next td = %p", (void *)next_td); usb_err = uhci_parse_td_error(uhcip, pp, next_td); } else { usb_err = uhci_parse_td_error(uhcip, pp, td); } /* * Update the link pointer. */ SetQH32(uhcip, pp->pp_qh->element_ptr, TD_PADDR(pp->pp_qh->td_tailp)); /* * Send up number of bytes transferred before the error condition. */ if ((tw->tw_direction == PID_OUT) && tw->tw_data) { tw->tw_data->b_rptr += tw->tw_bytes_xfered; } uhci_remove_bulk_tds_tws(uhcip, tw->tw_pipe_private, UHCI_IN_ERROR); /* * When there are no pending bulk commands, point the end of the * lattice tree to NULL. This will make sure that the HC control * does not loop anymore and PCI bus is not affected. */ if (--uhcip->uhci_pending_bulk_cmds == 0) { uhcip->uhci_bulk_xfers_q_tail->link_ptr = HC_END_OF_LIST; USB_DPRINTF_L3(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_handle_bulk_td_errors: count = %d", uhcip->uhci_pending_bulk_cmds); } uhci_hcdi_callback(uhcip, pp, ph, tw, usb_err); uhci_deallocate_tw(uhcip, pp, tw); } /* * uhci_get_bulk_td_by_paddr: * Obtain the address of the TD pool the physical address falls in. * * td_pool_pp - pointer to the address of the TD pool containing the paddr */ /* ARGSUSED */ static void uhci_get_bulk_td_by_paddr( uhci_state_t *uhcip, uhci_bulk_isoc_xfer_t *info, uint32_t paddr, uhci_bulk_isoc_td_pool_t **td_pool_pp) { uint_t i = 0; while (i < info->num_pools) { *td_pool_pp = &info->td_pools[i]; if (((*td_pool_pp)->cookie.dmac_address <= paddr) && (((*td_pool_pp)->cookie.dmac_address + (*td_pool_pp)->cookie.dmac_size) > paddr)) { break; } i++; } ASSERT(i < info->num_pools); } void uhci_remove_bulk_tds_tws( uhci_state_t *uhcip, uhci_pipe_private_t *pp, int what) { uint_t rval, i; uhci_td_t *head; uhci_td_t *head_next; usb_opaque_t curr_reqp; uhci_bulk_isoc_xfer_t *info; uhci_bulk_isoc_td_pool_t *td_pool_ptr; ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); if ((info = pp->pp_qh->bulk_xfer_info) == NULL) { return; } head = uhcip->uhci_outst_tds_head; while (head) { uhci_trans_wrapper_t *tw_tmp = head->tw; head_next = head->outst_td_next; if (pp->pp_qh == tw_tmp->tw_pipe_private->pp_qh) { curr_reqp = tw_tmp->tw_curr_xfer_reqp; if (curr_reqp && ((what == UHCI_IN_CLOSE) || (what == UHCI_IN_RESET))) { uhci_hcdi_callback(uhcip, pp, pp->pp_pipe_handle, tw_tmp, USB_CR_FLUSHED); } /* end of curr_reqp */ uhci_delete_td(uhcip, head); if (what == UHCI_IN_CLOSE || what == UHCI_IN_RESET) { ASSERT(info->num_tds > 0); if (--info->num_tds == 0) { uhci_deallocate_tw(uhcip, pp, tw_tmp); /* * This will make sure that the HC * does not loop anymore when there * are no pending bulk commands. */ if (--uhcip->uhci_pending_bulk_cmds == 0) { uhcip->uhci_bulk_xfers_q_tail-> link_ptr = HC_END_OF_LIST; USB_DPRINTF_L3(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "uhci_remove_bulk_tds_tws:" " count = %d", uhcip-> uhci_pending_bulk_cmds); } } } } head = head_next; } if (what == UHCI_IN_CLOSE || what == UHCI_IN_RESET) { ASSERT(info->num_tds == 0); } for (i = 0; i < info->num_pools; i++) { td_pool_ptr = &info->td_pools[i]; rval = ddi_dma_unbind_handle(td_pool_ptr->dma_handle); ASSERT(rval == DDI_SUCCESS); ddi_dma_mem_free(&td_pool_ptr->mem_handle); ddi_dma_free_handle(&td_pool_ptr->dma_handle); } kmem_free(info->td_pools, (sizeof (uhci_bulk_isoc_td_pool_t) * info->num_pools)); kmem_free(info, sizeof (uhci_bulk_isoc_xfer_t)); pp->pp_qh->bulk_xfer_info = NULL; } /* * uhci_save_data_toggle () * Save the data toggle in the usba_device structure */ void uhci_save_data_toggle(uhci_pipe_private_t *pp) { usba_pipe_handle_data_t *ph = pp->pp_pipe_handle; /* Save the data toggle in the usb devices structure. */ mutex_enter(&ph->p_mutex); usba_hcdi_set_data_toggle(ph->p_usba_device, ph->p_ep.bEndpointAddress, pp->pp_data_toggle); mutex_exit(&ph->p_mutex); } /* * uhci_create_isoc_transfer_wrapper: * Create a Transaction Wrapper (TW) for isoc transfer. * This involves the allocating of DMA resources. * * For isoc transfers, one isoc transfer includes multiple packets * and each packet may have a different length. So each packet is * transfered by one TD. We only know the individual packet length * won't exceed 1023 bytes, but we don't know exactly the lengths. * It is hard to make one physically discontiguous DMA buffer which * can fit in all the TDs like what can be done to the ctrl/bulk/ * intr transfers. It is also undesirable to make one physically * contiguous DMA buffer for all the packets, since this may easily * fail when the system is in low memory. So an individual DMA * buffer is allocated for an individual isoc packet and each DMA * buffer is physically contiguous. An extra structure is allocated * to save the multiple DMA handles. */ static uhci_trans_wrapper_t * uhci_create_isoc_transfer_wrapper( uhci_state_t *uhcip, uhci_pipe_private_t *pp, usb_isoc_req_t *req, size_t length, usb_flags_t usb_flags) { int result; size_t real_length, strtlen, xfer_size; uhci_trans_wrapper_t *tw; ddi_device_acc_attr_t dev_attr; ddi_dma_attr_t dma_attr; int kmem_flag; int (*dmamem_wait)(caddr_t); uint_t i, j, ccount; usb_isoc_req_t *tmp_req = req; ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); if (UHCI_XFER_TYPE(&pp->pp_pipe_handle->p_ep) != USB_EP_ATTR_ISOCH) { return (NULL); } if ((req == NULL) && (UHCI_XFER_DIR(&pp->pp_pipe_handle->p_ep) == USB_EP_DIR_IN)) { tmp_req = (usb_isoc_req_t *)pp->pp_client_periodic_in_reqp; } if (tmp_req == NULL) { return (NULL); } USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_create_isoc_transfer_wrapper: length = 0x%lx flags = 0x%x", length, usb_flags); /* SLEEP flag should not be used in interrupt context */ if (servicing_interrupt()) { kmem_flag = KM_NOSLEEP; dmamem_wait = DDI_DMA_DONTWAIT; } else { kmem_flag = KM_SLEEP; dmamem_wait = DDI_DMA_SLEEP; } /* Allocate space for the transfer wrapper */ if ((tw = kmem_zalloc(sizeof (uhci_trans_wrapper_t), kmem_flag)) == NULL) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_create_isoc_transfer_wrapper: kmem_alloc failed"); return (NULL); } /* Allocate space for the isoc buffer handles */ strtlen = sizeof (uhci_isoc_buf_t) * tmp_req->isoc_pkts_count; if ((tw->tw_isoc_bufs = kmem_zalloc(strtlen, kmem_flag)) == NULL) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_create_isoc_transfer_wrapper: kmem_alloc " "isoc buffer failed"); kmem_free(tw, sizeof (uhci_trans_wrapper_t)); return (NULL); } bcopy(&uhcip->uhci_dma_attr, &dma_attr, sizeof (ddi_dma_attr_t)); dma_attr.dma_attr_sgllen = 1; dev_attr.devacc_attr_version = DDI_DEVICE_ATTR_V0; dev_attr.devacc_attr_endian_flags = DDI_STRUCTURE_LE_ACC; dev_attr.devacc_attr_dataorder = DDI_STRICTORDER_ACC; /* Store the transfer length */ tw->tw_length = length; for (i = 0; i < tmp_req->isoc_pkts_count; i++) { tw->tw_isoc_bufs[i].index = (ushort_t)i; /* Allocate the DMA handle */ if ((result = ddi_dma_alloc_handle(uhcip->uhci_dip, &dma_attr, dmamem_wait, 0, &tw->tw_isoc_bufs[i].dma_handle)) != DDI_SUCCESS) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_create_isoc_transfer_wrapper: " "Alloc handle %d failed", i); for (j = 0; j < i; j++) { result = ddi_dma_unbind_handle( tw->tw_isoc_bufs[j].dma_handle); ASSERT(result == USB_SUCCESS); ddi_dma_mem_free(&tw->tw_isoc_bufs[j]. mem_handle); ddi_dma_free_handle(&tw->tw_isoc_bufs[j]. dma_handle); } kmem_free(tw->tw_isoc_bufs, strtlen); kmem_free(tw, sizeof (uhci_trans_wrapper_t)); return (NULL); } /* Allocate the memory */ xfer_size = tmp_req->isoc_pkt_descr[i].isoc_pkt_length; if ((result = ddi_dma_mem_alloc(tw->tw_isoc_bufs[i].dma_handle, xfer_size, &dev_attr, DDI_DMA_CONSISTENT, dmamem_wait, NULL, (caddr_t *)&tw->tw_isoc_bufs[i].buf_addr, &real_length, &tw->tw_isoc_bufs[i].mem_handle)) != DDI_SUCCESS) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_create_isoc_transfer_wrapper: " "dma_mem_alloc %d fail", i); ddi_dma_free_handle(&tw->tw_isoc_bufs[i].dma_handle); for (j = 0; j < i; j++) { result = ddi_dma_unbind_handle( tw->tw_isoc_bufs[j].dma_handle); ASSERT(result == USB_SUCCESS); ddi_dma_mem_free(&tw->tw_isoc_bufs[j]. mem_handle); ddi_dma_free_handle(&tw->tw_isoc_bufs[j]. dma_handle); } kmem_free(tw->tw_isoc_bufs, strtlen); kmem_free(tw, sizeof (uhci_trans_wrapper_t)); return (NULL); } ASSERT(real_length >= xfer_size); /* Bind the handle */ result = ddi_dma_addr_bind_handle( tw->tw_isoc_bufs[i].dma_handle, NULL, (caddr_t)tw->tw_isoc_bufs[i].buf_addr, real_length, DDI_DMA_RDWR|DDI_DMA_CONSISTENT, dmamem_wait, NULL, &tw->tw_isoc_bufs[i].cookie, &ccount); if ((result == DDI_DMA_MAPPED) && (ccount == 1)) { tw->tw_isoc_bufs[i].length = xfer_size; continue; } else { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_create_isoc_transfer_wrapper: " "Bind handle %d failed", i); if (result == DDI_DMA_MAPPED) { result = ddi_dma_unbind_handle( tw->tw_isoc_bufs[i].dma_handle); ASSERT(result == USB_SUCCESS); } ddi_dma_mem_free(&tw->tw_isoc_bufs[i].mem_handle); ddi_dma_free_handle(&tw->tw_isoc_bufs[i].dma_handle); for (j = 0; j < i; j++) { result = ddi_dma_unbind_handle( tw->tw_isoc_bufs[j].dma_handle); ASSERT(result == USB_SUCCESS); ddi_dma_mem_free(&tw->tw_isoc_bufs[j]. mem_handle); ddi_dma_free_handle(&tw->tw_isoc_bufs[j]. dma_handle); } kmem_free(tw->tw_isoc_bufs, strtlen); kmem_free(tw, sizeof (uhci_trans_wrapper_t)); return (NULL); } } tw->tw_ncookies = tmp_req->isoc_pkts_count; tw->tw_isoc_strtlen = strtlen; /* * Only allow one wrapper to be added at a time. Insert the * new transaction wrapper into the list for this pipe. */ if (pp->pp_tw_head == NULL) { pp->pp_tw_head = tw; pp->pp_tw_tail = tw; } else { pp->pp_tw_tail->tw_next = tw; pp->pp_tw_tail = tw; ASSERT(tw->tw_next == NULL); } /* Store a back pointer to the pipe private structure */ tw->tw_pipe_private = pp; /* Store the transfer type - synchronous or asynchronous */ tw->tw_flags = usb_flags; USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_create_isoc_transfer_wrapper: tw = 0x%p, ncookies = %u", (void *)tw, tw->tw_ncookies); return (tw); } /* * uhci_insert_isoc_td: * - Create transfer wrapper * - Allocate memory for the isoc td's * - Fill up all the TD's and submit to the HC * - Update all the linked lists */ int uhci_insert_isoc_td( uhci_state_t *uhcip, usba_pipe_handle_data_t *ph, usb_isoc_req_t *isoc_req, size_t length, usb_flags_t flags) { int rval = USB_SUCCESS; int error; uint_t ddic; uint32_t i, j, index; uint32_t bytes_to_xfer; uint32_t expired_frames = 0; usb_frame_number_t start_frame, end_frame, current_frame; uhci_td_t *td_ptr; uhci_pipe_private_t *pp = (uhci_pipe_private_t *)ph->p_hcd_private; uhci_trans_wrapper_t *tw; uhci_bulk_isoc_xfer_t *isoc_xfer_info; uhci_bulk_isoc_td_pool_t *td_pool_ptr; USB_DPRINTF_L4(PRINT_MASK_ISOC, uhcip->uhci_log_hdl, "uhci_insert_isoc_td: ph = 0x%p isoc req = %p length = %lu", (void *)ph, (void *)isoc_req, length); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); /* Allocate a transfer wrapper */ if ((tw = uhci_create_isoc_transfer_wrapper(uhcip, pp, isoc_req, length, flags)) == NULL) { USB_DPRINTF_L2(PRINT_MASK_ISOC, uhcip->uhci_log_hdl, "uhci_insert_isoc_td: TW allocation failed"); return (USB_NO_RESOURCES); } /* Save current isochronous request pointer */ tw->tw_curr_xfer_reqp = (usb_opaque_t)isoc_req; /* * Initialize the transfer wrapper. These values are useful * for sending back the reply. */ tw->tw_handle_td = uhci_handle_isoc_td; tw->tw_handle_callback_value = NULL; tw->tw_direction = (UHCI_XFER_DIR(&ph->p_ep) == USB_EP_DIR_OUT) ? PID_OUT : PID_IN; /* * If the transfer isoc send, then copy the data from the request * to the transfer wrapper. */ if ((tw->tw_direction == PID_OUT) && length) { uchar_t *p; ASSERT(isoc_req->isoc_data != NULL); p = isoc_req->isoc_data->b_rptr; /* Copy the data into the message */ for (i = 0; i < isoc_req->isoc_pkts_count; i++) { ddi_rep_put8(tw->tw_isoc_bufs[i].mem_handle, p, (uint8_t *)tw->tw_isoc_bufs[i].buf_addr, isoc_req->isoc_pkt_descr[i].isoc_pkt_length, DDI_DEV_AUTOINCR); p += isoc_req->isoc_pkt_descr[i].isoc_pkt_length; } } if (tw->tw_direction == PID_IN) { if ((rval = uhci_allocate_periodic_in_resource(uhcip, pp, tw, flags)) != USB_SUCCESS) { USB_DPRINTF_L2(PRINT_MASK_ISOC, uhcip->uhci_log_hdl, "uhci_insert_isoc_td: isoc_req_t alloc failed"); uhci_deallocate_tw(uhcip, pp, tw); return (rval); } isoc_req = (usb_isoc_req_t *)tw->tw_curr_xfer_reqp; } tw->tw_isoc_req = (usb_isoc_req_t *)tw->tw_curr_xfer_reqp; /* Get the pointer to the isoc_xfer_info structure */ isoc_xfer_info = (uhci_bulk_isoc_xfer_t *)&tw->tw_xfer_info; isoc_xfer_info->num_tds = isoc_req->isoc_pkts_count; /* * Allocate memory for isoc tds */ if ((rval = uhci_alloc_bulk_isoc_tds(uhcip, isoc_req->isoc_pkts_count, isoc_xfer_info)) != USB_SUCCESS) { USB_DPRINTF_L2(PRINT_MASK_ISOC, uhcip->uhci_log_hdl, "uhci_alloc_bulk_isoc_td: Memory allocation failure"); if (tw->tw_direction == PID_IN) { uhci_deallocate_periodic_in_resource(uhcip, pp, tw); } uhci_deallocate_tw(uhcip, pp, tw); return (rval); } /* * Get the isoc td pool address, buffer address and * max packet size that the device supports. */ td_pool_ptr = &isoc_xfer_info->td_pools[0]; td_ptr = (uhci_td_t *)td_pool_ptr->pool_addr; index = 0; /* * Fill up the isoc tds */ USB_DPRINTF_L3(PRINT_MASK_ISOC, uhcip->uhci_log_hdl, "uhci_insert_isoc_td : isoc pkts %d", isoc_req->isoc_pkts_count); for (i = 0; i < isoc_xfer_info->num_pools; i++) { for (j = 0; j < td_pool_ptr->num_tds; j++) { bytes_to_xfer = isoc_req->isoc_pkt_descr[index].isoc_pkt_length; uhci_fill_in_bulk_isoc_td(uhcip, &td_ptr[j], (uhci_td_t *)NULL, HC_END_OF_LIST, ph, index, bytes_to_xfer, tw); td_ptr[j].isoc_pkt_index = (ushort_t)index; index++; } if (i < (isoc_xfer_info->num_pools - 1)) { td_pool_ptr = &isoc_xfer_info->td_pools[i + 1]; td_ptr = (uhci_td_t *)td_pool_ptr->pool_addr; } } /* * Get the starting frame number. * The client drivers sets the flag USB_ATTRS_ISOC_XFER_ASAP to inform * the HCD to care of starting frame number. * * Following code is very time critical. So, perform atomic execution. */ ddic = ddi_enter_critical(); current_frame = uhci_get_sw_frame_number(uhcip); if (isoc_req->isoc_attributes & USB_ATTRS_ISOC_START_FRAME) { start_frame = isoc_req->isoc_frame_no; end_frame = start_frame + isoc_req->isoc_pkts_count; /* Check available frames */ if ((end_frame - current_frame) < UHCI_MAX_ISOC_FRAMES) { if (current_frame > start_frame) { if ((current_frame + FRNUM_OFFSET) < end_frame) { expired_frames = current_frame + FRNUM_OFFSET - start_frame; start_frame = current_frame + FRNUM_OFFSET; } else { rval = USB_INVALID_START_FRAME; } } } else { rval = USB_INVALID_START_FRAME; } } else if (isoc_req->isoc_attributes & USB_ATTRS_ISOC_XFER_ASAP) { start_frame = pp->pp_frame_num; if (start_frame == INVALID_FRNUM) { start_frame = current_frame + FRNUM_OFFSET; } else if (current_frame > start_frame) { start_frame = current_frame + FRNUM_OFFSET; } end_frame = start_frame + isoc_req->isoc_pkts_count; isoc_req->isoc_frame_no = start_frame; } if (rval != USB_SUCCESS) { /* Exit the critical */ ddi_exit_critical(ddic); USB_DPRINTF_L2(PRINT_MASK_ISOC, uhcip->uhci_log_hdl, "uhci_insert_isoc_td: Invalid starting frame number"); if (tw->tw_direction == PID_IN) { uhci_deallocate_periodic_in_resource(uhcip, pp, tw); } while (tw->tw_hctd_head) { uhci_delete_td(uhcip, tw->tw_hctd_head); } for (i = 0; i < isoc_xfer_info->num_pools; i++) { td_pool_ptr = &isoc_xfer_info->td_pools[i]; error = ddi_dma_unbind_handle(td_pool_ptr->dma_handle); ASSERT(error == DDI_SUCCESS); ddi_dma_mem_free(&td_pool_ptr->mem_handle); ddi_dma_free_handle(&td_pool_ptr->dma_handle); } kmem_free(isoc_xfer_info->td_pools, (sizeof (uhci_bulk_isoc_td_pool_t) * isoc_xfer_info->num_pools)); uhci_deallocate_tw(uhcip, pp, tw); return (rval); } for (i = 0; i < expired_frames; i++) { isoc_req->isoc_pkt_descr[i].isoc_pkt_status = USB_CR_NOT_ACCESSED; isoc_req->isoc_pkt_descr[i].isoc_pkt_actual_length = isoc_req->isoc_pkt_descr[i].isoc_pkt_length; uhci_get_isoc_td_by_index(uhcip, isoc_xfer_info, i, &td_ptr, &td_pool_ptr); uhci_delete_td(uhcip, td_ptr); --isoc_xfer_info->num_tds; } /* * Add the TD's to the HC list */ start_frame = (start_frame & 0x3ff); for (; i < isoc_req->isoc_pkts_count; i++) { uhci_get_isoc_td_by_index(uhcip, isoc_xfer_info, i, &td_ptr, &td_pool_ptr); if (uhcip->uhci_isoc_q_tailp[start_frame]) { td_ptr->isoc_prev = uhcip->uhci_isoc_q_tailp[start_frame]; td_ptr->isoc_next = NULL; td_ptr->link_ptr = uhcip->uhci_isoc_q_tailp[start_frame]->link_ptr; uhcip->uhci_isoc_q_tailp[start_frame]->isoc_next = td_ptr; SetTD32(uhcip, uhcip->uhci_isoc_q_tailp[start_frame]->link_ptr, ISOCTD_PADDR(td_pool_ptr, td_ptr)); uhcip->uhci_isoc_q_tailp[start_frame] = td_ptr; } else { uhcip->uhci_isoc_q_tailp[start_frame] = td_ptr; td_ptr->isoc_next = NULL; td_ptr->isoc_prev = NULL; SetTD32(uhcip, td_ptr->link_ptr, GetFL32(uhcip, uhcip->uhci_frame_lst_tablep[start_frame])); SetFL32(uhcip, uhcip->uhci_frame_lst_tablep[start_frame], ISOCTD_PADDR(td_pool_ptr, td_ptr)); } td_ptr->starting_frame = (uint_t)start_frame; if (++start_frame == NUM_FRAME_LST_ENTRIES) start_frame = 0; } ddi_exit_critical(ddic); pp->pp_frame_num = end_frame; USB_DPRINTF_L4(PRINT_MASK_ISOC, uhcip->uhci_log_hdl, "uhci_insert_isoc_td: current frame number 0x%llx, pipe frame num" " 0x%llx", (unsigned long long)current_frame, (unsigned long long)(pp->pp_frame_num)); return (rval); } /* * uhci_get_isoc_td_by_index: * Obtain the addresses of the TD pool and the TD at the index. * * tdpp - pointer to the address of the TD at the isoc packet index * td_pool_pp - pointer to the address of the TD pool containing * the specified TD */ /* ARGSUSED */ static void uhci_get_isoc_td_by_index( uhci_state_t *uhcip, uhci_bulk_isoc_xfer_t *info, uint_t index, uhci_td_t **tdpp, uhci_bulk_isoc_td_pool_t **td_pool_pp) { uint_t i = 0, j = 0; uhci_td_t *td_ptr; while (j < info->num_pools) { if ((i + info->td_pools[j].num_tds) <= index) { i += info->td_pools[j].num_tds; j++; } else { i = index - i; break; } } ASSERT(j < info->num_pools); *td_pool_pp = &info->td_pools[j]; td_ptr = (uhci_td_t *)((*td_pool_pp)->pool_addr); *tdpp = &td_ptr[i]; } /* * uhci_handle_isoc_td: * Handles the completed isoc tds */ void uhci_handle_isoc_td(uhci_state_t *uhcip, uhci_td_t *td) { uint_t rval, i; uint32_t pkt_index = td->isoc_pkt_index; usb_cr_t cr; uhci_trans_wrapper_t *tw = td->tw; usb_isoc_req_t *isoc_req = (usb_isoc_req_t *)tw->tw_isoc_req; uhci_pipe_private_t *pp = tw->tw_pipe_private; uhci_bulk_isoc_xfer_t *isoc_xfer_info = &tw->tw_xfer_info; usba_pipe_handle_data_t *usb_pp; uhci_bulk_isoc_td_pool_t *td_pool_ptr; USB_DPRINTF_L4(PRINT_MASK_ISOC, uhcip->uhci_log_hdl, "uhci_handle_isoc_td: td = 0x%p, pp = 0x%p, tw = 0x%p, req = 0x%p, " "index = %x", (void *)td, (void *)pp, (void *)tw, (void *)isoc_req, pkt_index); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); usb_pp = pp->pp_pipe_handle; /* * Check whether there are any errors occurred. If so, update error * count and return it to the upper.But never return a non zero * completion reason. */ cr = USB_CR_OK; if (GetTD_status(uhcip, td) & TD_STATUS_MASK) { USB_DPRINTF_L2(PRINT_MASK_ISOC, uhcip->uhci_log_hdl, "uhci_handle_isoc_td: Error Occurred: TD Status = %x", GetTD_status(uhcip, td)); isoc_req->isoc_error_count++; } if (isoc_req != NULL) { isoc_req->isoc_pkt_descr[pkt_index].isoc_pkt_status = cr; isoc_req->isoc_pkt_descr[pkt_index].isoc_pkt_actual_length = (GetTD_alen(uhcip, td) == ZERO_LENGTH) ? 0 : GetTD_alen(uhcip, td) + 1; } uhci_delete_isoc_td(uhcip, td); if (--isoc_xfer_info->num_tds != 0) { USB_DPRINTF_L3(PRINT_MASK_ISOC, uhcip->uhci_log_hdl, "uhci_handle_isoc_td: Number of TDs %d", isoc_xfer_info->num_tds); return; } tw->tw_claim = UHCI_INTR_HDLR_CLAIMED; if (tw->tw_direction == PID_IN) { uhci_sendup_td_message(uhcip, cr, tw); if ((uhci_handle_isoc_receive(uhcip, pp, tw)) != USB_SUCCESS) { USB_DPRINTF_L3(PRINT_MASK_ISOC, uhcip->uhci_log_hdl, "uhci_handle_isoc_td: Drop message"); } } else { /* update kstats only for OUT. sendup_td_msg() does it for IN */ uhci_do_byte_stats(uhcip, tw->tw_length, usb_pp->p_ep.bmAttributes, usb_pp->p_ep.bEndpointAddress); uhci_hcdi_callback(uhcip, pp, usb_pp, tw, USB_CR_OK); } for (i = 0; i < isoc_xfer_info->num_pools; i++) { td_pool_ptr = &isoc_xfer_info->td_pools[i]; rval = ddi_dma_unbind_handle(td_pool_ptr->dma_handle); ASSERT(rval == DDI_SUCCESS); ddi_dma_mem_free(&td_pool_ptr->mem_handle); ddi_dma_free_handle(&td_pool_ptr->dma_handle); } kmem_free(isoc_xfer_info->td_pools, (sizeof (uhci_bulk_isoc_td_pool_t) * isoc_xfer_info->num_pools)); uhci_deallocate_tw(uhcip, pp, tw); } /* * uhci_handle_isoc_receive: * - Sends the isoc data to the client * - Inserts another isoc receive request */ static int uhci_handle_isoc_receive( uhci_state_t *uhcip, uhci_pipe_private_t *pp, uhci_trans_wrapper_t *tw) { USB_DPRINTF_L4(PRINT_MASK_ISOC, uhcip->uhci_log_hdl, "uhci_handle_isoc_receive: tw = 0x%p", (void *)tw); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); /* * -- check for pipe state being polling before * inserting a new request. Check when is TD * de-allocation being done? (so we can reuse the same TD) */ if (uhci_start_isoc_receive_polling(uhcip, pp->pp_pipe_handle, (usb_isoc_req_t *)tw->tw_curr_xfer_reqp, 0) != USB_SUCCESS) { USB_DPRINTF_L2(PRINT_MASK_ISOC, uhcip->uhci_log_hdl, "uhci_handle_isoc_receive: receive polling failed"); return (USB_FAILURE); } return (USB_SUCCESS); } /* * uhci_delete_isoc_td: * - Delete from the outstanding command queue * - Delete from the tw queue * - Delete from the isoc queue * - Delete from the HOST CONTROLLER list */ static void uhci_delete_isoc_td(uhci_state_t *uhcip, uhci_td_t *td) { uint32_t starting_frame = td->starting_frame; if ((td->isoc_next == NULL) && (td->isoc_prev == NULL)) { SetFL32(uhcip, uhcip->uhci_frame_lst_tablep[starting_frame], GetTD32(uhcip, td->link_ptr)); uhcip->uhci_isoc_q_tailp[starting_frame] = 0; } else if (td->isoc_next == NULL) { td->isoc_prev->link_ptr = td->link_ptr; td->isoc_prev->isoc_next = NULL; uhcip->uhci_isoc_q_tailp[starting_frame] = td->isoc_prev; } else if (td->isoc_prev == NULL) { td->isoc_next->isoc_prev = NULL; SetFL32(uhcip, uhcip->uhci_frame_lst_tablep[starting_frame], GetTD32(uhcip, td->link_ptr)); } else { td->isoc_prev->isoc_next = td->isoc_next; td->isoc_next->isoc_prev = td->isoc_prev; td->isoc_prev->link_ptr = td->link_ptr; } uhci_delete_td(uhcip, td); } /* * uhci_send_isoc_receive * - Allocates usb_isoc_request * - Updates the isoc request * - Inserts the isoc td's into the HC processing list. */ int uhci_start_isoc_receive_polling( uhci_state_t *uhcip, usba_pipe_handle_data_t *ph, usb_isoc_req_t *isoc_req, usb_flags_t usb_flags) { int ii, error; size_t max_isoc_xfer_size, length, isoc_pkts_length; ushort_t isoc_pkt_count; uhci_pipe_private_t *pp = (uhci_pipe_private_t *)ph->p_hcd_private; usb_isoc_pkt_descr_t *isoc_pkt_descr; USB_DPRINTF_L4(PRINT_MASK_ISOC, uhcip->uhci_log_hdl, "uhci_start_isoc_receive_polling: usb_flags = %x", usb_flags); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); max_isoc_xfer_size = ph->p_ep.wMaxPacketSize * UHCI_MAX_ISOC_PKTS; if (isoc_req) { isoc_pkt_descr = isoc_req->isoc_pkt_descr; isoc_pkt_count = isoc_req->isoc_pkts_count; isoc_pkts_length = isoc_req->isoc_pkts_length; } else { isoc_pkt_descr = ((usb_isoc_req_t *) pp->pp_client_periodic_in_reqp)->isoc_pkt_descr; isoc_pkt_count = ((usb_isoc_req_t *) pp->pp_client_periodic_in_reqp)->isoc_pkts_count; isoc_pkts_length = ((usb_isoc_req_t *) pp->pp_client_periodic_in_reqp)->isoc_pkts_length; } for (ii = 0, length = 0; ii < isoc_pkt_count; ii++) { length += isoc_pkt_descr->isoc_pkt_length; isoc_pkt_descr++; } if ((isoc_pkts_length) && (isoc_pkts_length != length)) { USB_DPRINTF_L2(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_start_isoc_receive_polling: isoc_pkts_length 0x%lx " "is not equal to the sum of all pkt lengths 0x%lx in " "an isoc request", isoc_pkts_length, length); return (USB_FAILURE); } /* Check the size of isochronous request */ if (length > max_isoc_xfer_size) { USB_DPRINTF_L4(PRINT_MASK_ISOC, uhcip->uhci_log_hdl, "uhci_start_isoc_receive_polling: " "Max isoc request size = %lx, Given isoc req size = %lx", max_isoc_xfer_size, length); return (USB_FAILURE); } /* Add the TD into the Host Controller's isoc list */ error = uhci_insert_isoc_td(uhcip, ph, isoc_req, length, usb_flags); return (error); } /* * uhci_remove_isoc_tds_tws * This routine scans the pipe and removes all the td's * and transfer wrappers and deallocates the memory * associated with those td's and tw's. */ void uhci_remove_isoc_tds_tws(uhci_state_t *uhcip, uhci_pipe_private_t *pp) { uint_t rval, i; uhci_td_t *tmp_td, *td_head; usb_isoc_req_t *isoc_req; uhci_trans_wrapper_t *tmp_tw, *tw_head; uhci_bulk_isoc_xfer_t *isoc_xfer_info; uhci_bulk_isoc_td_pool_t *td_pool_ptr; USB_DPRINTF_L4(PRINT_MASK_ISOC, uhcip->uhci_log_hdl, "uhci_remove_isoc_tds_tws: pp = %p", (void *)pp); tw_head = pp->pp_tw_head; while (tw_head) { tmp_tw = tw_head; tw_head = tw_head->tw_next; td_head = tmp_tw->tw_hctd_head; if (tmp_tw->tw_direction == PID_IN) { uhci_deallocate_periodic_in_resource(uhcip, pp, tmp_tw); } else if (tmp_tw->tw_direction == PID_OUT) { uhci_hcdi_callback(uhcip, pp, pp->pp_pipe_handle, tmp_tw, USB_CR_FLUSHED); } while (td_head) { tmp_td = td_head; td_head = td_head->tw_td_next; uhci_delete_isoc_td(uhcip, tmp_td); } isoc_req = (usb_isoc_req_t *)tmp_tw->tw_isoc_req; if (isoc_req) { usb_free_isoc_req(isoc_req); } ASSERT(tmp_tw->tw_hctd_head == NULL); if (tmp_tw->tw_xfer_info.td_pools) { isoc_xfer_info = (uhci_bulk_isoc_xfer_t *)&tmp_tw->tw_xfer_info; for (i = 0; i < isoc_xfer_info->num_pools; i++) { td_pool_ptr = &isoc_xfer_info->td_pools[i]; rval = ddi_dma_unbind_handle( td_pool_ptr->dma_handle); ASSERT(rval == DDI_SUCCESS); ddi_dma_mem_free(&td_pool_ptr->mem_handle); ddi_dma_free_handle(&td_pool_ptr->dma_handle); } kmem_free(isoc_xfer_info->td_pools, (sizeof (uhci_bulk_isoc_td_pool_t) * isoc_xfer_info->num_pools)); } uhci_deallocate_tw(uhcip, pp, tmp_tw); } } /* * uhci_isoc_update_sw_frame_number() * to avoid code duplication, call uhci_get_sw_frame_number() */ void uhci_isoc_update_sw_frame_number(uhci_state_t *uhcip) { (void) uhci_get_sw_frame_number(uhcip); } /* * uhci_get_sw_frame_number: * Hold the uhci_int_mutex before calling this routine. */ uint64_t uhci_get_sw_frame_number(uhci_state_t *uhcip) { uint64_t sw_frnum, hw_frnum, current_frnum; ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); sw_frnum = uhcip->uhci_sw_frnum; hw_frnum = Get_OpReg16(FRNUM); /* * Check bit 10 in the software counter and hardware frame counter. * If both are same, then don't increment the software frame counter * (Bit 10 of hw frame counter toggle for every 1024 frames) * The lower 11 bits of software counter contains the hardware frame * counter value. The MSB (bit 10) of software counter is incremented * for every 1024 frames either here or in get frame number routine. */ if ((sw_frnum & UHCI_BIT_10_MASK) == (hw_frnum & UHCI_BIT_10_MASK)) { /* The MSB of hw counter did not toggle */ current_frnum = ((sw_frnum & (SW_FRNUM_MASK)) | hw_frnum); } else { /* * The hw counter wrapped around. And the interrupt handler * did not get a chance to update the sw frame counter. * So, update the sw frame counter and return correct frame no. */ sw_frnum >>= UHCI_SIZE_OF_HW_FRNUM - 1; current_frnum = ((++sw_frnum << (UHCI_SIZE_OF_HW_FRNUM - 1)) | hw_frnum); } uhcip->uhci_sw_frnum = current_frnum; USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_get_sw_frame_number: sw=%lld hd=%lld", (unsigned long long)(uhcip->uhci_sw_frnum), (unsigned long long)hw_frnum); return (current_frnum); } /* * uhci_cmd_timeout_hdlr: * This routine will get called for every second. It checks for * timed out control commands/bulk commands. Timeout any commands * that exceeds the time out period specified by the pipe policy. */ void uhci_cmd_timeout_hdlr(void *arg) { uint_t flag = B_FALSE; uhci_td_t *head, *tmp_td; uhci_state_t *uhcip = (uhci_state_t *)arg; uhci_pipe_private_t *pp; /* * Check whether any of the control xfers are timed out. * If so, complete those commands with time out as reason. */ mutex_enter(&uhcip->uhci_int_mutex); head = uhcip->uhci_outst_tds_head; while (head) { /* * If timeout out is zero, then dont timeout command. */ if (head->tw->tw_timeout_cnt == 0) { head = head->outst_td_next; continue; } if (!(head->tw->tw_flags & TW_TIMEOUT_FLAG)) { head->tw->tw_flags |= TW_TIMEOUT_FLAG; --head->tw->tw_timeout_cnt; } /* only do it for bulk and control TDs */ if ((head->tw->tw_timeout_cnt == 0) && (head->tw->tw_handle_td != uhci_handle_isoc_td)) { USB_DPRINTF_L3(PRINT_MASK_ATTA, uhcip->uhci_log_hdl, "Command timed out: td = %p", (void *)head); head->tw->tw_claim = UHCI_TIMEOUT_HDLR_CLAIMED; /* * Check finally whether the command completed */ if (GetTD_status(uhcip, head) & UHCI_TD_ACTIVE) { SetTD32(uhcip, head->link_ptr, GetTD32(uhcip, head->link_ptr) | HC_END_OF_LIST); pp = head->tw->tw_pipe_private; SetQH32(uhcip, pp->pp_qh->element_ptr, GetQH32(uhcip, pp->pp_qh->element_ptr) | HC_END_OF_LIST); } flag = B_TRUE; } head = head->outst_td_next; } if (flag) { (void) uhci_wait_for_sof(uhcip); } head = uhcip->uhci_outst_tds_head; while (head) { if (head->tw->tw_flags & TW_TIMEOUT_FLAG) { head->tw->tw_flags &= ~TW_TIMEOUT_FLAG; } if (head->tw->tw_claim == UHCI_TIMEOUT_HDLR_CLAIMED) { head->tw->tw_claim = UHCI_NOT_CLAIMED; tmp_td = head->tw->tw_hctd_head; while (tmp_td) { SetTD_status(uhcip, tmp_td, UHCI_TD_CRC_TIMEOUT); tmp_td = tmp_td->tw_td_next; } } head = head->outst_td_next; } /* * Process the td which was completed before shifting from normal * mode to polled mode */ if (uhcip->uhci_polled_flag == UHCI_POLLED_FLAG_TRUE) { uhci_process_submitted_td_queue(uhcip); uhcip->uhci_polled_flag = UHCI_POLLED_FLAG_FALSE; } else if (flag) { /* Process the completed/timed out commands */ uhci_process_submitted_td_queue(uhcip); } /* Re-register the control/bulk/intr commands' timeout handler */ if (uhcip->uhci_cmd_timeout_id) { uhcip->uhci_cmd_timeout_id = timeout(uhci_cmd_timeout_hdlr, (void *)uhcip, UHCI_ONE_SECOND); } mutex_exit(&uhcip->uhci_int_mutex); } /* * uhci_wait_for_sof: * Wait for the start of the next frame (implying any changes made in the * lattice have now taken effect). * To be sure this is the case, we wait for the completion of the current * frame (which might have already been pending), then another complete * frame to ensure everything has taken effect. */ int uhci_wait_for_sof(uhci_state_t *uhcip) { int n, error; ushort_t cmd_reg; usb_frame_number_t before_frame_number, after_frame_number; clock_t time, rval; USB_DPRINTF_L4(PRINT_MASK_LISTS, uhcip->uhci_log_hdl, "uhci_wait_for_sof: uhcip = %p", (void *)uhcip); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); error = uhci_state_is_operational(uhcip); if (error != USB_SUCCESS) { return (error); } before_frame_number = uhci_get_sw_frame_number(uhcip); for (n = 0; n < MAX_SOF_WAIT_COUNT; n++) { SetTD_ioc(uhcip, uhcip->uhci_sof_td, 1); uhcip->uhci_cv_signal = B_TRUE; time = ddi_get_lbolt() + UHCI_ONE_SECOND; rval = cv_timedwait(&uhcip->uhci_cv_SOF, &uhcip->uhci_int_mutex, time); after_frame_number = uhci_get_sw_frame_number(uhcip); if ((rval == -1) && (after_frame_number <= before_frame_number)) { cmd_reg = Get_OpReg16(USBCMD); Set_OpReg16(USBCMD, (cmd_reg | USBCMD_REG_HC_RUN)); Set_OpReg16(USBINTR, ENABLE_ALL_INTRS); after_frame_number = uhci_get_sw_frame_number(uhcip); } before_frame_number = after_frame_number; } SetTD_ioc(uhcip, uhcip->uhci_sof_td, 0); return (uhcip->uhci_cv_signal ? USB_FAILURE : USB_SUCCESS); } /* * uhci_allocate_periodic_in_resource: * Allocate interrupt/isochronous request structure for the * interrupt/isochronous IN transfer. */ int uhci_allocate_periodic_in_resource( uhci_state_t *uhcip, uhci_pipe_private_t *pp, uhci_trans_wrapper_t *tw, usb_flags_t flags) { size_t length = 0; usb_opaque_t client_periodic_in_reqp; usb_intr_req_t *cur_intr_req; usb_isoc_req_t *curr_isoc_reqp; usba_pipe_handle_data_t *ph = pp->pp_pipe_handle; USB_DPRINTF_L4(PRINT_MASK_HCDI, uhcip->uhci_log_hdl, "uhci_allocate_periodic_in_resource:\n\t" "ph = 0x%p, pp = 0x%p, tw = 0x%p, flags = 0x%x", (void *)ph, (void *)pp, (void *)tw, flags); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); /* Check the current periodic in request pointer */ if (tw->tw_curr_xfer_reqp) { USB_DPRINTF_L2(PRINT_MASK_HCDI, uhcip->uhci_log_hdl, "uhci_allocate_periodic_in_resource: Interrupt " "request structure already exists: " "allocation failed"); return (USB_SUCCESS); } /* Get the client periodic in request pointer */ client_periodic_in_reqp = pp->pp_client_periodic_in_reqp; /* * If it a periodic IN request and periodic request is NULL, * allocate corresponding usb periodic IN request for the * current periodic polling request and copy the information * from the saved periodic request structure. */ if (UHCI_XFER_TYPE(&ph->p_ep) == USB_EP_ATTR_INTR) { /* Get the interrupt transfer length */ length = ((usb_intr_req_t *)client_periodic_in_reqp)-> intr_len; cur_intr_req = usba_hcdi_dup_intr_req(ph->p_dip, (usb_intr_req_t *)client_periodic_in_reqp, length, flags); if (cur_intr_req == NULL) { USB_DPRINTF_L2(PRINT_MASK_HCDI, uhcip->uhci_log_hdl, "uhci_allocate_periodic_in_resource: Interrupt " "request structure allocation failed"); return (USB_NO_RESOURCES); } /* Check and save the timeout value */ tw->tw_timeout_cnt = (cur_intr_req->intr_attributes & USB_ATTRS_ONE_XFER) ? cur_intr_req->intr_timeout: 0; tw->tw_curr_xfer_reqp = (usb_opaque_t)cur_intr_req; tw->tw_length = cur_intr_req->intr_len; } else { ASSERT(client_periodic_in_reqp != NULL); if ((curr_isoc_reqp = usba_hcdi_dup_isoc_req(ph->p_dip, (usb_isoc_req_t *)client_periodic_in_reqp, flags)) == NULL) { USB_DPRINTF_L2(PRINT_MASK_HCDI, uhcip->uhci_log_hdl, "uhci_allocate_periodic_in_resource: Isochronous " "request structure allocation failed"); return (USB_NO_RESOURCES); } /* * Save the client's isochronous request pointer and * length of isochronous transfer in transfer wrapper. * The dup'ed request is saved in pp_client_periodic_in_reqp */ tw->tw_curr_xfer_reqp = (usb_opaque_t)pp->pp_client_periodic_in_reqp; pp->pp_client_periodic_in_reqp = (usb_opaque_t)curr_isoc_reqp; } mutex_enter(&ph->p_mutex); ph->p_req_count++; mutex_exit(&ph->p_mutex); return (USB_SUCCESS); } /* * uhci_deallocate_periodic_in_resource: * Deallocate interrupt/isochronous request structure for the * interrupt/isochronous IN transfer. */ void uhci_deallocate_periodic_in_resource( uhci_state_t *uhcip, uhci_pipe_private_t *pp, uhci_trans_wrapper_t *tw) { usb_opaque_t curr_xfer_reqp; usba_pipe_handle_data_t *ph = pp->pp_pipe_handle; USB_DPRINTF_L4(PRINT_MASK_HCDI, uhcip->uhci_log_hdl, "uhci_deallocate_periodic_in_resource: " "pp = 0x%p tw = 0x%p", (void *)pp, (void *)tw); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); curr_xfer_reqp = tw->tw_curr_xfer_reqp; if (curr_xfer_reqp) { /* * Reset periodic in request usb isoch * packet request pointers to null. */ tw->tw_curr_xfer_reqp = NULL; tw->tw_isoc_req = NULL; mutex_enter(&ph->p_mutex); ph->p_req_count--; mutex_exit(&ph->p_mutex); /* * Free pre-allocated interrupt or isochronous requests. */ switch (UHCI_XFER_TYPE(&ph->p_ep)) { case USB_EP_ATTR_INTR: usb_free_intr_req((usb_intr_req_t *)curr_xfer_reqp); break; case USB_EP_ATTR_ISOCH: usb_free_isoc_req((usb_isoc_req_t *)curr_xfer_reqp); break; } } } /* * uhci_hcdi_callback() * convenience wrapper around usba_hcdi_callback() */ void uhci_hcdi_callback(uhci_state_t *uhcip, uhci_pipe_private_t *pp, usba_pipe_handle_data_t *ph, uhci_trans_wrapper_t *tw, usb_cr_t cr) { usb_opaque_t curr_xfer_reqp; USB_DPRINTF_L4(PRINT_MASK_HCDI, uhcip->uhci_log_hdl, "uhci_hcdi_callback: ph = 0x%p, tw = 0x%p, cr = 0x%x", (void *)ph, (void *)tw, cr); ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); if (tw && tw->tw_curr_xfer_reqp) { curr_xfer_reqp = tw->tw_curr_xfer_reqp; tw->tw_curr_xfer_reqp = NULL; tw->tw_isoc_req = NULL; } else { ASSERT(pp->pp_client_periodic_in_reqp != NULL); curr_xfer_reqp = pp->pp_client_periodic_in_reqp; pp->pp_client_periodic_in_reqp = NULL; } ASSERT(curr_xfer_reqp != NULL); mutex_exit(&uhcip->uhci_int_mutex); usba_hcdi_cb(ph, curr_xfer_reqp, cr); mutex_enter(&uhcip->uhci_int_mutex); } /* * uhci_state_is_operational: * * Check the Host controller state and return proper values. */ int uhci_state_is_operational(uhci_state_t *uhcip) { int val; ASSERT(mutex_owned(&uhcip->uhci_int_mutex)); switch (uhcip->uhci_hc_soft_state) { case UHCI_CTLR_INIT_STATE: case UHCI_CTLR_SUSPEND_STATE: val = USB_FAILURE; break; case UHCI_CTLR_OPERATIONAL_STATE: val = USB_SUCCESS; break; case UHCI_CTLR_ERROR_STATE: val = USB_HC_HARDWARE_ERROR; break; default: val = USB_FAILURE; break; } return (val); } #ifdef DEBUG static void uhci_print_td(uhci_state_t *uhcip, uhci_td_t *td) { uint_t *ptr = (uint_t *)td; #ifndef lint _NOTE(NO_COMPETING_THREADS_NOW); #endif USB_DPRINTF_L3(PRINT_MASK_DUMPING, uhcip->uhci_log_hdl, "\tDWORD 1 0x%x\t DWORD 2 0x%x", ptr[0], ptr[1]); USB_DPRINTF_L3(PRINT_MASK_DUMPING, uhcip->uhci_log_hdl, "\tDWORD 3 0x%x\t DWORD 4 0x%x", ptr[2], ptr[3]); USB_DPRINTF_L3(PRINT_MASK_DUMPING, uhcip->uhci_log_hdl, "\tBytes xfered = %d", td->tw->tw_bytes_xfered); USB_DPRINTF_L3(PRINT_MASK_DUMPING, uhcip->uhci_log_hdl, "\tBytes Pending = %d", td->tw->tw_bytes_pending); USB_DPRINTF_L3(PRINT_MASK_DUMPING, uhcip->uhci_log_hdl, "Queue Head Details:"); uhci_print_qh(uhcip, td->tw->tw_pipe_private->pp_qh); #ifndef lint _NOTE(COMPETING_THREADS_NOW); #endif } static void uhci_print_qh(uhci_state_t *uhcip, queue_head_t *qh) { uint_t *ptr = (uint_t *)qh; USB_DPRINTF_L3(PRINT_MASK_DUMPING, uhcip->uhci_log_hdl, "\tLink Ptr = %x Element Ptr = %x", ptr[0], ptr[1]); } #endif