/titanic_44/usr/src/uts/common/io/fibre-channel/fca/emlxs/ |
H A D | emlxs_sli3.c | 32 static void emlxs_sli3_issue_iocb(emlxs_hba_t *hba, RING *rp, IOCBQ *iocbq); 33 static void emlxs_sli3_handle_link_event(emlxs_hba_t *hba); 34 static void emlxs_sli3_handle_ring_event(emlxs_hba_t *hba, int32_t ring_no, 44 static int emlxs_sli3_map_hdw(emlxs_hba_t *hba); 46 static void emlxs_sli3_unmap_hdw(emlxs_hba_t *hba); 48 static int32_t emlxs_sli3_online(emlxs_hba_t *hba); 50 static void emlxs_sli3_offline(emlxs_hba_t *hba, 53 static uint32_t emlxs_sli3_hba_reset(emlxs_hba_t *hba, 57 static void emlxs_sli3_hba_kill(emlxs_hba_t *hba); 58 static void emlxs_sli3_hba_kill4quiesce(emlxs_hba_t *hba); [all …]
|
H A D | emlxs_clock.c | 37 static void emlxs_timer_check_loopback(emlxs_hba_t *hba); 43 static void emlxs_timer_check_pools(emlxs_hba_t *hba); 45 static void emlxs_timer_check_fw_update(emlxs_hba_t *hba); 46 static void emlxs_timer_check_heartbeat(emlxs_hba_t *hba); 47 static uint32_t emlxs_timer_check_pkts(emlxs_hba_t *hba, uint8_t *flag); 49 static void emlxs_timer_check_linkup(emlxs_hba_t *hba); 53 static void emlxs_timer_check_channels(emlxs_hba_t *hba, uint8_t *flag); 58 static void emlxs_tx_watchdog(emlxs_hba_t *hba); 62 emlxs_timeout(emlxs_hba_t *hba, uint32_t timeout) in emlxs_timeout() argument 83 emlxs_hba_t *hba = (emlxs_hba_t *)arg; in emlxs_timer() local [all …]
|
H A D | emlxs_sli4.c | 33 static int emlxs_sli4_init_extents(emlxs_hba_t *hba, 35 static uint32_t emlxs_sli4_read_status(emlxs_hba_t *hba); 37 static int emlxs_init_bootstrap_mb(emlxs_hba_t *hba); 39 static uint32_t emlxs_sli4_read_sema(emlxs_hba_t *hba); 41 static uint32_t emlxs_sli4_read_mbdb(emlxs_hba_t *hba); 43 static void emlxs_sli4_write_mbdb(emlxs_hba_t *hba, uint32_t value); 45 static void emlxs_sli4_write_wqdb(emlxs_hba_t *hba, uint32_t value); 47 static void emlxs_sli4_write_mqdb(emlxs_hba_t *hba, uint32_t value); 49 static void emlxs_sli4_write_rqdb(emlxs_hba_t *hba, uint32_t value); 51 static void emlxs_sli4_write_cqdb(emlxs_hba_t *hba, uint32_t value); [all …]
|
H A D | emlxs_hba.c | 36 static void emlxs_handle_async_event(emlxs_hba_t *hba, CHANNEL *cp, 39 static void emlxs_pci_cap_offsets(emlxs_hba_t *hba); 128 emlxs_msi_init(emlxs_hba_t *hba, uint32_t max) in emlxs_msi_init() argument 149 if (!(hba->intr_flags & EMLXS_MSI_ENABLED)) { in emlxs_msi_init() 150 return (emlxs_intx_init(hba, max)); in emlxs_msi_init() 153 if (hba->intr_flags & EMLXS_MSI_INITED) { in emlxs_msi_init() 168 if (hba->model_info.intr_limit && (max > hba->model_info.intr_limit)) { in emlxs_msi_init() 169 max = hba->model_info.intr_limit; in emlxs_msi_init() 174 ret = ddi_intr_get_supported_types(hba->dip, &types); in emlxs_msi_init() 202 (hba->model_info.flags & EMLXS_MSIX_SUPPORTED)) { in emlxs_msi_init() [all …]
|
H A D | emlxs_solaris.c | 46 static void emlxs_fca_attach(emlxs_hba_t *hba); 47 static void emlxs_fca_detach(emlxs_hba_t *hba); 48 static void emlxs_drv_banner(emlxs_hba_t *hba); 50 static int32_t emlxs_get_props(emlxs_hba_t *hba); 67 static uint32_t emlxs_integrity_check(emlxs_hba_t *hba); 68 static uint32_t emlxs_test(emlxs_hba_t *hba, uint32_t test_code, 72 static void emlxs_read_vport_prop(emlxs_hba_t *hba); 75 static void emlxs_mode_init_masks(emlxs_hba_t *hba); 79 emlxs_msiid_to_chan(emlxs_hba_t *hba, int msi_id); 81 emlxs_select_msiid(emlxs_hba_t *hba); [all …]
|
H A D | emlxs_fcp.c | 32 #define EMLXS_GET_VADDR(hba, rp, icmd) emlxs_mem_get_vaddr(hba, rp, \ argument 51 emlxs_handle_fcp_event(emlxs_hba_t *hba, CHANNEL *cp, IOCBQ *iocbq) in emlxs_handle_fcp_event() argument 123 if (hba->underrun_counter && (iostat == IOSTAT_SUCCESS) && in emlxs_handle_fcp_event() 125 hba->underrun_counter--; in emlxs_handle_fcp_event() 163 emlxs_check_dma(hba, sbp); in emlxs_handle_fcp_event() 170 if (hba->flag & FC_DMA_CHECK_ERROR) { in emlxs_handle_fcp_event() 171 emlxs_thread_spawn(hba, emlxs_restart_thread, in emlxs_handle_fcp_event() 479 } else if ((hba->sli_mode == EMLXS_HBA_SLI4_MODE) && in emlxs_handle_fcp_event() 723 emlxs_post_buffer(emlxs_hba_t *hba, RING *rp, int16_t cnt) in emlxs_post_buffer() argument 741 if (rp->ringno == hba->channel_els) { in emlxs_post_buffer() [all …]
|
H A D | emlxs_mbox.c | 211 emlxs_mb_resetport(emlxs_hba_t *hba, MAILBOXQ *mbq) in emlxs_mb_resetport() argument 242 emlxs_mb_request_features(emlxs_hba_t *hba, MAILBOXQ *mbq, uint32_t mask) in emlxs_mb_request_features() argument 246 hba->flag &= ~FC_NPIV_ENABLED; in emlxs_mb_request_features() 247 hba->sli.sli4.flag &= ~(EMLXS_SLI4_PHON | EMLXS_SLI4_PHWQ); in emlxs_mb_request_features() 266 emlxs_mb_noop(emlxs_hba_t *hba, MAILBOXQ *mbq) in emlxs_mb_noop() argument 302 emlxs_mbext_noop(emlxs_hba_t *hba, MAILBOXQ *mbq) in emlxs_mbext_noop() argument 311 if ((mp = (MATCHMAP *)emlxs_mem_get(hba, MEM_BUF)) == 0) { in emlxs_mbext_noop() 342 emlxs_mb_eq_create(emlxs_hba_t *hba, MAILBOXQ *mbq, uint32_t num) in emlxs_mb_eq_create() argument 377 addr = hba->sli.sli4.eq[num].addr.phys; in emlxs_mb_eq_create() 390 emlxs_mb_cq_create(emlxs_hba_t *hba, MAILBOXQ *mbq, uint32_t num) in emlxs_mb_cq_create() argument [all …]
|
H A D | emlxs_mem.c | 34 static uint32_t emlxs_mem_pool_alloc(emlxs_hba_t *hba, MEMSEG *seg, 36 static void emlxs_mem_pool_free(emlxs_hba_t *hba, MEMSEG *seg, uint32_t count); 40 emlxs_mem_alloc_buffer(emlxs_hba_t *hba) in emlxs_mem_alloc_buffer() argument 54 bzero(hba->memseg, sizeof (hba->memseg)); in emlxs_mem_alloc_buffer() 58 buf_info->size = (hba->max_iotag * sizeof (emlxs_buf_t *)); in emlxs_mem_alloc_buffer() 60 (void) emlxs_mem_alloc(hba, buf_info); in emlxs_mem_alloc_buffer() 68 hba->fc_table = buf_info->virt; in emlxs_mem_alloc_buffer() 69 bzero(hba->fc_table, buf_info->size); in emlxs_mem_alloc_buffer() 73 seg = &hba->memseg[i]; in emlxs_mem_alloc_buffer() 81 seg->fc_hi_water = hba->max_nodes + 2; in emlxs_mem_alloc_buffer() [all …]
|
H A D | emlxs_dump.c | 40 emlxs_hba_t *hba, in emlxs_menlo_set_mode() argument 50 if (hba->model_info.device_id != PCI_DEVICE_ID_HORNET) { in emlxs_menlo_set_mode() 68 if (rval = emlxs_send_menlo_cmd(hba, (uint8_t *)cmd_buf, cmd_size, in emlxs_menlo_set_mode() 103 emlxs_hba_t *hba, in emlxs_menlo_reset() argument 113 if (hba->model_info.device_id != PCI_DEVICE_ID_HORNET) { in emlxs_menlo_reset() 130 if (rval = emlxs_send_menlo_cmd(hba, (uint8_t *)cmd_buf, cmd_size, in emlxs_menlo_reset() 165 emlxs_hba_t *hba, in emlxs_menlo_get_cfg() argument 174 if (hba->model_info.device_id != PCI_DEVICE_ID_HORNET) { in emlxs_menlo_get_cfg() 191 if (rval = emlxs_send_menlo_cmd(hba, (uint8_t *)cmd_buf, cmd_size, in emlxs_menlo_get_cfg() 223 emlxs_hba_t *hba, in emlxs_menlo_get_logcfg() argument [all …]
|
H A D | emlxs_dfc.c | 32 static int32_t emlxs_dfc_get_rev(emlxs_hba_t *hba, dfc_t *dfc, 34 static int32_t emlxs_dfc_get_hbainfo(emlxs_hba_t *hba, dfc_t *dfc, 36 static int32_t emlxs_dfc_get_hbastats(emlxs_hba_t *hba, dfc_t *dfc, 38 static int32_t emlxs_dfc_get_drvstats(emlxs_hba_t *hba, dfc_t *dfc, 40 static int32_t emlxs_dfc_set_diag(emlxs_hba_t *hba, dfc_t *dfc, 42 static int32_t emlxs_dfc_send_mbox(emlxs_hba_t *hba, dfc_t *dfc, 44 static int32_t emlxs_dfc_read_pci(emlxs_hba_t *hba, dfc_t *dfc, 46 static int32_t emlxs_dfc_write_pci(emlxs_hba_t *hba, dfc_t *dfc, 48 static int32_t emlxs_dfc_get_cfg(emlxs_hba_t *hba, dfc_t *dfc, 50 static int32_t emlxs_dfc_set_cfg(emlxs_hba_t *hba, dfc_t *dfc, [all …]
|
H A D | emlxs_download.c | 34 static uint32_t emlxs_erase_fcode_flash(emlxs_hba_t *hba); 36 static uint32_t emlxs_write_fcode_flash(emlxs_hba_t *hba, 41 static uint32_t emlxs_validate_image(emlxs_hba_t *hba, caddr_t Buffer, 43 static void emlxs_format_dump(emlxs_hba_t *hba, MAILBOXQ *mbq, 46 static uint32_t emlxs_start_abs_download(emlxs_hba_t *hba, PAIF_HDR AifHdr, 49 static uint32_t emlxs_start_abs_download_2mb(emlxs_hba_t *hba, caddr_t buffer, 52 static uint32_t emlxs_proc_abs_2mb(emlxs_hba_t *hba, 59 static uint32_t emlxs_build_parms_2mb_bwc(emlxs_hba_t *hba, PAIF_HDR AifHdr, 61 static uint32_t emlxs_update_exp_rom(emlxs_hba_t *hba, 63 extern uint32_t emlxs_get_max_sram(emlxs_hba_t *hba, uint32_t *MaxRbusSize, [all …]
|
H A D | emlxs_thread.c | 67 func(taskq->hba, arg); in emlxs_taskq_thread() 143 emlxs_taskq_create(emlxs_hba_t *hba, emlxs_taskq_t *taskq) in emlxs_taskq_create() argument 158 DDI_INTR_PRI(hba->intr_arg)); in emlxs_taskq_create() 162 taskq->hba = hba; in emlxs_taskq_create() 165 DDI_INTR_PRI(hba->intr_arg)); in emlxs_taskq_create() 172 DDI_INTR_PRI(hba->intr_arg)); in emlxs_taskq_create() 259 emlxs_hba_t *hba; in emlxs_thread() local 265 hba = ethread->hba; in emlxs_thread() 273 func(hba, arg1, arg2); in emlxs_thread() 281 if (hba->spawn_thread_head == ethread) in emlxs_thread() [all …]
|
H A D | emlxs_event.c | 36 static void emlxs_event_destroy(emlxs_hba_t *hba, emlxs_event_entry_t *entry); 45 emlxs_hba_t *hba = HBA; in emlxs_event_check() local 48 if ((hba->event_mask & evt->mask)) { in emlxs_event_check() 64 emlxs_event_queue_create(emlxs_hba_t *hba) in emlxs_event_queue_create() argument 74 if (!(hba->intr_flags & EMLXS_MSI_ENABLED)) { in emlxs_event_queue_create() 76 (void) ddi_get_iblock_cookie(hba->dip, (uint_t)EMLXS_INUMBER, in emlxs_event_queue_create() 86 DDI_INTR_PRI(hba->intr_arg)); in emlxs_event_queue_create() 96 emlxs_event_queue_destroy(emlxs_hba_t *hba) in emlxs_event_queue_destroy() argument 107 if (hba->event_mask) { in emlxs_event_queue_destroy() 108 hba->event_mask = 0; in emlxs_event_queue_destroy() [all …]
|
H A D | emlxs_diag.c | 136 emlxs_hba_t *hba = HBA; in emlxs_diag_echo_run() local 156 if ((hba->state < FC_LINK_UP) || (port->did == 0)) { in emlxs_diag_echo_run() 234 timeout = emlxs_timeout(hba, (pkt->pkt_timeout + 15)); in emlxs_diag_echo_run() 295 emlxs_diag_biu_run(emlxs_hba_t *hba, uint32_t pattern) in emlxs_diag_biu_run() argument 308 if (hba->state < FC_LINK_DOWN) { in emlxs_diag_biu_run() 318 if ((mbq = (MAILBOXQ *) emlxs_mem_get(hba, MEM_MBOX)) == 0) { in emlxs_diag_biu_run() 329 if (((mp = (MATCHMAP *) emlxs_mem_get(hba, MEM_BUF)) == 0) || in emlxs_diag_biu_run() 330 ((mp1 = (MATCHMAP *) emlxs_mem_get(hba, MEM_BUF)) == 0)) { in emlxs_diag_biu_run() 358 (void) emlxs_mb_run_biu_diag(hba, mbq, mp->phys, mp1->phys); in emlxs_diag_biu_run() 360 rval = EMLXS_SLI_ISSUE_MBOX_CMD(hba, mbq, MBX_WAIT, 60); in emlxs_diag_biu_run() [all …]
|
H A D | emlxs_node.c | 42 emlxs_hba_t *hba = HBA; in emlxs_node_close() local 69 if (channelno == hba->channel_ip) { in emlxs_node_close() 82 ndlp->nlp_tics[channelno] = hba->timer_tics + timeout; in emlxs_node_close() 91 ndlp->nlp_tics[channelno] = hba->timer_tics + timeout; in emlxs_node_close() 108 ndlp->nlp_tics[channelno] = hba->timer_tics + timeout; in emlxs_node_close() 116 ndlp->nlp_tics[channelno] = hba->timer_tics + timeout; in emlxs_node_close() 134 cp = &hba->chan[channelno]; in emlxs_node_close() 179 emlxs_hba_t *hba = HBA; in emlxs_node_timeout() local 214 (void) emlxs_tx_node_flush(port, ndlp, &hba->chan[channelno], 0, 0); in emlxs_node_timeout() 217 (void) emlxs_chipq_node_flush(port, &hba->chan[channelno], ndlp, 0); in emlxs_node_timeout() [all …]
|
H A D | emlxs_fcf.c | 221 static void emlxs_fc_fcftab_online_timer(emlxs_hba_t *hba); 525 static void emlxs_fcoe_fcftab_read_timer(emlxs_hba_t *hba); 526 static void emlxs_fcoe_fcftab_sol_timer(emlxs_hba_t *hba); 527 static void emlxs_fcoe_fcftab_offline_timer(emlxs_hba_t *hba); 1689 static void emlxs_rpi_idle_timer(emlxs_hba_t *hba); 1853 emlxs_hba_t *hba = HBA; in emlxs_fcf_linkdown() local 1855 if (hba->state <= FC_LINK_DOWN) { in emlxs_fcf_linkdown() 1861 if (hba->state <= FC_LINK_DOWN) { in emlxs_fcf_linkdown() 1867 EMLXS_STATE_CHANGE_LOCKED(hba, FC_LINK_DOWN); in emlxs_fcf_linkdown() 1869 hba->flag &= FC_LINKDOWN_MASK; in emlxs_fcf_linkdown() [all …]
|
H A D | emlxs_els.c | 85 emlxs_els_handle_event(emlxs_hba_t *hba, CHANNEL *cp, IOCBQ *iocbq) in emlxs_els_handle_event() argument 118 if (cp->channelno != hba->channel_els) { in emlxs_els_handle_event() 284 hba->channel_fcp); in emlxs_els_handle_event() 286 hba->channel_ip); in emlxs_els_handle_event() 306 hba->channel_fcp); in emlxs_els_handle_event() 331 if (hba->sli_mode == EMLXS_HBA_SLI4_MODE) { in emlxs_els_handle_event() 379 emlxs_hba_t *hba = HBA; in emlxs_els_handle_unsol_req() local 449 emlxs_close_els_exchange(hba, port, iocb->ULPCONTEXT); in emlxs_els_handle_unsol_req() 460 emlxs_close_els_exchange(hba, port, iocb->ULPCONTEXT); in emlxs_els_handle_unsol_req() 472 emlxs_close_els_exchange(hba, port, iocb->ULPCONTEXT); in emlxs_els_handle_unsol_req() [all …]
|
H A D | emlxs_fct.c | 35 static void emlxs_fct_memseg_init(emlxs_hba_t *hba); 83 static int emlxs_fct_dbuf_dma_sync(emlxs_hba_t *hba, stmf_data_buf_t *dbuf, 88 static void emlxs_fct_unsol_flush_thread(emlxs_hba_t *hba, void *arg1, 621 emlxs_hba_t *hba = HBA; local 743 emlxs_fct_unsol_flush_thread(emlxs_hba_t *hba, void *arg1, void *arg2) 757 emlxs_hba_t *hba = HBA; local 869 emlxs_fct_init(emlxs_hba_t *hba) 887 emlxs_fct_attach(emlxs_hba_t *hba) 900 if (hba->flag & FC_NPIV_ENABLED) { 901 for (vpi = 1; vpi <= hba->vpi_high; vpi++) { [all …]
|
H A D | emlxs_ip.c | 34 emlxs_ip_handle_event(emlxs_hba_t *hba, CHANNEL *cp, IOCBQ *iocbq) in emlxs_ip_handle_event() argument 58 if (cp->channelno != hba->channel_ip) { in emlxs_ip_handle_event() 145 !(ndlp->nlp_flag[hba->channel_ip] & in emlxs_ip_handle_event() 181 emlxs_hba_t *hba = HBA; in emlxs_ip_handle_unsol_req() local 234 !(ndlp->nlp_flag[hba->channel_ip] & NLP_RPI_XRI)) { in emlxs_ip_handle_unsol_req() 301 emlxs_ip_handle_rcv_seq_list(emlxs_hba_t *hba, CHANNEL *cp, IOCBQ *iocbq) in emlxs_ip_handle_rcv_seq_list() argument 316 rp = &hba->sli.sli3.ring[cp->channelno]; in emlxs_ip_handle_rcv_seq_list() 333 if (hba->flag & FC_HBQ_ENABLED) { in emlxs_ip_handle_rcv_seq_list() 336 hbq = &hba->sli.sli3.hbq_table[hbq_id]; in emlxs_ip_handle_rcv_seq_list() 343 mp = hba->sli.sli3.hbq_table in emlxs_ip_handle_rcv_seq_list() [all …]
|
H A D | emlxs_msg.c | 42 emlxs_msg_log_create(emlxs_hba_t *hba) in emlxs_msg_log_create() argument 51 DRIVER_NAME, hba->ddiinst, (void *)log); in emlxs_msg_log_create() 63 log->instance = hba->ddiinst; in emlxs_msg_log_create() 66 if (!(hba->intr_flags & EMLXS_MSI_ENABLED)) { in emlxs_msg_log_create() 68 (void) ddi_get_iblock_cookie(hba->dip, (uint_t)EMLXS_INUMBER, in emlxs_msg_log_create() 87 emlxs_msg_lock_reinit(emlxs_hba_t *hba) in emlxs_msg_lock_reinit() argument 95 DRIVER_NAME, hba->ddiinst, (void *)log); in emlxs_msg_lock_reinit() 104 mutex_init(&log->lock, NULL, MUTEX_DRIVER, DDI_INTR_PRI(hba->intr_arg)); in emlxs_msg_lock_reinit() 111 emlxs_msg_log_destroy(emlxs_hba_t *hba) in emlxs_msg_log_destroy() argument 120 DRIVER_NAME, hba->ddiinst, (void *)log); in emlxs_msg_log_destroy() [all …]
|
H A D | emlxs_dhchap.c | 59 static void emlxs_auth_cfg_init(emlxs_hba_t *hba); 60 static void emlxs_auth_cfg_fini(emlxs_hba_t *hba); 61 static void emlxs_auth_cfg_read(emlxs_hba_t *hba); 62 static uint32_t emlxs_auth_cfg_parse(emlxs_hba_t *hba, 64 static emlxs_auth_cfg_t *emlxs_auth_cfg_get(emlxs_hba_t *hba, 66 static emlxs_auth_cfg_t *emlxs_auth_cfg_create(emlxs_hba_t *hba, 68 static void emlxs_auth_cfg_destroy(emlxs_hba_t *hba, 70 static void emlxs_auth_cfg_print(emlxs_hba_t *hba, 73 static void emlxs_auth_key_init(emlxs_hba_t *hba); 74 static void emlxs_auth_key_fini(emlxs_hba_t *hba); [all …]
|
H A D | emlxs_pkt.c | 45 emlxs_pkt_thread(emlxs_hba_t *hba, void *arg1, void *arg2) in emlxs_pkt_thread() argument 82 emlxs_hba_t *hba = HBA; in emlxs_pkt_send() local 89 emlxs_thread_spawn(hba, emlxs_pkt_thread, (char *)pkt, NULL); in emlxs_pkt_send() 151 emlxs_hba_t *hba = HBA; in emlxs_pkt_alloc() local 200 if (ddi_dma_alloc_handle(hba->dip, &hba->dma_attr_1sg, cb, in emlxs_pkt_alloc() 267 if (ddi_dma_alloc_handle(hba->dip, &hba->dma_attr_1sg, cb, in emlxs_pkt_alloc() 331 if (ddi_dma_alloc_handle(hba->dip, &hba->dma_attr_1sg, cb, in emlxs_pkt_alloc() 389 mutex_init(&sbp->mtx, NULL, MUTEX_DRIVER, DDI_INTR_PRI(hba->intr_arg)); in emlxs_pkt_alloc()
|
/titanic_44/usr/src/uts/common/sys/fibre-channel/fca/emlxs/ |
H A D | emlxs_extern.h | 61 extern uint32_t emlxs_msg_log_create(emlxs_hba_t *hba); 62 extern void emlxs_msg_lock_reinit(emlxs_hba_t *hba); 63 extern void emlxs_msg_log_destroy(emlxs_hba_t *hba); 64 extern uint32_t emlxs_msg_log_get(emlxs_hba_t *hba, 71 extern void emlxs_timer_check_events(emlxs_hba_t *hba); 73 extern uint32_t emlxs_event_queue_create(emlxs_hba_t *hba); 75 extern void emlxs_event_queue_destroy(emlxs_hba_t *hba); 150 extern void emlxs_mem_free(emlxs_hba_t *hba, 152 extern uint8_t *emlxs_mem_alloc(emlxs_hba_t *hba, 154 extern int emlxs_map_bus(emlxs_hba_t *hba); [all …]
|
H A D | emlxs_fc.h | 321 struct emlxs_hba *hba; /* ptr to hba for channel */ member 371 struct emlxs_hba *hba; /* ptr to hba for ring */ member 518 #define MEM_ELSBUF_COUNT hba->max_nodes 989 struct emlxs_hba *hba; member 1611 #define SLI4_FCOE_MODE (hba->sli.sli4.flag & EMLXS_SLI4_FCOE_MODE) 2097 #define EMLXS_DUMP_LOCK hba->dump_lock 2118 #define EMLXS_SLI_MAP_HDW (hba->sli_api.sli_map_hdw) 2119 #define EMLXS_SLI_UNMAP_HDW (hba->sli_api.sli_unmap_hdw) 2120 #define EMLXS_SLI_ONLINE (hba->sli_api.sli_online) 2121 #define EMLXS_SLI_OFFLINE (hba->sli_api.sli_offline) [all …]
|
/titanic_44/usr/src/lib/sun_fc/common/ |
H A D | Handle.cc | 100 hba = myhba; in Handle() 153 hba = myhba; 332 HBAPort* port = hba->getPortByIndex(index); in getHandlePortByIndex() 350 HBAPort* port = hba->getPort(wwn); in getHandlePort() 351 portHandles[wwn] = new HandlePort(this, hba, port); in getHandlePort() 371 HBA_ADAPTERATTRIBUTES attributes = hba->getHBAAttributes(); in getHBAAttributes() 389 int rval = hba->doForceLip(); in doForceLip() 402 HBA_ADAPTERATTRIBUTES attributes = hba->npivGetHBAAttributes(); in npivGetHBAAttributes() 429 if (hba->containsWWN(wwn)) { in getPortAttributes() 430 attributes = hba->getPort(wwn)->getPortAttributes(tmp); in getPortAttributes() [all …]
|