Home
last modified time | relevance | path

Searched refs:hba (Results 1 – 25 of 90) sorted by relevance

1234

/titanic_44/usr/src/uts/common/io/fibre-channel/fca/emlxs/
H A Demlxs_sli3.c32 static void emlxs_sli3_issue_iocb(emlxs_hba_t *hba, RING *rp, IOCBQ *iocbq);
33 static void emlxs_sli3_handle_link_event(emlxs_hba_t *hba);
34 static void emlxs_sli3_handle_ring_event(emlxs_hba_t *hba, int32_t ring_no,
44 static int emlxs_sli3_map_hdw(emlxs_hba_t *hba);
46 static void emlxs_sli3_unmap_hdw(emlxs_hba_t *hba);
48 static int32_t emlxs_sli3_online(emlxs_hba_t *hba);
50 static void emlxs_sli3_offline(emlxs_hba_t *hba,
53 static uint32_t emlxs_sli3_hba_reset(emlxs_hba_t *hba,
57 static void emlxs_sli3_hba_kill(emlxs_hba_t *hba);
58 static void emlxs_sli3_hba_kill4quiesce(emlxs_hba_t *hba);
[all …]
H A Demlxs_clock.c37 static void emlxs_timer_check_loopback(emlxs_hba_t *hba);
43 static void emlxs_timer_check_pools(emlxs_hba_t *hba);
45 static void emlxs_timer_check_fw_update(emlxs_hba_t *hba);
46 static void emlxs_timer_check_heartbeat(emlxs_hba_t *hba);
47 static uint32_t emlxs_timer_check_pkts(emlxs_hba_t *hba, uint8_t *flag);
49 static void emlxs_timer_check_linkup(emlxs_hba_t *hba);
53 static void emlxs_timer_check_channels(emlxs_hba_t *hba, uint8_t *flag);
58 static void emlxs_tx_watchdog(emlxs_hba_t *hba);
62 emlxs_timeout(emlxs_hba_t *hba, uint32_t timeout) in emlxs_timeout() argument
83 emlxs_hba_t *hba = (emlxs_hba_t *)arg; in emlxs_timer() local
[all …]
H A Demlxs_sli4.c33 static int emlxs_sli4_init_extents(emlxs_hba_t *hba,
35 static uint32_t emlxs_sli4_read_status(emlxs_hba_t *hba);
37 static int emlxs_init_bootstrap_mb(emlxs_hba_t *hba);
39 static uint32_t emlxs_sli4_read_sema(emlxs_hba_t *hba);
41 static uint32_t emlxs_sli4_read_mbdb(emlxs_hba_t *hba);
43 static void emlxs_sli4_write_mbdb(emlxs_hba_t *hba, uint32_t value);
45 static void emlxs_sli4_write_wqdb(emlxs_hba_t *hba, uint32_t value);
47 static void emlxs_sli4_write_mqdb(emlxs_hba_t *hba, uint32_t value);
49 static void emlxs_sli4_write_rqdb(emlxs_hba_t *hba, uint32_t value);
51 static void emlxs_sli4_write_cqdb(emlxs_hba_t *hba, uint32_t value);
[all …]
H A Demlxs_hba.c36 static void emlxs_handle_async_event(emlxs_hba_t *hba, CHANNEL *cp,
39 static void emlxs_pci_cap_offsets(emlxs_hba_t *hba);
128 emlxs_msi_init(emlxs_hba_t *hba, uint32_t max) in emlxs_msi_init() argument
149 if (!(hba->intr_flags & EMLXS_MSI_ENABLED)) { in emlxs_msi_init()
150 return (emlxs_intx_init(hba, max)); in emlxs_msi_init()
153 if (hba->intr_flags & EMLXS_MSI_INITED) { in emlxs_msi_init()
168 if (hba->model_info.intr_limit && (max > hba->model_info.intr_limit)) { in emlxs_msi_init()
169 max = hba->model_info.intr_limit; in emlxs_msi_init()
174 ret = ddi_intr_get_supported_types(hba->dip, &types); in emlxs_msi_init()
202 (hba->model_info.flags & EMLXS_MSIX_SUPPORTED)) { in emlxs_msi_init()
[all …]
H A Demlxs_solaris.c46 static void emlxs_fca_attach(emlxs_hba_t *hba);
47 static void emlxs_fca_detach(emlxs_hba_t *hba);
48 static void emlxs_drv_banner(emlxs_hba_t *hba);
50 static int32_t emlxs_get_props(emlxs_hba_t *hba);
67 static uint32_t emlxs_integrity_check(emlxs_hba_t *hba);
68 static uint32_t emlxs_test(emlxs_hba_t *hba, uint32_t test_code,
72 static void emlxs_read_vport_prop(emlxs_hba_t *hba);
75 static void emlxs_mode_init_masks(emlxs_hba_t *hba);
79 emlxs_msiid_to_chan(emlxs_hba_t *hba, int msi_id);
81 emlxs_select_msiid(emlxs_hba_t *hba);
[all …]
H A Demlxs_fcp.c32 #define EMLXS_GET_VADDR(hba, rp, icmd) emlxs_mem_get_vaddr(hba, rp, \ argument
51 emlxs_handle_fcp_event(emlxs_hba_t *hba, CHANNEL *cp, IOCBQ *iocbq) in emlxs_handle_fcp_event() argument
123 if (hba->underrun_counter && (iostat == IOSTAT_SUCCESS) && in emlxs_handle_fcp_event()
125 hba->underrun_counter--; in emlxs_handle_fcp_event()
163 emlxs_check_dma(hba, sbp); in emlxs_handle_fcp_event()
170 if (hba->flag & FC_DMA_CHECK_ERROR) { in emlxs_handle_fcp_event()
171 emlxs_thread_spawn(hba, emlxs_restart_thread, in emlxs_handle_fcp_event()
479 } else if ((hba->sli_mode == EMLXS_HBA_SLI4_MODE) && in emlxs_handle_fcp_event()
723 emlxs_post_buffer(emlxs_hba_t *hba, RING *rp, int16_t cnt) in emlxs_post_buffer() argument
741 if (rp->ringno == hba->channel_els) { in emlxs_post_buffer()
[all …]
H A Demlxs_mbox.c211 emlxs_mb_resetport(emlxs_hba_t *hba, MAILBOXQ *mbq) in emlxs_mb_resetport() argument
242 emlxs_mb_request_features(emlxs_hba_t *hba, MAILBOXQ *mbq, uint32_t mask) in emlxs_mb_request_features() argument
246 hba->flag &= ~FC_NPIV_ENABLED; in emlxs_mb_request_features()
247 hba->sli.sli4.flag &= ~(EMLXS_SLI4_PHON | EMLXS_SLI4_PHWQ); in emlxs_mb_request_features()
266 emlxs_mb_noop(emlxs_hba_t *hba, MAILBOXQ *mbq) in emlxs_mb_noop() argument
302 emlxs_mbext_noop(emlxs_hba_t *hba, MAILBOXQ *mbq) in emlxs_mbext_noop() argument
311 if ((mp = (MATCHMAP *)emlxs_mem_get(hba, MEM_BUF)) == 0) { in emlxs_mbext_noop()
342 emlxs_mb_eq_create(emlxs_hba_t *hba, MAILBOXQ *mbq, uint32_t num) in emlxs_mb_eq_create() argument
377 addr = hba->sli.sli4.eq[num].addr.phys; in emlxs_mb_eq_create()
390 emlxs_mb_cq_create(emlxs_hba_t *hba, MAILBOXQ *mbq, uint32_t num) in emlxs_mb_cq_create() argument
[all …]
H A Demlxs_mem.c34 static uint32_t emlxs_mem_pool_alloc(emlxs_hba_t *hba, MEMSEG *seg,
36 static void emlxs_mem_pool_free(emlxs_hba_t *hba, MEMSEG *seg, uint32_t count);
40 emlxs_mem_alloc_buffer(emlxs_hba_t *hba) in emlxs_mem_alloc_buffer() argument
54 bzero(hba->memseg, sizeof (hba->memseg)); in emlxs_mem_alloc_buffer()
58 buf_info->size = (hba->max_iotag * sizeof (emlxs_buf_t *)); in emlxs_mem_alloc_buffer()
60 (void) emlxs_mem_alloc(hba, buf_info); in emlxs_mem_alloc_buffer()
68 hba->fc_table = buf_info->virt; in emlxs_mem_alloc_buffer()
69 bzero(hba->fc_table, buf_info->size); in emlxs_mem_alloc_buffer()
73 seg = &hba->memseg[i]; in emlxs_mem_alloc_buffer()
81 seg->fc_hi_water = hba->max_nodes + 2; in emlxs_mem_alloc_buffer()
[all …]
H A Demlxs_dump.c40 emlxs_hba_t *hba, in emlxs_menlo_set_mode() argument
50 if (hba->model_info.device_id != PCI_DEVICE_ID_HORNET) { in emlxs_menlo_set_mode()
68 if (rval = emlxs_send_menlo_cmd(hba, (uint8_t *)cmd_buf, cmd_size, in emlxs_menlo_set_mode()
103 emlxs_hba_t *hba, in emlxs_menlo_reset() argument
113 if (hba->model_info.device_id != PCI_DEVICE_ID_HORNET) { in emlxs_menlo_reset()
130 if (rval = emlxs_send_menlo_cmd(hba, (uint8_t *)cmd_buf, cmd_size, in emlxs_menlo_reset()
165 emlxs_hba_t *hba, in emlxs_menlo_get_cfg() argument
174 if (hba->model_info.device_id != PCI_DEVICE_ID_HORNET) { in emlxs_menlo_get_cfg()
191 if (rval = emlxs_send_menlo_cmd(hba, (uint8_t *)cmd_buf, cmd_size, in emlxs_menlo_get_cfg()
223 emlxs_hba_t *hba, in emlxs_menlo_get_logcfg() argument
[all …]
H A Demlxs_dfc.c32 static int32_t emlxs_dfc_get_rev(emlxs_hba_t *hba, dfc_t *dfc,
34 static int32_t emlxs_dfc_get_hbainfo(emlxs_hba_t *hba, dfc_t *dfc,
36 static int32_t emlxs_dfc_get_hbastats(emlxs_hba_t *hba, dfc_t *dfc,
38 static int32_t emlxs_dfc_get_drvstats(emlxs_hba_t *hba, dfc_t *dfc,
40 static int32_t emlxs_dfc_set_diag(emlxs_hba_t *hba, dfc_t *dfc,
42 static int32_t emlxs_dfc_send_mbox(emlxs_hba_t *hba, dfc_t *dfc,
44 static int32_t emlxs_dfc_read_pci(emlxs_hba_t *hba, dfc_t *dfc,
46 static int32_t emlxs_dfc_write_pci(emlxs_hba_t *hba, dfc_t *dfc,
48 static int32_t emlxs_dfc_get_cfg(emlxs_hba_t *hba, dfc_t *dfc,
50 static int32_t emlxs_dfc_set_cfg(emlxs_hba_t *hba, dfc_t *dfc,
[all …]
H A Demlxs_download.c34 static uint32_t emlxs_erase_fcode_flash(emlxs_hba_t *hba);
36 static uint32_t emlxs_write_fcode_flash(emlxs_hba_t *hba,
41 static uint32_t emlxs_validate_image(emlxs_hba_t *hba, caddr_t Buffer,
43 static void emlxs_format_dump(emlxs_hba_t *hba, MAILBOXQ *mbq,
46 static uint32_t emlxs_start_abs_download(emlxs_hba_t *hba, PAIF_HDR AifHdr,
49 static uint32_t emlxs_start_abs_download_2mb(emlxs_hba_t *hba, caddr_t buffer,
52 static uint32_t emlxs_proc_abs_2mb(emlxs_hba_t *hba,
59 static uint32_t emlxs_build_parms_2mb_bwc(emlxs_hba_t *hba, PAIF_HDR AifHdr,
61 static uint32_t emlxs_update_exp_rom(emlxs_hba_t *hba,
63 extern uint32_t emlxs_get_max_sram(emlxs_hba_t *hba, uint32_t *MaxRbusSize,
[all …]
H A Demlxs_thread.c67 func(taskq->hba, arg); in emlxs_taskq_thread()
143 emlxs_taskq_create(emlxs_hba_t *hba, emlxs_taskq_t *taskq) in emlxs_taskq_create() argument
158 DDI_INTR_PRI(hba->intr_arg)); in emlxs_taskq_create()
162 taskq->hba = hba; in emlxs_taskq_create()
165 DDI_INTR_PRI(hba->intr_arg)); in emlxs_taskq_create()
172 DDI_INTR_PRI(hba->intr_arg)); in emlxs_taskq_create()
259 emlxs_hba_t *hba; in emlxs_thread() local
265 hba = ethread->hba; in emlxs_thread()
273 func(hba, arg1, arg2); in emlxs_thread()
281 if (hba->spawn_thread_head == ethread) in emlxs_thread()
[all …]
H A Demlxs_event.c36 static void emlxs_event_destroy(emlxs_hba_t *hba, emlxs_event_entry_t *entry);
45 emlxs_hba_t *hba = HBA; in emlxs_event_check() local
48 if ((hba->event_mask & evt->mask)) { in emlxs_event_check()
64 emlxs_event_queue_create(emlxs_hba_t *hba) in emlxs_event_queue_create() argument
74 if (!(hba->intr_flags & EMLXS_MSI_ENABLED)) { in emlxs_event_queue_create()
76 (void) ddi_get_iblock_cookie(hba->dip, (uint_t)EMLXS_INUMBER, in emlxs_event_queue_create()
86 DDI_INTR_PRI(hba->intr_arg)); in emlxs_event_queue_create()
96 emlxs_event_queue_destroy(emlxs_hba_t *hba) in emlxs_event_queue_destroy() argument
107 if (hba->event_mask) { in emlxs_event_queue_destroy()
108 hba->event_mask = 0; in emlxs_event_queue_destroy()
[all …]
H A Demlxs_diag.c136 emlxs_hba_t *hba = HBA; in emlxs_diag_echo_run() local
156 if ((hba->state < FC_LINK_UP) || (port->did == 0)) { in emlxs_diag_echo_run()
234 timeout = emlxs_timeout(hba, (pkt->pkt_timeout + 15)); in emlxs_diag_echo_run()
295 emlxs_diag_biu_run(emlxs_hba_t *hba, uint32_t pattern) in emlxs_diag_biu_run() argument
308 if (hba->state < FC_LINK_DOWN) { in emlxs_diag_biu_run()
318 if ((mbq = (MAILBOXQ *) emlxs_mem_get(hba, MEM_MBOX)) == 0) { in emlxs_diag_biu_run()
329 if (((mp = (MATCHMAP *) emlxs_mem_get(hba, MEM_BUF)) == 0) || in emlxs_diag_biu_run()
330 ((mp1 = (MATCHMAP *) emlxs_mem_get(hba, MEM_BUF)) == 0)) { in emlxs_diag_biu_run()
358 (void) emlxs_mb_run_biu_diag(hba, mbq, mp->phys, mp1->phys); in emlxs_diag_biu_run()
360 rval = EMLXS_SLI_ISSUE_MBOX_CMD(hba, mbq, MBX_WAIT, 60); in emlxs_diag_biu_run()
[all …]
H A Demlxs_node.c42 emlxs_hba_t *hba = HBA; in emlxs_node_close() local
69 if (channelno == hba->channel_ip) { in emlxs_node_close()
82 ndlp->nlp_tics[channelno] = hba->timer_tics + timeout; in emlxs_node_close()
91 ndlp->nlp_tics[channelno] = hba->timer_tics + timeout; in emlxs_node_close()
108 ndlp->nlp_tics[channelno] = hba->timer_tics + timeout; in emlxs_node_close()
116 ndlp->nlp_tics[channelno] = hba->timer_tics + timeout; in emlxs_node_close()
134 cp = &hba->chan[channelno]; in emlxs_node_close()
179 emlxs_hba_t *hba = HBA; in emlxs_node_timeout() local
214 (void) emlxs_tx_node_flush(port, ndlp, &hba->chan[channelno], 0, 0); in emlxs_node_timeout()
217 (void) emlxs_chipq_node_flush(port, &hba->chan[channelno], ndlp, 0); in emlxs_node_timeout()
[all …]
H A Demlxs_fcf.c221 static void emlxs_fc_fcftab_online_timer(emlxs_hba_t *hba);
525 static void emlxs_fcoe_fcftab_read_timer(emlxs_hba_t *hba);
526 static void emlxs_fcoe_fcftab_sol_timer(emlxs_hba_t *hba);
527 static void emlxs_fcoe_fcftab_offline_timer(emlxs_hba_t *hba);
1689 static void emlxs_rpi_idle_timer(emlxs_hba_t *hba);
1853 emlxs_hba_t *hba = HBA; in emlxs_fcf_linkdown() local
1855 if (hba->state <= FC_LINK_DOWN) { in emlxs_fcf_linkdown()
1861 if (hba->state <= FC_LINK_DOWN) { in emlxs_fcf_linkdown()
1867 EMLXS_STATE_CHANGE_LOCKED(hba, FC_LINK_DOWN); in emlxs_fcf_linkdown()
1869 hba->flag &= FC_LINKDOWN_MASK; in emlxs_fcf_linkdown()
[all …]
H A Demlxs_els.c85 emlxs_els_handle_event(emlxs_hba_t *hba, CHANNEL *cp, IOCBQ *iocbq) in emlxs_els_handle_event() argument
118 if (cp->channelno != hba->channel_els) { in emlxs_els_handle_event()
284 hba->channel_fcp); in emlxs_els_handle_event()
286 hba->channel_ip); in emlxs_els_handle_event()
306 hba->channel_fcp); in emlxs_els_handle_event()
331 if (hba->sli_mode == EMLXS_HBA_SLI4_MODE) { in emlxs_els_handle_event()
379 emlxs_hba_t *hba = HBA; in emlxs_els_handle_unsol_req() local
449 emlxs_close_els_exchange(hba, port, iocb->ULPCONTEXT); in emlxs_els_handle_unsol_req()
460 emlxs_close_els_exchange(hba, port, iocb->ULPCONTEXT); in emlxs_els_handle_unsol_req()
472 emlxs_close_els_exchange(hba, port, iocb->ULPCONTEXT); in emlxs_els_handle_unsol_req()
[all …]
H A Demlxs_fct.c35 static void emlxs_fct_memseg_init(emlxs_hba_t *hba);
83 static int emlxs_fct_dbuf_dma_sync(emlxs_hba_t *hba, stmf_data_buf_t *dbuf,
88 static void emlxs_fct_unsol_flush_thread(emlxs_hba_t *hba, void *arg1,
621 emlxs_hba_t *hba = HBA; local
743 emlxs_fct_unsol_flush_thread(emlxs_hba_t *hba, void *arg1, void *arg2)
757 emlxs_hba_t *hba = HBA; local
869 emlxs_fct_init(emlxs_hba_t *hba)
887 emlxs_fct_attach(emlxs_hba_t *hba)
900 if (hba->flag & FC_NPIV_ENABLED) {
901 for (vpi = 1; vpi <= hba->vpi_high; vpi++) {
[all …]
H A Demlxs_ip.c34 emlxs_ip_handle_event(emlxs_hba_t *hba, CHANNEL *cp, IOCBQ *iocbq) in emlxs_ip_handle_event() argument
58 if (cp->channelno != hba->channel_ip) { in emlxs_ip_handle_event()
145 !(ndlp->nlp_flag[hba->channel_ip] & in emlxs_ip_handle_event()
181 emlxs_hba_t *hba = HBA; in emlxs_ip_handle_unsol_req() local
234 !(ndlp->nlp_flag[hba->channel_ip] & NLP_RPI_XRI)) { in emlxs_ip_handle_unsol_req()
301 emlxs_ip_handle_rcv_seq_list(emlxs_hba_t *hba, CHANNEL *cp, IOCBQ *iocbq) in emlxs_ip_handle_rcv_seq_list() argument
316 rp = &hba->sli.sli3.ring[cp->channelno]; in emlxs_ip_handle_rcv_seq_list()
333 if (hba->flag & FC_HBQ_ENABLED) { in emlxs_ip_handle_rcv_seq_list()
336 hbq = &hba->sli.sli3.hbq_table[hbq_id]; in emlxs_ip_handle_rcv_seq_list()
343 mp = hba->sli.sli3.hbq_table in emlxs_ip_handle_rcv_seq_list()
[all …]
H A Demlxs_msg.c42 emlxs_msg_log_create(emlxs_hba_t *hba) in emlxs_msg_log_create() argument
51 DRIVER_NAME, hba->ddiinst, (void *)log); in emlxs_msg_log_create()
63 log->instance = hba->ddiinst; in emlxs_msg_log_create()
66 if (!(hba->intr_flags & EMLXS_MSI_ENABLED)) { in emlxs_msg_log_create()
68 (void) ddi_get_iblock_cookie(hba->dip, (uint_t)EMLXS_INUMBER, in emlxs_msg_log_create()
87 emlxs_msg_lock_reinit(emlxs_hba_t *hba) in emlxs_msg_lock_reinit() argument
95 DRIVER_NAME, hba->ddiinst, (void *)log); in emlxs_msg_lock_reinit()
104 mutex_init(&log->lock, NULL, MUTEX_DRIVER, DDI_INTR_PRI(hba->intr_arg)); in emlxs_msg_lock_reinit()
111 emlxs_msg_log_destroy(emlxs_hba_t *hba) in emlxs_msg_log_destroy() argument
120 DRIVER_NAME, hba->ddiinst, (void *)log); in emlxs_msg_log_destroy()
[all …]
H A Demlxs_dhchap.c59 static void emlxs_auth_cfg_init(emlxs_hba_t *hba);
60 static void emlxs_auth_cfg_fini(emlxs_hba_t *hba);
61 static void emlxs_auth_cfg_read(emlxs_hba_t *hba);
62 static uint32_t emlxs_auth_cfg_parse(emlxs_hba_t *hba,
64 static emlxs_auth_cfg_t *emlxs_auth_cfg_get(emlxs_hba_t *hba,
66 static emlxs_auth_cfg_t *emlxs_auth_cfg_create(emlxs_hba_t *hba,
68 static void emlxs_auth_cfg_destroy(emlxs_hba_t *hba,
70 static void emlxs_auth_cfg_print(emlxs_hba_t *hba,
73 static void emlxs_auth_key_init(emlxs_hba_t *hba);
74 static void emlxs_auth_key_fini(emlxs_hba_t *hba);
[all …]
H A Demlxs_pkt.c45 emlxs_pkt_thread(emlxs_hba_t *hba, void *arg1, void *arg2) in emlxs_pkt_thread() argument
82 emlxs_hba_t *hba = HBA; in emlxs_pkt_send() local
89 emlxs_thread_spawn(hba, emlxs_pkt_thread, (char *)pkt, NULL); in emlxs_pkt_send()
151 emlxs_hba_t *hba = HBA; in emlxs_pkt_alloc() local
200 if (ddi_dma_alloc_handle(hba->dip, &hba->dma_attr_1sg, cb, in emlxs_pkt_alloc()
267 if (ddi_dma_alloc_handle(hba->dip, &hba->dma_attr_1sg, cb, in emlxs_pkt_alloc()
331 if (ddi_dma_alloc_handle(hba->dip, &hba->dma_attr_1sg, cb, in emlxs_pkt_alloc()
389 mutex_init(&sbp->mtx, NULL, MUTEX_DRIVER, DDI_INTR_PRI(hba->intr_arg)); in emlxs_pkt_alloc()
/titanic_44/usr/src/uts/common/sys/fibre-channel/fca/emlxs/
H A Demlxs_extern.h61 extern uint32_t emlxs_msg_log_create(emlxs_hba_t *hba);
62 extern void emlxs_msg_lock_reinit(emlxs_hba_t *hba);
63 extern void emlxs_msg_log_destroy(emlxs_hba_t *hba);
64 extern uint32_t emlxs_msg_log_get(emlxs_hba_t *hba,
71 extern void emlxs_timer_check_events(emlxs_hba_t *hba);
73 extern uint32_t emlxs_event_queue_create(emlxs_hba_t *hba);
75 extern void emlxs_event_queue_destroy(emlxs_hba_t *hba);
150 extern void emlxs_mem_free(emlxs_hba_t *hba,
152 extern uint8_t *emlxs_mem_alloc(emlxs_hba_t *hba,
154 extern int emlxs_map_bus(emlxs_hba_t *hba);
[all …]
H A Demlxs_fc.h321 struct emlxs_hba *hba; /* ptr to hba for channel */ member
371 struct emlxs_hba *hba; /* ptr to hba for ring */ member
518 #define MEM_ELSBUF_COUNT hba->max_nodes
989 struct emlxs_hba *hba; member
1611 #define SLI4_FCOE_MODE (hba->sli.sli4.flag & EMLXS_SLI4_FCOE_MODE)
2097 #define EMLXS_DUMP_LOCK hba->dump_lock
2118 #define EMLXS_SLI_MAP_HDW (hba->sli_api.sli_map_hdw)
2119 #define EMLXS_SLI_UNMAP_HDW (hba->sli_api.sli_unmap_hdw)
2120 #define EMLXS_SLI_ONLINE (hba->sli_api.sli_online)
2121 #define EMLXS_SLI_OFFLINE (hba->sli_api.sli_offline)
[all …]
/titanic_44/usr/src/lib/sun_fc/common/
H A DHandle.cc100 hba = myhba; in Handle()
153 hba = myhba;
332 HBAPort* port = hba->getPortByIndex(index); in getHandlePortByIndex()
350 HBAPort* port = hba->getPort(wwn); in getHandlePort()
351 portHandles[wwn] = new HandlePort(this, hba, port); in getHandlePort()
371 HBA_ADAPTERATTRIBUTES attributes = hba->getHBAAttributes(); in getHBAAttributes()
389 int rval = hba->doForceLip(); in doForceLip()
402 HBA_ADAPTERATTRIBUTES attributes = hba->npivGetHBAAttributes(); in npivGetHBAAttributes()
429 if (hba->containsWWN(wwn)) { in getPortAttributes()
430 attributes = hba->getPort(wwn)->getPortAttributes(tmp); in getPortAttributes()
[all …]

1234