1 /* 2 * Copyright (c) 2005-2010 Brocade Communications Systems, Inc. 3 * All rights reserved 4 * www.brocade.com 5 * 6 * Linux driver for Brocade Fibre Channel Host Bus Adapter. 7 * 8 * This program is free software; you can redistribute it and/or modify it 9 * under the terms of the GNU General Public License (GPL) Version 2 as 10 * published by the Free Software Foundation 11 * 12 * This program is distributed in the hope that it will be useful, but 13 * WITHOUT ANY WARRANTY; without even the implied warranty of 14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 15 * General Public License for more details. 16 */ 17 18 #include "bfad_drv.h" 19 #include "bfad_im.h" 20 #include "bfa_plog.h" 21 #include "bfa_cs.h" 22 #include "bfa_modules.h" 23 24 BFA_TRC_FILE(HAL, FCXP); 25 BFA_MODULE(fcdiag); 26 BFA_MODULE(fcxp); 27 BFA_MODULE(sgpg); 28 BFA_MODULE(lps); 29 BFA_MODULE(fcport); 30 BFA_MODULE(rport); 31 BFA_MODULE(uf); 32 33 /* 34 * LPS related definitions 35 */ 36 #define BFA_LPS_MIN_LPORTS (1) 37 #define BFA_LPS_MAX_LPORTS (256) 38 39 /* 40 * Maximum Vports supported per physical port or vf. 41 */ 42 #define BFA_LPS_MAX_VPORTS_SUPP_CB 255 43 #define BFA_LPS_MAX_VPORTS_SUPP_CT 190 44 45 46 /* 47 * FC PORT related definitions 48 */ 49 /* 50 * The port is considered disabled if corresponding physical port or IOC are 51 * disabled explicitly 52 */ 53 #define BFA_PORT_IS_DISABLED(bfa) \ 54 ((bfa_fcport_is_disabled(bfa) == BFA_TRUE) || \ 55 (bfa_ioc_is_disabled(&bfa->ioc) == BFA_TRUE)) 56 57 /* 58 * BFA port state machine events 59 */ 60 enum bfa_fcport_sm_event { 61 BFA_FCPORT_SM_START = 1, /* start port state machine */ 62 BFA_FCPORT_SM_STOP = 2, /* stop port state machine */ 63 BFA_FCPORT_SM_ENABLE = 3, /* enable port */ 64 BFA_FCPORT_SM_DISABLE = 4, /* disable port state machine */ 65 BFA_FCPORT_SM_FWRSP = 5, /* firmware enable/disable rsp */ 66 BFA_FCPORT_SM_LINKUP = 6, /* firmware linkup event */ 67 BFA_FCPORT_SM_LINKDOWN = 7, /* firmware linkup down */ 68 BFA_FCPORT_SM_QRESUME = 8, /* CQ space available */ 69 BFA_FCPORT_SM_HWFAIL = 9, /* IOC h/w failure */ 70 BFA_FCPORT_SM_DPORTENABLE = 10, /* enable dport */ 71 BFA_FCPORT_SM_DPORTDISABLE = 11,/* disable dport */ 72 BFA_FCPORT_SM_FAA_MISCONFIG = 12, /* FAA misconfiguratin */ 73 BFA_FCPORT_SM_DDPORTENABLE = 13, /* enable ddport */ 74 BFA_FCPORT_SM_DDPORTDISABLE = 14, /* disable ddport */ 75 }; 76 77 /* 78 * BFA port link notification state machine events 79 */ 80 81 enum bfa_fcport_ln_sm_event { 82 BFA_FCPORT_LN_SM_LINKUP = 1, /* linkup event */ 83 BFA_FCPORT_LN_SM_LINKDOWN = 2, /* linkdown event */ 84 BFA_FCPORT_LN_SM_NOTIFICATION = 3 /* done notification */ 85 }; 86 87 /* 88 * RPORT related definitions 89 */ 90 #define bfa_rport_offline_cb(__rp) do { \ 91 if ((__rp)->bfa->fcs) \ 92 bfa_cb_rport_offline((__rp)->rport_drv); \ 93 else { \ 94 bfa_cb_queue((__rp)->bfa, &(__rp)->hcb_qe, \ 95 __bfa_cb_rport_offline, (__rp)); \ 96 } \ 97 } while (0) 98 99 #define bfa_rport_online_cb(__rp) do { \ 100 if ((__rp)->bfa->fcs) \ 101 bfa_cb_rport_online((__rp)->rport_drv); \ 102 else { \ 103 bfa_cb_queue((__rp)->bfa, &(__rp)->hcb_qe, \ 104 __bfa_cb_rport_online, (__rp)); \ 105 } \ 106 } while (0) 107 108 /* 109 * forward declarations FCXP related functions 110 */ 111 static void __bfa_fcxp_send_cbfn(void *cbarg, bfa_boolean_t complete); 112 static void hal_fcxp_rx_plog(struct bfa_s *bfa, struct bfa_fcxp_s *fcxp, 113 struct bfi_fcxp_send_rsp_s *fcxp_rsp); 114 static void hal_fcxp_tx_plog(struct bfa_s *bfa, u32 reqlen, 115 struct bfa_fcxp_s *fcxp, struct fchs_s *fchs); 116 static void bfa_fcxp_qresume(void *cbarg); 117 static void bfa_fcxp_queue(struct bfa_fcxp_s *fcxp, 118 struct bfi_fcxp_send_req_s *send_req); 119 120 /* 121 * forward declarations for LPS functions 122 */ 123 static void bfa_lps_meminfo(struct bfa_iocfc_cfg_s *cfg, 124 struct bfa_meminfo_s *minfo, struct bfa_s *bfa); 125 static void bfa_lps_attach(struct bfa_s *bfa, void *bfad, 126 struct bfa_iocfc_cfg_s *cfg, 127 struct bfa_pcidev_s *pcidev); 128 static void bfa_lps_detach(struct bfa_s *bfa); 129 static void bfa_lps_start(struct bfa_s *bfa); 130 static void bfa_lps_stop(struct bfa_s *bfa); 131 static void bfa_lps_iocdisable(struct bfa_s *bfa); 132 static void bfa_lps_login_rsp(struct bfa_s *bfa, 133 struct bfi_lps_login_rsp_s *rsp); 134 static void bfa_lps_no_res(struct bfa_lps_s *first_lps, u8 count); 135 static void bfa_lps_logout_rsp(struct bfa_s *bfa, 136 struct bfi_lps_logout_rsp_s *rsp); 137 static void bfa_lps_reqq_resume(void *lps_arg); 138 static void bfa_lps_free(struct bfa_lps_s *lps); 139 static void bfa_lps_send_login(struct bfa_lps_s *lps); 140 static void bfa_lps_send_logout(struct bfa_lps_s *lps); 141 static void bfa_lps_send_set_n2n_pid(struct bfa_lps_s *lps); 142 static void bfa_lps_login_comp(struct bfa_lps_s *lps); 143 static void bfa_lps_logout_comp(struct bfa_lps_s *lps); 144 static void bfa_lps_cvl_event(struct bfa_lps_s *lps); 145 146 /* 147 * forward declaration for LPS state machine 148 */ 149 static void bfa_lps_sm_init(struct bfa_lps_s *lps, enum bfa_lps_event event); 150 static void bfa_lps_sm_login(struct bfa_lps_s *lps, enum bfa_lps_event event); 151 static void bfa_lps_sm_loginwait(struct bfa_lps_s *lps, enum bfa_lps_event 152 event); 153 static void bfa_lps_sm_online(struct bfa_lps_s *lps, enum bfa_lps_event event); 154 static void bfa_lps_sm_online_n2n_pid_wait(struct bfa_lps_s *lps, 155 enum bfa_lps_event event); 156 static void bfa_lps_sm_logout(struct bfa_lps_s *lps, enum bfa_lps_event event); 157 static void bfa_lps_sm_logowait(struct bfa_lps_s *lps, enum bfa_lps_event 158 event); 159 160 /* 161 * forward declaration for FC Port functions 162 */ 163 static bfa_boolean_t bfa_fcport_send_enable(struct bfa_fcport_s *fcport); 164 static bfa_boolean_t bfa_fcport_send_disable(struct bfa_fcport_s *fcport); 165 static void bfa_fcport_update_linkinfo(struct bfa_fcport_s *fcport); 166 static void bfa_fcport_reset_linkinfo(struct bfa_fcport_s *fcport); 167 static void bfa_fcport_set_wwns(struct bfa_fcport_s *fcport); 168 static void __bfa_cb_fcport_event(void *cbarg, bfa_boolean_t complete); 169 static void bfa_fcport_scn(struct bfa_fcport_s *fcport, 170 enum bfa_port_linkstate event, bfa_boolean_t trunk); 171 static void bfa_fcport_queue_cb(struct bfa_fcport_ln_s *ln, 172 enum bfa_port_linkstate event); 173 static void __bfa_cb_fcport_stats_clr(void *cbarg, bfa_boolean_t complete); 174 static void bfa_fcport_stats_get_timeout(void *cbarg); 175 static void bfa_fcport_stats_clr_timeout(void *cbarg); 176 static void bfa_trunk_iocdisable(struct bfa_s *bfa); 177 178 /* 179 * forward declaration for FC PORT state machine 180 */ 181 static void bfa_fcport_sm_uninit(struct bfa_fcport_s *fcport, 182 enum bfa_fcport_sm_event event); 183 static void bfa_fcport_sm_enabling_qwait(struct bfa_fcport_s *fcport, 184 enum bfa_fcport_sm_event event); 185 static void bfa_fcport_sm_enabling(struct bfa_fcport_s *fcport, 186 enum bfa_fcport_sm_event event); 187 static void bfa_fcport_sm_linkdown(struct bfa_fcport_s *fcport, 188 enum bfa_fcport_sm_event event); 189 static void bfa_fcport_sm_linkup(struct bfa_fcport_s *fcport, 190 enum bfa_fcport_sm_event event); 191 static void bfa_fcport_sm_disabling(struct bfa_fcport_s *fcport, 192 enum bfa_fcport_sm_event event); 193 static void bfa_fcport_sm_disabling_qwait(struct bfa_fcport_s *fcport, 194 enum bfa_fcport_sm_event event); 195 static void bfa_fcport_sm_toggling_qwait(struct bfa_fcport_s *fcport, 196 enum bfa_fcport_sm_event event); 197 static void bfa_fcport_sm_disabled(struct bfa_fcport_s *fcport, 198 enum bfa_fcport_sm_event event); 199 static void bfa_fcport_sm_stopped(struct bfa_fcport_s *fcport, 200 enum bfa_fcport_sm_event event); 201 static void bfa_fcport_sm_iocdown(struct bfa_fcport_s *fcport, 202 enum bfa_fcport_sm_event event); 203 static void bfa_fcport_sm_iocfail(struct bfa_fcport_s *fcport, 204 enum bfa_fcport_sm_event event); 205 static void bfa_fcport_sm_dport(struct bfa_fcport_s *fcport, 206 enum bfa_fcport_sm_event event); 207 static void bfa_fcport_sm_ddport(struct bfa_fcport_s *fcport, 208 enum bfa_fcport_sm_event event); 209 static void bfa_fcport_sm_faa_misconfig(struct bfa_fcport_s *fcport, 210 enum bfa_fcport_sm_event event); 211 212 static void bfa_fcport_ln_sm_dn(struct bfa_fcport_ln_s *ln, 213 enum bfa_fcport_ln_sm_event event); 214 static void bfa_fcport_ln_sm_dn_nf(struct bfa_fcport_ln_s *ln, 215 enum bfa_fcport_ln_sm_event event); 216 static void bfa_fcport_ln_sm_dn_up_nf(struct bfa_fcport_ln_s *ln, 217 enum bfa_fcport_ln_sm_event event); 218 static void bfa_fcport_ln_sm_up(struct bfa_fcport_ln_s *ln, 219 enum bfa_fcport_ln_sm_event event); 220 static void bfa_fcport_ln_sm_up_nf(struct bfa_fcport_ln_s *ln, 221 enum bfa_fcport_ln_sm_event event); 222 static void bfa_fcport_ln_sm_up_dn_nf(struct bfa_fcport_ln_s *ln, 223 enum bfa_fcport_ln_sm_event event); 224 static void bfa_fcport_ln_sm_up_dn_up_nf(struct bfa_fcport_ln_s *ln, 225 enum bfa_fcport_ln_sm_event event); 226 227 static struct bfa_sm_table_s hal_port_sm_table[] = { 228 {BFA_SM(bfa_fcport_sm_uninit), BFA_PORT_ST_UNINIT}, 229 {BFA_SM(bfa_fcport_sm_enabling_qwait), BFA_PORT_ST_ENABLING_QWAIT}, 230 {BFA_SM(bfa_fcport_sm_enabling), BFA_PORT_ST_ENABLING}, 231 {BFA_SM(bfa_fcport_sm_linkdown), BFA_PORT_ST_LINKDOWN}, 232 {BFA_SM(bfa_fcport_sm_linkup), BFA_PORT_ST_LINKUP}, 233 {BFA_SM(bfa_fcport_sm_disabling_qwait), BFA_PORT_ST_DISABLING_QWAIT}, 234 {BFA_SM(bfa_fcport_sm_toggling_qwait), BFA_PORT_ST_TOGGLING_QWAIT}, 235 {BFA_SM(bfa_fcport_sm_disabling), BFA_PORT_ST_DISABLING}, 236 {BFA_SM(bfa_fcport_sm_disabled), BFA_PORT_ST_DISABLED}, 237 {BFA_SM(bfa_fcport_sm_stopped), BFA_PORT_ST_STOPPED}, 238 {BFA_SM(bfa_fcport_sm_iocdown), BFA_PORT_ST_IOCDOWN}, 239 {BFA_SM(bfa_fcport_sm_iocfail), BFA_PORT_ST_IOCDOWN}, 240 {BFA_SM(bfa_fcport_sm_dport), BFA_PORT_ST_DPORT}, 241 {BFA_SM(bfa_fcport_sm_ddport), BFA_PORT_ST_DDPORT}, 242 {BFA_SM(bfa_fcport_sm_faa_misconfig), BFA_PORT_ST_FAA_MISCONFIG}, 243 }; 244 245 246 /* 247 * forward declaration for RPORT related functions 248 */ 249 static struct bfa_rport_s *bfa_rport_alloc(struct bfa_rport_mod_s *rp_mod); 250 static void bfa_rport_free(struct bfa_rport_s *rport); 251 static bfa_boolean_t bfa_rport_send_fwcreate(struct bfa_rport_s *rp); 252 static bfa_boolean_t bfa_rport_send_fwdelete(struct bfa_rport_s *rp); 253 static bfa_boolean_t bfa_rport_send_fwspeed(struct bfa_rport_s *rp); 254 static void __bfa_cb_rport_online(void *cbarg, 255 bfa_boolean_t complete); 256 static void __bfa_cb_rport_offline(void *cbarg, 257 bfa_boolean_t complete); 258 259 /* 260 * forward declaration for RPORT state machine 261 */ 262 static void bfa_rport_sm_uninit(struct bfa_rport_s *rp, 263 enum bfa_rport_event event); 264 static void bfa_rport_sm_created(struct bfa_rport_s *rp, 265 enum bfa_rport_event event); 266 static void bfa_rport_sm_fwcreate(struct bfa_rport_s *rp, 267 enum bfa_rport_event event); 268 static void bfa_rport_sm_online(struct bfa_rport_s *rp, 269 enum bfa_rport_event event); 270 static void bfa_rport_sm_fwdelete(struct bfa_rport_s *rp, 271 enum bfa_rport_event event); 272 static void bfa_rport_sm_offline(struct bfa_rport_s *rp, 273 enum bfa_rport_event event); 274 static void bfa_rport_sm_deleting(struct bfa_rport_s *rp, 275 enum bfa_rport_event event); 276 static void bfa_rport_sm_offline_pending(struct bfa_rport_s *rp, 277 enum bfa_rport_event event); 278 static void bfa_rport_sm_delete_pending(struct bfa_rport_s *rp, 279 enum bfa_rport_event event); 280 static void bfa_rport_sm_iocdisable(struct bfa_rport_s *rp, 281 enum bfa_rport_event event); 282 static void bfa_rport_sm_fwcreate_qfull(struct bfa_rport_s *rp, 283 enum bfa_rport_event event); 284 static void bfa_rport_sm_fwdelete_qfull(struct bfa_rport_s *rp, 285 enum bfa_rport_event event); 286 static void bfa_rport_sm_deleting_qfull(struct bfa_rport_s *rp, 287 enum bfa_rport_event event); 288 289 /* 290 * PLOG related definitions 291 */ 292 static int 293 plkd_validate_logrec(struct bfa_plog_rec_s *pl_rec) 294 { 295 if ((pl_rec->log_type != BFA_PL_LOG_TYPE_INT) && 296 (pl_rec->log_type != BFA_PL_LOG_TYPE_STRING)) 297 return 1; 298 299 if ((pl_rec->log_type != BFA_PL_LOG_TYPE_INT) && 300 (pl_rec->log_num_ints > BFA_PL_INT_LOG_SZ)) 301 return 1; 302 303 return 0; 304 } 305 306 static u64 307 bfa_get_log_time(void) 308 { 309 u64 system_time = 0; 310 struct timeval tv; 311 do_gettimeofday(&tv); 312 313 /* We are interested in seconds only. */ 314 system_time = tv.tv_sec; 315 return system_time; 316 } 317 318 static void 319 bfa_plog_add(struct bfa_plog_s *plog, struct bfa_plog_rec_s *pl_rec) 320 { 321 u16 tail; 322 struct bfa_plog_rec_s *pl_recp; 323 324 if (plog->plog_enabled == 0) 325 return; 326 327 if (plkd_validate_logrec(pl_rec)) { 328 WARN_ON(1); 329 return; 330 } 331 332 tail = plog->tail; 333 334 pl_recp = &(plog->plog_recs[tail]); 335 336 memcpy(pl_recp, pl_rec, sizeof(struct bfa_plog_rec_s)); 337 338 pl_recp->tv = bfa_get_log_time(); 339 BFA_PL_LOG_REC_INCR(plog->tail); 340 341 if (plog->head == plog->tail) 342 BFA_PL_LOG_REC_INCR(plog->head); 343 } 344 345 void 346 bfa_plog_init(struct bfa_plog_s *plog) 347 { 348 memset((char *)plog, 0, sizeof(struct bfa_plog_s)); 349 350 memcpy(plog->plog_sig, BFA_PL_SIG_STR, BFA_PL_SIG_LEN); 351 plog->head = plog->tail = 0; 352 plog->plog_enabled = 1; 353 } 354 355 void 356 bfa_plog_str(struct bfa_plog_s *plog, enum bfa_plog_mid mid, 357 enum bfa_plog_eid event, 358 u16 misc, char *log_str) 359 { 360 struct bfa_plog_rec_s lp; 361 362 if (plog->plog_enabled) { 363 memset(&lp, 0, sizeof(struct bfa_plog_rec_s)); 364 lp.mid = mid; 365 lp.eid = event; 366 lp.log_type = BFA_PL_LOG_TYPE_STRING; 367 lp.misc = misc; 368 strncpy(lp.log_entry.string_log, log_str, 369 BFA_PL_STRING_LOG_SZ - 1); 370 lp.log_entry.string_log[BFA_PL_STRING_LOG_SZ - 1] = '\0'; 371 bfa_plog_add(plog, &lp); 372 } 373 } 374 375 void 376 bfa_plog_intarr(struct bfa_plog_s *plog, enum bfa_plog_mid mid, 377 enum bfa_plog_eid event, 378 u16 misc, u32 *intarr, u32 num_ints) 379 { 380 struct bfa_plog_rec_s lp; 381 u32 i; 382 383 if (num_ints > BFA_PL_INT_LOG_SZ) 384 num_ints = BFA_PL_INT_LOG_SZ; 385 386 if (plog->plog_enabled) { 387 memset(&lp, 0, sizeof(struct bfa_plog_rec_s)); 388 lp.mid = mid; 389 lp.eid = event; 390 lp.log_type = BFA_PL_LOG_TYPE_INT; 391 lp.misc = misc; 392 393 for (i = 0; i < num_ints; i++) 394 lp.log_entry.int_log[i] = intarr[i]; 395 396 lp.log_num_ints = (u8) num_ints; 397 398 bfa_plog_add(plog, &lp); 399 } 400 } 401 402 void 403 bfa_plog_fchdr(struct bfa_plog_s *plog, enum bfa_plog_mid mid, 404 enum bfa_plog_eid event, 405 u16 misc, struct fchs_s *fchdr) 406 { 407 struct bfa_plog_rec_s lp; 408 u32 *tmp_int = (u32 *) fchdr; 409 u32 ints[BFA_PL_INT_LOG_SZ]; 410 411 if (plog->plog_enabled) { 412 memset(&lp, 0, sizeof(struct bfa_plog_rec_s)); 413 414 ints[0] = tmp_int[0]; 415 ints[1] = tmp_int[1]; 416 ints[2] = tmp_int[4]; 417 418 bfa_plog_intarr(plog, mid, event, misc, ints, 3); 419 } 420 } 421 422 void 423 bfa_plog_fchdr_and_pl(struct bfa_plog_s *plog, enum bfa_plog_mid mid, 424 enum bfa_plog_eid event, u16 misc, struct fchs_s *fchdr, 425 u32 pld_w0) 426 { 427 struct bfa_plog_rec_s lp; 428 u32 *tmp_int = (u32 *) fchdr; 429 u32 ints[BFA_PL_INT_LOG_SZ]; 430 431 if (plog->plog_enabled) { 432 memset(&lp, 0, sizeof(struct bfa_plog_rec_s)); 433 434 ints[0] = tmp_int[0]; 435 ints[1] = tmp_int[1]; 436 ints[2] = tmp_int[4]; 437 ints[3] = pld_w0; 438 439 bfa_plog_intarr(plog, mid, event, misc, ints, 4); 440 } 441 } 442 443 444 /* 445 * fcxp_pvt BFA FCXP private functions 446 */ 447 448 static void 449 claim_fcxps_mem(struct bfa_fcxp_mod_s *mod) 450 { 451 u16 i; 452 struct bfa_fcxp_s *fcxp; 453 454 fcxp = (struct bfa_fcxp_s *) bfa_mem_kva_curp(mod); 455 memset(fcxp, 0, sizeof(struct bfa_fcxp_s) * mod->num_fcxps); 456 457 INIT_LIST_HEAD(&mod->fcxp_req_free_q); 458 INIT_LIST_HEAD(&mod->fcxp_rsp_free_q); 459 INIT_LIST_HEAD(&mod->fcxp_active_q); 460 INIT_LIST_HEAD(&mod->fcxp_req_unused_q); 461 INIT_LIST_HEAD(&mod->fcxp_rsp_unused_q); 462 463 mod->fcxp_list = fcxp; 464 465 for (i = 0; i < mod->num_fcxps; i++) { 466 fcxp->fcxp_mod = mod; 467 fcxp->fcxp_tag = i; 468 469 if (i < (mod->num_fcxps / 2)) { 470 list_add_tail(&fcxp->qe, &mod->fcxp_req_free_q); 471 fcxp->req_rsp = BFA_TRUE; 472 } else { 473 list_add_tail(&fcxp->qe, &mod->fcxp_rsp_free_q); 474 fcxp->req_rsp = BFA_FALSE; 475 } 476 477 bfa_reqq_winit(&fcxp->reqq_wqe, bfa_fcxp_qresume, fcxp); 478 fcxp->reqq_waiting = BFA_FALSE; 479 480 fcxp = fcxp + 1; 481 } 482 483 bfa_mem_kva_curp(mod) = (void *)fcxp; 484 } 485 486 static void 487 bfa_fcxp_meminfo(struct bfa_iocfc_cfg_s *cfg, struct bfa_meminfo_s *minfo, 488 struct bfa_s *bfa) 489 { 490 struct bfa_fcxp_mod_s *fcxp_mod = BFA_FCXP_MOD(bfa); 491 struct bfa_mem_kva_s *fcxp_kva = BFA_MEM_FCXP_KVA(bfa); 492 struct bfa_mem_dma_s *seg_ptr; 493 u16 nsegs, idx, per_seg_fcxp; 494 u16 num_fcxps = cfg->fwcfg.num_fcxp_reqs; 495 u32 per_fcxp_sz; 496 497 if (num_fcxps == 0) 498 return; 499 500 if (cfg->drvcfg.min_cfg) 501 per_fcxp_sz = 2 * BFA_FCXP_MAX_IBUF_SZ; 502 else 503 per_fcxp_sz = BFA_FCXP_MAX_IBUF_SZ + BFA_FCXP_MAX_LBUF_SZ; 504 505 /* dma memory */ 506 nsegs = BFI_MEM_DMA_NSEGS(num_fcxps, per_fcxp_sz); 507 per_seg_fcxp = BFI_MEM_NREQS_SEG(per_fcxp_sz); 508 509 bfa_mem_dma_seg_iter(fcxp_mod, seg_ptr, nsegs, idx) { 510 if (num_fcxps >= per_seg_fcxp) { 511 num_fcxps -= per_seg_fcxp; 512 bfa_mem_dma_setup(minfo, seg_ptr, 513 per_seg_fcxp * per_fcxp_sz); 514 } else 515 bfa_mem_dma_setup(minfo, seg_ptr, 516 num_fcxps * per_fcxp_sz); 517 } 518 519 /* kva memory */ 520 bfa_mem_kva_setup(minfo, fcxp_kva, 521 cfg->fwcfg.num_fcxp_reqs * sizeof(struct bfa_fcxp_s)); 522 } 523 524 static void 525 bfa_fcxp_attach(struct bfa_s *bfa, void *bfad, struct bfa_iocfc_cfg_s *cfg, 526 struct bfa_pcidev_s *pcidev) 527 { 528 struct bfa_fcxp_mod_s *mod = BFA_FCXP_MOD(bfa); 529 530 mod->bfa = bfa; 531 mod->num_fcxps = cfg->fwcfg.num_fcxp_reqs; 532 533 /* 534 * Initialize FCXP request and response payload sizes. 535 */ 536 mod->req_pld_sz = mod->rsp_pld_sz = BFA_FCXP_MAX_IBUF_SZ; 537 if (!cfg->drvcfg.min_cfg) 538 mod->rsp_pld_sz = BFA_FCXP_MAX_LBUF_SZ; 539 540 INIT_LIST_HEAD(&mod->req_wait_q); 541 INIT_LIST_HEAD(&mod->rsp_wait_q); 542 543 claim_fcxps_mem(mod); 544 } 545 546 static void 547 bfa_fcxp_detach(struct bfa_s *bfa) 548 { 549 } 550 551 static void 552 bfa_fcxp_start(struct bfa_s *bfa) 553 { 554 } 555 556 static void 557 bfa_fcxp_stop(struct bfa_s *bfa) 558 { 559 } 560 561 static void 562 bfa_fcxp_iocdisable(struct bfa_s *bfa) 563 { 564 struct bfa_fcxp_mod_s *mod = BFA_FCXP_MOD(bfa); 565 struct bfa_fcxp_s *fcxp; 566 struct list_head *qe, *qen; 567 568 /* Enqueue unused fcxp resources to free_q */ 569 list_splice_tail_init(&mod->fcxp_req_unused_q, &mod->fcxp_req_free_q); 570 list_splice_tail_init(&mod->fcxp_rsp_unused_q, &mod->fcxp_rsp_free_q); 571 572 list_for_each_safe(qe, qen, &mod->fcxp_active_q) { 573 fcxp = (struct bfa_fcxp_s *) qe; 574 if (fcxp->caller == NULL) { 575 fcxp->send_cbfn(fcxp->caller, fcxp, fcxp->send_cbarg, 576 BFA_STATUS_IOC_FAILURE, 0, 0, NULL); 577 bfa_fcxp_free(fcxp); 578 } else { 579 fcxp->rsp_status = BFA_STATUS_IOC_FAILURE; 580 bfa_cb_queue(bfa, &fcxp->hcb_qe, 581 __bfa_fcxp_send_cbfn, fcxp); 582 } 583 } 584 } 585 586 static struct bfa_fcxp_s * 587 bfa_fcxp_get(struct bfa_fcxp_mod_s *fm, bfa_boolean_t req) 588 { 589 struct bfa_fcxp_s *fcxp; 590 591 if (req) 592 bfa_q_deq(&fm->fcxp_req_free_q, &fcxp); 593 else 594 bfa_q_deq(&fm->fcxp_rsp_free_q, &fcxp); 595 596 if (fcxp) 597 list_add_tail(&fcxp->qe, &fm->fcxp_active_q); 598 599 return fcxp; 600 } 601 602 static void 603 bfa_fcxp_init_reqrsp(struct bfa_fcxp_s *fcxp, 604 struct bfa_s *bfa, 605 u8 *use_ibuf, 606 u32 *nr_sgles, 607 bfa_fcxp_get_sgaddr_t *r_sga_cbfn, 608 bfa_fcxp_get_sglen_t *r_sglen_cbfn, 609 struct list_head *r_sgpg_q, 610 int n_sgles, 611 bfa_fcxp_get_sgaddr_t sga_cbfn, 612 bfa_fcxp_get_sglen_t sglen_cbfn) 613 { 614 615 WARN_ON(bfa == NULL); 616 617 bfa_trc(bfa, fcxp->fcxp_tag); 618 619 if (n_sgles == 0) { 620 *use_ibuf = 1; 621 } else { 622 WARN_ON(*sga_cbfn == NULL); 623 WARN_ON(*sglen_cbfn == NULL); 624 625 *use_ibuf = 0; 626 *r_sga_cbfn = sga_cbfn; 627 *r_sglen_cbfn = sglen_cbfn; 628 629 *nr_sgles = n_sgles; 630 631 /* 632 * alloc required sgpgs 633 */ 634 if (n_sgles > BFI_SGE_INLINE) 635 WARN_ON(1); 636 } 637 638 } 639 640 static void 641 bfa_fcxp_init(struct bfa_fcxp_s *fcxp, 642 void *caller, struct bfa_s *bfa, int nreq_sgles, 643 int nrsp_sgles, bfa_fcxp_get_sgaddr_t req_sga_cbfn, 644 bfa_fcxp_get_sglen_t req_sglen_cbfn, 645 bfa_fcxp_get_sgaddr_t rsp_sga_cbfn, 646 bfa_fcxp_get_sglen_t rsp_sglen_cbfn) 647 { 648 649 WARN_ON(bfa == NULL); 650 651 bfa_trc(bfa, fcxp->fcxp_tag); 652 653 fcxp->caller = caller; 654 655 bfa_fcxp_init_reqrsp(fcxp, bfa, 656 &fcxp->use_ireqbuf, &fcxp->nreq_sgles, &fcxp->req_sga_cbfn, 657 &fcxp->req_sglen_cbfn, &fcxp->req_sgpg_q, 658 nreq_sgles, req_sga_cbfn, req_sglen_cbfn); 659 660 bfa_fcxp_init_reqrsp(fcxp, bfa, 661 &fcxp->use_irspbuf, &fcxp->nrsp_sgles, &fcxp->rsp_sga_cbfn, 662 &fcxp->rsp_sglen_cbfn, &fcxp->rsp_sgpg_q, 663 nrsp_sgles, rsp_sga_cbfn, rsp_sglen_cbfn); 664 665 } 666 667 static void 668 bfa_fcxp_put(struct bfa_fcxp_s *fcxp) 669 { 670 struct bfa_fcxp_mod_s *mod = fcxp->fcxp_mod; 671 struct bfa_fcxp_wqe_s *wqe; 672 673 if (fcxp->req_rsp) 674 bfa_q_deq(&mod->req_wait_q, &wqe); 675 else 676 bfa_q_deq(&mod->rsp_wait_q, &wqe); 677 678 if (wqe) { 679 bfa_trc(mod->bfa, fcxp->fcxp_tag); 680 681 bfa_fcxp_init(fcxp, wqe->caller, wqe->bfa, wqe->nreq_sgles, 682 wqe->nrsp_sgles, wqe->req_sga_cbfn, 683 wqe->req_sglen_cbfn, wqe->rsp_sga_cbfn, 684 wqe->rsp_sglen_cbfn); 685 686 wqe->alloc_cbfn(wqe->alloc_cbarg, fcxp); 687 return; 688 } 689 690 WARN_ON(!bfa_q_is_on_q(&mod->fcxp_active_q, fcxp)); 691 list_del(&fcxp->qe); 692 693 if (fcxp->req_rsp) 694 list_add_tail(&fcxp->qe, &mod->fcxp_req_free_q); 695 else 696 list_add_tail(&fcxp->qe, &mod->fcxp_rsp_free_q); 697 } 698 699 static void 700 bfa_fcxp_null_comp(void *bfad_fcxp, struct bfa_fcxp_s *fcxp, void *cbarg, 701 bfa_status_t req_status, u32 rsp_len, 702 u32 resid_len, struct fchs_s *rsp_fchs) 703 { 704 /* discarded fcxp completion */ 705 } 706 707 static void 708 __bfa_fcxp_send_cbfn(void *cbarg, bfa_boolean_t complete) 709 { 710 struct bfa_fcxp_s *fcxp = cbarg; 711 712 if (complete) { 713 fcxp->send_cbfn(fcxp->caller, fcxp, fcxp->send_cbarg, 714 fcxp->rsp_status, fcxp->rsp_len, 715 fcxp->residue_len, &fcxp->rsp_fchs); 716 } else { 717 bfa_fcxp_free(fcxp); 718 } 719 } 720 721 static void 722 hal_fcxp_send_comp(struct bfa_s *bfa, struct bfi_fcxp_send_rsp_s *fcxp_rsp) 723 { 724 struct bfa_fcxp_mod_s *mod = BFA_FCXP_MOD(bfa); 725 struct bfa_fcxp_s *fcxp; 726 u16 fcxp_tag = be16_to_cpu(fcxp_rsp->fcxp_tag); 727 728 bfa_trc(bfa, fcxp_tag); 729 730 fcxp_rsp->rsp_len = be32_to_cpu(fcxp_rsp->rsp_len); 731 732 /* 733 * @todo f/w should not set residue to non-0 when everything 734 * is received. 735 */ 736 if (fcxp_rsp->req_status == BFA_STATUS_OK) 737 fcxp_rsp->residue_len = 0; 738 else 739 fcxp_rsp->residue_len = be32_to_cpu(fcxp_rsp->residue_len); 740 741 fcxp = BFA_FCXP_FROM_TAG(mod, fcxp_tag); 742 743 WARN_ON(fcxp->send_cbfn == NULL); 744 745 hal_fcxp_rx_plog(mod->bfa, fcxp, fcxp_rsp); 746 747 if (fcxp->send_cbfn != NULL) { 748 bfa_trc(mod->bfa, (NULL == fcxp->caller)); 749 if (fcxp->caller == NULL) { 750 fcxp->send_cbfn(fcxp->caller, fcxp, fcxp->send_cbarg, 751 fcxp_rsp->req_status, fcxp_rsp->rsp_len, 752 fcxp_rsp->residue_len, &fcxp_rsp->fchs); 753 /* 754 * fcxp automatically freed on return from the callback 755 */ 756 bfa_fcxp_free(fcxp); 757 } else { 758 fcxp->rsp_status = fcxp_rsp->req_status; 759 fcxp->rsp_len = fcxp_rsp->rsp_len; 760 fcxp->residue_len = fcxp_rsp->residue_len; 761 fcxp->rsp_fchs = fcxp_rsp->fchs; 762 763 bfa_cb_queue(bfa, &fcxp->hcb_qe, 764 __bfa_fcxp_send_cbfn, fcxp); 765 } 766 } else { 767 bfa_trc(bfa, (NULL == fcxp->send_cbfn)); 768 } 769 } 770 771 static void 772 hal_fcxp_tx_plog(struct bfa_s *bfa, u32 reqlen, struct bfa_fcxp_s *fcxp, 773 struct fchs_s *fchs) 774 { 775 /* 776 * TODO: TX ox_id 777 */ 778 if (reqlen > 0) { 779 if (fcxp->use_ireqbuf) { 780 u32 pld_w0 = 781 *((u32 *) BFA_FCXP_REQ_PLD(fcxp)); 782 783 bfa_plog_fchdr_and_pl(bfa->plog, BFA_PL_MID_HAL_FCXP, 784 BFA_PL_EID_TX, 785 reqlen + sizeof(struct fchs_s), fchs, 786 pld_w0); 787 } else { 788 bfa_plog_fchdr(bfa->plog, BFA_PL_MID_HAL_FCXP, 789 BFA_PL_EID_TX, 790 reqlen + sizeof(struct fchs_s), 791 fchs); 792 } 793 } else { 794 bfa_plog_fchdr(bfa->plog, BFA_PL_MID_HAL_FCXP, BFA_PL_EID_TX, 795 reqlen + sizeof(struct fchs_s), fchs); 796 } 797 } 798 799 static void 800 hal_fcxp_rx_plog(struct bfa_s *bfa, struct bfa_fcxp_s *fcxp, 801 struct bfi_fcxp_send_rsp_s *fcxp_rsp) 802 { 803 if (fcxp_rsp->rsp_len > 0) { 804 if (fcxp->use_irspbuf) { 805 u32 pld_w0 = 806 *((u32 *) BFA_FCXP_RSP_PLD(fcxp)); 807 808 bfa_plog_fchdr_and_pl(bfa->plog, BFA_PL_MID_HAL_FCXP, 809 BFA_PL_EID_RX, 810 (u16) fcxp_rsp->rsp_len, 811 &fcxp_rsp->fchs, pld_w0); 812 } else { 813 bfa_plog_fchdr(bfa->plog, BFA_PL_MID_HAL_FCXP, 814 BFA_PL_EID_RX, 815 (u16) fcxp_rsp->rsp_len, 816 &fcxp_rsp->fchs); 817 } 818 } else { 819 bfa_plog_fchdr(bfa->plog, BFA_PL_MID_HAL_FCXP, BFA_PL_EID_RX, 820 (u16) fcxp_rsp->rsp_len, &fcxp_rsp->fchs); 821 } 822 } 823 824 /* 825 * Handler to resume sending fcxp when space in available in cpe queue. 826 */ 827 static void 828 bfa_fcxp_qresume(void *cbarg) 829 { 830 struct bfa_fcxp_s *fcxp = cbarg; 831 struct bfa_s *bfa = fcxp->fcxp_mod->bfa; 832 struct bfi_fcxp_send_req_s *send_req; 833 834 fcxp->reqq_waiting = BFA_FALSE; 835 send_req = bfa_reqq_next(bfa, BFA_REQQ_FCXP); 836 bfa_fcxp_queue(fcxp, send_req); 837 } 838 839 /* 840 * Queue fcxp send request to foimrware. 841 */ 842 static void 843 bfa_fcxp_queue(struct bfa_fcxp_s *fcxp, struct bfi_fcxp_send_req_s *send_req) 844 { 845 struct bfa_s *bfa = fcxp->fcxp_mod->bfa; 846 struct bfa_fcxp_req_info_s *reqi = &fcxp->req_info; 847 struct bfa_fcxp_rsp_info_s *rspi = &fcxp->rsp_info; 848 struct bfa_rport_s *rport = reqi->bfa_rport; 849 850 bfi_h2i_set(send_req->mh, BFI_MC_FCXP, BFI_FCXP_H2I_SEND_REQ, 851 bfa_fn_lpu(bfa)); 852 853 send_req->fcxp_tag = cpu_to_be16(fcxp->fcxp_tag); 854 if (rport) { 855 send_req->rport_fw_hndl = rport->fw_handle; 856 send_req->max_frmsz = cpu_to_be16(rport->rport_info.max_frmsz); 857 if (send_req->max_frmsz == 0) 858 send_req->max_frmsz = cpu_to_be16(FC_MAX_PDUSZ); 859 } else { 860 send_req->rport_fw_hndl = 0; 861 send_req->max_frmsz = cpu_to_be16(FC_MAX_PDUSZ); 862 } 863 864 send_req->vf_id = cpu_to_be16(reqi->vf_id); 865 send_req->lp_fwtag = bfa_lps_get_fwtag(bfa, reqi->lp_tag); 866 send_req->class = reqi->class; 867 send_req->rsp_timeout = rspi->rsp_timeout; 868 send_req->cts = reqi->cts; 869 send_req->fchs = reqi->fchs; 870 871 send_req->req_len = cpu_to_be32(reqi->req_tot_len); 872 send_req->rsp_maxlen = cpu_to_be32(rspi->rsp_maxlen); 873 874 /* 875 * setup req sgles 876 */ 877 if (fcxp->use_ireqbuf == 1) { 878 bfa_alen_set(&send_req->req_alen, reqi->req_tot_len, 879 BFA_FCXP_REQ_PLD_PA(fcxp)); 880 } else { 881 if (fcxp->nreq_sgles > 0) { 882 WARN_ON(fcxp->nreq_sgles != 1); 883 bfa_alen_set(&send_req->req_alen, reqi->req_tot_len, 884 fcxp->req_sga_cbfn(fcxp->caller, 0)); 885 } else { 886 WARN_ON(reqi->req_tot_len != 0); 887 bfa_alen_set(&send_req->rsp_alen, 0, 0); 888 } 889 } 890 891 /* 892 * setup rsp sgles 893 */ 894 if (fcxp->use_irspbuf == 1) { 895 WARN_ON(rspi->rsp_maxlen > BFA_FCXP_MAX_LBUF_SZ); 896 897 bfa_alen_set(&send_req->rsp_alen, rspi->rsp_maxlen, 898 BFA_FCXP_RSP_PLD_PA(fcxp)); 899 } else { 900 if (fcxp->nrsp_sgles > 0) { 901 WARN_ON(fcxp->nrsp_sgles != 1); 902 bfa_alen_set(&send_req->rsp_alen, rspi->rsp_maxlen, 903 fcxp->rsp_sga_cbfn(fcxp->caller, 0)); 904 905 } else { 906 WARN_ON(rspi->rsp_maxlen != 0); 907 bfa_alen_set(&send_req->rsp_alen, 0, 0); 908 } 909 } 910 911 hal_fcxp_tx_plog(bfa, reqi->req_tot_len, fcxp, &reqi->fchs); 912 913 bfa_reqq_produce(bfa, BFA_REQQ_FCXP, send_req->mh); 914 915 bfa_trc(bfa, bfa_reqq_pi(bfa, BFA_REQQ_FCXP)); 916 bfa_trc(bfa, bfa_reqq_ci(bfa, BFA_REQQ_FCXP)); 917 } 918 919 /* 920 * Allocate an FCXP instance to send a response or to send a request 921 * that has a response. Request/response buffers are allocated by caller. 922 * 923 * @param[in] bfa BFA bfa instance 924 * @param[in] nreq_sgles Number of SG elements required for request 925 * buffer. 0, if fcxp internal buffers are used. 926 * Use bfa_fcxp_get_reqbuf() to get the 927 * internal req buffer. 928 * @param[in] req_sgles SG elements describing request buffer. Will be 929 * copied in by BFA and hence can be freed on 930 * return from this function. 931 * @param[in] get_req_sga function ptr to be called to get a request SG 932 * Address (given the sge index). 933 * @param[in] get_req_sglen function ptr to be called to get a request SG 934 * len (given the sge index). 935 * @param[in] get_rsp_sga function ptr to be called to get a response SG 936 * Address (given the sge index). 937 * @param[in] get_rsp_sglen function ptr to be called to get a response SG 938 * len (given the sge index). 939 * @param[in] req Allocated FCXP is used to send req or rsp? 940 * request - BFA_TRUE, response - BFA_FALSE 941 * 942 * @return FCXP instance. NULL on failure. 943 */ 944 struct bfa_fcxp_s * 945 bfa_fcxp_req_rsp_alloc(void *caller, struct bfa_s *bfa, int nreq_sgles, 946 int nrsp_sgles, bfa_fcxp_get_sgaddr_t req_sga_cbfn, 947 bfa_fcxp_get_sglen_t req_sglen_cbfn, 948 bfa_fcxp_get_sgaddr_t rsp_sga_cbfn, 949 bfa_fcxp_get_sglen_t rsp_sglen_cbfn, bfa_boolean_t req) 950 { 951 struct bfa_fcxp_s *fcxp = NULL; 952 953 WARN_ON(bfa == NULL); 954 955 fcxp = bfa_fcxp_get(BFA_FCXP_MOD(bfa), req); 956 if (fcxp == NULL) 957 return NULL; 958 959 bfa_trc(bfa, fcxp->fcxp_tag); 960 961 bfa_fcxp_init(fcxp, caller, bfa, nreq_sgles, nrsp_sgles, req_sga_cbfn, 962 req_sglen_cbfn, rsp_sga_cbfn, rsp_sglen_cbfn); 963 964 return fcxp; 965 } 966 967 /* 968 * Get the internal request buffer pointer 969 * 970 * @param[in] fcxp BFA fcxp pointer 971 * 972 * @return pointer to the internal request buffer 973 */ 974 void * 975 bfa_fcxp_get_reqbuf(struct bfa_fcxp_s *fcxp) 976 { 977 struct bfa_fcxp_mod_s *mod = fcxp->fcxp_mod; 978 void *reqbuf; 979 980 WARN_ON(fcxp->use_ireqbuf != 1); 981 reqbuf = bfa_mem_get_dmabuf_kva(mod, fcxp->fcxp_tag, 982 mod->req_pld_sz + mod->rsp_pld_sz); 983 return reqbuf; 984 } 985 986 u32 987 bfa_fcxp_get_reqbufsz(struct bfa_fcxp_s *fcxp) 988 { 989 struct bfa_fcxp_mod_s *mod = fcxp->fcxp_mod; 990 991 return mod->req_pld_sz; 992 } 993 994 /* 995 * Get the internal response buffer pointer 996 * 997 * @param[in] fcxp BFA fcxp pointer 998 * 999 * @return pointer to the internal request buffer 1000 */ 1001 void * 1002 bfa_fcxp_get_rspbuf(struct bfa_fcxp_s *fcxp) 1003 { 1004 struct bfa_fcxp_mod_s *mod = fcxp->fcxp_mod; 1005 void *fcxp_buf; 1006 1007 WARN_ON(fcxp->use_irspbuf != 1); 1008 1009 fcxp_buf = bfa_mem_get_dmabuf_kva(mod, fcxp->fcxp_tag, 1010 mod->req_pld_sz + mod->rsp_pld_sz); 1011 1012 /* fcxp_buf = req_buf + rsp_buf :- add req_buf_sz to get to rsp_buf */ 1013 return ((u8 *) fcxp_buf) + mod->req_pld_sz; 1014 } 1015 1016 /* 1017 * Free the BFA FCXP 1018 * 1019 * @param[in] fcxp BFA fcxp pointer 1020 * 1021 * @return void 1022 */ 1023 void 1024 bfa_fcxp_free(struct bfa_fcxp_s *fcxp) 1025 { 1026 struct bfa_fcxp_mod_s *mod = fcxp->fcxp_mod; 1027 1028 WARN_ON(fcxp == NULL); 1029 bfa_trc(mod->bfa, fcxp->fcxp_tag); 1030 bfa_fcxp_put(fcxp); 1031 } 1032 1033 /* 1034 * Send a FCXP request 1035 * 1036 * @param[in] fcxp BFA fcxp pointer 1037 * @param[in] rport BFA rport pointer. Could be left NULL for WKA rports 1038 * @param[in] vf_id virtual Fabric ID 1039 * @param[in] lp_tag lport tag 1040 * @param[in] cts use Continuous sequence 1041 * @param[in] cos fc Class of Service 1042 * @param[in] reqlen request length, does not include FCHS length 1043 * @param[in] fchs fc Header Pointer. The header content will be copied 1044 * in by BFA. 1045 * 1046 * @param[in] cbfn call back function to be called on receiving 1047 * the response 1048 * @param[in] cbarg arg for cbfn 1049 * @param[in] rsp_timeout 1050 * response timeout 1051 * 1052 * @return bfa_status_t 1053 */ 1054 void 1055 bfa_fcxp_send(struct bfa_fcxp_s *fcxp, struct bfa_rport_s *rport, 1056 u16 vf_id, u8 lp_tag, bfa_boolean_t cts, enum fc_cos cos, 1057 u32 reqlen, struct fchs_s *fchs, bfa_cb_fcxp_send_t cbfn, 1058 void *cbarg, u32 rsp_maxlen, u8 rsp_timeout) 1059 { 1060 struct bfa_s *bfa = fcxp->fcxp_mod->bfa; 1061 struct bfa_fcxp_req_info_s *reqi = &fcxp->req_info; 1062 struct bfa_fcxp_rsp_info_s *rspi = &fcxp->rsp_info; 1063 struct bfi_fcxp_send_req_s *send_req; 1064 1065 bfa_trc(bfa, fcxp->fcxp_tag); 1066 1067 /* 1068 * setup request/response info 1069 */ 1070 reqi->bfa_rport = rport; 1071 reqi->vf_id = vf_id; 1072 reqi->lp_tag = lp_tag; 1073 reqi->class = cos; 1074 rspi->rsp_timeout = rsp_timeout; 1075 reqi->cts = cts; 1076 reqi->fchs = *fchs; 1077 reqi->req_tot_len = reqlen; 1078 rspi->rsp_maxlen = rsp_maxlen; 1079 fcxp->send_cbfn = cbfn ? cbfn : bfa_fcxp_null_comp; 1080 fcxp->send_cbarg = cbarg; 1081 1082 /* 1083 * If no room in CPE queue, wait for space in request queue 1084 */ 1085 send_req = bfa_reqq_next(bfa, BFA_REQQ_FCXP); 1086 if (!send_req) { 1087 bfa_trc(bfa, fcxp->fcxp_tag); 1088 fcxp->reqq_waiting = BFA_TRUE; 1089 bfa_reqq_wait(bfa, BFA_REQQ_FCXP, &fcxp->reqq_wqe); 1090 return; 1091 } 1092 1093 bfa_fcxp_queue(fcxp, send_req); 1094 } 1095 1096 /* 1097 * Abort a BFA FCXP 1098 * 1099 * @param[in] fcxp BFA fcxp pointer 1100 * 1101 * @return void 1102 */ 1103 bfa_status_t 1104 bfa_fcxp_abort(struct bfa_fcxp_s *fcxp) 1105 { 1106 bfa_trc(fcxp->fcxp_mod->bfa, fcxp->fcxp_tag); 1107 WARN_ON(1); 1108 return BFA_STATUS_OK; 1109 } 1110 1111 void 1112 bfa_fcxp_req_rsp_alloc_wait(struct bfa_s *bfa, struct bfa_fcxp_wqe_s *wqe, 1113 bfa_fcxp_alloc_cbfn_t alloc_cbfn, void *alloc_cbarg, 1114 void *caller, int nreq_sgles, 1115 int nrsp_sgles, bfa_fcxp_get_sgaddr_t req_sga_cbfn, 1116 bfa_fcxp_get_sglen_t req_sglen_cbfn, 1117 bfa_fcxp_get_sgaddr_t rsp_sga_cbfn, 1118 bfa_fcxp_get_sglen_t rsp_sglen_cbfn, bfa_boolean_t req) 1119 { 1120 struct bfa_fcxp_mod_s *mod = BFA_FCXP_MOD(bfa); 1121 1122 if (req) 1123 WARN_ON(!list_empty(&mod->fcxp_req_free_q)); 1124 else 1125 WARN_ON(!list_empty(&mod->fcxp_rsp_free_q)); 1126 1127 wqe->alloc_cbfn = alloc_cbfn; 1128 wqe->alloc_cbarg = alloc_cbarg; 1129 wqe->caller = caller; 1130 wqe->bfa = bfa; 1131 wqe->nreq_sgles = nreq_sgles; 1132 wqe->nrsp_sgles = nrsp_sgles; 1133 wqe->req_sga_cbfn = req_sga_cbfn; 1134 wqe->req_sglen_cbfn = req_sglen_cbfn; 1135 wqe->rsp_sga_cbfn = rsp_sga_cbfn; 1136 wqe->rsp_sglen_cbfn = rsp_sglen_cbfn; 1137 1138 if (req) 1139 list_add_tail(&wqe->qe, &mod->req_wait_q); 1140 else 1141 list_add_tail(&wqe->qe, &mod->rsp_wait_q); 1142 } 1143 1144 void 1145 bfa_fcxp_walloc_cancel(struct bfa_s *bfa, struct bfa_fcxp_wqe_s *wqe) 1146 { 1147 struct bfa_fcxp_mod_s *mod = BFA_FCXP_MOD(bfa); 1148 1149 WARN_ON(!bfa_q_is_on_q(&mod->req_wait_q, wqe) || 1150 !bfa_q_is_on_q(&mod->rsp_wait_q, wqe)); 1151 list_del(&wqe->qe); 1152 } 1153 1154 void 1155 bfa_fcxp_discard(struct bfa_fcxp_s *fcxp) 1156 { 1157 /* 1158 * If waiting for room in request queue, cancel reqq wait 1159 * and free fcxp. 1160 */ 1161 if (fcxp->reqq_waiting) { 1162 fcxp->reqq_waiting = BFA_FALSE; 1163 bfa_reqq_wcancel(&fcxp->reqq_wqe); 1164 bfa_fcxp_free(fcxp); 1165 return; 1166 } 1167 1168 fcxp->send_cbfn = bfa_fcxp_null_comp; 1169 } 1170 1171 void 1172 bfa_fcxp_isr(struct bfa_s *bfa, struct bfi_msg_s *msg) 1173 { 1174 switch (msg->mhdr.msg_id) { 1175 case BFI_FCXP_I2H_SEND_RSP: 1176 hal_fcxp_send_comp(bfa, (struct bfi_fcxp_send_rsp_s *) msg); 1177 break; 1178 1179 default: 1180 bfa_trc(bfa, msg->mhdr.msg_id); 1181 WARN_ON(1); 1182 } 1183 } 1184 1185 u32 1186 bfa_fcxp_get_maxrsp(struct bfa_s *bfa) 1187 { 1188 struct bfa_fcxp_mod_s *mod = BFA_FCXP_MOD(bfa); 1189 1190 return mod->rsp_pld_sz; 1191 } 1192 1193 void 1194 bfa_fcxp_res_recfg(struct bfa_s *bfa, u16 num_fcxp_fw) 1195 { 1196 struct bfa_fcxp_mod_s *mod = BFA_FCXP_MOD(bfa); 1197 struct list_head *qe; 1198 int i; 1199 1200 for (i = 0; i < (mod->num_fcxps - num_fcxp_fw); i++) { 1201 if (i < ((mod->num_fcxps - num_fcxp_fw) / 2)) { 1202 bfa_q_deq_tail(&mod->fcxp_req_free_q, &qe); 1203 list_add_tail(qe, &mod->fcxp_req_unused_q); 1204 } else { 1205 bfa_q_deq_tail(&mod->fcxp_rsp_free_q, &qe); 1206 list_add_tail(qe, &mod->fcxp_rsp_unused_q); 1207 } 1208 } 1209 } 1210 1211 /* 1212 * BFA LPS state machine functions 1213 */ 1214 1215 /* 1216 * Init state -- no login 1217 */ 1218 static void 1219 bfa_lps_sm_init(struct bfa_lps_s *lps, enum bfa_lps_event event) 1220 { 1221 bfa_trc(lps->bfa, lps->bfa_tag); 1222 bfa_trc(lps->bfa, event); 1223 1224 switch (event) { 1225 case BFA_LPS_SM_LOGIN: 1226 if (bfa_reqq_full(lps->bfa, lps->reqq)) { 1227 bfa_sm_set_state(lps, bfa_lps_sm_loginwait); 1228 bfa_reqq_wait(lps->bfa, lps->reqq, &lps->wqe); 1229 } else { 1230 bfa_sm_set_state(lps, bfa_lps_sm_login); 1231 bfa_lps_send_login(lps); 1232 } 1233 1234 if (lps->fdisc) 1235 bfa_plog_str(lps->bfa->plog, BFA_PL_MID_LPS, 1236 BFA_PL_EID_LOGIN, 0, "FDISC Request"); 1237 else 1238 bfa_plog_str(lps->bfa->plog, BFA_PL_MID_LPS, 1239 BFA_PL_EID_LOGIN, 0, "FLOGI Request"); 1240 break; 1241 1242 case BFA_LPS_SM_LOGOUT: 1243 bfa_lps_logout_comp(lps); 1244 break; 1245 1246 case BFA_LPS_SM_DELETE: 1247 bfa_lps_free(lps); 1248 break; 1249 1250 case BFA_LPS_SM_RX_CVL: 1251 case BFA_LPS_SM_OFFLINE: 1252 break; 1253 1254 case BFA_LPS_SM_FWRSP: 1255 /* 1256 * Could happen when fabric detects loopback and discards 1257 * the lps request. Fw will eventually sent out the timeout 1258 * Just ignore 1259 */ 1260 break; 1261 case BFA_LPS_SM_SET_N2N_PID: 1262 /* 1263 * When topology is set to loop, bfa_lps_set_n2n_pid() sends 1264 * this event. Ignore this event. 1265 */ 1266 break; 1267 1268 default: 1269 bfa_sm_fault(lps->bfa, event); 1270 } 1271 } 1272 1273 /* 1274 * login is in progress -- awaiting response from firmware 1275 */ 1276 static void 1277 bfa_lps_sm_login(struct bfa_lps_s *lps, enum bfa_lps_event event) 1278 { 1279 bfa_trc(lps->bfa, lps->bfa_tag); 1280 bfa_trc(lps->bfa, event); 1281 1282 switch (event) { 1283 case BFA_LPS_SM_FWRSP: 1284 if (lps->status == BFA_STATUS_OK) { 1285 bfa_sm_set_state(lps, bfa_lps_sm_online); 1286 if (lps->fdisc) 1287 bfa_plog_str(lps->bfa->plog, BFA_PL_MID_LPS, 1288 BFA_PL_EID_LOGIN, 0, "FDISC Accept"); 1289 else 1290 bfa_plog_str(lps->bfa->plog, BFA_PL_MID_LPS, 1291 BFA_PL_EID_LOGIN, 0, "FLOGI Accept"); 1292 /* If N2N, send the assigned PID to FW */ 1293 bfa_trc(lps->bfa, lps->fport); 1294 bfa_trc(lps->bfa, lps->lp_pid); 1295 1296 if (!lps->fport && lps->lp_pid) 1297 bfa_sm_send_event(lps, BFA_LPS_SM_SET_N2N_PID); 1298 } else { 1299 bfa_sm_set_state(lps, bfa_lps_sm_init); 1300 if (lps->fdisc) 1301 bfa_plog_str(lps->bfa->plog, BFA_PL_MID_LPS, 1302 BFA_PL_EID_LOGIN, 0, 1303 "FDISC Fail (RJT or timeout)"); 1304 else 1305 bfa_plog_str(lps->bfa->plog, BFA_PL_MID_LPS, 1306 BFA_PL_EID_LOGIN, 0, 1307 "FLOGI Fail (RJT or timeout)"); 1308 } 1309 bfa_lps_login_comp(lps); 1310 break; 1311 1312 case BFA_LPS_SM_OFFLINE: 1313 case BFA_LPS_SM_DELETE: 1314 bfa_sm_set_state(lps, bfa_lps_sm_init); 1315 break; 1316 1317 case BFA_LPS_SM_SET_N2N_PID: 1318 bfa_trc(lps->bfa, lps->fport); 1319 bfa_trc(lps->bfa, lps->lp_pid); 1320 break; 1321 1322 default: 1323 bfa_sm_fault(lps->bfa, event); 1324 } 1325 } 1326 1327 /* 1328 * login pending - awaiting space in request queue 1329 */ 1330 static void 1331 bfa_lps_sm_loginwait(struct bfa_lps_s *lps, enum bfa_lps_event event) 1332 { 1333 bfa_trc(lps->bfa, lps->bfa_tag); 1334 bfa_trc(lps->bfa, event); 1335 1336 switch (event) { 1337 case BFA_LPS_SM_RESUME: 1338 bfa_sm_set_state(lps, bfa_lps_sm_login); 1339 bfa_lps_send_login(lps); 1340 break; 1341 1342 case BFA_LPS_SM_OFFLINE: 1343 case BFA_LPS_SM_DELETE: 1344 bfa_sm_set_state(lps, bfa_lps_sm_init); 1345 bfa_reqq_wcancel(&lps->wqe); 1346 break; 1347 1348 case BFA_LPS_SM_RX_CVL: 1349 /* 1350 * Login was not even sent out; so when getting out 1351 * of this state, it will appear like a login retry 1352 * after Clear virtual link 1353 */ 1354 break; 1355 1356 default: 1357 bfa_sm_fault(lps->bfa, event); 1358 } 1359 } 1360 1361 /* 1362 * login complete 1363 */ 1364 static void 1365 bfa_lps_sm_online(struct bfa_lps_s *lps, enum bfa_lps_event event) 1366 { 1367 bfa_trc(lps->bfa, lps->bfa_tag); 1368 bfa_trc(lps->bfa, event); 1369 1370 switch (event) { 1371 case BFA_LPS_SM_LOGOUT: 1372 if (bfa_reqq_full(lps->bfa, lps->reqq)) { 1373 bfa_sm_set_state(lps, bfa_lps_sm_logowait); 1374 bfa_reqq_wait(lps->bfa, lps->reqq, &lps->wqe); 1375 } else { 1376 bfa_sm_set_state(lps, bfa_lps_sm_logout); 1377 bfa_lps_send_logout(lps); 1378 } 1379 bfa_plog_str(lps->bfa->plog, BFA_PL_MID_LPS, 1380 BFA_PL_EID_LOGO, 0, "Logout"); 1381 break; 1382 1383 case BFA_LPS_SM_RX_CVL: 1384 bfa_sm_set_state(lps, bfa_lps_sm_init); 1385 1386 /* Let the vport module know about this event */ 1387 bfa_lps_cvl_event(lps); 1388 bfa_plog_str(lps->bfa->plog, BFA_PL_MID_LPS, 1389 BFA_PL_EID_FIP_FCF_CVL, 0, "FCF Clear Virt. Link Rx"); 1390 break; 1391 1392 case BFA_LPS_SM_SET_N2N_PID: 1393 if (bfa_reqq_full(lps->bfa, lps->reqq)) { 1394 bfa_sm_set_state(lps, bfa_lps_sm_online_n2n_pid_wait); 1395 bfa_reqq_wait(lps->bfa, lps->reqq, &lps->wqe); 1396 } else 1397 bfa_lps_send_set_n2n_pid(lps); 1398 break; 1399 1400 case BFA_LPS_SM_OFFLINE: 1401 case BFA_LPS_SM_DELETE: 1402 bfa_sm_set_state(lps, bfa_lps_sm_init); 1403 break; 1404 1405 default: 1406 bfa_sm_fault(lps->bfa, event); 1407 } 1408 } 1409 1410 /* 1411 * login complete 1412 */ 1413 static void 1414 bfa_lps_sm_online_n2n_pid_wait(struct bfa_lps_s *lps, enum bfa_lps_event event) 1415 { 1416 bfa_trc(lps->bfa, lps->bfa_tag); 1417 bfa_trc(lps->bfa, event); 1418 1419 switch (event) { 1420 case BFA_LPS_SM_RESUME: 1421 bfa_sm_set_state(lps, bfa_lps_sm_online); 1422 bfa_lps_send_set_n2n_pid(lps); 1423 break; 1424 1425 case BFA_LPS_SM_LOGOUT: 1426 bfa_sm_set_state(lps, bfa_lps_sm_logowait); 1427 bfa_plog_str(lps->bfa->plog, BFA_PL_MID_LPS, 1428 BFA_PL_EID_LOGO, 0, "Logout"); 1429 break; 1430 1431 case BFA_LPS_SM_RX_CVL: 1432 bfa_sm_set_state(lps, bfa_lps_sm_init); 1433 bfa_reqq_wcancel(&lps->wqe); 1434 1435 /* Let the vport module know about this event */ 1436 bfa_lps_cvl_event(lps); 1437 bfa_plog_str(lps->bfa->plog, BFA_PL_MID_LPS, 1438 BFA_PL_EID_FIP_FCF_CVL, 0, "FCF Clear Virt. Link Rx"); 1439 break; 1440 1441 case BFA_LPS_SM_OFFLINE: 1442 case BFA_LPS_SM_DELETE: 1443 bfa_sm_set_state(lps, bfa_lps_sm_init); 1444 bfa_reqq_wcancel(&lps->wqe); 1445 break; 1446 1447 default: 1448 bfa_sm_fault(lps->bfa, event); 1449 } 1450 } 1451 1452 /* 1453 * logout in progress - awaiting firmware response 1454 */ 1455 static void 1456 bfa_lps_sm_logout(struct bfa_lps_s *lps, enum bfa_lps_event event) 1457 { 1458 bfa_trc(lps->bfa, lps->bfa_tag); 1459 bfa_trc(lps->bfa, event); 1460 1461 switch (event) { 1462 case BFA_LPS_SM_FWRSP: 1463 case BFA_LPS_SM_OFFLINE: 1464 bfa_sm_set_state(lps, bfa_lps_sm_init); 1465 bfa_lps_logout_comp(lps); 1466 break; 1467 1468 case BFA_LPS_SM_DELETE: 1469 bfa_sm_set_state(lps, bfa_lps_sm_init); 1470 break; 1471 1472 default: 1473 bfa_sm_fault(lps->bfa, event); 1474 } 1475 } 1476 1477 /* 1478 * logout pending -- awaiting space in request queue 1479 */ 1480 static void 1481 bfa_lps_sm_logowait(struct bfa_lps_s *lps, enum bfa_lps_event event) 1482 { 1483 bfa_trc(lps->bfa, lps->bfa_tag); 1484 bfa_trc(lps->bfa, event); 1485 1486 switch (event) { 1487 case BFA_LPS_SM_RESUME: 1488 bfa_sm_set_state(lps, bfa_lps_sm_logout); 1489 bfa_lps_send_logout(lps); 1490 break; 1491 1492 case BFA_LPS_SM_OFFLINE: 1493 case BFA_LPS_SM_DELETE: 1494 bfa_sm_set_state(lps, bfa_lps_sm_init); 1495 bfa_reqq_wcancel(&lps->wqe); 1496 break; 1497 1498 default: 1499 bfa_sm_fault(lps->bfa, event); 1500 } 1501 } 1502 1503 1504 1505 /* 1506 * lps_pvt BFA LPS private functions 1507 */ 1508 1509 /* 1510 * return memory requirement 1511 */ 1512 static void 1513 bfa_lps_meminfo(struct bfa_iocfc_cfg_s *cfg, struct bfa_meminfo_s *minfo, 1514 struct bfa_s *bfa) 1515 { 1516 struct bfa_mem_kva_s *lps_kva = BFA_MEM_LPS_KVA(bfa); 1517 1518 if (cfg->drvcfg.min_cfg) 1519 bfa_mem_kva_setup(minfo, lps_kva, 1520 sizeof(struct bfa_lps_s) * BFA_LPS_MIN_LPORTS); 1521 else 1522 bfa_mem_kva_setup(minfo, lps_kva, 1523 sizeof(struct bfa_lps_s) * BFA_LPS_MAX_LPORTS); 1524 } 1525 1526 /* 1527 * bfa module attach at initialization time 1528 */ 1529 static void 1530 bfa_lps_attach(struct bfa_s *bfa, void *bfad, struct bfa_iocfc_cfg_s *cfg, 1531 struct bfa_pcidev_s *pcidev) 1532 { 1533 struct bfa_lps_mod_s *mod = BFA_LPS_MOD(bfa); 1534 struct bfa_lps_s *lps; 1535 int i; 1536 1537 mod->num_lps = BFA_LPS_MAX_LPORTS; 1538 if (cfg->drvcfg.min_cfg) 1539 mod->num_lps = BFA_LPS_MIN_LPORTS; 1540 else 1541 mod->num_lps = BFA_LPS_MAX_LPORTS; 1542 mod->lps_arr = lps = (struct bfa_lps_s *) bfa_mem_kva_curp(mod); 1543 1544 bfa_mem_kva_curp(mod) += mod->num_lps * sizeof(struct bfa_lps_s); 1545 1546 INIT_LIST_HEAD(&mod->lps_free_q); 1547 INIT_LIST_HEAD(&mod->lps_active_q); 1548 INIT_LIST_HEAD(&mod->lps_login_q); 1549 1550 for (i = 0; i < mod->num_lps; i++, lps++) { 1551 lps->bfa = bfa; 1552 lps->bfa_tag = (u8) i; 1553 lps->reqq = BFA_REQQ_LPS; 1554 bfa_reqq_winit(&lps->wqe, bfa_lps_reqq_resume, lps); 1555 list_add_tail(&lps->qe, &mod->lps_free_q); 1556 } 1557 } 1558 1559 static void 1560 bfa_lps_detach(struct bfa_s *bfa) 1561 { 1562 } 1563 1564 static void 1565 bfa_lps_start(struct bfa_s *bfa) 1566 { 1567 } 1568 1569 static void 1570 bfa_lps_stop(struct bfa_s *bfa) 1571 { 1572 } 1573 1574 /* 1575 * IOC in disabled state -- consider all lps offline 1576 */ 1577 static void 1578 bfa_lps_iocdisable(struct bfa_s *bfa) 1579 { 1580 struct bfa_lps_mod_s *mod = BFA_LPS_MOD(bfa); 1581 struct bfa_lps_s *lps; 1582 struct list_head *qe, *qen; 1583 1584 list_for_each_safe(qe, qen, &mod->lps_active_q) { 1585 lps = (struct bfa_lps_s *) qe; 1586 bfa_sm_send_event(lps, BFA_LPS_SM_OFFLINE); 1587 } 1588 list_for_each_safe(qe, qen, &mod->lps_login_q) { 1589 lps = (struct bfa_lps_s *) qe; 1590 bfa_sm_send_event(lps, BFA_LPS_SM_OFFLINE); 1591 } 1592 list_splice_tail_init(&mod->lps_login_q, &mod->lps_active_q); 1593 } 1594 1595 /* 1596 * Firmware login response 1597 */ 1598 static void 1599 bfa_lps_login_rsp(struct bfa_s *bfa, struct bfi_lps_login_rsp_s *rsp) 1600 { 1601 struct bfa_lps_mod_s *mod = BFA_LPS_MOD(bfa); 1602 struct bfa_lps_s *lps; 1603 1604 WARN_ON(rsp->bfa_tag >= mod->num_lps); 1605 lps = BFA_LPS_FROM_TAG(mod, rsp->bfa_tag); 1606 1607 lps->status = rsp->status; 1608 switch (rsp->status) { 1609 case BFA_STATUS_OK: 1610 lps->fw_tag = rsp->fw_tag; 1611 lps->fport = rsp->f_port; 1612 if (lps->fport) 1613 lps->lp_pid = rsp->lp_pid; 1614 lps->npiv_en = rsp->npiv_en; 1615 lps->pr_bbcred = be16_to_cpu(rsp->bb_credit); 1616 lps->pr_pwwn = rsp->port_name; 1617 lps->pr_nwwn = rsp->node_name; 1618 lps->auth_req = rsp->auth_req; 1619 lps->lp_mac = rsp->lp_mac; 1620 lps->brcd_switch = rsp->brcd_switch; 1621 lps->fcf_mac = rsp->fcf_mac; 1622 1623 break; 1624 1625 case BFA_STATUS_FABRIC_RJT: 1626 lps->lsrjt_rsn = rsp->lsrjt_rsn; 1627 lps->lsrjt_expl = rsp->lsrjt_expl; 1628 1629 break; 1630 1631 case BFA_STATUS_EPROTOCOL: 1632 lps->ext_status = rsp->ext_status; 1633 1634 break; 1635 1636 case BFA_STATUS_VPORT_MAX: 1637 if (rsp->ext_status) 1638 bfa_lps_no_res(lps, rsp->ext_status); 1639 break; 1640 1641 default: 1642 /* Nothing to do with other status */ 1643 break; 1644 } 1645 1646 list_del(&lps->qe); 1647 list_add_tail(&lps->qe, &mod->lps_active_q); 1648 bfa_sm_send_event(lps, BFA_LPS_SM_FWRSP); 1649 } 1650 1651 static void 1652 bfa_lps_no_res(struct bfa_lps_s *first_lps, u8 count) 1653 { 1654 struct bfa_s *bfa = first_lps->bfa; 1655 struct bfa_lps_mod_s *mod = BFA_LPS_MOD(bfa); 1656 struct list_head *qe, *qe_next; 1657 struct bfa_lps_s *lps; 1658 1659 bfa_trc(bfa, count); 1660 1661 qe = bfa_q_next(first_lps); 1662 1663 while (count && qe) { 1664 qe_next = bfa_q_next(qe); 1665 lps = (struct bfa_lps_s *)qe; 1666 bfa_trc(bfa, lps->bfa_tag); 1667 lps->status = first_lps->status; 1668 list_del(&lps->qe); 1669 list_add_tail(&lps->qe, &mod->lps_active_q); 1670 bfa_sm_send_event(lps, BFA_LPS_SM_FWRSP); 1671 qe = qe_next; 1672 count--; 1673 } 1674 } 1675 1676 /* 1677 * Firmware logout response 1678 */ 1679 static void 1680 bfa_lps_logout_rsp(struct bfa_s *bfa, struct bfi_lps_logout_rsp_s *rsp) 1681 { 1682 struct bfa_lps_mod_s *mod = BFA_LPS_MOD(bfa); 1683 struct bfa_lps_s *lps; 1684 1685 WARN_ON(rsp->bfa_tag >= mod->num_lps); 1686 lps = BFA_LPS_FROM_TAG(mod, rsp->bfa_tag); 1687 1688 bfa_sm_send_event(lps, BFA_LPS_SM_FWRSP); 1689 } 1690 1691 /* 1692 * Firmware received a Clear virtual link request (for FCoE) 1693 */ 1694 static void 1695 bfa_lps_rx_cvl_event(struct bfa_s *bfa, struct bfi_lps_cvl_event_s *cvl) 1696 { 1697 struct bfa_lps_mod_s *mod = BFA_LPS_MOD(bfa); 1698 struct bfa_lps_s *lps; 1699 1700 lps = BFA_LPS_FROM_TAG(mod, cvl->bfa_tag); 1701 1702 bfa_sm_send_event(lps, BFA_LPS_SM_RX_CVL); 1703 } 1704 1705 /* 1706 * Space is available in request queue, resume queueing request to firmware. 1707 */ 1708 static void 1709 bfa_lps_reqq_resume(void *lps_arg) 1710 { 1711 struct bfa_lps_s *lps = lps_arg; 1712 1713 bfa_sm_send_event(lps, BFA_LPS_SM_RESUME); 1714 } 1715 1716 /* 1717 * lps is freed -- triggered by vport delete 1718 */ 1719 static void 1720 bfa_lps_free(struct bfa_lps_s *lps) 1721 { 1722 struct bfa_lps_mod_s *mod = BFA_LPS_MOD(lps->bfa); 1723 1724 lps->lp_pid = 0; 1725 list_del(&lps->qe); 1726 list_add_tail(&lps->qe, &mod->lps_free_q); 1727 } 1728 1729 /* 1730 * send login request to firmware 1731 */ 1732 static void 1733 bfa_lps_send_login(struct bfa_lps_s *lps) 1734 { 1735 struct bfa_lps_mod_s *mod = BFA_LPS_MOD(lps->bfa); 1736 struct bfi_lps_login_req_s *m; 1737 1738 m = bfa_reqq_next(lps->bfa, lps->reqq); 1739 WARN_ON(!m); 1740 1741 bfi_h2i_set(m->mh, BFI_MC_LPS, BFI_LPS_H2I_LOGIN_REQ, 1742 bfa_fn_lpu(lps->bfa)); 1743 1744 m->bfa_tag = lps->bfa_tag; 1745 m->alpa = lps->alpa; 1746 m->pdu_size = cpu_to_be16(lps->pdusz); 1747 m->pwwn = lps->pwwn; 1748 m->nwwn = lps->nwwn; 1749 m->fdisc = lps->fdisc; 1750 m->auth_en = lps->auth_en; 1751 1752 bfa_reqq_produce(lps->bfa, lps->reqq, m->mh); 1753 list_del(&lps->qe); 1754 list_add_tail(&lps->qe, &mod->lps_login_q); 1755 } 1756 1757 /* 1758 * send logout request to firmware 1759 */ 1760 static void 1761 bfa_lps_send_logout(struct bfa_lps_s *lps) 1762 { 1763 struct bfi_lps_logout_req_s *m; 1764 1765 m = bfa_reqq_next(lps->bfa, lps->reqq); 1766 WARN_ON(!m); 1767 1768 bfi_h2i_set(m->mh, BFI_MC_LPS, BFI_LPS_H2I_LOGOUT_REQ, 1769 bfa_fn_lpu(lps->bfa)); 1770 1771 m->fw_tag = lps->fw_tag; 1772 m->port_name = lps->pwwn; 1773 bfa_reqq_produce(lps->bfa, lps->reqq, m->mh); 1774 } 1775 1776 /* 1777 * send n2n pid set request to firmware 1778 */ 1779 static void 1780 bfa_lps_send_set_n2n_pid(struct bfa_lps_s *lps) 1781 { 1782 struct bfi_lps_n2n_pid_req_s *m; 1783 1784 m = bfa_reqq_next(lps->bfa, lps->reqq); 1785 WARN_ON(!m); 1786 1787 bfi_h2i_set(m->mh, BFI_MC_LPS, BFI_LPS_H2I_N2N_PID_REQ, 1788 bfa_fn_lpu(lps->bfa)); 1789 1790 m->fw_tag = lps->fw_tag; 1791 m->lp_pid = lps->lp_pid; 1792 bfa_reqq_produce(lps->bfa, lps->reqq, m->mh); 1793 } 1794 1795 /* 1796 * Indirect login completion handler for non-fcs 1797 */ 1798 static void 1799 bfa_lps_login_comp_cb(void *arg, bfa_boolean_t complete) 1800 { 1801 struct bfa_lps_s *lps = arg; 1802 1803 if (!complete) 1804 return; 1805 1806 if (lps->fdisc) 1807 bfa_cb_lps_fdisc_comp(lps->bfa->bfad, lps->uarg, lps->status); 1808 else 1809 bfa_cb_lps_flogi_comp(lps->bfa->bfad, lps->uarg, lps->status); 1810 } 1811 1812 /* 1813 * Login completion handler -- direct call for fcs, queue for others 1814 */ 1815 static void 1816 bfa_lps_login_comp(struct bfa_lps_s *lps) 1817 { 1818 if (!lps->bfa->fcs) { 1819 bfa_cb_queue(lps->bfa, &lps->hcb_qe, bfa_lps_login_comp_cb, 1820 lps); 1821 return; 1822 } 1823 1824 if (lps->fdisc) 1825 bfa_cb_lps_fdisc_comp(lps->bfa->bfad, lps->uarg, lps->status); 1826 else 1827 bfa_cb_lps_flogi_comp(lps->bfa->bfad, lps->uarg, lps->status); 1828 } 1829 1830 /* 1831 * Indirect logout completion handler for non-fcs 1832 */ 1833 static void 1834 bfa_lps_logout_comp_cb(void *arg, bfa_boolean_t complete) 1835 { 1836 struct bfa_lps_s *lps = arg; 1837 1838 if (!complete) 1839 return; 1840 1841 if (lps->fdisc) 1842 bfa_cb_lps_fdisclogo_comp(lps->bfa->bfad, lps->uarg); 1843 else 1844 bfa_cb_lps_flogo_comp(lps->bfa->bfad, lps->uarg); 1845 } 1846 1847 /* 1848 * Logout completion handler -- direct call for fcs, queue for others 1849 */ 1850 static void 1851 bfa_lps_logout_comp(struct bfa_lps_s *lps) 1852 { 1853 if (!lps->bfa->fcs) { 1854 bfa_cb_queue(lps->bfa, &lps->hcb_qe, bfa_lps_logout_comp_cb, 1855 lps); 1856 return; 1857 } 1858 if (lps->fdisc) 1859 bfa_cb_lps_fdisclogo_comp(lps->bfa->bfad, lps->uarg); 1860 } 1861 1862 /* 1863 * Clear virtual link completion handler for non-fcs 1864 */ 1865 static void 1866 bfa_lps_cvl_event_cb(void *arg, bfa_boolean_t complete) 1867 { 1868 struct bfa_lps_s *lps = arg; 1869 1870 if (!complete) 1871 return; 1872 1873 /* Clear virtual link to base port will result in link down */ 1874 if (lps->fdisc) 1875 bfa_cb_lps_cvl_event(lps->bfa->bfad, lps->uarg); 1876 } 1877 1878 /* 1879 * Received Clear virtual link event --direct call for fcs, 1880 * queue for others 1881 */ 1882 static void 1883 bfa_lps_cvl_event(struct bfa_lps_s *lps) 1884 { 1885 if (!lps->bfa->fcs) { 1886 bfa_cb_queue(lps->bfa, &lps->hcb_qe, bfa_lps_cvl_event_cb, 1887 lps); 1888 return; 1889 } 1890 1891 /* Clear virtual link to base port will result in link down */ 1892 if (lps->fdisc) 1893 bfa_cb_lps_cvl_event(lps->bfa->bfad, lps->uarg); 1894 } 1895 1896 1897 1898 /* 1899 * lps_public BFA LPS public functions 1900 */ 1901 1902 u32 1903 bfa_lps_get_max_vport(struct bfa_s *bfa) 1904 { 1905 if (bfa_ioc_devid(&bfa->ioc) == BFA_PCI_DEVICE_ID_CT) 1906 return BFA_LPS_MAX_VPORTS_SUPP_CT; 1907 else 1908 return BFA_LPS_MAX_VPORTS_SUPP_CB; 1909 } 1910 1911 /* 1912 * Allocate a lport srvice tag. 1913 */ 1914 struct bfa_lps_s * 1915 bfa_lps_alloc(struct bfa_s *bfa) 1916 { 1917 struct bfa_lps_mod_s *mod = BFA_LPS_MOD(bfa); 1918 struct bfa_lps_s *lps = NULL; 1919 1920 bfa_q_deq(&mod->lps_free_q, &lps); 1921 1922 if (lps == NULL) 1923 return NULL; 1924 1925 list_add_tail(&lps->qe, &mod->lps_active_q); 1926 1927 bfa_sm_set_state(lps, bfa_lps_sm_init); 1928 return lps; 1929 } 1930 1931 /* 1932 * Free lport service tag. This can be called anytime after an alloc. 1933 * No need to wait for any pending login/logout completions. 1934 */ 1935 void 1936 bfa_lps_delete(struct bfa_lps_s *lps) 1937 { 1938 bfa_sm_send_event(lps, BFA_LPS_SM_DELETE); 1939 } 1940 1941 /* 1942 * Initiate a lport login. 1943 */ 1944 void 1945 bfa_lps_flogi(struct bfa_lps_s *lps, void *uarg, u8 alpa, u16 pdusz, 1946 wwn_t pwwn, wwn_t nwwn, bfa_boolean_t auth_en) 1947 { 1948 lps->uarg = uarg; 1949 lps->alpa = alpa; 1950 lps->pdusz = pdusz; 1951 lps->pwwn = pwwn; 1952 lps->nwwn = nwwn; 1953 lps->fdisc = BFA_FALSE; 1954 lps->auth_en = auth_en; 1955 bfa_sm_send_event(lps, BFA_LPS_SM_LOGIN); 1956 } 1957 1958 /* 1959 * Initiate a lport fdisc login. 1960 */ 1961 void 1962 bfa_lps_fdisc(struct bfa_lps_s *lps, void *uarg, u16 pdusz, wwn_t pwwn, 1963 wwn_t nwwn) 1964 { 1965 lps->uarg = uarg; 1966 lps->alpa = 0; 1967 lps->pdusz = pdusz; 1968 lps->pwwn = pwwn; 1969 lps->nwwn = nwwn; 1970 lps->fdisc = BFA_TRUE; 1971 lps->auth_en = BFA_FALSE; 1972 bfa_sm_send_event(lps, BFA_LPS_SM_LOGIN); 1973 } 1974 1975 1976 /* 1977 * Initiate a lport FDSIC logout. 1978 */ 1979 void 1980 bfa_lps_fdisclogo(struct bfa_lps_s *lps) 1981 { 1982 bfa_sm_send_event(lps, BFA_LPS_SM_LOGOUT); 1983 } 1984 1985 u8 1986 bfa_lps_get_fwtag(struct bfa_s *bfa, u8 lp_tag) 1987 { 1988 struct bfa_lps_mod_s *mod = BFA_LPS_MOD(bfa); 1989 1990 return BFA_LPS_FROM_TAG(mod, lp_tag)->fw_tag; 1991 } 1992 1993 /* 1994 * Return lport services tag given the pid 1995 */ 1996 u8 1997 bfa_lps_get_tag_from_pid(struct bfa_s *bfa, u32 pid) 1998 { 1999 struct bfa_lps_mod_s *mod = BFA_LPS_MOD(bfa); 2000 struct bfa_lps_s *lps; 2001 int i; 2002 2003 for (i = 0, lps = mod->lps_arr; i < mod->num_lps; i++, lps++) { 2004 if (lps->lp_pid == pid) 2005 return lps->bfa_tag; 2006 } 2007 2008 /* Return base port tag anyway */ 2009 return 0; 2010 } 2011 2012 2013 /* 2014 * return port id assigned to the base lport 2015 */ 2016 u32 2017 bfa_lps_get_base_pid(struct bfa_s *bfa) 2018 { 2019 struct bfa_lps_mod_s *mod = BFA_LPS_MOD(bfa); 2020 2021 return BFA_LPS_FROM_TAG(mod, 0)->lp_pid; 2022 } 2023 2024 /* 2025 * Set PID in case of n2n (which is assigned during PLOGI) 2026 */ 2027 void 2028 bfa_lps_set_n2n_pid(struct bfa_lps_s *lps, uint32_t n2n_pid) 2029 { 2030 bfa_trc(lps->bfa, lps->bfa_tag); 2031 bfa_trc(lps->bfa, n2n_pid); 2032 2033 lps->lp_pid = n2n_pid; 2034 bfa_sm_send_event(lps, BFA_LPS_SM_SET_N2N_PID); 2035 } 2036 2037 /* 2038 * LPS firmware message class handler. 2039 */ 2040 void 2041 bfa_lps_isr(struct bfa_s *bfa, struct bfi_msg_s *m) 2042 { 2043 union bfi_lps_i2h_msg_u msg; 2044 2045 bfa_trc(bfa, m->mhdr.msg_id); 2046 msg.msg = m; 2047 2048 switch (m->mhdr.msg_id) { 2049 case BFI_LPS_I2H_LOGIN_RSP: 2050 bfa_lps_login_rsp(bfa, msg.login_rsp); 2051 break; 2052 2053 case BFI_LPS_I2H_LOGOUT_RSP: 2054 bfa_lps_logout_rsp(bfa, msg.logout_rsp); 2055 break; 2056 2057 case BFI_LPS_I2H_CVL_EVENT: 2058 bfa_lps_rx_cvl_event(bfa, msg.cvl_event); 2059 break; 2060 2061 default: 2062 bfa_trc(bfa, m->mhdr.msg_id); 2063 WARN_ON(1); 2064 } 2065 } 2066 2067 static void 2068 bfa_fcport_aen_post(struct bfa_fcport_s *fcport, enum bfa_port_aen_event event) 2069 { 2070 struct bfad_s *bfad = (struct bfad_s *)fcport->bfa->bfad; 2071 struct bfa_aen_entry_s *aen_entry; 2072 2073 bfad_get_aen_entry(bfad, aen_entry); 2074 if (!aen_entry) 2075 return; 2076 2077 aen_entry->aen_data.port.ioc_type = bfa_get_type(fcport->bfa); 2078 aen_entry->aen_data.port.pwwn = fcport->pwwn; 2079 2080 /* Send the AEN notification */ 2081 bfad_im_post_vendor_event(aen_entry, bfad, ++fcport->bfa->bfa_aen_seq, 2082 BFA_AEN_CAT_PORT, event); 2083 } 2084 2085 /* 2086 * FC PORT state machine functions 2087 */ 2088 static void 2089 bfa_fcport_sm_uninit(struct bfa_fcport_s *fcport, 2090 enum bfa_fcport_sm_event event) 2091 { 2092 bfa_trc(fcport->bfa, event); 2093 2094 switch (event) { 2095 case BFA_FCPORT_SM_START: 2096 /* 2097 * Start event after IOC is configured and BFA is started. 2098 */ 2099 fcport->use_flash_cfg = BFA_TRUE; 2100 2101 if (bfa_fcport_send_enable(fcport)) { 2102 bfa_trc(fcport->bfa, BFA_TRUE); 2103 bfa_sm_set_state(fcport, bfa_fcport_sm_enabling); 2104 } else { 2105 bfa_trc(fcport->bfa, BFA_FALSE); 2106 bfa_sm_set_state(fcport, 2107 bfa_fcport_sm_enabling_qwait); 2108 } 2109 break; 2110 2111 case BFA_FCPORT_SM_ENABLE: 2112 /* 2113 * Port is persistently configured to be in enabled state. Do 2114 * not change state. Port enabling is done when START event is 2115 * received. 2116 */ 2117 break; 2118 2119 case BFA_FCPORT_SM_DISABLE: 2120 /* 2121 * If a port is persistently configured to be disabled, the 2122 * first event will a port disable request. 2123 */ 2124 bfa_sm_set_state(fcport, bfa_fcport_sm_disabled); 2125 break; 2126 2127 case BFA_FCPORT_SM_HWFAIL: 2128 bfa_sm_set_state(fcport, bfa_fcport_sm_iocdown); 2129 break; 2130 2131 default: 2132 bfa_sm_fault(fcport->bfa, event); 2133 } 2134 } 2135 2136 static void 2137 bfa_fcport_sm_enabling_qwait(struct bfa_fcport_s *fcport, 2138 enum bfa_fcport_sm_event event) 2139 { 2140 char pwwn_buf[BFA_STRING_32]; 2141 struct bfad_s *bfad = (struct bfad_s *)fcport->bfa->bfad; 2142 bfa_trc(fcport->bfa, event); 2143 2144 switch (event) { 2145 case BFA_FCPORT_SM_QRESUME: 2146 bfa_sm_set_state(fcport, bfa_fcport_sm_enabling); 2147 bfa_fcport_send_enable(fcport); 2148 break; 2149 2150 case BFA_FCPORT_SM_STOP: 2151 bfa_reqq_wcancel(&fcport->reqq_wait); 2152 bfa_sm_set_state(fcport, bfa_fcport_sm_stopped); 2153 break; 2154 2155 case BFA_FCPORT_SM_ENABLE: 2156 /* 2157 * Already enable is in progress. 2158 */ 2159 break; 2160 2161 case BFA_FCPORT_SM_DISABLE: 2162 /* 2163 * Just send disable request to firmware when room becomes 2164 * available in request queue. 2165 */ 2166 bfa_sm_set_state(fcport, bfa_fcport_sm_disabled); 2167 bfa_reqq_wcancel(&fcport->reqq_wait); 2168 bfa_plog_str(fcport->bfa->plog, BFA_PL_MID_HAL, 2169 BFA_PL_EID_PORT_DISABLE, 0, "Port Disable"); 2170 wwn2str(pwwn_buf, fcport->pwwn); 2171 BFA_LOG(KERN_INFO, bfad, bfa_log_level, 2172 "Base port disabled: WWN = %s\n", pwwn_buf); 2173 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_DISABLE); 2174 break; 2175 2176 case BFA_FCPORT_SM_LINKUP: 2177 case BFA_FCPORT_SM_LINKDOWN: 2178 /* 2179 * Possible to get link events when doing back-to-back 2180 * enable/disables. 2181 */ 2182 break; 2183 2184 case BFA_FCPORT_SM_HWFAIL: 2185 bfa_reqq_wcancel(&fcport->reqq_wait); 2186 bfa_sm_set_state(fcport, bfa_fcport_sm_iocdown); 2187 break; 2188 2189 case BFA_FCPORT_SM_FAA_MISCONFIG: 2190 bfa_fcport_reset_linkinfo(fcport); 2191 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_DISCONNECT); 2192 bfa_sm_set_state(fcport, bfa_fcport_sm_faa_misconfig); 2193 break; 2194 2195 default: 2196 bfa_sm_fault(fcport->bfa, event); 2197 } 2198 } 2199 2200 static void 2201 bfa_fcport_sm_enabling(struct bfa_fcport_s *fcport, 2202 enum bfa_fcport_sm_event event) 2203 { 2204 char pwwn_buf[BFA_STRING_32]; 2205 struct bfad_s *bfad = (struct bfad_s *)fcport->bfa->bfad; 2206 bfa_trc(fcport->bfa, event); 2207 2208 switch (event) { 2209 case BFA_FCPORT_SM_FWRSP: 2210 case BFA_FCPORT_SM_LINKDOWN: 2211 bfa_sm_set_state(fcport, bfa_fcport_sm_linkdown); 2212 break; 2213 2214 case BFA_FCPORT_SM_LINKUP: 2215 bfa_fcport_update_linkinfo(fcport); 2216 bfa_sm_set_state(fcport, bfa_fcport_sm_linkup); 2217 2218 WARN_ON(!fcport->event_cbfn); 2219 bfa_fcport_scn(fcport, BFA_PORT_LINKUP, BFA_FALSE); 2220 break; 2221 2222 case BFA_FCPORT_SM_ENABLE: 2223 /* 2224 * Already being enabled. 2225 */ 2226 break; 2227 2228 case BFA_FCPORT_SM_DISABLE: 2229 if (bfa_fcport_send_disable(fcport)) 2230 bfa_sm_set_state(fcport, bfa_fcport_sm_disabling); 2231 else 2232 bfa_sm_set_state(fcport, 2233 bfa_fcport_sm_disabling_qwait); 2234 2235 bfa_plog_str(fcport->bfa->plog, BFA_PL_MID_HAL, 2236 BFA_PL_EID_PORT_DISABLE, 0, "Port Disable"); 2237 wwn2str(pwwn_buf, fcport->pwwn); 2238 BFA_LOG(KERN_INFO, bfad, bfa_log_level, 2239 "Base port disabled: WWN = %s\n", pwwn_buf); 2240 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_DISABLE); 2241 break; 2242 2243 case BFA_FCPORT_SM_STOP: 2244 bfa_sm_set_state(fcport, bfa_fcport_sm_stopped); 2245 break; 2246 2247 case BFA_FCPORT_SM_HWFAIL: 2248 bfa_sm_set_state(fcport, bfa_fcport_sm_iocdown); 2249 break; 2250 2251 case BFA_FCPORT_SM_FAA_MISCONFIG: 2252 bfa_fcport_reset_linkinfo(fcport); 2253 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_DISCONNECT); 2254 bfa_sm_set_state(fcport, bfa_fcport_sm_faa_misconfig); 2255 break; 2256 2257 default: 2258 bfa_sm_fault(fcport->bfa, event); 2259 } 2260 } 2261 2262 static void 2263 bfa_fcport_sm_linkdown(struct bfa_fcport_s *fcport, 2264 enum bfa_fcport_sm_event event) 2265 { 2266 struct bfi_fcport_event_s *pevent = fcport->event_arg.i2hmsg.event; 2267 char pwwn_buf[BFA_STRING_32]; 2268 struct bfad_s *bfad = (struct bfad_s *)fcport->bfa->bfad; 2269 2270 bfa_trc(fcport->bfa, event); 2271 2272 switch (event) { 2273 case BFA_FCPORT_SM_LINKUP: 2274 bfa_fcport_update_linkinfo(fcport); 2275 bfa_sm_set_state(fcport, bfa_fcport_sm_linkup); 2276 WARN_ON(!fcport->event_cbfn); 2277 bfa_plog_str(fcport->bfa->plog, BFA_PL_MID_HAL, 2278 BFA_PL_EID_PORT_ST_CHANGE, 0, "Port Linkup"); 2279 if (!bfa_ioc_get_fcmode(&fcport->bfa->ioc)) { 2280 2281 bfa_trc(fcport->bfa, 2282 pevent->link_state.attr.vc_fcf.fcf.fipenabled); 2283 bfa_trc(fcport->bfa, 2284 pevent->link_state.attr.vc_fcf.fcf.fipfailed); 2285 2286 if (pevent->link_state.attr.vc_fcf.fcf.fipfailed) 2287 bfa_plog_str(fcport->bfa->plog, BFA_PL_MID_HAL, 2288 BFA_PL_EID_FIP_FCF_DISC, 0, 2289 "FIP FCF Discovery Failed"); 2290 else 2291 bfa_plog_str(fcport->bfa->plog, BFA_PL_MID_HAL, 2292 BFA_PL_EID_FIP_FCF_DISC, 0, 2293 "FIP FCF Discovered"); 2294 } 2295 2296 bfa_fcport_scn(fcport, BFA_PORT_LINKUP, BFA_FALSE); 2297 wwn2str(pwwn_buf, fcport->pwwn); 2298 BFA_LOG(KERN_INFO, bfad, bfa_log_level, 2299 "Base port online: WWN = %s\n", pwwn_buf); 2300 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_ONLINE); 2301 2302 /* If QoS is enabled and it is not online, send AEN */ 2303 if (fcport->cfg.qos_enabled && 2304 fcport->qos_attr.state != BFA_QOS_ONLINE) 2305 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_QOS_NEG); 2306 break; 2307 2308 case BFA_FCPORT_SM_LINKDOWN: 2309 /* 2310 * Possible to get link down event. 2311 */ 2312 break; 2313 2314 case BFA_FCPORT_SM_ENABLE: 2315 /* 2316 * Already enabled. 2317 */ 2318 break; 2319 2320 case BFA_FCPORT_SM_DISABLE: 2321 if (bfa_fcport_send_disable(fcport)) 2322 bfa_sm_set_state(fcport, bfa_fcport_sm_disabling); 2323 else 2324 bfa_sm_set_state(fcport, 2325 bfa_fcport_sm_disabling_qwait); 2326 2327 bfa_plog_str(fcport->bfa->plog, BFA_PL_MID_HAL, 2328 BFA_PL_EID_PORT_DISABLE, 0, "Port Disable"); 2329 wwn2str(pwwn_buf, fcport->pwwn); 2330 BFA_LOG(KERN_INFO, bfad, bfa_log_level, 2331 "Base port disabled: WWN = %s\n", pwwn_buf); 2332 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_DISABLE); 2333 break; 2334 2335 case BFA_FCPORT_SM_STOP: 2336 bfa_sm_set_state(fcport, bfa_fcport_sm_stopped); 2337 break; 2338 2339 case BFA_FCPORT_SM_HWFAIL: 2340 bfa_sm_set_state(fcport, bfa_fcport_sm_iocdown); 2341 break; 2342 2343 case BFA_FCPORT_SM_FAA_MISCONFIG: 2344 bfa_fcport_reset_linkinfo(fcport); 2345 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_DISCONNECT); 2346 bfa_sm_set_state(fcport, bfa_fcport_sm_faa_misconfig); 2347 break; 2348 2349 default: 2350 bfa_sm_fault(fcport->bfa, event); 2351 } 2352 } 2353 2354 static void 2355 bfa_fcport_sm_linkup(struct bfa_fcport_s *fcport, 2356 enum bfa_fcport_sm_event event) 2357 { 2358 char pwwn_buf[BFA_STRING_32]; 2359 struct bfad_s *bfad = (struct bfad_s *)fcport->bfa->bfad; 2360 2361 bfa_trc(fcport->bfa, event); 2362 2363 switch (event) { 2364 case BFA_FCPORT_SM_ENABLE: 2365 /* 2366 * Already enabled. 2367 */ 2368 break; 2369 2370 case BFA_FCPORT_SM_DISABLE: 2371 if (bfa_fcport_send_disable(fcport)) 2372 bfa_sm_set_state(fcport, bfa_fcport_sm_disabling); 2373 else 2374 bfa_sm_set_state(fcport, 2375 bfa_fcport_sm_disabling_qwait); 2376 2377 bfa_fcport_reset_linkinfo(fcport); 2378 bfa_fcport_scn(fcport, BFA_PORT_LINKDOWN, BFA_FALSE); 2379 bfa_plog_str(fcport->bfa->plog, BFA_PL_MID_HAL, 2380 BFA_PL_EID_PORT_DISABLE, 0, "Port Disable"); 2381 wwn2str(pwwn_buf, fcport->pwwn); 2382 BFA_LOG(KERN_INFO, bfad, bfa_log_level, 2383 "Base port offline: WWN = %s\n", pwwn_buf); 2384 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_OFFLINE); 2385 BFA_LOG(KERN_INFO, bfad, bfa_log_level, 2386 "Base port disabled: WWN = %s\n", pwwn_buf); 2387 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_DISABLE); 2388 break; 2389 2390 case BFA_FCPORT_SM_LINKDOWN: 2391 bfa_sm_set_state(fcport, bfa_fcport_sm_linkdown); 2392 bfa_fcport_reset_linkinfo(fcport); 2393 bfa_fcport_scn(fcport, BFA_PORT_LINKDOWN, BFA_FALSE); 2394 bfa_plog_str(fcport->bfa->plog, BFA_PL_MID_HAL, 2395 BFA_PL_EID_PORT_ST_CHANGE, 0, "Port Linkdown"); 2396 wwn2str(pwwn_buf, fcport->pwwn); 2397 if (BFA_PORT_IS_DISABLED(fcport->bfa)) { 2398 BFA_LOG(KERN_INFO, bfad, bfa_log_level, 2399 "Base port offline: WWN = %s\n", pwwn_buf); 2400 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_OFFLINE); 2401 } else { 2402 BFA_LOG(KERN_ERR, bfad, bfa_log_level, 2403 "Base port (WWN = %s) " 2404 "lost fabric connectivity\n", pwwn_buf); 2405 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_DISCONNECT); 2406 } 2407 break; 2408 2409 case BFA_FCPORT_SM_STOP: 2410 bfa_sm_set_state(fcport, bfa_fcport_sm_stopped); 2411 bfa_fcport_reset_linkinfo(fcport); 2412 wwn2str(pwwn_buf, fcport->pwwn); 2413 if (BFA_PORT_IS_DISABLED(fcport->bfa)) { 2414 BFA_LOG(KERN_INFO, bfad, bfa_log_level, 2415 "Base port offline: WWN = %s\n", pwwn_buf); 2416 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_OFFLINE); 2417 } else { 2418 BFA_LOG(KERN_ERR, bfad, bfa_log_level, 2419 "Base port (WWN = %s) " 2420 "lost fabric connectivity\n", pwwn_buf); 2421 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_DISCONNECT); 2422 } 2423 break; 2424 2425 case BFA_FCPORT_SM_HWFAIL: 2426 bfa_sm_set_state(fcport, bfa_fcport_sm_iocdown); 2427 bfa_fcport_reset_linkinfo(fcport); 2428 bfa_fcport_scn(fcport, BFA_PORT_LINKDOWN, BFA_FALSE); 2429 wwn2str(pwwn_buf, fcport->pwwn); 2430 if (BFA_PORT_IS_DISABLED(fcport->bfa)) { 2431 BFA_LOG(KERN_INFO, bfad, bfa_log_level, 2432 "Base port offline: WWN = %s\n", pwwn_buf); 2433 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_OFFLINE); 2434 } else { 2435 BFA_LOG(KERN_ERR, bfad, bfa_log_level, 2436 "Base port (WWN = %s) " 2437 "lost fabric connectivity\n", pwwn_buf); 2438 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_DISCONNECT); 2439 } 2440 break; 2441 2442 case BFA_FCPORT_SM_FAA_MISCONFIG: 2443 bfa_fcport_reset_linkinfo(fcport); 2444 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_DISCONNECT); 2445 bfa_sm_set_state(fcport, bfa_fcport_sm_faa_misconfig); 2446 break; 2447 2448 default: 2449 bfa_sm_fault(fcport->bfa, event); 2450 } 2451 } 2452 2453 static void 2454 bfa_fcport_sm_disabling_qwait(struct bfa_fcport_s *fcport, 2455 enum bfa_fcport_sm_event event) 2456 { 2457 bfa_trc(fcport->bfa, event); 2458 2459 switch (event) { 2460 case BFA_FCPORT_SM_QRESUME: 2461 bfa_sm_set_state(fcport, bfa_fcport_sm_disabling); 2462 bfa_fcport_send_disable(fcport); 2463 break; 2464 2465 case BFA_FCPORT_SM_STOP: 2466 bfa_sm_set_state(fcport, bfa_fcport_sm_stopped); 2467 bfa_reqq_wcancel(&fcport->reqq_wait); 2468 break; 2469 2470 case BFA_FCPORT_SM_ENABLE: 2471 bfa_sm_set_state(fcport, bfa_fcport_sm_toggling_qwait); 2472 break; 2473 2474 case BFA_FCPORT_SM_DISABLE: 2475 /* 2476 * Already being disabled. 2477 */ 2478 break; 2479 2480 case BFA_FCPORT_SM_LINKUP: 2481 case BFA_FCPORT_SM_LINKDOWN: 2482 /* 2483 * Possible to get link events when doing back-to-back 2484 * enable/disables. 2485 */ 2486 break; 2487 2488 case BFA_FCPORT_SM_HWFAIL: 2489 bfa_sm_set_state(fcport, bfa_fcport_sm_iocfail); 2490 bfa_reqq_wcancel(&fcport->reqq_wait); 2491 break; 2492 2493 case BFA_FCPORT_SM_FAA_MISCONFIG: 2494 bfa_fcport_reset_linkinfo(fcport); 2495 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_DISCONNECT); 2496 bfa_sm_set_state(fcport, bfa_fcport_sm_faa_misconfig); 2497 break; 2498 2499 default: 2500 bfa_sm_fault(fcport->bfa, event); 2501 } 2502 } 2503 2504 static void 2505 bfa_fcport_sm_toggling_qwait(struct bfa_fcport_s *fcport, 2506 enum bfa_fcport_sm_event event) 2507 { 2508 bfa_trc(fcport->bfa, event); 2509 2510 switch (event) { 2511 case BFA_FCPORT_SM_QRESUME: 2512 bfa_sm_set_state(fcport, bfa_fcport_sm_disabling); 2513 bfa_fcport_send_disable(fcport); 2514 if (bfa_fcport_send_enable(fcport)) 2515 bfa_sm_set_state(fcport, bfa_fcport_sm_enabling); 2516 else 2517 bfa_sm_set_state(fcport, 2518 bfa_fcport_sm_enabling_qwait); 2519 break; 2520 2521 case BFA_FCPORT_SM_STOP: 2522 bfa_sm_set_state(fcport, bfa_fcport_sm_stopped); 2523 bfa_reqq_wcancel(&fcport->reqq_wait); 2524 break; 2525 2526 case BFA_FCPORT_SM_ENABLE: 2527 break; 2528 2529 case BFA_FCPORT_SM_DISABLE: 2530 bfa_sm_set_state(fcport, bfa_fcport_sm_disabling_qwait); 2531 break; 2532 2533 case BFA_FCPORT_SM_LINKUP: 2534 case BFA_FCPORT_SM_LINKDOWN: 2535 /* 2536 * Possible to get link events when doing back-to-back 2537 * enable/disables. 2538 */ 2539 break; 2540 2541 case BFA_FCPORT_SM_HWFAIL: 2542 bfa_sm_set_state(fcport, bfa_fcport_sm_iocfail); 2543 bfa_reqq_wcancel(&fcport->reqq_wait); 2544 break; 2545 2546 default: 2547 bfa_sm_fault(fcport->bfa, event); 2548 } 2549 } 2550 2551 static void 2552 bfa_fcport_sm_disabling(struct bfa_fcport_s *fcport, 2553 enum bfa_fcport_sm_event event) 2554 { 2555 char pwwn_buf[BFA_STRING_32]; 2556 struct bfad_s *bfad = (struct bfad_s *)fcport->bfa->bfad; 2557 bfa_trc(fcport->bfa, event); 2558 2559 switch (event) { 2560 case BFA_FCPORT_SM_FWRSP: 2561 bfa_sm_set_state(fcport, bfa_fcport_sm_disabled); 2562 break; 2563 2564 case BFA_FCPORT_SM_DISABLE: 2565 /* 2566 * Already being disabled. 2567 */ 2568 break; 2569 2570 case BFA_FCPORT_SM_ENABLE: 2571 if (bfa_fcport_send_enable(fcport)) 2572 bfa_sm_set_state(fcport, bfa_fcport_sm_enabling); 2573 else 2574 bfa_sm_set_state(fcport, 2575 bfa_fcport_sm_enabling_qwait); 2576 2577 bfa_plog_str(fcport->bfa->plog, BFA_PL_MID_HAL, 2578 BFA_PL_EID_PORT_ENABLE, 0, "Port Enable"); 2579 wwn2str(pwwn_buf, fcport->pwwn); 2580 BFA_LOG(KERN_INFO, bfad, bfa_log_level, 2581 "Base port enabled: WWN = %s\n", pwwn_buf); 2582 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_ENABLE); 2583 break; 2584 2585 case BFA_FCPORT_SM_STOP: 2586 bfa_sm_set_state(fcport, bfa_fcport_sm_stopped); 2587 break; 2588 2589 case BFA_FCPORT_SM_LINKUP: 2590 case BFA_FCPORT_SM_LINKDOWN: 2591 /* 2592 * Possible to get link events when doing back-to-back 2593 * enable/disables. 2594 */ 2595 break; 2596 2597 case BFA_FCPORT_SM_HWFAIL: 2598 bfa_sm_set_state(fcport, bfa_fcport_sm_iocfail); 2599 break; 2600 2601 default: 2602 bfa_sm_fault(fcport->bfa, event); 2603 } 2604 } 2605 2606 static void 2607 bfa_fcport_sm_disabled(struct bfa_fcport_s *fcport, 2608 enum bfa_fcport_sm_event event) 2609 { 2610 char pwwn_buf[BFA_STRING_32]; 2611 struct bfad_s *bfad = (struct bfad_s *)fcport->bfa->bfad; 2612 bfa_trc(fcport->bfa, event); 2613 2614 switch (event) { 2615 case BFA_FCPORT_SM_START: 2616 /* 2617 * Ignore start event for a port that is disabled. 2618 */ 2619 break; 2620 2621 case BFA_FCPORT_SM_STOP: 2622 bfa_sm_set_state(fcport, bfa_fcport_sm_stopped); 2623 break; 2624 2625 case BFA_FCPORT_SM_ENABLE: 2626 if (bfa_fcport_send_enable(fcport)) 2627 bfa_sm_set_state(fcport, bfa_fcport_sm_enabling); 2628 else 2629 bfa_sm_set_state(fcport, 2630 bfa_fcport_sm_enabling_qwait); 2631 2632 bfa_plog_str(fcport->bfa->plog, BFA_PL_MID_HAL, 2633 BFA_PL_EID_PORT_ENABLE, 0, "Port Enable"); 2634 wwn2str(pwwn_buf, fcport->pwwn); 2635 BFA_LOG(KERN_INFO, bfad, bfa_log_level, 2636 "Base port enabled: WWN = %s\n", pwwn_buf); 2637 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_ENABLE); 2638 break; 2639 2640 case BFA_FCPORT_SM_DISABLE: 2641 /* 2642 * Already disabled. 2643 */ 2644 break; 2645 2646 case BFA_FCPORT_SM_HWFAIL: 2647 bfa_sm_set_state(fcport, bfa_fcport_sm_iocfail); 2648 break; 2649 2650 case BFA_FCPORT_SM_DPORTENABLE: 2651 bfa_sm_set_state(fcport, bfa_fcport_sm_dport); 2652 break; 2653 2654 case BFA_FCPORT_SM_DDPORTENABLE: 2655 bfa_sm_set_state(fcport, bfa_fcport_sm_ddport); 2656 break; 2657 2658 default: 2659 bfa_sm_fault(fcport->bfa, event); 2660 } 2661 } 2662 2663 static void 2664 bfa_fcport_sm_stopped(struct bfa_fcport_s *fcport, 2665 enum bfa_fcport_sm_event event) 2666 { 2667 bfa_trc(fcport->bfa, event); 2668 2669 switch (event) { 2670 case BFA_FCPORT_SM_START: 2671 if (bfa_fcport_send_enable(fcport)) 2672 bfa_sm_set_state(fcport, bfa_fcport_sm_enabling); 2673 else 2674 bfa_sm_set_state(fcport, 2675 bfa_fcport_sm_enabling_qwait); 2676 break; 2677 2678 default: 2679 /* 2680 * Ignore all other events. 2681 */ 2682 ; 2683 } 2684 } 2685 2686 /* 2687 * Port is enabled. IOC is down/failed. 2688 */ 2689 static void 2690 bfa_fcport_sm_iocdown(struct bfa_fcport_s *fcport, 2691 enum bfa_fcport_sm_event event) 2692 { 2693 bfa_trc(fcport->bfa, event); 2694 2695 switch (event) { 2696 case BFA_FCPORT_SM_START: 2697 if (bfa_fcport_send_enable(fcport)) 2698 bfa_sm_set_state(fcport, bfa_fcport_sm_enabling); 2699 else 2700 bfa_sm_set_state(fcport, 2701 bfa_fcport_sm_enabling_qwait); 2702 break; 2703 2704 default: 2705 /* 2706 * Ignore all events. 2707 */ 2708 ; 2709 } 2710 } 2711 2712 /* 2713 * Port is disabled. IOC is down/failed. 2714 */ 2715 static void 2716 bfa_fcport_sm_iocfail(struct bfa_fcport_s *fcport, 2717 enum bfa_fcport_sm_event event) 2718 { 2719 bfa_trc(fcport->bfa, event); 2720 2721 switch (event) { 2722 case BFA_FCPORT_SM_START: 2723 bfa_sm_set_state(fcport, bfa_fcport_sm_disabled); 2724 break; 2725 2726 case BFA_FCPORT_SM_ENABLE: 2727 bfa_sm_set_state(fcport, bfa_fcport_sm_iocdown); 2728 break; 2729 2730 default: 2731 /* 2732 * Ignore all events. 2733 */ 2734 ; 2735 } 2736 } 2737 2738 static void 2739 bfa_fcport_sm_dport(struct bfa_fcport_s *fcport, enum bfa_fcport_sm_event event) 2740 { 2741 bfa_trc(fcport->bfa, event); 2742 2743 switch (event) { 2744 case BFA_FCPORT_SM_DPORTENABLE: 2745 case BFA_FCPORT_SM_DISABLE: 2746 case BFA_FCPORT_SM_ENABLE: 2747 case BFA_FCPORT_SM_START: 2748 /* 2749 * Ignore event for a port that is dport 2750 */ 2751 break; 2752 2753 case BFA_FCPORT_SM_STOP: 2754 bfa_sm_set_state(fcport, bfa_fcport_sm_stopped); 2755 break; 2756 2757 case BFA_FCPORT_SM_HWFAIL: 2758 bfa_sm_set_state(fcport, bfa_fcport_sm_iocfail); 2759 break; 2760 2761 case BFA_FCPORT_SM_DPORTDISABLE: 2762 bfa_sm_set_state(fcport, bfa_fcport_sm_disabled); 2763 break; 2764 2765 default: 2766 bfa_sm_fault(fcport->bfa, event); 2767 } 2768 } 2769 2770 static void 2771 bfa_fcport_sm_ddport(struct bfa_fcport_s *fcport, 2772 enum bfa_fcport_sm_event event) 2773 { 2774 bfa_trc(fcport->bfa, event); 2775 2776 switch (event) { 2777 case BFA_FCPORT_SM_DISABLE: 2778 case BFA_FCPORT_SM_DDPORTDISABLE: 2779 bfa_sm_set_state(fcport, bfa_fcport_sm_disabled); 2780 break; 2781 2782 case BFA_FCPORT_SM_DPORTENABLE: 2783 case BFA_FCPORT_SM_DPORTDISABLE: 2784 case BFA_FCPORT_SM_ENABLE: 2785 case BFA_FCPORT_SM_START: 2786 /** 2787 * Ignore event for a port that is ddport 2788 */ 2789 break; 2790 2791 case BFA_FCPORT_SM_STOP: 2792 bfa_sm_set_state(fcport, bfa_fcport_sm_stopped); 2793 break; 2794 2795 case BFA_FCPORT_SM_HWFAIL: 2796 bfa_sm_set_state(fcport, bfa_fcport_sm_iocfail); 2797 break; 2798 2799 default: 2800 bfa_sm_fault(fcport->bfa, event); 2801 } 2802 } 2803 2804 static void 2805 bfa_fcport_sm_faa_misconfig(struct bfa_fcport_s *fcport, 2806 enum bfa_fcport_sm_event event) 2807 { 2808 bfa_trc(fcport->bfa, event); 2809 2810 switch (event) { 2811 case BFA_FCPORT_SM_DPORTENABLE: 2812 case BFA_FCPORT_SM_ENABLE: 2813 case BFA_FCPORT_SM_START: 2814 /* 2815 * Ignore event for a port as there is FAA misconfig 2816 */ 2817 break; 2818 2819 case BFA_FCPORT_SM_DISABLE: 2820 if (bfa_fcport_send_disable(fcport)) 2821 bfa_sm_set_state(fcport, bfa_fcport_sm_disabling); 2822 else 2823 bfa_sm_set_state(fcport, bfa_fcport_sm_disabling_qwait); 2824 2825 bfa_fcport_reset_linkinfo(fcport); 2826 bfa_fcport_scn(fcport, BFA_PORT_LINKDOWN, BFA_FALSE); 2827 bfa_plog_str(fcport->bfa->plog, BFA_PL_MID_HAL, 2828 BFA_PL_EID_PORT_DISABLE, 0, "Port Disable"); 2829 bfa_fcport_aen_post(fcport, BFA_PORT_AEN_DISABLE); 2830 break; 2831 2832 case BFA_FCPORT_SM_STOP: 2833 bfa_sm_set_state(fcport, bfa_fcport_sm_stopped); 2834 break; 2835 2836 case BFA_FCPORT_SM_HWFAIL: 2837 bfa_fcport_reset_linkinfo(fcport); 2838 bfa_fcport_scn(fcport, BFA_PORT_LINKDOWN, BFA_FALSE); 2839 bfa_sm_set_state(fcport, bfa_fcport_sm_iocdown); 2840 break; 2841 2842 default: 2843 bfa_sm_fault(fcport->bfa, event); 2844 } 2845 } 2846 2847 /* 2848 * Link state is down 2849 */ 2850 static void 2851 bfa_fcport_ln_sm_dn(struct bfa_fcport_ln_s *ln, 2852 enum bfa_fcport_ln_sm_event event) 2853 { 2854 bfa_trc(ln->fcport->bfa, event); 2855 2856 switch (event) { 2857 case BFA_FCPORT_LN_SM_LINKUP: 2858 bfa_sm_set_state(ln, bfa_fcport_ln_sm_up_nf); 2859 bfa_fcport_queue_cb(ln, BFA_PORT_LINKUP); 2860 break; 2861 2862 default: 2863 bfa_sm_fault(ln->fcport->bfa, event); 2864 } 2865 } 2866 2867 /* 2868 * Link state is waiting for down notification 2869 */ 2870 static void 2871 bfa_fcport_ln_sm_dn_nf(struct bfa_fcport_ln_s *ln, 2872 enum bfa_fcport_ln_sm_event event) 2873 { 2874 bfa_trc(ln->fcport->bfa, event); 2875 2876 switch (event) { 2877 case BFA_FCPORT_LN_SM_LINKUP: 2878 bfa_sm_set_state(ln, bfa_fcport_ln_sm_dn_up_nf); 2879 break; 2880 2881 case BFA_FCPORT_LN_SM_NOTIFICATION: 2882 bfa_sm_set_state(ln, bfa_fcport_ln_sm_dn); 2883 break; 2884 2885 default: 2886 bfa_sm_fault(ln->fcport->bfa, event); 2887 } 2888 } 2889 2890 /* 2891 * Link state is waiting for down notification and there is a pending up 2892 */ 2893 static void 2894 bfa_fcport_ln_sm_dn_up_nf(struct bfa_fcport_ln_s *ln, 2895 enum bfa_fcport_ln_sm_event event) 2896 { 2897 bfa_trc(ln->fcport->bfa, event); 2898 2899 switch (event) { 2900 case BFA_FCPORT_LN_SM_LINKDOWN: 2901 bfa_sm_set_state(ln, bfa_fcport_ln_sm_dn_nf); 2902 break; 2903 2904 case BFA_FCPORT_LN_SM_NOTIFICATION: 2905 bfa_sm_set_state(ln, bfa_fcport_ln_sm_up_nf); 2906 bfa_fcport_queue_cb(ln, BFA_PORT_LINKUP); 2907 break; 2908 2909 default: 2910 bfa_sm_fault(ln->fcport->bfa, event); 2911 } 2912 } 2913 2914 /* 2915 * Link state is up 2916 */ 2917 static void 2918 bfa_fcport_ln_sm_up(struct bfa_fcport_ln_s *ln, 2919 enum bfa_fcport_ln_sm_event event) 2920 { 2921 bfa_trc(ln->fcport->bfa, event); 2922 2923 switch (event) { 2924 case BFA_FCPORT_LN_SM_LINKDOWN: 2925 bfa_sm_set_state(ln, bfa_fcport_ln_sm_dn_nf); 2926 bfa_fcport_queue_cb(ln, BFA_PORT_LINKDOWN); 2927 break; 2928 2929 default: 2930 bfa_sm_fault(ln->fcport->bfa, event); 2931 } 2932 } 2933 2934 /* 2935 * Link state is waiting for up notification 2936 */ 2937 static void 2938 bfa_fcport_ln_sm_up_nf(struct bfa_fcport_ln_s *ln, 2939 enum bfa_fcport_ln_sm_event event) 2940 { 2941 bfa_trc(ln->fcport->bfa, event); 2942 2943 switch (event) { 2944 case BFA_FCPORT_LN_SM_LINKDOWN: 2945 bfa_sm_set_state(ln, bfa_fcport_ln_sm_up_dn_nf); 2946 break; 2947 2948 case BFA_FCPORT_LN_SM_NOTIFICATION: 2949 bfa_sm_set_state(ln, bfa_fcport_ln_sm_up); 2950 break; 2951 2952 default: 2953 bfa_sm_fault(ln->fcport->bfa, event); 2954 } 2955 } 2956 2957 /* 2958 * Link state is waiting for up notification and there is a pending down 2959 */ 2960 static void 2961 bfa_fcport_ln_sm_up_dn_nf(struct bfa_fcport_ln_s *ln, 2962 enum bfa_fcport_ln_sm_event event) 2963 { 2964 bfa_trc(ln->fcport->bfa, event); 2965 2966 switch (event) { 2967 case BFA_FCPORT_LN_SM_LINKUP: 2968 bfa_sm_set_state(ln, bfa_fcport_ln_sm_up_dn_up_nf); 2969 break; 2970 2971 case BFA_FCPORT_LN_SM_NOTIFICATION: 2972 bfa_sm_set_state(ln, bfa_fcport_ln_sm_dn_nf); 2973 bfa_fcport_queue_cb(ln, BFA_PORT_LINKDOWN); 2974 break; 2975 2976 default: 2977 bfa_sm_fault(ln->fcport->bfa, event); 2978 } 2979 } 2980 2981 /* 2982 * Link state is waiting for up notification and there are pending down and up 2983 */ 2984 static void 2985 bfa_fcport_ln_sm_up_dn_up_nf(struct bfa_fcport_ln_s *ln, 2986 enum bfa_fcport_ln_sm_event event) 2987 { 2988 bfa_trc(ln->fcport->bfa, event); 2989 2990 switch (event) { 2991 case BFA_FCPORT_LN_SM_LINKDOWN: 2992 bfa_sm_set_state(ln, bfa_fcport_ln_sm_up_dn_nf); 2993 break; 2994 2995 case BFA_FCPORT_LN_SM_NOTIFICATION: 2996 bfa_sm_set_state(ln, bfa_fcport_ln_sm_dn_up_nf); 2997 bfa_fcport_queue_cb(ln, BFA_PORT_LINKDOWN); 2998 break; 2999 3000 default: 3001 bfa_sm_fault(ln->fcport->bfa, event); 3002 } 3003 } 3004 3005 static void 3006 __bfa_cb_fcport_event(void *cbarg, bfa_boolean_t complete) 3007 { 3008 struct bfa_fcport_ln_s *ln = cbarg; 3009 3010 if (complete) 3011 ln->fcport->event_cbfn(ln->fcport->event_cbarg, ln->ln_event); 3012 else 3013 bfa_sm_send_event(ln, BFA_FCPORT_LN_SM_NOTIFICATION); 3014 } 3015 3016 /* 3017 * Send SCN notification to upper layers. 3018 * trunk - false if caller is fcport to ignore fcport event in trunked mode 3019 */ 3020 static void 3021 bfa_fcport_scn(struct bfa_fcport_s *fcport, enum bfa_port_linkstate event, 3022 bfa_boolean_t trunk) 3023 { 3024 if (fcport->cfg.trunked && !trunk) 3025 return; 3026 3027 switch (event) { 3028 case BFA_PORT_LINKUP: 3029 bfa_sm_send_event(&fcport->ln, BFA_FCPORT_LN_SM_LINKUP); 3030 break; 3031 case BFA_PORT_LINKDOWN: 3032 bfa_sm_send_event(&fcport->ln, BFA_FCPORT_LN_SM_LINKDOWN); 3033 break; 3034 default: 3035 WARN_ON(1); 3036 } 3037 } 3038 3039 static void 3040 bfa_fcport_queue_cb(struct bfa_fcport_ln_s *ln, enum bfa_port_linkstate event) 3041 { 3042 struct bfa_fcport_s *fcport = ln->fcport; 3043 3044 if (fcport->bfa->fcs) { 3045 fcport->event_cbfn(fcport->event_cbarg, event); 3046 bfa_sm_send_event(ln, BFA_FCPORT_LN_SM_NOTIFICATION); 3047 } else { 3048 ln->ln_event = event; 3049 bfa_cb_queue(fcport->bfa, &ln->ln_qe, 3050 __bfa_cb_fcport_event, ln); 3051 } 3052 } 3053 3054 #define FCPORT_STATS_DMA_SZ (BFA_ROUNDUP(sizeof(union bfa_fcport_stats_u), \ 3055 BFA_CACHELINE_SZ)) 3056 3057 static void 3058 bfa_fcport_meminfo(struct bfa_iocfc_cfg_s *cfg, struct bfa_meminfo_s *minfo, 3059 struct bfa_s *bfa) 3060 { 3061 struct bfa_mem_dma_s *fcport_dma = BFA_MEM_FCPORT_DMA(bfa); 3062 3063 bfa_mem_dma_setup(minfo, fcport_dma, FCPORT_STATS_DMA_SZ); 3064 } 3065 3066 static void 3067 bfa_fcport_qresume(void *cbarg) 3068 { 3069 struct bfa_fcport_s *fcport = cbarg; 3070 3071 bfa_sm_send_event(fcport, BFA_FCPORT_SM_QRESUME); 3072 } 3073 3074 static void 3075 bfa_fcport_mem_claim(struct bfa_fcport_s *fcport) 3076 { 3077 struct bfa_mem_dma_s *fcport_dma = &fcport->fcport_dma; 3078 3079 fcport->stats_kva = bfa_mem_dma_virt(fcport_dma); 3080 fcport->stats_pa = bfa_mem_dma_phys(fcport_dma); 3081 fcport->stats = (union bfa_fcport_stats_u *) 3082 bfa_mem_dma_virt(fcport_dma); 3083 } 3084 3085 /* 3086 * Memory initialization. 3087 */ 3088 static void 3089 bfa_fcport_attach(struct bfa_s *bfa, void *bfad, struct bfa_iocfc_cfg_s *cfg, 3090 struct bfa_pcidev_s *pcidev) 3091 { 3092 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3093 struct bfa_port_cfg_s *port_cfg = &fcport->cfg; 3094 struct bfa_fcport_ln_s *ln = &fcport->ln; 3095 struct timeval tv; 3096 3097 fcport->bfa = bfa; 3098 ln->fcport = fcport; 3099 3100 bfa_fcport_mem_claim(fcport); 3101 3102 bfa_sm_set_state(fcport, bfa_fcport_sm_uninit); 3103 bfa_sm_set_state(ln, bfa_fcport_ln_sm_dn); 3104 3105 /* 3106 * initialize time stamp for stats reset 3107 */ 3108 do_gettimeofday(&tv); 3109 fcport->stats_reset_time = tv.tv_sec; 3110 fcport->stats_dma_ready = BFA_FALSE; 3111 3112 /* 3113 * initialize and set default configuration 3114 */ 3115 port_cfg->topology = BFA_PORT_TOPOLOGY_P2P; 3116 port_cfg->speed = BFA_PORT_SPEED_AUTO; 3117 port_cfg->trunked = BFA_FALSE; 3118 port_cfg->maxfrsize = 0; 3119 3120 port_cfg->trl_def_speed = BFA_PORT_SPEED_1GBPS; 3121 port_cfg->qos_bw.high = BFA_QOS_BW_HIGH; 3122 port_cfg->qos_bw.med = BFA_QOS_BW_MED; 3123 port_cfg->qos_bw.low = BFA_QOS_BW_LOW; 3124 3125 fcport->fec_state = BFA_FEC_OFFLINE; 3126 3127 INIT_LIST_HEAD(&fcport->stats_pending_q); 3128 INIT_LIST_HEAD(&fcport->statsclr_pending_q); 3129 3130 bfa_reqq_winit(&fcport->reqq_wait, bfa_fcport_qresume, fcport); 3131 } 3132 3133 static void 3134 bfa_fcport_detach(struct bfa_s *bfa) 3135 { 3136 } 3137 3138 /* 3139 * Called when IOC is ready. 3140 */ 3141 static void 3142 bfa_fcport_start(struct bfa_s *bfa) 3143 { 3144 bfa_sm_send_event(BFA_FCPORT_MOD(bfa), BFA_FCPORT_SM_START); 3145 } 3146 3147 /* 3148 * Called before IOC is stopped. 3149 */ 3150 static void 3151 bfa_fcport_stop(struct bfa_s *bfa) 3152 { 3153 bfa_sm_send_event(BFA_FCPORT_MOD(bfa), BFA_FCPORT_SM_STOP); 3154 bfa_trunk_iocdisable(bfa); 3155 } 3156 3157 /* 3158 * Called when IOC failure is detected. 3159 */ 3160 static void 3161 bfa_fcport_iocdisable(struct bfa_s *bfa) 3162 { 3163 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3164 3165 bfa_sm_send_event(fcport, BFA_FCPORT_SM_HWFAIL); 3166 bfa_trunk_iocdisable(bfa); 3167 } 3168 3169 /* 3170 * Update loop info in fcport for SCN online 3171 */ 3172 static void 3173 bfa_fcport_update_loop_info(struct bfa_fcport_s *fcport, 3174 struct bfa_fcport_loop_info_s *loop_info) 3175 { 3176 fcport->myalpa = loop_info->myalpa; 3177 fcport->alpabm_valid = 3178 loop_info->alpabm_val; 3179 memcpy(fcport->alpabm.alpa_bm, 3180 loop_info->alpabm.alpa_bm, 3181 sizeof(struct fc_alpabm_s)); 3182 } 3183 3184 static void 3185 bfa_fcport_update_linkinfo(struct bfa_fcport_s *fcport) 3186 { 3187 struct bfi_fcport_event_s *pevent = fcport->event_arg.i2hmsg.event; 3188 struct bfa_fcport_trunk_s *trunk = &fcport->trunk; 3189 3190 fcport->speed = pevent->link_state.speed; 3191 fcport->topology = pevent->link_state.topology; 3192 3193 if (fcport->topology == BFA_PORT_TOPOLOGY_LOOP) { 3194 bfa_fcport_update_loop_info(fcport, 3195 &pevent->link_state.attr.loop_info); 3196 return; 3197 } 3198 3199 /* QoS Details */ 3200 fcport->qos_attr = pevent->link_state.qos_attr; 3201 fcport->qos_vc_attr = pevent->link_state.attr.vc_fcf.qos_vc_attr; 3202 3203 if (fcport->cfg.bb_cr_enabled) 3204 fcport->bbcr_attr = pevent->link_state.attr.bbcr_attr; 3205 3206 fcport->fec_state = pevent->link_state.fec_state; 3207 3208 /* 3209 * update trunk state if applicable 3210 */ 3211 if (!fcport->cfg.trunked) 3212 trunk->attr.state = BFA_TRUNK_DISABLED; 3213 3214 /* update FCoE specific */ 3215 fcport->fcoe_vlan = 3216 be16_to_cpu(pevent->link_state.attr.vc_fcf.fcf.vlan); 3217 3218 bfa_trc(fcport->bfa, fcport->speed); 3219 bfa_trc(fcport->bfa, fcport->topology); 3220 } 3221 3222 static void 3223 bfa_fcport_reset_linkinfo(struct bfa_fcport_s *fcport) 3224 { 3225 fcport->speed = BFA_PORT_SPEED_UNKNOWN; 3226 fcport->topology = BFA_PORT_TOPOLOGY_NONE; 3227 fcport->fec_state = BFA_FEC_OFFLINE; 3228 } 3229 3230 /* 3231 * Send port enable message to firmware. 3232 */ 3233 static bfa_boolean_t 3234 bfa_fcport_send_enable(struct bfa_fcport_s *fcport) 3235 { 3236 struct bfi_fcport_enable_req_s *m; 3237 3238 /* 3239 * Increment message tag before queue check, so that responses to old 3240 * requests are discarded. 3241 */ 3242 fcport->msgtag++; 3243 3244 /* 3245 * check for room in queue to send request now 3246 */ 3247 m = bfa_reqq_next(fcport->bfa, BFA_REQQ_PORT); 3248 if (!m) { 3249 bfa_reqq_wait(fcport->bfa, BFA_REQQ_PORT, 3250 &fcport->reqq_wait); 3251 return BFA_FALSE; 3252 } 3253 3254 bfi_h2i_set(m->mh, BFI_MC_FCPORT, BFI_FCPORT_H2I_ENABLE_REQ, 3255 bfa_fn_lpu(fcport->bfa)); 3256 m->nwwn = fcport->nwwn; 3257 m->pwwn = fcport->pwwn; 3258 m->port_cfg = fcport->cfg; 3259 m->msgtag = fcport->msgtag; 3260 m->port_cfg.maxfrsize = cpu_to_be16(fcport->cfg.maxfrsize); 3261 m->use_flash_cfg = fcport->use_flash_cfg; 3262 bfa_dma_be_addr_set(m->stats_dma_addr, fcport->stats_pa); 3263 bfa_trc(fcport->bfa, m->stats_dma_addr.a32.addr_lo); 3264 bfa_trc(fcport->bfa, m->stats_dma_addr.a32.addr_hi); 3265 3266 /* 3267 * queue I/O message to firmware 3268 */ 3269 bfa_reqq_produce(fcport->bfa, BFA_REQQ_PORT, m->mh); 3270 return BFA_TRUE; 3271 } 3272 3273 /* 3274 * Send port disable message to firmware. 3275 */ 3276 static bfa_boolean_t 3277 bfa_fcport_send_disable(struct bfa_fcport_s *fcport) 3278 { 3279 struct bfi_fcport_req_s *m; 3280 3281 /* 3282 * Increment message tag before queue check, so that responses to old 3283 * requests are discarded. 3284 */ 3285 fcport->msgtag++; 3286 3287 /* 3288 * check for room in queue to send request now 3289 */ 3290 m = bfa_reqq_next(fcport->bfa, BFA_REQQ_PORT); 3291 if (!m) { 3292 bfa_reqq_wait(fcport->bfa, BFA_REQQ_PORT, 3293 &fcport->reqq_wait); 3294 return BFA_FALSE; 3295 } 3296 3297 bfi_h2i_set(m->mh, BFI_MC_FCPORT, BFI_FCPORT_H2I_DISABLE_REQ, 3298 bfa_fn_lpu(fcport->bfa)); 3299 m->msgtag = fcport->msgtag; 3300 3301 /* 3302 * queue I/O message to firmware 3303 */ 3304 bfa_reqq_produce(fcport->bfa, BFA_REQQ_PORT, m->mh); 3305 3306 return BFA_TRUE; 3307 } 3308 3309 static void 3310 bfa_fcport_set_wwns(struct bfa_fcport_s *fcport) 3311 { 3312 fcport->pwwn = fcport->bfa->ioc.attr->pwwn; 3313 fcport->nwwn = fcport->bfa->ioc.attr->nwwn; 3314 3315 bfa_trc(fcport->bfa, fcport->pwwn); 3316 bfa_trc(fcport->bfa, fcport->nwwn); 3317 } 3318 3319 static void 3320 bfa_fcport_qos_stats_swap(struct bfa_qos_stats_s *d, 3321 struct bfa_qos_stats_s *s) 3322 { 3323 u32 *dip = (u32 *) d; 3324 __be32 *sip = (__be32 *) s; 3325 int i; 3326 3327 /* Now swap the 32 bit fields */ 3328 for (i = 0; i < (sizeof(struct bfa_qos_stats_s)/sizeof(u32)); ++i) 3329 dip[i] = be32_to_cpu(sip[i]); 3330 } 3331 3332 static void 3333 bfa_fcport_fcoe_stats_swap(struct bfa_fcoe_stats_s *d, 3334 struct bfa_fcoe_stats_s *s) 3335 { 3336 u32 *dip = (u32 *) d; 3337 __be32 *sip = (__be32 *) s; 3338 int i; 3339 3340 for (i = 0; i < ((sizeof(struct bfa_fcoe_stats_s))/sizeof(u32)); 3341 i = i + 2) { 3342 #ifdef __BIG_ENDIAN 3343 dip[i] = be32_to_cpu(sip[i]); 3344 dip[i + 1] = be32_to_cpu(sip[i + 1]); 3345 #else 3346 dip[i] = be32_to_cpu(sip[i + 1]); 3347 dip[i + 1] = be32_to_cpu(sip[i]); 3348 #endif 3349 } 3350 } 3351 3352 static void 3353 __bfa_cb_fcport_stats_get(void *cbarg, bfa_boolean_t complete) 3354 { 3355 struct bfa_fcport_s *fcport = (struct bfa_fcport_s *)cbarg; 3356 struct bfa_cb_pending_q_s *cb; 3357 struct list_head *qe, *qen; 3358 union bfa_fcport_stats_u *ret; 3359 3360 if (complete) { 3361 struct timeval tv; 3362 if (fcport->stats_status == BFA_STATUS_OK) 3363 do_gettimeofday(&tv); 3364 3365 list_for_each_safe(qe, qen, &fcport->stats_pending_q) { 3366 bfa_q_deq(&fcport->stats_pending_q, &qe); 3367 cb = (struct bfa_cb_pending_q_s *)qe; 3368 if (fcport->stats_status == BFA_STATUS_OK) { 3369 ret = (union bfa_fcport_stats_u *)cb->data; 3370 /* Swap FC QoS or FCoE stats */ 3371 if (bfa_ioc_get_fcmode(&fcport->bfa->ioc)) 3372 bfa_fcport_qos_stats_swap(&ret->fcqos, 3373 &fcport->stats->fcqos); 3374 else { 3375 bfa_fcport_fcoe_stats_swap(&ret->fcoe, 3376 &fcport->stats->fcoe); 3377 ret->fcoe.secs_reset = 3378 tv.tv_sec - fcport->stats_reset_time; 3379 } 3380 } 3381 bfa_cb_queue_status(fcport->bfa, &cb->hcb_qe, 3382 fcport->stats_status); 3383 } 3384 fcport->stats_status = BFA_STATUS_OK; 3385 } else { 3386 INIT_LIST_HEAD(&fcport->stats_pending_q); 3387 fcport->stats_status = BFA_STATUS_OK; 3388 } 3389 } 3390 3391 static void 3392 bfa_fcport_stats_get_timeout(void *cbarg) 3393 { 3394 struct bfa_fcport_s *fcport = (struct bfa_fcport_s *) cbarg; 3395 3396 bfa_trc(fcport->bfa, fcport->stats_qfull); 3397 3398 if (fcport->stats_qfull) { 3399 bfa_reqq_wcancel(&fcport->stats_reqq_wait); 3400 fcport->stats_qfull = BFA_FALSE; 3401 } 3402 3403 fcport->stats_status = BFA_STATUS_ETIMER; 3404 __bfa_cb_fcport_stats_get(fcport, BFA_TRUE); 3405 } 3406 3407 static void 3408 bfa_fcport_send_stats_get(void *cbarg) 3409 { 3410 struct bfa_fcport_s *fcport = (struct bfa_fcport_s *) cbarg; 3411 struct bfi_fcport_req_s *msg; 3412 3413 msg = bfa_reqq_next(fcport->bfa, BFA_REQQ_PORT); 3414 3415 if (!msg) { 3416 fcport->stats_qfull = BFA_TRUE; 3417 bfa_reqq_winit(&fcport->stats_reqq_wait, 3418 bfa_fcport_send_stats_get, fcport); 3419 bfa_reqq_wait(fcport->bfa, BFA_REQQ_PORT, 3420 &fcport->stats_reqq_wait); 3421 return; 3422 } 3423 fcport->stats_qfull = BFA_FALSE; 3424 3425 memset(msg, 0, sizeof(struct bfi_fcport_req_s)); 3426 bfi_h2i_set(msg->mh, BFI_MC_FCPORT, BFI_FCPORT_H2I_STATS_GET_REQ, 3427 bfa_fn_lpu(fcport->bfa)); 3428 bfa_reqq_produce(fcport->bfa, BFA_REQQ_PORT, msg->mh); 3429 } 3430 3431 static void 3432 __bfa_cb_fcport_stats_clr(void *cbarg, bfa_boolean_t complete) 3433 { 3434 struct bfa_fcport_s *fcport = (struct bfa_fcport_s *) cbarg; 3435 struct bfa_cb_pending_q_s *cb; 3436 struct list_head *qe, *qen; 3437 3438 if (complete) { 3439 struct timeval tv; 3440 3441 /* 3442 * re-initialize time stamp for stats reset 3443 */ 3444 do_gettimeofday(&tv); 3445 fcport->stats_reset_time = tv.tv_sec; 3446 list_for_each_safe(qe, qen, &fcport->statsclr_pending_q) { 3447 bfa_q_deq(&fcport->statsclr_pending_q, &qe); 3448 cb = (struct bfa_cb_pending_q_s *)qe; 3449 bfa_cb_queue_status(fcport->bfa, &cb->hcb_qe, 3450 fcport->stats_status); 3451 } 3452 fcport->stats_status = BFA_STATUS_OK; 3453 } else { 3454 INIT_LIST_HEAD(&fcport->statsclr_pending_q); 3455 fcport->stats_status = BFA_STATUS_OK; 3456 } 3457 } 3458 3459 static void 3460 bfa_fcport_stats_clr_timeout(void *cbarg) 3461 { 3462 struct bfa_fcport_s *fcport = (struct bfa_fcport_s *) cbarg; 3463 3464 bfa_trc(fcport->bfa, fcport->stats_qfull); 3465 3466 if (fcport->stats_qfull) { 3467 bfa_reqq_wcancel(&fcport->stats_reqq_wait); 3468 fcport->stats_qfull = BFA_FALSE; 3469 } 3470 3471 fcport->stats_status = BFA_STATUS_ETIMER; 3472 __bfa_cb_fcport_stats_clr(fcport, BFA_TRUE); 3473 } 3474 3475 static void 3476 bfa_fcport_send_stats_clear(void *cbarg) 3477 { 3478 struct bfa_fcport_s *fcport = (struct bfa_fcport_s *) cbarg; 3479 struct bfi_fcport_req_s *msg; 3480 3481 msg = bfa_reqq_next(fcport->bfa, BFA_REQQ_PORT); 3482 3483 if (!msg) { 3484 fcport->stats_qfull = BFA_TRUE; 3485 bfa_reqq_winit(&fcport->stats_reqq_wait, 3486 bfa_fcport_send_stats_clear, fcport); 3487 bfa_reqq_wait(fcport->bfa, BFA_REQQ_PORT, 3488 &fcport->stats_reqq_wait); 3489 return; 3490 } 3491 fcport->stats_qfull = BFA_FALSE; 3492 3493 memset(msg, 0, sizeof(struct bfi_fcport_req_s)); 3494 bfi_h2i_set(msg->mh, BFI_MC_FCPORT, BFI_FCPORT_H2I_STATS_CLEAR_REQ, 3495 bfa_fn_lpu(fcport->bfa)); 3496 bfa_reqq_produce(fcport->bfa, BFA_REQQ_PORT, msg->mh); 3497 } 3498 3499 /* 3500 * Handle trunk SCN event from firmware. 3501 */ 3502 static void 3503 bfa_trunk_scn(struct bfa_fcport_s *fcport, struct bfi_fcport_trunk_scn_s *scn) 3504 { 3505 struct bfa_fcport_trunk_s *trunk = &fcport->trunk; 3506 struct bfi_fcport_trunk_link_s *tlink; 3507 struct bfa_trunk_link_attr_s *lattr; 3508 enum bfa_trunk_state state_prev; 3509 int i; 3510 int link_bm = 0; 3511 3512 bfa_trc(fcport->bfa, fcport->cfg.trunked); 3513 WARN_ON(scn->trunk_state != BFA_TRUNK_ONLINE && 3514 scn->trunk_state != BFA_TRUNK_OFFLINE); 3515 3516 bfa_trc(fcport->bfa, trunk->attr.state); 3517 bfa_trc(fcport->bfa, scn->trunk_state); 3518 bfa_trc(fcport->bfa, scn->trunk_speed); 3519 3520 /* 3521 * Save off new state for trunk attribute query 3522 */ 3523 state_prev = trunk->attr.state; 3524 if (fcport->cfg.trunked && (trunk->attr.state != BFA_TRUNK_DISABLED)) 3525 trunk->attr.state = scn->trunk_state; 3526 trunk->attr.speed = scn->trunk_speed; 3527 for (i = 0; i < BFA_TRUNK_MAX_PORTS; i++) { 3528 lattr = &trunk->attr.link_attr[i]; 3529 tlink = &scn->tlink[i]; 3530 3531 lattr->link_state = tlink->state; 3532 lattr->trunk_wwn = tlink->trunk_wwn; 3533 lattr->fctl = tlink->fctl; 3534 lattr->speed = tlink->speed; 3535 lattr->deskew = be32_to_cpu(tlink->deskew); 3536 3537 if (tlink->state == BFA_TRUNK_LINK_STATE_UP) { 3538 fcport->speed = tlink->speed; 3539 fcport->topology = BFA_PORT_TOPOLOGY_P2P; 3540 link_bm |= 1 << i; 3541 } 3542 3543 bfa_trc(fcport->bfa, lattr->link_state); 3544 bfa_trc(fcport->bfa, lattr->trunk_wwn); 3545 bfa_trc(fcport->bfa, lattr->fctl); 3546 bfa_trc(fcport->bfa, lattr->speed); 3547 bfa_trc(fcport->bfa, lattr->deskew); 3548 } 3549 3550 switch (link_bm) { 3551 case 3: 3552 bfa_plog_str(fcport->bfa->plog, BFA_PL_MID_HAL, 3553 BFA_PL_EID_TRUNK_SCN, 0, "Trunk up(0,1)"); 3554 break; 3555 case 2: 3556 bfa_plog_str(fcport->bfa->plog, BFA_PL_MID_HAL, 3557 BFA_PL_EID_TRUNK_SCN, 0, "Trunk up(-,1)"); 3558 break; 3559 case 1: 3560 bfa_plog_str(fcport->bfa->plog, BFA_PL_MID_HAL, 3561 BFA_PL_EID_TRUNK_SCN, 0, "Trunk up(0,-)"); 3562 break; 3563 default: 3564 bfa_plog_str(fcport->bfa->plog, BFA_PL_MID_HAL, 3565 BFA_PL_EID_TRUNK_SCN, 0, "Trunk down"); 3566 } 3567 3568 /* 3569 * Notify upper layers if trunk state changed. 3570 */ 3571 if ((state_prev != trunk->attr.state) || 3572 (scn->trunk_state == BFA_TRUNK_OFFLINE)) { 3573 bfa_fcport_scn(fcport, (scn->trunk_state == BFA_TRUNK_ONLINE) ? 3574 BFA_PORT_LINKUP : BFA_PORT_LINKDOWN, BFA_TRUE); 3575 } 3576 } 3577 3578 static void 3579 bfa_trunk_iocdisable(struct bfa_s *bfa) 3580 { 3581 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3582 int i = 0; 3583 3584 /* 3585 * In trunked mode, notify upper layers that link is down 3586 */ 3587 if (fcport->cfg.trunked) { 3588 if (fcport->trunk.attr.state == BFA_TRUNK_ONLINE) 3589 bfa_fcport_scn(fcport, BFA_PORT_LINKDOWN, BFA_TRUE); 3590 3591 fcport->trunk.attr.state = BFA_TRUNK_OFFLINE; 3592 fcport->trunk.attr.speed = BFA_PORT_SPEED_UNKNOWN; 3593 for (i = 0; i < BFA_TRUNK_MAX_PORTS; i++) { 3594 fcport->trunk.attr.link_attr[i].trunk_wwn = 0; 3595 fcport->trunk.attr.link_attr[i].fctl = 3596 BFA_TRUNK_LINK_FCTL_NORMAL; 3597 fcport->trunk.attr.link_attr[i].link_state = 3598 BFA_TRUNK_LINK_STATE_DN_LINKDN; 3599 fcport->trunk.attr.link_attr[i].speed = 3600 BFA_PORT_SPEED_UNKNOWN; 3601 fcport->trunk.attr.link_attr[i].deskew = 0; 3602 } 3603 } 3604 } 3605 3606 /* 3607 * Called to initialize port attributes 3608 */ 3609 void 3610 bfa_fcport_init(struct bfa_s *bfa) 3611 { 3612 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3613 3614 /* 3615 * Initialize port attributes from IOC hardware data. 3616 */ 3617 bfa_fcport_set_wwns(fcport); 3618 if (fcport->cfg.maxfrsize == 0) 3619 fcport->cfg.maxfrsize = bfa_ioc_maxfrsize(&bfa->ioc); 3620 fcport->cfg.rx_bbcredit = bfa_ioc_rx_bbcredit(&bfa->ioc); 3621 fcport->speed_sup = bfa_ioc_speed_sup(&bfa->ioc); 3622 3623 if (bfa_fcport_is_pbcdisabled(bfa)) 3624 bfa->modules.port.pbc_disabled = BFA_TRUE; 3625 3626 WARN_ON(!fcport->cfg.maxfrsize); 3627 WARN_ON(!fcport->cfg.rx_bbcredit); 3628 WARN_ON(!fcport->speed_sup); 3629 } 3630 3631 /* 3632 * Firmware message handler. 3633 */ 3634 void 3635 bfa_fcport_isr(struct bfa_s *bfa, struct bfi_msg_s *msg) 3636 { 3637 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3638 union bfi_fcport_i2h_msg_u i2hmsg; 3639 3640 i2hmsg.msg = msg; 3641 fcport->event_arg.i2hmsg = i2hmsg; 3642 3643 bfa_trc(bfa, msg->mhdr.msg_id); 3644 bfa_trc(bfa, bfa_sm_to_state(hal_port_sm_table, fcport->sm)); 3645 3646 switch (msg->mhdr.msg_id) { 3647 case BFI_FCPORT_I2H_ENABLE_RSP: 3648 if (fcport->msgtag == i2hmsg.penable_rsp->msgtag) { 3649 3650 fcport->stats_dma_ready = BFA_TRUE; 3651 if (fcport->use_flash_cfg) { 3652 fcport->cfg = i2hmsg.penable_rsp->port_cfg; 3653 fcport->cfg.maxfrsize = 3654 cpu_to_be16(fcport->cfg.maxfrsize); 3655 fcport->cfg.path_tov = 3656 cpu_to_be16(fcport->cfg.path_tov); 3657 fcport->cfg.q_depth = 3658 cpu_to_be16(fcport->cfg.q_depth); 3659 3660 if (fcport->cfg.trunked) 3661 fcport->trunk.attr.state = 3662 BFA_TRUNK_OFFLINE; 3663 else 3664 fcport->trunk.attr.state = 3665 BFA_TRUNK_DISABLED; 3666 fcport->qos_attr.qos_bw = 3667 i2hmsg.penable_rsp->port_cfg.qos_bw; 3668 fcport->use_flash_cfg = BFA_FALSE; 3669 } 3670 3671 if (fcport->cfg.qos_enabled) 3672 fcport->qos_attr.state = BFA_QOS_OFFLINE; 3673 else 3674 fcport->qos_attr.state = BFA_QOS_DISABLED; 3675 3676 fcport->qos_attr.qos_bw_op = 3677 i2hmsg.penable_rsp->port_cfg.qos_bw; 3678 3679 if (fcport->cfg.bb_cr_enabled) 3680 fcport->bbcr_attr.state = BFA_BBCR_OFFLINE; 3681 else 3682 fcport->bbcr_attr.state = BFA_BBCR_DISABLED; 3683 3684 bfa_sm_send_event(fcport, BFA_FCPORT_SM_FWRSP); 3685 } 3686 break; 3687 3688 case BFI_FCPORT_I2H_DISABLE_RSP: 3689 if (fcport->msgtag == i2hmsg.penable_rsp->msgtag) 3690 bfa_sm_send_event(fcport, BFA_FCPORT_SM_FWRSP); 3691 break; 3692 3693 case BFI_FCPORT_I2H_EVENT: 3694 if (fcport->cfg.bb_cr_enabled) 3695 fcport->bbcr_attr.state = BFA_BBCR_OFFLINE; 3696 else 3697 fcport->bbcr_attr.state = BFA_BBCR_DISABLED; 3698 3699 if (i2hmsg.event->link_state.linkstate == BFA_PORT_LINKUP) 3700 bfa_sm_send_event(fcport, BFA_FCPORT_SM_LINKUP); 3701 else { 3702 if (i2hmsg.event->link_state.linkstate_rsn == 3703 BFA_PORT_LINKSTATE_RSN_FAA_MISCONFIG) 3704 bfa_sm_send_event(fcport, 3705 BFA_FCPORT_SM_FAA_MISCONFIG); 3706 else 3707 bfa_sm_send_event(fcport, 3708 BFA_FCPORT_SM_LINKDOWN); 3709 } 3710 fcport->qos_attr.qos_bw_op = 3711 i2hmsg.event->link_state.qos_attr.qos_bw_op; 3712 break; 3713 3714 case BFI_FCPORT_I2H_TRUNK_SCN: 3715 bfa_trunk_scn(fcport, i2hmsg.trunk_scn); 3716 break; 3717 3718 case BFI_FCPORT_I2H_STATS_GET_RSP: 3719 /* 3720 * check for timer pop before processing the rsp 3721 */ 3722 if (list_empty(&fcport->stats_pending_q) || 3723 (fcport->stats_status == BFA_STATUS_ETIMER)) 3724 break; 3725 3726 bfa_timer_stop(&fcport->timer); 3727 fcport->stats_status = i2hmsg.pstatsget_rsp->status; 3728 __bfa_cb_fcport_stats_get(fcport, BFA_TRUE); 3729 break; 3730 3731 case BFI_FCPORT_I2H_STATS_CLEAR_RSP: 3732 /* 3733 * check for timer pop before processing the rsp 3734 */ 3735 if (list_empty(&fcport->statsclr_pending_q) || 3736 (fcport->stats_status == BFA_STATUS_ETIMER)) 3737 break; 3738 3739 bfa_timer_stop(&fcport->timer); 3740 fcport->stats_status = BFA_STATUS_OK; 3741 __bfa_cb_fcport_stats_clr(fcport, BFA_TRUE); 3742 break; 3743 3744 case BFI_FCPORT_I2H_ENABLE_AEN: 3745 bfa_sm_send_event(fcport, BFA_FCPORT_SM_ENABLE); 3746 break; 3747 3748 case BFI_FCPORT_I2H_DISABLE_AEN: 3749 bfa_sm_send_event(fcport, BFA_FCPORT_SM_DISABLE); 3750 break; 3751 3752 default: 3753 WARN_ON(1); 3754 break; 3755 } 3756 } 3757 3758 /* 3759 * Registered callback for port events. 3760 */ 3761 void 3762 bfa_fcport_event_register(struct bfa_s *bfa, 3763 void (*cbfn) (void *cbarg, 3764 enum bfa_port_linkstate event), 3765 void *cbarg) 3766 { 3767 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3768 3769 fcport->event_cbfn = cbfn; 3770 fcport->event_cbarg = cbarg; 3771 } 3772 3773 bfa_status_t 3774 bfa_fcport_enable(struct bfa_s *bfa) 3775 { 3776 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3777 3778 if (bfa_fcport_is_pbcdisabled(bfa)) 3779 return BFA_STATUS_PBC; 3780 3781 if (bfa_ioc_is_disabled(&bfa->ioc)) 3782 return BFA_STATUS_IOC_DISABLED; 3783 3784 if (fcport->diag_busy) 3785 return BFA_STATUS_DIAG_BUSY; 3786 3787 bfa_sm_send_event(BFA_FCPORT_MOD(bfa), BFA_FCPORT_SM_ENABLE); 3788 return BFA_STATUS_OK; 3789 } 3790 3791 bfa_status_t 3792 bfa_fcport_disable(struct bfa_s *bfa) 3793 { 3794 if (bfa_fcport_is_pbcdisabled(bfa)) 3795 return BFA_STATUS_PBC; 3796 3797 if (bfa_ioc_is_disabled(&bfa->ioc)) 3798 return BFA_STATUS_IOC_DISABLED; 3799 3800 bfa_sm_send_event(BFA_FCPORT_MOD(bfa), BFA_FCPORT_SM_DISABLE); 3801 return BFA_STATUS_OK; 3802 } 3803 3804 /* If PBC is disabled on port, return error */ 3805 bfa_status_t 3806 bfa_fcport_is_pbcdisabled(struct bfa_s *bfa) 3807 { 3808 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3809 struct bfa_iocfc_s *iocfc = &bfa->iocfc; 3810 struct bfi_iocfc_cfgrsp_s *cfgrsp = iocfc->cfgrsp; 3811 3812 if (cfgrsp->pbc_cfg.port_enabled == BFI_PBC_PORT_DISABLED) { 3813 bfa_trc(bfa, fcport->pwwn); 3814 return BFA_STATUS_PBC; 3815 } 3816 return BFA_STATUS_OK; 3817 } 3818 3819 /* 3820 * Configure port speed. 3821 */ 3822 bfa_status_t 3823 bfa_fcport_cfg_speed(struct bfa_s *bfa, enum bfa_port_speed speed) 3824 { 3825 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3826 3827 bfa_trc(bfa, speed); 3828 3829 if (fcport->cfg.trunked == BFA_TRUE) 3830 return BFA_STATUS_TRUNK_ENABLED; 3831 if ((fcport->cfg.topology == BFA_PORT_TOPOLOGY_LOOP) && 3832 (speed == BFA_PORT_SPEED_16GBPS)) 3833 return BFA_STATUS_UNSUPP_SPEED; 3834 if ((speed != BFA_PORT_SPEED_AUTO) && (speed > fcport->speed_sup)) { 3835 bfa_trc(bfa, fcport->speed_sup); 3836 return BFA_STATUS_UNSUPP_SPEED; 3837 } 3838 3839 /* Port speed entered needs to be checked */ 3840 if (bfa_ioc_get_type(&fcport->bfa->ioc) == BFA_IOC_TYPE_FC) { 3841 /* For CT2, 1G is not supported */ 3842 if ((speed == BFA_PORT_SPEED_1GBPS) && 3843 (bfa_asic_id_ct2(bfa->ioc.pcidev.device_id))) 3844 return BFA_STATUS_UNSUPP_SPEED; 3845 3846 /* Already checked for Auto Speed and Max Speed supp */ 3847 if (!(speed == BFA_PORT_SPEED_1GBPS || 3848 speed == BFA_PORT_SPEED_2GBPS || 3849 speed == BFA_PORT_SPEED_4GBPS || 3850 speed == BFA_PORT_SPEED_8GBPS || 3851 speed == BFA_PORT_SPEED_16GBPS || 3852 speed == BFA_PORT_SPEED_AUTO)) 3853 return BFA_STATUS_UNSUPP_SPEED; 3854 } else { 3855 if (speed != BFA_PORT_SPEED_10GBPS) 3856 return BFA_STATUS_UNSUPP_SPEED; 3857 } 3858 3859 fcport->cfg.speed = speed; 3860 3861 return BFA_STATUS_OK; 3862 } 3863 3864 /* 3865 * Get current speed. 3866 */ 3867 enum bfa_port_speed 3868 bfa_fcport_get_speed(struct bfa_s *bfa) 3869 { 3870 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3871 3872 return fcport->speed; 3873 } 3874 3875 /* 3876 * Configure port topology. 3877 */ 3878 bfa_status_t 3879 bfa_fcport_cfg_topology(struct bfa_s *bfa, enum bfa_port_topology topology) 3880 { 3881 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3882 3883 bfa_trc(bfa, topology); 3884 bfa_trc(bfa, fcport->cfg.topology); 3885 3886 switch (topology) { 3887 case BFA_PORT_TOPOLOGY_P2P: 3888 break; 3889 3890 case BFA_PORT_TOPOLOGY_LOOP: 3891 if ((bfa_fcport_is_qos_enabled(bfa) != BFA_FALSE) || 3892 (fcport->qos_attr.state != BFA_QOS_DISABLED)) 3893 return BFA_STATUS_ERROR_QOS_ENABLED; 3894 if (fcport->cfg.ratelimit != BFA_FALSE) 3895 return BFA_STATUS_ERROR_TRL_ENABLED; 3896 if ((bfa_fcport_is_trunk_enabled(bfa) != BFA_FALSE) || 3897 (fcport->trunk.attr.state != BFA_TRUNK_DISABLED)) 3898 return BFA_STATUS_ERROR_TRUNK_ENABLED; 3899 if ((bfa_fcport_get_speed(bfa) == BFA_PORT_SPEED_16GBPS) || 3900 (fcport->cfg.speed == BFA_PORT_SPEED_16GBPS)) 3901 return BFA_STATUS_UNSUPP_SPEED; 3902 if (bfa_mfg_is_mezz(bfa->ioc.attr->card_type)) 3903 return BFA_STATUS_LOOP_UNSUPP_MEZZ; 3904 if (bfa_fcport_is_dport(bfa) != BFA_FALSE) 3905 return BFA_STATUS_DPORT_ERR; 3906 if (bfa_fcport_is_ddport(bfa) != BFA_FALSE) 3907 return BFA_STATUS_DPORT_ERR; 3908 break; 3909 3910 case BFA_PORT_TOPOLOGY_AUTO: 3911 break; 3912 3913 default: 3914 return BFA_STATUS_EINVAL; 3915 } 3916 3917 fcport->cfg.topology = topology; 3918 return BFA_STATUS_OK; 3919 } 3920 3921 /* 3922 * Get current topology. 3923 */ 3924 enum bfa_port_topology 3925 bfa_fcport_get_topology(struct bfa_s *bfa) 3926 { 3927 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3928 3929 return fcport->topology; 3930 } 3931 3932 /** 3933 * Get config topology. 3934 */ 3935 enum bfa_port_topology 3936 bfa_fcport_get_cfg_topology(struct bfa_s *bfa) 3937 { 3938 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3939 3940 return fcport->cfg.topology; 3941 } 3942 3943 bfa_status_t 3944 bfa_fcport_cfg_hardalpa(struct bfa_s *bfa, u8 alpa) 3945 { 3946 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3947 3948 bfa_trc(bfa, alpa); 3949 bfa_trc(bfa, fcport->cfg.cfg_hardalpa); 3950 bfa_trc(bfa, fcport->cfg.hardalpa); 3951 3952 fcport->cfg.cfg_hardalpa = BFA_TRUE; 3953 fcport->cfg.hardalpa = alpa; 3954 3955 return BFA_STATUS_OK; 3956 } 3957 3958 bfa_status_t 3959 bfa_fcport_clr_hardalpa(struct bfa_s *bfa) 3960 { 3961 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3962 3963 bfa_trc(bfa, fcport->cfg.cfg_hardalpa); 3964 bfa_trc(bfa, fcport->cfg.hardalpa); 3965 3966 fcport->cfg.cfg_hardalpa = BFA_FALSE; 3967 return BFA_STATUS_OK; 3968 } 3969 3970 bfa_boolean_t 3971 bfa_fcport_get_hardalpa(struct bfa_s *bfa, u8 *alpa) 3972 { 3973 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3974 3975 *alpa = fcport->cfg.hardalpa; 3976 return fcport->cfg.cfg_hardalpa; 3977 } 3978 3979 u8 3980 bfa_fcport_get_myalpa(struct bfa_s *bfa) 3981 { 3982 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3983 3984 return fcport->myalpa; 3985 } 3986 3987 bfa_status_t 3988 bfa_fcport_cfg_maxfrsize(struct bfa_s *bfa, u16 maxfrsize) 3989 { 3990 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 3991 3992 bfa_trc(bfa, maxfrsize); 3993 bfa_trc(bfa, fcport->cfg.maxfrsize); 3994 3995 /* with in range */ 3996 if ((maxfrsize > FC_MAX_PDUSZ) || (maxfrsize < FC_MIN_PDUSZ)) 3997 return BFA_STATUS_INVLD_DFSZ; 3998 3999 /* power of 2, if not the max frame size of 2112 */ 4000 if ((maxfrsize != FC_MAX_PDUSZ) && (maxfrsize & (maxfrsize - 1))) 4001 return BFA_STATUS_INVLD_DFSZ; 4002 4003 fcport->cfg.maxfrsize = maxfrsize; 4004 return BFA_STATUS_OK; 4005 } 4006 4007 u16 4008 bfa_fcport_get_maxfrsize(struct bfa_s *bfa) 4009 { 4010 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4011 4012 return fcport->cfg.maxfrsize; 4013 } 4014 4015 u8 4016 bfa_fcport_get_rx_bbcredit(struct bfa_s *bfa) 4017 { 4018 if (bfa_fcport_get_topology(bfa) != BFA_PORT_TOPOLOGY_LOOP) 4019 return (BFA_FCPORT_MOD(bfa))->cfg.rx_bbcredit; 4020 4021 else 4022 return 0; 4023 } 4024 4025 void 4026 bfa_fcport_set_tx_bbcredit(struct bfa_s *bfa, u16 tx_bbcredit) 4027 { 4028 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4029 4030 fcport->cfg.tx_bbcredit = (u8)tx_bbcredit; 4031 } 4032 4033 /* 4034 * Get port attributes. 4035 */ 4036 4037 wwn_t 4038 bfa_fcport_get_wwn(struct bfa_s *bfa, bfa_boolean_t node) 4039 { 4040 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4041 if (node) 4042 return fcport->nwwn; 4043 else 4044 return fcport->pwwn; 4045 } 4046 4047 void 4048 bfa_fcport_get_attr(struct bfa_s *bfa, struct bfa_port_attr_s *attr) 4049 { 4050 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4051 4052 memset(attr, 0, sizeof(struct bfa_port_attr_s)); 4053 4054 attr->nwwn = fcport->nwwn; 4055 attr->pwwn = fcport->pwwn; 4056 4057 attr->factorypwwn = bfa->ioc.attr->mfg_pwwn; 4058 attr->factorynwwn = bfa->ioc.attr->mfg_nwwn; 4059 4060 memcpy(&attr->pport_cfg, &fcport->cfg, 4061 sizeof(struct bfa_port_cfg_s)); 4062 /* speed attributes */ 4063 attr->pport_cfg.speed = fcport->cfg.speed; 4064 attr->speed_supported = fcport->speed_sup; 4065 attr->speed = fcport->speed; 4066 attr->cos_supported = FC_CLASS_3; 4067 4068 /* topology attributes */ 4069 attr->pport_cfg.topology = fcport->cfg.topology; 4070 attr->topology = fcport->topology; 4071 attr->pport_cfg.trunked = fcport->cfg.trunked; 4072 4073 /* beacon attributes */ 4074 attr->beacon = fcport->beacon; 4075 attr->link_e2e_beacon = fcport->link_e2e_beacon; 4076 4077 attr->pport_cfg.path_tov = bfa_fcpim_path_tov_get(bfa); 4078 attr->pport_cfg.q_depth = bfa_fcpim_qdepth_get(bfa); 4079 attr->port_state = bfa_sm_to_state(hal_port_sm_table, fcport->sm); 4080 4081 attr->fec_state = fcport->fec_state; 4082 4083 /* PBC Disabled State */ 4084 if (bfa_fcport_is_pbcdisabled(bfa)) 4085 attr->port_state = BFA_PORT_ST_PREBOOT_DISABLED; 4086 else { 4087 if (bfa_ioc_is_disabled(&fcport->bfa->ioc)) 4088 attr->port_state = BFA_PORT_ST_IOCDIS; 4089 else if (bfa_ioc_fw_mismatch(&fcport->bfa->ioc)) 4090 attr->port_state = BFA_PORT_ST_FWMISMATCH; 4091 } 4092 4093 /* FCoE vlan */ 4094 attr->fcoe_vlan = fcport->fcoe_vlan; 4095 } 4096 4097 #define BFA_FCPORT_STATS_TOV 1000 4098 4099 /* 4100 * Fetch port statistics (FCQoS or FCoE). 4101 */ 4102 bfa_status_t 4103 bfa_fcport_get_stats(struct bfa_s *bfa, struct bfa_cb_pending_q_s *cb) 4104 { 4105 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4106 4107 if (!bfa_iocfc_is_operational(bfa) || 4108 !fcport->stats_dma_ready) 4109 return BFA_STATUS_IOC_NON_OP; 4110 4111 if (!list_empty(&fcport->statsclr_pending_q)) 4112 return BFA_STATUS_DEVBUSY; 4113 4114 if (list_empty(&fcport->stats_pending_q)) { 4115 list_add_tail(&cb->hcb_qe.qe, &fcport->stats_pending_q); 4116 bfa_fcport_send_stats_get(fcport); 4117 bfa_timer_start(bfa, &fcport->timer, 4118 bfa_fcport_stats_get_timeout, 4119 fcport, BFA_FCPORT_STATS_TOV); 4120 } else 4121 list_add_tail(&cb->hcb_qe.qe, &fcport->stats_pending_q); 4122 4123 return BFA_STATUS_OK; 4124 } 4125 4126 /* 4127 * Reset port statistics (FCQoS or FCoE). 4128 */ 4129 bfa_status_t 4130 bfa_fcport_clear_stats(struct bfa_s *bfa, struct bfa_cb_pending_q_s *cb) 4131 { 4132 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4133 4134 if (!bfa_iocfc_is_operational(bfa) || 4135 !fcport->stats_dma_ready) 4136 return BFA_STATUS_IOC_NON_OP; 4137 4138 if (!list_empty(&fcport->stats_pending_q)) 4139 return BFA_STATUS_DEVBUSY; 4140 4141 if (list_empty(&fcport->statsclr_pending_q)) { 4142 list_add_tail(&cb->hcb_qe.qe, &fcport->statsclr_pending_q); 4143 bfa_fcport_send_stats_clear(fcport); 4144 bfa_timer_start(bfa, &fcport->timer, 4145 bfa_fcport_stats_clr_timeout, 4146 fcport, BFA_FCPORT_STATS_TOV); 4147 } else 4148 list_add_tail(&cb->hcb_qe.qe, &fcport->statsclr_pending_q); 4149 4150 return BFA_STATUS_OK; 4151 } 4152 4153 /* 4154 * Fetch port attributes. 4155 */ 4156 bfa_boolean_t 4157 bfa_fcport_is_disabled(struct bfa_s *bfa) 4158 { 4159 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4160 4161 return bfa_sm_to_state(hal_port_sm_table, fcport->sm) == 4162 BFA_PORT_ST_DISABLED; 4163 4164 } 4165 4166 bfa_boolean_t 4167 bfa_fcport_is_dport(struct bfa_s *bfa) 4168 { 4169 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4170 4171 return (bfa_sm_to_state(hal_port_sm_table, fcport->sm) == 4172 BFA_PORT_ST_DPORT); 4173 } 4174 4175 bfa_boolean_t 4176 bfa_fcport_is_ddport(struct bfa_s *bfa) 4177 { 4178 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4179 4180 return (bfa_sm_to_state(hal_port_sm_table, fcport->sm) == 4181 BFA_PORT_ST_DDPORT); 4182 } 4183 4184 bfa_status_t 4185 bfa_fcport_set_qos_bw(struct bfa_s *bfa, struct bfa_qos_bw_s *qos_bw) 4186 { 4187 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4188 enum bfa_ioc_type_e ioc_type = bfa_get_type(bfa); 4189 4190 bfa_trc(bfa, ioc_type); 4191 4192 if ((qos_bw->high == 0) || (qos_bw->med == 0) || (qos_bw->low == 0)) 4193 return BFA_STATUS_QOS_BW_INVALID; 4194 4195 if ((qos_bw->high + qos_bw->med + qos_bw->low) != 100) 4196 return BFA_STATUS_QOS_BW_INVALID; 4197 4198 if ((qos_bw->med > qos_bw->high) || (qos_bw->low > qos_bw->med) || 4199 (qos_bw->low > qos_bw->high)) 4200 return BFA_STATUS_QOS_BW_INVALID; 4201 4202 if ((ioc_type == BFA_IOC_TYPE_FC) && 4203 (fcport->cfg.topology != BFA_PORT_TOPOLOGY_LOOP)) 4204 fcport->cfg.qos_bw = *qos_bw; 4205 4206 return BFA_STATUS_OK; 4207 } 4208 4209 bfa_boolean_t 4210 bfa_fcport_is_ratelim(struct bfa_s *bfa) 4211 { 4212 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4213 4214 return fcport->cfg.ratelimit ? BFA_TRUE : BFA_FALSE; 4215 4216 } 4217 4218 /* 4219 * Enable/Disable FAA feature in port config 4220 */ 4221 void 4222 bfa_fcport_cfg_faa(struct bfa_s *bfa, u8 state) 4223 { 4224 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4225 4226 bfa_trc(bfa, state); 4227 fcport->cfg.faa_state = state; 4228 } 4229 4230 /* 4231 * Get default minimum ratelim speed 4232 */ 4233 enum bfa_port_speed 4234 bfa_fcport_get_ratelim_speed(struct bfa_s *bfa) 4235 { 4236 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4237 4238 bfa_trc(bfa, fcport->cfg.trl_def_speed); 4239 return fcport->cfg.trl_def_speed; 4240 4241 } 4242 4243 void 4244 bfa_fcport_beacon(void *dev, bfa_boolean_t beacon, 4245 bfa_boolean_t link_e2e_beacon) 4246 { 4247 struct bfa_s *bfa = dev; 4248 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4249 4250 bfa_trc(bfa, beacon); 4251 bfa_trc(bfa, link_e2e_beacon); 4252 bfa_trc(bfa, fcport->beacon); 4253 bfa_trc(bfa, fcport->link_e2e_beacon); 4254 4255 fcport->beacon = beacon; 4256 fcport->link_e2e_beacon = link_e2e_beacon; 4257 } 4258 4259 bfa_boolean_t 4260 bfa_fcport_is_linkup(struct bfa_s *bfa) 4261 { 4262 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4263 4264 return (!fcport->cfg.trunked && 4265 bfa_sm_cmp_state(fcport, bfa_fcport_sm_linkup)) || 4266 (fcport->cfg.trunked && 4267 fcport->trunk.attr.state == BFA_TRUNK_ONLINE); 4268 } 4269 4270 bfa_boolean_t 4271 bfa_fcport_is_qos_enabled(struct bfa_s *bfa) 4272 { 4273 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4274 4275 return fcport->cfg.qos_enabled; 4276 } 4277 4278 bfa_boolean_t 4279 bfa_fcport_is_trunk_enabled(struct bfa_s *bfa) 4280 { 4281 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4282 4283 return fcport->cfg.trunked; 4284 } 4285 4286 bfa_status_t 4287 bfa_fcport_cfg_bbcr(struct bfa_s *bfa, bfa_boolean_t on_off, u8 bb_scn) 4288 { 4289 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4290 4291 bfa_trc(bfa, on_off); 4292 4293 if (bfa_ioc_get_type(&fcport->bfa->ioc) != BFA_IOC_TYPE_FC) 4294 return BFA_STATUS_BBCR_FC_ONLY; 4295 4296 if (bfa_mfg_is_mezz(bfa->ioc.attr->card_type) && 4297 (bfa->ioc.attr->card_type != BFA_MFG_TYPE_CHINOOK)) 4298 return BFA_STATUS_CMD_NOTSUPP_MEZZ; 4299 4300 if (on_off) { 4301 if (fcport->cfg.topology == BFA_PORT_TOPOLOGY_LOOP) 4302 return BFA_STATUS_TOPOLOGY_LOOP; 4303 4304 if (fcport->cfg.qos_enabled) 4305 return BFA_STATUS_ERROR_QOS_ENABLED; 4306 4307 if (fcport->cfg.trunked) 4308 return BFA_STATUS_TRUNK_ENABLED; 4309 4310 if ((fcport->cfg.speed != BFA_PORT_SPEED_AUTO) && 4311 (fcport->cfg.speed < bfa_ioc_speed_sup(&bfa->ioc))) 4312 return BFA_STATUS_ERR_BBCR_SPEED_UNSUPPORT; 4313 4314 if (bfa_ioc_speed_sup(&bfa->ioc) < BFA_PORT_SPEED_8GBPS) 4315 return BFA_STATUS_FEATURE_NOT_SUPPORTED; 4316 4317 if (fcport->cfg.bb_cr_enabled) { 4318 if (bb_scn != fcport->cfg.bb_scn) 4319 return BFA_STATUS_BBCR_CFG_NO_CHANGE; 4320 else 4321 return BFA_STATUS_NO_CHANGE; 4322 } 4323 4324 if ((bb_scn == 0) || (bb_scn > BFA_BB_SCN_MAX)) 4325 bb_scn = BFA_BB_SCN_DEF; 4326 4327 fcport->cfg.bb_cr_enabled = on_off; 4328 fcport->cfg.bb_scn = bb_scn; 4329 } else { 4330 if (!fcport->cfg.bb_cr_enabled) 4331 return BFA_STATUS_NO_CHANGE; 4332 4333 fcport->cfg.bb_cr_enabled = on_off; 4334 fcport->cfg.bb_scn = 0; 4335 } 4336 4337 return BFA_STATUS_OK; 4338 } 4339 4340 bfa_status_t 4341 bfa_fcport_get_bbcr_attr(struct bfa_s *bfa, 4342 struct bfa_bbcr_attr_s *bbcr_attr) 4343 { 4344 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(bfa); 4345 4346 if (bfa_ioc_get_type(&fcport->bfa->ioc) != BFA_IOC_TYPE_FC) 4347 return BFA_STATUS_BBCR_FC_ONLY; 4348 4349 if (fcport->cfg.topology == BFA_PORT_TOPOLOGY_LOOP) 4350 return BFA_STATUS_TOPOLOGY_LOOP; 4351 4352 *bbcr_attr = fcport->bbcr_attr; 4353 4354 return BFA_STATUS_OK; 4355 } 4356 4357 void 4358 bfa_fcport_dportenable(struct bfa_s *bfa) 4359 { 4360 /* 4361 * Assume caller check for port is in disable state 4362 */ 4363 bfa_sm_send_event(BFA_FCPORT_MOD(bfa), BFA_FCPORT_SM_DPORTENABLE); 4364 bfa_port_set_dportenabled(&bfa->modules.port, BFA_TRUE); 4365 } 4366 4367 void 4368 bfa_fcport_dportdisable(struct bfa_s *bfa) 4369 { 4370 /* 4371 * Assume caller check for port is in disable state 4372 */ 4373 bfa_sm_send_event(BFA_FCPORT_MOD(bfa), BFA_FCPORT_SM_DPORTDISABLE); 4374 bfa_port_set_dportenabled(&bfa->modules.port, BFA_FALSE); 4375 } 4376 4377 void 4378 bfa_fcport_ddportenable(struct bfa_s *bfa) 4379 { 4380 /* 4381 * Assume caller check for port is in disable state 4382 */ 4383 bfa_sm_send_event(BFA_FCPORT_MOD(bfa), BFA_FCPORT_SM_DDPORTENABLE); 4384 } 4385 4386 void 4387 bfa_fcport_ddportdisable(struct bfa_s *bfa) 4388 { 4389 /* 4390 * Assume caller check for port is in disable state 4391 */ 4392 bfa_sm_send_event(BFA_FCPORT_MOD(bfa), BFA_FCPORT_SM_DDPORTDISABLE); 4393 } 4394 4395 /* 4396 * Rport State machine functions 4397 */ 4398 /* 4399 * Beginning state, only online event expected. 4400 */ 4401 static void 4402 bfa_rport_sm_uninit(struct bfa_rport_s *rp, enum bfa_rport_event event) 4403 { 4404 bfa_trc(rp->bfa, rp->rport_tag); 4405 bfa_trc(rp->bfa, event); 4406 4407 switch (event) { 4408 case BFA_RPORT_SM_CREATE: 4409 bfa_stats(rp, sm_un_cr); 4410 bfa_sm_set_state(rp, bfa_rport_sm_created); 4411 break; 4412 4413 default: 4414 bfa_stats(rp, sm_un_unexp); 4415 bfa_sm_fault(rp->bfa, event); 4416 } 4417 } 4418 4419 static void 4420 bfa_rport_sm_created(struct bfa_rport_s *rp, enum bfa_rport_event event) 4421 { 4422 bfa_trc(rp->bfa, rp->rport_tag); 4423 bfa_trc(rp->bfa, event); 4424 4425 switch (event) { 4426 case BFA_RPORT_SM_ONLINE: 4427 bfa_stats(rp, sm_cr_on); 4428 if (bfa_rport_send_fwcreate(rp)) 4429 bfa_sm_set_state(rp, bfa_rport_sm_fwcreate); 4430 else 4431 bfa_sm_set_state(rp, bfa_rport_sm_fwcreate_qfull); 4432 break; 4433 4434 case BFA_RPORT_SM_DELETE: 4435 bfa_stats(rp, sm_cr_del); 4436 bfa_sm_set_state(rp, bfa_rport_sm_uninit); 4437 bfa_rport_free(rp); 4438 break; 4439 4440 case BFA_RPORT_SM_HWFAIL: 4441 bfa_stats(rp, sm_cr_hwf); 4442 bfa_sm_set_state(rp, bfa_rport_sm_iocdisable); 4443 break; 4444 4445 default: 4446 bfa_stats(rp, sm_cr_unexp); 4447 bfa_sm_fault(rp->bfa, event); 4448 } 4449 } 4450 4451 /* 4452 * Waiting for rport create response from firmware. 4453 */ 4454 static void 4455 bfa_rport_sm_fwcreate(struct bfa_rport_s *rp, enum bfa_rport_event event) 4456 { 4457 bfa_trc(rp->bfa, rp->rport_tag); 4458 bfa_trc(rp->bfa, event); 4459 4460 switch (event) { 4461 case BFA_RPORT_SM_FWRSP: 4462 bfa_stats(rp, sm_fwc_rsp); 4463 bfa_sm_set_state(rp, bfa_rport_sm_online); 4464 bfa_rport_online_cb(rp); 4465 break; 4466 4467 case BFA_RPORT_SM_DELETE: 4468 bfa_stats(rp, sm_fwc_del); 4469 bfa_sm_set_state(rp, bfa_rport_sm_delete_pending); 4470 break; 4471 4472 case BFA_RPORT_SM_OFFLINE: 4473 bfa_stats(rp, sm_fwc_off); 4474 bfa_sm_set_state(rp, bfa_rport_sm_offline_pending); 4475 break; 4476 4477 case BFA_RPORT_SM_HWFAIL: 4478 bfa_stats(rp, sm_fwc_hwf); 4479 bfa_sm_set_state(rp, bfa_rport_sm_iocdisable); 4480 break; 4481 4482 default: 4483 bfa_stats(rp, sm_fwc_unexp); 4484 bfa_sm_fault(rp->bfa, event); 4485 } 4486 } 4487 4488 /* 4489 * Request queue is full, awaiting queue resume to send create request. 4490 */ 4491 static void 4492 bfa_rport_sm_fwcreate_qfull(struct bfa_rport_s *rp, enum bfa_rport_event event) 4493 { 4494 bfa_trc(rp->bfa, rp->rport_tag); 4495 bfa_trc(rp->bfa, event); 4496 4497 switch (event) { 4498 case BFA_RPORT_SM_QRESUME: 4499 bfa_sm_set_state(rp, bfa_rport_sm_fwcreate); 4500 bfa_rport_send_fwcreate(rp); 4501 break; 4502 4503 case BFA_RPORT_SM_DELETE: 4504 bfa_stats(rp, sm_fwc_del); 4505 bfa_sm_set_state(rp, bfa_rport_sm_uninit); 4506 bfa_reqq_wcancel(&rp->reqq_wait); 4507 bfa_rport_free(rp); 4508 break; 4509 4510 case BFA_RPORT_SM_OFFLINE: 4511 bfa_stats(rp, sm_fwc_off); 4512 bfa_sm_set_state(rp, bfa_rport_sm_offline); 4513 bfa_reqq_wcancel(&rp->reqq_wait); 4514 bfa_rport_offline_cb(rp); 4515 break; 4516 4517 case BFA_RPORT_SM_HWFAIL: 4518 bfa_stats(rp, sm_fwc_hwf); 4519 bfa_sm_set_state(rp, bfa_rport_sm_iocdisable); 4520 bfa_reqq_wcancel(&rp->reqq_wait); 4521 break; 4522 4523 default: 4524 bfa_stats(rp, sm_fwc_unexp); 4525 bfa_sm_fault(rp->bfa, event); 4526 } 4527 } 4528 4529 /* 4530 * Online state - normal parking state. 4531 */ 4532 static void 4533 bfa_rport_sm_online(struct bfa_rport_s *rp, enum bfa_rport_event event) 4534 { 4535 struct bfi_rport_qos_scn_s *qos_scn; 4536 4537 bfa_trc(rp->bfa, rp->rport_tag); 4538 bfa_trc(rp->bfa, event); 4539 4540 switch (event) { 4541 case BFA_RPORT_SM_OFFLINE: 4542 bfa_stats(rp, sm_on_off); 4543 if (bfa_rport_send_fwdelete(rp)) 4544 bfa_sm_set_state(rp, bfa_rport_sm_fwdelete); 4545 else 4546 bfa_sm_set_state(rp, bfa_rport_sm_fwdelete_qfull); 4547 break; 4548 4549 case BFA_RPORT_SM_DELETE: 4550 bfa_stats(rp, sm_on_del); 4551 if (bfa_rport_send_fwdelete(rp)) 4552 bfa_sm_set_state(rp, bfa_rport_sm_deleting); 4553 else 4554 bfa_sm_set_state(rp, bfa_rport_sm_deleting_qfull); 4555 break; 4556 4557 case BFA_RPORT_SM_HWFAIL: 4558 bfa_stats(rp, sm_on_hwf); 4559 bfa_sm_set_state(rp, bfa_rport_sm_iocdisable); 4560 break; 4561 4562 case BFA_RPORT_SM_SET_SPEED: 4563 bfa_rport_send_fwspeed(rp); 4564 break; 4565 4566 case BFA_RPORT_SM_QOS_SCN: 4567 qos_scn = (struct bfi_rport_qos_scn_s *) rp->event_arg.fw_msg; 4568 rp->qos_attr = qos_scn->new_qos_attr; 4569 bfa_trc(rp->bfa, qos_scn->old_qos_attr.qos_flow_id); 4570 bfa_trc(rp->bfa, qos_scn->new_qos_attr.qos_flow_id); 4571 bfa_trc(rp->bfa, qos_scn->old_qos_attr.qos_priority); 4572 bfa_trc(rp->bfa, qos_scn->new_qos_attr.qos_priority); 4573 4574 qos_scn->old_qos_attr.qos_flow_id = 4575 be32_to_cpu(qos_scn->old_qos_attr.qos_flow_id); 4576 qos_scn->new_qos_attr.qos_flow_id = 4577 be32_to_cpu(qos_scn->new_qos_attr.qos_flow_id); 4578 4579 if (qos_scn->old_qos_attr.qos_flow_id != 4580 qos_scn->new_qos_attr.qos_flow_id) 4581 bfa_cb_rport_qos_scn_flowid(rp->rport_drv, 4582 qos_scn->old_qos_attr, 4583 qos_scn->new_qos_attr); 4584 if (qos_scn->old_qos_attr.qos_priority != 4585 qos_scn->new_qos_attr.qos_priority) 4586 bfa_cb_rport_qos_scn_prio(rp->rport_drv, 4587 qos_scn->old_qos_attr, 4588 qos_scn->new_qos_attr); 4589 break; 4590 4591 default: 4592 bfa_stats(rp, sm_on_unexp); 4593 bfa_sm_fault(rp->bfa, event); 4594 } 4595 } 4596 4597 /* 4598 * Firmware rport is being deleted - awaiting f/w response. 4599 */ 4600 static void 4601 bfa_rport_sm_fwdelete(struct bfa_rport_s *rp, enum bfa_rport_event event) 4602 { 4603 bfa_trc(rp->bfa, rp->rport_tag); 4604 bfa_trc(rp->bfa, event); 4605 4606 switch (event) { 4607 case BFA_RPORT_SM_FWRSP: 4608 bfa_stats(rp, sm_fwd_rsp); 4609 bfa_sm_set_state(rp, bfa_rport_sm_offline); 4610 bfa_rport_offline_cb(rp); 4611 break; 4612 4613 case BFA_RPORT_SM_DELETE: 4614 bfa_stats(rp, sm_fwd_del); 4615 bfa_sm_set_state(rp, bfa_rport_sm_deleting); 4616 break; 4617 4618 case BFA_RPORT_SM_HWFAIL: 4619 bfa_stats(rp, sm_fwd_hwf); 4620 bfa_sm_set_state(rp, bfa_rport_sm_iocdisable); 4621 bfa_rport_offline_cb(rp); 4622 break; 4623 4624 default: 4625 bfa_stats(rp, sm_fwd_unexp); 4626 bfa_sm_fault(rp->bfa, event); 4627 } 4628 } 4629 4630 static void 4631 bfa_rport_sm_fwdelete_qfull(struct bfa_rport_s *rp, enum bfa_rport_event event) 4632 { 4633 bfa_trc(rp->bfa, rp->rport_tag); 4634 bfa_trc(rp->bfa, event); 4635 4636 switch (event) { 4637 case BFA_RPORT_SM_QRESUME: 4638 bfa_sm_set_state(rp, bfa_rport_sm_fwdelete); 4639 bfa_rport_send_fwdelete(rp); 4640 break; 4641 4642 case BFA_RPORT_SM_DELETE: 4643 bfa_stats(rp, sm_fwd_del); 4644 bfa_sm_set_state(rp, bfa_rport_sm_deleting_qfull); 4645 break; 4646 4647 case BFA_RPORT_SM_HWFAIL: 4648 bfa_stats(rp, sm_fwd_hwf); 4649 bfa_sm_set_state(rp, bfa_rport_sm_iocdisable); 4650 bfa_reqq_wcancel(&rp->reqq_wait); 4651 bfa_rport_offline_cb(rp); 4652 break; 4653 4654 default: 4655 bfa_stats(rp, sm_fwd_unexp); 4656 bfa_sm_fault(rp->bfa, event); 4657 } 4658 } 4659 4660 /* 4661 * Offline state. 4662 */ 4663 static void 4664 bfa_rport_sm_offline(struct bfa_rport_s *rp, enum bfa_rport_event event) 4665 { 4666 bfa_trc(rp->bfa, rp->rport_tag); 4667 bfa_trc(rp->bfa, event); 4668 4669 switch (event) { 4670 case BFA_RPORT_SM_DELETE: 4671 bfa_stats(rp, sm_off_del); 4672 bfa_sm_set_state(rp, bfa_rport_sm_uninit); 4673 bfa_rport_free(rp); 4674 break; 4675 4676 case BFA_RPORT_SM_ONLINE: 4677 bfa_stats(rp, sm_off_on); 4678 if (bfa_rport_send_fwcreate(rp)) 4679 bfa_sm_set_state(rp, bfa_rport_sm_fwcreate); 4680 else 4681 bfa_sm_set_state(rp, bfa_rport_sm_fwcreate_qfull); 4682 break; 4683 4684 case BFA_RPORT_SM_HWFAIL: 4685 bfa_stats(rp, sm_off_hwf); 4686 bfa_sm_set_state(rp, bfa_rport_sm_iocdisable); 4687 break; 4688 4689 case BFA_RPORT_SM_OFFLINE: 4690 bfa_rport_offline_cb(rp); 4691 break; 4692 4693 default: 4694 bfa_stats(rp, sm_off_unexp); 4695 bfa_sm_fault(rp->bfa, event); 4696 } 4697 } 4698 4699 /* 4700 * Rport is deleted, waiting for firmware response to delete. 4701 */ 4702 static void 4703 bfa_rport_sm_deleting(struct bfa_rport_s *rp, enum bfa_rport_event event) 4704 { 4705 bfa_trc(rp->bfa, rp->rport_tag); 4706 bfa_trc(rp->bfa, event); 4707 4708 switch (event) { 4709 case BFA_RPORT_SM_FWRSP: 4710 bfa_stats(rp, sm_del_fwrsp); 4711 bfa_sm_set_state(rp, bfa_rport_sm_uninit); 4712 bfa_rport_free(rp); 4713 break; 4714 4715 case BFA_RPORT_SM_HWFAIL: 4716 bfa_stats(rp, sm_del_hwf); 4717 bfa_sm_set_state(rp, bfa_rport_sm_uninit); 4718 bfa_rport_free(rp); 4719 break; 4720 4721 default: 4722 bfa_sm_fault(rp->bfa, event); 4723 } 4724 } 4725 4726 static void 4727 bfa_rport_sm_deleting_qfull(struct bfa_rport_s *rp, enum bfa_rport_event event) 4728 { 4729 bfa_trc(rp->bfa, rp->rport_tag); 4730 bfa_trc(rp->bfa, event); 4731 4732 switch (event) { 4733 case BFA_RPORT_SM_QRESUME: 4734 bfa_stats(rp, sm_del_fwrsp); 4735 bfa_sm_set_state(rp, bfa_rport_sm_deleting); 4736 bfa_rport_send_fwdelete(rp); 4737 break; 4738 4739 case BFA_RPORT_SM_HWFAIL: 4740 bfa_stats(rp, sm_del_hwf); 4741 bfa_sm_set_state(rp, bfa_rport_sm_uninit); 4742 bfa_reqq_wcancel(&rp->reqq_wait); 4743 bfa_rport_free(rp); 4744 break; 4745 4746 default: 4747 bfa_sm_fault(rp->bfa, event); 4748 } 4749 } 4750 4751 /* 4752 * Waiting for rport create response from firmware. A delete is pending. 4753 */ 4754 static void 4755 bfa_rport_sm_delete_pending(struct bfa_rport_s *rp, 4756 enum bfa_rport_event event) 4757 { 4758 bfa_trc(rp->bfa, rp->rport_tag); 4759 bfa_trc(rp->bfa, event); 4760 4761 switch (event) { 4762 case BFA_RPORT_SM_FWRSP: 4763 bfa_stats(rp, sm_delp_fwrsp); 4764 if (bfa_rport_send_fwdelete(rp)) 4765 bfa_sm_set_state(rp, bfa_rport_sm_deleting); 4766 else 4767 bfa_sm_set_state(rp, bfa_rport_sm_deleting_qfull); 4768 break; 4769 4770 case BFA_RPORT_SM_HWFAIL: 4771 bfa_stats(rp, sm_delp_hwf); 4772 bfa_sm_set_state(rp, bfa_rport_sm_uninit); 4773 bfa_rport_free(rp); 4774 break; 4775 4776 default: 4777 bfa_stats(rp, sm_delp_unexp); 4778 bfa_sm_fault(rp->bfa, event); 4779 } 4780 } 4781 4782 /* 4783 * Waiting for rport create response from firmware. Rport offline is pending. 4784 */ 4785 static void 4786 bfa_rport_sm_offline_pending(struct bfa_rport_s *rp, 4787 enum bfa_rport_event event) 4788 { 4789 bfa_trc(rp->bfa, rp->rport_tag); 4790 bfa_trc(rp->bfa, event); 4791 4792 switch (event) { 4793 case BFA_RPORT_SM_FWRSP: 4794 bfa_stats(rp, sm_offp_fwrsp); 4795 if (bfa_rport_send_fwdelete(rp)) 4796 bfa_sm_set_state(rp, bfa_rport_sm_fwdelete); 4797 else 4798 bfa_sm_set_state(rp, bfa_rport_sm_fwdelete_qfull); 4799 break; 4800 4801 case BFA_RPORT_SM_DELETE: 4802 bfa_stats(rp, sm_offp_del); 4803 bfa_sm_set_state(rp, bfa_rport_sm_delete_pending); 4804 break; 4805 4806 case BFA_RPORT_SM_HWFAIL: 4807 bfa_stats(rp, sm_offp_hwf); 4808 bfa_sm_set_state(rp, bfa_rport_sm_iocdisable); 4809 bfa_rport_offline_cb(rp); 4810 break; 4811 4812 default: 4813 bfa_stats(rp, sm_offp_unexp); 4814 bfa_sm_fault(rp->bfa, event); 4815 } 4816 } 4817 4818 /* 4819 * IOC h/w failed. 4820 */ 4821 static void 4822 bfa_rport_sm_iocdisable(struct bfa_rport_s *rp, enum bfa_rport_event event) 4823 { 4824 bfa_trc(rp->bfa, rp->rport_tag); 4825 bfa_trc(rp->bfa, event); 4826 4827 switch (event) { 4828 case BFA_RPORT_SM_OFFLINE: 4829 bfa_stats(rp, sm_iocd_off); 4830 bfa_rport_offline_cb(rp); 4831 break; 4832 4833 case BFA_RPORT_SM_DELETE: 4834 bfa_stats(rp, sm_iocd_del); 4835 bfa_sm_set_state(rp, bfa_rport_sm_uninit); 4836 bfa_rport_free(rp); 4837 break; 4838 4839 case BFA_RPORT_SM_ONLINE: 4840 bfa_stats(rp, sm_iocd_on); 4841 if (bfa_rport_send_fwcreate(rp)) 4842 bfa_sm_set_state(rp, bfa_rport_sm_fwcreate); 4843 else 4844 bfa_sm_set_state(rp, bfa_rport_sm_fwcreate_qfull); 4845 break; 4846 4847 case BFA_RPORT_SM_HWFAIL: 4848 break; 4849 4850 default: 4851 bfa_stats(rp, sm_iocd_unexp); 4852 bfa_sm_fault(rp->bfa, event); 4853 } 4854 } 4855 4856 4857 4858 /* 4859 * bfa_rport_private BFA rport private functions 4860 */ 4861 4862 static void 4863 __bfa_cb_rport_online(void *cbarg, bfa_boolean_t complete) 4864 { 4865 struct bfa_rport_s *rp = cbarg; 4866 4867 if (complete) 4868 bfa_cb_rport_online(rp->rport_drv); 4869 } 4870 4871 static void 4872 __bfa_cb_rport_offline(void *cbarg, bfa_boolean_t complete) 4873 { 4874 struct bfa_rport_s *rp = cbarg; 4875 4876 if (complete) 4877 bfa_cb_rport_offline(rp->rport_drv); 4878 } 4879 4880 static void 4881 bfa_rport_qresume(void *cbarg) 4882 { 4883 struct bfa_rport_s *rp = cbarg; 4884 4885 bfa_sm_send_event(rp, BFA_RPORT_SM_QRESUME); 4886 } 4887 4888 static void 4889 bfa_rport_meminfo(struct bfa_iocfc_cfg_s *cfg, struct bfa_meminfo_s *minfo, 4890 struct bfa_s *bfa) 4891 { 4892 struct bfa_mem_kva_s *rport_kva = BFA_MEM_RPORT_KVA(bfa); 4893 4894 if (cfg->fwcfg.num_rports < BFA_RPORT_MIN) 4895 cfg->fwcfg.num_rports = BFA_RPORT_MIN; 4896 4897 /* kva memory */ 4898 bfa_mem_kva_setup(minfo, rport_kva, 4899 cfg->fwcfg.num_rports * sizeof(struct bfa_rport_s)); 4900 } 4901 4902 static void 4903 bfa_rport_attach(struct bfa_s *bfa, void *bfad, struct bfa_iocfc_cfg_s *cfg, 4904 struct bfa_pcidev_s *pcidev) 4905 { 4906 struct bfa_rport_mod_s *mod = BFA_RPORT_MOD(bfa); 4907 struct bfa_rport_s *rp; 4908 u16 i; 4909 4910 INIT_LIST_HEAD(&mod->rp_free_q); 4911 INIT_LIST_HEAD(&mod->rp_active_q); 4912 INIT_LIST_HEAD(&mod->rp_unused_q); 4913 4914 rp = (struct bfa_rport_s *) bfa_mem_kva_curp(mod); 4915 mod->rps_list = rp; 4916 mod->num_rports = cfg->fwcfg.num_rports; 4917 4918 WARN_ON(!mod->num_rports || 4919 (mod->num_rports & (mod->num_rports - 1))); 4920 4921 for (i = 0; i < mod->num_rports; i++, rp++) { 4922 memset(rp, 0, sizeof(struct bfa_rport_s)); 4923 rp->bfa = bfa; 4924 rp->rport_tag = i; 4925 bfa_sm_set_state(rp, bfa_rport_sm_uninit); 4926 4927 /* 4928 * - is unused 4929 */ 4930 if (i) 4931 list_add_tail(&rp->qe, &mod->rp_free_q); 4932 4933 bfa_reqq_winit(&rp->reqq_wait, bfa_rport_qresume, rp); 4934 } 4935 4936 /* 4937 * consume memory 4938 */ 4939 bfa_mem_kva_curp(mod) = (u8 *) rp; 4940 } 4941 4942 static void 4943 bfa_rport_detach(struct bfa_s *bfa) 4944 { 4945 } 4946 4947 static void 4948 bfa_rport_start(struct bfa_s *bfa) 4949 { 4950 } 4951 4952 static void 4953 bfa_rport_stop(struct bfa_s *bfa) 4954 { 4955 } 4956 4957 static void 4958 bfa_rport_iocdisable(struct bfa_s *bfa) 4959 { 4960 struct bfa_rport_mod_s *mod = BFA_RPORT_MOD(bfa); 4961 struct bfa_rport_s *rport; 4962 struct list_head *qe, *qen; 4963 4964 /* Enqueue unused rport resources to free_q */ 4965 list_splice_tail_init(&mod->rp_unused_q, &mod->rp_free_q); 4966 4967 list_for_each_safe(qe, qen, &mod->rp_active_q) { 4968 rport = (struct bfa_rport_s *) qe; 4969 bfa_sm_send_event(rport, BFA_RPORT_SM_HWFAIL); 4970 } 4971 } 4972 4973 static struct bfa_rport_s * 4974 bfa_rport_alloc(struct bfa_rport_mod_s *mod) 4975 { 4976 struct bfa_rport_s *rport; 4977 4978 bfa_q_deq(&mod->rp_free_q, &rport); 4979 if (rport) 4980 list_add_tail(&rport->qe, &mod->rp_active_q); 4981 4982 return rport; 4983 } 4984 4985 static void 4986 bfa_rport_free(struct bfa_rport_s *rport) 4987 { 4988 struct bfa_rport_mod_s *mod = BFA_RPORT_MOD(rport->bfa); 4989 4990 WARN_ON(!bfa_q_is_on_q(&mod->rp_active_q, rport)); 4991 list_del(&rport->qe); 4992 list_add_tail(&rport->qe, &mod->rp_free_q); 4993 } 4994 4995 static bfa_boolean_t 4996 bfa_rport_send_fwcreate(struct bfa_rport_s *rp) 4997 { 4998 struct bfi_rport_create_req_s *m; 4999 5000 /* 5001 * check for room in queue to send request now 5002 */ 5003 m = bfa_reqq_next(rp->bfa, BFA_REQQ_RPORT); 5004 if (!m) { 5005 bfa_reqq_wait(rp->bfa, BFA_REQQ_RPORT, &rp->reqq_wait); 5006 return BFA_FALSE; 5007 } 5008 5009 bfi_h2i_set(m->mh, BFI_MC_RPORT, BFI_RPORT_H2I_CREATE_REQ, 5010 bfa_fn_lpu(rp->bfa)); 5011 m->bfa_handle = rp->rport_tag; 5012 m->max_frmsz = cpu_to_be16(rp->rport_info.max_frmsz); 5013 m->pid = rp->rport_info.pid; 5014 m->lp_fwtag = bfa_lps_get_fwtag(rp->bfa, (u8)rp->rport_info.lp_tag); 5015 m->local_pid = rp->rport_info.local_pid; 5016 m->fc_class = rp->rport_info.fc_class; 5017 m->vf_en = rp->rport_info.vf_en; 5018 m->vf_id = rp->rport_info.vf_id; 5019 m->cisc = rp->rport_info.cisc; 5020 5021 /* 5022 * queue I/O message to firmware 5023 */ 5024 bfa_reqq_produce(rp->bfa, BFA_REQQ_RPORT, m->mh); 5025 return BFA_TRUE; 5026 } 5027 5028 static bfa_boolean_t 5029 bfa_rport_send_fwdelete(struct bfa_rport_s *rp) 5030 { 5031 struct bfi_rport_delete_req_s *m; 5032 5033 /* 5034 * check for room in queue to send request now 5035 */ 5036 m = bfa_reqq_next(rp->bfa, BFA_REQQ_RPORT); 5037 if (!m) { 5038 bfa_reqq_wait(rp->bfa, BFA_REQQ_RPORT, &rp->reqq_wait); 5039 return BFA_FALSE; 5040 } 5041 5042 bfi_h2i_set(m->mh, BFI_MC_RPORT, BFI_RPORT_H2I_DELETE_REQ, 5043 bfa_fn_lpu(rp->bfa)); 5044 m->fw_handle = rp->fw_handle; 5045 5046 /* 5047 * queue I/O message to firmware 5048 */ 5049 bfa_reqq_produce(rp->bfa, BFA_REQQ_RPORT, m->mh); 5050 return BFA_TRUE; 5051 } 5052 5053 static bfa_boolean_t 5054 bfa_rport_send_fwspeed(struct bfa_rport_s *rp) 5055 { 5056 struct bfa_rport_speed_req_s *m; 5057 5058 /* 5059 * check for room in queue to send request now 5060 */ 5061 m = bfa_reqq_next(rp->bfa, BFA_REQQ_RPORT); 5062 if (!m) { 5063 bfa_trc(rp->bfa, rp->rport_info.speed); 5064 return BFA_FALSE; 5065 } 5066 5067 bfi_h2i_set(m->mh, BFI_MC_RPORT, BFI_RPORT_H2I_SET_SPEED_REQ, 5068 bfa_fn_lpu(rp->bfa)); 5069 m->fw_handle = rp->fw_handle; 5070 m->speed = (u8)rp->rport_info.speed; 5071 5072 /* 5073 * queue I/O message to firmware 5074 */ 5075 bfa_reqq_produce(rp->bfa, BFA_REQQ_RPORT, m->mh); 5076 return BFA_TRUE; 5077 } 5078 5079 5080 5081 /* 5082 * bfa_rport_public 5083 */ 5084 5085 /* 5086 * Rport interrupt processing. 5087 */ 5088 void 5089 bfa_rport_isr(struct bfa_s *bfa, struct bfi_msg_s *m) 5090 { 5091 union bfi_rport_i2h_msg_u msg; 5092 struct bfa_rport_s *rp; 5093 5094 bfa_trc(bfa, m->mhdr.msg_id); 5095 5096 msg.msg = m; 5097 5098 switch (m->mhdr.msg_id) { 5099 case BFI_RPORT_I2H_CREATE_RSP: 5100 rp = BFA_RPORT_FROM_TAG(bfa, msg.create_rsp->bfa_handle); 5101 rp->fw_handle = msg.create_rsp->fw_handle; 5102 rp->qos_attr = msg.create_rsp->qos_attr; 5103 bfa_rport_set_lunmask(bfa, rp); 5104 WARN_ON(msg.create_rsp->status != BFA_STATUS_OK); 5105 bfa_sm_send_event(rp, BFA_RPORT_SM_FWRSP); 5106 break; 5107 5108 case BFI_RPORT_I2H_DELETE_RSP: 5109 rp = BFA_RPORT_FROM_TAG(bfa, msg.delete_rsp->bfa_handle); 5110 WARN_ON(msg.delete_rsp->status != BFA_STATUS_OK); 5111 bfa_rport_unset_lunmask(bfa, rp); 5112 bfa_sm_send_event(rp, BFA_RPORT_SM_FWRSP); 5113 break; 5114 5115 case BFI_RPORT_I2H_QOS_SCN: 5116 rp = BFA_RPORT_FROM_TAG(bfa, msg.qos_scn_evt->bfa_handle); 5117 rp->event_arg.fw_msg = msg.qos_scn_evt; 5118 bfa_sm_send_event(rp, BFA_RPORT_SM_QOS_SCN); 5119 break; 5120 5121 case BFI_RPORT_I2H_LIP_SCN_ONLINE: 5122 bfa_fcport_update_loop_info(BFA_FCPORT_MOD(bfa), 5123 &msg.lip_scn->loop_info); 5124 bfa_cb_rport_scn_online(bfa); 5125 break; 5126 5127 case BFI_RPORT_I2H_LIP_SCN_OFFLINE: 5128 bfa_cb_rport_scn_offline(bfa); 5129 break; 5130 5131 case BFI_RPORT_I2H_NO_DEV: 5132 rp = BFA_RPORT_FROM_TAG(bfa, msg.lip_scn->bfa_handle); 5133 bfa_cb_rport_scn_no_dev(rp->rport_drv); 5134 break; 5135 5136 default: 5137 bfa_trc(bfa, m->mhdr.msg_id); 5138 WARN_ON(1); 5139 } 5140 } 5141 5142 void 5143 bfa_rport_res_recfg(struct bfa_s *bfa, u16 num_rport_fw) 5144 { 5145 struct bfa_rport_mod_s *mod = BFA_RPORT_MOD(bfa); 5146 struct list_head *qe; 5147 int i; 5148 5149 for (i = 0; i < (mod->num_rports - num_rport_fw); i++) { 5150 bfa_q_deq_tail(&mod->rp_free_q, &qe); 5151 list_add_tail(qe, &mod->rp_unused_q); 5152 } 5153 } 5154 5155 /* 5156 * bfa_rport_api 5157 */ 5158 5159 struct bfa_rport_s * 5160 bfa_rport_create(struct bfa_s *bfa, void *rport_drv) 5161 { 5162 struct bfa_rport_s *rp; 5163 5164 rp = bfa_rport_alloc(BFA_RPORT_MOD(bfa)); 5165 5166 if (rp == NULL) 5167 return NULL; 5168 5169 rp->bfa = bfa; 5170 rp->rport_drv = rport_drv; 5171 memset(&rp->stats, 0, sizeof(rp->stats)); 5172 5173 WARN_ON(!bfa_sm_cmp_state(rp, bfa_rport_sm_uninit)); 5174 bfa_sm_send_event(rp, BFA_RPORT_SM_CREATE); 5175 5176 return rp; 5177 } 5178 5179 void 5180 bfa_rport_online(struct bfa_rport_s *rport, struct bfa_rport_info_s *rport_info) 5181 { 5182 WARN_ON(rport_info->max_frmsz == 0); 5183 5184 /* 5185 * Some JBODs are seen to be not setting PDU size correctly in PLOGI 5186 * responses. Default to minimum size. 5187 */ 5188 if (rport_info->max_frmsz == 0) { 5189 bfa_trc(rport->bfa, rport->rport_tag); 5190 rport_info->max_frmsz = FC_MIN_PDUSZ; 5191 } 5192 5193 rport->rport_info = *rport_info; 5194 bfa_sm_send_event(rport, BFA_RPORT_SM_ONLINE); 5195 } 5196 5197 void 5198 bfa_rport_speed(struct bfa_rport_s *rport, enum bfa_port_speed speed) 5199 { 5200 WARN_ON(speed == 0); 5201 WARN_ON(speed == BFA_PORT_SPEED_AUTO); 5202 5203 if (rport) { 5204 rport->rport_info.speed = speed; 5205 bfa_sm_send_event(rport, BFA_RPORT_SM_SET_SPEED); 5206 } 5207 } 5208 5209 /* Set Rport LUN Mask */ 5210 void 5211 bfa_rport_set_lunmask(struct bfa_s *bfa, struct bfa_rport_s *rp) 5212 { 5213 struct bfa_lps_mod_s *lps_mod = BFA_LPS_MOD(bfa); 5214 wwn_t lp_wwn, rp_wwn; 5215 u8 lp_tag = (u8)rp->rport_info.lp_tag; 5216 5217 rp_wwn = ((struct bfa_fcs_rport_s *)rp->rport_drv)->pwwn; 5218 lp_wwn = (BFA_LPS_FROM_TAG(lps_mod, rp->rport_info.lp_tag))->pwwn; 5219 5220 BFA_LPS_FROM_TAG(lps_mod, rp->rport_info.lp_tag)->lun_mask = 5221 rp->lun_mask = BFA_TRUE; 5222 bfa_fcpim_lunmask_rp_update(bfa, lp_wwn, rp_wwn, rp->rport_tag, lp_tag); 5223 } 5224 5225 /* Unset Rport LUN mask */ 5226 void 5227 bfa_rport_unset_lunmask(struct bfa_s *bfa, struct bfa_rport_s *rp) 5228 { 5229 struct bfa_lps_mod_s *lps_mod = BFA_LPS_MOD(bfa); 5230 wwn_t lp_wwn, rp_wwn; 5231 5232 rp_wwn = ((struct bfa_fcs_rport_s *)rp->rport_drv)->pwwn; 5233 lp_wwn = (BFA_LPS_FROM_TAG(lps_mod, rp->rport_info.lp_tag))->pwwn; 5234 5235 BFA_LPS_FROM_TAG(lps_mod, rp->rport_info.lp_tag)->lun_mask = 5236 rp->lun_mask = BFA_FALSE; 5237 bfa_fcpim_lunmask_rp_update(bfa, lp_wwn, rp_wwn, 5238 BFA_RPORT_TAG_INVALID, BFA_LP_TAG_INVALID); 5239 } 5240 5241 /* 5242 * SGPG related functions 5243 */ 5244 5245 /* 5246 * Compute and return memory needed by FCP(im) module. 5247 */ 5248 static void 5249 bfa_sgpg_meminfo(struct bfa_iocfc_cfg_s *cfg, struct bfa_meminfo_s *minfo, 5250 struct bfa_s *bfa) 5251 { 5252 struct bfa_sgpg_mod_s *sgpg_mod = BFA_SGPG_MOD(bfa); 5253 struct bfa_mem_kva_s *sgpg_kva = BFA_MEM_SGPG_KVA(bfa); 5254 struct bfa_mem_dma_s *seg_ptr; 5255 u16 nsegs, idx, per_seg_sgpg, num_sgpg; 5256 u32 sgpg_sz = sizeof(struct bfi_sgpg_s); 5257 5258 if (cfg->drvcfg.num_sgpgs < BFA_SGPG_MIN) 5259 cfg->drvcfg.num_sgpgs = BFA_SGPG_MIN; 5260 else if (cfg->drvcfg.num_sgpgs > BFA_SGPG_MAX) 5261 cfg->drvcfg.num_sgpgs = BFA_SGPG_MAX; 5262 5263 num_sgpg = cfg->drvcfg.num_sgpgs; 5264 5265 nsegs = BFI_MEM_DMA_NSEGS(num_sgpg, sgpg_sz); 5266 per_seg_sgpg = BFI_MEM_NREQS_SEG(sgpg_sz); 5267 5268 bfa_mem_dma_seg_iter(sgpg_mod, seg_ptr, nsegs, idx) { 5269 if (num_sgpg >= per_seg_sgpg) { 5270 num_sgpg -= per_seg_sgpg; 5271 bfa_mem_dma_setup(minfo, seg_ptr, 5272 per_seg_sgpg * sgpg_sz); 5273 } else 5274 bfa_mem_dma_setup(minfo, seg_ptr, 5275 num_sgpg * sgpg_sz); 5276 } 5277 5278 /* kva memory */ 5279 bfa_mem_kva_setup(minfo, sgpg_kva, 5280 cfg->drvcfg.num_sgpgs * sizeof(struct bfa_sgpg_s)); 5281 } 5282 5283 static void 5284 bfa_sgpg_attach(struct bfa_s *bfa, void *bfad, struct bfa_iocfc_cfg_s *cfg, 5285 struct bfa_pcidev_s *pcidev) 5286 { 5287 struct bfa_sgpg_mod_s *mod = BFA_SGPG_MOD(bfa); 5288 struct bfa_sgpg_s *hsgpg; 5289 struct bfi_sgpg_s *sgpg; 5290 u64 align_len; 5291 struct bfa_mem_dma_s *seg_ptr; 5292 u32 sgpg_sz = sizeof(struct bfi_sgpg_s); 5293 u16 i, idx, nsegs, per_seg_sgpg, num_sgpg; 5294 5295 union { 5296 u64 pa; 5297 union bfi_addr_u addr; 5298 } sgpg_pa, sgpg_pa_tmp; 5299 5300 INIT_LIST_HEAD(&mod->sgpg_q); 5301 INIT_LIST_HEAD(&mod->sgpg_wait_q); 5302 5303 bfa_trc(bfa, cfg->drvcfg.num_sgpgs); 5304 5305 mod->free_sgpgs = mod->num_sgpgs = cfg->drvcfg.num_sgpgs; 5306 5307 num_sgpg = cfg->drvcfg.num_sgpgs; 5308 nsegs = BFI_MEM_DMA_NSEGS(num_sgpg, sgpg_sz); 5309 5310 /* dma/kva mem claim */ 5311 hsgpg = (struct bfa_sgpg_s *) bfa_mem_kva_curp(mod); 5312 5313 bfa_mem_dma_seg_iter(mod, seg_ptr, nsegs, idx) { 5314 5315 if (!bfa_mem_dma_virt(seg_ptr)) 5316 break; 5317 5318 align_len = BFA_SGPG_ROUNDUP(bfa_mem_dma_phys(seg_ptr)) - 5319 bfa_mem_dma_phys(seg_ptr); 5320 5321 sgpg = (struct bfi_sgpg_s *) 5322 (((u8 *) bfa_mem_dma_virt(seg_ptr)) + align_len); 5323 sgpg_pa.pa = bfa_mem_dma_phys(seg_ptr) + align_len; 5324 WARN_ON(sgpg_pa.pa & (sgpg_sz - 1)); 5325 5326 per_seg_sgpg = (seg_ptr->mem_len - (u32)align_len) / sgpg_sz; 5327 5328 for (i = 0; num_sgpg > 0 && i < per_seg_sgpg; i++, num_sgpg--) { 5329 memset(hsgpg, 0, sizeof(*hsgpg)); 5330 memset(sgpg, 0, sizeof(*sgpg)); 5331 5332 hsgpg->sgpg = sgpg; 5333 sgpg_pa_tmp.pa = bfa_sgaddr_le(sgpg_pa.pa); 5334 hsgpg->sgpg_pa = sgpg_pa_tmp.addr; 5335 list_add_tail(&hsgpg->qe, &mod->sgpg_q); 5336 5337 sgpg++; 5338 hsgpg++; 5339 sgpg_pa.pa += sgpg_sz; 5340 } 5341 } 5342 5343 bfa_mem_kva_curp(mod) = (u8 *) hsgpg; 5344 } 5345 5346 static void 5347 bfa_sgpg_detach(struct bfa_s *bfa) 5348 { 5349 } 5350 5351 static void 5352 bfa_sgpg_start(struct bfa_s *bfa) 5353 { 5354 } 5355 5356 static void 5357 bfa_sgpg_stop(struct bfa_s *bfa) 5358 { 5359 } 5360 5361 static void 5362 bfa_sgpg_iocdisable(struct bfa_s *bfa) 5363 { 5364 } 5365 5366 bfa_status_t 5367 bfa_sgpg_malloc(struct bfa_s *bfa, struct list_head *sgpg_q, int nsgpgs) 5368 { 5369 struct bfa_sgpg_mod_s *mod = BFA_SGPG_MOD(bfa); 5370 struct bfa_sgpg_s *hsgpg; 5371 int i; 5372 5373 if (mod->free_sgpgs < nsgpgs) 5374 return BFA_STATUS_ENOMEM; 5375 5376 for (i = 0; i < nsgpgs; i++) { 5377 bfa_q_deq(&mod->sgpg_q, &hsgpg); 5378 WARN_ON(!hsgpg); 5379 list_add_tail(&hsgpg->qe, sgpg_q); 5380 } 5381 5382 mod->free_sgpgs -= nsgpgs; 5383 return BFA_STATUS_OK; 5384 } 5385 5386 void 5387 bfa_sgpg_mfree(struct bfa_s *bfa, struct list_head *sgpg_q, int nsgpg) 5388 { 5389 struct bfa_sgpg_mod_s *mod = BFA_SGPG_MOD(bfa); 5390 struct bfa_sgpg_wqe_s *wqe; 5391 5392 mod->free_sgpgs += nsgpg; 5393 WARN_ON(mod->free_sgpgs > mod->num_sgpgs); 5394 5395 list_splice_tail_init(sgpg_q, &mod->sgpg_q); 5396 5397 if (list_empty(&mod->sgpg_wait_q)) 5398 return; 5399 5400 /* 5401 * satisfy as many waiting requests as possible 5402 */ 5403 do { 5404 wqe = bfa_q_first(&mod->sgpg_wait_q); 5405 if (mod->free_sgpgs < wqe->nsgpg) 5406 nsgpg = mod->free_sgpgs; 5407 else 5408 nsgpg = wqe->nsgpg; 5409 bfa_sgpg_malloc(bfa, &wqe->sgpg_q, nsgpg); 5410 wqe->nsgpg -= nsgpg; 5411 if (wqe->nsgpg == 0) { 5412 list_del(&wqe->qe); 5413 wqe->cbfn(wqe->cbarg); 5414 } 5415 } while (mod->free_sgpgs && !list_empty(&mod->sgpg_wait_q)); 5416 } 5417 5418 void 5419 bfa_sgpg_wait(struct bfa_s *bfa, struct bfa_sgpg_wqe_s *wqe, int nsgpg) 5420 { 5421 struct bfa_sgpg_mod_s *mod = BFA_SGPG_MOD(bfa); 5422 5423 WARN_ON(nsgpg <= 0); 5424 WARN_ON(nsgpg <= mod->free_sgpgs); 5425 5426 wqe->nsgpg_total = wqe->nsgpg = nsgpg; 5427 5428 /* 5429 * allocate any left to this one first 5430 */ 5431 if (mod->free_sgpgs) { 5432 /* 5433 * no one else is waiting for SGPG 5434 */ 5435 WARN_ON(!list_empty(&mod->sgpg_wait_q)); 5436 list_splice_tail_init(&mod->sgpg_q, &wqe->sgpg_q); 5437 wqe->nsgpg -= mod->free_sgpgs; 5438 mod->free_sgpgs = 0; 5439 } 5440 5441 list_add_tail(&wqe->qe, &mod->sgpg_wait_q); 5442 } 5443 5444 void 5445 bfa_sgpg_wcancel(struct bfa_s *bfa, struct bfa_sgpg_wqe_s *wqe) 5446 { 5447 struct bfa_sgpg_mod_s *mod = BFA_SGPG_MOD(bfa); 5448 5449 WARN_ON(!bfa_q_is_on_q(&mod->sgpg_wait_q, wqe)); 5450 list_del(&wqe->qe); 5451 5452 if (wqe->nsgpg_total != wqe->nsgpg) 5453 bfa_sgpg_mfree(bfa, &wqe->sgpg_q, 5454 wqe->nsgpg_total - wqe->nsgpg); 5455 } 5456 5457 void 5458 bfa_sgpg_winit(struct bfa_sgpg_wqe_s *wqe, void (*cbfn) (void *cbarg), 5459 void *cbarg) 5460 { 5461 INIT_LIST_HEAD(&wqe->sgpg_q); 5462 wqe->cbfn = cbfn; 5463 wqe->cbarg = cbarg; 5464 } 5465 5466 /* 5467 * UF related functions 5468 */ 5469 /* 5470 ***************************************************************************** 5471 * Internal functions 5472 ***************************************************************************** 5473 */ 5474 static void 5475 __bfa_cb_uf_recv(void *cbarg, bfa_boolean_t complete) 5476 { 5477 struct bfa_uf_s *uf = cbarg; 5478 struct bfa_uf_mod_s *ufm = BFA_UF_MOD(uf->bfa); 5479 5480 if (complete) 5481 ufm->ufrecv(ufm->cbarg, uf); 5482 } 5483 5484 static void 5485 claim_uf_post_msgs(struct bfa_uf_mod_s *ufm) 5486 { 5487 struct bfi_uf_buf_post_s *uf_bp_msg; 5488 u16 i; 5489 u16 buf_len; 5490 5491 ufm->uf_buf_posts = (struct bfi_uf_buf_post_s *) bfa_mem_kva_curp(ufm); 5492 uf_bp_msg = ufm->uf_buf_posts; 5493 5494 for (i = 0, uf_bp_msg = ufm->uf_buf_posts; i < ufm->num_ufs; 5495 i++, uf_bp_msg++) { 5496 memset(uf_bp_msg, 0, sizeof(struct bfi_uf_buf_post_s)); 5497 5498 uf_bp_msg->buf_tag = i; 5499 buf_len = sizeof(struct bfa_uf_buf_s); 5500 uf_bp_msg->buf_len = cpu_to_be16(buf_len); 5501 bfi_h2i_set(uf_bp_msg->mh, BFI_MC_UF, BFI_UF_H2I_BUF_POST, 5502 bfa_fn_lpu(ufm->bfa)); 5503 bfa_alen_set(&uf_bp_msg->alen, buf_len, ufm_pbs_pa(ufm, i)); 5504 } 5505 5506 /* 5507 * advance pointer beyond consumed memory 5508 */ 5509 bfa_mem_kva_curp(ufm) = (u8 *) uf_bp_msg; 5510 } 5511 5512 static void 5513 claim_ufs(struct bfa_uf_mod_s *ufm) 5514 { 5515 u16 i; 5516 struct bfa_uf_s *uf; 5517 5518 /* 5519 * Claim block of memory for UF list 5520 */ 5521 ufm->uf_list = (struct bfa_uf_s *) bfa_mem_kva_curp(ufm); 5522 5523 /* 5524 * Initialize UFs and queue it in UF free queue 5525 */ 5526 for (i = 0, uf = ufm->uf_list; i < ufm->num_ufs; i++, uf++) { 5527 memset(uf, 0, sizeof(struct bfa_uf_s)); 5528 uf->bfa = ufm->bfa; 5529 uf->uf_tag = i; 5530 uf->pb_len = BFA_PER_UF_DMA_SZ; 5531 uf->buf_kva = bfa_mem_get_dmabuf_kva(ufm, i, BFA_PER_UF_DMA_SZ); 5532 uf->buf_pa = ufm_pbs_pa(ufm, i); 5533 list_add_tail(&uf->qe, &ufm->uf_free_q); 5534 } 5535 5536 /* 5537 * advance memory pointer 5538 */ 5539 bfa_mem_kva_curp(ufm) = (u8 *) uf; 5540 } 5541 5542 static void 5543 uf_mem_claim(struct bfa_uf_mod_s *ufm) 5544 { 5545 claim_ufs(ufm); 5546 claim_uf_post_msgs(ufm); 5547 } 5548 5549 static void 5550 bfa_uf_meminfo(struct bfa_iocfc_cfg_s *cfg, struct bfa_meminfo_s *minfo, 5551 struct bfa_s *bfa) 5552 { 5553 struct bfa_uf_mod_s *ufm = BFA_UF_MOD(bfa); 5554 struct bfa_mem_kva_s *uf_kva = BFA_MEM_UF_KVA(bfa); 5555 u32 num_ufs = cfg->fwcfg.num_uf_bufs; 5556 struct bfa_mem_dma_s *seg_ptr; 5557 u16 nsegs, idx, per_seg_uf = 0; 5558 5559 nsegs = BFI_MEM_DMA_NSEGS(num_ufs, BFA_PER_UF_DMA_SZ); 5560 per_seg_uf = BFI_MEM_NREQS_SEG(BFA_PER_UF_DMA_SZ); 5561 5562 bfa_mem_dma_seg_iter(ufm, seg_ptr, nsegs, idx) { 5563 if (num_ufs >= per_seg_uf) { 5564 num_ufs -= per_seg_uf; 5565 bfa_mem_dma_setup(minfo, seg_ptr, 5566 per_seg_uf * BFA_PER_UF_DMA_SZ); 5567 } else 5568 bfa_mem_dma_setup(minfo, seg_ptr, 5569 num_ufs * BFA_PER_UF_DMA_SZ); 5570 } 5571 5572 /* kva memory */ 5573 bfa_mem_kva_setup(minfo, uf_kva, cfg->fwcfg.num_uf_bufs * 5574 (sizeof(struct bfa_uf_s) + sizeof(struct bfi_uf_buf_post_s))); 5575 } 5576 5577 static void 5578 bfa_uf_attach(struct bfa_s *bfa, void *bfad, struct bfa_iocfc_cfg_s *cfg, 5579 struct bfa_pcidev_s *pcidev) 5580 { 5581 struct bfa_uf_mod_s *ufm = BFA_UF_MOD(bfa); 5582 5583 ufm->bfa = bfa; 5584 ufm->num_ufs = cfg->fwcfg.num_uf_bufs; 5585 INIT_LIST_HEAD(&ufm->uf_free_q); 5586 INIT_LIST_HEAD(&ufm->uf_posted_q); 5587 INIT_LIST_HEAD(&ufm->uf_unused_q); 5588 5589 uf_mem_claim(ufm); 5590 } 5591 5592 static void 5593 bfa_uf_detach(struct bfa_s *bfa) 5594 { 5595 } 5596 5597 static struct bfa_uf_s * 5598 bfa_uf_get(struct bfa_uf_mod_s *uf_mod) 5599 { 5600 struct bfa_uf_s *uf; 5601 5602 bfa_q_deq(&uf_mod->uf_free_q, &uf); 5603 return uf; 5604 } 5605 5606 static void 5607 bfa_uf_put(struct bfa_uf_mod_s *uf_mod, struct bfa_uf_s *uf) 5608 { 5609 list_add_tail(&uf->qe, &uf_mod->uf_free_q); 5610 } 5611 5612 static bfa_status_t 5613 bfa_uf_post(struct bfa_uf_mod_s *ufm, struct bfa_uf_s *uf) 5614 { 5615 struct bfi_uf_buf_post_s *uf_post_msg; 5616 5617 uf_post_msg = bfa_reqq_next(ufm->bfa, BFA_REQQ_FCXP); 5618 if (!uf_post_msg) 5619 return BFA_STATUS_FAILED; 5620 5621 memcpy(uf_post_msg, &ufm->uf_buf_posts[uf->uf_tag], 5622 sizeof(struct bfi_uf_buf_post_s)); 5623 bfa_reqq_produce(ufm->bfa, BFA_REQQ_FCXP, uf_post_msg->mh); 5624 5625 bfa_trc(ufm->bfa, uf->uf_tag); 5626 5627 list_add_tail(&uf->qe, &ufm->uf_posted_q); 5628 return BFA_STATUS_OK; 5629 } 5630 5631 static void 5632 bfa_uf_post_all(struct bfa_uf_mod_s *uf_mod) 5633 { 5634 struct bfa_uf_s *uf; 5635 5636 while ((uf = bfa_uf_get(uf_mod)) != NULL) { 5637 if (bfa_uf_post(uf_mod, uf) != BFA_STATUS_OK) 5638 break; 5639 } 5640 } 5641 5642 static void 5643 uf_recv(struct bfa_s *bfa, struct bfi_uf_frm_rcvd_s *m) 5644 { 5645 struct bfa_uf_mod_s *ufm = BFA_UF_MOD(bfa); 5646 u16 uf_tag = m->buf_tag; 5647 struct bfa_uf_s *uf = &ufm->uf_list[uf_tag]; 5648 struct bfa_uf_buf_s *uf_buf; 5649 uint8_t *buf; 5650 struct fchs_s *fchs; 5651 5652 uf_buf = (struct bfa_uf_buf_s *) 5653 bfa_mem_get_dmabuf_kva(ufm, uf_tag, uf->pb_len); 5654 buf = &uf_buf->d[0]; 5655 5656 m->frm_len = be16_to_cpu(m->frm_len); 5657 m->xfr_len = be16_to_cpu(m->xfr_len); 5658 5659 fchs = (struct fchs_s *)uf_buf; 5660 5661 list_del(&uf->qe); /* dequeue from posted queue */ 5662 5663 uf->data_ptr = buf; 5664 uf->data_len = m->xfr_len; 5665 5666 WARN_ON(uf->data_len < sizeof(struct fchs_s)); 5667 5668 if (uf->data_len == sizeof(struct fchs_s)) { 5669 bfa_plog_fchdr(bfa->plog, BFA_PL_MID_HAL_UF, BFA_PL_EID_RX, 5670 uf->data_len, (struct fchs_s *)buf); 5671 } else { 5672 u32 pld_w0 = *((u32 *) (buf + sizeof(struct fchs_s))); 5673 bfa_plog_fchdr_and_pl(bfa->plog, BFA_PL_MID_HAL_UF, 5674 BFA_PL_EID_RX, uf->data_len, 5675 (struct fchs_s *)buf, pld_w0); 5676 } 5677 5678 if (bfa->fcs) 5679 __bfa_cb_uf_recv(uf, BFA_TRUE); 5680 else 5681 bfa_cb_queue(bfa, &uf->hcb_qe, __bfa_cb_uf_recv, uf); 5682 } 5683 5684 static void 5685 bfa_uf_stop(struct bfa_s *bfa) 5686 { 5687 } 5688 5689 static void 5690 bfa_uf_iocdisable(struct bfa_s *bfa) 5691 { 5692 struct bfa_uf_mod_s *ufm = BFA_UF_MOD(bfa); 5693 struct bfa_uf_s *uf; 5694 struct list_head *qe, *qen; 5695 5696 /* Enqueue unused uf resources to free_q */ 5697 list_splice_tail_init(&ufm->uf_unused_q, &ufm->uf_free_q); 5698 5699 list_for_each_safe(qe, qen, &ufm->uf_posted_q) { 5700 uf = (struct bfa_uf_s *) qe; 5701 list_del(&uf->qe); 5702 bfa_uf_put(ufm, uf); 5703 } 5704 } 5705 5706 static void 5707 bfa_uf_start(struct bfa_s *bfa) 5708 { 5709 bfa_uf_post_all(BFA_UF_MOD(bfa)); 5710 } 5711 5712 /* 5713 * Register handler for all unsolicted receive frames. 5714 * 5715 * @param[in] bfa BFA instance 5716 * @param[in] ufrecv receive handler function 5717 * @param[in] cbarg receive handler arg 5718 */ 5719 void 5720 bfa_uf_recv_register(struct bfa_s *bfa, bfa_cb_uf_recv_t ufrecv, void *cbarg) 5721 { 5722 struct bfa_uf_mod_s *ufm = BFA_UF_MOD(bfa); 5723 5724 ufm->ufrecv = ufrecv; 5725 ufm->cbarg = cbarg; 5726 } 5727 5728 /* 5729 * Free an unsolicited frame back to BFA. 5730 * 5731 * @param[in] uf unsolicited frame to be freed 5732 * 5733 * @return None 5734 */ 5735 void 5736 bfa_uf_free(struct bfa_uf_s *uf) 5737 { 5738 bfa_uf_put(BFA_UF_MOD(uf->bfa), uf); 5739 bfa_uf_post_all(BFA_UF_MOD(uf->bfa)); 5740 } 5741 5742 5743 5744 /* 5745 * uf_pub BFA uf module public functions 5746 */ 5747 void 5748 bfa_uf_isr(struct bfa_s *bfa, struct bfi_msg_s *msg) 5749 { 5750 bfa_trc(bfa, msg->mhdr.msg_id); 5751 5752 switch (msg->mhdr.msg_id) { 5753 case BFI_UF_I2H_FRM_RCVD: 5754 uf_recv(bfa, (struct bfi_uf_frm_rcvd_s *) msg); 5755 break; 5756 5757 default: 5758 bfa_trc(bfa, msg->mhdr.msg_id); 5759 WARN_ON(1); 5760 } 5761 } 5762 5763 void 5764 bfa_uf_res_recfg(struct bfa_s *bfa, u16 num_uf_fw) 5765 { 5766 struct bfa_uf_mod_s *mod = BFA_UF_MOD(bfa); 5767 struct list_head *qe; 5768 int i; 5769 5770 for (i = 0; i < (mod->num_ufs - num_uf_fw); i++) { 5771 bfa_q_deq_tail(&mod->uf_free_q, &qe); 5772 list_add_tail(qe, &mod->uf_unused_q); 5773 } 5774 } 5775 5776 /* 5777 * Dport forward declaration 5778 */ 5779 5780 enum bfa_dport_test_state_e { 5781 BFA_DPORT_ST_DISABLED = 0, /*!< dport is disabled */ 5782 BFA_DPORT_ST_INP = 1, /*!< test in progress */ 5783 BFA_DPORT_ST_COMP = 2, /*!< test complete successfully */ 5784 BFA_DPORT_ST_NO_SFP = 3, /*!< sfp is not present */ 5785 BFA_DPORT_ST_NOTSTART = 4, /*!< test not start dport is enabled */ 5786 }; 5787 5788 /* 5789 * BFA DPORT state machine events 5790 */ 5791 enum bfa_dport_sm_event { 5792 BFA_DPORT_SM_ENABLE = 1, /* dport enable event */ 5793 BFA_DPORT_SM_DISABLE = 2, /* dport disable event */ 5794 BFA_DPORT_SM_FWRSP = 3, /* fw enable/disable rsp */ 5795 BFA_DPORT_SM_QRESUME = 4, /* CQ space available */ 5796 BFA_DPORT_SM_HWFAIL = 5, /* IOC h/w failure */ 5797 BFA_DPORT_SM_START = 6, /* re-start dport test */ 5798 BFA_DPORT_SM_REQFAIL = 7, /* request failure */ 5799 BFA_DPORT_SM_SCN = 8, /* state change notify frm fw */ 5800 }; 5801 5802 static void bfa_dport_sm_disabled(struct bfa_dport_s *dport, 5803 enum bfa_dport_sm_event event); 5804 static void bfa_dport_sm_enabling_qwait(struct bfa_dport_s *dport, 5805 enum bfa_dport_sm_event event); 5806 static void bfa_dport_sm_enabling(struct bfa_dport_s *dport, 5807 enum bfa_dport_sm_event event); 5808 static void bfa_dport_sm_enabled(struct bfa_dport_s *dport, 5809 enum bfa_dport_sm_event event); 5810 static void bfa_dport_sm_disabling_qwait(struct bfa_dport_s *dport, 5811 enum bfa_dport_sm_event event); 5812 static void bfa_dport_sm_disabling(struct bfa_dport_s *dport, 5813 enum bfa_dport_sm_event event); 5814 static void bfa_dport_sm_starting_qwait(struct bfa_dport_s *dport, 5815 enum bfa_dport_sm_event event); 5816 static void bfa_dport_sm_starting(struct bfa_dport_s *dport, 5817 enum bfa_dport_sm_event event); 5818 static void bfa_dport_sm_dynamic_disabling(struct bfa_dport_s *dport, 5819 enum bfa_dport_sm_event event); 5820 static void bfa_dport_sm_dynamic_disabling_qwait(struct bfa_dport_s *dport, 5821 enum bfa_dport_sm_event event); 5822 static void bfa_dport_qresume(void *cbarg); 5823 static void bfa_dport_req_comp(struct bfa_dport_s *dport, 5824 struct bfi_diag_dport_rsp_s *msg); 5825 static void bfa_dport_scn(struct bfa_dport_s *dport, 5826 struct bfi_diag_dport_scn_s *msg); 5827 5828 /* 5829 * BFA fcdiag module 5830 */ 5831 #define BFA_DIAG_QTEST_TOV 1000 /* msec */ 5832 5833 /* 5834 * Set port status to busy 5835 */ 5836 static void 5837 bfa_fcdiag_set_busy_status(struct bfa_fcdiag_s *fcdiag) 5838 { 5839 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(fcdiag->bfa); 5840 5841 if (fcdiag->lb.lock) 5842 fcport->diag_busy = BFA_TRUE; 5843 else 5844 fcport->diag_busy = BFA_FALSE; 5845 } 5846 5847 static void 5848 bfa_fcdiag_meminfo(struct bfa_iocfc_cfg_s *cfg, struct bfa_meminfo_s *meminfo, 5849 struct bfa_s *bfa) 5850 { 5851 } 5852 5853 static void 5854 bfa_fcdiag_attach(struct bfa_s *bfa, void *bfad, struct bfa_iocfc_cfg_s *cfg, 5855 struct bfa_pcidev_s *pcidev) 5856 { 5857 struct bfa_fcdiag_s *fcdiag = BFA_FCDIAG_MOD(bfa); 5858 struct bfa_dport_s *dport = &fcdiag->dport; 5859 5860 fcdiag->bfa = bfa; 5861 fcdiag->trcmod = bfa->trcmod; 5862 /* The common DIAG attach bfa_diag_attach() will do all memory claim */ 5863 dport->bfa = bfa; 5864 bfa_sm_set_state(dport, bfa_dport_sm_disabled); 5865 bfa_reqq_winit(&dport->reqq_wait, bfa_dport_qresume, dport); 5866 dport->cbfn = NULL; 5867 dport->cbarg = NULL; 5868 dport->test_state = BFA_DPORT_ST_DISABLED; 5869 memset(&dport->result, 0, sizeof(struct bfa_diag_dport_result_s)); 5870 } 5871 5872 static void 5873 bfa_fcdiag_iocdisable(struct bfa_s *bfa) 5874 { 5875 struct bfa_fcdiag_s *fcdiag = BFA_FCDIAG_MOD(bfa); 5876 struct bfa_dport_s *dport = &fcdiag->dport; 5877 5878 bfa_trc(fcdiag, fcdiag->lb.lock); 5879 if (fcdiag->lb.lock) { 5880 fcdiag->lb.status = BFA_STATUS_IOC_FAILURE; 5881 fcdiag->lb.cbfn(fcdiag->lb.cbarg, fcdiag->lb.status); 5882 fcdiag->lb.lock = 0; 5883 bfa_fcdiag_set_busy_status(fcdiag); 5884 } 5885 5886 bfa_sm_send_event(dport, BFA_DPORT_SM_HWFAIL); 5887 } 5888 5889 static void 5890 bfa_fcdiag_detach(struct bfa_s *bfa) 5891 { 5892 } 5893 5894 static void 5895 bfa_fcdiag_start(struct bfa_s *bfa) 5896 { 5897 } 5898 5899 static void 5900 bfa_fcdiag_stop(struct bfa_s *bfa) 5901 { 5902 } 5903 5904 static void 5905 bfa_fcdiag_queuetest_timeout(void *cbarg) 5906 { 5907 struct bfa_fcdiag_s *fcdiag = cbarg; 5908 struct bfa_diag_qtest_result_s *res = fcdiag->qtest.result; 5909 5910 bfa_trc(fcdiag, fcdiag->qtest.all); 5911 bfa_trc(fcdiag, fcdiag->qtest.count); 5912 5913 fcdiag->qtest.timer_active = 0; 5914 5915 res->status = BFA_STATUS_ETIMER; 5916 res->count = QTEST_CNT_DEFAULT - fcdiag->qtest.count; 5917 if (fcdiag->qtest.all) 5918 res->queue = fcdiag->qtest.all; 5919 5920 bfa_trc(fcdiag, BFA_STATUS_ETIMER); 5921 fcdiag->qtest.status = BFA_STATUS_ETIMER; 5922 fcdiag->qtest.cbfn(fcdiag->qtest.cbarg, fcdiag->qtest.status); 5923 fcdiag->qtest.lock = 0; 5924 } 5925 5926 static bfa_status_t 5927 bfa_fcdiag_queuetest_send(struct bfa_fcdiag_s *fcdiag) 5928 { 5929 u32 i; 5930 struct bfi_diag_qtest_req_s *req; 5931 5932 req = bfa_reqq_next(fcdiag->bfa, fcdiag->qtest.queue); 5933 if (!req) 5934 return BFA_STATUS_DEVBUSY; 5935 5936 /* build host command */ 5937 bfi_h2i_set(req->mh, BFI_MC_DIAG, BFI_DIAG_H2I_QTEST, 5938 bfa_fn_lpu(fcdiag->bfa)); 5939 5940 for (i = 0; i < BFI_LMSG_PL_WSZ; i++) 5941 req->data[i] = QTEST_PAT_DEFAULT; 5942 5943 bfa_trc(fcdiag, fcdiag->qtest.queue); 5944 /* ring door bell */ 5945 bfa_reqq_produce(fcdiag->bfa, fcdiag->qtest.queue, req->mh); 5946 return BFA_STATUS_OK; 5947 } 5948 5949 static void 5950 bfa_fcdiag_queuetest_comp(struct bfa_fcdiag_s *fcdiag, 5951 bfi_diag_qtest_rsp_t *rsp) 5952 { 5953 struct bfa_diag_qtest_result_s *res = fcdiag->qtest.result; 5954 bfa_status_t status = BFA_STATUS_OK; 5955 int i; 5956 5957 /* Check timer, should still be active */ 5958 if (!fcdiag->qtest.timer_active) { 5959 bfa_trc(fcdiag, fcdiag->qtest.timer_active); 5960 return; 5961 } 5962 5963 /* update count */ 5964 fcdiag->qtest.count--; 5965 5966 /* Check result */ 5967 for (i = 0; i < BFI_LMSG_PL_WSZ; i++) { 5968 if (rsp->data[i] != ~(QTEST_PAT_DEFAULT)) { 5969 res->status = BFA_STATUS_DATACORRUPTED; 5970 break; 5971 } 5972 } 5973 5974 if (res->status == BFA_STATUS_OK) { 5975 if (fcdiag->qtest.count > 0) { 5976 status = bfa_fcdiag_queuetest_send(fcdiag); 5977 if (status == BFA_STATUS_OK) 5978 return; 5979 else 5980 res->status = status; 5981 } else if (fcdiag->qtest.all > 0 && 5982 fcdiag->qtest.queue < (BFI_IOC_MAX_CQS - 1)) { 5983 fcdiag->qtest.count = QTEST_CNT_DEFAULT; 5984 fcdiag->qtest.queue++; 5985 status = bfa_fcdiag_queuetest_send(fcdiag); 5986 if (status == BFA_STATUS_OK) 5987 return; 5988 else 5989 res->status = status; 5990 } 5991 } 5992 5993 /* Stop timer when we comp all queue */ 5994 if (fcdiag->qtest.timer_active) { 5995 bfa_timer_stop(&fcdiag->qtest.timer); 5996 fcdiag->qtest.timer_active = 0; 5997 } 5998 res->queue = fcdiag->qtest.queue; 5999 res->count = QTEST_CNT_DEFAULT - fcdiag->qtest.count; 6000 bfa_trc(fcdiag, res->count); 6001 bfa_trc(fcdiag, res->status); 6002 fcdiag->qtest.status = res->status; 6003 fcdiag->qtest.cbfn(fcdiag->qtest.cbarg, fcdiag->qtest.status); 6004 fcdiag->qtest.lock = 0; 6005 } 6006 6007 static void 6008 bfa_fcdiag_loopback_comp(struct bfa_fcdiag_s *fcdiag, 6009 struct bfi_diag_lb_rsp_s *rsp) 6010 { 6011 struct bfa_diag_loopback_result_s *res = fcdiag->lb.result; 6012 6013 res->numtxmfrm = be32_to_cpu(rsp->res.numtxmfrm); 6014 res->numosffrm = be32_to_cpu(rsp->res.numosffrm); 6015 res->numrcvfrm = be32_to_cpu(rsp->res.numrcvfrm); 6016 res->badfrminf = be32_to_cpu(rsp->res.badfrminf); 6017 res->badfrmnum = be32_to_cpu(rsp->res.badfrmnum); 6018 res->status = rsp->res.status; 6019 fcdiag->lb.status = rsp->res.status; 6020 bfa_trc(fcdiag, fcdiag->lb.status); 6021 fcdiag->lb.cbfn(fcdiag->lb.cbarg, fcdiag->lb.status); 6022 fcdiag->lb.lock = 0; 6023 bfa_fcdiag_set_busy_status(fcdiag); 6024 } 6025 6026 static bfa_status_t 6027 bfa_fcdiag_loopback_send(struct bfa_fcdiag_s *fcdiag, 6028 struct bfa_diag_loopback_s *loopback) 6029 { 6030 struct bfi_diag_lb_req_s *lb_req; 6031 6032 lb_req = bfa_reqq_next(fcdiag->bfa, BFA_REQQ_DIAG); 6033 if (!lb_req) 6034 return BFA_STATUS_DEVBUSY; 6035 6036 /* build host command */ 6037 bfi_h2i_set(lb_req->mh, BFI_MC_DIAG, BFI_DIAG_H2I_LOOPBACK, 6038 bfa_fn_lpu(fcdiag->bfa)); 6039 6040 lb_req->lb_mode = loopback->lb_mode; 6041 lb_req->speed = loopback->speed; 6042 lb_req->loopcnt = loopback->loopcnt; 6043 lb_req->pattern = loopback->pattern; 6044 6045 /* ring door bell */ 6046 bfa_reqq_produce(fcdiag->bfa, BFA_REQQ_DIAG, lb_req->mh); 6047 6048 bfa_trc(fcdiag, loopback->lb_mode); 6049 bfa_trc(fcdiag, loopback->speed); 6050 bfa_trc(fcdiag, loopback->loopcnt); 6051 bfa_trc(fcdiag, loopback->pattern); 6052 return BFA_STATUS_OK; 6053 } 6054 6055 /* 6056 * cpe/rme intr handler 6057 */ 6058 void 6059 bfa_fcdiag_intr(struct bfa_s *bfa, struct bfi_msg_s *msg) 6060 { 6061 struct bfa_fcdiag_s *fcdiag = BFA_FCDIAG_MOD(bfa); 6062 6063 switch (msg->mhdr.msg_id) { 6064 case BFI_DIAG_I2H_LOOPBACK: 6065 bfa_fcdiag_loopback_comp(fcdiag, 6066 (struct bfi_diag_lb_rsp_s *) msg); 6067 break; 6068 case BFI_DIAG_I2H_QTEST: 6069 bfa_fcdiag_queuetest_comp(fcdiag, (bfi_diag_qtest_rsp_t *)msg); 6070 break; 6071 case BFI_DIAG_I2H_DPORT: 6072 bfa_dport_req_comp(&fcdiag->dport, 6073 (struct bfi_diag_dport_rsp_s *)msg); 6074 break; 6075 case BFI_DIAG_I2H_DPORT_SCN: 6076 bfa_dport_scn(&fcdiag->dport, 6077 (struct bfi_diag_dport_scn_s *)msg); 6078 break; 6079 default: 6080 bfa_trc(fcdiag, msg->mhdr.msg_id); 6081 WARN_ON(1); 6082 } 6083 } 6084 6085 /* 6086 * Loopback test 6087 * 6088 * @param[in] *bfa - bfa data struct 6089 * @param[in] opmode - port operation mode 6090 * @param[in] speed - port speed 6091 * @param[in] lpcnt - loop count 6092 * @param[in] pat - pattern to build packet 6093 * @param[in] *result - pt to bfa_diag_loopback_result_t data struct 6094 * @param[in] cbfn - callback function 6095 * @param[in] cbarg - callback functioin arg 6096 * 6097 * @param[out] 6098 */ 6099 bfa_status_t 6100 bfa_fcdiag_loopback(struct bfa_s *bfa, enum bfa_port_opmode opmode, 6101 enum bfa_port_speed speed, u32 lpcnt, u32 pat, 6102 struct bfa_diag_loopback_result_s *result, bfa_cb_diag_t cbfn, 6103 void *cbarg) 6104 { 6105 struct bfa_diag_loopback_s loopback; 6106 struct bfa_port_attr_s attr; 6107 bfa_status_t status; 6108 struct bfa_fcdiag_s *fcdiag = BFA_FCDIAG_MOD(bfa); 6109 6110 if (!bfa_iocfc_is_operational(bfa)) 6111 return BFA_STATUS_IOC_NON_OP; 6112 6113 /* if port is PBC disabled, return error */ 6114 if (bfa_fcport_is_pbcdisabled(bfa)) { 6115 bfa_trc(fcdiag, BFA_STATUS_PBC); 6116 return BFA_STATUS_PBC; 6117 } 6118 6119 if (bfa_fcport_is_disabled(bfa) == BFA_FALSE) { 6120 bfa_trc(fcdiag, opmode); 6121 return BFA_STATUS_PORT_NOT_DISABLED; 6122 } 6123 6124 /* 6125 * Check if input speed is supported by the port mode 6126 */ 6127 if (bfa_ioc_get_type(&bfa->ioc) == BFA_IOC_TYPE_FC) { 6128 if (!(speed == BFA_PORT_SPEED_1GBPS || 6129 speed == BFA_PORT_SPEED_2GBPS || 6130 speed == BFA_PORT_SPEED_4GBPS || 6131 speed == BFA_PORT_SPEED_8GBPS || 6132 speed == BFA_PORT_SPEED_16GBPS || 6133 speed == BFA_PORT_SPEED_AUTO)) { 6134 bfa_trc(fcdiag, speed); 6135 return BFA_STATUS_UNSUPP_SPEED; 6136 } 6137 bfa_fcport_get_attr(bfa, &attr); 6138 bfa_trc(fcdiag, attr.speed_supported); 6139 if (speed > attr.speed_supported) 6140 return BFA_STATUS_UNSUPP_SPEED; 6141 } else { 6142 if (speed != BFA_PORT_SPEED_10GBPS) { 6143 bfa_trc(fcdiag, speed); 6144 return BFA_STATUS_UNSUPP_SPEED; 6145 } 6146 } 6147 6148 /* 6149 * For CT2, 1G is not supported 6150 */ 6151 if ((speed == BFA_PORT_SPEED_1GBPS) && 6152 (bfa_asic_id_ct2(bfa->ioc.pcidev.device_id))) { 6153 bfa_trc(fcdiag, speed); 6154 return BFA_STATUS_UNSUPP_SPEED; 6155 } 6156 6157 /* For Mezz card, port speed entered needs to be checked */ 6158 if (bfa_mfg_is_mezz(bfa->ioc.attr->card_type)) { 6159 if (bfa_ioc_get_type(&bfa->ioc) == BFA_IOC_TYPE_FC) { 6160 if (!(speed == BFA_PORT_SPEED_1GBPS || 6161 speed == BFA_PORT_SPEED_2GBPS || 6162 speed == BFA_PORT_SPEED_4GBPS || 6163 speed == BFA_PORT_SPEED_8GBPS || 6164 speed == BFA_PORT_SPEED_16GBPS || 6165 speed == BFA_PORT_SPEED_AUTO)) 6166 return BFA_STATUS_UNSUPP_SPEED; 6167 } else { 6168 if (speed != BFA_PORT_SPEED_10GBPS) 6169 return BFA_STATUS_UNSUPP_SPEED; 6170 } 6171 } 6172 /* check to see if fcport is dport */ 6173 if (bfa_fcport_is_dport(bfa)) { 6174 bfa_trc(fcdiag, fcdiag->lb.lock); 6175 return BFA_STATUS_DPORT_ENABLED; 6176 } 6177 /* check to see if there is another destructive diag cmd running */ 6178 if (fcdiag->lb.lock) { 6179 bfa_trc(fcdiag, fcdiag->lb.lock); 6180 return BFA_STATUS_DEVBUSY; 6181 } 6182 6183 fcdiag->lb.lock = 1; 6184 loopback.lb_mode = opmode; 6185 loopback.speed = speed; 6186 loopback.loopcnt = lpcnt; 6187 loopback.pattern = pat; 6188 fcdiag->lb.result = result; 6189 fcdiag->lb.cbfn = cbfn; 6190 fcdiag->lb.cbarg = cbarg; 6191 memset(result, 0, sizeof(struct bfa_diag_loopback_result_s)); 6192 bfa_fcdiag_set_busy_status(fcdiag); 6193 6194 /* Send msg to fw */ 6195 status = bfa_fcdiag_loopback_send(fcdiag, &loopback); 6196 return status; 6197 } 6198 6199 /* 6200 * DIAG queue test command 6201 * 6202 * @param[in] *bfa - bfa data struct 6203 * @param[in] force - 1: don't do ioc op checking 6204 * @param[in] queue - queue no. to test 6205 * @param[in] *result - pt to bfa_diag_qtest_result_t data struct 6206 * @param[in] cbfn - callback function 6207 * @param[in] *cbarg - callback functioin arg 6208 * 6209 * @param[out] 6210 */ 6211 bfa_status_t 6212 bfa_fcdiag_queuetest(struct bfa_s *bfa, u32 force, u32 queue, 6213 struct bfa_diag_qtest_result_s *result, bfa_cb_diag_t cbfn, 6214 void *cbarg) 6215 { 6216 struct bfa_fcdiag_s *fcdiag = BFA_FCDIAG_MOD(bfa); 6217 bfa_status_t status; 6218 bfa_trc(fcdiag, force); 6219 bfa_trc(fcdiag, queue); 6220 6221 if (!force && !bfa_iocfc_is_operational(bfa)) 6222 return BFA_STATUS_IOC_NON_OP; 6223 6224 /* check to see if there is another destructive diag cmd running */ 6225 if (fcdiag->qtest.lock) { 6226 bfa_trc(fcdiag, fcdiag->qtest.lock); 6227 return BFA_STATUS_DEVBUSY; 6228 } 6229 6230 /* Initialization */ 6231 fcdiag->qtest.lock = 1; 6232 fcdiag->qtest.cbfn = cbfn; 6233 fcdiag->qtest.cbarg = cbarg; 6234 fcdiag->qtest.result = result; 6235 fcdiag->qtest.count = QTEST_CNT_DEFAULT; 6236 6237 /* Init test results */ 6238 fcdiag->qtest.result->status = BFA_STATUS_OK; 6239 fcdiag->qtest.result->count = 0; 6240 6241 /* send */ 6242 if (queue < BFI_IOC_MAX_CQS) { 6243 fcdiag->qtest.result->queue = (u8)queue; 6244 fcdiag->qtest.queue = (u8)queue; 6245 fcdiag->qtest.all = 0; 6246 } else { 6247 fcdiag->qtest.result->queue = 0; 6248 fcdiag->qtest.queue = 0; 6249 fcdiag->qtest.all = 1; 6250 } 6251 status = bfa_fcdiag_queuetest_send(fcdiag); 6252 6253 /* Start a timer */ 6254 if (status == BFA_STATUS_OK) { 6255 bfa_timer_start(bfa, &fcdiag->qtest.timer, 6256 bfa_fcdiag_queuetest_timeout, fcdiag, 6257 BFA_DIAG_QTEST_TOV); 6258 fcdiag->qtest.timer_active = 1; 6259 } 6260 return status; 6261 } 6262 6263 /* 6264 * DIAG PLB is running 6265 * 6266 * @param[in] *bfa - bfa data struct 6267 * 6268 * @param[out] 6269 */ 6270 bfa_status_t 6271 bfa_fcdiag_lb_is_running(struct bfa_s *bfa) 6272 { 6273 struct bfa_fcdiag_s *fcdiag = BFA_FCDIAG_MOD(bfa); 6274 return fcdiag->lb.lock ? BFA_STATUS_DIAG_BUSY : BFA_STATUS_OK; 6275 } 6276 6277 /* 6278 * D-port 6279 */ 6280 #define bfa_dport_result_start(__dport, __mode) do { \ 6281 (__dport)->result.start_time = bfa_get_log_time(); \ 6282 (__dport)->result.status = DPORT_TEST_ST_INPRG; \ 6283 (__dport)->result.mode = (__mode); \ 6284 (__dport)->result.rp_pwwn = (__dport)->rp_pwwn; \ 6285 (__dport)->result.rp_nwwn = (__dport)->rp_nwwn; \ 6286 (__dport)->result.lpcnt = (__dport)->lpcnt; \ 6287 } while (0) 6288 6289 static bfa_boolean_t bfa_dport_send_req(struct bfa_dport_s *dport, 6290 enum bfi_dport_req req); 6291 static void 6292 bfa_cb_fcdiag_dport(struct bfa_dport_s *dport, bfa_status_t bfa_status) 6293 { 6294 if (dport->cbfn != NULL) { 6295 dport->cbfn(dport->cbarg, bfa_status); 6296 dport->cbfn = NULL; 6297 dport->cbarg = NULL; 6298 } 6299 } 6300 6301 static void 6302 bfa_dport_sm_disabled(struct bfa_dport_s *dport, enum bfa_dport_sm_event event) 6303 { 6304 bfa_trc(dport->bfa, event); 6305 6306 switch (event) { 6307 case BFA_DPORT_SM_ENABLE: 6308 bfa_fcport_dportenable(dport->bfa); 6309 if (bfa_dport_send_req(dport, BFI_DPORT_ENABLE)) 6310 bfa_sm_set_state(dport, bfa_dport_sm_enabling); 6311 else 6312 bfa_sm_set_state(dport, bfa_dport_sm_enabling_qwait); 6313 break; 6314 6315 case BFA_DPORT_SM_DISABLE: 6316 /* Already disabled */ 6317 break; 6318 6319 case BFA_DPORT_SM_HWFAIL: 6320 /* ignore */ 6321 break; 6322 6323 case BFA_DPORT_SM_SCN: 6324 if (dport->i2hmsg.scn.state == BFI_DPORT_SCN_DDPORT_ENABLE) { 6325 bfa_fcport_ddportenable(dport->bfa); 6326 dport->dynamic = BFA_TRUE; 6327 dport->test_state = BFA_DPORT_ST_NOTSTART; 6328 bfa_sm_set_state(dport, bfa_dport_sm_enabled); 6329 } else { 6330 bfa_trc(dport->bfa, dport->i2hmsg.scn.state); 6331 WARN_ON(1); 6332 } 6333 break; 6334 6335 default: 6336 bfa_sm_fault(dport->bfa, event); 6337 } 6338 } 6339 6340 static void 6341 bfa_dport_sm_enabling_qwait(struct bfa_dport_s *dport, 6342 enum bfa_dport_sm_event event) 6343 { 6344 bfa_trc(dport->bfa, event); 6345 6346 switch (event) { 6347 case BFA_DPORT_SM_QRESUME: 6348 bfa_sm_set_state(dport, bfa_dport_sm_enabling); 6349 bfa_dport_send_req(dport, BFI_DPORT_ENABLE); 6350 break; 6351 6352 case BFA_DPORT_SM_HWFAIL: 6353 bfa_reqq_wcancel(&dport->reqq_wait); 6354 bfa_sm_set_state(dport, bfa_dport_sm_disabled); 6355 bfa_cb_fcdiag_dport(dport, BFA_STATUS_FAILED); 6356 break; 6357 6358 default: 6359 bfa_sm_fault(dport->bfa, event); 6360 } 6361 } 6362 6363 static void 6364 bfa_dport_sm_enabling(struct bfa_dport_s *dport, enum bfa_dport_sm_event event) 6365 { 6366 bfa_trc(dport->bfa, event); 6367 6368 switch (event) { 6369 case BFA_DPORT_SM_FWRSP: 6370 memset(&dport->result, 0, 6371 sizeof(struct bfa_diag_dport_result_s)); 6372 if (dport->i2hmsg.rsp.status == BFA_STATUS_DPORT_INV_SFP) { 6373 dport->test_state = BFA_DPORT_ST_NO_SFP; 6374 } else { 6375 dport->test_state = BFA_DPORT_ST_INP; 6376 bfa_dport_result_start(dport, BFA_DPORT_OPMODE_AUTO); 6377 } 6378 bfa_sm_set_state(dport, bfa_dport_sm_enabled); 6379 break; 6380 6381 case BFA_DPORT_SM_REQFAIL: 6382 dport->test_state = BFA_DPORT_ST_DISABLED; 6383 bfa_fcport_dportdisable(dport->bfa); 6384 bfa_sm_set_state(dport, bfa_dport_sm_disabled); 6385 break; 6386 6387 case BFA_DPORT_SM_HWFAIL: 6388 bfa_sm_set_state(dport, bfa_dport_sm_disabled); 6389 bfa_cb_fcdiag_dport(dport, BFA_STATUS_FAILED); 6390 break; 6391 6392 default: 6393 bfa_sm_fault(dport->bfa, event); 6394 } 6395 } 6396 6397 static void 6398 bfa_dport_sm_enabled(struct bfa_dport_s *dport, enum bfa_dport_sm_event event) 6399 { 6400 bfa_trc(dport->bfa, event); 6401 6402 switch (event) { 6403 case BFA_DPORT_SM_START: 6404 if (bfa_dport_send_req(dport, BFI_DPORT_START)) 6405 bfa_sm_set_state(dport, bfa_dport_sm_starting); 6406 else 6407 bfa_sm_set_state(dport, bfa_dport_sm_starting_qwait); 6408 break; 6409 6410 case BFA_DPORT_SM_DISABLE: 6411 bfa_fcport_dportdisable(dport->bfa); 6412 if (bfa_dport_send_req(dport, BFI_DPORT_DISABLE)) 6413 bfa_sm_set_state(dport, bfa_dport_sm_disabling); 6414 else 6415 bfa_sm_set_state(dport, bfa_dport_sm_disabling_qwait); 6416 break; 6417 6418 case BFA_DPORT_SM_HWFAIL: 6419 bfa_sm_set_state(dport, bfa_dport_sm_disabled); 6420 break; 6421 6422 case BFA_DPORT_SM_SCN: 6423 switch (dport->i2hmsg.scn.state) { 6424 case BFI_DPORT_SCN_TESTCOMP: 6425 dport->test_state = BFA_DPORT_ST_COMP; 6426 break; 6427 6428 case BFI_DPORT_SCN_TESTSTART: 6429 dport->test_state = BFA_DPORT_ST_INP; 6430 break; 6431 6432 case BFI_DPORT_SCN_TESTSKIP: 6433 case BFI_DPORT_SCN_SUBTESTSTART: 6434 /* no state change */ 6435 break; 6436 6437 case BFI_DPORT_SCN_SFP_REMOVED: 6438 dport->test_state = BFA_DPORT_ST_NO_SFP; 6439 break; 6440 6441 case BFI_DPORT_SCN_DDPORT_DISABLE: 6442 bfa_fcport_ddportdisable(dport->bfa); 6443 6444 if (bfa_dport_send_req(dport, BFI_DPORT_DYN_DISABLE)) 6445 bfa_sm_set_state(dport, 6446 bfa_dport_sm_dynamic_disabling); 6447 else 6448 bfa_sm_set_state(dport, 6449 bfa_dport_sm_dynamic_disabling_qwait); 6450 break; 6451 6452 case BFI_DPORT_SCN_FCPORT_DISABLE: 6453 bfa_fcport_ddportdisable(dport->bfa); 6454 6455 bfa_sm_set_state(dport, bfa_dport_sm_disabled); 6456 dport->dynamic = BFA_FALSE; 6457 break; 6458 6459 default: 6460 bfa_trc(dport->bfa, dport->i2hmsg.scn.state); 6461 bfa_sm_fault(dport->bfa, event); 6462 } 6463 break; 6464 default: 6465 bfa_sm_fault(dport->bfa, event); 6466 } 6467 } 6468 6469 static void 6470 bfa_dport_sm_disabling_qwait(struct bfa_dport_s *dport, 6471 enum bfa_dport_sm_event event) 6472 { 6473 bfa_trc(dport->bfa, event); 6474 6475 switch (event) { 6476 case BFA_DPORT_SM_QRESUME: 6477 bfa_sm_set_state(dport, bfa_dport_sm_disabling); 6478 bfa_dport_send_req(dport, BFI_DPORT_DISABLE); 6479 break; 6480 6481 case BFA_DPORT_SM_HWFAIL: 6482 bfa_sm_set_state(dport, bfa_dport_sm_disabled); 6483 bfa_reqq_wcancel(&dport->reqq_wait); 6484 bfa_cb_fcdiag_dport(dport, BFA_STATUS_OK); 6485 break; 6486 6487 case BFA_DPORT_SM_SCN: 6488 /* ignore */ 6489 break; 6490 6491 default: 6492 bfa_sm_fault(dport->bfa, event); 6493 } 6494 } 6495 6496 static void 6497 bfa_dport_sm_disabling(struct bfa_dport_s *dport, enum bfa_dport_sm_event event) 6498 { 6499 bfa_trc(dport->bfa, event); 6500 6501 switch (event) { 6502 case BFA_DPORT_SM_FWRSP: 6503 dport->test_state = BFA_DPORT_ST_DISABLED; 6504 bfa_sm_set_state(dport, bfa_dport_sm_disabled); 6505 break; 6506 6507 case BFA_DPORT_SM_HWFAIL: 6508 bfa_sm_set_state(dport, bfa_dport_sm_disabled); 6509 bfa_cb_fcdiag_dport(dport, BFA_STATUS_OK); 6510 break; 6511 6512 case BFA_DPORT_SM_SCN: 6513 /* no state change */ 6514 break; 6515 6516 default: 6517 bfa_sm_fault(dport->bfa, event); 6518 } 6519 } 6520 6521 static void 6522 bfa_dport_sm_starting_qwait(struct bfa_dport_s *dport, 6523 enum bfa_dport_sm_event event) 6524 { 6525 bfa_trc(dport->bfa, event); 6526 6527 switch (event) { 6528 case BFA_DPORT_SM_QRESUME: 6529 bfa_sm_set_state(dport, bfa_dport_sm_starting); 6530 bfa_dport_send_req(dport, BFI_DPORT_START); 6531 break; 6532 6533 case BFA_DPORT_SM_HWFAIL: 6534 bfa_reqq_wcancel(&dport->reqq_wait); 6535 bfa_sm_set_state(dport, bfa_dport_sm_disabled); 6536 bfa_cb_fcdiag_dport(dport, BFA_STATUS_FAILED); 6537 break; 6538 6539 default: 6540 bfa_sm_fault(dport->bfa, event); 6541 } 6542 } 6543 6544 static void 6545 bfa_dport_sm_starting(struct bfa_dport_s *dport, enum bfa_dport_sm_event event) 6546 { 6547 bfa_trc(dport->bfa, event); 6548 6549 switch (event) { 6550 case BFA_DPORT_SM_FWRSP: 6551 memset(&dport->result, 0, 6552 sizeof(struct bfa_diag_dport_result_s)); 6553 if (dport->i2hmsg.rsp.status == BFA_STATUS_DPORT_INV_SFP) { 6554 dport->test_state = BFA_DPORT_ST_NO_SFP; 6555 } else { 6556 dport->test_state = BFA_DPORT_ST_INP; 6557 bfa_dport_result_start(dport, BFA_DPORT_OPMODE_MANU); 6558 } 6559 /* fall thru */ 6560 6561 case BFA_DPORT_SM_REQFAIL: 6562 bfa_sm_set_state(dport, bfa_dport_sm_enabled); 6563 break; 6564 6565 case BFA_DPORT_SM_HWFAIL: 6566 bfa_sm_set_state(dport, bfa_dport_sm_disabled); 6567 bfa_cb_fcdiag_dport(dport, BFA_STATUS_FAILED); 6568 break; 6569 6570 default: 6571 bfa_sm_fault(dport->bfa, event); 6572 } 6573 } 6574 6575 static void 6576 bfa_dport_sm_dynamic_disabling(struct bfa_dport_s *dport, 6577 enum bfa_dport_sm_event event) 6578 { 6579 bfa_trc(dport->bfa, event); 6580 6581 switch (event) { 6582 case BFA_DPORT_SM_SCN: 6583 switch (dport->i2hmsg.scn.state) { 6584 case BFI_DPORT_SCN_DDPORT_DISABLED: 6585 bfa_sm_set_state(dport, bfa_dport_sm_disabled); 6586 dport->dynamic = BFA_FALSE; 6587 bfa_fcport_enable(dport->bfa); 6588 break; 6589 6590 default: 6591 bfa_trc(dport->bfa, dport->i2hmsg.scn.state); 6592 bfa_sm_fault(dport->bfa, event); 6593 6594 } 6595 break; 6596 6597 case BFA_DPORT_SM_HWFAIL: 6598 bfa_sm_set_state(dport, bfa_dport_sm_disabled); 6599 bfa_cb_fcdiag_dport(dport, BFA_STATUS_OK); 6600 break; 6601 6602 default: 6603 bfa_sm_fault(dport->bfa, event); 6604 } 6605 } 6606 6607 static void 6608 bfa_dport_sm_dynamic_disabling_qwait(struct bfa_dport_s *dport, 6609 enum bfa_dport_sm_event event) 6610 { 6611 bfa_trc(dport->bfa, event); 6612 6613 switch (event) { 6614 case BFA_DPORT_SM_QRESUME: 6615 bfa_sm_set_state(dport, bfa_dport_sm_dynamic_disabling); 6616 bfa_dport_send_req(dport, BFI_DPORT_DYN_DISABLE); 6617 break; 6618 6619 case BFA_DPORT_SM_HWFAIL: 6620 bfa_sm_set_state(dport, bfa_dport_sm_disabled); 6621 bfa_reqq_wcancel(&dport->reqq_wait); 6622 bfa_cb_fcdiag_dport(dport, BFA_STATUS_OK); 6623 break; 6624 6625 case BFA_DPORT_SM_SCN: 6626 /* ignore */ 6627 break; 6628 6629 default: 6630 bfa_sm_fault(dport->bfa, event); 6631 } 6632 } 6633 6634 static bfa_boolean_t 6635 bfa_dport_send_req(struct bfa_dport_s *dport, enum bfi_dport_req req) 6636 { 6637 struct bfi_diag_dport_req_s *m; 6638 6639 /* 6640 * check for room in queue to send request now 6641 */ 6642 m = bfa_reqq_next(dport->bfa, BFA_REQQ_DIAG); 6643 if (!m) { 6644 bfa_reqq_wait(dport->bfa, BFA_REQQ_PORT, &dport->reqq_wait); 6645 return BFA_FALSE; 6646 } 6647 6648 bfi_h2i_set(m->mh, BFI_MC_DIAG, BFI_DIAG_H2I_DPORT, 6649 bfa_fn_lpu(dport->bfa)); 6650 m->req = req; 6651 if ((req == BFI_DPORT_ENABLE) || (req == BFI_DPORT_START)) { 6652 m->lpcnt = cpu_to_be32(dport->lpcnt); 6653 m->payload = cpu_to_be32(dport->payload); 6654 } 6655 6656 /* 6657 * queue I/O message to firmware 6658 */ 6659 bfa_reqq_produce(dport->bfa, BFA_REQQ_DIAG, m->mh); 6660 6661 return BFA_TRUE; 6662 } 6663 6664 static void 6665 bfa_dport_qresume(void *cbarg) 6666 { 6667 struct bfa_dport_s *dport = cbarg; 6668 6669 bfa_sm_send_event(dport, BFA_DPORT_SM_QRESUME); 6670 } 6671 6672 static void 6673 bfa_dport_req_comp(struct bfa_dport_s *dport, struct bfi_diag_dport_rsp_s *msg) 6674 { 6675 msg->status = cpu_to_be32(msg->status); 6676 dport->i2hmsg.rsp.status = msg->status; 6677 dport->rp_pwwn = msg->pwwn; 6678 dport->rp_nwwn = msg->nwwn; 6679 6680 if ((msg->status == BFA_STATUS_OK) || 6681 (msg->status == BFA_STATUS_DPORT_NO_SFP)) { 6682 bfa_trc(dport->bfa, msg->status); 6683 bfa_trc(dport->bfa, dport->rp_pwwn); 6684 bfa_trc(dport->bfa, dport->rp_nwwn); 6685 bfa_sm_send_event(dport, BFA_DPORT_SM_FWRSP); 6686 6687 } else { 6688 bfa_trc(dport->bfa, msg->status); 6689 bfa_sm_send_event(dport, BFA_DPORT_SM_REQFAIL); 6690 } 6691 bfa_cb_fcdiag_dport(dport, msg->status); 6692 } 6693 6694 static bfa_boolean_t 6695 bfa_dport_is_sending_req(struct bfa_dport_s *dport) 6696 { 6697 if (bfa_sm_cmp_state(dport, bfa_dport_sm_enabling) || 6698 bfa_sm_cmp_state(dport, bfa_dport_sm_enabling_qwait) || 6699 bfa_sm_cmp_state(dport, bfa_dport_sm_disabling) || 6700 bfa_sm_cmp_state(dport, bfa_dport_sm_disabling_qwait) || 6701 bfa_sm_cmp_state(dport, bfa_dport_sm_starting) || 6702 bfa_sm_cmp_state(dport, bfa_dport_sm_starting_qwait)) { 6703 return BFA_TRUE; 6704 } else { 6705 return BFA_FALSE; 6706 } 6707 } 6708 6709 static void 6710 bfa_dport_scn(struct bfa_dport_s *dport, struct bfi_diag_dport_scn_s *msg) 6711 { 6712 int i; 6713 uint8_t subtesttype; 6714 6715 bfa_trc(dport->bfa, msg->state); 6716 dport->i2hmsg.scn.state = msg->state; 6717 6718 switch (dport->i2hmsg.scn.state) { 6719 case BFI_DPORT_SCN_TESTCOMP: 6720 dport->result.end_time = bfa_get_log_time(); 6721 bfa_trc(dport->bfa, dport->result.end_time); 6722 6723 dport->result.status = msg->info.testcomp.status; 6724 bfa_trc(dport->bfa, dport->result.status); 6725 6726 dport->result.roundtrip_latency = 6727 cpu_to_be32(msg->info.testcomp.latency); 6728 dport->result.est_cable_distance = 6729 cpu_to_be32(msg->info.testcomp.distance); 6730 dport->result.buffer_required = 6731 be16_to_cpu(msg->info.testcomp.numbuffer); 6732 6733 dport->result.frmsz = be16_to_cpu(msg->info.testcomp.frm_sz); 6734 dport->result.speed = msg->info.testcomp.speed; 6735 6736 bfa_trc(dport->bfa, dport->result.roundtrip_latency); 6737 bfa_trc(dport->bfa, dport->result.est_cable_distance); 6738 bfa_trc(dport->bfa, dport->result.buffer_required); 6739 bfa_trc(dport->bfa, dport->result.frmsz); 6740 bfa_trc(dport->bfa, dport->result.speed); 6741 6742 for (i = DPORT_TEST_ELOOP; i < DPORT_TEST_MAX; i++) { 6743 dport->result.subtest[i].status = 6744 msg->info.testcomp.subtest_status[i]; 6745 bfa_trc(dport->bfa, dport->result.subtest[i].status); 6746 } 6747 break; 6748 6749 case BFI_DPORT_SCN_TESTSKIP: 6750 case BFI_DPORT_SCN_DDPORT_ENABLE: 6751 memset(&dport->result, 0, 6752 sizeof(struct bfa_diag_dport_result_s)); 6753 break; 6754 6755 case BFI_DPORT_SCN_TESTSTART: 6756 memset(&dport->result, 0, 6757 sizeof(struct bfa_diag_dport_result_s)); 6758 dport->rp_pwwn = msg->info.teststart.pwwn; 6759 dport->rp_nwwn = msg->info.teststart.nwwn; 6760 dport->lpcnt = cpu_to_be32(msg->info.teststart.numfrm); 6761 bfa_dport_result_start(dport, msg->info.teststart.mode); 6762 break; 6763 6764 case BFI_DPORT_SCN_SUBTESTSTART: 6765 subtesttype = msg->info.teststart.type; 6766 dport->result.subtest[subtesttype].start_time = 6767 bfa_get_log_time(); 6768 dport->result.subtest[subtesttype].status = 6769 DPORT_TEST_ST_INPRG; 6770 6771 bfa_trc(dport->bfa, subtesttype); 6772 bfa_trc(dport->bfa, 6773 dport->result.subtest[subtesttype].start_time); 6774 break; 6775 6776 case BFI_DPORT_SCN_SFP_REMOVED: 6777 case BFI_DPORT_SCN_DDPORT_DISABLED: 6778 case BFI_DPORT_SCN_DDPORT_DISABLE: 6779 case BFI_DPORT_SCN_FCPORT_DISABLE: 6780 dport->result.status = DPORT_TEST_ST_IDLE; 6781 break; 6782 6783 default: 6784 bfa_sm_fault(dport->bfa, msg->state); 6785 } 6786 6787 bfa_sm_send_event(dport, BFA_DPORT_SM_SCN); 6788 } 6789 6790 /* 6791 * Dport enable 6792 * 6793 * @param[in] *bfa - bfa data struct 6794 */ 6795 bfa_status_t 6796 bfa_dport_enable(struct bfa_s *bfa, u32 lpcnt, u32 pat, 6797 bfa_cb_diag_t cbfn, void *cbarg) 6798 { 6799 struct bfa_fcdiag_s *fcdiag = BFA_FCDIAG_MOD(bfa); 6800 struct bfa_dport_s *dport = &fcdiag->dport; 6801 6802 /* 6803 * Dport is not support in MEZZ card 6804 */ 6805 if (bfa_mfg_is_mezz(dport->bfa->ioc.attr->card_type)) { 6806 bfa_trc(dport->bfa, BFA_STATUS_PBC); 6807 return BFA_STATUS_CMD_NOTSUPP_MEZZ; 6808 } 6809 6810 /* 6811 * Dport is supported in CT2 or above 6812 */ 6813 if (!(bfa_asic_id_ct2(dport->bfa->ioc.pcidev.device_id))) { 6814 bfa_trc(dport->bfa, dport->bfa->ioc.pcidev.device_id); 6815 return BFA_STATUS_FEATURE_NOT_SUPPORTED; 6816 } 6817 6818 /* 6819 * Check to see if IOC is down 6820 */ 6821 if (!bfa_iocfc_is_operational(bfa)) 6822 return BFA_STATUS_IOC_NON_OP; 6823 6824 /* if port is PBC disabled, return error */ 6825 if (bfa_fcport_is_pbcdisabled(bfa)) { 6826 bfa_trc(dport->bfa, BFA_STATUS_PBC); 6827 return BFA_STATUS_PBC; 6828 } 6829 6830 /* 6831 * Check if port mode is FC port 6832 */ 6833 if (bfa_ioc_get_type(&bfa->ioc) != BFA_IOC_TYPE_FC) { 6834 bfa_trc(dport->bfa, bfa_ioc_get_type(&bfa->ioc)); 6835 return BFA_STATUS_CMD_NOTSUPP_CNA; 6836 } 6837 6838 /* 6839 * Check if port is in LOOP mode 6840 */ 6841 if ((bfa_fcport_get_cfg_topology(bfa) == BFA_PORT_TOPOLOGY_LOOP) || 6842 (bfa_fcport_get_topology(bfa) == BFA_PORT_TOPOLOGY_LOOP)) { 6843 bfa_trc(dport->bfa, 0); 6844 return BFA_STATUS_TOPOLOGY_LOOP; 6845 } 6846 6847 /* 6848 * Check if port is TRUNK mode 6849 */ 6850 if (bfa_fcport_is_trunk_enabled(bfa)) { 6851 bfa_trc(dport->bfa, 0); 6852 return BFA_STATUS_ERROR_TRUNK_ENABLED; 6853 } 6854 6855 /* 6856 * Check if diag loopback is running 6857 */ 6858 if (bfa_fcdiag_lb_is_running(bfa)) { 6859 bfa_trc(dport->bfa, 0); 6860 return BFA_STATUS_DIAG_BUSY; 6861 } 6862 6863 /* 6864 * Check to see if port is disable or in dport state 6865 */ 6866 if ((bfa_fcport_is_disabled(bfa) == BFA_FALSE) && 6867 (bfa_fcport_is_dport(bfa) == BFA_FALSE)) { 6868 bfa_trc(dport->bfa, 0); 6869 return BFA_STATUS_PORT_NOT_DISABLED; 6870 } 6871 6872 /* 6873 * Check if dport is in dynamic mode 6874 */ 6875 if (dport->dynamic) 6876 return BFA_STATUS_DDPORT_ERR; 6877 6878 /* 6879 * Check if dport is busy 6880 */ 6881 if (bfa_dport_is_sending_req(dport)) 6882 return BFA_STATUS_DEVBUSY; 6883 6884 /* 6885 * Check if dport is already enabled 6886 */ 6887 if (bfa_sm_cmp_state(dport, bfa_dport_sm_enabled)) { 6888 bfa_trc(dport->bfa, 0); 6889 return BFA_STATUS_DPORT_ENABLED; 6890 } 6891 6892 bfa_trc(dport->bfa, lpcnt); 6893 bfa_trc(dport->bfa, pat); 6894 dport->lpcnt = (lpcnt) ? lpcnt : DPORT_ENABLE_LOOPCNT_DEFAULT; 6895 dport->payload = (pat) ? pat : LB_PATTERN_DEFAULT; 6896 dport->cbfn = cbfn; 6897 dport->cbarg = cbarg; 6898 6899 bfa_sm_send_event(dport, BFA_DPORT_SM_ENABLE); 6900 return BFA_STATUS_OK; 6901 } 6902 6903 /* 6904 * Dport disable 6905 * 6906 * @param[in] *bfa - bfa data struct 6907 */ 6908 bfa_status_t 6909 bfa_dport_disable(struct bfa_s *bfa, bfa_cb_diag_t cbfn, void *cbarg) 6910 { 6911 struct bfa_fcdiag_s *fcdiag = BFA_FCDIAG_MOD(bfa); 6912 struct bfa_dport_s *dport = &fcdiag->dport; 6913 6914 if (bfa_ioc_is_disabled(&bfa->ioc)) 6915 return BFA_STATUS_IOC_DISABLED; 6916 6917 /* if port is PBC disabled, return error */ 6918 if (bfa_fcport_is_pbcdisabled(bfa)) { 6919 bfa_trc(dport->bfa, BFA_STATUS_PBC); 6920 return BFA_STATUS_PBC; 6921 } 6922 6923 /* 6924 * Check if dport is in dynamic mode 6925 */ 6926 if (dport->dynamic) { 6927 return BFA_STATUS_DDPORT_ERR; 6928 } 6929 6930 /* 6931 * Check to see if port is disable or in dport state 6932 */ 6933 if ((bfa_fcport_is_disabled(bfa) == BFA_FALSE) && 6934 (bfa_fcport_is_dport(bfa) == BFA_FALSE)) { 6935 bfa_trc(dport->bfa, 0); 6936 return BFA_STATUS_PORT_NOT_DISABLED; 6937 } 6938 6939 /* 6940 * Check if dport is busy 6941 */ 6942 if (bfa_dport_is_sending_req(dport)) 6943 return BFA_STATUS_DEVBUSY; 6944 6945 /* 6946 * Check if dport is already disabled 6947 */ 6948 if (bfa_sm_cmp_state(dport, bfa_dport_sm_disabled)) { 6949 bfa_trc(dport->bfa, 0); 6950 return BFA_STATUS_DPORT_DISABLED; 6951 } 6952 6953 dport->cbfn = cbfn; 6954 dport->cbarg = cbarg; 6955 6956 bfa_sm_send_event(dport, BFA_DPORT_SM_DISABLE); 6957 return BFA_STATUS_OK; 6958 } 6959 6960 /* 6961 * Dport start -- restart dport test 6962 * 6963 * @param[in] *bfa - bfa data struct 6964 */ 6965 bfa_status_t 6966 bfa_dport_start(struct bfa_s *bfa, u32 lpcnt, u32 pat, 6967 bfa_cb_diag_t cbfn, void *cbarg) 6968 { 6969 struct bfa_fcdiag_s *fcdiag = BFA_FCDIAG_MOD(bfa); 6970 struct bfa_dport_s *dport = &fcdiag->dport; 6971 6972 /* 6973 * Check to see if IOC is down 6974 */ 6975 if (!bfa_iocfc_is_operational(bfa)) 6976 return BFA_STATUS_IOC_NON_OP; 6977 6978 /* 6979 * Check if dport is in dynamic mode 6980 */ 6981 if (dport->dynamic) 6982 return BFA_STATUS_DDPORT_ERR; 6983 6984 /* 6985 * Check if dport is busy 6986 */ 6987 if (bfa_dport_is_sending_req(dport)) 6988 return BFA_STATUS_DEVBUSY; 6989 6990 /* 6991 * Check if dport is in enabled state. 6992 * Test can only be restart when previous test has completed 6993 */ 6994 if (!bfa_sm_cmp_state(dport, bfa_dport_sm_enabled)) { 6995 bfa_trc(dport->bfa, 0); 6996 return BFA_STATUS_DPORT_DISABLED; 6997 6998 } else { 6999 if (dport->test_state == BFA_DPORT_ST_NO_SFP) 7000 return BFA_STATUS_DPORT_INV_SFP; 7001 7002 if (dport->test_state == BFA_DPORT_ST_INP) 7003 return BFA_STATUS_DEVBUSY; 7004 7005 WARN_ON(dport->test_state != BFA_DPORT_ST_COMP); 7006 } 7007 7008 bfa_trc(dport->bfa, lpcnt); 7009 bfa_trc(dport->bfa, pat); 7010 7011 dport->lpcnt = (lpcnt) ? lpcnt : DPORT_ENABLE_LOOPCNT_DEFAULT; 7012 dport->payload = (pat) ? pat : LB_PATTERN_DEFAULT; 7013 7014 dport->cbfn = cbfn; 7015 dport->cbarg = cbarg; 7016 7017 bfa_sm_send_event(dport, BFA_DPORT_SM_START); 7018 return BFA_STATUS_OK; 7019 } 7020 7021 /* 7022 * Dport show -- return dport test result 7023 * 7024 * @param[in] *bfa - bfa data struct 7025 */ 7026 bfa_status_t 7027 bfa_dport_show(struct bfa_s *bfa, struct bfa_diag_dport_result_s *result) 7028 { 7029 struct bfa_fcdiag_s *fcdiag = BFA_FCDIAG_MOD(bfa); 7030 struct bfa_dport_s *dport = &fcdiag->dport; 7031 7032 /* 7033 * Check to see if IOC is down 7034 */ 7035 if (!bfa_iocfc_is_operational(bfa)) 7036 return BFA_STATUS_IOC_NON_OP; 7037 7038 /* 7039 * Check if dport is busy 7040 */ 7041 if (bfa_dport_is_sending_req(dport)) 7042 return BFA_STATUS_DEVBUSY; 7043 7044 /* 7045 * Check if dport is in enabled state. 7046 */ 7047 if (!bfa_sm_cmp_state(dport, bfa_dport_sm_enabled)) { 7048 bfa_trc(dport->bfa, 0); 7049 return BFA_STATUS_DPORT_DISABLED; 7050 7051 } 7052 7053 /* 7054 * Check if there is SFP 7055 */ 7056 if (dport->test_state == BFA_DPORT_ST_NO_SFP) 7057 return BFA_STATUS_DPORT_INV_SFP; 7058 7059 memcpy(result, &dport->result, sizeof(struct bfa_diag_dport_result_s)); 7060 7061 return BFA_STATUS_OK; 7062 } 7063