xref: /freebsd/sys/dev/ocs_fc/ocs_sport.c (revision 8e9b3e707151d136ec95e7f1d37556e39c1e228c)
1 /*-
2  * Copyright (c) 2017 Broadcom. All rights reserved.
3  * The term "Broadcom" refers to Broadcom Limited and/or its subsidiaries.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions are met:
7  *
8  * 1. Redistributions of source code must retain the above copyright notice,
9  *    this list of conditions and the following disclaimer.
10  *
11  * 2. Redistributions in binary form must reproduce the above copyright notice,
12  *    this list of conditions and the following disclaimer in the documentation
13  *    and/or other materials provided with the distribution.
14  *
15  * 3. Neither the name of the copyright holder nor the names of its contributors
16  *    may be used to endorse or promote products derived from this software
17  *    without specific prior written permission.
18  *
19  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
20  * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
21  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
22  * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
23  * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29  * POSSIBILITY OF SUCH DAMAGE.
30  *
31  * $FreeBSD$
32  */
33 
34 /**
35  * @file
36  * Details SLI port (sport) functions.
37  */
38 
39 
40 #include "ocs.h"
41 #include "ocs_fabric.h"
42 #include "ocs_els.h"
43 #include "ocs_device.h"
44 
45 static void ocs_vport_update_spec(ocs_sport_t *sport);
46 static void ocs_vport_link_down(ocs_sport_t *sport);
47 
48 void ocs_mgmt_sport_list(ocs_textbuf_t *textbuf, void *sport);
49 void ocs_mgmt_sport_get_all(ocs_textbuf_t *textbuf, void *sport);
50 int ocs_mgmt_sport_get(ocs_textbuf_t *textbuf, char *parent, char *name, void *sport);
51 int ocs_mgmt_sport_set(char *parent, char *name, char *value, void *sport);
52 int ocs_mgmt_sport_exec(char *parent, char *action, void *arg_in, uint32_t arg_in_length,
53 		void *arg_out, uint32_t arg_out_length, void *sport);
54 static ocs_mgmt_functions_t sport_mgmt_functions = {
55 	.get_list_handler = ocs_mgmt_sport_list,
56 	.get_handler = ocs_mgmt_sport_get,
57 	.get_all_handler = ocs_mgmt_sport_get_all,
58 	.set_handler = ocs_mgmt_sport_set,
59 	.exec_handler = ocs_mgmt_sport_exec,
60 };
61 
62 /*!
63 @defgroup sport_sm SLI Port (sport) State Machine: States
64 */
65 
66 /**
67  * @ingroup sport_sm
68  * @brief SLI port HW callback.
69  *
70  * @par Description
71  * This function is called in response to a HW sport event. This code resolves
72  * the reference to the sport object, and posts the corresponding event.
73  *
74  * @param arg Pointer to the OCS context.
75  * @param event HW sport event.
76  * @param data Application-specific event (pointer to the sport).
77  *
78  * @return Returns 0 on success, or a negative error value on failure.
79  */
80 
81 int32_t
82 ocs_port_cb(void *arg, ocs_hw_port_event_e event, void *data)
83 {
84 	ocs_t *ocs = arg;
85 	ocs_sli_port_t *sport = data;
86 
87 	switch (event) {
88 	case OCS_HW_PORT_ALLOC_OK:
89 		ocs_log_debug(ocs, "OCS_HW_PORT_ALLOC_OK\n");
90 		ocs_sm_post_event(&sport->sm, OCS_EVT_SPORT_ALLOC_OK, NULL);
91 		break;
92 	case OCS_HW_PORT_ALLOC_FAIL:
93 		ocs_log_debug(ocs, "OCS_HW_PORT_ALLOC_FAIL\n");
94 		ocs_sm_post_event(&sport->sm, OCS_EVT_SPORT_ALLOC_FAIL, NULL);
95 		break;
96 	case OCS_HW_PORT_ATTACH_OK:
97 		ocs_log_debug(ocs, "OCS_HW_PORT_ATTACH_OK\n");
98 		ocs_sm_post_event(&sport->sm, OCS_EVT_SPORT_ATTACH_OK, NULL);
99 		break;
100 	case OCS_HW_PORT_ATTACH_FAIL:
101 		ocs_log_debug(ocs, "OCS_HW_PORT_ATTACH_FAIL\n");
102 		ocs_sm_post_event(&sport->sm, OCS_EVT_SPORT_ATTACH_FAIL, NULL);
103 		break;
104 	case OCS_HW_PORT_FREE_OK:
105 		ocs_log_debug(ocs, "OCS_HW_PORT_FREE_OK\n");
106 		ocs_sm_post_event(&sport->sm, OCS_EVT_SPORT_FREE_OK, NULL);
107 		break;
108 	case OCS_HW_PORT_FREE_FAIL:
109 		ocs_log_debug(ocs, "OCS_HW_PORT_FREE_FAIL\n");
110 		ocs_sm_post_event(&sport->sm, OCS_EVT_SPORT_FREE_FAIL, NULL);
111 		break;
112 	default:
113 		ocs_log_test(ocs, "unknown event %#x\n", event);
114 	}
115 
116 	return 0;
117 }
118 
119 /**
120  * @ingroup sport_sm
121  * @brief Allocate a SLI port object.
122  *
123  * @par Description
124  * A sport object is allocated and associated with the domain. Various
125  * structure members are initialized.
126  *
127  * @param domain Pointer to the domain structure.
128  * @param wwpn World wide port name in host endian.
129  * @param wwnn World wide node name in host endian.
130  * @param fc_id Port ID of sport may be specified, use UINT32_MAX to fabric choose
131  * @param enable_ini Enables initiator capability on this port using a non-zero value.
132  * @param enable_tgt Enables target capability on this port using a non-zero value.
133  *
134  * @return Pointer to an ocs_sport_t object; or NULL.
135  */
136 
137 ocs_sport_t *
138 ocs_sport_alloc(ocs_domain_t *domain, uint64_t wwpn, uint64_t wwnn, uint32_t fc_id, uint8_t enable_ini, uint8_t enable_tgt)
139 {
140 	ocs_sport_t *sport;
141 
142 	if (domain->ocs->ctrlmask & OCS_CTRLMASK_INHIBIT_INITIATOR) {
143 		enable_ini = 0;
144 	}
145 
146 	/* Return a failure if this sport has already been allocated */
147 	if (wwpn != 0) {
148 		sport = ocs_sport_find_wwn(domain, wwnn, wwpn);
149 		if (sport != NULL) {
150 			ocs_log_test(domain->ocs, "Failed: SPORT %016llx  %016llx already allocated\n",
151 				     (unsigned long long)wwnn, (unsigned long long)wwpn);
152 			return NULL;
153 		}
154 	}
155 
156 	sport = ocs_malloc(domain->ocs, sizeof(*sport), OCS_M_NOWAIT | OCS_M_ZERO);
157 	if (sport) {
158 		sport->ocs = domain->ocs;
159 		ocs_snprintf(sport->display_name, sizeof(sport->display_name), "------");
160 		sport->domain = domain;
161 		sport->lookup = spv_new(domain->ocs);
162 		sport->instance_index = domain->sport_instance_count++;
163 		ocs_sport_lock_init(sport);
164 		ocs_list_init(&sport->node_list, ocs_node_t, link);
165 		sport->sm.app = sport;
166 		sport->enable_ini = enable_ini;
167 		sport->enable_tgt = enable_tgt;
168 		sport->enable_rscn = (sport->enable_ini || (sport->enable_tgt && enable_target_rscn(sport->ocs)));
169 
170 		/* Copy service parameters from domain */
171 		ocs_memcpy(sport->service_params, domain->service_params, sizeof(fc_plogi_payload_t));
172 
173 		/* Update requested fc_id */
174 		sport->fc_id = fc_id;
175 
176 		/* Update the sport's service parameters for the new wwn's */
177 		sport->wwpn = wwpn;
178 		sport->wwnn = wwnn;
179 		ocs_snprintf(sport->wwnn_str, sizeof(sport->wwnn_str), "%016llx" , (unsigned long long)wwnn);
180 
181 		/* Initialize node group list */
182 		ocs_lock_init(sport->ocs, &sport->node_group_lock, "node_group_lock[%d]", sport->instance_index);
183 		ocs_list_init(&sport->node_group_dir_list, ocs_node_group_dir_t, link);
184 
185 		/* if this is the "first" sport of the domain, then make it the "phys" sport */
186 		ocs_domain_lock(domain);
187 			if (ocs_list_empty(&domain->sport_list)) {
188 				domain->sport = sport;
189 			}
190 
191 			ocs_list_add_tail(&domain->sport_list, sport);
192 		ocs_domain_unlock(domain);
193 
194 		sport->mgmt_functions = &sport_mgmt_functions;
195 
196 		ocs_log_debug(domain->ocs, "[%s] allocate sport\n", sport->display_name);
197 	}
198 	return sport;
199 }
200 
201 /**
202  * @ingroup sport_sm
203  * @brief Free a SLI port object.
204  *
205  * @par Description
206  * The sport object is freed.
207  *
208  * @param sport Pointer to the SLI port object.
209  *
210  * @return None.
211  */
212 
213 void
214 ocs_sport_free(ocs_sport_t *sport)
215 {
216 	ocs_domain_t *domain;
217 	ocs_node_group_dir_t *node_group_dir;
218 	ocs_node_group_dir_t *node_group_dir_next;
219 	int post_all_free = FALSE;
220 
221 	if (sport) {
222 		domain = sport->domain;
223 		ocs_log_debug(domain->ocs, "[%s] free sport\n", sport->display_name);
224 		ocs_domain_lock(domain);
225 			ocs_list_remove(&domain->sport_list, sport);
226 			ocs_sport_lock(sport);
227 				spv_del(sport->lookup);
228 				sport->lookup = NULL;
229 
230 				ocs_lock(&domain->lookup_lock);
231 					/* Remove the sport from the domain's sparse vector lookup table */
232 					spv_set(domain->lookup, sport->fc_id, NULL);
233 				ocs_unlock(&domain->lookup_lock);
234 
235 				/* if this is the physical sport, then clear it out of the domain */
236 				if (sport == domain->sport) {
237 					domain->sport = NULL;
238 				}
239 
240 				/*
241 				 * If the domain's sport_list is empty, then post the ALL_NODES_FREE event to the domain,
242 				 * after the lock is released. The domain may be free'd as a result of the event.
243 				 */
244 				if (ocs_list_empty(&domain->sport_list)) {
245 					post_all_free = TRUE;
246 				}
247 
248 				/* Free any node group directories */
249 				ocs_lock(&sport->node_group_lock);
250 					ocs_list_foreach_safe(&sport->node_group_dir_list, node_group_dir, node_group_dir_next) {
251 						ocs_unlock(&sport->node_group_lock);
252 							ocs_node_group_dir_free(node_group_dir);
253 						ocs_lock(&sport->node_group_lock);
254 					}
255 				ocs_unlock(&sport->node_group_lock);
256 			ocs_sport_unlock(sport);
257 		ocs_domain_unlock(domain);
258 
259 		if (post_all_free) {
260 			ocs_domain_post_event(domain, OCS_EVT_ALL_CHILD_NODES_FREE, NULL);
261 		}
262 
263 		ocs_sport_lock_free(sport);
264 		ocs_lock_free(&sport->node_group_lock);
265 		ocs_scsi_sport_deleted(sport);
266 
267 		ocs_free(domain->ocs, sport, sizeof(*sport));
268 
269 	}
270 }
271 
272 /**
273  * @ingroup sport_sm
274  * @brief Free memory resources of a SLI port object.
275  *
276  * @par Description
277  * After the sport object is freed, its child objects are freed.
278  *
279  * @param sport Pointer to the SLI port object.
280  *
281  * @return None.
282  */
283 
284 void ocs_sport_force_free(ocs_sport_t *sport)
285 {
286 	ocs_node_t *node;
287 	ocs_node_t *next;
288 
289 	/* shutdown sm processing */
290 	ocs_sm_disable(&sport->sm);
291 
292 	ocs_scsi_notify_sport_force_free(sport);
293 
294 	ocs_sport_lock(sport);
295 		ocs_list_foreach_safe(&sport->node_list, node, next) {
296 			ocs_node_force_free(node);
297 		}
298 	ocs_sport_unlock(sport);
299 	ocs_sport_free(sport);
300 }
301 
302 /**
303  * @ingroup sport_sm
304  * @brief Return a SLI port object, given an instance index.
305  *
306  * @par Description
307  * A pointer to a sport object is returned, given its instance @c index.
308  *
309  * @param domain Pointer to the domain.
310  * @param index Instance index value to find.
311  *
312  * @return Returns a pointer to the ocs_sport_t object; or NULL.
313  */
314 
315 ocs_sport_t *
316 ocs_sport_get_instance(ocs_domain_t *domain, uint32_t index)
317 {
318 	ocs_sport_t *sport;
319 
320 	ocs_domain_lock(domain);
321 		ocs_list_foreach(&domain->sport_list, sport) {
322 			if (sport->instance_index == index) {
323 				ocs_domain_unlock(domain);
324 				return sport;
325 			}
326 		}
327 	ocs_domain_unlock(domain);
328 	return NULL;
329 }
330 
331 /**
332  * @ingroup sport_sm
333  * @brief Find a SLI port object, given an FC_ID.
334  *
335  * @par Description
336  * Returns a pointer to the sport object, given an FC_ID.
337  *
338  * @param domain Pointer to the domain.
339  * @param d_id FC_ID to find.
340  *
341  * @return Returns a pointer to the ocs_sport_t; or NULL.
342  */
343 
344 ocs_sport_t *
345 ocs_sport_find(ocs_domain_t *domain, uint32_t d_id)
346 {
347 	ocs_sport_t *sport;
348 
349 	ocs_assert(domain, NULL);
350 	ocs_lock(&domain->lookup_lock);
351 		if (domain->lookup == NULL) {
352 			ocs_log_test(domain->ocs, "assertion failed: domain->lookup is not valid\n");
353 			ocs_unlock(&domain->lookup_lock);
354 			return NULL;
355 		}
356 
357 		sport = spv_get(domain->lookup, d_id);
358 	ocs_unlock(&domain->lookup_lock);
359 	return sport;
360 }
361 
362 /**
363  * @ingroup sport_sm
364  * @brief Find a SLI port, given the WWNN and WWPN.
365  *
366  * @par Description
367  * Return a pointer to a sport, given the WWNN and WWPN.
368  *
369  * @param domain Pointer to the domain.
370  * @param wwnn World wide node name.
371  * @param wwpn World wide port name.
372  *
373  * @return Returns a pointer to a SLI port, if found; or NULL.
374  */
375 
376 ocs_sport_t *
377 ocs_sport_find_wwn(ocs_domain_t *domain, uint64_t wwnn, uint64_t wwpn)
378 {
379 	ocs_sport_t *sport = NULL;
380 
381 	ocs_domain_lock(domain);
382 		ocs_list_foreach(&domain->sport_list, sport) {
383 			if ((sport->wwnn == wwnn) && (sport->wwpn == wwpn)) {
384 				ocs_domain_unlock(domain);
385 				return sport;
386 			}
387 		}
388 	ocs_domain_unlock(domain);
389 	return NULL;
390 }
391 
392 /**
393  * @ingroup sport_sm
394  * @brief Request a SLI port attach.
395  *
396  * @par Description
397  * External call to request an attach for a sport, given an FC_ID.
398  *
399  * @param sport Pointer to the sport context.
400  * @param fc_id FC_ID of which to attach.
401  *
402  * @return Returns 0 on success, or a negative error value on failure.
403  */
404 
405 int32_t
406 ocs_sport_attach(ocs_sport_t *sport, uint32_t fc_id)
407 {
408 	ocs_hw_rtn_e rc;
409 	ocs_node_t *node;
410 
411 	/* Set our lookup */
412 	ocs_lock(&sport->domain->lookup_lock);
413 		spv_set(sport->domain->lookup, fc_id, sport);
414 	ocs_unlock(&sport->domain->lookup_lock);
415 
416 	/* Update our display_name */
417 	ocs_node_fcid_display(fc_id, sport->display_name, sizeof(sport->display_name));
418 	ocs_sport_lock(sport);
419 		ocs_list_foreach(&sport->node_list, node) {
420 			ocs_node_update_display_name(node);
421 		}
422 	ocs_sport_unlock(sport);
423 	ocs_log_debug(sport->ocs, "[%s] attach sport: fc_id x%06x\n", sport->display_name, fc_id);
424 
425 	rc = ocs_hw_port_attach(&sport->ocs->hw, sport, fc_id);
426 	if (rc != OCS_HW_RTN_SUCCESS) {
427 		ocs_log_err(sport->ocs, "ocs_hw_port_attach failed: %d\n", rc);
428 		return -1;
429 	}
430 	return 0;
431 }
432 
433 /**
434  * @brief Common SLI port state machine declarations and initialization.
435  */
436 #define std_sport_state_decl() \
437 	ocs_sport_t *sport = NULL; \
438 	ocs_domain_t *domain = NULL; \
439 	ocs_t *ocs = NULL; \
440 	\
441 	ocs_assert(ctx, NULL); \
442 	sport = ctx->app; \
443 	ocs_assert(sport, NULL); \
444 	\
445 	domain = sport->domain; \
446 	ocs_assert(domain, NULL); \
447 	ocs = sport->ocs; \
448 	ocs_assert(ocs, NULL);
449 
450 /**
451  * @brief Common SLI port state machine trace logging.
452  */
453 #define sport_sm_trace(sport)  \
454 	do { \
455 		if (OCS_LOG_ENABLE_DOMAIN_SM_TRACE(ocs)) \
456 			ocs_log_debug(ocs, "[%s] %-20s\n", sport->display_name, ocs_sm_event_name(evt)); \
457 	} while (0)
458 
459 
460 /**
461  * @brief SLI port state machine: Common event handler.
462  *
463  * @par Description
464  * Handle common sport events.
465  *
466  * @param funcname Function name to display.
467  * @param ctx Sport state machine context.
468  * @param evt Event to process.
469  * @param arg Per event optional argument.
470  *
471  * @return Returns NULL.
472  */
473 
474 static void *
475 __ocs_sport_common(const char *funcname, ocs_sm_ctx_t *ctx, ocs_sm_event_t evt, void *arg)
476 {
477 	std_sport_state_decl();
478 
479 	switch(evt) {
480 	case OCS_EVT_ENTER:
481 	case OCS_EVT_REENTER:
482 	case OCS_EVT_EXIT:
483 	case OCS_EVT_ALL_CHILD_NODES_FREE:
484 		break;
485 	case OCS_EVT_SPORT_ATTACH_OK:
486 			ocs_sm_transition(ctx, __ocs_sport_attached, NULL);
487 		break;
488 	case OCS_EVT_SHUTDOWN: {
489 		ocs_node_t *node;
490 		ocs_node_t *node_next;
491 		int node_list_empty;
492 
493 		/* Flag this sport as shutting down */
494 		sport->shutting_down = 1;
495 
496 		if (sport->is_vport) {
497 			ocs_vport_link_down(sport);
498 		}
499 
500 		ocs_sport_lock(sport);
501 			node_list_empty = ocs_list_empty(&sport->node_list);
502 		ocs_sport_unlock(sport);
503 
504 		if (node_list_empty) {
505 			/* sm: node list is empty / ocs_hw_port_free
506 			 * Remove the sport from the domain's sparse vector lookup table */
507 			ocs_lock(&domain->lookup_lock);
508 				spv_set(domain->lookup, sport->fc_id, NULL);
509 			ocs_unlock(&domain->lookup_lock);
510 			ocs_sm_transition(ctx, __ocs_sport_wait_port_free, NULL);
511 			if (ocs_hw_port_free(&ocs->hw, sport)) {
512 				ocs_log_test(sport->ocs, "ocs_hw_port_free failed\n");
513 				/* Not much we can do, free the sport anyways */
514 				ocs_sport_free(sport);
515 			}
516 		} else {
517 			/* sm: node list is not empty / shutdown nodes */
518 			ocs_sm_transition(ctx, __ocs_sport_wait_shutdown, NULL);
519 			ocs_sport_lock(sport);
520 				ocs_list_foreach_safe(&sport->node_list, node, node_next) {
521 					/*
522 					 * If this is a vport, logout of the fabric controller so that it
523 					 * deletes the vport on the switch.
524 					 */
525 					if((node->rnode.fc_id == FC_ADDR_FABRIC) && (sport->is_vport)) {
526 						/* if link is down, don't send logo */
527 						if (sport->ocs->hw.link.status == SLI_LINK_STATUS_DOWN) {
528 							ocs_node_post_event(node, OCS_EVT_SHUTDOWN, NULL);
529 						} else {
530 							ocs_log_debug(ocs,"[%s] sport shutdown vport,sending logo to node\n",
531 								      node->display_name);
532 
533 							if (ocs_send_logo(node, OCS_FC_ELS_SEND_DEFAULT_TIMEOUT,
534 								  0, NULL, NULL) == NULL) {
535 								/* failed to send LOGO, go ahead and cleanup node anyways */
536 								node_printf(node, "Failed to send LOGO\n");
537 								ocs_node_post_event(node, OCS_EVT_SHUTDOWN_EXPLICIT_LOGO, NULL);
538 							} else {
539 								/* sent LOGO, wait for response */
540 								ocs_node_transition(node, __ocs_d_wait_logo_rsp, NULL);
541 							}
542 						}
543 					} else {
544 						ocs_node_post_event(node, OCS_EVT_SHUTDOWN, NULL);
545 					}
546 				}
547 			ocs_sport_unlock(sport);
548 		}
549 		break;
550 	}
551 	default:
552 		ocs_log_test(sport->ocs, "[%s] %-20s %-20s not handled\n", sport->display_name, funcname, ocs_sm_event_name(evt));
553 		break;
554 	}
555 
556 	return NULL;
557 }
558 
559 /**
560  * @ingroup sport_sm
561  * @brief SLI port state machine: Physical sport allocated.
562  *
563  * @par Description
564  * This is the initial state for sport objects.
565  *
566  * @param ctx Remote node state machine context.
567  * @param evt Event to process.
568  * @param arg Per event optional argument.
569  *
570  * @return Returns NULL.
571  */
572 
573 void *
574 __ocs_sport_allocated(ocs_sm_ctx_t *ctx, ocs_sm_event_t evt, void *arg)
575 {
576 	std_sport_state_decl();
577 
578 	sport_sm_trace(sport);
579 
580 	switch(evt) {
581 	/* the physical sport is attached */
582 	case OCS_EVT_SPORT_ATTACH_OK:
583 		ocs_assert(sport == domain->sport, NULL);
584 		ocs_sm_transition(ctx, __ocs_sport_attached, NULL);
585 		break;
586 
587 	case OCS_EVT_SPORT_ALLOC_OK:
588 		/* ignore */
589 		break;
590 	default:
591 		__ocs_sport_common(__func__, ctx, evt, arg);
592 		return NULL;
593 	}
594 	return NULL;
595 }
596 
597 /**
598  * @ingroup sport_sm
599  * @brief SLI port state machine: Handle initial virtual port events.
600  *
601  * @par Description
602  * This state is entered when a virtual port is instantiated,
603  *
604  * @param ctx Remote node state machine context.
605  * @param evt Event to process.
606  * @param arg Per event optional argument.
607  *
608  * @return Returns NULL.
609  */
610 
611 void *
612 __ocs_sport_vport_init(ocs_sm_ctx_t *ctx, ocs_sm_event_t evt, void *arg)
613 {
614 	std_sport_state_decl();
615 
616 	sport_sm_trace(sport);
617 
618 	switch(evt) {
619 	case OCS_EVT_ENTER: {
620 		uint64_t be_wwpn = ocs_htobe64(sport->wwpn);
621 
622 		if (sport->wwpn == 0) {
623 			ocs_log_debug(ocs, "vport: letting f/w select WWN\n");
624 		}
625 
626 		if (sport->fc_id != UINT32_MAX) {
627 			ocs_log_debug(ocs, "vport: hard coding port id: %x\n", sport->fc_id);
628 		}
629 
630 		ocs_sm_transition(ctx, __ocs_sport_vport_wait_alloc, NULL);
631 		/* If wwpn is zero, then we'll let the f/w */
632 		if (ocs_hw_port_alloc(&ocs->hw, sport, sport->domain,
633 			(sport->wwpn == 0) ? NULL : (uint8_t *)&be_wwpn)) {
634 			ocs_log_err(ocs, "Can't allocate port\n");
635 			break;
636 		}
637 
638 
639 		break;
640 	}
641 	default:
642 		__ocs_sport_common(__func__, ctx, evt, arg);
643 		return NULL;
644 	}
645 	return NULL;
646 }
647 
648 /**
649  * @ingroup sport_sm
650  * @brief SLI port state machine: Wait for the HW SLI port allocation to complete.
651  *
652  * @par Description
653  * Waits for the HW sport allocation request to complete.
654  *
655  * @param ctx Remote node state machine context.
656  * @param evt Event to process.
657  * @param arg Per event optional argument.
658  *
659  * @return Returns NULL.
660  */
661 
662 void *
663 __ocs_sport_vport_wait_alloc(ocs_sm_ctx_t *ctx, ocs_sm_event_t evt, void *arg)
664 {
665 	std_sport_state_decl();
666 
667 	sport_sm_trace(sport);
668 
669 	switch(evt) {
670 	case OCS_EVT_SPORT_ALLOC_OK: {
671 		fc_plogi_payload_t *sp = (fc_plogi_payload_t*) sport->service_params;
672 		ocs_node_t *fabric;
673 
674 		/* If we let f/w assign wwn's, then sport wwn's with those returned by hw */
675 		if (sport->wwnn == 0) {
676 			sport->wwnn = ocs_be64toh(sport->sli_wwnn);
677 			sport->wwpn = ocs_be64toh(sport->sli_wwpn);
678 			ocs_snprintf(sport->wwnn_str, sizeof(sport->wwnn_str), "%016llx", (unsigned long long) sport->wwpn);
679 		}
680 
681 		/* Update the sport's service parameters */
682 		sp->port_name_hi = ocs_htobe32((uint32_t) (sport->wwpn >> 32ll));
683 		sp->port_name_lo = ocs_htobe32((uint32_t) sport->wwpn);
684 		sp->node_name_hi = ocs_htobe32((uint32_t) (sport->wwnn >> 32ll));
685 		sp->node_name_lo = ocs_htobe32((uint32_t) sport->wwnn);
686 
687 		/* if sport->fc_id is uninitialized, then request that the fabric node use FDISC
688 		 * to find an fc_id.   Otherwise we're restoring vports, or we're in
689 		 * fabric emulation mode, so attach the fc_id
690 		 */
691 		if (sport->fc_id == UINT32_MAX) {
692 			fabric = ocs_node_alloc(sport, FC_ADDR_FABRIC, FALSE, FALSE);
693 			if (fabric == NULL) {
694 				ocs_log_err(ocs, "ocs_node_alloc() failed\n");
695 				return NULL;
696 			}
697 			ocs_node_transition(fabric, __ocs_vport_fabric_init, NULL);
698 		} else {
699 			ocs_snprintf(sport->wwnn_str, sizeof(sport->wwnn_str), "%016llx", (unsigned long long)sport->wwpn);
700 			ocs_sport_attach(sport, sport->fc_id);
701 		}
702 		ocs_sm_transition(ctx, __ocs_sport_vport_allocated, NULL);
703 		break;
704 	}
705 	default:
706 		__ocs_sport_common(__func__, ctx, evt, arg);
707 		return NULL;
708 	}
709 	return NULL;
710 }
711 
712 /**
713  * @ingroup sport_sm
714  * @brief SLI port state machine: virtual sport allocated.
715  *
716  * @par Description
717  * This state is entered after the sport is allocated; it then waits for a fabric node
718  * FDISC to complete, which requests a sport attach.
719  * The sport attach complete is handled in this state.
720  *
721  * @param ctx Remote node state machine context.
722  * @param evt Event to process.
723  * @param arg Per event optional argument.
724  *
725  * @return Returns NULL.
726  */
727 
728 void *
729 __ocs_sport_vport_allocated(ocs_sm_ctx_t *ctx, ocs_sm_event_t evt, void *arg)
730 {
731 	std_sport_state_decl();
732 
733 	sport_sm_trace(sport);
734 
735 	switch(evt) {
736 	case OCS_EVT_SPORT_ATTACH_OK: {
737 		ocs_node_t *node;
738 
739 		if (!(domain->femul_enable)) {
740 			/* Find our fabric node, and forward this event */
741 			node = ocs_node_find(sport, FC_ADDR_FABRIC);
742 			if (node == NULL) {
743 				ocs_log_test(ocs, "can't find node %06x\n", FC_ADDR_FABRIC);
744 				break;
745 			}
746 			/* sm: / forward sport attach to fabric node */
747 			ocs_node_post_event(node, evt, NULL);
748 		}
749 		ocs_sm_transition(ctx, __ocs_sport_attached, NULL);
750 		break;
751 	}
752 	default:
753 		__ocs_sport_common(__func__, ctx, evt, arg);
754 		return NULL;
755 	}
756 	return NULL;
757 }
758 
759 /**
760  * @ingroup sport_sm
761  * @brief SLI port state machine: Attached.
762  *
763  * @par Description
764  * State entered after the sport attach has completed.
765  *
766  * @param ctx Remote node state machine context.
767  * @param evt Event to process.
768  * @param arg Per event optional argument.
769  *
770  * @return Returns NULL.
771  */
772 
773 void *
774 __ocs_sport_attached(ocs_sm_ctx_t *ctx, ocs_sm_event_t evt, void *arg)
775 {
776 	std_sport_state_decl();
777 
778 	sport_sm_trace(sport);
779 
780 	switch(evt) {
781 	case OCS_EVT_ENTER: {
782 		ocs_node_t *node;
783 
784 		ocs_log_debug(ocs, "[%s] SPORT attached WWPN %016llx WWNN %016llx \n", sport->display_name,
785 			sport->wwpn, sport->wwnn);
786 		ocs_sport_lock(sport);
787 			ocs_list_foreach(&sport->node_list, node) {
788 				ocs_node_update_display_name(node);
789 			}
790 		ocs_sport_unlock(sport);
791 		sport->tgt_id = sport->fc_id;
792 		if (sport->enable_ini) {
793 			ocs_scsi_ini_new_sport(sport);
794 		}
795 		if (sport->enable_tgt) {
796 			ocs_scsi_tgt_new_sport(sport);
797 		}
798 
799 		/* Update the vport (if its not the physical sport) parameters */
800 		if (sport->is_vport) {
801 			ocs_vport_update_spec(sport);
802 		}
803 
804 		break;
805 	}
806 
807 	case OCS_EVT_EXIT:
808 		ocs_log_debug(ocs, "[%s] SPORT deattached WWPN %016llx WWNN %016llx \n", sport->display_name,
809 			sport->wwpn, sport->wwnn);
810 		if (sport->enable_ini) {
811 			ocs_scsi_ini_del_sport(sport);
812 		}
813 		if (sport->enable_tgt) {
814 			ocs_scsi_tgt_del_sport(sport);
815 		}
816 		break;
817 	default:
818 		__ocs_sport_common(__func__, ctx, evt, arg);
819 		return NULL;
820 	}
821 	return NULL;
822 }
823 
824 /**
825  * @ingroup sport_sm
826  * @brief SLI port state machine: Wait for the node shutdowns to complete.
827  *
828  * @par Description
829  * Waits for the ALL_CHILD_NODES_FREE event to be posted from the node
830  * shutdown process.
831  *
832  * @param ctx Remote node state machine context.
833  * @param evt Event to process.
834  * @param arg Per event optional argument.
835  *
836  * @return Returns NULL.
837  */
838 
839 void *
840 __ocs_sport_wait_shutdown(ocs_sm_ctx_t *ctx, ocs_sm_event_t evt, void *arg)
841 {
842 	std_sport_state_decl();
843 
844 	sport_sm_trace(sport);
845 
846 	switch(evt) {
847 	case OCS_EVT_SPORT_ALLOC_OK:
848 	case OCS_EVT_SPORT_ALLOC_FAIL:
849 	case OCS_EVT_SPORT_ATTACH_OK:
850 	case OCS_EVT_SPORT_ATTACH_FAIL:
851 		/* ignore these events - just wait for the all free event */
852 		break;
853 
854 	case OCS_EVT_ALL_CHILD_NODES_FREE: {
855 		/* Remove the sport from the domain's sparse vector lookup table */
856 		ocs_lock(&domain->lookup_lock);
857 			spv_set(domain->lookup, sport->fc_id, NULL);
858 		ocs_unlock(&domain->lookup_lock);
859 		ocs_sm_transition(ctx, __ocs_sport_wait_port_free, NULL);
860 		if (ocs_hw_port_free(&ocs->hw, sport)) {
861 			ocs_log_err(sport->ocs, "ocs_hw_port_free failed\n");
862 			/* Not much we can do, free the sport anyways */
863 			ocs_sport_free(sport);
864 		}
865 		break;
866 	}
867 	default:
868 		__ocs_sport_common(__func__, ctx, evt, arg);
869 		return NULL;
870 	}
871 	return NULL;
872 }
873 
874 /**
875  * @ingroup sport_sm
876  * @brief SLI port state machine: Wait for the HW's port free to complete.
877  *
878  * @par Description
879  * Waits for the HW's port free to complete.
880  *
881  * @param ctx Remote node state machine context.
882  * @param evt Event to process.
883  * @param arg Per event optional argument.
884  *
885  * @return Returns NULL.
886  */
887 
888 void *
889 __ocs_sport_wait_port_free(ocs_sm_ctx_t *ctx, ocs_sm_event_t evt, void *arg)
890 {
891 	std_sport_state_decl();
892 
893 	sport_sm_trace(sport);
894 
895 	switch(evt) {
896 	case OCS_EVT_SPORT_ATTACH_OK:
897 		/* Ignore as we are waiting for the free CB */
898 		break;
899 	case OCS_EVT_SPORT_FREE_OK: {
900 		/* All done, free myself */
901 		ocs_sport_free(sport);
902 		break;
903 	}
904 	default:
905 		__ocs_sport_common(__func__, ctx, evt, arg);
906 		return NULL;
907 	}
908 	return NULL;
909 }
910 
911 /**
912  * @ingroup sport_sm
913  * @brief Start the vports on a domain
914  *
915  * @par Description
916  * Use the vport specification to find the associated vports and start them.
917  *
918  * @param domain Pointer to the domain context.
919  *
920  * @return Returns 0 on success, or a negative error value on failure.
921  */
922 int32_t
923 ocs_vport_start(ocs_domain_t *domain)
924 {
925 	ocs_t *ocs = domain->ocs;
926 	ocs_xport_t *xport = ocs->xport;
927 	ocs_vport_spec_t *vport;
928 	ocs_vport_spec_t *next;
929 	ocs_sport_t *sport;
930 	int32_t rc = 0;
931 
932 	ocs_device_lock(ocs);
933 	ocs_list_foreach_safe(&xport->vport_list, vport, next) {
934 		if (vport->domain_instance == domain->instance_index &&
935 		    vport->sport == NULL) {
936 			/* If role not set, skip this vport */
937 			if (!(vport->enable_ini || vport->enable_tgt)) {
938 				continue;
939 			}
940 
941 			/* Allocate a sport */
942 			vport->sport = sport = ocs_sport_alloc(domain, vport->wwpn, vport->wwnn, vport->fc_id,
943 							       vport->enable_ini, vport->enable_tgt);
944 			if (sport == NULL) {
945 				rc = -1;
946 			} else {
947 				sport->is_vport = 1;
948 				sport->tgt_data = vport->tgt_data;
949 				sport->ini_data = vport->ini_data;
950 
951 				/* Transition to vport_init */
952 				ocs_sm_transition(&sport->sm, __ocs_sport_vport_init, NULL);
953 			}
954 		}
955 	}
956 	ocs_device_unlock(ocs);
957 	return rc;
958 }
959 
960 /**
961  * @ingroup sport_sm
962  * @brief Clear the sport reference in the vport specification.
963  *
964  * @par Description
965  * Clear the sport pointer on the vport specification when the vport is torn down. This allows it to be
966  * re-created when the link is re-established.
967  *
968  * @param sport Pointer to the sport context.
969  */
970 static void
971 ocs_vport_link_down(ocs_sport_t *sport)
972 {
973 	ocs_t *ocs = sport->ocs;
974 	ocs_xport_t *xport = ocs->xport;
975 	ocs_vport_spec_t *vport;
976 
977 	ocs_device_lock(ocs);
978 	ocs_list_foreach(&xport->vport_list, vport) {
979 		if (vport->sport == sport) {
980 			vport->sport = NULL;
981 			break;
982 		}
983 	}
984 	ocs_device_unlock(ocs);
985 }
986 
987 /**
988  * @ingroup sport_sm
989  * @brief Allocate a new virtual SLI port.
990  *
991  * @par Description
992  * A new sport is created, in response to an external management request.
993  *
994  * @n @b Note: If the WWPN is zero, the firmware will assign the WWNs.
995  *
996  * @param domain Pointer to the domain context.
997  * @param wwpn World wide port name.
998  * @param wwnn World wide node name
999  * @param fc_id Requested port ID (used in fabric emulation mode).
1000  * @param ini TRUE, if port is created as an initiator node.
1001  * @param tgt TRUE, if port is created as a target node.
1002  * @param tgt_data Pointer to target specific data
1003  * @param ini_data Pointer to initiator specific data
1004  * @param restore_vport If TRUE, then the vport will be re-created automatically
1005  *                      on link disruption.
1006  *
1007  * @return Returns 0 on success; or a negative error value on failure.
1008  */
1009 
1010 int32_t
1011 ocs_sport_vport_new(ocs_domain_t *domain, uint64_t wwpn, uint64_t wwnn,
1012 		    uint32_t fc_id, uint8_t ini, uint8_t tgt, void *tgt_data,
1013 		    void *ini_data, uint8_t restore_vport)
1014 {
1015 	ocs_sport_t *sport;
1016 
1017 	if (ini && (domain->ocs->enable_ini == 0)) {
1018 		ocs_log_test(domain->ocs, "driver initiator functionality not enabled\n");
1019 		return -1;
1020 	}
1021 
1022 	if (tgt && (domain->ocs->enable_tgt == 0)) {
1023 		ocs_log_test(domain->ocs, "driver target functionality not enabled\n");
1024 		return -1;
1025 	}
1026 
1027 	/* Create a vport spec if we need to recreate this vport after a link up event */
1028 	if (restore_vport) {
1029 		if (ocs_vport_create_spec(domain->ocs, wwnn, wwpn, fc_id, ini, tgt, tgt_data, ini_data)) {
1030 			ocs_log_test(domain->ocs, "failed to create vport object entry\n");
1031 			return -1;
1032 		}
1033 		return ocs_vport_start(domain);
1034 	}
1035 
1036 	/* Allocate a sport */
1037 	sport = ocs_sport_alloc(domain, wwpn, wwnn, fc_id, ini, tgt);
1038 
1039 	if (sport == NULL) {
1040 		return -1;
1041 	}
1042 
1043 	sport->is_vport = 1;
1044 	sport->tgt_data = tgt_data;
1045 	sport->ini_data = ini_data;
1046 
1047 	/* Transition to vport_init */
1048 	ocs_sm_transition(&sport->sm, __ocs_sport_vport_init, NULL);
1049 
1050 	return 0;
1051 }
1052 
1053 int32_t
1054 ocs_sport_vport_alloc(ocs_domain_t *domain, ocs_vport_spec_t *vport)
1055 {
1056 	ocs_sport_t *sport = NULL;
1057 
1058 	if (domain == NULL) {
1059 		return (0);
1060 	}
1061 
1062 	ocs_assert((vport->sport == NULL), -1);
1063 
1064 	/* Allocate a sport */
1065 	vport->sport = sport = ocs_sport_alloc(domain, vport->wwpn, vport->wwnn, UINT32_MAX, vport->enable_ini, vport->enable_tgt);
1066 
1067 	if (sport == NULL) {
1068 		return -1;
1069 	}
1070 
1071 	sport->is_vport = 1;
1072 	sport->tgt_data = vport->tgt_data;
1073 	sport->ini_data = vport->tgt_data;
1074 
1075 	/* Transition to vport_init */
1076 	ocs_sm_transition(&sport->sm, __ocs_sport_vport_init, NULL);
1077 
1078 	return (0);
1079 }
1080 
1081 /**
1082  * @ingroup sport_sm
1083  * @brief Remove a previously-allocated virtual port.
1084  *
1085  * @par Description
1086  * A previously-allocated virtual port is removed by posting the shutdown event to the
1087  * sport with a matching WWN.
1088  *
1089  * @param ocs Pointer to the device object.
1090  * @param domain Pointer to the domain structure (may be NULL).
1091  * @param wwpn World wide port name of the port to delete (host endian).
1092  * @param wwnn World wide node name of the port to delete (host endian).
1093  *
1094  * @return Returns 0 on success, or a negative error value on failure.
1095  */
1096 
1097 int32_t ocs_sport_vport_del(ocs_t *ocs, ocs_domain_t *domain, uint64_t wwpn, uint64_t wwnn)
1098 {
1099 	ocs_xport_t *xport = ocs->xport;
1100 	ocs_sport_t *sport;
1101 	int found = 0;
1102 	ocs_vport_spec_t *vport;
1103 	ocs_vport_spec_t *next;
1104 	uint32_t instance;
1105 
1106 	/* If no domain is given, use instance 0, otherwise use domain instance */
1107 	if (domain == NULL) {
1108 		instance = 0;
1109 	} else {
1110 		instance = domain->instance_index;
1111 	}
1112 
1113 	/* walk the ocs_vport_list and remove from there */
1114 
1115 	ocs_device_lock(ocs);
1116 		ocs_list_foreach_safe(&xport->vport_list, vport, next) {
1117 			if ((vport->domain_instance == instance) &&
1118 				(vport->wwpn == wwpn) && (vport->wwnn == wwnn)) {
1119 				vport->sport = NULL;
1120 				break;
1121 			}
1122 		}
1123 	ocs_device_unlock(ocs);
1124 
1125 	if (domain == NULL) {
1126 		/* No domain means no sport to look for */
1127 		return 0;
1128 	}
1129 
1130 	ocs_domain_lock(domain);
1131 		ocs_list_foreach(&domain->sport_list, sport) {
1132 			if ((sport->wwpn == wwpn) && (sport->wwnn == wwnn)) {
1133 				found = 1;
1134 				break;
1135 			}
1136 		}
1137 		if (found) {
1138 			/* Shutdown this SPORT */
1139 			ocs_sm_post_event(&sport->sm, OCS_EVT_SHUTDOWN, NULL);
1140 		}
1141 	ocs_domain_unlock(domain);
1142 	return 0;
1143 }
1144 
1145 /**
1146  * @brief Force free all saved vports.
1147  *
1148  * @par Description
1149  * Delete all device vports.
1150  *
1151  * @param ocs Pointer to the device object.
1152  *
1153  * @return None.
1154  */
1155 
1156 void
1157 ocs_vport_del_all(ocs_t *ocs)
1158 {
1159 	ocs_xport_t *xport = ocs->xport;
1160 	ocs_vport_spec_t *vport;
1161 	ocs_vport_spec_t *next;
1162 
1163 	ocs_device_lock(ocs);
1164 		ocs_list_foreach_safe(&xport->vport_list, vport, next) {
1165 			ocs_list_remove(&xport->vport_list, vport);
1166 			ocs_free(ocs, vport, sizeof(*vport));
1167 		}
1168 	ocs_device_unlock(ocs);
1169 }
1170 
1171 /**
1172  * @ingroup sport_sm
1173  * @brief Generate a SLI port ddump.
1174  *
1175  * @par Description
1176  * Generates the SLI port ddump data.
1177  *
1178  * @param textbuf Pointer to the text buffer.
1179  * @param sport Pointer to the SLI-4 port.
1180  *
1181  * @return Returns 0 on success, or a negative value on failure.
1182  */
1183 
1184 int
1185 ocs_ddump_sport(ocs_textbuf_t *textbuf, ocs_sli_port_t *sport)
1186 {
1187 	ocs_node_t *node;
1188 	ocs_node_group_dir_t *node_group_dir;
1189 	int retval = 0;
1190 
1191 	ocs_ddump_section(textbuf, "sport", sport->instance_index);
1192 	ocs_ddump_value(textbuf, "display_name", "%s", sport->display_name);
1193 
1194 	ocs_ddump_value(textbuf, "is_vport", "%d", sport->is_vport);
1195 	ocs_ddump_value(textbuf, "enable_ini", "%d", sport->enable_ini);
1196 	ocs_ddump_value(textbuf, "enable_tgt", "%d", sport->enable_tgt);
1197 	ocs_ddump_value(textbuf, "shutting_down", "%d", sport->shutting_down);
1198 	ocs_ddump_value(textbuf, "topology", "%d", sport->topology);
1199 	ocs_ddump_value(textbuf, "p2p_winner", "%d", sport->p2p_winner);
1200 	ocs_ddump_value(textbuf, "p2p_port_id", "%06x", sport->p2p_port_id);
1201 	ocs_ddump_value(textbuf, "p2p_remote_port_id", "%06x", sport->p2p_remote_port_id);
1202 	ocs_ddump_value(textbuf, "wwpn", "%016llx", (unsigned long long)sport->wwpn);
1203 	ocs_ddump_value(textbuf, "wwnn", "%016llx", (unsigned long long)sport->wwnn);
1204 	/*TODO: service_params */
1205 
1206 	ocs_ddump_value(textbuf, "indicator", "x%x", sport->indicator);
1207 	ocs_ddump_value(textbuf, "fc_id", "x%06x", sport->fc_id);
1208 	ocs_ddump_value(textbuf, "index", "%d", sport->index);
1209 
1210 	ocs_display_sparams(NULL, "sport_sparams", 1, textbuf, sport->service_params+4);
1211 
1212 	/* HLM dump */
1213 	ocs_ddump_section(textbuf, "hlm", sport->instance_index);
1214 	ocs_lock(&sport->node_group_lock);
1215 		ocs_list_foreach(&sport->node_group_dir_list, node_group_dir) {
1216 			ocs_remote_node_group_t *remote_node_group;
1217 
1218 			ocs_ddump_section(textbuf, "node_group_dir", node_group_dir->instance_index);
1219 
1220 			ocs_ddump_value(textbuf, "node_group_list_count", "%d", node_group_dir->node_group_list_count);
1221 			ocs_ddump_value(textbuf, "next_idx", "%d", node_group_dir->next_idx);
1222 			ocs_list_foreach(&node_group_dir->node_group_list, remote_node_group) {
1223 				ocs_ddump_section(textbuf, "node_group", remote_node_group->instance_index);
1224 				ocs_ddump_value(textbuf, "indicator", "x%x", remote_node_group->indicator);
1225 				ocs_ddump_value(textbuf, "index", "x%x", remote_node_group->index);
1226 				ocs_ddump_value(textbuf, "instance_index", "x%x", remote_node_group->instance_index);
1227 				ocs_ddump_endsection(textbuf, "node_group", 0);
1228 			}
1229 			ocs_ddump_endsection(textbuf, "node_group_dir", 0);
1230 		}
1231 	ocs_unlock(&sport->node_group_lock);
1232 	ocs_ddump_endsection(textbuf, "hlm", sport->instance_index);
1233 
1234 	ocs_scsi_ini_ddump(textbuf, OCS_SCSI_DDUMP_SPORT, sport);
1235 	ocs_scsi_tgt_ddump(textbuf, OCS_SCSI_DDUMP_SPORT, sport);
1236 
1237 	/* Dump all the nodes */
1238 	if (ocs_sport_lock_try(sport) != TRUE) {
1239 		/* Didn't get lock */
1240 		return -1;
1241 	}
1242 		/* Here the sport lock is held */
1243 		ocs_list_foreach(&sport->node_list, node) {
1244 			retval = ocs_ddump_node(textbuf, node);
1245 			if (retval != 0) {
1246 				break;
1247 			}
1248 		}
1249 	ocs_sport_unlock(sport);
1250 
1251 	ocs_ddump_endsection(textbuf, "sport", sport->index);
1252 
1253 	return retval;
1254 }
1255 
1256 
1257 void
1258 ocs_mgmt_sport_list(ocs_textbuf_t *textbuf, void *object)
1259 {
1260 	ocs_node_t *node;
1261 	ocs_sport_t *sport = (ocs_sport_t *)object;
1262 
1263 	ocs_mgmt_start_section(textbuf, "sport", sport->instance_index);
1264 
1265 	/* Add my status values to textbuf */
1266 	ocs_mgmt_emit_property_name(textbuf, MGMT_MODE_RD, "indicator");
1267 	ocs_mgmt_emit_property_name(textbuf, MGMT_MODE_RD, "fc_id");
1268 	ocs_mgmt_emit_property_name(textbuf, MGMT_MODE_RD, "index");
1269 	ocs_mgmt_emit_property_name(textbuf, MGMT_MODE_RD, "display_name");
1270 	ocs_mgmt_emit_property_name(textbuf, MGMT_MODE_RD, "is_vport");
1271 	ocs_mgmt_emit_property_name(textbuf, MGMT_MODE_RD, "enable_ini");
1272 	ocs_mgmt_emit_property_name(textbuf, MGMT_MODE_RD, "enable_tgt");
1273 	ocs_mgmt_emit_property_name(textbuf, MGMT_MODE_RD, "p2p");
1274 	ocs_mgmt_emit_property_name(textbuf, MGMT_MODE_RD, "p2p_winner");
1275 	ocs_mgmt_emit_property_name(textbuf, MGMT_MODE_RD, "p2p_port_id");
1276 	ocs_mgmt_emit_property_name(textbuf, MGMT_MODE_RD, "p2p_remote_port_id");
1277 	ocs_mgmt_emit_property_name(textbuf, MGMT_MODE_RD, "wwpn");
1278 	ocs_mgmt_emit_property_name(textbuf, MGMT_MODE_RD, "wwnn");
1279 
1280 	if (ocs_sport_lock_try(sport) == TRUE) {
1281 
1282 		/* If we get here, then we are holding the sport lock */
1283 		ocs_list_foreach(&sport->node_list, node) {
1284 			if ((node->mgmt_functions) && (node->mgmt_functions->get_list_handler)) {
1285 				node->mgmt_functions->get_list_handler(textbuf, node);
1286 			}
1287 
1288 		}
1289 		ocs_sport_unlock(sport);
1290 	}
1291 
1292 	ocs_mgmt_end_section(textbuf, "sport", sport->instance_index);
1293 }
1294 
1295 int
1296 ocs_mgmt_sport_get(ocs_textbuf_t *textbuf, char *parent, char *name, void *object)
1297 {
1298 	ocs_node_t *node;
1299 	ocs_sport_t *sport = (ocs_sport_t *)object;
1300 	char qualifier[80];
1301 	int retval = -1;
1302 
1303 	ocs_mgmt_start_section(textbuf, "sport", sport->instance_index);
1304 
1305 	snprintf(qualifier, sizeof(qualifier), "%s/sport[%d]", parent, sport->instance_index);
1306 
1307 	/* If it doesn't start with my qualifier I don't know what to do with it */
1308 	if (ocs_strncmp(name, qualifier, strlen(qualifier)) == 0) {
1309 		char *unqualified_name = name + strlen(qualifier) +1;
1310 
1311 		/* See if it's a value I can supply */
1312 		if (ocs_strcmp(unqualified_name, "indicator") == 0) {
1313 			ocs_mgmt_emit_int(textbuf, MGMT_MODE_RD, "indicator", "0x%x", sport->indicator);
1314 			retval = 0;
1315 		} else if (ocs_strcmp(unqualified_name, "fc_id") == 0) {
1316 			ocs_mgmt_emit_int(textbuf, MGMT_MODE_RD, "fc_id", "0x%06x", sport->fc_id);
1317 			retval = 0;
1318 		} else if (ocs_strcmp(unqualified_name, "index") == 0) {
1319 			ocs_mgmt_emit_int(textbuf, MGMT_MODE_RD, "index", "%d", sport->index);
1320 			retval = 0;
1321 		} else if (ocs_strcmp(unqualified_name, "display_name") == 0) {
1322 			ocs_mgmt_emit_string(textbuf, MGMT_MODE_RD, "display_name", sport->display_name);
1323 			retval = 0;
1324 		} else if (ocs_strcmp(unqualified_name, "is_vport") == 0) {
1325 			ocs_mgmt_emit_boolean(textbuf, MGMT_MODE_RD, "is_vport",  sport->is_vport);
1326 			retval = 0;
1327 		} else if (ocs_strcmp(unqualified_name, "enable_ini") == 0) {
1328 			ocs_mgmt_emit_boolean(textbuf, MGMT_MODE_RD, "enable_ini",  sport->enable_ini);
1329 			retval = 0;
1330 		} else if (ocs_strcmp(unqualified_name, "enable_tgt") == 0) {
1331 			ocs_mgmt_emit_boolean(textbuf, MGMT_MODE_RD, "enable_tgt",  sport->enable_tgt);
1332 			retval = 0;
1333 		} else if (ocs_strcmp(unqualified_name, "p2p_winner") == 0) {
1334 			ocs_mgmt_emit_boolean(textbuf, MGMT_MODE_RD, "p2p_winner",  sport->p2p_winner);
1335 			retval = 0;
1336 		} else if (ocs_strcmp(unqualified_name, "p2p_port_id") == 0) {
1337 			ocs_mgmt_emit_int(textbuf, MGMT_MODE_RD, "p2p_port_id", "0x%06x", sport->p2p_port_id);
1338 			retval = 0;
1339 		} else if (ocs_strcmp(unqualified_name, "p2p_remote_port_id") == 0) {
1340 			ocs_mgmt_emit_int(textbuf, MGMT_MODE_RD, "p2p_remote_port_id", "0x%06x", sport->p2p_remote_port_id);
1341 			retval = 0;
1342 		} else if (ocs_strcmp(unqualified_name, "wwpn") == 0) {
1343 			ocs_mgmt_emit_int(textbuf, MGMT_MODE_RD, "wwpn", "0x%016llx", (unsigned long long)sport->wwpn);
1344 			retval = 0;
1345 		} else if (ocs_strcmp(unqualified_name, "wwnn") == 0) {
1346 			ocs_mgmt_emit_int(textbuf, MGMT_MODE_RD, "wwnn", "0x%016llx", (unsigned long long)sport->wwnn);
1347 			retval = 0;
1348 		} else {
1349 			/* If I didn't know the value of this status pass the request to each of my children */
1350 			ocs_sport_lock(sport);
1351 				ocs_list_foreach(&sport->node_list, node) {
1352 					if ((node->mgmt_functions) && (node->mgmt_functions->get_handler)) {
1353 						retval = node->mgmt_functions->get_handler(textbuf, qualifier, name, node);
1354 					}
1355 
1356 					if (retval == 0) {
1357 						break;
1358 					}
1359 				}
1360 			ocs_sport_unlock(sport);
1361 		}
1362 	}
1363 
1364 	ocs_mgmt_end_section(textbuf, "sport", sport->instance_index);
1365 
1366 	return retval;
1367 }
1368 
1369 void
1370 ocs_mgmt_sport_get_all(ocs_textbuf_t *textbuf, void *object)
1371 {
1372 	ocs_node_t *node;
1373 	ocs_sport_t *sport = (ocs_sport_t *)object;
1374 
1375 	ocs_mgmt_start_section(textbuf, "sport", sport->instance_index);
1376 
1377 	ocs_mgmt_emit_int(textbuf, MGMT_MODE_RD, "indicator", "0x%x", sport->indicator);
1378 	ocs_mgmt_emit_int(textbuf, MGMT_MODE_RD, "fc_id", "0x%06x", sport->fc_id);
1379 	ocs_mgmt_emit_int(textbuf, MGMT_MODE_RD, "index", "%d", sport->index);
1380 	ocs_mgmt_emit_string(textbuf, MGMT_MODE_RD, "display_name", sport->display_name);
1381 	ocs_mgmt_emit_boolean(textbuf, MGMT_MODE_RD, "is_vport",  sport->is_vport);
1382 	ocs_mgmt_emit_boolean(textbuf, MGMT_MODE_RD, "enable_ini",  sport->enable_ini);
1383 	ocs_mgmt_emit_boolean(textbuf, MGMT_MODE_RD, "enable_tgt",  sport->enable_tgt);
1384 	ocs_mgmt_emit_boolean(textbuf, MGMT_MODE_RD, "p2p_winner",  sport->p2p_winner);
1385 	ocs_mgmt_emit_int(textbuf, MGMT_MODE_RD, "p2p_port_id", "0x%06x", sport->p2p_port_id);
1386 	ocs_mgmt_emit_int(textbuf, MGMT_MODE_RD, "p2p_remote_port_id", "0x%06x", sport->p2p_remote_port_id);
1387 	ocs_mgmt_emit_int(textbuf, MGMT_MODE_RD, "wwpn", "0x%016llx" , (unsigned long long)sport->wwpn);
1388 	ocs_mgmt_emit_int(textbuf, MGMT_MODE_RD, "wwnn", "0x%016llx", (unsigned long long)sport->wwnn);
1389 
1390 	ocs_sport_lock(sport);
1391 	ocs_list_foreach(&sport->node_list, node) {
1392 		if ((node->mgmt_functions) && (node->mgmt_functions->get_all_handler)) {
1393 			node->mgmt_functions->get_all_handler(textbuf, node);
1394 		}
1395 	}
1396 	ocs_sport_unlock(sport);
1397 
1398 	ocs_mgmt_end_section(textbuf, "sport", sport->instance_index);
1399 }
1400 
1401 int
1402 ocs_mgmt_sport_set(char *parent, char *name, char *value, void *object)
1403 {
1404 	ocs_node_t *node;
1405 	ocs_sport_t *sport = (ocs_sport_t *)object;
1406 	char qualifier[80];
1407 	int retval = -1;
1408 
1409 	snprintf(qualifier, sizeof(qualifier), "%s/sport[%d]", parent, sport->instance_index);
1410 
1411 	/* If it doesn't start with my qualifier I don't know what to do with it */
1412 	if (ocs_strncmp(name, qualifier, strlen(qualifier)) == 0) {
1413 		/* The sport has no settable values.  Pass the request to each node. */
1414 
1415 		ocs_sport_lock(sport);
1416 		ocs_list_foreach(&sport->node_list, node) {
1417 			if ((node->mgmt_functions) && (node->mgmt_functions->set_handler)) {
1418 				retval = node->mgmt_functions->set_handler(qualifier, name, value, node);
1419 			}
1420 			if (retval == 0) {
1421 				break;
1422 			}
1423 		}
1424 		ocs_sport_unlock(sport);
1425 	}
1426 
1427 	return retval;
1428 }
1429 
1430 
1431 int
1432 ocs_mgmt_sport_exec(char *parent, char *action, void *arg_in, uint32_t arg_in_length,
1433 		    void *arg_out, uint32_t arg_out_length, void *object)
1434 {
1435 	ocs_node_t *node;
1436 	ocs_sport_t *sport = (ocs_sport_t *)object;
1437 	char qualifier[80];
1438 	int retval = -1;
1439 
1440 	snprintf(qualifier, sizeof(qualifier), "%s.sport%d", parent, sport->instance_index);
1441 
1442 	/* If it doesn't start with my qualifier I don't know what to do with it */
1443 	if (ocs_strncmp(action, qualifier, strlen(qualifier)) == 0) {
1444 
1445 		/* See if it's an action I can perform */
1446 
1447 		/* if (ocs_strcmp ....
1448 		 * {
1449 		 * } else
1450 		 */
1451 
1452 		{
1453 			/* If I didn't know how to do this action pass the request to each of my children */
1454 			ocs_sport_lock(sport);
1455 				ocs_list_foreach(&sport->node_list, node) {
1456 					if ((node->mgmt_functions) && (node->mgmt_functions->exec_handler)) {
1457 						retval = node->mgmt_functions->exec_handler(qualifier, action, arg_in, arg_in_length,
1458 											    arg_out, arg_out_length, node);
1459 					}
1460 
1461 					if (retval == 0) {
1462 						break;
1463 					}
1464 
1465 				}
1466 			ocs_sport_unlock(sport);
1467 		}
1468 	}
1469 
1470 	return retval;
1471 }
1472 
1473 /**
1474  * @brief Save the virtual port's parameters.
1475  *
1476  * @par Description
1477  * The information required to restore a virtual port is saved.
1478  *
1479  * @param sport Pointer to the sport context.
1480  *
1481  * @return None.
1482  */
1483 
1484 static void
1485 ocs_vport_update_spec(ocs_sport_t *sport)
1486 {
1487 	ocs_t *ocs = sport->ocs;
1488 	ocs_xport_t *xport = ocs->xport;
1489 	ocs_vport_spec_t *vport;
1490 
1491 	ocs_device_lock(ocs);
1492 	ocs_list_foreach(&xport->vport_list, vport) {
1493 		if (vport->sport == sport) {
1494 			vport->wwnn = sport->wwnn;
1495 			vport->wwpn = sport->wwpn;
1496 			vport->tgt_data = sport->tgt_data;
1497 			vport->ini_data = sport->ini_data;
1498 			break;
1499 		}
1500 	}
1501 	ocs_device_unlock(ocs);
1502 }
1503 
1504 /**
1505  * @brief Create a saved vport entry.
1506  *
1507  * A saved vport entry is added to the vport list, which is restored following
1508  * a link up. This function is used to allow vports to be created the first time
1509  * the link comes up without having to go through the ioctl() API.
1510  *
1511  * @param ocs Pointer to device context.
1512  * @param wwnn World wide node name (may be zero for auto-select).
1513  * @param wwpn World wide port name (may be zero for auto-select).
1514  * @param fc_id Requested port ID (used in fabric emulation mode).
1515  * @param enable_ini TRUE if vport is to be an initiator port.
1516  * @param enable_tgt TRUE if vport is to be a target port.
1517  * @param tgt_data Pointer to target specific data.
1518  * @param ini_data Pointer to initiator specific data.
1519  *
1520  * @return None.
1521  */
1522 
1523 int8_t
1524 ocs_vport_create_spec(ocs_t *ocs, uint64_t wwnn, uint64_t wwpn, uint32_t fc_id, uint32_t enable_ini, uint32_t enable_tgt, void *tgt_data, void *ini_data)
1525 {
1526 	ocs_xport_t *xport = ocs->xport;
1527 	ocs_vport_spec_t *vport;
1528 
1529 	/* walk the ocs_vport_list and return failure if a valid(vport with non zero WWPN and WWNN) vport entry
1530 	   is already created */
1531 	ocs_list_foreach(&xport->vport_list, vport) {
1532 		if ((wwpn && (vport->wwpn == wwpn)) && (wwnn && (vport->wwnn == wwnn))) {
1533 			ocs_log_test(ocs, "Failed: VPORT %016llx  %016llx already allocated\n",
1534 				     (unsigned long long)wwnn, (unsigned long long)wwpn);
1535 			return -1;
1536 		}
1537 	}
1538 
1539 	vport = ocs_malloc(ocs, sizeof(*vport), OCS_M_ZERO | OCS_M_NOWAIT);
1540 	if (vport == NULL) {
1541 		ocs_log_err(ocs, "ocs_malloc failed\n");
1542 		return -1;
1543 	}
1544 
1545 	vport->wwnn = wwnn;
1546 	vport->wwpn = wwpn;
1547 	vport->fc_id = fc_id;
1548 	vport->domain_instance = 0;	/*TODO: may need to change this */
1549 	vport->enable_tgt = enable_tgt;
1550 	vport->enable_ini = enable_ini;
1551 	vport->tgt_data = tgt_data;
1552 	vport->ini_data = ini_data;
1553 
1554 	ocs_device_lock(ocs);
1555 		ocs_list_add_tail(&xport->vport_list, vport);
1556 	ocs_device_unlock(ocs);
1557 	return 0;
1558 }
1559 
1560 /* node group api */
1561 
1562 /**
1563  * @brief Perform the AND operation on source vectors.
1564  *
1565  * @par Description
1566  * Performs an AND operation on the 8-bit values in source vectors @c b and @c c.
1567  * The resulting value is stored in @c a.
1568  *
1569  * @param a Destination-byte vector.
1570  * @param b Source-byte vector.
1571  * @param c Source-byte vector.
1572  * @param n Byte count.
1573  *
1574  * @return None.
1575  */
1576 
1577 static void
1578 and8(uint8_t *a, uint8_t *b, uint8_t *c, uint32_t n)
1579 {
1580 	uint32_t i;
1581 
1582 	for (i = 0; i < n; i ++) {
1583 		*a = *b & *c;
1584 		a++;
1585 		b++;
1586 		c++;
1587 	}
1588 }
1589 
1590 /**
1591  * @brief Service parameters mask data.
1592  */
1593 static fc_sparms_t sparms_cmp_mask = {
1594 	0,			/*uint32_t	command_code: 8, */
1595 	0,			/*		resv1: 24; */
1596 	{~0, ~0, ~0, ~0},	/* uint32_t	common_service_parameters[4]; */
1597 	0,			/* uint32_t	port_name_hi; */
1598 	0,			/* uint32_t	port_name_lo; */
1599 	0,			/* uint32_t	node_name_hi; */
1600 	0,			/* uint32_t	node_name_lo; */
1601 	{~0, ~0, ~0, ~0},	/* uint32_t	class1_service_parameters[4]; */
1602 	{~0, ~0, ~0, ~0},	/* uint32_t	class2_service_parameters[4]; */
1603 	{~0, ~0, ~0, ~0},	/* uint32_t	class3_service_parameters[4]; */
1604 	{~0, ~0, ~0, ~0},	/* uint32_t	class4_service_parameters[4]; */
1605 	{~0, ~0, ~0, ~0}};	/* uint32_t	vendor_version_level[4]; */
1606 
1607 /**
1608  * @brief Compare service parameters.
1609  *
1610  * @par Description
1611  * Returns 0 if the two service parameters are the same, excluding the port/node name
1612  * elements.
1613  *
1614  * @param sp1 Pointer to service parameters 1.
1615  * @param sp2 Pointer to service parameters 2.
1616  *
1617  * @return Returns 0 if parameters match; otherwise, returns a positive or negative value,
1618  * depending on the arithmetic magnitude of the first mismatching byte.
1619  */
1620 
1621 int
1622 ocs_sparm_cmp(uint8_t *sp1, uint8_t *sp2)
1623 {
1624 	int i;
1625 	int v;
1626 	uint8_t *sp3 = (uint8_t*) &sparms_cmp_mask;
1627 
1628 	for (i = 0; i < OCS_SERVICE_PARMS_LENGTH; i ++) {
1629 		v = ((int)(sp1[i] & sp3[i])) - ((int)(sp2[i] & sp3[i]));
1630 		if (v) {
1631 			break;
1632 		}
1633 	}
1634 	return v;
1635 }
1636 
1637 /**
1638  * @brief Allocate a node group directory entry.
1639  *
1640  * @par Description
1641  * A node group directory entry is allocated, initialized, and added to the sport's
1642  * node group directory list.
1643  *
1644  * @param sport Pointer to the sport object.
1645  * @param sparms Pointer to the service parameters.
1646  *
1647  * @return Returns a pointer to the allocated ocs_node_group_dir_t; or NULL.
1648  */
1649 
1650 ocs_node_group_dir_t *
1651 ocs_node_group_dir_alloc(ocs_sport_t *sport, uint8_t *sparms)
1652 {
1653 	ocs_node_group_dir_t *node_group_dir;
1654 
1655 	node_group_dir = ocs_malloc(sport->ocs, sizeof(*node_group_dir), OCS_M_ZERO | OCS_M_NOWAIT);
1656 	if (node_group_dir != NULL) {
1657 		node_group_dir->sport = sport;
1658 
1659 		ocs_lock(&sport->node_group_lock);
1660 			node_group_dir->instance_index = sport->node_group_dir_next_instance++;
1661 			and8(node_group_dir->service_params, sparms, (uint8_t*)&sparms_cmp_mask, OCS_SERVICE_PARMS_LENGTH);
1662 			ocs_list_init(&node_group_dir->node_group_list, ocs_remote_node_group_t, link);
1663 
1664 			node_group_dir->node_group_list_count = 0;
1665 			node_group_dir->next_idx = 0;
1666 			ocs_list_add_tail(&sport->node_group_dir_list, node_group_dir);
1667 		ocs_unlock(&sport->node_group_lock);
1668 
1669 		ocs_log_debug(sport->ocs, "[%s] [%d] allocating node group directory\n", sport->display_name,
1670 			node_group_dir->instance_index);
1671 	}
1672 	return node_group_dir;
1673 }
1674 
1675 /**
1676  * @brief Free a node group directory entry.
1677  *
1678  * @par Description
1679  * The node group directory entry @c node_group_dir is removed
1680  * from the sport's node group directory list and freed.
1681  *
1682  * @param node_group_dir Pointer to the node group directory entry.
1683  *
1684  * @return None.
1685  */
1686 
1687 void
1688 ocs_node_group_dir_free(ocs_node_group_dir_t *node_group_dir)
1689 {
1690 	ocs_sport_t *sport;
1691 	if (node_group_dir != NULL) {
1692 		sport = node_group_dir->sport;
1693 		ocs_log_debug(sport->ocs, "[%s] [%d] freeing node group directory\n", sport->display_name,
1694 			node_group_dir->instance_index);
1695 		ocs_lock(&sport->node_group_lock);
1696 			if (!ocs_list_empty(&node_group_dir->node_group_list)) {
1697 				ocs_log_test(sport->ocs, "[%s] WARNING: node group list not empty\n", sport->display_name);
1698 			}
1699 			ocs_list_remove(&sport->node_group_dir_list, node_group_dir);
1700 		ocs_unlock(&sport->node_group_lock);
1701 		ocs_free(sport->ocs, node_group_dir, sizeof(*node_group_dir));
1702 	}
1703 }
1704 
1705 /**
1706  * @brief Find a matching node group directory entry.
1707  *
1708  * @par Description
1709  * The sport's node group directory list is searched for a matching set of
1710  * service parameters. The first matching entry is returned; otherwise
1711  * NULL is returned.
1712  *
1713  * @param sport Pointer to the sport object.
1714  * @param sparms Pointer to the sparams to match.
1715  *
1716  * @return Returns a pointer to the first matching entry found; or NULL.
1717  */
1718 
1719 ocs_node_group_dir_t *
1720 ocs_node_group_dir_find(ocs_sport_t *sport, uint8_t *sparms)
1721 {
1722 	ocs_node_group_dir_t *node_dir = NULL;
1723 
1724 	ocs_lock(&sport->node_group_lock);
1725 		ocs_list_foreach(&sport->node_group_dir_list, node_dir) {
1726 			if (ocs_sparm_cmp(sparms, node_dir->service_params) == 0) {
1727 				ocs_unlock(&sport->node_group_lock);
1728 				return node_dir;
1729 			}
1730 		}
1731 	ocs_unlock(&sport->node_group_lock);
1732 	return NULL;
1733 }
1734 
1735 /**
1736  * @brief Allocate a remote node group object.
1737  *
1738  * @par Description
1739  * A remote node group object is allocated, initialized, and placed on the node group
1740  * list of @c node_group_dir. The HW remote node group @b alloc function is called.
1741  *
1742  * @param node_group_dir Pointer to the node group directory.
1743  *
1744  * @return Returns a pointer to the allocated remote node group object; or NULL.
1745  */
1746 
1747 ocs_remote_node_group_t *
1748 ocs_remote_node_group_alloc(ocs_node_group_dir_t *node_group_dir)
1749 {
1750 	ocs_t *ocs;
1751 	ocs_sport_t *sport;
1752 	ocs_remote_node_group_t *node_group;
1753 	ocs_hw_rtn_e hrc;
1754 
1755 	ocs_assert(node_group_dir, NULL);
1756 	ocs_assert(node_group_dir->sport, NULL);
1757 	ocs_assert(node_group_dir->sport->ocs, NULL);
1758 
1759 	sport = node_group_dir->sport;
1760 	ocs = sport->ocs;
1761 
1762 
1763 	node_group = ocs_malloc(ocs, sizeof(*node_group), OCS_M_ZERO | OCS_M_NOWAIT);
1764 	if (node_group != NULL) {
1765 
1766 		/* set pointer to node group directory */
1767 		node_group->node_group_dir = node_group_dir;
1768 
1769 		ocs_lock(&node_group_dir->sport->node_group_lock);
1770 			node_group->instance_index = sport->node_group_next_instance++;
1771 		ocs_unlock(&node_group_dir->sport->node_group_lock);
1772 
1773 		/* invoke HW node group inialization */
1774 		hrc = ocs_hw_node_group_alloc(&ocs->hw, node_group);
1775 		if (hrc != OCS_HW_RTN_SUCCESS) {
1776 			ocs_log_err(ocs, "ocs_hw_node_group_alloc() failed: %d\n", hrc);
1777 			ocs_free(ocs, node_group, sizeof(*node_group));
1778 			return NULL;
1779 		}
1780 
1781 		ocs_log_debug(ocs, "[%s] [%d] indicator x%03x allocating node group\n", sport->display_name,
1782 			node_group->indicator, node_group->instance_index);
1783 
1784 			/* add to the node group directory entry node group list */
1785 		ocs_lock(&node_group_dir->sport->node_group_lock);
1786 			ocs_list_add_tail(&node_group_dir->node_group_list, node_group);
1787 			node_group_dir->node_group_list_count ++;
1788 		ocs_unlock(&node_group_dir->sport->node_group_lock);
1789 	}
1790 	return node_group;
1791 }
1792 
1793 /**
1794  * @brief Free a remote node group object.
1795  *
1796  * @par Description
1797  * The remote node group object @c node_group is removed from its
1798  * node group directory entry and freed.
1799  *
1800  * @param node_group Pointer to the remote node group object.
1801  *
1802  * @return None.
1803  */
1804 
1805 void
1806 ocs_remote_node_group_free(ocs_remote_node_group_t *node_group)
1807 {
1808 	ocs_sport_t *sport;
1809 	ocs_node_group_dir_t *node_group_dir;
1810 
1811 	if (node_group != NULL) {
1812 
1813 		ocs_assert(node_group->node_group_dir);
1814 		ocs_assert(node_group->node_group_dir->sport);
1815 		ocs_assert(node_group->node_group_dir->sport->ocs);
1816 
1817 		node_group_dir = node_group->node_group_dir;
1818 		sport = node_group_dir->sport;
1819 
1820 		ocs_log_debug(sport->ocs, "[%s] [%d] freeing node group\n", sport->display_name, node_group->instance_index);
1821 
1822 		/* Remove from node group directory node group list */
1823 		ocs_lock(&sport->node_group_lock);
1824 			ocs_list_remove(&node_group_dir->node_group_list, node_group);
1825 			node_group_dir->node_group_list_count --;
1826 		/* TODO: note that we're going to have the node_group_dir entry persist forever ... we could delete it if
1827 		 * the group_list_count goes to zero (or the linked list is empty */
1828 		ocs_unlock(&sport->node_group_lock);
1829 		ocs_free(sport->ocs, node_group, sizeof(*node_group));
1830 	}
1831 }
1832 
1833 /**
1834  * @brief Initialize a node for high login mode.
1835  *
1836  * @par Description
1837  * The @c node is initialized for high login mode. The following steps are performed:
1838  * 1. The sports node group directory is searched for a matching set of service parameters.
1839  * 2. If a matching set is not found, a node group directory entry is allocated.
1840  * 3. If less than the @c hlm_group_size number of remote node group objects is present in the
1841  *   node group directory, a new remote node group object is allocated and added to the list.
1842  * 4. A remote node group object is selected, and the node is attached to the node group.
1843  *
1844  * @param node Pointer to the node.
1845  *
1846  * @return Returns 0 on success, or a negative error value on failure.
1847  */
1848 
1849 int
1850 ocs_node_group_init(ocs_node_t *node)
1851 {
1852 	ocs_t *ocs;
1853 	ocs_sport_t *sport;
1854 	ocs_node_group_dir_t *node_group_dir;
1855 	ocs_remote_node_group_t *node_group;
1856 	ocs_hw_rtn_e hrc;
1857 
1858 	ocs_assert(node, -1);
1859 	ocs_assert(node->sport, -1);
1860 	ocs_assert(node->ocs, -1);
1861 
1862 	ocs = node->ocs;
1863 	sport = node->sport;
1864 
1865 	ocs_assert(ocs->enable_hlm, -1);
1866 
1867 	/* see if there's a node group directory allocated for this service parameter set */
1868 	node_group_dir = ocs_node_group_dir_find(sport, node->service_params);
1869 	if (node_group_dir == NULL) {
1870 		/* not found, so allocate one */
1871 		node_group_dir = ocs_node_group_dir_alloc(sport, node->service_params);
1872 		if (node_group_dir == NULL) {
1873 			/* node group directory allocation failed ... can't continue, however,
1874 			 * the node will be allocated with a normal (not shared) RPI
1875 			 */
1876 			ocs_log_err(ocs, "ocs_node_group_dir_alloc() failed\n");
1877 			return -1;
1878 		}
1879 	}
1880 
1881 	/* check to see if we've allocated hlm_group_size's worth of node group structures for this
1882 	 * directory entry, if not, then allocate and use a new one, otherwise pick the next one.
1883 	 */
1884 	ocs_lock(&node->sport->node_group_lock);
1885 		if (node_group_dir->node_group_list_count < ocs->hlm_group_size) {
1886 			ocs_unlock(&node->sport->node_group_lock);
1887 				node_group = ocs_remote_node_group_alloc(node_group_dir);
1888 			if (node_group == NULL) {
1889 				ocs_log_err(ocs, "ocs_remote_node_group_alloc() failed\n");
1890 				return -1;
1891 			}
1892 			ocs_lock(&node->sport->node_group_lock);
1893 		} else {
1894 			uint32_t idx = 0;
1895 
1896 			ocs_list_foreach(&node_group_dir->node_group_list, node_group) {
1897 				if (idx >= ocs->hlm_group_size) {
1898 					ocs_log_err(node->ocs, "assertion failed: idx >= ocs->hlm_group_size\n");
1899 					ocs_unlock(&node->sport->node_group_lock);
1900 					return -1;
1901 				}
1902 
1903 				if (idx == node_group_dir->next_idx) {
1904 					break;
1905 				}
1906 				idx ++;
1907 			}
1908 			if (idx == ocs->hlm_group_size) {
1909 				node_group = ocs_list_get_head(&node_group_dir->node_group_list);
1910 			}
1911 			if (++node_group_dir->next_idx >= node_group_dir->node_group_list_count) {
1912 				node_group_dir->next_idx = 0;
1913 			}
1914 		}
1915 	ocs_unlock(&node->sport->node_group_lock);
1916 
1917 	/* Initialize a pointer in the node back to the node group */
1918 	node->node_group = node_group;
1919 
1920 	/* Join this node into the group */
1921 	hrc = ocs_hw_node_group_attach(&ocs->hw, node_group, &node->rnode);
1922 
1923 	return (hrc == OCS_HW_RTN_SUCCESS) ? 0 : -1;
1924 }
1925 
1926 
1927