1 /* 2 * net/tipc/node.h: Include file for TIPC node management routines 3 * 4 * Copyright (c) 2000-2006, Ericsson AB 5 * Copyright (c) 2005, Wind River Systems 6 * All rights reserved. 7 * 8 * Redistribution and use in source and binary forms, with or without 9 * modification, are permitted provided that the following conditions are met: 10 * 11 * 1. Redistributions of source code must retain the above copyright 12 * notice, this list of conditions and the following disclaimer. 13 * 2. Redistributions in binary form must reproduce the above copyright 14 * notice, this list of conditions and the following disclaimer in the 15 * documentation and/or other materials provided with the distribution. 16 * 3. Neither the names of the copyright holders nor the names of its 17 * contributors may be used to endorse or promote products derived from 18 * this software without specific prior written permission. 19 * 20 * Alternatively, this software may be distributed under the terms of the 21 * GNU General Public License ("GPL") version 2 as published by the Free 22 * Software Foundation. 23 * 24 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" 25 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 26 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 27 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE 28 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 29 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 30 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 31 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 32 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 33 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 34 * POSSIBILITY OF SUCH DAMAGE. 35 */ 36 37 #ifndef _TIPC_NODE_H 38 #define _TIPC_NODE_H 39 40 #include "node_subscr.h" 41 #include "addr.h" 42 #include "cluster.h" 43 #include "bearer.h" 44 45 /** 46 * struct node - TIPC node structure 47 * @addr: network address of node 48 * @lock: spinlock governing access to structure 49 * @owner: pointer to cluster that node belongs to 50 * @next: pointer to next node in sorted list of cluster's nodes 51 * @nsub: list of "node down" subscriptions monitoring node 52 * @active_links: pointers to active links to node 53 * @links: pointers to all links to node 54 * @link_cnt: number of links to node 55 * @permit_changeover: non-zero if node has redundant links to this system 56 * @routers: bitmap (used for multicluster communication) 57 * @last_router: (used for multicluster communication) 58 * @bclink: broadcast-related info 59 * @supported: non-zero if node supports TIPC b'cast capability 60 * @acked: sequence # of last outbound b'cast message acknowledged by node 61 * @last_in: sequence # of last in-sequence b'cast message received from node 62 * @gap_after: sequence # of last message not requiring a NAK request 63 * @gap_to: sequence # of last message requiring a NAK request 64 * @nack_sync: counter that determines when NAK requests should be sent 65 * @deferred_head: oldest OOS b'cast message received from node 66 * @deferred_tail: newest OOS b'cast message received from node 67 * @defragm: list of partially reassembled b'cast message fragments from node 68 */ 69 70 struct node { 71 u32 addr; 72 spinlock_t lock; 73 struct cluster *owner; 74 struct node *next; 75 struct list_head nsub; 76 struct link *active_links[2]; 77 struct link *links[MAX_BEARERS]; 78 int link_cnt; 79 int permit_changeover; 80 u32 routers[512/32]; 81 int last_router; 82 struct { 83 int supported; 84 u32 acked; 85 u32 last_in; 86 u32 gap_after; 87 u32 gap_to; 88 u32 nack_sync; 89 struct sk_buff *deferred_head; 90 struct sk_buff *deferred_tail; 91 struct sk_buff *defragm; 92 } bclink; 93 }; 94 95 extern struct node *tipc_nodes; 96 extern u32 tipc_own_tag; 97 98 struct node *tipc_node_create(u32 addr); 99 void tipc_node_delete(struct node *n_ptr); 100 struct node *tipc_node_attach_link(struct link *l_ptr); 101 void tipc_node_detach_link(struct node *n_ptr, struct link *l_ptr); 102 void tipc_node_link_down(struct node *n_ptr, struct link *l_ptr); 103 void tipc_node_link_up(struct node *n_ptr, struct link *l_ptr); 104 int tipc_node_has_active_links(struct node *n_ptr); 105 int tipc_node_has_redundant_links(struct node *n_ptr); 106 u32 tipc_node_select_router(struct node *n_ptr, u32 ref); 107 struct node *tipc_node_select_next_hop(u32 addr, u32 selector); 108 int tipc_node_is_up(struct node *n_ptr); 109 void tipc_node_add_router(struct node *n_ptr, u32 router); 110 void tipc_node_remove_router(struct node *n_ptr, u32 router); 111 struct sk_buff *tipc_node_get_links(const void *req_tlv_area, int req_tlv_space); 112 struct sk_buff *tipc_node_get_nodes(const void *req_tlv_area, int req_tlv_space); 113 114 static inline struct node *tipc_node_find(u32 addr) 115 { 116 if (likely(in_own_cluster(addr))) 117 return tipc_local_nodes[tipc_node(addr)]; 118 else if (tipc_addr_domain_valid(addr)) { 119 struct cluster *c_ptr = tipc_cltr_find(addr); 120 121 if (c_ptr) 122 return c_ptr->nodes[tipc_node(addr)]; 123 } 124 return 0; 125 } 126 127 static inline struct node *tipc_node_select(u32 addr, u32 selector) 128 { 129 if (likely(in_own_cluster(addr))) 130 return tipc_local_nodes[tipc_node(addr)]; 131 return tipc_node_select_next_hop(addr, selector); 132 } 133 134 static inline void tipc_node_lock(struct node *n_ptr) 135 { 136 spin_lock_bh(&n_ptr->lock); 137 } 138 139 static inline void tipc_node_unlock(struct node *n_ptr) 140 { 141 spin_unlock_bh(&n_ptr->lock); 142 } 143 144 #endif 145