xref: /freebsd/sys/netinet/sctp_lock_bsd.h (revision 1e413cf93298b5b97441a21d9a50fdcd0ee9945e)
1 #ifndef __sctp_lock_bsd_h__
2 #define __sctp_lock_bsd_h__
3 /*-
4  * Copyright (c) 2001-2007, by Cisco Systems, Inc. All rights reserved.
5  *
6  * Redistribution and use in source and binary forms, with or without
7  * modification, are permitted provided that the following conditions are met:
8  *
9  * a) Redistributions of source code must retain the above copyright notice,
10  *   this list of conditions and the following disclaimer.
11  *
12  * b) Redistributions in binary form must reproduce the above copyright
13  *    notice, this list of conditions and the following disclaimer in
14  *   the documentation and/or other materials provided with the distribution.
15  *
16  * c) Neither the name of Cisco Systems, Inc. nor the names of its
17  *    contributors may be used to endorse or promote products derived
18  *    from this software without specific prior written permission.
19  *
20  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
21  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
22  * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23  * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
24  * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
30  * THE POSSIBILITY OF SUCH DAMAGE.
31  */
32 
33 /*
34  * General locking concepts: The goal of our locking is to of course provide
35  * consistency and yet minimize overhead. We will attempt to use
36  * non-recursive locks which are supposed to be quite inexpensive. Now in
37  * order to do this the goal is that most functions are not aware of locking.
38  * Once we have a TCB we lock it and unlock when we are through. This means
39  * that the TCB lock is kind-of a "global" lock when working on an
40  * association. Caution must be used when asserting a TCB_LOCK since if we
41  * recurse we deadlock.
42  *
43  * Most other locks (INP and INFO) attempt to localize the locking i.e. we try
44  * to contain the lock and unlock within the function that needs to lock it.
45  * This sometimes mean we do extra locks and unlocks and lose a bit of
46  * efficency, but if the performance statements about non-recursive locks are
47  * true this should not be a problem.  One issue that arises with this only
48  * lock when needed is that if an implicit association setup is done we have
49  * a problem. If at the time I lookup an association I have NULL in the tcb
50  * return, by the time I call to create the association some other processor
51  * could have created it. This is what the CREATE lock on the endpoint.
52  * Places where we will be implicitly creating the association OR just
53  * creating an association (the connect call) will assert the CREATE_INP
54  * lock. This will assure us that during all the lookup of INP and INFO if
55  * another creator is also locking/looking up we can gate the two to
56  * synchronize. So the CREATE_INP lock is also another one we must use
57  * extreme caution in locking to make sure we don't hit a re-entrancy issue.
58  *
59  * For non FreeBSD 5.x we provide a bunch of EMPTY lock macros so we can
60  * blatantly put locks everywhere and they reduce to nothing on
61  * NetBSD/OpenBSD and FreeBSD 4.x
62  *
63  */
64 
65 /*
66  * When working with the global SCTP lists we lock and unlock the INP_INFO
67  * lock. So when we go to lookup an association we will want to do a
68  * SCTP_INP_INFO_RLOCK() and then when we want to add a new association to
69  * the sctppcbinfo list's we will do a SCTP_INP_INFO_WLOCK().
70  */
71 #include <sys/cdefs.h>
72 __FBSDID("$FreeBSD$");
73 
74 
75 extern struct sctp_foo_stuff sctp_logoff[];
76 extern int sctp_logoff_stuff;
77 
78 #define SCTP_IPI_COUNT_INIT()
79 
80 #define SCTP_STATLOG_INIT_LOCK()
81 #define SCTP_STATLOG_LOCK()
82 #define SCTP_STATLOG_UNLOCK()
83 #define SCTP_STATLOG_DESTROY()
84 
85 #define SCTP_INP_INFO_LOCK_DESTROY() do { \
86         if(rw_wowned(sctppcbinfo.ipi_ep_mtx)) { \
87              rw_wunlock(&sctppcbinfo.ipi_ep_mtx); \
88         } \
89         rw_destroy(sctppcbinfo.ipi_ep_mtx); \
90       }  while (0)
91 
92 #define SCTP_INP_INFO_LOCK_INIT() \
93         rw_init(&sctppcbinfo.ipi_ep_mtx, "sctp-info");
94 
95 
96 #define SCTP_INP_INFO_RLOCK()	do { 					\
97              rw_rlock(&sctppcbinfo.ipi_ep_mtx);                         \
98 } while (0)
99 
100 
101 #define SCTP_INP_INFO_WLOCK()	do { 					\
102             rw_wlock(&sctppcbinfo.ipi_ep_mtx);                         \
103 } while (0)
104 
105 
106 #define SCTP_INP_INFO_RUNLOCK()		rw_runlock(&sctppcbinfo.ipi_ep_mtx)
107 #define SCTP_INP_INFO_WUNLOCK()		rw_wunlock(&sctppcbinfo.ipi_ep_mtx)
108 
109 
110 #define SCTP_IPI_ADDR_INIT() \
111         rw_init(&sctppcbinfo.ipi_addr_mtx, "sctp-addr")
112 
113 #define SCTP_IPI_ADDR_DESTROY() do  { \
114         if(rw_wowned(sctppcbinfo.ipi_addr_mtx)) { \
115              rw_wunlock(&sctppcbinfo.ipi_addr_mtx); \
116         } \
117 	rw_destroy(&sctppcbinfo.ipi_addr_mtx) \
118       }  while (0)
119 
120 
121 
122 #define SCTP_IPI_ADDR_RLOCK()	do { 					\
123              rw_rlock(&sctppcbinfo.ipi_addr_mtx);                         \
124 } while (0)
125 
126 #define SCTP_IPI_ADDR_WLOCK()	do { 					\
127              rw_wlock(&sctppcbinfo.ipi_addr_mtx);                         \
128 } while (0)
129 
130 
131 #define SCTP_IPI_ADDR_RUNLOCK()		rw_runlock(&sctppcbinfo.ipi_addr_mtx)
132 #define SCTP_IPI_ADDR_WUNLOCK()		rw_wunlock(&sctppcbinfo.ipi_addr_mtx)
133 
134 
135 #define SCTP_IPI_ITERATOR_WQ_INIT() \
136         mtx_init(&sctppcbinfo.ipi_iterator_wq_mtx, "sctp-it-wq", "sctp_it_wq", MTX_DEF)
137 
138 #define SCTP_IPI_ITERATOR_WQ_DESTROY() \
139 	mtx_destroy(&sctppcbinfo.ipi_iterator_wq_mtx)
140 
141 #define SCTP_IPI_ITERATOR_WQ_LOCK()	do { 					\
142              mtx_lock(&sctppcbinfo.ipi_iterator_wq_mtx);                \
143 } while (0)
144 
145 #define SCTP_IPI_ITERATOR_WQ_UNLOCK()		mtx_unlock(&sctppcbinfo.ipi_iterator_wq_mtx)
146 
147 
148 #define SCTP_IP_PKTLOG_INIT() \
149         mtx_init(&sctppcbinfo.ipi_pktlog_mtx, "sctp-pktlog", "packetlog", MTX_DEF)
150 
151 
152 #define SCTP_IP_PKTLOG_LOCK()	do { 			\
153              mtx_lock(&sctppcbinfo.ipi_pktlog_mtx);     \
154 } while (0)
155 
156 #define SCTP_IP_PKTLOG_UNLOCK()	mtx_unlock(&sctppcbinfo.ipi_pktlog_mtx)
157 
158 #define SCTP_IP_PKTLOG_DESTROY() \
159 	mtx_destroy(&sctppcbinfo.ipi_pktlog_mtx)
160 
161 
162 
163 
164 
165 /*
166  * The INP locks we will use for locking an SCTP endpoint, so for example if
167  * we want to change something at the endpoint level for example random_store
168  * or cookie secrets we lock the INP level.
169  */
170 
171 #define SCTP_INP_READ_INIT(_inp) \
172 	mtx_init(&(_inp)->inp_rdata_mtx, "sctp-read", "inpr", MTX_DEF | MTX_DUPOK)
173 
174 #define SCTP_INP_READ_DESTROY(_inp) \
175 	mtx_destroy(&(_inp)->inp_rdata_mtx)
176 
177 #define SCTP_INP_READ_LOCK(_inp)	do { \
178         mtx_lock(&(_inp)->inp_rdata_mtx);    \
179 } while (0)
180 
181 
182 #define SCTP_INP_READ_UNLOCK(_inp) mtx_unlock(&(_inp)->inp_rdata_mtx)
183 
184 
185 #define SCTP_INP_LOCK_INIT(_inp) \
186 	mtx_init(&(_inp)->inp_mtx, "sctp-inp", "inp", MTX_DEF | MTX_DUPOK)
187 #define SCTP_ASOC_CREATE_LOCK_INIT(_inp) \
188 	mtx_init(&(_inp)->inp_create_mtx, "sctp-create", "inp_create", \
189 		 MTX_DEF | MTX_DUPOK)
190 
191 #define SCTP_INP_LOCK_DESTROY(_inp) \
192 	mtx_destroy(&(_inp)->inp_mtx)
193 
194 #define SCTP_ASOC_CREATE_LOCK_DESTROY(_inp) \
195 	mtx_destroy(&(_inp)->inp_create_mtx)
196 
197 
198 #ifdef SCTP_LOCK_LOGGING
199 #define SCTP_INP_RLOCK(_inp)	do { 					\
200 	if(sctp_logging_level & SCTP_LOCK_LOGGING_ENABLE) sctp_log_lock(_inp, (struct sctp_tcb *)NULL, SCTP_LOG_LOCK_INP);\
201         mtx_lock(&(_inp)->inp_mtx);                                     \
202 } while (0)
203 
204 #define SCTP_INP_WLOCK(_inp)	do { 					\
205 	if(sctp_logging_level & SCTP_LOCK_LOGGING_ENABLE) sctp_log_lock(_inp, (struct sctp_tcb *)NULL, SCTP_LOG_LOCK_INP);\
206         mtx_lock(&(_inp)->inp_mtx);                                     \
207 } while (0)
208 
209 #else
210 
211 #define SCTP_INP_RLOCK(_inp)	do { 					\
212         mtx_lock(&(_inp)->inp_mtx);                                     \
213 } while (0)
214 
215 #define SCTP_INP_WLOCK(_inp)	do { 					\
216         mtx_lock(&(_inp)->inp_mtx);                                     \
217 } while (0)
218 
219 #endif
220 
221 
222 #define SCTP_TCB_SEND_LOCK_INIT(_tcb) \
223 	mtx_init(&(_tcb)->tcb_send_mtx, "sctp-send-tcb", "tcbs", MTX_DEF | MTX_DUPOK)
224 
225 #define SCTP_TCB_SEND_LOCK_DESTROY(_tcb) mtx_destroy(&(_tcb)->tcb_send_mtx)
226 
227 #define SCTP_TCB_SEND_LOCK(_tcb)  do { \
228 	mtx_lock(&(_tcb)->tcb_send_mtx); \
229 } while (0)
230 
231 #define SCTP_TCB_SEND_UNLOCK(_tcb) mtx_unlock(&(_tcb)->tcb_send_mtx)
232 
233 #define SCTP_INP_INCR_REF(_inp) atomic_add_int(&((_inp)->refcount), 1)
234 #define SCTP_INP_DECR_REF(_inp) atomic_add_int(&((_inp)->refcount), -1)
235 
236 
237 #ifdef SCTP_LOCK_LOGGING
238 #define SCTP_ASOC_CREATE_LOCK(_inp) \
239 	do {								\
240 	if(sctp_logging_level & SCTP_LOCK_LOGGING_ENABLE) sctp_log_lock(_inp, (struct sctp_tcb *)NULL, SCTP_LOG_LOCK_CREATE); \
241 		mtx_lock(&(_inp)->inp_create_mtx);			\
242 	} while (0)
243 #else
244 
245 #define SCTP_ASOC_CREATE_LOCK(_inp) \
246 	do {								\
247 		mtx_lock(&(_inp)->inp_create_mtx);			\
248 	} while (0)
249 #endif
250 
251 #define SCTP_INP_RUNLOCK(_inp)		mtx_unlock(&(_inp)->inp_mtx)
252 #define SCTP_INP_WUNLOCK(_inp)		mtx_unlock(&(_inp)->inp_mtx)
253 #define SCTP_ASOC_CREATE_UNLOCK(_inp)	mtx_unlock(&(_inp)->inp_create_mtx)
254 
255 /*
256  * For the majority of things (once we have found the association) we will
257  * lock the actual association mutex. This will protect all the assoiciation
258  * level queues and streams and such. We will need to lock the socket layer
259  * when we stuff data up into the receiving sb_mb. I.e. we will need to do an
260  * extra SOCKBUF_LOCK(&so->so_rcv) even though the association is locked.
261  */
262 
263 #define SCTP_TCB_LOCK_INIT(_tcb) \
264 	mtx_init(&(_tcb)->tcb_mtx, "sctp-tcb", "tcb", MTX_DEF | MTX_DUPOK)
265 
266 #define SCTP_TCB_LOCK_DESTROY(_tcb)	mtx_destroy(&(_tcb)->tcb_mtx)
267 
268 #ifdef SCTP_LOCK_LOGGING
269 #define SCTP_TCB_LOCK(_tcb)  do {					\
270 	if(sctp_logging_level & SCTP_LOCK_LOGGING_ENABLE)  sctp_log_lock(_tcb->sctp_ep, _tcb, SCTP_LOG_LOCK_TCB);          \
271 	mtx_lock(&(_tcb)->tcb_mtx);                                     \
272 } while (0)
273 
274 #else
275 #define SCTP_TCB_LOCK(_tcb)  do {					\
276 	mtx_lock(&(_tcb)->tcb_mtx);                                     \
277 } while (0)
278 
279 #endif
280 
281 
282 #define SCTP_TCB_TRYLOCK(_tcb) 	mtx_trylock(&(_tcb)->tcb_mtx)
283 
284 #define SCTP_TCB_UNLOCK(_tcb)		mtx_unlock(&(_tcb)->tcb_mtx)
285 
286 #define SCTP_TCB_UNLOCK_IFOWNED(_tcb)	      do { \
287                                                 if (mtx_owned(&(_tcb)->tcb_mtx)) \
288                                                      mtx_unlock(&(_tcb)->tcb_mtx); \
289                                               } while (0)
290 
291 
292 
293 #ifdef INVARIANTS
294 #define SCTP_TCB_LOCK_ASSERT(_tcb) do { \
295                             if (mtx_owned(&(_tcb)->tcb_mtx) == 0) \
296                                 panic("Don't own TCB lock"); \
297                             } while (0)
298 #else
299 #define SCTP_TCB_LOCK_ASSERT(_tcb)
300 #endif
301 
302 #define SCTP_ITERATOR_LOCK_INIT() \
303         mtx_init(&sctppcbinfo.it_mtx, "sctp-it", "iterator", MTX_DEF)
304 
305 #ifdef INVARIANTS
306 #define SCTP_ITERATOR_LOCK() \
307 	do {								\
308 		if (mtx_owned(&sctppcbinfo.it_mtx))			\
309 			panic("Iterator Lock");				\
310 		mtx_lock(&sctppcbinfo.it_mtx);				\
311 	} while (0)
312 #else
313 #define SCTP_ITERATOR_LOCK() \
314 	do {								\
315 		mtx_lock(&sctppcbinfo.it_mtx);				\
316 	} while (0)
317 
318 #endif
319 
320 #define SCTP_ITERATOR_UNLOCK()	        mtx_unlock(&sctppcbinfo.it_mtx)
321 #define SCTP_ITERATOR_LOCK_DESTROY()	mtx_destroy(&sctppcbinfo.it_mtx)
322 
323 
324 #define SCTP_INCR_EP_COUNT() \
325                 do { \
326 		       atomic_add_int(&sctppcbinfo.ipi_count_ep, 1); \
327 	        } while (0)
328 
329 #define SCTP_DECR_EP_COUNT() \
330                 do { \
331 		       atomic_subtract_int(&sctppcbinfo.ipi_count_ep, 1); \
332 	        } while (0)
333 
334 #define SCTP_INCR_ASOC_COUNT() \
335                 do { \
336 	               atomic_add_int(&sctppcbinfo.ipi_count_asoc, 1); \
337 	        } while (0)
338 
339 #define SCTP_DECR_ASOC_COUNT() \
340                 do { \
341 	               atomic_subtract_int(&sctppcbinfo.ipi_count_asoc, 1); \
342 	        } while (0)
343 
344 #define SCTP_INCR_LADDR_COUNT() \
345                 do { \
346 	               atomic_add_int(&sctppcbinfo.ipi_count_laddr, 1); \
347 	        } while (0)
348 
349 #define SCTP_DECR_LADDR_COUNT() \
350                 do { \
351 	               atomic_subtract_int(&sctppcbinfo.ipi_count_laddr, 1); \
352 	        } while (0)
353 
354 #define SCTP_INCR_RADDR_COUNT() \
355                 do { \
356  	               atomic_add_int(&sctppcbinfo.ipi_count_raddr, 1); \
357 	        } while (0)
358 
359 #define SCTP_DECR_RADDR_COUNT() \
360                 do { \
361  	               atomic_subtract_int(&sctppcbinfo.ipi_count_raddr,1); \
362 	        } while (0)
363 
364 #define SCTP_INCR_CHK_COUNT() \
365                 do { \
366   	               atomic_add_int(&sctppcbinfo.ipi_count_chunk, 1); \
367 	        } while (0)
368 
369 #define SCTP_DECR_CHK_COUNT() \
370                 do { \
371                        if(sctppcbinfo.ipi_count_chunk == 0) \
372                              panic("chunk count to 0?");    \
373   	               atomic_subtract_int(&sctppcbinfo.ipi_count_chunk, 1); \
374 	        } while (0)
375 
376 #define SCTP_INCR_READQ_COUNT() \
377                 do { \
378 		       atomic_add_int(&sctppcbinfo.ipi_count_readq,1); \
379 	        } while (0)
380 
381 #define SCTP_DECR_READQ_COUNT() \
382                 do { \
383 		       atomic_subtract_int(&sctppcbinfo.ipi_count_readq, 1); \
384 	        } while (0)
385 
386 #define SCTP_INCR_STRMOQ_COUNT() \
387                 do { \
388 		       atomic_add_int(&sctppcbinfo.ipi_count_strmoq, 1); \
389 	        } while (0)
390 
391 #define SCTP_DECR_STRMOQ_COUNT() \
392                 do { \
393 		       atomic_subtract_int(&sctppcbinfo.ipi_count_strmoq, 1); \
394 	        } while (0)
395 
396 
397 #if defined(SCTP_SO_LOCK_TESTING)
398 #define SCTP_INP_SO(sctpinp)	(sctpinp)->ip_inp.inp.inp_socket
399 #define SCTP_SOCKET_LOCK(so, refcnt)
400 #define SCTP_SOCKET_UNLOCK(so, refcnt)
401 #endif
402 
403 #endif
404