1 #ifndef __sctp_lock_bsd_h__
2 #define __sctp_lock_bsd_h__
3 /*-
4 * Copyright (c) 2001-2007, by Cisco Systems, Inc. All rights reserved.
5 *
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions are met:
8 *
9 * a) Redistributions of source code must retain the above copyright notice,
10 * this list of conditions and the following disclaimer.
11 *
12 * b) Redistributions in binary form must reproduce the above copyright
13 * notice, this list of conditions and the following disclaimer in
14 * the documentation and/or other materials provided with the distribution.
15 *
16 * c) Neither the name of Cisco Systems, Inc. nor the names of its
17 * contributors may be used to endorse or promote products derived
18 * from this software without specific prior written permission.
19 *
20 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
21 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
22 * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
24 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
30 * THE POSSIBILITY OF SUCH DAMAGE.
31 */
32
33 /*
34 * General locking concepts: The goal of our locking is to of course provide
35 * consistency and yet minimize overhead. We will attempt to use
36 * non-recursive locks which are supposed to be quite inexpensive. Now in
37 * order to do this the goal is that most functions are not aware of locking.
38 * Once we have a TCB we lock it and unlock when we are through. This means
39 * that the TCB lock is kind-of a "global" lock when working on an
40 * association. Caution must be used when asserting a TCB_LOCK since if we
41 * recurse we deadlock.
42 *
43 * Most other locks (INP and INFO) attempt to localize the locking i.e. we try
44 * to contain the lock and unlock within the function that needs to lock it.
45 * This sometimes mean we do extra locks and unlocks and lose a bit of
46 * efficency, but if the performance statements about non-recursive locks are
47 * true this should not be a problem. One issue that arises with this only
48 * lock when needed is that if an implicit association setup is done we have
49 * a problem. If at the time I lookup an association I have NULL in the tcb
50 * return, by the time I call to create the association some other processor
51 * could have created it. This is what the CREATE lock on the endpoint.
52 * Places where we will be implicitly creating the association OR just
53 * creating an association (the connect call) will assert the CREATE_INP
54 * lock. This will assure us that during all the lookup of INP and INFO if
55 * another creator is also locking/looking up we can gate the two to
56 * synchronize. So the CREATE_INP lock is also another one we must use
57 * extreme caution in locking to make sure we don't hit a re-entrancy issue.
58 *
59 * For non FreeBSD 5.x we provide a bunch of EMPTY lock macros so we can
60 * blatantly put locks everywhere and they reduce to nothing on
61 * NetBSD/OpenBSD and FreeBSD 4.x
62 *
63 */
64
65 /*
66 * When working with the global SCTP lists we lock and unlock the INP_INFO
67 * lock. So when we go to lookup an association we will want to do a
68 * SCTP_INP_INFO_RLOCK() and then when we want to add a new association to
69 * the SCTP_BASE_INFO() list's we will do a SCTP_INP_INFO_WLOCK().
70 */
71 #include <sys/cdefs.h>
72 __FBSDID("$FreeBSD$");
73
74
75 extern struct sctp_foo_stuff sctp_logoff[];
76 extern int sctp_logoff_stuff;
77
78 #define SCTP_IPI_COUNT_INIT()
79
80 #define SCTP_STATLOG_INIT_LOCK()
81 #define SCTP_STATLOG_LOCK()
82 #define SCTP_STATLOG_UNLOCK()
83 #define SCTP_STATLOG_DESTROY()
84
85 #define SCTP_INP_INFO_LOCK_DESTROY() do { \
86 if(rw_wowned(&SCTP_BASE_INFO(ipi_ep_mtx))) { \
87 rw_wunlock(&SCTP_BASE_INFO(ipi_ep_mtx)); \
88 } \
89 rw_destroy(&SCTP_BASE_INFO(ipi_ep_mtx)); \
90 } while (0)
91
92 #define SCTP_INP_INFO_LOCK_INIT() \
93 rw_init(&SCTP_BASE_INFO(ipi_ep_mtx), "sctp-info");
94
95
96 #define SCTP_INP_INFO_RLOCK() do { \
97 rw_rlock(&SCTP_BASE_INFO(ipi_ep_mtx)); \
98 } while (0)
99
100
101 #define SCTP_INP_INFO_WLOCK() do { \
102 rw_wlock(&SCTP_BASE_INFO(ipi_ep_mtx)); \
103 } while (0)
104
105
106 #define SCTP_INP_INFO_RUNLOCK() rw_runlock(&SCTP_BASE_INFO(ipi_ep_mtx))
107 #define SCTP_INP_INFO_WUNLOCK() rw_wunlock(&SCTP_BASE_INFO(ipi_ep_mtx))
108
109
110 #define SCTP_IPI_ADDR_INIT() \
111 rw_init(&SCTP_BASE_INFO(ipi_addr_mtx), "sctp-addr")
112
113 #define SCTP_IPI_ADDR_DESTROY() do { \
114 if(rw_wowned(&SCTP_BASE_INFO(ipi_addr_mtx))) { \
115 rw_wunlock(&SCTP_BASE_INFO(ipi_addr_mtx)); \
116 } \
117 rw_destroy(&SCTP_BASE_INFO(ipi_addr_mtx)); \
118 } while (0)
119
120
121
122 #define SCTP_IPI_ADDR_RLOCK() do { \
123 rw_rlock(&SCTP_BASE_INFO(ipi_addr_mtx)); \
124 } while (0)
125
126 #define SCTP_IPI_ADDR_WLOCK() do { \
127 rw_wlock(&SCTP_BASE_INFO(ipi_addr_mtx)); \
128 } while (0)
129
130
131 #define SCTP_IPI_ADDR_RUNLOCK() rw_runlock(&SCTP_BASE_INFO(ipi_addr_mtx))
132 #define SCTP_IPI_ADDR_WUNLOCK() rw_wunlock(&SCTP_BASE_INFO(ipi_addr_mtx))
133
134
135 #define SCTP_IPI_ITERATOR_WQ_INIT() \
136 mtx_init(&SCTP_BASE_INFO(ipi_iterator_wq_mtx), "sctp-it-wq", "sctp_it_wq", MTX_DEF)
137
138 #define SCTP_IPI_ITERATOR_WQ_DESTROY() \
139 mtx_destroy(&SCTP_BASE_INFO(ipi_iterator_wq_mtx))
140
141 #define SCTP_IPI_ITERATOR_WQ_LOCK() do { \
142 mtx_lock(&SCTP_BASE_INFO(ipi_iterator_wq_mtx)); \
143 } while (0)
144
145 #define SCTP_IPI_ITERATOR_WQ_UNLOCK() mtx_unlock(&SCTP_BASE_INFO(ipi_iterator_wq_mtx))
146
147
148 #define SCTP_IP_PKTLOG_INIT() \
149 mtx_init(&SCTP_BASE_INFO(ipi_pktlog_mtx), "sctp-pktlog", "packetlog", MTX_DEF)
150
151
152 #define SCTP_IP_PKTLOG_LOCK() do { \
153 mtx_lock(&SCTP_BASE_INFO(ipi_pktlog_mtx)); \
154 } while (0)
155
156 #define SCTP_IP_PKTLOG_UNLOCK() mtx_unlock(&SCTP_BASE_INFO(ipi_pktlog_mtx))
157
158 #define SCTP_IP_PKTLOG_DESTROY() \
159 mtx_destroy(&SCTP_BASE_INFO(ipi_pktlog_mtx))
160
161
162
163
164
165 /*
166 * The INP locks we will use for locking an SCTP endpoint, so for example if
167 * we want to change something at the endpoint level for example random_store
168 * or cookie secrets we lock the INP level.
169 */
170
171 #define SCTP_INP_READ_INIT(_inp) \
172 mtx_init(&(_inp)->inp_rdata_mtx, "sctp-read", "inpr", MTX_DEF | MTX_DUPOK)
173
174 #define SCTP_INP_READ_DESTROY(_inp) \
175 mtx_destroy(&(_inp)->inp_rdata_mtx)
176
177 #define SCTP_INP_READ_LOCK(_inp) do { \
178 mtx_lock(&(_inp)->inp_rdata_mtx); \
179 } while (0)
180
181
182 #define SCTP_INP_READ_UNLOCK(_inp) mtx_unlock(&(_inp)->inp_rdata_mtx)
183
184
185 #define SCTP_INP_LOCK_INIT(_inp) \
186 mtx_init(&(_inp)->inp_mtx, "sctp-inp", "inp", MTX_DEF | MTX_DUPOK)
187 #define SCTP_ASOC_CREATE_LOCK_INIT(_inp) \
188 mtx_init(&(_inp)->inp_create_mtx, "sctp-create", "inp_create", \
189 MTX_DEF | MTX_DUPOK)
190
191 #define SCTP_INP_LOCK_DESTROY(_inp) \
192 mtx_destroy(&(_inp)->inp_mtx)
193
194 #define SCTP_ASOC_CREATE_LOCK_DESTROY(_inp) \
195 mtx_destroy(&(_inp)->inp_create_mtx)
196
197
198 #ifdef SCTP_LOCK_LOGGING
199 #define SCTP_INP_RLOCK(_inp) do { \
200 if(SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LOCK_LOGGING_ENABLE) sctp_log_lock(_inp, (struct sctp_tcb *)NULL, SCTP_LOG_LOCK_INP);\
201 mtx_lock(&(_inp)->inp_mtx); \
202 } while (0)
203
204 #define SCTP_INP_WLOCK(_inp) do { \
205 if(SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LOCK_LOGGING_ENABLE) sctp_log_lock(_inp, (struct sctp_tcb *)NULL, SCTP_LOG_LOCK_INP);\
206 mtx_lock(&(_inp)->inp_mtx); \
207 } while (0)
208
209 #else
210
211 #define SCTP_INP_RLOCK(_inp) do { \
212 mtx_lock(&(_inp)->inp_mtx); \
213 } while (0)
214
215 #define SCTP_INP_WLOCK(_inp) do { \
216 mtx_lock(&(_inp)->inp_mtx); \
217 } while (0)
218
219 #endif
220
221
222 #define SCTP_TCB_SEND_LOCK_INIT(_tcb) \
223 mtx_init(&(_tcb)->tcb_send_mtx, "sctp-send-tcb", "tcbs", MTX_DEF | MTX_DUPOK)
224
225 #define SCTP_TCB_SEND_LOCK_DESTROY(_tcb) mtx_destroy(&(_tcb)->tcb_send_mtx)
226
227 #define SCTP_TCB_SEND_LOCK(_tcb) do { \
228 mtx_lock(&(_tcb)->tcb_send_mtx); \
229 } while (0)
230
231 #define SCTP_TCB_SEND_UNLOCK(_tcb) mtx_unlock(&(_tcb)->tcb_send_mtx)
232
233 #define SCTP_INP_INCR_REF(_inp) atomic_add_int(&((_inp)->refcount), 1)
234 #define SCTP_INP_DECR_REF(_inp) atomic_add_int(&((_inp)->refcount), -1)
235
236
237 #ifdef SCTP_LOCK_LOGGING
238 #define SCTP_ASOC_CREATE_LOCK(_inp) \
239 do { \
240 if(SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LOCK_LOGGING_ENABLE) sctp_log_lock(_inp, (struct sctp_tcb *)NULL, SCTP_LOG_LOCK_CREATE); \
241 mtx_lock(&(_inp)->inp_create_mtx); \
242 } while (0)
243 #else
244
245 #define SCTP_ASOC_CREATE_LOCK(_inp) \
246 do { \
247 mtx_lock(&(_inp)->inp_create_mtx); \
248 } while (0)
249 #endif
250
251 #define SCTP_INP_RUNLOCK(_inp) mtx_unlock(&(_inp)->inp_mtx)
252 #define SCTP_INP_WUNLOCK(_inp) mtx_unlock(&(_inp)->inp_mtx)
253 #define SCTP_ASOC_CREATE_UNLOCK(_inp) mtx_unlock(&(_inp)->inp_create_mtx)
254
255 /*
256 * For the majority of things (once we have found the association) we will
257 * lock the actual association mutex. This will protect all the assoiciation
258 * level queues and streams and such. We will need to lock the socket layer
259 * when we stuff data up into the receiving sb_mb. I.e. we will need to do an
260 * extra SOCKBUF_LOCK(&so->so_rcv) even though the association is locked.
261 */
262
263 #define SCTP_TCB_LOCK_INIT(_tcb) \
264 mtx_init(&(_tcb)->tcb_mtx, "sctp-tcb", "tcb", MTX_DEF | MTX_DUPOK)
265
266 #define SCTP_TCB_LOCK_DESTROY(_tcb) mtx_destroy(&(_tcb)->tcb_mtx)
267
268 #ifdef SCTP_LOCK_LOGGING
269 #define SCTP_TCB_LOCK(_tcb) do { \
270 if(SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LOCK_LOGGING_ENABLE) sctp_log_lock(_tcb->sctp_ep, _tcb, SCTP_LOG_LOCK_TCB); \
271 mtx_lock(&(_tcb)->tcb_mtx); \
272 } while (0)
273
274 #else
275 #define SCTP_TCB_LOCK(_tcb) do { \
276 mtx_lock(&(_tcb)->tcb_mtx); \
277 } while (0)
278
279 #endif
280
281
282 #define SCTP_TCB_TRYLOCK(_tcb) mtx_trylock(&(_tcb)->tcb_mtx)
283
284 #define SCTP_TCB_UNLOCK(_tcb) mtx_unlock(&(_tcb)->tcb_mtx)
285
286 #define SCTP_TCB_UNLOCK_IFOWNED(_tcb) do { \
287 if (mtx_owned(&(_tcb)->tcb_mtx)) \
288 mtx_unlock(&(_tcb)->tcb_mtx); \
289 } while (0)
290
291
292
293 #ifdef INVARIANTS
294 #define SCTP_TCB_LOCK_ASSERT(_tcb) do { \
295 if (mtx_owned(&(_tcb)->tcb_mtx) == 0) \
296 panic("Don't own TCB lock"); \
297 } while (0)
298 #else
299 #define SCTP_TCB_LOCK_ASSERT(_tcb)
300 #endif
301
302 #define SCTP_ITERATOR_LOCK_INIT() \
303 mtx_init(&SCTP_BASE_INFO(it_mtx), "sctp-it", "iterator", MTX_DEF)
304
305 #ifdef INVARIANTS
306 #define SCTP_ITERATOR_LOCK() \
307 do { \
308 if (mtx_owned(&SCTP_BASE_INFO(it_mtx))) \
309 panic("Iterator Lock"); \
310 mtx_lock(&SCTP_BASE_INFO(it_mtx)); \
311 } while (0)
312 #else
313 #define SCTP_ITERATOR_LOCK() \
314 do { \
315 mtx_lock(&SCTP_BASE_INFO(it_mtx)); \
316 } while (0)
317
318 #endif
319
320 #define SCTP_ITERATOR_UNLOCK() mtx_unlock(&SCTP_BASE_INFO(it_mtx))
321 #define SCTP_ITERATOR_LOCK_DESTROY() mtx_destroy(&SCTP_BASE_INFO(it_mtx))
322
323
324 #define SCTP_INCR_EP_COUNT() \
325 do { \
326 atomic_add_int(&SCTP_BASE_INFO(ipi_count_ep), 1); \
327 } while (0)
328
329 #define SCTP_DECR_EP_COUNT() \
330 do { \
331 atomic_subtract_int(&SCTP_BASE_INFO(ipi_count_ep), 1); \
332 } while (0)
333
334 #define SCTP_INCR_ASOC_COUNT() \
335 do { \
336 atomic_add_int(&SCTP_BASE_INFO(ipi_count_asoc), 1); \
337 } while (0)
338
339 #define SCTP_DECR_ASOC_COUNT() \
340 do { \
341 atomic_subtract_int(&SCTP_BASE_INFO(ipi_count_asoc), 1); \
342 } while (0)
343
344 #define SCTP_INCR_LADDR_COUNT() \
345 do { \
346 atomic_add_int(&SCTP_BASE_INFO(ipi_count_laddr), 1); \
347 } while (0)
348
349 #define SCTP_DECR_LADDR_COUNT() \
350 do { \
351 atomic_subtract_int(&SCTP_BASE_INFO(ipi_count_laddr), 1); \
352 } while (0)
353
354 #define SCTP_INCR_RADDR_COUNT() \
355 do { \
356 atomic_add_int(&SCTP_BASE_INFO(ipi_count_raddr), 1); \
357 } while (0)
358
359 #define SCTP_DECR_RADDR_COUNT() \
360 do { \
361 atomic_subtract_int(&SCTP_BASE_INFO(ipi_count_raddr),1); \
362 } while (0)
363
364 #define SCTP_INCR_CHK_COUNT() \
365 do { \
366 atomic_add_int(&SCTP_BASE_INFO(ipi_count_chunk), 1); \
367 } while (0)
368
369 #define SCTP_DECR_CHK_COUNT() \
370 do { \
371 if(SCTP_BASE_INFO(ipi_count_chunk) == 0) \
372 panic("chunk count to 0?"); \
373 atomic_subtract_int(&SCTP_BASE_INFO(ipi_count_chunk), 1); \
374 } while (0)
375
376 #define SCTP_INCR_READQ_COUNT() \
377 do { \
378 atomic_add_int(&SCTP_BASE_INFO(ipi_count_readq),1); \
379 } while (0)
380
381 #define SCTP_DECR_READQ_COUNT() \
382 do { \
383 atomic_subtract_int(&SCTP_BASE_INFO(ipi_count_readq), 1); \
384 } while (0)
385
386 #define SCTP_INCR_STRMOQ_COUNT() \
387 do { \
388 atomic_add_int(&SCTP_BASE_INFO(ipi_count_strmoq), 1); \
389 } while (0)
390
391 #define SCTP_DECR_STRMOQ_COUNT() \
392 do { \
393 atomic_subtract_int(&SCTP_BASE_INFO(ipi_count_strmoq), 1); \
394 } while (0)
395
396
397 #if defined(SCTP_SO_LOCK_TESTING)
398 #define SCTP_INP_SO(sctpinp) (sctpinp)->ip_inp.inp.inp_socket
399 #define SCTP_SOCKET_LOCK(so, refcnt)
400 #define SCTP_SOCKET_UNLOCK(so, refcnt)
401 #endif
402
403 #endif
Cache object: 9366ef8edbf7b30b5250f851e5f54421
|