cxgbe/t4_tom: Read the chip's DDP page sizes and save them in a
[freebsd-src.git] / sys / netinet / sctp_lock_bsd.h
blob96e35214594a4d82ef894192f25657a10cf1f1ec
1 /*-
2 * Copyright (c) 2001-2007, by Cisco Systems, Inc. All rights reserved.
3 * Copyright (c) 2008-2012, by Randall Stewart. All rights reserved.
4 * Copyright (c) 2008-2012, by Michael Tuexen. All rights reserved.
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions are met:
9 * a) Redistributions of source code must retain the above copyright notice,
10 * this list of conditions and the following disclaimer.
12 * b) Redistributions in binary form must reproduce the above copyright
13 * notice, this list of conditions and the following disclaimer in
14 * the documentation and/or other materials provided with the distribution.
16 * c) Neither the name of Cisco Systems, Inc. nor the names of its
17 * contributors may be used to endorse or promote products derived
18 * from this software without specific prior written permission.
20 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
21 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
22 * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
24 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
30 * THE POSSIBILITY OF SUCH DAMAGE.
33 #include <sys/cdefs.h>
34 __FBSDID("$FreeBSD$");
36 #ifndef _NETINET_SCTP_LOCK_BSD_H_
37 #define _NETINET_SCTP_LOCK_BSD_H_
40 * General locking concepts: The goal of our locking is to of course provide
41 * consistency and yet minimize overhead. We will attempt to use
42 * non-recursive locks which are supposed to be quite inexpensive. Now in
43 * order to do this the goal is that most functions are not aware of locking.
44 * Once we have a TCB we lock it and unlock when we are through. This means
45 * that the TCB lock is kind-of a "global" lock when working on an
46 * association. Caution must be used when asserting a TCB_LOCK since if we
47 * recurse we deadlock.
49 * Most other locks (INP and INFO) attempt to localize the locking i.e. we try
50 * to contain the lock and unlock within the function that needs to lock it.
51 * This sometimes mean we do extra locks and unlocks and lose a bit of
52 * efficiency, but if the performance statements about non-recursive locks are
53 * true this should not be a problem. One issue that arises with this only
54 * lock when needed is that if an implicit association setup is done we have
55 * a problem. If at the time I lookup an association I have NULL in the tcb
56 * return, by the time I call to create the association some other processor
57 * could have created it. This is what the CREATE lock on the endpoint.
58 * Places where we will be implicitly creating the association OR just
59 * creating an association (the connect call) will assert the CREATE_INP
60 * lock. This will assure us that during all the lookup of INP and INFO if
61 * another creator is also locking/looking up we can gate the two to
62 * synchronize. So the CREATE_INP lock is also another one we must use
63 * extreme caution in locking to make sure we don't hit a re-entrancy issue.
65 * For non FreeBSD 5.x we provide a bunch of EMPTY lock macros so we can
66 * blatantly put locks everywhere and they reduce to nothing on
67 * NetBSD/OpenBSD and FreeBSD 4.x
72 * When working with the global SCTP lists we lock and unlock the INP_INFO
73 * lock. So when we go to lookup an association we will want to do a
74 * SCTP_INP_INFO_RLOCK() and then when we want to add a new association to
75 * the SCTP_BASE_INFO() list's we will do a SCTP_INP_INFO_WLOCK().
78 extern struct sctp_foo_stuff sctp_logoff[];
79 extern int sctp_logoff_stuff;
81 #define SCTP_IPI_COUNT_INIT()
83 #define SCTP_STATLOG_INIT_LOCK()
84 #define SCTP_STATLOG_LOCK()
85 #define SCTP_STATLOG_UNLOCK()
86 #define SCTP_STATLOG_DESTROY()
88 #define SCTP_INP_INFO_LOCK_DESTROY() do { \
89 if(rw_wowned(&SCTP_BASE_INFO(ipi_ep_mtx))) { \
90 rw_wunlock(&SCTP_BASE_INFO(ipi_ep_mtx)); \
91 } \
92 rw_destroy(&SCTP_BASE_INFO(ipi_ep_mtx)); \
93 } while (0)
95 #define SCTP_INP_INFO_LOCK_INIT() \
96 rw_init(&SCTP_BASE_INFO(ipi_ep_mtx), "sctp-info");
99 #define SCTP_INP_INFO_RLOCK() do { \
100 rw_rlock(&SCTP_BASE_INFO(ipi_ep_mtx)); \
101 } while (0)
103 #define SCTP_MCORE_QLOCK_INIT(cpstr) do { \
104 mtx_init(&(cpstr)->que_mtx, \
105 "sctp-mcore_queue","queue_lock", \
106 MTX_DEF|MTX_DUPOK); \
107 } while (0)
109 #define SCTP_MCORE_QLOCK(cpstr) do { \
110 mtx_lock(&(cpstr)->que_mtx); \
111 } while (0)
113 #define SCTP_MCORE_QUNLOCK(cpstr) do { \
114 mtx_unlock(&(cpstr)->que_mtx); \
115 } while (0)
117 #define SCTP_MCORE_QDESTROY(cpstr) do { \
118 if(mtx_owned(&(cpstr)->core_mtx)) { \
119 mtx_unlock(&(cpstr)->que_mtx); \
121 mtx_destroy(&(cpstr)->que_mtx); \
122 } while (0)
125 #define SCTP_MCORE_LOCK_INIT(cpstr) do { \
126 mtx_init(&(cpstr)->core_mtx, \
127 "sctp-cpulck","cpu_proc_lock", \
128 MTX_DEF|MTX_DUPOK); \
129 } while (0)
131 #define SCTP_MCORE_LOCK(cpstr) do { \
132 mtx_lock(&(cpstr)->core_mtx); \
133 } while (0)
135 #define SCTP_MCORE_UNLOCK(cpstr) do { \
136 mtx_unlock(&(cpstr)->core_mtx); \
137 } while (0)
139 #define SCTP_MCORE_DESTROY(cpstr) do { \
140 if(mtx_owned(&(cpstr)->core_mtx)) { \
141 mtx_unlock(&(cpstr)->core_mtx); \
143 mtx_destroy(&(cpstr)->core_mtx); \
144 } while (0)
146 #define SCTP_INP_INFO_WLOCK() do { \
147 rw_wlock(&SCTP_BASE_INFO(ipi_ep_mtx)); \
148 } while (0)
151 #define SCTP_INP_INFO_RUNLOCK() rw_runlock(&SCTP_BASE_INFO(ipi_ep_mtx))
152 #define SCTP_INP_INFO_WUNLOCK() rw_wunlock(&SCTP_BASE_INFO(ipi_ep_mtx))
155 #define SCTP_IPI_ADDR_INIT() \
156 rw_init(&SCTP_BASE_INFO(ipi_addr_mtx), "sctp-addr")
157 #define SCTP_IPI_ADDR_DESTROY() do { \
158 if(rw_wowned(&SCTP_BASE_INFO(ipi_addr_mtx))) { \
159 rw_wunlock(&SCTP_BASE_INFO(ipi_addr_mtx)); \
161 rw_destroy(&SCTP_BASE_INFO(ipi_addr_mtx)); \
162 } while (0)
163 #define SCTP_IPI_ADDR_RLOCK() do { \
164 rw_rlock(&SCTP_BASE_INFO(ipi_addr_mtx)); \
165 } while (0)
166 #define SCTP_IPI_ADDR_WLOCK() do { \
167 rw_wlock(&SCTP_BASE_INFO(ipi_addr_mtx)); \
168 } while (0)
170 #define SCTP_IPI_ADDR_RUNLOCK() rw_runlock(&SCTP_BASE_INFO(ipi_addr_mtx))
171 #define SCTP_IPI_ADDR_WUNLOCK() rw_wunlock(&SCTP_BASE_INFO(ipi_addr_mtx))
174 #define SCTP_IPI_ITERATOR_WQ_INIT() \
175 mtx_init(&sctp_it_ctl.ipi_iterator_wq_mtx, "sctp-it-wq", "sctp_it_wq", MTX_DEF)
177 #define SCTP_IPI_ITERATOR_WQ_DESTROY() \
178 mtx_destroy(&sctp_it_ctl.ipi_iterator_wq_mtx)
180 #define SCTP_IPI_ITERATOR_WQ_LOCK() do { \
181 mtx_lock(&sctp_it_ctl.ipi_iterator_wq_mtx); \
182 } while (0)
184 #define SCTP_IPI_ITERATOR_WQ_UNLOCK() mtx_unlock(&sctp_it_ctl.ipi_iterator_wq_mtx)
187 #define SCTP_IP_PKTLOG_INIT() \
188 mtx_init(&SCTP_BASE_INFO(ipi_pktlog_mtx), "sctp-pktlog", "packetlog", MTX_DEF)
191 #define SCTP_IP_PKTLOG_LOCK() do { \
192 mtx_lock(&SCTP_BASE_INFO(ipi_pktlog_mtx)); \
193 } while (0)
195 #define SCTP_IP_PKTLOG_UNLOCK() mtx_unlock(&SCTP_BASE_INFO(ipi_pktlog_mtx))
197 #define SCTP_IP_PKTLOG_DESTROY() \
198 mtx_destroy(&SCTP_BASE_INFO(ipi_pktlog_mtx))
205 * The INP locks we will use for locking an SCTP endpoint, so for example if
206 * we want to change something at the endpoint level for example random_store
207 * or cookie secrets we lock the INP level.
210 #define SCTP_INP_READ_INIT(_inp) \
211 mtx_init(&(_inp)->inp_rdata_mtx, "sctp-read", "inpr", MTX_DEF | MTX_DUPOK)
213 #define SCTP_INP_READ_DESTROY(_inp) \
214 mtx_destroy(&(_inp)->inp_rdata_mtx)
216 #define SCTP_INP_READ_LOCK(_inp) do { \
217 mtx_lock(&(_inp)->inp_rdata_mtx); \
218 } while (0)
221 #define SCTP_INP_READ_UNLOCK(_inp) mtx_unlock(&(_inp)->inp_rdata_mtx)
224 #define SCTP_INP_LOCK_INIT(_inp) \
225 mtx_init(&(_inp)->inp_mtx, "sctp-inp", "inp", MTX_DEF | MTX_DUPOK)
226 #define SCTP_ASOC_CREATE_LOCK_INIT(_inp) \
227 mtx_init(&(_inp)->inp_create_mtx, "sctp-create", "inp_create", \
228 MTX_DEF | MTX_DUPOK)
230 #define SCTP_INP_LOCK_DESTROY(_inp) \
231 mtx_destroy(&(_inp)->inp_mtx)
233 #define SCTP_INP_LOCK_CONTENDED(_inp) ((_inp)->inp_mtx.mtx_lock & MTX_CONTESTED)
235 #define SCTP_INP_READ_CONTENDED(_inp) ((_inp)->inp_rdata_mtx.mtx_lock & MTX_CONTESTED)
237 #define SCTP_ASOC_CREATE_LOCK_CONTENDED(_inp) ((_inp)->inp_create_mtx.mtx_lock & MTX_CONTESTED)
240 #define SCTP_ASOC_CREATE_LOCK_DESTROY(_inp) \
241 mtx_destroy(&(_inp)->inp_create_mtx)
244 #ifdef SCTP_LOCK_LOGGING
245 #define SCTP_INP_RLOCK(_inp) do { \
246 if(SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LOCK_LOGGING_ENABLE) sctp_log_lock(_inp, (struct sctp_tcb *)NULL, SCTP_LOG_LOCK_INP);\
247 mtx_lock(&(_inp)->inp_mtx); \
248 } while (0)
250 #define SCTP_INP_WLOCK(_inp) do { \
251 if(SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LOCK_LOGGING_ENABLE) sctp_log_lock(_inp, (struct sctp_tcb *)NULL, SCTP_LOG_LOCK_INP);\
252 mtx_lock(&(_inp)->inp_mtx); \
253 } while (0)
255 #else
257 #define SCTP_INP_RLOCK(_inp) do { \
258 mtx_lock(&(_inp)->inp_mtx); \
259 } while (0)
261 #define SCTP_INP_WLOCK(_inp) do { \
262 mtx_lock(&(_inp)->inp_mtx); \
263 } while (0)
265 #endif
268 #define SCTP_TCB_SEND_LOCK_INIT(_tcb) \
269 mtx_init(&(_tcb)->tcb_send_mtx, "sctp-send-tcb", "tcbs", MTX_DEF | MTX_DUPOK)
271 #define SCTP_TCB_SEND_LOCK_DESTROY(_tcb) mtx_destroy(&(_tcb)->tcb_send_mtx)
273 #define SCTP_TCB_SEND_LOCK(_tcb) do { \
274 mtx_lock(&(_tcb)->tcb_send_mtx); \
275 } while (0)
277 #define SCTP_TCB_SEND_UNLOCK(_tcb) mtx_unlock(&(_tcb)->tcb_send_mtx)
279 #define SCTP_INP_INCR_REF(_inp) atomic_add_int(&((_inp)->refcount), 1)
280 #define SCTP_INP_DECR_REF(_inp) atomic_add_int(&((_inp)->refcount), -1)
283 #ifdef SCTP_LOCK_LOGGING
284 #define SCTP_ASOC_CREATE_LOCK(_inp) \
285 do { \
286 if(SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LOCK_LOGGING_ENABLE) sctp_log_lock(_inp, (struct sctp_tcb *)NULL, SCTP_LOG_LOCK_CREATE); \
287 mtx_lock(&(_inp)->inp_create_mtx); \
288 } while (0)
289 #else
291 #define SCTP_ASOC_CREATE_LOCK(_inp) \
292 do { \
293 mtx_lock(&(_inp)->inp_create_mtx); \
294 } while (0)
295 #endif
297 #define SCTP_INP_RUNLOCK(_inp) mtx_unlock(&(_inp)->inp_mtx)
298 #define SCTP_INP_WUNLOCK(_inp) mtx_unlock(&(_inp)->inp_mtx)
299 #define SCTP_ASOC_CREATE_UNLOCK(_inp) mtx_unlock(&(_inp)->inp_create_mtx)
302 * For the majority of things (once we have found the association) we will
303 * lock the actual association mutex. This will protect all the assoiciation
304 * level queues and streams and such. We will need to lock the socket layer
305 * when we stuff data up into the receiving sb_mb. I.e. we will need to do an
306 * extra SOCKBUF_LOCK(&so->so_rcv) even though the association is locked.
309 #define SCTP_TCB_LOCK_INIT(_tcb) \
310 mtx_init(&(_tcb)->tcb_mtx, "sctp-tcb", "tcb", MTX_DEF | MTX_DUPOK)
312 #define SCTP_TCB_LOCK_DESTROY(_tcb) mtx_destroy(&(_tcb)->tcb_mtx)
314 #ifdef SCTP_LOCK_LOGGING
315 #define SCTP_TCB_LOCK(_tcb) do { \
316 if(SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LOCK_LOGGING_ENABLE) sctp_log_lock(_tcb->sctp_ep, _tcb, SCTP_LOG_LOCK_TCB); \
317 mtx_lock(&(_tcb)->tcb_mtx); \
318 } while (0)
320 #else
321 #define SCTP_TCB_LOCK(_tcb) do { \
322 mtx_lock(&(_tcb)->tcb_mtx); \
323 } while (0)
325 #endif
328 #define SCTP_TCB_TRYLOCK(_tcb) mtx_trylock(&(_tcb)->tcb_mtx)
330 #define SCTP_TCB_UNLOCK(_tcb) mtx_unlock(&(_tcb)->tcb_mtx)
332 #define SCTP_TCB_UNLOCK_IFOWNED(_tcb) do { \
333 if (mtx_owned(&(_tcb)->tcb_mtx)) \
334 mtx_unlock(&(_tcb)->tcb_mtx); \
335 } while (0)
339 #ifdef INVARIANTS
340 #define SCTP_TCB_LOCK_ASSERT(_tcb) do { \
341 if (mtx_owned(&(_tcb)->tcb_mtx) == 0) \
342 panic("Don't own TCB lock"); \
343 } while (0)
344 #else
345 #define SCTP_TCB_LOCK_ASSERT(_tcb)
346 #endif
348 #define SCTP_ITERATOR_LOCK_INIT() \
349 mtx_init(&sctp_it_ctl.it_mtx, "sctp-it", "iterator", MTX_DEF)
351 #ifdef INVARIANTS
352 #define SCTP_ITERATOR_LOCK() \
353 do { \
354 if (mtx_owned(&sctp_it_ctl.it_mtx)) \
355 panic("Iterator Lock"); \
356 mtx_lock(&sctp_it_ctl.it_mtx); \
357 } while (0)
358 #else
359 #define SCTP_ITERATOR_LOCK() \
360 do { \
361 mtx_lock(&sctp_it_ctl.it_mtx); \
362 } while (0)
364 #endif
366 #define SCTP_ITERATOR_UNLOCK() mtx_unlock(&sctp_it_ctl.it_mtx)
367 #define SCTP_ITERATOR_LOCK_DESTROY() mtx_destroy(&sctp_it_ctl.it_mtx)
370 #define SCTP_WQ_ADDR_INIT() do { \
371 mtx_init(&SCTP_BASE_INFO(wq_addr_mtx), "sctp-addr-wq","sctp_addr_wq",MTX_DEF); \
372 } while (0)
374 #define SCTP_WQ_ADDR_DESTROY() do { \
375 if(mtx_owned(&SCTP_BASE_INFO(wq_addr_mtx))) { \
376 mtx_unlock(&SCTP_BASE_INFO(wq_addr_mtx)); \
378 mtx_destroy(&SCTP_BASE_INFO(wq_addr_mtx)); \
379 } while (0)
381 #define SCTP_WQ_ADDR_LOCK() do { \
382 mtx_lock(&SCTP_BASE_INFO(wq_addr_mtx)); \
383 } while (0)
384 #define SCTP_WQ_ADDR_UNLOCK() do { \
385 mtx_unlock(&SCTP_BASE_INFO(wq_addr_mtx)); \
386 } while (0)
390 #define SCTP_INCR_EP_COUNT() \
391 do { \
392 atomic_add_int(&SCTP_BASE_INFO(ipi_count_ep), 1); \
393 } while (0)
395 #define SCTP_DECR_EP_COUNT() \
396 do { \
397 atomic_subtract_int(&SCTP_BASE_INFO(ipi_count_ep), 1); \
398 } while (0)
400 #define SCTP_INCR_ASOC_COUNT() \
401 do { \
402 atomic_add_int(&SCTP_BASE_INFO(ipi_count_asoc), 1); \
403 } while (0)
405 #define SCTP_DECR_ASOC_COUNT() \
406 do { \
407 atomic_subtract_int(&SCTP_BASE_INFO(ipi_count_asoc), 1); \
408 } while (0)
410 #define SCTP_INCR_LADDR_COUNT() \
411 do { \
412 atomic_add_int(&SCTP_BASE_INFO(ipi_count_laddr), 1); \
413 } while (0)
415 #define SCTP_DECR_LADDR_COUNT() \
416 do { \
417 atomic_subtract_int(&SCTP_BASE_INFO(ipi_count_laddr), 1); \
418 } while (0)
420 #define SCTP_INCR_RADDR_COUNT() \
421 do { \
422 atomic_add_int(&SCTP_BASE_INFO(ipi_count_raddr), 1); \
423 } while (0)
425 #define SCTP_DECR_RADDR_COUNT() \
426 do { \
427 atomic_subtract_int(&SCTP_BASE_INFO(ipi_count_raddr),1); \
428 } while (0)
430 #define SCTP_INCR_CHK_COUNT() \
431 do { \
432 atomic_add_int(&SCTP_BASE_INFO(ipi_count_chunk), 1); \
433 } while (0)
434 #ifdef INVARIANTS
435 #define SCTP_DECR_CHK_COUNT() \
436 do { \
437 if(SCTP_BASE_INFO(ipi_count_chunk) == 0) \
438 panic("chunk count to 0?"); \
439 atomic_subtract_int(&SCTP_BASE_INFO(ipi_count_chunk), 1); \
440 } while (0)
441 #else
442 #define SCTP_DECR_CHK_COUNT() \
443 do { \
444 if(SCTP_BASE_INFO(ipi_count_chunk) != 0) \
445 atomic_subtract_int(&SCTP_BASE_INFO(ipi_count_chunk), 1); \
446 } while (0)
447 #endif
448 #define SCTP_INCR_READQ_COUNT() \
449 do { \
450 atomic_add_int(&SCTP_BASE_INFO(ipi_count_readq),1); \
451 } while (0)
453 #define SCTP_DECR_READQ_COUNT() \
454 do { \
455 atomic_subtract_int(&SCTP_BASE_INFO(ipi_count_readq), 1); \
456 } while (0)
458 #define SCTP_INCR_STRMOQ_COUNT() \
459 do { \
460 atomic_add_int(&SCTP_BASE_INFO(ipi_count_strmoq), 1); \
461 } while (0)
463 #define SCTP_DECR_STRMOQ_COUNT() \
464 do { \
465 atomic_subtract_int(&SCTP_BASE_INFO(ipi_count_strmoq), 1); \
466 } while (0)
469 #if defined(SCTP_SO_LOCK_TESTING)
470 #define SCTP_INP_SO(sctpinp) (sctpinp)->ip_inp.inp.inp_socket
471 #define SCTP_SOCKET_LOCK(so, refcnt)
472 #define SCTP_SOCKET_UNLOCK(so, refcnt)
473 #endif
475 #endif