arcmsr(4): Use MSI if it is supported by the device.
[dragonfly.git] / sys / netinet / sctp_pcb.c
blob699f52d96a644cfc7569ad3c1918b5ace01902fb
1 /* $KAME: sctp_pcb.c,v 1.37 2004/08/17 06:28:02 t-momose Exp $ */
3 /*
4 * Copyright (c) 2001, 2002, 2003, 2004 Cisco Systems, Inc.
5 * All rights reserved.
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
9 * are met:
10 * 1. Redistributions of source code must retain the above copyright
11 * notice, this list of conditions and the following disclaimer.
12 * 2. Redistributions in binary form must reproduce the above copyright
13 * notice, this list of conditions and the following disclaimer in the
14 * documentation and/or other materials provided with the distribution.
15 * 3. All advertising materials mentioning features or use of this software
16 * must display the following acknowledgement:
17 * This product includes software developed by Cisco Systems, Inc.
18 * 4. Neither the name of the project nor the names of its contributors
19 * may be used to endorse or promote products derived from this software
20 * without specific prior written permission.
22 * THIS SOFTWARE IS PROVIDED BY CISCO SYSTEMS AND CONTRIBUTORS ``AS IS'' AND
23 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
24 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
25 * ARE DISCLAIMED. IN NO EVENT SHALL CISCO SYSTEMS OR CONTRIBUTORS BE LIABLE
26 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
27 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
28 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
29 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
30 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
31 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
32 * SUCH DAMAGE.
34 #if !(defined(__OpenBSD__) || defined(__APPLE__))
35 #include "opt_ipsec.h"
36 #endif
37 #if defined(__FreeBSD__) || defined(__DragonFly__)
38 #include "opt_compat.h"
39 #include "opt_inet6.h"
40 #include "opt_inet.h"
41 #endif
42 #if defined(__NetBSD__)
43 #include "opt_inet.h"
44 #endif
45 #ifdef __APPLE__
46 #include <sctp.h>
47 #elif !defined(__OpenBSD__)
48 #include "opt_sctp.h"
49 #endif
51 #include <sys/param.h>
52 #include <sys/systm.h>
53 #include <sys/malloc.h>
54 #include <sys/mbuf.h>
55 #include <sys/domain.h>
56 #include <sys/protosw.h>
57 #include <sys/socket.h>
58 #include <sys/socketvar.h>
59 #include <sys/proc.h>
60 #include <sys/priv.h>
61 #include <sys/kernel.h>
62 #include <sys/sysctl.h>
63 #include <sys/thread2.h>
64 #if defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__)
65 #include <sys/random.h>
66 #endif
67 #if defined(__NetBSD__)
68 #include <sys/rnd.h>
69 #endif
70 #if defined(__OpenBSD__)
71 #include <dev/rndvar.h>
72 #endif
74 #if defined(__APPLE__)
75 #include <netinet/sctp_callout.h>
76 #elif defined(__OpenBSD__)
77 #include <sys/timeout.h>
78 #else
79 #include <sys/callout.h>
80 #endif
82 #if (defined(__FreeBSD__) && __FreeBSD_version >= 500000)
83 #include <sys/limits.h>
84 #else
85 #include <machine/limits.h>
86 #endif
87 #include <machine/cpu.h>
89 #include <net/if.h>
90 #include <net/if_types.h>
91 #include <net/route.h>
92 #include <netinet/in.h>
93 #include <netinet/in_systm.h>
94 #include <netinet/ip.h>
95 #include <netinet/in_pcb.h>
96 #include <netinet/in_var.h>
97 #include <netinet/ip_var.h>
99 #ifdef INET6
100 #include <netinet/ip6.h>
101 #include <netinet6/ip6_var.h>
102 #include <netinet6/scope6_var.h>
103 #if defined(__FreeBSD__) || (__NetBSD__) || defined(__DragonFly__)
104 #include <netinet6/in6_pcb.h>
105 #elif defined(__OpenBSD__)
106 #include <netinet/in_pcb.h>
107 #endif
108 #endif /* INET6 */
110 #ifdef IPSEC
111 #ifndef __OpenBSD__
112 #include <netinet6/ipsec.h>
113 #include <netproto/key/key.h>
114 #else
115 #undef IPSEC
116 #endif
117 #endif /* IPSEC */
119 #include <netinet/sctp_var.h>
120 #include <netinet/sctp_pcb.h>
121 #include <netinet/sctputil.h>
122 #include <netinet/sctp.h>
123 #include <netinet/sctp_header.h>
124 #include <netinet/sctp_asconf.h>
125 #include <netinet/sctp_output.h>
126 #include <netinet/sctp_timer.h>
128 #ifndef SCTP_PCBHASHSIZE
129 /* default number of association hash buckets in each endpoint */
130 #define SCTP_PCBHASHSIZE 256
131 #endif
133 #ifdef SCTP_DEBUG
134 u_int32_t sctp_debug_on = 0;
135 #endif /* SCTP_DEBUG */
137 u_int32_t sctp_pegs[SCTP_NUMBER_OF_PEGS];
139 int sctp_pcbtblsize = SCTP_PCBHASHSIZE;
141 struct sctp_epinfo sctppcbinfo;
143 /* FIX: we don't handle multiple link local scopes */
144 /* "scopeless" replacement IN6_ARE_ADDR_EQUAL */
146 SCTP6_ARE_ADDR_EQUAL(struct in6_addr *a, struct in6_addr *b)
148 struct in6_addr tmp_a, tmp_b;
149 /* use a copy of a and b */
150 tmp_a = *a;
151 tmp_b = *b;
152 in6_clearscope(&tmp_a);
153 in6_clearscope(&tmp_b);
154 return (IN6_ARE_ADDR_EQUAL(&tmp_a, &tmp_b));
157 #ifdef __OpenBSD__
158 extern int ipport_firstauto;
159 extern int ipport_lastauto;
160 extern int ipport_hifirstauto;
161 extern int ipport_hilastauto;
162 #endif
164 #if defined(__FreeBSD__) && __FreeBSD_version > 500000
166 #ifndef xyzzy
167 void sctp_validate_no_locks(void);
169 void
170 SCTP_INP_RLOCK(struct sctp_inpcb *inp)
172 struct sctp_tcb *stcb;
173 LIST_FOREACH(stcb, &inp->sctp_asoc_list, sctp_tcblist) {
174 if (mtx_owned(&(stcb)->tcb_mtx))
175 panic("I own TCB lock?");
177 if (mtx_owned(&(inp)->inp_mtx))
178 panic("INP Recursive Lock-R");
179 mtx_lock(&(inp)->inp_mtx);
182 void
183 SCTP_INP_WLOCK(struct sctp_inpcb *inp)
185 SCTP_INP_RLOCK(inp);
188 void
189 SCTP_INP_INFO_RLOCK(void)
191 struct sctp_inpcb *inp;
192 struct sctp_tcb *stcb;
193 LIST_FOREACH(inp, &sctppcbinfo.listhead, sctp_list) {
194 if (mtx_owned(&(inp)->inp_mtx))
195 panic("info-lock and own inp lock?");
196 LIST_FOREACH(stcb, &inp->sctp_asoc_list, sctp_tcblist) {
197 if (mtx_owned(&(stcb)->tcb_mtx))
198 panic("Info lock and own a tcb lock?");
201 if (mtx_owned(&sctppcbinfo.ipi_ep_mtx))
202 panic("INP INFO Recursive Lock-R");
203 mtx_lock(&sctppcbinfo.ipi_ep_mtx);
206 void
207 SCTP_INP_INFO_WLOCK(void)
209 SCTP_INP_INFO_RLOCK();
213 void sctp_validate_no_locks(void)
215 struct sctp_inpcb *inp;
216 struct sctp_tcb *stcb;
218 if (mtx_owned(&sctppcbinfo.ipi_ep_mtx))
219 panic("INP INFO lock is owned?");
221 LIST_FOREACH(inp, &sctppcbinfo.listhead, sctp_list) {
222 if (mtx_owned(&(inp)->inp_mtx))
223 panic("You own an INP lock?");
224 LIST_FOREACH(stcb, &inp->sctp_asoc_list, sctp_tcblist) {
225 if (mtx_owned(&(stcb)->tcb_mtx))
226 panic("You own a TCB lock?");
231 #endif
232 #endif
234 void
235 sctp_fill_pcbinfo(struct sctp_pcbinfo *spcb)
237 /* We really don't need
238 * to lock this, but I will
239 * just because it does not hurt.
241 SCTP_INP_INFO_RLOCK();
242 spcb->ep_count = sctppcbinfo.ipi_count_ep;
243 spcb->asoc_count = sctppcbinfo.ipi_count_asoc;
244 spcb->laddr_count = sctppcbinfo.ipi_count_laddr;
245 spcb->raddr_count = sctppcbinfo.ipi_count_raddr;
246 spcb->chk_count = sctppcbinfo.ipi_count_chunk;
247 spcb->sockq_count = sctppcbinfo.ipi_count_sockq;
248 spcb->mbuf_track = sctppcbinfo.mbuf_track;
249 SCTP_INP_INFO_RUNLOCK();
254 * Notes on locks for FreeBSD 5 and up. All association
255 * lookups that have a definte ep, the INP structure is
256 * assumed to be locked for reading. If we need to go
257 * find the INP (ususally when a **inp is passed) then
258 * we must lock the INFO structure first and if needed
259 * lock the INP too. Note that if we lock it we must
265 * Given a endpoint, look and find in its association list any association
266 * with the "to" address given. This can be a "from" address, too, for
267 * inbound packets. For outbound packets it is a true "to" address.
269 static struct sctp_tcb *
270 sctp_tcb_special_locate(struct sctp_inpcb **inp_p, struct sockaddr *from,
271 struct sockaddr *to, struct sctp_nets **netp)
273 /**** ASSUMSES THE CALLER holds the INP_INFO_RLOCK */
276 * Note for this module care must be taken when observing what to is
277 * for. In most of the rest of the code the TO field represents my
278 * peer and the FROM field represents my address. For this module it
279 * is reversed of that.
282 * If we support the TCP model, then we must now dig through to
283 * see if we can find our endpoint in the list of tcp ep's.
285 uint16_t lport, rport;
286 struct sctppcbhead *ephead;
287 struct sctp_inpcb *inp;
288 struct sctp_laddr *laddr;
289 struct sctp_tcb *stcb;
290 struct sctp_nets *net;
292 if ((to == NULL) || (from == NULL)) {
293 return (NULL);
296 if (to->sa_family == AF_INET && from->sa_family == AF_INET) {
297 lport = ((struct sockaddr_in *)to)->sin_port;
298 rport = ((struct sockaddr_in *)from)->sin_port;
299 } else if (to->sa_family == AF_INET6 && from->sa_family == AF_INET6) {
300 lport = ((struct sockaddr_in6 *)to)->sin6_port;
301 rport = ((struct sockaddr_in6 *)from)->sin6_port;
302 } else {
303 return NULL;
305 ephead = &sctppcbinfo.sctp_tcpephash[SCTP_PCBHASH_ALLADDR(
306 (lport + rport), sctppcbinfo.hashtcpmark)];
308 * Ok now for each of the guys in this bucket we must look
309 * and see:
310 * - Does the remote port match.
311 * - Does there single association's addresses match this
312 * address (to).
313 * If so we update p_ep to point to this ep and return the
314 * tcb from it.
316 LIST_FOREACH(inp, ephead, sctp_hash) {
317 if (lport != inp->sctp_lport) {
318 continue;
320 SCTP_INP_RLOCK(inp);
321 /* check to see if the ep has one of the addresses */
322 if ((inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) == 0) {
323 /* We are NOT bound all, so look further */
324 int match = 0;
326 LIST_FOREACH(laddr, &inp->sctp_addr_list, sctp_nxt_addr) {
327 if (laddr->ifa == NULL) {
328 #ifdef SCTP_DEBUG
329 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
330 kprintf("An ounce of prevention is worth a pound of cure\n");
332 #endif
333 continue;
335 if (laddr->ifa->ifa_addr == NULL) {
336 #ifdef SCTP_DEBUG
337 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
338 kprintf("ifa with a NULL address\n");
340 #endif
341 continue;
343 if (laddr->ifa->ifa_addr->sa_family ==
344 to->sa_family) {
345 /* see if it matches */
346 struct sockaddr_in *intf_addr, *sin;
347 intf_addr = (struct sockaddr_in *)
348 laddr->ifa->ifa_addr;
349 sin = (struct sockaddr_in *)to;
350 if (from->sa_family == AF_INET) {
351 if (sin->sin_addr.s_addr ==
352 intf_addr->sin_addr.s_addr) {
353 match = 1;
354 SCTP_INP_RUNLOCK(inp);
355 break;
357 } else {
358 struct sockaddr_in6 *intf_addr6;
359 struct sockaddr_in6 *sin6;
360 sin6 = (struct sockaddr_in6 *)
362 intf_addr6 = (struct sockaddr_in6 *)
363 laddr->ifa->ifa_addr;
365 if (SCTP6_ARE_ADDR_EQUAL(&sin6->sin6_addr,
366 &intf_addr6->sin6_addr)) {
367 match = 1;
368 SCTP_INP_RUNLOCK(inp);
369 break;
374 if (match == 0) {
375 /* This endpoint does not have this address */
376 SCTP_INP_RUNLOCK(inp);
377 continue;
381 * Ok if we hit here the ep has the address, does it hold the
382 * tcb?
385 stcb = LIST_FIRST(&inp->sctp_asoc_list);
386 if (stcb == NULL) {
387 SCTP_INP_RUNLOCK(inp);
388 continue;
390 SCTP_TCB_LOCK(stcb);
391 if (stcb->rport != rport) {
392 /* remote port does not match. */
393 SCTP_TCB_UNLOCK(stcb);
394 SCTP_INP_RUNLOCK(inp);
395 continue;
397 /* Does this TCB have a matching address? */
398 TAILQ_FOREACH(net, &stcb->asoc.nets, sctp_next) {
399 if (net->ro._l_addr.sa.sa_family != from->sa_family) {
400 /* not the same family, can't be a match */
401 continue;
403 if (from->sa_family == AF_INET) {
404 struct sockaddr_in *sin, *rsin;
405 sin = (struct sockaddr_in *)&net->ro._l_addr;
406 rsin = (struct sockaddr_in *)from;
407 if (sin->sin_addr.s_addr ==
408 rsin->sin_addr.s_addr) {
409 /* found it */
410 if (netp != NULL) {
411 *netp = net;
413 /* Update the endpoint pointer */
414 *inp_p = inp;
415 SCTP_INP_RUNLOCK(inp);
416 return (stcb);
418 } else {
419 struct sockaddr_in6 *sin6, *rsin6;
420 sin6 = (struct sockaddr_in6 *)&net->ro._l_addr;
421 rsin6 = (struct sockaddr_in6 *)from;
422 if (SCTP6_ARE_ADDR_EQUAL(&sin6->sin6_addr,
423 &rsin6->sin6_addr)) {
424 /* found it */
425 if (netp != NULL) {
426 *netp = net;
428 /* Update the endpoint pointer */
429 *inp_p = inp;
430 SCTP_INP_RUNLOCK(inp);
431 return (stcb);
435 SCTP_TCB_UNLOCK(stcb);
437 SCTP_INP_RUNLOCK(inp);
439 return (NULL);
442 struct sctp_tcb *
443 sctp_findassociation_ep_asconf(struct mbuf *m, int iphlen, int offset,
444 struct sctphdr *sh, struct sctp_inpcb **inp_p, struct sctp_nets **netp)
446 struct sctp_tcb *stcb;
447 struct sockaddr_in *sin;
448 struct sockaddr_in6 *sin6;
449 struct sockaddr_storage local_store, remote_store;
450 struct ip *iph;
451 struct sctp_paramhdr parm_buf, *phdr;
452 int ptype;
454 memset(&local_store, 0, sizeof(local_store));
455 memset(&remote_store, 0, sizeof(remote_store));
457 /* First get the destination address setup too. */
458 iph = mtod(m, struct ip *);
459 if (iph->ip_v == IPVERSION) {
460 /* its IPv4 */
461 sin = (struct sockaddr_in *)&local_store;
462 sin->sin_family = AF_INET;
463 sin->sin_len = sizeof(*sin);
464 sin->sin_port = sh->dest_port;
465 sin->sin_addr.s_addr = iph->ip_dst.s_addr ;
466 } else if (iph->ip_v == (IPV6_VERSION >> 4)) {
467 /* its IPv6 */
468 struct ip6_hdr *ip6;
469 ip6 = mtod(m, struct ip6_hdr *);
470 sin6 = (struct sockaddr_in6 *)&local_store;
471 sin6->sin6_family = AF_INET6;
472 sin6->sin6_len = sizeof(*sin6);
473 sin6->sin6_port = sh->dest_port;
474 sin6->sin6_addr = ip6->ip6_dst;
475 } else {
476 return NULL;
479 phdr = sctp_get_next_param(m, offset + sizeof(struct sctp_asconf_chunk),
480 &parm_buf, sizeof(struct sctp_paramhdr));
481 if (phdr == NULL) {
482 #ifdef SCTP_DEBUG
483 if (sctp_debug_on & SCTP_DEBUG_INPUT3) {
484 kprintf("sctp_process_control: failed to get asconf lookup addr\n");
486 #endif /* SCTP_DEBUG */
487 return NULL;
489 ptype = (int)((u_int)ntohs(phdr->param_type));
490 /* get the correlation address */
491 if (ptype == SCTP_IPV6_ADDRESS) {
492 /* ipv6 address param */
493 struct sctp_ipv6addr_param *p6, p6_buf;
494 if (ntohs(phdr->param_length) != sizeof(struct sctp_ipv6addr_param)) {
495 return NULL;
498 p6 = (struct sctp_ipv6addr_param *)sctp_get_next_param(m,
499 offset + sizeof(struct sctp_asconf_chunk),
500 &p6_buf.ph, sizeof(*p6));
501 if (p6 == NULL) {
502 #ifdef SCTP_DEBUG
503 if (sctp_debug_on & SCTP_DEBUG_INPUT3) {
504 kprintf("sctp_process_control: failed to get asconf v6 lookup addr\n");
506 #endif /* SCTP_DEBUG */
507 return (NULL);
509 sin6 = (struct sockaddr_in6 *)&remote_store;
510 sin6->sin6_family = AF_INET6;
511 sin6->sin6_len = sizeof(*sin6);
512 sin6->sin6_port = sh->src_port;
513 memcpy(&sin6->sin6_addr, &p6->addr, sizeof(struct in6_addr));
514 } else if (ptype == SCTP_IPV4_ADDRESS) {
515 /* ipv4 address param */
516 struct sctp_ipv4addr_param *p4, p4_buf;
517 if (ntohs(phdr->param_length) != sizeof(struct sctp_ipv4addr_param)) {
518 return NULL;
521 p4 = (struct sctp_ipv4addr_param *)sctp_get_next_param(m,
522 offset + sizeof(struct sctp_asconf_chunk),
523 &p4_buf.ph, sizeof(*p4));
524 if (p4 == NULL) {
525 #ifdef SCTP_DEBUG
526 if (sctp_debug_on & SCTP_DEBUG_INPUT3) {
527 kprintf("sctp_process_control: failed to get asconf v4 lookup addr\n");
529 #endif /* SCTP_DEBUG */
530 return (NULL);
532 sin = (struct sockaddr_in *)&remote_store;
533 sin->sin_family = AF_INET;
534 sin->sin_len = sizeof(*sin);
535 sin->sin_port = sh->src_port;
536 memcpy(&sin->sin_addr, &p4->addr, sizeof(struct in_addr));
537 } else {
538 /* invalid address param type */
539 return NULL;
542 stcb = sctp_findassociation_ep_addr(inp_p,
543 (struct sockaddr *)&remote_store, netp,
544 (struct sockaddr *)&local_store, NULL);
545 return (stcb);
548 struct sctp_tcb *
549 sctp_findassociation_ep_addr(struct sctp_inpcb **inp_p, struct sockaddr *remote,
550 struct sctp_nets **netp, struct sockaddr *local, struct sctp_tcb *locked_tcb)
552 struct sctpasochead *head;
553 struct sctp_inpcb *inp;
554 struct sctp_tcb *stcb;
555 struct sctp_nets *net;
556 uint16_t rport;
558 inp = *inp_p;
559 if (remote->sa_family == AF_INET) {
560 rport = (((struct sockaddr_in *)remote)->sin_port);
561 } else if (remote->sa_family == AF_INET6) {
562 rport = (((struct sockaddr_in6 *)remote)->sin6_port);
563 } else {
564 return (NULL);
566 if (locked_tcb) {
567 /* UN-lock so we can do proper locking here
568 * this occurs when called from load_addresses_from_init.
570 SCTP_TCB_UNLOCK(locked_tcb);
572 SCTP_INP_INFO_RLOCK();
573 if (inp->sctp_flags & SCTP_PCB_FLAGS_TCPTYPE) {
575 * Now either this guy is our listner or it's the connector.
576 * If it is the one that issued the connect, then it's only
577 * chance is to be the first TCB in the list. If it is the
578 * acceptor, then do the special_lookup to hash and find the
579 * real inp.
581 if (inp->sctp_flags & SCTP_PCB_FLAGS_ACCEPTING) {
582 /* to is peer addr, from is my addr */
583 stcb = sctp_tcb_special_locate(inp_p, remote, local,
584 netp);
585 if ((stcb != NULL) && (locked_tcb == NULL)){
586 /* we have a locked tcb, lower refcount */
587 SCTP_INP_WLOCK(inp);
588 SCTP_INP_DECR_REF(inp);
589 SCTP_INP_WUNLOCK(inp);
591 if (locked_tcb != NULL) {
592 SCTP_INP_RLOCK(locked_tcb->sctp_ep);
593 SCTP_TCB_LOCK(locked_tcb);
594 SCTP_INP_RUNLOCK(locked_tcb->sctp_ep);
595 if (stcb != NULL)
596 SCTP_TCB_UNLOCK(stcb);
598 SCTP_INP_INFO_RUNLOCK();
599 return (stcb);
600 } else {
601 SCTP_INP_WLOCK(inp);
602 stcb = LIST_FIRST(&inp->sctp_asoc_list);
603 if (stcb == NULL) {
604 goto null_return;
606 SCTP_TCB_LOCK(stcb);
607 if (stcb->rport != rport) {
608 /* remote port does not match. */
609 SCTP_TCB_UNLOCK(stcb);
610 goto null_return;
612 /* now look at the list of remote addresses */
613 TAILQ_FOREACH(net, &stcb->asoc.nets, sctp_next) {
614 if (net->ro._l_addr.sa.sa_family !=
615 remote->sa_family) {
616 /* not the same family */
617 continue;
619 if (remote->sa_family == AF_INET) {
620 struct sockaddr_in *sin, *rsin;
621 sin = (struct sockaddr_in *)
622 &net->ro._l_addr;
623 rsin = (struct sockaddr_in *)remote;
624 if (sin->sin_addr.s_addr ==
625 rsin->sin_addr.s_addr) {
626 /* found it */
627 if (netp != NULL) {
628 *netp = net;
630 if (locked_tcb == NULL) {
631 SCTP_INP_DECR_REF(inp);
633 SCTP_INP_WUNLOCK(inp);
634 SCTP_INP_INFO_RUNLOCK();
635 return (stcb);
637 } else if (remote->sa_family == AF_INET6) {
638 struct sockaddr_in6 *sin6, *rsin6;
639 sin6 = (struct sockaddr_in6 *)&net->ro._l_addr;
640 rsin6 = (struct sockaddr_in6 *)remote;
641 if (SCTP6_ARE_ADDR_EQUAL(&sin6->sin6_addr,
642 &rsin6->sin6_addr)) {
643 /* found it */
644 if (netp != NULL) {
645 *netp = net;
647 if (locked_tcb == NULL) {
648 SCTP_INP_DECR_REF(inp);
650 SCTP_INP_WUNLOCK(inp);
651 SCTP_INP_INFO_RUNLOCK();
652 return (stcb);
656 SCTP_TCB_UNLOCK(stcb);
658 } else {
659 SCTP_INP_WLOCK(inp);
660 head = &inp->sctp_tcbhash[SCTP_PCBHASH_ALLADDR(rport,
661 inp->sctp_hashmark)];
662 if (head == NULL) {
663 goto null_return;
665 LIST_FOREACH(stcb, head, sctp_tcbhash) {
666 if (stcb->rport != rport) {
667 /* remote port does not match */
668 continue;
670 /* now look at the list of remote addresses */
671 SCTP_TCB_LOCK(stcb);
672 TAILQ_FOREACH(net, &stcb->asoc.nets, sctp_next) {
673 if (net->ro._l_addr.sa.sa_family !=
674 remote->sa_family) {
675 /* not the same family */
676 continue;
678 if (remote->sa_family == AF_INET) {
679 struct sockaddr_in *sin, *rsin;
680 sin = (struct sockaddr_in *)
681 &net->ro._l_addr;
682 rsin = (struct sockaddr_in *)remote;
683 if (sin->sin_addr.s_addr ==
684 rsin->sin_addr.s_addr) {
685 /* found it */
686 if (netp != NULL) {
687 *netp = net;
689 if (locked_tcb == NULL) {
690 SCTP_INP_DECR_REF(inp);
692 SCTP_INP_WUNLOCK(inp);
693 SCTP_INP_INFO_RUNLOCK();
694 return (stcb);
696 } else if (remote->sa_family == AF_INET6) {
697 struct sockaddr_in6 *sin6, *rsin6;
698 sin6 = (struct sockaddr_in6 *)
699 &net->ro._l_addr;
700 rsin6 = (struct sockaddr_in6 *)remote;
701 if (SCTP6_ARE_ADDR_EQUAL(&sin6->sin6_addr,
702 &rsin6->sin6_addr)) {
703 /* found it */
704 if (netp != NULL) {
705 *netp = net;
707 if (locked_tcb == NULL) {
708 SCTP_INP_DECR_REF(inp);
710 SCTP_INP_WUNLOCK(inp);
711 SCTP_INP_INFO_RUNLOCK();
712 return (stcb);
716 SCTP_TCB_UNLOCK(stcb);
719 null_return:
720 /* clean up for returning null */
721 if (locked_tcb){
722 if (locked_tcb->sctp_ep != inp) {
723 SCTP_INP_RLOCK(locked_tcb->sctp_ep);
724 SCTP_TCB_LOCK(locked_tcb);
725 SCTP_INP_RUNLOCK(locked_tcb->sctp_ep);
726 } else
727 SCTP_TCB_LOCK(locked_tcb);
729 SCTP_INP_WUNLOCK(inp);
730 SCTP_INP_INFO_RUNLOCK();
731 /* not found */
732 return (NULL);
736 * Find an association for a specific endpoint using the association id
737 * given out in the COMM_UP notification
739 struct sctp_tcb *
740 sctp_findassociation_ep_asocid(struct sctp_inpcb *inp, caddr_t asoc_id)
743 * Use my the assoc_id to find a endpoint
745 struct sctpasochead *head;
746 struct sctp_tcb *stcb;
747 u_int32_t vtag;
749 if (asoc_id == 0 || inp == NULL) {
750 return (NULL);
752 SCTP_INP_INFO_RLOCK();
753 vtag = (u_int32_t)(uintptr_t)asoc_id;
754 head = &sctppcbinfo.sctp_asochash[SCTP_PCBHASH_ASOC(vtag,
755 sctppcbinfo.hashasocmark)];
756 if (head == NULL) {
757 /* invalid vtag */
758 SCTP_INP_INFO_RUNLOCK();
759 return (NULL);
761 LIST_FOREACH(stcb, head, sctp_asocs) {
762 SCTP_INP_RLOCK(stcb->sctp_ep);
763 SCTP_TCB_LOCK(stcb);
764 SCTP_INP_RUNLOCK(stcb->sctp_ep);
765 if (stcb->asoc.my_vtag == vtag) {
766 /* candidate */
767 if (inp != stcb->sctp_ep) {
768 /* some other guy has the
769 * same vtag active (vtag collision).
771 sctp_pegs[SCTP_VTAG_BOGUS]++;
772 SCTP_TCB_UNLOCK(stcb);
773 continue;
775 sctp_pegs[SCTP_VTAG_EXPR]++;
776 SCTP_INP_INFO_RUNLOCK();
777 return (stcb);
779 SCTP_TCB_UNLOCK(stcb);
781 SCTP_INP_INFO_RUNLOCK();
782 return (NULL);
785 static struct sctp_inpcb *
786 sctp_endpoint_probe(struct sockaddr *nam, struct sctppcbhead *head,
787 uint16_t lport)
789 struct sctp_inpcb *inp;
790 struct sockaddr_in *sin;
791 struct sockaddr_in6 *sin6;
792 struct sctp_laddr *laddr;
794 /* Endpoing probe expects
795 * that the INP_INFO is locked.
797 if (nam->sa_family == AF_INET) {
798 sin = (struct sockaddr_in *)nam;
799 sin6 = NULL;
800 } else if (nam->sa_family == AF_INET6) {
801 sin6 = (struct sockaddr_in6 *)nam;
802 sin = NULL;
803 } else {
804 /* unsupported family */
805 return (NULL);
807 if (head == NULL)
808 return (NULL);
810 LIST_FOREACH(inp, head, sctp_hash) {
811 SCTP_INP_RLOCK(inp);
813 if ((inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) &&
814 (inp->sctp_lport == lport)) {
815 /* got it */
816 if ((nam->sa_family == AF_INET) &&
817 (inp->sctp_flags & SCTP_PCB_FLAGS_BOUND_V6) &&
818 #if defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__)
819 (((struct inpcb *)inp)->inp_flags & IN6P_IPV6_V6ONLY)
820 #else
821 #if defined(__OpenBSD__)
822 (0) /* For open bsd we do dual bind only */
823 #else
824 (((struct in6pcb *)inp)->in6p_flags & IN6P_IPV6_V6ONLY)
825 #endif
826 #endif
828 /* IPv4 on a IPv6 socket with ONLY IPv6 set */
829 SCTP_INP_RUNLOCK(inp);
830 continue;
832 /* A V6 address and the endpoint is NOT bound V6 */
833 if (nam->sa_family == AF_INET6 &&
834 (inp->sctp_flags & SCTP_PCB_FLAGS_BOUND_V6) == 0) {
835 SCTP_INP_RUNLOCK(inp);
836 continue;
838 SCTP_INP_RUNLOCK(inp);
839 return (inp);
841 SCTP_INP_RUNLOCK(inp);
844 if ((nam->sa_family == AF_INET) &&
845 (sin->sin_addr.s_addr == INADDR_ANY)) {
846 /* Can't hunt for one that has no address specified */
847 return (NULL);
848 } else if ((nam->sa_family == AF_INET6) &&
849 (IN6_IS_ADDR_UNSPECIFIED(&sin6->sin6_addr))) {
850 /* Can't hunt for one that has no address specified */
851 return (NULL);
854 * ok, not bound to all so see if we can find a EP bound to this
855 * address.
857 #ifdef SCTP_DEBUG
858 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
859 kprintf("Ok, there is NO bound-all available for port:%x\n", ntohs(lport));
861 #endif
862 LIST_FOREACH(inp, head, sctp_hash) {
863 SCTP_INP_RLOCK(inp);
864 if ((inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL)) {
865 SCTP_INP_RUNLOCK(inp);
866 continue;
869 * Ok this could be a likely candidate, look at all of
870 * its addresses
872 if (inp->sctp_lport != lport) {
873 SCTP_INP_RUNLOCK(inp);
874 continue;
876 #ifdef SCTP_DEBUG
877 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
878 kprintf("Ok, found matching local port\n");
880 #endif
881 LIST_FOREACH(laddr, &inp->sctp_addr_list, sctp_nxt_addr) {
882 if (laddr->ifa == NULL) {
883 #ifdef SCTP_DEBUG
884 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
885 kprintf("An ounce of prevention is worth a pound of cure\n");
887 #endif
888 continue;
890 #ifdef SCTP_DEBUG
891 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
892 kprintf("Ok laddr->ifa:%p is possible, ",
893 laddr->ifa);
895 #endif
896 if (laddr->ifa->ifa_addr == NULL) {
897 #ifdef SCTP_DEBUG
898 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
899 kprintf("Huh IFA as an ifa_addr=NULL, ");
901 #endif
902 continue;
904 #ifdef SCTP_DEBUG
905 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
906 kprintf("Ok laddr->ifa:%p is possible, ",
907 laddr->ifa->ifa_addr);
908 sctp_print_address(laddr->ifa->ifa_addr);
909 kprintf("looking for ");
910 sctp_print_address(nam);
912 #endif
913 if (laddr->ifa->ifa_addr->sa_family == nam->sa_family) {
914 /* possible, see if it matches */
915 struct sockaddr_in *intf_addr;
916 intf_addr = (struct sockaddr_in *)
917 laddr->ifa->ifa_addr;
918 if (nam->sa_family == AF_INET) {
919 if (sin->sin_addr.s_addr ==
920 intf_addr->sin_addr.s_addr) {
921 #ifdef SCTP_DEBUG
922 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
923 kprintf("YES, return ep:%p\n", inp);
925 #endif
926 SCTP_INP_RUNLOCK(inp);
927 return (inp);
929 } else if (nam->sa_family == AF_INET6) {
930 struct sockaddr_in6 *intf_addr6;
931 intf_addr6 = (struct sockaddr_in6 *)
932 laddr->ifa->ifa_addr;
933 if (SCTP6_ARE_ADDR_EQUAL(&sin6->sin6_addr,
934 &intf_addr6->sin6_addr)) {
935 #ifdef SCTP_DEBUG
936 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
937 kprintf("YES, return ep:%p\n", inp);
939 #endif
940 SCTP_INP_RUNLOCK(inp);
941 return (inp);
945 SCTP_INP_RUNLOCK(inp);
948 #ifdef SCTP_DEBUG
949 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
950 kprintf("NO, Falls out to NULL\n");
952 #endif
953 return (NULL);
957 struct sctp_inpcb *
958 sctp_pcb_findep(struct sockaddr *nam, int find_tcp_pool, int have_lock)
961 * First we check the hash table to see if someone has this port
962 * bound with just the port.
964 struct sctp_inpcb *inp;
965 struct sctppcbhead *head;
966 struct sockaddr_in *sin;
967 struct sockaddr_in6 *sin6;
968 int lport;
969 #ifdef SCTP_DEBUG
970 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
971 kprintf("Looking for endpoint %d :",
972 ntohs(((struct sockaddr_in *)nam)->sin_port));
973 sctp_print_address(nam);
975 #endif
976 if (nam->sa_family == AF_INET) {
977 sin = (struct sockaddr_in *)nam;
978 lport = ((struct sockaddr_in *)nam)->sin_port;
979 } else if (nam->sa_family == AF_INET6) {
980 sin6 = (struct sockaddr_in6 *)nam;
981 lport = ((struct sockaddr_in6 *)nam)->sin6_port;
982 } else {
983 /* unsupported family */
984 return (NULL);
987 * I could cheat here and just cast to one of the types but we will
988 * do it right. It also provides the check against an Unsupported
989 * type too.
991 /* Find the head of the ALLADDR chain */
992 if (have_lock == 0)
993 SCTP_INP_INFO_RLOCK();
994 head = &sctppcbinfo.sctp_ephash[SCTP_PCBHASH_ALLADDR(lport,
995 sctppcbinfo.hashmark)];
996 #ifdef SCTP_DEBUG
997 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
998 kprintf("Main hash to lookup at head:%p\n", head);
1000 #endif
1001 inp = sctp_endpoint_probe(nam, head, lport);
1004 * If the TCP model exists it could be that the main listening
1005 * endpoint is gone but there exists a connected socket for this
1006 * guy yet. If so we can return the first one that we find. This
1007 * may NOT be the correct one but the sctp_findassociation_ep_addr
1008 * has further code to look at all TCP models.
1010 if (inp == NULL && find_tcp_pool) {
1011 unsigned int i;
1012 #ifdef SCTP_DEBUG
1013 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1014 kprintf("EP was NULL and TCP model is supported\n");
1016 #endif
1017 for (i = 0; i < sctppcbinfo.hashtblsize; i++) {
1019 * This is real gross, but we do NOT have a remote
1020 * port at this point depending on who is calling. We
1021 * must therefore look for ANY one that matches our
1022 * local port :/
1024 head = &sctppcbinfo.sctp_tcpephash[i];
1025 if (LIST_FIRST(head)) {
1026 inp = sctp_endpoint_probe(nam, head, lport);
1027 if (inp) {
1028 /* Found one */
1029 break;
1034 #ifdef SCTP_DEBUG
1035 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1036 kprintf("EP to return is %p\n", inp);
1038 #endif
1039 if (have_lock == 0) {
1040 if (inp) {
1041 SCTP_INP_WLOCK(inp);
1042 SCTP_INP_INCR_REF(inp);
1043 SCTP_INP_WUNLOCK(inp);
1045 SCTP_INP_INFO_RUNLOCK();
1046 } else {
1047 if (inp) {
1048 SCTP_INP_WLOCK(inp);
1049 SCTP_INP_INCR_REF(inp);
1050 SCTP_INP_WUNLOCK(inp);
1053 return (inp);
1057 * Find an association for an endpoint with the pointer to whom you want
1058 * to send to and the endpoint pointer. The address can be IPv4 or IPv6.
1059 * We may need to change the *to to some other struct like a mbuf...
1061 struct sctp_tcb *
1062 sctp_findassociation_addr_sa(struct sockaddr *to, struct sockaddr *from,
1063 struct sctp_inpcb **inp_p, struct sctp_nets **netp, int find_tcp_pool)
1065 struct sctp_inpcb *inp;
1066 struct sctp_tcb *retval;
1068 SCTP_INP_INFO_RLOCK();
1069 if (find_tcp_pool) {
1070 if (inp_p != NULL) {
1071 retval = sctp_tcb_special_locate(inp_p, from, to, netp);
1072 } else {
1073 retval = sctp_tcb_special_locate(&inp, from, to, netp);
1075 if (retval != NULL) {
1076 SCTP_INP_INFO_RUNLOCK();
1077 return (retval);
1080 inp = sctp_pcb_findep(to, 0, 1);
1081 if (inp_p != NULL) {
1082 *inp_p = inp;
1084 SCTP_INP_INFO_RUNLOCK();
1086 if (inp == NULL) {
1087 return (NULL);
1091 * ok, we have an endpoint, now lets find the assoc for it (if any)
1092 * we now place the source address or from in the to of the find
1093 * endpoint call. Since in reality this chain is used from the
1094 * inbound packet side.
1096 if (inp_p != NULL) {
1097 return (sctp_findassociation_ep_addr(inp_p, from, netp, to, NULL));
1098 } else {
1099 return (sctp_findassociation_ep_addr(&inp, from, netp, to, NULL));
1105 * This routine will grub through the mbuf that is a INIT or INIT-ACK and
1106 * find all addresses that the sender has specified in any address list.
1107 * Each address will be used to lookup the TCB and see if one exits.
1109 static struct sctp_tcb *
1110 sctp_findassociation_special_addr(struct mbuf *m, int iphlen, int offset,
1111 struct sctphdr *sh, struct sctp_inpcb **inp_p, struct sctp_nets **netp,
1112 struct sockaddr *dest)
1114 struct sockaddr_in sin4;
1115 struct sockaddr_in6 sin6;
1116 struct sctp_paramhdr *phdr, parm_buf;
1117 struct sctp_tcb *retval;
1118 u_int32_t ptype, plen;
1120 memset(&sin4, 0, sizeof(sin4));
1121 memset(&sin6, 0, sizeof(sin6));
1122 sin4.sin_len = sizeof(sin4);
1123 sin4.sin_family = AF_INET;
1124 sin4.sin_port = sh->src_port;
1125 sin6.sin6_len = sizeof(sin6);
1126 sin6.sin6_family = AF_INET6;
1127 sin6.sin6_port = sh->src_port;
1129 retval = NULL;
1130 offset += sizeof(struct sctp_init_chunk);
1132 phdr = sctp_get_next_param(m, offset, &parm_buf, sizeof(parm_buf));
1133 while (phdr != NULL) {
1134 /* now we must see if we want the parameter */
1135 ptype = ntohs(phdr->param_type);
1136 plen = ntohs(phdr->param_length);
1137 if (plen == 0) {
1138 #ifdef SCTP_DEBUG
1139 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1140 kprintf("sctp_findassociation_special_addr: Impossible length in parameter\n");
1142 #endif /* SCTP_DEBUG */
1143 break;
1145 if (ptype == SCTP_IPV4_ADDRESS &&
1146 plen == sizeof(struct sctp_ipv4addr_param)) {
1147 /* Get the rest of the address */
1148 struct sctp_ipv4addr_param ip4_parm, *p4;
1150 phdr = sctp_get_next_param(m, offset,
1151 (struct sctp_paramhdr *)&ip4_parm, plen);
1152 if (phdr == NULL) {
1153 return (NULL);
1155 p4 = (struct sctp_ipv4addr_param *)phdr;
1156 memcpy(&sin4.sin_addr, &p4->addr, sizeof(p4->addr));
1157 /* look it up */
1158 retval = sctp_findassociation_ep_addr(inp_p,
1159 (struct sockaddr *)&sin4, netp, dest, NULL);
1160 if (retval != NULL) {
1161 return (retval);
1163 } else if (ptype == SCTP_IPV6_ADDRESS &&
1164 plen == sizeof(struct sctp_ipv6addr_param)) {
1165 /* Get the rest of the address */
1166 struct sctp_ipv6addr_param ip6_parm, *p6;
1168 phdr = sctp_get_next_param(m, offset,
1169 (struct sctp_paramhdr *)&ip6_parm, plen);
1170 if (phdr == NULL) {
1171 return (NULL);
1173 p6 = (struct sctp_ipv6addr_param *)phdr;
1174 memcpy(&sin6.sin6_addr, &p6->addr, sizeof(p6->addr));
1175 /* look it up */
1176 retval = sctp_findassociation_ep_addr(inp_p,
1177 (struct sockaddr *)&sin6, netp, dest, NULL);
1178 if (retval != NULL) {
1179 return (retval);
1182 offset += SCTP_SIZE32(plen);
1183 phdr = sctp_get_next_param(m, offset, &parm_buf,
1184 sizeof(parm_buf));
1186 return (NULL);
1189 static struct sctp_tcb *
1190 sctp_findassoc_by_vtag(struct sockaddr *from, uint32_t vtag,
1191 struct sctp_inpcb **inp_p, struct sctp_nets **netp, uint16_t rport,
1192 uint16_t lport)
1195 * Use my vtag to hash. If we find it we then verify the source addr
1196 * is in the assoc. If all goes well we save a bit on rec of a packet.
1198 struct sctpasochead *head;
1199 struct sctp_nets *net;
1200 struct sctp_tcb *stcb;
1202 SCTP_INP_INFO_RLOCK();
1203 head = &sctppcbinfo.sctp_asochash[SCTP_PCBHASH_ASOC(vtag,
1204 sctppcbinfo.hashasocmark)];
1205 if (head == NULL) {
1206 /* invalid vtag */
1207 SCTP_INP_INFO_RUNLOCK();
1208 return (NULL);
1210 LIST_FOREACH(stcb, head, sctp_asocs) {
1211 SCTP_INP_RLOCK(stcb->sctp_ep);
1212 SCTP_TCB_LOCK(stcb);
1213 SCTP_INP_RUNLOCK(stcb->sctp_ep);
1214 if (stcb->asoc.my_vtag == vtag) {
1215 /* candidate */
1216 if (stcb->rport != rport) {
1218 * we could remove this if vtags are unique
1219 * across the system.
1221 SCTP_TCB_UNLOCK(stcb);
1222 continue;
1224 if (stcb->sctp_ep->sctp_lport != lport) {
1226 * we could remove this if vtags are unique
1227 * across the system.
1229 SCTP_TCB_UNLOCK(stcb);
1230 continue;
1232 net = sctp_findnet(stcb, from);
1233 if (net) {
1234 /* yep its him. */
1235 *netp = net;
1236 sctp_pegs[SCTP_VTAG_EXPR]++;
1237 *inp_p = stcb->sctp_ep;
1238 SCTP_INP_INFO_RUNLOCK();
1239 return (stcb);
1240 } else {
1241 /* not him, this should only
1242 * happen in rare cases so
1243 * I peg it.
1245 sctp_pegs[SCTP_VTAG_BOGUS]++;
1248 SCTP_TCB_UNLOCK(stcb);
1250 SCTP_INP_INFO_RUNLOCK();
1251 return (NULL);
1255 * Find an association with the pointer to the inbound IP packet. This
1256 * can be a IPv4 or IPv6 packet.
1258 struct sctp_tcb *
1259 sctp_findassociation_addr(struct mbuf *m, int iphlen, int offset,
1260 struct sctphdr *sh, struct sctp_chunkhdr *ch,
1261 struct sctp_inpcb **inp_p, struct sctp_nets **netp)
1263 int find_tcp_pool;
1264 struct ip *iph;
1265 struct sctp_tcb *retval;
1266 struct sockaddr_storage to_store, from_store;
1267 struct sockaddr *to = (struct sockaddr *)&to_store;
1268 struct sockaddr *from = (struct sockaddr *)&from_store;
1269 struct sctp_inpcb *inp;
1272 iph = mtod(m, struct ip *);
1273 if (iph->ip_v == IPVERSION) {
1274 /* its IPv4 */
1275 struct sockaddr_in *to4, *from4;
1277 to4 = (struct sockaddr_in *)&to_store;
1278 from4 = (struct sockaddr_in *)&from_store;
1279 bzero(to4, sizeof(*to4));
1280 bzero(from4, sizeof(*from4));
1281 from4->sin_family = to4->sin_family = AF_INET;
1282 from4->sin_len = to4->sin_len = sizeof(struct sockaddr_in);
1283 from4->sin_addr.s_addr = iph->ip_src.s_addr;
1284 to4->sin_addr.s_addr = iph->ip_dst.s_addr ;
1285 from4->sin_port = sh->src_port;
1286 to4->sin_port = sh->dest_port;
1287 } else if (iph->ip_v == (IPV6_VERSION >> 4)) {
1288 /* its IPv6 */
1289 struct ip6_hdr *ip6;
1290 struct sockaddr_in6 *to6, *from6;
1292 ip6 = mtod(m, struct ip6_hdr *);
1293 to6 = (struct sockaddr_in6 *)&to_store;
1294 from6 = (struct sockaddr_in6 *)&from_store;
1295 bzero(to6, sizeof(*to6));
1296 bzero(from6, sizeof(*from6));
1297 from6->sin6_family = to6->sin6_family = AF_INET6;
1298 from6->sin6_len = to6->sin6_len = sizeof(struct sockaddr_in6);
1299 to6->sin6_addr = ip6->ip6_dst;
1300 from6->sin6_addr = ip6->ip6_src;
1301 from6->sin6_port = sh->src_port;
1302 to6->sin6_port = sh->dest_port;
1303 /* Get the scopes in properly to the sin6 addr's */
1304 in6_recoverscope(to6, &to6->sin6_addr, NULL);
1305 #if defined(SCTP_BASE_FREEBSD) || defined(__APPLE__) || defined(__DragonFly__)
1306 in6_embedscope(&to6->sin6_addr, to6, NULL, NULL);
1307 #else
1308 in6_embedscope(&to6->sin6_addr, to6);
1309 #endif
1311 in6_recoverscope(from6, &from6->sin6_addr, NULL);
1312 #if defined(SCTP_BASE_FREEBSD) || defined(__APPLE__) || defined(__DragonFly__)
1313 in6_embedscope(&from6->sin6_addr, from6, NULL, NULL);
1314 #else
1315 in6_embedscope(&from6->sin6_addr, from6);
1316 #endif
1317 } else {
1318 /* Currently not supported. */
1319 return (NULL);
1321 #ifdef SCTP_DEBUG
1322 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1323 kprintf("Looking for port %d address :",
1324 ntohs(((struct sockaddr_in *)to)->sin_port));
1325 sctp_print_address(to);
1326 kprintf("From for port %d address :",
1327 ntohs(((struct sockaddr_in *)from)->sin_port));
1328 sctp_print_address(from);
1330 #endif
1332 if (sh->v_tag) {
1333 /* we only go down this path if vtag is non-zero */
1334 retval = sctp_findassoc_by_vtag(from, ntohl(sh->v_tag),
1335 inp_p, netp, sh->src_port, sh->dest_port);
1336 if (retval) {
1337 return (retval);
1340 find_tcp_pool = 0;
1341 if ((ch->chunk_type != SCTP_INITIATION) &&
1342 (ch->chunk_type != SCTP_INITIATION_ACK) &&
1343 (ch->chunk_type != SCTP_COOKIE_ACK) &&
1344 (ch->chunk_type != SCTP_COOKIE_ECHO)) {
1345 /* Other chunk types go to the tcp pool. */
1346 find_tcp_pool = 1;
1348 if (inp_p) {
1349 retval = sctp_findassociation_addr_sa(to, from, inp_p, netp,
1350 find_tcp_pool);
1351 inp = *inp_p;
1352 } else {
1353 retval = sctp_findassociation_addr_sa(to, from, &inp, netp,
1354 find_tcp_pool);
1356 #ifdef SCTP_DEBUG
1357 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1358 kprintf("retval:%p inp:%p\n", retval, inp);
1360 #endif
1361 if (retval == NULL && inp) {
1362 /* Found a EP but not this address */
1363 #ifdef SCTP_DEBUG
1364 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1365 kprintf("Found endpoint %p but no asoc - ep state:%x\n",
1366 inp, inp->sctp_flags);
1368 #endif
1369 if ((ch->chunk_type == SCTP_INITIATION) ||
1370 (ch->chunk_type == SCTP_INITIATION_ACK)) {
1372 * special hook, we do NOT return linp or an
1373 * association that is linked to an existing
1374 * association that is under the TCP pool (i.e. no
1375 * listener exists). The endpoint finding routine
1376 * will always find a listner before examining the
1377 * TCP pool.
1379 if (inp->sctp_flags & SCTP_PCB_FLAGS_IN_TCPPOOL) {
1380 #ifdef SCTP_DEBUG
1381 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1382 kprintf("Gak, its in the TCP pool... return NULL");
1384 #endif
1385 if (inp_p) {
1386 *inp_p = NULL;
1388 return (NULL);
1390 #ifdef SCTP_DEBUG
1391 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1392 kprintf("Now doing SPECIAL find\n");
1394 #endif
1395 retval = sctp_findassociation_special_addr(m, iphlen,
1396 offset, sh, inp_p, netp, to);
1399 #ifdef SCTP_DEBUG
1400 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1401 kprintf("retval is %p\n", retval);
1403 #endif
1404 return (retval);
1407 extern int sctp_max_burst_default;
1409 extern unsigned int sctp_delayed_sack_time_default;
1410 extern unsigned int sctp_heartbeat_interval_default;
1411 extern unsigned int sctp_pmtu_raise_time_default;
1412 extern unsigned int sctp_shutdown_guard_time_default;
1413 extern unsigned int sctp_secret_lifetime_default;
1415 extern unsigned int sctp_rto_max_default;
1416 extern unsigned int sctp_rto_min_default;
1417 extern unsigned int sctp_rto_initial_default;
1418 extern unsigned int sctp_init_rto_max_default;
1419 extern unsigned int sctp_valid_cookie_life_default;
1420 extern unsigned int sctp_init_rtx_max_default;
1421 extern unsigned int sctp_assoc_rtx_max_default;
1422 extern unsigned int sctp_path_rtx_max_default;
1423 extern unsigned int sctp_nr_outgoing_streams_default;
1426 * allocate a sctp_inpcb and setup a temporary binding to a port/all
1427 * addresses. This way if we don't get a bind we by default pick a ephemeral
1428 * port with all addresses bound.
1431 sctp_inpcb_alloc(struct socket *so)
1434 * we get called when a new endpoint starts up. We need to allocate
1435 * the sctp_inpcb structure from the zone and init it. Mark it as
1436 * unbound and find a port that we can use as an ephemeral with
1437 * INADDR_ANY. If the user binds later no problem we can then add
1438 * in the specific addresses. And setup the default parameters for
1439 * the EP.
1441 int i, error;
1442 struct sctp_inpcb *inp, *n_inp;
1443 struct sctp_pcb *m;
1444 struct timeval time;
1446 error = 0;
1448 /* Hack alert:
1450 * This code audits the entire INP list to see if
1451 * any ep's that are in the GONE state are now
1452 * all free. This should not happen really since when
1453 * the last association if freed we should end up deleting
1454 * the inpcb. This code including the locks should
1455 * be taken out ... since the last set of fixes I
1456 * have not seen the "Found a GONE on list" has not
1457 * came out. But i am paranoid and we will leave this
1458 * in at the cost of efficency on allocation of PCB's.
1459 * Probably we should move this to the invariant
1460 * compile options
1462 /* #ifdef INVARIANTS*/
1463 SCTP_INP_INFO_RLOCK();
1464 inp = LIST_FIRST(&sctppcbinfo.listhead);
1465 while (inp) {
1466 n_inp = LIST_NEXT(inp, sctp_list);
1467 if (inp->sctp_flags & SCTP_PCB_FLAGS_SOCKET_GONE) {
1468 if (LIST_FIRST(&inp->sctp_asoc_list) == NULL) {
1469 /* finish the job now */
1470 kprintf("Found a GONE on list\n");
1471 SCTP_INP_INFO_RUNLOCK();
1472 sctp_inpcb_free(inp, 1);
1473 SCTP_INP_INFO_RLOCK();
1476 inp = n_inp;
1478 SCTP_INP_INFO_RUNLOCK();
1479 /* #endif INVARIANTS*/
1481 SCTP_INP_INFO_WLOCK();
1482 inp = (struct sctp_inpcb *)SCTP_ZONE_GET(sctppcbinfo.ipi_zone_ep);
1483 if (inp == NULL) {
1484 kprintf("Out of SCTP-INPCB structures - no resources\n");
1485 SCTP_INP_INFO_WUNLOCK();
1486 return (ENOBUFS);
1489 /* zap it */
1490 bzero(inp, sizeof(*inp));
1492 /* bump generations */
1493 inp->ip_inp.inp.inp_socket = so;
1495 /* setup socket pointers */
1496 inp->sctp_socket = so;
1498 /* setup inpcb socket too */
1499 inp->ip_inp.inp.inp_socket = so;
1500 inp->sctp_frag_point = SCTP_DEFAULT_MAXSEGMENT;
1501 #ifdef IPSEC
1502 #if !(defined(__OpenBSD__) || defined(__APPLE__))
1504 struct inpcbpolicy *pcb_sp = NULL;
1505 error = ipsec_init_policy(so, &pcb_sp);
1506 /* Arrange to share the policy */
1507 inp->ip_inp.inp.inp_sp = pcb_sp;
1508 ((struct in6pcb *)(&inp->ip_inp.inp))->in6p_sp = pcb_sp;
1510 #else
1511 /* not sure what to do for openbsd here */
1512 error = 0;
1513 #endif
1514 if (error != 0) {
1515 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_ep, inp);
1516 SCTP_INP_INFO_WUNLOCK();
1517 return error;
1519 #endif /* IPSEC */
1520 sctppcbinfo.ipi_count_ep++;
1521 #if defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__)
1522 inp->ip_inp.inp.inp_gencnt = ++sctppcbinfo.ipi_gencnt_ep;
1523 inp->ip_inp.inp.inp_ip_ttl = ip_defttl;
1524 #else
1525 inp->inp_ip_ttl = ip_defttl;
1526 inp->inp_ip_tos = 0;
1527 #endif
1529 so->so_pcb = (caddr_t)inp;
1531 if ((so->so_type == SOCK_DGRAM) ||
1532 (so->so_type == SOCK_SEQPACKET)) {
1533 /* UDP style socket */
1534 inp->sctp_flags = (SCTP_PCB_FLAGS_UDPTYPE |
1535 SCTP_PCB_FLAGS_UNBOUND);
1536 inp->sctp_flags |= (SCTP_PCB_FLAGS_RECVDATAIOEVNT);
1537 } else if (so->so_type == SOCK_STREAM) {
1538 /* TCP style socket */
1539 inp->sctp_flags = (SCTP_PCB_FLAGS_TCPTYPE |
1540 SCTP_PCB_FLAGS_UNBOUND);
1541 inp->sctp_flags |= (SCTP_PCB_FLAGS_RECVDATAIOEVNT);
1542 } else {
1544 * unsupported socket type (RAW, etc)- in case we missed
1545 * it in protosw
1547 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_ep, inp);
1548 SCTP_INP_INFO_WUNLOCK();
1549 return (EOPNOTSUPP);
1551 inp->sctp_tcbhash = hashinit(sctp_pcbtblsize,
1552 #ifdef __NetBSD__
1553 HASH_LIST,
1554 #endif
1555 M_PCB,
1556 #if defined(__NetBSD__) || defined(__OpenBSD__)
1557 M_WAITOK,
1558 #endif
1559 &inp->sctp_hashmark);
1560 if (inp->sctp_tcbhash == NULL) {
1561 kprintf("Out of SCTP-INPCB->hashinit - no resources\n");
1562 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_ep, inp);
1563 SCTP_INP_INFO_WUNLOCK();
1564 return (ENOBUFS);
1566 /* LOCK init's */
1567 SCTP_INP_LOCK_INIT(inp);
1568 SCTP_ASOC_CREATE_LOCK_INIT(inp);
1569 /* lock the new ep */
1570 SCTP_INP_WLOCK(inp);
1572 /* add it to the info area */
1573 LIST_INSERT_HEAD(&sctppcbinfo.listhead, inp, sctp_list);
1574 SCTP_INP_INFO_WUNLOCK();
1576 LIST_INIT(&inp->sctp_addr_list);
1577 LIST_INIT(&inp->sctp_asoc_list);
1578 TAILQ_INIT(&inp->sctp_queue_list);
1579 /* Init the timer structure for signature change */
1580 #if defined (__FreeBSD__) && __FreeBSD_version >= 500000
1581 callout_init(&inp->sctp_ep.signature_change.timer, 0);
1582 #else
1583 callout_init(&inp->sctp_ep.signature_change.timer);
1584 #endif
1585 inp->sctp_ep.signature_change.type = SCTP_TIMER_TYPE_NEWCOOKIE;
1587 /* now init the actual endpoint default data */
1588 m = &inp->sctp_ep;
1590 /* setup the base timeout information */
1591 m->sctp_timeoutticks[SCTP_TIMER_SEND] = SEC_TO_TICKS(SCTP_SEND_SEC); /* needed ? */
1592 m->sctp_timeoutticks[SCTP_TIMER_INIT] = SEC_TO_TICKS(SCTP_INIT_SEC); /* needed ? */
1593 m->sctp_timeoutticks[SCTP_TIMER_RECV] = MSEC_TO_TICKS(sctp_delayed_sack_time_default);
1594 m->sctp_timeoutticks[SCTP_TIMER_HEARTBEAT] = sctp_heartbeat_interval_default; /* this is in MSEC */
1595 m->sctp_timeoutticks[SCTP_TIMER_PMTU] = SEC_TO_TICKS(sctp_pmtu_raise_time_default);
1596 m->sctp_timeoutticks[SCTP_TIMER_MAXSHUTDOWN] = SEC_TO_TICKS(sctp_shutdown_guard_time_default);
1597 m->sctp_timeoutticks[SCTP_TIMER_SIGNATURE] = SEC_TO_TICKS(sctp_secret_lifetime_default);
1598 /* all max/min max are in ms */
1599 m->sctp_maxrto = sctp_rto_max_default;
1600 m->sctp_minrto = sctp_rto_min_default;
1601 m->initial_rto = sctp_rto_initial_default;
1602 m->initial_init_rto_max = sctp_init_rto_max_default;
1604 m->max_open_streams_intome = MAX_SCTP_STREAMS;
1606 m->max_init_times = sctp_init_rtx_max_default;
1607 m->max_send_times = sctp_assoc_rtx_max_default;
1608 m->def_net_failure = sctp_path_rtx_max_default;
1609 m->sctp_sws_sender = SCTP_SWS_SENDER_DEF;
1610 m->sctp_sws_receiver = SCTP_SWS_RECEIVER_DEF;
1611 m->max_burst = sctp_max_burst_default;
1612 /* number of streams to pre-open on a association */
1613 m->pre_open_stream_count = sctp_nr_outgoing_streams_default;
1615 /* Add adaption cookie */
1616 m->adaption_layer_indicator = 0x504C5253;
1618 /* seed random number generator */
1619 m->random_counter = 1;
1620 m->store_at = SCTP_SIGNATURE_SIZE;
1621 #if (defined(__FreeBSD__) && (__FreeBSD_version < 500000)) || defined(__DragonFly__)
1622 read_random_unlimited(m->random_numbers, sizeof(m->random_numbers));
1623 #elif defined(__APPLE__) || (__FreeBSD_version > 500000)
1624 read_random(m->random_numbers, sizeof(m->random_numbers));
1625 #elif defined(__OpenBSD__)
1626 get_random_bytes(m->random_numbers, sizeof(m->random_numbers));
1627 #elif defined(__NetBSD__) && NRND > 0
1628 rnd_extract_data(m->random_numbers, sizeof(m->random_numbers),
1629 RND_EXTRACT_ANY);
1630 #else
1632 u_int32_t *ranm, *ranp;
1633 ranp = (u_int32_t *)&m->random_numbers;
1634 ranm = ranp + (SCTP_SIGNATURE_ALOC_SIZE/sizeof(u_int32_t));
1635 if ((u_long)ranp % 4) {
1636 /* not a even boundary? */
1637 ranp = (u_int32_t *)SCTP_SIZE32((u_long)ranp);
1639 while (ranp < ranm) {
1640 *ranp = random();
1641 ranp++;
1644 #endif
1645 sctp_fill_random_store(m);
1647 /* Minimum cookie size */
1648 m->size_of_a_cookie = (sizeof(struct sctp_init_msg) * 2) +
1649 sizeof(struct sctp_state_cookie);
1650 m->size_of_a_cookie += SCTP_SIGNATURE_SIZE;
1652 /* Setup the initial secret */
1653 SCTP_GETTIME_TIMEVAL(&time);
1654 m->time_of_secret_change = time.tv_sec;
1656 for (i = 0; i < SCTP_NUMBER_OF_SECRETS; i++) {
1657 m->secret_key[0][i] = sctp_select_initial_TSN(m);
1659 sctp_timer_start(SCTP_TIMER_TYPE_NEWCOOKIE, inp, NULL, NULL);
1661 /* How long is a cookie good for ? */
1662 m->def_cookie_life = sctp_valid_cookie_life_default;
1663 SCTP_INP_WUNLOCK(inp);
1664 return (error);
1668 void
1669 sctp_move_pcb_and_assoc(struct sctp_inpcb *old_inp, struct sctp_inpcb *new_inp,
1670 struct sctp_tcb *stcb)
1672 uint16_t lport, rport;
1673 struct sctppcbhead *head;
1674 struct sctp_laddr *laddr, *oladdr;
1676 SCTP_TCB_UNLOCK(stcb);
1677 SCTP_INP_INFO_WLOCK();
1678 SCTP_INP_WLOCK(old_inp);
1679 SCTP_INP_WLOCK(new_inp);
1680 SCTP_TCB_LOCK(stcb);
1682 new_inp->sctp_ep.time_of_secret_change =
1683 old_inp->sctp_ep.time_of_secret_change;
1684 memcpy(new_inp->sctp_ep.secret_key, old_inp->sctp_ep.secret_key,
1685 sizeof(old_inp->sctp_ep.secret_key));
1686 new_inp->sctp_ep.current_secret_number =
1687 old_inp->sctp_ep.current_secret_number;
1688 new_inp->sctp_ep.last_secret_number =
1689 old_inp->sctp_ep.last_secret_number;
1690 new_inp->sctp_ep.size_of_a_cookie = old_inp->sctp_ep.size_of_a_cookie;
1692 /* Copy the port across */
1693 lport = new_inp->sctp_lport = old_inp->sctp_lport;
1694 rport = stcb->rport;
1695 /* Pull the tcb from the old association */
1696 LIST_REMOVE(stcb, sctp_tcbhash);
1697 LIST_REMOVE(stcb, sctp_tcblist);
1699 /* Now insert the new_inp into the TCP connected hash */
1700 head = &sctppcbinfo.sctp_tcpephash[SCTP_PCBHASH_ALLADDR((lport + rport),
1701 sctppcbinfo.hashtcpmark)];
1703 LIST_INSERT_HEAD(head, new_inp, sctp_hash);
1705 /* Now move the tcb into the endpoint list */
1706 LIST_INSERT_HEAD(&new_inp->sctp_asoc_list, stcb, sctp_tcblist);
1708 * Question, do we even need to worry about the ep-hash since
1709 * we only have one connection? Probably not :> so lets
1710 * get rid of it and not suck up any kernel memory in that.
1712 SCTP_INP_INFO_WUNLOCK();
1713 stcb->sctp_socket = new_inp->sctp_socket;
1714 stcb->sctp_ep = new_inp;
1715 if (new_inp->sctp_tcbhash != NULL) {
1716 kfree(new_inp->sctp_tcbhash, M_PCB);
1717 new_inp->sctp_tcbhash = NULL;
1719 if ((new_inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) == 0) {
1720 /* Subset bound, so copy in the laddr list from the old_inp */
1721 LIST_FOREACH(oladdr, &old_inp->sctp_addr_list, sctp_nxt_addr) {
1722 laddr = (struct sctp_laddr *)SCTP_ZONE_GET(
1723 sctppcbinfo.ipi_zone_laddr);
1724 if (laddr == NULL) {
1726 * Gak, what can we do? This assoc is really
1727 * HOSED. We probably should send an abort
1728 * here.
1730 #ifdef SCTP_DEBUG
1731 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1732 kprintf("Association hosed in TCP model, out of laddr memory\n");
1734 #endif /* SCTP_DEBUG */
1735 continue;
1737 sctppcbinfo.ipi_count_laddr++;
1738 sctppcbinfo.ipi_gencnt_laddr++;
1739 bzero(laddr, sizeof(*laddr));
1740 laddr->ifa = oladdr->ifa;
1741 LIST_INSERT_HEAD(&new_inp->sctp_addr_list, laddr,
1742 sctp_nxt_addr);
1743 new_inp->laddr_count++;
1746 SCTP_INP_WUNLOCK(new_inp);
1747 SCTP_INP_WUNLOCK(old_inp);
1750 static int
1751 sctp_isport_inuse(struct sctp_inpcb *inp, uint16_t lport)
1753 struct sctppcbhead *head;
1754 struct sctp_inpcb *t_inp;
1756 head = &sctppcbinfo.sctp_ephash[SCTP_PCBHASH_ALLADDR(lport,
1757 sctppcbinfo.hashmark)];
1758 LIST_FOREACH(t_inp, head, sctp_hash) {
1759 if (t_inp->sctp_lport != lport) {
1760 continue;
1762 /* This one is in use. */
1763 /* check the v6/v4 binding issue */
1764 if ((t_inp->sctp_flags & SCTP_PCB_FLAGS_BOUND_V6) &&
1765 #if defined(__FreeBSD__)
1766 (((struct inpcb *)t_inp)->inp_flags & IN6P_IPV6_V6ONLY)
1767 #else
1768 #if defined(__OpenBSD__)
1769 (0) /* For open bsd we do dual bind only */
1770 #else
1771 (((struct in6pcb *)t_inp)->in6p_flags & IN6P_IPV6_V6ONLY)
1772 #endif
1773 #endif
1775 if (inp->sctp_flags & SCTP_PCB_FLAGS_BOUND_V6) {
1776 /* collision in V6 space */
1777 return (1);
1778 } else {
1779 /* inp is BOUND_V4 no conflict */
1780 continue;
1782 } else if (t_inp->sctp_flags & SCTP_PCB_FLAGS_BOUND_V6) {
1783 /* t_inp is bound v4 and v6, conflict always */
1784 return (1);
1785 } else {
1786 /* t_inp is bound only V4 */
1787 if ((inp->sctp_flags & SCTP_PCB_FLAGS_BOUND_V6) &&
1788 #if defined(__FreeBSD__)
1789 (((struct inpcb *)inp)->inp_flags & IN6P_IPV6_V6ONLY)
1790 #else
1791 #if defined(__OpenBSD__)
1792 (0) /* For open bsd we do dual bind only */
1793 #else
1794 (((struct in6pcb *)inp)->in6p_flags & IN6P_IPV6_V6ONLY)
1795 #endif
1796 #endif
1798 /* no conflict */
1799 continue;
1801 /* else fall through to conflict */
1803 return (1);
1805 return (0);
1808 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
1810 * Don't know why, but without this there is an unknown reference when
1811 * compiling NetBSD... hmm
1813 extern void in6_sin6_2_sin (struct sockaddr_in *, struct sockaddr_in6 *sin6);
1814 #endif
1818 #if (defined(__FreeBSD__) && __FreeBSD_version >= 500000) || defined(__DragonFly__)
1819 sctp_inpcb_bind(struct socket *so, struct sockaddr *addr, struct thread *p)
1820 #else
1821 sctp_inpcb_bind(struct socket *so, struct sockaddr *addr, struct proc *p)
1822 #endif
1824 /* bind a ep to a socket address */
1825 struct sctppcbhead *head;
1826 struct sctp_inpcb *inp, *inp_tmp;
1827 struct inpcb *ip_inp;
1828 int bindall;
1829 uint16_t lport;
1830 int error;
1832 lport = 0;
1833 error = 0;
1834 bindall = 1;
1835 inp = (struct sctp_inpcb *)so->so_pcb;
1836 ip_inp = (struct inpcb *)so->so_pcb;
1837 #ifdef SCTP_DEBUG
1838 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1839 if (addr) {
1840 kprintf("Bind called port:%d\n",
1841 ntohs(((struct sockaddr_in *)addr)->sin_port));
1842 kprintf("Addr :");
1843 sctp_print_address(addr);
1846 #endif /* SCTP_DEBUG */
1847 if ((inp->sctp_flags & SCTP_PCB_FLAGS_UNBOUND) == 0) {
1848 /* already did a bind, subsequent binds NOT allowed ! */
1849 return (EINVAL);
1852 if (addr != NULL) {
1853 if (addr->sa_family == AF_INET) {
1854 struct sockaddr_in *sin;
1856 /* IPV6_V6ONLY socket? */
1857 if (
1858 #if defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__)
1859 (ip_inp->inp_flags & IN6P_IPV6_V6ONLY)
1860 #else
1861 #if defined(__OpenBSD__)
1862 (0) /* For openbsd we do dual bind only */
1863 #else
1864 (((struct in6pcb *)inp)->in6p_flags & IN6P_IPV6_V6ONLY)
1865 #endif
1866 #endif
1868 return (EINVAL);
1871 if (addr->sa_len != sizeof(*sin))
1872 return (EINVAL);
1874 sin = (struct sockaddr_in *)addr;
1875 lport = sin->sin_port;
1877 if (sin->sin_addr.s_addr != INADDR_ANY) {
1878 bindall = 0;
1880 } else if (addr->sa_family == AF_INET6) {
1881 /* Only for pure IPv6 Address. (No IPv4 Mapped!) */
1882 struct sockaddr_in6 *sin6;
1884 sin6 = (struct sockaddr_in6 *)addr;
1886 if (addr->sa_len != sizeof(*sin6))
1887 return (EINVAL);
1889 lport = sin6->sin6_port;
1890 if (!IN6_IS_ADDR_UNSPECIFIED(&sin6->sin6_addr)) {
1891 bindall = 0;
1892 /* KAME hack: embed scopeid */
1893 #if defined(SCTP_BASE_FREEBSD) || defined(__APPLE__) || defined(__DragonFly__)
1894 if (in6_embedscope(&sin6->sin6_addr, sin6,
1895 ip_inp, NULL) != 0)
1896 return (EINVAL);
1897 #elif defined(__FreeBSD__)
1898 error = scope6_check_id(sin6, ip6_use_defzone);
1899 if (error != 0)
1900 return (error);
1901 #else
1902 if (in6_embedscope(&sin6->sin6_addr, sin6) != 0) {
1903 return (EINVAL);
1905 #endif
1907 #ifndef SCOPEDROUTING
1908 /* this must be cleared for ifa_ifwithaddr() */
1909 sin6->sin6_scope_id = 0;
1910 #endif /* SCOPEDROUTING */
1911 } else {
1912 return (EAFNOSUPPORT);
1915 SCTP_INP_INFO_WLOCK();
1916 SCTP_INP_WLOCK(inp);
1917 /* increase our count due to the unlock we do */
1918 SCTP_INP_INCR_REF(inp);
1919 if (lport) {
1921 * Did the caller specify a port? if so we must see if a
1922 * ep already has this one bound.
1924 /* got to be root to get at low ports */
1925 if (ntohs(lport) < IPPORT_RESERVED) {
1926 if (p && (error =
1927 #ifdef __FreeBSD__
1928 #if __FreeBSD_version >= 500000
1929 suser_cred(p->td_ucred, 0)
1930 #else
1931 suser(p)
1932 #endif
1933 #elif defined(__NetBSD__) || defined(__APPLE__)
1934 suser(p->p_ucred, &p->p_acflag)
1935 #elif defined(__DragonFly__)
1936 priv_check(p, PRIV_ROOT)
1937 #else
1938 suser(p, 0)
1939 #endif
1940 )) {
1941 SCTP_INP_DECR_REF(inp);
1942 SCTP_INP_WUNLOCK(inp);
1943 SCTP_INP_INFO_WUNLOCK();
1944 return (error);
1947 if (p == NULL) {
1948 SCTP_INP_DECR_REF(inp);
1949 SCTP_INP_WUNLOCK(inp);
1950 SCTP_INP_INFO_WUNLOCK();
1951 return (error);
1953 SCTP_INP_WUNLOCK(inp);
1954 inp_tmp = sctp_pcb_findep(addr, 0, 1);
1955 if (inp_tmp != NULL) {
1956 /* lock guy returned and lower count
1957 * note that we are not bound so inp_tmp
1958 * should NEVER be inp. And it is this
1959 * inp (inp_tmp) that gets the reference
1960 * bump, so we must lower it.
1962 SCTP_INP_WLOCK(inp_tmp);
1963 SCTP_INP_DECR_REF(inp_tmp);
1964 SCTP_INP_WUNLOCK(inp_tmp);
1966 /* unlock info */
1967 SCTP_INP_INFO_WUNLOCK();
1968 return (EADDRNOTAVAIL);
1970 SCTP_INP_WLOCK(inp);
1971 if (bindall) {
1972 /* verify that no lport is not used by a singleton */
1973 if (sctp_isport_inuse(inp, lport)) {
1974 /* Sorry someone already has this one bound */
1975 SCTP_INP_DECR_REF(inp);
1976 SCTP_INP_WUNLOCK(inp);
1977 SCTP_INP_INFO_WUNLOCK();
1978 return (EADDRNOTAVAIL);
1981 } else {
1983 * get any port but lets make sure no one has any address
1984 * with this port bound
1988 * setup the inp to the top (I could use the union but this
1989 * is just as easy
1991 uint32_t port_guess;
1992 uint16_t port_attempt;
1993 int not_done=1;
1995 while (not_done) {
1996 port_guess = sctp_select_initial_TSN(&inp->sctp_ep);
1997 port_attempt = (port_guess & 0x0000ffff);
1998 if (port_attempt == 0) {
1999 goto next_half;
2001 if (port_attempt < IPPORT_RESERVED) {
2002 port_attempt += IPPORT_RESERVED;
2005 if (sctp_isport_inuse(inp, htons(port_attempt)) == 0) {
2006 /* got a port we can use */
2007 not_done = 0;
2008 continue;
2010 /* try upper half */
2011 next_half:
2012 port_attempt = ((port_guess >> 16) & 0x0000ffff);
2013 if (port_attempt == 0) {
2014 goto last_try;
2016 if (port_attempt < IPPORT_RESERVED) {
2017 port_attempt += IPPORT_RESERVED;
2019 if (sctp_isport_inuse(inp, htons(port_attempt)) == 0) {
2020 /* got a port we can use */
2021 not_done = 0;
2022 continue;
2024 /* try two half's added together */
2025 last_try:
2026 port_attempt = (((port_guess >> 16) & 0x0000ffff) + (port_guess & 0x0000ffff));
2027 if (port_attempt == 0) {
2028 /* get a new random number */
2029 continue;
2031 if (port_attempt < IPPORT_RESERVED) {
2032 port_attempt += IPPORT_RESERVED;
2034 if (sctp_isport_inuse(inp, htons(port_attempt)) == 0) {
2035 /* got a port we can use */
2036 not_done = 0;
2037 continue;
2040 /* we don't get out of the loop until we have a port */
2041 lport = htons(port_attempt);
2043 SCTP_INP_DECR_REF(inp);
2044 if (inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
2045 /* this really should not happen. The guy
2046 * did a non-blocking bind and then did a close
2047 * at the same time.
2049 SCTP_INP_WUNLOCK(inp);
2050 SCTP_INP_INFO_WUNLOCK();
2051 return (EINVAL);
2053 /* ok we look clear to give out this port, so lets setup the binding */
2054 if (bindall) {
2055 /* binding to all addresses, so just set in the proper flags */
2056 inp->sctp_flags |= (SCTP_PCB_FLAGS_BOUNDALL |
2057 SCTP_PCB_FLAGS_DO_ASCONF);
2058 /* set the automatic addr changes from kernel flag */
2059 if (sctp_auto_asconf == 0) {
2060 inp->sctp_flags &= ~SCTP_PCB_FLAGS_AUTO_ASCONF;
2061 } else {
2062 inp->sctp_flags |= SCTP_PCB_FLAGS_AUTO_ASCONF;
2064 } else {
2066 * bind specific, make sure flags is off and add a new address
2067 * structure to the sctp_addr_list inside the ep structure.
2069 * We will need to allocate one and insert it at the head.
2070 * The socketopt call can just insert new addresses in there
2071 * as well. It will also have to do the embed scope kame hack
2072 * too (before adding).
2074 struct ifaddr *ifa;
2075 struct sockaddr_storage store_sa;
2077 memset(&store_sa, 0, sizeof(store_sa));
2078 if (addr->sa_family == AF_INET) {
2079 struct sockaddr_in *sin;
2081 sin = (struct sockaddr_in *)&store_sa;
2082 memcpy(sin, addr, sizeof(struct sockaddr_in));
2083 sin->sin_port = 0;
2084 } else if (addr->sa_family == AF_INET6) {
2085 struct sockaddr_in6 *sin6;
2087 sin6 = (struct sockaddr_in6 *)&store_sa;
2088 memcpy(sin6, addr, sizeof(struct sockaddr_in6));
2089 sin6->sin6_port = 0;
2092 * first find the interface with the bound address
2093 * need to zero out the port to find the address! yuck!
2094 * can't do this earlier since need port for sctp_pcb_findep()
2096 ifa = sctp_find_ifa_by_addr((struct sockaddr *)&store_sa);
2097 if (ifa == NULL) {
2098 /* Can't find an interface with that address */
2099 SCTP_INP_WUNLOCK(inp);
2100 SCTP_INP_INFO_WUNLOCK();
2101 return (EADDRNOTAVAIL);
2103 if (addr->sa_family == AF_INET6) {
2104 struct in6_ifaddr *ifa6;
2105 ifa6 = (struct in6_ifaddr *)ifa;
2107 * allow binding of deprecated addresses as per
2108 * RFC 2462 and ipng discussion
2110 if (ifa6->ia6_flags & (IN6_IFF_DETACHED |
2111 IN6_IFF_ANYCAST |
2112 IN6_IFF_NOTREADY)) {
2113 /* Can't bind a non-existent addr. */
2114 SCTP_INP_WUNLOCK(inp);
2115 SCTP_INP_INFO_WUNLOCK();
2116 return (EINVAL);
2119 /* we're not bound all */
2120 inp->sctp_flags &= ~SCTP_PCB_FLAGS_BOUNDALL;
2121 #if 0 /* use sysctl now */
2122 /* don't allow automatic addr changes from kernel */
2123 inp->sctp_flags &= ~SCTP_PCB_FLAGS_AUTO_ASCONF;
2124 #endif
2125 /* set the automatic addr changes from kernel flag */
2126 if (sctp_auto_asconf == 0) {
2127 inp->sctp_flags &= ~SCTP_PCB_FLAGS_AUTO_ASCONF;
2128 } else {
2129 inp->sctp_flags |= SCTP_PCB_FLAGS_AUTO_ASCONF;
2131 /* allow bindx() to send ASCONF's for binding changes */
2132 inp->sctp_flags |= SCTP_PCB_FLAGS_DO_ASCONF;
2133 /* add this address to the endpoint list */
2134 error = sctp_insert_laddr(&inp->sctp_addr_list, ifa);
2135 if (error != 0) {
2136 SCTP_INP_WUNLOCK(inp);
2137 SCTP_INP_INFO_WUNLOCK();
2138 return (error);
2140 inp->laddr_count++;
2142 /* find the bucket */
2143 head = &sctppcbinfo.sctp_ephash[SCTP_PCBHASH_ALLADDR(lport,
2144 sctppcbinfo.hashmark)];
2145 /* put it in the bucket */
2146 LIST_INSERT_HEAD(head, inp, sctp_hash);
2147 #ifdef SCTP_DEBUG
2148 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
2149 kprintf("Main hash to bind at head:%p, bound port:%d\n", head, ntohs(lport));
2151 #endif
2152 /* set in the port */
2153 inp->sctp_lport = lport;
2155 /* turn off just the unbound flag */
2156 inp->sctp_flags &= ~SCTP_PCB_FLAGS_UNBOUND;
2157 SCTP_INP_WUNLOCK(inp);
2158 SCTP_INP_INFO_WUNLOCK();
2159 return (0);
2163 static void
2164 sctp_iterator_inp_being_freed(struct sctp_inpcb *inp, struct sctp_inpcb *inp_next)
2166 struct sctp_iterator *it;
2167 /* We enter with the only the ITERATOR_LOCK in place and
2168 * A write lock on the inp_info stuff.
2171 /* Go through all iterators, we must do this since
2172 * it is possible that some iterator does NOT have
2173 * the lock, but is waiting for it. And the one that
2174 * had the lock has either moved in the last iteration
2175 * or we just cleared it above. We need to find all
2176 * of those guys. The list of iterators should never
2177 * be very big though.
2179 LIST_FOREACH(it, &sctppcbinfo.iteratorhead, sctp_nxt_itr) {
2180 if (it == inp->inp_starting_point_for_iterator)
2181 /* skip this guy, he's special */
2182 continue;
2183 if (it->inp == inp) {
2184 /* This is tricky and we DON'T lock the iterator.
2185 * Reason is he's running but waiting for me since
2186 * inp->inp_starting_point_for_iterator has the lock
2187 * on me (the guy above we skipped). This tells us
2188 * its is not running but waiting for inp->inp_starting_point_for_iterator
2189 * to be released by the guy that does have our INP in a lock.
2191 if (it->iterator_flags & SCTP_ITERATOR_DO_SINGLE_INP) {
2192 it->inp = NULL;
2193 it->stcb = NULL;
2194 } else {
2195 /* set him up to do the next guy not me */
2196 it->inp = inp_next;
2197 it->stcb = NULL;
2201 it = inp->inp_starting_point_for_iterator;
2202 if (it) {
2203 if (it->iterator_flags & SCTP_ITERATOR_DO_SINGLE_INP) {
2204 it->inp = NULL;
2205 } else {
2206 it->inp = inp_next;
2208 it->stcb = NULL;
2212 /* release sctp_inpcb unbind the port */
2213 void
2214 sctp_inpcb_free(struct sctp_inpcb *inp, int immediate)
2217 * Here we free a endpoint. We must find it (if it is in the Hash
2218 * table) and remove it from there. Then we must also find it in
2219 * the overall list and remove it from there. After all removals are
2220 * complete then any timer has to be stopped. Then start the actual
2221 * freeing.
2222 * a) Any local lists.
2223 * b) Any associations.
2224 * c) The hash of all associations.
2225 * d) finally the ep itself.
2227 struct sctp_pcb *m;
2228 struct sctp_inpcb *inp_save;
2229 struct sctp_tcb *asoc, *nasoc;
2230 struct sctp_laddr *laddr, *nladdr;
2231 struct inpcb *ip_pcb;
2232 struct socket *so;
2233 struct sctp_socket_q_list *sq;
2234 #if !defined(__FreeBSD__) || __FreeBSD_version < 500000
2235 struct rtentry *rt;
2236 #endif
2237 int cnt;
2239 SCTP_ASOC_CREATE_LOCK(inp);
2240 SCTP_INP_WLOCK(inp);
2242 if (inp->sctp_flags & SCTP_PCB_FLAGS_SOCKET_ALLGONE) {
2243 /* been here before */
2244 kprintf("Endpoint was all gone (dup free)?\n");
2245 SCTP_INP_WUNLOCK(inp);
2246 SCTP_ASOC_CREATE_UNLOCK(inp);
2247 return;
2249 sctp_timer_stop(SCTP_TIMER_TYPE_NEWCOOKIE, inp, NULL, NULL);
2251 if (inp->control) {
2252 sctp_m_freem(inp->control);
2253 inp->control = NULL;
2255 if (inp->pkt) {
2256 sctp_m_freem(inp->pkt);
2257 inp->pkt = NULL;
2259 so = inp->sctp_socket;
2260 m = &inp->sctp_ep;
2261 ip_pcb = &inp->ip_inp.inp; /* we could just cast the main
2262 * pointer here but I will
2263 * be nice :> (i.e. ip_pcb = ep;)
2266 if (immediate == 0) {
2267 int cnt_in_sd;
2268 cnt_in_sd = 0;
2269 for ((asoc = LIST_FIRST(&inp->sctp_asoc_list)); asoc != NULL;
2270 asoc = nasoc) {
2271 nasoc = LIST_NEXT(asoc, sctp_tcblist);
2272 if ((SCTP_GET_STATE(&asoc->asoc) == SCTP_STATE_COOKIE_WAIT) ||
2273 (SCTP_GET_STATE(&asoc->asoc) == SCTP_STATE_COOKIE_ECHOED)) {
2274 /* Just abandon things in the front states */
2275 SCTP_TCB_LOCK(asoc);
2276 SCTP_INP_WUNLOCK(inp);
2277 sctp_free_assoc(inp, asoc);
2278 SCTP_INP_WLOCK(inp);
2279 continue;
2280 } else {
2281 asoc->asoc.state |= SCTP_STATE_CLOSED_SOCKET;
2283 if ((asoc->asoc.size_on_delivery_queue > 0) ||
2284 (asoc->asoc.size_on_reasm_queue > 0) ||
2285 (asoc->asoc.size_on_all_streams > 0) ||
2286 (so && (so->so_rcv.ssb_cc > 0))
2288 /* Left with Data unread */
2289 struct mbuf *op_err;
2290 MGET(op_err, MB_DONTWAIT, MT_DATA);
2291 if (op_err) {
2292 /* Fill in the user initiated abort */
2293 struct sctp_paramhdr *ph;
2294 op_err->m_len =
2295 sizeof(struct sctp_paramhdr);
2296 ph = mtod(op_err,
2297 struct sctp_paramhdr *);
2298 ph->param_type = htons(
2299 SCTP_CAUSE_USER_INITIATED_ABT);
2300 ph->param_length = htons(op_err->m_len);
2302 SCTP_TCB_LOCK(asoc);
2303 sctp_send_abort_tcb(asoc, op_err);
2305 SCTP_INP_WUNLOCK(inp);
2306 sctp_free_assoc(inp, asoc);
2307 SCTP_INP_WLOCK(inp);
2308 continue;
2309 } else if (TAILQ_EMPTY(&asoc->asoc.send_queue) &&
2310 TAILQ_EMPTY(&asoc->asoc.sent_queue)) {
2311 if ((SCTP_GET_STATE(&asoc->asoc) != SCTP_STATE_SHUTDOWN_SENT) &&
2312 (SCTP_GET_STATE(&asoc->asoc) != SCTP_STATE_SHUTDOWN_ACK_SENT)) {
2313 /* there is nothing queued to send, so I send shutdown */
2314 SCTP_TCB_LOCK(asoc);
2315 sctp_send_shutdown(asoc, asoc->asoc.primary_destination);
2316 asoc->asoc.state = SCTP_STATE_SHUTDOWN_SENT;
2317 sctp_timer_start(SCTP_TIMER_TYPE_SHUTDOWN, asoc->sctp_ep, asoc,
2318 asoc->asoc.primary_destination);
2319 sctp_timer_start(SCTP_TIMER_TYPE_SHUTDOWNGUARD, asoc->sctp_ep, asoc,
2320 asoc->asoc.primary_destination);
2321 sctp_chunk_output(inp, asoc, 1);
2322 SCTP_TCB_UNLOCK(asoc);
2324 } else {
2325 /* mark into shutdown pending */
2326 asoc->asoc.state |= SCTP_STATE_SHUTDOWN_PENDING;
2328 cnt_in_sd++;
2330 /* now is there some left in our SHUTDOWN state? */
2331 if (cnt_in_sd) {
2332 inp->sctp_flags |= SCTP_PCB_FLAGS_SOCKET_GONE;
2333 SCTP_INP_WUNLOCK(inp);
2334 SCTP_ASOC_CREATE_UNLOCK(inp);
2335 return;
2338 #if defined(__FreeBSD__) && __FreeBSD_version >= 503000
2339 if (inp->refcount) {
2340 sctp_timer_start(SCTP_TIMER_TYPE_INPKILL, inp, NULL, NULL);
2341 SCTP_INP_WUNLOCK(inp);
2342 SCTP_ASOC_CREATE_UNLOCK(inp);
2343 return;
2345 #endif
2346 inp->sctp_flags |= SCTP_PCB_FLAGS_SOCKET_ALLGONE;
2347 #if !defined(__FreeBSD__) || __FreeBSD_version < 500000
2348 rt = ip_pcb->inp_route.ro_rt;
2349 #endif
2351 callout_stop(&inp->sctp_ep.signature_change.timer);
2353 if (so) {
2354 /* First take care of socket level things */
2355 #ifdef IPSEC
2356 #ifdef __OpenBSD__
2357 /* XXX IPsec cleanup here */
2358 if (ip_pcb->inp_tdb_in)
2359 TAILQ_REMOVE(&ip_pcb->inp_tdb_in->tdb_inp_in,
2360 ip_pcb, inp_tdb_in_next);
2361 if (ip_pcb->inp_tdb_out)
2362 TAILQ_REMOVE(&ip_pcb->inp_tdb_out->tdb_inp_out, ip_pcb,
2363 inp_tdb_out_next);
2364 if (ip_pcb->inp_ipsec_localid)
2365 ipsp_reffree(ip_pcb->inp_ipsec_localid);
2366 if (ip_pcb->inp_ipsec_remoteid)
2367 ipsp_reffree(ip_pcb->inp_ipsec_remoteid);
2368 if (ip_pcb->inp_ipsec_localcred)
2369 ipsp_reffree(ip_pcb->inp_ipsec_localcred);
2370 if (ip_pcb->inp_ipsec_remotecred)
2371 ipsp_reffree(ip_pcb->inp_ipsec_remotecred);
2372 if (ip_pcb->inp_ipsec_localauth)
2373 ipsp_reffree(ip_pcb->inp_ipsec_localauth);
2374 if (ip_pcb->inp_ipsec_remoteauth)
2375 ipsp_reffree(ip_pcb->inp_ipsec_remoteauth);
2376 #else
2377 ipsec4_delete_pcbpolicy(ip_pcb);
2378 #endif
2379 #endif /*IPSEC*/
2380 #if defined(__FreeBSD__) && __FreeBSD_version > 500000
2381 ACCEPT_LOCK();
2382 SOCK_LOCK(so);
2383 #endif
2384 so->so_pcb = NULL;
2385 sofree(so);
2388 if (ip_pcb->inp_options) {
2389 m_free(ip_pcb->inp_options);
2390 ip_pcb->inp_options = 0;
2392 #if !defined(__FreeBSD__) || __FreeBSD_version < 500000
2393 if (rt) {
2394 RTFREE(rt);
2395 ip_pcb->inp_route.ro_rt = 0;
2397 #endif
2398 if (ip_pcb->inp_moptions) {
2399 ip_freemoptions(ip_pcb->inp_moptions);
2400 ip_pcb->inp_moptions = 0;
2402 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
2403 inp->inp_vflag = 0;
2404 #else
2405 ip_pcb->inp_vflag = 0;
2406 #endif
2408 /* Now the sctp_pcb things */
2411 * free each asoc if it is not already closed/free. we can't use
2412 * the macro here since le_next will get freed as part of the
2413 * sctp_free_assoc() call.
2415 cnt = 0;
2416 for ((asoc = LIST_FIRST(&inp->sctp_asoc_list)); asoc != NULL;
2417 asoc = nasoc) {
2418 nasoc = LIST_NEXT(asoc, sctp_tcblist);
2419 SCTP_TCB_LOCK(asoc);
2420 if (SCTP_GET_STATE(&asoc->asoc) != SCTP_STATE_COOKIE_WAIT) {
2421 struct mbuf *op_err;
2422 MGET(op_err, MB_DONTWAIT, MT_DATA);
2423 if (op_err) {
2424 /* Fill in the user initiated abort */
2425 struct sctp_paramhdr *ph;
2426 op_err->m_len = sizeof(struct sctp_paramhdr);
2427 ph = mtod(op_err, struct sctp_paramhdr *);
2428 ph->param_type = htons(
2429 SCTP_CAUSE_USER_INITIATED_ABT);
2430 ph->param_length = htons(op_err->m_len);
2432 sctp_send_abort_tcb(asoc, op_err);
2434 cnt++;
2436 * sctp_free_assoc() will call sctp_inpcb_free(),
2437 * if SCTP_PCB_FLAGS_SOCKET_GONE set.
2438 * So, we clear it before sctp_free_assoc() making sure
2439 * no double sctp_inpcb_free().
2441 inp->sctp_flags &= ~SCTP_PCB_FLAGS_SOCKET_GONE;
2442 SCTP_INP_WUNLOCK(inp);
2443 sctp_free_assoc(inp, asoc);
2444 SCTP_INP_WLOCK(inp);
2446 while ((sq = TAILQ_FIRST(&inp->sctp_queue_list)) != NULL) {
2447 TAILQ_REMOVE(&inp->sctp_queue_list, sq, next_sq);
2448 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_sockq, sq);
2449 sctppcbinfo.ipi_count_sockq--;
2450 sctppcbinfo.ipi_gencnt_sockq++;
2452 inp->sctp_socket = 0;
2453 /* Now first we remove ourselves from the overall list of all EP's */
2455 /* Unlock inp first, need correct order */
2456 SCTP_INP_WUNLOCK(inp);
2457 /* now iterator lock */
2458 SCTP_ITERATOR_LOCK();
2459 /* now info lock */
2460 SCTP_INP_INFO_WLOCK();
2461 /* now reget the inp lock */
2462 SCTP_INP_WLOCK(inp);
2464 inp_save = LIST_NEXT(inp, sctp_list);
2465 LIST_REMOVE(inp, sctp_list);
2467 * Now the question comes as to if this EP was ever bound at all.
2468 * If it was, then we must pull it out of the EP hash list.
2470 if ((inp->sctp_flags & SCTP_PCB_FLAGS_UNBOUND) !=
2471 SCTP_PCB_FLAGS_UNBOUND) {
2473 * ok, this guy has been bound. It's port is somewhere
2474 * in the sctppcbinfo hash table. Remove it!
2476 LIST_REMOVE(inp, sctp_hash);
2478 /* fix any iterators only after out of the list */
2479 sctp_iterator_inp_being_freed(inp, inp_save);
2480 SCTP_ITERATOR_UNLOCK();
2482 * if we have an address list the following will free the list of
2483 * ifaddr's that are set into this ep. Again macro limitations here,
2484 * since the LIST_FOREACH could be a bad idea.
2486 for ((laddr = LIST_FIRST(&inp->sctp_addr_list)); laddr != NULL;
2487 laddr = nladdr) {
2488 nladdr = LIST_NEXT(laddr, sctp_nxt_addr);
2489 LIST_REMOVE(laddr, sctp_nxt_addr);
2490 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_laddr, laddr);
2491 sctppcbinfo.ipi_gencnt_laddr++;
2492 sctppcbinfo.ipi_count_laddr--;
2494 /* Now lets see about freeing the EP hash table. */
2495 if (inp->sctp_tcbhash != NULL) {
2496 kfree(inp->sctp_tcbhash, M_PCB);
2497 inp->sctp_tcbhash = 0;
2499 SCTP_INP_WUNLOCK(inp);
2500 SCTP_ASOC_CREATE_UNLOCK(inp);
2501 SCTP_INP_LOCK_DESTROY(inp);
2502 SCTP_ASOC_CREATE_LOCK_DESTROY(inp);
2504 /* Now we must put the ep memory back into the zone pool */
2505 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_ep, inp);
2506 sctppcbinfo.ipi_count_ep--;
2508 SCTP_INP_INFO_WUNLOCK();
2512 struct sctp_nets *
2513 sctp_findnet(struct sctp_tcb *stcb, struct sockaddr *addr)
2515 struct sctp_nets *net;
2516 struct sockaddr_in *sin;
2517 struct sockaddr_in6 *sin6;
2518 /* use the peer's/remote port for lookup if unspecified */
2519 sin = (struct sockaddr_in *)addr;
2520 sin6 = (struct sockaddr_in6 *)addr;
2521 #if 0 /* why do we need to check the port for a nets list on an assoc? */
2522 if (stcb->rport != sin->sin_port) {
2523 /* we cheat and just a sin for this test */
2524 return (NULL);
2526 #endif
2527 /* locate the address */
2528 TAILQ_FOREACH(net, &stcb->asoc.nets, sctp_next) {
2529 if (sctp_cmpaddr(addr, (struct sockaddr *)&net->ro._l_addr))
2530 return (net);
2532 return (NULL);
2537 * add's a remote endpoint address, done with the INIT/INIT-ACK
2538 * as well as when a ASCONF arrives that adds it. It will also
2539 * initialize all the cwnd stats of stuff.
2542 sctp_is_address_on_local_host(struct sockaddr *addr)
2544 struct ifnet *ifn;
2546 TAILQ_FOREACH(ifn, &ifnet, if_list) {
2547 struct ifaddr_container *ifac;
2549 TAILQ_FOREACH(ifac, &ifn->if_addrheads[mycpuid], ifa_link) {
2550 struct ifaddr *ifa = ifac->ifa;
2552 if (addr->sa_family == ifa->ifa_addr->sa_family) {
2553 /* same family */
2554 if (addr->sa_family == AF_INET) {
2555 struct sockaddr_in *sin, *sin_c;
2556 sin = (struct sockaddr_in *)addr;
2557 sin_c = (struct sockaddr_in *)
2558 ifa->ifa_addr;
2559 if (sin->sin_addr.s_addr ==
2560 sin_c->sin_addr.s_addr) {
2561 /* we are on the same machine */
2562 return (1);
2564 } else if (addr->sa_family == AF_INET6) {
2565 struct sockaddr_in6 *sin6, *sin_c6;
2566 sin6 = (struct sockaddr_in6 *)addr;
2567 sin_c6 = (struct sockaddr_in6 *)
2568 ifa->ifa_addr;
2569 if (SCTP6_ARE_ADDR_EQUAL(&sin6->sin6_addr,
2570 &sin_c6->sin6_addr)) {
2571 /* we are on the same machine */
2572 return (1);
2578 return (0);
2582 sctp_add_remote_addr(struct sctp_tcb *stcb, struct sockaddr *newaddr,
2583 int set_scope, int from)
2586 * The following is redundant to the same lines in the
2587 * sctp_aloc_assoc() but is needed since other's call the add
2588 * address function
2590 struct sctp_nets *net, *netfirst;
2591 int addr_inscope;
2593 #ifdef SCTP_DEBUG
2594 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
2595 kprintf("Adding an address (from:%d) to the peer: ", from);
2596 sctp_print_address(newaddr);
2598 #endif
2599 netfirst = sctp_findnet(stcb, newaddr);
2600 if (netfirst) {
2602 * Lie and return ok, we don't want to make the association
2603 * go away for this behavior. It will happen in the TCP model
2604 * in a connected socket. It does not reach the hash table
2605 * until after the association is built so it can't be found.
2606 * Mark as reachable, since the initial creation will have
2607 * been cleared and the NOT_IN_ASSOC flag will have been
2608 * added... and we don't want to end up removing it back out.
2610 if (netfirst->dest_state & SCTP_ADDR_UNCONFIRMED) {
2611 netfirst->dest_state = (SCTP_ADDR_REACHABLE|
2612 SCTP_ADDR_UNCONFIRMED);
2613 } else {
2614 netfirst->dest_state = SCTP_ADDR_REACHABLE;
2617 return (0);
2619 addr_inscope = 1;
2620 if (newaddr->sa_family == AF_INET) {
2621 struct sockaddr_in *sin;
2622 sin = (struct sockaddr_in *)newaddr;
2623 if (sin->sin_addr.s_addr == 0) {
2624 /* Invalid address */
2625 return (-1);
2627 /* zero out the bzero area */
2628 memset(&sin->sin_zero, 0, sizeof(sin->sin_zero));
2630 /* assure len is set */
2631 sin->sin_len = sizeof(struct sockaddr_in);
2632 if (set_scope) {
2633 #ifdef SCTP_DONT_DO_PRIVADDR_SCOPE
2634 stcb->ipv4_local_scope = 1;
2635 #else
2636 if (IN4_ISPRIVATE_ADDRESS(&sin->sin_addr)) {
2637 stcb->asoc.ipv4_local_scope = 1;
2639 #endif /* SCTP_DONT_DO_PRIVADDR_SCOPE */
2641 if (sctp_is_address_on_local_host(newaddr)) {
2642 stcb->asoc.loopback_scope = 1;
2643 stcb->asoc.ipv4_local_scope = 1;
2644 stcb->asoc.local_scope = 1;
2645 stcb->asoc.site_scope = 1;
2647 } else {
2648 if (from == 8) {
2649 /* From connectx */
2650 if (sctp_is_address_on_local_host(newaddr)) {
2651 stcb->asoc.loopback_scope = 1;
2652 stcb->asoc.ipv4_local_scope = 1;
2653 stcb->asoc.local_scope = 1;
2654 stcb->asoc.site_scope = 1;
2657 /* Validate the address is in scope */
2658 if ((IN4_ISPRIVATE_ADDRESS(&sin->sin_addr)) &&
2659 (stcb->asoc.ipv4_local_scope == 0)) {
2660 addr_inscope = 0;
2663 } else if (newaddr->sa_family == AF_INET6) {
2664 struct sockaddr_in6 *sin6;
2665 sin6 = (struct sockaddr_in6 *)newaddr;
2666 if (IN6_IS_ADDR_UNSPECIFIED(&sin6->sin6_addr)) {
2667 /* Invalid address */
2668 return (-1);
2670 /* assure len is set */
2671 sin6->sin6_len = sizeof(struct sockaddr_in6);
2672 if (set_scope) {
2673 if (sctp_is_address_on_local_host(newaddr)) {
2674 stcb->asoc.loopback_scope = 1;
2675 stcb->asoc.local_scope = 1;
2676 stcb->asoc.ipv4_local_scope = 1;
2677 stcb->asoc.site_scope = 1;
2678 } else if (IN6_IS_ADDR_LINKLOCAL(&sin6->sin6_addr)) {
2680 * If the new destination is a LINK_LOCAL
2681 * we must have common site scope. Don't set
2682 * the local scope since we may not share all
2683 * links, only loopback can do this.
2684 * Links on the local network would also
2685 * be on our private network for v4 too.
2687 stcb->asoc.ipv4_local_scope = 1;
2688 stcb->asoc.site_scope = 1;
2689 } else if (IN6_IS_ADDR_SITELOCAL(&sin6->sin6_addr)) {
2691 * If the new destination is SITE_LOCAL
2692 * then we must have site scope in common.
2694 stcb->asoc.site_scope = 1;
2696 } else {
2697 if (from == 8) {
2698 /* From connectx */
2699 if (sctp_is_address_on_local_host(newaddr)) {
2700 stcb->asoc.loopback_scope = 1;
2701 stcb->asoc.ipv4_local_scope = 1;
2702 stcb->asoc.local_scope = 1;
2703 stcb->asoc.site_scope = 1;
2706 /* Validate the address is in scope */
2707 if (IN6_IS_ADDR_LOOPBACK(&sin6->sin6_addr) &&
2708 (stcb->asoc.loopback_scope == 0)) {
2709 addr_inscope = 0;
2710 } else if (IN6_IS_ADDR_LINKLOCAL(&sin6->sin6_addr) &&
2711 (stcb->asoc.local_scope == 0)) {
2712 addr_inscope = 0;
2713 } else if (IN6_IS_ADDR_SITELOCAL(&sin6->sin6_addr) &&
2714 (stcb->asoc.site_scope == 0)) {
2715 addr_inscope = 0;
2718 } else {
2719 /* not supported family type */
2720 return (-1);
2722 net = (struct sctp_nets *)SCTP_ZONE_GET(sctppcbinfo.ipi_zone_net);
2723 if (net == NULL) {
2724 return (-1);
2726 sctppcbinfo.ipi_count_raddr++;
2727 sctppcbinfo.ipi_gencnt_raddr++;
2728 bzero(net, sizeof(*net));
2729 memcpy(&net->ro._l_addr, newaddr, newaddr->sa_len);
2730 if (newaddr->sa_family == AF_INET) {
2731 ((struct sockaddr_in *)&net->ro._l_addr)->sin_port = stcb->rport;
2732 } else if (newaddr->sa_family == AF_INET6) {
2733 ((struct sockaddr_in6 *)&net->ro._l_addr)->sin6_port = stcb->rport;
2735 net->addr_is_local = sctp_is_address_on_local_host(newaddr);
2736 net->failure_threshold = stcb->asoc.def_net_failure;
2737 if (addr_inscope == 0) {
2738 #ifdef SCTP_DEBUG
2739 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
2740 kprintf("Adding an address which is OUT OF SCOPE\n");
2742 #endif /* SCTP_DEBUG */
2743 net->dest_state = (SCTP_ADDR_REACHABLE |
2744 SCTP_ADDR_OUT_OF_SCOPE);
2745 } else {
2746 if (from == 8)
2747 /* 8 is passed by connect_x */
2748 net->dest_state = SCTP_ADDR_REACHABLE;
2749 else
2750 net->dest_state = SCTP_ADDR_REACHABLE |
2751 SCTP_ADDR_UNCONFIRMED;
2753 net->RTO = stcb->asoc.initial_rto;
2754 stcb->asoc.numnets++;
2755 net->ref_count = 1;
2757 /* Init the timer structure */
2758 #if defined(__FreeBSD__) && __FreeBSD_version >= 500000
2759 callout_init(&net->rxt_timer.timer, 0);
2760 callout_init(&net->pmtu_timer.timer, 0);
2761 #else
2762 callout_init(&net->rxt_timer.timer);
2763 callout_init(&net->pmtu_timer.timer);
2764 #endif
2766 /* Now generate a route for this guy */
2767 /* KAME hack: embed scopeid */
2768 if (newaddr->sa_family == AF_INET6) {
2769 struct sockaddr_in6 *sin6;
2770 sin6 = (struct sockaddr_in6 *)&net->ro._l_addr;
2771 #if defined(SCTP_BASE_FREEBSD) || defined(__APPLE__) || defined(__DragonFly__)
2772 in6_embedscope(&sin6->sin6_addr, sin6,
2773 &stcb->sctp_ep->ip_inp.inp, NULL);
2774 #else
2775 in6_embedscope(&sin6->sin6_addr, sin6);
2776 #endif
2777 #ifndef SCOPEDROUTING
2778 sin6->sin6_scope_id = 0;
2779 #endif
2781 #if defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__)
2782 rtalloc_ign((struct route *)&net->ro, 0UL);
2783 #else
2784 rtalloc((struct route *)&net->ro);
2785 #endif
2786 if (newaddr->sa_family == AF_INET6) {
2787 struct sockaddr_in6 *sin6;
2788 sin6 = (struct sockaddr_in6 *)&net->ro._l_addr;
2789 in6_recoverscope(sin6, &sin6->sin6_addr, NULL);
2791 if ((net->ro.ro_rt) &&
2792 (net->ro.ro_rt->rt_ifp)) {
2793 net->mtu = net->ro.ro_rt->rt_ifp->if_mtu;
2794 if (from == 1) {
2795 stcb->asoc.smallest_mtu = net->mtu;
2797 /* start things off to match mtu of interface please. */
2798 net->ro.ro_rt->rt_rmx.rmx_mtu = net->ro.ro_rt->rt_ifp->if_mtu;
2799 } else {
2800 net->mtu = stcb->asoc.smallest_mtu;
2802 if (stcb->asoc.smallest_mtu > net->mtu) {
2803 stcb->asoc.smallest_mtu = net->mtu;
2805 /* We take the max of the burst limit times a MTU or the INITIAL_CWND.
2806 * We then limit this to 4 MTU's of sending.
2808 net->cwnd = min((net->mtu * 4), max((stcb->asoc.max_burst * net->mtu), SCTP_INITIAL_CWND));
2810 /* we always get at LEAST 2 MTU's */
2811 if (net->cwnd < (2 * net->mtu)) {
2812 net->cwnd = 2 * net->mtu;
2815 net->ssthresh = stcb->asoc.peers_rwnd;
2817 net->src_addr_selected = 0;
2818 netfirst = TAILQ_FIRST(&stcb->asoc.nets);
2819 if (net->ro.ro_rt == NULL) {
2820 /* Since we have no route put it at the back */
2821 TAILQ_INSERT_TAIL(&stcb->asoc.nets, net, sctp_next);
2822 } else if (netfirst == NULL) {
2823 /* We are the first one in the pool. */
2824 TAILQ_INSERT_HEAD(&stcb->asoc.nets, net, sctp_next);
2825 } else if (netfirst->ro.ro_rt == NULL) {
2827 * First one has NO route. Place this one ahead of the
2828 * first one.
2830 TAILQ_INSERT_HEAD(&stcb->asoc.nets, net, sctp_next);
2831 } else if (net->ro.ro_rt->rt_ifp != netfirst->ro.ro_rt->rt_ifp) {
2833 * This one has a different interface than the one at the
2834 * top of the list. Place it ahead.
2836 TAILQ_INSERT_HEAD(&stcb->asoc.nets, net, sctp_next);
2837 } else {
2839 * Ok we have the same interface as the first one. Move
2840 * forward until we find either
2841 * a) one with a NULL route... insert ahead of that
2842 * b) one with a different ifp.. insert after that.
2843 * c) end of the list.. insert at the tail.
2845 struct sctp_nets *netlook;
2846 do {
2847 netlook = TAILQ_NEXT(netfirst, sctp_next);
2848 if (netlook == NULL) {
2849 /* End of the list */
2850 TAILQ_INSERT_TAIL(&stcb->asoc.nets, net,
2851 sctp_next);
2852 break;
2853 } else if (netlook->ro.ro_rt == NULL) {
2854 /* next one has NO route */
2855 TAILQ_INSERT_BEFORE(netfirst, net, sctp_next);
2856 break;
2857 } else if (netlook->ro.ro_rt->rt_ifp !=
2858 net->ro.ro_rt->rt_ifp) {
2859 TAILQ_INSERT_AFTER(&stcb->asoc.nets, netlook,
2860 net, sctp_next);
2861 break;
2863 /* Shift forward */
2864 netfirst = netlook;
2865 } while (netlook != NULL);
2867 /* got to have a primary set */
2868 if (stcb->asoc.primary_destination == 0) {
2869 stcb->asoc.primary_destination = net;
2870 } else if ((stcb->asoc.primary_destination->ro.ro_rt == NULL) &&
2871 (net->ro.ro_rt)) {
2872 /* No route to current primary adopt new primary */
2873 stcb->asoc.primary_destination = net;
2875 sctp_timer_start(SCTP_TIMER_TYPE_PATHMTURAISE, stcb->sctp_ep, stcb,
2876 net);
2878 return (0);
2883 * allocate an association and add it to the endpoint. The caller must
2884 * be careful to add all additional addresses once they are know right
2885 * away or else the assoc will be may experience a blackout scenario.
2887 struct sctp_tcb *
2888 sctp_aloc_assoc(struct sctp_inpcb *inp, struct sockaddr *firstaddr,
2889 int for_a_init, int *error, uint32_t override_tag)
2891 struct sctp_tcb *stcb;
2892 struct sctp_association *asoc;
2893 struct sctpasochead *head;
2894 uint16_t rport;
2895 int err;
2898 * Assumption made here:
2899 * Caller has done a sctp_findassociation_ep_addr(ep, addr's);
2900 * to make sure the address does not exist already.
2902 if (sctppcbinfo.ipi_count_asoc >= SCTP_MAX_NUM_OF_ASOC) {
2903 /* Hit max assoc, sorry no more */
2904 *error = ENOBUFS;
2905 return (NULL);
2907 SCTP_INP_RLOCK(inp);
2908 if (inp->sctp_flags & SCTP_PCB_FLAGS_IN_TCPPOOL) {
2910 * If its in the TCP pool, its NOT allowed to create an
2911 * association. The parent listener needs to call
2912 * sctp_aloc_assoc.. or the one-2-many socket. If a
2913 * peeled off, or connected one does this.. its an error.
2915 SCTP_INP_RUNLOCK(inp);
2916 *error = EINVAL;
2917 return (NULL);
2920 #ifdef SCTP_DEBUG
2921 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
2922 kprintf("Allocate an association for peer:");
2923 if (firstaddr)
2924 sctp_print_address(firstaddr);
2925 else
2926 kprintf("None\n");
2927 kprintf("Port:%d\n",
2928 ntohs(((struct sockaddr_in *)firstaddr)->sin_port));
2930 #endif /* SCTP_DEBUG */
2931 if (firstaddr->sa_family == AF_INET) {
2932 struct sockaddr_in *sin;
2933 sin = (struct sockaddr_in *)firstaddr;
2934 if ((sin->sin_port == 0) || (sin->sin_addr.s_addr == 0)) {
2935 /* Invalid address */
2936 #ifdef SCTP_DEBUG
2937 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
2938 kprintf("peer address invalid\n");
2940 #endif
2941 SCTP_INP_RUNLOCK(inp);
2942 *error = EINVAL;
2943 return (NULL);
2945 rport = sin->sin_port;
2946 } else if (firstaddr->sa_family == AF_INET6) {
2947 struct sockaddr_in6 *sin6;
2948 sin6 = (struct sockaddr_in6 *)firstaddr;
2949 if ((sin6->sin6_port == 0) ||
2950 (IN6_IS_ADDR_UNSPECIFIED(&sin6->sin6_addr))) {
2951 /* Invalid address */
2952 #ifdef SCTP_DEBUG
2953 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
2954 kprintf("peer address invalid\n");
2956 #endif
2957 SCTP_INP_RUNLOCK(inp);
2958 *error = EINVAL;
2959 return (NULL);
2961 rport = sin6->sin6_port;
2962 } else {
2963 /* not supported family type */
2964 #ifdef SCTP_DEBUG
2965 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
2966 kprintf("BAD family %d\n", firstaddr->sa_family);
2968 #endif
2969 SCTP_INP_RUNLOCK(inp);
2970 *error = EINVAL;
2971 return (NULL);
2973 SCTP_INP_RUNLOCK(inp);
2974 if (inp->sctp_flags & SCTP_PCB_FLAGS_UNBOUND) {
2976 * If you have not performed a bind, then we need to do
2977 * the ephemerial bind for you.
2979 #ifdef SCTP_DEBUG
2980 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
2981 kprintf("Doing implicit BIND\n");
2983 #endif
2985 if ((err = sctp_inpcb_bind(inp->sctp_socket,
2986 NULL,
2987 #if (defined(__FreeBSD__) && __FreeBSD_version >= 500000) || defined(__DragonFly__)
2988 NULL
2989 #else
2990 NULL
2991 #endif
2992 ))){
2993 /* bind error, probably perm */
2994 #ifdef SCTP_DEBUG
2995 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
2996 kprintf("BIND FAILS ret:%d\n", err);
2998 #endif
3000 *error = err;
3001 return (NULL);
3004 stcb = (struct sctp_tcb *)SCTP_ZONE_GET(sctppcbinfo.ipi_zone_asoc);
3005 if (stcb == NULL) {
3006 /* out of memory? */
3007 #ifdef SCTP_DEBUG
3008 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
3009 kprintf("aloc_assoc: no assoc mem left, stcb=NULL\n");
3011 #endif
3012 *error = ENOMEM;
3013 return (NULL);
3015 sctppcbinfo.ipi_count_asoc++;
3016 sctppcbinfo.ipi_gencnt_asoc++;
3018 bzero(stcb, sizeof(*stcb));
3019 asoc = &stcb->asoc;
3020 SCTP_TCB_LOCK_INIT(stcb);
3021 /* setup back pointer's */
3022 stcb->sctp_ep = inp;
3023 stcb->sctp_socket = inp->sctp_socket;
3024 if ((err = sctp_init_asoc(inp, asoc, for_a_init, override_tag))) {
3025 /* failed */
3026 SCTP_TCB_LOCK_DESTROY (stcb);
3027 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_asoc, stcb);
3028 sctppcbinfo.ipi_count_asoc--;
3029 #ifdef SCTP_DEBUG
3030 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
3031 kprintf("aloc_assoc: couldn't init asoc, out of mem?!\n");
3033 #endif
3034 *error = err;
3035 return (NULL);
3037 /* and the port */
3038 stcb->rport = rport;
3039 SCTP_INP_INFO_WLOCK();
3040 SCTP_INP_WLOCK(inp);
3041 if (inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
3042 /* inpcb freed while alloc going on */
3043 SCTP_TCB_LOCK_DESTROY (stcb);
3044 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_asoc, stcb);
3045 SCTP_INP_WUNLOCK(inp);
3046 SCTP_INP_INFO_WUNLOCK();
3047 sctppcbinfo.ipi_count_asoc--;
3048 #ifdef SCTP_DEBUG
3049 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
3050 kprintf("aloc_assoc: couldn't init asoc, out of mem?!\n");
3052 #endif
3053 *error = EINVAL;
3054 return (NULL);
3056 SCTP_TCB_LOCK(stcb);
3058 /* now that my_vtag is set, add it to the hash */
3059 head = &sctppcbinfo.sctp_asochash[SCTP_PCBHASH_ASOC(stcb->asoc.my_vtag,
3060 sctppcbinfo.hashasocmark)];
3061 /* put it in the bucket in the vtag hash of assoc's for the system */
3062 LIST_INSERT_HEAD(head, stcb, sctp_asocs);
3063 SCTP_INP_INFO_WUNLOCK();
3066 if ((err = sctp_add_remote_addr(stcb, firstaddr, 1, 1))) {
3067 /* failure.. memory error? */
3068 if (asoc->strmout)
3069 kfree(asoc->strmout, M_PCB);
3070 if (asoc->mapping_array)
3071 kfree(asoc->mapping_array, M_PCB);
3073 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_asoc, stcb);
3074 sctppcbinfo.ipi_count_asoc--;
3075 #ifdef SCTP_DEBUG
3076 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
3077 kprintf("aloc_assoc: couldn't add remote addr!\n");
3079 #endif
3080 SCTP_TCB_LOCK_DESTROY (stcb);
3081 *error = ENOBUFS;
3082 return (NULL);
3084 /* Init all the timers */
3085 #if defined(__FreeBSD__) && __FreeBSD_version >= 500000
3086 callout_init(&asoc->hb_timer.timer, 0);
3087 callout_init(&asoc->dack_timer.timer, 0);
3088 callout_init(&asoc->asconf_timer.timer, 0);
3089 callout_init(&asoc->shut_guard_timer.timer, 0);
3090 callout_init(&asoc->autoclose_timer.timer, 0);
3091 callout_init(&asoc->delayed_event_timer.timer, 0);
3092 #else
3093 callout_init(&asoc->hb_timer.timer);
3094 callout_init(&asoc->dack_timer.timer);
3095 callout_init(&asoc->asconf_timer.timer);
3096 callout_init(&asoc->shut_guard_timer.timer);
3097 callout_init(&asoc->autoclose_timer.timer);
3098 callout_init(&asoc->delayed_event_timer.timer);
3099 #endif
3100 LIST_INSERT_HEAD(&inp->sctp_asoc_list, stcb, sctp_tcblist);
3101 /* now file the port under the hash as well */
3102 if (inp->sctp_tcbhash != NULL) {
3103 head = &inp->sctp_tcbhash[SCTP_PCBHASH_ALLADDR(stcb->rport,
3104 inp->sctp_hashmark)];
3105 LIST_INSERT_HEAD(head, stcb, sctp_tcbhash);
3107 SCTP_INP_WUNLOCK(inp);
3108 #ifdef SCTP_DEBUG
3109 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
3110 kprintf("Association %p now allocated\n", stcb);
3112 #endif
3113 return (stcb);
3116 void
3117 sctp_free_remote_addr(struct sctp_nets *net)
3119 if (net == NULL)
3120 return;
3121 net->ref_count--;
3122 if (net->ref_count <= 0) {
3123 /* stop timer if running */
3124 callout_stop(&net->rxt_timer.timer);
3125 callout_stop(&net->pmtu_timer.timer);
3126 net->dest_state = SCTP_ADDR_NOT_REACHABLE;
3127 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_net, net);
3128 sctppcbinfo.ipi_count_raddr--;
3133 * remove a remote endpoint address from an association, it
3134 * will fail if the address does not exist.
3137 sctp_del_remote_addr(struct sctp_tcb *stcb, struct sockaddr *remaddr)
3140 * Here we need to remove a remote address. This is quite simple, we
3141 * first find it in the list of address for the association
3142 * (tasoc->asoc.nets) and then if it is there, we do a LIST_REMOVE on
3143 * that item.
3144 * Note we do not allow it to be removed if there are no other
3145 * addresses.
3147 struct sctp_association *asoc;
3148 struct sctp_nets *net, *net_tmp;
3149 asoc = &stcb->asoc;
3150 if (asoc->numnets < 2) {
3151 /* Must have at LEAST two remote addresses */
3152 return (-1);
3154 /* locate the address */
3155 for (net = TAILQ_FIRST(&asoc->nets); net != NULL; net = net_tmp) {
3156 net_tmp = TAILQ_NEXT(net, sctp_next);
3157 if (net->ro._l_addr.sa.sa_family != remaddr->sa_family) {
3158 continue;
3160 if (sctp_cmpaddr((struct sockaddr *)&net->ro._l_addr,
3161 remaddr)) {
3162 /* we found the guy */
3163 asoc->numnets--;
3164 TAILQ_REMOVE(&asoc->nets, net, sctp_next);
3165 sctp_free_remote_addr(net);
3166 if (net == asoc->primary_destination) {
3167 /* Reset primary */
3168 struct sctp_nets *lnet;
3169 lnet = TAILQ_FIRST(&asoc->nets);
3170 /* Try to find a confirmed primary */
3171 asoc->primary_destination =
3172 sctp_find_alternate_net(stcb, lnet);
3174 if (net == asoc->last_data_chunk_from) {
3175 /* Reset primary */
3176 asoc->last_data_chunk_from =
3177 TAILQ_FIRST(&asoc->nets);
3179 if (net == asoc->last_control_chunk_from) {
3180 /* Reset primary */
3181 asoc->last_control_chunk_from =
3182 TAILQ_FIRST(&asoc->nets);
3184 if (net == asoc->asconf_last_sent_to) {
3185 /* Reset primary */
3186 asoc->asconf_last_sent_to =
3187 TAILQ_FIRST(&asoc->nets);
3189 return (0);
3192 /* not found. */
3193 return (-2);
3197 static void
3198 sctp_add_vtag_to_timewait(struct sctp_inpcb *inp, u_int32_t tag)
3200 struct sctpvtaghead *chain;
3201 struct sctp_tagblock *twait_block;
3202 struct timeval now;
3203 int set, i;
3204 SCTP_GETTIME_TIMEVAL(&now);
3205 chain = &sctppcbinfo.vtag_timewait[(tag % SCTP_STACK_VTAG_HASH_SIZE)];
3206 set = 0;
3207 if (!LIST_EMPTY(chain)) {
3208 /* Block(s) present, lets find space, and expire on the fly */
3209 LIST_FOREACH(twait_block, chain, sctp_nxt_tagblock) {
3210 for (i = 0; i < SCTP_NUMBER_IN_VTAG_BLOCK; i++) {
3211 if ((twait_block->vtag_block[i].v_tag == 0) &&
3212 !set) {
3213 twait_block->vtag_block[0].tv_sec_at_expire =
3214 now.tv_sec + SCTP_TIME_WAIT;
3215 twait_block->vtag_block[0].v_tag = tag;
3216 set = 1;
3217 } else if ((twait_block->vtag_block[i].v_tag) &&
3218 ((long)twait_block->vtag_block[i].tv_sec_at_expire >
3219 now.tv_sec)) {
3220 /* Audit expires this guy */
3221 twait_block->vtag_block[i].tv_sec_at_expire = 0;
3222 twait_block->vtag_block[i].v_tag = 0;
3223 if (set == 0) {
3224 /* Reuse it for my new tag */
3225 twait_block->vtag_block[0].tv_sec_at_expire = now.tv_sec + SCTP_TIME_WAIT;
3226 twait_block->vtag_block[0].v_tag = tag;
3227 set = 1;
3231 if (set) {
3233 * We only do up to the block where we can
3234 * place our tag for audits
3236 break;
3240 /* Need to add a new block to chain */
3241 if (!set) {
3242 twait_block = kmalloc(sizeof(struct sctp_tagblock), M_PCB,
3243 M_NOWAIT);
3244 if (twait_block == NULL) {
3245 return;
3247 memset(twait_block, 0, sizeof(struct sctp_timewait));
3248 LIST_INSERT_HEAD(chain, twait_block, sctp_nxt_tagblock);
3249 twait_block->vtag_block[0].tv_sec_at_expire = now.tv_sec +
3250 SCTP_TIME_WAIT;
3251 twait_block->vtag_block[0].v_tag = tag;
3256 static void
3257 sctp_iterator_asoc_being_freed(struct sctp_inpcb *inp, struct sctp_tcb *stcb)
3259 struct sctp_iterator *it;
3263 /* Unlock the tcb lock we do this so
3264 * we avoid a dead lock scenario where
3265 * the iterator is waiting on the TCB lock
3266 * and the TCB lock is waiting on the iterator
3267 * lock.
3269 SCTP_ITERATOR_LOCK();
3270 SCTP_INP_INFO_WLOCK();
3271 SCTP_INP_WLOCK(inp);
3272 SCTP_TCB_LOCK(stcb);
3274 it = stcb->asoc.stcb_starting_point_for_iterator;
3275 if (it == NULL) {
3276 return;
3278 if (it->inp != stcb->sctp_ep) {
3279 /* hm, focused on the wrong one? */
3280 return;
3282 if (it->stcb != stcb) {
3283 return;
3285 it->stcb = LIST_NEXT(stcb, sctp_tcblist);
3286 if (it->stcb == NULL) {
3287 /* done with all asoc's in this assoc */
3288 if (it->iterator_flags & SCTP_ITERATOR_DO_SINGLE_INP) {
3289 it->inp = NULL;
3290 } else {
3292 it->inp = LIST_NEXT(inp, sctp_list);
3298 * Free the association after un-hashing the remote port.
3300 void
3301 sctp_free_assoc(struct sctp_inpcb *inp, struct sctp_tcb *stcb)
3303 struct sctp_association *asoc;
3304 struct sctp_nets *net, *prev;
3305 struct sctp_laddr *laddr;
3306 struct sctp_tmit_chunk *chk;
3307 struct sctp_asconf_addr *aparam;
3308 struct sctp_socket_q_list *sq;
3310 /* first, lets purge the entry from the hash table. */
3311 if (stcb->asoc.state == 0) {
3312 kprintf("Freeing already free association:%p - huh??\n",
3313 stcb);
3314 return;
3316 asoc = &stcb->asoc;
3317 asoc->state = 0;
3318 /* now clean up any other timers */
3319 callout_stop(&asoc->hb_timer.timer);
3320 callout_stop(&asoc->dack_timer.timer);
3321 callout_stop(&asoc->asconf_timer.timer);
3322 callout_stop(&asoc->shut_guard_timer.timer);
3323 callout_stop(&asoc->autoclose_timer.timer);
3324 callout_stop(&asoc->delayed_event_timer.timer);
3325 TAILQ_FOREACH(net, &asoc->nets, sctp_next) {
3326 callout_stop(&net->rxt_timer.timer);
3327 callout_stop(&net->pmtu_timer.timer);
3330 /* Iterator asoc being freed we send an
3331 * unlocked TCB. It returns with INP_INFO
3332 * and INP write locked and the TCB locked
3333 * too and of course the iterator lock
3334 * in place as well..
3336 SCTP_TCB_UNLOCK(stcb);
3337 sctp_iterator_asoc_being_freed(inp, stcb);
3339 /* Null all of my entry's on the socket q */
3340 TAILQ_FOREACH(sq, &inp->sctp_queue_list, next_sq) {
3341 if (sq->tcb == stcb) {
3342 sq->tcb = NULL;
3346 if (inp->sctp_tcb_at_block == (void *)stcb) {
3347 inp->error_on_block = ECONNRESET;
3350 if (inp->sctp_tcbhash) {
3351 LIST_REMOVE(stcb, sctp_tcbhash);
3353 /* Now lets remove it from the list of ALL associations in the EP */
3354 LIST_REMOVE(stcb, sctp_tcblist);
3355 SCTP_INP_WUNLOCK(inp);
3356 SCTP_ITERATOR_UNLOCK();
3359 /* pull from vtag hash */
3360 LIST_REMOVE(stcb, sctp_asocs);
3363 * Now before we can free the assoc, we must remove all of the
3364 * networks and any other allocated space.. i.e. add removes here
3365 * before the SCTP_ZONE_FREE() of the tasoc entry.
3368 sctp_add_vtag_to_timewait(inp, asoc->my_vtag);
3369 SCTP_INP_INFO_WUNLOCK();
3370 prev = NULL;
3371 while (!TAILQ_EMPTY(&asoc->nets)) {
3372 net = TAILQ_FIRST(&asoc->nets);
3373 /* pull from list */
3374 if ((sctppcbinfo.ipi_count_raddr == 0) || (prev == net)) {
3375 break;
3377 prev = net;
3378 TAILQ_REMOVE(&asoc->nets, net, sctp_next);
3379 /* free it */
3380 net->ref_count = 0;
3381 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_net, net);
3382 sctppcbinfo.ipi_count_raddr--;
3385 * The chunk lists and such SHOULD be empty but we check them
3386 * just in case.
3388 /* anything on the wheel needs to be removed */
3389 while (!TAILQ_EMPTY(&asoc->out_wheel)) {
3390 struct sctp_stream_out *outs;
3391 outs = TAILQ_FIRST(&asoc->out_wheel);
3392 TAILQ_REMOVE(&asoc->out_wheel, outs, next_spoke);
3393 /* now clean up any chunks here */
3394 chk = TAILQ_FIRST(&outs->outqueue);
3395 while (chk) {
3396 TAILQ_REMOVE(&outs->outqueue, chk, sctp_next);
3397 if (chk->data) {
3398 sctp_m_freem(chk->data);
3399 chk->data = NULL;
3401 chk->whoTo = NULL;
3402 chk->asoc = NULL;
3403 /* Free the chunk */
3404 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
3405 sctppcbinfo.ipi_count_chunk--;
3406 sctppcbinfo.ipi_gencnt_chunk++;
3407 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
3408 panic("Chunk count is negative");
3410 chk = TAILQ_FIRST(&outs->outqueue);
3412 outs = TAILQ_FIRST(&asoc->out_wheel);
3415 if (asoc->pending_reply) {
3416 kfree(asoc->pending_reply, M_PCB);
3417 asoc->pending_reply = NULL;
3419 chk = TAILQ_FIRST(&asoc->pending_reply_queue);
3420 while (chk) {
3421 TAILQ_REMOVE(&asoc->pending_reply_queue, chk, sctp_next);
3422 if (chk->data) {
3423 sctp_m_freem(chk->data);
3424 chk->data = NULL;
3426 chk->whoTo = NULL;
3427 chk->asoc = NULL;
3428 /* Free the chunk */
3429 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
3430 sctppcbinfo.ipi_count_chunk--;
3431 sctppcbinfo.ipi_gencnt_chunk++;
3432 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
3433 panic("Chunk count is negative");
3435 chk = TAILQ_FIRST(&asoc->pending_reply_queue);
3437 /* pending send queue SHOULD be empty */
3438 if (!TAILQ_EMPTY(&asoc->send_queue)) {
3439 chk = TAILQ_FIRST(&asoc->send_queue);
3440 while (chk) {
3441 TAILQ_REMOVE(&asoc->send_queue, chk, sctp_next);
3442 if (chk->data) {
3443 sctp_m_freem(chk->data);
3444 chk->data = NULL;
3446 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
3447 sctppcbinfo.ipi_count_chunk--;
3448 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
3449 panic("Chunk count is negative");
3451 sctppcbinfo.ipi_gencnt_chunk++;
3452 chk = TAILQ_FIRST(&asoc->send_queue);
3455 /* sent queue SHOULD be empty */
3456 if (!TAILQ_EMPTY(&asoc->sent_queue)) {
3457 chk = TAILQ_FIRST(&asoc->sent_queue);
3458 while (chk) {
3459 TAILQ_REMOVE(&asoc->sent_queue, chk, sctp_next);
3460 if (chk->data) {
3461 sctp_m_freem(chk->data);
3462 chk->data = NULL;
3464 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
3465 sctppcbinfo.ipi_count_chunk--;
3466 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
3467 panic("Chunk count is negative");
3469 sctppcbinfo.ipi_gencnt_chunk++;
3470 chk = TAILQ_FIRST(&asoc->sent_queue);
3473 /* control queue MAY not be empty */
3474 if (!TAILQ_EMPTY(&asoc->control_send_queue)) {
3475 chk = TAILQ_FIRST(&asoc->control_send_queue);
3476 while (chk) {
3477 TAILQ_REMOVE(&asoc->control_send_queue, chk, sctp_next);
3478 if (chk->data) {
3479 sctp_m_freem(chk->data);
3480 chk->data = NULL;
3482 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
3483 sctppcbinfo.ipi_count_chunk--;
3484 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
3485 panic("Chunk count is negative");
3487 sctppcbinfo.ipi_gencnt_chunk++;
3488 chk = TAILQ_FIRST(&asoc->control_send_queue);
3491 if (!TAILQ_EMPTY(&asoc->reasmqueue)) {
3492 chk = TAILQ_FIRST(&asoc->reasmqueue);
3493 while (chk) {
3494 TAILQ_REMOVE(&asoc->reasmqueue, chk, sctp_next);
3495 if (chk->data) {
3496 sctp_m_freem(chk->data);
3497 chk->data = NULL;
3499 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
3500 sctppcbinfo.ipi_count_chunk--;
3501 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
3502 panic("Chunk count is negative");
3504 sctppcbinfo.ipi_gencnt_chunk++;
3505 chk = TAILQ_FIRST(&asoc->reasmqueue);
3508 if (!TAILQ_EMPTY(&asoc->delivery_queue)) {
3509 chk = TAILQ_FIRST(&asoc->delivery_queue);
3510 while (chk) {
3511 TAILQ_REMOVE(&asoc->delivery_queue, chk, sctp_next);
3512 if (chk->data) {
3513 sctp_m_freem(chk->data);
3514 chk->data = NULL;
3516 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
3517 sctppcbinfo.ipi_count_chunk--;
3518 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
3519 panic("Chunk count is negative");
3521 sctppcbinfo.ipi_gencnt_chunk++;
3522 chk = TAILQ_FIRST(&asoc->delivery_queue);
3525 if (asoc->mapping_array) {
3526 kfree(asoc->mapping_array, M_PCB);
3527 asoc->mapping_array = NULL;
3530 /* the stream outs */
3531 if (asoc->strmout) {
3532 kfree(asoc->strmout, M_PCB);
3533 asoc->strmout = NULL;
3535 asoc->streamoutcnt = 0;
3536 if (asoc->strmin) {
3537 int i;
3538 for (i = 0; i < asoc->streamincnt; i++) {
3539 if (!TAILQ_EMPTY(&asoc->strmin[i].inqueue)) {
3540 /* We have somethings on the streamin queue */
3541 chk = TAILQ_FIRST(&asoc->strmin[i].inqueue);
3542 while (chk) {
3543 TAILQ_REMOVE(&asoc->strmin[i].inqueue,
3544 chk, sctp_next);
3545 if (chk->data) {
3546 sctp_m_freem(chk->data);
3547 chk->data = NULL;
3549 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk,
3550 chk);
3551 sctppcbinfo.ipi_count_chunk--;
3552 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
3553 panic("Chunk count is negative");
3555 sctppcbinfo.ipi_gencnt_chunk++;
3556 chk = TAILQ_FIRST(&asoc->strmin[i].inqueue);
3560 kfree(asoc->strmin, M_PCB);
3561 asoc->strmin = NULL;
3563 asoc->streamincnt = 0;
3564 /* local addresses, if any */
3565 while (!LIST_EMPTY(&asoc->sctp_local_addr_list)) {
3566 laddr = LIST_FIRST(&asoc->sctp_local_addr_list);
3567 LIST_REMOVE(laddr, sctp_nxt_addr);
3568 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_laddr, laddr);
3569 sctppcbinfo.ipi_count_laddr--;
3571 /* pending asconf (address) parameters */
3572 while (!TAILQ_EMPTY(&asoc->asconf_queue)) {
3573 aparam = TAILQ_FIRST(&asoc->asconf_queue);
3574 TAILQ_REMOVE(&asoc->asconf_queue, aparam, next);
3575 kfree(aparam, M_PCB);
3577 if (asoc->last_asconf_ack_sent != NULL) {
3578 sctp_m_freem(asoc->last_asconf_ack_sent);
3579 asoc->last_asconf_ack_sent = NULL;
3581 /* Insert new items here :> */
3583 /* Get rid of LOCK */
3584 SCTP_TCB_LOCK_DESTROY(stcb);
3586 /* now clean up the tasoc itself */
3587 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_asoc, stcb);
3588 sctppcbinfo.ipi_count_asoc--;
3589 if ((inp->sctp_socket->so_snd.ssb_cc) ||
3590 (inp->sctp_socket->so_snd.ssb_mbcnt)) {
3591 /* This will happen when a abort is done */
3592 inp->sctp_socket->so_snd.ssb_cc = 0;
3593 inp->sctp_socket->so_snd.ssb_mbcnt = 0;
3595 if (inp->sctp_flags & SCTP_PCB_FLAGS_TCPTYPE) {
3596 if ((inp->sctp_flags & SCTP_PCB_FLAGS_IN_TCPPOOL) == 0) {
3597 if (inp->sctp_flags & SCTP_PCB_FLAGS_CONNECTED) {
3599 * For the base fd, that is NOT in TCP pool we
3600 * turn off the connected flag. This allows
3601 * non-listening endpoints to connect/shutdown/
3602 * connect.
3604 inp->sctp_flags &= ~SCTP_PCB_FLAGS_CONNECTED;
3605 soisdisconnected(inp->sctp_socket);
3608 * For those that are in the TCP pool we just leave
3609 * so it cannot be used. When they close the fd we
3610 * will free it all.
3614 if (inp->sctp_flags & SCTP_PCB_FLAGS_SOCKET_GONE) {
3615 sctp_inpcb_free(inp, 0);
3621 * determine if a destination is "reachable" based upon the addresses
3622 * bound to the current endpoint (e.g. only v4 or v6 currently bound)
3625 * FIX: if we allow assoc-level bindx(), then this needs to be fixed
3626 * to use assoc level v4/v6 flags, as the assoc *may* not have the
3627 * same address types bound as its endpoint
3630 sctp_destination_is_reachable(struct sctp_tcb *stcb, struct sockaddr *destaddr)
3632 struct sctp_inpcb *inp;
3633 int answer;
3635 /* No locks here, the TCB, in all cases is already
3636 * locked and an assoc is up. There is either a
3637 * INP lock by the caller applied (in asconf case when
3638 * deleting an address) or NOT in the HB case, however
3639 * if HB then the INP increment is up and the INP
3640 * will not be removed (on top of the fact that
3641 * we have a TCB lock). So we only want to
3642 * read the sctp_flags, which is either bound-all
3643 * or not.. no protection needed since once an
3644 * assoc is up you can't be changing your binding.
3646 inp = stcb->sctp_ep;
3647 if (inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) {
3648 /* if bound all, destination is not restricted */
3649 /* RRS: Question during lock work: Is this
3650 * correct? If you are bound-all you still
3651 * might need to obey the V4--V6 flags???
3652 * IMO this bound-all stuff needs to be removed!
3654 return (1);
3656 /* NOTE: all "scope" checks are done when local addresses are added */
3657 if (destaddr->sa_family == AF_INET6) {
3658 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
3659 answer = inp->inp_vflag & INP_IPV6;
3660 #else
3661 answer = inp->ip_inp.inp.inp_vflag & INP_IPV6;
3662 #endif
3663 } else if (destaddr->sa_family == AF_INET) {
3664 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
3665 answer = inp->inp_vflag & INP_IPV4;
3666 #else
3667 answer = inp->ip_inp.inp.inp_vflag & INP_IPV4;
3668 #endif
3669 } else {
3670 /* invalid family, so it's unreachable */
3671 answer = 0;
3673 return (answer);
3677 * update the inp_vflags on an endpoint
3679 static void
3680 sctp_update_ep_vflag(struct sctp_inpcb *inp) {
3681 struct sctp_laddr *laddr;
3683 /* first clear the flag */
3684 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
3685 inp->inp_vflag = 0;
3686 #else
3687 inp->ip_inp.inp.inp_vflag = 0;
3688 #endif
3689 /* set the flag based on addresses on the ep list */
3690 LIST_FOREACH(laddr, &inp->sctp_addr_list, sctp_nxt_addr) {
3691 if (laddr->ifa == NULL) {
3692 #ifdef SCTP_DEBUG
3693 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
3694 kprintf("An ounce of prevention is worth a pound of cure\n");
3696 #endif /* SCTP_DEBUG */
3697 continue;
3699 if (laddr->ifa->ifa_addr) {
3700 continue;
3702 if (laddr->ifa->ifa_addr->sa_family == AF_INET6) {
3703 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
3704 inp->inp_vflag |= INP_IPV6;
3705 #else
3706 inp->ip_inp.inp.inp_vflag |= INP_IPV6;
3707 #endif
3708 } else if (laddr->ifa->ifa_addr->sa_family == AF_INET) {
3709 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
3710 inp->inp_vflag |= INP_IPV4;
3711 #else
3712 inp->ip_inp.inp.inp_vflag |= INP_IPV4;
3713 #endif
3719 * Add the address to the endpoint local address list
3720 * There is nothing to be done if we are bound to all addresses
3723 sctp_add_local_addr_ep(struct sctp_inpcb *inp, struct ifaddr *ifa)
3725 struct sctp_laddr *laddr;
3726 int fnd, error;
3727 fnd = 0;
3729 if (inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) {
3730 /* You are already bound to all. You have it already */
3731 return (0);
3733 if (ifa->ifa_addr->sa_family == AF_INET6) {
3734 struct in6_ifaddr *ifa6;
3735 ifa6 = (struct in6_ifaddr *)ifa;
3736 if (ifa6->ia6_flags & (IN6_IFF_DETACHED |
3737 IN6_IFF_DEPRECATED | IN6_IFF_ANYCAST | IN6_IFF_NOTREADY))
3738 /* Can't bind a non-existent addr. */
3739 return (-1);
3741 /* first, is it already present? */
3742 LIST_FOREACH(laddr, &inp->sctp_addr_list, sctp_nxt_addr) {
3743 if (laddr->ifa == ifa) {
3744 fnd = 1;
3745 break;
3749 if (((inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) == 0) && (fnd == 0)) {
3750 /* Not bound to all */
3751 error = sctp_insert_laddr(&inp->sctp_addr_list, ifa);
3752 if (error != 0)
3753 return (error);
3754 inp->laddr_count++;
3755 /* update inp_vflag flags */
3756 if (ifa->ifa_addr->sa_family == AF_INET6) {
3757 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
3758 inp->inp_vflag |= INP_IPV6;
3759 #else
3760 inp->ip_inp.inp.inp_vflag |= INP_IPV6;
3761 #endif
3762 } else if (ifa->ifa_addr->sa_family == AF_INET) {
3763 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
3764 inp->inp_vflag |= INP_IPV4;
3765 #else
3766 inp->ip_inp.inp.inp_vflag |= INP_IPV4;
3767 #endif
3770 return (0);
3775 * select a new (hopefully reachable) destination net
3776 * (should only be used when we deleted an ep addr that is the
3777 * only usable source address to reach the destination net)
3779 static void
3780 sctp_select_primary_destination(struct sctp_tcb *stcb)
3782 struct sctp_nets *net;
3784 TAILQ_FOREACH(net, &stcb->asoc.nets, sctp_next) {
3785 /* for now, we'll just pick the first reachable one we find */
3786 if (net->dest_state & SCTP_ADDR_UNCONFIRMED)
3787 continue;
3788 if (sctp_destination_is_reachable(stcb,
3789 (struct sockaddr *)&net->ro._l_addr)) {
3790 /* found a reachable destination */
3791 stcb->asoc.primary_destination = net;
3794 /* I can't there from here! ...we're gonna die shortly... */
3799 * Delete the address from the endpoint local address list
3800 * There is nothing to be done if we are bound to all addresses
3803 sctp_del_local_addr_ep(struct sctp_inpcb *inp, struct ifaddr *ifa)
3805 struct sctp_laddr *laddr;
3806 int fnd;
3807 fnd = 0;
3808 if (inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) {
3809 /* You are already bound to all. You have it already */
3810 return (EINVAL);
3813 LIST_FOREACH(laddr, &inp->sctp_addr_list, sctp_nxt_addr) {
3814 if (laddr->ifa == ifa) {
3815 fnd = 1;
3816 break;
3819 if (fnd && (inp->laddr_count < 2)) {
3820 /* can't delete unless there are at LEAST 2 addresses */
3821 return (-1);
3823 if (((inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) == 0) && (fnd)) {
3825 * clean up any use of this address
3826 * go through our associations and clear any
3827 * last_used_address that match this one
3828 * for each assoc, see if a new primary_destination is needed
3830 struct sctp_tcb *stcb;
3832 /* clean up "next_addr_touse" */
3833 if (inp->next_addr_touse == laddr)
3834 /* delete this address */
3835 inp->next_addr_touse = NULL;
3837 /* clean up "last_used_address" */
3838 LIST_FOREACH(stcb, &inp->sctp_asoc_list, sctp_tcblist) {
3839 if (stcb->asoc.last_used_address == laddr)
3840 /* delete this address */
3841 stcb->asoc.last_used_address = NULL;
3842 } /* for each tcb */
3844 /* remove it from the ep list */
3845 sctp_remove_laddr(laddr);
3846 inp->laddr_count--;
3847 /* update inp_vflag flags */
3848 sctp_update_ep_vflag(inp);
3849 /* select a new primary destination if needed */
3850 LIST_FOREACH(stcb, &inp->sctp_asoc_list, sctp_tcblist) {
3851 /* presume caller (sctp_asconf.c) already owns INP lock */
3852 SCTP_TCB_LOCK(stcb);
3853 if (sctp_destination_is_reachable(stcb,
3854 (struct sockaddr *)&stcb->asoc.primary_destination->ro._l_addr) == 0) {
3855 sctp_select_primary_destination(stcb);
3857 SCTP_TCB_UNLOCK(stcb);
3858 } /* for each tcb */
3860 return (0);
3864 * Add the addr to the TCB local address list
3865 * For the BOUNDALL or dynamic case, this is a "pending" address list
3866 * (eg. addresses waiting for an ASCONF-ACK response)
3867 * For the subset binding, static case, this is a "valid" address list
3870 sctp_add_local_addr_assoc(struct sctp_tcb *stcb, struct ifaddr *ifa)
3872 struct sctp_inpcb *inp;
3873 struct sctp_laddr *laddr;
3874 int error;
3876 /* Assumes TCP is locked.. and possiblye
3877 * the INP. May need to confirm/fix that if
3878 * we need it and is not the case.
3880 inp = stcb->sctp_ep;
3881 if (ifa->ifa_addr->sa_family == AF_INET6) {
3882 struct in6_ifaddr *ifa6;
3883 ifa6 = (struct in6_ifaddr *)ifa;
3884 if (ifa6->ia6_flags & (IN6_IFF_DETACHED |
3885 /* IN6_IFF_DEPRECATED | */
3886 IN6_IFF_ANYCAST |
3887 IN6_IFF_NOTREADY))
3888 /* Can't bind a non-existent addr. */
3889 return (-1);
3891 /* does the address already exist? */
3892 LIST_FOREACH(laddr, &stcb->asoc.sctp_local_addr_list, sctp_nxt_addr) {
3893 if (laddr->ifa == ifa) {
3894 return (-1);
3898 /* add to the list */
3899 error = sctp_insert_laddr(&stcb->asoc.sctp_local_addr_list, ifa);
3900 if (error != 0)
3901 return (error);
3902 return (0);
3906 * insert an laddr entry with the given ifa for the desired list
3909 sctp_insert_laddr(struct sctpladdr *list, struct ifaddr *ifa) {
3910 struct sctp_laddr *laddr;
3912 laddr = (struct sctp_laddr *)SCTP_ZONE_GET(sctppcbinfo.ipi_zone_laddr);
3913 if (laddr == NULL) {
3914 /* out of memory? */
3915 return (EINVAL);
3917 sctppcbinfo.ipi_count_laddr++;
3918 sctppcbinfo.ipi_gencnt_laddr++;
3919 bzero(laddr, sizeof(*laddr));
3920 laddr->ifa = ifa;
3921 /* insert it */
3922 LIST_INSERT_HEAD(list, laddr, sctp_nxt_addr);
3924 return (0);
3928 * Remove an laddr entry from the local address list (on an assoc)
3930 void
3931 sctp_remove_laddr(struct sctp_laddr *laddr)
3933 /* remove from the list */
3934 LIST_REMOVE(laddr, sctp_nxt_addr);
3935 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_laddr, laddr);
3936 sctppcbinfo.ipi_count_laddr--;
3937 sctppcbinfo.ipi_gencnt_laddr++;
3941 * Remove an address from the TCB local address list
3944 sctp_del_local_addr_assoc(struct sctp_tcb *stcb, struct ifaddr *ifa)
3946 struct sctp_inpcb *inp;
3947 struct sctp_laddr *laddr;
3949 /* This is called by asconf work. It is assumed that
3950 * a) The TCB is locked
3951 * and
3952 * b) The INP is locked.
3953 * This is true in as much as I can trace through
3954 * the entry asconf code where I did these locks.
3955 * Again, the ASCONF code is a bit different in
3956 * that it does lock the INP during its work often
3957 * times. This must be since we don't want other
3958 * proc's looking up things while what they are
3959 * looking up is changing :-D
3962 inp = stcb->sctp_ep;
3963 /* if subset bound and don't allow ASCONF's, can't delete last */
3964 if (((inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) == 0) &&
3965 ((inp->sctp_flags & SCTP_PCB_FLAGS_DO_ASCONF) == 0)) {
3966 if (stcb->asoc.numnets < 2) {
3967 /* can't delete last address */
3968 return (-1);
3972 LIST_FOREACH(laddr, &stcb->asoc.sctp_local_addr_list, sctp_nxt_addr) {
3973 /* remove the address if it exists */
3974 if (laddr->ifa == NULL)
3975 continue;
3976 if (laddr->ifa == ifa) {
3977 sctp_remove_laddr(laddr);
3978 return (0);
3982 /* address not found! */
3983 return (-1);
3987 * Remove an address from the TCB local address list
3988 * lookup using a sockaddr addr
3991 sctp_del_local_addr_assoc_sa(struct sctp_tcb *stcb, struct sockaddr *sa)
3993 struct sctp_inpcb *inp;
3994 struct sctp_laddr *laddr;
3995 struct sockaddr *l_sa;
3998 * This function I find does not seem to have a caller.
3999 * As such we NEED TO DELETE this code. If we do
4000 * find a caller, the caller MUST have locked the TCB
4001 * at the least and probably the INP as well.
4003 inp = stcb->sctp_ep;
4004 /* if subset bound and don't allow ASCONF's, can't delete last */
4005 if (((inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) == 0) &&
4006 ((inp->sctp_flags & SCTP_PCB_FLAGS_DO_ASCONF) == 0)) {
4007 if (stcb->asoc.numnets < 2) {
4008 /* can't delete last address */
4009 return (-1);
4013 LIST_FOREACH(laddr, &stcb->asoc.sctp_local_addr_list, sctp_nxt_addr) {
4014 /* make sure the address exists */
4015 if (laddr->ifa == NULL)
4016 continue;
4017 if (laddr->ifa->ifa_addr == NULL)
4018 continue;
4020 l_sa = laddr->ifa->ifa_addr;
4021 if (l_sa->sa_family == AF_INET6) {
4022 /* IPv6 address */
4023 struct sockaddr_in6 *sin1, *sin2;
4024 sin1 = (struct sockaddr_in6 *)l_sa;
4025 sin2 = (struct sockaddr_in6 *)sa;
4026 if (memcmp(&sin1->sin6_addr, &sin2->sin6_addr,
4027 sizeof(struct in6_addr)) == 0) {
4028 /* matched */
4029 sctp_remove_laddr(laddr);
4030 return (0);
4032 } else if (l_sa->sa_family == AF_INET) {
4033 /* IPv4 address */
4034 struct sockaddr_in *sin1, *sin2;
4035 sin1 = (struct sockaddr_in *)l_sa;
4036 sin2 = (struct sockaddr_in *)sa;
4037 if (sin1->sin_addr.s_addr == sin2->sin_addr.s_addr) {
4038 /* matched */
4039 sctp_remove_laddr(laddr);
4040 return (0);
4042 } else {
4043 /* invalid family */
4044 return (-1);
4046 } /* end foreach */
4047 /* address not found! */
4048 return (-1);
4051 static char sctp_pcb_initialized = 0;
4053 #if defined(__FreeBSD__) || defined(__APPLE__)
4054 /* sysctl */
4055 /* not used by DragonFly SCTP_ZONE_INIT macro */
4056 static int sctp_max_number_of_assoc = SCTP_MAX_NUM_OF_ASOC;
4057 static int sctp_scale_up_for_address = SCTP_SCALE_FOR_ADDR;
4059 #endif /* FreeBSD || APPLE || DragonFly */
4061 #ifndef SCTP_TCBHASHSIZE
4062 #define SCTP_TCBHASHSIZE 1024
4063 #endif
4065 #ifndef SCTP_CHUNKQUEUE_SCALE
4066 #define SCTP_CHUNKQUEUE_SCALE 10
4067 #endif
4069 void
4070 sctp_pcb_init(void)
4073 * SCTP initialization for the PCB structures
4074 * should be called by the sctp_init() funciton.
4076 int i;
4077 int hashtblsize = SCTP_TCBHASHSIZE;
4079 #if defined(__FreeBSD__) || defined(__APPLE__)
4080 /* not used by DragonFly SCTP_ZONE_INIT macro */
4081 int sctp_chunkscale = SCTP_CHUNKQUEUE_SCALE;
4082 #endif
4084 if (sctp_pcb_initialized != 0) {
4085 /* error I was called twice */
4086 return;
4088 sctp_pcb_initialized = 1;
4090 /* Init all peg counts */
4091 for (i = 0; i < SCTP_NUMBER_OF_PEGS; i++) {
4092 sctp_pegs[i] = 0;
4095 /* init the empty list of (All) Endpoints */
4096 LIST_INIT(&sctppcbinfo.listhead);
4098 /* init the iterator head */
4099 LIST_INIT(&sctppcbinfo.iteratorhead);
4101 /* init the hash table of endpoints */
4102 #if defined(__FreeBSD__)
4103 #if defined(__FreeBSD_cc_version) && __FreeBSD_cc_version >= 440000
4104 TUNABLE_INT_FETCH("net.inet.sctp.tcbhashsize", &hashtblsize);
4105 TUNABLE_INT_FETCH("net.inet.sctp.pcbhashsize", &sctp_pcbtblsize);
4106 TUNABLE_INT_FETCH("net.inet.sctp.chunkscale", &sctp_chunkscale);
4107 #else
4108 TUNABLE_INT_FETCH("net.inet.sctp.tcbhashsize", SCTP_TCBHASHSIZE,
4109 hashtblsize);
4110 TUNABLE_INT_FETCH("net.inet.sctp.pcbhashsize", SCTP_PCBHASHSIZE,
4111 sctp_pcbtblsize);
4112 TUNABLE_INT_FETCH("net.inet.sctp.chunkscale", SCTP_CHUNKQUEUE_SCALE,
4113 sctp_chunkscale);
4114 #endif
4115 #endif
4117 sctppcbinfo.sctp_asochash = hashinit((hashtblsize * 31),
4118 #ifdef __NetBSD__
4119 HASH_LIST,
4120 #endif
4121 M_PCB,
4122 #if defined(__NetBSD__) || defined(__OpenBSD__)
4123 M_WAITOK,
4124 #endif
4125 &sctppcbinfo.hashasocmark);
4127 sctppcbinfo.sctp_ephash = hashinit(hashtblsize,
4128 #ifdef __NetBSD__
4129 HASH_LIST,
4130 #endif
4131 M_PCB,
4132 #if defined(__NetBSD__) || defined(__OpenBSD__)
4133 M_WAITOK,
4134 #endif
4135 &sctppcbinfo.hashmark);
4137 sctppcbinfo.sctp_tcpephash = hashinit(hashtblsize,
4138 #ifdef __NetBSD__
4139 HASH_LIST,
4140 #endif
4141 M_PCB,
4142 #if defined(__NetBSD__) || defined(__OpenBSD__)
4143 M_WAITOK,
4144 #endif
4145 &sctppcbinfo.hashtcpmark);
4147 sctppcbinfo.hashtblsize = hashtblsize;
4149 /* init the zones */
4151 * FIX ME: Should check for NULL returns, but if it does fail we
4152 * are doomed to panic anyways... add later maybe.
4154 SCTP_ZONE_INIT(sctppcbinfo.ipi_zone_ep, "sctp_ep",
4155 sizeof(struct sctp_inpcb), maxsockets);
4157 SCTP_ZONE_INIT(sctppcbinfo.ipi_zone_asoc, "sctp_asoc",
4158 sizeof(struct sctp_tcb), sctp_max_number_of_assoc);
4160 SCTP_ZONE_INIT(sctppcbinfo.ipi_zone_laddr, "sctp_laddr",
4161 sizeof(struct sctp_laddr),
4162 (sctp_max_number_of_assoc * sctp_scale_up_for_address));
4164 SCTP_ZONE_INIT(sctppcbinfo.ipi_zone_net, "sctp_raddr",
4165 sizeof(struct sctp_nets),
4166 (sctp_max_number_of_assoc * sctp_scale_up_for_address));
4168 SCTP_ZONE_INIT(sctppcbinfo.ipi_zone_chunk, "sctp_chunk",
4169 sizeof(struct sctp_tmit_chunk),
4170 (sctp_max_number_of_assoc * sctp_scale_up_for_address *
4171 sctp_chunkscale));
4173 SCTP_ZONE_INIT(sctppcbinfo.ipi_zone_sockq, "sctp_sockq",
4174 sizeof(struct sctp_socket_q_list),
4175 (sctp_max_number_of_assoc * sctp_scale_up_for_address *
4176 sctp_chunkscale));
4178 /* Master Lock INIT for info structure */
4179 SCTP_INP_INFO_LOCK_INIT();
4180 SCTP_ITERATOR_LOCK_INIT();
4181 /* not sure if we need all the counts */
4182 sctppcbinfo.ipi_count_ep = 0;
4183 sctppcbinfo.ipi_gencnt_ep = 0;
4184 /* assoc/tcb zone info */
4185 sctppcbinfo.ipi_count_asoc = 0;
4186 sctppcbinfo.ipi_gencnt_asoc = 0;
4187 /* local addrlist zone info */
4188 sctppcbinfo.ipi_count_laddr = 0;
4189 sctppcbinfo.ipi_gencnt_laddr = 0;
4190 /* remote addrlist zone info */
4191 sctppcbinfo.ipi_count_raddr = 0;
4192 sctppcbinfo.ipi_gencnt_raddr = 0;
4193 /* chunk info */
4194 sctppcbinfo.ipi_count_chunk = 0;
4195 sctppcbinfo.ipi_gencnt_chunk = 0;
4197 /* socket queue zone info */
4198 sctppcbinfo.ipi_count_sockq = 0;
4199 sctppcbinfo.ipi_gencnt_sockq = 0;
4201 /* mbuf tracker */
4202 sctppcbinfo.mbuf_track = 0;
4203 /* port stuff */
4204 #if defined(__FreeBSD__) || defined(__OpenBSD__) || defined(__APPLE__) \
4205 || defined(__DragonFly__)
4206 sctppcbinfo.lastlow = ipport_firstauto;
4207 #else
4208 sctppcbinfo.lastlow = anonportmin;
4209 #endif
4210 /* Init the TIMEWAIT list */
4211 for (i = 0; i < SCTP_STACK_VTAG_HASH_SIZE; i++) {
4212 LIST_INIT(&sctppcbinfo.vtag_timewait[i]);
4215 #if defined(_SCTP_NEEDS_CALLOUT_) && !defined(__APPLE__)
4216 TAILQ_INIT(&sctppcbinfo.callqueue);
4217 #endif
4222 sctp_load_addresses_from_init(struct sctp_tcb *stcb, struct mbuf *m,
4223 int iphlen, int offset, int limit, struct sctphdr *sh,
4224 struct sockaddr *altsa)
4227 * grub through the INIT pulling addresses and
4228 * loading them to the nets structure in the asoc.
4229 * The from address in the mbuf should also be loaded
4230 * (if it is not already). This routine can be called
4231 * with either INIT or INIT-ACK's as long as the
4232 * m points to the IP packet and the offset points
4233 * to the beginning of the parameters.
4235 struct sctp_inpcb *inp, *l_inp;
4236 struct sctp_nets *net, *net_tmp;
4237 struct ip *iph;
4238 struct sctp_paramhdr *phdr, parm_buf;
4239 struct sctp_tcb *stcb_tmp;
4240 u_int16_t ptype, plen;
4241 struct sockaddr *sa;
4242 struct sockaddr_storage dest_store;
4243 struct sockaddr *local_sa = (struct sockaddr *)&dest_store;
4244 struct sockaddr_in sin;
4245 struct sockaddr_in6 sin6;
4247 /* First get the destination address setup too. */
4248 memset(&sin, 0, sizeof(sin));
4249 memset(&sin6, 0, sizeof(sin6));
4251 sin.sin_family = AF_INET;
4252 sin.sin_len = sizeof(sin);
4253 sin.sin_port = stcb->rport;
4255 sin6.sin6_family = AF_INET6;
4256 sin6.sin6_len = sizeof(struct sockaddr_in6);
4257 sin6.sin6_port = stcb->rport;
4258 if (altsa == NULL) {
4259 iph = mtod(m, struct ip *);
4260 if (iph->ip_v == IPVERSION) {
4261 /* its IPv4 */
4262 struct sockaddr_in *sin_2;
4263 sin_2 = (struct sockaddr_in *)(local_sa);
4264 memset(sin_2, 0, sizeof(sin));
4265 sin_2->sin_family = AF_INET;
4266 sin_2->sin_len = sizeof(sin);
4267 sin_2->sin_port = sh->dest_port;
4268 sin_2->sin_addr.s_addr = iph->ip_dst.s_addr ;
4269 sin.sin_addr = iph->ip_src;
4270 sa = (struct sockaddr *)&sin;
4271 } else if (iph->ip_v == (IPV6_VERSION >> 4)) {
4272 /* its IPv6 */
4273 struct ip6_hdr *ip6;
4274 struct sockaddr_in6 *sin6_2;
4276 ip6 = mtod(m, struct ip6_hdr *);
4277 sin6_2 = (struct sockaddr_in6 *)(local_sa);
4278 memset(sin6_2, 0, sizeof(sin6));
4279 sin6_2->sin6_family = AF_INET6;
4280 sin6_2->sin6_len = sizeof(struct sockaddr_in6);
4281 sin6_2->sin6_port = sh->dest_port;
4282 sin6.sin6_addr = ip6->ip6_src;
4283 sa = (struct sockaddr *)&sin6;
4284 } else {
4285 sa = NULL;
4287 } else {
4289 * For cookies we use the src address NOT from the packet
4290 * but from the original INIT
4292 sa = altsa;
4294 /* Turn off ECN until we get through all params */
4295 stcb->asoc.ecn_allowed = 0;
4297 TAILQ_FOREACH(net, &stcb->asoc.nets, sctp_next) {
4298 /* mark all addresses that we have currently on the list */
4299 net->dest_state |= SCTP_ADDR_NOT_IN_ASSOC;
4301 /* does the source address already exist? if so skip it */
4302 l_inp = inp = stcb->sctp_ep;
4303 stcb_tmp = sctp_findassociation_ep_addr(&inp, sa, &net_tmp, local_sa, stcb);
4304 if ((stcb_tmp == NULL && inp == stcb->sctp_ep) || inp == NULL) {
4305 /* we must add the source address */
4306 /* no scope set here since we have a tcb already. */
4307 if ((sa->sa_family == AF_INET) &&
4308 (stcb->asoc.ipv4_addr_legal)) {
4309 if (sctp_add_remote_addr(stcb, sa, 0, 2)) {
4310 return (-1);
4312 } else if ((sa->sa_family == AF_INET6) &&
4313 (stcb->asoc.ipv6_addr_legal)) {
4314 if (sctp_add_remote_addr(stcb, sa, 0, 3)) {
4315 return (-1);
4318 } else {
4319 if (net_tmp != NULL && stcb_tmp == stcb) {
4320 net_tmp->dest_state &= ~SCTP_ADDR_NOT_IN_ASSOC;
4321 } else if (stcb_tmp != stcb) {
4322 /* It belongs to another association? */
4323 return (-1);
4326 /* since a unlock occured we must check the
4327 * TCB's state and the pcb's gone flags.
4329 if (l_inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
4330 /* the user freed the ep */
4331 return (-1);
4333 if (stcb->asoc.state == 0) {
4334 /* the assoc was freed? */
4335 return (-1);
4338 /* now we must go through each of the params. */
4339 phdr = sctp_get_next_param(m, offset, &parm_buf, sizeof(parm_buf));
4340 while (phdr) {
4341 ptype = ntohs(phdr->param_type);
4342 plen = ntohs(phdr->param_length);
4343 /*kprintf("ptype => %d, plen => %d\n", ptype, plen);*/
4344 if (offset + plen > limit) {
4345 break;
4347 if (plen == 0) {
4348 break;
4350 if ((ptype == SCTP_IPV4_ADDRESS) &&
4351 (stcb->asoc.ipv4_addr_legal)) {
4352 struct sctp_ipv4addr_param *p4, p4_buf;
4353 /* ok get the v4 address and check/add */
4354 phdr = sctp_get_next_param(m, offset,
4355 (struct sctp_paramhdr *)&p4_buf, sizeof(p4_buf));
4356 if (plen != sizeof(struct sctp_ipv4addr_param) ||
4357 phdr == NULL) {
4358 return (-1);
4360 p4 = (struct sctp_ipv4addr_param *)phdr;
4361 sin.sin_addr.s_addr = p4->addr;
4362 sa = (struct sockaddr *)&sin;
4363 inp = stcb->sctp_ep;
4364 stcb_tmp = sctp_findassociation_ep_addr(&inp, sa, &net,
4365 local_sa, stcb);
4367 if ((stcb_tmp== NULL && inp == stcb->sctp_ep) ||
4368 inp == NULL) {
4369 /* we must add the source address */
4370 /* no scope set since we have a tcb already */
4372 /* we must validate the state again here */
4373 if (l_inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
4374 /* the user freed the ep */
4375 return (-1);
4377 if (stcb->asoc.state == 0) {
4378 /* the assoc was freed? */
4379 return (-1);
4381 if (sctp_add_remote_addr(stcb, sa, 0, 4)) {
4382 return (-1);
4384 } else if (stcb_tmp == stcb) {
4385 if (l_inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
4386 /* the user freed the ep */
4387 return (-1);
4389 if (stcb->asoc.state == 0) {
4390 /* the assoc was freed? */
4391 return (-1);
4393 if (net != NULL) {
4394 /* clear flag */
4395 net->dest_state &=
4396 ~SCTP_ADDR_NOT_IN_ASSOC;
4398 } else {
4399 /* strange, address is in another assoc?
4400 * straighten out locks.
4402 SCTP_TCB_UNLOCK(stcb_tmp);
4403 SCTP_INP_RLOCK(inp);
4404 if (l_inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
4405 /* the user freed the ep */
4406 SCTP_INP_RUNLOCK(l_inp);
4407 return (-1);
4409 if (stcb->asoc.state == 0) {
4410 /* the assoc was freed? */
4411 SCTP_INP_RUNLOCK(l_inp);
4412 return (-1);
4414 SCTP_TCB_LOCK(stcb);
4415 SCTP_INP_RUNLOCK(stcb->sctp_ep);
4416 return (-1);
4418 } else if ((ptype == SCTP_IPV6_ADDRESS) &&
4419 (stcb->asoc.ipv6_addr_legal)) {
4420 /* ok get the v6 address and check/add */
4421 struct sctp_ipv6addr_param *p6, p6_buf;
4422 phdr = sctp_get_next_param(m, offset,
4423 (struct sctp_paramhdr *)&p6_buf, sizeof(p6_buf));
4424 if (plen != sizeof(struct sctp_ipv6addr_param) ||
4425 phdr == NULL) {
4426 return (-1);
4428 p6 = (struct sctp_ipv6addr_param *)phdr;
4429 memcpy((caddr_t)&sin6.sin6_addr, p6->addr,
4430 sizeof(p6->addr));
4431 sa = (struct sockaddr *)&sin6;
4432 inp = stcb->sctp_ep;
4433 stcb_tmp= sctp_findassociation_ep_addr(&inp, sa, &net,
4434 local_sa, stcb);
4435 if (stcb_tmp == NULL && (inp == stcb->sctp_ep ||
4436 inp == NULL)) {
4437 /* we must validate the state again here */
4438 if (l_inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
4439 /* the user freed the ep */
4440 return (-1);
4442 if (stcb->asoc.state == 0) {
4443 /* the assoc was freed? */
4444 return (-1);
4446 /* we must add the address, no scope set */
4447 if (sctp_add_remote_addr(stcb, sa, 0, 5)) {
4448 return (-1);
4450 } else if (stcb_tmp == stcb) {
4451 /* we must validate the state again here */
4452 if (l_inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
4453 /* the user freed the ep */
4454 return (-1);
4456 if (stcb->asoc.state == 0) {
4457 /* the assoc was freed? */
4458 return (-1);
4460 if (net != NULL) {
4461 /* clear flag */
4462 net->dest_state &=
4463 ~SCTP_ADDR_NOT_IN_ASSOC;
4465 } else {
4466 /* strange, address is in another assoc?
4467 * straighten out locks.
4469 SCTP_TCB_UNLOCK(stcb_tmp);
4470 SCTP_INP_RLOCK(l_inp);
4471 /* we must validate the state again here */
4472 if (l_inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
4473 /* the user freed the ep */
4474 SCTP_INP_RUNLOCK(l_inp);
4475 return (-1);
4477 if (stcb->asoc.state == 0) {
4478 /* the assoc was freed? */
4479 SCTP_INP_RUNLOCK(l_inp);
4480 return (-1);
4482 SCTP_TCB_LOCK(stcb);
4483 SCTP_INP_RUNLOCK(l_inp);
4484 return (-1);
4486 } else if (ptype == SCTP_ECN_CAPABLE) {
4487 stcb->asoc.ecn_allowed = 1;
4488 } else if (ptype == SCTP_ULP_ADAPTION) {
4489 if (stcb->asoc.state != SCTP_STATE_OPEN) {
4490 struct sctp_adaption_layer_indication ai, *aip;
4492 phdr = sctp_get_next_param(m, offset,
4493 (struct sctp_paramhdr *)&ai, sizeof(ai));
4494 aip = (struct sctp_adaption_layer_indication *)phdr;
4495 sctp_ulp_notify(SCTP_NOTIFY_ADAPTION_INDICATION,
4496 stcb, ntohl(aip->indication), NULL);
4498 } else if (ptype == SCTP_SET_PRIM_ADDR) {
4499 struct sctp_asconf_addr_param lstore, *fee;
4500 struct sctp_asconf_addrv4_param *fii;
4501 int lptype;
4502 struct sockaddr *lsa = NULL;
4504 stcb->asoc.peer_supports_asconf = 1;
4505 stcb->asoc.peer_supports_asconf_setprim = 1;
4506 if (plen > sizeof(lstore)) {
4507 return (-1);
4509 phdr = sctp_get_next_param(m, offset,
4510 (struct sctp_paramhdr *)&lstore, plen);
4511 if (phdr == NULL) {
4512 return (-1);
4515 fee = (struct sctp_asconf_addr_param *)phdr;
4516 lptype = ntohs(fee->addrp.ph.param_type);
4517 if (lptype == SCTP_IPV4_ADDRESS) {
4518 if (plen !=
4519 sizeof(struct sctp_asconf_addrv4_param)) {
4520 kprintf("Sizeof setprim in init/init ack not %d but %d - ignored\n",
4521 (int)sizeof(struct sctp_asconf_addrv4_param),
4522 plen);
4523 } else {
4524 fii = (struct sctp_asconf_addrv4_param *)fee;
4525 sin.sin_addr.s_addr = fii->addrp.addr;
4526 lsa = (struct sockaddr *)&sin;
4528 } else if (lptype == SCTP_IPV6_ADDRESS) {
4529 if (plen !=
4530 sizeof(struct sctp_asconf_addr_param)) {
4531 kprintf("Sizeof setprim (v6) in init/init ack not %d but %d - ignored\n",
4532 (int)sizeof(struct sctp_asconf_addr_param),
4533 plen);
4534 } else {
4535 memcpy(sin6.sin6_addr.s6_addr,
4536 fee->addrp.addr,
4537 sizeof(fee->addrp.addr));
4538 lsa = (struct sockaddr *)&sin6;
4541 if (lsa) {
4542 sctp_set_primary_addr(stcb, sa, NULL);
4545 } else if (ptype == SCTP_PRSCTP_SUPPORTED) {
4546 /* Peer supports pr-sctp */
4547 stcb->asoc.peer_supports_prsctp = 1;
4548 } else if (ptype == SCTP_SUPPORTED_CHUNK_EXT) {
4549 /* A supported extension chunk */
4550 struct sctp_supported_chunk_types_param *pr_supported;
4551 uint8_t local_store[128];
4552 int num_ent, i;
4554 phdr = sctp_get_next_param(m, offset,
4555 (struct sctp_paramhdr *)&local_store, plen);
4556 if (phdr == NULL) {
4557 return (-1);
4559 stcb->asoc.peer_supports_asconf = 0;
4560 stcb->asoc.peer_supports_asconf_setprim = 0;
4561 stcb->asoc.peer_supports_prsctp = 0;
4562 stcb->asoc.peer_supports_pktdrop = 0;
4563 stcb->asoc.peer_supports_strreset = 0;
4564 pr_supported = (struct sctp_supported_chunk_types_param *)phdr;
4565 num_ent = plen - sizeof(struct sctp_paramhdr);
4566 for (i=0; i<num_ent; i++) {
4567 switch (pr_supported->chunk_types[i]) {
4568 case SCTP_ASCONF:
4569 stcb->asoc.peer_supports_asconf = 1;
4570 stcb->asoc.peer_supports_asconf_setprim = 1;
4571 break;
4572 case SCTP_ASCONF_ACK:
4573 stcb->asoc.peer_supports_asconf = 1;
4574 stcb->asoc.peer_supports_asconf_setprim = 1;
4575 break;
4576 case SCTP_FORWARD_CUM_TSN:
4577 stcb->asoc.peer_supports_prsctp = 1;
4578 break;
4579 case SCTP_PACKET_DROPPED:
4580 stcb->asoc.peer_supports_pktdrop = 1;
4581 break;
4582 case SCTP_STREAM_RESET:
4583 stcb->asoc.peer_supports_strreset = 1;
4584 break;
4585 default:
4586 /* one I have not learned yet */
4587 break;
4591 } else if (ptype == SCTP_ECN_NONCE_SUPPORTED) {
4592 /* Peer supports ECN-nonce */
4593 stcb->asoc.peer_supports_ecn_nonce = 1;
4594 stcb->asoc.ecn_nonce_allowed = 1;
4595 } else if ((ptype == SCTP_HEARTBEAT_INFO) ||
4596 (ptype == SCTP_STATE_COOKIE) ||
4597 (ptype == SCTP_UNRECOG_PARAM) ||
4598 (ptype == SCTP_COOKIE_PRESERVE) ||
4599 (ptype == SCTP_SUPPORTED_ADDRTYPE) ||
4600 (ptype == SCTP_ADD_IP_ADDRESS) ||
4601 (ptype == SCTP_DEL_IP_ADDRESS) ||
4602 (ptype == SCTP_ERROR_CAUSE_IND) ||
4603 (ptype == SCTP_SUCCESS_REPORT)) {
4604 /* don't care */;
4605 } else {
4606 if ((ptype & 0x8000) == 0x0000) {
4607 /* must stop processing the rest of
4608 * the param's. Any report bits were
4609 * handled with the call to sctp_arethere_unrecognized_parameters()
4610 * when the INIT or INIT-ACK was first seen.
4612 break;
4615 offset += SCTP_SIZE32(plen);
4616 if (offset >= limit) {
4617 break;
4619 phdr = sctp_get_next_param(m, offset, &parm_buf,
4620 sizeof(parm_buf));
4622 /* Now check to see if we need to purge any addresses */
4623 for (net = TAILQ_FIRST(&stcb->asoc.nets); net != NULL; net = net_tmp) {
4624 net_tmp = TAILQ_NEXT(net, sctp_next);
4625 if ((net->dest_state & SCTP_ADDR_NOT_IN_ASSOC) ==
4626 SCTP_ADDR_NOT_IN_ASSOC) {
4627 /* This address has been removed from the asoc */
4628 /* remove and free it */
4629 stcb->asoc.numnets--;
4630 TAILQ_REMOVE(&stcb->asoc.nets, net, sctp_next);
4631 sctp_free_remote_addr(net);
4632 if (net == stcb->asoc.primary_destination) {
4633 stcb->asoc.primary_destination = NULL;
4634 sctp_select_primary_destination(stcb);
4638 return (0);
4642 sctp_set_primary_addr(struct sctp_tcb *stcb, struct sockaddr *sa,
4643 struct sctp_nets *net)
4645 /* make sure the requested primary address exists in the assoc */
4646 if (net == NULL && sa)
4647 net = sctp_findnet(stcb, sa);
4649 if (net == NULL) {
4650 /* didn't find the requested primary address! */
4651 return (-1);
4652 } else {
4653 /* set the primary address */
4654 if (net->dest_state & SCTP_ADDR_UNCONFIRMED) {
4655 /* Must be confirmed */
4656 return (-1);
4658 stcb->asoc.primary_destination = net;
4659 net->dest_state &= ~SCTP_ADDR_WAS_PRIMARY;
4660 return (0);
4666 sctp_is_vtag_good(struct sctp_inpcb *inp, u_int32_t tag, struct timeval *now)
4669 * This function serves two purposes. It will see if a TAG can be
4670 * re-used and return 1 for yes it is ok and 0 for don't use that
4671 * tag.
4672 * A secondary function it will do is purge out old tags that can
4673 * be removed.
4675 struct sctpasochead *head;
4676 struct sctpvtaghead *chain;
4677 struct sctp_tagblock *twait_block;
4678 struct sctp_tcb *stcb;
4680 int i;
4681 SCTP_INP_INFO_WLOCK();
4682 chain = &sctppcbinfo.vtag_timewait[(tag % SCTP_STACK_VTAG_HASH_SIZE)];
4683 /* First is the vtag in use ? */
4685 head = &sctppcbinfo.sctp_asochash[SCTP_PCBHASH_ASOC(tag,
4686 sctppcbinfo.hashasocmark)];
4687 if (head == NULL) {
4688 SCTP_INP_INFO_WUNLOCK();
4689 return (0);
4691 LIST_FOREACH(stcb, head, sctp_asocs) {
4692 if (stcb->asoc.my_vtag == tag) {
4693 /* We should remove this if and
4694 * return 0 always if we want vtags
4695 * unique across all endpoints. For
4696 * now within a endpoint is ok.
4698 if (inp == stcb->sctp_ep) {
4699 /* bad tag, in use */
4700 SCTP_INP_INFO_WUNLOCK();
4701 return (0);
4705 if (!LIST_EMPTY(chain)) {
4707 * Block(s) are present, lets see if we have this tag in
4708 * the list
4710 LIST_FOREACH(twait_block, chain, sctp_nxt_tagblock) {
4711 for (i = 0; i < SCTP_NUMBER_IN_VTAG_BLOCK; i++) {
4712 if (twait_block->vtag_block[i].v_tag == 0) {
4713 /* not used */
4714 continue;
4715 } else if ((long)twait_block->vtag_block[i].tv_sec_at_expire >
4716 now->tv_sec) {
4717 /* Audit expires this guy */
4718 twait_block->vtag_block[i].tv_sec_at_expire = 0;
4719 twait_block->vtag_block[i].v_tag = 0;
4720 } else if (twait_block->vtag_block[i].v_tag ==
4721 tag) {
4722 /* Bad tag, sorry :< */
4723 SCTP_INP_INFO_WUNLOCK();
4724 return (0);
4729 /* Not found, ok to use the tag */
4730 SCTP_INP_INFO_WUNLOCK();
4731 return (1);
4736 * Delete the address from the endpoint local address list
4737 * Lookup using a sockaddr address (ie. not an ifaddr)
4740 sctp_del_local_addr_ep_sa(struct sctp_inpcb *inp, struct sockaddr *sa)
4742 struct sctp_laddr *laddr;
4743 struct sockaddr *l_sa;
4744 int found = 0;
4745 /* Here is another function I cannot find a
4746 * caller for. As such we SHOULD delete it
4747 * if we have no users. If we find a user that
4748 * user MUST have the INP locked.
4752 if (inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) {
4753 /* You are already bound to all. You have it already */
4754 return (EINVAL);
4757 LIST_FOREACH(laddr, &inp->sctp_addr_list, sctp_nxt_addr) {
4758 /* make sure the address exists */
4759 if (laddr->ifa == NULL)
4760 continue;
4761 if (laddr->ifa->ifa_addr == NULL)
4762 continue;
4764 l_sa = laddr->ifa->ifa_addr;
4765 if (l_sa->sa_family == AF_INET6) {
4766 /* IPv6 address */
4767 struct sockaddr_in6 *sin1, *sin2;
4768 sin1 = (struct sockaddr_in6 *)l_sa;
4769 sin2 = (struct sockaddr_in6 *)sa;
4770 if (memcmp(&sin1->sin6_addr, &sin2->sin6_addr,
4771 sizeof(struct in6_addr)) == 0) {
4772 /* matched */
4773 found = 1;
4774 break;
4776 } else if (l_sa->sa_family == AF_INET) {
4777 /* IPv4 address */
4778 struct sockaddr_in *sin1, *sin2;
4779 sin1 = (struct sockaddr_in *)l_sa;
4780 sin2 = (struct sockaddr_in *)sa;
4781 if (sin1->sin_addr.s_addr == sin2->sin_addr.s_addr) {
4782 /* matched */
4783 found = 1;
4784 break;
4786 } else {
4787 /* invalid family */
4788 return (-1);
4792 if (found && inp->laddr_count < 2) {
4793 /* can't delete unless there are at LEAST 2 addresses */
4794 return (-1);
4797 if (found && (inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) == 0) {
4799 * remove it from the ep list, this should NOT be
4800 * done until its really gone from the interface list and
4801 * we won't be receiving more of these. Probably right
4802 * away. If we do allow a removal of an address from
4803 * an association (sub-set bind) than this should NOT
4804 * be called until the all ASCONF come back from this
4805 * association.
4807 sctp_remove_laddr(laddr);
4808 return (0);
4809 } else {
4810 return (-1);
4814 static void
4815 sctp_drain_mbufs(struct sctp_inpcb *inp, struct sctp_tcb *stcb)
4818 * We must hunt this association for MBUF's past the cumack
4819 * (i.e. out of order data that we can renege on).
4821 struct sctp_association *asoc;
4822 struct sctp_tmit_chunk *chk, *nchk;
4823 u_int32_t cumulative_tsn_p1, tsn;
4824 int cnt, strmat, gap;
4825 /* We look for anything larger than the cum-ack + 1 */
4827 asoc = &stcb->asoc;
4828 cumulative_tsn_p1 = asoc->cumulative_tsn + 1;
4829 cnt = 0;
4830 /* First look in the re-assembly queue */
4831 chk = TAILQ_FIRST(&asoc->reasmqueue);
4832 while (chk) {
4833 /* Get the next one */
4834 nchk = TAILQ_NEXT(chk, sctp_next);
4835 if (compare_with_wrap(chk->rec.data.TSN_seq,
4836 cumulative_tsn_p1, MAX_TSN)) {
4837 /* Yep it is above cum-ack */
4838 cnt++;
4839 tsn = chk->rec.data.TSN_seq;
4840 if (tsn >= asoc->mapping_array_base_tsn) {
4841 gap = tsn - asoc->mapping_array_base_tsn;
4842 } else {
4843 gap = (MAX_TSN - asoc->mapping_array_base_tsn) +
4844 tsn + 1;
4846 asoc->size_on_reasm_queue -= chk->send_size;
4847 asoc->cnt_on_reasm_queue--;
4848 SCTP_UNSET_TSN_PRESENT(asoc->mapping_array, gap);
4849 TAILQ_REMOVE(&asoc->reasmqueue, chk, sctp_next);
4850 if (chk->data) {
4851 sctp_m_freem(chk->data);
4852 chk->data = NULL;
4854 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
4855 sctppcbinfo.ipi_count_chunk--;
4856 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
4857 panic("Chunk count is negative");
4859 sctppcbinfo.ipi_gencnt_chunk++;
4861 chk = nchk;
4863 /* Ok that was fun, now we will drain all the inbound streams? */
4864 for (strmat = 0; strmat < asoc->streamincnt; strmat++) {
4865 chk = TAILQ_FIRST(&asoc->strmin[strmat].inqueue);
4866 while (chk) {
4867 nchk = TAILQ_NEXT(chk, sctp_next);
4868 if (compare_with_wrap(chk->rec.data.TSN_seq,
4869 cumulative_tsn_p1, MAX_TSN)) {
4870 /* Yep it is above cum-ack */
4871 cnt++;
4872 tsn = chk->rec.data.TSN_seq;
4873 if (tsn >= asoc->mapping_array_base_tsn) {
4874 gap = tsn -
4875 asoc->mapping_array_base_tsn;
4876 } else {
4877 gap = (MAX_TSN -
4878 asoc->mapping_array_base_tsn) +
4879 tsn + 1;
4881 asoc->size_on_all_streams -= chk->send_size;
4882 asoc->cnt_on_all_streams--;
4884 SCTP_UNSET_TSN_PRESENT(asoc->mapping_array,
4885 gap);
4886 TAILQ_REMOVE(&asoc->strmin[strmat].inqueue,
4887 chk, sctp_next);
4888 if (chk->data) {
4889 sctp_m_freem(chk->data);
4890 chk->data = NULL;
4892 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
4893 sctppcbinfo.ipi_count_chunk--;
4894 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
4895 panic("Chunk count is negative");
4897 sctppcbinfo.ipi_gencnt_chunk++;
4899 chk = nchk;
4903 * Question, should we go through the delivery queue?
4904 * The only reason things are on here is the app not reading OR a
4905 * p-d-api up. An attacker COULD send enough in to initiate the
4906 * PD-API and then send a bunch of stuff to other streams... these
4907 * would wind up on the delivery queue.. and then we would not get
4908 * to them. But in order to do this I then have to back-track and
4909 * un-deliver sequence numbers in streams.. el-yucko. I think for
4910 * now we will NOT look at the delivery queue and leave it to be
4911 * something to consider later. An alternative would be to abort
4912 * the P-D-API with a notification and then deliver the data....
4913 * Or another method might be to keep track of how many times the
4914 * situation occurs and if we see a possible attack underway just
4915 * abort the association.
4917 #ifdef SCTP_DEBUG
4918 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
4919 if (cnt) {
4920 kprintf("Freed %d chunks from reneg harvest\n", cnt);
4923 #endif /* SCTP_DEBUG */
4926 * Another issue, in un-setting the TSN's in the mapping array we
4927 * DID NOT adjust the higest_tsn marker. This will cause one of
4928 * two things to occur. It may cause us to do extra work in checking
4929 * for our mapping array movement. More importantly it may cause us
4930 * to SACK every datagram. This may not be a bad thing though since
4931 * we will recover once we get our cum-ack above and all this stuff
4932 * we dumped recovered.
4936 void
4937 sctp_drain(void)
4940 * We must walk the PCB lists for ALL associations here. The system
4941 * is LOW on MBUF's and needs help. This is where reneging will
4942 * occur. We really hope this does NOT happen!
4944 struct sctp_inpcb *inp;
4945 struct sctp_tcb *stcb;
4947 SCTP_INP_INFO_RLOCK();
4948 LIST_FOREACH(inp, &sctppcbinfo.listhead, sctp_list) {
4949 /* For each endpoint */
4950 SCTP_INP_RLOCK(inp);
4951 LIST_FOREACH(stcb, &inp->sctp_asoc_list, sctp_tcblist) {
4952 /* For each association */
4953 SCTP_TCB_LOCK(stcb);
4954 sctp_drain_mbufs(inp, stcb);
4955 SCTP_TCB_UNLOCK(stcb);
4957 SCTP_INP_RUNLOCK(inp);
4959 SCTP_INP_INFO_RUNLOCK();
4963 sctp_add_to_socket_q(struct sctp_inpcb *inp, struct sctp_tcb *stcb)
4965 struct sctp_socket_q_list *sq;
4967 /* write lock on INP assumed */
4968 if ((inp == NULL) || (stcb == NULL)) {
4969 /* I am paranoid */
4970 return (0);
4972 sq = (struct sctp_socket_q_list *)SCTP_ZONE_GET(
4973 sctppcbinfo.ipi_zone_sockq);
4974 if (sq == NULL) {
4975 /* out of sq structs */
4976 return (0);
4978 sctppcbinfo.ipi_count_sockq++;
4979 sctppcbinfo.ipi_gencnt_sockq++;
4980 if (stcb)
4981 stcb->asoc.cnt_msg_on_sb++;
4982 sq->tcb = stcb;
4983 TAILQ_INSERT_TAIL(&inp->sctp_queue_list, sq, next_sq);
4984 return (1);
4988 struct sctp_tcb *
4989 sctp_remove_from_socket_q(struct sctp_inpcb *inp)
4991 struct sctp_tcb *stcb = NULL;
4992 struct sctp_socket_q_list *sq;
4994 /* W-Lock on INP assumed held */
4995 sq = TAILQ_FIRST(&inp->sctp_queue_list);
4996 if (sq == NULL)
4997 return (NULL);
4999 stcb = sq->tcb;
5000 TAILQ_REMOVE(&inp->sctp_queue_list, sq, next_sq);
5001 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_sockq, sq);
5002 sctppcbinfo.ipi_count_sockq--;
5003 sctppcbinfo.ipi_gencnt_sockq++;
5004 if (stcb) {
5005 stcb->asoc.cnt_msg_on_sb--;
5007 return (stcb);
5011 sctp_initiate_iterator(asoc_func af, uint32_t pcb_state, uint32_t asoc_state,
5012 void *argp, uint32_t argi, end_func ef,
5013 struct sctp_inpcb *s_inp)
5015 struct sctp_iterator *it=NULL;
5017 if (af == NULL) {
5018 return (-1);
5020 it = kmalloc(sizeof(struct sctp_iterator), M_PCB, M_WAITOK);
5021 memset(it, 0, sizeof(*it));
5022 it->function_toapply = af;
5023 it->function_atend = ef;
5024 it->pointer = argp;
5025 it->val = argi;
5026 it->pcb_flags = pcb_state;
5027 it->asoc_state = asoc_state;
5028 if (s_inp) {
5029 it->inp = s_inp;
5030 it->iterator_flags = SCTP_ITERATOR_DO_SINGLE_INP;
5031 } else {
5032 SCTP_INP_INFO_RLOCK();
5033 it->inp = LIST_FIRST(&sctppcbinfo.listhead);
5034 SCTP_INP_INFO_RUNLOCK();
5035 it->iterator_flags = SCTP_ITERATOR_DO_ALL_INP;
5038 /* Init the timer */
5039 #if defined(__FreeBSD__) && __FreeBSD_version >= 500000
5040 callout_init(&it->tmr.timer, 0);
5041 #else
5042 callout_init(&it->tmr.timer);
5043 #endif
5044 /* add to the list of all iterators */
5045 SCTP_INP_INFO_WLOCK();
5046 LIST_INSERT_HEAD(&sctppcbinfo.iteratorhead, it, sctp_nxt_itr);
5047 SCTP_INP_INFO_WUNLOCK();
5048 sctp_iterator_timer(it);
5049 return (0);
5054 * Callout/Timer routines for OS that doesn't have them
5056 #ifdef _SCTP_NEEDS_CALLOUT_
5057 #ifndef __APPLE__
5058 extern int ticks;
5059 #endif
5061 void
5062 callout_init(struct callout *c)
5064 bzero(c, sizeof(*c));
5067 void
5068 callout_reset(struct callout *c, int to_ticks, void (*ftn)(void *), void *arg)
5070 if (c->c_flags & CALLOUT_PENDING)
5071 callout_stop(c);
5074 * We could spl down here and back up at the TAILQ_INSERT_TAIL,
5075 * but there's no point since doing this setup doesn't take much
5076 * time.
5078 if (to_ticks <= 0)
5079 to_ticks = 1;
5081 c->c_arg = arg;
5082 c->c_flags = (CALLOUT_ACTIVE | CALLOUT_PENDING);
5083 c->c_func = ftn;
5084 #ifdef __APPLE__
5085 c->c_time = to_ticks; /* just store the requested timeout */
5086 timeout(ftn, arg, to_ticks);
5087 #else
5088 c->c_time = ticks + to_ticks;
5089 TAILQ_INSERT_TAIL(&sctppcbinfo.callqueue, c, tqe);
5090 #endif
5094 callout_stop(struct callout *c)
5097 * Don't attempt to delete a callout that's not on the queue.
5099 if (!(c->c_flags & CALLOUT_PENDING)) {
5100 c->c_flags &= ~CALLOUT_ACTIVE;
5101 return (0);
5103 c->c_flags &= ~(CALLOUT_ACTIVE | CALLOUT_PENDING| CALLOUT_FIRED);
5104 #ifdef __APPLE__
5105 /* thread_call_cancel(c->c_call); */
5106 untimeout(c->c_func, c->c_arg);
5107 #else
5108 TAILQ_REMOVE(&sctppcbinfo.callqueue, c, tqe);
5109 c->c_func = NULL;
5110 #endif
5111 return (1);
5114 #if !defined(__APPLE__)
5115 void
5116 sctp_fasttim(void)
5118 struct callout *c, *n;
5119 struct calloutlist locallist;
5120 int inited = 0;
5122 /* run through and subtract and mark all callouts */
5123 c = TAILQ_FIRST(&sctppcbinfo.callqueue);
5124 while (c) {
5125 n = TAILQ_NEXT(c, tqe);
5126 if (c->c_time <= ticks) {
5127 c->c_flags |= CALLOUT_FIRED;
5128 c->c_time = 0;
5129 TAILQ_REMOVE(&sctppcbinfo.callqueue, c, tqe);
5130 if (inited == 0) {
5131 TAILQ_INIT(&locallist);
5132 inited = 1;
5134 /* move off of main list */
5135 TAILQ_INSERT_TAIL(&locallist, c, tqe);
5137 c = n;
5139 /* Now all the ones on the locallist must be called */
5140 if (inited) {
5141 c = TAILQ_FIRST(&locallist);
5142 while (c) {
5143 /* remove it */
5144 TAILQ_REMOVE(&locallist, c, tqe);
5145 /* now validate that it did not get canceled */
5146 if (c->c_flags & CALLOUT_FIRED) {
5147 c->c_flags &= ~CALLOUT_PENDING;
5148 (*c->c_func)(c->c_arg);
5150 c = TAILQ_FIRST(&locallist);
5154 #endif
5155 #endif /* _SCTP_NEEDS_CALLOUT_ */