ipflow: Unroll the first iteration of the hash generation loop.
[dragonfly.git] / sys / netinet / ip_flow.c
blob3286b9f0758880c62d26273bb179ad3bb1c68b58
1 /*-
2 * Copyright (c) 1998 The NetBSD Foundation, Inc.
3 * All rights reserved.
5 * This code is derived from software contributed to The NetBSD Foundation
6 * by the 3am Software Foundry ("3am"). It was developed by Matt Thomas.
8 * Redistribution and use in source and binary forms, with or without
9 * modification, are permitted provided that the following conditions
10 * are met:
11 * 1. Redistributions of source code must retain the above copyright
12 * notice, this list of conditions and the following disclaimer.
13 * 2. Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in the
15 * documentation and/or other materials provided with the distribution.
16 * 3. All advertising materials mentioning features or use of this software
17 * must display the following acknowledgement:
18 * This product includes software developed by the NetBSD
19 * Foundation, Inc. and its contributors.
20 * 4. Neither the name of The NetBSD Foundation nor the names of its
21 * contributors may be used to endorse or promote products derived
22 * from this software without specific prior written permission.
24 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
25 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
26 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
27 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
28 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
29 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
30 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
31 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
32 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
33 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
34 * POSSIBILITY OF SUCH DAMAGE.
36 * $FreeBSD: src/sys/netinet/ip_flow.c,v 1.9.2.2 2001/11/04 17:35:31 luigi Exp $
39 #include <sys/param.h>
40 #include <sys/kernel.h>
41 #include <sys/malloc.h>
42 #include <sys/mbuf.h>
43 #include <sys/protosw.h>
44 #include <sys/socket.h>
45 #include <sys/sysctl.h>
46 #include <sys/thread2.h>
47 #include <sys/in_cksum.h>
49 #include <machine/smp.h>
51 #include <net/if.h>
52 #include <net/if_var.h>
53 #include <net/route.h>
54 #include <net/netisr2.h>
55 #include <net/netmsg2.h>
57 #include <netinet/in.h>
58 #include <netinet/ip.h>
59 #include <netinet/in_var.h>
60 #include <netinet/ip_var.h>
61 #include <netinet/ip_flow.h>
63 #define IPFLOW_TIMER (5 * PR_SLOWHZ)
64 #define IPFLOW_HASHBITS 6 /* should not be a multiple of 8 */
65 #define IPFLOW_HASHSIZE (1 << IPFLOW_HASHBITS)
66 #define IPFLOW_MAX 256
68 #define IPFLOW_RTENTRY_ISDOWN(rt) \
69 (((rt)->rt_flags & RTF_UP) == 0 || \
70 ((rt)->rt_ifp->if_flags & IFF_UP) == 0)
72 struct netmsg_ipfaddr {
73 struct netmsg_base base;
74 struct in_addr ipf_addr;
77 struct ipflow {
78 LIST_ENTRY(ipflow) ipf_hash; /* next ipflow in hash bucket */
79 LIST_ENTRY(ipflow) ipf_list; /* next ipflow in list */
81 struct in_addr ipf_dst; /* destination address */
82 struct in_addr ipf_src; /* source address */
83 uint8_t ipf_tos; /* type-of-service */
85 uint8_t ipf_flags; /* see IPFLOW_FLAG_ */
86 uint8_t ipf_pad[2]; /* explicit pad */
87 int ipf_refcnt; /* reference count */
89 struct route ipf_ro; /* associated route entry */
90 u_long ipf_uses; /* number of uses in this period */
92 int ipf_timer; /* remaining lifetime of this entry */
93 u_long ipf_dropped; /* ENOBUFS returned by if_output */
94 u_long ipf_errors; /* other errors returned by if_output */
95 u_long ipf_last_uses; /* number of uses in last period */
97 LIST_HEAD(ipflowhead, ipflow);
99 #define IPFLOW_FLAG_ONLIST 0x1
101 #define ipflow_inuse ipflow_inuse_pcpu[mycpuid]
102 #define ipflowtable ipflowtable_pcpu[mycpuid]
103 #define ipflowlist ipflowlist_pcpu[mycpuid]
105 static struct ipflowhead ipflowtable_pcpu[MAXCPU][IPFLOW_HASHSIZE];
106 static struct ipflowhead ipflowlist_pcpu[MAXCPU];
107 static int ipflow_inuse_pcpu[MAXCPU];
108 static struct netmsg_base ipflow_timo_netmsgs[MAXCPU];
109 static int ipflow_active = 0;
111 #define IPFLOW_REFCNT_INIT 1
113 /* ipflow is alive and active */
114 #define IPFLOW_IS_ACTIVE(ipf) ((ipf)->ipf_refcnt > IPFLOW_REFCNT_INIT)
115 /* ipflow is alive but not active */
116 #define IPFLOW_NOT_ACTIVE(ipf) ((ipf)->ipf_refcnt == IPFLOW_REFCNT_INIT)
118 #define IPFLOW_REF(ipf) \
119 do { \
120 KKASSERT((ipf)->ipf_refcnt > 0); \
121 (ipf)->ipf_refcnt++; \
122 } while (0)
124 #define IPFLOW_FREE(ipf) \
125 do { \
126 KKASSERT((ipf)->ipf_refcnt > 0); \
127 (ipf)->ipf_refcnt--; \
128 if ((ipf)->ipf_refcnt == 0) \
129 ipflow_free((ipf)); \
130 } while (0)
132 #define IPFLOW_INSERT(bucket, ipf) \
133 do { \
134 KKASSERT(((ipf)->ipf_flags & IPFLOW_FLAG_ONLIST) == 0); \
135 (ipf)->ipf_flags |= IPFLOW_FLAG_ONLIST; \
136 LIST_INSERT_HEAD((bucket), (ipf), ipf_hash); \
137 LIST_INSERT_HEAD(&ipflowlist, (ipf), ipf_list); \
138 } while (0)
140 #define IPFLOW_REMOVE(ipf) \
141 do { \
142 KKASSERT((ipf)->ipf_flags & IPFLOW_FLAG_ONLIST); \
143 (ipf)->ipf_flags &= ~IPFLOW_FLAG_ONLIST; \
144 LIST_REMOVE((ipf), ipf_hash); \
145 LIST_REMOVE((ipf), ipf_list); \
146 } while (0)
148 SYSCTL_NODE(_net_inet_ip, OID_AUTO, ipflow, CTLFLAG_RW, 0, "ip flow");
149 SYSCTL_INT(_net_inet_ip, IPCTL_FASTFORWARDING, fastforwarding, CTLFLAG_RW,
150 &ipflow_active, 0, "Enable flow-based IP forwarding");
152 static MALLOC_DEFINE(M_IPFLOW, "ip_flow", "IP flow");
154 static void ipflow_free(struct ipflow *);
156 static unsigned
157 ipflow_hash(struct in_addr dst, struct in_addr src, unsigned tos)
159 unsigned hash = tos + src.s_addr;
160 int idx;
162 for (idx = IPFLOW_HASHBITS; idx < 32; idx += IPFLOW_HASHBITS)
163 hash += (dst.s_addr >> (32 - idx)) + (src.s_addr >> idx);
164 return hash & (IPFLOW_HASHSIZE-1);
167 static struct ipflow *
168 ipflow_lookup(const struct ip *ip)
170 unsigned hash;
171 struct ipflow *ipf;
173 hash = ipflow_hash(ip->ip_dst, ip->ip_src, ip->ip_tos);
174 LIST_FOREACH(ipf, &ipflowtable[hash], ipf_hash) {
175 if (ip->ip_dst.s_addr == ipf->ipf_dst.s_addr &&
176 ip->ip_src.s_addr == ipf->ipf_src.s_addr &&
177 ip->ip_tos == ipf->ipf_tos)
178 break;
180 return ipf;
184 ipflow_fastforward(struct mbuf *m)
186 struct ip *ip;
187 struct ipflow *ipf;
188 struct rtentry *rt;
189 struct sockaddr *dst;
190 struct ifnet *ifp;
191 int error, iplen;
194 * Are we forwarding packets?
196 if (!ipforwarding || !ipflow_active)
197 return 0;
200 * Was packet received as a link-level multicast or broadcast?
201 * If so, don't try to fast forward..
203 if (m->m_flags & (M_BCAST | M_MCAST))
204 return 0;
206 /* length checks already done in ip_hashfn() */
207 KASSERT(m->m_len >= sizeof(struct ip), ("IP header not in one mbuf"));
208 ip = mtod(m, struct ip *);
211 * IP header with no option and valid version
213 if (ip->ip_v != IPVERSION || ip->ip_hl != (sizeof(struct ip) >> 2))
214 return 0;
216 iplen = ntohs(ip->ip_len);
217 /* length checks already done in ip_hashfn() */
218 KASSERT(iplen >= sizeof(struct ip),
219 ("total length less then header length"));
220 KASSERT(m->m_pkthdr.len >= iplen, ("mbuf too short"));
223 * Find a flow.
225 ipf = ipflow_lookup(ip);
226 if (ipf == NULL)
227 return 0;
230 * Verify the IP header checksum.
232 if (m->m_pkthdr.csum_flags & CSUM_IP_CHECKED) {
233 if (!(m->m_pkthdr.csum_flags & CSUM_IP_VALID))
234 return 0;
235 } else {
236 /* Must compute it ourselves. */
237 if (in_cksum_hdr(ip) != 0)
238 return 0;
242 * Route and interface still up?
244 rt = ipf->ipf_ro.ro_rt;
245 if (IPFLOW_RTENTRY_ISDOWN(rt))
246 return 0;
247 ifp = rt->rt_ifp;
250 * Packet size OK? TTL?
252 if (m->m_pkthdr.len > ifp->if_mtu || ip->ip_ttl <= IPTTLDEC)
253 return 0;
256 * Clear any in-bound checksum flags for this packet.
258 m->m_pkthdr.csum_flags = 0;
261 * Everything checks out and so we can forward this packet.
262 * Modify the TTL and incrementally change the checksum.
264 * This method of adding the checksum works on either endian CPU.
265 * If htons() is inlined, all the arithmetic is folded; otherwise
266 * the htons()s are combined by CSE due to the __const__ attribute.
268 * Don't bother using HW checksumming here -- the incremental
269 * update is pretty fast.
271 ip->ip_ttl -= IPTTLDEC;
272 if (ip->ip_sum >= (uint16_t)~htons(IPTTLDEC << 8))
273 ip->ip_sum -= ~htons(IPTTLDEC << 8);
274 else
275 ip->ip_sum += htons(IPTTLDEC << 8);
278 * Trim the packet in case it's too long..
280 if (m->m_pkthdr.len > iplen) {
281 if (m->m_len == m->m_pkthdr.len) {
282 m->m_len = iplen;
283 m->m_pkthdr.len = iplen;
284 } else {
285 m_adj(m, iplen - m->m_pkthdr.len);
290 * Send the packet on its way. All we can get back is ENOBUFS
292 ipf->ipf_uses++;
293 ipf->ipf_timer = IPFLOW_TIMER;
295 if (rt->rt_flags & RTF_GATEWAY)
296 dst = rt->rt_gateway;
297 else
298 dst = &ipf->ipf_ro.ro_dst;
301 * Reference count this ipflow, before the possible blocking
302 * ifnet.if_output(), so this ipflow will not be changed or
303 * reaped behind our back.
305 IPFLOW_REF(ipf);
307 error = ifp->if_output(ifp, m, dst, rt);
308 if (error) {
309 if (error == ENOBUFS)
310 ipf->ipf_dropped++;
311 else
312 ipf->ipf_errors++;
315 IPFLOW_FREE(ipf);
316 return 1;
319 static void
320 ipflow_addstats(struct ipflow *ipf)
322 ipf->ipf_ro.ro_rt->rt_use += ipf->ipf_uses;
323 ipstat.ips_cantforward += ipf->ipf_errors + ipf->ipf_dropped;
324 ipstat.ips_total += ipf->ipf_uses;
325 ipstat.ips_forward += ipf->ipf_uses;
326 ipstat.ips_fastforward += ipf->ipf_uses;
329 static void
330 ipflow_free(struct ipflow *ipf)
332 KKASSERT(ipf->ipf_refcnt == 0);
333 KKASSERT((ipf->ipf_flags & IPFLOW_FLAG_ONLIST) == 0);
335 KKASSERT(ipflow_inuse > 0);
336 ipflow_inuse--;
338 ipflow_addstats(ipf);
339 RTFREE(ipf->ipf_ro.ro_rt);
340 kfree(ipf, M_IPFLOW);
343 static void
344 ipflow_reset(struct ipflow *ipf)
346 ipflow_addstats(ipf);
347 RTFREE(ipf->ipf_ro.ro_rt);
348 ipf->ipf_uses = ipf->ipf_last_uses = 0;
349 ipf->ipf_errors = ipf->ipf_dropped = 0;
352 static struct ipflow *
353 ipflow_reap(void)
355 struct ipflow *ipf, *maybe_ipf = NULL;
357 LIST_FOREACH(ipf, &ipflowlist, ipf_list) {
359 * Skip actively used ipflow
361 if (IPFLOW_IS_ACTIVE(ipf))
362 continue;
365 * If this no longer points to a valid route
366 * reclaim it.
368 if ((ipf->ipf_ro.ro_rt->rt_flags & RTF_UP) == 0)
369 goto done;
372 * choose the one that's been least recently used
373 * or has had the least uses in the last 1.5
374 * intervals.
376 if (maybe_ipf == NULL ||
377 ipf->ipf_timer < maybe_ipf->ipf_timer ||
378 (ipf->ipf_timer == maybe_ipf->ipf_timer &&
379 ipf->ipf_last_uses + ipf->ipf_uses <
380 maybe_ipf->ipf_last_uses + maybe_ipf->ipf_uses))
381 maybe_ipf = ipf;
383 if (maybe_ipf == NULL)
384 return NULL;
386 ipf = maybe_ipf;
387 done:
389 * Remove the entry from the flow table and reset its states
391 IPFLOW_REMOVE(ipf);
392 ipflow_reset(ipf);
393 return ipf;
396 static void
397 ipflow_timo_dispatch(netmsg_t nmsg)
399 struct ipflow *ipf, *next_ipf;
401 crit_enter();
402 lwkt_replymsg(&nmsg->lmsg, 0); /* reply ASAP */
403 crit_exit();
405 LIST_FOREACH_MUTABLE(ipf, &ipflowlist, ipf_list, next_ipf) {
406 if (--ipf->ipf_timer == 0) {
407 IPFLOW_REMOVE(ipf);
408 IPFLOW_FREE(ipf);
409 } else {
410 ipf->ipf_last_uses = ipf->ipf_uses;
411 ipf->ipf_ro.ro_rt->rt_use += ipf->ipf_uses;
412 ipstat.ips_total += ipf->ipf_uses;
413 ipstat.ips_forward += ipf->ipf_uses;
414 ipstat.ips_fastforward += ipf->ipf_uses;
415 ipf->ipf_uses = 0;
420 static void
421 ipflow_timo_ipi(void *arg __unused)
423 struct lwkt_msg *msg = &ipflow_timo_netmsgs[mycpuid].lmsg;
425 crit_enter();
426 if (msg->ms_flags & MSGF_DONE)
427 lwkt_sendmsg_oncpu(netisr_cpuport(mycpuid), msg);
428 crit_exit();
431 void
432 ipflow_slowtimo(void)
434 cpumask_t mask;
435 int i;
437 CPUMASK_ASSZERO(mask);
438 for (i = 0; i < ncpus; ++i) {
439 if (ipflow_inuse_pcpu[i])
440 CPUMASK_ORBIT(mask, i);
442 CPUMASK_ANDMASK(mask, smp_active_mask);
443 if (CPUMASK_TESTNZERO(mask))
444 lwkt_send_ipiq_mask(mask, ipflow_timo_ipi, NULL);
447 void
448 ipflow_create(const struct route *ro, struct mbuf *m)
450 const struct ip *const ip = mtod(m, struct ip *);
451 struct ipflow *ipf;
452 unsigned hash;
455 * Don't create cache entries for ICMP messages.
457 if (!ipflow_active || ip->ip_p == IPPROTO_ICMP)
458 return;
461 * See if an existing flow struct exists. If so remove it from it's
462 * list and free the old route. If not, try to malloc a new one
463 * (if we aren't at our limit).
465 ipf = ipflow_lookup(ip);
466 if (ipf == NULL) {
467 if (ipflow_inuse == IPFLOW_MAX) {
468 ipf = ipflow_reap();
469 if (ipf == NULL)
470 return;
471 } else {
472 ipf = kmalloc(sizeof(*ipf), M_IPFLOW,
473 M_NOWAIT | M_ZERO);
474 if (ipf == NULL)
475 return;
476 ipf->ipf_refcnt = IPFLOW_REFCNT_INIT;
478 ipflow_inuse++;
480 } else {
481 if (IPFLOW_NOT_ACTIVE(ipf)) {
482 IPFLOW_REMOVE(ipf);
483 ipflow_reset(ipf);
484 } else {
485 /* This ipflow is being used; don't change it */
486 KKASSERT(IPFLOW_IS_ACTIVE(ipf));
487 return;
490 /* This ipflow should not be actively used */
491 KKASSERT(IPFLOW_NOT_ACTIVE(ipf));
494 * Fill in the updated information.
496 ipf->ipf_ro = *ro;
497 ro->ro_rt->rt_refcnt++;
498 ipf->ipf_dst = ip->ip_dst;
499 ipf->ipf_src = ip->ip_src;
500 ipf->ipf_tos = ip->ip_tos;
501 ipf->ipf_timer = IPFLOW_TIMER;
504 * Insert into the approriate bucket of the flow table.
506 hash = ipflow_hash(ip->ip_dst, ip->ip_src, ip->ip_tos);
507 IPFLOW_INSERT(&ipflowtable[hash], ipf);
510 void
511 ipflow_flush_oncpu(void)
513 struct ipflow *ipf;
515 while ((ipf = LIST_FIRST(&ipflowlist)) != NULL) {
516 IPFLOW_REMOVE(ipf);
517 IPFLOW_FREE(ipf);
521 static void
522 ipflow_ifaddr_handler(netmsg_t nmsg)
524 struct netmsg_ipfaddr *amsg = (struct netmsg_ipfaddr *)nmsg;
525 struct ipflow *ipf, *next_ipf;
527 LIST_FOREACH_MUTABLE(ipf, &ipflowlist, ipf_list, next_ipf) {
528 if (ipf->ipf_dst.s_addr == amsg->ipf_addr.s_addr ||
529 ipf->ipf_src.s_addr == amsg->ipf_addr.s_addr) {
530 IPFLOW_REMOVE(ipf);
531 IPFLOW_FREE(ipf);
534 netisr_forwardmsg(&nmsg->base, mycpuid + 1);
537 static void
538 ipflow_ifaddr(void *arg __unused, struct ifnet *ifp __unused,
539 enum ifaddr_event event, struct ifaddr *ifa)
541 struct netmsg_ipfaddr amsg;
543 if (ifa->ifa_addr->sa_family != AF_INET)
544 return;
546 /* Only add/change events need to be handled */
547 switch (event) {
548 case IFADDR_EVENT_ADD:
549 case IFADDR_EVENT_CHANGE:
550 break;
552 case IFADDR_EVENT_DELETE:
553 return;
556 netmsg_init(&amsg.base, NULL, &curthread->td_msgport,
557 MSGF_PRIORITY, ipflow_ifaddr_handler);
558 amsg.ipf_addr = ifatoia(ifa)->ia_addr.sin_addr;
560 netisr_domsg(&amsg.base, 0);
563 static void
564 ipflow_init(void)
566 char oid_name[32];
567 int i;
569 for (i = 0; i < ncpus; ++i) {
570 netmsg_init(&ipflow_timo_netmsgs[i], NULL, &netisr_adone_rport,
571 MSGF_PRIORITY, ipflow_timo_dispatch);
573 ksnprintf(oid_name, sizeof(oid_name), "inuse%d", i);
575 SYSCTL_ADD_INT(NULL,
576 SYSCTL_STATIC_CHILDREN(_net_inet_ip_ipflow),
577 OID_AUTO, oid_name, CTLFLAG_RD, &ipflow_inuse_pcpu[i], 0,
578 "# of ip flow being used");
580 EVENTHANDLER_REGISTER(ifaddr_event, ipflow_ifaddr, NULL,
581 EVENTHANDLER_PRI_ANY);
583 SYSINIT(arp, SI_SUB_PROTO_DOMAIN, SI_ORDER_ANY, ipflow_init, 0);