KVM: VMX: Fix reporting of unhandled EPT violations
[linux-2.6/linux-2.6-openrd.git] / net / ipv6 / icmp.c
blobeab62a7a8f06e54fd7e161b8312565f8da59ed60
1 /*
2 * Internet Control Message Protocol (ICMPv6)
3 * Linux INET6 implementation
5 * Authors:
6 * Pedro Roque <roque@di.fc.ul.pt>
8 * Based on net/ipv4/icmp.c
10 * RFC 1885
12 * This program is free software; you can redistribute it and/or
13 * modify it under the terms of the GNU General Public License
14 * as published by the Free Software Foundation; either version
15 * 2 of the License, or (at your option) any later version.
19 * Changes:
21 * Andi Kleen : exception handling
22 * Andi Kleen add rate limits. never reply to a icmp.
23 * add more length checks and other fixes.
24 * yoshfuji : ensure to sent parameter problem for
25 * fragments.
26 * YOSHIFUJI Hideaki @USAGI: added sysctl for icmp rate limit.
27 * Randy Dunlap and
28 * YOSHIFUJI Hideaki @USAGI: Per-interface statistics support
29 * Kazunori MIYAZAWA @USAGI: change output process to use ip6_append_data
32 #include <linux/module.h>
33 #include <linux/errno.h>
34 #include <linux/types.h>
35 #include <linux/socket.h>
36 #include <linux/in.h>
37 #include <linux/kernel.h>
38 #include <linux/sockios.h>
39 #include <linux/net.h>
40 #include <linux/skbuff.h>
41 #include <linux/init.h>
42 #include <linux/netfilter.h>
44 #ifdef CONFIG_SYSCTL
45 #include <linux/sysctl.h>
46 #endif
48 #include <linux/inet.h>
49 #include <linux/netdevice.h>
50 #include <linux/icmpv6.h>
52 #include <net/ip.h>
53 #include <net/sock.h>
55 #include <net/ipv6.h>
56 #include <net/ip6_checksum.h>
57 #include <net/protocol.h>
58 #include <net/raw.h>
59 #include <net/rawv6.h>
60 #include <net/transp_v6.h>
61 #include <net/ip6_route.h>
62 #include <net/addrconf.h>
63 #include <net/icmp.h>
64 #include <net/xfrm.h>
65 #include <net/inet_common.h>
67 #include <asm/uaccess.h>
68 #include <asm/system.h>
70 DEFINE_SNMP_STAT(struct icmpv6_mib, icmpv6_statistics) __read_mostly;
71 EXPORT_SYMBOL(icmpv6_statistics);
72 DEFINE_SNMP_STAT(struct icmpv6msg_mib, icmpv6msg_statistics) __read_mostly;
73 EXPORT_SYMBOL(icmpv6msg_statistics);
76 * The ICMP socket(s). This is the most convenient way to flow control
77 * our ICMP output as well as maintain a clean interface throughout
78 * all layers. All Socketless IP sends will soon be gone.
80 * On SMP we have one ICMP socket per-cpu.
82 static inline struct sock *icmpv6_sk(struct net *net)
84 return net->ipv6.icmp_sk[smp_processor_id()];
87 static int icmpv6_rcv(struct sk_buff *skb);
89 static struct inet6_protocol icmpv6_protocol = {
90 .handler = icmpv6_rcv,
91 .flags = INET6_PROTO_NOPOLICY|INET6_PROTO_FINAL,
94 static __inline__ struct sock *icmpv6_xmit_lock(struct net *net)
96 struct sock *sk;
98 local_bh_disable();
100 sk = icmpv6_sk(net);
101 if (unlikely(!spin_trylock(&sk->sk_lock.slock))) {
102 /* This can happen if the output path (f.e. SIT or
103 * ip6ip6 tunnel) signals dst_link_failure() for an
104 * outgoing ICMP6 packet.
106 local_bh_enable();
107 return NULL;
109 return sk;
112 static __inline__ void icmpv6_xmit_unlock(struct sock *sk)
114 spin_unlock_bh(&sk->sk_lock.slock);
118 * Slightly more convenient version of icmpv6_send.
120 void icmpv6_param_prob(struct sk_buff *skb, u8 code, int pos)
122 icmpv6_send(skb, ICMPV6_PARAMPROB, code, pos, skb->dev);
123 kfree_skb(skb);
127 * Figure out, may we reply to this packet with icmp error.
129 * We do not reply, if:
130 * - it was icmp error message.
131 * - it is truncated, so that it is known, that protocol is ICMPV6
132 * (i.e. in the middle of some exthdr)
134 * --ANK (980726)
137 static int is_ineligible(struct sk_buff *skb)
139 int ptr = (u8 *)(ipv6_hdr(skb) + 1) - skb->data;
140 int len = skb->len - ptr;
141 __u8 nexthdr = ipv6_hdr(skb)->nexthdr;
143 if (len < 0)
144 return 1;
146 ptr = ipv6_skip_exthdr(skb, ptr, &nexthdr);
147 if (ptr < 0)
148 return 0;
149 if (nexthdr == IPPROTO_ICMPV6) {
150 u8 _type, *tp;
151 tp = skb_header_pointer(skb,
152 ptr+offsetof(struct icmp6hdr, icmp6_type),
153 sizeof(_type), &_type);
154 if (tp == NULL ||
155 !(*tp & ICMPV6_INFOMSG_MASK))
156 return 1;
158 return 0;
162 * Check the ICMP output rate limit
164 static inline int icmpv6_xrlim_allow(struct sock *sk, u8 type,
165 struct flowi *fl)
167 struct dst_entry *dst;
168 struct net *net = sock_net(sk);
169 int res = 0;
171 /* Informational messages are not limited. */
172 if (type & ICMPV6_INFOMSG_MASK)
173 return 1;
175 /* Do not limit pmtu discovery, it would break it. */
176 if (type == ICMPV6_PKT_TOOBIG)
177 return 1;
180 * Look up the output route.
181 * XXX: perhaps the expire for routing entries cloned by
182 * this lookup should be more aggressive (not longer than timeout).
184 dst = ip6_route_output(net, sk, fl);
185 if (dst->error) {
186 IP6_INC_STATS(net, ip6_dst_idev(dst),
187 IPSTATS_MIB_OUTNOROUTES);
188 } else if (dst->dev && (dst->dev->flags&IFF_LOOPBACK)) {
189 res = 1;
190 } else {
191 struct rt6_info *rt = (struct rt6_info *)dst;
192 int tmo = net->ipv6.sysctl.icmpv6_time;
194 /* Give more bandwidth to wider prefixes. */
195 if (rt->rt6i_dst.plen < 128)
196 tmo >>= ((128 - rt->rt6i_dst.plen)>>5);
198 res = xrlim_allow(dst, tmo);
200 dst_release(dst);
201 return res;
205 * an inline helper for the "simple" if statement below
206 * checks if parameter problem report is caused by an
207 * unrecognized IPv6 option that has the Option Type
208 * highest-order two bits set to 10
211 static __inline__ int opt_unrec(struct sk_buff *skb, __u32 offset)
213 u8 _optval, *op;
215 offset += skb_network_offset(skb);
216 op = skb_header_pointer(skb, offset, sizeof(_optval), &_optval);
217 if (op == NULL)
218 return 1;
219 return (*op & 0xC0) == 0x80;
222 static int icmpv6_push_pending_frames(struct sock *sk, struct flowi *fl, struct icmp6hdr *thdr, int len)
224 struct sk_buff *skb;
225 struct icmp6hdr *icmp6h;
226 int err = 0;
228 if ((skb = skb_peek(&sk->sk_write_queue)) == NULL)
229 goto out;
231 icmp6h = icmp6_hdr(skb);
232 memcpy(icmp6h, thdr, sizeof(struct icmp6hdr));
233 icmp6h->icmp6_cksum = 0;
235 if (skb_queue_len(&sk->sk_write_queue) == 1) {
236 skb->csum = csum_partial(icmp6h,
237 sizeof(struct icmp6hdr), skb->csum);
238 icmp6h->icmp6_cksum = csum_ipv6_magic(&fl->fl6_src,
239 &fl->fl6_dst,
240 len, fl->proto,
241 skb->csum);
242 } else {
243 __wsum tmp_csum = 0;
245 skb_queue_walk(&sk->sk_write_queue, skb) {
246 tmp_csum = csum_add(tmp_csum, skb->csum);
249 tmp_csum = csum_partial(icmp6h,
250 sizeof(struct icmp6hdr), tmp_csum);
251 icmp6h->icmp6_cksum = csum_ipv6_magic(&fl->fl6_src,
252 &fl->fl6_dst,
253 len, fl->proto,
254 tmp_csum);
256 ip6_push_pending_frames(sk);
257 out:
258 return err;
261 struct icmpv6_msg {
262 struct sk_buff *skb;
263 int offset;
264 uint8_t type;
267 static int icmpv6_getfrag(void *from, char *to, int offset, int len, int odd, struct sk_buff *skb)
269 struct icmpv6_msg *msg = (struct icmpv6_msg *) from;
270 struct sk_buff *org_skb = msg->skb;
271 __wsum csum = 0;
273 csum = skb_copy_and_csum_bits(org_skb, msg->offset + offset,
274 to, len, csum);
275 skb->csum = csum_block_add(skb->csum, csum, odd);
276 if (!(msg->type & ICMPV6_INFOMSG_MASK))
277 nf_ct_attach(skb, org_skb);
278 return 0;
281 #if defined(CONFIG_IPV6_MIP6) || defined(CONFIG_IPV6_MIP6_MODULE)
282 static void mip6_addr_swap(struct sk_buff *skb)
284 struct ipv6hdr *iph = ipv6_hdr(skb);
285 struct inet6_skb_parm *opt = IP6CB(skb);
286 struct ipv6_destopt_hao *hao;
287 struct in6_addr tmp;
288 int off;
290 if (opt->dsthao) {
291 off = ipv6_find_tlv(skb, opt->dsthao, IPV6_TLV_HAO);
292 if (likely(off >= 0)) {
293 hao = (struct ipv6_destopt_hao *)
294 (skb_network_header(skb) + off);
295 ipv6_addr_copy(&tmp, &iph->saddr);
296 ipv6_addr_copy(&iph->saddr, &hao->addr);
297 ipv6_addr_copy(&hao->addr, &tmp);
301 #else
302 static inline void mip6_addr_swap(struct sk_buff *skb) {}
303 #endif
306 * Send an ICMP message in response to a packet in error
308 void icmpv6_send(struct sk_buff *skb, u8 type, u8 code, __u32 info,
309 struct net_device *dev)
311 struct net *net = dev_net(skb->dev);
312 struct inet6_dev *idev = NULL;
313 struct ipv6hdr *hdr = ipv6_hdr(skb);
314 struct sock *sk;
315 struct ipv6_pinfo *np;
316 struct in6_addr *saddr = NULL;
317 struct dst_entry *dst;
318 struct dst_entry *dst2;
319 struct icmp6hdr tmp_hdr;
320 struct flowi fl;
321 struct flowi fl2;
322 struct icmpv6_msg msg;
323 int iif = 0;
324 int addr_type = 0;
325 int len;
326 int hlimit, tclass;
327 int err = 0;
329 if ((u8 *)hdr < skb->head ||
330 (skb->network_header + sizeof(*hdr)) > skb->tail)
331 return;
334 * Make sure we respect the rules
335 * i.e. RFC 1885 2.4(e)
336 * Rule (e.1) is enforced by not using icmpv6_send
337 * in any code that processes icmp errors.
339 addr_type = ipv6_addr_type(&hdr->daddr);
341 if (ipv6_chk_addr(net, &hdr->daddr, skb->dev, 0))
342 saddr = &hdr->daddr;
345 * Dest addr check
348 if ((addr_type & IPV6_ADDR_MULTICAST || skb->pkt_type != PACKET_HOST)) {
349 if (type != ICMPV6_PKT_TOOBIG &&
350 !(type == ICMPV6_PARAMPROB &&
351 code == ICMPV6_UNK_OPTION &&
352 (opt_unrec(skb, info))))
353 return;
355 saddr = NULL;
358 addr_type = ipv6_addr_type(&hdr->saddr);
361 * Source addr check
364 if (addr_type & IPV6_ADDR_LINKLOCAL)
365 iif = skb->dev->ifindex;
368 * Must not send error if the source does not uniquely
369 * identify a single node (RFC2463 Section 2.4).
370 * We check unspecified / multicast addresses here,
371 * and anycast addresses will be checked later.
373 if ((addr_type == IPV6_ADDR_ANY) || (addr_type & IPV6_ADDR_MULTICAST)) {
374 LIMIT_NETDEBUG(KERN_DEBUG "icmpv6_send: addr_any/mcast source\n");
375 return;
379 * Never answer to a ICMP packet.
381 if (is_ineligible(skb)) {
382 LIMIT_NETDEBUG(KERN_DEBUG "icmpv6_send: no reply to icmp error\n");
383 return;
386 mip6_addr_swap(skb);
388 memset(&fl, 0, sizeof(fl));
389 fl.proto = IPPROTO_ICMPV6;
390 ipv6_addr_copy(&fl.fl6_dst, &hdr->saddr);
391 if (saddr)
392 ipv6_addr_copy(&fl.fl6_src, saddr);
393 fl.oif = iif;
394 fl.fl_icmp_type = type;
395 fl.fl_icmp_code = code;
396 security_skb_classify_flow(skb, &fl);
398 sk = icmpv6_xmit_lock(net);
399 if (sk == NULL)
400 return;
401 np = inet6_sk(sk);
403 if (!icmpv6_xrlim_allow(sk, type, &fl))
404 goto out;
406 tmp_hdr.icmp6_type = type;
407 tmp_hdr.icmp6_code = code;
408 tmp_hdr.icmp6_cksum = 0;
409 tmp_hdr.icmp6_pointer = htonl(info);
411 if (!fl.oif && ipv6_addr_is_multicast(&fl.fl6_dst))
412 fl.oif = np->mcast_oif;
414 err = ip6_dst_lookup(sk, &dst, &fl);
415 if (err)
416 goto out;
419 * We won't send icmp if the destination is known
420 * anycast.
422 if (((struct rt6_info *)dst)->rt6i_flags & RTF_ANYCAST) {
423 LIMIT_NETDEBUG(KERN_DEBUG "icmpv6_send: acast source\n");
424 goto out_dst_release;
427 /* No need to clone since we're just using its address. */
428 dst2 = dst;
430 err = xfrm_lookup(net, &dst, &fl, sk, 0);
431 switch (err) {
432 case 0:
433 if (dst != dst2)
434 goto route_done;
435 break;
436 case -EPERM:
437 dst = NULL;
438 break;
439 default:
440 goto out;
443 if (xfrm_decode_session_reverse(skb, &fl2, AF_INET6))
444 goto relookup_failed;
446 if (ip6_dst_lookup(sk, &dst2, &fl2))
447 goto relookup_failed;
449 err = xfrm_lookup(net, &dst2, &fl2, sk, XFRM_LOOKUP_ICMP);
450 switch (err) {
451 case 0:
452 dst_release(dst);
453 dst = dst2;
454 break;
455 case -EPERM:
456 goto out_dst_release;
457 default:
458 relookup_failed:
459 if (!dst)
460 goto out;
461 break;
464 route_done:
465 if (ipv6_addr_is_multicast(&fl.fl6_dst))
466 hlimit = np->mcast_hops;
467 else
468 hlimit = np->hop_limit;
469 if (hlimit < 0)
470 hlimit = ip6_dst_hoplimit(dst);
472 tclass = np->tclass;
473 if (tclass < 0)
474 tclass = 0;
476 msg.skb = skb;
477 msg.offset = skb_network_offset(skb);
478 msg.type = type;
480 len = skb->len - msg.offset;
481 len = min_t(unsigned int, len, IPV6_MIN_MTU - sizeof(struct ipv6hdr) -sizeof(struct icmp6hdr));
482 if (len < 0) {
483 LIMIT_NETDEBUG(KERN_DEBUG "icmp: len problem\n");
484 goto out_dst_release;
487 idev = in6_dev_get(skb->dev);
489 err = ip6_append_data(sk, icmpv6_getfrag, &msg,
490 len + sizeof(struct icmp6hdr),
491 sizeof(struct icmp6hdr),
492 hlimit, tclass, NULL, &fl, (struct rt6_info*)dst,
493 MSG_DONTWAIT);
494 if (err) {
495 ip6_flush_pending_frames(sk);
496 goto out_put;
498 err = icmpv6_push_pending_frames(sk, &fl, &tmp_hdr, len + sizeof(struct icmp6hdr));
500 out_put:
501 if (likely(idev != NULL))
502 in6_dev_put(idev);
503 out_dst_release:
504 dst_release(dst);
505 out:
506 icmpv6_xmit_unlock(sk);
509 EXPORT_SYMBOL(icmpv6_send);
511 static void icmpv6_echo_reply(struct sk_buff *skb)
513 struct net *net = dev_net(skb->dev);
514 struct sock *sk;
515 struct inet6_dev *idev;
516 struct ipv6_pinfo *np;
517 struct in6_addr *saddr = NULL;
518 struct icmp6hdr *icmph = icmp6_hdr(skb);
519 struct icmp6hdr tmp_hdr;
520 struct flowi fl;
521 struct icmpv6_msg msg;
522 struct dst_entry *dst;
523 int err = 0;
524 int hlimit;
525 int tclass;
527 saddr = &ipv6_hdr(skb)->daddr;
529 if (!ipv6_unicast_destination(skb))
530 saddr = NULL;
532 memcpy(&tmp_hdr, icmph, sizeof(tmp_hdr));
533 tmp_hdr.icmp6_type = ICMPV6_ECHO_REPLY;
535 memset(&fl, 0, sizeof(fl));
536 fl.proto = IPPROTO_ICMPV6;
537 ipv6_addr_copy(&fl.fl6_dst, &ipv6_hdr(skb)->saddr);
538 if (saddr)
539 ipv6_addr_copy(&fl.fl6_src, saddr);
540 fl.oif = skb->dev->ifindex;
541 fl.fl_icmp_type = ICMPV6_ECHO_REPLY;
542 security_skb_classify_flow(skb, &fl);
544 sk = icmpv6_xmit_lock(net);
545 if (sk == NULL)
546 return;
547 np = inet6_sk(sk);
549 if (!fl.oif && ipv6_addr_is_multicast(&fl.fl6_dst))
550 fl.oif = np->mcast_oif;
552 err = ip6_dst_lookup(sk, &dst, &fl);
553 if (err)
554 goto out;
555 if ((err = xfrm_lookup(net, &dst, &fl, sk, 0)) < 0)
556 goto out;
558 if (ipv6_addr_is_multicast(&fl.fl6_dst))
559 hlimit = np->mcast_hops;
560 else
561 hlimit = np->hop_limit;
562 if (hlimit < 0)
563 hlimit = ip6_dst_hoplimit(dst);
565 tclass = np->tclass;
566 if (tclass < 0)
567 tclass = 0;
569 idev = in6_dev_get(skb->dev);
571 msg.skb = skb;
572 msg.offset = 0;
573 msg.type = ICMPV6_ECHO_REPLY;
575 err = ip6_append_data(sk, icmpv6_getfrag, &msg, skb->len + sizeof(struct icmp6hdr),
576 sizeof(struct icmp6hdr), hlimit, tclass, NULL, &fl,
577 (struct rt6_info*)dst, MSG_DONTWAIT);
579 if (err) {
580 ip6_flush_pending_frames(sk);
581 goto out_put;
583 err = icmpv6_push_pending_frames(sk, &fl, &tmp_hdr, skb->len + sizeof(struct icmp6hdr));
585 out_put:
586 if (likely(idev != NULL))
587 in6_dev_put(idev);
588 dst_release(dst);
589 out:
590 icmpv6_xmit_unlock(sk);
593 static void icmpv6_notify(struct sk_buff *skb, u8 type, u8 code, __be32 info)
595 struct inet6_protocol *ipprot;
596 int inner_offset;
597 int hash;
598 u8 nexthdr;
600 if (!pskb_may_pull(skb, sizeof(struct ipv6hdr)))
601 return;
603 nexthdr = ((struct ipv6hdr *)skb->data)->nexthdr;
604 if (ipv6_ext_hdr(nexthdr)) {
605 /* now skip over extension headers */
606 inner_offset = ipv6_skip_exthdr(skb, sizeof(struct ipv6hdr), &nexthdr);
607 if (inner_offset<0)
608 return;
609 } else {
610 inner_offset = sizeof(struct ipv6hdr);
613 /* Checkin header including 8 bytes of inner protocol header. */
614 if (!pskb_may_pull(skb, inner_offset+8))
615 return;
617 /* BUGGG_FUTURE: we should try to parse exthdrs in this packet.
618 Without this we will not able f.e. to make source routed
619 pmtu discovery.
620 Corresponding argument (opt) to notifiers is already added.
621 --ANK (980726)
624 hash = nexthdr & (MAX_INET_PROTOS - 1);
626 rcu_read_lock();
627 ipprot = rcu_dereference(inet6_protos[hash]);
628 if (ipprot && ipprot->err_handler)
629 ipprot->err_handler(skb, NULL, type, code, inner_offset, info);
630 rcu_read_unlock();
632 raw6_icmp_error(skb, nexthdr, type, code, inner_offset, info);
636 * Handle icmp messages
639 static int icmpv6_rcv(struct sk_buff *skb)
641 struct net_device *dev = skb->dev;
642 struct inet6_dev *idev = __in6_dev_get(dev);
643 struct in6_addr *saddr, *daddr;
644 struct ipv6hdr *orig_hdr;
645 struct icmp6hdr *hdr;
646 u8 type;
648 if (!xfrm6_policy_check(NULL, XFRM_POLICY_IN, skb)) {
649 struct sec_path *sp = skb_sec_path(skb);
650 int nh;
652 if (!(sp && sp->xvec[sp->len - 1]->props.flags &
653 XFRM_STATE_ICMP))
654 goto drop_no_count;
656 if (!pskb_may_pull(skb, sizeof(*hdr) + sizeof(*orig_hdr)))
657 goto drop_no_count;
659 nh = skb_network_offset(skb);
660 skb_set_network_header(skb, sizeof(*hdr));
662 if (!xfrm6_policy_check_reverse(NULL, XFRM_POLICY_IN, skb))
663 goto drop_no_count;
665 skb_set_network_header(skb, nh);
668 ICMP6_INC_STATS_BH(dev_net(dev), idev, ICMP6_MIB_INMSGS);
670 saddr = &ipv6_hdr(skb)->saddr;
671 daddr = &ipv6_hdr(skb)->daddr;
673 /* Perform checksum. */
674 switch (skb->ip_summed) {
675 case CHECKSUM_COMPLETE:
676 if (!csum_ipv6_magic(saddr, daddr, skb->len, IPPROTO_ICMPV6,
677 skb->csum))
678 break;
679 /* fall through */
680 case CHECKSUM_NONE:
681 skb->csum = ~csum_unfold(csum_ipv6_magic(saddr, daddr, skb->len,
682 IPPROTO_ICMPV6, 0));
683 if (__skb_checksum_complete(skb)) {
684 LIMIT_NETDEBUG(KERN_DEBUG "ICMPv6 checksum failed [%pI6 > %pI6]\n",
685 saddr, daddr);
686 goto discard_it;
690 if (!pskb_pull(skb, sizeof(*hdr)))
691 goto discard_it;
693 hdr = icmp6_hdr(skb);
695 type = hdr->icmp6_type;
697 ICMP6MSGIN_INC_STATS_BH(dev_net(dev), idev, type);
699 switch (type) {
700 case ICMPV6_ECHO_REQUEST:
701 icmpv6_echo_reply(skb);
702 break;
704 case ICMPV6_ECHO_REPLY:
705 /* we couldn't care less */
706 break;
708 case ICMPV6_PKT_TOOBIG:
709 /* BUGGG_FUTURE: if packet contains rthdr, we cannot update
710 standard destination cache. Seems, only "advanced"
711 destination cache will allow to solve this problem
712 --ANK (980726)
714 if (!pskb_may_pull(skb, sizeof(struct ipv6hdr)))
715 goto discard_it;
716 hdr = icmp6_hdr(skb);
717 orig_hdr = (struct ipv6hdr *) (hdr + 1);
718 rt6_pmtu_discovery(&orig_hdr->daddr, &orig_hdr->saddr, dev,
719 ntohl(hdr->icmp6_mtu));
722 * Drop through to notify
725 case ICMPV6_DEST_UNREACH:
726 case ICMPV6_TIME_EXCEED:
727 case ICMPV6_PARAMPROB:
728 icmpv6_notify(skb, type, hdr->icmp6_code, hdr->icmp6_mtu);
729 break;
731 case NDISC_ROUTER_SOLICITATION:
732 case NDISC_ROUTER_ADVERTISEMENT:
733 case NDISC_NEIGHBOUR_SOLICITATION:
734 case NDISC_NEIGHBOUR_ADVERTISEMENT:
735 case NDISC_REDIRECT:
736 ndisc_rcv(skb);
737 break;
739 case ICMPV6_MGM_QUERY:
740 igmp6_event_query(skb);
741 break;
743 case ICMPV6_MGM_REPORT:
744 igmp6_event_report(skb);
745 break;
747 case ICMPV6_MGM_REDUCTION:
748 case ICMPV6_NI_QUERY:
749 case ICMPV6_NI_REPLY:
750 case ICMPV6_MLD2_REPORT:
751 case ICMPV6_DHAAD_REQUEST:
752 case ICMPV6_DHAAD_REPLY:
753 case ICMPV6_MOBILE_PREFIX_SOL:
754 case ICMPV6_MOBILE_PREFIX_ADV:
755 break;
757 default:
758 LIMIT_NETDEBUG(KERN_DEBUG "icmpv6: msg of unknown type\n");
760 /* informational */
761 if (type & ICMPV6_INFOMSG_MASK)
762 break;
765 * error of unknown type.
766 * must pass to upper level
769 icmpv6_notify(skb, type, hdr->icmp6_code, hdr->icmp6_mtu);
772 kfree_skb(skb);
773 return 0;
775 discard_it:
776 ICMP6_INC_STATS_BH(dev_net(dev), idev, ICMP6_MIB_INERRORS);
777 drop_no_count:
778 kfree_skb(skb);
779 return 0;
782 void icmpv6_flow_init(struct sock *sk, struct flowi *fl,
783 u8 type,
784 const struct in6_addr *saddr,
785 const struct in6_addr *daddr,
786 int oif)
788 memset(fl, 0, sizeof(*fl));
789 ipv6_addr_copy(&fl->fl6_src, saddr);
790 ipv6_addr_copy(&fl->fl6_dst, daddr);
791 fl->proto = IPPROTO_ICMPV6;
792 fl->fl_icmp_type = type;
793 fl->fl_icmp_code = 0;
794 fl->oif = oif;
795 security_sk_classify_flow(sk, fl);
799 * Special lock-class for __icmpv6_sk:
801 static struct lock_class_key icmpv6_socket_sk_dst_lock_key;
803 static int __net_init icmpv6_sk_init(struct net *net)
805 struct sock *sk;
806 int err, i, j;
808 net->ipv6.icmp_sk =
809 kzalloc(nr_cpu_ids * sizeof(struct sock *), GFP_KERNEL);
810 if (net->ipv6.icmp_sk == NULL)
811 return -ENOMEM;
813 for_each_possible_cpu(i) {
814 err = inet_ctl_sock_create(&sk, PF_INET6,
815 SOCK_RAW, IPPROTO_ICMPV6, net);
816 if (err < 0) {
817 printk(KERN_ERR
818 "Failed to initialize the ICMP6 control socket "
819 "(err %d).\n",
820 err);
821 goto fail;
824 net->ipv6.icmp_sk[i] = sk;
827 * Split off their lock-class, because sk->sk_dst_lock
828 * gets used from softirqs, which is safe for
829 * __icmpv6_sk (because those never get directly used
830 * via userspace syscalls), but unsafe for normal sockets.
832 lockdep_set_class(&sk->sk_dst_lock,
833 &icmpv6_socket_sk_dst_lock_key);
835 /* Enough space for 2 64K ICMP packets, including
836 * sk_buff struct overhead.
838 sk->sk_sndbuf =
839 (2 * ((64 * 1024) + sizeof(struct sk_buff)));
841 return 0;
843 fail:
844 for (j = 0; j < i; j++)
845 inet_ctl_sock_destroy(net->ipv6.icmp_sk[j]);
846 kfree(net->ipv6.icmp_sk);
847 return err;
850 static void __net_exit icmpv6_sk_exit(struct net *net)
852 int i;
854 for_each_possible_cpu(i) {
855 inet_ctl_sock_destroy(net->ipv6.icmp_sk[i]);
857 kfree(net->ipv6.icmp_sk);
860 static struct pernet_operations icmpv6_sk_ops = {
861 .init = icmpv6_sk_init,
862 .exit = icmpv6_sk_exit,
865 int __init icmpv6_init(void)
867 int err;
869 err = register_pernet_subsys(&icmpv6_sk_ops);
870 if (err < 0)
871 return err;
873 err = -EAGAIN;
874 if (inet6_add_protocol(&icmpv6_protocol, IPPROTO_ICMPV6) < 0)
875 goto fail;
876 return 0;
878 fail:
879 printk(KERN_ERR "Failed to register ICMP6 protocol\n");
880 unregister_pernet_subsys(&icmpv6_sk_ops);
881 return err;
884 void icmpv6_cleanup(void)
886 unregister_pernet_subsys(&icmpv6_sk_ops);
887 inet6_del_protocol(&icmpv6_protocol, IPPROTO_ICMPV6);
891 static const struct icmp6_err {
892 int err;
893 int fatal;
894 } tab_unreach[] = {
895 { /* NOROUTE */
896 .err = ENETUNREACH,
897 .fatal = 0,
899 { /* ADM_PROHIBITED */
900 .err = EACCES,
901 .fatal = 1,
903 { /* Was NOT_NEIGHBOUR, now reserved */
904 .err = EHOSTUNREACH,
905 .fatal = 0,
907 { /* ADDR_UNREACH */
908 .err = EHOSTUNREACH,
909 .fatal = 0,
911 { /* PORT_UNREACH */
912 .err = ECONNREFUSED,
913 .fatal = 1,
917 int icmpv6_err_convert(u8 type, u8 code, int *err)
919 int fatal = 0;
921 *err = EPROTO;
923 switch (type) {
924 case ICMPV6_DEST_UNREACH:
925 fatal = 1;
926 if (code <= ICMPV6_PORT_UNREACH) {
927 *err = tab_unreach[code].err;
928 fatal = tab_unreach[code].fatal;
930 break;
932 case ICMPV6_PKT_TOOBIG:
933 *err = EMSGSIZE;
934 break;
936 case ICMPV6_PARAMPROB:
937 *err = EPROTO;
938 fatal = 1;
939 break;
941 case ICMPV6_TIME_EXCEED:
942 *err = EHOSTUNREACH;
943 break;
946 return fatal;
949 EXPORT_SYMBOL(icmpv6_err_convert);
951 #ifdef CONFIG_SYSCTL
952 ctl_table ipv6_icmp_table_template[] = {
954 .ctl_name = NET_IPV6_ICMP_RATELIMIT,
955 .procname = "ratelimit",
956 .data = &init_net.ipv6.sysctl.icmpv6_time,
957 .maxlen = sizeof(int),
958 .mode = 0644,
959 .proc_handler = proc_dointvec_ms_jiffies,
960 .strategy = sysctl_ms_jiffies
962 { .ctl_name = 0 },
965 struct ctl_table *ipv6_icmp_sysctl_init(struct net *net)
967 struct ctl_table *table;
969 table = kmemdup(ipv6_icmp_table_template,
970 sizeof(ipv6_icmp_table_template),
971 GFP_KERNEL);
973 if (table)
974 table[0].data = &net->ipv6.sysctl.icmpv6_time;
976 return table;
978 #endif