2 * Extension Header handling for IPv6
3 * Linux INET6 implementation
6 * Pedro Roque <roque@di.fc.ul.pt>
7 * Andi Kleen <ak@muc.de>
8 * Alexey Kuznetsov <kuznet@ms2.inr.ac.ru>
10 * This program is free software; you can redistribute it and/or
11 * modify it under the terms of the GNU General Public License
12 * as published by the Free Software Foundation; either version
13 * 2 of the License, or (at your option) any later version.
17 * yoshfuji : ensure not to overrun while parsing
19 * Mitsuru KANDA @USAGI and: Remove ipv6_parse_exthdrs().
20 * YOSHIFUJI Hideaki @USAGI Register inbound extension header
21 * handlers as inet6_protocol{}.
24 #include <linux/errno.h>
25 #include <linux/types.h>
26 #include <linux/socket.h>
27 #include <linux/sockios.h>
28 #include <linux/net.h>
29 #include <linux/netdevice.h>
30 #include <linux/in6.h>
31 #include <linux/icmpv6.h>
38 #include <net/protocol.h>
39 #include <net/transp_v6.h>
40 #include <net/rawv6.h>
41 #include <net/ndisc.h>
42 #include <net/ip6_route.h>
43 #include <net/addrconf.h>
44 #if defined(CONFIG_IPV6_MIP6) || defined(CONFIG_IPV6_MIP6_MODULE)
48 #include <asm/uaccess.h>
50 int ipv6_find_tlv(struct sk_buff
*skb
, int offset
, int type
)
52 const unsigned char *nh
= skb_network_header(skb
);
53 int packet_len
= skb
->tail
- skb
->network_header
;
54 struct ipv6_opt_hdr
*hdr
;
57 if (offset
+ 2 > packet_len
)
59 hdr
= (struct ipv6_opt_hdr
*)(nh
+ offset
);
60 len
= ((hdr
->hdrlen
+ 1) << 3);
62 if (offset
+ len
> packet_len
)
69 int opttype
= nh
[offset
];
80 optlen
= nh
[offset
+ 1] + 2;
92 EXPORT_SYMBOL_GPL(ipv6_find_tlv
);
95 * Parsing tlv encoded headers.
97 * Parsing function "func" returns 1, if parsing succeed
98 * and 0, if it failed.
99 * It MUST NOT touch skb->h.
102 struct tlvtype_proc
{
104 int (*func
)(struct sk_buff
*skb
, int offset
);
107 /*********************
109 *********************/
111 /* An unknown option is detected, decide what to do */
113 static int ip6_tlvopt_unknown(struct sk_buff
*skb
, int optoff
)
115 switch ((skb_network_header(skb
)[optoff
] & 0xC0) >> 6) {
119 case 1: /* drop packet */
122 case 3: /* Send ICMP if not a multicast address and drop packet */
123 /* Actually, it is redundant check. icmp_send
124 will recheck in any case.
126 if (ipv6_addr_is_multicast(&ipv6_hdr(skb
)->daddr
))
128 case 2: /* send ICMP PARM PROB regardless and drop packet */
129 icmpv6_param_prob(skb
, ICMPV6_UNK_OPTION
, optoff
);
137 /* Parse tlv encoded option header (hop-by-hop or destination) */
139 static int ip6_parse_tlv(struct tlvtype_proc
*procs
, struct sk_buff
*skb
)
141 struct tlvtype_proc
*curr
;
142 const unsigned char *nh
= skb_network_header(skb
);
143 int off
= skb_network_header_len(skb
);
144 int len
= (skb_transport_header(skb
)[1] + 1) << 3;
146 if (skb_transport_offset(skb
) + len
> skb_headlen(skb
))
153 int optlen
= nh
[off
+ 1] + 2;
163 default: /* Other TLV code so scan list */
166 for (curr
=procs
; curr
->type
>= 0; curr
++) {
167 if (curr
->type
== nh
[off
]) {
168 /* type specific length/alignment
169 checks will be performed in the
171 if (curr
->func(skb
, off
) == 0)
176 if (curr
->type
< 0) {
177 if (ip6_tlvopt_unknown(skb
, off
) == 0)
192 /*****************************
193 Destination options header.
194 *****************************/
196 #if defined(CONFIG_IPV6_MIP6) || defined(CONFIG_IPV6_MIP6_MODULE)
197 static int ipv6_dest_hao(struct sk_buff
*skb
, int optoff
)
199 struct ipv6_destopt_hao
*hao
;
200 struct inet6_skb_parm
*opt
= IP6CB(skb
);
201 struct ipv6hdr
*ipv6h
= ipv6_hdr(skb
);
202 struct in6_addr tmp_addr
;
206 LIMIT_NETDEBUG(KERN_DEBUG
"hao duplicated\n");
209 opt
->dsthao
= opt
->dst1
;
212 hao
= (struct ipv6_destopt_hao
*)(skb_network_header(skb
) + optoff
);
214 if (hao
->length
!= 16) {
216 KERN_DEBUG
"hao invalid option length = %d\n", hao
->length
);
220 if (!(ipv6_addr_type(&hao
->addr
) & IPV6_ADDR_UNICAST
)) {
222 KERN_DEBUG
"hao is not an unicast addr: " NIP6_FMT
"\n", NIP6(hao
->addr
));
226 ret
= xfrm6_input_addr(skb
, (xfrm_address_t
*)&ipv6h
->daddr
,
227 (xfrm_address_t
*)&hao
->addr
, IPPROTO_DSTOPTS
);
228 if (unlikely(ret
< 0))
231 if (skb_cloned(skb
)) {
232 if (pskb_expand_head(skb
, 0, 0, GFP_ATOMIC
))
235 /* update all variable using below by copied skbuff */
236 hao
= (struct ipv6_destopt_hao
*)(skb_network_header(skb
) +
238 ipv6h
= ipv6_hdr(skb
);
241 if (skb
->ip_summed
== CHECKSUM_COMPLETE
)
242 skb
->ip_summed
= CHECKSUM_NONE
;
244 ipv6_addr_copy(&tmp_addr
, &ipv6h
->saddr
);
245 ipv6_addr_copy(&ipv6h
->saddr
, &hao
->addr
);
246 ipv6_addr_copy(&hao
->addr
, &tmp_addr
);
248 if (skb
->tstamp
.tv64
== 0)
249 __net_timestamp(skb
);
259 static struct tlvtype_proc tlvprocdestopt_lst
[] = {
260 #if defined(CONFIG_IPV6_MIP6) || defined(CONFIG_IPV6_MIP6_MODULE)
262 .type
= IPV6_TLV_HAO
,
263 .func
= ipv6_dest_hao
,
269 static int ipv6_destopt_rcv(struct sk_buff
*skb
)
271 struct inet6_skb_parm
*opt
= IP6CB(skb
);
272 #if defined(CONFIG_IPV6_MIP6) || defined(CONFIG_IPV6_MIP6_MODULE)
275 struct dst_entry
*dst
;
277 if (!pskb_may_pull(skb
, skb_transport_offset(skb
) + 8) ||
278 !pskb_may_pull(skb
, (skb_transport_offset(skb
) +
279 ((skb_transport_header(skb
)[1] + 1) << 3)))) {
280 IP6_INC_STATS_BH(ip6_dst_idev(skb
->dst
),
281 IPSTATS_MIB_INHDRERRORS
);
286 opt
->lastopt
= opt
->dst1
= skb_network_header_len(skb
);
287 #if defined(CONFIG_IPV6_MIP6) || defined(CONFIG_IPV6_MIP6_MODULE)
291 dst
= dst_clone(skb
->dst
);
292 if (ip6_parse_tlv(tlvprocdestopt_lst
, skb
)) {
294 skb
->transport_header
+= (skb_transport_header(skb
)[1] + 1) << 3;
296 #if defined(CONFIG_IPV6_MIP6) || defined(CONFIG_IPV6_MIP6_MODULE)
299 opt
->nhoff
= opt
->dst1
;
304 IP6_INC_STATS_BH(ip6_dst_idev(dst
), IPSTATS_MIB_INHDRERRORS
);
309 /********************************
311 ********************************/
313 static int ipv6_rthdr_rcv(struct sk_buff
*skb
)
315 struct inet6_skb_parm
*opt
= IP6CB(skb
);
316 struct in6_addr
*addr
= NULL
;
317 struct in6_addr daddr
;
318 struct inet6_dev
*idev
;
320 struct ipv6_rt_hdr
*hdr
;
321 struct rt0_hdr
*rthdr
;
322 int accept_source_route
= dev_net(skb
->dev
)->ipv6
.devconf_all
->accept_source_route
;
324 idev
= in6_dev_get(skb
->dev
);
326 if (accept_source_route
> idev
->cnf
.accept_source_route
)
327 accept_source_route
= idev
->cnf
.accept_source_route
;
331 if (!pskb_may_pull(skb
, skb_transport_offset(skb
) + 8) ||
332 !pskb_may_pull(skb
, (skb_transport_offset(skb
) +
333 ((skb_transport_header(skb
)[1] + 1) << 3)))) {
334 IP6_INC_STATS_BH(ip6_dst_idev(skb
->dst
),
335 IPSTATS_MIB_INHDRERRORS
);
340 hdr
= (struct ipv6_rt_hdr
*)skb_transport_header(skb
);
342 if (ipv6_addr_is_multicast(&ipv6_hdr(skb
)->daddr
) ||
343 skb
->pkt_type
!= PACKET_HOST
) {
344 IP6_INC_STATS_BH(ip6_dst_idev(skb
->dst
),
345 IPSTATS_MIB_INADDRERRORS
);
351 if (hdr
->segments_left
== 0) {
353 #if defined(CONFIG_IPV6_MIP6) || defined(CONFIG_IPV6_MIP6_MODULE)
354 case IPV6_SRCRT_TYPE_2
:
355 /* Silently discard type 2 header unless it was
359 IP6_INC_STATS_BH(ip6_dst_idev(skb
->dst
),
360 IPSTATS_MIB_INADDRERRORS
);
370 opt
->lastopt
= opt
->srcrt
= skb_network_header_len(skb
);
371 skb
->transport_header
+= (hdr
->hdrlen
+ 1) << 3;
372 opt
->dst0
= opt
->dst1
;
374 opt
->nhoff
= (&hdr
->nexthdr
) - skb_network_header(skb
);
379 #if defined(CONFIG_IPV6_MIP6) || defined(CONFIG_IPV6_MIP6_MODULE)
380 case IPV6_SRCRT_TYPE_2
:
381 if (accept_source_route
< 0)
383 /* Silently discard invalid RTH type 2 */
384 if (hdr
->hdrlen
!= 2 || hdr
->segments_left
!= 1) {
385 IP6_INC_STATS_BH(ip6_dst_idev(skb
->dst
),
386 IPSTATS_MIB_INHDRERRORS
);
397 * This is the routing header forwarding algorithm from
401 n
= hdr
->hdrlen
>> 1;
403 if (hdr
->segments_left
> n
) {
404 IP6_INC_STATS_BH(ip6_dst_idev(skb
->dst
),
405 IPSTATS_MIB_INHDRERRORS
);
406 icmpv6_param_prob(skb
, ICMPV6_HDR_FIELD
,
407 ((&hdr
->segments_left
) -
408 skb_network_header(skb
)));
412 /* We are about to mangle packet header. Be careful!
413 Do not damage packets queued somewhere.
415 if (skb_cloned(skb
)) {
416 /* the copy is a forwarded packet */
417 if (pskb_expand_head(skb
, 0, 0, GFP_ATOMIC
)) {
418 IP6_INC_STATS_BH(ip6_dst_idev(skb
->dst
),
419 IPSTATS_MIB_OUTDISCARDS
);
423 hdr
= (struct ipv6_rt_hdr
*)skb_transport_header(skb
);
426 if (skb
->ip_summed
== CHECKSUM_COMPLETE
)
427 skb
->ip_summed
= CHECKSUM_NONE
;
429 i
= n
- --hdr
->segments_left
;
431 rthdr
= (struct rt0_hdr
*) hdr
;
436 #if defined(CONFIG_IPV6_MIP6) || defined(CONFIG_IPV6_MIP6_MODULE)
437 case IPV6_SRCRT_TYPE_2
:
438 if (xfrm6_input_addr(skb
, (xfrm_address_t
*)addr
,
439 (xfrm_address_t
*)&ipv6_hdr(skb
)->saddr
,
440 IPPROTO_ROUTING
) < 0) {
441 IP6_INC_STATS_BH(ip6_dst_idev(skb
->dst
),
442 IPSTATS_MIB_INADDRERRORS
);
446 if (!ipv6_chk_home_addr(dev_net(skb
->dst
->dev
), addr
)) {
447 IP6_INC_STATS_BH(ip6_dst_idev(skb
->dst
),
448 IPSTATS_MIB_INADDRERRORS
);
458 if (ipv6_addr_is_multicast(addr
)) {
459 IP6_INC_STATS_BH(ip6_dst_idev(skb
->dst
),
460 IPSTATS_MIB_INADDRERRORS
);
465 ipv6_addr_copy(&daddr
, addr
);
466 ipv6_addr_copy(addr
, &ipv6_hdr(skb
)->daddr
);
467 ipv6_addr_copy(&ipv6_hdr(skb
)->daddr
, &daddr
);
469 dst_release(xchg(&skb
->dst
, NULL
));
470 ip6_route_input(skb
);
471 if (skb
->dst
->error
) {
472 skb_push(skb
, skb
->data
- skb_network_header(skb
));
477 if (skb
->dst
->dev
->flags
&IFF_LOOPBACK
) {
478 if (ipv6_hdr(skb
)->hop_limit
<= 1) {
479 IP6_INC_STATS_BH(ip6_dst_idev(skb
->dst
),
480 IPSTATS_MIB_INHDRERRORS
);
481 icmpv6_send(skb
, ICMPV6_TIME_EXCEED
, ICMPV6_EXC_HOPLIMIT
,
486 ipv6_hdr(skb
)->hop_limit
--;
490 skb_push(skb
, skb
->data
- skb_network_header(skb
));
495 IP6_INC_STATS_BH(ip6_dst_idev(skb
->dst
), IPSTATS_MIB_INHDRERRORS
);
496 icmpv6_param_prob(skb
, ICMPV6_HDR_FIELD
,
497 (&hdr
->type
) - skb_network_header(skb
));
501 static struct inet6_protocol rthdr_protocol
= {
502 .handler
= ipv6_rthdr_rcv
,
503 .flags
= INET6_PROTO_NOPOLICY
| INET6_PROTO_GSO_EXTHDR
,
506 static struct inet6_protocol destopt_protocol
= {
507 .handler
= ipv6_destopt_rcv
,
508 .flags
= INET6_PROTO_NOPOLICY
| INET6_PROTO_GSO_EXTHDR
,
511 static struct inet6_protocol nodata_protocol
= {
512 .handler
= dst_discard
,
513 .flags
= INET6_PROTO_NOPOLICY
,
516 int __init
ipv6_exthdrs_init(void)
520 ret
= inet6_add_protocol(&rthdr_protocol
, IPPROTO_ROUTING
);
524 ret
= inet6_add_protocol(&destopt_protocol
, IPPROTO_DSTOPTS
);
528 ret
= inet6_add_protocol(&nodata_protocol
, IPPROTO_NONE
);
535 inet6_del_protocol(&rthdr_protocol
, IPPROTO_ROUTING
);
537 inet6_del_protocol(&destopt_protocol
, IPPROTO_DSTOPTS
);
541 void ipv6_exthdrs_exit(void)
543 inet6_del_protocol(&nodata_protocol
, IPPROTO_NONE
);
544 inet6_del_protocol(&destopt_protocol
, IPPROTO_DSTOPTS
);
545 inet6_del_protocol(&rthdr_protocol
, IPPROTO_ROUTING
);
548 /**********************************
550 **********************************/
553 * Note: we cannot rely on skb->dst before we assign it in ip6_route_input().
555 static inline struct inet6_dev
*ipv6_skb_idev(struct sk_buff
*skb
)
557 return skb
->dst
? ip6_dst_idev(skb
->dst
) : __in6_dev_get(skb
->dev
);
560 /* Router Alert as of RFC 2711 */
562 static int ipv6_hop_ra(struct sk_buff
*skb
, int optoff
)
564 const unsigned char *nh
= skb_network_header(skb
);
566 if (nh
[optoff
+ 1] == 2) {
567 IP6CB(skb
)->ra
= optoff
;
570 LIMIT_NETDEBUG(KERN_DEBUG
"ipv6_hop_ra: wrong RA length %d\n",
578 static int ipv6_hop_jumbo(struct sk_buff
*skb
, int optoff
)
580 const unsigned char *nh
= skb_network_header(skb
);
583 if (nh
[optoff
+ 1] != 4 || (optoff
& 3) != 2) {
584 LIMIT_NETDEBUG(KERN_DEBUG
"ipv6_hop_jumbo: wrong jumbo opt length/alignment %d\n",
586 IP6_INC_STATS_BH(ipv6_skb_idev(skb
),
587 IPSTATS_MIB_INHDRERRORS
);
591 pkt_len
= ntohl(*(__be32
*)(nh
+ optoff
+ 2));
592 if (pkt_len
<= IPV6_MAXPLEN
) {
593 IP6_INC_STATS_BH(ipv6_skb_idev(skb
), IPSTATS_MIB_INHDRERRORS
);
594 icmpv6_param_prob(skb
, ICMPV6_HDR_FIELD
, optoff
+2);
597 if (ipv6_hdr(skb
)->payload_len
) {
598 IP6_INC_STATS_BH(ipv6_skb_idev(skb
), IPSTATS_MIB_INHDRERRORS
);
599 icmpv6_param_prob(skb
, ICMPV6_HDR_FIELD
, optoff
);
603 if (pkt_len
> skb
->len
- sizeof(struct ipv6hdr
)) {
604 IP6_INC_STATS_BH(ipv6_skb_idev(skb
), IPSTATS_MIB_INTRUNCATEDPKTS
);
608 if (pskb_trim_rcsum(skb
, pkt_len
+ sizeof(struct ipv6hdr
)))
618 static struct tlvtype_proc tlvprochopopt_lst
[] = {
620 .type
= IPV6_TLV_ROUTERALERT
,
624 .type
= IPV6_TLV_JUMBO
,
625 .func
= ipv6_hop_jumbo
,
630 int ipv6_parse_hopopts(struct sk_buff
*skb
)
632 struct inet6_skb_parm
*opt
= IP6CB(skb
);
635 * skb_network_header(skb) is equal to skb->data, and
636 * skb_network_header_len(skb) is always equal to
637 * sizeof(struct ipv6hdr) by definition of
638 * hop-by-hop options.
640 if (!pskb_may_pull(skb
, sizeof(struct ipv6hdr
) + 8) ||
641 !pskb_may_pull(skb
, (sizeof(struct ipv6hdr
) +
642 ((skb_transport_header(skb
)[1] + 1) << 3)))) {
647 opt
->hop
= sizeof(struct ipv6hdr
);
648 if (ip6_parse_tlv(tlvprochopopt_lst
, skb
)) {
649 skb
->transport_header
+= (skb_transport_header(skb
)[1] + 1) << 3;
651 opt
->nhoff
= sizeof(struct ipv6hdr
);
658 * Creating outbound headers.
660 * "build" functions work when skb is filled from head to tail (datagram)
661 * "push" functions work when headers are added from tail to head (tcp)
663 * In both cases we assume, that caller reserved enough room
667 static void ipv6_push_rthdr(struct sk_buff
*skb
, u8
*proto
,
668 struct ipv6_rt_hdr
*opt
,
669 struct in6_addr
**addr_p
)
671 struct rt0_hdr
*phdr
, *ihdr
;
674 ihdr
= (struct rt0_hdr
*) opt
;
676 phdr
= (struct rt0_hdr
*) skb_push(skb
, (ihdr
->rt_hdr
.hdrlen
+ 1) << 3);
677 memcpy(phdr
, ihdr
, sizeof(struct rt0_hdr
));
679 hops
= ihdr
->rt_hdr
.hdrlen
>> 1;
682 memcpy(phdr
->addr
, ihdr
->addr
+ 1,
683 (hops
- 1) * sizeof(struct in6_addr
));
685 ipv6_addr_copy(phdr
->addr
+ (hops
- 1), *addr_p
);
686 *addr_p
= ihdr
->addr
;
688 phdr
->rt_hdr
.nexthdr
= *proto
;
689 *proto
= NEXTHDR_ROUTING
;
692 static void ipv6_push_exthdr(struct sk_buff
*skb
, u8
*proto
, u8 type
, struct ipv6_opt_hdr
*opt
)
694 struct ipv6_opt_hdr
*h
= (struct ipv6_opt_hdr
*)skb_push(skb
, ipv6_optlen(opt
));
696 memcpy(h
, opt
, ipv6_optlen(opt
));
701 void ipv6_push_nfrag_opts(struct sk_buff
*skb
, struct ipv6_txoptions
*opt
,
703 struct in6_addr
**daddr
)
706 ipv6_push_rthdr(skb
, proto
, opt
->srcrt
, daddr
);
708 * IPV6_RTHDRDSTOPTS is ignored
709 * unless IPV6_RTHDR is set (RFC3542).
712 ipv6_push_exthdr(skb
, proto
, NEXTHDR_DEST
, opt
->dst0opt
);
715 ipv6_push_exthdr(skb
, proto
, NEXTHDR_HOP
, opt
->hopopt
);
718 EXPORT_SYMBOL(ipv6_push_nfrag_opts
);
720 void ipv6_push_frag_opts(struct sk_buff
*skb
, struct ipv6_txoptions
*opt
, u8
*proto
)
723 ipv6_push_exthdr(skb
, proto
, NEXTHDR_DEST
, opt
->dst1opt
);
726 struct ipv6_txoptions
*
727 ipv6_dup_options(struct sock
*sk
, struct ipv6_txoptions
*opt
)
729 struct ipv6_txoptions
*opt2
;
731 opt2
= sock_kmalloc(sk
, opt
->tot_len
, GFP_ATOMIC
);
733 long dif
= (char*)opt2
- (char*)opt
;
734 memcpy(opt2
, opt
, opt
->tot_len
);
736 *((char**)&opt2
->hopopt
) += dif
;
738 *((char**)&opt2
->dst0opt
) += dif
;
740 *((char**)&opt2
->dst1opt
) += dif
;
742 *((char**)&opt2
->srcrt
) += dif
;
747 EXPORT_SYMBOL_GPL(ipv6_dup_options
);
749 static int ipv6_renew_option(void *ohdr
,
750 struct ipv6_opt_hdr __user
*newopt
, int newoptlen
,
752 struct ipv6_opt_hdr
**hdr
,
757 memcpy(*p
, ohdr
, ipv6_optlen((struct ipv6_opt_hdr
*)ohdr
));
758 *hdr
= (struct ipv6_opt_hdr
*)*p
;
759 *p
+= CMSG_ALIGN(ipv6_optlen(*(struct ipv6_opt_hdr
**)hdr
));
763 if (copy_from_user(*p
, newopt
, newoptlen
))
765 *hdr
= (struct ipv6_opt_hdr
*)*p
;
766 if (ipv6_optlen(*(struct ipv6_opt_hdr
**)hdr
) > newoptlen
)
768 *p
+= CMSG_ALIGN(newoptlen
);
774 struct ipv6_txoptions
*
775 ipv6_renew_options(struct sock
*sk
, struct ipv6_txoptions
*opt
,
777 struct ipv6_opt_hdr __user
*newopt
, int newoptlen
)
781 struct ipv6_txoptions
*opt2
;
785 if (newtype
!= IPV6_HOPOPTS
&& opt
->hopopt
)
786 tot_len
+= CMSG_ALIGN(ipv6_optlen(opt
->hopopt
));
787 if (newtype
!= IPV6_RTHDRDSTOPTS
&& opt
->dst0opt
)
788 tot_len
+= CMSG_ALIGN(ipv6_optlen(opt
->dst0opt
));
789 if (newtype
!= IPV6_RTHDR
&& opt
->srcrt
)
790 tot_len
+= CMSG_ALIGN(ipv6_optlen(opt
->srcrt
));
791 if (newtype
!= IPV6_DSTOPTS
&& opt
->dst1opt
)
792 tot_len
+= CMSG_ALIGN(ipv6_optlen(opt
->dst1opt
));
795 if (newopt
&& newoptlen
)
796 tot_len
+= CMSG_ALIGN(newoptlen
);
801 tot_len
+= sizeof(*opt2
);
802 opt2
= sock_kmalloc(sk
, tot_len
, GFP_ATOMIC
);
804 return ERR_PTR(-ENOBUFS
);
806 memset(opt2
, 0, tot_len
);
808 opt2
->tot_len
= tot_len
;
809 p
= (char *)(opt2
+ 1);
811 err
= ipv6_renew_option(opt
? opt
->hopopt
: NULL
, newopt
, newoptlen
,
812 newtype
!= IPV6_HOPOPTS
,
817 err
= ipv6_renew_option(opt
? opt
->dst0opt
: NULL
, newopt
, newoptlen
,
818 newtype
!= IPV6_RTHDRDSTOPTS
,
823 err
= ipv6_renew_option(opt
? opt
->srcrt
: NULL
, newopt
, newoptlen
,
824 newtype
!= IPV6_RTHDR
,
825 (struct ipv6_opt_hdr
**)&opt2
->srcrt
, &p
);
829 err
= ipv6_renew_option(opt
? opt
->dst1opt
: NULL
, newopt
, newoptlen
,
830 newtype
!= IPV6_DSTOPTS
,
835 opt2
->opt_nflen
= (opt2
->hopopt
? ipv6_optlen(opt2
->hopopt
) : 0) +
836 (opt2
->dst0opt
? ipv6_optlen(opt2
->dst0opt
) : 0) +
837 (opt2
->srcrt
? ipv6_optlen(opt2
->srcrt
) : 0);
838 opt2
->opt_flen
= (opt2
->dst1opt
? ipv6_optlen(opt2
->dst1opt
) : 0);
842 sock_kfree_s(sk
, opt2
, opt2
->tot_len
);
846 struct ipv6_txoptions
*ipv6_fixup_options(struct ipv6_txoptions
*opt_space
,
847 struct ipv6_txoptions
*opt
)
850 * ignore the dest before srcrt unless srcrt is being included.
853 if (opt
&& opt
->dst0opt
&& !opt
->srcrt
) {
854 if (opt_space
!= opt
) {
855 memcpy(opt_space
, opt
, sizeof(*opt_space
));
858 opt
->opt_nflen
-= ipv6_optlen(opt
->dst0opt
);