Loading...
1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * Linux INET6 implementation
4 * FIB front-end.
5 *
6 * Authors:
7 * Pedro Roque <roque@di.fc.ul.pt>
8 */
9
10/* Changes:
11 *
12 * YOSHIFUJI Hideaki @USAGI
13 * reworked default router selection.
14 * - respect outgoing interface
15 * - select from (probably) reachable routers (i.e.
16 * routers in REACHABLE, STALE, DELAY or PROBE states).
17 * - always select the same router if it is (probably)
18 * reachable. otherwise, round-robin the list.
19 * Ville Nuorvala
20 * Fixed routing subtrees.
21 */
22
23#define pr_fmt(fmt) "IPv6: " fmt
24
25#include <linux/capability.h>
26#include <linux/errno.h>
27#include <linux/export.h>
28#include <linux/types.h>
29#include <linux/times.h>
30#include <linux/socket.h>
31#include <linux/sockios.h>
32#include <linux/net.h>
33#include <linux/route.h>
34#include <linux/netdevice.h>
35#include <linux/in6.h>
36#include <linux/mroute6.h>
37#include <linux/init.h>
38#include <linux/if_arp.h>
39#include <linux/proc_fs.h>
40#include <linux/seq_file.h>
41#include <linux/nsproxy.h>
42#include <linux/slab.h>
43#include <linux/jhash.h>
44#include <linux/siphash.h>
45#include <net/net_namespace.h>
46#include <net/snmp.h>
47#include <net/ipv6.h>
48#include <net/ip6_fib.h>
49#include <net/ip6_route.h>
50#include <net/ndisc.h>
51#include <net/addrconf.h>
52#include <net/tcp.h>
53#include <linux/rtnetlink.h>
54#include <net/dst.h>
55#include <net/dst_metadata.h>
56#include <net/xfrm.h>
57#include <net/netevent.h>
58#include <net/netlink.h>
59#include <net/rtnh.h>
60#include <net/lwtunnel.h>
61#include <net/ip_tunnels.h>
62#include <net/l3mdev.h>
63#include <net/ip.h>
64#include <linux/uaccess.h>
65#include <linux/btf_ids.h>
66
67#ifdef CONFIG_SYSCTL
68#include <linux/sysctl.h>
69#endif
70
71static int ip6_rt_type_to_error(u8 fib6_type);
72
73#define CREATE_TRACE_POINTS
74#include <trace/events/fib6.h>
75EXPORT_TRACEPOINT_SYMBOL_GPL(fib6_table_lookup);
76#undef CREATE_TRACE_POINTS
77
78enum rt6_nud_state {
79 RT6_NUD_FAIL_HARD = -3,
80 RT6_NUD_FAIL_PROBE = -2,
81 RT6_NUD_FAIL_DO_RR = -1,
82 RT6_NUD_SUCCEED = 1
83};
84
85INDIRECT_CALLABLE_SCOPE
86struct dst_entry *ip6_dst_check(struct dst_entry *dst, u32 cookie);
87static unsigned int ip6_default_advmss(const struct dst_entry *dst);
88INDIRECT_CALLABLE_SCOPE
89unsigned int ip6_mtu(const struct dst_entry *dst);
90static struct dst_entry *ip6_negative_advice(struct dst_entry *);
91static void ip6_dst_destroy(struct dst_entry *);
92static void ip6_dst_ifdown(struct dst_entry *,
93 struct net_device *dev, int how);
94static int ip6_dst_gc(struct dst_ops *ops);
95
96static int ip6_pkt_discard(struct sk_buff *skb);
97static int ip6_pkt_discard_out(struct net *net, struct sock *sk, struct sk_buff *skb);
98static int ip6_pkt_prohibit(struct sk_buff *skb);
99static int ip6_pkt_prohibit_out(struct net *net, struct sock *sk, struct sk_buff *skb);
100static void ip6_link_failure(struct sk_buff *skb);
101static void ip6_rt_update_pmtu(struct dst_entry *dst, struct sock *sk,
102 struct sk_buff *skb, u32 mtu,
103 bool confirm_neigh);
104static void rt6_do_redirect(struct dst_entry *dst, struct sock *sk,
105 struct sk_buff *skb);
106static int rt6_score_route(const struct fib6_nh *nh, u32 fib6_flags, int oif,
107 int strict);
108static size_t rt6_nlmsg_size(struct fib6_info *f6i);
109static int rt6_fill_node(struct net *net, struct sk_buff *skb,
110 struct fib6_info *rt, struct dst_entry *dst,
111 struct in6_addr *dest, struct in6_addr *src,
112 int iif, int type, u32 portid, u32 seq,
113 unsigned int flags);
114static struct rt6_info *rt6_find_cached_rt(const struct fib6_result *res,
115 const struct in6_addr *daddr,
116 const struct in6_addr *saddr);
117
118#ifdef CONFIG_IPV6_ROUTE_INFO
119static struct fib6_info *rt6_add_route_info(struct net *net,
120 const struct in6_addr *prefix, int prefixlen,
121 const struct in6_addr *gwaddr,
122 struct net_device *dev,
123 unsigned int pref);
124static struct fib6_info *rt6_get_route_info(struct net *net,
125 const struct in6_addr *prefix, int prefixlen,
126 const struct in6_addr *gwaddr,
127 struct net_device *dev);
128#endif
129
130struct uncached_list {
131 spinlock_t lock;
132 struct list_head head;
133};
134
135static DEFINE_PER_CPU_ALIGNED(struct uncached_list, rt6_uncached_list);
136
137void rt6_uncached_list_add(struct rt6_info *rt)
138{
139 struct uncached_list *ul = raw_cpu_ptr(&rt6_uncached_list);
140
141 rt->rt6i_uncached_list = ul;
142
143 spin_lock_bh(&ul->lock);
144 list_add_tail(&rt->rt6i_uncached, &ul->head);
145 spin_unlock_bh(&ul->lock);
146}
147
148void rt6_uncached_list_del(struct rt6_info *rt)
149{
150 if (!list_empty(&rt->rt6i_uncached)) {
151 struct uncached_list *ul = rt->rt6i_uncached_list;
152 struct net *net = dev_net(rt->dst.dev);
153
154 spin_lock_bh(&ul->lock);
155 list_del(&rt->rt6i_uncached);
156 atomic_dec(&net->ipv6.rt6_stats->fib_rt_uncache);
157 spin_unlock_bh(&ul->lock);
158 }
159}
160
161static void rt6_uncached_list_flush_dev(struct net *net, struct net_device *dev)
162{
163 struct net_device *loopback_dev = net->loopback_dev;
164 int cpu;
165
166 if (dev == loopback_dev)
167 return;
168
169 for_each_possible_cpu(cpu) {
170 struct uncached_list *ul = per_cpu_ptr(&rt6_uncached_list, cpu);
171 struct rt6_info *rt;
172
173 spin_lock_bh(&ul->lock);
174 list_for_each_entry(rt, &ul->head, rt6i_uncached) {
175 struct inet6_dev *rt_idev = rt->rt6i_idev;
176 struct net_device *rt_dev = rt->dst.dev;
177
178 if (rt_idev->dev == dev) {
179 rt->rt6i_idev = in6_dev_get(loopback_dev);
180 in6_dev_put(rt_idev);
181 }
182
183 if (rt_dev == dev) {
184 rt->dst.dev = blackhole_netdev;
185 dev_hold(rt->dst.dev);
186 dev_put(rt_dev);
187 }
188 }
189 spin_unlock_bh(&ul->lock);
190 }
191}
192
193static inline const void *choose_neigh_daddr(const struct in6_addr *p,
194 struct sk_buff *skb,
195 const void *daddr)
196{
197 if (!ipv6_addr_any(p))
198 return (const void *) p;
199 else if (skb)
200 return &ipv6_hdr(skb)->daddr;
201 return daddr;
202}
203
204struct neighbour *ip6_neigh_lookup(const struct in6_addr *gw,
205 struct net_device *dev,
206 struct sk_buff *skb,
207 const void *daddr)
208{
209 struct neighbour *n;
210
211 daddr = choose_neigh_daddr(gw, skb, daddr);
212 n = __ipv6_neigh_lookup(dev, daddr);
213 if (n)
214 return n;
215
216 n = neigh_create(&nd_tbl, daddr, dev);
217 return IS_ERR(n) ? NULL : n;
218}
219
220static struct neighbour *ip6_dst_neigh_lookup(const struct dst_entry *dst,
221 struct sk_buff *skb,
222 const void *daddr)
223{
224 const struct rt6_info *rt = container_of(dst, struct rt6_info, dst);
225
226 return ip6_neigh_lookup(rt6_nexthop(rt, &in6addr_any),
227 dst->dev, skb, daddr);
228}
229
230static void ip6_confirm_neigh(const struct dst_entry *dst, const void *daddr)
231{
232 struct net_device *dev = dst->dev;
233 struct rt6_info *rt = (struct rt6_info *)dst;
234
235 daddr = choose_neigh_daddr(rt6_nexthop(rt, &in6addr_any), NULL, daddr);
236 if (!daddr)
237 return;
238 if (dev->flags & (IFF_NOARP | IFF_LOOPBACK))
239 return;
240 if (ipv6_addr_is_multicast((const struct in6_addr *)daddr))
241 return;
242 __ipv6_confirm_neigh(dev, daddr);
243}
244
245static struct dst_ops ip6_dst_ops_template = {
246 .family = AF_INET6,
247 .gc = ip6_dst_gc,
248 .gc_thresh = 1024,
249 .check = ip6_dst_check,
250 .default_advmss = ip6_default_advmss,
251 .mtu = ip6_mtu,
252 .cow_metrics = dst_cow_metrics_generic,
253 .destroy = ip6_dst_destroy,
254 .ifdown = ip6_dst_ifdown,
255 .negative_advice = ip6_negative_advice,
256 .link_failure = ip6_link_failure,
257 .update_pmtu = ip6_rt_update_pmtu,
258 .redirect = rt6_do_redirect,
259 .local_out = __ip6_local_out,
260 .neigh_lookup = ip6_dst_neigh_lookup,
261 .confirm_neigh = ip6_confirm_neigh,
262};
263
264static struct dst_ops ip6_dst_blackhole_ops = {
265 .family = AF_INET6,
266 .default_advmss = ip6_default_advmss,
267 .neigh_lookup = ip6_dst_neigh_lookup,
268 .check = ip6_dst_check,
269 .destroy = ip6_dst_destroy,
270 .cow_metrics = dst_cow_metrics_generic,
271 .update_pmtu = dst_blackhole_update_pmtu,
272 .redirect = dst_blackhole_redirect,
273 .mtu = dst_blackhole_mtu,
274};
275
276static const u32 ip6_template_metrics[RTAX_MAX] = {
277 [RTAX_HOPLIMIT - 1] = 0,
278};
279
280static const struct fib6_info fib6_null_entry_template = {
281 .fib6_flags = (RTF_REJECT | RTF_NONEXTHOP),
282 .fib6_protocol = RTPROT_KERNEL,
283 .fib6_metric = ~(u32)0,
284 .fib6_ref = REFCOUNT_INIT(1),
285 .fib6_type = RTN_UNREACHABLE,
286 .fib6_metrics = (struct dst_metrics *)&dst_default_metrics,
287};
288
289static const struct rt6_info ip6_null_entry_template = {
290 .dst = {
291 .__refcnt = ATOMIC_INIT(1),
292 .__use = 1,
293 .obsolete = DST_OBSOLETE_FORCE_CHK,
294 .error = -ENETUNREACH,
295 .input = ip6_pkt_discard,
296 .output = ip6_pkt_discard_out,
297 },
298 .rt6i_flags = (RTF_REJECT | RTF_NONEXTHOP),
299};
300
301#ifdef CONFIG_IPV6_MULTIPLE_TABLES
302
303static const struct rt6_info ip6_prohibit_entry_template = {
304 .dst = {
305 .__refcnt = ATOMIC_INIT(1),
306 .__use = 1,
307 .obsolete = DST_OBSOLETE_FORCE_CHK,
308 .error = -EACCES,
309 .input = ip6_pkt_prohibit,
310 .output = ip6_pkt_prohibit_out,
311 },
312 .rt6i_flags = (RTF_REJECT | RTF_NONEXTHOP),
313};
314
315static const struct rt6_info ip6_blk_hole_entry_template = {
316 .dst = {
317 .__refcnt = ATOMIC_INIT(1),
318 .__use = 1,
319 .obsolete = DST_OBSOLETE_FORCE_CHK,
320 .error = -EINVAL,
321 .input = dst_discard,
322 .output = dst_discard_out,
323 },
324 .rt6i_flags = (RTF_REJECT | RTF_NONEXTHOP),
325};
326
327#endif
328
329static void rt6_info_init(struct rt6_info *rt)
330{
331 struct dst_entry *dst = &rt->dst;
332
333 memset(dst + 1, 0, sizeof(*rt) - sizeof(*dst));
334 INIT_LIST_HEAD(&rt->rt6i_uncached);
335}
336
337/* allocate dst with ip6_dst_ops */
338struct rt6_info *ip6_dst_alloc(struct net *net, struct net_device *dev,
339 int flags)
340{
341 struct rt6_info *rt = dst_alloc(&net->ipv6.ip6_dst_ops, dev,
342 1, DST_OBSOLETE_FORCE_CHK, flags);
343
344 if (rt) {
345 rt6_info_init(rt);
346 atomic_inc(&net->ipv6.rt6_stats->fib_rt_alloc);
347 }
348
349 return rt;
350}
351EXPORT_SYMBOL(ip6_dst_alloc);
352
353static void ip6_dst_destroy(struct dst_entry *dst)
354{
355 struct rt6_info *rt = (struct rt6_info *)dst;
356 struct fib6_info *from;
357 struct inet6_dev *idev;
358
359 ip_dst_metrics_put(dst);
360 rt6_uncached_list_del(rt);
361
362 idev = rt->rt6i_idev;
363 if (idev) {
364 rt->rt6i_idev = NULL;
365 in6_dev_put(idev);
366 }
367
368 from = xchg((__force struct fib6_info **)&rt->from, NULL);
369 fib6_info_release(from);
370}
371
372static void ip6_dst_ifdown(struct dst_entry *dst, struct net_device *dev,
373 int how)
374{
375 struct rt6_info *rt = (struct rt6_info *)dst;
376 struct inet6_dev *idev = rt->rt6i_idev;
377 struct net_device *loopback_dev =
378 dev_net(dev)->loopback_dev;
379
380 if (idev && idev->dev != loopback_dev) {
381 struct inet6_dev *loopback_idev = in6_dev_get(loopback_dev);
382 if (loopback_idev) {
383 rt->rt6i_idev = loopback_idev;
384 in6_dev_put(idev);
385 }
386 }
387}
388
389static bool __rt6_check_expired(const struct rt6_info *rt)
390{
391 if (rt->rt6i_flags & RTF_EXPIRES)
392 return time_after(jiffies, rt->dst.expires);
393 else
394 return false;
395}
396
397static bool rt6_check_expired(const struct rt6_info *rt)
398{
399 struct fib6_info *from;
400
401 from = rcu_dereference(rt->from);
402
403 if (rt->rt6i_flags & RTF_EXPIRES) {
404 if (time_after(jiffies, rt->dst.expires))
405 return true;
406 } else if (from) {
407 return rt->dst.obsolete != DST_OBSOLETE_FORCE_CHK ||
408 fib6_check_expired(from);
409 }
410 return false;
411}
412
413void fib6_select_path(const struct net *net, struct fib6_result *res,
414 struct flowi6 *fl6, int oif, bool have_oif_match,
415 const struct sk_buff *skb, int strict)
416{
417 struct fib6_info *sibling, *next_sibling;
418 struct fib6_info *match = res->f6i;
419
420 if (!match->nh && (!match->fib6_nsiblings || have_oif_match))
421 goto out;
422
423 if (match->nh && have_oif_match && res->nh)
424 return;
425
426 /* We might have already computed the hash for ICMPv6 errors. In such
427 * case it will always be non-zero. Otherwise now is the time to do it.
428 */
429 if (!fl6->mp_hash &&
430 (!match->nh || nexthop_is_multipath(match->nh)))
431 fl6->mp_hash = rt6_multipath_hash(net, fl6, skb, NULL);
432
433 if (unlikely(match->nh)) {
434 nexthop_path_fib6_result(res, fl6->mp_hash);
435 return;
436 }
437
438 if (fl6->mp_hash <= atomic_read(&match->fib6_nh->fib_nh_upper_bound))
439 goto out;
440
441 list_for_each_entry_safe(sibling, next_sibling, &match->fib6_siblings,
442 fib6_siblings) {
443 const struct fib6_nh *nh = sibling->fib6_nh;
444 int nh_upper_bound;
445
446 nh_upper_bound = atomic_read(&nh->fib_nh_upper_bound);
447 if (fl6->mp_hash > nh_upper_bound)
448 continue;
449 if (rt6_score_route(nh, sibling->fib6_flags, oif, strict) < 0)
450 break;
451 match = sibling;
452 break;
453 }
454
455out:
456 res->f6i = match;
457 res->nh = match->fib6_nh;
458}
459
460/*
461 * Route lookup. rcu_read_lock() should be held.
462 */
463
464static bool __rt6_device_match(struct net *net, const struct fib6_nh *nh,
465 const struct in6_addr *saddr, int oif, int flags)
466{
467 const struct net_device *dev;
468
469 if (nh->fib_nh_flags & RTNH_F_DEAD)
470 return false;
471
472 dev = nh->fib_nh_dev;
473 if (oif) {
474 if (dev->ifindex == oif)
475 return true;
476 } else {
477 if (ipv6_chk_addr(net, saddr, dev,
478 flags & RT6_LOOKUP_F_IFACE))
479 return true;
480 }
481
482 return false;
483}
484
485struct fib6_nh_dm_arg {
486 struct net *net;
487 const struct in6_addr *saddr;
488 int oif;
489 int flags;
490 struct fib6_nh *nh;
491};
492
493static int __rt6_nh_dev_match(struct fib6_nh *nh, void *_arg)
494{
495 struct fib6_nh_dm_arg *arg = _arg;
496
497 arg->nh = nh;
498 return __rt6_device_match(arg->net, nh, arg->saddr, arg->oif,
499 arg->flags);
500}
501
502/* returns fib6_nh from nexthop or NULL */
503static struct fib6_nh *rt6_nh_dev_match(struct net *net, struct nexthop *nh,
504 struct fib6_result *res,
505 const struct in6_addr *saddr,
506 int oif, int flags)
507{
508 struct fib6_nh_dm_arg arg = {
509 .net = net,
510 .saddr = saddr,
511 .oif = oif,
512 .flags = flags,
513 };
514
515 if (nexthop_is_blackhole(nh))
516 return NULL;
517
518 if (nexthop_for_each_fib6_nh(nh, __rt6_nh_dev_match, &arg))
519 return arg.nh;
520
521 return NULL;
522}
523
524static void rt6_device_match(struct net *net, struct fib6_result *res,
525 const struct in6_addr *saddr, int oif, int flags)
526{
527 struct fib6_info *f6i = res->f6i;
528 struct fib6_info *spf6i;
529 struct fib6_nh *nh;
530
531 if (!oif && ipv6_addr_any(saddr)) {
532 if (unlikely(f6i->nh)) {
533 nh = nexthop_fib6_nh(f6i->nh);
534 if (nexthop_is_blackhole(f6i->nh))
535 goto out_blackhole;
536 } else {
537 nh = f6i->fib6_nh;
538 }
539 if (!(nh->fib_nh_flags & RTNH_F_DEAD))
540 goto out;
541 }
542
543 for (spf6i = f6i; spf6i; spf6i = rcu_dereference(spf6i->fib6_next)) {
544 bool matched = false;
545
546 if (unlikely(spf6i->nh)) {
547 nh = rt6_nh_dev_match(net, spf6i->nh, res, saddr,
548 oif, flags);
549 if (nh)
550 matched = true;
551 } else {
552 nh = spf6i->fib6_nh;
553 if (__rt6_device_match(net, nh, saddr, oif, flags))
554 matched = true;
555 }
556 if (matched) {
557 res->f6i = spf6i;
558 goto out;
559 }
560 }
561
562 if (oif && flags & RT6_LOOKUP_F_IFACE) {
563 res->f6i = net->ipv6.fib6_null_entry;
564 nh = res->f6i->fib6_nh;
565 goto out;
566 }
567
568 if (unlikely(f6i->nh)) {
569 nh = nexthop_fib6_nh(f6i->nh);
570 if (nexthop_is_blackhole(f6i->nh))
571 goto out_blackhole;
572 } else {
573 nh = f6i->fib6_nh;
574 }
575
576 if (nh->fib_nh_flags & RTNH_F_DEAD) {
577 res->f6i = net->ipv6.fib6_null_entry;
578 nh = res->f6i->fib6_nh;
579 }
580out:
581 res->nh = nh;
582 res->fib6_type = res->f6i->fib6_type;
583 res->fib6_flags = res->f6i->fib6_flags;
584 return;
585
586out_blackhole:
587 res->fib6_flags |= RTF_REJECT;
588 res->fib6_type = RTN_BLACKHOLE;
589 res->nh = nh;
590}
591
592#ifdef CONFIG_IPV6_ROUTER_PREF
593struct __rt6_probe_work {
594 struct work_struct work;
595 struct in6_addr target;
596 struct net_device *dev;
597};
598
599static void rt6_probe_deferred(struct work_struct *w)
600{
601 struct in6_addr mcaddr;
602 struct __rt6_probe_work *work =
603 container_of(w, struct __rt6_probe_work, work);
604
605 addrconf_addr_solict_mult(&work->target, &mcaddr);
606 ndisc_send_ns(work->dev, &work->target, &mcaddr, NULL, 0);
607 dev_put(work->dev);
608 kfree(work);
609}
610
611static void rt6_probe(struct fib6_nh *fib6_nh)
612{
613 struct __rt6_probe_work *work = NULL;
614 const struct in6_addr *nh_gw;
615 unsigned long last_probe;
616 struct neighbour *neigh;
617 struct net_device *dev;
618 struct inet6_dev *idev;
619
620 /*
621 * Okay, this does not seem to be appropriate
622 * for now, however, we need to check if it
623 * is really so; aka Router Reachability Probing.
624 *
625 * Router Reachability Probe MUST be rate-limited
626 * to no more than one per minute.
627 */
628 if (!fib6_nh->fib_nh_gw_family)
629 return;
630
631 nh_gw = &fib6_nh->fib_nh_gw6;
632 dev = fib6_nh->fib_nh_dev;
633 rcu_read_lock_bh();
634 last_probe = READ_ONCE(fib6_nh->last_probe);
635 idev = __in6_dev_get(dev);
636 neigh = __ipv6_neigh_lookup_noref(dev, nh_gw);
637 if (neigh) {
638 if (neigh->nud_state & NUD_VALID)
639 goto out;
640
641 write_lock(&neigh->lock);
642 if (!(neigh->nud_state & NUD_VALID) &&
643 time_after(jiffies,
644 neigh->updated + idev->cnf.rtr_probe_interval)) {
645 work = kmalloc(sizeof(*work), GFP_ATOMIC);
646 if (work)
647 __neigh_set_probe_once(neigh);
648 }
649 write_unlock(&neigh->lock);
650 } else if (time_after(jiffies, last_probe +
651 idev->cnf.rtr_probe_interval)) {
652 work = kmalloc(sizeof(*work), GFP_ATOMIC);
653 }
654
655 if (!work || cmpxchg(&fib6_nh->last_probe,
656 last_probe, jiffies) != last_probe) {
657 kfree(work);
658 } else {
659 INIT_WORK(&work->work, rt6_probe_deferred);
660 work->target = *nh_gw;
661 dev_hold(dev);
662 work->dev = dev;
663 schedule_work(&work->work);
664 }
665
666out:
667 rcu_read_unlock_bh();
668}
669#else
670static inline void rt6_probe(struct fib6_nh *fib6_nh)
671{
672}
673#endif
674
675/*
676 * Default Router Selection (RFC 2461 6.3.6)
677 */
678static enum rt6_nud_state rt6_check_neigh(const struct fib6_nh *fib6_nh)
679{
680 enum rt6_nud_state ret = RT6_NUD_FAIL_HARD;
681 struct neighbour *neigh;
682
683 rcu_read_lock_bh();
684 neigh = __ipv6_neigh_lookup_noref(fib6_nh->fib_nh_dev,
685 &fib6_nh->fib_nh_gw6);
686 if (neigh) {
687 read_lock(&neigh->lock);
688 if (neigh->nud_state & NUD_VALID)
689 ret = RT6_NUD_SUCCEED;
690#ifdef CONFIG_IPV6_ROUTER_PREF
691 else if (!(neigh->nud_state & NUD_FAILED))
692 ret = RT6_NUD_SUCCEED;
693 else
694 ret = RT6_NUD_FAIL_PROBE;
695#endif
696 read_unlock(&neigh->lock);
697 } else {
698 ret = IS_ENABLED(CONFIG_IPV6_ROUTER_PREF) ?
699 RT6_NUD_SUCCEED : RT6_NUD_FAIL_DO_RR;
700 }
701 rcu_read_unlock_bh();
702
703 return ret;
704}
705
706static int rt6_score_route(const struct fib6_nh *nh, u32 fib6_flags, int oif,
707 int strict)
708{
709 int m = 0;
710
711 if (!oif || nh->fib_nh_dev->ifindex == oif)
712 m = 2;
713
714 if (!m && (strict & RT6_LOOKUP_F_IFACE))
715 return RT6_NUD_FAIL_HARD;
716#ifdef CONFIG_IPV6_ROUTER_PREF
717 m |= IPV6_DECODE_PREF(IPV6_EXTRACT_PREF(fib6_flags)) << 2;
718#endif
719 if ((strict & RT6_LOOKUP_F_REACHABLE) &&
720 !(fib6_flags & RTF_NONEXTHOP) && nh->fib_nh_gw_family) {
721 int n = rt6_check_neigh(nh);
722 if (n < 0)
723 return n;
724 }
725 return m;
726}
727
728static bool find_match(struct fib6_nh *nh, u32 fib6_flags,
729 int oif, int strict, int *mpri, bool *do_rr)
730{
731 bool match_do_rr = false;
732 bool rc = false;
733 int m;
734
735 if (nh->fib_nh_flags & RTNH_F_DEAD)
736 goto out;
737
738 if (ip6_ignore_linkdown(nh->fib_nh_dev) &&
739 nh->fib_nh_flags & RTNH_F_LINKDOWN &&
740 !(strict & RT6_LOOKUP_F_IGNORE_LINKSTATE))
741 goto out;
742
743 m = rt6_score_route(nh, fib6_flags, oif, strict);
744 if (m == RT6_NUD_FAIL_DO_RR) {
745 match_do_rr = true;
746 m = 0; /* lowest valid score */
747 } else if (m == RT6_NUD_FAIL_HARD) {
748 goto out;
749 }
750
751 if (strict & RT6_LOOKUP_F_REACHABLE)
752 rt6_probe(nh);
753
754 /* note that m can be RT6_NUD_FAIL_PROBE at this point */
755 if (m > *mpri) {
756 *do_rr = match_do_rr;
757 *mpri = m;
758 rc = true;
759 }
760out:
761 return rc;
762}
763
764struct fib6_nh_frl_arg {
765 u32 flags;
766 int oif;
767 int strict;
768 int *mpri;
769 bool *do_rr;
770 struct fib6_nh *nh;
771};
772
773static int rt6_nh_find_match(struct fib6_nh *nh, void *_arg)
774{
775 struct fib6_nh_frl_arg *arg = _arg;
776
777 arg->nh = nh;
778 return find_match(nh, arg->flags, arg->oif, arg->strict,
779 arg->mpri, arg->do_rr);
780}
781
782static void __find_rr_leaf(struct fib6_info *f6i_start,
783 struct fib6_info *nomatch, u32 metric,
784 struct fib6_result *res, struct fib6_info **cont,
785 int oif, int strict, bool *do_rr, int *mpri)
786{
787 struct fib6_info *f6i;
788
789 for (f6i = f6i_start;
790 f6i && f6i != nomatch;
791 f6i = rcu_dereference(f6i->fib6_next)) {
792 bool matched = false;
793 struct fib6_nh *nh;
794
795 if (cont && f6i->fib6_metric != metric) {
796 *cont = f6i;
797 return;
798 }
799
800 if (fib6_check_expired(f6i))
801 continue;
802
803 if (unlikely(f6i->nh)) {
804 struct fib6_nh_frl_arg arg = {
805 .flags = f6i->fib6_flags,
806 .oif = oif,
807 .strict = strict,
808 .mpri = mpri,
809 .do_rr = do_rr
810 };
811
812 if (nexthop_is_blackhole(f6i->nh)) {
813 res->fib6_flags = RTF_REJECT;
814 res->fib6_type = RTN_BLACKHOLE;
815 res->f6i = f6i;
816 res->nh = nexthop_fib6_nh(f6i->nh);
817 return;
818 }
819 if (nexthop_for_each_fib6_nh(f6i->nh, rt6_nh_find_match,
820 &arg)) {
821 matched = true;
822 nh = arg.nh;
823 }
824 } else {
825 nh = f6i->fib6_nh;
826 if (find_match(nh, f6i->fib6_flags, oif, strict,
827 mpri, do_rr))
828 matched = true;
829 }
830 if (matched) {
831 res->f6i = f6i;
832 res->nh = nh;
833 res->fib6_flags = f6i->fib6_flags;
834 res->fib6_type = f6i->fib6_type;
835 }
836 }
837}
838
839static void find_rr_leaf(struct fib6_node *fn, struct fib6_info *leaf,
840 struct fib6_info *rr_head, int oif, int strict,
841 bool *do_rr, struct fib6_result *res)
842{
843 u32 metric = rr_head->fib6_metric;
844 struct fib6_info *cont = NULL;
845 int mpri = -1;
846
847 __find_rr_leaf(rr_head, NULL, metric, res, &cont,
848 oif, strict, do_rr, &mpri);
849
850 __find_rr_leaf(leaf, rr_head, metric, res, &cont,
851 oif, strict, do_rr, &mpri);
852
853 if (res->f6i || !cont)
854 return;
855
856 __find_rr_leaf(cont, NULL, metric, res, NULL,
857 oif, strict, do_rr, &mpri);
858}
859
860static void rt6_select(struct net *net, struct fib6_node *fn, int oif,
861 struct fib6_result *res, int strict)
862{
863 struct fib6_info *leaf = rcu_dereference(fn->leaf);
864 struct fib6_info *rt0;
865 bool do_rr = false;
866 int key_plen;
867
868 /* make sure this function or its helpers sets f6i */
869 res->f6i = NULL;
870
871 if (!leaf || leaf == net->ipv6.fib6_null_entry)
872 goto out;
873
874 rt0 = rcu_dereference(fn->rr_ptr);
875 if (!rt0)
876 rt0 = leaf;
877
878 /* Double check to make sure fn is not an intermediate node
879 * and fn->leaf does not points to its child's leaf
880 * (This might happen if all routes under fn are deleted from
881 * the tree and fib6_repair_tree() is called on the node.)
882 */
883 key_plen = rt0->fib6_dst.plen;
884#ifdef CONFIG_IPV6_SUBTREES
885 if (rt0->fib6_src.plen)
886 key_plen = rt0->fib6_src.plen;
887#endif
888 if (fn->fn_bit != key_plen)
889 goto out;
890
891 find_rr_leaf(fn, leaf, rt0, oif, strict, &do_rr, res);
892 if (do_rr) {
893 struct fib6_info *next = rcu_dereference(rt0->fib6_next);
894
895 /* no entries matched; do round-robin */
896 if (!next || next->fib6_metric != rt0->fib6_metric)
897 next = leaf;
898
899 if (next != rt0) {
900 spin_lock_bh(&leaf->fib6_table->tb6_lock);
901 /* make sure next is not being deleted from the tree */
902 if (next->fib6_node)
903 rcu_assign_pointer(fn->rr_ptr, next);
904 spin_unlock_bh(&leaf->fib6_table->tb6_lock);
905 }
906 }
907
908out:
909 if (!res->f6i) {
910 res->f6i = net->ipv6.fib6_null_entry;
911 res->nh = res->f6i->fib6_nh;
912 res->fib6_flags = res->f6i->fib6_flags;
913 res->fib6_type = res->f6i->fib6_type;
914 }
915}
916
917static bool rt6_is_gw_or_nonexthop(const struct fib6_result *res)
918{
919 return (res->f6i->fib6_flags & RTF_NONEXTHOP) ||
920 res->nh->fib_nh_gw_family;
921}
922
923#ifdef CONFIG_IPV6_ROUTE_INFO
924int rt6_route_rcv(struct net_device *dev, u8 *opt, int len,
925 const struct in6_addr *gwaddr)
926{
927 struct net *net = dev_net(dev);
928 struct route_info *rinfo = (struct route_info *) opt;
929 struct in6_addr prefix_buf, *prefix;
930 unsigned int pref;
931 unsigned long lifetime;
932 struct fib6_info *rt;
933
934 if (len < sizeof(struct route_info)) {
935 return -EINVAL;
936 }
937
938 /* Sanity check for prefix_len and length */
939 if (rinfo->length > 3) {
940 return -EINVAL;
941 } else if (rinfo->prefix_len > 128) {
942 return -EINVAL;
943 } else if (rinfo->prefix_len > 64) {
944 if (rinfo->length < 2) {
945 return -EINVAL;
946 }
947 } else if (rinfo->prefix_len > 0) {
948 if (rinfo->length < 1) {
949 return -EINVAL;
950 }
951 }
952
953 pref = rinfo->route_pref;
954 if (pref == ICMPV6_ROUTER_PREF_INVALID)
955 return -EINVAL;
956
957 lifetime = addrconf_timeout_fixup(ntohl(rinfo->lifetime), HZ);
958
959 if (rinfo->length == 3)
960 prefix = (struct in6_addr *)rinfo->prefix;
961 else {
962 /* this function is safe */
963 ipv6_addr_prefix(&prefix_buf,
964 (struct in6_addr *)rinfo->prefix,
965 rinfo->prefix_len);
966 prefix = &prefix_buf;
967 }
968
969 if (rinfo->prefix_len == 0)
970 rt = rt6_get_dflt_router(net, gwaddr, dev);
971 else
972 rt = rt6_get_route_info(net, prefix, rinfo->prefix_len,
973 gwaddr, dev);
974
975 if (rt && !lifetime) {
976 ip6_del_rt(net, rt, false);
977 rt = NULL;
978 }
979
980 if (!rt && lifetime)
981 rt = rt6_add_route_info(net, prefix, rinfo->prefix_len, gwaddr,
982 dev, pref);
983 else if (rt)
984 rt->fib6_flags = RTF_ROUTEINFO |
985 (rt->fib6_flags & ~RTF_PREF_MASK) | RTF_PREF(pref);
986
987 if (rt) {
988 if (!addrconf_finite_timeout(lifetime))
989 fib6_clean_expires(rt);
990 else
991 fib6_set_expires(rt, jiffies + HZ * lifetime);
992
993 fib6_info_release(rt);
994 }
995 return 0;
996}
997#endif
998
999/*
1000 * Misc support functions
1001 */
1002
1003/* called with rcu_lock held */
1004static struct net_device *ip6_rt_get_dev_rcu(const struct fib6_result *res)
1005{
1006 struct net_device *dev = res->nh->fib_nh_dev;
1007
1008 if (res->fib6_flags & (RTF_LOCAL | RTF_ANYCAST)) {
1009 /* for copies of local routes, dst->dev needs to be the
1010 * device if it is a master device, the master device if
1011 * device is enslaved, and the loopback as the default
1012 */
1013 if (netif_is_l3_slave(dev) &&
1014 !rt6_need_strict(&res->f6i->fib6_dst.addr))
1015 dev = l3mdev_master_dev_rcu(dev);
1016 else if (!netif_is_l3_master(dev))
1017 dev = dev_net(dev)->loopback_dev;
1018 /* last case is netif_is_l3_master(dev) is true in which
1019 * case we want dev returned to be dev
1020 */
1021 }
1022
1023 return dev;
1024}
1025
1026static const int fib6_prop[RTN_MAX + 1] = {
1027 [RTN_UNSPEC] = 0,
1028 [RTN_UNICAST] = 0,
1029 [RTN_LOCAL] = 0,
1030 [RTN_BROADCAST] = 0,
1031 [RTN_ANYCAST] = 0,
1032 [RTN_MULTICAST] = 0,
1033 [RTN_BLACKHOLE] = -EINVAL,
1034 [RTN_UNREACHABLE] = -EHOSTUNREACH,
1035 [RTN_PROHIBIT] = -EACCES,
1036 [RTN_THROW] = -EAGAIN,
1037 [RTN_NAT] = -EINVAL,
1038 [RTN_XRESOLVE] = -EINVAL,
1039};
1040
1041static int ip6_rt_type_to_error(u8 fib6_type)
1042{
1043 return fib6_prop[fib6_type];
1044}
1045
1046static unsigned short fib6_info_dst_flags(struct fib6_info *rt)
1047{
1048 unsigned short flags = 0;
1049
1050 if (rt->dst_nocount)
1051 flags |= DST_NOCOUNT;
1052 if (rt->dst_nopolicy)
1053 flags |= DST_NOPOLICY;
1054
1055 return flags;
1056}
1057
1058static void ip6_rt_init_dst_reject(struct rt6_info *rt, u8 fib6_type)
1059{
1060 rt->dst.error = ip6_rt_type_to_error(fib6_type);
1061
1062 switch (fib6_type) {
1063 case RTN_BLACKHOLE:
1064 rt->dst.output = dst_discard_out;
1065 rt->dst.input = dst_discard;
1066 break;
1067 case RTN_PROHIBIT:
1068 rt->dst.output = ip6_pkt_prohibit_out;
1069 rt->dst.input = ip6_pkt_prohibit;
1070 break;
1071 case RTN_THROW:
1072 case RTN_UNREACHABLE:
1073 default:
1074 rt->dst.output = ip6_pkt_discard_out;
1075 rt->dst.input = ip6_pkt_discard;
1076 break;
1077 }
1078}
1079
1080static void ip6_rt_init_dst(struct rt6_info *rt, const struct fib6_result *res)
1081{
1082 struct fib6_info *f6i = res->f6i;
1083
1084 if (res->fib6_flags & RTF_REJECT) {
1085 ip6_rt_init_dst_reject(rt, res->fib6_type);
1086 return;
1087 }
1088
1089 rt->dst.error = 0;
1090 rt->dst.output = ip6_output;
1091
1092 if (res->fib6_type == RTN_LOCAL || res->fib6_type == RTN_ANYCAST) {
1093 rt->dst.input = ip6_input;
1094 } else if (ipv6_addr_type(&f6i->fib6_dst.addr) & IPV6_ADDR_MULTICAST) {
1095 rt->dst.input = ip6_mc_input;
1096 } else {
1097 rt->dst.input = ip6_forward;
1098 }
1099
1100 if (res->nh->fib_nh_lws) {
1101 rt->dst.lwtstate = lwtstate_get(res->nh->fib_nh_lws);
1102 lwtunnel_set_redirect(&rt->dst);
1103 }
1104
1105 rt->dst.lastuse = jiffies;
1106}
1107
1108/* Caller must already hold reference to @from */
1109static void rt6_set_from(struct rt6_info *rt, struct fib6_info *from)
1110{
1111 rt->rt6i_flags &= ~RTF_EXPIRES;
1112 rcu_assign_pointer(rt->from, from);
1113 ip_dst_init_metrics(&rt->dst, from->fib6_metrics);
1114}
1115
1116/* Caller must already hold reference to f6i in result */
1117static void ip6_rt_copy_init(struct rt6_info *rt, const struct fib6_result *res)
1118{
1119 const struct fib6_nh *nh = res->nh;
1120 const struct net_device *dev = nh->fib_nh_dev;
1121 struct fib6_info *f6i = res->f6i;
1122
1123 ip6_rt_init_dst(rt, res);
1124
1125 rt->rt6i_dst = f6i->fib6_dst;
1126 rt->rt6i_idev = dev ? in6_dev_get(dev) : NULL;
1127 rt->rt6i_flags = res->fib6_flags;
1128 if (nh->fib_nh_gw_family) {
1129 rt->rt6i_gateway = nh->fib_nh_gw6;
1130 rt->rt6i_flags |= RTF_GATEWAY;
1131 }
1132 rt6_set_from(rt, f6i);
1133#ifdef CONFIG_IPV6_SUBTREES
1134 rt->rt6i_src = f6i->fib6_src;
1135#endif
1136}
1137
1138static struct fib6_node* fib6_backtrack(struct fib6_node *fn,
1139 struct in6_addr *saddr)
1140{
1141 struct fib6_node *pn, *sn;
1142 while (1) {
1143 if (fn->fn_flags & RTN_TL_ROOT)
1144 return NULL;
1145 pn = rcu_dereference(fn->parent);
1146 sn = FIB6_SUBTREE(pn);
1147 if (sn && sn != fn)
1148 fn = fib6_node_lookup(sn, NULL, saddr);
1149 else
1150 fn = pn;
1151 if (fn->fn_flags & RTN_RTINFO)
1152 return fn;
1153 }
1154}
1155
1156static bool ip6_hold_safe(struct net *net, struct rt6_info **prt)
1157{
1158 struct rt6_info *rt = *prt;
1159
1160 if (dst_hold_safe(&rt->dst))
1161 return true;
1162 if (net) {
1163 rt = net->ipv6.ip6_null_entry;
1164 dst_hold(&rt->dst);
1165 } else {
1166 rt = NULL;
1167 }
1168 *prt = rt;
1169 return false;
1170}
1171
1172/* called with rcu_lock held */
1173static struct rt6_info *ip6_create_rt_rcu(const struct fib6_result *res)
1174{
1175 struct net_device *dev = res->nh->fib_nh_dev;
1176 struct fib6_info *f6i = res->f6i;
1177 unsigned short flags;
1178 struct rt6_info *nrt;
1179
1180 if (!fib6_info_hold_safe(f6i))
1181 goto fallback;
1182
1183 flags = fib6_info_dst_flags(f6i);
1184 nrt = ip6_dst_alloc(dev_net(dev), dev, flags);
1185 if (!nrt) {
1186 fib6_info_release(f6i);
1187 goto fallback;
1188 }
1189
1190 ip6_rt_copy_init(nrt, res);
1191 return nrt;
1192
1193fallback:
1194 nrt = dev_net(dev)->ipv6.ip6_null_entry;
1195 dst_hold(&nrt->dst);
1196 return nrt;
1197}
1198
1199INDIRECT_CALLABLE_SCOPE struct rt6_info *ip6_pol_route_lookup(struct net *net,
1200 struct fib6_table *table,
1201 struct flowi6 *fl6,
1202 const struct sk_buff *skb,
1203 int flags)
1204{
1205 struct fib6_result res = {};
1206 struct fib6_node *fn;
1207 struct rt6_info *rt;
1208
1209 if (fl6->flowi6_flags & FLOWI_FLAG_SKIP_NH_OIF)
1210 flags &= ~RT6_LOOKUP_F_IFACE;
1211
1212 rcu_read_lock();
1213 fn = fib6_node_lookup(&table->tb6_root, &fl6->daddr, &fl6->saddr);
1214restart:
1215 res.f6i = rcu_dereference(fn->leaf);
1216 if (!res.f6i)
1217 res.f6i = net->ipv6.fib6_null_entry;
1218 else
1219 rt6_device_match(net, &res, &fl6->saddr, fl6->flowi6_oif,
1220 flags);
1221
1222 if (res.f6i == net->ipv6.fib6_null_entry) {
1223 fn = fib6_backtrack(fn, &fl6->saddr);
1224 if (fn)
1225 goto restart;
1226
1227 rt = net->ipv6.ip6_null_entry;
1228 dst_hold(&rt->dst);
1229 goto out;
1230 } else if (res.fib6_flags & RTF_REJECT) {
1231 goto do_create;
1232 }
1233
1234 fib6_select_path(net, &res, fl6, fl6->flowi6_oif,
1235 fl6->flowi6_oif != 0, skb, flags);
1236
1237 /* Search through exception table */
1238 rt = rt6_find_cached_rt(&res, &fl6->daddr, &fl6->saddr);
1239 if (rt) {
1240 if (ip6_hold_safe(net, &rt))
1241 dst_use_noref(&rt->dst, jiffies);
1242 } else {
1243do_create:
1244 rt = ip6_create_rt_rcu(&res);
1245 }
1246
1247out:
1248 trace_fib6_table_lookup(net, &res, table, fl6);
1249
1250 rcu_read_unlock();
1251
1252 return rt;
1253}
1254
1255struct dst_entry *ip6_route_lookup(struct net *net, struct flowi6 *fl6,
1256 const struct sk_buff *skb, int flags)
1257{
1258 return fib6_rule_lookup(net, fl6, skb, flags, ip6_pol_route_lookup);
1259}
1260EXPORT_SYMBOL_GPL(ip6_route_lookup);
1261
1262struct rt6_info *rt6_lookup(struct net *net, const struct in6_addr *daddr,
1263 const struct in6_addr *saddr, int oif,
1264 const struct sk_buff *skb, int strict)
1265{
1266 struct flowi6 fl6 = {
1267 .flowi6_oif = oif,
1268 .daddr = *daddr,
1269 };
1270 struct dst_entry *dst;
1271 int flags = strict ? RT6_LOOKUP_F_IFACE : 0;
1272
1273 if (saddr) {
1274 memcpy(&fl6.saddr, saddr, sizeof(*saddr));
1275 flags |= RT6_LOOKUP_F_HAS_SADDR;
1276 }
1277
1278 dst = fib6_rule_lookup(net, &fl6, skb, flags, ip6_pol_route_lookup);
1279 if (dst->error == 0)
1280 return (struct rt6_info *) dst;
1281
1282 dst_release(dst);
1283
1284 return NULL;
1285}
1286EXPORT_SYMBOL(rt6_lookup);
1287
1288/* ip6_ins_rt is called with FREE table->tb6_lock.
1289 * It takes new route entry, the addition fails by any reason the
1290 * route is released.
1291 * Caller must hold dst before calling it.
1292 */
1293
1294static int __ip6_ins_rt(struct fib6_info *rt, struct nl_info *info,
1295 struct netlink_ext_ack *extack)
1296{
1297 int err;
1298 struct fib6_table *table;
1299
1300 table = rt->fib6_table;
1301 spin_lock_bh(&table->tb6_lock);
1302 err = fib6_add(&table->tb6_root, rt, info, extack);
1303 spin_unlock_bh(&table->tb6_lock);
1304
1305 return err;
1306}
1307
1308int ip6_ins_rt(struct net *net, struct fib6_info *rt)
1309{
1310 struct nl_info info = { .nl_net = net, };
1311
1312 return __ip6_ins_rt(rt, &info, NULL);
1313}
1314
1315static struct rt6_info *ip6_rt_cache_alloc(const struct fib6_result *res,
1316 const struct in6_addr *daddr,
1317 const struct in6_addr *saddr)
1318{
1319 struct fib6_info *f6i = res->f6i;
1320 struct net_device *dev;
1321 struct rt6_info *rt;
1322
1323 /*
1324 * Clone the route.
1325 */
1326
1327 if (!fib6_info_hold_safe(f6i))
1328 return NULL;
1329
1330 dev = ip6_rt_get_dev_rcu(res);
1331 rt = ip6_dst_alloc(dev_net(dev), dev, 0);
1332 if (!rt) {
1333 fib6_info_release(f6i);
1334 return NULL;
1335 }
1336
1337 ip6_rt_copy_init(rt, res);
1338 rt->rt6i_flags |= RTF_CACHE;
1339 rt->rt6i_dst.addr = *daddr;
1340 rt->rt6i_dst.plen = 128;
1341
1342 if (!rt6_is_gw_or_nonexthop(res)) {
1343 if (f6i->fib6_dst.plen != 128 &&
1344 ipv6_addr_equal(&f6i->fib6_dst.addr, daddr))
1345 rt->rt6i_flags |= RTF_ANYCAST;
1346#ifdef CONFIG_IPV6_SUBTREES
1347 if (rt->rt6i_src.plen && saddr) {
1348 rt->rt6i_src.addr = *saddr;
1349 rt->rt6i_src.plen = 128;
1350 }
1351#endif
1352 }
1353
1354 return rt;
1355}
1356
1357static struct rt6_info *ip6_rt_pcpu_alloc(const struct fib6_result *res)
1358{
1359 struct fib6_info *f6i = res->f6i;
1360 unsigned short flags = fib6_info_dst_flags(f6i);
1361 struct net_device *dev;
1362 struct rt6_info *pcpu_rt;
1363
1364 if (!fib6_info_hold_safe(f6i))
1365 return NULL;
1366
1367 rcu_read_lock();
1368 dev = ip6_rt_get_dev_rcu(res);
1369 pcpu_rt = ip6_dst_alloc(dev_net(dev), dev, flags | DST_NOCOUNT);
1370 rcu_read_unlock();
1371 if (!pcpu_rt) {
1372 fib6_info_release(f6i);
1373 return NULL;
1374 }
1375 ip6_rt_copy_init(pcpu_rt, res);
1376 pcpu_rt->rt6i_flags |= RTF_PCPU;
1377
1378 if (f6i->nh)
1379 pcpu_rt->sernum = rt_genid_ipv6(dev_net(dev));
1380
1381 return pcpu_rt;
1382}
1383
1384static bool rt6_is_valid(const struct rt6_info *rt6)
1385{
1386 return rt6->sernum == rt_genid_ipv6(dev_net(rt6->dst.dev));
1387}
1388
1389/* It should be called with rcu_read_lock() acquired */
1390static struct rt6_info *rt6_get_pcpu_route(const struct fib6_result *res)
1391{
1392 struct rt6_info *pcpu_rt;
1393
1394 pcpu_rt = this_cpu_read(*res->nh->rt6i_pcpu);
1395
1396 if (pcpu_rt && pcpu_rt->sernum && !rt6_is_valid(pcpu_rt)) {
1397 struct rt6_info *prev, **p;
1398
1399 p = this_cpu_ptr(res->nh->rt6i_pcpu);
1400 prev = xchg(p, NULL);
1401 if (prev) {
1402 dst_dev_put(&prev->dst);
1403 dst_release(&prev->dst);
1404 }
1405
1406 pcpu_rt = NULL;
1407 }
1408
1409 return pcpu_rt;
1410}
1411
1412static struct rt6_info *rt6_make_pcpu_route(struct net *net,
1413 const struct fib6_result *res)
1414{
1415 struct rt6_info *pcpu_rt, *prev, **p;
1416
1417 pcpu_rt = ip6_rt_pcpu_alloc(res);
1418 if (!pcpu_rt)
1419 return NULL;
1420
1421 p = this_cpu_ptr(res->nh->rt6i_pcpu);
1422 prev = cmpxchg(p, NULL, pcpu_rt);
1423 BUG_ON(prev);
1424
1425 if (res->f6i->fib6_destroying) {
1426 struct fib6_info *from;
1427
1428 from = xchg((__force struct fib6_info **)&pcpu_rt->from, NULL);
1429 fib6_info_release(from);
1430 }
1431
1432 return pcpu_rt;
1433}
1434
1435/* exception hash table implementation
1436 */
1437static DEFINE_SPINLOCK(rt6_exception_lock);
1438
1439/* Remove rt6_ex from hash table and free the memory
1440 * Caller must hold rt6_exception_lock
1441 */
1442static void rt6_remove_exception(struct rt6_exception_bucket *bucket,
1443 struct rt6_exception *rt6_ex)
1444{
1445 struct fib6_info *from;
1446 struct net *net;
1447
1448 if (!bucket || !rt6_ex)
1449 return;
1450
1451 net = dev_net(rt6_ex->rt6i->dst.dev);
1452 net->ipv6.rt6_stats->fib_rt_cache--;
1453
1454 /* purge completely the exception to allow releasing the held resources:
1455 * some [sk] cache may keep the dst around for unlimited time
1456 */
1457 from = xchg((__force struct fib6_info **)&rt6_ex->rt6i->from, NULL);
1458 fib6_info_release(from);
1459 dst_dev_put(&rt6_ex->rt6i->dst);
1460
1461 hlist_del_rcu(&rt6_ex->hlist);
1462 dst_release(&rt6_ex->rt6i->dst);
1463 kfree_rcu(rt6_ex, rcu);
1464 WARN_ON_ONCE(!bucket->depth);
1465 bucket->depth--;
1466}
1467
1468/* Remove oldest rt6_ex in bucket and free the memory
1469 * Caller must hold rt6_exception_lock
1470 */
1471static void rt6_exception_remove_oldest(struct rt6_exception_bucket *bucket)
1472{
1473 struct rt6_exception *rt6_ex, *oldest = NULL;
1474
1475 if (!bucket)
1476 return;
1477
1478 hlist_for_each_entry(rt6_ex, &bucket->chain, hlist) {
1479 if (!oldest || time_before(rt6_ex->stamp, oldest->stamp))
1480 oldest = rt6_ex;
1481 }
1482 rt6_remove_exception(bucket, oldest);
1483}
1484
1485static u32 rt6_exception_hash(const struct in6_addr *dst,
1486 const struct in6_addr *src)
1487{
1488 static siphash_key_t rt6_exception_key __read_mostly;
1489 struct {
1490 struct in6_addr dst;
1491 struct in6_addr src;
1492 } __aligned(SIPHASH_ALIGNMENT) combined = {
1493 .dst = *dst,
1494 };
1495 u64 val;
1496
1497 net_get_random_once(&rt6_exception_key, sizeof(rt6_exception_key));
1498
1499#ifdef CONFIG_IPV6_SUBTREES
1500 if (src)
1501 combined.src = *src;
1502#endif
1503 val = siphash(&combined, sizeof(combined), &rt6_exception_key);
1504
1505 return hash_64(val, FIB6_EXCEPTION_BUCKET_SIZE_SHIFT);
1506}
1507
1508/* Helper function to find the cached rt in the hash table
1509 * and update bucket pointer to point to the bucket for this
1510 * (daddr, saddr) pair
1511 * Caller must hold rt6_exception_lock
1512 */
1513static struct rt6_exception *
1514__rt6_find_exception_spinlock(struct rt6_exception_bucket **bucket,
1515 const struct in6_addr *daddr,
1516 const struct in6_addr *saddr)
1517{
1518 struct rt6_exception *rt6_ex;
1519 u32 hval;
1520
1521 if (!(*bucket) || !daddr)
1522 return NULL;
1523
1524 hval = rt6_exception_hash(daddr, saddr);
1525 *bucket += hval;
1526
1527 hlist_for_each_entry(rt6_ex, &(*bucket)->chain, hlist) {
1528 struct rt6_info *rt6 = rt6_ex->rt6i;
1529 bool matched = ipv6_addr_equal(daddr, &rt6->rt6i_dst.addr);
1530
1531#ifdef CONFIG_IPV6_SUBTREES
1532 if (matched && saddr)
1533 matched = ipv6_addr_equal(saddr, &rt6->rt6i_src.addr);
1534#endif
1535 if (matched)
1536 return rt6_ex;
1537 }
1538 return NULL;
1539}
1540
1541/* Helper function to find the cached rt in the hash table
1542 * and update bucket pointer to point to the bucket for this
1543 * (daddr, saddr) pair
1544 * Caller must hold rcu_read_lock()
1545 */
1546static struct rt6_exception *
1547__rt6_find_exception_rcu(struct rt6_exception_bucket **bucket,
1548 const struct in6_addr *daddr,
1549 const struct in6_addr *saddr)
1550{
1551 struct rt6_exception *rt6_ex;
1552 u32 hval;
1553
1554 WARN_ON_ONCE(!rcu_read_lock_held());
1555
1556 if (!(*bucket) || !daddr)
1557 return NULL;
1558
1559 hval = rt6_exception_hash(daddr, saddr);
1560 *bucket += hval;
1561
1562 hlist_for_each_entry_rcu(rt6_ex, &(*bucket)->chain, hlist) {
1563 struct rt6_info *rt6 = rt6_ex->rt6i;
1564 bool matched = ipv6_addr_equal(daddr, &rt6->rt6i_dst.addr);
1565
1566#ifdef CONFIG_IPV6_SUBTREES
1567 if (matched && saddr)
1568 matched = ipv6_addr_equal(saddr, &rt6->rt6i_src.addr);
1569#endif
1570 if (matched)
1571 return rt6_ex;
1572 }
1573 return NULL;
1574}
1575
1576static unsigned int fib6_mtu(const struct fib6_result *res)
1577{
1578 const struct fib6_nh *nh = res->nh;
1579 unsigned int mtu;
1580
1581 if (res->f6i->fib6_pmtu) {
1582 mtu = res->f6i->fib6_pmtu;
1583 } else {
1584 struct net_device *dev = nh->fib_nh_dev;
1585 struct inet6_dev *idev;
1586
1587 rcu_read_lock();
1588 idev = __in6_dev_get(dev);
1589 mtu = idev->cnf.mtu6;
1590 rcu_read_unlock();
1591 }
1592
1593 mtu = min_t(unsigned int, mtu, IP6_MAX_MTU);
1594
1595 return mtu - lwtunnel_headroom(nh->fib_nh_lws, mtu);
1596}
1597
1598#define FIB6_EXCEPTION_BUCKET_FLUSHED 0x1UL
1599
1600/* used when the flushed bit is not relevant, only access to the bucket
1601 * (ie., all bucket users except rt6_insert_exception);
1602 *
1603 * called under rcu lock; sometimes called with rt6_exception_lock held
1604 */
1605static
1606struct rt6_exception_bucket *fib6_nh_get_excptn_bucket(const struct fib6_nh *nh,
1607 spinlock_t *lock)
1608{
1609 struct rt6_exception_bucket *bucket;
1610
1611 if (lock)
1612 bucket = rcu_dereference_protected(nh->rt6i_exception_bucket,
1613 lockdep_is_held(lock));
1614 else
1615 bucket = rcu_dereference(nh->rt6i_exception_bucket);
1616
1617 /* remove bucket flushed bit if set */
1618 if (bucket) {
1619 unsigned long p = (unsigned long)bucket;
1620
1621 p &= ~FIB6_EXCEPTION_BUCKET_FLUSHED;
1622 bucket = (struct rt6_exception_bucket *)p;
1623 }
1624
1625 return bucket;
1626}
1627
1628static bool fib6_nh_excptn_bucket_flushed(struct rt6_exception_bucket *bucket)
1629{
1630 unsigned long p = (unsigned long)bucket;
1631
1632 return !!(p & FIB6_EXCEPTION_BUCKET_FLUSHED);
1633}
1634
1635/* called with rt6_exception_lock held */
1636static void fib6_nh_excptn_bucket_set_flushed(struct fib6_nh *nh,
1637 spinlock_t *lock)
1638{
1639 struct rt6_exception_bucket *bucket;
1640 unsigned long p;
1641
1642 bucket = rcu_dereference_protected(nh->rt6i_exception_bucket,
1643 lockdep_is_held(lock));
1644
1645 p = (unsigned long)bucket;
1646 p |= FIB6_EXCEPTION_BUCKET_FLUSHED;
1647 bucket = (struct rt6_exception_bucket *)p;
1648 rcu_assign_pointer(nh->rt6i_exception_bucket, bucket);
1649}
1650
1651static int rt6_insert_exception(struct rt6_info *nrt,
1652 const struct fib6_result *res)
1653{
1654 struct net *net = dev_net(nrt->dst.dev);
1655 struct rt6_exception_bucket *bucket;
1656 struct fib6_info *f6i = res->f6i;
1657 struct in6_addr *src_key = NULL;
1658 struct rt6_exception *rt6_ex;
1659 struct fib6_nh *nh = res->nh;
1660 int max_depth;
1661 int err = 0;
1662
1663 spin_lock_bh(&rt6_exception_lock);
1664
1665 bucket = rcu_dereference_protected(nh->rt6i_exception_bucket,
1666 lockdep_is_held(&rt6_exception_lock));
1667 if (!bucket) {
1668 bucket = kcalloc(FIB6_EXCEPTION_BUCKET_SIZE, sizeof(*bucket),
1669 GFP_ATOMIC);
1670 if (!bucket) {
1671 err = -ENOMEM;
1672 goto out;
1673 }
1674 rcu_assign_pointer(nh->rt6i_exception_bucket, bucket);
1675 } else if (fib6_nh_excptn_bucket_flushed(bucket)) {
1676 err = -EINVAL;
1677 goto out;
1678 }
1679
1680#ifdef CONFIG_IPV6_SUBTREES
1681 /* fib6_src.plen != 0 indicates f6i is in subtree
1682 * and exception table is indexed by a hash of
1683 * both fib6_dst and fib6_src.
1684 * Otherwise, the exception table is indexed by
1685 * a hash of only fib6_dst.
1686 */
1687 if (f6i->fib6_src.plen)
1688 src_key = &nrt->rt6i_src.addr;
1689#endif
1690 /* rt6_mtu_change() might lower mtu on f6i.
1691 * Only insert this exception route if its mtu
1692 * is less than f6i's mtu value.
1693 */
1694 if (dst_metric_raw(&nrt->dst, RTAX_MTU) >= fib6_mtu(res)) {
1695 err = -EINVAL;
1696 goto out;
1697 }
1698
1699 rt6_ex = __rt6_find_exception_spinlock(&bucket, &nrt->rt6i_dst.addr,
1700 src_key);
1701 if (rt6_ex)
1702 rt6_remove_exception(bucket, rt6_ex);
1703
1704 rt6_ex = kzalloc(sizeof(*rt6_ex), GFP_ATOMIC);
1705 if (!rt6_ex) {
1706 err = -ENOMEM;
1707 goto out;
1708 }
1709 rt6_ex->rt6i = nrt;
1710 rt6_ex->stamp = jiffies;
1711 hlist_add_head_rcu(&rt6_ex->hlist, &bucket->chain);
1712 bucket->depth++;
1713 net->ipv6.rt6_stats->fib_rt_cache++;
1714
1715 /* Randomize max depth to avoid some side channels attacks. */
1716 max_depth = FIB6_MAX_DEPTH + prandom_u32_max(FIB6_MAX_DEPTH);
1717 while (bucket->depth > max_depth)
1718 rt6_exception_remove_oldest(bucket);
1719
1720out:
1721 spin_unlock_bh(&rt6_exception_lock);
1722
1723 /* Update fn->fn_sernum to invalidate all cached dst */
1724 if (!err) {
1725 spin_lock_bh(&f6i->fib6_table->tb6_lock);
1726 fib6_update_sernum(net, f6i);
1727 spin_unlock_bh(&f6i->fib6_table->tb6_lock);
1728 fib6_force_start_gc(net);
1729 }
1730
1731 return err;
1732}
1733
1734static void fib6_nh_flush_exceptions(struct fib6_nh *nh, struct fib6_info *from)
1735{
1736 struct rt6_exception_bucket *bucket;
1737 struct rt6_exception *rt6_ex;
1738 struct hlist_node *tmp;
1739 int i;
1740
1741 spin_lock_bh(&rt6_exception_lock);
1742
1743 bucket = fib6_nh_get_excptn_bucket(nh, &rt6_exception_lock);
1744 if (!bucket)
1745 goto out;
1746
1747 /* Prevent rt6_insert_exception() to recreate the bucket list */
1748 if (!from)
1749 fib6_nh_excptn_bucket_set_flushed(nh, &rt6_exception_lock);
1750
1751 for (i = 0; i < FIB6_EXCEPTION_BUCKET_SIZE; i++) {
1752 hlist_for_each_entry_safe(rt6_ex, tmp, &bucket->chain, hlist) {
1753 if (!from ||
1754 rcu_access_pointer(rt6_ex->rt6i->from) == from)
1755 rt6_remove_exception(bucket, rt6_ex);
1756 }
1757 WARN_ON_ONCE(!from && bucket->depth);
1758 bucket++;
1759 }
1760out:
1761 spin_unlock_bh(&rt6_exception_lock);
1762}
1763
1764static int rt6_nh_flush_exceptions(struct fib6_nh *nh, void *arg)
1765{
1766 struct fib6_info *f6i = arg;
1767
1768 fib6_nh_flush_exceptions(nh, f6i);
1769
1770 return 0;
1771}
1772
1773void rt6_flush_exceptions(struct fib6_info *f6i)
1774{
1775 if (f6i->nh)
1776 nexthop_for_each_fib6_nh(f6i->nh, rt6_nh_flush_exceptions,
1777 f6i);
1778 else
1779 fib6_nh_flush_exceptions(f6i->fib6_nh, f6i);
1780}
1781
1782/* Find cached rt in the hash table inside passed in rt
1783 * Caller has to hold rcu_read_lock()
1784 */
1785static struct rt6_info *rt6_find_cached_rt(const struct fib6_result *res,
1786 const struct in6_addr *daddr,
1787 const struct in6_addr *saddr)
1788{
1789 const struct in6_addr *src_key = NULL;
1790 struct rt6_exception_bucket *bucket;
1791 struct rt6_exception *rt6_ex;
1792 struct rt6_info *ret = NULL;
1793
1794#ifdef CONFIG_IPV6_SUBTREES
1795 /* fib6i_src.plen != 0 indicates f6i is in subtree
1796 * and exception table is indexed by a hash of
1797 * both fib6_dst and fib6_src.
1798 * However, the src addr used to create the hash
1799 * might not be exactly the passed in saddr which
1800 * is a /128 addr from the flow.
1801 * So we need to use f6i->fib6_src to redo lookup
1802 * if the passed in saddr does not find anything.
1803 * (See the logic in ip6_rt_cache_alloc() on how
1804 * rt->rt6i_src is updated.)
1805 */
1806 if (res->f6i->fib6_src.plen)
1807 src_key = saddr;
1808find_ex:
1809#endif
1810 bucket = fib6_nh_get_excptn_bucket(res->nh, NULL);
1811 rt6_ex = __rt6_find_exception_rcu(&bucket, daddr, src_key);
1812
1813 if (rt6_ex && !rt6_check_expired(rt6_ex->rt6i))
1814 ret = rt6_ex->rt6i;
1815
1816#ifdef CONFIG_IPV6_SUBTREES
1817 /* Use fib6_src as src_key and redo lookup */
1818 if (!ret && src_key && src_key != &res->f6i->fib6_src.addr) {
1819 src_key = &res->f6i->fib6_src.addr;
1820 goto find_ex;
1821 }
1822#endif
1823
1824 return ret;
1825}
1826
1827/* Remove the passed in cached rt from the hash table that contains it */
1828static int fib6_nh_remove_exception(const struct fib6_nh *nh, int plen,
1829 const struct rt6_info *rt)
1830{
1831 const struct in6_addr *src_key = NULL;
1832 struct rt6_exception_bucket *bucket;
1833 struct rt6_exception *rt6_ex;
1834 int err;
1835
1836 if (!rcu_access_pointer(nh->rt6i_exception_bucket))
1837 return -ENOENT;
1838
1839 spin_lock_bh(&rt6_exception_lock);
1840 bucket = fib6_nh_get_excptn_bucket(nh, &rt6_exception_lock);
1841
1842#ifdef CONFIG_IPV6_SUBTREES
1843 /* rt6i_src.plen != 0 indicates 'from' is in subtree
1844 * and exception table is indexed by a hash of
1845 * both rt6i_dst and rt6i_src.
1846 * Otherwise, the exception table is indexed by
1847 * a hash of only rt6i_dst.
1848 */
1849 if (plen)
1850 src_key = &rt->rt6i_src.addr;
1851#endif
1852 rt6_ex = __rt6_find_exception_spinlock(&bucket,
1853 &rt->rt6i_dst.addr,
1854 src_key);
1855 if (rt6_ex) {
1856 rt6_remove_exception(bucket, rt6_ex);
1857 err = 0;
1858 } else {
1859 err = -ENOENT;
1860 }
1861
1862 spin_unlock_bh(&rt6_exception_lock);
1863 return err;
1864}
1865
1866struct fib6_nh_excptn_arg {
1867 struct rt6_info *rt;
1868 int plen;
1869};
1870
1871static int rt6_nh_remove_exception_rt(struct fib6_nh *nh, void *_arg)
1872{
1873 struct fib6_nh_excptn_arg *arg = _arg;
1874 int err;
1875
1876 err = fib6_nh_remove_exception(nh, arg->plen, arg->rt);
1877 if (err == 0)
1878 return 1;
1879
1880 return 0;
1881}
1882
1883static int rt6_remove_exception_rt(struct rt6_info *rt)
1884{
1885 struct fib6_info *from;
1886
1887 from = rcu_dereference(rt->from);
1888 if (!from || !(rt->rt6i_flags & RTF_CACHE))
1889 return -EINVAL;
1890
1891 if (from->nh) {
1892 struct fib6_nh_excptn_arg arg = {
1893 .rt = rt,
1894 .plen = from->fib6_src.plen
1895 };
1896 int rc;
1897
1898 /* rc = 1 means an entry was found */
1899 rc = nexthop_for_each_fib6_nh(from->nh,
1900 rt6_nh_remove_exception_rt,
1901 &arg);
1902 return rc ? 0 : -ENOENT;
1903 }
1904
1905 return fib6_nh_remove_exception(from->fib6_nh,
1906 from->fib6_src.plen, rt);
1907}
1908
1909/* Find rt6_ex which contains the passed in rt cache and
1910 * refresh its stamp
1911 */
1912static void fib6_nh_update_exception(const struct fib6_nh *nh, int plen,
1913 const struct rt6_info *rt)
1914{
1915 const struct in6_addr *src_key = NULL;
1916 struct rt6_exception_bucket *bucket;
1917 struct rt6_exception *rt6_ex;
1918
1919 bucket = fib6_nh_get_excptn_bucket(nh, NULL);
1920#ifdef CONFIG_IPV6_SUBTREES
1921 /* rt6i_src.plen != 0 indicates 'from' is in subtree
1922 * and exception table is indexed by a hash of
1923 * both rt6i_dst and rt6i_src.
1924 * Otherwise, the exception table is indexed by
1925 * a hash of only rt6i_dst.
1926 */
1927 if (plen)
1928 src_key = &rt->rt6i_src.addr;
1929#endif
1930 rt6_ex = __rt6_find_exception_rcu(&bucket, &rt->rt6i_dst.addr, src_key);
1931 if (rt6_ex)
1932 rt6_ex->stamp = jiffies;
1933}
1934
1935struct fib6_nh_match_arg {
1936 const struct net_device *dev;
1937 const struct in6_addr *gw;
1938 struct fib6_nh *match;
1939};
1940
1941/* determine if fib6_nh has given device and gateway */
1942static int fib6_nh_find_match(struct fib6_nh *nh, void *_arg)
1943{
1944 struct fib6_nh_match_arg *arg = _arg;
1945
1946 if (arg->dev != nh->fib_nh_dev ||
1947 (arg->gw && !nh->fib_nh_gw_family) ||
1948 (!arg->gw && nh->fib_nh_gw_family) ||
1949 (arg->gw && !ipv6_addr_equal(arg->gw, &nh->fib_nh_gw6)))
1950 return 0;
1951
1952 arg->match = nh;
1953
1954 /* found a match, break the loop */
1955 return 1;
1956}
1957
1958static void rt6_update_exception_stamp_rt(struct rt6_info *rt)
1959{
1960 struct fib6_info *from;
1961 struct fib6_nh *fib6_nh;
1962
1963 rcu_read_lock();
1964
1965 from = rcu_dereference(rt->from);
1966 if (!from || !(rt->rt6i_flags & RTF_CACHE))
1967 goto unlock;
1968
1969 if (from->nh) {
1970 struct fib6_nh_match_arg arg = {
1971 .dev = rt->dst.dev,
1972 .gw = &rt->rt6i_gateway,
1973 };
1974
1975 nexthop_for_each_fib6_nh(from->nh, fib6_nh_find_match, &arg);
1976
1977 if (!arg.match)
1978 goto unlock;
1979 fib6_nh = arg.match;
1980 } else {
1981 fib6_nh = from->fib6_nh;
1982 }
1983 fib6_nh_update_exception(fib6_nh, from->fib6_src.plen, rt);
1984unlock:
1985 rcu_read_unlock();
1986}
1987
1988static bool rt6_mtu_change_route_allowed(struct inet6_dev *idev,
1989 struct rt6_info *rt, int mtu)
1990{
1991 /* If the new MTU is lower than the route PMTU, this new MTU will be the
1992 * lowest MTU in the path: always allow updating the route PMTU to
1993 * reflect PMTU decreases.
1994 *
1995 * If the new MTU is higher, and the route PMTU is equal to the local
1996 * MTU, this means the old MTU is the lowest in the path, so allow
1997 * updating it: if other nodes now have lower MTUs, PMTU discovery will
1998 * handle this.
1999 */
2000
2001 if (dst_mtu(&rt->dst) >= mtu)
2002 return true;
2003
2004 if (dst_mtu(&rt->dst) == idev->cnf.mtu6)
2005 return true;
2006
2007 return false;
2008}
2009
2010static void rt6_exceptions_update_pmtu(struct inet6_dev *idev,
2011 const struct fib6_nh *nh, int mtu)
2012{
2013 struct rt6_exception_bucket *bucket;
2014 struct rt6_exception *rt6_ex;
2015 int i;
2016
2017 bucket = fib6_nh_get_excptn_bucket(nh, &rt6_exception_lock);
2018 if (!bucket)
2019 return;
2020
2021 for (i = 0; i < FIB6_EXCEPTION_BUCKET_SIZE; i++) {
2022 hlist_for_each_entry(rt6_ex, &bucket->chain, hlist) {
2023 struct rt6_info *entry = rt6_ex->rt6i;
2024
2025 /* For RTF_CACHE with rt6i_pmtu == 0 (i.e. a redirected
2026 * route), the metrics of its rt->from have already
2027 * been updated.
2028 */
2029 if (dst_metric_raw(&entry->dst, RTAX_MTU) &&
2030 rt6_mtu_change_route_allowed(idev, entry, mtu))
2031 dst_metric_set(&entry->dst, RTAX_MTU, mtu);
2032 }
2033 bucket++;
2034 }
2035}
2036
2037#define RTF_CACHE_GATEWAY (RTF_GATEWAY | RTF_CACHE)
2038
2039static void fib6_nh_exceptions_clean_tohost(const struct fib6_nh *nh,
2040 const struct in6_addr *gateway)
2041{
2042 struct rt6_exception_bucket *bucket;
2043 struct rt6_exception *rt6_ex;
2044 struct hlist_node *tmp;
2045 int i;
2046
2047 if (!rcu_access_pointer(nh->rt6i_exception_bucket))
2048 return;
2049
2050 spin_lock_bh(&rt6_exception_lock);
2051 bucket = fib6_nh_get_excptn_bucket(nh, &rt6_exception_lock);
2052 if (bucket) {
2053 for (i = 0; i < FIB6_EXCEPTION_BUCKET_SIZE; i++) {
2054 hlist_for_each_entry_safe(rt6_ex, tmp,
2055 &bucket->chain, hlist) {
2056 struct rt6_info *entry = rt6_ex->rt6i;
2057
2058 if ((entry->rt6i_flags & RTF_CACHE_GATEWAY) ==
2059 RTF_CACHE_GATEWAY &&
2060 ipv6_addr_equal(gateway,
2061 &entry->rt6i_gateway)) {
2062 rt6_remove_exception(bucket, rt6_ex);
2063 }
2064 }
2065 bucket++;
2066 }
2067 }
2068
2069 spin_unlock_bh(&rt6_exception_lock);
2070}
2071
2072static void rt6_age_examine_exception(struct rt6_exception_bucket *bucket,
2073 struct rt6_exception *rt6_ex,
2074 struct fib6_gc_args *gc_args,
2075 unsigned long now)
2076{
2077 struct rt6_info *rt = rt6_ex->rt6i;
2078
2079 /* we are pruning and obsoleting aged-out and non gateway exceptions
2080 * even if others have still references to them, so that on next
2081 * dst_check() such references can be dropped.
2082 * EXPIRES exceptions - e.g. pmtu-generated ones are pruned when
2083 * expired, independently from their aging, as per RFC 8201 section 4
2084 */
2085 if (!(rt->rt6i_flags & RTF_EXPIRES)) {
2086 if (time_after_eq(now, rt->dst.lastuse + gc_args->timeout)) {
2087 RT6_TRACE("aging clone %p\n", rt);
2088 rt6_remove_exception(bucket, rt6_ex);
2089 return;
2090 }
2091 } else if (time_after(jiffies, rt->dst.expires)) {
2092 RT6_TRACE("purging expired route %p\n", rt);
2093 rt6_remove_exception(bucket, rt6_ex);
2094 return;
2095 }
2096
2097 if (rt->rt6i_flags & RTF_GATEWAY) {
2098 struct neighbour *neigh;
2099
2100 neigh = __ipv6_neigh_lookup_noref(rt->dst.dev, &rt->rt6i_gateway);
2101
2102 if (!(neigh && (neigh->flags & NTF_ROUTER))) {
2103 RT6_TRACE("purging route %p via non-router but gateway\n",
2104 rt);
2105 rt6_remove_exception(bucket, rt6_ex);
2106 return;
2107 }
2108 }
2109
2110 gc_args->more++;
2111}
2112
2113static void fib6_nh_age_exceptions(const struct fib6_nh *nh,
2114 struct fib6_gc_args *gc_args,
2115 unsigned long now)
2116{
2117 struct rt6_exception_bucket *bucket;
2118 struct rt6_exception *rt6_ex;
2119 struct hlist_node *tmp;
2120 int i;
2121
2122 if (!rcu_access_pointer(nh->rt6i_exception_bucket))
2123 return;
2124
2125 rcu_read_lock_bh();
2126 spin_lock(&rt6_exception_lock);
2127 bucket = fib6_nh_get_excptn_bucket(nh, &rt6_exception_lock);
2128 if (bucket) {
2129 for (i = 0; i < FIB6_EXCEPTION_BUCKET_SIZE; i++) {
2130 hlist_for_each_entry_safe(rt6_ex, tmp,
2131 &bucket->chain, hlist) {
2132 rt6_age_examine_exception(bucket, rt6_ex,
2133 gc_args, now);
2134 }
2135 bucket++;
2136 }
2137 }
2138 spin_unlock(&rt6_exception_lock);
2139 rcu_read_unlock_bh();
2140}
2141
2142struct fib6_nh_age_excptn_arg {
2143 struct fib6_gc_args *gc_args;
2144 unsigned long now;
2145};
2146
2147static int rt6_nh_age_exceptions(struct fib6_nh *nh, void *_arg)
2148{
2149 struct fib6_nh_age_excptn_arg *arg = _arg;
2150
2151 fib6_nh_age_exceptions(nh, arg->gc_args, arg->now);
2152 return 0;
2153}
2154
2155void rt6_age_exceptions(struct fib6_info *f6i,
2156 struct fib6_gc_args *gc_args,
2157 unsigned long now)
2158{
2159 if (f6i->nh) {
2160 struct fib6_nh_age_excptn_arg arg = {
2161 .gc_args = gc_args,
2162 .now = now
2163 };
2164
2165 nexthop_for_each_fib6_nh(f6i->nh, rt6_nh_age_exceptions,
2166 &arg);
2167 } else {
2168 fib6_nh_age_exceptions(f6i->fib6_nh, gc_args, now);
2169 }
2170}
2171
2172/* must be called with rcu lock held */
2173int fib6_table_lookup(struct net *net, struct fib6_table *table, int oif,
2174 struct flowi6 *fl6, struct fib6_result *res, int strict)
2175{
2176 struct fib6_node *fn, *saved_fn;
2177
2178 fn = fib6_node_lookup(&table->tb6_root, &fl6->daddr, &fl6->saddr);
2179 saved_fn = fn;
2180
2181 if (fl6->flowi6_flags & FLOWI_FLAG_SKIP_NH_OIF)
2182 oif = 0;
2183
2184redo_rt6_select:
2185 rt6_select(net, fn, oif, res, strict);
2186 if (res->f6i == net->ipv6.fib6_null_entry) {
2187 fn = fib6_backtrack(fn, &fl6->saddr);
2188 if (fn)
2189 goto redo_rt6_select;
2190 else if (strict & RT6_LOOKUP_F_REACHABLE) {
2191 /* also consider unreachable route */
2192 strict &= ~RT6_LOOKUP_F_REACHABLE;
2193 fn = saved_fn;
2194 goto redo_rt6_select;
2195 }
2196 }
2197
2198 trace_fib6_table_lookup(net, res, table, fl6);
2199
2200 return 0;
2201}
2202
2203struct rt6_info *ip6_pol_route(struct net *net, struct fib6_table *table,
2204 int oif, struct flowi6 *fl6,
2205 const struct sk_buff *skb, int flags)
2206{
2207 struct fib6_result res = {};
2208 struct rt6_info *rt = NULL;
2209 int strict = 0;
2210
2211 WARN_ON_ONCE((flags & RT6_LOOKUP_F_DST_NOREF) &&
2212 !rcu_read_lock_held());
2213
2214 strict |= flags & RT6_LOOKUP_F_IFACE;
2215 strict |= flags & RT6_LOOKUP_F_IGNORE_LINKSTATE;
2216 if (net->ipv6.devconf_all->forwarding == 0)
2217 strict |= RT6_LOOKUP_F_REACHABLE;
2218
2219 rcu_read_lock();
2220
2221 fib6_table_lookup(net, table, oif, fl6, &res, strict);
2222 if (res.f6i == net->ipv6.fib6_null_entry)
2223 goto out;
2224
2225 fib6_select_path(net, &res, fl6, oif, false, skb, strict);
2226
2227 /*Search through exception table */
2228 rt = rt6_find_cached_rt(&res, &fl6->daddr, &fl6->saddr);
2229 if (rt) {
2230 goto out;
2231 } else if (unlikely((fl6->flowi6_flags & FLOWI_FLAG_KNOWN_NH) &&
2232 !res.nh->fib_nh_gw_family)) {
2233 /* Create a RTF_CACHE clone which will not be
2234 * owned by the fib6 tree. It is for the special case where
2235 * the daddr in the skb during the neighbor look-up is different
2236 * from the fl6->daddr used to look-up route here.
2237 */
2238 rt = ip6_rt_cache_alloc(&res, &fl6->daddr, NULL);
2239
2240 if (rt) {
2241 /* 1 refcnt is taken during ip6_rt_cache_alloc().
2242 * As rt6_uncached_list_add() does not consume refcnt,
2243 * this refcnt is always returned to the caller even
2244 * if caller sets RT6_LOOKUP_F_DST_NOREF flag.
2245 */
2246 rt6_uncached_list_add(rt);
2247 atomic_inc(&net->ipv6.rt6_stats->fib_rt_uncache);
2248 rcu_read_unlock();
2249
2250 return rt;
2251 }
2252 } else {
2253 /* Get a percpu copy */
2254 local_bh_disable();
2255 rt = rt6_get_pcpu_route(&res);
2256
2257 if (!rt)
2258 rt = rt6_make_pcpu_route(net, &res);
2259
2260 local_bh_enable();
2261 }
2262out:
2263 if (!rt)
2264 rt = net->ipv6.ip6_null_entry;
2265 if (!(flags & RT6_LOOKUP_F_DST_NOREF))
2266 ip6_hold_safe(net, &rt);
2267 rcu_read_unlock();
2268
2269 return rt;
2270}
2271EXPORT_SYMBOL_GPL(ip6_pol_route);
2272
2273INDIRECT_CALLABLE_SCOPE struct rt6_info *ip6_pol_route_input(struct net *net,
2274 struct fib6_table *table,
2275 struct flowi6 *fl6,
2276 const struct sk_buff *skb,
2277 int flags)
2278{
2279 return ip6_pol_route(net, table, fl6->flowi6_iif, fl6, skb, flags);
2280}
2281
2282struct dst_entry *ip6_route_input_lookup(struct net *net,
2283 struct net_device *dev,
2284 struct flowi6 *fl6,
2285 const struct sk_buff *skb,
2286 int flags)
2287{
2288 if (rt6_need_strict(&fl6->daddr) && dev->type != ARPHRD_PIMREG)
2289 flags |= RT6_LOOKUP_F_IFACE;
2290
2291 return fib6_rule_lookup(net, fl6, skb, flags, ip6_pol_route_input);
2292}
2293EXPORT_SYMBOL_GPL(ip6_route_input_lookup);
2294
2295static void ip6_multipath_l3_keys(const struct sk_buff *skb,
2296 struct flow_keys *keys,
2297 struct flow_keys *flkeys)
2298{
2299 const struct ipv6hdr *outer_iph = ipv6_hdr(skb);
2300 const struct ipv6hdr *key_iph = outer_iph;
2301 struct flow_keys *_flkeys = flkeys;
2302 const struct ipv6hdr *inner_iph;
2303 const struct icmp6hdr *icmph;
2304 struct ipv6hdr _inner_iph;
2305 struct icmp6hdr _icmph;
2306
2307 if (likely(outer_iph->nexthdr != IPPROTO_ICMPV6))
2308 goto out;
2309
2310 icmph = skb_header_pointer(skb, skb_transport_offset(skb),
2311 sizeof(_icmph), &_icmph);
2312 if (!icmph)
2313 goto out;
2314
2315 if (!icmpv6_is_err(icmph->icmp6_type))
2316 goto out;
2317
2318 inner_iph = skb_header_pointer(skb,
2319 skb_transport_offset(skb) + sizeof(*icmph),
2320 sizeof(_inner_iph), &_inner_iph);
2321 if (!inner_iph)
2322 goto out;
2323
2324 key_iph = inner_iph;
2325 _flkeys = NULL;
2326out:
2327 if (_flkeys) {
2328 keys->addrs.v6addrs.src = _flkeys->addrs.v6addrs.src;
2329 keys->addrs.v6addrs.dst = _flkeys->addrs.v6addrs.dst;
2330 keys->tags.flow_label = _flkeys->tags.flow_label;
2331 keys->basic.ip_proto = _flkeys->basic.ip_proto;
2332 } else {
2333 keys->addrs.v6addrs.src = key_iph->saddr;
2334 keys->addrs.v6addrs.dst = key_iph->daddr;
2335 keys->tags.flow_label = ip6_flowlabel(key_iph);
2336 keys->basic.ip_proto = key_iph->nexthdr;
2337 }
2338}
2339
2340static u32 rt6_multipath_custom_hash_outer(const struct net *net,
2341 const struct sk_buff *skb,
2342 bool *p_has_inner)
2343{
2344 u32 hash_fields = ip6_multipath_hash_fields(net);
2345 struct flow_keys keys, hash_keys;
2346
2347 if (!(hash_fields & FIB_MULTIPATH_HASH_FIELD_OUTER_MASK))
2348 return 0;
2349
2350 memset(&hash_keys, 0, sizeof(hash_keys));
2351 skb_flow_dissect_flow_keys(skb, &keys, FLOW_DISSECTOR_F_STOP_AT_ENCAP);
2352
2353 hash_keys.control.addr_type = FLOW_DISSECTOR_KEY_IPV6_ADDRS;
2354 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_SRC_IP)
2355 hash_keys.addrs.v6addrs.src = keys.addrs.v6addrs.src;
2356 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_DST_IP)
2357 hash_keys.addrs.v6addrs.dst = keys.addrs.v6addrs.dst;
2358 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_IP_PROTO)
2359 hash_keys.basic.ip_proto = keys.basic.ip_proto;
2360 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_FLOWLABEL)
2361 hash_keys.tags.flow_label = keys.tags.flow_label;
2362 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_SRC_PORT)
2363 hash_keys.ports.src = keys.ports.src;
2364 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_DST_PORT)
2365 hash_keys.ports.dst = keys.ports.dst;
2366
2367 *p_has_inner = !!(keys.control.flags & FLOW_DIS_ENCAPSULATION);
2368 return flow_hash_from_keys(&hash_keys);
2369}
2370
2371static u32 rt6_multipath_custom_hash_inner(const struct net *net,
2372 const struct sk_buff *skb,
2373 bool has_inner)
2374{
2375 u32 hash_fields = ip6_multipath_hash_fields(net);
2376 struct flow_keys keys, hash_keys;
2377
2378 /* We assume the packet carries an encapsulation, but if none was
2379 * encountered during dissection of the outer flow, then there is no
2380 * point in calling the flow dissector again.
2381 */
2382 if (!has_inner)
2383 return 0;
2384
2385 if (!(hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_MASK))
2386 return 0;
2387
2388 memset(&hash_keys, 0, sizeof(hash_keys));
2389 skb_flow_dissect_flow_keys(skb, &keys, 0);
2390
2391 if (!(keys.control.flags & FLOW_DIS_ENCAPSULATION))
2392 return 0;
2393
2394 if (keys.control.addr_type == FLOW_DISSECTOR_KEY_IPV4_ADDRS) {
2395 hash_keys.control.addr_type = FLOW_DISSECTOR_KEY_IPV4_ADDRS;
2396 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_SRC_IP)
2397 hash_keys.addrs.v4addrs.src = keys.addrs.v4addrs.src;
2398 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_DST_IP)
2399 hash_keys.addrs.v4addrs.dst = keys.addrs.v4addrs.dst;
2400 } else if (keys.control.addr_type == FLOW_DISSECTOR_KEY_IPV6_ADDRS) {
2401 hash_keys.control.addr_type = FLOW_DISSECTOR_KEY_IPV6_ADDRS;
2402 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_SRC_IP)
2403 hash_keys.addrs.v6addrs.src = keys.addrs.v6addrs.src;
2404 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_DST_IP)
2405 hash_keys.addrs.v6addrs.dst = keys.addrs.v6addrs.dst;
2406 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_FLOWLABEL)
2407 hash_keys.tags.flow_label = keys.tags.flow_label;
2408 }
2409
2410 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_IP_PROTO)
2411 hash_keys.basic.ip_proto = keys.basic.ip_proto;
2412 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_SRC_PORT)
2413 hash_keys.ports.src = keys.ports.src;
2414 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_DST_PORT)
2415 hash_keys.ports.dst = keys.ports.dst;
2416
2417 return flow_hash_from_keys(&hash_keys);
2418}
2419
2420static u32 rt6_multipath_custom_hash_skb(const struct net *net,
2421 const struct sk_buff *skb)
2422{
2423 u32 mhash, mhash_inner;
2424 bool has_inner = true;
2425
2426 mhash = rt6_multipath_custom_hash_outer(net, skb, &has_inner);
2427 mhash_inner = rt6_multipath_custom_hash_inner(net, skb, has_inner);
2428
2429 return jhash_2words(mhash, mhash_inner, 0);
2430}
2431
2432static u32 rt6_multipath_custom_hash_fl6(const struct net *net,
2433 const struct flowi6 *fl6)
2434{
2435 u32 hash_fields = ip6_multipath_hash_fields(net);
2436 struct flow_keys hash_keys;
2437
2438 if (!(hash_fields & FIB_MULTIPATH_HASH_FIELD_OUTER_MASK))
2439 return 0;
2440
2441 memset(&hash_keys, 0, sizeof(hash_keys));
2442 hash_keys.control.addr_type = FLOW_DISSECTOR_KEY_IPV6_ADDRS;
2443 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_SRC_IP)
2444 hash_keys.addrs.v6addrs.src = fl6->saddr;
2445 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_DST_IP)
2446 hash_keys.addrs.v6addrs.dst = fl6->daddr;
2447 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_IP_PROTO)
2448 hash_keys.basic.ip_proto = fl6->flowi6_proto;
2449 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_FLOWLABEL)
2450 hash_keys.tags.flow_label = (__force u32)flowi6_get_flowlabel(fl6);
2451 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_SRC_PORT)
2452 hash_keys.ports.src = fl6->fl6_sport;
2453 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_DST_PORT)
2454 hash_keys.ports.dst = fl6->fl6_dport;
2455
2456 return flow_hash_from_keys(&hash_keys);
2457}
2458
2459/* if skb is set it will be used and fl6 can be NULL */
2460u32 rt6_multipath_hash(const struct net *net, const struct flowi6 *fl6,
2461 const struct sk_buff *skb, struct flow_keys *flkeys)
2462{
2463 struct flow_keys hash_keys;
2464 u32 mhash = 0;
2465
2466 switch (ip6_multipath_hash_policy(net)) {
2467 case 0:
2468 memset(&hash_keys, 0, sizeof(hash_keys));
2469 hash_keys.control.addr_type = FLOW_DISSECTOR_KEY_IPV6_ADDRS;
2470 if (skb) {
2471 ip6_multipath_l3_keys(skb, &hash_keys, flkeys);
2472 } else {
2473 hash_keys.addrs.v6addrs.src = fl6->saddr;
2474 hash_keys.addrs.v6addrs.dst = fl6->daddr;
2475 hash_keys.tags.flow_label = (__force u32)flowi6_get_flowlabel(fl6);
2476 hash_keys.basic.ip_proto = fl6->flowi6_proto;
2477 }
2478 mhash = flow_hash_from_keys(&hash_keys);
2479 break;
2480 case 1:
2481 if (skb) {
2482 unsigned int flag = FLOW_DISSECTOR_F_STOP_AT_ENCAP;
2483 struct flow_keys keys;
2484
2485 /* short-circuit if we already have L4 hash present */
2486 if (skb->l4_hash)
2487 return skb_get_hash_raw(skb) >> 1;
2488
2489 memset(&hash_keys, 0, sizeof(hash_keys));
2490
2491 if (!flkeys) {
2492 skb_flow_dissect_flow_keys(skb, &keys, flag);
2493 flkeys = &keys;
2494 }
2495 hash_keys.control.addr_type = FLOW_DISSECTOR_KEY_IPV6_ADDRS;
2496 hash_keys.addrs.v6addrs.src = flkeys->addrs.v6addrs.src;
2497 hash_keys.addrs.v6addrs.dst = flkeys->addrs.v6addrs.dst;
2498 hash_keys.ports.src = flkeys->ports.src;
2499 hash_keys.ports.dst = flkeys->ports.dst;
2500 hash_keys.basic.ip_proto = flkeys->basic.ip_proto;
2501 } else {
2502 memset(&hash_keys, 0, sizeof(hash_keys));
2503 hash_keys.control.addr_type = FLOW_DISSECTOR_KEY_IPV6_ADDRS;
2504 hash_keys.addrs.v6addrs.src = fl6->saddr;
2505 hash_keys.addrs.v6addrs.dst = fl6->daddr;
2506 hash_keys.ports.src = fl6->fl6_sport;
2507 hash_keys.ports.dst = fl6->fl6_dport;
2508 hash_keys.basic.ip_proto = fl6->flowi6_proto;
2509 }
2510 mhash = flow_hash_from_keys(&hash_keys);
2511 break;
2512 case 2:
2513 memset(&hash_keys, 0, sizeof(hash_keys));
2514 hash_keys.control.addr_type = FLOW_DISSECTOR_KEY_IPV6_ADDRS;
2515 if (skb) {
2516 struct flow_keys keys;
2517
2518 if (!flkeys) {
2519 skb_flow_dissect_flow_keys(skb, &keys, 0);
2520 flkeys = &keys;
2521 }
2522
2523 /* Inner can be v4 or v6 */
2524 if (flkeys->control.addr_type == FLOW_DISSECTOR_KEY_IPV4_ADDRS) {
2525 hash_keys.control.addr_type = FLOW_DISSECTOR_KEY_IPV4_ADDRS;
2526 hash_keys.addrs.v4addrs.src = flkeys->addrs.v4addrs.src;
2527 hash_keys.addrs.v4addrs.dst = flkeys->addrs.v4addrs.dst;
2528 } else if (flkeys->control.addr_type == FLOW_DISSECTOR_KEY_IPV6_ADDRS) {
2529 hash_keys.control.addr_type = FLOW_DISSECTOR_KEY_IPV6_ADDRS;
2530 hash_keys.addrs.v6addrs.src = flkeys->addrs.v6addrs.src;
2531 hash_keys.addrs.v6addrs.dst = flkeys->addrs.v6addrs.dst;
2532 hash_keys.tags.flow_label = flkeys->tags.flow_label;
2533 hash_keys.basic.ip_proto = flkeys->basic.ip_proto;
2534 } else {
2535 /* Same as case 0 */
2536 hash_keys.control.addr_type = FLOW_DISSECTOR_KEY_IPV6_ADDRS;
2537 ip6_multipath_l3_keys(skb, &hash_keys, flkeys);
2538 }
2539 } else {
2540 /* Same as case 0 */
2541 hash_keys.control.addr_type = FLOW_DISSECTOR_KEY_IPV6_ADDRS;
2542 hash_keys.addrs.v6addrs.src = fl6->saddr;
2543 hash_keys.addrs.v6addrs.dst = fl6->daddr;
2544 hash_keys.tags.flow_label = (__force u32)flowi6_get_flowlabel(fl6);
2545 hash_keys.basic.ip_proto = fl6->flowi6_proto;
2546 }
2547 mhash = flow_hash_from_keys(&hash_keys);
2548 break;
2549 case 3:
2550 if (skb)
2551 mhash = rt6_multipath_custom_hash_skb(net, skb);
2552 else
2553 mhash = rt6_multipath_custom_hash_fl6(net, fl6);
2554 break;
2555 }
2556
2557 return mhash >> 1;
2558}
2559
2560/* Called with rcu held */
2561void ip6_route_input(struct sk_buff *skb)
2562{
2563 const struct ipv6hdr *iph = ipv6_hdr(skb);
2564 struct net *net = dev_net(skb->dev);
2565 int flags = RT6_LOOKUP_F_HAS_SADDR | RT6_LOOKUP_F_DST_NOREF;
2566 struct ip_tunnel_info *tun_info;
2567 struct flowi6 fl6 = {
2568 .flowi6_iif = skb->dev->ifindex,
2569 .daddr = iph->daddr,
2570 .saddr = iph->saddr,
2571 .flowlabel = ip6_flowinfo(iph),
2572 .flowi6_mark = skb->mark,
2573 .flowi6_proto = iph->nexthdr,
2574 };
2575 struct flow_keys *flkeys = NULL, _flkeys;
2576
2577 tun_info = skb_tunnel_info(skb);
2578 if (tun_info && !(tun_info->mode & IP_TUNNEL_INFO_TX))
2579 fl6.flowi6_tun_key.tun_id = tun_info->key.tun_id;
2580
2581 if (fib6_rules_early_flow_dissect(net, skb, &fl6, &_flkeys))
2582 flkeys = &_flkeys;
2583
2584 if (unlikely(fl6.flowi6_proto == IPPROTO_ICMPV6))
2585 fl6.mp_hash = rt6_multipath_hash(net, &fl6, skb, flkeys);
2586 skb_dst_drop(skb);
2587 skb_dst_set_noref(skb, ip6_route_input_lookup(net, skb->dev,
2588 &fl6, skb, flags));
2589}
2590
2591INDIRECT_CALLABLE_SCOPE struct rt6_info *ip6_pol_route_output(struct net *net,
2592 struct fib6_table *table,
2593 struct flowi6 *fl6,
2594 const struct sk_buff *skb,
2595 int flags)
2596{
2597 return ip6_pol_route(net, table, fl6->flowi6_oif, fl6, skb, flags);
2598}
2599
2600struct dst_entry *ip6_route_output_flags_noref(struct net *net,
2601 const struct sock *sk,
2602 struct flowi6 *fl6, int flags)
2603{
2604 bool any_src;
2605
2606 if (ipv6_addr_type(&fl6->daddr) &
2607 (IPV6_ADDR_MULTICAST | IPV6_ADDR_LINKLOCAL)) {
2608 struct dst_entry *dst;
2609
2610 /* This function does not take refcnt on the dst */
2611 dst = l3mdev_link_scope_lookup(net, fl6);
2612 if (dst)
2613 return dst;
2614 }
2615
2616 fl6->flowi6_iif = LOOPBACK_IFINDEX;
2617
2618 flags |= RT6_LOOKUP_F_DST_NOREF;
2619 any_src = ipv6_addr_any(&fl6->saddr);
2620 if ((sk && sk->sk_bound_dev_if) || rt6_need_strict(&fl6->daddr) ||
2621 (fl6->flowi6_oif && any_src))
2622 flags |= RT6_LOOKUP_F_IFACE;
2623
2624 if (!any_src)
2625 flags |= RT6_LOOKUP_F_HAS_SADDR;
2626 else if (sk)
2627 flags |= rt6_srcprefs2flags(inet6_sk(sk)->srcprefs);
2628
2629 return fib6_rule_lookup(net, fl6, NULL, flags, ip6_pol_route_output);
2630}
2631EXPORT_SYMBOL_GPL(ip6_route_output_flags_noref);
2632
2633struct dst_entry *ip6_route_output_flags(struct net *net,
2634 const struct sock *sk,
2635 struct flowi6 *fl6,
2636 int flags)
2637{
2638 struct dst_entry *dst;
2639 struct rt6_info *rt6;
2640
2641 rcu_read_lock();
2642 dst = ip6_route_output_flags_noref(net, sk, fl6, flags);
2643 rt6 = (struct rt6_info *)dst;
2644 /* For dst cached in uncached_list, refcnt is already taken. */
2645 if (list_empty(&rt6->rt6i_uncached) && !dst_hold_safe(dst)) {
2646 dst = &net->ipv6.ip6_null_entry->dst;
2647 dst_hold(dst);
2648 }
2649 rcu_read_unlock();
2650
2651 return dst;
2652}
2653EXPORT_SYMBOL_GPL(ip6_route_output_flags);
2654
2655struct dst_entry *ip6_blackhole_route(struct net *net, struct dst_entry *dst_orig)
2656{
2657 struct rt6_info *rt, *ort = (struct rt6_info *) dst_orig;
2658 struct net_device *loopback_dev = net->loopback_dev;
2659 struct dst_entry *new = NULL;
2660
2661 rt = dst_alloc(&ip6_dst_blackhole_ops, loopback_dev, 1,
2662 DST_OBSOLETE_DEAD, 0);
2663 if (rt) {
2664 rt6_info_init(rt);
2665 atomic_inc(&net->ipv6.rt6_stats->fib_rt_alloc);
2666
2667 new = &rt->dst;
2668 new->__use = 1;
2669 new->input = dst_discard;
2670 new->output = dst_discard_out;
2671
2672 dst_copy_metrics(new, &ort->dst);
2673
2674 rt->rt6i_idev = in6_dev_get(loopback_dev);
2675 rt->rt6i_gateway = ort->rt6i_gateway;
2676 rt->rt6i_flags = ort->rt6i_flags & ~RTF_PCPU;
2677
2678 memcpy(&rt->rt6i_dst, &ort->rt6i_dst, sizeof(struct rt6key));
2679#ifdef CONFIG_IPV6_SUBTREES
2680 memcpy(&rt->rt6i_src, &ort->rt6i_src, sizeof(struct rt6key));
2681#endif
2682 }
2683
2684 dst_release(dst_orig);
2685 return new ? new : ERR_PTR(-ENOMEM);
2686}
2687
2688/*
2689 * Destination cache support functions
2690 */
2691
2692static bool fib6_check(struct fib6_info *f6i, u32 cookie)
2693{
2694 u32 rt_cookie = 0;
2695
2696 if (!fib6_get_cookie_safe(f6i, &rt_cookie) || rt_cookie != cookie)
2697 return false;
2698
2699 if (fib6_check_expired(f6i))
2700 return false;
2701
2702 return true;
2703}
2704
2705static struct dst_entry *rt6_check(struct rt6_info *rt,
2706 struct fib6_info *from,
2707 u32 cookie)
2708{
2709 u32 rt_cookie = 0;
2710
2711 if (!from || !fib6_get_cookie_safe(from, &rt_cookie) ||
2712 rt_cookie != cookie)
2713 return NULL;
2714
2715 if (rt6_check_expired(rt))
2716 return NULL;
2717
2718 return &rt->dst;
2719}
2720
2721static struct dst_entry *rt6_dst_from_check(struct rt6_info *rt,
2722 struct fib6_info *from,
2723 u32 cookie)
2724{
2725 if (!__rt6_check_expired(rt) &&
2726 rt->dst.obsolete == DST_OBSOLETE_FORCE_CHK &&
2727 fib6_check(from, cookie))
2728 return &rt->dst;
2729 else
2730 return NULL;
2731}
2732
2733INDIRECT_CALLABLE_SCOPE struct dst_entry *ip6_dst_check(struct dst_entry *dst,
2734 u32 cookie)
2735{
2736 struct dst_entry *dst_ret;
2737 struct fib6_info *from;
2738 struct rt6_info *rt;
2739
2740 rt = container_of(dst, struct rt6_info, dst);
2741
2742 if (rt->sernum)
2743 return rt6_is_valid(rt) ? dst : NULL;
2744
2745 rcu_read_lock();
2746
2747 /* All IPV6 dsts are created with ->obsolete set to the value
2748 * DST_OBSOLETE_FORCE_CHK which forces validation calls down
2749 * into this function always.
2750 */
2751
2752 from = rcu_dereference(rt->from);
2753
2754 if (from && (rt->rt6i_flags & RTF_PCPU ||
2755 unlikely(!list_empty(&rt->rt6i_uncached))))
2756 dst_ret = rt6_dst_from_check(rt, from, cookie);
2757 else
2758 dst_ret = rt6_check(rt, from, cookie);
2759
2760 rcu_read_unlock();
2761
2762 return dst_ret;
2763}
2764EXPORT_INDIRECT_CALLABLE(ip6_dst_check);
2765
2766static struct dst_entry *ip6_negative_advice(struct dst_entry *dst)
2767{
2768 struct rt6_info *rt = (struct rt6_info *) dst;
2769
2770 if (rt) {
2771 if (rt->rt6i_flags & RTF_CACHE) {
2772 rcu_read_lock();
2773 if (rt6_check_expired(rt)) {
2774 rt6_remove_exception_rt(rt);
2775 dst = NULL;
2776 }
2777 rcu_read_unlock();
2778 } else {
2779 dst_release(dst);
2780 dst = NULL;
2781 }
2782 }
2783 return dst;
2784}
2785
2786static void ip6_link_failure(struct sk_buff *skb)
2787{
2788 struct rt6_info *rt;
2789
2790 icmpv6_send(skb, ICMPV6_DEST_UNREACH, ICMPV6_ADDR_UNREACH, 0);
2791
2792 rt = (struct rt6_info *) skb_dst(skb);
2793 if (rt) {
2794 rcu_read_lock();
2795 if (rt->rt6i_flags & RTF_CACHE) {
2796 rt6_remove_exception_rt(rt);
2797 } else {
2798 struct fib6_info *from;
2799 struct fib6_node *fn;
2800
2801 from = rcu_dereference(rt->from);
2802 if (from) {
2803 fn = rcu_dereference(from->fib6_node);
2804 if (fn && (rt->rt6i_flags & RTF_DEFAULT))
2805 fn->fn_sernum = -1;
2806 }
2807 }
2808 rcu_read_unlock();
2809 }
2810}
2811
2812static void rt6_update_expires(struct rt6_info *rt0, int timeout)
2813{
2814 if (!(rt0->rt6i_flags & RTF_EXPIRES)) {
2815 struct fib6_info *from;
2816
2817 rcu_read_lock();
2818 from = rcu_dereference(rt0->from);
2819 if (from)
2820 rt0->dst.expires = from->expires;
2821 rcu_read_unlock();
2822 }
2823
2824 dst_set_expires(&rt0->dst, timeout);
2825 rt0->rt6i_flags |= RTF_EXPIRES;
2826}
2827
2828static void rt6_do_update_pmtu(struct rt6_info *rt, u32 mtu)
2829{
2830 struct net *net = dev_net(rt->dst.dev);
2831
2832 dst_metric_set(&rt->dst, RTAX_MTU, mtu);
2833 rt->rt6i_flags |= RTF_MODIFIED;
2834 rt6_update_expires(rt, net->ipv6.sysctl.ip6_rt_mtu_expires);
2835}
2836
2837static bool rt6_cache_allowed_for_pmtu(const struct rt6_info *rt)
2838{
2839 return !(rt->rt6i_flags & RTF_CACHE) &&
2840 (rt->rt6i_flags & RTF_PCPU || rcu_access_pointer(rt->from));
2841}
2842
2843static void __ip6_rt_update_pmtu(struct dst_entry *dst, const struct sock *sk,
2844 const struct ipv6hdr *iph, u32 mtu,
2845 bool confirm_neigh)
2846{
2847 const struct in6_addr *daddr, *saddr;
2848 struct rt6_info *rt6 = (struct rt6_info *)dst;
2849
2850 /* Note: do *NOT* check dst_metric_locked(dst, RTAX_MTU)
2851 * IPv6 pmtu discovery isn't optional, so 'mtu lock' cannot disable it.
2852 * [see also comment in rt6_mtu_change_route()]
2853 */
2854
2855 if (iph) {
2856 daddr = &iph->daddr;
2857 saddr = &iph->saddr;
2858 } else if (sk) {
2859 daddr = &sk->sk_v6_daddr;
2860 saddr = &inet6_sk(sk)->saddr;
2861 } else {
2862 daddr = NULL;
2863 saddr = NULL;
2864 }
2865
2866 if (confirm_neigh)
2867 dst_confirm_neigh(dst, daddr);
2868
2869 if (mtu < IPV6_MIN_MTU)
2870 return;
2871 if (mtu >= dst_mtu(dst))
2872 return;
2873
2874 if (!rt6_cache_allowed_for_pmtu(rt6)) {
2875 rt6_do_update_pmtu(rt6, mtu);
2876 /* update rt6_ex->stamp for cache */
2877 if (rt6->rt6i_flags & RTF_CACHE)
2878 rt6_update_exception_stamp_rt(rt6);
2879 } else if (daddr) {
2880 struct fib6_result res = {};
2881 struct rt6_info *nrt6;
2882
2883 rcu_read_lock();
2884 res.f6i = rcu_dereference(rt6->from);
2885 if (!res.f6i)
2886 goto out_unlock;
2887
2888 res.fib6_flags = res.f6i->fib6_flags;
2889 res.fib6_type = res.f6i->fib6_type;
2890
2891 if (res.f6i->nh) {
2892 struct fib6_nh_match_arg arg = {
2893 .dev = dst->dev,
2894 .gw = &rt6->rt6i_gateway,
2895 };
2896
2897 nexthop_for_each_fib6_nh(res.f6i->nh,
2898 fib6_nh_find_match, &arg);
2899
2900 /* fib6_info uses a nexthop that does not have fib6_nh
2901 * using the dst->dev + gw. Should be impossible.
2902 */
2903 if (!arg.match)
2904 goto out_unlock;
2905
2906 res.nh = arg.match;
2907 } else {
2908 res.nh = res.f6i->fib6_nh;
2909 }
2910
2911 nrt6 = ip6_rt_cache_alloc(&res, daddr, saddr);
2912 if (nrt6) {
2913 rt6_do_update_pmtu(nrt6, mtu);
2914 if (rt6_insert_exception(nrt6, &res))
2915 dst_release_immediate(&nrt6->dst);
2916 }
2917out_unlock:
2918 rcu_read_unlock();
2919 }
2920}
2921
2922static void ip6_rt_update_pmtu(struct dst_entry *dst, struct sock *sk,
2923 struct sk_buff *skb, u32 mtu,
2924 bool confirm_neigh)
2925{
2926 __ip6_rt_update_pmtu(dst, sk, skb ? ipv6_hdr(skb) : NULL, mtu,
2927 confirm_neigh);
2928}
2929
2930void ip6_update_pmtu(struct sk_buff *skb, struct net *net, __be32 mtu,
2931 int oif, u32 mark, kuid_t uid)
2932{
2933 const struct ipv6hdr *iph = (struct ipv6hdr *) skb->data;
2934 struct dst_entry *dst;
2935 struct flowi6 fl6 = {
2936 .flowi6_oif = oif,
2937 .flowi6_mark = mark ? mark : IP6_REPLY_MARK(net, skb->mark),
2938 .daddr = iph->daddr,
2939 .saddr = iph->saddr,
2940 .flowlabel = ip6_flowinfo(iph),
2941 .flowi6_uid = uid,
2942 };
2943
2944 dst = ip6_route_output(net, NULL, &fl6);
2945 if (!dst->error)
2946 __ip6_rt_update_pmtu(dst, NULL, iph, ntohl(mtu), true);
2947 dst_release(dst);
2948}
2949EXPORT_SYMBOL_GPL(ip6_update_pmtu);
2950
2951void ip6_sk_update_pmtu(struct sk_buff *skb, struct sock *sk, __be32 mtu)
2952{
2953 int oif = sk->sk_bound_dev_if;
2954 struct dst_entry *dst;
2955
2956 if (!oif && skb->dev)
2957 oif = l3mdev_master_ifindex(skb->dev);
2958
2959 ip6_update_pmtu(skb, sock_net(sk), mtu, oif, sk->sk_mark, sk->sk_uid);
2960
2961 dst = __sk_dst_get(sk);
2962 if (!dst || !dst->obsolete ||
2963 dst->ops->check(dst, inet6_sk(sk)->dst_cookie))
2964 return;
2965
2966 bh_lock_sock(sk);
2967 if (!sock_owned_by_user(sk) && !ipv6_addr_v4mapped(&sk->sk_v6_daddr))
2968 ip6_datagram_dst_update(sk, false);
2969 bh_unlock_sock(sk);
2970}
2971EXPORT_SYMBOL_GPL(ip6_sk_update_pmtu);
2972
2973void ip6_sk_dst_store_flow(struct sock *sk, struct dst_entry *dst,
2974 const struct flowi6 *fl6)
2975{
2976#ifdef CONFIG_IPV6_SUBTREES
2977 struct ipv6_pinfo *np = inet6_sk(sk);
2978#endif
2979
2980 ip6_dst_store(sk, dst,
2981 ipv6_addr_equal(&fl6->daddr, &sk->sk_v6_daddr) ?
2982 &sk->sk_v6_daddr : NULL,
2983#ifdef CONFIG_IPV6_SUBTREES
2984 ipv6_addr_equal(&fl6->saddr, &np->saddr) ?
2985 &np->saddr :
2986#endif
2987 NULL);
2988}
2989
2990static bool ip6_redirect_nh_match(const struct fib6_result *res,
2991 struct flowi6 *fl6,
2992 const struct in6_addr *gw,
2993 struct rt6_info **ret)
2994{
2995 const struct fib6_nh *nh = res->nh;
2996
2997 if (nh->fib_nh_flags & RTNH_F_DEAD || !nh->fib_nh_gw_family ||
2998 fl6->flowi6_oif != nh->fib_nh_dev->ifindex)
2999 return false;
3000
3001 /* rt_cache's gateway might be different from its 'parent'
3002 * in the case of an ip redirect.
3003 * So we keep searching in the exception table if the gateway
3004 * is different.
3005 */
3006 if (!ipv6_addr_equal(gw, &nh->fib_nh_gw6)) {
3007 struct rt6_info *rt_cache;
3008
3009 rt_cache = rt6_find_cached_rt(res, &fl6->daddr, &fl6->saddr);
3010 if (rt_cache &&
3011 ipv6_addr_equal(gw, &rt_cache->rt6i_gateway)) {
3012 *ret = rt_cache;
3013 return true;
3014 }
3015 return false;
3016 }
3017 return true;
3018}
3019
3020struct fib6_nh_rd_arg {
3021 struct fib6_result *res;
3022 struct flowi6 *fl6;
3023 const struct in6_addr *gw;
3024 struct rt6_info **ret;
3025};
3026
3027static int fib6_nh_redirect_match(struct fib6_nh *nh, void *_arg)
3028{
3029 struct fib6_nh_rd_arg *arg = _arg;
3030
3031 arg->res->nh = nh;
3032 return ip6_redirect_nh_match(arg->res, arg->fl6, arg->gw, arg->ret);
3033}
3034
3035/* Handle redirects */
3036struct ip6rd_flowi {
3037 struct flowi6 fl6;
3038 struct in6_addr gateway;
3039};
3040
3041INDIRECT_CALLABLE_SCOPE struct rt6_info *__ip6_route_redirect(struct net *net,
3042 struct fib6_table *table,
3043 struct flowi6 *fl6,
3044 const struct sk_buff *skb,
3045 int flags)
3046{
3047 struct ip6rd_flowi *rdfl = (struct ip6rd_flowi *)fl6;
3048 struct rt6_info *ret = NULL;
3049 struct fib6_result res = {};
3050 struct fib6_nh_rd_arg arg = {
3051 .res = &res,
3052 .fl6 = fl6,
3053 .gw = &rdfl->gateway,
3054 .ret = &ret
3055 };
3056 struct fib6_info *rt;
3057 struct fib6_node *fn;
3058
3059 /* l3mdev_update_flow overrides oif if the device is enslaved; in
3060 * this case we must match on the real ingress device, so reset it
3061 */
3062 if (fl6->flowi6_flags & FLOWI_FLAG_SKIP_NH_OIF)
3063 fl6->flowi6_oif = skb->dev->ifindex;
3064
3065 /* Get the "current" route for this destination and
3066 * check if the redirect has come from appropriate router.
3067 *
3068 * RFC 4861 specifies that redirects should only be
3069 * accepted if they come from the nexthop to the target.
3070 * Due to the way the routes are chosen, this notion
3071 * is a bit fuzzy and one might need to check all possible
3072 * routes.
3073 */
3074
3075 rcu_read_lock();
3076 fn = fib6_node_lookup(&table->tb6_root, &fl6->daddr, &fl6->saddr);
3077restart:
3078 for_each_fib6_node_rt_rcu(fn) {
3079 res.f6i = rt;
3080 if (fib6_check_expired(rt))
3081 continue;
3082 if (rt->fib6_flags & RTF_REJECT)
3083 break;
3084 if (unlikely(rt->nh)) {
3085 if (nexthop_is_blackhole(rt->nh))
3086 continue;
3087 /* on match, res->nh is filled in and potentially ret */
3088 if (nexthop_for_each_fib6_nh(rt->nh,
3089 fib6_nh_redirect_match,
3090 &arg))
3091 goto out;
3092 } else {
3093 res.nh = rt->fib6_nh;
3094 if (ip6_redirect_nh_match(&res, fl6, &rdfl->gateway,
3095 &ret))
3096 goto out;
3097 }
3098 }
3099
3100 if (!rt)
3101 rt = net->ipv6.fib6_null_entry;
3102 else if (rt->fib6_flags & RTF_REJECT) {
3103 ret = net->ipv6.ip6_null_entry;
3104 goto out;
3105 }
3106
3107 if (rt == net->ipv6.fib6_null_entry) {
3108 fn = fib6_backtrack(fn, &fl6->saddr);
3109 if (fn)
3110 goto restart;
3111 }
3112
3113 res.f6i = rt;
3114 res.nh = rt->fib6_nh;
3115out:
3116 if (ret) {
3117 ip6_hold_safe(net, &ret);
3118 } else {
3119 res.fib6_flags = res.f6i->fib6_flags;
3120 res.fib6_type = res.f6i->fib6_type;
3121 ret = ip6_create_rt_rcu(&res);
3122 }
3123
3124 rcu_read_unlock();
3125
3126 trace_fib6_table_lookup(net, &res, table, fl6);
3127 return ret;
3128};
3129
3130static struct dst_entry *ip6_route_redirect(struct net *net,
3131 const struct flowi6 *fl6,
3132 const struct sk_buff *skb,
3133 const struct in6_addr *gateway)
3134{
3135 int flags = RT6_LOOKUP_F_HAS_SADDR;
3136 struct ip6rd_flowi rdfl;
3137
3138 rdfl.fl6 = *fl6;
3139 rdfl.gateway = *gateway;
3140
3141 return fib6_rule_lookup(net, &rdfl.fl6, skb,
3142 flags, __ip6_route_redirect);
3143}
3144
3145void ip6_redirect(struct sk_buff *skb, struct net *net, int oif, u32 mark,
3146 kuid_t uid)
3147{
3148 const struct ipv6hdr *iph = (struct ipv6hdr *) skb->data;
3149 struct dst_entry *dst;
3150 struct flowi6 fl6 = {
3151 .flowi6_iif = LOOPBACK_IFINDEX,
3152 .flowi6_oif = oif,
3153 .flowi6_mark = mark,
3154 .daddr = iph->daddr,
3155 .saddr = iph->saddr,
3156 .flowlabel = ip6_flowinfo(iph),
3157 .flowi6_uid = uid,
3158 };
3159
3160 dst = ip6_route_redirect(net, &fl6, skb, &ipv6_hdr(skb)->saddr);
3161 rt6_do_redirect(dst, NULL, skb);
3162 dst_release(dst);
3163}
3164EXPORT_SYMBOL_GPL(ip6_redirect);
3165
3166void ip6_redirect_no_header(struct sk_buff *skb, struct net *net, int oif)
3167{
3168 const struct ipv6hdr *iph = ipv6_hdr(skb);
3169 const struct rd_msg *msg = (struct rd_msg *)icmp6_hdr(skb);
3170 struct dst_entry *dst;
3171 struct flowi6 fl6 = {
3172 .flowi6_iif = LOOPBACK_IFINDEX,
3173 .flowi6_oif = oif,
3174 .daddr = msg->dest,
3175 .saddr = iph->daddr,
3176 .flowi6_uid = sock_net_uid(net, NULL),
3177 };
3178
3179 dst = ip6_route_redirect(net, &fl6, skb, &iph->saddr);
3180 rt6_do_redirect(dst, NULL, skb);
3181 dst_release(dst);
3182}
3183
3184void ip6_sk_redirect(struct sk_buff *skb, struct sock *sk)
3185{
3186 ip6_redirect(skb, sock_net(sk), sk->sk_bound_dev_if, sk->sk_mark,
3187 sk->sk_uid);
3188}
3189EXPORT_SYMBOL_GPL(ip6_sk_redirect);
3190
3191static unsigned int ip6_default_advmss(const struct dst_entry *dst)
3192{
3193 struct net_device *dev = dst->dev;
3194 unsigned int mtu = dst_mtu(dst);
3195 struct net *net = dev_net(dev);
3196
3197 mtu -= sizeof(struct ipv6hdr) + sizeof(struct tcphdr);
3198
3199 if (mtu < net->ipv6.sysctl.ip6_rt_min_advmss)
3200 mtu = net->ipv6.sysctl.ip6_rt_min_advmss;
3201
3202 /*
3203 * Maximal non-jumbo IPv6 payload is IPV6_MAXPLEN and
3204 * corresponding MSS is IPV6_MAXPLEN - tcp_header_size.
3205 * IPV6_MAXPLEN is also valid and means: "any MSS,
3206 * rely only on pmtu discovery"
3207 */
3208 if (mtu > IPV6_MAXPLEN - sizeof(struct tcphdr))
3209 mtu = IPV6_MAXPLEN;
3210 return mtu;
3211}
3212
3213INDIRECT_CALLABLE_SCOPE unsigned int ip6_mtu(const struct dst_entry *dst)
3214{
3215 struct inet6_dev *idev;
3216 unsigned int mtu;
3217
3218 mtu = dst_metric_raw(dst, RTAX_MTU);
3219 if (mtu)
3220 goto out;
3221
3222 mtu = IPV6_MIN_MTU;
3223
3224 rcu_read_lock();
3225 idev = __in6_dev_get(dst->dev);
3226 if (idev)
3227 mtu = idev->cnf.mtu6;
3228 rcu_read_unlock();
3229
3230out:
3231 mtu = min_t(unsigned int, mtu, IP6_MAX_MTU);
3232
3233 return mtu - lwtunnel_headroom(dst->lwtstate, mtu);
3234}
3235EXPORT_INDIRECT_CALLABLE(ip6_mtu);
3236
3237/* MTU selection:
3238 * 1. mtu on route is locked - use it
3239 * 2. mtu from nexthop exception
3240 * 3. mtu from egress device
3241 *
3242 * based on ip6_dst_mtu_forward and exception logic of
3243 * rt6_find_cached_rt; called with rcu_read_lock
3244 */
3245u32 ip6_mtu_from_fib6(const struct fib6_result *res,
3246 const struct in6_addr *daddr,
3247 const struct in6_addr *saddr)
3248{
3249 const struct fib6_nh *nh = res->nh;
3250 struct fib6_info *f6i = res->f6i;
3251 struct inet6_dev *idev;
3252 struct rt6_info *rt;
3253 u32 mtu = 0;
3254
3255 if (unlikely(fib6_metric_locked(f6i, RTAX_MTU))) {
3256 mtu = f6i->fib6_pmtu;
3257 if (mtu)
3258 goto out;
3259 }
3260
3261 rt = rt6_find_cached_rt(res, daddr, saddr);
3262 if (unlikely(rt)) {
3263 mtu = dst_metric_raw(&rt->dst, RTAX_MTU);
3264 } else {
3265 struct net_device *dev = nh->fib_nh_dev;
3266
3267 mtu = IPV6_MIN_MTU;
3268 idev = __in6_dev_get(dev);
3269 if (idev && idev->cnf.mtu6 > mtu)
3270 mtu = idev->cnf.mtu6;
3271 }
3272
3273 mtu = min_t(unsigned int, mtu, IP6_MAX_MTU);
3274out:
3275 return mtu - lwtunnel_headroom(nh->fib_nh_lws, mtu);
3276}
3277
3278struct dst_entry *icmp6_dst_alloc(struct net_device *dev,
3279 struct flowi6 *fl6)
3280{
3281 struct dst_entry *dst;
3282 struct rt6_info *rt;
3283 struct inet6_dev *idev = in6_dev_get(dev);
3284 struct net *net = dev_net(dev);
3285
3286 if (unlikely(!idev))
3287 return ERR_PTR(-ENODEV);
3288
3289 rt = ip6_dst_alloc(net, dev, 0);
3290 if (unlikely(!rt)) {
3291 in6_dev_put(idev);
3292 dst = ERR_PTR(-ENOMEM);
3293 goto out;
3294 }
3295
3296 rt->dst.input = ip6_input;
3297 rt->dst.output = ip6_output;
3298 rt->rt6i_gateway = fl6->daddr;
3299 rt->rt6i_dst.addr = fl6->daddr;
3300 rt->rt6i_dst.plen = 128;
3301 rt->rt6i_idev = idev;
3302 dst_metric_set(&rt->dst, RTAX_HOPLIMIT, 0);
3303
3304 /* Add this dst into uncached_list so that rt6_disable_ip() can
3305 * do proper release of the net_device
3306 */
3307 rt6_uncached_list_add(rt);
3308 atomic_inc(&net->ipv6.rt6_stats->fib_rt_uncache);
3309
3310 dst = xfrm_lookup(net, &rt->dst, flowi6_to_flowi(fl6), NULL, 0);
3311
3312out:
3313 return dst;
3314}
3315
3316static int ip6_dst_gc(struct dst_ops *ops)
3317{
3318 struct net *net = container_of(ops, struct net, ipv6.ip6_dst_ops);
3319 int rt_min_interval = net->ipv6.sysctl.ip6_rt_gc_min_interval;
3320 int rt_max_size = net->ipv6.sysctl.ip6_rt_max_size;
3321 int rt_elasticity = net->ipv6.sysctl.ip6_rt_gc_elasticity;
3322 int rt_gc_timeout = net->ipv6.sysctl.ip6_rt_gc_timeout;
3323 unsigned long rt_last_gc = net->ipv6.ip6_rt_last_gc;
3324 int entries;
3325
3326 entries = dst_entries_get_fast(ops);
3327 if (entries > rt_max_size)
3328 entries = dst_entries_get_slow(ops);
3329
3330 if (time_after(rt_last_gc + rt_min_interval, jiffies) &&
3331 entries <= rt_max_size)
3332 goto out;
3333
3334 net->ipv6.ip6_rt_gc_expire++;
3335 fib6_run_gc(net->ipv6.ip6_rt_gc_expire, net, true);
3336 entries = dst_entries_get_slow(ops);
3337 if (entries < ops->gc_thresh)
3338 net->ipv6.ip6_rt_gc_expire = rt_gc_timeout>>1;
3339out:
3340 net->ipv6.ip6_rt_gc_expire -= net->ipv6.ip6_rt_gc_expire>>rt_elasticity;
3341 return entries > rt_max_size;
3342}
3343
3344static int ip6_nh_lookup_table(struct net *net, struct fib6_config *cfg,
3345 const struct in6_addr *gw_addr, u32 tbid,
3346 int flags, struct fib6_result *res)
3347{
3348 struct flowi6 fl6 = {
3349 .flowi6_oif = cfg->fc_ifindex,
3350 .daddr = *gw_addr,
3351 .saddr = cfg->fc_prefsrc,
3352 };
3353 struct fib6_table *table;
3354 int err;
3355
3356 table = fib6_get_table(net, tbid);
3357 if (!table)
3358 return -EINVAL;
3359
3360 if (!ipv6_addr_any(&cfg->fc_prefsrc))
3361 flags |= RT6_LOOKUP_F_HAS_SADDR;
3362
3363 flags |= RT6_LOOKUP_F_IGNORE_LINKSTATE;
3364
3365 err = fib6_table_lookup(net, table, cfg->fc_ifindex, &fl6, res, flags);
3366 if (!err && res->f6i != net->ipv6.fib6_null_entry)
3367 fib6_select_path(net, res, &fl6, cfg->fc_ifindex,
3368 cfg->fc_ifindex != 0, NULL, flags);
3369
3370 return err;
3371}
3372
3373static int ip6_route_check_nh_onlink(struct net *net,
3374 struct fib6_config *cfg,
3375 const struct net_device *dev,
3376 struct netlink_ext_ack *extack)
3377{
3378 u32 tbid = l3mdev_fib_table_rcu(dev) ? : RT_TABLE_MAIN;
3379 const struct in6_addr *gw_addr = &cfg->fc_gateway;
3380 struct fib6_result res = {};
3381 int err;
3382
3383 err = ip6_nh_lookup_table(net, cfg, gw_addr, tbid, 0, &res);
3384 if (!err && !(res.fib6_flags & RTF_REJECT) &&
3385 /* ignore match if it is the default route */
3386 !ipv6_addr_any(&res.f6i->fib6_dst.addr) &&
3387 (res.fib6_type != RTN_UNICAST || dev != res.nh->fib_nh_dev)) {
3388 NL_SET_ERR_MSG(extack,
3389 "Nexthop has invalid gateway or device mismatch");
3390 err = -EINVAL;
3391 }
3392
3393 return err;
3394}
3395
3396static int ip6_route_check_nh(struct net *net,
3397 struct fib6_config *cfg,
3398 struct net_device **_dev,
3399 struct inet6_dev **idev)
3400{
3401 const struct in6_addr *gw_addr = &cfg->fc_gateway;
3402 struct net_device *dev = _dev ? *_dev : NULL;
3403 int flags = RT6_LOOKUP_F_IFACE;
3404 struct fib6_result res = {};
3405 int err = -EHOSTUNREACH;
3406
3407 if (cfg->fc_table) {
3408 err = ip6_nh_lookup_table(net, cfg, gw_addr,
3409 cfg->fc_table, flags, &res);
3410 /* gw_addr can not require a gateway or resolve to a reject
3411 * route. If a device is given, it must match the result.
3412 */
3413 if (err || res.fib6_flags & RTF_REJECT ||
3414 res.nh->fib_nh_gw_family ||
3415 (dev && dev != res.nh->fib_nh_dev))
3416 err = -EHOSTUNREACH;
3417 }
3418
3419 if (err < 0) {
3420 struct flowi6 fl6 = {
3421 .flowi6_oif = cfg->fc_ifindex,
3422 .daddr = *gw_addr,
3423 };
3424
3425 err = fib6_lookup(net, cfg->fc_ifindex, &fl6, &res, flags);
3426 if (err || res.fib6_flags & RTF_REJECT ||
3427 res.nh->fib_nh_gw_family)
3428 err = -EHOSTUNREACH;
3429
3430 if (err)
3431 return err;
3432
3433 fib6_select_path(net, &res, &fl6, cfg->fc_ifindex,
3434 cfg->fc_ifindex != 0, NULL, flags);
3435 }
3436
3437 err = 0;
3438 if (dev) {
3439 if (dev != res.nh->fib_nh_dev)
3440 err = -EHOSTUNREACH;
3441 } else {
3442 *_dev = dev = res.nh->fib_nh_dev;
3443 dev_hold(dev);
3444 *idev = in6_dev_get(dev);
3445 }
3446
3447 return err;
3448}
3449
3450static int ip6_validate_gw(struct net *net, struct fib6_config *cfg,
3451 struct net_device **_dev, struct inet6_dev **idev,
3452 struct netlink_ext_ack *extack)
3453{
3454 const struct in6_addr *gw_addr = &cfg->fc_gateway;
3455 int gwa_type = ipv6_addr_type(gw_addr);
3456 bool skip_dev = gwa_type & IPV6_ADDR_LINKLOCAL ? false : true;
3457 const struct net_device *dev = *_dev;
3458 bool need_addr_check = !dev;
3459 int err = -EINVAL;
3460
3461 /* if gw_addr is local we will fail to detect this in case
3462 * address is still TENTATIVE (DAD in progress). rt6_lookup()
3463 * will return already-added prefix route via interface that
3464 * prefix route was assigned to, which might be non-loopback.
3465 */
3466 if (dev &&
3467 ipv6_chk_addr_and_flags(net, gw_addr, dev, skip_dev, 0, 0)) {
3468 NL_SET_ERR_MSG(extack, "Gateway can not be a local address");
3469 goto out;
3470 }
3471
3472 if (gwa_type != (IPV6_ADDR_LINKLOCAL | IPV6_ADDR_UNICAST)) {
3473 /* IPv6 strictly inhibits using not link-local
3474 * addresses as nexthop address.
3475 * Otherwise, router will not able to send redirects.
3476 * It is very good, but in some (rare!) circumstances
3477 * (SIT, PtP, NBMA NOARP links) it is handy to allow
3478 * some exceptions. --ANK
3479 * We allow IPv4-mapped nexthops to support RFC4798-type
3480 * addressing
3481 */
3482 if (!(gwa_type & (IPV6_ADDR_UNICAST | IPV6_ADDR_MAPPED))) {
3483 NL_SET_ERR_MSG(extack, "Invalid gateway address");
3484 goto out;
3485 }
3486
3487 rcu_read_lock();
3488
3489 if (cfg->fc_flags & RTNH_F_ONLINK)
3490 err = ip6_route_check_nh_onlink(net, cfg, dev, extack);
3491 else
3492 err = ip6_route_check_nh(net, cfg, _dev, idev);
3493
3494 rcu_read_unlock();
3495
3496 if (err)
3497 goto out;
3498 }
3499
3500 /* reload in case device was changed */
3501 dev = *_dev;
3502
3503 err = -EINVAL;
3504 if (!dev) {
3505 NL_SET_ERR_MSG(extack, "Egress device not specified");
3506 goto out;
3507 } else if (dev->flags & IFF_LOOPBACK) {
3508 NL_SET_ERR_MSG(extack,
3509 "Egress device can not be loopback device for this route");
3510 goto out;
3511 }
3512
3513 /* if we did not check gw_addr above, do so now that the
3514 * egress device has been resolved.
3515 */
3516 if (need_addr_check &&
3517 ipv6_chk_addr_and_flags(net, gw_addr, dev, skip_dev, 0, 0)) {
3518 NL_SET_ERR_MSG(extack, "Gateway can not be a local address");
3519 goto out;
3520 }
3521
3522 err = 0;
3523out:
3524 return err;
3525}
3526
3527static bool fib6_is_reject(u32 flags, struct net_device *dev, int addr_type)
3528{
3529 if ((flags & RTF_REJECT) ||
3530 (dev && (dev->flags & IFF_LOOPBACK) &&
3531 !(addr_type & IPV6_ADDR_LOOPBACK) &&
3532 !(flags & (RTF_ANYCAST | RTF_LOCAL))))
3533 return true;
3534
3535 return false;
3536}
3537
3538int fib6_nh_init(struct net *net, struct fib6_nh *fib6_nh,
3539 struct fib6_config *cfg, gfp_t gfp_flags,
3540 struct netlink_ext_ack *extack)
3541{
3542 struct net_device *dev = NULL;
3543 struct inet6_dev *idev = NULL;
3544 int addr_type;
3545 int err;
3546
3547 fib6_nh->fib_nh_family = AF_INET6;
3548#ifdef CONFIG_IPV6_ROUTER_PREF
3549 fib6_nh->last_probe = jiffies;
3550#endif
3551 if (cfg->fc_is_fdb) {
3552 fib6_nh->fib_nh_gw6 = cfg->fc_gateway;
3553 fib6_nh->fib_nh_gw_family = AF_INET6;
3554 return 0;
3555 }
3556
3557 err = -ENODEV;
3558 if (cfg->fc_ifindex) {
3559 dev = dev_get_by_index(net, cfg->fc_ifindex);
3560 if (!dev)
3561 goto out;
3562 idev = in6_dev_get(dev);
3563 if (!idev)
3564 goto out;
3565 }
3566
3567 if (cfg->fc_flags & RTNH_F_ONLINK) {
3568 if (!dev) {
3569 NL_SET_ERR_MSG(extack,
3570 "Nexthop device required for onlink");
3571 goto out;
3572 }
3573
3574 if (!(dev->flags & IFF_UP)) {
3575 NL_SET_ERR_MSG(extack, "Nexthop device is not up");
3576 err = -ENETDOWN;
3577 goto out;
3578 }
3579
3580 fib6_nh->fib_nh_flags |= RTNH_F_ONLINK;
3581 }
3582
3583 fib6_nh->fib_nh_weight = 1;
3584
3585 /* We cannot add true routes via loopback here,
3586 * they would result in kernel looping; promote them to reject routes
3587 */
3588 addr_type = ipv6_addr_type(&cfg->fc_dst);
3589 if (fib6_is_reject(cfg->fc_flags, dev, addr_type)) {
3590 /* hold loopback dev/idev if we haven't done so. */
3591 if (dev != net->loopback_dev) {
3592 if (dev) {
3593 dev_put(dev);
3594 in6_dev_put(idev);
3595 }
3596 dev = net->loopback_dev;
3597 dev_hold(dev);
3598 idev = in6_dev_get(dev);
3599 if (!idev) {
3600 err = -ENODEV;
3601 goto out;
3602 }
3603 }
3604 goto pcpu_alloc;
3605 }
3606
3607 if (cfg->fc_flags & RTF_GATEWAY) {
3608 err = ip6_validate_gw(net, cfg, &dev, &idev, extack);
3609 if (err)
3610 goto out;
3611
3612 fib6_nh->fib_nh_gw6 = cfg->fc_gateway;
3613 fib6_nh->fib_nh_gw_family = AF_INET6;
3614 }
3615
3616 err = -ENODEV;
3617 if (!dev)
3618 goto out;
3619
3620 if (idev->cnf.disable_ipv6) {
3621 NL_SET_ERR_MSG(extack, "IPv6 is disabled on nexthop device");
3622 err = -EACCES;
3623 goto out;
3624 }
3625
3626 if (!(dev->flags & IFF_UP) && !cfg->fc_ignore_dev_down) {
3627 NL_SET_ERR_MSG(extack, "Nexthop device is not up");
3628 err = -ENETDOWN;
3629 goto out;
3630 }
3631
3632 if (!(cfg->fc_flags & (RTF_LOCAL | RTF_ANYCAST)) &&
3633 !netif_carrier_ok(dev))
3634 fib6_nh->fib_nh_flags |= RTNH_F_LINKDOWN;
3635
3636 err = fib_nh_common_init(net, &fib6_nh->nh_common, cfg->fc_encap,
3637 cfg->fc_encap_type, cfg, gfp_flags, extack);
3638 if (err)
3639 goto out;
3640
3641pcpu_alloc:
3642 fib6_nh->rt6i_pcpu = alloc_percpu_gfp(struct rt6_info *, gfp_flags);
3643 if (!fib6_nh->rt6i_pcpu) {
3644 err = -ENOMEM;
3645 goto out;
3646 }
3647
3648 fib6_nh->fib_nh_dev = dev;
3649 fib6_nh->fib_nh_oif = dev->ifindex;
3650 err = 0;
3651out:
3652 if (idev)
3653 in6_dev_put(idev);
3654
3655 if (err) {
3656 lwtstate_put(fib6_nh->fib_nh_lws);
3657 fib6_nh->fib_nh_lws = NULL;
3658 if (dev)
3659 dev_put(dev);
3660 }
3661
3662 return err;
3663}
3664
3665void fib6_nh_release(struct fib6_nh *fib6_nh)
3666{
3667 struct rt6_exception_bucket *bucket;
3668
3669 rcu_read_lock();
3670
3671 fib6_nh_flush_exceptions(fib6_nh, NULL);
3672 bucket = fib6_nh_get_excptn_bucket(fib6_nh, NULL);
3673 if (bucket) {
3674 rcu_assign_pointer(fib6_nh->rt6i_exception_bucket, NULL);
3675 kfree(bucket);
3676 }
3677
3678 rcu_read_unlock();
3679
3680 if (fib6_nh->rt6i_pcpu) {
3681 int cpu;
3682
3683 for_each_possible_cpu(cpu) {
3684 struct rt6_info **ppcpu_rt;
3685 struct rt6_info *pcpu_rt;
3686
3687 ppcpu_rt = per_cpu_ptr(fib6_nh->rt6i_pcpu, cpu);
3688 pcpu_rt = *ppcpu_rt;
3689 if (pcpu_rt) {
3690 dst_dev_put(&pcpu_rt->dst);
3691 dst_release(&pcpu_rt->dst);
3692 *ppcpu_rt = NULL;
3693 }
3694 }
3695
3696 free_percpu(fib6_nh->rt6i_pcpu);
3697 }
3698
3699 fib_nh_common_release(&fib6_nh->nh_common);
3700}
3701
3702static struct fib6_info *ip6_route_info_create(struct fib6_config *cfg,
3703 gfp_t gfp_flags,
3704 struct netlink_ext_ack *extack)
3705{
3706 struct net *net = cfg->fc_nlinfo.nl_net;
3707 struct fib6_info *rt = NULL;
3708 struct nexthop *nh = NULL;
3709 struct fib6_table *table;
3710 struct fib6_nh *fib6_nh;
3711 int err = -EINVAL;
3712 int addr_type;
3713
3714 /* RTF_PCPU is an internal flag; can not be set by userspace */
3715 if (cfg->fc_flags & RTF_PCPU) {
3716 NL_SET_ERR_MSG(extack, "Userspace can not set RTF_PCPU");
3717 goto out;
3718 }
3719
3720 /* RTF_CACHE is an internal flag; can not be set by userspace */
3721 if (cfg->fc_flags & RTF_CACHE) {
3722 NL_SET_ERR_MSG(extack, "Userspace can not set RTF_CACHE");
3723 goto out;
3724 }
3725
3726 if (cfg->fc_type > RTN_MAX) {
3727 NL_SET_ERR_MSG(extack, "Invalid route type");
3728 goto out;
3729 }
3730
3731 if (cfg->fc_dst_len > 128) {
3732 NL_SET_ERR_MSG(extack, "Invalid prefix length");
3733 goto out;
3734 }
3735 if (cfg->fc_src_len > 128) {
3736 NL_SET_ERR_MSG(extack, "Invalid source address length");
3737 goto out;
3738 }
3739#ifndef CONFIG_IPV6_SUBTREES
3740 if (cfg->fc_src_len) {
3741 NL_SET_ERR_MSG(extack,
3742 "Specifying source address requires IPV6_SUBTREES to be enabled");
3743 goto out;
3744 }
3745#endif
3746 if (cfg->fc_nh_id) {
3747 nh = nexthop_find_by_id(net, cfg->fc_nh_id);
3748 if (!nh) {
3749 NL_SET_ERR_MSG(extack, "Nexthop id does not exist");
3750 goto out;
3751 }
3752 err = fib6_check_nexthop(nh, cfg, extack);
3753 if (err)
3754 goto out;
3755 }
3756
3757 err = -ENOBUFS;
3758 if (cfg->fc_nlinfo.nlh &&
3759 !(cfg->fc_nlinfo.nlh->nlmsg_flags & NLM_F_CREATE)) {
3760 table = fib6_get_table(net, cfg->fc_table);
3761 if (!table) {
3762 pr_warn("NLM_F_CREATE should be specified when creating new route\n");
3763 table = fib6_new_table(net, cfg->fc_table);
3764 }
3765 } else {
3766 table = fib6_new_table(net, cfg->fc_table);
3767 }
3768
3769 if (!table)
3770 goto out;
3771
3772 err = -ENOMEM;
3773 rt = fib6_info_alloc(gfp_flags, !nh);
3774 if (!rt)
3775 goto out;
3776
3777 rt->fib6_metrics = ip_fib_metrics_init(net, cfg->fc_mx, cfg->fc_mx_len,
3778 extack);
3779 if (IS_ERR(rt->fib6_metrics)) {
3780 err = PTR_ERR(rt->fib6_metrics);
3781 /* Do not leave garbage there. */
3782 rt->fib6_metrics = (struct dst_metrics *)&dst_default_metrics;
3783 goto out_free;
3784 }
3785
3786 if (cfg->fc_flags & RTF_ADDRCONF)
3787 rt->dst_nocount = true;
3788
3789 if (cfg->fc_flags & RTF_EXPIRES)
3790 fib6_set_expires(rt, jiffies +
3791 clock_t_to_jiffies(cfg->fc_expires));
3792 else
3793 fib6_clean_expires(rt);
3794
3795 if (cfg->fc_protocol == RTPROT_UNSPEC)
3796 cfg->fc_protocol = RTPROT_BOOT;
3797 rt->fib6_protocol = cfg->fc_protocol;
3798
3799 rt->fib6_table = table;
3800 rt->fib6_metric = cfg->fc_metric;
3801 rt->fib6_type = cfg->fc_type ? : RTN_UNICAST;
3802 rt->fib6_flags = cfg->fc_flags & ~RTF_GATEWAY;
3803
3804 ipv6_addr_prefix(&rt->fib6_dst.addr, &cfg->fc_dst, cfg->fc_dst_len);
3805 rt->fib6_dst.plen = cfg->fc_dst_len;
3806
3807#ifdef CONFIG_IPV6_SUBTREES
3808 ipv6_addr_prefix(&rt->fib6_src.addr, &cfg->fc_src, cfg->fc_src_len);
3809 rt->fib6_src.plen = cfg->fc_src_len;
3810#endif
3811 if (nh) {
3812 if (rt->fib6_src.plen) {
3813 NL_SET_ERR_MSG(extack, "Nexthops can not be used with source routing");
3814 goto out_free;
3815 }
3816 if (!nexthop_get(nh)) {
3817 NL_SET_ERR_MSG(extack, "Nexthop has been deleted");
3818 goto out_free;
3819 }
3820 rt->nh = nh;
3821 fib6_nh = nexthop_fib6_nh(rt->nh);
3822 } else {
3823 err = fib6_nh_init(net, rt->fib6_nh, cfg, gfp_flags, extack);
3824 if (err)
3825 goto out;
3826
3827 fib6_nh = rt->fib6_nh;
3828
3829 /* We cannot add true routes via loopback here, they would
3830 * result in kernel looping; promote them to reject routes
3831 */
3832 addr_type = ipv6_addr_type(&cfg->fc_dst);
3833 if (fib6_is_reject(cfg->fc_flags, rt->fib6_nh->fib_nh_dev,
3834 addr_type))
3835 rt->fib6_flags = RTF_REJECT | RTF_NONEXTHOP;
3836 }
3837
3838 if (!ipv6_addr_any(&cfg->fc_prefsrc)) {
3839 struct net_device *dev = fib6_nh->fib_nh_dev;
3840
3841 if (!ipv6_chk_addr(net, &cfg->fc_prefsrc, dev, 0)) {
3842 NL_SET_ERR_MSG(extack, "Invalid source address");
3843 err = -EINVAL;
3844 goto out;
3845 }
3846 rt->fib6_prefsrc.addr = cfg->fc_prefsrc;
3847 rt->fib6_prefsrc.plen = 128;
3848 } else
3849 rt->fib6_prefsrc.plen = 0;
3850
3851 return rt;
3852out:
3853 fib6_info_release(rt);
3854 return ERR_PTR(err);
3855out_free:
3856 ip_fib_metrics_put(rt->fib6_metrics);
3857 kfree(rt);
3858 return ERR_PTR(err);
3859}
3860
3861int ip6_route_add(struct fib6_config *cfg, gfp_t gfp_flags,
3862 struct netlink_ext_ack *extack)
3863{
3864 struct fib6_info *rt;
3865 int err;
3866
3867 rt = ip6_route_info_create(cfg, gfp_flags, extack);
3868 if (IS_ERR(rt))
3869 return PTR_ERR(rt);
3870
3871 err = __ip6_ins_rt(rt, &cfg->fc_nlinfo, extack);
3872 fib6_info_release(rt);
3873
3874 return err;
3875}
3876
3877static int __ip6_del_rt(struct fib6_info *rt, struct nl_info *info)
3878{
3879 struct net *net = info->nl_net;
3880 struct fib6_table *table;
3881 int err;
3882
3883 if (rt == net->ipv6.fib6_null_entry) {
3884 err = -ENOENT;
3885 goto out;
3886 }
3887
3888 table = rt->fib6_table;
3889 spin_lock_bh(&table->tb6_lock);
3890 err = fib6_del(rt, info);
3891 spin_unlock_bh(&table->tb6_lock);
3892
3893out:
3894 fib6_info_release(rt);
3895 return err;
3896}
3897
3898int ip6_del_rt(struct net *net, struct fib6_info *rt, bool skip_notify)
3899{
3900 struct nl_info info = {
3901 .nl_net = net,
3902 .skip_notify = skip_notify
3903 };
3904
3905 return __ip6_del_rt(rt, &info);
3906}
3907
3908static int __ip6_del_rt_siblings(struct fib6_info *rt, struct fib6_config *cfg)
3909{
3910 struct nl_info *info = &cfg->fc_nlinfo;
3911 struct net *net = info->nl_net;
3912 struct sk_buff *skb = NULL;
3913 struct fib6_table *table;
3914 int err = -ENOENT;
3915
3916 if (rt == net->ipv6.fib6_null_entry)
3917 goto out_put;
3918 table = rt->fib6_table;
3919 spin_lock_bh(&table->tb6_lock);
3920
3921 if (rt->fib6_nsiblings && cfg->fc_delete_all_nh) {
3922 struct fib6_info *sibling, *next_sibling;
3923 struct fib6_node *fn;
3924
3925 /* prefer to send a single notification with all hops */
3926 skb = nlmsg_new(rt6_nlmsg_size(rt), gfp_any());
3927 if (skb) {
3928 u32 seq = info->nlh ? info->nlh->nlmsg_seq : 0;
3929
3930 if (rt6_fill_node(net, skb, rt, NULL,
3931 NULL, NULL, 0, RTM_DELROUTE,
3932 info->portid, seq, 0) < 0) {
3933 kfree_skb(skb);
3934 skb = NULL;
3935 } else
3936 info->skip_notify = 1;
3937 }
3938
3939 /* 'rt' points to the first sibling route. If it is not the
3940 * leaf, then we do not need to send a notification. Otherwise,
3941 * we need to check if the last sibling has a next route or not
3942 * and emit a replace or delete notification, respectively.
3943 */
3944 info->skip_notify_kernel = 1;
3945 fn = rcu_dereference_protected(rt->fib6_node,
3946 lockdep_is_held(&table->tb6_lock));
3947 if (rcu_access_pointer(fn->leaf) == rt) {
3948 struct fib6_info *last_sibling, *replace_rt;
3949
3950 last_sibling = list_last_entry(&rt->fib6_siblings,
3951 struct fib6_info,
3952 fib6_siblings);
3953 replace_rt = rcu_dereference_protected(
3954 last_sibling->fib6_next,
3955 lockdep_is_held(&table->tb6_lock));
3956 if (replace_rt)
3957 call_fib6_entry_notifiers_replace(net,
3958 replace_rt);
3959 else
3960 call_fib6_multipath_entry_notifiers(net,
3961 FIB_EVENT_ENTRY_DEL,
3962 rt, rt->fib6_nsiblings,
3963 NULL);
3964 }
3965 list_for_each_entry_safe(sibling, next_sibling,
3966 &rt->fib6_siblings,
3967 fib6_siblings) {
3968 err = fib6_del(sibling, info);
3969 if (err)
3970 goto out_unlock;
3971 }
3972 }
3973
3974 err = fib6_del(rt, info);
3975out_unlock:
3976 spin_unlock_bh(&table->tb6_lock);
3977out_put:
3978 fib6_info_release(rt);
3979
3980 if (skb) {
3981 rtnl_notify(skb, net, info->portid, RTNLGRP_IPV6_ROUTE,
3982 info->nlh, gfp_any());
3983 }
3984 return err;
3985}
3986
3987static int __ip6_del_cached_rt(struct rt6_info *rt, struct fib6_config *cfg)
3988{
3989 int rc = -ESRCH;
3990
3991 if (cfg->fc_ifindex && rt->dst.dev->ifindex != cfg->fc_ifindex)
3992 goto out;
3993
3994 if (cfg->fc_flags & RTF_GATEWAY &&
3995 !ipv6_addr_equal(&cfg->fc_gateway, &rt->rt6i_gateway))
3996 goto out;
3997
3998 rc = rt6_remove_exception_rt(rt);
3999out:
4000 return rc;
4001}
4002
4003static int ip6_del_cached_rt(struct fib6_config *cfg, struct fib6_info *rt,
4004 struct fib6_nh *nh)
4005{
4006 struct fib6_result res = {
4007 .f6i = rt,
4008 .nh = nh,
4009 };
4010 struct rt6_info *rt_cache;
4011
4012 rt_cache = rt6_find_cached_rt(&res, &cfg->fc_dst, &cfg->fc_src);
4013 if (rt_cache)
4014 return __ip6_del_cached_rt(rt_cache, cfg);
4015
4016 return 0;
4017}
4018
4019struct fib6_nh_del_cached_rt_arg {
4020 struct fib6_config *cfg;
4021 struct fib6_info *f6i;
4022};
4023
4024static int fib6_nh_del_cached_rt(struct fib6_nh *nh, void *_arg)
4025{
4026 struct fib6_nh_del_cached_rt_arg *arg = _arg;
4027 int rc;
4028
4029 rc = ip6_del_cached_rt(arg->cfg, arg->f6i, nh);
4030 return rc != -ESRCH ? rc : 0;
4031}
4032
4033static int ip6_del_cached_rt_nh(struct fib6_config *cfg, struct fib6_info *f6i)
4034{
4035 struct fib6_nh_del_cached_rt_arg arg = {
4036 .cfg = cfg,
4037 .f6i = f6i
4038 };
4039
4040 return nexthop_for_each_fib6_nh(f6i->nh, fib6_nh_del_cached_rt, &arg);
4041}
4042
4043static int ip6_route_del(struct fib6_config *cfg,
4044 struct netlink_ext_ack *extack)
4045{
4046 struct fib6_table *table;
4047 struct fib6_info *rt;
4048 struct fib6_node *fn;
4049 int err = -ESRCH;
4050
4051 table = fib6_get_table(cfg->fc_nlinfo.nl_net, cfg->fc_table);
4052 if (!table) {
4053 NL_SET_ERR_MSG(extack, "FIB table does not exist");
4054 return err;
4055 }
4056
4057 rcu_read_lock();
4058
4059 fn = fib6_locate(&table->tb6_root,
4060 &cfg->fc_dst, cfg->fc_dst_len,
4061 &cfg->fc_src, cfg->fc_src_len,
4062 !(cfg->fc_flags & RTF_CACHE));
4063
4064 if (fn) {
4065 for_each_fib6_node_rt_rcu(fn) {
4066 struct fib6_nh *nh;
4067
4068 if (rt->nh && cfg->fc_nh_id &&
4069 rt->nh->id != cfg->fc_nh_id)
4070 continue;
4071
4072 if (cfg->fc_flags & RTF_CACHE) {
4073 int rc = 0;
4074
4075 if (rt->nh) {
4076 rc = ip6_del_cached_rt_nh(cfg, rt);
4077 } else if (cfg->fc_nh_id) {
4078 continue;
4079 } else {
4080 nh = rt->fib6_nh;
4081 rc = ip6_del_cached_rt(cfg, rt, nh);
4082 }
4083 if (rc != -ESRCH) {
4084 rcu_read_unlock();
4085 return rc;
4086 }
4087 continue;
4088 }
4089
4090 if (cfg->fc_metric && cfg->fc_metric != rt->fib6_metric)
4091 continue;
4092 if (cfg->fc_protocol &&
4093 cfg->fc_protocol != rt->fib6_protocol)
4094 continue;
4095
4096 if (rt->nh) {
4097 if (!fib6_info_hold_safe(rt))
4098 continue;
4099 rcu_read_unlock();
4100
4101 return __ip6_del_rt(rt, &cfg->fc_nlinfo);
4102 }
4103 if (cfg->fc_nh_id)
4104 continue;
4105
4106 nh = rt->fib6_nh;
4107 if (cfg->fc_ifindex &&
4108 (!nh->fib_nh_dev ||
4109 nh->fib_nh_dev->ifindex != cfg->fc_ifindex))
4110 continue;
4111 if (cfg->fc_flags & RTF_GATEWAY &&
4112 !ipv6_addr_equal(&cfg->fc_gateway, &nh->fib_nh_gw6))
4113 continue;
4114 if (!fib6_info_hold_safe(rt))
4115 continue;
4116 rcu_read_unlock();
4117
4118 /* if gateway was specified only delete the one hop */
4119 if (cfg->fc_flags & RTF_GATEWAY)
4120 return __ip6_del_rt(rt, &cfg->fc_nlinfo);
4121
4122 return __ip6_del_rt_siblings(rt, cfg);
4123 }
4124 }
4125 rcu_read_unlock();
4126
4127 return err;
4128}
4129
4130static void rt6_do_redirect(struct dst_entry *dst, struct sock *sk, struct sk_buff *skb)
4131{
4132 struct netevent_redirect netevent;
4133 struct rt6_info *rt, *nrt = NULL;
4134 struct fib6_result res = {};
4135 struct ndisc_options ndopts;
4136 struct inet6_dev *in6_dev;
4137 struct neighbour *neigh;
4138 struct rd_msg *msg;
4139 int optlen, on_link;
4140 u8 *lladdr;
4141
4142 optlen = skb_tail_pointer(skb) - skb_transport_header(skb);
4143 optlen -= sizeof(*msg);
4144
4145 if (optlen < 0) {
4146 net_dbg_ratelimited("rt6_do_redirect: packet too short\n");
4147 return;
4148 }
4149
4150 msg = (struct rd_msg *)icmp6_hdr(skb);
4151
4152 if (ipv6_addr_is_multicast(&msg->dest)) {
4153 net_dbg_ratelimited("rt6_do_redirect: destination address is multicast\n");
4154 return;
4155 }
4156
4157 on_link = 0;
4158 if (ipv6_addr_equal(&msg->dest, &msg->target)) {
4159 on_link = 1;
4160 } else if (ipv6_addr_type(&msg->target) !=
4161 (IPV6_ADDR_UNICAST|IPV6_ADDR_LINKLOCAL)) {
4162 net_dbg_ratelimited("rt6_do_redirect: target address is not link-local unicast\n");
4163 return;
4164 }
4165
4166 in6_dev = __in6_dev_get(skb->dev);
4167 if (!in6_dev)
4168 return;
4169 if (in6_dev->cnf.forwarding || !in6_dev->cnf.accept_redirects)
4170 return;
4171
4172 /* RFC2461 8.1:
4173 * The IP source address of the Redirect MUST be the same as the current
4174 * first-hop router for the specified ICMP Destination Address.
4175 */
4176
4177 if (!ndisc_parse_options(skb->dev, msg->opt, optlen, &ndopts)) {
4178 net_dbg_ratelimited("rt6_redirect: invalid ND options\n");
4179 return;
4180 }
4181
4182 lladdr = NULL;
4183 if (ndopts.nd_opts_tgt_lladdr) {
4184 lladdr = ndisc_opt_addr_data(ndopts.nd_opts_tgt_lladdr,
4185 skb->dev);
4186 if (!lladdr) {
4187 net_dbg_ratelimited("rt6_redirect: invalid link-layer address length\n");
4188 return;
4189 }
4190 }
4191
4192 rt = (struct rt6_info *) dst;
4193 if (rt->rt6i_flags & RTF_REJECT) {
4194 net_dbg_ratelimited("rt6_redirect: source isn't a valid nexthop for redirect target\n");
4195 return;
4196 }
4197
4198 /* Redirect received -> path was valid.
4199 * Look, redirects are sent only in response to data packets,
4200 * so that this nexthop apparently is reachable. --ANK
4201 */
4202 dst_confirm_neigh(&rt->dst, &ipv6_hdr(skb)->saddr);
4203
4204 neigh = __neigh_lookup(&nd_tbl, &msg->target, skb->dev, 1);
4205 if (!neigh)
4206 return;
4207
4208 /*
4209 * We have finally decided to accept it.
4210 */
4211
4212 ndisc_update(skb->dev, neigh, lladdr, NUD_STALE,
4213 NEIGH_UPDATE_F_WEAK_OVERRIDE|
4214 NEIGH_UPDATE_F_OVERRIDE|
4215 (on_link ? 0 : (NEIGH_UPDATE_F_OVERRIDE_ISROUTER|
4216 NEIGH_UPDATE_F_ISROUTER)),
4217 NDISC_REDIRECT, &ndopts);
4218
4219 rcu_read_lock();
4220 res.f6i = rcu_dereference(rt->from);
4221 if (!res.f6i)
4222 goto out;
4223
4224 if (res.f6i->nh) {
4225 struct fib6_nh_match_arg arg = {
4226 .dev = dst->dev,
4227 .gw = &rt->rt6i_gateway,
4228 };
4229
4230 nexthop_for_each_fib6_nh(res.f6i->nh,
4231 fib6_nh_find_match, &arg);
4232
4233 /* fib6_info uses a nexthop that does not have fib6_nh
4234 * using the dst->dev. Should be impossible
4235 */
4236 if (!arg.match)
4237 goto out;
4238 res.nh = arg.match;
4239 } else {
4240 res.nh = res.f6i->fib6_nh;
4241 }
4242
4243 res.fib6_flags = res.f6i->fib6_flags;
4244 res.fib6_type = res.f6i->fib6_type;
4245 nrt = ip6_rt_cache_alloc(&res, &msg->dest, NULL);
4246 if (!nrt)
4247 goto out;
4248
4249 nrt->rt6i_flags = RTF_GATEWAY|RTF_UP|RTF_DYNAMIC|RTF_CACHE;
4250 if (on_link)
4251 nrt->rt6i_flags &= ~RTF_GATEWAY;
4252
4253 nrt->rt6i_gateway = *(struct in6_addr *)neigh->primary_key;
4254
4255 /* rt6_insert_exception() will take care of duplicated exceptions */
4256 if (rt6_insert_exception(nrt, &res)) {
4257 dst_release_immediate(&nrt->dst);
4258 goto out;
4259 }
4260
4261 netevent.old = &rt->dst;
4262 netevent.new = &nrt->dst;
4263 netevent.daddr = &msg->dest;
4264 netevent.neigh = neigh;
4265 call_netevent_notifiers(NETEVENT_REDIRECT, &netevent);
4266
4267out:
4268 rcu_read_unlock();
4269 neigh_release(neigh);
4270}
4271
4272#ifdef CONFIG_IPV6_ROUTE_INFO
4273static struct fib6_info *rt6_get_route_info(struct net *net,
4274 const struct in6_addr *prefix, int prefixlen,
4275 const struct in6_addr *gwaddr,
4276 struct net_device *dev)
4277{
4278 u32 tb_id = l3mdev_fib_table(dev) ? : RT6_TABLE_INFO;
4279 int ifindex = dev->ifindex;
4280 struct fib6_node *fn;
4281 struct fib6_info *rt = NULL;
4282 struct fib6_table *table;
4283
4284 table = fib6_get_table(net, tb_id);
4285 if (!table)
4286 return NULL;
4287
4288 rcu_read_lock();
4289 fn = fib6_locate(&table->tb6_root, prefix, prefixlen, NULL, 0, true);
4290 if (!fn)
4291 goto out;
4292
4293 for_each_fib6_node_rt_rcu(fn) {
4294 /* these routes do not use nexthops */
4295 if (rt->nh)
4296 continue;
4297 if (rt->fib6_nh->fib_nh_dev->ifindex != ifindex)
4298 continue;
4299 if (!(rt->fib6_flags & RTF_ROUTEINFO) ||
4300 !rt->fib6_nh->fib_nh_gw_family)
4301 continue;
4302 if (!ipv6_addr_equal(&rt->fib6_nh->fib_nh_gw6, gwaddr))
4303 continue;
4304 if (!fib6_info_hold_safe(rt))
4305 continue;
4306 break;
4307 }
4308out:
4309 rcu_read_unlock();
4310 return rt;
4311}
4312
4313static struct fib6_info *rt6_add_route_info(struct net *net,
4314 const struct in6_addr *prefix, int prefixlen,
4315 const struct in6_addr *gwaddr,
4316 struct net_device *dev,
4317 unsigned int pref)
4318{
4319 struct fib6_config cfg = {
4320 .fc_metric = IP6_RT_PRIO_USER,
4321 .fc_ifindex = dev->ifindex,
4322 .fc_dst_len = prefixlen,
4323 .fc_flags = RTF_GATEWAY | RTF_ADDRCONF | RTF_ROUTEINFO |
4324 RTF_UP | RTF_PREF(pref),
4325 .fc_protocol = RTPROT_RA,
4326 .fc_type = RTN_UNICAST,
4327 .fc_nlinfo.portid = 0,
4328 .fc_nlinfo.nlh = NULL,
4329 .fc_nlinfo.nl_net = net,
4330 };
4331
4332 cfg.fc_table = l3mdev_fib_table(dev) ? : RT6_TABLE_INFO;
4333 cfg.fc_dst = *prefix;
4334 cfg.fc_gateway = *gwaddr;
4335
4336 /* We should treat it as a default route if prefix length is 0. */
4337 if (!prefixlen)
4338 cfg.fc_flags |= RTF_DEFAULT;
4339
4340 ip6_route_add(&cfg, GFP_ATOMIC, NULL);
4341
4342 return rt6_get_route_info(net, prefix, prefixlen, gwaddr, dev);
4343}
4344#endif
4345
4346struct fib6_info *rt6_get_dflt_router(struct net *net,
4347 const struct in6_addr *addr,
4348 struct net_device *dev)
4349{
4350 u32 tb_id = l3mdev_fib_table(dev) ? : RT6_TABLE_DFLT;
4351 struct fib6_info *rt;
4352 struct fib6_table *table;
4353
4354 table = fib6_get_table(net, tb_id);
4355 if (!table)
4356 return NULL;
4357
4358 rcu_read_lock();
4359 for_each_fib6_node_rt_rcu(&table->tb6_root) {
4360 struct fib6_nh *nh;
4361
4362 /* RA routes do not use nexthops */
4363 if (rt->nh)
4364 continue;
4365
4366 nh = rt->fib6_nh;
4367 if (dev == nh->fib_nh_dev &&
4368 ((rt->fib6_flags & (RTF_ADDRCONF | RTF_DEFAULT)) == (RTF_ADDRCONF | RTF_DEFAULT)) &&
4369 ipv6_addr_equal(&nh->fib_nh_gw6, addr))
4370 break;
4371 }
4372 if (rt && !fib6_info_hold_safe(rt))
4373 rt = NULL;
4374 rcu_read_unlock();
4375 return rt;
4376}
4377
4378struct fib6_info *rt6_add_dflt_router(struct net *net,
4379 const struct in6_addr *gwaddr,
4380 struct net_device *dev,
4381 unsigned int pref,
4382 u32 defrtr_usr_metric)
4383{
4384 struct fib6_config cfg = {
4385 .fc_table = l3mdev_fib_table(dev) ? : RT6_TABLE_DFLT,
4386 .fc_metric = defrtr_usr_metric,
4387 .fc_ifindex = dev->ifindex,
4388 .fc_flags = RTF_GATEWAY | RTF_ADDRCONF | RTF_DEFAULT |
4389 RTF_UP | RTF_EXPIRES | RTF_PREF(pref),
4390 .fc_protocol = RTPROT_RA,
4391 .fc_type = RTN_UNICAST,
4392 .fc_nlinfo.portid = 0,
4393 .fc_nlinfo.nlh = NULL,
4394 .fc_nlinfo.nl_net = net,
4395 };
4396
4397 cfg.fc_gateway = *gwaddr;
4398
4399 if (!ip6_route_add(&cfg, GFP_ATOMIC, NULL)) {
4400 struct fib6_table *table;
4401
4402 table = fib6_get_table(dev_net(dev), cfg.fc_table);
4403 if (table)
4404 table->flags |= RT6_TABLE_HAS_DFLT_ROUTER;
4405 }
4406
4407 return rt6_get_dflt_router(net, gwaddr, dev);
4408}
4409
4410static void __rt6_purge_dflt_routers(struct net *net,
4411 struct fib6_table *table)
4412{
4413 struct fib6_info *rt;
4414
4415restart:
4416 rcu_read_lock();
4417 for_each_fib6_node_rt_rcu(&table->tb6_root) {
4418 struct net_device *dev = fib6_info_nh_dev(rt);
4419 struct inet6_dev *idev = dev ? __in6_dev_get(dev) : NULL;
4420
4421 if (rt->fib6_flags & (RTF_DEFAULT | RTF_ADDRCONF) &&
4422 (!idev || idev->cnf.accept_ra != 2) &&
4423 fib6_info_hold_safe(rt)) {
4424 rcu_read_unlock();
4425 ip6_del_rt(net, rt, false);
4426 goto restart;
4427 }
4428 }
4429 rcu_read_unlock();
4430
4431 table->flags &= ~RT6_TABLE_HAS_DFLT_ROUTER;
4432}
4433
4434void rt6_purge_dflt_routers(struct net *net)
4435{
4436 struct fib6_table *table;
4437 struct hlist_head *head;
4438 unsigned int h;
4439
4440 rcu_read_lock();
4441
4442 for (h = 0; h < FIB6_TABLE_HASHSZ; h++) {
4443 head = &net->ipv6.fib_table_hash[h];
4444 hlist_for_each_entry_rcu(table, head, tb6_hlist) {
4445 if (table->flags & RT6_TABLE_HAS_DFLT_ROUTER)
4446 __rt6_purge_dflt_routers(net, table);
4447 }
4448 }
4449
4450 rcu_read_unlock();
4451}
4452
4453static void rtmsg_to_fib6_config(struct net *net,
4454 struct in6_rtmsg *rtmsg,
4455 struct fib6_config *cfg)
4456{
4457 *cfg = (struct fib6_config){
4458 .fc_table = l3mdev_fib_table_by_index(net, rtmsg->rtmsg_ifindex) ?
4459 : RT6_TABLE_MAIN,
4460 .fc_ifindex = rtmsg->rtmsg_ifindex,
4461 .fc_metric = rtmsg->rtmsg_metric ? : IP6_RT_PRIO_USER,
4462 .fc_expires = rtmsg->rtmsg_info,
4463 .fc_dst_len = rtmsg->rtmsg_dst_len,
4464 .fc_src_len = rtmsg->rtmsg_src_len,
4465 .fc_flags = rtmsg->rtmsg_flags,
4466 .fc_type = rtmsg->rtmsg_type,
4467
4468 .fc_nlinfo.nl_net = net,
4469
4470 .fc_dst = rtmsg->rtmsg_dst,
4471 .fc_src = rtmsg->rtmsg_src,
4472 .fc_gateway = rtmsg->rtmsg_gateway,
4473 };
4474}
4475
4476int ipv6_route_ioctl(struct net *net, unsigned int cmd, struct in6_rtmsg *rtmsg)
4477{
4478 struct fib6_config cfg;
4479 int err;
4480
4481 if (cmd != SIOCADDRT && cmd != SIOCDELRT)
4482 return -EINVAL;
4483 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
4484 return -EPERM;
4485
4486 rtmsg_to_fib6_config(net, rtmsg, &cfg);
4487
4488 rtnl_lock();
4489 switch (cmd) {
4490 case SIOCADDRT:
4491 err = ip6_route_add(&cfg, GFP_KERNEL, NULL);
4492 break;
4493 case SIOCDELRT:
4494 err = ip6_route_del(&cfg, NULL);
4495 break;
4496 }
4497 rtnl_unlock();
4498 return err;
4499}
4500
4501/*
4502 * Drop the packet on the floor
4503 */
4504
4505static int ip6_pkt_drop(struct sk_buff *skb, u8 code, int ipstats_mib_noroutes)
4506{
4507 struct dst_entry *dst = skb_dst(skb);
4508 struct net *net = dev_net(dst->dev);
4509 struct inet6_dev *idev;
4510 int type;
4511
4512 if (netif_is_l3_master(skb->dev) &&
4513 dst->dev == net->loopback_dev)
4514 idev = __in6_dev_get_safely(dev_get_by_index_rcu(net, IP6CB(skb)->iif));
4515 else
4516 idev = ip6_dst_idev(dst);
4517
4518 switch (ipstats_mib_noroutes) {
4519 case IPSTATS_MIB_INNOROUTES:
4520 type = ipv6_addr_type(&ipv6_hdr(skb)->daddr);
4521 if (type == IPV6_ADDR_ANY) {
4522 IP6_INC_STATS(net, idev, IPSTATS_MIB_INADDRERRORS);
4523 break;
4524 }
4525 fallthrough;
4526 case IPSTATS_MIB_OUTNOROUTES:
4527 IP6_INC_STATS(net, idev, ipstats_mib_noroutes);
4528 break;
4529 }
4530
4531 /* Start over by dropping the dst for l3mdev case */
4532 if (netif_is_l3_master(skb->dev))
4533 skb_dst_drop(skb);
4534
4535 icmpv6_send(skb, ICMPV6_DEST_UNREACH, code, 0);
4536 kfree_skb(skb);
4537 return 0;
4538}
4539
4540static int ip6_pkt_discard(struct sk_buff *skb)
4541{
4542 return ip6_pkt_drop(skb, ICMPV6_NOROUTE, IPSTATS_MIB_INNOROUTES);
4543}
4544
4545static int ip6_pkt_discard_out(struct net *net, struct sock *sk, struct sk_buff *skb)
4546{
4547 skb->dev = skb_dst(skb)->dev;
4548 return ip6_pkt_drop(skb, ICMPV6_NOROUTE, IPSTATS_MIB_OUTNOROUTES);
4549}
4550
4551static int ip6_pkt_prohibit(struct sk_buff *skb)
4552{
4553 return ip6_pkt_drop(skb, ICMPV6_ADM_PROHIBITED, IPSTATS_MIB_INNOROUTES);
4554}
4555
4556static int ip6_pkt_prohibit_out(struct net *net, struct sock *sk, struct sk_buff *skb)
4557{
4558 skb->dev = skb_dst(skb)->dev;
4559 return ip6_pkt_drop(skb, ICMPV6_ADM_PROHIBITED, IPSTATS_MIB_OUTNOROUTES);
4560}
4561
4562/*
4563 * Allocate a dst for local (unicast / anycast) address.
4564 */
4565
4566struct fib6_info *addrconf_f6i_alloc(struct net *net,
4567 struct inet6_dev *idev,
4568 const struct in6_addr *addr,
4569 bool anycast, gfp_t gfp_flags)
4570{
4571 struct fib6_config cfg = {
4572 .fc_table = l3mdev_fib_table(idev->dev) ? : RT6_TABLE_LOCAL,
4573 .fc_ifindex = idev->dev->ifindex,
4574 .fc_flags = RTF_UP | RTF_NONEXTHOP,
4575 .fc_dst = *addr,
4576 .fc_dst_len = 128,
4577 .fc_protocol = RTPROT_KERNEL,
4578 .fc_nlinfo.nl_net = net,
4579 .fc_ignore_dev_down = true,
4580 };
4581 struct fib6_info *f6i;
4582
4583 if (anycast) {
4584 cfg.fc_type = RTN_ANYCAST;
4585 cfg.fc_flags |= RTF_ANYCAST;
4586 } else {
4587 cfg.fc_type = RTN_LOCAL;
4588 cfg.fc_flags |= RTF_LOCAL;
4589 }
4590
4591 f6i = ip6_route_info_create(&cfg, gfp_flags, NULL);
4592 if (!IS_ERR(f6i))
4593 f6i->dst_nocount = true;
4594 return f6i;
4595}
4596
4597/* remove deleted ip from prefsrc entries */
4598struct arg_dev_net_ip {
4599 struct net_device *dev;
4600 struct net *net;
4601 struct in6_addr *addr;
4602};
4603
4604static int fib6_remove_prefsrc(struct fib6_info *rt, void *arg)
4605{
4606 struct net_device *dev = ((struct arg_dev_net_ip *)arg)->dev;
4607 struct net *net = ((struct arg_dev_net_ip *)arg)->net;
4608 struct in6_addr *addr = ((struct arg_dev_net_ip *)arg)->addr;
4609
4610 if (!rt->nh &&
4611 ((void *)rt->fib6_nh->fib_nh_dev == dev || !dev) &&
4612 rt != net->ipv6.fib6_null_entry &&
4613 ipv6_addr_equal(addr, &rt->fib6_prefsrc.addr)) {
4614 spin_lock_bh(&rt6_exception_lock);
4615 /* remove prefsrc entry */
4616 rt->fib6_prefsrc.plen = 0;
4617 spin_unlock_bh(&rt6_exception_lock);
4618 }
4619 return 0;
4620}
4621
4622void rt6_remove_prefsrc(struct inet6_ifaddr *ifp)
4623{
4624 struct net *net = dev_net(ifp->idev->dev);
4625 struct arg_dev_net_ip adni = {
4626 .dev = ifp->idev->dev,
4627 .net = net,
4628 .addr = &ifp->addr,
4629 };
4630 fib6_clean_all(net, fib6_remove_prefsrc, &adni);
4631}
4632
4633#define RTF_RA_ROUTER (RTF_ADDRCONF | RTF_DEFAULT)
4634
4635/* Remove routers and update dst entries when gateway turn into host. */
4636static int fib6_clean_tohost(struct fib6_info *rt, void *arg)
4637{
4638 struct in6_addr *gateway = (struct in6_addr *)arg;
4639 struct fib6_nh *nh;
4640
4641 /* RA routes do not use nexthops */
4642 if (rt->nh)
4643 return 0;
4644
4645 nh = rt->fib6_nh;
4646 if (((rt->fib6_flags & RTF_RA_ROUTER) == RTF_RA_ROUTER) &&
4647 nh->fib_nh_gw_family && ipv6_addr_equal(gateway, &nh->fib_nh_gw6))
4648 return -1;
4649
4650 /* Further clean up cached routes in exception table.
4651 * This is needed because cached route may have a different
4652 * gateway than its 'parent' in the case of an ip redirect.
4653 */
4654 fib6_nh_exceptions_clean_tohost(nh, gateway);
4655
4656 return 0;
4657}
4658
4659void rt6_clean_tohost(struct net *net, struct in6_addr *gateway)
4660{
4661 fib6_clean_all(net, fib6_clean_tohost, gateway);
4662}
4663
4664struct arg_netdev_event {
4665 const struct net_device *dev;
4666 union {
4667 unsigned char nh_flags;
4668 unsigned long event;
4669 };
4670};
4671
4672static struct fib6_info *rt6_multipath_first_sibling(const struct fib6_info *rt)
4673{
4674 struct fib6_info *iter;
4675 struct fib6_node *fn;
4676
4677 fn = rcu_dereference_protected(rt->fib6_node,
4678 lockdep_is_held(&rt->fib6_table->tb6_lock));
4679 iter = rcu_dereference_protected(fn->leaf,
4680 lockdep_is_held(&rt->fib6_table->tb6_lock));
4681 while (iter) {
4682 if (iter->fib6_metric == rt->fib6_metric &&
4683 rt6_qualify_for_ecmp(iter))
4684 return iter;
4685 iter = rcu_dereference_protected(iter->fib6_next,
4686 lockdep_is_held(&rt->fib6_table->tb6_lock));
4687 }
4688
4689 return NULL;
4690}
4691
4692/* only called for fib entries with builtin fib6_nh */
4693static bool rt6_is_dead(const struct fib6_info *rt)
4694{
4695 if (rt->fib6_nh->fib_nh_flags & RTNH_F_DEAD ||
4696 (rt->fib6_nh->fib_nh_flags & RTNH_F_LINKDOWN &&
4697 ip6_ignore_linkdown(rt->fib6_nh->fib_nh_dev)))
4698 return true;
4699
4700 return false;
4701}
4702
4703static int rt6_multipath_total_weight(const struct fib6_info *rt)
4704{
4705 struct fib6_info *iter;
4706 int total = 0;
4707
4708 if (!rt6_is_dead(rt))
4709 total += rt->fib6_nh->fib_nh_weight;
4710
4711 list_for_each_entry(iter, &rt->fib6_siblings, fib6_siblings) {
4712 if (!rt6_is_dead(iter))
4713 total += iter->fib6_nh->fib_nh_weight;
4714 }
4715
4716 return total;
4717}
4718
4719static void rt6_upper_bound_set(struct fib6_info *rt, int *weight, int total)
4720{
4721 int upper_bound = -1;
4722
4723 if (!rt6_is_dead(rt)) {
4724 *weight += rt->fib6_nh->fib_nh_weight;
4725 upper_bound = DIV_ROUND_CLOSEST_ULL((u64) (*weight) << 31,
4726 total) - 1;
4727 }
4728 atomic_set(&rt->fib6_nh->fib_nh_upper_bound, upper_bound);
4729}
4730
4731static void rt6_multipath_upper_bound_set(struct fib6_info *rt, int total)
4732{
4733 struct fib6_info *iter;
4734 int weight = 0;
4735
4736 rt6_upper_bound_set(rt, &weight, total);
4737
4738 list_for_each_entry(iter, &rt->fib6_siblings, fib6_siblings)
4739 rt6_upper_bound_set(iter, &weight, total);
4740}
4741
4742void rt6_multipath_rebalance(struct fib6_info *rt)
4743{
4744 struct fib6_info *first;
4745 int total;
4746
4747 /* In case the entire multipath route was marked for flushing,
4748 * then there is no need to rebalance upon the removal of every
4749 * sibling route.
4750 */
4751 if (!rt->fib6_nsiblings || rt->should_flush)
4752 return;
4753
4754 /* During lookup routes are evaluated in order, so we need to
4755 * make sure upper bounds are assigned from the first sibling
4756 * onwards.
4757 */
4758 first = rt6_multipath_first_sibling(rt);
4759 if (WARN_ON_ONCE(!first))
4760 return;
4761
4762 total = rt6_multipath_total_weight(first);
4763 rt6_multipath_upper_bound_set(first, total);
4764}
4765
4766static int fib6_ifup(struct fib6_info *rt, void *p_arg)
4767{
4768 const struct arg_netdev_event *arg = p_arg;
4769 struct net *net = dev_net(arg->dev);
4770
4771 if (rt != net->ipv6.fib6_null_entry && !rt->nh &&
4772 rt->fib6_nh->fib_nh_dev == arg->dev) {
4773 rt->fib6_nh->fib_nh_flags &= ~arg->nh_flags;
4774 fib6_update_sernum_upto_root(net, rt);
4775 rt6_multipath_rebalance(rt);
4776 }
4777
4778 return 0;
4779}
4780
4781void rt6_sync_up(struct net_device *dev, unsigned char nh_flags)
4782{
4783 struct arg_netdev_event arg = {
4784 .dev = dev,
4785 {
4786 .nh_flags = nh_flags,
4787 },
4788 };
4789
4790 if (nh_flags & RTNH_F_DEAD && netif_carrier_ok(dev))
4791 arg.nh_flags |= RTNH_F_LINKDOWN;
4792
4793 fib6_clean_all(dev_net(dev), fib6_ifup, &arg);
4794}
4795
4796/* only called for fib entries with inline fib6_nh */
4797static bool rt6_multipath_uses_dev(const struct fib6_info *rt,
4798 const struct net_device *dev)
4799{
4800 struct fib6_info *iter;
4801
4802 if (rt->fib6_nh->fib_nh_dev == dev)
4803 return true;
4804 list_for_each_entry(iter, &rt->fib6_siblings, fib6_siblings)
4805 if (iter->fib6_nh->fib_nh_dev == dev)
4806 return true;
4807
4808 return false;
4809}
4810
4811static void rt6_multipath_flush(struct fib6_info *rt)
4812{
4813 struct fib6_info *iter;
4814
4815 rt->should_flush = 1;
4816 list_for_each_entry(iter, &rt->fib6_siblings, fib6_siblings)
4817 iter->should_flush = 1;
4818}
4819
4820static unsigned int rt6_multipath_dead_count(const struct fib6_info *rt,
4821 const struct net_device *down_dev)
4822{
4823 struct fib6_info *iter;
4824 unsigned int dead = 0;
4825
4826 if (rt->fib6_nh->fib_nh_dev == down_dev ||
4827 rt->fib6_nh->fib_nh_flags & RTNH_F_DEAD)
4828 dead++;
4829 list_for_each_entry(iter, &rt->fib6_siblings, fib6_siblings)
4830 if (iter->fib6_nh->fib_nh_dev == down_dev ||
4831 iter->fib6_nh->fib_nh_flags & RTNH_F_DEAD)
4832 dead++;
4833
4834 return dead;
4835}
4836
4837static void rt6_multipath_nh_flags_set(struct fib6_info *rt,
4838 const struct net_device *dev,
4839 unsigned char nh_flags)
4840{
4841 struct fib6_info *iter;
4842
4843 if (rt->fib6_nh->fib_nh_dev == dev)
4844 rt->fib6_nh->fib_nh_flags |= nh_flags;
4845 list_for_each_entry(iter, &rt->fib6_siblings, fib6_siblings)
4846 if (iter->fib6_nh->fib_nh_dev == dev)
4847 iter->fib6_nh->fib_nh_flags |= nh_flags;
4848}
4849
4850/* called with write lock held for table with rt */
4851static int fib6_ifdown(struct fib6_info *rt, void *p_arg)
4852{
4853 const struct arg_netdev_event *arg = p_arg;
4854 const struct net_device *dev = arg->dev;
4855 struct net *net = dev_net(dev);
4856
4857 if (rt == net->ipv6.fib6_null_entry || rt->nh)
4858 return 0;
4859
4860 switch (arg->event) {
4861 case NETDEV_UNREGISTER:
4862 return rt->fib6_nh->fib_nh_dev == dev ? -1 : 0;
4863 case NETDEV_DOWN:
4864 if (rt->should_flush)
4865 return -1;
4866 if (!rt->fib6_nsiblings)
4867 return rt->fib6_nh->fib_nh_dev == dev ? -1 : 0;
4868 if (rt6_multipath_uses_dev(rt, dev)) {
4869 unsigned int count;
4870
4871 count = rt6_multipath_dead_count(rt, dev);
4872 if (rt->fib6_nsiblings + 1 == count) {
4873 rt6_multipath_flush(rt);
4874 return -1;
4875 }
4876 rt6_multipath_nh_flags_set(rt, dev, RTNH_F_DEAD |
4877 RTNH_F_LINKDOWN);
4878 fib6_update_sernum(net, rt);
4879 rt6_multipath_rebalance(rt);
4880 }
4881 return -2;
4882 case NETDEV_CHANGE:
4883 if (rt->fib6_nh->fib_nh_dev != dev ||
4884 rt->fib6_flags & (RTF_LOCAL | RTF_ANYCAST))
4885 break;
4886 rt->fib6_nh->fib_nh_flags |= RTNH_F_LINKDOWN;
4887 rt6_multipath_rebalance(rt);
4888 break;
4889 }
4890
4891 return 0;
4892}
4893
4894void rt6_sync_down_dev(struct net_device *dev, unsigned long event)
4895{
4896 struct arg_netdev_event arg = {
4897 .dev = dev,
4898 {
4899 .event = event,
4900 },
4901 };
4902 struct net *net = dev_net(dev);
4903
4904 if (net->ipv6.sysctl.skip_notify_on_dev_down)
4905 fib6_clean_all_skip_notify(net, fib6_ifdown, &arg);
4906 else
4907 fib6_clean_all(net, fib6_ifdown, &arg);
4908}
4909
4910void rt6_disable_ip(struct net_device *dev, unsigned long event)
4911{
4912 rt6_sync_down_dev(dev, event);
4913 rt6_uncached_list_flush_dev(dev_net(dev), dev);
4914 neigh_ifdown(&nd_tbl, dev);
4915}
4916
4917struct rt6_mtu_change_arg {
4918 struct net_device *dev;
4919 unsigned int mtu;
4920 struct fib6_info *f6i;
4921};
4922
4923static int fib6_nh_mtu_change(struct fib6_nh *nh, void *_arg)
4924{
4925 struct rt6_mtu_change_arg *arg = (struct rt6_mtu_change_arg *)_arg;
4926 struct fib6_info *f6i = arg->f6i;
4927
4928 /* For administrative MTU increase, there is no way to discover
4929 * IPv6 PMTU increase, so PMTU increase should be updated here.
4930 * Since RFC 1981 doesn't include administrative MTU increase
4931 * update PMTU increase is a MUST. (i.e. jumbo frame)
4932 */
4933 if (nh->fib_nh_dev == arg->dev) {
4934 struct inet6_dev *idev = __in6_dev_get(arg->dev);
4935 u32 mtu = f6i->fib6_pmtu;
4936
4937 if (mtu >= arg->mtu ||
4938 (mtu < arg->mtu && mtu == idev->cnf.mtu6))
4939 fib6_metric_set(f6i, RTAX_MTU, arg->mtu);
4940
4941 spin_lock_bh(&rt6_exception_lock);
4942 rt6_exceptions_update_pmtu(idev, nh, arg->mtu);
4943 spin_unlock_bh(&rt6_exception_lock);
4944 }
4945
4946 return 0;
4947}
4948
4949static int rt6_mtu_change_route(struct fib6_info *f6i, void *p_arg)
4950{
4951 struct rt6_mtu_change_arg *arg = (struct rt6_mtu_change_arg *) p_arg;
4952 struct inet6_dev *idev;
4953
4954 /* In IPv6 pmtu discovery is not optional,
4955 so that RTAX_MTU lock cannot disable it.
4956 We still use this lock to block changes
4957 caused by addrconf/ndisc.
4958 */
4959
4960 idev = __in6_dev_get(arg->dev);
4961 if (!idev)
4962 return 0;
4963
4964 if (fib6_metric_locked(f6i, RTAX_MTU))
4965 return 0;
4966
4967 arg->f6i = f6i;
4968 if (f6i->nh) {
4969 /* fib6_nh_mtu_change only returns 0, so this is safe */
4970 return nexthop_for_each_fib6_nh(f6i->nh, fib6_nh_mtu_change,
4971 arg);
4972 }
4973
4974 return fib6_nh_mtu_change(f6i->fib6_nh, arg);
4975}
4976
4977void rt6_mtu_change(struct net_device *dev, unsigned int mtu)
4978{
4979 struct rt6_mtu_change_arg arg = {
4980 .dev = dev,
4981 .mtu = mtu,
4982 };
4983
4984 fib6_clean_all(dev_net(dev), rt6_mtu_change_route, &arg);
4985}
4986
4987static const struct nla_policy rtm_ipv6_policy[RTA_MAX+1] = {
4988 [RTA_UNSPEC] = { .strict_start_type = RTA_DPORT + 1 },
4989 [RTA_GATEWAY] = { .len = sizeof(struct in6_addr) },
4990 [RTA_PREFSRC] = { .len = sizeof(struct in6_addr) },
4991 [RTA_OIF] = { .type = NLA_U32 },
4992 [RTA_IIF] = { .type = NLA_U32 },
4993 [RTA_PRIORITY] = { .type = NLA_U32 },
4994 [RTA_METRICS] = { .type = NLA_NESTED },
4995 [RTA_MULTIPATH] = { .len = sizeof(struct rtnexthop) },
4996 [RTA_PREF] = { .type = NLA_U8 },
4997 [RTA_ENCAP_TYPE] = { .type = NLA_U16 },
4998 [RTA_ENCAP] = { .type = NLA_NESTED },
4999 [RTA_EXPIRES] = { .type = NLA_U32 },
5000 [RTA_UID] = { .type = NLA_U32 },
5001 [RTA_MARK] = { .type = NLA_U32 },
5002 [RTA_TABLE] = { .type = NLA_U32 },
5003 [RTA_IP_PROTO] = { .type = NLA_U8 },
5004 [RTA_SPORT] = { .type = NLA_U16 },
5005 [RTA_DPORT] = { .type = NLA_U16 },
5006 [RTA_NH_ID] = { .type = NLA_U32 },
5007};
5008
5009static int rtm_to_fib6_config(struct sk_buff *skb, struct nlmsghdr *nlh,
5010 struct fib6_config *cfg,
5011 struct netlink_ext_ack *extack)
5012{
5013 struct rtmsg *rtm;
5014 struct nlattr *tb[RTA_MAX+1];
5015 unsigned int pref;
5016 int err;
5017
5018 err = nlmsg_parse_deprecated(nlh, sizeof(*rtm), tb, RTA_MAX,
5019 rtm_ipv6_policy, extack);
5020 if (err < 0)
5021 goto errout;
5022
5023 err = -EINVAL;
5024 rtm = nlmsg_data(nlh);
5025
5026 *cfg = (struct fib6_config){
5027 .fc_table = rtm->rtm_table,
5028 .fc_dst_len = rtm->rtm_dst_len,
5029 .fc_src_len = rtm->rtm_src_len,
5030 .fc_flags = RTF_UP,
5031 .fc_protocol = rtm->rtm_protocol,
5032 .fc_type = rtm->rtm_type,
5033
5034 .fc_nlinfo.portid = NETLINK_CB(skb).portid,
5035 .fc_nlinfo.nlh = nlh,
5036 .fc_nlinfo.nl_net = sock_net(skb->sk),
5037 };
5038
5039 if (rtm->rtm_type == RTN_UNREACHABLE ||
5040 rtm->rtm_type == RTN_BLACKHOLE ||
5041 rtm->rtm_type == RTN_PROHIBIT ||
5042 rtm->rtm_type == RTN_THROW)
5043 cfg->fc_flags |= RTF_REJECT;
5044
5045 if (rtm->rtm_type == RTN_LOCAL)
5046 cfg->fc_flags |= RTF_LOCAL;
5047
5048 if (rtm->rtm_flags & RTM_F_CLONED)
5049 cfg->fc_flags |= RTF_CACHE;
5050
5051 cfg->fc_flags |= (rtm->rtm_flags & RTNH_F_ONLINK);
5052
5053 if (tb[RTA_NH_ID]) {
5054 if (tb[RTA_GATEWAY] || tb[RTA_OIF] ||
5055 tb[RTA_MULTIPATH] || tb[RTA_ENCAP]) {
5056 NL_SET_ERR_MSG(extack,
5057 "Nexthop specification and nexthop id are mutually exclusive");
5058 goto errout;
5059 }
5060 cfg->fc_nh_id = nla_get_u32(tb[RTA_NH_ID]);
5061 }
5062
5063 if (tb[RTA_GATEWAY]) {
5064 cfg->fc_gateway = nla_get_in6_addr(tb[RTA_GATEWAY]);
5065 cfg->fc_flags |= RTF_GATEWAY;
5066 }
5067 if (tb[RTA_VIA]) {
5068 NL_SET_ERR_MSG(extack, "IPv6 does not support RTA_VIA attribute");
5069 goto errout;
5070 }
5071
5072 if (tb[RTA_DST]) {
5073 int plen = (rtm->rtm_dst_len + 7) >> 3;
5074
5075 if (nla_len(tb[RTA_DST]) < plen)
5076 goto errout;
5077
5078 nla_memcpy(&cfg->fc_dst, tb[RTA_DST], plen);
5079 }
5080
5081 if (tb[RTA_SRC]) {
5082 int plen = (rtm->rtm_src_len + 7) >> 3;
5083
5084 if (nla_len(tb[RTA_SRC]) < plen)
5085 goto errout;
5086
5087 nla_memcpy(&cfg->fc_src, tb[RTA_SRC], plen);
5088 }
5089
5090 if (tb[RTA_PREFSRC])
5091 cfg->fc_prefsrc = nla_get_in6_addr(tb[RTA_PREFSRC]);
5092
5093 if (tb[RTA_OIF])
5094 cfg->fc_ifindex = nla_get_u32(tb[RTA_OIF]);
5095
5096 if (tb[RTA_PRIORITY])
5097 cfg->fc_metric = nla_get_u32(tb[RTA_PRIORITY]);
5098
5099 if (tb[RTA_METRICS]) {
5100 cfg->fc_mx = nla_data(tb[RTA_METRICS]);
5101 cfg->fc_mx_len = nla_len(tb[RTA_METRICS]);
5102 }
5103
5104 if (tb[RTA_TABLE])
5105 cfg->fc_table = nla_get_u32(tb[RTA_TABLE]);
5106
5107 if (tb[RTA_MULTIPATH]) {
5108 cfg->fc_mp = nla_data(tb[RTA_MULTIPATH]);
5109 cfg->fc_mp_len = nla_len(tb[RTA_MULTIPATH]);
5110
5111 err = lwtunnel_valid_encap_type_attr(cfg->fc_mp,
5112 cfg->fc_mp_len, extack);
5113 if (err < 0)
5114 goto errout;
5115 }
5116
5117 if (tb[RTA_PREF]) {
5118 pref = nla_get_u8(tb[RTA_PREF]);
5119 if (pref != ICMPV6_ROUTER_PREF_LOW &&
5120 pref != ICMPV6_ROUTER_PREF_HIGH)
5121 pref = ICMPV6_ROUTER_PREF_MEDIUM;
5122 cfg->fc_flags |= RTF_PREF(pref);
5123 }
5124
5125 if (tb[RTA_ENCAP])
5126 cfg->fc_encap = tb[RTA_ENCAP];
5127
5128 if (tb[RTA_ENCAP_TYPE]) {
5129 cfg->fc_encap_type = nla_get_u16(tb[RTA_ENCAP_TYPE]);
5130
5131 err = lwtunnel_valid_encap_type(cfg->fc_encap_type, extack);
5132 if (err < 0)
5133 goto errout;
5134 }
5135
5136 if (tb[RTA_EXPIRES]) {
5137 unsigned long timeout = addrconf_timeout_fixup(nla_get_u32(tb[RTA_EXPIRES]), HZ);
5138
5139 if (addrconf_finite_timeout(timeout)) {
5140 cfg->fc_expires = jiffies_to_clock_t(timeout * HZ);
5141 cfg->fc_flags |= RTF_EXPIRES;
5142 }
5143 }
5144
5145 err = 0;
5146errout:
5147 return err;
5148}
5149
5150struct rt6_nh {
5151 struct fib6_info *fib6_info;
5152 struct fib6_config r_cfg;
5153 struct list_head next;
5154};
5155
5156static int ip6_route_info_append(struct net *net,
5157 struct list_head *rt6_nh_list,
5158 struct fib6_info *rt,
5159 struct fib6_config *r_cfg)
5160{
5161 struct rt6_nh *nh;
5162 int err = -EEXIST;
5163
5164 list_for_each_entry(nh, rt6_nh_list, next) {
5165 /* check if fib6_info already exists */
5166 if (rt6_duplicate_nexthop(nh->fib6_info, rt))
5167 return err;
5168 }
5169
5170 nh = kzalloc(sizeof(*nh), GFP_KERNEL);
5171 if (!nh)
5172 return -ENOMEM;
5173 nh->fib6_info = rt;
5174 memcpy(&nh->r_cfg, r_cfg, sizeof(*r_cfg));
5175 list_add_tail(&nh->next, rt6_nh_list);
5176
5177 return 0;
5178}
5179
5180static void ip6_route_mpath_notify(struct fib6_info *rt,
5181 struct fib6_info *rt_last,
5182 struct nl_info *info,
5183 __u16 nlflags)
5184{
5185 /* if this is an APPEND route, then rt points to the first route
5186 * inserted and rt_last points to last route inserted. Userspace
5187 * wants a consistent dump of the route which starts at the first
5188 * nexthop. Since sibling routes are always added at the end of
5189 * the list, find the first sibling of the last route appended
5190 */
5191 if ((nlflags & NLM_F_APPEND) && rt_last && rt_last->fib6_nsiblings) {
5192 rt = list_first_entry(&rt_last->fib6_siblings,
5193 struct fib6_info,
5194 fib6_siblings);
5195 }
5196
5197 if (rt)
5198 inet6_rt_notify(RTM_NEWROUTE, rt, info, nlflags);
5199}
5200
5201static bool ip6_route_mpath_should_notify(const struct fib6_info *rt)
5202{
5203 bool rt_can_ecmp = rt6_qualify_for_ecmp(rt);
5204 bool should_notify = false;
5205 struct fib6_info *leaf;
5206 struct fib6_node *fn;
5207
5208 rcu_read_lock();
5209 fn = rcu_dereference(rt->fib6_node);
5210 if (!fn)
5211 goto out;
5212
5213 leaf = rcu_dereference(fn->leaf);
5214 if (!leaf)
5215 goto out;
5216
5217 if (rt == leaf ||
5218 (rt_can_ecmp && rt->fib6_metric == leaf->fib6_metric &&
5219 rt6_qualify_for_ecmp(leaf)))
5220 should_notify = true;
5221out:
5222 rcu_read_unlock();
5223
5224 return should_notify;
5225}
5226
5227static int ip6_route_multipath_add(struct fib6_config *cfg,
5228 struct netlink_ext_ack *extack)
5229{
5230 struct fib6_info *rt_notif = NULL, *rt_last = NULL;
5231 struct nl_info *info = &cfg->fc_nlinfo;
5232 struct fib6_config r_cfg;
5233 struct rtnexthop *rtnh;
5234 struct fib6_info *rt;
5235 struct rt6_nh *err_nh;
5236 struct rt6_nh *nh, *nh_safe;
5237 __u16 nlflags;
5238 int remaining;
5239 int attrlen;
5240 int err = 1;
5241 int nhn = 0;
5242 int replace = (cfg->fc_nlinfo.nlh &&
5243 (cfg->fc_nlinfo.nlh->nlmsg_flags & NLM_F_REPLACE));
5244 LIST_HEAD(rt6_nh_list);
5245
5246 nlflags = replace ? NLM_F_REPLACE : NLM_F_CREATE;
5247 if (info->nlh && info->nlh->nlmsg_flags & NLM_F_APPEND)
5248 nlflags |= NLM_F_APPEND;
5249
5250 remaining = cfg->fc_mp_len;
5251 rtnh = (struct rtnexthop *)cfg->fc_mp;
5252
5253 /* Parse a Multipath Entry and build a list (rt6_nh_list) of
5254 * fib6_info structs per nexthop
5255 */
5256 while (rtnh_ok(rtnh, remaining)) {
5257 memcpy(&r_cfg, cfg, sizeof(*cfg));
5258 if (rtnh->rtnh_ifindex)
5259 r_cfg.fc_ifindex = rtnh->rtnh_ifindex;
5260
5261 attrlen = rtnh_attrlen(rtnh);
5262 if (attrlen > 0) {
5263 struct nlattr *nla, *attrs = rtnh_attrs(rtnh);
5264
5265 nla = nla_find(attrs, attrlen, RTA_GATEWAY);
5266 if (nla) {
5267 r_cfg.fc_gateway = nla_get_in6_addr(nla);
5268 r_cfg.fc_flags |= RTF_GATEWAY;
5269 }
5270 r_cfg.fc_encap = nla_find(attrs, attrlen, RTA_ENCAP);
5271 nla = nla_find(attrs, attrlen, RTA_ENCAP_TYPE);
5272 if (nla)
5273 r_cfg.fc_encap_type = nla_get_u16(nla);
5274 }
5275
5276 r_cfg.fc_flags |= (rtnh->rtnh_flags & RTNH_F_ONLINK);
5277 rt = ip6_route_info_create(&r_cfg, GFP_KERNEL, extack);
5278 if (IS_ERR(rt)) {
5279 err = PTR_ERR(rt);
5280 rt = NULL;
5281 goto cleanup;
5282 }
5283 if (!rt6_qualify_for_ecmp(rt)) {
5284 err = -EINVAL;
5285 NL_SET_ERR_MSG(extack,
5286 "Device only routes can not be added for IPv6 using the multipath API.");
5287 fib6_info_release(rt);
5288 goto cleanup;
5289 }
5290
5291 rt->fib6_nh->fib_nh_weight = rtnh->rtnh_hops + 1;
5292
5293 err = ip6_route_info_append(info->nl_net, &rt6_nh_list,
5294 rt, &r_cfg);
5295 if (err) {
5296 fib6_info_release(rt);
5297 goto cleanup;
5298 }
5299
5300 rtnh = rtnh_next(rtnh, &remaining);
5301 }
5302
5303 if (list_empty(&rt6_nh_list)) {
5304 NL_SET_ERR_MSG(extack,
5305 "Invalid nexthop configuration - no valid nexthops");
5306 return -EINVAL;
5307 }
5308
5309 /* for add and replace send one notification with all nexthops.
5310 * Skip the notification in fib6_add_rt2node and send one with
5311 * the full route when done
5312 */
5313 info->skip_notify = 1;
5314
5315 /* For add and replace, send one notification with all nexthops. For
5316 * append, send one notification with all appended nexthops.
5317 */
5318 info->skip_notify_kernel = 1;
5319
5320 err_nh = NULL;
5321 list_for_each_entry(nh, &rt6_nh_list, next) {
5322 err = __ip6_ins_rt(nh->fib6_info, info, extack);
5323 fib6_info_release(nh->fib6_info);
5324
5325 if (!err) {
5326 /* save reference to last route successfully inserted */
5327 rt_last = nh->fib6_info;
5328
5329 /* save reference to first route for notification */
5330 if (!rt_notif)
5331 rt_notif = nh->fib6_info;
5332 }
5333
5334 /* nh->fib6_info is used or freed at this point, reset to NULL*/
5335 nh->fib6_info = NULL;
5336 if (err) {
5337 if (replace && nhn)
5338 NL_SET_ERR_MSG_MOD(extack,
5339 "multipath route replace failed (check consistency of installed routes)");
5340 err_nh = nh;
5341 goto add_errout;
5342 }
5343
5344 /* Because each route is added like a single route we remove
5345 * these flags after the first nexthop: if there is a collision,
5346 * we have already failed to add the first nexthop:
5347 * fib6_add_rt2node() has rejected it; when replacing, old
5348 * nexthops have been replaced by first new, the rest should
5349 * be added to it.
5350 */
5351 if (cfg->fc_nlinfo.nlh) {
5352 cfg->fc_nlinfo.nlh->nlmsg_flags &= ~(NLM_F_EXCL |
5353 NLM_F_REPLACE);
5354 cfg->fc_nlinfo.nlh->nlmsg_flags |= NLM_F_CREATE;
5355 }
5356 nhn++;
5357 }
5358
5359 /* An in-kernel notification should only be sent in case the new
5360 * multipath route is added as the first route in the node, or if
5361 * it was appended to it. We pass 'rt_notif' since it is the first
5362 * sibling and might allow us to skip some checks in the replace case.
5363 */
5364 if (ip6_route_mpath_should_notify(rt_notif)) {
5365 enum fib_event_type fib_event;
5366
5367 if (rt_notif->fib6_nsiblings != nhn - 1)
5368 fib_event = FIB_EVENT_ENTRY_APPEND;
5369 else
5370 fib_event = FIB_EVENT_ENTRY_REPLACE;
5371
5372 err = call_fib6_multipath_entry_notifiers(info->nl_net,
5373 fib_event, rt_notif,
5374 nhn - 1, extack);
5375 if (err) {
5376 /* Delete all the siblings that were just added */
5377 err_nh = NULL;
5378 goto add_errout;
5379 }
5380 }
5381
5382 /* success ... tell user about new route */
5383 ip6_route_mpath_notify(rt_notif, rt_last, info, nlflags);
5384 goto cleanup;
5385
5386add_errout:
5387 /* send notification for routes that were added so that
5388 * the delete notifications sent by ip6_route_del are
5389 * coherent
5390 */
5391 if (rt_notif)
5392 ip6_route_mpath_notify(rt_notif, rt_last, info, nlflags);
5393
5394 /* Delete routes that were already added */
5395 list_for_each_entry(nh, &rt6_nh_list, next) {
5396 if (err_nh == nh)
5397 break;
5398 ip6_route_del(&nh->r_cfg, extack);
5399 }
5400
5401cleanup:
5402 list_for_each_entry_safe(nh, nh_safe, &rt6_nh_list, next) {
5403 if (nh->fib6_info)
5404 fib6_info_release(nh->fib6_info);
5405 list_del(&nh->next);
5406 kfree(nh);
5407 }
5408
5409 return err;
5410}
5411
5412static int ip6_route_multipath_del(struct fib6_config *cfg,
5413 struct netlink_ext_ack *extack)
5414{
5415 struct fib6_config r_cfg;
5416 struct rtnexthop *rtnh;
5417 int last_err = 0;
5418 int remaining;
5419 int attrlen;
5420 int err;
5421
5422 remaining = cfg->fc_mp_len;
5423 rtnh = (struct rtnexthop *)cfg->fc_mp;
5424
5425 /* Parse a Multipath Entry */
5426 while (rtnh_ok(rtnh, remaining)) {
5427 memcpy(&r_cfg, cfg, sizeof(*cfg));
5428 if (rtnh->rtnh_ifindex)
5429 r_cfg.fc_ifindex = rtnh->rtnh_ifindex;
5430
5431 attrlen = rtnh_attrlen(rtnh);
5432 if (attrlen > 0) {
5433 struct nlattr *nla, *attrs = rtnh_attrs(rtnh);
5434
5435 nla = nla_find(attrs, attrlen, RTA_GATEWAY);
5436 if (nla) {
5437 nla_memcpy(&r_cfg.fc_gateway, nla, 16);
5438 r_cfg.fc_flags |= RTF_GATEWAY;
5439 }
5440 }
5441 err = ip6_route_del(&r_cfg, extack);
5442 if (err)
5443 last_err = err;
5444
5445 rtnh = rtnh_next(rtnh, &remaining);
5446 }
5447
5448 return last_err;
5449}
5450
5451static int inet6_rtm_delroute(struct sk_buff *skb, struct nlmsghdr *nlh,
5452 struct netlink_ext_ack *extack)
5453{
5454 struct fib6_config cfg;
5455 int err;
5456
5457 err = rtm_to_fib6_config(skb, nlh, &cfg, extack);
5458 if (err < 0)
5459 return err;
5460
5461 if (cfg.fc_nh_id &&
5462 !nexthop_find_by_id(sock_net(skb->sk), cfg.fc_nh_id)) {
5463 NL_SET_ERR_MSG(extack, "Nexthop id does not exist");
5464 return -EINVAL;
5465 }
5466
5467 if (cfg.fc_mp)
5468 return ip6_route_multipath_del(&cfg, extack);
5469 else {
5470 cfg.fc_delete_all_nh = 1;
5471 return ip6_route_del(&cfg, extack);
5472 }
5473}
5474
5475static int inet6_rtm_newroute(struct sk_buff *skb, struct nlmsghdr *nlh,
5476 struct netlink_ext_ack *extack)
5477{
5478 struct fib6_config cfg;
5479 int err;
5480
5481 err = rtm_to_fib6_config(skb, nlh, &cfg, extack);
5482 if (err < 0)
5483 return err;
5484
5485 if (cfg.fc_metric == 0)
5486 cfg.fc_metric = IP6_RT_PRIO_USER;
5487
5488 if (cfg.fc_mp)
5489 return ip6_route_multipath_add(&cfg, extack);
5490 else
5491 return ip6_route_add(&cfg, GFP_KERNEL, extack);
5492}
5493
5494/* add the overhead of this fib6_nh to nexthop_len */
5495static int rt6_nh_nlmsg_size(struct fib6_nh *nh, void *arg)
5496{
5497 int *nexthop_len = arg;
5498
5499 *nexthop_len += nla_total_size(0) /* RTA_MULTIPATH */
5500 + NLA_ALIGN(sizeof(struct rtnexthop))
5501 + nla_total_size(16); /* RTA_GATEWAY */
5502
5503 if (nh->fib_nh_lws) {
5504 /* RTA_ENCAP_TYPE */
5505 *nexthop_len += lwtunnel_get_encap_size(nh->fib_nh_lws);
5506 /* RTA_ENCAP */
5507 *nexthop_len += nla_total_size(2);
5508 }
5509
5510 return 0;
5511}
5512
5513static size_t rt6_nlmsg_size(struct fib6_info *f6i)
5514{
5515 int nexthop_len;
5516
5517 if (f6i->nh) {
5518 nexthop_len = nla_total_size(4); /* RTA_NH_ID */
5519 nexthop_for_each_fib6_nh(f6i->nh, rt6_nh_nlmsg_size,
5520 &nexthop_len);
5521 } else {
5522 struct fib6_nh *nh = f6i->fib6_nh;
5523
5524 nexthop_len = 0;
5525 if (f6i->fib6_nsiblings) {
5526 nexthop_len = nla_total_size(0) /* RTA_MULTIPATH */
5527 + NLA_ALIGN(sizeof(struct rtnexthop))
5528 + nla_total_size(16) /* RTA_GATEWAY */
5529 + lwtunnel_get_encap_size(nh->fib_nh_lws);
5530
5531 nexthop_len *= f6i->fib6_nsiblings;
5532 }
5533 nexthop_len += lwtunnel_get_encap_size(nh->fib_nh_lws);
5534 }
5535
5536 return NLMSG_ALIGN(sizeof(struct rtmsg))
5537 + nla_total_size(16) /* RTA_SRC */
5538 + nla_total_size(16) /* RTA_DST */
5539 + nla_total_size(16) /* RTA_GATEWAY */
5540 + nla_total_size(16) /* RTA_PREFSRC */
5541 + nla_total_size(4) /* RTA_TABLE */
5542 + nla_total_size(4) /* RTA_IIF */
5543 + nla_total_size(4) /* RTA_OIF */
5544 + nla_total_size(4) /* RTA_PRIORITY */
5545 + RTAX_MAX * nla_total_size(4) /* RTA_METRICS */
5546 + nla_total_size(sizeof(struct rta_cacheinfo))
5547 + nla_total_size(TCP_CA_NAME_MAX) /* RTAX_CC_ALGO */
5548 + nla_total_size(1) /* RTA_PREF */
5549 + nexthop_len;
5550}
5551
5552static int rt6_fill_node_nexthop(struct sk_buff *skb, struct nexthop *nh,
5553 unsigned char *flags)
5554{
5555 if (nexthop_is_multipath(nh)) {
5556 struct nlattr *mp;
5557
5558 mp = nla_nest_start_noflag(skb, RTA_MULTIPATH);
5559 if (!mp)
5560 goto nla_put_failure;
5561
5562 if (nexthop_mpath_fill_node(skb, nh, AF_INET6))
5563 goto nla_put_failure;
5564
5565 nla_nest_end(skb, mp);
5566 } else {
5567 struct fib6_nh *fib6_nh;
5568
5569 fib6_nh = nexthop_fib6_nh(nh);
5570 if (fib_nexthop_info(skb, &fib6_nh->nh_common, AF_INET6,
5571 flags, false) < 0)
5572 goto nla_put_failure;
5573 }
5574
5575 return 0;
5576
5577nla_put_failure:
5578 return -EMSGSIZE;
5579}
5580
5581static int rt6_fill_node(struct net *net, struct sk_buff *skb,
5582 struct fib6_info *rt, struct dst_entry *dst,
5583 struct in6_addr *dest, struct in6_addr *src,
5584 int iif, int type, u32 portid, u32 seq,
5585 unsigned int flags)
5586{
5587 struct rt6_info *rt6 = (struct rt6_info *)dst;
5588 struct rt6key *rt6_dst, *rt6_src;
5589 u32 *pmetrics, table, rt6_flags;
5590 unsigned char nh_flags = 0;
5591 struct nlmsghdr *nlh;
5592 struct rtmsg *rtm;
5593 long expires = 0;
5594
5595 nlh = nlmsg_put(skb, portid, seq, type, sizeof(*rtm), flags);
5596 if (!nlh)
5597 return -EMSGSIZE;
5598
5599 if (rt6) {
5600 rt6_dst = &rt6->rt6i_dst;
5601 rt6_src = &rt6->rt6i_src;
5602 rt6_flags = rt6->rt6i_flags;
5603 } else {
5604 rt6_dst = &rt->fib6_dst;
5605 rt6_src = &rt->fib6_src;
5606 rt6_flags = rt->fib6_flags;
5607 }
5608
5609 rtm = nlmsg_data(nlh);
5610 rtm->rtm_family = AF_INET6;
5611 rtm->rtm_dst_len = rt6_dst->plen;
5612 rtm->rtm_src_len = rt6_src->plen;
5613 rtm->rtm_tos = 0;
5614 if (rt->fib6_table)
5615 table = rt->fib6_table->tb6_id;
5616 else
5617 table = RT6_TABLE_UNSPEC;
5618 rtm->rtm_table = table < 256 ? table : RT_TABLE_COMPAT;
5619 if (nla_put_u32(skb, RTA_TABLE, table))
5620 goto nla_put_failure;
5621
5622 rtm->rtm_type = rt->fib6_type;
5623 rtm->rtm_flags = 0;
5624 rtm->rtm_scope = RT_SCOPE_UNIVERSE;
5625 rtm->rtm_protocol = rt->fib6_protocol;
5626
5627 if (rt6_flags & RTF_CACHE)
5628 rtm->rtm_flags |= RTM_F_CLONED;
5629
5630 if (dest) {
5631 if (nla_put_in6_addr(skb, RTA_DST, dest))
5632 goto nla_put_failure;
5633 rtm->rtm_dst_len = 128;
5634 } else if (rtm->rtm_dst_len)
5635 if (nla_put_in6_addr(skb, RTA_DST, &rt6_dst->addr))
5636 goto nla_put_failure;
5637#ifdef CONFIG_IPV6_SUBTREES
5638 if (src) {
5639 if (nla_put_in6_addr(skb, RTA_SRC, src))
5640 goto nla_put_failure;
5641 rtm->rtm_src_len = 128;
5642 } else if (rtm->rtm_src_len &&
5643 nla_put_in6_addr(skb, RTA_SRC, &rt6_src->addr))
5644 goto nla_put_failure;
5645#endif
5646 if (iif) {
5647#ifdef CONFIG_IPV6_MROUTE
5648 if (ipv6_addr_is_multicast(&rt6_dst->addr)) {
5649 int err = ip6mr_get_route(net, skb, rtm, portid);
5650
5651 if (err == 0)
5652 return 0;
5653 if (err < 0)
5654 goto nla_put_failure;
5655 } else
5656#endif
5657 if (nla_put_u32(skb, RTA_IIF, iif))
5658 goto nla_put_failure;
5659 } else if (dest) {
5660 struct in6_addr saddr_buf;
5661 if (ip6_route_get_saddr(net, rt, dest, 0, &saddr_buf) == 0 &&
5662 nla_put_in6_addr(skb, RTA_PREFSRC, &saddr_buf))
5663 goto nla_put_failure;
5664 }
5665
5666 if (rt->fib6_prefsrc.plen) {
5667 struct in6_addr saddr_buf;
5668 saddr_buf = rt->fib6_prefsrc.addr;
5669 if (nla_put_in6_addr(skb, RTA_PREFSRC, &saddr_buf))
5670 goto nla_put_failure;
5671 }
5672
5673 pmetrics = dst ? dst_metrics_ptr(dst) : rt->fib6_metrics->metrics;
5674 if (rtnetlink_put_metrics(skb, pmetrics) < 0)
5675 goto nla_put_failure;
5676
5677 if (nla_put_u32(skb, RTA_PRIORITY, rt->fib6_metric))
5678 goto nla_put_failure;
5679
5680 /* For multipath routes, walk the siblings list and add
5681 * each as a nexthop within RTA_MULTIPATH.
5682 */
5683 if (rt6) {
5684 if (rt6_flags & RTF_GATEWAY &&
5685 nla_put_in6_addr(skb, RTA_GATEWAY, &rt6->rt6i_gateway))
5686 goto nla_put_failure;
5687
5688 if (dst->dev && nla_put_u32(skb, RTA_OIF, dst->dev->ifindex))
5689 goto nla_put_failure;
5690
5691 if (dst->lwtstate &&
5692 lwtunnel_fill_encap(skb, dst->lwtstate, RTA_ENCAP, RTA_ENCAP_TYPE) < 0)
5693 goto nla_put_failure;
5694 } else if (rt->fib6_nsiblings) {
5695 struct fib6_info *sibling, *next_sibling;
5696 struct nlattr *mp;
5697
5698 mp = nla_nest_start_noflag(skb, RTA_MULTIPATH);
5699 if (!mp)
5700 goto nla_put_failure;
5701
5702 if (fib_add_nexthop(skb, &rt->fib6_nh->nh_common,
5703 rt->fib6_nh->fib_nh_weight, AF_INET6,
5704 0) < 0)
5705 goto nla_put_failure;
5706
5707 list_for_each_entry_safe(sibling, next_sibling,
5708 &rt->fib6_siblings, fib6_siblings) {
5709 if (fib_add_nexthop(skb, &sibling->fib6_nh->nh_common,
5710 sibling->fib6_nh->fib_nh_weight,
5711 AF_INET6, 0) < 0)
5712 goto nla_put_failure;
5713 }
5714
5715 nla_nest_end(skb, mp);
5716 } else if (rt->nh) {
5717 if (nla_put_u32(skb, RTA_NH_ID, rt->nh->id))
5718 goto nla_put_failure;
5719
5720 if (nexthop_is_blackhole(rt->nh))
5721 rtm->rtm_type = RTN_BLACKHOLE;
5722
5723 if (net->ipv4.sysctl_nexthop_compat_mode &&
5724 rt6_fill_node_nexthop(skb, rt->nh, &nh_flags) < 0)
5725 goto nla_put_failure;
5726
5727 rtm->rtm_flags |= nh_flags;
5728 } else {
5729 if (fib_nexthop_info(skb, &rt->fib6_nh->nh_common, AF_INET6,
5730 &nh_flags, false) < 0)
5731 goto nla_put_failure;
5732
5733 rtm->rtm_flags |= nh_flags;
5734 }
5735
5736 if (rt6_flags & RTF_EXPIRES) {
5737 expires = dst ? dst->expires : rt->expires;
5738 expires -= jiffies;
5739 }
5740
5741 if (!dst) {
5742 if (rt->offload)
5743 rtm->rtm_flags |= RTM_F_OFFLOAD;
5744 if (rt->trap)
5745 rtm->rtm_flags |= RTM_F_TRAP;
5746 if (rt->offload_failed)
5747 rtm->rtm_flags |= RTM_F_OFFLOAD_FAILED;
5748 }
5749
5750 if (rtnl_put_cacheinfo(skb, dst, 0, expires, dst ? dst->error : 0) < 0)
5751 goto nla_put_failure;
5752
5753 if (nla_put_u8(skb, RTA_PREF, IPV6_EXTRACT_PREF(rt6_flags)))
5754 goto nla_put_failure;
5755
5756
5757 nlmsg_end(skb, nlh);
5758 return 0;
5759
5760nla_put_failure:
5761 nlmsg_cancel(skb, nlh);
5762 return -EMSGSIZE;
5763}
5764
5765static int fib6_info_nh_uses_dev(struct fib6_nh *nh, void *arg)
5766{
5767 const struct net_device *dev = arg;
5768
5769 if (nh->fib_nh_dev == dev)
5770 return 1;
5771
5772 return 0;
5773}
5774
5775static bool fib6_info_uses_dev(const struct fib6_info *f6i,
5776 const struct net_device *dev)
5777{
5778 if (f6i->nh) {
5779 struct net_device *_dev = (struct net_device *)dev;
5780
5781 return !!nexthop_for_each_fib6_nh(f6i->nh,
5782 fib6_info_nh_uses_dev,
5783 _dev);
5784 }
5785
5786 if (f6i->fib6_nh->fib_nh_dev == dev)
5787 return true;
5788
5789 if (f6i->fib6_nsiblings) {
5790 struct fib6_info *sibling, *next_sibling;
5791
5792 list_for_each_entry_safe(sibling, next_sibling,
5793 &f6i->fib6_siblings, fib6_siblings) {
5794 if (sibling->fib6_nh->fib_nh_dev == dev)
5795 return true;
5796 }
5797 }
5798
5799 return false;
5800}
5801
5802struct fib6_nh_exception_dump_walker {
5803 struct rt6_rtnl_dump_arg *dump;
5804 struct fib6_info *rt;
5805 unsigned int flags;
5806 unsigned int skip;
5807 unsigned int count;
5808};
5809
5810static int rt6_nh_dump_exceptions(struct fib6_nh *nh, void *arg)
5811{
5812 struct fib6_nh_exception_dump_walker *w = arg;
5813 struct rt6_rtnl_dump_arg *dump = w->dump;
5814 struct rt6_exception_bucket *bucket;
5815 struct rt6_exception *rt6_ex;
5816 int i, err;
5817
5818 bucket = fib6_nh_get_excptn_bucket(nh, NULL);
5819 if (!bucket)
5820 return 0;
5821
5822 for (i = 0; i < FIB6_EXCEPTION_BUCKET_SIZE; i++) {
5823 hlist_for_each_entry(rt6_ex, &bucket->chain, hlist) {
5824 if (w->skip) {
5825 w->skip--;
5826 continue;
5827 }
5828
5829 /* Expiration of entries doesn't bump sernum, insertion
5830 * does. Removal is triggered by insertion, so we can
5831 * rely on the fact that if entries change between two
5832 * partial dumps, this node is scanned again completely,
5833 * see rt6_insert_exception() and fib6_dump_table().
5834 *
5835 * Count expired entries we go through as handled
5836 * entries that we'll skip next time, in case of partial
5837 * node dump. Otherwise, if entries expire meanwhile,
5838 * we'll skip the wrong amount.
5839 */
5840 if (rt6_check_expired(rt6_ex->rt6i)) {
5841 w->count++;
5842 continue;
5843 }
5844
5845 err = rt6_fill_node(dump->net, dump->skb, w->rt,
5846 &rt6_ex->rt6i->dst, NULL, NULL, 0,
5847 RTM_NEWROUTE,
5848 NETLINK_CB(dump->cb->skb).portid,
5849 dump->cb->nlh->nlmsg_seq, w->flags);
5850 if (err)
5851 return err;
5852
5853 w->count++;
5854 }
5855 bucket++;
5856 }
5857
5858 return 0;
5859}
5860
5861/* Return -1 if done with node, number of handled routes on partial dump */
5862int rt6_dump_route(struct fib6_info *rt, void *p_arg, unsigned int skip)
5863{
5864 struct rt6_rtnl_dump_arg *arg = (struct rt6_rtnl_dump_arg *) p_arg;
5865 struct fib_dump_filter *filter = &arg->filter;
5866 unsigned int flags = NLM_F_MULTI;
5867 struct net *net = arg->net;
5868 int count = 0;
5869
5870 if (rt == net->ipv6.fib6_null_entry)
5871 return -1;
5872
5873 if ((filter->flags & RTM_F_PREFIX) &&
5874 !(rt->fib6_flags & RTF_PREFIX_RT)) {
5875 /* success since this is not a prefix route */
5876 return -1;
5877 }
5878 if (filter->filter_set &&
5879 ((filter->rt_type && rt->fib6_type != filter->rt_type) ||
5880 (filter->dev && !fib6_info_uses_dev(rt, filter->dev)) ||
5881 (filter->protocol && rt->fib6_protocol != filter->protocol))) {
5882 return -1;
5883 }
5884
5885 if (filter->filter_set ||
5886 !filter->dump_routes || !filter->dump_exceptions) {
5887 flags |= NLM_F_DUMP_FILTERED;
5888 }
5889
5890 if (filter->dump_routes) {
5891 if (skip) {
5892 skip--;
5893 } else {
5894 if (rt6_fill_node(net, arg->skb, rt, NULL, NULL, NULL,
5895 0, RTM_NEWROUTE,
5896 NETLINK_CB(arg->cb->skb).portid,
5897 arg->cb->nlh->nlmsg_seq, flags)) {
5898 return 0;
5899 }
5900 count++;
5901 }
5902 }
5903
5904 if (filter->dump_exceptions) {
5905 struct fib6_nh_exception_dump_walker w = { .dump = arg,
5906 .rt = rt,
5907 .flags = flags,
5908 .skip = skip,
5909 .count = 0 };
5910 int err;
5911
5912 rcu_read_lock();
5913 if (rt->nh) {
5914 err = nexthop_for_each_fib6_nh(rt->nh,
5915 rt6_nh_dump_exceptions,
5916 &w);
5917 } else {
5918 err = rt6_nh_dump_exceptions(rt->fib6_nh, &w);
5919 }
5920 rcu_read_unlock();
5921
5922 if (err)
5923 return count += w.count;
5924 }
5925
5926 return -1;
5927}
5928
5929static int inet6_rtm_valid_getroute_req(struct sk_buff *skb,
5930 const struct nlmsghdr *nlh,
5931 struct nlattr **tb,
5932 struct netlink_ext_ack *extack)
5933{
5934 struct rtmsg *rtm;
5935 int i, err;
5936
5937 if (nlh->nlmsg_len < nlmsg_msg_size(sizeof(*rtm))) {
5938 NL_SET_ERR_MSG_MOD(extack,
5939 "Invalid header for get route request");
5940 return -EINVAL;
5941 }
5942
5943 if (!netlink_strict_get_check(skb))
5944 return nlmsg_parse_deprecated(nlh, sizeof(*rtm), tb, RTA_MAX,
5945 rtm_ipv6_policy, extack);
5946
5947 rtm = nlmsg_data(nlh);
5948 if ((rtm->rtm_src_len && rtm->rtm_src_len != 128) ||
5949 (rtm->rtm_dst_len && rtm->rtm_dst_len != 128) ||
5950 rtm->rtm_table || rtm->rtm_protocol || rtm->rtm_scope ||
5951 rtm->rtm_type) {
5952 NL_SET_ERR_MSG_MOD(extack, "Invalid values in header for get route request");
5953 return -EINVAL;
5954 }
5955 if (rtm->rtm_flags & ~RTM_F_FIB_MATCH) {
5956 NL_SET_ERR_MSG_MOD(extack,
5957 "Invalid flags for get route request");
5958 return -EINVAL;
5959 }
5960
5961 err = nlmsg_parse_deprecated_strict(nlh, sizeof(*rtm), tb, RTA_MAX,
5962 rtm_ipv6_policy, extack);
5963 if (err)
5964 return err;
5965
5966 if ((tb[RTA_SRC] && !rtm->rtm_src_len) ||
5967 (tb[RTA_DST] && !rtm->rtm_dst_len)) {
5968 NL_SET_ERR_MSG_MOD(extack, "rtm_src_len and rtm_dst_len must be 128 for IPv6");
5969 return -EINVAL;
5970 }
5971
5972 for (i = 0; i <= RTA_MAX; i++) {
5973 if (!tb[i])
5974 continue;
5975
5976 switch (i) {
5977 case RTA_SRC:
5978 case RTA_DST:
5979 case RTA_IIF:
5980 case RTA_OIF:
5981 case RTA_MARK:
5982 case RTA_UID:
5983 case RTA_SPORT:
5984 case RTA_DPORT:
5985 case RTA_IP_PROTO:
5986 break;
5987 default:
5988 NL_SET_ERR_MSG_MOD(extack, "Unsupported attribute in get route request");
5989 return -EINVAL;
5990 }
5991 }
5992
5993 return 0;
5994}
5995
5996static int inet6_rtm_getroute(struct sk_buff *in_skb, struct nlmsghdr *nlh,
5997 struct netlink_ext_ack *extack)
5998{
5999 struct net *net = sock_net(in_skb->sk);
6000 struct nlattr *tb[RTA_MAX+1];
6001 int err, iif = 0, oif = 0;
6002 struct fib6_info *from;
6003 struct dst_entry *dst;
6004 struct rt6_info *rt;
6005 struct sk_buff *skb;
6006 struct rtmsg *rtm;
6007 struct flowi6 fl6 = {};
6008 bool fibmatch;
6009
6010 err = inet6_rtm_valid_getroute_req(in_skb, nlh, tb, extack);
6011 if (err < 0)
6012 goto errout;
6013
6014 err = -EINVAL;
6015 rtm = nlmsg_data(nlh);
6016 fl6.flowlabel = ip6_make_flowinfo(rtm->rtm_tos, 0);
6017 fibmatch = !!(rtm->rtm_flags & RTM_F_FIB_MATCH);
6018
6019 if (tb[RTA_SRC]) {
6020 if (nla_len(tb[RTA_SRC]) < sizeof(struct in6_addr))
6021 goto errout;
6022
6023 fl6.saddr = *(struct in6_addr *)nla_data(tb[RTA_SRC]);
6024 }
6025
6026 if (tb[RTA_DST]) {
6027 if (nla_len(tb[RTA_DST]) < sizeof(struct in6_addr))
6028 goto errout;
6029
6030 fl6.daddr = *(struct in6_addr *)nla_data(tb[RTA_DST]);
6031 }
6032
6033 if (tb[RTA_IIF])
6034 iif = nla_get_u32(tb[RTA_IIF]);
6035
6036 if (tb[RTA_OIF])
6037 oif = nla_get_u32(tb[RTA_OIF]);
6038
6039 if (tb[RTA_MARK])
6040 fl6.flowi6_mark = nla_get_u32(tb[RTA_MARK]);
6041
6042 if (tb[RTA_UID])
6043 fl6.flowi6_uid = make_kuid(current_user_ns(),
6044 nla_get_u32(tb[RTA_UID]));
6045 else
6046 fl6.flowi6_uid = iif ? INVALID_UID : current_uid();
6047
6048 if (tb[RTA_SPORT])
6049 fl6.fl6_sport = nla_get_be16(tb[RTA_SPORT]);
6050
6051 if (tb[RTA_DPORT])
6052 fl6.fl6_dport = nla_get_be16(tb[RTA_DPORT]);
6053
6054 if (tb[RTA_IP_PROTO]) {
6055 err = rtm_getroute_parse_ip_proto(tb[RTA_IP_PROTO],
6056 &fl6.flowi6_proto, AF_INET6,
6057 extack);
6058 if (err)
6059 goto errout;
6060 }
6061
6062 if (iif) {
6063 struct net_device *dev;
6064 int flags = 0;
6065
6066 rcu_read_lock();
6067
6068 dev = dev_get_by_index_rcu(net, iif);
6069 if (!dev) {
6070 rcu_read_unlock();
6071 err = -ENODEV;
6072 goto errout;
6073 }
6074
6075 fl6.flowi6_iif = iif;
6076
6077 if (!ipv6_addr_any(&fl6.saddr))
6078 flags |= RT6_LOOKUP_F_HAS_SADDR;
6079
6080 dst = ip6_route_input_lookup(net, dev, &fl6, NULL, flags);
6081
6082 rcu_read_unlock();
6083 } else {
6084 fl6.flowi6_oif = oif;
6085
6086 dst = ip6_route_output(net, NULL, &fl6);
6087 }
6088
6089
6090 rt = container_of(dst, struct rt6_info, dst);
6091 if (rt->dst.error) {
6092 err = rt->dst.error;
6093 ip6_rt_put(rt);
6094 goto errout;
6095 }
6096
6097 if (rt == net->ipv6.ip6_null_entry) {
6098 err = rt->dst.error;
6099 ip6_rt_put(rt);
6100 goto errout;
6101 }
6102
6103 skb = alloc_skb(NLMSG_GOODSIZE, GFP_KERNEL);
6104 if (!skb) {
6105 ip6_rt_put(rt);
6106 err = -ENOBUFS;
6107 goto errout;
6108 }
6109
6110 skb_dst_set(skb, &rt->dst);
6111
6112 rcu_read_lock();
6113 from = rcu_dereference(rt->from);
6114 if (from) {
6115 if (fibmatch)
6116 err = rt6_fill_node(net, skb, from, NULL, NULL, NULL,
6117 iif, RTM_NEWROUTE,
6118 NETLINK_CB(in_skb).portid,
6119 nlh->nlmsg_seq, 0);
6120 else
6121 err = rt6_fill_node(net, skb, from, dst, &fl6.daddr,
6122 &fl6.saddr, iif, RTM_NEWROUTE,
6123 NETLINK_CB(in_skb).portid,
6124 nlh->nlmsg_seq, 0);
6125 } else {
6126 err = -ENETUNREACH;
6127 }
6128 rcu_read_unlock();
6129
6130 if (err < 0) {
6131 kfree_skb(skb);
6132 goto errout;
6133 }
6134
6135 err = rtnl_unicast(skb, net, NETLINK_CB(in_skb).portid);
6136errout:
6137 return err;
6138}
6139
6140void inet6_rt_notify(int event, struct fib6_info *rt, struct nl_info *info,
6141 unsigned int nlm_flags)
6142{
6143 struct sk_buff *skb;
6144 struct net *net = info->nl_net;
6145 u32 seq;
6146 int err;
6147
6148 err = -ENOBUFS;
6149 seq = info->nlh ? info->nlh->nlmsg_seq : 0;
6150
6151 skb = nlmsg_new(rt6_nlmsg_size(rt), gfp_any());
6152 if (!skb)
6153 goto errout;
6154
6155 err = rt6_fill_node(net, skb, rt, NULL, NULL, NULL, 0,
6156 event, info->portid, seq, nlm_flags);
6157 if (err < 0) {
6158 /* -EMSGSIZE implies BUG in rt6_nlmsg_size() */
6159 WARN_ON(err == -EMSGSIZE);
6160 kfree_skb(skb);
6161 goto errout;
6162 }
6163 rtnl_notify(skb, net, info->portid, RTNLGRP_IPV6_ROUTE,
6164 info->nlh, gfp_any());
6165 return;
6166errout:
6167 if (err < 0)
6168 rtnl_set_sk_err(net, RTNLGRP_IPV6_ROUTE, err);
6169}
6170
6171void fib6_rt_update(struct net *net, struct fib6_info *rt,
6172 struct nl_info *info)
6173{
6174 u32 seq = info->nlh ? info->nlh->nlmsg_seq : 0;
6175 struct sk_buff *skb;
6176 int err = -ENOBUFS;
6177
6178 skb = nlmsg_new(rt6_nlmsg_size(rt), gfp_any());
6179 if (!skb)
6180 goto errout;
6181
6182 err = rt6_fill_node(net, skb, rt, NULL, NULL, NULL, 0,
6183 RTM_NEWROUTE, info->portid, seq, NLM_F_REPLACE);
6184 if (err < 0) {
6185 /* -EMSGSIZE implies BUG in rt6_nlmsg_size() */
6186 WARN_ON(err == -EMSGSIZE);
6187 kfree_skb(skb);
6188 goto errout;
6189 }
6190 rtnl_notify(skb, net, info->portid, RTNLGRP_IPV6_ROUTE,
6191 info->nlh, gfp_any());
6192 return;
6193errout:
6194 if (err < 0)
6195 rtnl_set_sk_err(net, RTNLGRP_IPV6_ROUTE, err);
6196}
6197
6198void fib6_info_hw_flags_set(struct net *net, struct fib6_info *f6i,
6199 bool offload, bool trap, bool offload_failed)
6200{
6201 struct sk_buff *skb;
6202 int err;
6203
6204 if (f6i->offload == offload && f6i->trap == trap &&
6205 f6i->offload_failed == offload_failed)
6206 return;
6207
6208 f6i->offload = offload;
6209 f6i->trap = trap;
6210
6211 /* 2 means send notifications only if offload_failed was changed. */
6212 if (net->ipv6.sysctl.fib_notify_on_flag_change == 2 &&
6213 f6i->offload_failed == offload_failed)
6214 return;
6215
6216 f6i->offload_failed = offload_failed;
6217
6218 if (!rcu_access_pointer(f6i->fib6_node))
6219 /* The route was removed from the tree, do not send
6220 * notification.
6221 */
6222 return;
6223
6224 if (!net->ipv6.sysctl.fib_notify_on_flag_change)
6225 return;
6226
6227 skb = nlmsg_new(rt6_nlmsg_size(f6i), GFP_KERNEL);
6228 if (!skb) {
6229 err = -ENOBUFS;
6230 goto errout;
6231 }
6232
6233 err = rt6_fill_node(net, skb, f6i, NULL, NULL, NULL, 0, RTM_NEWROUTE, 0,
6234 0, 0);
6235 if (err < 0) {
6236 /* -EMSGSIZE implies BUG in rt6_nlmsg_size() */
6237 WARN_ON(err == -EMSGSIZE);
6238 kfree_skb(skb);
6239 goto errout;
6240 }
6241
6242 rtnl_notify(skb, net, 0, RTNLGRP_IPV6_ROUTE, NULL, GFP_KERNEL);
6243 return;
6244
6245errout:
6246 rtnl_set_sk_err(net, RTNLGRP_IPV6_ROUTE, err);
6247}
6248EXPORT_SYMBOL(fib6_info_hw_flags_set);
6249
6250static int ip6_route_dev_notify(struct notifier_block *this,
6251 unsigned long event, void *ptr)
6252{
6253 struct net_device *dev = netdev_notifier_info_to_dev(ptr);
6254 struct net *net = dev_net(dev);
6255
6256 if (!(dev->flags & IFF_LOOPBACK))
6257 return NOTIFY_OK;
6258
6259 if (event == NETDEV_REGISTER) {
6260 net->ipv6.fib6_null_entry->fib6_nh->fib_nh_dev = dev;
6261 net->ipv6.ip6_null_entry->dst.dev = dev;
6262 net->ipv6.ip6_null_entry->rt6i_idev = in6_dev_get(dev);
6263#ifdef CONFIG_IPV6_MULTIPLE_TABLES
6264 net->ipv6.ip6_prohibit_entry->dst.dev = dev;
6265 net->ipv6.ip6_prohibit_entry->rt6i_idev = in6_dev_get(dev);
6266 net->ipv6.ip6_blk_hole_entry->dst.dev = dev;
6267 net->ipv6.ip6_blk_hole_entry->rt6i_idev = in6_dev_get(dev);
6268#endif
6269 } else if (event == NETDEV_UNREGISTER &&
6270 dev->reg_state != NETREG_UNREGISTERED) {
6271 /* NETDEV_UNREGISTER could be fired for multiple times by
6272 * netdev_wait_allrefs(). Make sure we only call this once.
6273 */
6274 in6_dev_put_clear(&net->ipv6.ip6_null_entry->rt6i_idev);
6275#ifdef CONFIG_IPV6_MULTIPLE_TABLES
6276 in6_dev_put_clear(&net->ipv6.ip6_prohibit_entry->rt6i_idev);
6277 in6_dev_put_clear(&net->ipv6.ip6_blk_hole_entry->rt6i_idev);
6278#endif
6279 }
6280
6281 return NOTIFY_OK;
6282}
6283
6284/*
6285 * /proc
6286 */
6287
6288#ifdef CONFIG_PROC_FS
6289static int rt6_stats_seq_show(struct seq_file *seq, void *v)
6290{
6291 struct net *net = (struct net *)seq->private;
6292 seq_printf(seq, "%04x %04x %04x %04x %04x %04x %04x\n",
6293 net->ipv6.rt6_stats->fib_nodes,
6294 net->ipv6.rt6_stats->fib_route_nodes,
6295 atomic_read(&net->ipv6.rt6_stats->fib_rt_alloc),
6296 net->ipv6.rt6_stats->fib_rt_entries,
6297 net->ipv6.rt6_stats->fib_rt_cache,
6298 dst_entries_get_slow(&net->ipv6.ip6_dst_ops),
6299 net->ipv6.rt6_stats->fib_discarded_routes);
6300
6301 return 0;
6302}
6303#endif /* CONFIG_PROC_FS */
6304
6305#ifdef CONFIG_SYSCTL
6306
6307static int ipv6_sysctl_rtcache_flush(struct ctl_table *ctl, int write,
6308 void *buffer, size_t *lenp, loff_t *ppos)
6309{
6310 struct net *net;
6311 int delay;
6312 int ret;
6313 if (!write)
6314 return -EINVAL;
6315
6316 net = (struct net *)ctl->extra1;
6317 delay = net->ipv6.sysctl.flush_delay;
6318 ret = proc_dointvec(ctl, write, buffer, lenp, ppos);
6319 if (ret)
6320 return ret;
6321
6322 fib6_run_gc(delay <= 0 ? 0 : (unsigned long)delay, net, delay > 0);
6323 return 0;
6324}
6325
6326static struct ctl_table ipv6_route_table_template[] = {
6327 {
6328 .procname = "flush",
6329 .data = &init_net.ipv6.sysctl.flush_delay,
6330 .maxlen = sizeof(int),
6331 .mode = 0200,
6332 .proc_handler = ipv6_sysctl_rtcache_flush
6333 },
6334 {
6335 .procname = "gc_thresh",
6336 .data = &ip6_dst_ops_template.gc_thresh,
6337 .maxlen = sizeof(int),
6338 .mode = 0644,
6339 .proc_handler = proc_dointvec,
6340 },
6341 {
6342 .procname = "max_size",
6343 .data = &init_net.ipv6.sysctl.ip6_rt_max_size,
6344 .maxlen = sizeof(int),
6345 .mode = 0644,
6346 .proc_handler = proc_dointvec,
6347 },
6348 {
6349 .procname = "gc_min_interval",
6350 .data = &init_net.ipv6.sysctl.ip6_rt_gc_min_interval,
6351 .maxlen = sizeof(int),
6352 .mode = 0644,
6353 .proc_handler = proc_dointvec_jiffies,
6354 },
6355 {
6356 .procname = "gc_timeout",
6357 .data = &init_net.ipv6.sysctl.ip6_rt_gc_timeout,
6358 .maxlen = sizeof(int),
6359 .mode = 0644,
6360 .proc_handler = proc_dointvec_jiffies,
6361 },
6362 {
6363 .procname = "gc_interval",
6364 .data = &init_net.ipv6.sysctl.ip6_rt_gc_interval,
6365 .maxlen = sizeof(int),
6366 .mode = 0644,
6367 .proc_handler = proc_dointvec_jiffies,
6368 },
6369 {
6370 .procname = "gc_elasticity",
6371 .data = &init_net.ipv6.sysctl.ip6_rt_gc_elasticity,
6372 .maxlen = sizeof(int),
6373 .mode = 0644,
6374 .proc_handler = proc_dointvec,
6375 },
6376 {
6377 .procname = "mtu_expires",
6378 .data = &init_net.ipv6.sysctl.ip6_rt_mtu_expires,
6379 .maxlen = sizeof(int),
6380 .mode = 0644,
6381 .proc_handler = proc_dointvec_jiffies,
6382 },
6383 {
6384 .procname = "min_adv_mss",
6385 .data = &init_net.ipv6.sysctl.ip6_rt_min_advmss,
6386 .maxlen = sizeof(int),
6387 .mode = 0644,
6388 .proc_handler = proc_dointvec,
6389 },
6390 {
6391 .procname = "gc_min_interval_ms",
6392 .data = &init_net.ipv6.sysctl.ip6_rt_gc_min_interval,
6393 .maxlen = sizeof(int),
6394 .mode = 0644,
6395 .proc_handler = proc_dointvec_ms_jiffies,
6396 },
6397 {
6398 .procname = "skip_notify_on_dev_down",
6399 .data = &init_net.ipv6.sysctl.skip_notify_on_dev_down,
6400 .maxlen = sizeof(int),
6401 .mode = 0644,
6402 .proc_handler = proc_dointvec_minmax,
6403 .extra1 = SYSCTL_ZERO,
6404 .extra2 = SYSCTL_ONE,
6405 },
6406 { }
6407};
6408
6409struct ctl_table * __net_init ipv6_route_sysctl_init(struct net *net)
6410{
6411 struct ctl_table *table;
6412
6413 table = kmemdup(ipv6_route_table_template,
6414 sizeof(ipv6_route_table_template),
6415 GFP_KERNEL);
6416
6417 if (table) {
6418 table[0].data = &net->ipv6.sysctl.flush_delay;
6419 table[0].extra1 = net;
6420 table[1].data = &net->ipv6.ip6_dst_ops.gc_thresh;
6421 table[2].data = &net->ipv6.sysctl.ip6_rt_max_size;
6422 table[3].data = &net->ipv6.sysctl.ip6_rt_gc_min_interval;
6423 table[4].data = &net->ipv6.sysctl.ip6_rt_gc_timeout;
6424 table[5].data = &net->ipv6.sysctl.ip6_rt_gc_interval;
6425 table[6].data = &net->ipv6.sysctl.ip6_rt_gc_elasticity;
6426 table[7].data = &net->ipv6.sysctl.ip6_rt_mtu_expires;
6427 table[8].data = &net->ipv6.sysctl.ip6_rt_min_advmss;
6428 table[9].data = &net->ipv6.sysctl.ip6_rt_gc_min_interval;
6429 table[10].data = &net->ipv6.sysctl.skip_notify_on_dev_down;
6430
6431 /* Don't export sysctls to unprivileged users */
6432 if (net->user_ns != &init_user_ns)
6433 table[0].procname = NULL;
6434 }
6435
6436 return table;
6437}
6438#endif
6439
6440static int __net_init ip6_route_net_init(struct net *net)
6441{
6442 int ret = -ENOMEM;
6443
6444 memcpy(&net->ipv6.ip6_dst_ops, &ip6_dst_ops_template,
6445 sizeof(net->ipv6.ip6_dst_ops));
6446
6447 if (dst_entries_init(&net->ipv6.ip6_dst_ops) < 0)
6448 goto out_ip6_dst_ops;
6449
6450 net->ipv6.fib6_null_entry = fib6_info_alloc(GFP_KERNEL, true);
6451 if (!net->ipv6.fib6_null_entry)
6452 goto out_ip6_dst_entries;
6453 memcpy(net->ipv6.fib6_null_entry, &fib6_null_entry_template,
6454 sizeof(*net->ipv6.fib6_null_entry));
6455
6456 net->ipv6.ip6_null_entry = kmemdup(&ip6_null_entry_template,
6457 sizeof(*net->ipv6.ip6_null_entry),
6458 GFP_KERNEL);
6459 if (!net->ipv6.ip6_null_entry)
6460 goto out_fib6_null_entry;
6461 net->ipv6.ip6_null_entry->dst.ops = &net->ipv6.ip6_dst_ops;
6462 dst_init_metrics(&net->ipv6.ip6_null_entry->dst,
6463 ip6_template_metrics, true);
6464 INIT_LIST_HEAD(&net->ipv6.ip6_null_entry->rt6i_uncached);
6465
6466#ifdef CONFIG_IPV6_MULTIPLE_TABLES
6467 net->ipv6.fib6_has_custom_rules = false;
6468 net->ipv6.ip6_prohibit_entry = kmemdup(&ip6_prohibit_entry_template,
6469 sizeof(*net->ipv6.ip6_prohibit_entry),
6470 GFP_KERNEL);
6471 if (!net->ipv6.ip6_prohibit_entry)
6472 goto out_ip6_null_entry;
6473 net->ipv6.ip6_prohibit_entry->dst.ops = &net->ipv6.ip6_dst_ops;
6474 dst_init_metrics(&net->ipv6.ip6_prohibit_entry->dst,
6475 ip6_template_metrics, true);
6476 INIT_LIST_HEAD(&net->ipv6.ip6_prohibit_entry->rt6i_uncached);
6477
6478 net->ipv6.ip6_blk_hole_entry = kmemdup(&ip6_blk_hole_entry_template,
6479 sizeof(*net->ipv6.ip6_blk_hole_entry),
6480 GFP_KERNEL);
6481 if (!net->ipv6.ip6_blk_hole_entry)
6482 goto out_ip6_prohibit_entry;
6483 net->ipv6.ip6_blk_hole_entry->dst.ops = &net->ipv6.ip6_dst_ops;
6484 dst_init_metrics(&net->ipv6.ip6_blk_hole_entry->dst,
6485 ip6_template_metrics, true);
6486 INIT_LIST_HEAD(&net->ipv6.ip6_blk_hole_entry->rt6i_uncached);
6487#ifdef CONFIG_IPV6_SUBTREES
6488 net->ipv6.fib6_routes_require_src = 0;
6489#endif
6490#endif
6491
6492 net->ipv6.sysctl.flush_delay = 0;
6493 net->ipv6.sysctl.ip6_rt_max_size = 4096;
6494 net->ipv6.sysctl.ip6_rt_gc_min_interval = HZ / 2;
6495 net->ipv6.sysctl.ip6_rt_gc_timeout = 60*HZ;
6496 net->ipv6.sysctl.ip6_rt_gc_interval = 30*HZ;
6497 net->ipv6.sysctl.ip6_rt_gc_elasticity = 9;
6498 net->ipv6.sysctl.ip6_rt_mtu_expires = 10*60*HZ;
6499 net->ipv6.sysctl.ip6_rt_min_advmss = IPV6_MIN_MTU - 20 - 40;
6500 net->ipv6.sysctl.skip_notify_on_dev_down = 0;
6501
6502 net->ipv6.ip6_rt_gc_expire = 30*HZ;
6503
6504 ret = 0;
6505out:
6506 return ret;
6507
6508#ifdef CONFIG_IPV6_MULTIPLE_TABLES
6509out_ip6_prohibit_entry:
6510 kfree(net->ipv6.ip6_prohibit_entry);
6511out_ip6_null_entry:
6512 kfree(net->ipv6.ip6_null_entry);
6513#endif
6514out_fib6_null_entry:
6515 kfree(net->ipv6.fib6_null_entry);
6516out_ip6_dst_entries:
6517 dst_entries_destroy(&net->ipv6.ip6_dst_ops);
6518out_ip6_dst_ops:
6519 goto out;
6520}
6521
6522static void __net_exit ip6_route_net_exit(struct net *net)
6523{
6524 kfree(net->ipv6.fib6_null_entry);
6525 kfree(net->ipv6.ip6_null_entry);
6526#ifdef CONFIG_IPV6_MULTIPLE_TABLES
6527 kfree(net->ipv6.ip6_prohibit_entry);
6528 kfree(net->ipv6.ip6_blk_hole_entry);
6529#endif
6530 dst_entries_destroy(&net->ipv6.ip6_dst_ops);
6531}
6532
6533static int __net_init ip6_route_net_init_late(struct net *net)
6534{
6535#ifdef CONFIG_PROC_FS
6536 proc_create_net("ipv6_route", 0, net->proc_net, &ipv6_route_seq_ops,
6537 sizeof(struct ipv6_route_iter));
6538 proc_create_net_single("rt6_stats", 0444, net->proc_net,
6539 rt6_stats_seq_show, NULL);
6540#endif
6541 return 0;
6542}
6543
6544static void __net_exit ip6_route_net_exit_late(struct net *net)
6545{
6546#ifdef CONFIG_PROC_FS
6547 remove_proc_entry("ipv6_route", net->proc_net);
6548 remove_proc_entry("rt6_stats", net->proc_net);
6549#endif
6550}
6551
6552static struct pernet_operations ip6_route_net_ops = {
6553 .init = ip6_route_net_init,
6554 .exit = ip6_route_net_exit,
6555};
6556
6557static int __net_init ipv6_inetpeer_init(struct net *net)
6558{
6559 struct inet_peer_base *bp = kmalloc(sizeof(*bp), GFP_KERNEL);
6560
6561 if (!bp)
6562 return -ENOMEM;
6563 inet_peer_base_init(bp);
6564 net->ipv6.peers = bp;
6565 return 0;
6566}
6567
6568static void __net_exit ipv6_inetpeer_exit(struct net *net)
6569{
6570 struct inet_peer_base *bp = net->ipv6.peers;
6571
6572 net->ipv6.peers = NULL;
6573 inetpeer_invalidate_tree(bp);
6574 kfree(bp);
6575}
6576
6577static struct pernet_operations ipv6_inetpeer_ops = {
6578 .init = ipv6_inetpeer_init,
6579 .exit = ipv6_inetpeer_exit,
6580};
6581
6582static struct pernet_operations ip6_route_net_late_ops = {
6583 .init = ip6_route_net_init_late,
6584 .exit = ip6_route_net_exit_late,
6585};
6586
6587static struct notifier_block ip6_route_dev_notifier = {
6588 .notifier_call = ip6_route_dev_notify,
6589 .priority = ADDRCONF_NOTIFY_PRIORITY - 10,
6590};
6591
6592void __init ip6_route_init_special_entries(void)
6593{
6594 /* Registering of the loopback is done before this portion of code,
6595 * the loopback reference in rt6_info will not be taken, do it
6596 * manually for init_net */
6597 init_net.ipv6.fib6_null_entry->fib6_nh->fib_nh_dev = init_net.loopback_dev;
6598 init_net.ipv6.ip6_null_entry->dst.dev = init_net.loopback_dev;
6599 init_net.ipv6.ip6_null_entry->rt6i_idev = in6_dev_get(init_net.loopback_dev);
6600 #ifdef CONFIG_IPV6_MULTIPLE_TABLES
6601 init_net.ipv6.ip6_prohibit_entry->dst.dev = init_net.loopback_dev;
6602 init_net.ipv6.ip6_prohibit_entry->rt6i_idev = in6_dev_get(init_net.loopback_dev);
6603 init_net.ipv6.ip6_blk_hole_entry->dst.dev = init_net.loopback_dev;
6604 init_net.ipv6.ip6_blk_hole_entry->rt6i_idev = in6_dev_get(init_net.loopback_dev);
6605 #endif
6606}
6607
6608#if IS_BUILTIN(CONFIG_IPV6)
6609#if defined(CONFIG_BPF_SYSCALL) && defined(CONFIG_PROC_FS)
6610DEFINE_BPF_ITER_FUNC(ipv6_route, struct bpf_iter_meta *meta, struct fib6_info *rt)
6611
6612BTF_ID_LIST(btf_fib6_info_id)
6613BTF_ID(struct, fib6_info)
6614
6615static const struct bpf_iter_seq_info ipv6_route_seq_info = {
6616 .seq_ops = &ipv6_route_seq_ops,
6617 .init_seq_private = bpf_iter_init_seq_net,
6618 .fini_seq_private = bpf_iter_fini_seq_net,
6619 .seq_priv_size = sizeof(struct ipv6_route_iter),
6620};
6621
6622static struct bpf_iter_reg ipv6_route_reg_info = {
6623 .target = "ipv6_route",
6624 .ctx_arg_info_size = 1,
6625 .ctx_arg_info = {
6626 { offsetof(struct bpf_iter__ipv6_route, rt),
6627 PTR_TO_BTF_ID_OR_NULL },
6628 },
6629 .seq_info = &ipv6_route_seq_info,
6630};
6631
6632static int __init bpf_iter_register(void)
6633{
6634 ipv6_route_reg_info.ctx_arg_info[0].btf_id = *btf_fib6_info_id;
6635 return bpf_iter_reg_target(&ipv6_route_reg_info);
6636}
6637
6638static void bpf_iter_unregister(void)
6639{
6640 bpf_iter_unreg_target(&ipv6_route_reg_info);
6641}
6642#endif
6643#endif
6644
6645int __init ip6_route_init(void)
6646{
6647 int ret;
6648 int cpu;
6649
6650 ret = -ENOMEM;
6651 ip6_dst_ops_template.kmem_cachep =
6652 kmem_cache_create("ip6_dst_cache", sizeof(struct rt6_info), 0,
6653 SLAB_HWCACHE_ALIGN, NULL);
6654 if (!ip6_dst_ops_template.kmem_cachep)
6655 goto out;
6656
6657 ret = dst_entries_init(&ip6_dst_blackhole_ops);
6658 if (ret)
6659 goto out_kmem_cache;
6660
6661 ret = register_pernet_subsys(&ipv6_inetpeer_ops);
6662 if (ret)
6663 goto out_dst_entries;
6664
6665 ret = register_pernet_subsys(&ip6_route_net_ops);
6666 if (ret)
6667 goto out_register_inetpeer;
6668
6669 ip6_dst_blackhole_ops.kmem_cachep = ip6_dst_ops_template.kmem_cachep;
6670
6671 ret = fib6_init();
6672 if (ret)
6673 goto out_register_subsys;
6674
6675 ret = xfrm6_init();
6676 if (ret)
6677 goto out_fib6_init;
6678
6679 ret = fib6_rules_init();
6680 if (ret)
6681 goto xfrm6_init;
6682
6683 ret = register_pernet_subsys(&ip6_route_net_late_ops);
6684 if (ret)
6685 goto fib6_rules_init;
6686
6687 ret = rtnl_register_module(THIS_MODULE, PF_INET6, RTM_NEWROUTE,
6688 inet6_rtm_newroute, NULL, 0);
6689 if (ret < 0)
6690 goto out_register_late_subsys;
6691
6692 ret = rtnl_register_module(THIS_MODULE, PF_INET6, RTM_DELROUTE,
6693 inet6_rtm_delroute, NULL, 0);
6694 if (ret < 0)
6695 goto out_register_late_subsys;
6696
6697 ret = rtnl_register_module(THIS_MODULE, PF_INET6, RTM_GETROUTE,
6698 inet6_rtm_getroute, NULL,
6699 RTNL_FLAG_DOIT_UNLOCKED);
6700 if (ret < 0)
6701 goto out_register_late_subsys;
6702
6703 ret = register_netdevice_notifier(&ip6_route_dev_notifier);
6704 if (ret)
6705 goto out_register_late_subsys;
6706
6707#if IS_BUILTIN(CONFIG_IPV6)
6708#if defined(CONFIG_BPF_SYSCALL) && defined(CONFIG_PROC_FS)
6709 ret = bpf_iter_register();
6710 if (ret)
6711 goto out_register_late_subsys;
6712#endif
6713#endif
6714
6715 for_each_possible_cpu(cpu) {
6716 struct uncached_list *ul = per_cpu_ptr(&rt6_uncached_list, cpu);
6717
6718 INIT_LIST_HEAD(&ul->head);
6719 spin_lock_init(&ul->lock);
6720 }
6721
6722out:
6723 return ret;
6724
6725out_register_late_subsys:
6726 rtnl_unregister_all(PF_INET6);
6727 unregister_pernet_subsys(&ip6_route_net_late_ops);
6728fib6_rules_init:
6729 fib6_rules_cleanup();
6730xfrm6_init:
6731 xfrm6_fini();
6732out_fib6_init:
6733 fib6_gc_cleanup();
6734out_register_subsys:
6735 unregister_pernet_subsys(&ip6_route_net_ops);
6736out_register_inetpeer:
6737 unregister_pernet_subsys(&ipv6_inetpeer_ops);
6738out_dst_entries:
6739 dst_entries_destroy(&ip6_dst_blackhole_ops);
6740out_kmem_cache:
6741 kmem_cache_destroy(ip6_dst_ops_template.kmem_cachep);
6742 goto out;
6743}
6744
6745void ip6_route_cleanup(void)
6746{
6747#if IS_BUILTIN(CONFIG_IPV6)
6748#if defined(CONFIG_BPF_SYSCALL) && defined(CONFIG_PROC_FS)
6749 bpf_iter_unregister();
6750#endif
6751#endif
6752 unregister_netdevice_notifier(&ip6_route_dev_notifier);
6753 unregister_pernet_subsys(&ip6_route_net_late_ops);
6754 fib6_rules_cleanup();
6755 xfrm6_fini();
6756 fib6_gc_cleanup();
6757 unregister_pernet_subsys(&ipv6_inetpeer_ops);
6758 unregister_pernet_subsys(&ip6_route_net_ops);
6759 dst_entries_destroy(&ip6_dst_blackhole_ops);
6760 kmem_cache_destroy(ip6_dst_ops_template.kmem_cachep);
6761}
1/*
2 * Linux INET6 implementation
3 * FIB front-end.
4 *
5 * Authors:
6 * Pedro Roque <roque@di.fc.ul.pt>
7 *
8 * This program is free software; you can redistribute it and/or
9 * modify it under the terms of the GNU General Public License
10 * as published by the Free Software Foundation; either version
11 * 2 of the License, or (at your option) any later version.
12 */
13
14/* Changes:
15 *
16 * YOSHIFUJI Hideaki @USAGI
17 * reworked default router selection.
18 * - respect outgoing interface
19 * - select from (probably) reachable routers (i.e.
20 * routers in REACHABLE, STALE, DELAY or PROBE states).
21 * - always select the same router if it is (probably)
22 * reachable. otherwise, round-robin the list.
23 * Ville Nuorvala
24 * Fixed routing subtrees.
25 */
26
27#define pr_fmt(fmt) "IPv6: " fmt
28
29#include <linux/capability.h>
30#include <linux/errno.h>
31#include <linux/export.h>
32#include <linux/types.h>
33#include <linux/times.h>
34#include <linux/socket.h>
35#include <linux/sockios.h>
36#include <linux/net.h>
37#include <linux/route.h>
38#include <linux/netdevice.h>
39#include <linux/in6.h>
40#include <linux/mroute6.h>
41#include <linux/init.h>
42#include <linux/if_arp.h>
43#include <linux/proc_fs.h>
44#include <linux/seq_file.h>
45#include <linux/nsproxy.h>
46#include <linux/slab.h>
47#include <net/net_namespace.h>
48#include <net/snmp.h>
49#include <net/ipv6.h>
50#include <net/ip6_fib.h>
51#include <net/ip6_route.h>
52#include <net/ndisc.h>
53#include <net/addrconf.h>
54#include <net/tcp.h>
55#include <linux/rtnetlink.h>
56#include <net/dst.h>
57#include <net/xfrm.h>
58#include <net/netevent.h>
59#include <net/netlink.h>
60
61#include <asm/uaccess.h>
62
63#ifdef CONFIG_SYSCTL
64#include <linux/sysctl.h>
65#endif
66
67static struct rt6_info *ip6_rt_copy(struct rt6_info *ort,
68 const struct in6_addr *dest);
69static struct dst_entry *ip6_dst_check(struct dst_entry *dst, u32 cookie);
70static unsigned int ip6_default_advmss(const struct dst_entry *dst);
71static unsigned int ip6_mtu(const struct dst_entry *dst);
72static struct dst_entry *ip6_negative_advice(struct dst_entry *);
73static void ip6_dst_destroy(struct dst_entry *);
74static void ip6_dst_ifdown(struct dst_entry *,
75 struct net_device *dev, int how);
76static int ip6_dst_gc(struct dst_ops *ops);
77
78static int ip6_pkt_discard(struct sk_buff *skb);
79static int ip6_pkt_discard_out(struct sk_buff *skb);
80static void ip6_link_failure(struct sk_buff *skb);
81static void ip6_rt_update_pmtu(struct dst_entry *dst, u32 mtu);
82
83#ifdef CONFIG_IPV6_ROUTE_INFO
84static struct rt6_info *rt6_add_route_info(struct net *net,
85 const struct in6_addr *prefix, int prefixlen,
86 const struct in6_addr *gwaddr, int ifindex,
87 unsigned int pref);
88static struct rt6_info *rt6_get_route_info(struct net *net,
89 const struct in6_addr *prefix, int prefixlen,
90 const struct in6_addr *gwaddr, int ifindex);
91#endif
92
93static u32 *ipv6_cow_metrics(struct dst_entry *dst, unsigned long old)
94{
95 struct rt6_info *rt = (struct rt6_info *) dst;
96 struct inet_peer *peer;
97 u32 *p = NULL;
98
99 if (!(rt->dst.flags & DST_HOST))
100 return NULL;
101
102 if (!rt->rt6i_peer)
103 rt6_bind_peer(rt, 1);
104
105 peer = rt->rt6i_peer;
106 if (peer) {
107 u32 *old_p = __DST_METRICS_PTR(old);
108 unsigned long prev, new;
109
110 p = peer->metrics;
111 if (inet_metrics_new(peer))
112 memcpy(p, old_p, sizeof(u32) * RTAX_MAX);
113
114 new = (unsigned long) p;
115 prev = cmpxchg(&dst->_metrics, old, new);
116
117 if (prev != old) {
118 p = __DST_METRICS_PTR(prev);
119 if (prev & DST_METRICS_READ_ONLY)
120 p = NULL;
121 }
122 }
123 return p;
124}
125
126static inline const void *choose_neigh_daddr(struct rt6_info *rt, const void *daddr)
127{
128 struct in6_addr *p = &rt->rt6i_gateway;
129
130 if (!ipv6_addr_any(p))
131 return (const void *) p;
132 return daddr;
133}
134
135static struct neighbour *ip6_neigh_lookup(const struct dst_entry *dst, const void *daddr)
136{
137 struct rt6_info *rt = (struct rt6_info *) dst;
138 struct neighbour *n;
139
140 daddr = choose_neigh_daddr(rt, daddr);
141 n = __ipv6_neigh_lookup(&nd_tbl, dst->dev, daddr);
142 if (n)
143 return n;
144 return neigh_create(&nd_tbl, daddr, dst->dev);
145}
146
147static int rt6_bind_neighbour(struct rt6_info *rt, struct net_device *dev)
148{
149 struct neighbour *n = __ipv6_neigh_lookup(&nd_tbl, dev, &rt->rt6i_gateway);
150 if (!n) {
151 n = neigh_create(&nd_tbl, &rt->rt6i_gateway, dev);
152 if (IS_ERR(n))
153 return PTR_ERR(n);
154 }
155 dst_set_neighbour(&rt->dst, n);
156
157 return 0;
158}
159
160static struct dst_ops ip6_dst_ops_template = {
161 .family = AF_INET6,
162 .protocol = cpu_to_be16(ETH_P_IPV6),
163 .gc = ip6_dst_gc,
164 .gc_thresh = 1024,
165 .check = ip6_dst_check,
166 .default_advmss = ip6_default_advmss,
167 .mtu = ip6_mtu,
168 .cow_metrics = ipv6_cow_metrics,
169 .destroy = ip6_dst_destroy,
170 .ifdown = ip6_dst_ifdown,
171 .negative_advice = ip6_negative_advice,
172 .link_failure = ip6_link_failure,
173 .update_pmtu = ip6_rt_update_pmtu,
174 .local_out = __ip6_local_out,
175 .neigh_lookup = ip6_neigh_lookup,
176};
177
178static unsigned int ip6_blackhole_mtu(const struct dst_entry *dst)
179{
180 unsigned int mtu = dst_metric_raw(dst, RTAX_MTU);
181
182 return mtu ? : dst->dev->mtu;
183}
184
185static void ip6_rt_blackhole_update_pmtu(struct dst_entry *dst, u32 mtu)
186{
187}
188
189static u32 *ip6_rt_blackhole_cow_metrics(struct dst_entry *dst,
190 unsigned long old)
191{
192 return NULL;
193}
194
195static struct dst_ops ip6_dst_blackhole_ops = {
196 .family = AF_INET6,
197 .protocol = cpu_to_be16(ETH_P_IPV6),
198 .destroy = ip6_dst_destroy,
199 .check = ip6_dst_check,
200 .mtu = ip6_blackhole_mtu,
201 .default_advmss = ip6_default_advmss,
202 .update_pmtu = ip6_rt_blackhole_update_pmtu,
203 .cow_metrics = ip6_rt_blackhole_cow_metrics,
204 .neigh_lookup = ip6_neigh_lookup,
205};
206
207static const u32 ip6_template_metrics[RTAX_MAX] = {
208 [RTAX_HOPLIMIT - 1] = 255,
209};
210
211static struct rt6_info ip6_null_entry_template = {
212 .dst = {
213 .__refcnt = ATOMIC_INIT(1),
214 .__use = 1,
215 .obsolete = -1,
216 .error = -ENETUNREACH,
217 .input = ip6_pkt_discard,
218 .output = ip6_pkt_discard_out,
219 },
220 .rt6i_flags = (RTF_REJECT | RTF_NONEXTHOP),
221 .rt6i_protocol = RTPROT_KERNEL,
222 .rt6i_metric = ~(u32) 0,
223 .rt6i_ref = ATOMIC_INIT(1),
224};
225
226#ifdef CONFIG_IPV6_MULTIPLE_TABLES
227
228static int ip6_pkt_prohibit(struct sk_buff *skb);
229static int ip6_pkt_prohibit_out(struct sk_buff *skb);
230
231static struct rt6_info ip6_prohibit_entry_template = {
232 .dst = {
233 .__refcnt = ATOMIC_INIT(1),
234 .__use = 1,
235 .obsolete = -1,
236 .error = -EACCES,
237 .input = ip6_pkt_prohibit,
238 .output = ip6_pkt_prohibit_out,
239 },
240 .rt6i_flags = (RTF_REJECT | RTF_NONEXTHOP),
241 .rt6i_protocol = RTPROT_KERNEL,
242 .rt6i_metric = ~(u32) 0,
243 .rt6i_ref = ATOMIC_INIT(1),
244};
245
246static struct rt6_info ip6_blk_hole_entry_template = {
247 .dst = {
248 .__refcnt = ATOMIC_INIT(1),
249 .__use = 1,
250 .obsolete = -1,
251 .error = -EINVAL,
252 .input = dst_discard,
253 .output = dst_discard,
254 },
255 .rt6i_flags = (RTF_REJECT | RTF_NONEXTHOP),
256 .rt6i_protocol = RTPROT_KERNEL,
257 .rt6i_metric = ~(u32) 0,
258 .rt6i_ref = ATOMIC_INIT(1),
259};
260
261#endif
262
263/* allocate dst with ip6_dst_ops */
264static inline struct rt6_info *ip6_dst_alloc(struct dst_ops *ops,
265 struct net_device *dev,
266 int flags)
267{
268 struct rt6_info *rt = dst_alloc(ops, dev, 0, 0, flags);
269
270 if (rt)
271 memset(&rt->rt6i_table, 0,
272 sizeof(*rt) - sizeof(struct dst_entry));
273
274 return rt;
275}
276
277static void ip6_dst_destroy(struct dst_entry *dst)
278{
279 struct rt6_info *rt = (struct rt6_info *)dst;
280 struct inet6_dev *idev = rt->rt6i_idev;
281 struct inet_peer *peer = rt->rt6i_peer;
282
283 if (!(rt->dst.flags & DST_HOST))
284 dst_destroy_metrics_generic(dst);
285
286 if (idev) {
287 rt->rt6i_idev = NULL;
288 in6_dev_put(idev);
289 }
290
291 if (!(rt->rt6i_flags & RTF_EXPIRES) && dst->from)
292 dst_release(dst->from);
293
294 if (peer) {
295 rt->rt6i_peer = NULL;
296 inet_putpeer(peer);
297 }
298}
299
300static atomic_t __rt6_peer_genid = ATOMIC_INIT(0);
301
302static u32 rt6_peer_genid(void)
303{
304 return atomic_read(&__rt6_peer_genid);
305}
306
307void rt6_bind_peer(struct rt6_info *rt, int create)
308{
309 struct inet_peer *peer;
310
311 peer = inet_getpeer_v6(&rt->rt6i_dst.addr, create);
312 if (peer && cmpxchg(&rt->rt6i_peer, NULL, peer) != NULL)
313 inet_putpeer(peer);
314 else
315 rt->rt6i_peer_genid = rt6_peer_genid();
316}
317
318static void ip6_dst_ifdown(struct dst_entry *dst, struct net_device *dev,
319 int how)
320{
321 struct rt6_info *rt = (struct rt6_info *)dst;
322 struct inet6_dev *idev = rt->rt6i_idev;
323 struct net_device *loopback_dev =
324 dev_net(dev)->loopback_dev;
325
326 if (dev != loopback_dev && idev && idev->dev == dev) {
327 struct inet6_dev *loopback_idev =
328 in6_dev_get(loopback_dev);
329 if (loopback_idev) {
330 rt->rt6i_idev = loopback_idev;
331 in6_dev_put(idev);
332 }
333 }
334}
335
336static bool rt6_check_expired(const struct rt6_info *rt)
337{
338 struct rt6_info *ort = NULL;
339
340 if (rt->rt6i_flags & RTF_EXPIRES) {
341 if (time_after(jiffies, rt->dst.expires))
342 return true;
343 } else if (rt->dst.from) {
344 ort = (struct rt6_info *) rt->dst.from;
345 return (ort->rt6i_flags & RTF_EXPIRES) &&
346 time_after(jiffies, ort->dst.expires);
347 }
348 return false;
349}
350
351static bool rt6_need_strict(const struct in6_addr *daddr)
352{
353 return ipv6_addr_type(daddr) &
354 (IPV6_ADDR_MULTICAST | IPV6_ADDR_LINKLOCAL | IPV6_ADDR_LOOPBACK);
355}
356
357/*
358 * Route lookup. Any table->tb6_lock is implied.
359 */
360
361static inline struct rt6_info *rt6_device_match(struct net *net,
362 struct rt6_info *rt,
363 const struct in6_addr *saddr,
364 int oif,
365 int flags)
366{
367 struct rt6_info *local = NULL;
368 struct rt6_info *sprt;
369
370 if (!oif && ipv6_addr_any(saddr))
371 goto out;
372
373 for (sprt = rt; sprt; sprt = sprt->dst.rt6_next) {
374 struct net_device *dev = sprt->dst.dev;
375
376 if (oif) {
377 if (dev->ifindex == oif)
378 return sprt;
379 if (dev->flags & IFF_LOOPBACK) {
380 if (!sprt->rt6i_idev ||
381 sprt->rt6i_idev->dev->ifindex != oif) {
382 if (flags & RT6_LOOKUP_F_IFACE && oif)
383 continue;
384 if (local && (!oif ||
385 local->rt6i_idev->dev->ifindex == oif))
386 continue;
387 }
388 local = sprt;
389 }
390 } else {
391 if (ipv6_chk_addr(net, saddr, dev,
392 flags & RT6_LOOKUP_F_IFACE))
393 return sprt;
394 }
395 }
396
397 if (oif) {
398 if (local)
399 return local;
400
401 if (flags & RT6_LOOKUP_F_IFACE)
402 return net->ipv6.ip6_null_entry;
403 }
404out:
405 return rt;
406}
407
408#ifdef CONFIG_IPV6_ROUTER_PREF
409static void rt6_probe(struct rt6_info *rt)
410{
411 struct neighbour *neigh;
412 /*
413 * Okay, this does not seem to be appropriate
414 * for now, however, we need to check if it
415 * is really so; aka Router Reachability Probing.
416 *
417 * Router Reachability Probe MUST be rate-limited
418 * to no more than one per minute.
419 */
420 rcu_read_lock();
421 neigh = rt ? dst_get_neighbour_noref(&rt->dst) : NULL;
422 if (!neigh || (neigh->nud_state & NUD_VALID))
423 goto out;
424 read_lock_bh(&neigh->lock);
425 if (!(neigh->nud_state & NUD_VALID) &&
426 time_after(jiffies, neigh->updated + rt->rt6i_idev->cnf.rtr_probe_interval)) {
427 struct in6_addr mcaddr;
428 struct in6_addr *target;
429
430 neigh->updated = jiffies;
431 read_unlock_bh(&neigh->lock);
432
433 target = (struct in6_addr *)&neigh->primary_key;
434 addrconf_addr_solict_mult(target, &mcaddr);
435 ndisc_send_ns(rt->dst.dev, NULL, target, &mcaddr, NULL);
436 } else {
437 read_unlock_bh(&neigh->lock);
438 }
439out:
440 rcu_read_unlock();
441}
442#else
443static inline void rt6_probe(struct rt6_info *rt)
444{
445}
446#endif
447
448/*
449 * Default Router Selection (RFC 2461 6.3.6)
450 */
451static inline int rt6_check_dev(struct rt6_info *rt, int oif)
452{
453 struct net_device *dev = rt->dst.dev;
454 if (!oif || dev->ifindex == oif)
455 return 2;
456 if ((dev->flags & IFF_LOOPBACK) &&
457 rt->rt6i_idev && rt->rt6i_idev->dev->ifindex == oif)
458 return 1;
459 return 0;
460}
461
462static inline int rt6_check_neigh(struct rt6_info *rt)
463{
464 struct neighbour *neigh;
465 int m;
466
467 rcu_read_lock();
468 neigh = dst_get_neighbour_noref(&rt->dst);
469 if (rt->rt6i_flags & RTF_NONEXTHOP ||
470 !(rt->rt6i_flags & RTF_GATEWAY))
471 m = 1;
472 else if (neigh) {
473 read_lock_bh(&neigh->lock);
474 if (neigh->nud_state & NUD_VALID)
475 m = 2;
476#ifdef CONFIG_IPV6_ROUTER_PREF
477 else if (neigh->nud_state & NUD_FAILED)
478 m = 0;
479#endif
480 else
481 m = 1;
482 read_unlock_bh(&neigh->lock);
483 } else
484 m = 0;
485 rcu_read_unlock();
486 return m;
487}
488
489static int rt6_score_route(struct rt6_info *rt, int oif,
490 int strict)
491{
492 int m, n;
493
494 m = rt6_check_dev(rt, oif);
495 if (!m && (strict & RT6_LOOKUP_F_IFACE))
496 return -1;
497#ifdef CONFIG_IPV6_ROUTER_PREF
498 m |= IPV6_DECODE_PREF(IPV6_EXTRACT_PREF(rt->rt6i_flags)) << 2;
499#endif
500 n = rt6_check_neigh(rt);
501 if (!n && (strict & RT6_LOOKUP_F_REACHABLE))
502 return -1;
503 return m;
504}
505
506static struct rt6_info *find_match(struct rt6_info *rt, int oif, int strict,
507 int *mpri, struct rt6_info *match)
508{
509 int m;
510
511 if (rt6_check_expired(rt))
512 goto out;
513
514 m = rt6_score_route(rt, oif, strict);
515 if (m < 0)
516 goto out;
517
518 if (m > *mpri) {
519 if (strict & RT6_LOOKUP_F_REACHABLE)
520 rt6_probe(match);
521 *mpri = m;
522 match = rt;
523 } else if (strict & RT6_LOOKUP_F_REACHABLE) {
524 rt6_probe(rt);
525 }
526
527out:
528 return match;
529}
530
531static struct rt6_info *find_rr_leaf(struct fib6_node *fn,
532 struct rt6_info *rr_head,
533 u32 metric, int oif, int strict)
534{
535 struct rt6_info *rt, *match;
536 int mpri = -1;
537
538 match = NULL;
539 for (rt = rr_head; rt && rt->rt6i_metric == metric;
540 rt = rt->dst.rt6_next)
541 match = find_match(rt, oif, strict, &mpri, match);
542 for (rt = fn->leaf; rt && rt != rr_head && rt->rt6i_metric == metric;
543 rt = rt->dst.rt6_next)
544 match = find_match(rt, oif, strict, &mpri, match);
545
546 return match;
547}
548
549static struct rt6_info *rt6_select(struct fib6_node *fn, int oif, int strict)
550{
551 struct rt6_info *match, *rt0;
552 struct net *net;
553
554 rt0 = fn->rr_ptr;
555 if (!rt0)
556 fn->rr_ptr = rt0 = fn->leaf;
557
558 match = find_rr_leaf(fn, rt0, rt0->rt6i_metric, oif, strict);
559
560 if (!match &&
561 (strict & RT6_LOOKUP_F_REACHABLE)) {
562 struct rt6_info *next = rt0->dst.rt6_next;
563
564 /* no entries matched; do round-robin */
565 if (!next || next->rt6i_metric != rt0->rt6i_metric)
566 next = fn->leaf;
567
568 if (next != rt0)
569 fn->rr_ptr = next;
570 }
571
572 net = dev_net(rt0->dst.dev);
573 return match ? match : net->ipv6.ip6_null_entry;
574}
575
576#ifdef CONFIG_IPV6_ROUTE_INFO
577int rt6_route_rcv(struct net_device *dev, u8 *opt, int len,
578 const struct in6_addr *gwaddr)
579{
580 struct net *net = dev_net(dev);
581 struct route_info *rinfo = (struct route_info *) opt;
582 struct in6_addr prefix_buf, *prefix;
583 unsigned int pref;
584 unsigned long lifetime;
585 struct rt6_info *rt;
586
587 if (len < sizeof(struct route_info)) {
588 return -EINVAL;
589 }
590
591 /* Sanity check for prefix_len and length */
592 if (rinfo->length > 3) {
593 return -EINVAL;
594 } else if (rinfo->prefix_len > 128) {
595 return -EINVAL;
596 } else if (rinfo->prefix_len > 64) {
597 if (rinfo->length < 2) {
598 return -EINVAL;
599 }
600 } else if (rinfo->prefix_len > 0) {
601 if (rinfo->length < 1) {
602 return -EINVAL;
603 }
604 }
605
606 pref = rinfo->route_pref;
607 if (pref == ICMPV6_ROUTER_PREF_INVALID)
608 return -EINVAL;
609
610 lifetime = addrconf_timeout_fixup(ntohl(rinfo->lifetime), HZ);
611
612 if (rinfo->length == 3)
613 prefix = (struct in6_addr *)rinfo->prefix;
614 else {
615 /* this function is safe */
616 ipv6_addr_prefix(&prefix_buf,
617 (struct in6_addr *)rinfo->prefix,
618 rinfo->prefix_len);
619 prefix = &prefix_buf;
620 }
621
622 rt = rt6_get_route_info(net, prefix, rinfo->prefix_len, gwaddr,
623 dev->ifindex);
624
625 if (rt && !lifetime) {
626 ip6_del_rt(rt);
627 rt = NULL;
628 }
629
630 if (!rt && lifetime)
631 rt = rt6_add_route_info(net, prefix, rinfo->prefix_len, gwaddr, dev->ifindex,
632 pref);
633 else if (rt)
634 rt->rt6i_flags = RTF_ROUTEINFO |
635 (rt->rt6i_flags & ~RTF_PREF_MASK) | RTF_PREF(pref);
636
637 if (rt) {
638 if (!addrconf_finite_timeout(lifetime))
639 rt6_clean_expires(rt);
640 else
641 rt6_set_expires(rt, jiffies + HZ * lifetime);
642
643 dst_release(&rt->dst);
644 }
645 return 0;
646}
647#endif
648
649#define BACKTRACK(__net, saddr) \
650do { \
651 if (rt == __net->ipv6.ip6_null_entry) { \
652 struct fib6_node *pn; \
653 while (1) { \
654 if (fn->fn_flags & RTN_TL_ROOT) \
655 goto out; \
656 pn = fn->parent; \
657 if (FIB6_SUBTREE(pn) && FIB6_SUBTREE(pn) != fn) \
658 fn = fib6_lookup(FIB6_SUBTREE(pn), NULL, saddr); \
659 else \
660 fn = pn; \
661 if (fn->fn_flags & RTN_RTINFO) \
662 goto restart; \
663 } \
664 } \
665} while (0)
666
667static struct rt6_info *ip6_pol_route_lookup(struct net *net,
668 struct fib6_table *table,
669 struct flowi6 *fl6, int flags)
670{
671 struct fib6_node *fn;
672 struct rt6_info *rt;
673
674 read_lock_bh(&table->tb6_lock);
675 fn = fib6_lookup(&table->tb6_root, &fl6->daddr, &fl6->saddr);
676restart:
677 rt = fn->leaf;
678 rt = rt6_device_match(net, rt, &fl6->saddr, fl6->flowi6_oif, flags);
679 BACKTRACK(net, &fl6->saddr);
680out:
681 dst_use(&rt->dst, jiffies);
682 read_unlock_bh(&table->tb6_lock);
683 return rt;
684
685}
686
687struct dst_entry * ip6_route_lookup(struct net *net, struct flowi6 *fl6,
688 int flags)
689{
690 return fib6_rule_lookup(net, fl6, flags, ip6_pol_route_lookup);
691}
692EXPORT_SYMBOL_GPL(ip6_route_lookup);
693
694struct rt6_info *rt6_lookup(struct net *net, const struct in6_addr *daddr,
695 const struct in6_addr *saddr, int oif, int strict)
696{
697 struct flowi6 fl6 = {
698 .flowi6_oif = oif,
699 .daddr = *daddr,
700 };
701 struct dst_entry *dst;
702 int flags = strict ? RT6_LOOKUP_F_IFACE : 0;
703
704 if (saddr) {
705 memcpy(&fl6.saddr, saddr, sizeof(*saddr));
706 flags |= RT6_LOOKUP_F_HAS_SADDR;
707 }
708
709 dst = fib6_rule_lookup(net, &fl6, flags, ip6_pol_route_lookup);
710 if (dst->error == 0)
711 return (struct rt6_info *) dst;
712
713 dst_release(dst);
714
715 return NULL;
716}
717
718EXPORT_SYMBOL(rt6_lookup);
719
720/* ip6_ins_rt is called with FREE table->tb6_lock.
721 It takes new route entry, the addition fails by any reason the
722 route is freed. In any case, if caller does not hold it, it may
723 be destroyed.
724 */
725
726static int __ip6_ins_rt(struct rt6_info *rt, struct nl_info *info)
727{
728 int err;
729 struct fib6_table *table;
730
731 table = rt->rt6i_table;
732 write_lock_bh(&table->tb6_lock);
733 err = fib6_add(&table->tb6_root, rt, info);
734 write_unlock_bh(&table->tb6_lock);
735
736 return err;
737}
738
739int ip6_ins_rt(struct rt6_info *rt)
740{
741 struct nl_info info = {
742 .nl_net = dev_net(rt->dst.dev),
743 };
744 return __ip6_ins_rt(rt, &info);
745}
746
747static struct rt6_info *rt6_alloc_cow(struct rt6_info *ort,
748 const struct in6_addr *daddr,
749 const struct in6_addr *saddr)
750{
751 struct rt6_info *rt;
752
753 /*
754 * Clone the route.
755 */
756
757 rt = ip6_rt_copy(ort, daddr);
758
759 if (rt) {
760 int attempts = !in_softirq();
761
762 if (!(rt->rt6i_flags & RTF_GATEWAY)) {
763 if (ort->rt6i_dst.plen != 128 &&
764 ipv6_addr_equal(&ort->rt6i_dst.addr, daddr))
765 rt->rt6i_flags |= RTF_ANYCAST;
766 rt->rt6i_gateway = *daddr;
767 }
768
769 rt->rt6i_flags |= RTF_CACHE;
770
771#ifdef CONFIG_IPV6_SUBTREES
772 if (rt->rt6i_src.plen && saddr) {
773 rt->rt6i_src.addr = *saddr;
774 rt->rt6i_src.plen = 128;
775 }
776#endif
777
778 retry:
779 if (rt6_bind_neighbour(rt, rt->dst.dev)) {
780 struct net *net = dev_net(rt->dst.dev);
781 int saved_rt_min_interval =
782 net->ipv6.sysctl.ip6_rt_gc_min_interval;
783 int saved_rt_elasticity =
784 net->ipv6.sysctl.ip6_rt_gc_elasticity;
785
786 if (attempts-- > 0) {
787 net->ipv6.sysctl.ip6_rt_gc_elasticity = 1;
788 net->ipv6.sysctl.ip6_rt_gc_min_interval = 0;
789
790 ip6_dst_gc(&net->ipv6.ip6_dst_ops);
791
792 net->ipv6.sysctl.ip6_rt_gc_elasticity =
793 saved_rt_elasticity;
794 net->ipv6.sysctl.ip6_rt_gc_min_interval =
795 saved_rt_min_interval;
796 goto retry;
797 }
798
799 net_warn_ratelimited("Neighbour table overflow\n");
800 dst_free(&rt->dst);
801 return NULL;
802 }
803 }
804
805 return rt;
806}
807
808static struct rt6_info *rt6_alloc_clone(struct rt6_info *ort,
809 const struct in6_addr *daddr)
810{
811 struct rt6_info *rt = ip6_rt_copy(ort, daddr);
812
813 if (rt) {
814 rt->rt6i_flags |= RTF_CACHE;
815 dst_set_neighbour(&rt->dst, neigh_clone(dst_get_neighbour_noref_raw(&ort->dst)));
816 }
817 return rt;
818}
819
820static struct rt6_info *ip6_pol_route(struct net *net, struct fib6_table *table, int oif,
821 struct flowi6 *fl6, int flags)
822{
823 struct fib6_node *fn;
824 struct rt6_info *rt, *nrt;
825 int strict = 0;
826 int attempts = 3;
827 int err;
828 int reachable = net->ipv6.devconf_all->forwarding ? 0 : RT6_LOOKUP_F_REACHABLE;
829
830 strict |= flags & RT6_LOOKUP_F_IFACE;
831
832relookup:
833 read_lock_bh(&table->tb6_lock);
834
835restart_2:
836 fn = fib6_lookup(&table->tb6_root, &fl6->daddr, &fl6->saddr);
837
838restart:
839 rt = rt6_select(fn, oif, strict | reachable);
840
841 BACKTRACK(net, &fl6->saddr);
842 if (rt == net->ipv6.ip6_null_entry ||
843 rt->rt6i_flags & RTF_CACHE)
844 goto out;
845
846 dst_hold(&rt->dst);
847 read_unlock_bh(&table->tb6_lock);
848
849 if (!dst_get_neighbour_noref_raw(&rt->dst) && !(rt->rt6i_flags & RTF_NONEXTHOP))
850 nrt = rt6_alloc_cow(rt, &fl6->daddr, &fl6->saddr);
851 else if (!(rt->dst.flags & DST_HOST))
852 nrt = rt6_alloc_clone(rt, &fl6->daddr);
853 else
854 goto out2;
855
856 dst_release(&rt->dst);
857 rt = nrt ? : net->ipv6.ip6_null_entry;
858
859 dst_hold(&rt->dst);
860 if (nrt) {
861 err = ip6_ins_rt(nrt);
862 if (!err)
863 goto out2;
864 }
865
866 if (--attempts <= 0)
867 goto out2;
868
869 /*
870 * Race condition! In the gap, when table->tb6_lock was
871 * released someone could insert this route. Relookup.
872 */
873 dst_release(&rt->dst);
874 goto relookup;
875
876out:
877 if (reachable) {
878 reachable = 0;
879 goto restart_2;
880 }
881 dst_hold(&rt->dst);
882 read_unlock_bh(&table->tb6_lock);
883out2:
884 rt->dst.lastuse = jiffies;
885 rt->dst.__use++;
886
887 return rt;
888}
889
890static struct rt6_info *ip6_pol_route_input(struct net *net, struct fib6_table *table,
891 struct flowi6 *fl6, int flags)
892{
893 return ip6_pol_route(net, table, fl6->flowi6_iif, fl6, flags);
894}
895
896static struct dst_entry *ip6_route_input_lookup(struct net *net,
897 struct net_device *dev,
898 struct flowi6 *fl6, int flags)
899{
900 if (rt6_need_strict(&fl6->daddr) && dev->type != ARPHRD_PIMREG)
901 flags |= RT6_LOOKUP_F_IFACE;
902
903 return fib6_rule_lookup(net, fl6, flags, ip6_pol_route_input);
904}
905
906void ip6_route_input(struct sk_buff *skb)
907{
908 const struct ipv6hdr *iph = ipv6_hdr(skb);
909 struct net *net = dev_net(skb->dev);
910 int flags = RT6_LOOKUP_F_HAS_SADDR;
911 struct flowi6 fl6 = {
912 .flowi6_iif = skb->dev->ifindex,
913 .daddr = iph->daddr,
914 .saddr = iph->saddr,
915 .flowlabel = (* (__be32 *) iph) & IPV6_FLOWINFO_MASK,
916 .flowi6_mark = skb->mark,
917 .flowi6_proto = iph->nexthdr,
918 };
919
920 skb_dst_set(skb, ip6_route_input_lookup(net, skb->dev, &fl6, flags));
921}
922
923static struct rt6_info *ip6_pol_route_output(struct net *net, struct fib6_table *table,
924 struct flowi6 *fl6, int flags)
925{
926 return ip6_pol_route(net, table, fl6->flowi6_oif, fl6, flags);
927}
928
929struct dst_entry * ip6_route_output(struct net *net, const struct sock *sk,
930 struct flowi6 *fl6)
931{
932 int flags = 0;
933
934 if ((sk && sk->sk_bound_dev_if) || rt6_need_strict(&fl6->daddr))
935 flags |= RT6_LOOKUP_F_IFACE;
936
937 if (!ipv6_addr_any(&fl6->saddr))
938 flags |= RT6_LOOKUP_F_HAS_SADDR;
939 else if (sk)
940 flags |= rt6_srcprefs2flags(inet6_sk(sk)->srcprefs);
941
942 return fib6_rule_lookup(net, fl6, flags, ip6_pol_route_output);
943}
944
945EXPORT_SYMBOL(ip6_route_output);
946
947struct dst_entry *ip6_blackhole_route(struct net *net, struct dst_entry *dst_orig)
948{
949 struct rt6_info *rt, *ort = (struct rt6_info *) dst_orig;
950 struct dst_entry *new = NULL;
951
952 rt = dst_alloc(&ip6_dst_blackhole_ops, ort->dst.dev, 1, 0, 0);
953 if (rt) {
954 memset(&rt->rt6i_table, 0, sizeof(*rt) - sizeof(struct dst_entry));
955
956 new = &rt->dst;
957
958 new->__use = 1;
959 new->input = dst_discard;
960 new->output = dst_discard;
961
962 if (dst_metrics_read_only(&ort->dst))
963 new->_metrics = ort->dst._metrics;
964 else
965 dst_copy_metrics(new, &ort->dst);
966 rt->rt6i_idev = ort->rt6i_idev;
967 if (rt->rt6i_idev)
968 in6_dev_hold(rt->rt6i_idev);
969
970 rt->rt6i_gateway = ort->rt6i_gateway;
971 rt->rt6i_flags = ort->rt6i_flags;
972 rt6_clean_expires(rt);
973 rt->rt6i_metric = 0;
974
975 memcpy(&rt->rt6i_dst, &ort->rt6i_dst, sizeof(struct rt6key));
976#ifdef CONFIG_IPV6_SUBTREES
977 memcpy(&rt->rt6i_src, &ort->rt6i_src, sizeof(struct rt6key));
978#endif
979
980 dst_free(new);
981 }
982
983 dst_release(dst_orig);
984 return new ? new : ERR_PTR(-ENOMEM);
985}
986
987/*
988 * Destination cache support functions
989 */
990
991static struct dst_entry *ip6_dst_check(struct dst_entry *dst, u32 cookie)
992{
993 struct rt6_info *rt;
994
995 rt = (struct rt6_info *) dst;
996
997 if (rt->rt6i_node && (rt->rt6i_node->fn_sernum == cookie)) {
998 if (rt->rt6i_peer_genid != rt6_peer_genid()) {
999 if (!rt->rt6i_peer)
1000 rt6_bind_peer(rt, 0);
1001 rt->rt6i_peer_genid = rt6_peer_genid();
1002 }
1003 return dst;
1004 }
1005 return NULL;
1006}
1007
1008static struct dst_entry *ip6_negative_advice(struct dst_entry *dst)
1009{
1010 struct rt6_info *rt = (struct rt6_info *) dst;
1011
1012 if (rt) {
1013 if (rt->rt6i_flags & RTF_CACHE) {
1014 if (rt6_check_expired(rt)) {
1015 ip6_del_rt(rt);
1016 dst = NULL;
1017 }
1018 } else {
1019 dst_release(dst);
1020 dst = NULL;
1021 }
1022 }
1023 return dst;
1024}
1025
1026static void ip6_link_failure(struct sk_buff *skb)
1027{
1028 struct rt6_info *rt;
1029
1030 icmpv6_send(skb, ICMPV6_DEST_UNREACH, ICMPV6_ADDR_UNREACH, 0);
1031
1032 rt = (struct rt6_info *) skb_dst(skb);
1033 if (rt) {
1034 if (rt->rt6i_flags & RTF_CACHE)
1035 rt6_update_expires(rt, 0);
1036 else if (rt->rt6i_node && (rt->rt6i_flags & RTF_DEFAULT))
1037 rt->rt6i_node->fn_sernum = -1;
1038 }
1039}
1040
1041static void ip6_rt_update_pmtu(struct dst_entry *dst, u32 mtu)
1042{
1043 struct rt6_info *rt6 = (struct rt6_info*)dst;
1044
1045 if (mtu < dst_mtu(dst) && rt6->rt6i_dst.plen == 128) {
1046 rt6->rt6i_flags |= RTF_MODIFIED;
1047 if (mtu < IPV6_MIN_MTU) {
1048 u32 features = dst_metric(dst, RTAX_FEATURES);
1049 mtu = IPV6_MIN_MTU;
1050 features |= RTAX_FEATURE_ALLFRAG;
1051 dst_metric_set(dst, RTAX_FEATURES, features);
1052 }
1053 dst_metric_set(dst, RTAX_MTU, mtu);
1054 }
1055}
1056
1057static unsigned int ip6_default_advmss(const struct dst_entry *dst)
1058{
1059 struct net_device *dev = dst->dev;
1060 unsigned int mtu = dst_mtu(dst);
1061 struct net *net = dev_net(dev);
1062
1063 mtu -= sizeof(struct ipv6hdr) + sizeof(struct tcphdr);
1064
1065 if (mtu < net->ipv6.sysctl.ip6_rt_min_advmss)
1066 mtu = net->ipv6.sysctl.ip6_rt_min_advmss;
1067
1068 /*
1069 * Maximal non-jumbo IPv6 payload is IPV6_MAXPLEN and
1070 * corresponding MSS is IPV6_MAXPLEN - tcp_header_size.
1071 * IPV6_MAXPLEN is also valid and means: "any MSS,
1072 * rely only on pmtu discovery"
1073 */
1074 if (mtu > IPV6_MAXPLEN - sizeof(struct tcphdr))
1075 mtu = IPV6_MAXPLEN;
1076 return mtu;
1077}
1078
1079static unsigned int ip6_mtu(const struct dst_entry *dst)
1080{
1081 struct inet6_dev *idev;
1082 unsigned int mtu = dst_metric_raw(dst, RTAX_MTU);
1083
1084 if (mtu)
1085 return mtu;
1086
1087 mtu = IPV6_MIN_MTU;
1088
1089 rcu_read_lock();
1090 idev = __in6_dev_get(dst->dev);
1091 if (idev)
1092 mtu = idev->cnf.mtu6;
1093 rcu_read_unlock();
1094
1095 return mtu;
1096}
1097
1098static struct dst_entry *icmp6_dst_gc_list;
1099static DEFINE_SPINLOCK(icmp6_dst_lock);
1100
1101struct dst_entry *icmp6_dst_alloc(struct net_device *dev,
1102 struct neighbour *neigh,
1103 struct flowi6 *fl6)
1104{
1105 struct dst_entry *dst;
1106 struct rt6_info *rt;
1107 struct inet6_dev *idev = in6_dev_get(dev);
1108 struct net *net = dev_net(dev);
1109
1110 if (unlikely(!idev))
1111 return ERR_PTR(-ENODEV);
1112
1113 rt = ip6_dst_alloc(&net->ipv6.ip6_dst_ops, dev, 0);
1114 if (unlikely(!rt)) {
1115 in6_dev_put(idev);
1116 dst = ERR_PTR(-ENOMEM);
1117 goto out;
1118 }
1119
1120 if (neigh)
1121 neigh_hold(neigh);
1122 else {
1123 neigh = ip6_neigh_lookup(&rt->dst, &fl6->daddr);
1124 if (IS_ERR(neigh)) {
1125 in6_dev_put(idev);
1126 dst_free(&rt->dst);
1127 return ERR_CAST(neigh);
1128 }
1129 }
1130
1131 rt->dst.flags |= DST_HOST;
1132 rt->dst.output = ip6_output;
1133 dst_set_neighbour(&rt->dst, neigh);
1134 atomic_set(&rt->dst.__refcnt, 1);
1135 rt->rt6i_dst.addr = fl6->daddr;
1136 rt->rt6i_dst.plen = 128;
1137 rt->rt6i_idev = idev;
1138 dst_metric_set(&rt->dst, RTAX_HOPLIMIT, 255);
1139
1140 spin_lock_bh(&icmp6_dst_lock);
1141 rt->dst.next = icmp6_dst_gc_list;
1142 icmp6_dst_gc_list = &rt->dst;
1143 spin_unlock_bh(&icmp6_dst_lock);
1144
1145 fib6_force_start_gc(net);
1146
1147 dst = xfrm_lookup(net, &rt->dst, flowi6_to_flowi(fl6), NULL, 0);
1148
1149out:
1150 return dst;
1151}
1152
1153int icmp6_dst_gc(void)
1154{
1155 struct dst_entry *dst, **pprev;
1156 int more = 0;
1157
1158 spin_lock_bh(&icmp6_dst_lock);
1159 pprev = &icmp6_dst_gc_list;
1160
1161 while ((dst = *pprev) != NULL) {
1162 if (!atomic_read(&dst->__refcnt)) {
1163 *pprev = dst->next;
1164 dst_free(dst);
1165 } else {
1166 pprev = &dst->next;
1167 ++more;
1168 }
1169 }
1170
1171 spin_unlock_bh(&icmp6_dst_lock);
1172
1173 return more;
1174}
1175
1176static void icmp6_clean_all(int (*func)(struct rt6_info *rt, void *arg),
1177 void *arg)
1178{
1179 struct dst_entry *dst, **pprev;
1180
1181 spin_lock_bh(&icmp6_dst_lock);
1182 pprev = &icmp6_dst_gc_list;
1183 while ((dst = *pprev) != NULL) {
1184 struct rt6_info *rt = (struct rt6_info *) dst;
1185 if (func(rt, arg)) {
1186 *pprev = dst->next;
1187 dst_free(dst);
1188 } else {
1189 pprev = &dst->next;
1190 }
1191 }
1192 spin_unlock_bh(&icmp6_dst_lock);
1193}
1194
1195static int ip6_dst_gc(struct dst_ops *ops)
1196{
1197 unsigned long now = jiffies;
1198 struct net *net = container_of(ops, struct net, ipv6.ip6_dst_ops);
1199 int rt_min_interval = net->ipv6.sysctl.ip6_rt_gc_min_interval;
1200 int rt_max_size = net->ipv6.sysctl.ip6_rt_max_size;
1201 int rt_elasticity = net->ipv6.sysctl.ip6_rt_gc_elasticity;
1202 int rt_gc_timeout = net->ipv6.sysctl.ip6_rt_gc_timeout;
1203 unsigned long rt_last_gc = net->ipv6.ip6_rt_last_gc;
1204 int entries;
1205
1206 entries = dst_entries_get_fast(ops);
1207 if (time_after(rt_last_gc + rt_min_interval, now) &&
1208 entries <= rt_max_size)
1209 goto out;
1210
1211 net->ipv6.ip6_rt_gc_expire++;
1212 fib6_run_gc(net->ipv6.ip6_rt_gc_expire, net);
1213 net->ipv6.ip6_rt_last_gc = now;
1214 entries = dst_entries_get_slow(ops);
1215 if (entries < ops->gc_thresh)
1216 net->ipv6.ip6_rt_gc_expire = rt_gc_timeout>>1;
1217out:
1218 net->ipv6.ip6_rt_gc_expire -= net->ipv6.ip6_rt_gc_expire>>rt_elasticity;
1219 return entries > rt_max_size;
1220}
1221
1222/* Clean host part of a prefix. Not necessary in radix tree,
1223 but results in cleaner routing tables.
1224
1225 Remove it only when all the things will work!
1226 */
1227
1228int ip6_dst_hoplimit(struct dst_entry *dst)
1229{
1230 int hoplimit = dst_metric_raw(dst, RTAX_HOPLIMIT);
1231 if (hoplimit == 0) {
1232 struct net_device *dev = dst->dev;
1233 struct inet6_dev *idev;
1234
1235 rcu_read_lock();
1236 idev = __in6_dev_get(dev);
1237 if (idev)
1238 hoplimit = idev->cnf.hop_limit;
1239 else
1240 hoplimit = dev_net(dev)->ipv6.devconf_all->hop_limit;
1241 rcu_read_unlock();
1242 }
1243 return hoplimit;
1244}
1245EXPORT_SYMBOL(ip6_dst_hoplimit);
1246
1247/*
1248 *
1249 */
1250
1251int ip6_route_add(struct fib6_config *cfg)
1252{
1253 int err;
1254 struct net *net = cfg->fc_nlinfo.nl_net;
1255 struct rt6_info *rt = NULL;
1256 struct net_device *dev = NULL;
1257 struct inet6_dev *idev = NULL;
1258 struct fib6_table *table;
1259 int addr_type;
1260
1261 if (cfg->fc_dst_len > 128 || cfg->fc_src_len > 128)
1262 return -EINVAL;
1263#ifndef CONFIG_IPV6_SUBTREES
1264 if (cfg->fc_src_len)
1265 return -EINVAL;
1266#endif
1267 if (cfg->fc_ifindex) {
1268 err = -ENODEV;
1269 dev = dev_get_by_index(net, cfg->fc_ifindex);
1270 if (!dev)
1271 goto out;
1272 idev = in6_dev_get(dev);
1273 if (!idev)
1274 goto out;
1275 }
1276
1277 if (cfg->fc_metric == 0)
1278 cfg->fc_metric = IP6_RT_PRIO_USER;
1279
1280 err = -ENOBUFS;
1281 if (cfg->fc_nlinfo.nlh &&
1282 !(cfg->fc_nlinfo.nlh->nlmsg_flags & NLM_F_CREATE)) {
1283 table = fib6_get_table(net, cfg->fc_table);
1284 if (!table) {
1285 pr_warn("NLM_F_CREATE should be specified when creating new route\n");
1286 table = fib6_new_table(net, cfg->fc_table);
1287 }
1288 } else {
1289 table = fib6_new_table(net, cfg->fc_table);
1290 }
1291
1292 if (!table)
1293 goto out;
1294
1295 rt = ip6_dst_alloc(&net->ipv6.ip6_dst_ops, NULL, DST_NOCOUNT);
1296
1297 if (!rt) {
1298 err = -ENOMEM;
1299 goto out;
1300 }
1301
1302 rt->dst.obsolete = -1;
1303
1304 if (cfg->fc_flags & RTF_EXPIRES)
1305 rt6_set_expires(rt, jiffies +
1306 clock_t_to_jiffies(cfg->fc_expires));
1307 else
1308 rt6_clean_expires(rt);
1309
1310 if (cfg->fc_protocol == RTPROT_UNSPEC)
1311 cfg->fc_protocol = RTPROT_BOOT;
1312 rt->rt6i_protocol = cfg->fc_protocol;
1313
1314 addr_type = ipv6_addr_type(&cfg->fc_dst);
1315
1316 if (addr_type & IPV6_ADDR_MULTICAST)
1317 rt->dst.input = ip6_mc_input;
1318 else if (cfg->fc_flags & RTF_LOCAL)
1319 rt->dst.input = ip6_input;
1320 else
1321 rt->dst.input = ip6_forward;
1322
1323 rt->dst.output = ip6_output;
1324
1325 ipv6_addr_prefix(&rt->rt6i_dst.addr, &cfg->fc_dst, cfg->fc_dst_len);
1326 rt->rt6i_dst.plen = cfg->fc_dst_len;
1327 if (rt->rt6i_dst.plen == 128)
1328 rt->dst.flags |= DST_HOST;
1329
1330 if (!(rt->dst.flags & DST_HOST) && cfg->fc_mx) {
1331 u32 *metrics = kzalloc(sizeof(u32) * RTAX_MAX, GFP_KERNEL);
1332 if (!metrics) {
1333 err = -ENOMEM;
1334 goto out;
1335 }
1336 dst_init_metrics(&rt->dst, metrics, 0);
1337 }
1338#ifdef CONFIG_IPV6_SUBTREES
1339 ipv6_addr_prefix(&rt->rt6i_src.addr, &cfg->fc_src, cfg->fc_src_len);
1340 rt->rt6i_src.plen = cfg->fc_src_len;
1341#endif
1342
1343 rt->rt6i_metric = cfg->fc_metric;
1344
1345 /* We cannot add true routes via loopback here,
1346 they would result in kernel looping; promote them to reject routes
1347 */
1348 if ((cfg->fc_flags & RTF_REJECT) ||
1349 (dev && (dev->flags & IFF_LOOPBACK) &&
1350 !(addr_type & IPV6_ADDR_LOOPBACK) &&
1351 !(cfg->fc_flags & RTF_LOCAL))) {
1352 /* hold loopback dev/idev if we haven't done so. */
1353 if (dev != net->loopback_dev) {
1354 if (dev) {
1355 dev_put(dev);
1356 in6_dev_put(idev);
1357 }
1358 dev = net->loopback_dev;
1359 dev_hold(dev);
1360 idev = in6_dev_get(dev);
1361 if (!idev) {
1362 err = -ENODEV;
1363 goto out;
1364 }
1365 }
1366 rt->dst.output = ip6_pkt_discard_out;
1367 rt->dst.input = ip6_pkt_discard;
1368 rt->dst.error = -ENETUNREACH;
1369 rt->rt6i_flags = RTF_REJECT|RTF_NONEXTHOP;
1370 goto install_route;
1371 }
1372
1373 if (cfg->fc_flags & RTF_GATEWAY) {
1374 const struct in6_addr *gw_addr;
1375 int gwa_type;
1376
1377 gw_addr = &cfg->fc_gateway;
1378 rt->rt6i_gateway = *gw_addr;
1379 gwa_type = ipv6_addr_type(gw_addr);
1380
1381 if (gwa_type != (IPV6_ADDR_LINKLOCAL|IPV6_ADDR_UNICAST)) {
1382 struct rt6_info *grt;
1383
1384 /* IPv6 strictly inhibits using not link-local
1385 addresses as nexthop address.
1386 Otherwise, router will not able to send redirects.
1387 It is very good, but in some (rare!) circumstances
1388 (SIT, PtP, NBMA NOARP links) it is handy to allow
1389 some exceptions. --ANK
1390 */
1391 err = -EINVAL;
1392 if (!(gwa_type & IPV6_ADDR_UNICAST))
1393 goto out;
1394
1395 grt = rt6_lookup(net, gw_addr, NULL, cfg->fc_ifindex, 1);
1396
1397 err = -EHOSTUNREACH;
1398 if (!grt)
1399 goto out;
1400 if (dev) {
1401 if (dev != grt->dst.dev) {
1402 dst_release(&grt->dst);
1403 goto out;
1404 }
1405 } else {
1406 dev = grt->dst.dev;
1407 idev = grt->rt6i_idev;
1408 dev_hold(dev);
1409 in6_dev_hold(grt->rt6i_idev);
1410 }
1411 if (!(grt->rt6i_flags & RTF_GATEWAY))
1412 err = 0;
1413 dst_release(&grt->dst);
1414
1415 if (err)
1416 goto out;
1417 }
1418 err = -EINVAL;
1419 if (!dev || (dev->flags & IFF_LOOPBACK))
1420 goto out;
1421 }
1422
1423 err = -ENODEV;
1424 if (!dev)
1425 goto out;
1426
1427 if (!ipv6_addr_any(&cfg->fc_prefsrc)) {
1428 if (!ipv6_chk_addr(net, &cfg->fc_prefsrc, dev, 0)) {
1429 err = -EINVAL;
1430 goto out;
1431 }
1432 rt->rt6i_prefsrc.addr = cfg->fc_prefsrc;
1433 rt->rt6i_prefsrc.plen = 128;
1434 } else
1435 rt->rt6i_prefsrc.plen = 0;
1436
1437 if (cfg->fc_flags & (RTF_GATEWAY | RTF_NONEXTHOP)) {
1438 err = rt6_bind_neighbour(rt, dev);
1439 if (err)
1440 goto out;
1441 }
1442
1443 rt->rt6i_flags = cfg->fc_flags;
1444
1445install_route:
1446 if (cfg->fc_mx) {
1447 struct nlattr *nla;
1448 int remaining;
1449
1450 nla_for_each_attr(nla, cfg->fc_mx, cfg->fc_mx_len, remaining) {
1451 int type = nla_type(nla);
1452
1453 if (type) {
1454 if (type > RTAX_MAX) {
1455 err = -EINVAL;
1456 goto out;
1457 }
1458
1459 dst_metric_set(&rt->dst, type, nla_get_u32(nla));
1460 }
1461 }
1462 }
1463
1464 rt->dst.dev = dev;
1465 rt->rt6i_idev = idev;
1466 rt->rt6i_table = table;
1467
1468 cfg->fc_nlinfo.nl_net = dev_net(dev);
1469
1470 return __ip6_ins_rt(rt, &cfg->fc_nlinfo);
1471
1472out:
1473 if (dev)
1474 dev_put(dev);
1475 if (idev)
1476 in6_dev_put(idev);
1477 if (rt)
1478 dst_free(&rt->dst);
1479 return err;
1480}
1481
1482static int __ip6_del_rt(struct rt6_info *rt, struct nl_info *info)
1483{
1484 int err;
1485 struct fib6_table *table;
1486 struct net *net = dev_net(rt->dst.dev);
1487
1488 if (rt == net->ipv6.ip6_null_entry)
1489 return -ENOENT;
1490
1491 table = rt->rt6i_table;
1492 write_lock_bh(&table->tb6_lock);
1493
1494 err = fib6_del(rt, info);
1495 dst_release(&rt->dst);
1496
1497 write_unlock_bh(&table->tb6_lock);
1498
1499 return err;
1500}
1501
1502int ip6_del_rt(struct rt6_info *rt)
1503{
1504 struct nl_info info = {
1505 .nl_net = dev_net(rt->dst.dev),
1506 };
1507 return __ip6_del_rt(rt, &info);
1508}
1509
1510static int ip6_route_del(struct fib6_config *cfg)
1511{
1512 struct fib6_table *table;
1513 struct fib6_node *fn;
1514 struct rt6_info *rt;
1515 int err = -ESRCH;
1516
1517 table = fib6_get_table(cfg->fc_nlinfo.nl_net, cfg->fc_table);
1518 if (!table)
1519 return err;
1520
1521 read_lock_bh(&table->tb6_lock);
1522
1523 fn = fib6_locate(&table->tb6_root,
1524 &cfg->fc_dst, cfg->fc_dst_len,
1525 &cfg->fc_src, cfg->fc_src_len);
1526
1527 if (fn) {
1528 for (rt = fn->leaf; rt; rt = rt->dst.rt6_next) {
1529 if (cfg->fc_ifindex &&
1530 (!rt->dst.dev ||
1531 rt->dst.dev->ifindex != cfg->fc_ifindex))
1532 continue;
1533 if (cfg->fc_flags & RTF_GATEWAY &&
1534 !ipv6_addr_equal(&cfg->fc_gateway, &rt->rt6i_gateway))
1535 continue;
1536 if (cfg->fc_metric && cfg->fc_metric != rt->rt6i_metric)
1537 continue;
1538 dst_hold(&rt->dst);
1539 read_unlock_bh(&table->tb6_lock);
1540
1541 return __ip6_del_rt(rt, &cfg->fc_nlinfo);
1542 }
1543 }
1544 read_unlock_bh(&table->tb6_lock);
1545
1546 return err;
1547}
1548
1549/*
1550 * Handle redirects
1551 */
1552struct ip6rd_flowi {
1553 struct flowi6 fl6;
1554 struct in6_addr gateway;
1555};
1556
1557static struct rt6_info *__ip6_route_redirect(struct net *net,
1558 struct fib6_table *table,
1559 struct flowi6 *fl6,
1560 int flags)
1561{
1562 struct ip6rd_flowi *rdfl = (struct ip6rd_flowi *)fl6;
1563 struct rt6_info *rt;
1564 struct fib6_node *fn;
1565
1566 /*
1567 * Get the "current" route for this destination and
1568 * check if the redirect has come from approriate router.
1569 *
1570 * RFC 2461 specifies that redirects should only be
1571 * accepted if they come from the nexthop to the target.
1572 * Due to the way the routes are chosen, this notion
1573 * is a bit fuzzy and one might need to check all possible
1574 * routes.
1575 */
1576
1577 read_lock_bh(&table->tb6_lock);
1578 fn = fib6_lookup(&table->tb6_root, &fl6->daddr, &fl6->saddr);
1579restart:
1580 for (rt = fn->leaf; rt; rt = rt->dst.rt6_next) {
1581 /*
1582 * Current route is on-link; redirect is always invalid.
1583 *
1584 * Seems, previous statement is not true. It could
1585 * be node, which looks for us as on-link (f.e. proxy ndisc)
1586 * But then router serving it might decide, that we should
1587 * know truth 8)8) --ANK (980726).
1588 */
1589 if (rt6_check_expired(rt))
1590 continue;
1591 if (!(rt->rt6i_flags & RTF_GATEWAY))
1592 continue;
1593 if (fl6->flowi6_oif != rt->dst.dev->ifindex)
1594 continue;
1595 if (!ipv6_addr_equal(&rdfl->gateway, &rt->rt6i_gateway))
1596 continue;
1597 break;
1598 }
1599
1600 if (!rt)
1601 rt = net->ipv6.ip6_null_entry;
1602 BACKTRACK(net, &fl6->saddr);
1603out:
1604 dst_hold(&rt->dst);
1605
1606 read_unlock_bh(&table->tb6_lock);
1607
1608 return rt;
1609};
1610
1611static struct rt6_info *ip6_route_redirect(const struct in6_addr *dest,
1612 const struct in6_addr *src,
1613 const struct in6_addr *gateway,
1614 struct net_device *dev)
1615{
1616 int flags = RT6_LOOKUP_F_HAS_SADDR;
1617 struct net *net = dev_net(dev);
1618 struct ip6rd_flowi rdfl = {
1619 .fl6 = {
1620 .flowi6_oif = dev->ifindex,
1621 .daddr = *dest,
1622 .saddr = *src,
1623 },
1624 };
1625
1626 rdfl.gateway = *gateway;
1627
1628 if (rt6_need_strict(dest))
1629 flags |= RT6_LOOKUP_F_IFACE;
1630
1631 return (struct rt6_info *)fib6_rule_lookup(net, &rdfl.fl6,
1632 flags, __ip6_route_redirect);
1633}
1634
1635void rt6_redirect(const struct in6_addr *dest, const struct in6_addr *src,
1636 const struct in6_addr *saddr,
1637 struct neighbour *neigh, u8 *lladdr, int on_link)
1638{
1639 struct rt6_info *rt, *nrt = NULL;
1640 struct netevent_redirect netevent;
1641 struct net *net = dev_net(neigh->dev);
1642
1643 rt = ip6_route_redirect(dest, src, saddr, neigh->dev);
1644
1645 if (rt == net->ipv6.ip6_null_entry) {
1646 net_dbg_ratelimited("rt6_redirect: source isn't a valid nexthop for redirect target\n");
1647 goto out;
1648 }
1649
1650 /*
1651 * We have finally decided to accept it.
1652 */
1653
1654 neigh_update(neigh, lladdr, NUD_STALE,
1655 NEIGH_UPDATE_F_WEAK_OVERRIDE|
1656 NEIGH_UPDATE_F_OVERRIDE|
1657 (on_link ? 0 : (NEIGH_UPDATE_F_OVERRIDE_ISROUTER|
1658 NEIGH_UPDATE_F_ISROUTER))
1659 );
1660
1661 /*
1662 * Redirect received -> path was valid.
1663 * Look, redirects are sent only in response to data packets,
1664 * so that this nexthop apparently is reachable. --ANK
1665 */
1666 dst_confirm(&rt->dst);
1667
1668 /* Duplicate redirect: silently ignore. */
1669 if (neigh == dst_get_neighbour_noref_raw(&rt->dst))
1670 goto out;
1671
1672 nrt = ip6_rt_copy(rt, dest);
1673 if (!nrt)
1674 goto out;
1675
1676 nrt->rt6i_flags = RTF_GATEWAY|RTF_UP|RTF_DYNAMIC|RTF_CACHE;
1677 if (on_link)
1678 nrt->rt6i_flags &= ~RTF_GATEWAY;
1679
1680 nrt->rt6i_gateway = *(struct in6_addr *)neigh->primary_key;
1681 dst_set_neighbour(&nrt->dst, neigh_clone(neigh));
1682
1683 if (ip6_ins_rt(nrt))
1684 goto out;
1685
1686 netevent.old = &rt->dst;
1687 netevent.new = &nrt->dst;
1688 call_netevent_notifiers(NETEVENT_REDIRECT, &netevent);
1689
1690 if (rt->rt6i_flags & RTF_CACHE) {
1691 ip6_del_rt(rt);
1692 return;
1693 }
1694
1695out:
1696 dst_release(&rt->dst);
1697}
1698
1699/*
1700 * Handle ICMP "packet too big" messages
1701 * i.e. Path MTU discovery
1702 */
1703
1704static void rt6_do_pmtu_disc(const struct in6_addr *daddr, const struct in6_addr *saddr,
1705 struct net *net, u32 pmtu, int ifindex)
1706{
1707 struct rt6_info *rt, *nrt;
1708 int allfrag = 0;
1709again:
1710 rt = rt6_lookup(net, daddr, saddr, ifindex, 0);
1711 if (!rt)
1712 return;
1713
1714 if (rt6_check_expired(rt)) {
1715 ip6_del_rt(rt);
1716 goto again;
1717 }
1718
1719 if (pmtu >= dst_mtu(&rt->dst))
1720 goto out;
1721
1722 if (pmtu < IPV6_MIN_MTU) {
1723 /*
1724 * According to RFC2460, PMTU is set to the IPv6 Minimum Link
1725 * MTU (1280) and a fragment header should always be included
1726 * after a node receiving Too Big message reporting PMTU is
1727 * less than the IPv6 Minimum Link MTU.
1728 */
1729 pmtu = IPV6_MIN_MTU;
1730 allfrag = 1;
1731 }
1732
1733 /* New mtu received -> path was valid.
1734 They are sent only in response to data packets,
1735 so that this nexthop apparently is reachable. --ANK
1736 */
1737 dst_confirm(&rt->dst);
1738
1739 /* Host route. If it is static, it would be better
1740 not to override it, but add new one, so that
1741 when cache entry will expire old pmtu
1742 would return automatically.
1743 */
1744 if (rt->rt6i_flags & RTF_CACHE) {
1745 dst_metric_set(&rt->dst, RTAX_MTU, pmtu);
1746 if (allfrag) {
1747 u32 features = dst_metric(&rt->dst, RTAX_FEATURES);
1748 features |= RTAX_FEATURE_ALLFRAG;
1749 dst_metric_set(&rt->dst, RTAX_FEATURES, features);
1750 }
1751 rt6_update_expires(rt, net->ipv6.sysctl.ip6_rt_mtu_expires);
1752 rt->rt6i_flags |= RTF_MODIFIED;
1753 goto out;
1754 }
1755
1756 /* Network route.
1757 Two cases are possible:
1758 1. It is connected route. Action: COW
1759 2. It is gatewayed route or NONEXTHOP route. Action: clone it.
1760 */
1761 if (!dst_get_neighbour_noref_raw(&rt->dst) && !(rt->rt6i_flags & RTF_NONEXTHOP))
1762 nrt = rt6_alloc_cow(rt, daddr, saddr);
1763 else
1764 nrt = rt6_alloc_clone(rt, daddr);
1765
1766 if (nrt) {
1767 dst_metric_set(&nrt->dst, RTAX_MTU, pmtu);
1768 if (allfrag) {
1769 u32 features = dst_metric(&nrt->dst, RTAX_FEATURES);
1770 features |= RTAX_FEATURE_ALLFRAG;
1771 dst_metric_set(&nrt->dst, RTAX_FEATURES, features);
1772 }
1773
1774 /* According to RFC 1981, detecting PMTU increase shouldn't be
1775 * happened within 5 mins, the recommended timer is 10 mins.
1776 * Here this route expiration time is set to ip6_rt_mtu_expires
1777 * which is 10 mins. After 10 mins the decreased pmtu is expired
1778 * and detecting PMTU increase will be automatically happened.
1779 */
1780 rt6_update_expires(nrt, net->ipv6.sysctl.ip6_rt_mtu_expires);
1781 nrt->rt6i_flags |= RTF_DYNAMIC;
1782 ip6_ins_rt(nrt);
1783 }
1784out:
1785 dst_release(&rt->dst);
1786}
1787
1788void rt6_pmtu_discovery(const struct in6_addr *daddr, const struct in6_addr *saddr,
1789 struct net_device *dev, u32 pmtu)
1790{
1791 struct net *net = dev_net(dev);
1792
1793 /*
1794 * RFC 1981 states that a node "MUST reduce the size of the packets it
1795 * is sending along the path" that caused the Packet Too Big message.
1796 * Since it's not possible in the general case to determine which
1797 * interface was used to send the original packet, we update the MTU
1798 * on the interface that will be used to send future packets. We also
1799 * update the MTU on the interface that received the Packet Too Big in
1800 * case the original packet was forced out that interface with
1801 * SO_BINDTODEVICE or similar. This is the next best thing to the
1802 * correct behaviour, which would be to update the MTU on all
1803 * interfaces.
1804 */
1805 rt6_do_pmtu_disc(daddr, saddr, net, pmtu, 0);
1806 rt6_do_pmtu_disc(daddr, saddr, net, pmtu, dev->ifindex);
1807}
1808
1809/*
1810 * Misc support functions
1811 */
1812
1813static struct rt6_info *ip6_rt_copy(struct rt6_info *ort,
1814 const struct in6_addr *dest)
1815{
1816 struct net *net = dev_net(ort->dst.dev);
1817 struct rt6_info *rt = ip6_dst_alloc(&net->ipv6.ip6_dst_ops,
1818 ort->dst.dev, 0);
1819
1820 if (rt) {
1821 rt->dst.input = ort->dst.input;
1822 rt->dst.output = ort->dst.output;
1823 rt->dst.flags |= DST_HOST;
1824
1825 rt->rt6i_dst.addr = *dest;
1826 rt->rt6i_dst.plen = 128;
1827 dst_copy_metrics(&rt->dst, &ort->dst);
1828 rt->dst.error = ort->dst.error;
1829 rt->rt6i_idev = ort->rt6i_idev;
1830 if (rt->rt6i_idev)
1831 in6_dev_hold(rt->rt6i_idev);
1832 rt->dst.lastuse = jiffies;
1833
1834 rt->rt6i_gateway = ort->rt6i_gateway;
1835 rt->rt6i_flags = ort->rt6i_flags;
1836 if ((ort->rt6i_flags & (RTF_DEFAULT | RTF_ADDRCONF)) ==
1837 (RTF_DEFAULT | RTF_ADDRCONF))
1838 rt6_set_from(rt, ort);
1839 else
1840 rt6_clean_expires(rt);
1841 rt->rt6i_metric = 0;
1842
1843#ifdef CONFIG_IPV6_SUBTREES
1844 memcpy(&rt->rt6i_src, &ort->rt6i_src, sizeof(struct rt6key));
1845#endif
1846 memcpy(&rt->rt6i_prefsrc, &ort->rt6i_prefsrc, sizeof(struct rt6key));
1847 rt->rt6i_table = ort->rt6i_table;
1848 }
1849 return rt;
1850}
1851
1852#ifdef CONFIG_IPV6_ROUTE_INFO
1853static struct rt6_info *rt6_get_route_info(struct net *net,
1854 const struct in6_addr *prefix, int prefixlen,
1855 const struct in6_addr *gwaddr, int ifindex)
1856{
1857 struct fib6_node *fn;
1858 struct rt6_info *rt = NULL;
1859 struct fib6_table *table;
1860
1861 table = fib6_get_table(net, RT6_TABLE_INFO);
1862 if (!table)
1863 return NULL;
1864
1865 write_lock_bh(&table->tb6_lock);
1866 fn = fib6_locate(&table->tb6_root, prefix ,prefixlen, NULL, 0);
1867 if (!fn)
1868 goto out;
1869
1870 for (rt = fn->leaf; rt; rt = rt->dst.rt6_next) {
1871 if (rt->dst.dev->ifindex != ifindex)
1872 continue;
1873 if ((rt->rt6i_flags & (RTF_ROUTEINFO|RTF_GATEWAY)) != (RTF_ROUTEINFO|RTF_GATEWAY))
1874 continue;
1875 if (!ipv6_addr_equal(&rt->rt6i_gateway, gwaddr))
1876 continue;
1877 dst_hold(&rt->dst);
1878 break;
1879 }
1880out:
1881 write_unlock_bh(&table->tb6_lock);
1882 return rt;
1883}
1884
1885static struct rt6_info *rt6_add_route_info(struct net *net,
1886 const struct in6_addr *prefix, int prefixlen,
1887 const struct in6_addr *gwaddr, int ifindex,
1888 unsigned int pref)
1889{
1890 struct fib6_config cfg = {
1891 .fc_table = RT6_TABLE_INFO,
1892 .fc_metric = IP6_RT_PRIO_USER,
1893 .fc_ifindex = ifindex,
1894 .fc_dst_len = prefixlen,
1895 .fc_flags = RTF_GATEWAY | RTF_ADDRCONF | RTF_ROUTEINFO |
1896 RTF_UP | RTF_PREF(pref),
1897 .fc_nlinfo.pid = 0,
1898 .fc_nlinfo.nlh = NULL,
1899 .fc_nlinfo.nl_net = net,
1900 };
1901
1902 cfg.fc_dst = *prefix;
1903 cfg.fc_gateway = *gwaddr;
1904
1905 /* We should treat it as a default route if prefix length is 0. */
1906 if (!prefixlen)
1907 cfg.fc_flags |= RTF_DEFAULT;
1908
1909 ip6_route_add(&cfg);
1910
1911 return rt6_get_route_info(net, prefix, prefixlen, gwaddr, ifindex);
1912}
1913#endif
1914
1915struct rt6_info *rt6_get_dflt_router(const struct in6_addr *addr, struct net_device *dev)
1916{
1917 struct rt6_info *rt;
1918 struct fib6_table *table;
1919
1920 table = fib6_get_table(dev_net(dev), RT6_TABLE_DFLT);
1921 if (!table)
1922 return NULL;
1923
1924 write_lock_bh(&table->tb6_lock);
1925 for (rt = table->tb6_root.leaf; rt; rt=rt->dst.rt6_next) {
1926 if (dev == rt->dst.dev &&
1927 ((rt->rt6i_flags & (RTF_ADDRCONF | RTF_DEFAULT)) == (RTF_ADDRCONF | RTF_DEFAULT)) &&
1928 ipv6_addr_equal(&rt->rt6i_gateway, addr))
1929 break;
1930 }
1931 if (rt)
1932 dst_hold(&rt->dst);
1933 write_unlock_bh(&table->tb6_lock);
1934 return rt;
1935}
1936
1937struct rt6_info *rt6_add_dflt_router(const struct in6_addr *gwaddr,
1938 struct net_device *dev,
1939 unsigned int pref)
1940{
1941 struct fib6_config cfg = {
1942 .fc_table = RT6_TABLE_DFLT,
1943 .fc_metric = IP6_RT_PRIO_USER,
1944 .fc_ifindex = dev->ifindex,
1945 .fc_flags = RTF_GATEWAY | RTF_ADDRCONF | RTF_DEFAULT |
1946 RTF_UP | RTF_EXPIRES | RTF_PREF(pref),
1947 .fc_nlinfo.pid = 0,
1948 .fc_nlinfo.nlh = NULL,
1949 .fc_nlinfo.nl_net = dev_net(dev),
1950 };
1951
1952 cfg.fc_gateway = *gwaddr;
1953
1954 ip6_route_add(&cfg);
1955
1956 return rt6_get_dflt_router(gwaddr, dev);
1957}
1958
1959void rt6_purge_dflt_routers(struct net *net)
1960{
1961 struct rt6_info *rt;
1962 struct fib6_table *table;
1963
1964 /* NOTE: Keep consistent with rt6_get_dflt_router */
1965 table = fib6_get_table(net, RT6_TABLE_DFLT);
1966 if (!table)
1967 return;
1968
1969restart:
1970 read_lock_bh(&table->tb6_lock);
1971 for (rt = table->tb6_root.leaf; rt; rt = rt->dst.rt6_next) {
1972 if (rt->rt6i_flags & (RTF_DEFAULT | RTF_ADDRCONF)) {
1973 dst_hold(&rt->dst);
1974 read_unlock_bh(&table->tb6_lock);
1975 ip6_del_rt(rt);
1976 goto restart;
1977 }
1978 }
1979 read_unlock_bh(&table->tb6_lock);
1980}
1981
1982static void rtmsg_to_fib6_config(struct net *net,
1983 struct in6_rtmsg *rtmsg,
1984 struct fib6_config *cfg)
1985{
1986 memset(cfg, 0, sizeof(*cfg));
1987
1988 cfg->fc_table = RT6_TABLE_MAIN;
1989 cfg->fc_ifindex = rtmsg->rtmsg_ifindex;
1990 cfg->fc_metric = rtmsg->rtmsg_metric;
1991 cfg->fc_expires = rtmsg->rtmsg_info;
1992 cfg->fc_dst_len = rtmsg->rtmsg_dst_len;
1993 cfg->fc_src_len = rtmsg->rtmsg_src_len;
1994 cfg->fc_flags = rtmsg->rtmsg_flags;
1995
1996 cfg->fc_nlinfo.nl_net = net;
1997
1998 cfg->fc_dst = rtmsg->rtmsg_dst;
1999 cfg->fc_src = rtmsg->rtmsg_src;
2000 cfg->fc_gateway = rtmsg->rtmsg_gateway;
2001}
2002
2003int ipv6_route_ioctl(struct net *net, unsigned int cmd, void __user *arg)
2004{
2005 struct fib6_config cfg;
2006 struct in6_rtmsg rtmsg;
2007 int err;
2008
2009 switch(cmd) {
2010 case SIOCADDRT: /* Add a route */
2011 case SIOCDELRT: /* Delete a route */
2012 if (!capable(CAP_NET_ADMIN))
2013 return -EPERM;
2014 err = copy_from_user(&rtmsg, arg,
2015 sizeof(struct in6_rtmsg));
2016 if (err)
2017 return -EFAULT;
2018
2019 rtmsg_to_fib6_config(net, &rtmsg, &cfg);
2020
2021 rtnl_lock();
2022 switch (cmd) {
2023 case SIOCADDRT:
2024 err = ip6_route_add(&cfg);
2025 break;
2026 case SIOCDELRT:
2027 err = ip6_route_del(&cfg);
2028 break;
2029 default:
2030 err = -EINVAL;
2031 }
2032 rtnl_unlock();
2033
2034 return err;
2035 }
2036
2037 return -EINVAL;
2038}
2039
2040/*
2041 * Drop the packet on the floor
2042 */
2043
2044static int ip6_pkt_drop(struct sk_buff *skb, u8 code, int ipstats_mib_noroutes)
2045{
2046 int type;
2047 struct dst_entry *dst = skb_dst(skb);
2048 switch (ipstats_mib_noroutes) {
2049 case IPSTATS_MIB_INNOROUTES:
2050 type = ipv6_addr_type(&ipv6_hdr(skb)->daddr);
2051 if (type == IPV6_ADDR_ANY) {
2052 IP6_INC_STATS(dev_net(dst->dev), ip6_dst_idev(dst),
2053 IPSTATS_MIB_INADDRERRORS);
2054 break;
2055 }
2056 /* FALLTHROUGH */
2057 case IPSTATS_MIB_OUTNOROUTES:
2058 IP6_INC_STATS(dev_net(dst->dev), ip6_dst_idev(dst),
2059 ipstats_mib_noroutes);
2060 break;
2061 }
2062 icmpv6_send(skb, ICMPV6_DEST_UNREACH, code, 0);
2063 kfree_skb(skb);
2064 return 0;
2065}
2066
2067static int ip6_pkt_discard(struct sk_buff *skb)
2068{
2069 return ip6_pkt_drop(skb, ICMPV6_NOROUTE, IPSTATS_MIB_INNOROUTES);
2070}
2071
2072static int ip6_pkt_discard_out(struct sk_buff *skb)
2073{
2074 skb->dev = skb_dst(skb)->dev;
2075 return ip6_pkt_drop(skb, ICMPV6_NOROUTE, IPSTATS_MIB_OUTNOROUTES);
2076}
2077
2078#ifdef CONFIG_IPV6_MULTIPLE_TABLES
2079
2080static int ip6_pkt_prohibit(struct sk_buff *skb)
2081{
2082 return ip6_pkt_drop(skb, ICMPV6_ADM_PROHIBITED, IPSTATS_MIB_INNOROUTES);
2083}
2084
2085static int ip6_pkt_prohibit_out(struct sk_buff *skb)
2086{
2087 skb->dev = skb_dst(skb)->dev;
2088 return ip6_pkt_drop(skb, ICMPV6_ADM_PROHIBITED, IPSTATS_MIB_OUTNOROUTES);
2089}
2090
2091#endif
2092
2093/*
2094 * Allocate a dst for local (unicast / anycast) address.
2095 */
2096
2097struct rt6_info *addrconf_dst_alloc(struct inet6_dev *idev,
2098 const struct in6_addr *addr,
2099 bool anycast)
2100{
2101 struct net *net = dev_net(idev->dev);
2102 struct rt6_info *rt = ip6_dst_alloc(&net->ipv6.ip6_dst_ops,
2103 net->loopback_dev, 0);
2104 int err;
2105
2106 if (!rt) {
2107 net_warn_ratelimited("Maximum number of routes reached, consider increasing route/max_size\n");
2108 return ERR_PTR(-ENOMEM);
2109 }
2110
2111 in6_dev_hold(idev);
2112
2113 rt->dst.flags |= DST_HOST;
2114 rt->dst.input = ip6_input;
2115 rt->dst.output = ip6_output;
2116 rt->rt6i_idev = idev;
2117 rt->dst.obsolete = -1;
2118
2119 rt->rt6i_flags = RTF_UP | RTF_NONEXTHOP;
2120 if (anycast)
2121 rt->rt6i_flags |= RTF_ANYCAST;
2122 else
2123 rt->rt6i_flags |= RTF_LOCAL;
2124 err = rt6_bind_neighbour(rt, rt->dst.dev);
2125 if (err) {
2126 dst_free(&rt->dst);
2127 return ERR_PTR(err);
2128 }
2129
2130 rt->rt6i_dst.addr = *addr;
2131 rt->rt6i_dst.plen = 128;
2132 rt->rt6i_table = fib6_get_table(net, RT6_TABLE_LOCAL);
2133
2134 atomic_set(&rt->dst.__refcnt, 1);
2135
2136 return rt;
2137}
2138
2139int ip6_route_get_saddr(struct net *net,
2140 struct rt6_info *rt,
2141 const struct in6_addr *daddr,
2142 unsigned int prefs,
2143 struct in6_addr *saddr)
2144{
2145 struct inet6_dev *idev = ip6_dst_idev((struct dst_entry*)rt);
2146 int err = 0;
2147 if (rt->rt6i_prefsrc.plen)
2148 *saddr = rt->rt6i_prefsrc.addr;
2149 else
2150 err = ipv6_dev_get_saddr(net, idev ? idev->dev : NULL,
2151 daddr, prefs, saddr);
2152 return err;
2153}
2154
2155/* remove deleted ip from prefsrc entries */
2156struct arg_dev_net_ip {
2157 struct net_device *dev;
2158 struct net *net;
2159 struct in6_addr *addr;
2160};
2161
2162static int fib6_remove_prefsrc(struct rt6_info *rt, void *arg)
2163{
2164 struct net_device *dev = ((struct arg_dev_net_ip *)arg)->dev;
2165 struct net *net = ((struct arg_dev_net_ip *)arg)->net;
2166 struct in6_addr *addr = ((struct arg_dev_net_ip *)arg)->addr;
2167
2168 if (((void *)rt->dst.dev == dev || !dev) &&
2169 rt != net->ipv6.ip6_null_entry &&
2170 ipv6_addr_equal(addr, &rt->rt6i_prefsrc.addr)) {
2171 /* remove prefsrc entry */
2172 rt->rt6i_prefsrc.plen = 0;
2173 }
2174 return 0;
2175}
2176
2177void rt6_remove_prefsrc(struct inet6_ifaddr *ifp)
2178{
2179 struct net *net = dev_net(ifp->idev->dev);
2180 struct arg_dev_net_ip adni = {
2181 .dev = ifp->idev->dev,
2182 .net = net,
2183 .addr = &ifp->addr,
2184 };
2185 fib6_clean_all(net, fib6_remove_prefsrc, 0, &adni);
2186}
2187
2188struct arg_dev_net {
2189 struct net_device *dev;
2190 struct net *net;
2191};
2192
2193static int fib6_ifdown(struct rt6_info *rt, void *arg)
2194{
2195 const struct arg_dev_net *adn = arg;
2196 const struct net_device *dev = adn->dev;
2197
2198 if ((rt->dst.dev == dev || !dev) &&
2199 rt != adn->net->ipv6.ip6_null_entry)
2200 return -1;
2201
2202 return 0;
2203}
2204
2205void rt6_ifdown(struct net *net, struct net_device *dev)
2206{
2207 struct arg_dev_net adn = {
2208 .dev = dev,
2209 .net = net,
2210 };
2211
2212 fib6_clean_all(net, fib6_ifdown, 0, &adn);
2213 icmp6_clean_all(fib6_ifdown, &adn);
2214}
2215
2216struct rt6_mtu_change_arg {
2217 struct net_device *dev;
2218 unsigned int mtu;
2219};
2220
2221static int rt6_mtu_change_route(struct rt6_info *rt, void *p_arg)
2222{
2223 struct rt6_mtu_change_arg *arg = (struct rt6_mtu_change_arg *) p_arg;
2224 struct inet6_dev *idev;
2225
2226 /* In IPv6 pmtu discovery is not optional,
2227 so that RTAX_MTU lock cannot disable it.
2228 We still use this lock to block changes
2229 caused by addrconf/ndisc.
2230 */
2231
2232 idev = __in6_dev_get(arg->dev);
2233 if (!idev)
2234 return 0;
2235
2236 /* For administrative MTU increase, there is no way to discover
2237 IPv6 PMTU increase, so PMTU increase should be updated here.
2238 Since RFC 1981 doesn't include administrative MTU increase
2239 update PMTU increase is a MUST. (i.e. jumbo frame)
2240 */
2241 /*
2242 If new MTU is less than route PMTU, this new MTU will be the
2243 lowest MTU in the path, update the route PMTU to reflect PMTU
2244 decreases; if new MTU is greater than route PMTU, and the
2245 old MTU is the lowest MTU in the path, update the route PMTU
2246 to reflect the increase. In this case if the other nodes' MTU
2247 also have the lowest MTU, TOO BIG MESSAGE will be lead to
2248 PMTU discouvery.
2249 */
2250 if (rt->dst.dev == arg->dev &&
2251 !dst_metric_locked(&rt->dst, RTAX_MTU) &&
2252 (dst_mtu(&rt->dst) >= arg->mtu ||
2253 (dst_mtu(&rt->dst) < arg->mtu &&
2254 dst_mtu(&rt->dst) == idev->cnf.mtu6))) {
2255 dst_metric_set(&rt->dst, RTAX_MTU, arg->mtu);
2256 }
2257 return 0;
2258}
2259
2260void rt6_mtu_change(struct net_device *dev, unsigned int mtu)
2261{
2262 struct rt6_mtu_change_arg arg = {
2263 .dev = dev,
2264 .mtu = mtu,
2265 };
2266
2267 fib6_clean_all(dev_net(dev), rt6_mtu_change_route, 0, &arg);
2268}
2269
2270static const struct nla_policy rtm_ipv6_policy[RTA_MAX+1] = {
2271 [RTA_GATEWAY] = { .len = sizeof(struct in6_addr) },
2272 [RTA_OIF] = { .type = NLA_U32 },
2273 [RTA_IIF] = { .type = NLA_U32 },
2274 [RTA_PRIORITY] = { .type = NLA_U32 },
2275 [RTA_METRICS] = { .type = NLA_NESTED },
2276};
2277
2278static int rtm_to_fib6_config(struct sk_buff *skb, struct nlmsghdr *nlh,
2279 struct fib6_config *cfg)
2280{
2281 struct rtmsg *rtm;
2282 struct nlattr *tb[RTA_MAX+1];
2283 int err;
2284
2285 err = nlmsg_parse(nlh, sizeof(*rtm), tb, RTA_MAX, rtm_ipv6_policy);
2286 if (err < 0)
2287 goto errout;
2288
2289 err = -EINVAL;
2290 rtm = nlmsg_data(nlh);
2291 memset(cfg, 0, sizeof(*cfg));
2292
2293 cfg->fc_table = rtm->rtm_table;
2294 cfg->fc_dst_len = rtm->rtm_dst_len;
2295 cfg->fc_src_len = rtm->rtm_src_len;
2296 cfg->fc_flags = RTF_UP;
2297 cfg->fc_protocol = rtm->rtm_protocol;
2298
2299 if (rtm->rtm_type == RTN_UNREACHABLE)
2300 cfg->fc_flags |= RTF_REJECT;
2301
2302 if (rtm->rtm_type == RTN_LOCAL)
2303 cfg->fc_flags |= RTF_LOCAL;
2304
2305 cfg->fc_nlinfo.pid = NETLINK_CB(skb).pid;
2306 cfg->fc_nlinfo.nlh = nlh;
2307 cfg->fc_nlinfo.nl_net = sock_net(skb->sk);
2308
2309 if (tb[RTA_GATEWAY]) {
2310 nla_memcpy(&cfg->fc_gateway, tb[RTA_GATEWAY], 16);
2311 cfg->fc_flags |= RTF_GATEWAY;
2312 }
2313
2314 if (tb[RTA_DST]) {
2315 int plen = (rtm->rtm_dst_len + 7) >> 3;
2316
2317 if (nla_len(tb[RTA_DST]) < plen)
2318 goto errout;
2319
2320 nla_memcpy(&cfg->fc_dst, tb[RTA_DST], plen);
2321 }
2322
2323 if (tb[RTA_SRC]) {
2324 int plen = (rtm->rtm_src_len + 7) >> 3;
2325
2326 if (nla_len(tb[RTA_SRC]) < plen)
2327 goto errout;
2328
2329 nla_memcpy(&cfg->fc_src, tb[RTA_SRC], plen);
2330 }
2331
2332 if (tb[RTA_PREFSRC])
2333 nla_memcpy(&cfg->fc_prefsrc, tb[RTA_PREFSRC], 16);
2334
2335 if (tb[RTA_OIF])
2336 cfg->fc_ifindex = nla_get_u32(tb[RTA_OIF]);
2337
2338 if (tb[RTA_PRIORITY])
2339 cfg->fc_metric = nla_get_u32(tb[RTA_PRIORITY]);
2340
2341 if (tb[RTA_METRICS]) {
2342 cfg->fc_mx = nla_data(tb[RTA_METRICS]);
2343 cfg->fc_mx_len = nla_len(tb[RTA_METRICS]);
2344 }
2345
2346 if (tb[RTA_TABLE])
2347 cfg->fc_table = nla_get_u32(tb[RTA_TABLE]);
2348
2349 err = 0;
2350errout:
2351 return err;
2352}
2353
2354static int inet6_rtm_delroute(struct sk_buff *skb, struct nlmsghdr* nlh, void *arg)
2355{
2356 struct fib6_config cfg;
2357 int err;
2358
2359 err = rtm_to_fib6_config(skb, nlh, &cfg);
2360 if (err < 0)
2361 return err;
2362
2363 return ip6_route_del(&cfg);
2364}
2365
2366static int inet6_rtm_newroute(struct sk_buff *skb, struct nlmsghdr* nlh, void *arg)
2367{
2368 struct fib6_config cfg;
2369 int err;
2370
2371 err = rtm_to_fib6_config(skb, nlh, &cfg);
2372 if (err < 0)
2373 return err;
2374
2375 return ip6_route_add(&cfg);
2376}
2377
2378static inline size_t rt6_nlmsg_size(void)
2379{
2380 return NLMSG_ALIGN(sizeof(struct rtmsg))
2381 + nla_total_size(16) /* RTA_SRC */
2382 + nla_total_size(16) /* RTA_DST */
2383 + nla_total_size(16) /* RTA_GATEWAY */
2384 + nla_total_size(16) /* RTA_PREFSRC */
2385 + nla_total_size(4) /* RTA_TABLE */
2386 + nla_total_size(4) /* RTA_IIF */
2387 + nla_total_size(4) /* RTA_OIF */
2388 + nla_total_size(4) /* RTA_PRIORITY */
2389 + RTAX_MAX * nla_total_size(4) /* RTA_METRICS */
2390 + nla_total_size(sizeof(struct rta_cacheinfo));
2391}
2392
2393static int rt6_fill_node(struct net *net,
2394 struct sk_buff *skb, struct rt6_info *rt,
2395 struct in6_addr *dst, struct in6_addr *src,
2396 int iif, int type, u32 pid, u32 seq,
2397 int prefix, int nowait, unsigned int flags)
2398{
2399 const struct inet_peer *peer;
2400 struct rtmsg *rtm;
2401 struct nlmsghdr *nlh;
2402 long expires;
2403 u32 table;
2404 struct neighbour *n;
2405 u32 ts, tsage;
2406
2407 if (prefix) { /* user wants prefix routes only */
2408 if (!(rt->rt6i_flags & RTF_PREFIX_RT)) {
2409 /* success since this is not a prefix route */
2410 return 1;
2411 }
2412 }
2413
2414 nlh = nlmsg_put(skb, pid, seq, type, sizeof(*rtm), flags);
2415 if (!nlh)
2416 return -EMSGSIZE;
2417
2418 rtm = nlmsg_data(nlh);
2419 rtm->rtm_family = AF_INET6;
2420 rtm->rtm_dst_len = rt->rt6i_dst.plen;
2421 rtm->rtm_src_len = rt->rt6i_src.plen;
2422 rtm->rtm_tos = 0;
2423 if (rt->rt6i_table)
2424 table = rt->rt6i_table->tb6_id;
2425 else
2426 table = RT6_TABLE_UNSPEC;
2427 rtm->rtm_table = table;
2428 if (nla_put_u32(skb, RTA_TABLE, table))
2429 goto nla_put_failure;
2430 if (rt->rt6i_flags & RTF_REJECT)
2431 rtm->rtm_type = RTN_UNREACHABLE;
2432 else if (rt->rt6i_flags & RTF_LOCAL)
2433 rtm->rtm_type = RTN_LOCAL;
2434 else if (rt->dst.dev && (rt->dst.dev->flags & IFF_LOOPBACK))
2435 rtm->rtm_type = RTN_LOCAL;
2436 else
2437 rtm->rtm_type = RTN_UNICAST;
2438 rtm->rtm_flags = 0;
2439 rtm->rtm_scope = RT_SCOPE_UNIVERSE;
2440 rtm->rtm_protocol = rt->rt6i_protocol;
2441 if (rt->rt6i_flags & RTF_DYNAMIC)
2442 rtm->rtm_protocol = RTPROT_REDIRECT;
2443 else if (rt->rt6i_flags & RTF_ADDRCONF)
2444 rtm->rtm_protocol = RTPROT_KERNEL;
2445 else if (rt->rt6i_flags & RTF_DEFAULT)
2446 rtm->rtm_protocol = RTPROT_RA;
2447
2448 if (rt->rt6i_flags & RTF_CACHE)
2449 rtm->rtm_flags |= RTM_F_CLONED;
2450
2451 if (dst) {
2452 if (nla_put(skb, RTA_DST, 16, dst))
2453 goto nla_put_failure;
2454 rtm->rtm_dst_len = 128;
2455 } else if (rtm->rtm_dst_len)
2456 if (nla_put(skb, RTA_DST, 16, &rt->rt6i_dst.addr))
2457 goto nla_put_failure;
2458#ifdef CONFIG_IPV6_SUBTREES
2459 if (src) {
2460 if (nla_put(skb, RTA_SRC, 16, src))
2461 goto nla_put_failure;
2462 rtm->rtm_src_len = 128;
2463 } else if (rtm->rtm_src_len &&
2464 nla_put(skb, RTA_SRC, 16, &rt->rt6i_src.addr))
2465 goto nla_put_failure;
2466#endif
2467 if (iif) {
2468#ifdef CONFIG_IPV6_MROUTE
2469 if (ipv6_addr_is_multicast(&rt->rt6i_dst.addr)) {
2470 int err = ip6mr_get_route(net, skb, rtm, nowait);
2471 if (err <= 0) {
2472 if (!nowait) {
2473 if (err == 0)
2474 return 0;
2475 goto nla_put_failure;
2476 } else {
2477 if (err == -EMSGSIZE)
2478 goto nla_put_failure;
2479 }
2480 }
2481 } else
2482#endif
2483 if (nla_put_u32(skb, RTA_IIF, iif))
2484 goto nla_put_failure;
2485 } else if (dst) {
2486 struct in6_addr saddr_buf;
2487 if (ip6_route_get_saddr(net, rt, dst, 0, &saddr_buf) == 0 &&
2488 nla_put(skb, RTA_PREFSRC, 16, &saddr_buf))
2489 goto nla_put_failure;
2490 }
2491
2492 if (rt->rt6i_prefsrc.plen) {
2493 struct in6_addr saddr_buf;
2494 saddr_buf = rt->rt6i_prefsrc.addr;
2495 if (nla_put(skb, RTA_PREFSRC, 16, &saddr_buf))
2496 goto nla_put_failure;
2497 }
2498
2499 if (rtnetlink_put_metrics(skb, dst_metrics_ptr(&rt->dst)) < 0)
2500 goto nla_put_failure;
2501
2502 rcu_read_lock();
2503 n = dst_get_neighbour_noref(&rt->dst);
2504 if (n) {
2505 if (nla_put(skb, RTA_GATEWAY, 16, &n->primary_key) < 0) {
2506 rcu_read_unlock();
2507 goto nla_put_failure;
2508 }
2509 }
2510 rcu_read_unlock();
2511
2512 if (rt->dst.dev &&
2513 nla_put_u32(skb, RTA_OIF, rt->dst.dev->ifindex))
2514 goto nla_put_failure;
2515 if (nla_put_u32(skb, RTA_PRIORITY, rt->rt6i_metric))
2516 goto nla_put_failure;
2517 if (!(rt->rt6i_flags & RTF_EXPIRES))
2518 expires = 0;
2519 else if (rt->dst.expires - jiffies < INT_MAX)
2520 expires = rt->dst.expires - jiffies;
2521 else
2522 expires = INT_MAX;
2523
2524 peer = rt->rt6i_peer;
2525 ts = tsage = 0;
2526 if (peer && peer->tcp_ts_stamp) {
2527 ts = peer->tcp_ts;
2528 tsage = get_seconds() - peer->tcp_ts_stamp;
2529 }
2530
2531 if (rtnl_put_cacheinfo(skb, &rt->dst, 0, ts, tsage,
2532 expires, rt->dst.error) < 0)
2533 goto nla_put_failure;
2534
2535 return nlmsg_end(skb, nlh);
2536
2537nla_put_failure:
2538 nlmsg_cancel(skb, nlh);
2539 return -EMSGSIZE;
2540}
2541
2542int rt6_dump_route(struct rt6_info *rt, void *p_arg)
2543{
2544 struct rt6_rtnl_dump_arg *arg = (struct rt6_rtnl_dump_arg *) p_arg;
2545 int prefix;
2546
2547 if (nlmsg_len(arg->cb->nlh) >= sizeof(struct rtmsg)) {
2548 struct rtmsg *rtm = nlmsg_data(arg->cb->nlh);
2549 prefix = (rtm->rtm_flags & RTM_F_PREFIX) != 0;
2550 } else
2551 prefix = 0;
2552
2553 return rt6_fill_node(arg->net,
2554 arg->skb, rt, NULL, NULL, 0, RTM_NEWROUTE,
2555 NETLINK_CB(arg->cb->skb).pid, arg->cb->nlh->nlmsg_seq,
2556 prefix, 0, NLM_F_MULTI);
2557}
2558
2559static int inet6_rtm_getroute(struct sk_buff *in_skb, struct nlmsghdr* nlh, void *arg)
2560{
2561 struct net *net = sock_net(in_skb->sk);
2562 struct nlattr *tb[RTA_MAX+1];
2563 struct rt6_info *rt;
2564 struct sk_buff *skb;
2565 struct rtmsg *rtm;
2566 struct flowi6 fl6;
2567 int err, iif = 0, oif = 0;
2568
2569 err = nlmsg_parse(nlh, sizeof(*rtm), tb, RTA_MAX, rtm_ipv6_policy);
2570 if (err < 0)
2571 goto errout;
2572
2573 err = -EINVAL;
2574 memset(&fl6, 0, sizeof(fl6));
2575
2576 if (tb[RTA_SRC]) {
2577 if (nla_len(tb[RTA_SRC]) < sizeof(struct in6_addr))
2578 goto errout;
2579
2580 fl6.saddr = *(struct in6_addr *)nla_data(tb[RTA_SRC]);
2581 }
2582
2583 if (tb[RTA_DST]) {
2584 if (nla_len(tb[RTA_DST]) < sizeof(struct in6_addr))
2585 goto errout;
2586
2587 fl6.daddr = *(struct in6_addr *)nla_data(tb[RTA_DST]);
2588 }
2589
2590 if (tb[RTA_IIF])
2591 iif = nla_get_u32(tb[RTA_IIF]);
2592
2593 if (tb[RTA_OIF])
2594 oif = nla_get_u32(tb[RTA_OIF]);
2595
2596 if (iif) {
2597 struct net_device *dev;
2598 int flags = 0;
2599
2600 dev = __dev_get_by_index(net, iif);
2601 if (!dev) {
2602 err = -ENODEV;
2603 goto errout;
2604 }
2605
2606 fl6.flowi6_iif = iif;
2607
2608 if (!ipv6_addr_any(&fl6.saddr))
2609 flags |= RT6_LOOKUP_F_HAS_SADDR;
2610
2611 rt = (struct rt6_info *)ip6_route_input_lookup(net, dev, &fl6,
2612 flags);
2613 } else {
2614 fl6.flowi6_oif = oif;
2615
2616 rt = (struct rt6_info *)ip6_route_output(net, NULL, &fl6);
2617 }
2618
2619 skb = alloc_skb(NLMSG_GOODSIZE, GFP_KERNEL);
2620 if (!skb) {
2621 dst_release(&rt->dst);
2622 err = -ENOBUFS;
2623 goto errout;
2624 }
2625
2626 /* Reserve room for dummy headers, this skb can pass
2627 through good chunk of routing engine.
2628 */
2629 skb_reset_mac_header(skb);
2630 skb_reserve(skb, MAX_HEADER + sizeof(struct ipv6hdr));
2631
2632 skb_dst_set(skb, &rt->dst);
2633
2634 err = rt6_fill_node(net, skb, rt, &fl6.daddr, &fl6.saddr, iif,
2635 RTM_NEWROUTE, NETLINK_CB(in_skb).pid,
2636 nlh->nlmsg_seq, 0, 0, 0);
2637 if (err < 0) {
2638 kfree_skb(skb);
2639 goto errout;
2640 }
2641
2642 err = rtnl_unicast(skb, net, NETLINK_CB(in_skb).pid);
2643errout:
2644 return err;
2645}
2646
2647void inet6_rt_notify(int event, struct rt6_info *rt, struct nl_info *info)
2648{
2649 struct sk_buff *skb;
2650 struct net *net = info->nl_net;
2651 u32 seq;
2652 int err;
2653
2654 err = -ENOBUFS;
2655 seq = info->nlh ? info->nlh->nlmsg_seq : 0;
2656
2657 skb = nlmsg_new(rt6_nlmsg_size(), gfp_any());
2658 if (!skb)
2659 goto errout;
2660
2661 err = rt6_fill_node(net, skb, rt, NULL, NULL, 0,
2662 event, info->pid, seq, 0, 0, 0);
2663 if (err < 0) {
2664 /* -EMSGSIZE implies BUG in rt6_nlmsg_size() */
2665 WARN_ON(err == -EMSGSIZE);
2666 kfree_skb(skb);
2667 goto errout;
2668 }
2669 rtnl_notify(skb, net, info->pid, RTNLGRP_IPV6_ROUTE,
2670 info->nlh, gfp_any());
2671 return;
2672errout:
2673 if (err < 0)
2674 rtnl_set_sk_err(net, RTNLGRP_IPV6_ROUTE, err);
2675}
2676
2677static int ip6_route_dev_notify(struct notifier_block *this,
2678 unsigned long event, void *data)
2679{
2680 struct net_device *dev = (struct net_device *)data;
2681 struct net *net = dev_net(dev);
2682
2683 if (event == NETDEV_REGISTER && (dev->flags & IFF_LOOPBACK)) {
2684 net->ipv6.ip6_null_entry->dst.dev = dev;
2685 net->ipv6.ip6_null_entry->rt6i_idev = in6_dev_get(dev);
2686#ifdef CONFIG_IPV6_MULTIPLE_TABLES
2687 net->ipv6.ip6_prohibit_entry->dst.dev = dev;
2688 net->ipv6.ip6_prohibit_entry->rt6i_idev = in6_dev_get(dev);
2689 net->ipv6.ip6_blk_hole_entry->dst.dev = dev;
2690 net->ipv6.ip6_blk_hole_entry->rt6i_idev = in6_dev_get(dev);
2691#endif
2692 }
2693
2694 return NOTIFY_OK;
2695}
2696
2697/*
2698 * /proc
2699 */
2700
2701#ifdef CONFIG_PROC_FS
2702
2703struct rt6_proc_arg
2704{
2705 char *buffer;
2706 int offset;
2707 int length;
2708 int skip;
2709 int len;
2710};
2711
2712static int rt6_info_route(struct rt6_info *rt, void *p_arg)
2713{
2714 struct seq_file *m = p_arg;
2715 struct neighbour *n;
2716
2717 seq_printf(m, "%pi6 %02x ", &rt->rt6i_dst.addr, rt->rt6i_dst.plen);
2718
2719#ifdef CONFIG_IPV6_SUBTREES
2720 seq_printf(m, "%pi6 %02x ", &rt->rt6i_src.addr, rt->rt6i_src.plen);
2721#else
2722 seq_puts(m, "00000000000000000000000000000000 00 ");
2723#endif
2724 rcu_read_lock();
2725 n = dst_get_neighbour_noref(&rt->dst);
2726 if (n) {
2727 seq_printf(m, "%pi6", n->primary_key);
2728 } else {
2729 seq_puts(m, "00000000000000000000000000000000");
2730 }
2731 rcu_read_unlock();
2732 seq_printf(m, " %08x %08x %08x %08x %8s\n",
2733 rt->rt6i_metric, atomic_read(&rt->dst.__refcnt),
2734 rt->dst.__use, rt->rt6i_flags,
2735 rt->dst.dev ? rt->dst.dev->name : "");
2736 return 0;
2737}
2738
2739static int ipv6_route_show(struct seq_file *m, void *v)
2740{
2741 struct net *net = (struct net *)m->private;
2742 fib6_clean_all_ro(net, rt6_info_route, 0, m);
2743 return 0;
2744}
2745
2746static int ipv6_route_open(struct inode *inode, struct file *file)
2747{
2748 return single_open_net(inode, file, ipv6_route_show);
2749}
2750
2751static const struct file_operations ipv6_route_proc_fops = {
2752 .owner = THIS_MODULE,
2753 .open = ipv6_route_open,
2754 .read = seq_read,
2755 .llseek = seq_lseek,
2756 .release = single_release_net,
2757};
2758
2759static int rt6_stats_seq_show(struct seq_file *seq, void *v)
2760{
2761 struct net *net = (struct net *)seq->private;
2762 seq_printf(seq, "%04x %04x %04x %04x %04x %04x %04x\n",
2763 net->ipv6.rt6_stats->fib_nodes,
2764 net->ipv6.rt6_stats->fib_route_nodes,
2765 net->ipv6.rt6_stats->fib_rt_alloc,
2766 net->ipv6.rt6_stats->fib_rt_entries,
2767 net->ipv6.rt6_stats->fib_rt_cache,
2768 dst_entries_get_slow(&net->ipv6.ip6_dst_ops),
2769 net->ipv6.rt6_stats->fib_discarded_routes);
2770
2771 return 0;
2772}
2773
2774static int rt6_stats_seq_open(struct inode *inode, struct file *file)
2775{
2776 return single_open_net(inode, file, rt6_stats_seq_show);
2777}
2778
2779static const struct file_operations rt6_stats_seq_fops = {
2780 .owner = THIS_MODULE,
2781 .open = rt6_stats_seq_open,
2782 .read = seq_read,
2783 .llseek = seq_lseek,
2784 .release = single_release_net,
2785};
2786#endif /* CONFIG_PROC_FS */
2787
2788#ifdef CONFIG_SYSCTL
2789
2790static
2791int ipv6_sysctl_rtcache_flush(ctl_table *ctl, int write,
2792 void __user *buffer, size_t *lenp, loff_t *ppos)
2793{
2794 struct net *net;
2795 int delay;
2796 if (!write)
2797 return -EINVAL;
2798
2799 net = (struct net *)ctl->extra1;
2800 delay = net->ipv6.sysctl.flush_delay;
2801 proc_dointvec(ctl, write, buffer, lenp, ppos);
2802 fib6_run_gc(delay <= 0 ? ~0UL : (unsigned long)delay, net);
2803 return 0;
2804}
2805
2806ctl_table ipv6_route_table_template[] = {
2807 {
2808 .procname = "flush",
2809 .data = &init_net.ipv6.sysctl.flush_delay,
2810 .maxlen = sizeof(int),
2811 .mode = 0200,
2812 .proc_handler = ipv6_sysctl_rtcache_flush
2813 },
2814 {
2815 .procname = "gc_thresh",
2816 .data = &ip6_dst_ops_template.gc_thresh,
2817 .maxlen = sizeof(int),
2818 .mode = 0644,
2819 .proc_handler = proc_dointvec,
2820 },
2821 {
2822 .procname = "max_size",
2823 .data = &init_net.ipv6.sysctl.ip6_rt_max_size,
2824 .maxlen = sizeof(int),
2825 .mode = 0644,
2826 .proc_handler = proc_dointvec,
2827 },
2828 {
2829 .procname = "gc_min_interval",
2830 .data = &init_net.ipv6.sysctl.ip6_rt_gc_min_interval,
2831 .maxlen = sizeof(int),
2832 .mode = 0644,
2833 .proc_handler = proc_dointvec_jiffies,
2834 },
2835 {
2836 .procname = "gc_timeout",
2837 .data = &init_net.ipv6.sysctl.ip6_rt_gc_timeout,
2838 .maxlen = sizeof(int),
2839 .mode = 0644,
2840 .proc_handler = proc_dointvec_jiffies,
2841 },
2842 {
2843 .procname = "gc_interval",
2844 .data = &init_net.ipv6.sysctl.ip6_rt_gc_interval,
2845 .maxlen = sizeof(int),
2846 .mode = 0644,
2847 .proc_handler = proc_dointvec_jiffies,
2848 },
2849 {
2850 .procname = "gc_elasticity",
2851 .data = &init_net.ipv6.sysctl.ip6_rt_gc_elasticity,
2852 .maxlen = sizeof(int),
2853 .mode = 0644,
2854 .proc_handler = proc_dointvec,
2855 },
2856 {
2857 .procname = "mtu_expires",
2858 .data = &init_net.ipv6.sysctl.ip6_rt_mtu_expires,
2859 .maxlen = sizeof(int),
2860 .mode = 0644,
2861 .proc_handler = proc_dointvec_jiffies,
2862 },
2863 {
2864 .procname = "min_adv_mss",
2865 .data = &init_net.ipv6.sysctl.ip6_rt_min_advmss,
2866 .maxlen = sizeof(int),
2867 .mode = 0644,
2868 .proc_handler = proc_dointvec,
2869 },
2870 {
2871 .procname = "gc_min_interval_ms",
2872 .data = &init_net.ipv6.sysctl.ip6_rt_gc_min_interval,
2873 .maxlen = sizeof(int),
2874 .mode = 0644,
2875 .proc_handler = proc_dointvec_ms_jiffies,
2876 },
2877 { }
2878};
2879
2880struct ctl_table * __net_init ipv6_route_sysctl_init(struct net *net)
2881{
2882 struct ctl_table *table;
2883
2884 table = kmemdup(ipv6_route_table_template,
2885 sizeof(ipv6_route_table_template),
2886 GFP_KERNEL);
2887
2888 if (table) {
2889 table[0].data = &net->ipv6.sysctl.flush_delay;
2890 table[0].extra1 = net;
2891 table[1].data = &net->ipv6.ip6_dst_ops.gc_thresh;
2892 table[2].data = &net->ipv6.sysctl.ip6_rt_max_size;
2893 table[3].data = &net->ipv6.sysctl.ip6_rt_gc_min_interval;
2894 table[4].data = &net->ipv6.sysctl.ip6_rt_gc_timeout;
2895 table[5].data = &net->ipv6.sysctl.ip6_rt_gc_interval;
2896 table[6].data = &net->ipv6.sysctl.ip6_rt_gc_elasticity;
2897 table[7].data = &net->ipv6.sysctl.ip6_rt_mtu_expires;
2898 table[8].data = &net->ipv6.sysctl.ip6_rt_min_advmss;
2899 table[9].data = &net->ipv6.sysctl.ip6_rt_gc_min_interval;
2900 }
2901
2902 return table;
2903}
2904#endif
2905
2906static int __net_init ip6_route_net_init(struct net *net)
2907{
2908 int ret = -ENOMEM;
2909
2910 memcpy(&net->ipv6.ip6_dst_ops, &ip6_dst_ops_template,
2911 sizeof(net->ipv6.ip6_dst_ops));
2912
2913 if (dst_entries_init(&net->ipv6.ip6_dst_ops) < 0)
2914 goto out_ip6_dst_ops;
2915
2916 net->ipv6.ip6_null_entry = kmemdup(&ip6_null_entry_template,
2917 sizeof(*net->ipv6.ip6_null_entry),
2918 GFP_KERNEL);
2919 if (!net->ipv6.ip6_null_entry)
2920 goto out_ip6_dst_entries;
2921 net->ipv6.ip6_null_entry->dst.path =
2922 (struct dst_entry *)net->ipv6.ip6_null_entry;
2923 net->ipv6.ip6_null_entry->dst.ops = &net->ipv6.ip6_dst_ops;
2924 dst_init_metrics(&net->ipv6.ip6_null_entry->dst,
2925 ip6_template_metrics, true);
2926
2927#ifdef CONFIG_IPV6_MULTIPLE_TABLES
2928 net->ipv6.ip6_prohibit_entry = kmemdup(&ip6_prohibit_entry_template,
2929 sizeof(*net->ipv6.ip6_prohibit_entry),
2930 GFP_KERNEL);
2931 if (!net->ipv6.ip6_prohibit_entry)
2932 goto out_ip6_null_entry;
2933 net->ipv6.ip6_prohibit_entry->dst.path =
2934 (struct dst_entry *)net->ipv6.ip6_prohibit_entry;
2935 net->ipv6.ip6_prohibit_entry->dst.ops = &net->ipv6.ip6_dst_ops;
2936 dst_init_metrics(&net->ipv6.ip6_prohibit_entry->dst,
2937 ip6_template_metrics, true);
2938
2939 net->ipv6.ip6_blk_hole_entry = kmemdup(&ip6_blk_hole_entry_template,
2940 sizeof(*net->ipv6.ip6_blk_hole_entry),
2941 GFP_KERNEL);
2942 if (!net->ipv6.ip6_blk_hole_entry)
2943 goto out_ip6_prohibit_entry;
2944 net->ipv6.ip6_blk_hole_entry->dst.path =
2945 (struct dst_entry *)net->ipv6.ip6_blk_hole_entry;
2946 net->ipv6.ip6_blk_hole_entry->dst.ops = &net->ipv6.ip6_dst_ops;
2947 dst_init_metrics(&net->ipv6.ip6_blk_hole_entry->dst,
2948 ip6_template_metrics, true);
2949#endif
2950
2951 net->ipv6.sysctl.flush_delay = 0;
2952 net->ipv6.sysctl.ip6_rt_max_size = 4096;
2953 net->ipv6.sysctl.ip6_rt_gc_min_interval = HZ / 2;
2954 net->ipv6.sysctl.ip6_rt_gc_timeout = 60*HZ;
2955 net->ipv6.sysctl.ip6_rt_gc_interval = 30*HZ;
2956 net->ipv6.sysctl.ip6_rt_gc_elasticity = 9;
2957 net->ipv6.sysctl.ip6_rt_mtu_expires = 10*60*HZ;
2958 net->ipv6.sysctl.ip6_rt_min_advmss = IPV6_MIN_MTU - 20 - 40;
2959
2960 net->ipv6.ip6_rt_gc_expire = 30*HZ;
2961
2962 ret = 0;
2963out:
2964 return ret;
2965
2966#ifdef CONFIG_IPV6_MULTIPLE_TABLES
2967out_ip6_prohibit_entry:
2968 kfree(net->ipv6.ip6_prohibit_entry);
2969out_ip6_null_entry:
2970 kfree(net->ipv6.ip6_null_entry);
2971#endif
2972out_ip6_dst_entries:
2973 dst_entries_destroy(&net->ipv6.ip6_dst_ops);
2974out_ip6_dst_ops:
2975 goto out;
2976}
2977
2978static void __net_exit ip6_route_net_exit(struct net *net)
2979{
2980 kfree(net->ipv6.ip6_null_entry);
2981#ifdef CONFIG_IPV6_MULTIPLE_TABLES
2982 kfree(net->ipv6.ip6_prohibit_entry);
2983 kfree(net->ipv6.ip6_blk_hole_entry);
2984#endif
2985 dst_entries_destroy(&net->ipv6.ip6_dst_ops);
2986}
2987
2988static int __net_init ip6_route_net_init_late(struct net *net)
2989{
2990#ifdef CONFIG_PROC_FS
2991 proc_net_fops_create(net, "ipv6_route", 0, &ipv6_route_proc_fops);
2992 proc_net_fops_create(net, "rt6_stats", S_IRUGO, &rt6_stats_seq_fops);
2993#endif
2994 return 0;
2995}
2996
2997static void __net_exit ip6_route_net_exit_late(struct net *net)
2998{
2999#ifdef CONFIG_PROC_FS
3000 proc_net_remove(net, "ipv6_route");
3001 proc_net_remove(net, "rt6_stats");
3002#endif
3003}
3004
3005static struct pernet_operations ip6_route_net_ops = {
3006 .init = ip6_route_net_init,
3007 .exit = ip6_route_net_exit,
3008};
3009
3010static struct pernet_operations ip6_route_net_late_ops = {
3011 .init = ip6_route_net_init_late,
3012 .exit = ip6_route_net_exit_late,
3013};
3014
3015static struct notifier_block ip6_route_dev_notifier = {
3016 .notifier_call = ip6_route_dev_notify,
3017 .priority = 0,
3018};
3019
3020int __init ip6_route_init(void)
3021{
3022 int ret;
3023
3024 ret = -ENOMEM;
3025 ip6_dst_ops_template.kmem_cachep =
3026 kmem_cache_create("ip6_dst_cache", sizeof(struct rt6_info), 0,
3027 SLAB_HWCACHE_ALIGN, NULL);
3028 if (!ip6_dst_ops_template.kmem_cachep)
3029 goto out;
3030
3031 ret = dst_entries_init(&ip6_dst_blackhole_ops);
3032 if (ret)
3033 goto out_kmem_cache;
3034
3035 ret = register_pernet_subsys(&ip6_route_net_ops);
3036 if (ret)
3037 goto out_dst_entries;
3038
3039 ip6_dst_blackhole_ops.kmem_cachep = ip6_dst_ops_template.kmem_cachep;
3040
3041 /* Registering of the loopback is done before this portion of code,
3042 * the loopback reference in rt6_info will not be taken, do it
3043 * manually for init_net */
3044 init_net.ipv6.ip6_null_entry->dst.dev = init_net.loopback_dev;
3045 init_net.ipv6.ip6_null_entry->rt6i_idev = in6_dev_get(init_net.loopback_dev);
3046 #ifdef CONFIG_IPV6_MULTIPLE_TABLES
3047 init_net.ipv6.ip6_prohibit_entry->dst.dev = init_net.loopback_dev;
3048 init_net.ipv6.ip6_prohibit_entry->rt6i_idev = in6_dev_get(init_net.loopback_dev);
3049 init_net.ipv6.ip6_blk_hole_entry->dst.dev = init_net.loopback_dev;
3050 init_net.ipv6.ip6_blk_hole_entry->rt6i_idev = in6_dev_get(init_net.loopback_dev);
3051 #endif
3052 ret = fib6_init();
3053 if (ret)
3054 goto out_register_subsys;
3055
3056 ret = xfrm6_init();
3057 if (ret)
3058 goto out_fib6_init;
3059
3060 ret = fib6_rules_init();
3061 if (ret)
3062 goto xfrm6_init;
3063
3064 ret = register_pernet_subsys(&ip6_route_net_late_ops);
3065 if (ret)
3066 goto fib6_rules_init;
3067
3068 ret = -ENOBUFS;
3069 if (__rtnl_register(PF_INET6, RTM_NEWROUTE, inet6_rtm_newroute, NULL, NULL) ||
3070 __rtnl_register(PF_INET6, RTM_DELROUTE, inet6_rtm_delroute, NULL, NULL) ||
3071 __rtnl_register(PF_INET6, RTM_GETROUTE, inet6_rtm_getroute, NULL, NULL))
3072 goto out_register_late_subsys;
3073
3074 ret = register_netdevice_notifier(&ip6_route_dev_notifier);
3075 if (ret)
3076 goto out_register_late_subsys;
3077
3078out:
3079 return ret;
3080
3081out_register_late_subsys:
3082 unregister_pernet_subsys(&ip6_route_net_late_ops);
3083fib6_rules_init:
3084 fib6_rules_cleanup();
3085xfrm6_init:
3086 xfrm6_fini();
3087out_fib6_init:
3088 fib6_gc_cleanup();
3089out_register_subsys:
3090 unregister_pernet_subsys(&ip6_route_net_ops);
3091out_dst_entries:
3092 dst_entries_destroy(&ip6_dst_blackhole_ops);
3093out_kmem_cache:
3094 kmem_cache_destroy(ip6_dst_ops_template.kmem_cachep);
3095 goto out;
3096}
3097
3098void ip6_route_cleanup(void)
3099{
3100 unregister_netdevice_notifier(&ip6_route_dev_notifier);
3101 unregister_pernet_subsys(&ip6_route_net_late_ops);
3102 fib6_rules_cleanup();
3103 xfrm6_fini();
3104 fib6_gc_cleanup();
3105 unregister_pernet_subsys(&ip6_route_net_ops);
3106 dst_entries_destroy(&ip6_dst_blackhole_ops);
3107 kmem_cache_destroy(ip6_dst_ops_template.kmem_cachep);
3108}