Linux Audio

Check our new training course

Loading...
v4.6
 
   1/*
   2 *	NET3	IP device support routines.
   3 *
   4 *		This program is free software; you can redistribute it and/or
   5 *		modify it under the terms of the GNU General Public License
   6 *		as published by the Free Software Foundation; either version
   7 *		2 of the License, or (at your option) any later version.
   8 *
   9 *	Derived from the IP parts of dev.c 1.0.19
  10 * 		Authors:	Ross Biro
  11 *				Fred N. van Kempen, <waltje@uWalt.NL.Mugnet.ORG>
  12 *				Mark Evans, <evansmp@uhura.aston.ac.uk>
  13 *
  14 *	Additional Authors:
  15 *		Alan Cox, <gw4pts@gw4pts.ampr.org>
  16 *		Alexey Kuznetsov, <kuznet@ms2.inr.ac.ru>
  17 *
  18 *	Changes:
  19 *		Alexey Kuznetsov:	pa_* fields are replaced with ifaddr
  20 *					lists.
  21 *		Cyrus Durgin:		updated for kmod
  22 *		Matthias Andree:	in devinet_ioctl, compare label and
  23 *					address (4.4BSD alias style support),
  24 *					fall back to comparing just the label
  25 *					if no match found.
  26 */
  27
  28
  29#include <asm/uaccess.h>
  30#include <linux/bitops.h>
  31#include <linux/capability.h>
  32#include <linux/module.h>
  33#include <linux/types.h>
  34#include <linux/kernel.h>
 
  35#include <linux/string.h>
  36#include <linux/mm.h>
  37#include <linux/socket.h>
  38#include <linux/sockios.h>
  39#include <linux/in.h>
  40#include <linux/errno.h>
  41#include <linux/interrupt.h>
  42#include <linux/if_addr.h>
  43#include <linux/if_ether.h>
  44#include <linux/inet.h>
  45#include <linux/netdevice.h>
  46#include <linux/etherdevice.h>
  47#include <linux/skbuff.h>
  48#include <linux/init.h>
  49#include <linux/notifier.h>
  50#include <linux/inetdevice.h>
  51#include <linux/igmp.h>
  52#include <linux/slab.h>
  53#include <linux/hash.h>
  54#ifdef CONFIG_SYSCTL
  55#include <linux/sysctl.h>
  56#endif
  57#include <linux/kmod.h>
  58#include <linux/netconf.h>
  59
  60#include <net/arp.h>
  61#include <net/ip.h>
  62#include <net/route.h>
  63#include <net/ip_fib.h>
  64#include <net/rtnetlink.h>
  65#include <net/net_namespace.h>
  66#include <net/addrconf.h>
  67
  68#include "fib_lookup.h"
 
 
 
  69
  70static struct ipv4_devconf ipv4_devconf = {
  71	.data = {
  72		[IPV4_DEVCONF_ACCEPT_REDIRECTS - 1] = 1,
  73		[IPV4_DEVCONF_SEND_REDIRECTS - 1] = 1,
  74		[IPV4_DEVCONF_SECURE_REDIRECTS - 1] = 1,
  75		[IPV4_DEVCONF_SHARED_MEDIA - 1] = 1,
  76		[IPV4_DEVCONF_IGMPV2_UNSOLICITED_REPORT_INTERVAL - 1] = 10000 /*ms*/,
  77		[IPV4_DEVCONF_IGMPV3_UNSOLICITED_REPORT_INTERVAL - 1] =  1000 /*ms*/,
  78	},
  79};
  80
  81static struct ipv4_devconf ipv4_devconf_dflt = {
  82	.data = {
  83		[IPV4_DEVCONF_ACCEPT_REDIRECTS - 1] = 1,
  84		[IPV4_DEVCONF_SEND_REDIRECTS - 1] = 1,
  85		[IPV4_DEVCONF_SECURE_REDIRECTS - 1] = 1,
  86		[IPV4_DEVCONF_SHARED_MEDIA - 1] = 1,
  87		[IPV4_DEVCONF_ACCEPT_SOURCE_ROUTE - 1] = 1,
  88		[IPV4_DEVCONF_IGMPV2_UNSOLICITED_REPORT_INTERVAL - 1] = 10000 /*ms*/,
  89		[IPV4_DEVCONF_IGMPV3_UNSOLICITED_REPORT_INTERVAL - 1] =  1000 /*ms*/,
  90	},
  91};
  92
  93#define IPV4_DEVCONF_DFLT(net, attr) \
  94	IPV4_DEVCONF((*net->ipv4.devconf_dflt), attr)
  95
  96static const struct nla_policy ifa_ipv4_policy[IFA_MAX+1] = {
  97	[IFA_LOCAL]     	= { .type = NLA_U32 },
  98	[IFA_ADDRESS]   	= { .type = NLA_U32 },
  99	[IFA_BROADCAST] 	= { .type = NLA_U32 },
 100	[IFA_LABEL]     	= { .type = NLA_STRING, .len = IFNAMSIZ - 1 },
 101	[IFA_CACHEINFO]		= { .len = sizeof(struct ifa_cacheinfo) },
 102	[IFA_FLAGS]		= { .type = NLA_U32 },
 
 
 
 
 
 
 
 
 
 
 
 103};
 104
 105#define IN4_ADDR_HSIZE_SHIFT	8
 106#define IN4_ADDR_HSIZE		(1U << IN4_ADDR_HSIZE_SHIFT)
 107
 108static struct hlist_head inet_addr_lst[IN4_ADDR_HSIZE];
 109
 110static u32 inet_addr_hash(const struct net *net, __be32 addr)
 111{
 112	u32 val = (__force u32) addr ^ net_hash_mix(net);
 113
 114	return hash_32(val, IN4_ADDR_HSIZE_SHIFT);
 115}
 116
 117static void inet_hash_insert(struct net *net, struct in_ifaddr *ifa)
 118{
 119	u32 hash = inet_addr_hash(net, ifa->ifa_local);
 120
 121	ASSERT_RTNL();
 122	hlist_add_head_rcu(&ifa->hash, &inet_addr_lst[hash]);
 123}
 124
 125static void inet_hash_remove(struct in_ifaddr *ifa)
 126{
 127	ASSERT_RTNL();
 128	hlist_del_init_rcu(&ifa->hash);
 129}
 130
 131/**
 132 * __ip_dev_find - find the first device with a given source address.
 133 * @net: the net namespace
 134 * @addr: the source address
 135 * @devref: if true, take a reference on the found device
 136 *
 137 * If a caller uses devref=false, it should be protected by RCU, or RTNL
 138 */
 139struct net_device *__ip_dev_find(struct net *net, __be32 addr, bool devref)
 140{
 141	u32 hash = inet_addr_hash(net, addr);
 142	struct net_device *result = NULL;
 143	struct in_ifaddr *ifa;
 144
 145	rcu_read_lock();
 146	hlist_for_each_entry_rcu(ifa, &inet_addr_lst[hash], hash) {
 147		if (ifa->ifa_local == addr) {
 148			struct net_device *dev = ifa->ifa_dev->dev;
 149
 150			if (!net_eq(dev_net(dev), net))
 151				continue;
 152			result = dev;
 153			break;
 154		}
 155	}
 156	if (!result) {
 157		struct flowi4 fl4 = { .daddr = addr };
 158		struct fib_result res = { 0 };
 159		struct fib_table *local;
 160
 161		/* Fallback to FIB local table so that communication
 162		 * over loopback subnets work.
 163		 */
 164		local = fib_get_table(net, RT_TABLE_LOCAL);
 165		if (local &&
 166		    !fib_table_lookup(local, &fl4, &res, FIB_LOOKUP_NOREF) &&
 167		    res.type == RTN_LOCAL)
 168			result = FIB_RES_DEV(res);
 
 
 169	}
 170	if (result && devref)
 171		dev_hold(result);
 172	rcu_read_unlock();
 173	return result;
 174}
 175EXPORT_SYMBOL(__ip_dev_find);
 176
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 177static void rtmsg_ifa(int event, struct in_ifaddr *, struct nlmsghdr *, u32);
 178
 179static BLOCKING_NOTIFIER_HEAD(inetaddr_chain);
 180static void inet_del_ifa(struct in_device *in_dev, struct in_ifaddr **ifap,
 
 
 181			 int destroy);
 182#ifdef CONFIG_SYSCTL
 183static int devinet_sysctl_register(struct in_device *idev);
 184static void devinet_sysctl_unregister(struct in_device *idev);
 185#else
 186static int devinet_sysctl_register(struct in_device *idev)
 187{
 188	return 0;
 189}
 190static void devinet_sysctl_unregister(struct in_device *idev)
 191{
 192}
 193#endif
 194
 195/* Locks all the inet devices. */
 196
 197static struct in_ifaddr *inet_alloc_ifa(void)
 198{
 199	return kzalloc(sizeof(struct in_ifaddr), GFP_KERNEL);
 200}
 201
 202static void inet_rcu_free_ifa(struct rcu_head *head)
 203{
 204	struct in_ifaddr *ifa = container_of(head, struct in_ifaddr, rcu_head);
 205	if (ifa->ifa_dev)
 206		in_dev_put(ifa->ifa_dev);
 207	kfree(ifa);
 208}
 209
 210static void inet_free_ifa(struct in_ifaddr *ifa)
 211{
 212	call_rcu(&ifa->rcu_head, inet_rcu_free_ifa);
 213}
 214
 215void in_dev_finish_destroy(struct in_device *idev)
 216{
 217	struct net_device *dev = idev->dev;
 218
 219	WARN_ON(idev->ifa_list);
 220	WARN_ON(idev->mc_list);
 221	kfree(rcu_dereference_protected(idev->mc_hash, 1));
 222#ifdef NET_REFCNT_DEBUG
 223	pr_debug("%s: %p=%s\n", __func__, idev, dev ? dev->name : "NIL");
 224#endif
 225	dev_put(dev);
 226	if (!idev->dead)
 227		pr_err("Freeing alive in_device %p\n", idev);
 228	else
 229		kfree(idev);
 230}
 231EXPORT_SYMBOL(in_dev_finish_destroy);
 232
 233static struct in_device *inetdev_init(struct net_device *dev)
 234{
 235	struct in_device *in_dev;
 236	int err = -ENOMEM;
 237
 238	ASSERT_RTNL();
 239
 240	in_dev = kzalloc(sizeof(*in_dev), GFP_KERNEL);
 241	if (!in_dev)
 242		goto out;
 243	memcpy(&in_dev->cnf, dev_net(dev)->ipv4.devconf_dflt,
 244			sizeof(in_dev->cnf));
 245	in_dev->cnf.sysctl = NULL;
 246	in_dev->dev = dev;
 247	in_dev->arp_parms = neigh_parms_alloc(dev, &arp_tbl);
 248	if (!in_dev->arp_parms)
 249		goto out_kfree;
 250	if (IPV4_DEVCONF(in_dev->cnf, FORWARDING))
 251		dev_disable_lro(dev);
 252	/* Reference in_dev->dev */
 253	dev_hold(dev);
 254	/* Account for reference dev->ip_ptr (below) */
 255	in_dev_hold(in_dev);
 256
 257	err = devinet_sysctl_register(in_dev);
 258	if (err) {
 259		in_dev->dead = 1;
 
 260		in_dev_put(in_dev);
 261		in_dev = NULL;
 262		goto out;
 263	}
 264	ip_mc_init_dev(in_dev);
 265	if (dev->flags & IFF_UP)
 266		ip_mc_up(in_dev);
 267
 268	/* we can receive as soon as ip_ptr is set -- do this last */
 269	rcu_assign_pointer(dev->ip_ptr, in_dev);
 270out:
 271	return in_dev ?: ERR_PTR(err);
 272out_kfree:
 273	kfree(in_dev);
 274	in_dev = NULL;
 275	goto out;
 276}
 277
 278static void in_dev_rcu_put(struct rcu_head *head)
 279{
 280	struct in_device *idev = container_of(head, struct in_device, rcu_head);
 281	in_dev_put(idev);
 282}
 283
 284static void inetdev_destroy(struct in_device *in_dev)
 285{
 286	struct in_ifaddr *ifa;
 287	struct net_device *dev;
 
 288
 289	ASSERT_RTNL();
 290
 291	dev = in_dev->dev;
 292
 293	in_dev->dead = 1;
 294
 295	ip_mc_destroy_dev(in_dev);
 296
 297	while ((ifa = in_dev->ifa_list) != NULL) {
 298		inet_del_ifa(in_dev, &in_dev->ifa_list, 0);
 299		inet_free_ifa(ifa);
 300	}
 301
 302	RCU_INIT_POINTER(dev->ip_ptr, NULL);
 303
 304	devinet_sysctl_unregister(in_dev);
 305	neigh_parms_release(&arp_tbl, in_dev->arp_parms);
 306	arp_ifdown(dev);
 307
 308	call_rcu(&in_dev->rcu_head, in_dev_rcu_put);
 309}
 310
 311int inet_addr_onlink(struct in_device *in_dev, __be32 a, __be32 b)
 312{
 
 
 313	rcu_read_lock();
 314	for_primary_ifa(in_dev) {
 315		if (inet_ifa_match(a, ifa)) {
 316			if (!b || inet_ifa_match(b, ifa)) {
 317				rcu_read_unlock();
 318				return 1;
 319			}
 320		}
 321	} endfor_ifa(in_dev);
 322	rcu_read_unlock();
 323	return 0;
 324}
 325
 326static void __inet_del_ifa(struct in_device *in_dev, struct in_ifaddr **ifap,
 327			 int destroy, struct nlmsghdr *nlh, u32 portid)
 
 328{
 329	struct in_ifaddr *promote = NULL;
 330	struct in_ifaddr *ifa, *ifa1 = *ifap;
 331	struct in_ifaddr *last_prim = in_dev->ifa_list;
 332	struct in_ifaddr *prev_prom = NULL;
 333	int do_promote = IN_DEV_PROMOTE_SECONDARIES(in_dev);
 334
 335	ASSERT_RTNL();
 336
 
 
 337	if (in_dev->dead)
 338		goto no_promotions;
 339
 340	/* 1. Deleting primary ifaddr forces deletion all secondaries
 341	 * unless alias promotion is set
 342	 **/
 343
 344	if (!(ifa1->ifa_flags & IFA_F_SECONDARY)) {
 345		struct in_ifaddr **ifap1 = &ifa1->ifa_next;
 346
 347		while ((ifa = *ifap1) != NULL) {
 348			if (!(ifa->ifa_flags & IFA_F_SECONDARY) &&
 349			    ifa1->ifa_scope <= ifa->ifa_scope)
 350				last_prim = ifa;
 351
 352			if (!(ifa->ifa_flags & IFA_F_SECONDARY) ||
 353			    ifa1->ifa_mask != ifa->ifa_mask ||
 354			    !inet_ifa_match(ifa1->ifa_address, ifa)) {
 355				ifap1 = &ifa->ifa_next;
 356				prev_prom = ifa;
 357				continue;
 358			}
 359
 360			if (!do_promote) {
 361				inet_hash_remove(ifa);
 362				*ifap1 = ifa->ifa_next;
 363
 364				rtmsg_ifa(RTM_DELADDR, ifa, nlh, portid);
 365				blocking_notifier_call_chain(&inetaddr_chain,
 366						NETDEV_DOWN, ifa);
 367				inet_free_ifa(ifa);
 368			} else {
 369				promote = ifa;
 370				break;
 371			}
 372		}
 373	}
 374
 375	/* On promotion all secondaries from subnet are changing
 376	 * the primary IP, we must remove all their routes silently
 377	 * and later to add them back with new prefsrc. Do this
 378	 * while all addresses are on the device list.
 379	 */
 380	for (ifa = promote; ifa; ifa = ifa->ifa_next) {
 381		if (ifa1->ifa_mask == ifa->ifa_mask &&
 382		    inet_ifa_match(ifa1->ifa_address, ifa))
 383			fib_del_ifaddr(ifa, ifa1);
 384	}
 385
 386no_promotions:
 387	/* 2. Unlink it */
 388
 389	*ifap = ifa1->ifa_next;
 390	inet_hash_remove(ifa1);
 391
 392	/* 3. Announce address deletion */
 393
 394	/* Send message first, then call notifier.
 395	   At first sight, FIB update triggered by notifier
 396	   will refer to already deleted ifaddr, that could confuse
 397	   netlink listeners. It is not true: look, gated sees
 398	   that route deleted and if it still thinks that ifaddr
 399	   is valid, it will try to restore deleted routes... Grr.
 400	   So that, this order is correct.
 401	 */
 402	rtmsg_ifa(RTM_DELADDR, ifa1, nlh, portid);
 403	blocking_notifier_call_chain(&inetaddr_chain, NETDEV_DOWN, ifa1);
 404
 405	if (promote) {
 406		struct in_ifaddr *next_sec = promote->ifa_next;
 407
 
 408		if (prev_prom) {
 409			prev_prom->ifa_next = promote->ifa_next;
 410			promote->ifa_next = last_prim->ifa_next;
 411			last_prim->ifa_next = promote;
 
 
 
 
 412		}
 413
 414		promote->ifa_flags &= ~IFA_F_SECONDARY;
 415		rtmsg_ifa(RTM_NEWADDR, promote, nlh, portid);
 416		blocking_notifier_call_chain(&inetaddr_chain,
 417				NETDEV_UP, promote);
 418		for (ifa = next_sec; ifa; ifa = ifa->ifa_next) {
 
 419			if (ifa1->ifa_mask != ifa->ifa_mask ||
 420			    !inet_ifa_match(ifa1->ifa_address, ifa))
 421					continue;
 422			fib_add_ifaddr(ifa);
 423		}
 424
 425	}
 426	if (destroy)
 427		inet_free_ifa(ifa1);
 428}
 429
 430static void inet_del_ifa(struct in_device *in_dev, struct in_ifaddr **ifap,
 
 431			 int destroy)
 432{
 433	__inet_del_ifa(in_dev, ifap, destroy, NULL, 0);
 434}
 435
 436static void check_lifetime(struct work_struct *work);
 437
 438static DECLARE_DELAYED_WORK(check_lifetime_work, check_lifetime);
 439
 440static int __inet_insert_ifa(struct in_ifaddr *ifa, struct nlmsghdr *nlh,
 441			     u32 portid)
 442{
 
 443	struct in_device *in_dev = ifa->ifa_dev;
 444	struct in_ifaddr *ifa1, **ifap, **last_primary;
 
 
 445
 446	ASSERT_RTNL();
 447
 448	if (!ifa->ifa_local) {
 449		inet_free_ifa(ifa);
 450		return 0;
 451	}
 452
 453	ifa->ifa_flags &= ~IFA_F_SECONDARY;
 454	last_primary = &in_dev->ifa_list;
 455
 456	for (ifap = &in_dev->ifa_list; (ifa1 = *ifap) != NULL;
 457	     ifap = &ifa1->ifa_next) {
 
 
 
 
 
 458		if (!(ifa1->ifa_flags & IFA_F_SECONDARY) &&
 459		    ifa->ifa_scope <= ifa1->ifa_scope)
 460			last_primary = &ifa1->ifa_next;
 461		if (ifa1->ifa_mask == ifa->ifa_mask &&
 462		    inet_ifa_match(ifa1->ifa_address, ifa)) {
 463			if (ifa1->ifa_local == ifa->ifa_local) {
 464				inet_free_ifa(ifa);
 465				return -EEXIST;
 466			}
 467			if (ifa1->ifa_scope != ifa->ifa_scope) {
 468				inet_free_ifa(ifa);
 469				return -EINVAL;
 470			}
 471			ifa->ifa_flags |= IFA_F_SECONDARY;
 472		}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 473	}
 474
 475	if (!(ifa->ifa_flags & IFA_F_SECONDARY)) {
 476		prandom_seed((__force u32) ifa->ifa_local);
 477		ifap = last_primary;
 478	}
 479
 480	ifa->ifa_next = *ifap;
 481	*ifap = ifa;
 482
 483	inet_hash_insert(dev_net(in_dev->dev), ifa);
 484
 485	cancel_delayed_work(&check_lifetime_work);
 486	queue_delayed_work(system_power_efficient_wq, &check_lifetime_work, 0);
 487
 488	/* Send message first, then call notifier.
 489	   Notifier will trigger FIB update, so that
 490	   listeners of netlink will know about new ifaddr */
 491	rtmsg_ifa(RTM_NEWADDR, ifa, nlh, portid);
 492	blocking_notifier_call_chain(&inetaddr_chain, NETDEV_UP, ifa);
 493
 494	return 0;
 495}
 496
 497static int inet_insert_ifa(struct in_ifaddr *ifa)
 498{
 499	return __inet_insert_ifa(ifa, NULL, 0);
 500}
 501
 502static int inet_set_ifa(struct net_device *dev, struct in_ifaddr *ifa)
 503{
 504	struct in_device *in_dev = __in_dev_get_rtnl(dev);
 505
 506	ASSERT_RTNL();
 507
 508	if (!in_dev) {
 509		inet_free_ifa(ifa);
 510		return -ENOBUFS;
 511	}
 512	ipv4_devconf_setall(in_dev);
 513	neigh_parms_data_state_setall(in_dev->arp_parms);
 514	if (ifa->ifa_dev != in_dev) {
 515		WARN_ON(ifa->ifa_dev);
 516		in_dev_hold(in_dev);
 517		ifa->ifa_dev = in_dev;
 518	}
 519	if (ipv4_is_loopback(ifa->ifa_local))
 520		ifa->ifa_scope = RT_SCOPE_HOST;
 521	return inet_insert_ifa(ifa);
 522}
 523
 524/* Caller must hold RCU or RTNL :
 525 * We dont take a reference on found in_device
 526 */
 527struct in_device *inetdev_by_index(struct net *net, int ifindex)
 528{
 529	struct net_device *dev;
 530	struct in_device *in_dev = NULL;
 531
 532	rcu_read_lock();
 533	dev = dev_get_by_index_rcu(net, ifindex);
 534	if (dev)
 535		in_dev = rcu_dereference_rtnl(dev->ip_ptr);
 536	rcu_read_unlock();
 537	return in_dev;
 538}
 539EXPORT_SYMBOL(inetdev_by_index);
 540
 541/* Called only from RTNL semaphored context. No locks. */
 542
 543struct in_ifaddr *inet_ifa_byprefix(struct in_device *in_dev, __be32 prefix,
 544				    __be32 mask)
 545{
 
 
 546	ASSERT_RTNL();
 547
 548	for_primary_ifa(in_dev) {
 549		if (ifa->ifa_mask == mask && inet_ifa_match(prefix, ifa))
 550			return ifa;
 551	} endfor_ifa(in_dev);
 552	return NULL;
 553}
 554
 555static int ip_mc_config(struct sock *sk, bool join, const struct in_ifaddr *ifa)
 
 556{
 
 557	struct ip_mreqn mreq = {
 558		.imr_multiaddr.s_addr = ifa->ifa_address,
 559		.imr_ifindex = ifa->ifa_dev->dev->ifindex,
 560	};
 
 561	int ret;
 562
 563	ASSERT_RTNL();
 564
 565	lock_sock(sk);
 566	if (join)
 567		ret = ip_mc_join_group(sk, &mreq);
 568	else
 569		ret = ip_mc_leave_group(sk, &mreq);
 570	release_sock(sk);
 571
 572	return ret;
 
 
 
 573}
 574
 575static int inet_rtm_deladdr(struct sk_buff *skb, struct nlmsghdr *nlh)
 
 576{
 577	struct net *net = sock_net(skb->sk);
 
 578	struct nlattr *tb[IFA_MAX+1];
 579	struct in_device *in_dev;
 580	struct ifaddrmsg *ifm;
 581	struct in_ifaddr *ifa, **ifap;
 
 582	int err = -EINVAL;
 583
 584	ASSERT_RTNL();
 585
 586	err = nlmsg_parse(nlh, sizeof(*ifm), tb, IFA_MAX, ifa_ipv4_policy);
 
 587	if (err < 0)
 588		goto errout;
 589
 590	ifm = nlmsg_data(nlh);
 591	in_dev = inetdev_by_index(net, ifm->ifa_index);
 592	if (!in_dev) {
 593		err = -ENODEV;
 594		goto errout;
 595	}
 596
 597	for (ifap = &in_dev->ifa_list; (ifa = *ifap) != NULL;
 598	     ifap = &ifa->ifa_next) {
 599		if (tb[IFA_LOCAL] &&
 600		    ifa->ifa_local != nla_get_in_addr(tb[IFA_LOCAL]))
 601			continue;
 602
 603		if (tb[IFA_LABEL] && nla_strcmp(tb[IFA_LABEL], ifa->ifa_label))
 604			continue;
 605
 606		if (tb[IFA_ADDRESS] &&
 607		    (ifm->ifa_prefixlen != ifa->ifa_prefixlen ||
 608		    !inet_ifa_match(nla_get_in_addr(tb[IFA_ADDRESS]), ifa)))
 609			continue;
 610
 611		if (ipv4_is_multicast(ifa->ifa_address))
 612			ip_mc_config(net->ipv4.mc_autojoin_sk, false, ifa);
 613		__inet_del_ifa(in_dev, ifap, 1, nlh, NETLINK_CB(skb).portid);
 614		return 0;
 615	}
 616
 617	err = -EADDRNOTAVAIL;
 618errout:
 619	return err;
 620}
 621
 622#define INFINITY_LIFE_TIME	0xFFFFFFFF
 623
 624static void check_lifetime(struct work_struct *work)
 625{
 626	unsigned long now, next, next_sec, next_sched;
 627	struct in_ifaddr *ifa;
 628	struct hlist_node *n;
 629	int i;
 630
 631	now = jiffies;
 632	next = round_jiffies_up(now + ADDR_CHECK_FREQUENCY);
 633
 634	for (i = 0; i < IN4_ADDR_HSIZE; i++) {
 635		bool change_needed = false;
 636
 637		rcu_read_lock();
 638		hlist_for_each_entry_rcu(ifa, &inet_addr_lst[i], hash) {
 639			unsigned long age;
 640
 641			if (ifa->ifa_flags & IFA_F_PERMANENT)
 642				continue;
 643
 644			/* We try to batch several events at once. */
 645			age = (now - ifa->ifa_tstamp +
 646			       ADDRCONF_TIMER_FUZZ_MINUS) / HZ;
 647
 648			if (ifa->ifa_valid_lft != INFINITY_LIFE_TIME &&
 649			    age >= ifa->ifa_valid_lft) {
 650				change_needed = true;
 651			} else if (ifa->ifa_preferred_lft ==
 652				   INFINITY_LIFE_TIME) {
 653				continue;
 654			} else if (age >= ifa->ifa_preferred_lft) {
 655				if (time_before(ifa->ifa_tstamp +
 656						ifa->ifa_valid_lft * HZ, next))
 657					next = ifa->ifa_tstamp +
 658					       ifa->ifa_valid_lft * HZ;
 659
 660				if (!(ifa->ifa_flags & IFA_F_DEPRECATED))
 661					change_needed = true;
 662			} else if (time_before(ifa->ifa_tstamp +
 663					       ifa->ifa_preferred_lft * HZ,
 664					       next)) {
 665				next = ifa->ifa_tstamp +
 666				       ifa->ifa_preferred_lft * HZ;
 667			}
 668		}
 669		rcu_read_unlock();
 670		if (!change_needed)
 671			continue;
 672		rtnl_lock();
 673		hlist_for_each_entry_safe(ifa, n, &inet_addr_lst[i], hash) {
 674			unsigned long age;
 675
 676			if (ifa->ifa_flags & IFA_F_PERMANENT)
 677				continue;
 678
 679			/* We try to batch several events at once. */
 680			age = (now - ifa->ifa_tstamp +
 681			       ADDRCONF_TIMER_FUZZ_MINUS) / HZ;
 682
 683			if (ifa->ifa_valid_lft != INFINITY_LIFE_TIME &&
 684			    age >= ifa->ifa_valid_lft) {
 685				struct in_ifaddr **ifap;
 
 686
 687				for (ifap = &ifa->ifa_dev->ifa_list;
 688				     *ifap != NULL; ifap = &(*ifap)->ifa_next) {
 689					if (*ifap == ifa) {
 
 690						inet_del_ifa(ifa->ifa_dev,
 691							     ifap, 1);
 692						break;
 693					}
 
 
 694				}
 695			} else if (ifa->ifa_preferred_lft !=
 696				   INFINITY_LIFE_TIME &&
 697				   age >= ifa->ifa_preferred_lft &&
 698				   !(ifa->ifa_flags & IFA_F_DEPRECATED)) {
 699				ifa->ifa_flags |= IFA_F_DEPRECATED;
 700				rtmsg_ifa(RTM_NEWADDR, ifa, NULL, 0);
 701			}
 702		}
 703		rtnl_unlock();
 704	}
 705
 706	next_sec = round_jiffies_up(next);
 707	next_sched = next;
 708
 709	/* If rounded timeout is accurate enough, accept it. */
 710	if (time_before(next_sec, next + ADDRCONF_TIMER_FUZZ))
 711		next_sched = next_sec;
 712
 713	now = jiffies;
 714	/* And minimum interval is ADDRCONF_TIMER_FUZZ_MAX. */
 715	if (time_before(next_sched, now + ADDRCONF_TIMER_FUZZ_MAX))
 716		next_sched = now + ADDRCONF_TIMER_FUZZ_MAX;
 717
 718	queue_delayed_work(system_power_efficient_wq, &check_lifetime_work,
 719			next_sched - now);
 720}
 721
 722static void set_ifa_lifetime(struct in_ifaddr *ifa, __u32 valid_lft,
 723			     __u32 prefered_lft)
 724{
 725	unsigned long timeout;
 726
 727	ifa->ifa_flags &= ~(IFA_F_PERMANENT | IFA_F_DEPRECATED);
 728
 729	timeout = addrconf_timeout_fixup(valid_lft, HZ);
 730	if (addrconf_finite_timeout(timeout))
 731		ifa->ifa_valid_lft = timeout;
 732	else
 733		ifa->ifa_flags |= IFA_F_PERMANENT;
 734
 735	timeout = addrconf_timeout_fixup(prefered_lft, HZ);
 736	if (addrconf_finite_timeout(timeout)) {
 737		if (timeout == 0)
 738			ifa->ifa_flags |= IFA_F_DEPRECATED;
 739		ifa->ifa_preferred_lft = timeout;
 740	}
 741	ifa->ifa_tstamp = jiffies;
 742	if (!ifa->ifa_cstamp)
 743		ifa->ifa_cstamp = ifa->ifa_tstamp;
 744}
 745
 746static struct in_ifaddr *rtm_to_ifaddr(struct net *net, struct nlmsghdr *nlh,
 747				       __u32 *pvalid_lft, __u32 *pprefered_lft)
 
 748{
 749	struct nlattr *tb[IFA_MAX+1];
 750	struct in_ifaddr *ifa;
 751	struct ifaddrmsg *ifm;
 752	struct net_device *dev;
 753	struct in_device *in_dev;
 754	int err;
 755
 756	err = nlmsg_parse(nlh, sizeof(*ifm), tb, IFA_MAX, ifa_ipv4_policy);
 
 757	if (err < 0)
 758		goto errout;
 759
 760	ifm = nlmsg_data(nlh);
 761	err = -EINVAL;
 762	if (ifm->ifa_prefixlen > 32 || !tb[IFA_LOCAL])
 763		goto errout;
 764
 765	dev = __dev_get_by_index(net, ifm->ifa_index);
 766	err = -ENODEV;
 767	if (!dev)
 768		goto errout;
 769
 770	in_dev = __in_dev_get_rtnl(dev);
 771	err = -ENOBUFS;
 772	if (!in_dev)
 773		goto errout;
 774
 775	ifa = inet_alloc_ifa();
 776	if (!ifa)
 777		/*
 778		 * A potential indev allocation can be left alive, it stays
 779		 * assigned to its device and is destroy with it.
 780		 */
 781		goto errout;
 782
 783	ipv4_devconf_setall(in_dev);
 784	neigh_parms_data_state_setall(in_dev->arp_parms);
 785	in_dev_hold(in_dev);
 786
 787	if (!tb[IFA_ADDRESS])
 788		tb[IFA_ADDRESS] = tb[IFA_LOCAL];
 789
 790	INIT_HLIST_NODE(&ifa->hash);
 791	ifa->ifa_prefixlen = ifm->ifa_prefixlen;
 792	ifa->ifa_mask = inet_make_mask(ifm->ifa_prefixlen);
 793	ifa->ifa_flags = tb[IFA_FLAGS] ? nla_get_u32(tb[IFA_FLAGS]) :
 794					 ifm->ifa_flags;
 795	ifa->ifa_scope = ifm->ifa_scope;
 796	ifa->ifa_dev = in_dev;
 797
 798	ifa->ifa_local = nla_get_in_addr(tb[IFA_LOCAL]);
 799	ifa->ifa_address = nla_get_in_addr(tb[IFA_ADDRESS]);
 800
 801	if (tb[IFA_BROADCAST])
 802		ifa->ifa_broadcast = nla_get_in_addr(tb[IFA_BROADCAST]);
 803
 804	if (tb[IFA_LABEL])
 805		nla_strlcpy(ifa->ifa_label, tb[IFA_LABEL], IFNAMSIZ);
 806	else
 807		memcpy(ifa->ifa_label, dev->name, IFNAMSIZ);
 808
 
 
 
 809	if (tb[IFA_CACHEINFO]) {
 810		struct ifa_cacheinfo *ci;
 811
 812		ci = nla_data(tb[IFA_CACHEINFO]);
 813		if (!ci->ifa_valid || ci->ifa_prefered > ci->ifa_valid) {
 814			err = -EINVAL;
 815			goto errout_free;
 816		}
 817		*pvalid_lft = ci->ifa_valid;
 818		*pprefered_lft = ci->ifa_prefered;
 819	}
 820
 821	return ifa;
 822
 823errout_free:
 824	inet_free_ifa(ifa);
 825errout:
 826	return ERR_PTR(err);
 827}
 828
 829static struct in_ifaddr *find_matching_ifa(struct in_ifaddr *ifa)
 830{
 831	struct in_device *in_dev = ifa->ifa_dev;
 832	struct in_ifaddr *ifa1, **ifap;
 833
 834	if (!ifa->ifa_local)
 835		return NULL;
 836
 837	for (ifap = &in_dev->ifa_list; (ifa1 = *ifap) != NULL;
 838	     ifap = &ifa1->ifa_next) {
 839		if (ifa1->ifa_mask == ifa->ifa_mask &&
 840		    inet_ifa_match(ifa1->ifa_address, ifa) &&
 841		    ifa1->ifa_local == ifa->ifa_local)
 842			return ifa1;
 843	}
 844	return NULL;
 845}
 846
 847static int inet_rtm_newaddr(struct sk_buff *skb, struct nlmsghdr *nlh)
 
 848{
 849	struct net *net = sock_net(skb->sk);
 850	struct in_ifaddr *ifa;
 851	struct in_ifaddr *ifa_existing;
 852	__u32 valid_lft = INFINITY_LIFE_TIME;
 853	__u32 prefered_lft = INFINITY_LIFE_TIME;
 854
 855	ASSERT_RTNL();
 856
 857	ifa = rtm_to_ifaddr(net, nlh, &valid_lft, &prefered_lft);
 858	if (IS_ERR(ifa))
 859		return PTR_ERR(ifa);
 860
 861	ifa_existing = find_matching_ifa(ifa);
 862	if (!ifa_existing) {
 863		/* It would be best to check for !NLM_F_CREATE here but
 864		 * userspace already relies on not having to provide this.
 865		 */
 866		set_ifa_lifetime(ifa, valid_lft, prefered_lft);
 867		if (ifa->ifa_flags & IFA_F_MCAUTOJOIN) {
 868			int ret = ip_mc_config(net->ipv4.mc_autojoin_sk,
 869					       true, ifa);
 870
 871			if (ret < 0) {
 872				inet_free_ifa(ifa);
 873				return ret;
 874			}
 875		}
 876		return __inet_insert_ifa(ifa, nlh, NETLINK_CB(skb).portid);
 
 877	} else {
 
 
 878		inet_free_ifa(ifa);
 879
 880		if (nlh->nlmsg_flags & NLM_F_EXCL ||
 881		    !(nlh->nlmsg_flags & NLM_F_REPLACE))
 882			return -EEXIST;
 883		ifa = ifa_existing;
 
 
 
 
 
 
 884		set_ifa_lifetime(ifa, valid_lft, prefered_lft);
 885		cancel_delayed_work(&check_lifetime_work);
 886		queue_delayed_work(system_power_efficient_wq,
 887				&check_lifetime_work, 0);
 888		rtmsg_ifa(RTM_NEWADDR, ifa, nlh, NETLINK_CB(skb).portid);
 889	}
 890	return 0;
 891}
 892
 893/*
 894 *	Determine a default network mask, based on the IP address.
 895 */
 896
 897static int inet_abc_len(__be32 addr)
 898{
 899	int rc = -1;	/* Something else, probably a multicast. */
 900
 901	if (ipv4_is_zeronet(addr))
 902		rc = 0;
 903	else {
 904		__u32 haddr = ntohl(addr);
 905
 906		if (IN_CLASSA(haddr))
 907			rc = 8;
 908		else if (IN_CLASSB(haddr))
 909			rc = 16;
 910		else if (IN_CLASSC(haddr))
 911			rc = 24;
 
 
 912	}
 913
 914	return rc;
 915}
 916
 917
 918int devinet_ioctl(struct net *net, unsigned int cmd, void __user *arg)
 919{
 920	struct ifreq ifr;
 921	struct sockaddr_in sin_orig;
 922	struct sockaddr_in *sin = (struct sockaddr_in *)&ifr.ifr_addr;
 
 923	struct in_device *in_dev;
 924	struct in_ifaddr **ifap = NULL;
 925	struct in_ifaddr *ifa = NULL;
 926	struct net_device *dev;
 927	char *colon;
 928	int ret = -EFAULT;
 929	int tryaddrmatch = 0;
 930
 931	/*
 932	 *	Fetch the caller's info block into kernel space
 933	 */
 934
 935	if (copy_from_user(&ifr, arg, sizeof(struct ifreq)))
 936		goto out;
 937	ifr.ifr_name[IFNAMSIZ - 1] = 0;
 938
 939	/* save original address for comparison */
 940	memcpy(&sin_orig, sin, sizeof(*sin));
 941
 942	colon = strchr(ifr.ifr_name, ':');
 943	if (colon)
 944		*colon = 0;
 945
 946	dev_load(net, ifr.ifr_name);
 947
 948	switch (cmd) {
 949	case SIOCGIFADDR:	/* Get interface address */
 950	case SIOCGIFBRDADDR:	/* Get the broadcast address */
 951	case SIOCGIFDSTADDR:	/* Get the destination address */
 952	case SIOCGIFNETMASK:	/* Get the netmask for the interface */
 953		/* Note that these ioctls will not sleep,
 954		   so that we do not impose a lock.
 955		   One day we will be forced to put shlock here (I mean SMP)
 956		 */
 957		tryaddrmatch = (sin_orig.sin_family == AF_INET);
 958		memset(sin, 0, sizeof(*sin));
 959		sin->sin_family = AF_INET;
 960		break;
 961
 962	case SIOCSIFFLAGS:
 963		ret = -EPERM;
 964		if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
 965			goto out;
 966		break;
 967	case SIOCSIFADDR:	/* Set interface address (and family) */
 968	case SIOCSIFBRDADDR:	/* Set the broadcast address */
 969	case SIOCSIFDSTADDR:	/* Set the destination address */
 970	case SIOCSIFNETMASK: 	/* Set the netmask for the interface */
 971		ret = -EPERM;
 972		if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
 973			goto out;
 974		ret = -EINVAL;
 975		if (sin->sin_family != AF_INET)
 976			goto out;
 977		break;
 978	default:
 979		ret = -EINVAL;
 980		goto out;
 981	}
 982
 983	rtnl_lock();
 984
 985	ret = -ENODEV;
 986	dev = __dev_get_by_name(net, ifr.ifr_name);
 987	if (!dev)
 988		goto done;
 989
 990	if (colon)
 991		*colon = ':';
 992
 993	in_dev = __in_dev_get_rtnl(dev);
 994	if (in_dev) {
 995		if (tryaddrmatch) {
 996			/* Matthias Andree */
 997			/* compare label and address (4.4BSD style) */
 998			/* note: we only do this for a limited set of ioctls
 999			   and only if the original address family was AF_INET.
1000			   This is checked above. */
1001			for (ifap = &in_dev->ifa_list; (ifa = *ifap) != NULL;
 
 
1002			     ifap = &ifa->ifa_next) {
1003				if (!strcmp(ifr.ifr_name, ifa->ifa_label) &&
1004				    sin_orig.sin_addr.s_addr ==
1005							ifa->ifa_local) {
1006					break; /* found */
1007				}
1008			}
1009		}
1010		/* we didn't get a match, maybe the application is
1011		   4.3BSD-style and passed in junk so we fall back to
1012		   comparing just the label */
1013		if (!ifa) {
1014			for (ifap = &in_dev->ifa_list; (ifa = *ifap) != NULL;
 
1015			     ifap = &ifa->ifa_next)
1016				if (!strcmp(ifr.ifr_name, ifa->ifa_label))
1017					break;
1018		}
1019	}
1020
1021	ret = -EADDRNOTAVAIL;
1022	if (!ifa && cmd != SIOCSIFADDR && cmd != SIOCSIFFLAGS)
1023		goto done;
1024
1025	switch (cmd) {
1026	case SIOCGIFADDR:	/* Get interface address */
 
1027		sin->sin_addr.s_addr = ifa->ifa_local;
1028		goto rarok;
1029
1030	case SIOCGIFBRDADDR:	/* Get the broadcast address */
 
1031		sin->sin_addr.s_addr = ifa->ifa_broadcast;
1032		goto rarok;
1033
1034	case SIOCGIFDSTADDR:	/* Get the destination address */
 
1035		sin->sin_addr.s_addr = ifa->ifa_address;
1036		goto rarok;
1037
1038	case SIOCGIFNETMASK:	/* Get the netmask for the interface */
 
1039		sin->sin_addr.s_addr = ifa->ifa_mask;
1040		goto rarok;
1041
1042	case SIOCSIFFLAGS:
1043		if (colon) {
1044			ret = -EADDRNOTAVAIL;
1045			if (!ifa)
1046				break;
1047			ret = 0;
1048			if (!(ifr.ifr_flags & IFF_UP))
1049				inet_del_ifa(in_dev, ifap, 1);
1050			break;
1051		}
1052		ret = dev_change_flags(dev, ifr.ifr_flags);
1053		break;
1054
1055	case SIOCSIFADDR:	/* Set interface address (and family) */
1056		ret = -EINVAL;
1057		if (inet_abc_len(sin->sin_addr.s_addr) < 0)
1058			break;
1059
1060		if (!ifa) {
1061			ret = -ENOBUFS;
1062			ifa = inet_alloc_ifa();
1063			if (!ifa)
1064				break;
1065			INIT_HLIST_NODE(&ifa->hash);
1066			if (colon)
1067				memcpy(ifa->ifa_label, ifr.ifr_name, IFNAMSIZ);
1068			else
1069				memcpy(ifa->ifa_label, dev->name, IFNAMSIZ);
1070		} else {
1071			ret = 0;
1072			if (ifa->ifa_local == sin->sin_addr.s_addr)
1073				break;
1074			inet_del_ifa(in_dev, ifap, 0);
1075			ifa->ifa_broadcast = 0;
1076			ifa->ifa_scope = 0;
1077		}
1078
1079		ifa->ifa_address = ifa->ifa_local = sin->sin_addr.s_addr;
1080
1081		if (!(dev->flags & IFF_POINTOPOINT)) {
1082			ifa->ifa_prefixlen = inet_abc_len(ifa->ifa_address);
1083			ifa->ifa_mask = inet_make_mask(ifa->ifa_prefixlen);
1084			if ((dev->flags & IFF_BROADCAST) &&
1085			    ifa->ifa_prefixlen < 31)
1086				ifa->ifa_broadcast = ifa->ifa_address |
1087						     ~ifa->ifa_mask;
1088		} else {
1089			ifa->ifa_prefixlen = 32;
1090			ifa->ifa_mask = inet_make_mask(32);
1091		}
1092		set_ifa_lifetime(ifa, INFINITY_LIFE_TIME, INFINITY_LIFE_TIME);
1093		ret = inet_set_ifa(dev, ifa);
1094		break;
1095
1096	case SIOCSIFBRDADDR:	/* Set the broadcast address */
1097		ret = 0;
1098		if (ifa->ifa_broadcast != sin->sin_addr.s_addr) {
1099			inet_del_ifa(in_dev, ifap, 0);
1100			ifa->ifa_broadcast = sin->sin_addr.s_addr;
1101			inet_insert_ifa(ifa);
1102		}
1103		break;
1104
1105	case SIOCSIFDSTADDR:	/* Set the destination address */
1106		ret = 0;
1107		if (ifa->ifa_address == sin->sin_addr.s_addr)
1108			break;
1109		ret = -EINVAL;
1110		if (inet_abc_len(sin->sin_addr.s_addr) < 0)
1111			break;
1112		ret = 0;
1113		inet_del_ifa(in_dev, ifap, 0);
1114		ifa->ifa_address = sin->sin_addr.s_addr;
1115		inet_insert_ifa(ifa);
1116		break;
1117
1118	case SIOCSIFNETMASK: 	/* Set the netmask for the interface */
1119
1120		/*
1121		 *	The mask we set must be legal.
1122		 */
1123		ret = -EINVAL;
1124		if (bad_mask(sin->sin_addr.s_addr, 0))
1125			break;
1126		ret = 0;
1127		if (ifa->ifa_mask != sin->sin_addr.s_addr) {
1128			__be32 old_mask = ifa->ifa_mask;
1129			inet_del_ifa(in_dev, ifap, 0);
1130			ifa->ifa_mask = sin->sin_addr.s_addr;
1131			ifa->ifa_prefixlen = inet_mask_len(ifa->ifa_mask);
1132
1133			/* See if current broadcast address matches
1134			 * with current netmask, then recalculate
1135			 * the broadcast address. Otherwise it's a
1136			 * funny address, so don't touch it since
1137			 * the user seems to know what (s)he's doing...
1138			 */
1139			if ((dev->flags & IFF_BROADCAST) &&
1140			    (ifa->ifa_prefixlen < 31) &&
1141			    (ifa->ifa_broadcast ==
1142			     (ifa->ifa_local|~old_mask))) {
1143				ifa->ifa_broadcast = (ifa->ifa_local |
1144						      ~sin->sin_addr.s_addr);
1145			}
1146			inet_insert_ifa(ifa);
1147		}
1148		break;
1149	}
1150done:
1151	rtnl_unlock();
1152out:
1153	return ret;
1154rarok:
1155	rtnl_unlock();
1156	ret = copy_to_user(arg, &ifr, sizeof(struct ifreq)) ? -EFAULT : 0;
1157	goto out;
1158}
1159
1160static int inet_gifconf(struct net_device *dev, char __user *buf, int len)
1161{
1162	struct in_device *in_dev = __in_dev_get_rtnl(dev);
1163	struct in_ifaddr *ifa;
1164	struct ifreq ifr;
1165	int done = 0;
1166
 
 
 
1167	if (!in_dev)
1168		goto out;
1169
1170	for (ifa = in_dev->ifa_list; ifa; ifa = ifa->ifa_next) {
1171		if (!buf) {
1172			done += sizeof(ifr);
1173			continue;
1174		}
1175		if (len < (int) sizeof(ifr))
1176			break;
1177		memset(&ifr, 0, sizeof(struct ifreq));
1178		strcpy(ifr.ifr_name, ifa->ifa_label);
1179
1180		(*(struct sockaddr_in *)&ifr.ifr_addr).sin_family = AF_INET;
1181		(*(struct sockaddr_in *)&ifr.ifr_addr).sin_addr.s_addr =
1182								ifa->ifa_local;
1183
1184		if (copy_to_user(buf, &ifr, sizeof(struct ifreq))) {
1185			done = -EFAULT;
1186			break;
1187		}
1188		buf  += sizeof(struct ifreq);
1189		len  -= sizeof(struct ifreq);
1190		done += sizeof(struct ifreq);
1191	}
1192out:
1193	return done;
1194}
1195
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1196__be32 inet_select_addr(const struct net_device *dev, __be32 dst, int scope)
1197{
 
1198	__be32 addr = 0;
 
1199	struct in_device *in_dev;
1200	struct net *net = dev_net(dev);
1201	int master_idx;
1202
1203	rcu_read_lock();
1204	in_dev = __in_dev_get_rcu(dev);
1205	if (!in_dev)
1206		goto no_in_dev;
1207
1208	for_primary_ifa(in_dev) {
1209		if (ifa->ifa_scope > scope)
 
 
 
 
 
1210			continue;
1211		if (!dst || inet_ifa_match(dst, ifa)) {
1212			addr = ifa->ifa_local;
1213			break;
1214		}
1215		if (!addr)
1216			addr = ifa->ifa_local;
1217	} endfor_ifa(in_dev);
1218
1219	if (addr)
1220		goto out_unlock;
1221no_in_dev:
1222	master_idx = l3mdev_master_ifindex_rcu(dev);
1223
1224	/* For VRFs, the VRF device takes the place of the loopback device,
1225	 * with addresses on it being preferred.  Note in such cases the
1226	 * loopback device will be among the devices that fail the master_idx
1227	 * equality check in the loop below.
1228	 */
1229	if (master_idx &&
1230	    (dev = dev_get_by_index_rcu(net, master_idx)) &&
1231	    (in_dev = __in_dev_get_rcu(dev))) {
1232		for_primary_ifa(in_dev) {
1233			if (ifa->ifa_scope != RT_SCOPE_LINK &&
1234			    ifa->ifa_scope <= scope) {
1235				addr = ifa->ifa_local;
1236				goto out_unlock;
1237			}
1238		} endfor_ifa(in_dev);
1239	}
1240
1241	/* Not loopback addresses on loopback should be preferred
1242	   in this case. It is important that lo is the first interface
1243	   in dev_base list.
1244	 */
1245	for_each_netdev_rcu(net, dev) {
1246		if (l3mdev_master_ifindex_rcu(dev) != master_idx)
1247			continue;
1248
1249		in_dev = __in_dev_get_rcu(dev);
1250		if (!in_dev)
1251			continue;
1252
1253		for_primary_ifa(in_dev) {
1254			if (ifa->ifa_scope != RT_SCOPE_LINK &&
1255			    ifa->ifa_scope <= scope) {
1256				addr = ifa->ifa_local;
1257				goto out_unlock;
1258			}
1259		} endfor_ifa(in_dev);
1260	}
1261out_unlock:
1262	rcu_read_unlock();
1263	return addr;
1264}
1265EXPORT_SYMBOL(inet_select_addr);
1266
1267static __be32 confirm_addr_indev(struct in_device *in_dev, __be32 dst,
1268			      __be32 local, int scope)
1269{
1270	int same = 0;
 
1271	__be32 addr = 0;
 
 
 
 
 
 
 
1272
1273	for_ifa(in_dev) {
1274		if (!addr &&
1275		    (local == ifa->ifa_local || !local) &&
1276		    ifa->ifa_scope <= scope) {
1277			addr = ifa->ifa_local;
1278			if (same)
1279				break;
1280		}
1281		if (!same) {
1282			same = (!local || inet_ifa_match(local, ifa)) &&
1283				(!dst || inet_ifa_match(dst, ifa));
1284			if (same && addr) {
1285				if (local || !dst)
1286					break;
1287				/* Is the selected addr into dst subnet? */
1288				if (inet_ifa_match(addr, ifa))
1289					break;
1290				/* No, then can we use new local src? */
1291				if (ifa->ifa_scope <= scope) {
1292					addr = ifa->ifa_local;
1293					break;
1294				}
1295				/* search for large dst subnet for addr */
1296				same = 0;
1297			}
1298		}
1299	} endfor_ifa(in_dev);
1300
1301	return same ? addr : 0;
1302}
1303
1304/*
1305 * Confirm that local IP address exists using wildcards:
1306 * - net: netns to check, cannot be NULL
1307 * - in_dev: only on this interface, NULL=any interface
1308 * - dst: only in the same subnet as dst, 0=any dst
1309 * - local: address, 0=autoselect the local address
1310 * - scope: maximum allowed scope value for the local address
1311 */
1312__be32 inet_confirm_addr(struct net *net, struct in_device *in_dev,
1313			 __be32 dst, __be32 local, int scope)
1314{
1315	__be32 addr = 0;
1316	struct net_device *dev;
1317
1318	if (in_dev)
1319		return confirm_addr_indev(in_dev, dst, local, scope);
1320
1321	rcu_read_lock();
1322	for_each_netdev_rcu(net, dev) {
1323		in_dev = __in_dev_get_rcu(dev);
1324		if (in_dev) {
1325			addr = confirm_addr_indev(in_dev, dst, local, scope);
1326			if (addr)
1327				break;
1328		}
1329	}
1330	rcu_read_unlock();
1331
1332	return addr;
1333}
1334EXPORT_SYMBOL(inet_confirm_addr);
1335
1336/*
1337 *	Device notifier
1338 */
1339
1340int register_inetaddr_notifier(struct notifier_block *nb)
1341{
1342	return blocking_notifier_chain_register(&inetaddr_chain, nb);
1343}
1344EXPORT_SYMBOL(register_inetaddr_notifier);
1345
1346int unregister_inetaddr_notifier(struct notifier_block *nb)
1347{
1348	return blocking_notifier_chain_unregister(&inetaddr_chain, nb);
1349}
1350EXPORT_SYMBOL(unregister_inetaddr_notifier);
1351
 
 
 
 
 
 
 
 
 
 
 
 
 
1352/* Rename ifa_labels for a device name change. Make some effort to preserve
1353 * existing alias numbering and to create unique labels if possible.
1354*/
1355static void inetdev_changename(struct net_device *dev, struct in_device *in_dev)
1356{
1357	struct in_ifaddr *ifa;
1358	int named = 0;
1359
1360	for (ifa = in_dev->ifa_list; ifa; ifa = ifa->ifa_next) {
1361		char old[IFNAMSIZ], *dot;
1362
1363		memcpy(old, ifa->ifa_label, IFNAMSIZ);
1364		memcpy(ifa->ifa_label, dev->name, IFNAMSIZ);
1365		if (named++ == 0)
1366			goto skip;
1367		dot = strchr(old, ':');
1368		if (!dot) {
1369			sprintf(old, ":%d", named);
1370			dot = old;
1371		}
1372		if (strlen(dot) + strlen(dev->name) < IFNAMSIZ)
1373			strcat(ifa->ifa_label, dot);
1374		else
1375			strcpy(ifa->ifa_label + (IFNAMSIZ - strlen(dot) - 1), dot);
1376skip:
1377		rtmsg_ifa(RTM_NEWADDR, ifa, NULL, 0);
1378	}
1379}
1380
1381static bool inetdev_valid_mtu(unsigned int mtu)
1382{
1383	return mtu >= 68;
1384}
1385
1386static void inetdev_send_gratuitous_arp(struct net_device *dev,
1387					struct in_device *in_dev)
1388
1389{
1390	struct in_ifaddr *ifa;
1391
1392	for (ifa = in_dev->ifa_list; ifa;
1393	     ifa = ifa->ifa_next) {
1394		arp_send(ARPOP_REQUEST, ETH_P_ARP,
1395			 ifa->ifa_local, dev,
1396			 ifa->ifa_local, NULL,
1397			 dev->dev_addr, NULL);
1398	}
1399}
1400
1401/* Called only under RTNL semaphore */
1402
1403static int inetdev_event(struct notifier_block *this, unsigned long event,
1404			 void *ptr)
1405{
1406	struct net_device *dev = netdev_notifier_info_to_dev(ptr);
1407	struct in_device *in_dev = __in_dev_get_rtnl(dev);
1408
1409	ASSERT_RTNL();
1410
1411	if (!in_dev) {
1412		if (event == NETDEV_REGISTER) {
1413			in_dev = inetdev_init(dev);
1414			if (IS_ERR(in_dev))
1415				return notifier_from_errno(PTR_ERR(in_dev));
1416			if (dev->flags & IFF_LOOPBACK) {
1417				IN_DEV_CONF_SET(in_dev, NOXFRM, 1);
1418				IN_DEV_CONF_SET(in_dev, NOPOLICY, 1);
1419			}
1420		} else if (event == NETDEV_CHANGEMTU) {
1421			/* Re-enabling IP */
1422			if (inetdev_valid_mtu(dev->mtu))
1423				in_dev = inetdev_init(dev);
1424		}
1425		goto out;
1426	}
1427
1428	switch (event) {
1429	case NETDEV_REGISTER:
1430		pr_debug("%s: bug\n", __func__);
1431		RCU_INIT_POINTER(dev->ip_ptr, NULL);
1432		break;
1433	case NETDEV_UP:
1434		if (!inetdev_valid_mtu(dev->mtu))
1435			break;
1436		if (dev->flags & IFF_LOOPBACK) {
1437			struct in_ifaddr *ifa = inet_alloc_ifa();
1438
1439			if (ifa) {
1440				INIT_HLIST_NODE(&ifa->hash);
1441				ifa->ifa_local =
1442				  ifa->ifa_address = htonl(INADDR_LOOPBACK);
1443				ifa->ifa_prefixlen = 8;
1444				ifa->ifa_mask = inet_make_mask(8);
1445				in_dev_hold(in_dev);
1446				ifa->ifa_dev = in_dev;
1447				ifa->ifa_scope = RT_SCOPE_HOST;
1448				memcpy(ifa->ifa_label, dev->name, IFNAMSIZ);
1449				set_ifa_lifetime(ifa, INFINITY_LIFE_TIME,
1450						 INFINITY_LIFE_TIME);
1451				ipv4_devconf_setall(in_dev);
1452				neigh_parms_data_state_setall(in_dev->arp_parms);
1453				inet_insert_ifa(ifa);
1454			}
1455		}
1456		ip_mc_up(in_dev);
1457		/* fall through */
1458	case NETDEV_CHANGEADDR:
1459		if (!IN_DEV_ARP_NOTIFY(in_dev))
1460			break;
1461		/* fall through */
1462	case NETDEV_NOTIFY_PEERS:
1463		/* Send gratuitous ARP to notify of link change */
1464		inetdev_send_gratuitous_arp(dev, in_dev);
1465		break;
1466	case NETDEV_DOWN:
1467		ip_mc_down(in_dev);
1468		break;
1469	case NETDEV_PRE_TYPE_CHANGE:
1470		ip_mc_unmap(in_dev);
1471		break;
1472	case NETDEV_POST_TYPE_CHANGE:
1473		ip_mc_remap(in_dev);
1474		break;
1475	case NETDEV_CHANGEMTU:
1476		if (inetdev_valid_mtu(dev->mtu))
1477			break;
1478		/* disable IP when MTU is not enough */
 
1479	case NETDEV_UNREGISTER:
1480		inetdev_destroy(in_dev);
1481		break;
1482	case NETDEV_CHANGENAME:
1483		/* Do not notify about label change, this event is
1484		 * not interesting to applications using netlink.
1485		 */
1486		inetdev_changename(dev, in_dev);
1487
1488		devinet_sysctl_unregister(in_dev);
1489		devinet_sysctl_register(in_dev);
1490		break;
1491	}
1492out:
1493	return NOTIFY_DONE;
1494}
1495
1496static struct notifier_block ip_netdev_notifier = {
1497	.notifier_call = inetdev_event,
1498};
1499
1500static size_t inet_nlmsg_size(void)
1501{
1502	return NLMSG_ALIGN(sizeof(struct ifaddrmsg))
1503	       + nla_total_size(4) /* IFA_ADDRESS */
1504	       + nla_total_size(4) /* IFA_LOCAL */
1505	       + nla_total_size(4) /* IFA_BROADCAST */
1506	       + nla_total_size(IFNAMSIZ) /* IFA_LABEL */
1507	       + nla_total_size(4)  /* IFA_FLAGS */
 
1508	       + nla_total_size(sizeof(struct ifa_cacheinfo)); /* IFA_CACHEINFO */
1509}
1510
1511static inline u32 cstamp_delta(unsigned long cstamp)
1512{
1513	return (cstamp - INITIAL_JIFFIES) * 100UL / HZ;
1514}
1515
1516static int put_cacheinfo(struct sk_buff *skb, unsigned long cstamp,
1517			 unsigned long tstamp, u32 preferred, u32 valid)
1518{
1519	struct ifa_cacheinfo ci;
1520
1521	ci.cstamp = cstamp_delta(cstamp);
1522	ci.tstamp = cstamp_delta(tstamp);
1523	ci.ifa_prefered = preferred;
1524	ci.ifa_valid = valid;
1525
1526	return nla_put(skb, IFA_CACHEINFO, sizeof(ci), &ci);
1527}
1528
1529static int inet_fill_ifaddr(struct sk_buff *skb, struct in_ifaddr *ifa,
1530			    u32 portid, u32 seq, int event, unsigned int flags)
1531{
1532	struct ifaddrmsg *ifm;
1533	struct nlmsghdr  *nlh;
1534	u32 preferred, valid;
1535
1536	nlh = nlmsg_put(skb, portid, seq, event, sizeof(*ifm), flags);
 
1537	if (!nlh)
1538		return -EMSGSIZE;
1539
1540	ifm = nlmsg_data(nlh);
1541	ifm->ifa_family = AF_INET;
1542	ifm->ifa_prefixlen = ifa->ifa_prefixlen;
1543	ifm->ifa_flags = ifa->ifa_flags;
1544	ifm->ifa_scope = ifa->ifa_scope;
1545	ifm->ifa_index = ifa->ifa_dev->dev->ifindex;
1546
 
 
 
 
1547	if (!(ifm->ifa_flags & IFA_F_PERMANENT)) {
1548		preferred = ifa->ifa_preferred_lft;
1549		valid = ifa->ifa_valid_lft;
1550		if (preferred != INFINITY_LIFE_TIME) {
1551			long tval = (jiffies - ifa->ifa_tstamp) / HZ;
1552
1553			if (preferred > tval)
1554				preferred -= tval;
1555			else
1556				preferred = 0;
1557			if (valid != INFINITY_LIFE_TIME) {
1558				if (valid > tval)
1559					valid -= tval;
1560				else
1561					valid = 0;
1562			}
1563		}
1564	} else {
1565		preferred = INFINITY_LIFE_TIME;
1566		valid = INFINITY_LIFE_TIME;
1567	}
1568	if ((ifa->ifa_address &&
1569	     nla_put_in_addr(skb, IFA_ADDRESS, ifa->ifa_address)) ||
1570	    (ifa->ifa_local &&
1571	     nla_put_in_addr(skb, IFA_LOCAL, ifa->ifa_local)) ||
1572	    (ifa->ifa_broadcast &&
1573	     nla_put_in_addr(skb, IFA_BROADCAST, ifa->ifa_broadcast)) ||
1574	    (ifa->ifa_label[0] &&
1575	     nla_put_string(skb, IFA_LABEL, ifa->ifa_label)) ||
1576	    nla_put_u32(skb, IFA_FLAGS, ifa->ifa_flags) ||
 
 
1577	    put_cacheinfo(skb, ifa->ifa_cstamp, ifa->ifa_tstamp,
1578			  preferred, valid))
1579		goto nla_put_failure;
1580
1581	nlmsg_end(skb, nlh);
1582	return 0;
1583
1584nla_put_failure:
1585	nlmsg_cancel(skb, nlh);
1586	return -EMSGSIZE;
1587}
1588
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1589static int inet_dump_ifaddr(struct sk_buff *skb, struct netlink_callback *cb)
1590{
 
 
 
 
 
 
 
 
1591	struct net *net = sock_net(skb->sk);
 
1592	int h, s_h;
1593	int idx, s_idx;
1594	int ip_idx, s_ip_idx;
1595	struct net_device *dev;
1596	struct in_device *in_dev;
1597	struct in_ifaddr *ifa;
1598	struct hlist_head *head;
 
1599
1600	s_h = cb->args[0];
1601	s_idx = idx = cb->args[1];
1602	s_ip_idx = ip_idx = cb->args[2];
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1603
1604	for (h = s_h; h < NETDEV_HASHENTRIES; h++, s_idx = 0) {
1605		idx = 0;
1606		head = &net->dev_index_head[h];
1607		rcu_read_lock();
1608		cb->seq = atomic_read(&net->ipv4.dev_addr_genid) ^
1609			  net->dev_base_seq;
1610		hlist_for_each_entry_rcu(dev, head, index_hlist) {
1611			if (idx < s_idx)
1612				goto cont;
1613			if (h > s_h || idx > s_idx)
1614				s_ip_idx = 0;
1615			in_dev = __in_dev_get_rcu(dev);
1616			if (!in_dev)
1617				goto cont;
1618
1619			for (ifa = in_dev->ifa_list, ip_idx = 0; ifa;
1620			     ifa = ifa->ifa_next, ip_idx++) {
1621				if (ip_idx < s_ip_idx)
1622					continue;
1623				if (inet_fill_ifaddr(skb, ifa,
1624					     NETLINK_CB(cb->skb).portid,
1625					     cb->nlh->nlmsg_seq,
1626					     RTM_NEWADDR, NLM_F_MULTI) < 0) {
1627					rcu_read_unlock();
1628					goto done;
1629				}
1630				nl_dump_check_consistent(cb, nlmsg_hdr(skb));
1631			}
1632cont:
1633			idx++;
1634		}
1635		rcu_read_unlock();
1636	}
1637
1638done:
1639	cb->args[0] = h;
1640	cb->args[1] = idx;
1641	cb->args[2] = ip_idx;
 
 
1642
1643	return skb->len;
1644}
1645
1646static void rtmsg_ifa(int event, struct in_ifaddr *ifa, struct nlmsghdr *nlh,
1647		      u32 portid)
1648{
 
 
 
 
 
 
 
1649	struct sk_buff *skb;
1650	u32 seq = nlh ? nlh->nlmsg_seq : 0;
1651	int err = -ENOBUFS;
1652	struct net *net;
1653
1654	net = dev_net(ifa->ifa_dev->dev);
1655	skb = nlmsg_new(inet_nlmsg_size(), GFP_KERNEL);
1656	if (!skb)
1657		goto errout;
1658
1659	err = inet_fill_ifaddr(skb, ifa, portid, seq, event, 0);
1660	if (err < 0) {
1661		/* -EMSGSIZE implies BUG in inet_nlmsg_size() */
1662		WARN_ON(err == -EMSGSIZE);
1663		kfree_skb(skb);
1664		goto errout;
1665	}
1666	rtnl_notify(skb, net, portid, RTNLGRP_IPV4_IFADDR, nlh, GFP_KERNEL);
1667	return;
1668errout:
1669	if (err < 0)
1670		rtnl_set_sk_err(net, RTNLGRP_IPV4_IFADDR, err);
1671}
1672
1673static size_t inet_get_link_af_size(const struct net_device *dev,
1674				    u32 ext_filter_mask)
1675{
1676	struct in_device *in_dev = rcu_dereference_rtnl(dev->ip_ptr);
1677
1678	if (!in_dev)
1679		return 0;
1680
1681	return nla_total_size(IPV4_DEVCONF_MAX * 4); /* IFLA_INET_CONF */
1682}
1683
1684static int inet_fill_link_af(struct sk_buff *skb, const struct net_device *dev,
1685			     u32 ext_filter_mask)
1686{
1687	struct in_device *in_dev = rcu_dereference_rtnl(dev->ip_ptr);
1688	struct nlattr *nla;
1689	int i;
1690
1691	if (!in_dev)
1692		return -ENODATA;
1693
1694	nla = nla_reserve(skb, IFLA_INET_CONF, IPV4_DEVCONF_MAX * 4);
1695	if (!nla)
1696		return -EMSGSIZE;
1697
1698	for (i = 0; i < IPV4_DEVCONF_MAX; i++)
1699		((u32 *) nla_data(nla))[i] = in_dev->cnf.data[i];
1700
1701	return 0;
1702}
1703
1704static const struct nla_policy inet_af_policy[IFLA_INET_MAX+1] = {
1705	[IFLA_INET_CONF]	= { .type = NLA_NESTED },
1706};
1707
1708static int inet_validate_link_af(const struct net_device *dev,
1709				 const struct nlattr *nla)
1710{
1711	struct nlattr *a, *tb[IFLA_INET_MAX+1];
1712	int err, rem;
1713
1714	if (dev && !__in_dev_get_rtnl(dev))
1715		return -EAFNOSUPPORT;
1716
1717	err = nla_parse_nested(tb, IFLA_INET_MAX, nla, inet_af_policy);
 
1718	if (err < 0)
1719		return err;
1720
1721	if (tb[IFLA_INET_CONF]) {
1722		nla_for_each_nested(a, tb[IFLA_INET_CONF], rem) {
1723			int cfgid = nla_type(a);
1724
1725			if (nla_len(a) < 4)
1726				return -EINVAL;
1727
1728			if (cfgid <= 0 || cfgid > IPV4_DEVCONF_MAX)
1729				return -EINVAL;
1730		}
1731	}
1732
1733	return 0;
1734}
1735
1736static int inet_set_link_af(struct net_device *dev, const struct nlattr *nla)
1737{
1738	struct in_device *in_dev = __in_dev_get_rtnl(dev);
1739	struct nlattr *a, *tb[IFLA_INET_MAX+1];
1740	int rem;
1741
1742	if (!in_dev)
1743		return -EAFNOSUPPORT;
1744
1745	if (nla_parse_nested(tb, IFLA_INET_MAX, nla, NULL) < 0)
1746		BUG();
1747
1748	if (tb[IFLA_INET_CONF]) {
1749		nla_for_each_nested(a, tb[IFLA_INET_CONF], rem)
1750			ipv4_devconf_set(in_dev, nla_type(a), nla_get_u32(a));
1751	}
1752
1753	return 0;
1754}
1755
1756static int inet_netconf_msgsize_devconf(int type)
1757{
1758	int size = NLMSG_ALIGN(sizeof(struct netconfmsg))
1759		   + nla_total_size(4);	/* NETCONFA_IFINDEX */
1760	bool all = false;
1761
1762	if (type == NETCONFA_ALL)
1763		all = true;
1764
1765	if (all || type == NETCONFA_FORWARDING)
1766		size += nla_total_size(4);
1767	if (all || type == NETCONFA_RP_FILTER)
1768		size += nla_total_size(4);
1769	if (all || type == NETCONFA_MC_FORWARDING)
1770		size += nla_total_size(4);
 
 
1771	if (all || type == NETCONFA_PROXY_NEIGH)
1772		size += nla_total_size(4);
1773	if (all || type == NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN)
1774		size += nla_total_size(4);
1775
1776	return size;
1777}
1778
1779static int inet_netconf_fill_devconf(struct sk_buff *skb, int ifindex,
1780				     struct ipv4_devconf *devconf, u32 portid,
1781				     u32 seq, int event, unsigned int flags,
1782				     int type)
1783{
1784	struct nlmsghdr  *nlh;
1785	struct netconfmsg *ncm;
1786	bool all = false;
1787
1788	nlh = nlmsg_put(skb, portid, seq, event, sizeof(struct netconfmsg),
1789			flags);
1790	if (!nlh)
1791		return -EMSGSIZE;
1792
1793	if (type == NETCONFA_ALL)
1794		all = true;
1795
1796	ncm = nlmsg_data(nlh);
1797	ncm->ncm_family = AF_INET;
1798
1799	if (nla_put_s32(skb, NETCONFA_IFINDEX, ifindex) < 0)
1800		goto nla_put_failure;
1801
 
 
 
1802	if ((all || type == NETCONFA_FORWARDING) &&
1803	    nla_put_s32(skb, NETCONFA_FORWARDING,
1804			IPV4_DEVCONF(*devconf, FORWARDING)) < 0)
1805		goto nla_put_failure;
1806	if ((all || type == NETCONFA_RP_FILTER) &&
1807	    nla_put_s32(skb, NETCONFA_RP_FILTER,
1808			IPV4_DEVCONF(*devconf, RP_FILTER)) < 0)
1809		goto nla_put_failure;
1810	if ((all || type == NETCONFA_MC_FORWARDING) &&
1811	    nla_put_s32(skb, NETCONFA_MC_FORWARDING,
1812			IPV4_DEVCONF(*devconf, MC_FORWARDING)) < 0)
1813		goto nla_put_failure;
 
 
 
 
1814	if ((all || type == NETCONFA_PROXY_NEIGH) &&
1815	    nla_put_s32(skb, NETCONFA_PROXY_NEIGH,
1816			IPV4_DEVCONF(*devconf, PROXY_ARP)) < 0)
1817		goto nla_put_failure;
1818	if ((all || type == NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN) &&
1819	    nla_put_s32(skb, NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN,
1820			IPV4_DEVCONF(*devconf, IGNORE_ROUTES_WITH_LINKDOWN)) < 0)
1821		goto nla_put_failure;
1822
 
1823	nlmsg_end(skb, nlh);
1824	return 0;
1825
1826nla_put_failure:
1827	nlmsg_cancel(skb, nlh);
1828	return -EMSGSIZE;
1829}
1830
1831void inet_netconf_notify_devconf(struct net *net, int type, int ifindex,
1832				 struct ipv4_devconf *devconf)
1833{
1834	struct sk_buff *skb;
1835	int err = -ENOBUFS;
1836
1837	skb = nlmsg_new(inet_netconf_msgsize_devconf(type), GFP_ATOMIC);
1838	if (!skb)
1839		goto errout;
1840
1841	err = inet_netconf_fill_devconf(skb, ifindex, devconf, 0, 0,
1842					RTM_NEWNETCONF, 0, type);
1843	if (err < 0) {
1844		/* -EMSGSIZE implies BUG in inet_netconf_msgsize_devconf() */
1845		WARN_ON(err == -EMSGSIZE);
1846		kfree_skb(skb);
1847		goto errout;
1848	}
1849	rtnl_notify(skb, net, 0, RTNLGRP_IPV4_NETCONF, NULL, GFP_ATOMIC);
1850	return;
1851errout:
1852	if (err < 0)
1853		rtnl_set_sk_err(net, RTNLGRP_IPV4_NETCONF, err);
1854}
1855
1856static const struct nla_policy devconf_ipv4_policy[NETCONFA_MAX+1] = {
1857	[NETCONFA_IFINDEX]	= { .len = sizeof(int) },
1858	[NETCONFA_FORWARDING]	= { .len = sizeof(int) },
1859	[NETCONFA_RP_FILTER]	= { .len = sizeof(int) },
1860	[NETCONFA_PROXY_NEIGH]	= { .len = sizeof(int) },
1861	[NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN]	= { .len = sizeof(int) },
1862};
1863
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1864static int inet_netconf_get_devconf(struct sk_buff *in_skb,
1865				    struct nlmsghdr *nlh)
 
1866{
1867	struct net *net = sock_net(in_skb->sk);
1868	struct nlattr *tb[NETCONFA_MAX+1];
1869	struct netconfmsg *ncm;
1870	struct sk_buff *skb;
1871	struct ipv4_devconf *devconf;
1872	struct in_device *in_dev;
1873	struct net_device *dev;
1874	int ifindex;
1875	int err;
1876
1877	err = nlmsg_parse(nlh, sizeof(*ncm), tb, NETCONFA_MAX,
1878			  devconf_ipv4_policy);
1879	if (err < 0)
1880		goto errout;
1881
1882	err = -EINVAL;
1883	if (!tb[NETCONFA_IFINDEX])
1884		goto errout;
1885
1886	ifindex = nla_get_s32(tb[NETCONFA_IFINDEX]);
1887	switch (ifindex) {
1888	case NETCONFA_IFINDEX_ALL:
1889		devconf = net->ipv4.devconf_all;
1890		break;
1891	case NETCONFA_IFINDEX_DEFAULT:
1892		devconf = net->ipv4.devconf_dflt;
1893		break;
1894	default:
1895		dev = __dev_get_by_index(net, ifindex);
1896		if (!dev)
1897			goto errout;
1898		in_dev = __in_dev_get_rtnl(dev);
1899		if (!in_dev)
1900			goto errout;
1901		devconf = &in_dev->cnf;
1902		break;
1903	}
1904
1905	err = -ENOBUFS;
1906	skb = nlmsg_new(inet_netconf_msgsize_devconf(NETCONFA_ALL), GFP_ATOMIC);
1907	if (!skb)
1908		goto errout;
1909
1910	err = inet_netconf_fill_devconf(skb, ifindex, devconf,
1911					NETLINK_CB(in_skb).portid,
1912					nlh->nlmsg_seq, RTM_NEWNETCONF, 0,
1913					NETCONFA_ALL);
1914	if (err < 0) {
1915		/* -EMSGSIZE implies BUG in inet_netconf_msgsize_devconf() */
1916		WARN_ON(err == -EMSGSIZE);
1917		kfree_skb(skb);
1918		goto errout;
1919	}
1920	err = rtnl_unicast(skb, net, NETLINK_CB(in_skb).portid);
1921errout:
1922	return err;
1923}
1924
1925static int inet_netconf_dump_devconf(struct sk_buff *skb,
1926				     struct netlink_callback *cb)
1927{
 
1928	struct net *net = sock_net(skb->sk);
1929	int h, s_h;
1930	int idx, s_idx;
1931	struct net_device *dev;
1932	struct in_device *in_dev;
1933	struct hlist_head *head;
1934
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1935	s_h = cb->args[0];
1936	s_idx = idx = cb->args[1];
1937
1938	for (h = s_h; h < NETDEV_HASHENTRIES; h++, s_idx = 0) {
1939		idx = 0;
1940		head = &net->dev_index_head[h];
1941		rcu_read_lock();
1942		cb->seq = atomic_read(&net->ipv4.dev_addr_genid) ^
1943			  net->dev_base_seq;
1944		hlist_for_each_entry_rcu(dev, head, index_hlist) {
1945			if (idx < s_idx)
1946				goto cont;
1947			in_dev = __in_dev_get_rcu(dev);
1948			if (!in_dev)
1949				goto cont;
1950
1951			if (inet_netconf_fill_devconf(skb, dev->ifindex,
1952						      &in_dev->cnf,
1953						      NETLINK_CB(cb->skb).portid,
1954						      cb->nlh->nlmsg_seq,
1955						      RTM_NEWNETCONF,
1956						      NLM_F_MULTI,
1957						      NETCONFA_ALL) < 0) {
1958				rcu_read_unlock();
1959				goto done;
1960			}
1961			nl_dump_check_consistent(cb, nlmsg_hdr(skb));
1962cont:
1963			idx++;
1964		}
1965		rcu_read_unlock();
1966	}
1967	if (h == NETDEV_HASHENTRIES) {
1968		if (inet_netconf_fill_devconf(skb, NETCONFA_IFINDEX_ALL,
1969					      net->ipv4.devconf_all,
1970					      NETLINK_CB(cb->skb).portid,
1971					      cb->nlh->nlmsg_seq,
1972					      RTM_NEWNETCONF, NLM_F_MULTI,
1973					      NETCONFA_ALL) < 0)
1974			goto done;
1975		else
1976			h++;
1977	}
1978	if (h == NETDEV_HASHENTRIES + 1) {
1979		if (inet_netconf_fill_devconf(skb, NETCONFA_IFINDEX_DEFAULT,
1980					      net->ipv4.devconf_dflt,
1981					      NETLINK_CB(cb->skb).portid,
1982					      cb->nlh->nlmsg_seq,
1983					      RTM_NEWNETCONF, NLM_F_MULTI,
1984					      NETCONFA_ALL) < 0)
1985			goto done;
1986		else
1987			h++;
1988	}
1989done:
1990	cb->args[0] = h;
1991	cb->args[1] = idx;
1992
1993	return skb->len;
1994}
1995
1996#ifdef CONFIG_SYSCTL
1997
1998static void devinet_copy_dflt_conf(struct net *net, int i)
1999{
2000	struct net_device *dev;
2001
2002	rcu_read_lock();
2003	for_each_netdev_rcu(net, dev) {
2004		struct in_device *in_dev;
2005
2006		in_dev = __in_dev_get_rcu(dev);
2007		if (in_dev && !test_bit(i, in_dev->cnf.state))
2008			in_dev->cnf.data[i] = net->ipv4.devconf_dflt->data[i];
2009	}
2010	rcu_read_unlock();
2011}
2012
2013/* called with RTNL locked */
2014static void inet_forward_change(struct net *net)
2015{
2016	struct net_device *dev;
2017	int on = IPV4_DEVCONF_ALL(net, FORWARDING);
2018
2019	IPV4_DEVCONF_ALL(net, ACCEPT_REDIRECTS) = !on;
2020	IPV4_DEVCONF_DFLT(net, FORWARDING) = on;
2021	inet_netconf_notify_devconf(net, NETCONFA_FORWARDING,
 
2022				    NETCONFA_IFINDEX_ALL,
2023				    net->ipv4.devconf_all);
2024	inet_netconf_notify_devconf(net, NETCONFA_FORWARDING,
 
2025				    NETCONFA_IFINDEX_DEFAULT,
2026				    net->ipv4.devconf_dflt);
2027
2028	for_each_netdev(net, dev) {
2029		struct in_device *in_dev;
 
2030		if (on)
2031			dev_disable_lro(dev);
2032		rcu_read_lock();
2033		in_dev = __in_dev_get_rcu(dev);
2034		if (in_dev) {
2035			IN_DEV_CONF_SET(in_dev, FORWARDING, on);
2036			inet_netconf_notify_devconf(net, NETCONFA_FORWARDING,
 
2037						    dev->ifindex, &in_dev->cnf);
2038		}
2039		rcu_read_unlock();
2040	}
2041}
2042
2043static int devinet_conf_ifindex(struct net *net, struct ipv4_devconf *cnf)
2044{
2045	if (cnf == net->ipv4.devconf_dflt)
2046		return NETCONFA_IFINDEX_DEFAULT;
2047	else if (cnf == net->ipv4.devconf_all)
2048		return NETCONFA_IFINDEX_ALL;
2049	else {
2050		struct in_device *idev
2051			= container_of(cnf, struct in_device, cnf);
2052		return idev->dev->ifindex;
2053	}
2054}
2055
2056static int devinet_conf_proc(struct ctl_table *ctl, int write,
2057			     void __user *buffer,
2058			     size_t *lenp, loff_t *ppos)
2059{
2060	int old_value = *(int *)ctl->data;
2061	int ret = proc_dointvec(ctl, write, buffer, lenp, ppos);
2062	int new_value = *(int *)ctl->data;
2063
2064	if (write) {
2065		struct ipv4_devconf *cnf = ctl->extra1;
2066		struct net *net = ctl->extra2;
2067		int i = (int *)ctl->data - cnf->data;
2068		int ifindex;
2069
2070		set_bit(i, cnf->state);
2071
2072		if (cnf == net->ipv4.devconf_dflt)
2073			devinet_copy_dflt_conf(net, i);
2074		if (i == IPV4_DEVCONF_ACCEPT_LOCAL - 1 ||
2075		    i == IPV4_DEVCONF_ROUTE_LOCALNET - 1)
2076			if ((new_value == 0) && (old_value != 0))
2077				rt_cache_flush(net);
2078
 
 
 
 
2079		if (i == IPV4_DEVCONF_RP_FILTER - 1 &&
2080		    new_value != old_value) {
2081			ifindex = devinet_conf_ifindex(net, cnf);
2082			inet_netconf_notify_devconf(net, NETCONFA_RP_FILTER,
 
2083						    ifindex, cnf);
2084		}
2085		if (i == IPV4_DEVCONF_PROXY_ARP - 1 &&
2086		    new_value != old_value) {
2087			ifindex = devinet_conf_ifindex(net, cnf);
2088			inet_netconf_notify_devconf(net, NETCONFA_PROXY_NEIGH,
 
2089						    ifindex, cnf);
2090		}
2091		if (i == IPV4_DEVCONF_IGNORE_ROUTES_WITH_LINKDOWN - 1 &&
2092		    new_value != old_value) {
2093			ifindex = devinet_conf_ifindex(net, cnf);
2094			inet_netconf_notify_devconf(net, NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN,
 
2095						    ifindex, cnf);
2096		}
2097	}
2098
2099	return ret;
2100}
2101
2102static int devinet_sysctl_forward(struct ctl_table *ctl, int write,
2103				  void __user *buffer,
2104				  size_t *lenp, loff_t *ppos)
2105{
2106	int *valp = ctl->data;
2107	int val = *valp;
2108	loff_t pos = *ppos;
2109	int ret = proc_dointvec(ctl, write, buffer, lenp, ppos);
2110
2111	if (write && *valp != val) {
2112		struct net *net = ctl->extra2;
2113
2114		if (valp != &IPV4_DEVCONF_DFLT(net, FORWARDING)) {
2115			if (!rtnl_trylock()) {
2116				/* Restore the original values before restarting */
2117				*valp = val;
2118				*ppos = pos;
2119				return restart_syscall();
2120			}
2121			if (valp == &IPV4_DEVCONF_ALL(net, FORWARDING)) {
2122				inet_forward_change(net);
2123			} else {
2124				struct ipv4_devconf *cnf = ctl->extra1;
2125				struct in_device *idev =
2126					container_of(cnf, struct in_device, cnf);
2127				if (*valp)
2128					dev_disable_lro(idev->dev);
2129				inet_netconf_notify_devconf(net,
2130							    NETCONFA_FORWARDING,
2131							    idev->dev->ifindex,
2132							    cnf);
2133			}
2134			rtnl_unlock();
2135			rt_cache_flush(net);
2136		} else
2137			inet_netconf_notify_devconf(net, NETCONFA_FORWARDING,
 
2138						    NETCONFA_IFINDEX_DEFAULT,
2139						    net->ipv4.devconf_dflt);
2140	}
2141
2142	return ret;
2143}
2144
2145static int ipv4_doint_and_flush(struct ctl_table *ctl, int write,
2146				void __user *buffer,
2147				size_t *lenp, loff_t *ppos)
2148{
2149	int *valp = ctl->data;
2150	int val = *valp;
2151	int ret = proc_dointvec(ctl, write, buffer, lenp, ppos);
2152	struct net *net = ctl->extra2;
2153
2154	if (write && *valp != val)
2155		rt_cache_flush(net);
2156
2157	return ret;
2158}
2159
2160#define DEVINET_SYSCTL_ENTRY(attr, name, mval, proc) \
2161	{ \
2162		.procname	= name, \
2163		.data		= ipv4_devconf.data + \
2164				  IPV4_DEVCONF_ ## attr - 1, \
2165		.maxlen		= sizeof(int), \
2166		.mode		= mval, \
2167		.proc_handler	= proc, \
2168		.extra1		= &ipv4_devconf, \
2169	}
2170
2171#define DEVINET_SYSCTL_RW_ENTRY(attr, name) \
2172	DEVINET_SYSCTL_ENTRY(attr, name, 0644, devinet_conf_proc)
2173
2174#define DEVINET_SYSCTL_RO_ENTRY(attr, name) \
2175	DEVINET_SYSCTL_ENTRY(attr, name, 0444, devinet_conf_proc)
2176
2177#define DEVINET_SYSCTL_COMPLEX_ENTRY(attr, name, proc) \
2178	DEVINET_SYSCTL_ENTRY(attr, name, 0644, proc)
2179
2180#define DEVINET_SYSCTL_FLUSHING_ENTRY(attr, name) \
2181	DEVINET_SYSCTL_COMPLEX_ENTRY(attr, name, ipv4_doint_and_flush)
2182
2183static struct devinet_sysctl_table {
2184	struct ctl_table_header *sysctl_header;
2185	struct ctl_table devinet_vars[__IPV4_DEVCONF_MAX];
2186} devinet_sysctl = {
2187	.devinet_vars = {
2188		DEVINET_SYSCTL_COMPLEX_ENTRY(FORWARDING, "forwarding",
2189					     devinet_sysctl_forward),
2190		DEVINET_SYSCTL_RO_ENTRY(MC_FORWARDING, "mc_forwarding"),
 
2191
2192		DEVINET_SYSCTL_RW_ENTRY(ACCEPT_REDIRECTS, "accept_redirects"),
2193		DEVINET_SYSCTL_RW_ENTRY(SECURE_REDIRECTS, "secure_redirects"),
2194		DEVINET_SYSCTL_RW_ENTRY(SHARED_MEDIA, "shared_media"),
2195		DEVINET_SYSCTL_RW_ENTRY(RP_FILTER, "rp_filter"),
2196		DEVINET_SYSCTL_RW_ENTRY(SEND_REDIRECTS, "send_redirects"),
2197		DEVINET_SYSCTL_RW_ENTRY(ACCEPT_SOURCE_ROUTE,
2198					"accept_source_route"),
2199		DEVINET_SYSCTL_RW_ENTRY(ACCEPT_LOCAL, "accept_local"),
2200		DEVINET_SYSCTL_RW_ENTRY(SRC_VMARK, "src_valid_mark"),
2201		DEVINET_SYSCTL_RW_ENTRY(PROXY_ARP, "proxy_arp"),
2202		DEVINET_SYSCTL_RW_ENTRY(MEDIUM_ID, "medium_id"),
2203		DEVINET_SYSCTL_RW_ENTRY(BOOTP_RELAY, "bootp_relay"),
2204		DEVINET_SYSCTL_RW_ENTRY(LOG_MARTIANS, "log_martians"),
2205		DEVINET_SYSCTL_RW_ENTRY(TAG, "tag"),
2206		DEVINET_SYSCTL_RW_ENTRY(ARPFILTER, "arp_filter"),
2207		DEVINET_SYSCTL_RW_ENTRY(ARP_ANNOUNCE, "arp_announce"),
2208		DEVINET_SYSCTL_RW_ENTRY(ARP_IGNORE, "arp_ignore"),
2209		DEVINET_SYSCTL_RW_ENTRY(ARP_ACCEPT, "arp_accept"),
2210		DEVINET_SYSCTL_RW_ENTRY(ARP_NOTIFY, "arp_notify"),
2211		DEVINET_SYSCTL_RW_ENTRY(PROXY_ARP_PVLAN, "proxy_arp_pvlan"),
2212		DEVINET_SYSCTL_RW_ENTRY(FORCE_IGMP_VERSION,
2213					"force_igmp_version"),
2214		DEVINET_SYSCTL_RW_ENTRY(IGMPV2_UNSOLICITED_REPORT_INTERVAL,
2215					"igmpv2_unsolicited_report_interval"),
2216		DEVINET_SYSCTL_RW_ENTRY(IGMPV3_UNSOLICITED_REPORT_INTERVAL,
2217					"igmpv3_unsolicited_report_interval"),
2218		DEVINET_SYSCTL_RW_ENTRY(IGNORE_ROUTES_WITH_LINKDOWN,
2219					"ignore_routes_with_linkdown"),
2220		DEVINET_SYSCTL_RW_ENTRY(DROP_GRATUITOUS_ARP,
2221					"drop_gratuitous_arp"),
2222
2223		DEVINET_SYSCTL_FLUSHING_ENTRY(NOXFRM, "disable_xfrm"),
2224		DEVINET_SYSCTL_FLUSHING_ENTRY(NOPOLICY, "disable_policy"),
2225		DEVINET_SYSCTL_FLUSHING_ENTRY(PROMOTE_SECONDARIES,
2226					      "promote_secondaries"),
2227		DEVINET_SYSCTL_FLUSHING_ENTRY(ROUTE_LOCALNET,
2228					      "route_localnet"),
2229		DEVINET_SYSCTL_FLUSHING_ENTRY(DROP_UNICAST_IN_L2_MULTICAST,
2230					      "drop_unicast_in_l2_multicast"),
2231	},
2232};
2233
2234static int __devinet_sysctl_register(struct net *net, char *dev_name,
2235					struct ipv4_devconf *p)
2236{
2237	int i;
2238	struct devinet_sysctl_table *t;
2239	char path[sizeof("net/ipv4/conf/") + IFNAMSIZ];
2240
2241	t = kmemdup(&devinet_sysctl, sizeof(*t), GFP_KERNEL);
2242	if (!t)
2243		goto out;
2244
2245	for (i = 0; i < ARRAY_SIZE(t->devinet_vars) - 1; i++) {
2246		t->devinet_vars[i].data += (char *)p - (char *)&ipv4_devconf;
2247		t->devinet_vars[i].extra1 = p;
2248		t->devinet_vars[i].extra2 = net;
2249	}
2250
2251	snprintf(path, sizeof(path), "net/ipv4/conf/%s", dev_name);
2252
2253	t->sysctl_header = register_net_sysctl(net, path, t->devinet_vars);
2254	if (!t->sysctl_header)
2255		goto free;
2256
2257	p->sysctl = t;
 
 
 
2258	return 0;
2259
2260free:
2261	kfree(t);
2262out:
2263	return -ENOBUFS;
2264}
2265
2266static void __devinet_sysctl_unregister(struct ipv4_devconf *cnf)
 
2267{
2268	struct devinet_sysctl_table *t = cnf->sysctl;
2269
2270	if (!t)
2271		return;
 
 
 
2272
2273	cnf->sysctl = NULL;
2274	unregister_net_sysctl_table(t->sysctl_header);
2275	kfree(t);
2276}
2277
2278static int devinet_sysctl_register(struct in_device *idev)
2279{
2280	int err;
2281
2282	if (!sysctl_dev_name_is_allowed(idev->dev->name))
2283		return -EINVAL;
2284
2285	err = neigh_sysctl_register(idev->dev, idev->arp_parms, NULL);
2286	if (err)
2287		return err;
2288	err = __devinet_sysctl_register(dev_net(idev->dev), idev->dev->name,
2289					&idev->cnf);
2290	if (err)
2291		neigh_sysctl_unregister(idev->arp_parms);
2292	return err;
2293}
2294
2295static void devinet_sysctl_unregister(struct in_device *idev)
2296{
2297	__devinet_sysctl_unregister(&idev->cnf);
 
 
2298	neigh_sysctl_unregister(idev->arp_parms);
2299}
2300
2301static struct ctl_table ctl_forward_entry[] = {
2302	{
2303		.procname	= "ip_forward",
2304		.data		= &ipv4_devconf.data[
2305					IPV4_DEVCONF_FORWARDING - 1],
2306		.maxlen		= sizeof(int),
2307		.mode		= 0644,
2308		.proc_handler	= devinet_sysctl_forward,
2309		.extra1		= &ipv4_devconf,
2310		.extra2		= &init_net,
2311	},
2312	{ },
2313};
2314#endif
2315
2316static __net_init int devinet_init_net(struct net *net)
2317{
2318	int err;
2319	struct ipv4_devconf *all, *dflt;
2320#ifdef CONFIG_SYSCTL
2321	struct ctl_table *tbl = ctl_forward_entry;
2322	struct ctl_table_header *forw_hdr;
2323#endif
2324
2325	err = -ENOMEM;
2326	all = &ipv4_devconf;
2327	dflt = &ipv4_devconf_dflt;
2328
2329	if (!net_eq(net, &init_net)) {
2330		all = kmemdup(all, sizeof(ipv4_devconf), GFP_KERNEL);
2331		if (!all)
2332			goto err_alloc_all;
2333
2334		dflt = kmemdup(dflt, sizeof(ipv4_devconf_dflt), GFP_KERNEL);
2335		if (!dflt)
2336			goto err_alloc_dflt;
2337
2338#ifdef CONFIG_SYSCTL
2339		tbl = kmemdup(tbl, sizeof(ctl_forward_entry), GFP_KERNEL);
2340		if (!tbl)
2341			goto err_alloc_ctl;
2342
2343		tbl[0].data = &all->data[IPV4_DEVCONF_FORWARDING - 1];
2344		tbl[0].extra1 = all;
2345		tbl[0].extra2 = net;
2346#endif
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2347	}
2348
2349#ifdef CONFIG_SYSCTL
2350	err = __devinet_sysctl_register(net, "all", all);
2351	if (err < 0)
2352		goto err_reg_all;
2353
2354	err = __devinet_sysctl_register(net, "default", dflt);
 
2355	if (err < 0)
2356		goto err_reg_dflt;
2357
2358	err = -ENOMEM;
2359	forw_hdr = register_net_sysctl(net, "net/ipv4", tbl);
2360	if (!forw_hdr)
2361		goto err_reg_ctl;
2362	net->ipv4.forw_hdr = forw_hdr;
2363#endif
2364
2365	net->ipv4.devconf_all = all;
2366	net->ipv4.devconf_dflt = dflt;
2367	return 0;
2368
2369#ifdef CONFIG_SYSCTL
2370err_reg_ctl:
2371	__devinet_sysctl_unregister(dflt);
2372err_reg_dflt:
2373	__devinet_sysctl_unregister(all);
2374err_reg_all:
2375	if (tbl != ctl_forward_entry)
2376		kfree(tbl);
2377err_alloc_ctl:
2378#endif
2379	if (dflt != &ipv4_devconf_dflt)
2380		kfree(dflt);
2381err_alloc_dflt:
2382	if (all != &ipv4_devconf)
2383		kfree(all);
2384err_alloc_all:
2385	return err;
2386}
2387
2388static __net_exit void devinet_exit_net(struct net *net)
2389{
2390#ifdef CONFIG_SYSCTL
2391	struct ctl_table *tbl;
2392
2393	tbl = net->ipv4.forw_hdr->ctl_table_arg;
2394	unregister_net_sysctl_table(net->ipv4.forw_hdr);
2395	__devinet_sysctl_unregister(net->ipv4.devconf_dflt);
2396	__devinet_sysctl_unregister(net->ipv4.devconf_all);
 
 
2397	kfree(tbl);
2398#endif
2399	kfree(net->ipv4.devconf_dflt);
2400	kfree(net->ipv4.devconf_all);
2401}
2402
2403static __net_initdata struct pernet_operations devinet_ops = {
2404	.init = devinet_init_net,
2405	.exit = devinet_exit_net,
2406};
2407
2408static struct rtnl_af_ops inet_af_ops __read_mostly = {
2409	.family		  = AF_INET,
2410	.fill_link_af	  = inet_fill_link_af,
2411	.get_link_af_size = inet_get_link_af_size,
2412	.validate_link_af = inet_validate_link_af,
2413	.set_link_af	  = inet_set_link_af,
2414};
2415
2416void __init devinet_init(void)
2417{
2418	int i;
2419
2420	for (i = 0; i < IN4_ADDR_HSIZE; i++)
2421		INIT_HLIST_HEAD(&inet_addr_lst[i]);
2422
2423	register_pernet_subsys(&devinet_ops);
2424
2425	register_gifconf(PF_INET, inet_gifconf);
2426	register_netdevice_notifier(&ip_netdev_notifier);
2427
2428	queue_delayed_work(system_power_efficient_wq, &check_lifetime_work, 0);
2429
2430	rtnl_af_register(&inet_af_ops);
2431
2432	rtnl_register(PF_INET, RTM_NEWADDR, inet_rtm_newaddr, NULL, NULL);
2433	rtnl_register(PF_INET, RTM_DELADDR, inet_rtm_deladdr, NULL, NULL);
2434	rtnl_register(PF_INET, RTM_GETADDR, NULL, inet_dump_ifaddr, NULL);
2435	rtnl_register(PF_INET, RTM_GETNETCONF, inet_netconf_get_devconf,
2436		      inet_netconf_dump_devconf, NULL);
2437}
v5.9
   1// SPDX-License-Identifier: GPL-2.0-or-later
   2/*
   3 *	NET3	IP device support routines.
   4 *
 
 
 
 
 
   5 *	Derived from the IP parts of dev.c 1.0.19
   6 * 		Authors:	Ross Biro
   7 *				Fred N. van Kempen, <waltje@uWalt.NL.Mugnet.ORG>
   8 *				Mark Evans, <evansmp@uhura.aston.ac.uk>
   9 *
  10 *	Additional Authors:
  11 *		Alan Cox, <gw4pts@gw4pts.ampr.org>
  12 *		Alexey Kuznetsov, <kuznet@ms2.inr.ac.ru>
  13 *
  14 *	Changes:
  15 *		Alexey Kuznetsov:	pa_* fields are replaced with ifaddr
  16 *					lists.
  17 *		Cyrus Durgin:		updated for kmod
  18 *		Matthias Andree:	in devinet_ioctl, compare label and
  19 *					address (4.4BSD alias style support),
  20 *					fall back to comparing just the label
  21 *					if no match found.
  22 */
  23
  24
  25#include <linux/uaccess.h>
  26#include <linux/bitops.h>
  27#include <linux/capability.h>
  28#include <linux/module.h>
  29#include <linux/types.h>
  30#include <linux/kernel.h>
  31#include <linux/sched/signal.h>
  32#include <linux/string.h>
  33#include <linux/mm.h>
  34#include <linux/socket.h>
  35#include <linux/sockios.h>
  36#include <linux/in.h>
  37#include <linux/errno.h>
  38#include <linux/interrupt.h>
  39#include <linux/if_addr.h>
  40#include <linux/if_ether.h>
  41#include <linux/inet.h>
  42#include <linux/netdevice.h>
  43#include <linux/etherdevice.h>
  44#include <linux/skbuff.h>
  45#include <linux/init.h>
  46#include <linux/notifier.h>
  47#include <linux/inetdevice.h>
  48#include <linux/igmp.h>
  49#include <linux/slab.h>
  50#include <linux/hash.h>
  51#ifdef CONFIG_SYSCTL
  52#include <linux/sysctl.h>
  53#endif
  54#include <linux/kmod.h>
  55#include <linux/netconf.h>
  56
  57#include <net/arp.h>
  58#include <net/ip.h>
  59#include <net/route.h>
  60#include <net/ip_fib.h>
  61#include <net/rtnetlink.h>
  62#include <net/net_namespace.h>
  63#include <net/addrconf.h>
  64
  65#define IPV6ONLY_FLAGS	\
  66		(IFA_F_NODAD | IFA_F_OPTIMISTIC | IFA_F_DADFAILED | \
  67		 IFA_F_HOMEADDRESS | IFA_F_TENTATIVE | \
  68		 IFA_F_MANAGETEMPADDR | IFA_F_STABLE_PRIVACY)
  69
  70static struct ipv4_devconf ipv4_devconf = {
  71	.data = {
  72		[IPV4_DEVCONF_ACCEPT_REDIRECTS - 1] = 1,
  73		[IPV4_DEVCONF_SEND_REDIRECTS - 1] = 1,
  74		[IPV4_DEVCONF_SECURE_REDIRECTS - 1] = 1,
  75		[IPV4_DEVCONF_SHARED_MEDIA - 1] = 1,
  76		[IPV4_DEVCONF_IGMPV2_UNSOLICITED_REPORT_INTERVAL - 1] = 10000 /*ms*/,
  77		[IPV4_DEVCONF_IGMPV3_UNSOLICITED_REPORT_INTERVAL - 1] =  1000 /*ms*/,
  78	},
  79};
  80
  81static struct ipv4_devconf ipv4_devconf_dflt = {
  82	.data = {
  83		[IPV4_DEVCONF_ACCEPT_REDIRECTS - 1] = 1,
  84		[IPV4_DEVCONF_SEND_REDIRECTS - 1] = 1,
  85		[IPV4_DEVCONF_SECURE_REDIRECTS - 1] = 1,
  86		[IPV4_DEVCONF_SHARED_MEDIA - 1] = 1,
  87		[IPV4_DEVCONF_ACCEPT_SOURCE_ROUTE - 1] = 1,
  88		[IPV4_DEVCONF_IGMPV2_UNSOLICITED_REPORT_INTERVAL - 1] = 10000 /*ms*/,
  89		[IPV4_DEVCONF_IGMPV3_UNSOLICITED_REPORT_INTERVAL - 1] =  1000 /*ms*/,
  90	},
  91};
  92
  93#define IPV4_DEVCONF_DFLT(net, attr) \
  94	IPV4_DEVCONF((*net->ipv4.devconf_dflt), attr)
  95
  96static const struct nla_policy ifa_ipv4_policy[IFA_MAX+1] = {
  97	[IFA_LOCAL]     	= { .type = NLA_U32 },
  98	[IFA_ADDRESS]   	= { .type = NLA_U32 },
  99	[IFA_BROADCAST] 	= { .type = NLA_U32 },
 100	[IFA_LABEL]     	= { .type = NLA_STRING, .len = IFNAMSIZ - 1 },
 101	[IFA_CACHEINFO]		= { .len = sizeof(struct ifa_cacheinfo) },
 102	[IFA_FLAGS]		= { .type = NLA_U32 },
 103	[IFA_RT_PRIORITY]	= { .type = NLA_U32 },
 104	[IFA_TARGET_NETNSID]	= { .type = NLA_S32 },
 105};
 106
 107struct inet_fill_args {
 108	u32 portid;
 109	u32 seq;
 110	int event;
 111	unsigned int flags;
 112	int netnsid;
 113	int ifindex;
 114};
 115
 116#define IN4_ADDR_HSIZE_SHIFT	8
 117#define IN4_ADDR_HSIZE		(1U << IN4_ADDR_HSIZE_SHIFT)
 118
 119static struct hlist_head inet_addr_lst[IN4_ADDR_HSIZE];
 120
 121static u32 inet_addr_hash(const struct net *net, __be32 addr)
 122{
 123	u32 val = (__force u32) addr ^ net_hash_mix(net);
 124
 125	return hash_32(val, IN4_ADDR_HSIZE_SHIFT);
 126}
 127
 128static void inet_hash_insert(struct net *net, struct in_ifaddr *ifa)
 129{
 130	u32 hash = inet_addr_hash(net, ifa->ifa_local);
 131
 132	ASSERT_RTNL();
 133	hlist_add_head_rcu(&ifa->hash, &inet_addr_lst[hash]);
 134}
 135
 136static void inet_hash_remove(struct in_ifaddr *ifa)
 137{
 138	ASSERT_RTNL();
 139	hlist_del_init_rcu(&ifa->hash);
 140}
 141
 142/**
 143 * __ip_dev_find - find the first device with a given source address.
 144 * @net: the net namespace
 145 * @addr: the source address
 146 * @devref: if true, take a reference on the found device
 147 *
 148 * If a caller uses devref=false, it should be protected by RCU, or RTNL
 149 */
 150struct net_device *__ip_dev_find(struct net *net, __be32 addr, bool devref)
 151{
 
 152	struct net_device *result = NULL;
 153	struct in_ifaddr *ifa;
 154
 155	rcu_read_lock();
 156	ifa = inet_lookup_ifaddr_rcu(net, addr);
 157	if (!ifa) {
 
 
 
 
 
 
 
 
 
 158		struct flowi4 fl4 = { .daddr = addr };
 159		struct fib_result res = { 0 };
 160		struct fib_table *local;
 161
 162		/* Fallback to FIB local table so that communication
 163		 * over loopback subnets work.
 164		 */
 165		local = fib_get_table(net, RT_TABLE_LOCAL);
 166		if (local &&
 167		    !fib_table_lookup(local, &fl4, &res, FIB_LOOKUP_NOREF) &&
 168		    res.type == RTN_LOCAL)
 169			result = FIB_RES_DEV(res);
 170	} else {
 171		result = ifa->ifa_dev->dev;
 172	}
 173	if (result && devref)
 174		dev_hold(result);
 175	rcu_read_unlock();
 176	return result;
 177}
 178EXPORT_SYMBOL(__ip_dev_find);
 179
 180/* called under RCU lock */
 181struct in_ifaddr *inet_lookup_ifaddr_rcu(struct net *net, __be32 addr)
 182{
 183	u32 hash = inet_addr_hash(net, addr);
 184	struct in_ifaddr *ifa;
 185
 186	hlist_for_each_entry_rcu(ifa, &inet_addr_lst[hash], hash)
 187		if (ifa->ifa_local == addr &&
 188		    net_eq(dev_net(ifa->ifa_dev->dev), net))
 189			return ifa;
 190
 191	return NULL;
 192}
 193
 194static void rtmsg_ifa(int event, struct in_ifaddr *, struct nlmsghdr *, u32);
 195
 196static BLOCKING_NOTIFIER_HEAD(inetaddr_chain);
 197static BLOCKING_NOTIFIER_HEAD(inetaddr_validator_chain);
 198static void inet_del_ifa(struct in_device *in_dev,
 199			 struct in_ifaddr __rcu **ifap,
 200			 int destroy);
 201#ifdef CONFIG_SYSCTL
 202static int devinet_sysctl_register(struct in_device *idev);
 203static void devinet_sysctl_unregister(struct in_device *idev);
 204#else
 205static int devinet_sysctl_register(struct in_device *idev)
 206{
 207	return 0;
 208}
 209static void devinet_sysctl_unregister(struct in_device *idev)
 210{
 211}
 212#endif
 213
 214/* Locks all the inet devices. */
 215
 216static struct in_ifaddr *inet_alloc_ifa(void)
 217{
 218	return kzalloc(sizeof(struct in_ifaddr), GFP_KERNEL);
 219}
 220
 221static void inet_rcu_free_ifa(struct rcu_head *head)
 222{
 223	struct in_ifaddr *ifa = container_of(head, struct in_ifaddr, rcu_head);
 224	if (ifa->ifa_dev)
 225		in_dev_put(ifa->ifa_dev);
 226	kfree(ifa);
 227}
 228
 229static void inet_free_ifa(struct in_ifaddr *ifa)
 230{
 231	call_rcu(&ifa->rcu_head, inet_rcu_free_ifa);
 232}
 233
 234void in_dev_finish_destroy(struct in_device *idev)
 235{
 236	struct net_device *dev = idev->dev;
 237
 238	WARN_ON(idev->ifa_list);
 239	WARN_ON(idev->mc_list);
 240	kfree(rcu_dereference_protected(idev->mc_hash, 1));
 241#ifdef NET_REFCNT_DEBUG
 242	pr_debug("%s: %p=%s\n", __func__, idev, dev ? dev->name : "NIL");
 243#endif
 244	dev_put(dev);
 245	if (!idev->dead)
 246		pr_err("Freeing alive in_device %p\n", idev);
 247	else
 248		kfree(idev);
 249}
 250EXPORT_SYMBOL(in_dev_finish_destroy);
 251
 252static struct in_device *inetdev_init(struct net_device *dev)
 253{
 254	struct in_device *in_dev;
 255	int err = -ENOMEM;
 256
 257	ASSERT_RTNL();
 258
 259	in_dev = kzalloc(sizeof(*in_dev), GFP_KERNEL);
 260	if (!in_dev)
 261		goto out;
 262	memcpy(&in_dev->cnf, dev_net(dev)->ipv4.devconf_dflt,
 263			sizeof(in_dev->cnf));
 264	in_dev->cnf.sysctl = NULL;
 265	in_dev->dev = dev;
 266	in_dev->arp_parms = neigh_parms_alloc(dev, &arp_tbl);
 267	if (!in_dev->arp_parms)
 268		goto out_kfree;
 269	if (IPV4_DEVCONF(in_dev->cnf, FORWARDING))
 270		dev_disable_lro(dev);
 271	/* Reference in_dev->dev */
 272	dev_hold(dev);
 273	/* Account for reference dev->ip_ptr (below) */
 274	refcount_set(&in_dev->refcnt, 1);
 275
 276	err = devinet_sysctl_register(in_dev);
 277	if (err) {
 278		in_dev->dead = 1;
 279		neigh_parms_release(&arp_tbl, in_dev->arp_parms);
 280		in_dev_put(in_dev);
 281		in_dev = NULL;
 282		goto out;
 283	}
 284	ip_mc_init_dev(in_dev);
 285	if (dev->flags & IFF_UP)
 286		ip_mc_up(in_dev);
 287
 288	/* we can receive as soon as ip_ptr is set -- do this last */
 289	rcu_assign_pointer(dev->ip_ptr, in_dev);
 290out:
 291	return in_dev ?: ERR_PTR(err);
 292out_kfree:
 293	kfree(in_dev);
 294	in_dev = NULL;
 295	goto out;
 296}
 297
 298static void in_dev_rcu_put(struct rcu_head *head)
 299{
 300	struct in_device *idev = container_of(head, struct in_device, rcu_head);
 301	in_dev_put(idev);
 302}
 303
 304static void inetdev_destroy(struct in_device *in_dev)
 305{
 
 306	struct net_device *dev;
 307	struct in_ifaddr *ifa;
 308
 309	ASSERT_RTNL();
 310
 311	dev = in_dev->dev;
 312
 313	in_dev->dead = 1;
 314
 315	ip_mc_destroy_dev(in_dev);
 316
 317	while ((ifa = rtnl_dereference(in_dev->ifa_list)) != NULL) {
 318		inet_del_ifa(in_dev, &in_dev->ifa_list, 0);
 319		inet_free_ifa(ifa);
 320	}
 321
 322	RCU_INIT_POINTER(dev->ip_ptr, NULL);
 323
 324	devinet_sysctl_unregister(in_dev);
 325	neigh_parms_release(&arp_tbl, in_dev->arp_parms);
 326	arp_ifdown(dev);
 327
 328	call_rcu(&in_dev->rcu_head, in_dev_rcu_put);
 329}
 330
 331int inet_addr_onlink(struct in_device *in_dev, __be32 a, __be32 b)
 332{
 333	const struct in_ifaddr *ifa;
 334
 335	rcu_read_lock();
 336	in_dev_for_each_ifa_rcu(ifa, in_dev) {
 337		if (inet_ifa_match(a, ifa)) {
 338			if (!b || inet_ifa_match(b, ifa)) {
 339				rcu_read_unlock();
 340				return 1;
 341			}
 342		}
 343	}
 344	rcu_read_unlock();
 345	return 0;
 346}
 347
 348static void __inet_del_ifa(struct in_device *in_dev,
 349			   struct in_ifaddr __rcu **ifap,
 350			   int destroy, struct nlmsghdr *nlh, u32 portid)
 351{
 352	struct in_ifaddr *promote = NULL;
 353	struct in_ifaddr *ifa, *ifa1;
 354	struct in_ifaddr *last_prim;
 355	struct in_ifaddr *prev_prom = NULL;
 356	int do_promote = IN_DEV_PROMOTE_SECONDARIES(in_dev);
 357
 358	ASSERT_RTNL();
 359
 360	ifa1 = rtnl_dereference(*ifap);
 361	last_prim = rtnl_dereference(in_dev->ifa_list);
 362	if (in_dev->dead)
 363		goto no_promotions;
 364
 365	/* 1. Deleting primary ifaddr forces deletion all secondaries
 366	 * unless alias promotion is set
 367	 **/
 368
 369	if (!(ifa1->ifa_flags & IFA_F_SECONDARY)) {
 370		struct in_ifaddr __rcu **ifap1 = &ifa1->ifa_next;
 371
 372		while ((ifa = rtnl_dereference(*ifap1)) != NULL) {
 373			if (!(ifa->ifa_flags & IFA_F_SECONDARY) &&
 374			    ifa1->ifa_scope <= ifa->ifa_scope)
 375				last_prim = ifa;
 376
 377			if (!(ifa->ifa_flags & IFA_F_SECONDARY) ||
 378			    ifa1->ifa_mask != ifa->ifa_mask ||
 379			    !inet_ifa_match(ifa1->ifa_address, ifa)) {
 380				ifap1 = &ifa->ifa_next;
 381				prev_prom = ifa;
 382				continue;
 383			}
 384
 385			if (!do_promote) {
 386				inet_hash_remove(ifa);
 387				*ifap1 = ifa->ifa_next;
 388
 389				rtmsg_ifa(RTM_DELADDR, ifa, nlh, portid);
 390				blocking_notifier_call_chain(&inetaddr_chain,
 391						NETDEV_DOWN, ifa);
 392				inet_free_ifa(ifa);
 393			} else {
 394				promote = ifa;
 395				break;
 396			}
 397		}
 398	}
 399
 400	/* On promotion all secondaries from subnet are changing
 401	 * the primary IP, we must remove all their routes silently
 402	 * and later to add them back with new prefsrc. Do this
 403	 * while all addresses are on the device list.
 404	 */
 405	for (ifa = promote; ifa; ifa = rtnl_dereference(ifa->ifa_next)) {
 406		if (ifa1->ifa_mask == ifa->ifa_mask &&
 407		    inet_ifa_match(ifa1->ifa_address, ifa))
 408			fib_del_ifaddr(ifa, ifa1);
 409	}
 410
 411no_promotions:
 412	/* 2. Unlink it */
 413
 414	*ifap = ifa1->ifa_next;
 415	inet_hash_remove(ifa1);
 416
 417	/* 3. Announce address deletion */
 418
 419	/* Send message first, then call notifier.
 420	   At first sight, FIB update triggered by notifier
 421	   will refer to already deleted ifaddr, that could confuse
 422	   netlink listeners. It is not true: look, gated sees
 423	   that route deleted and if it still thinks that ifaddr
 424	   is valid, it will try to restore deleted routes... Grr.
 425	   So that, this order is correct.
 426	 */
 427	rtmsg_ifa(RTM_DELADDR, ifa1, nlh, portid);
 428	blocking_notifier_call_chain(&inetaddr_chain, NETDEV_DOWN, ifa1);
 429
 430	if (promote) {
 431		struct in_ifaddr *next_sec;
 432
 433		next_sec = rtnl_dereference(promote->ifa_next);
 434		if (prev_prom) {
 435			struct in_ifaddr *last_sec;
 436
 437			rcu_assign_pointer(prev_prom->ifa_next, next_sec);
 438
 439			last_sec = rtnl_dereference(last_prim->ifa_next);
 440			rcu_assign_pointer(promote->ifa_next, last_sec);
 441			rcu_assign_pointer(last_prim->ifa_next, promote);
 442		}
 443
 444		promote->ifa_flags &= ~IFA_F_SECONDARY;
 445		rtmsg_ifa(RTM_NEWADDR, promote, nlh, portid);
 446		blocking_notifier_call_chain(&inetaddr_chain,
 447				NETDEV_UP, promote);
 448		for (ifa = next_sec; ifa;
 449		     ifa = rtnl_dereference(ifa->ifa_next)) {
 450			if (ifa1->ifa_mask != ifa->ifa_mask ||
 451			    !inet_ifa_match(ifa1->ifa_address, ifa))
 452					continue;
 453			fib_add_ifaddr(ifa);
 454		}
 455
 456	}
 457	if (destroy)
 458		inet_free_ifa(ifa1);
 459}
 460
 461static void inet_del_ifa(struct in_device *in_dev,
 462			 struct in_ifaddr __rcu **ifap,
 463			 int destroy)
 464{
 465	__inet_del_ifa(in_dev, ifap, destroy, NULL, 0);
 466}
 467
 468static void check_lifetime(struct work_struct *work);
 469
 470static DECLARE_DELAYED_WORK(check_lifetime_work, check_lifetime);
 471
 472static int __inet_insert_ifa(struct in_ifaddr *ifa, struct nlmsghdr *nlh,
 473			     u32 portid, struct netlink_ext_ack *extack)
 474{
 475	struct in_ifaddr __rcu **last_primary, **ifap;
 476	struct in_device *in_dev = ifa->ifa_dev;
 477	struct in_validator_info ivi;
 478	struct in_ifaddr *ifa1;
 479	int ret;
 480
 481	ASSERT_RTNL();
 482
 483	if (!ifa->ifa_local) {
 484		inet_free_ifa(ifa);
 485		return 0;
 486	}
 487
 488	ifa->ifa_flags &= ~IFA_F_SECONDARY;
 489	last_primary = &in_dev->ifa_list;
 490
 491	/* Don't set IPv6 only flags to IPv4 addresses */
 492	ifa->ifa_flags &= ~IPV6ONLY_FLAGS;
 493
 494	ifap = &in_dev->ifa_list;
 495	ifa1 = rtnl_dereference(*ifap);
 496
 497	while (ifa1) {
 498		if (!(ifa1->ifa_flags & IFA_F_SECONDARY) &&
 499		    ifa->ifa_scope <= ifa1->ifa_scope)
 500			last_primary = &ifa1->ifa_next;
 501		if (ifa1->ifa_mask == ifa->ifa_mask &&
 502		    inet_ifa_match(ifa1->ifa_address, ifa)) {
 503			if (ifa1->ifa_local == ifa->ifa_local) {
 504				inet_free_ifa(ifa);
 505				return -EEXIST;
 506			}
 507			if (ifa1->ifa_scope != ifa->ifa_scope) {
 508				inet_free_ifa(ifa);
 509				return -EINVAL;
 510			}
 511			ifa->ifa_flags |= IFA_F_SECONDARY;
 512		}
 513
 514		ifap = &ifa1->ifa_next;
 515		ifa1 = rtnl_dereference(*ifap);
 516	}
 517
 518	/* Allow any devices that wish to register ifaddr validtors to weigh
 519	 * in now, before changes are committed.  The rntl lock is serializing
 520	 * access here, so the state should not change between a validator call
 521	 * and a final notify on commit.  This isn't invoked on promotion under
 522	 * the assumption that validators are checking the address itself, and
 523	 * not the flags.
 524	 */
 525	ivi.ivi_addr = ifa->ifa_address;
 526	ivi.ivi_dev = ifa->ifa_dev;
 527	ivi.extack = extack;
 528	ret = blocking_notifier_call_chain(&inetaddr_validator_chain,
 529					   NETDEV_UP, &ivi);
 530	ret = notifier_to_errno(ret);
 531	if (ret) {
 532		inet_free_ifa(ifa);
 533		return ret;
 534	}
 535
 536	if (!(ifa->ifa_flags & IFA_F_SECONDARY)) {
 537		prandom_seed((__force u32) ifa->ifa_local);
 538		ifap = last_primary;
 539	}
 540
 541	rcu_assign_pointer(ifa->ifa_next, *ifap);
 542	rcu_assign_pointer(*ifap, ifa);
 543
 544	inet_hash_insert(dev_net(in_dev->dev), ifa);
 545
 546	cancel_delayed_work(&check_lifetime_work);
 547	queue_delayed_work(system_power_efficient_wq, &check_lifetime_work, 0);
 548
 549	/* Send message first, then call notifier.
 550	   Notifier will trigger FIB update, so that
 551	   listeners of netlink will know about new ifaddr */
 552	rtmsg_ifa(RTM_NEWADDR, ifa, nlh, portid);
 553	blocking_notifier_call_chain(&inetaddr_chain, NETDEV_UP, ifa);
 554
 555	return 0;
 556}
 557
 558static int inet_insert_ifa(struct in_ifaddr *ifa)
 559{
 560	return __inet_insert_ifa(ifa, NULL, 0, NULL);
 561}
 562
 563static int inet_set_ifa(struct net_device *dev, struct in_ifaddr *ifa)
 564{
 565	struct in_device *in_dev = __in_dev_get_rtnl(dev);
 566
 567	ASSERT_RTNL();
 568
 569	if (!in_dev) {
 570		inet_free_ifa(ifa);
 571		return -ENOBUFS;
 572	}
 573	ipv4_devconf_setall(in_dev);
 574	neigh_parms_data_state_setall(in_dev->arp_parms);
 575	if (ifa->ifa_dev != in_dev) {
 576		WARN_ON(ifa->ifa_dev);
 577		in_dev_hold(in_dev);
 578		ifa->ifa_dev = in_dev;
 579	}
 580	if (ipv4_is_loopback(ifa->ifa_local))
 581		ifa->ifa_scope = RT_SCOPE_HOST;
 582	return inet_insert_ifa(ifa);
 583}
 584
 585/* Caller must hold RCU or RTNL :
 586 * We dont take a reference on found in_device
 587 */
 588struct in_device *inetdev_by_index(struct net *net, int ifindex)
 589{
 590	struct net_device *dev;
 591	struct in_device *in_dev = NULL;
 592
 593	rcu_read_lock();
 594	dev = dev_get_by_index_rcu(net, ifindex);
 595	if (dev)
 596		in_dev = rcu_dereference_rtnl(dev->ip_ptr);
 597	rcu_read_unlock();
 598	return in_dev;
 599}
 600EXPORT_SYMBOL(inetdev_by_index);
 601
 602/* Called only from RTNL semaphored context. No locks. */
 603
 604struct in_ifaddr *inet_ifa_byprefix(struct in_device *in_dev, __be32 prefix,
 605				    __be32 mask)
 606{
 607	struct in_ifaddr *ifa;
 608
 609	ASSERT_RTNL();
 610
 611	in_dev_for_each_ifa_rtnl(ifa, in_dev) {
 612		if (ifa->ifa_mask == mask && inet_ifa_match(prefix, ifa))
 613			return ifa;
 614	}
 615	return NULL;
 616}
 617
 618static int ip_mc_autojoin_config(struct net *net, bool join,
 619				 const struct in_ifaddr *ifa)
 620{
 621#if defined(CONFIG_IP_MULTICAST)
 622	struct ip_mreqn mreq = {
 623		.imr_multiaddr.s_addr = ifa->ifa_address,
 624		.imr_ifindex = ifa->ifa_dev->dev->ifindex,
 625	};
 626	struct sock *sk = net->ipv4.mc_autojoin_sk;
 627	int ret;
 628
 629	ASSERT_RTNL();
 630
 631	lock_sock(sk);
 632	if (join)
 633		ret = ip_mc_join_group(sk, &mreq);
 634	else
 635		ret = ip_mc_leave_group(sk, &mreq);
 636	release_sock(sk);
 637
 638	return ret;
 639#else
 640	return -EOPNOTSUPP;
 641#endif
 642}
 643
 644static int inet_rtm_deladdr(struct sk_buff *skb, struct nlmsghdr *nlh,
 645			    struct netlink_ext_ack *extack)
 646{
 647	struct net *net = sock_net(skb->sk);
 648	struct in_ifaddr __rcu **ifap;
 649	struct nlattr *tb[IFA_MAX+1];
 650	struct in_device *in_dev;
 651	struct ifaddrmsg *ifm;
 652	struct in_ifaddr *ifa;
 653
 654	int err = -EINVAL;
 655
 656	ASSERT_RTNL();
 657
 658	err = nlmsg_parse_deprecated(nlh, sizeof(*ifm), tb, IFA_MAX,
 659				     ifa_ipv4_policy, extack);
 660	if (err < 0)
 661		goto errout;
 662
 663	ifm = nlmsg_data(nlh);
 664	in_dev = inetdev_by_index(net, ifm->ifa_index);
 665	if (!in_dev) {
 666		err = -ENODEV;
 667		goto errout;
 668	}
 669
 670	for (ifap = &in_dev->ifa_list; (ifa = rtnl_dereference(*ifap)) != NULL;
 671	     ifap = &ifa->ifa_next) {
 672		if (tb[IFA_LOCAL] &&
 673		    ifa->ifa_local != nla_get_in_addr(tb[IFA_LOCAL]))
 674			continue;
 675
 676		if (tb[IFA_LABEL] && nla_strcmp(tb[IFA_LABEL], ifa->ifa_label))
 677			continue;
 678
 679		if (tb[IFA_ADDRESS] &&
 680		    (ifm->ifa_prefixlen != ifa->ifa_prefixlen ||
 681		    !inet_ifa_match(nla_get_in_addr(tb[IFA_ADDRESS]), ifa)))
 682			continue;
 683
 684		if (ipv4_is_multicast(ifa->ifa_address))
 685			ip_mc_autojoin_config(net, false, ifa);
 686		__inet_del_ifa(in_dev, ifap, 1, nlh, NETLINK_CB(skb).portid);
 687		return 0;
 688	}
 689
 690	err = -EADDRNOTAVAIL;
 691errout:
 692	return err;
 693}
 694
 695#define INFINITY_LIFE_TIME	0xFFFFFFFF
 696
 697static void check_lifetime(struct work_struct *work)
 698{
 699	unsigned long now, next, next_sec, next_sched;
 700	struct in_ifaddr *ifa;
 701	struct hlist_node *n;
 702	int i;
 703
 704	now = jiffies;
 705	next = round_jiffies_up(now + ADDR_CHECK_FREQUENCY);
 706
 707	for (i = 0; i < IN4_ADDR_HSIZE; i++) {
 708		bool change_needed = false;
 709
 710		rcu_read_lock();
 711		hlist_for_each_entry_rcu(ifa, &inet_addr_lst[i], hash) {
 712			unsigned long age;
 713
 714			if (ifa->ifa_flags & IFA_F_PERMANENT)
 715				continue;
 716
 717			/* We try to batch several events at once. */
 718			age = (now - ifa->ifa_tstamp +
 719			       ADDRCONF_TIMER_FUZZ_MINUS) / HZ;
 720
 721			if (ifa->ifa_valid_lft != INFINITY_LIFE_TIME &&
 722			    age >= ifa->ifa_valid_lft) {
 723				change_needed = true;
 724			} else if (ifa->ifa_preferred_lft ==
 725				   INFINITY_LIFE_TIME) {
 726				continue;
 727			} else if (age >= ifa->ifa_preferred_lft) {
 728				if (time_before(ifa->ifa_tstamp +
 729						ifa->ifa_valid_lft * HZ, next))
 730					next = ifa->ifa_tstamp +
 731					       ifa->ifa_valid_lft * HZ;
 732
 733				if (!(ifa->ifa_flags & IFA_F_DEPRECATED))
 734					change_needed = true;
 735			} else if (time_before(ifa->ifa_tstamp +
 736					       ifa->ifa_preferred_lft * HZ,
 737					       next)) {
 738				next = ifa->ifa_tstamp +
 739				       ifa->ifa_preferred_lft * HZ;
 740			}
 741		}
 742		rcu_read_unlock();
 743		if (!change_needed)
 744			continue;
 745		rtnl_lock();
 746		hlist_for_each_entry_safe(ifa, n, &inet_addr_lst[i], hash) {
 747			unsigned long age;
 748
 749			if (ifa->ifa_flags & IFA_F_PERMANENT)
 750				continue;
 751
 752			/* We try to batch several events at once. */
 753			age = (now - ifa->ifa_tstamp +
 754			       ADDRCONF_TIMER_FUZZ_MINUS) / HZ;
 755
 756			if (ifa->ifa_valid_lft != INFINITY_LIFE_TIME &&
 757			    age >= ifa->ifa_valid_lft) {
 758				struct in_ifaddr __rcu **ifap;
 759				struct in_ifaddr *tmp;
 760
 761				ifap = &ifa->ifa_dev->ifa_list;
 762				tmp = rtnl_dereference(*ifap);
 763				while (tmp) {
 764					if (tmp == ifa) {
 765						inet_del_ifa(ifa->ifa_dev,
 766							     ifap, 1);
 767						break;
 768					}
 769					ifap = &tmp->ifa_next;
 770					tmp = rtnl_dereference(*ifap);
 771				}
 772			} else if (ifa->ifa_preferred_lft !=
 773				   INFINITY_LIFE_TIME &&
 774				   age >= ifa->ifa_preferred_lft &&
 775				   !(ifa->ifa_flags & IFA_F_DEPRECATED)) {
 776				ifa->ifa_flags |= IFA_F_DEPRECATED;
 777				rtmsg_ifa(RTM_NEWADDR, ifa, NULL, 0);
 778			}
 779		}
 780		rtnl_unlock();
 781	}
 782
 783	next_sec = round_jiffies_up(next);
 784	next_sched = next;
 785
 786	/* If rounded timeout is accurate enough, accept it. */
 787	if (time_before(next_sec, next + ADDRCONF_TIMER_FUZZ))
 788		next_sched = next_sec;
 789
 790	now = jiffies;
 791	/* And minimum interval is ADDRCONF_TIMER_FUZZ_MAX. */
 792	if (time_before(next_sched, now + ADDRCONF_TIMER_FUZZ_MAX))
 793		next_sched = now + ADDRCONF_TIMER_FUZZ_MAX;
 794
 795	queue_delayed_work(system_power_efficient_wq, &check_lifetime_work,
 796			next_sched - now);
 797}
 798
 799static void set_ifa_lifetime(struct in_ifaddr *ifa, __u32 valid_lft,
 800			     __u32 prefered_lft)
 801{
 802	unsigned long timeout;
 803
 804	ifa->ifa_flags &= ~(IFA_F_PERMANENT | IFA_F_DEPRECATED);
 805
 806	timeout = addrconf_timeout_fixup(valid_lft, HZ);
 807	if (addrconf_finite_timeout(timeout))
 808		ifa->ifa_valid_lft = timeout;
 809	else
 810		ifa->ifa_flags |= IFA_F_PERMANENT;
 811
 812	timeout = addrconf_timeout_fixup(prefered_lft, HZ);
 813	if (addrconf_finite_timeout(timeout)) {
 814		if (timeout == 0)
 815			ifa->ifa_flags |= IFA_F_DEPRECATED;
 816		ifa->ifa_preferred_lft = timeout;
 817	}
 818	ifa->ifa_tstamp = jiffies;
 819	if (!ifa->ifa_cstamp)
 820		ifa->ifa_cstamp = ifa->ifa_tstamp;
 821}
 822
 823static struct in_ifaddr *rtm_to_ifaddr(struct net *net, struct nlmsghdr *nlh,
 824				       __u32 *pvalid_lft, __u32 *pprefered_lft,
 825				       struct netlink_ext_ack *extack)
 826{
 827	struct nlattr *tb[IFA_MAX+1];
 828	struct in_ifaddr *ifa;
 829	struct ifaddrmsg *ifm;
 830	struct net_device *dev;
 831	struct in_device *in_dev;
 832	int err;
 833
 834	err = nlmsg_parse_deprecated(nlh, sizeof(*ifm), tb, IFA_MAX,
 835				     ifa_ipv4_policy, extack);
 836	if (err < 0)
 837		goto errout;
 838
 839	ifm = nlmsg_data(nlh);
 840	err = -EINVAL;
 841	if (ifm->ifa_prefixlen > 32 || !tb[IFA_LOCAL])
 842		goto errout;
 843
 844	dev = __dev_get_by_index(net, ifm->ifa_index);
 845	err = -ENODEV;
 846	if (!dev)
 847		goto errout;
 848
 849	in_dev = __in_dev_get_rtnl(dev);
 850	err = -ENOBUFS;
 851	if (!in_dev)
 852		goto errout;
 853
 854	ifa = inet_alloc_ifa();
 855	if (!ifa)
 856		/*
 857		 * A potential indev allocation can be left alive, it stays
 858		 * assigned to its device and is destroy with it.
 859		 */
 860		goto errout;
 861
 862	ipv4_devconf_setall(in_dev);
 863	neigh_parms_data_state_setall(in_dev->arp_parms);
 864	in_dev_hold(in_dev);
 865
 866	if (!tb[IFA_ADDRESS])
 867		tb[IFA_ADDRESS] = tb[IFA_LOCAL];
 868
 869	INIT_HLIST_NODE(&ifa->hash);
 870	ifa->ifa_prefixlen = ifm->ifa_prefixlen;
 871	ifa->ifa_mask = inet_make_mask(ifm->ifa_prefixlen);
 872	ifa->ifa_flags = tb[IFA_FLAGS] ? nla_get_u32(tb[IFA_FLAGS]) :
 873					 ifm->ifa_flags;
 874	ifa->ifa_scope = ifm->ifa_scope;
 875	ifa->ifa_dev = in_dev;
 876
 877	ifa->ifa_local = nla_get_in_addr(tb[IFA_LOCAL]);
 878	ifa->ifa_address = nla_get_in_addr(tb[IFA_ADDRESS]);
 879
 880	if (tb[IFA_BROADCAST])
 881		ifa->ifa_broadcast = nla_get_in_addr(tb[IFA_BROADCAST]);
 882
 883	if (tb[IFA_LABEL])
 884		nla_strlcpy(ifa->ifa_label, tb[IFA_LABEL], IFNAMSIZ);
 885	else
 886		memcpy(ifa->ifa_label, dev->name, IFNAMSIZ);
 887
 888	if (tb[IFA_RT_PRIORITY])
 889		ifa->ifa_rt_priority = nla_get_u32(tb[IFA_RT_PRIORITY]);
 890
 891	if (tb[IFA_CACHEINFO]) {
 892		struct ifa_cacheinfo *ci;
 893
 894		ci = nla_data(tb[IFA_CACHEINFO]);
 895		if (!ci->ifa_valid || ci->ifa_prefered > ci->ifa_valid) {
 896			err = -EINVAL;
 897			goto errout_free;
 898		}
 899		*pvalid_lft = ci->ifa_valid;
 900		*pprefered_lft = ci->ifa_prefered;
 901	}
 902
 903	return ifa;
 904
 905errout_free:
 906	inet_free_ifa(ifa);
 907errout:
 908	return ERR_PTR(err);
 909}
 910
 911static struct in_ifaddr *find_matching_ifa(struct in_ifaddr *ifa)
 912{
 913	struct in_device *in_dev = ifa->ifa_dev;
 914	struct in_ifaddr *ifa1;
 915
 916	if (!ifa->ifa_local)
 917		return NULL;
 918
 919	in_dev_for_each_ifa_rtnl(ifa1, in_dev) {
 
 920		if (ifa1->ifa_mask == ifa->ifa_mask &&
 921		    inet_ifa_match(ifa1->ifa_address, ifa) &&
 922		    ifa1->ifa_local == ifa->ifa_local)
 923			return ifa1;
 924	}
 925	return NULL;
 926}
 927
 928static int inet_rtm_newaddr(struct sk_buff *skb, struct nlmsghdr *nlh,
 929			    struct netlink_ext_ack *extack)
 930{
 931	struct net *net = sock_net(skb->sk);
 932	struct in_ifaddr *ifa;
 933	struct in_ifaddr *ifa_existing;
 934	__u32 valid_lft = INFINITY_LIFE_TIME;
 935	__u32 prefered_lft = INFINITY_LIFE_TIME;
 936
 937	ASSERT_RTNL();
 938
 939	ifa = rtm_to_ifaddr(net, nlh, &valid_lft, &prefered_lft, extack);
 940	if (IS_ERR(ifa))
 941		return PTR_ERR(ifa);
 942
 943	ifa_existing = find_matching_ifa(ifa);
 944	if (!ifa_existing) {
 945		/* It would be best to check for !NLM_F_CREATE here but
 946		 * userspace already relies on not having to provide this.
 947		 */
 948		set_ifa_lifetime(ifa, valid_lft, prefered_lft);
 949		if (ifa->ifa_flags & IFA_F_MCAUTOJOIN) {
 950			int ret = ip_mc_autojoin_config(net, true, ifa);
 
 951
 952			if (ret < 0) {
 953				inet_free_ifa(ifa);
 954				return ret;
 955			}
 956		}
 957		return __inet_insert_ifa(ifa, nlh, NETLINK_CB(skb).portid,
 958					 extack);
 959	} else {
 960		u32 new_metric = ifa->ifa_rt_priority;
 961
 962		inet_free_ifa(ifa);
 963
 964		if (nlh->nlmsg_flags & NLM_F_EXCL ||
 965		    !(nlh->nlmsg_flags & NLM_F_REPLACE))
 966			return -EEXIST;
 967		ifa = ifa_existing;
 968
 969		if (ifa->ifa_rt_priority != new_metric) {
 970			fib_modify_prefix_metric(ifa, new_metric);
 971			ifa->ifa_rt_priority = new_metric;
 972		}
 973
 974		set_ifa_lifetime(ifa, valid_lft, prefered_lft);
 975		cancel_delayed_work(&check_lifetime_work);
 976		queue_delayed_work(system_power_efficient_wq,
 977				&check_lifetime_work, 0);
 978		rtmsg_ifa(RTM_NEWADDR, ifa, nlh, NETLINK_CB(skb).portid);
 979	}
 980	return 0;
 981}
 982
 983/*
 984 *	Determine a default network mask, based on the IP address.
 985 */
 986
 987static int inet_abc_len(__be32 addr)
 988{
 989	int rc = -1;	/* Something else, probably a multicast. */
 990
 991	if (ipv4_is_zeronet(addr) || ipv4_is_lbcast(addr))
 992		rc = 0;
 993	else {
 994		__u32 haddr = ntohl(addr);
 
 995		if (IN_CLASSA(haddr))
 996			rc = 8;
 997		else if (IN_CLASSB(haddr))
 998			rc = 16;
 999		else if (IN_CLASSC(haddr))
1000			rc = 24;
1001		else if (IN_CLASSE(haddr))
1002			rc = 32;
1003	}
1004
1005	return rc;
1006}
1007
1008
1009int devinet_ioctl(struct net *net, unsigned int cmd, struct ifreq *ifr)
1010{
 
1011	struct sockaddr_in sin_orig;
1012	struct sockaddr_in *sin = (struct sockaddr_in *)&ifr->ifr_addr;
1013	struct in_ifaddr __rcu **ifap = NULL;
1014	struct in_device *in_dev;
 
1015	struct in_ifaddr *ifa = NULL;
1016	struct net_device *dev;
1017	char *colon;
1018	int ret = -EFAULT;
1019	int tryaddrmatch = 0;
1020
1021	ifr->ifr_name[IFNAMSIZ - 1] = 0;
 
 
 
 
 
 
1022
1023	/* save original address for comparison */
1024	memcpy(&sin_orig, sin, sizeof(*sin));
1025
1026	colon = strchr(ifr->ifr_name, ':');
1027	if (colon)
1028		*colon = 0;
1029
1030	dev_load(net, ifr->ifr_name);
1031
1032	switch (cmd) {
1033	case SIOCGIFADDR:	/* Get interface address */
1034	case SIOCGIFBRDADDR:	/* Get the broadcast address */
1035	case SIOCGIFDSTADDR:	/* Get the destination address */
1036	case SIOCGIFNETMASK:	/* Get the netmask for the interface */
1037		/* Note that these ioctls will not sleep,
1038		   so that we do not impose a lock.
1039		   One day we will be forced to put shlock here (I mean SMP)
1040		 */
1041		tryaddrmatch = (sin_orig.sin_family == AF_INET);
1042		memset(sin, 0, sizeof(*sin));
1043		sin->sin_family = AF_INET;
1044		break;
1045
1046	case SIOCSIFFLAGS:
1047		ret = -EPERM;
1048		if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
1049			goto out;
1050		break;
1051	case SIOCSIFADDR:	/* Set interface address (and family) */
1052	case SIOCSIFBRDADDR:	/* Set the broadcast address */
1053	case SIOCSIFDSTADDR:	/* Set the destination address */
1054	case SIOCSIFNETMASK: 	/* Set the netmask for the interface */
1055		ret = -EPERM;
1056		if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
1057			goto out;
1058		ret = -EINVAL;
1059		if (sin->sin_family != AF_INET)
1060			goto out;
1061		break;
1062	default:
1063		ret = -EINVAL;
1064		goto out;
1065	}
1066
1067	rtnl_lock();
1068
1069	ret = -ENODEV;
1070	dev = __dev_get_by_name(net, ifr->ifr_name);
1071	if (!dev)
1072		goto done;
1073
1074	if (colon)
1075		*colon = ':';
1076
1077	in_dev = __in_dev_get_rtnl(dev);
1078	if (in_dev) {
1079		if (tryaddrmatch) {
1080			/* Matthias Andree */
1081			/* compare label and address (4.4BSD style) */
1082			/* note: we only do this for a limited set of ioctls
1083			   and only if the original address family was AF_INET.
1084			   This is checked above. */
1085
1086			for (ifap = &in_dev->ifa_list;
1087			     (ifa = rtnl_dereference(*ifap)) != NULL;
1088			     ifap = &ifa->ifa_next) {
1089				if (!strcmp(ifr->ifr_name, ifa->ifa_label) &&
1090				    sin_orig.sin_addr.s_addr ==
1091							ifa->ifa_local) {
1092					break; /* found */
1093				}
1094			}
1095		}
1096		/* we didn't get a match, maybe the application is
1097		   4.3BSD-style and passed in junk so we fall back to
1098		   comparing just the label */
1099		if (!ifa) {
1100			for (ifap = &in_dev->ifa_list;
1101			     (ifa = rtnl_dereference(*ifap)) != NULL;
1102			     ifap = &ifa->ifa_next)
1103				if (!strcmp(ifr->ifr_name, ifa->ifa_label))
1104					break;
1105		}
1106	}
1107
1108	ret = -EADDRNOTAVAIL;
1109	if (!ifa && cmd != SIOCSIFADDR && cmd != SIOCSIFFLAGS)
1110		goto done;
1111
1112	switch (cmd) {
1113	case SIOCGIFADDR:	/* Get interface address */
1114		ret = 0;
1115		sin->sin_addr.s_addr = ifa->ifa_local;
1116		break;
1117
1118	case SIOCGIFBRDADDR:	/* Get the broadcast address */
1119		ret = 0;
1120		sin->sin_addr.s_addr = ifa->ifa_broadcast;
1121		break;
1122
1123	case SIOCGIFDSTADDR:	/* Get the destination address */
1124		ret = 0;
1125		sin->sin_addr.s_addr = ifa->ifa_address;
1126		break;
1127
1128	case SIOCGIFNETMASK:	/* Get the netmask for the interface */
1129		ret = 0;
1130		sin->sin_addr.s_addr = ifa->ifa_mask;
1131		break;
1132
1133	case SIOCSIFFLAGS:
1134		if (colon) {
1135			ret = -EADDRNOTAVAIL;
1136			if (!ifa)
1137				break;
1138			ret = 0;
1139			if (!(ifr->ifr_flags & IFF_UP))
1140				inet_del_ifa(in_dev, ifap, 1);
1141			break;
1142		}
1143		ret = dev_change_flags(dev, ifr->ifr_flags, NULL);
1144		break;
1145
1146	case SIOCSIFADDR:	/* Set interface address (and family) */
1147		ret = -EINVAL;
1148		if (inet_abc_len(sin->sin_addr.s_addr) < 0)
1149			break;
1150
1151		if (!ifa) {
1152			ret = -ENOBUFS;
1153			ifa = inet_alloc_ifa();
1154			if (!ifa)
1155				break;
1156			INIT_HLIST_NODE(&ifa->hash);
1157			if (colon)
1158				memcpy(ifa->ifa_label, ifr->ifr_name, IFNAMSIZ);
1159			else
1160				memcpy(ifa->ifa_label, dev->name, IFNAMSIZ);
1161		} else {
1162			ret = 0;
1163			if (ifa->ifa_local == sin->sin_addr.s_addr)
1164				break;
1165			inet_del_ifa(in_dev, ifap, 0);
1166			ifa->ifa_broadcast = 0;
1167			ifa->ifa_scope = 0;
1168		}
1169
1170		ifa->ifa_address = ifa->ifa_local = sin->sin_addr.s_addr;
1171
1172		if (!(dev->flags & IFF_POINTOPOINT)) {
1173			ifa->ifa_prefixlen = inet_abc_len(ifa->ifa_address);
1174			ifa->ifa_mask = inet_make_mask(ifa->ifa_prefixlen);
1175			if ((dev->flags & IFF_BROADCAST) &&
1176			    ifa->ifa_prefixlen < 31)
1177				ifa->ifa_broadcast = ifa->ifa_address |
1178						     ~ifa->ifa_mask;
1179		} else {
1180			ifa->ifa_prefixlen = 32;
1181			ifa->ifa_mask = inet_make_mask(32);
1182		}
1183		set_ifa_lifetime(ifa, INFINITY_LIFE_TIME, INFINITY_LIFE_TIME);
1184		ret = inet_set_ifa(dev, ifa);
1185		break;
1186
1187	case SIOCSIFBRDADDR:	/* Set the broadcast address */
1188		ret = 0;
1189		if (ifa->ifa_broadcast != sin->sin_addr.s_addr) {
1190			inet_del_ifa(in_dev, ifap, 0);
1191			ifa->ifa_broadcast = sin->sin_addr.s_addr;
1192			inet_insert_ifa(ifa);
1193		}
1194		break;
1195
1196	case SIOCSIFDSTADDR:	/* Set the destination address */
1197		ret = 0;
1198		if (ifa->ifa_address == sin->sin_addr.s_addr)
1199			break;
1200		ret = -EINVAL;
1201		if (inet_abc_len(sin->sin_addr.s_addr) < 0)
1202			break;
1203		ret = 0;
1204		inet_del_ifa(in_dev, ifap, 0);
1205		ifa->ifa_address = sin->sin_addr.s_addr;
1206		inet_insert_ifa(ifa);
1207		break;
1208
1209	case SIOCSIFNETMASK: 	/* Set the netmask for the interface */
1210
1211		/*
1212		 *	The mask we set must be legal.
1213		 */
1214		ret = -EINVAL;
1215		if (bad_mask(sin->sin_addr.s_addr, 0))
1216			break;
1217		ret = 0;
1218		if (ifa->ifa_mask != sin->sin_addr.s_addr) {
1219			__be32 old_mask = ifa->ifa_mask;
1220			inet_del_ifa(in_dev, ifap, 0);
1221			ifa->ifa_mask = sin->sin_addr.s_addr;
1222			ifa->ifa_prefixlen = inet_mask_len(ifa->ifa_mask);
1223
1224			/* See if current broadcast address matches
1225			 * with current netmask, then recalculate
1226			 * the broadcast address. Otherwise it's a
1227			 * funny address, so don't touch it since
1228			 * the user seems to know what (s)he's doing...
1229			 */
1230			if ((dev->flags & IFF_BROADCAST) &&
1231			    (ifa->ifa_prefixlen < 31) &&
1232			    (ifa->ifa_broadcast ==
1233			     (ifa->ifa_local|~old_mask))) {
1234				ifa->ifa_broadcast = (ifa->ifa_local |
1235						      ~sin->sin_addr.s_addr);
1236			}
1237			inet_insert_ifa(ifa);
1238		}
1239		break;
1240	}
1241done:
1242	rtnl_unlock();
1243out:
1244	return ret;
 
 
 
 
1245}
1246
1247static int inet_gifconf(struct net_device *dev, char __user *buf, int len, int size)
1248{
1249	struct in_device *in_dev = __in_dev_get_rtnl(dev);
1250	const struct in_ifaddr *ifa;
1251	struct ifreq ifr;
1252	int done = 0;
1253
1254	if (WARN_ON(size > sizeof(struct ifreq)))
1255		goto out;
1256
1257	if (!in_dev)
1258		goto out;
1259
1260	in_dev_for_each_ifa_rtnl(ifa, in_dev) {
1261		if (!buf) {
1262			done += size;
1263			continue;
1264		}
1265		if (len < size)
1266			break;
1267		memset(&ifr, 0, sizeof(struct ifreq));
1268		strcpy(ifr.ifr_name, ifa->ifa_label);
1269
1270		(*(struct sockaddr_in *)&ifr.ifr_addr).sin_family = AF_INET;
1271		(*(struct sockaddr_in *)&ifr.ifr_addr).sin_addr.s_addr =
1272								ifa->ifa_local;
1273
1274		if (copy_to_user(buf + done, &ifr, size)) {
1275			done = -EFAULT;
1276			break;
1277		}
1278		len  -= size;
1279		done += size;
 
1280	}
1281out:
1282	return done;
1283}
1284
1285static __be32 in_dev_select_addr(const struct in_device *in_dev,
1286				 int scope)
1287{
1288	const struct in_ifaddr *ifa;
1289
1290	in_dev_for_each_ifa_rcu(ifa, in_dev) {
1291		if (ifa->ifa_flags & IFA_F_SECONDARY)
1292			continue;
1293		if (ifa->ifa_scope != RT_SCOPE_LINK &&
1294		    ifa->ifa_scope <= scope)
1295			return ifa->ifa_local;
1296	}
1297
1298	return 0;
1299}
1300
1301__be32 inet_select_addr(const struct net_device *dev, __be32 dst, int scope)
1302{
1303	const struct in_ifaddr *ifa;
1304	__be32 addr = 0;
1305	unsigned char localnet_scope = RT_SCOPE_HOST;
1306	struct in_device *in_dev;
1307	struct net *net = dev_net(dev);
1308	int master_idx;
1309
1310	rcu_read_lock();
1311	in_dev = __in_dev_get_rcu(dev);
1312	if (!in_dev)
1313		goto no_in_dev;
1314
1315	if (unlikely(IN_DEV_ROUTE_LOCALNET(in_dev)))
1316		localnet_scope = RT_SCOPE_LINK;
1317
1318	in_dev_for_each_ifa_rcu(ifa, in_dev) {
1319		if (ifa->ifa_flags & IFA_F_SECONDARY)
1320			continue;
1321		if (min(ifa->ifa_scope, localnet_scope) > scope)
1322			continue;
1323		if (!dst || inet_ifa_match(dst, ifa)) {
1324			addr = ifa->ifa_local;
1325			break;
1326		}
1327		if (!addr)
1328			addr = ifa->ifa_local;
1329	}
1330
1331	if (addr)
1332		goto out_unlock;
1333no_in_dev:
1334	master_idx = l3mdev_master_ifindex_rcu(dev);
1335
1336	/* For VRFs, the VRF device takes the place of the loopback device,
1337	 * with addresses on it being preferred.  Note in such cases the
1338	 * loopback device will be among the devices that fail the master_idx
1339	 * equality check in the loop below.
1340	 */
1341	if (master_idx &&
1342	    (dev = dev_get_by_index_rcu(net, master_idx)) &&
1343	    (in_dev = __in_dev_get_rcu(dev))) {
1344		addr = in_dev_select_addr(in_dev, scope);
1345		if (addr)
1346			goto out_unlock;
 
 
 
 
1347	}
1348
1349	/* Not loopback addresses on loopback should be preferred
1350	   in this case. It is important that lo is the first interface
1351	   in dev_base list.
1352	 */
1353	for_each_netdev_rcu(net, dev) {
1354		if (l3mdev_master_ifindex_rcu(dev) != master_idx)
1355			continue;
1356
1357		in_dev = __in_dev_get_rcu(dev);
1358		if (!in_dev)
1359			continue;
1360
1361		addr = in_dev_select_addr(in_dev, scope);
1362		if (addr)
1363			goto out_unlock;
 
 
 
 
1364	}
1365out_unlock:
1366	rcu_read_unlock();
1367	return addr;
1368}
1369EXPORT_SYMBOL(inet_select_addr);
1370
1371static __be32 confirm_addr_indev(struct in_device *in_dev, __be32 dst,
1372			      __be32 local, int scope)
1373{
1374	unsigned char localnet_scope = RT_SCOPE_HOST;
1375	const struct in_ifaddr *ifa;
1376	__be32 addr = 0;
1377	int same = 0;
1378
1379	if (unlikely(IN_DEV_ROUTE_LOCALNET(in_dev)))
1380		localnet_scope = RT_SCOPE_LINK;
1381
1382	in_dev_for_each_ifa_rcu(ifa, in_dev) {
1383		unsigned char min_scope = min(ifa->ifa_scope, localnet_scope);
1384
 
1385		if (!addr &&
1386		    (local == ifa->ifa_local || !local) &&
1387		    min_scope <= scope) {
1388			addr = ifa->ifa_local;
1389			if (same)
1390				break;
1391		}
1392		if (!same) {
1393			same = (!local || inet_ifa_match(local, ifa)) &&
1394				(!dst || inet_ifa_match(dst, ifa));
1395			if (same && addr) {
1396				if (local || !dst)
1397					break;
1398				/* Is the selected addr into dst subnet? */
1399				if (inet_ifa_match(addr, ifa))
1400					break;
1401				/* No, then can we use new local src? */
1402				if (min_scope <= scope) {
1403					addr = ifa->ifa_local;
1404					break;
1405				}
1406				/* search for large dst subnet for addr */
1407				same = 0;
1408			}
1409		}
1410	}
1411
1412	return same ? addr : 0;
1413}
1414
1415/*
1416 * Confirm that local IP address exists using wildcards:
1417 * - net: netns to check, cannot be NULL
1418 * - in_dev: only on this interface, NULL=any interface
1419 * - dst: only in the same subnet as dst, 0=any dst
1420 * - local: address, 0=autoselect the local address
1421 * - scope: maximum allowed scope value for the local address
1422 */
1423__be32 inet_confirm_addr(struct net *net, struct in_device *in_dev,
1424			 __be32 dst, __be32 local, int scope)
1425{
1426	__be32 addr = 0;
1427	struct net_device *dev;
1428
1429	if (in_dev)
1430		return confirm_addr_indev(in_dev, dst, local, scope);
1431
1432	rcu_read_lock();
1433	for_each_netdev_rcu(net, dev) {
1434		in_dev = __in_dev_get_rcu(dev);
1435		if (in_dev) {
1436			addr = confirm_addr_indev(in_dev, dst, local, scope);
1437			if (addr)
1438				break;
1439		}
1440	}
1441	rcu_read_unlock();
1442
1443	return addr;
1444}
1445EXPORT_SYMBOL(inet_confirm_addr);
1446
1447/*
1448 *	Device notifier
1449 */
1450
1451int register_inetaddr_notifier(struct notifier_block *nb)
1452{
1453	return blocking_notifier_chain_register(&inetaddr_chain, nb);
1454}
1455EXPORT_SYMBOL(register_inetaddr_notifier);
1456
1457int unregister_inetaddr_notifier(struct notifier_block *nb)
1458{
1459	return blocking_notifier_chain_unregister(&inetaddr_chain, nb);
1460}
1461EXPORT_SYMBOL(unregister_inetaddr_notifier);
1462
1463int register_inetaddr_validator_notifier(struct notifier_block *nb)
1464{
1465	return blocking_notifier_chain_register(&inetaddr_validator_chain, nb);
1466}
1467EXPORT_SYMBOL(register_inetaddr_validator_notifier);
1468
1469int unregister_inetaddr_validator_notifier(struct notifier_block *nb)
1470{
1471	return blocking_notifier_chain_unregister(&inetaddr_validator_chain,
1472	    nb);
1473}
1474EXPORT_SYMBOL(unregister_inetaddr_validator_notifier);
1475
1476/* Rename ifa_labels for a device name change. Make some effort to preserve
1477 * existing alias numbering and to create unique labels if possible.
1478*/
1479static void inetdev_changename(struct net_device *dev, struct in_device *in_dev)
1480{
1481	struct in_ifaddr *ifa;
1482	int named = 0;
1483
1484	in_dev_for_each_ifa_rtnl(ifa, in_dev) {
1485		char old[IFNAMSIZ], *dot;
1486
1487		memcpy(old, ifa->ifa_label, IFNAMSIZ);
1488		memcpy(ifa->ifa_label, dev->name, IFNAMSIZ);
1489		if (named++ == 0)
1490			goto skip;
1491		dot = strchr(old, ':');
1492		if (!dot) {
1493			sprintf(old, ":%d", named);
1494			dot = old;
1495		}
1496		if (strlen(dot) + strlen(dev->name) < IFNAMSIZ)
1497			strcat(ifa->ifa_label, dot);
1498		else
1499			strcpy(ifa->ifa_label + (IFNAMSIZ - strlen(dot) - 1), dot);
1500skip:
1501		rtmsg_ifa(RTM_NEWADDR, ifa, NULL, 0);
1502	}
1503}
1504
 
 
 
 
 
1505static void inetdev_send_gratuitous_arp(struct net_device *dev,
1506					struct in_device *in_dev)
1507
1508{
1509	const struct in_ifaddr *ifa;
1510
1511	in_dev_for_each_ifa_rtnl(ifa, in_dev) {
 
1512		arp_send(ARPOP_REQUEST, ETH_P_ARP,
1513			 ifa->ifa_local, dev,
1514			 ifa->ifa_local, NULL,
1515			 dev->dev_addr, NULL);
1516	}
1517}
1518
1519/* Called only under RTNL semaphore */
1520
1521static int inetdev_event(struct notifier_block *this, unsigned long event,
1522			 void *ptr)
1523{
1524	struct net_device *dev = netdev_notifier_info_to_dev(ptr);
1525	struct in_device *in_dev = __in_dev_get_rtnl(dev);
1526
1527	ASSERT_RTNL();
1528
1529	if (!in_dev) {
1530		if (event == NETDEV_REGISTER) {
1531			in_dev = inetdev_init(dev);
1532			if (IS_ERR(in_dev))
1533				return notifier_from_errno(PTR_ERR(in_dev));
1534			if (dev->flags & IFF_LOOPBACK) {
1535				IN_DEV_CONF_SET(in_dev, NOXFRM, 1);
1536				IN_DEV_CONF_SET(in_dev, NOPOLICY, 1);
1537			}
1538		} else if (event == NETDEV_CHANGEMTU) {
1539			/* Re-enabling IP */
1540			if (inetdev_valid_mtu(dev->mtu))
1541				in_dev = inetdev_init(dev);
1542		}
1543		goto out;
1544	}
1545
1546	switch (event) {
1547	case NETDEV_REGISTER:
1548		pr_debug("%s: bug\n", __func__);
1549		RCU_INIT_POINTER(dev->ip_ptr, NULL);
1550		break;
1551	case NETDEV_UP:
1552		if (!inetdev_valid_mtu(dev->mtu))
1553			break;
1554		if (dev->flags & IFF_LOOPBACK) {
1555			struct in_ifaddr *ifa = inet_alloc_ifa();
1556
1557			if (ifa) {
1558				INIT_HLIST_NODE(&ifa->hash);
1559				ifa->ifa_local =
1560				  ifa->ifa_address = htonl(INADDR_LOOPBACK);
1561				ifa->ifa_prefixlen = 8;
1562				ifa->ifa_mask = inet_make_mask(8);
1563				in_dev_hold(in_dev);
1564				ifa->ifa_dev = in_dev;
1565				ifa->ifa_scope = RT_SCOPE_HOST;
1566				memcpy(ifa->ifa_label, dev->name, IFNAMSIZ);
1567				set_ifa_lifetime(ifa, INFINITY_LIFE_TIME,
1568						 INFINITY_LIFE_TIME);
1569				ipv4_devconf_setall(in_dev);
1570				neigh_parms_data_state_setall(in_dev->arp_parms);
1571				inet_insert_ifa(ifa);
1572			}
1573		}
1574		ip_mc_up(in_dev);
1575		fallthrough;
1576	case NETDEV_CHANGEADDR:
1577		if (!IN_DEV_ARP_NOTIFY(in_dev))
1578			break;
1579		fallthrough;
1580	case NETDEV_NOTIFY_PEERS:
1581		/* Send gratuitous ARP to notify of link change */
1582		inetdev_send_gratuitous_arp(dev, in_dev);
1583		break;
1584	case NETDEV_DOWN:
1585		ip_mc_down(in_dev);
1586		break;
1587	case NETDEV_PRE_TYPE_CHANGE:
1588		ip_mc_unmap(in_dev);
1589		break;
1590	case NETDEV_POST_TYPE_CHANGE:
1591		ip_mc_remap(in_dev);
1592		break;
1593	case NETDEV_CHANGEMTU:
1594		if (inetdev_valid_mtu(dev->mtu))
1595			break;
1596		/* disable IP when MTU is not enough */
1597		fallthrough;
1598	case NETDEV_UNREGISTER:
1599		inetdev_destroy(in_dev);
1600		break;
1601	case NETDEV_CHANGENAME:
1602		/* Do not notify about label change, this event is
1603		 * not interesting to applications using netlink.
1604		 */
1605		inetdev_changename(dev, in_dev);
1606
1607		devinet_sysctl_unregister(in_dev);
1608		devinet_sysctl_register(in_dev);
1609		break;
1610	}
1611out:
1612	return NOTIFY_DONE;
1613}
1614
1615static struct notifier_block ip_netdev_notifier = {
1616	.notifier_call = inetdev_event,
1617};
1618
1619static size_t inet_nlmsg_size(void)
1620{
1621	return NLMSG_ALIGN(sizeof(struct ifaddrmsg))
1622	       + nla_total_size(4) /* IFA_ADDRESS */
1623	       + nla_total_size(4) /* IFA_LOCAL */
1624	       + nla_total_size(4) /* IFA_BROADCAST */
1625	       + nla_total_size(IFNAMSIZ) /* IFA_LABEL */
1626	       + nla_total_size(4)  /* IFA_FLAGS */
1627	       + nla_total_size(4)  /* IFA_RT_PRIORITY */
1628	       + nla_total_size(sizeof(struct ifa_cacheinfo)); /* IFA_CACHEINFO */
1629}
1630
1631static inline u32 cstamp_delta(unsigned long cstamp)
1632{
1633	return (cstamp - INITIAL_JIFFIES) * 100UL / HZ;
1634}
1635
1636static int put_cacheinfo(struct sk_buff *skb, unsigned long cstamp,
1637			 unsigned long tstamp, u32 preferred, u32 valid)
1638{
1639	struct ifa_cacheinfo ci;
1640
1641	ci.cstamp = cstamp_delta(cstamp);
1642	ci.tstamp = cstamp_delta(tstamp);
1643	ci.ifa_prefered = preferred;
1644	ci.ifa_valid = valid;
1645
1646	return nla_put(skb, IFA_CACHEINFO, sizeof(ci), &ci);
1647}
1648
1649static int inet_fill_ifaddr(struct sk_buff *skb, struct in_ifaddr *ifa,
1650			    struct inet_fill_args *args)
1651{
1652	struct ifaddrmsg *ifm;
1653	struct nlmsghdr  *nlh;
1654	u32 preferred, valid;
1655
1656	nlh = nlmsg_put(skb, args->portid, args->seq, args->event, sizeof(*ifm),
1657			args->flags);
1658	if (!nlh)
1659		return -EMSGSIZE;
1660
1661	ifm = nlmsg_data(nlh);
1662	ifm->ifa_family = AF_INET;
1663	ifm->ifa_prefixlen = ifa->ifa_prefixlen;
1664	ifm->ifa_flags = ifa->ifa_flags;
1665	ifm->ifa_scope = ifa->ifa_scope;
1666	ifm->ifa_index = ifa->ifa_dev->dev->ifindex;
1667
1668	if (args->netnsid >= 0 &&
1669	    nla_put_s32(skb, IFA_TARGET_NETNSID, args->netnsid))
1670		goto nla_put_failure;
1671
1672	if (!(ifm->ifa_flags & IFA_F_PERMANENT)) {
1673		preferred = ifa->ifa_preferred_lft;
1674		valid = ifa->ifa_valid_lft;
1675		if (preferred != INFINITY_LIFE_TIME) {
1676			long tval = (jiffies - ifa->ifa_tstamp) / HZ;
1677
1678			if (preferred > tval)
1679				preferred -= tval;
1680			else
1681				preferred = 0;
1682			if (valid != INFINITY_LIFE_TIME) {
1683				if (valid > tval)
1684					valid -= tval;
1685				else
1686					valid = 0;
1687			}
1688		}
1689	} else {
1690		preferred = INFINITY_LIFE_TIME;
1691		valid = INFINITY_LIFE_TIME;
1692	}
1693	if ((ifa->ifa_address &&
1694	     nla_put_in_addr(skb, IFA_ADDRESS, ifa->ifa_address)) ||
1695	    (ifa->ifa_local &&
1696	     nla_put_in_addr(skb, IFA_LOCAL, ifa->ifa_local)) ||
1697	    (ifa->ifa_broadcast &&
1698	     nla_put_in_addr(skb, IFA_BROADCAST, ifa->ifa_broadcast)) ||
1699	    (ifa->ifa_label[0] &&
1700	     nla_put_string(skb, IFA_LABEL, ifa->ifa_label)) ||
1701	    nla_put_u32(skb, IFA_FLAGS, ifa->ifa_flags) ||
1702	    (ifa->ifa_rt_priority &&
1703	     nla_put_u32(skb, IFA_RT_PRIORITY, ifa->ifa_rt_priority)) ||
1704	    put_cacheinfo(skb, ifa->ifa_cstamp, ifa->ifa_tstamp,
1705			  preferred, valid))
1706		goto nla_put_failure;
1707
1708	nlmsg_end(skb, nlh);
1709	return 0;
1710
1711nla_put_failure:
1712	nlmsg_cancel(skb, nlh);
1713	return -EMSGSIZE;
1714}
1715
1716static int inet_valid_dump_ifaddr_req(const struct nlmsghdr *nlh,
1717				      struct inet_fill_args *fillargs,
1718				      struct net **tgt_net, struct sock *sk,
1719				      struct netlink_callback *cb)
1720{
1721	struct netlink_ext_ack *extack = cb->extack;
1722	struct nlattr *tb[IFA_MAX+1];
1723	struct ifaddrmsg *ifm;
1724	int err, i;
1725
1726	if (nlh->nlmsg_len < nlmsg_msg_size(sizeof(*ifm))) {
1727		NL_SET_ERR_MSG(extack, "ipv4: Invalid header for address dump request");
1728		return -EINVAL;
1729	}
1730
1731	ifm = nlmsg_data(nlh);
1732	if (ifm->ifa_prefixlen || ifm->ifa_flags || ifm->ifa_scope) {
1733		NL_SET_ERR_MSG(extack, "ipv4: Invalid values in header for address dump request");
1734		return -EINVAL;
1735	}
1736
1737	fillargs->ifindex = ifm->ifa_index;
1738	if (fillargs->ifindex) {
1739		cb->answer_flags |= NLM_F_DUMP_FILTERED;
1740		fillargs->flags |= NLM_F_DUMP_FILTERED;
1741	}
1742
1743	err = nlmsg_parse_deprecated_strict(nlh, sizeof(*ifm), tb, IFA_MAX,
1744					    ifa_ipv4_policy, extack);
1745	if (err < 0)
1746		return err;
1747
1748	for (i = 0; i <= IFA_MAX; ++i) {
1749		if (!tb[i])
1750			continue;
1751
1752		if (i == IFA_TARGET_NETNSID) {
1753			struct net *net;
1754
1755			fillargs->netnsid = nla_get_s32(tb[i]);
1756
1757			net = rtnl_get_net_ns_capable(sk, fillargs->netnsid);
1758			if (IS_ERR(net)) {
1759				fillargs->netnsid = -1;
1760				NL_SET_ERR_MSG(extack, "ipv4: Invalid target network namespace id");
1761				return PTR_ERR(net);
1762			}
1763			*tgt_net = net;
1764		} else {
1765			NL_SET_ERR_MSG(extack, "ipv4: Unsupported attribute in dump request");
1766			return -EINVAL;
1767		}
1768	}
1769
1770	return 0;
1771}
1772
1773static int in_dev_dump_addr(struct in_device *in_dev, struct sk_buff *skb,
1774			    struct netlink_callback *cb, int s_ip_idx,
1775			    struct inet_fill_args *fillargs)
1776{
1777	struct in_ifaddr *ifa;
1778	int ip_idx = 0;
1779	int err;
1780
1781	in_dev_for_each_ifa_rtnl(ifa, in_dev) {
1782		if (ip_idx < s_ip_idx) {
1783			ip_idx++;
1784			continue;
1785		}
1786		err = inet_fill_ifaddr(skb, ifa, fillargs);
1787		if (err < 0)
1788			goto done;
1789
1790		nl_dump_check_consistent(cb, nlmsg_hdr(skb));
1791		ip_idx++;
1792	}
1793	err = 0;
1794
1795done:
1796	cb->args[2] = ip_idx;
1797
1798	return err;
1799}
1800
1801static int inet_dump_ifaddr(struct sk_buff *skb, struct netlink_callback *cb)
1802{
1803	const struct nlmsghdr *nlh = cb->nlh;
1804	struct inet_fill_args fillargs = {
1805		.portid = NETLINK_CB(cb->skb).portid,
1806		.seq = nlh->nlmsg_seq,
1807		.event = RTM_NEWADDR,
1808		.flags = NLM_F_MULTI,
1809		.netnsid = -1,
1810	};
1811	struct net *net = sock_net(skb->sk);
1812	struct net *tgt_net = net;
1813	int h, s_h;
1814	int idx, s_idx;
1815	int s_ip_idx;
1816	struct net_device *dev;
1817	struct in_device *in_dev;
 
1818	struct hlist_head *head;
1819	int err = 0;
1820
1821	s_h = cb->args[0];
1822	s_idx = idx = cb->args[1];
1823	s_ip_idx = cb->args[2];
1824
1825	if (cb->strict_check) {
1826		err = inet_valid_dump_ifaddr_req(nlh, &fillargs, &tgt_net,
1827						 skb->sk, cb);
1828		if (err < 0)
1829			goto put_tgt_net;
1830
1831		err = 0;
1832		if (fillargs.ifindex) {
1833			dev = __dev_get_by_index(tgt_net, fillargs.ifindex);
1834			if (!dev) {
1835				err = -ENODEV;
1836				goto put_tgt_net;
1837			}
1838
1839			in_dev = __in_dev_get_rtnl(dev);
1840			if (in_dev) {
1841				err = in_dev_dump_addr(in_dev, skb, cb, s_ip_idx,
1842						       &fillargs);
1843			}
1844			goto put_tgt_net;
1845		}
1846	}
1847
1848	for (h = s_h; h < NETDEV_HASHENTRIES; h++, s_idx = 0) {
1849		idx = 0;
1850		head = &tgt_net->dev_index_head[h];
1851		rcu_read_lock();
1852		cb->seq = atomic_read(&tgt_net->ipv4.dev_addr_genid) ^
1853			  tgt_net->dev_base_seq;
1854		hlist_for_each_entry_rcu(dev, head, index_hlist) {
1855			if (idx < s_idx)
1856				goto cont;
1857			if (h > s_h || idx > s_idx)
1858				s_ip_idx = 0;
1859			in_dev = __in_dev_get_rcu(dev);
1860			if (!in_dev)
1861				goto cont;
1862
1863			err = in_dev_dump_addr(in_dev, skb, cb, s_ip_idx,
1864					       &fillargs);
1865			if (err < 0) {
1866				rcu_read_unlock();
1867				goto done;
 
 
 
 
 
 
 
1868			}
1869cont:
1870			idx++;
1871		}
1872		rcu_read_unlock();
1873	}
1874
1875done:
1876	cb->args[0] = h;
1877	cb->args[1] = idx;
1878put_tgt_net:
1879	if (fillargs.netnsid >= 0)
1880		put_net(tgt_net);
1881
1882	return skb->len ? : err;
1883}
1884
1885static void rtmsg_ifa(int event, struct in_ifaddr *ifa, struct nlmsghdr *nlh,
1886		      u32 portid)
1887{
1888	struct inet_fill_args fillargs = {
1889		.portid = portid,
1890		.seq = nlh ? nlh->nlmsg_seq : 0,
1891		.event = event,
1892		.flags = 0,
1893		.netnsid = -1,
1894	};
1895	struct sk_buff *skb;
 
1896	int err = -ENOBUFS;
1897	struct net *net;
1898
1899	net = dev_net(ifa->ifa_dev->dev);
1900	skb = nlmsg_new(inet_nlmsg_size(), GFP_KERNEL);
1901	if (!skb)
1902		goto errout;
1903
1904	err = inet_fill_ifaddr(skb, ifa, &fillargs);
1905	if (err < 0) {
1906		/* -EMSGSIZE implies BUG in inet_nlmsg_size() */
1907		WARN_ON(err == -EMSGSIZE);
1908		kfree_skb(skb);
1909		goto errout;
1910	}
1911	rtnl_notify(skb, net, portid, RTNLGRP_IPV4_IFADDR, nlh, GFP_KERNEL);
1912	return;
1913errout:
1914	if (err < 0)
1915		rtnl_set_sk_err(net, RTNLGRP_IPV4_IFADDR, err);
1916}
1917
1918static size_t inet_get_link_af_size(const struct net_device *dev,
1919				    u32 ext_filter_mask)
1920{
1921	struct in_device *in_dev = rcu_dereference_rtnl(dev->ip_ptr);
1922
1923	if (!in_dev)
1924		return 0;
1925
1926	return nla_total_size(IPV4_DEVCONF_MAX * 4); /* IFLA_INET_CONF */
1927}
1928
1929static int inet_fill_link_af(struct sk_buff *skb, const struct net_device *dev,
1930			     u32 ext_filter_mask)
1931{
1932	struct in_device *in_dev = rcu_dereference_rtnl(dev->ip_ptr);
1933	struct nlattr *nla;
1934	int i;
1935
1936	if (!in_dev)
1937		return -ENODATA;
1938
1939	nla = nla_reserve(skb, IFLA_INET_CONF, IPV4_DEVCONF_MAX * 4);
1940	if (!nla)
1941		return -EMSGSIZE;
1942
1943	for (i = 0; i < IPV4_DEVCONF_MAX; i++)
1944		((u32 *) nla_data(nla))[i] = in_dev->cnf.data[i];
1945
1946	return 0;
1947}
1948
1949static const struct nla_policy inet_af_policy[IFLA_INET_MAX+1] = {
1950	[IFLA_INET_CONF]	= { .type = NLA_NESTED },
1951};
1952
1953static int inet_validate_link_af(const struct net_device *dev,
1954				 const struct nlattr *nla)
1955{
1956	struct nlattr *a, *tb[IFLA_INET_MAX+1];
1957	int err, rem;
1958
1959	if (dev && !__in_dev_get_rcu(dev))
1960		return -EAFNOSUPPORT;
1961
1962	err = nla_parse_nested_deprecated(tb, IFLA_INET_MAX, nla,
1963					  inet_af_policy, NULL);
1964	if (err < 0)
1965		return err;
1966
1967	if (tb[IFLA_INET_CONF]) {
1968		nla_for_each_nested(a, tb[IFLA_INET_CONF], rem) {
1969			int cfgid = nla_type(a);
1970
1971			if (nla_len(a) < 4)
1972				return -EINVAL;
1973
1974			if (cfgid <= 0 || cfgid > IPV4_DEVCONF_MAX)
1975				return -EINVAL;
1976		}
1977	}
1978
1979	return 0;
1980}
1981
1982static int inet_set_link_af(struct net_device *dev, const struct nlattr *nla)
1983{
1984	struct in_device *in_dev = __in_dev_get_rcu(dev);
1985	struct nlattr *a, *tb[IFLA_INET_MAX+1];
1986	int rem;
1987
1988	if (!in_dev)
1989		return -EAFNOSUPPORT;
1990
1991	if (nla_parse_nested_deprecated(tb, IFLA_INET_MAX, nla, NULL, NULL) < 0)
1992		BUG();
1993
1994	if (tb[IFLA_INET_CONF]) {
1995		nla_for_each_nested(a, tb[IFLA_INET_CONF], rem)
1996			ipv4_devconf_set(in_dev, nla_type(a), nla_get_u32(a));
1997	}
1998
1999	return 0;
2000}
2001
2002static int inet_netconf_msgsize_devconf(int type)
2003{
2004	int size = NLMSG_ALIGN(sizeof(struct netconfmsg))
2005		   + nla_total_size(4);	/* NETCONFA_IFINDEX */
2006	bool all = false;
2007
2008	if (type == NETCONFA_ALL)
2009		all = true;
2010
2011	if (all || type == NETCONFA_FORWARDING)
2012		size += nla_total_size(4);
2013	if (all || type == NETCONFA_RP_FILTER)
2014		size += nla_total_size(4);
2015	if (all || type == NETCONFA_MC_FORWARDING)
2016		size += nla_total_size(4);
2017	if (all || type == NETCONFA_BC_FORWARDING)
2018		size += nla_total_size(4);
2019	if (all || type == NETCONFA_PROXY_NEIGH)
2020		size += nla_total_size(4);
2021	if (all || type == NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN)
2022		size += nla_total_size(4);
2023
2024	return size;
2025}
2026
2027static int inet_netconf_fill_devconf(struct sk_buff *skb, int ifindex,
2028				     struct ipv4_devconf *devconf, u32 portid,
2029				     u32 seq, int event, unsigned int flags,
2030				     int type)
2031{
2032	struct nlmsghdr  *nlh;
2033	struct netconfmsg *ncm;
2034	bool all = false;
2035
2036	nlh = nlmsg_put(skb, portid, seq, event, sizeof(struct netconfmsg),
2037			flags);
2038	if (!nlh)
2039		return -EMSGSIZE;
2040
2041	if (type == NETCONFA_ALL)
2042		all = true;
2043
2044	ncm = nlmsg_data(nlh);
2045	ncm->ncm_family = AF_INET;
2046
2047	if (nla_put_s32(skb, NETCONFA_IFINDEX, ifindex) < 0)
2048		goto nla_put_failure;
2049
2050	if (!devconf)
2051		goto out;
2052
2053	if ((all || type == NETCONFA_FORWARDING) &&
2054	    nla_put_s32(skb, NETCONFA_FORWARDING,
2055			IPV4_DEVCONF(*devconf, FORWARDING)) < 0)
2056		goto nla_put_failure;
2057	if ((all || type == NETCONFA_RP_FILTER) &&
2058	    nla_put_s32(skb, NETCONFA_RP_FILTER,
2059			IPV4_DEVCONF(*devconf, RP_FILTER)) < 0)
2060		goto nla_put_failure;
2061	if ((all || type == NETCONFA_MC_FORWARDING) &&
2062	    nla_put_s32(skb, NETCONFA_MC_FORWARDING,
2063			IPV4_DEVCONF(*devconf, MC_FORWARDING)) < 0)
2064		goto nla_put_failure;
2065	if ((all || type == NETCONFA_BC_FORWARDING) &&
2066	    nla_put_s32(skb, NETCONFA_BC_FORWARDING,
2067			IPV4_DEVCONF(*devconf, BC_FORWARDING)) < 0)
2068		goto nla_put_failure;
2069	if ((all || type == NETCONFA_PROXY_NEIGH) &&
2070	    nla_put_s32(skb, NETCONFA_PROXY_NEIGH,
2071			IPV4_DEVCONF(*devconf, PROXY_ARP)) < 0)
2072		goto nla_put_failure;
2073	if ((all || type == NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN) &&
2074	    nla_put_s32(skb, NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN,
2075			IPV4_DEVCONF(*devconf, IGNORE_ROUTES_WITH_LINKDOWN)) < 0)
2076		goto nla_put_failure;
2077
2078out:
2079	nlmsg_end(skb, nlh);
2080	return 0;
2081
2082nla_put_failure:
2083	nlmsg_cancel(skb, nlh);
2084	return -EMSGSIZE;
2085}
2086
2087void inet_netconf_notify_devconf(struct net *net, int event, int type,
2088				 int ifindex, struct ipv4_devconf *devconf)
2089{
2090	struct sk_buff *skb;
2091	int err = -ENOBUFS;
2092
2093	skb = nlmsg_new(inet_netconf_msgsize_devconf(type), GFP_KERNEL);
2094	if (!skb)
2095		goto errout;
2096
2097	err = inet_netconf_fill_devconf(skb, ifindex, devconf, 0, 0,
2098					event, 0, type);
2099	if (err < 0) {
2100		/* -EMSGSIZE implies BUG in inet_netconf_msgsize_devconf() */
2101		WARN_ON(err == -EMSGSIZE);
2102		kfree_skb(skb);
2103		goto errout;
2104	}
2105	rtnl_notify(skb, net, 0, RTNLGRP_IPV4_NETCONF, NULL, GFP_KERNEL);
2106	return;
2107errout:
2108	if (err < 0)
2109		rtnl_set_sk_err(net, RTNLGRP_IPV4_NETCONF, err);
2110}
2111
2112static const struct nla_policy devconf_ipv4_policy[NETCONFA_MAX+1] = {
2113	[NETCONFA_IFINDEX]	= { .len = sizeof(int) },
2114	[NETCONFA_FORWARDING]	= { .len = sizeof(int) },
2115	[NETCONFA_RP_FILTER]	= { .len = sizeof(int) },
2116	[NETCONFA_PROXY_NEIGH]	= { .len = sizeof(int) },
2117	[NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN]	= { .len = sizeof(int) },
2118};
2119
2120static int inet_netconf_valid_get_req(struct sk_buff *skb,
2121				      const struct nlmsghdr *nlh,
2122				      struct nlattr **tb,
2123				      struct netlink_ext_ack *extack)
2124{
2125	int i, err;
2126
2127	if (nlh->nlmsg_len < nlmsg_msg_size(sizeof(struct netconfmsg))) {
2128		NL_SET_ERR_MSG(extack, "ipv4: Invalid header for netconf get request");
2129		return -EINVAL;
2130	}
2131
2132	if (!netlink_strict_get_check(skb))
2133		return nlmsg_parse_deprecated(nlh, sizeof(struct netconfmsg),
2134					      tb, NETCONFA_MAX,
2135					      devconf_ipv4_policy, extack);
2136
2137	err = nlmsg_parse_deprecated_strict(nlh, sizeof(struct netconfmsg),
2138					    tb, NETCONFA_MAX,
2139					    devconf_ipv4_policy, extack);
2140	if (err)
2141		return err;
2142
2143	for (i = 0; i <= NETCONFA_MAX; i++) {
2144		if (!tb[i])
2145			continue;
2146
2147		switch (i) {
2148		case NETCONFA_IFINDEX:
2149			break;
2150		default:
2151			NL_SET_ERR_MSG(extack, "ipv4: Unsupported attribute in netconf get request");
2152			return -EINVAL;
2153		}
2154	}
2155
2156	return 0;
2157}
2158
2159static int inet_netconf_get_devconf(struct sk_buff *in_skb,
2160				    struct nlmsghdr *nlh,
2161				    struct netlink_ext_ack *extack)
2162{
2163	struct net *net = sock_net(in_skb->sk);
2164	struct nlattr *tb[NETCONFA_MAX+1];
 
2165	struct sk_buff *skb;
2166	struct ipv4_devconf *devconf;
2167	struct in_device *in_dev;
2168	struct net_device *dev;
2169	int ifindex;
2170	int err;
2171
2172	err = inet_netconf_valid_get_req(in_skb, nlh, tb, extack);
2173	if (err)
 
2174		goto errout;
2175
2176	err = -EINVAL;
2177	if (!tb[NETCONFA_IFINDEX])
2178		goto errout;
2179
2180	ifindex = nla_get_s32(tb[NETCONFA_IFINDEX]);
2181	switch (ifindex) {
2182	case NETCONFA_IFINDEX_ALL:
2183		devconf = net->ipv4.devconf_all;
2184		break;
2185	case NETCONFA_IFINDEX_DEFAULT:
2186		devconf = net->ipv4.devconf_dflt;
2187		break;
2188	default:
2189		dev = __dev_get_by_index(net, ifindex);
2190		if (!dev)
2191			goto errout;
2192		in_dev = __in_dev_get_rtnl(dev);
2193		if (!in_dev)
2194			goto errout;
2195		devconf = &in_dev->cnf;
2196		break;
2197	}
2198
2199	err = -ENOBUFS;
2200	skb = nlmsg_new(inet_netconf_msgsize_devconf(NETCONFA_ALL), GFP_KERNEL);
2201	if (!skb)
2202		goto errout;
2203
2204	err = inet_netconf_fill_devconf(skb, ifindex, devconf,
2205					NETLINK_CB(in_skb).portid,
2206					nlh->nlmsg_seq, RTM_NEWNETCONF, 0,
2207					NETCONFA_ALL);
2208	if (err < 0) {
2209		/* -EMSGSIZE implies BUG in inet_netconf_msgsize_devconf() */
2210		WARN_ON(err == -EMSGSIZE);
2211		kfree_skb(skb);
2212		goto errout;
2213	}
2214	err = rtnl_unicast(skb, net, NETLINK_CB(in_skb).portid);
2215errout:
2216	return err;
2217}
2218
2219static int inet_netconf_dump_devconf(struct sk_buff *skb,
2220				     struct netlink_callback *cb)
2221{
2222	const struct nlmsghdr *nlh = cb->nlh;
2223	struct net *net = sock_net(skb->sk);
2224	int h, s_h;
2225	int idx, s_idx;
2226	struct net_device *dev;
2227	struct in_device *in_dev;
2228	struct hlist_head *head;
2229
2230	if (cb->strict_check) {
2231		struct netlink_ext_ack *extack = cb->extack;
2232		struct netconfmsg *ncm;
2233
2234		if (nlh->nlmsg_len < nlmsg_msg_size(sizeof(*ncm))) {
2235			NL_SET_ERR_MSG(extack, "ipv4: Invalid header for netconf dump request");
2236			return -EINVAL;
2237		}
2238
2239		if (nlmsg_attrlen(nlh, sizeof(*ncm))) {
2240			NL_SET_ERR_MSG(extack, "ipv4: Invalid data after header in netconf dump request");
2241			return -EINVAL;
2242		}
2243	}
2244
2245	s_h = cb->args[0];
2246	s_idx = idx = cb->args[1];
2247
2248	for (h = s_h; h < NETDEV_HASHENTRIES; h++, s_idx = 0) {
2249		idx = 0;
2250		head = &net->dev_index_head[h];
2251		rcu_read_lock();
2252		cb->seq = atomic_read(&net->ipv4.dev_addr_genid) ^
2253			  net->dev_base_seq;
2254		hlist_for_each_entry_rcu(dev, head, index_hlist) {
2255			if (idx < s_idx)
2256				goto cont;
2257			in_dev = __in_dev_get_rcu(dev);
2258			if (!in_dev)
2259				goto cont;
2260
2261			if (inet_netconf_fill_devconf(skb, dev->ifindex,
2262						      &in_dev->cnf,
2263						      NETLINK_CB(cb->skb).portid,
2264						      nlh->nlmsg_seq,
2265						      RTM_NEWNETCONF,
2266						      NLM_F_MULTI,
2267						      NETCONFA_ALL) < 0) {
2268				rcu_read_unlock();
2269				goto done;
2270			}
2271			nl_dump_check_consistent(cb, nlmsg_hdr(skb));
2272cont:
2273			idx++;
2274		}
2275		rcu_read_unlock();
2276	}
2277	if (h == NETDEV_HASHENTRIES) {
2278		if (inet_netconf_fill_devconf(skb, NETCONFA_IFINDEX_ALL,
2279					      net->ipv4.devconf_all,
2280					      NETLINK_CB(cb->skb).portid,
2281					      nlh->nlmsg_seq,
2282					      RTM_NEWNETCONF, NLM_F_MULTI,
2283					      NETCONFA_ALL) < 0)
2284			goto done;
2285		else
2286			h++;
2287	}
2288	if (h == NETDEV_HASHENTRIES + 1) {
2289		if (inet_netconf_fill_devconf(skb, NETCONFA_IFINDEX_DEFAULT,
2290					      net->ipv4.devconf_dflt,
2291					      NETLINK_CB(cb->skb).portid,
2292					      nlh->nlmsg_seq,
2293					      RTM_NEWNETCONF, NLM_F_MULTI,
2294					      NETCONFA_ALL) < 0)
2295			goto done;
2296		else
2297			h++;
2298	}
2299done:
2300	cb->args[0] = h;
2301	cb->args[1] = idx;
2302
2303	return skb->len;
2304}
2305
2306#ifdef CONFIG_SYSCTL
2307
2308static void devinet_copy_dflt_conf(struct net *net, int i)
2309{
2310	struct net_device *dev;
2311
2312	rcu_read_lock();
2313	for_each_netdev_rcu(net, dev) {
2314		struct in_device *in_dev;
2315
2316		in_dev = __in_dev_get_rcu(dev);
2317		if (in_dev && !test_bit(i, in_dev->cnf.state))
2318			in_dev->cnf.data[i] = net->ipv4.devconf_dflt->data[i];
2319	}
2320	rcu_read_unlock();
2321}
2322
2323/* called with RTNL locked */
2324static void inet_forward_change(struct net *net)
2325{
2326	struct net_device *dev;
2327	int on = IPV4_DEVCONF_ALL(net, FORWARDING);
2328
2329	IPV4_DEVCONF_ALL(net, ACCEPT_REDIRECTS) = !on;
2330	IPV4_DEVCONF_DFLT(net, FORWARDING) = on;
2331	inet_netconf_notify_devconf(net, RTM_NEWNETCONF,
2332				    NETCONFA_FORWARDING,
2333				    NETCONFA_IFINDEX_ALL,
2334				    net->ipv4.devconf_all);
2335	inet_netconf_notify_devconf(net, RTM_NEWNETCONF,
2336				    NETCONFA_FORWARDING,
2337				    NETCONFA_IFINDEX_DEFAULT,
2338				    net->ipv4.devconf_dflt);
2339
2340	for_each_netdev(net, dev) {
2341		struct in_device *in_dev;
2342
2343		if (on)
2344			dev_disable_lro(dev);
2345
2346		in_dev = __in_dev_get_rtnl(dev);
2347		if (in_dev) {
2348			IN_DEV_CONF_SET(in_dev, FORWARDING, on);
2349			inet_netconf_notify_devconf(net, RTM_NEWNETCONF,
2350						    NETCONFA_FORWARDING,
2351						    dev->ifindex, &in_dev->cnf);
2352		}
 
2353	}
2354}
2355
2356static int devinet_conf_ifindex(struct net *net, struct ipv4_devconf *cnf)
2357{
2358	if (cnf == net->ipv4.devconf_dflt)
2359		return NETCONFA_IFINDEX_DEFAULT;
2360	else if (cnf == net->ipv4.devconf_all)
2361		return NETCONFA_IFINDEX_ALL;
2362	else {
2363		struct in_device *idev
2364			= container_of(cnf, struct in_device, cnf);
2365		return idev->dev->ifindex;
2366	}
2367}
2368
2369static int devinet_conf_proc(struct ctl_table *ctl, int write,
2370			     void *buffer, size_t *lenp, loff_t *ppos)
 
2371{
2372	int old_value = *(int *)ctl->data;
2373	int ret = proc_dointvec(ctl, write, buffer, lenp, ppos);
2374	int new_value = *(int *)ctl->data;
2375
2376	if (write) {
2377		struct ipv4_devconf *cnf = ctl->extra1;
2378		struct net *net = ctl->extra2;
2379		int i = (int *)ctl->data - cnf->data;
2380		int ifindex;
2381
2382		set_bit(i, cnf->state);
2383
2384		if (cnf == net->ipv4.devconf_dflt)
2385			devinet_copy_dflt_conf(net, i);
2386		if (i == IPV4_DEVCONF_ACCEPT_LOCAL - 1 ||
2387		    i == IPV4_DEVCONF_ROUTE_LOCALNET - 1)
2388			if ((new_value == 0) && (old_value != 0))
2389				rt_cache_flush(net);
2390
2391		if (i == IPV4_DEVCONF_BC_FORWARDING - 1 &&
2392		    new_value != old_value)
2393			rt_cache_flush(net);
2394
2395		if (i == IPV4_DEVCONF_RP_FILTER - 1 &&
2396		    new_value != old_value) {
2397			ifindex = devinet_conf_ifindex(net, cnf);
2398			inet_netconf_notify_devconf(net, RTM_NEWNETCONF,
2399						    NETCONFA_RP_FILTER,
2400						    ifindex, cnf);
2401		}
2402		if (i == IPV4_DEVCONF_PROXY_ARP - 1 &&
2403		    new_value != old_value) {
2404			ifindex = devinet_conf_ifindex(net, cnf);
2405			inet_netconf_notify_devconf(net, RTM_NEWNETCONF,
2406						    NETCONFA_PROXY_NEIGH,
2407						    ifindex, cnf);
2408		}
2409		if (i == IPV4_DEVCONF_IGNORE_ROUTES_WITH_LINKDOWN - 1 &&
2410		    new_value != old_value) {
2411			ifindex = devinet_conf_ifindex(net, cnf);
2412			inet_netconf_notify_devconf(net, RTM_NEWNETCONF,
2413						    NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN,
2414						    ifindex, cnf);
2415		}
2416	}
2417
2418	return ret;
2419}
2420
2421static int devinet_sysctl_forward(struct ctl_table *ctl, int write,
2422				  void *buffer, size_t *lenp, loff_t *ppos)
 
2423{
2424	int *valp = ctl->data;
2425	int val = *valp;
2426	loff_t pos = *ppos;
2427	int ret = proc_dointvec(ctl, write, buffer, lenp, ppos);
2428
2429	if (write && *valp != val) {
2430		struct net *net = ctl->extra2;
2431
2432		if (valp != &IPV4_DEVCONF_DFLT(net, FORWARDING)) {
2433			if (!rtnl_trylock()) {
2434				/* Restore the original values before restarting */
2435				*valp = val;
2436				*ppos = pos;
2437				return restart_syscall();
2438			}
2439			if (valp == &IPV4_DEVCONF_ALL(net, FORWARDING)) {
2440				inet_forward_change(net);
2441			} else {
2442				struct ipv4_devconf *cnf = ctl->extra1;
2443				struct in_device *idev =
2444					container_of(cnf, struct in_device, cnf);
2445				if (*valp)
2446					dev_disable_lro(idev->dev);
2447				inet_netconf_notify_devconf(net, RTM_NEWNETCONF,
2448							    NETCONFA_FORWARDING,
2449							    idev->dev->ifindex,
2450							    cnf);
2451			}
2452			rtnl_unlock();
2453			rt_cache_flush(net);
2454		} else
2455			inet_netconf_notify_devconf(net, RTM_NEWNETCONF,
2456						    NETCONFA_FORWARDING,
2457						    NETCONFA_IFINDEX_DEFAULT,
2458						    net->ipv4.devconf_dflt);
2459	}
2460
2461	return ret;
2462}
2463
2464static int ipv4_doint_and_flush(struct ctl_table *ctl, int write,
2465				void *buffer, size_t *lenp, loff_t *ppos)
 
2466{
2467	int *valp = ctl->data;
2468	int val = *valp;
2469	int ret = proc_dointvec(ctl, write, buffer, lenp, ppos);
2470	struct net *net = ctl->extra2;
2471
2472	if (write && *valp != val)
2473		rt_cache_flush(net);
2474
2475	return ret;
2476}
2477
2478#define DEVINET_SYSCTL_ENTRY(attr, name, mval, proc) \
2479	{ \
2480		.procname	= name, \
2481		.data		= ipv4_devconf.data + \
2482				  IPV4_DEVCONF_ ## attr - 1, \
2483		.maxlen		= sizeof(int), \
2484		.mode		= mval, \
2485		.proc_handler	= proc, \
2486		.extra1		= &ipv4_devconf, \
2487	}
2488
2489#define DEVINET_SYSCTL_RW_ENTRY(attr, name) \
2490	DEVINET_SYSCTL_ENTRY(attr, name, 0644, devinet_conf_proc)
2491
2492#define DEVINET_SYSCTL_RO_ENTRY(attr, name) \
2493	DEVINET_SYSCTL_ENTRY(attr, name, 0444, devinet_conf_proc)
2494
2495#define DEVINET_SYSCTL_COMPLEX_ENTRY(attr, name, proc) \
2496	DEVINET_SYSCTL_ENTRY(attr, name, 0644, proc)
2497
2498#define DEVINET_SYSCTL_FLUSHING_ENTRY(attr, name) \
2499	DEVINET_SYSCTL_COMPLEX_ENTRY(attr, name, ipv4_doint_and_flush)
2500
2501static struct devinet_sysctl_table {
2502	struct ctl_table_header *sysctl_header;
2503	struct ctl_table devinet_vars[__IPV4_DEVCONF_MAX];
2504} devinet_sysctl = {
2505	.devinet_vars = {
2506		DEVINET_SYSCTL_COMPLEX_ENTRY(FORWARDING, "forwarding",
2507					     devinet_sysctl_forward),
2508		DEVINET_SYSCTL_RO_ENTRY(MC_FORWARDING, "mc_forwarding"),
2509		DEVINET_SYSCTL_RW_ENTRY(BC_FORWARDING, "bc_forwarding"),
2510
2511		DEVINET_SYSCTL_RW_ENTRY(ACCEPT_REDIRECTS, "accept_redirects"),
2512		DEVINET_SYSCTL_RW_ENTRY(SECURE_REDIRECTS, "secure_redirects"),
2513		DEVINET_SYSCTL_RW_ENTRY(SHARED_MEDIA, "shared_media"),
2514		DEVINET_SYSCTL_RW_ENTRY(RP_FILTER, "rp_filter"),
2515		DEVINET_SYSCTL_RW_ENTRY(SEND_REDIRECTS, "send_redirects"),
2516		DEVINET_SYSCTL_RW_ENTRY(ACCEPT_SOURCE_ROUTE,
2517					"accept_source_route"),
2518		DEVINET_SYSCTL_RW_ENTRY(ACCEPT_LOCAL, "accept_local"),
2519		DEVINET_SYSCTL_RW_ENTRY(SRC_VMARK, "src_valid_mark"),
2520		DEVINET_SYSCTL_RW_ENTRY(PROXY_ARP, "proxy_arp"),
2521		DEVINET_SYSCTL_RW_ENTRY(MEDIUM_ID, "medium_id"),
2522		DEVINET_SYSCTL_RW_ENTRY(BOOTP_RELAY, "bootp_relay"),
2523		DEVINET_SYSCTL_RW_ENTRY(LOG_MARTIANS, "log_martians"),
2524		DEVINET_SYSCTL_RW_ENTRY(TAG, "tag"),
2525		DEVINET_SYSCTL_RW_ENTRY(ARPFILTER, "arp_filter"),
2526		DEVINET_SYSCTL_RW_ENTRY(ARP_ANNOUNCE, "arp_announce"),
2527		DEVINET_SYSCTL_RW_ENTRY(ARP_IGNORE, "arp_ignore"),
2528		DEVINET_SYSCTL_RW_ENTRY(ARP_ACCEPT, "arp_accept"),
2529		DEVINET_SYSCTL_RW_ENTRY(ARP_NOTIFY, "arp_notify"),
2530		DEVINET_SYSCTL_RW_ENTRY(PROXY_ARP_PVLAN, "proxy_arp_pvlan"),
2531		DEVINET_SYSCTL_RW_ENTRY(FORCE_IGMP_VERSION,
2532					"force_igmp_version"),
2533		DEVINET_SYSCTL_RW_ENTRY(IGMPV2_UNSOLICITED_REPORT_INTERVAL,
2534					"igmpv2_unsolicited_report_interval"),
2535		DEVINET_SYSCTL_RW_ENTRY(IGMPV3_UNSOLICITED_REPORT_INTERVAL,
2536					"igmpv3_unsolicited_report_interval"),
2537		DEVINET_SYSCTL_RW_ENTRY(IGNORE_ROUTES_WITH_LINKDOWN,
2538					"ignore_routes_with_linkdown"),
2539		DEVINET_SYSCTL_RW_ENTRY(DROP_GRATUITOUS_ARP,
2540					"drop_gratuitous_arp"),
2541
2542		DEVINET_SYSCTL_FLUSHING_ENTRY(NOXFRM, "disable_xfrm"),
2543		DEVINET_SYSCTL_FLUSHING_ENTRY(NOPOLICY, "disable_policy"),
2544		DEVINET_SYSCTL_FLUSHING_ENTRY(PROMOTE_SECONDARIES,
2545					      "promote_secondaries"),
2546		DEVINET_SYSCTL_FLUSHING_ENTRY(ROUTE_LOCALNET,
2547					      "route_localnet"),
2548		DEVINET_SYSCTL_FLUSHING_ENTRY(DROP_UNICAST_IN_L2_MULTICAST,
2549					      "drop_unicast_in_l2_multicast"),
2550	},
2551};
2552
2553static int __devinet_sysctl_register(struct net *net, char *dev_name,
2554				     int ifindex, struct ipv4_devconf *p)
2555{
2556	int i;
2557	struct devinet_sysctl_table *t;
2558	char path[sizeof("net/ipv4/conf/") + IFNAMSIZ];
2559
2560	t = kmemdup(&devinet_sysctl, sizeof(*t), GFP_KERNEL);
2561	if (!t)
2562		goto out;
2563
2564	for (i = 0; i < ARRAY_SIZE(t->devinet_vars) - 1; i++) {
2565		t->devinet_vars[i].data += (char *)p - (char *)&ipv4_devconf;
2566		t->devinet_vars[i].extra1 = p;
2567		t->devinet_vars[i].extra2 = net;
2568	}
2569
2570	snprintf(path, sizeof(path), "net/ipv4/conf/%s", dev_name);
2571
2572	t->sysctl_header = register_net_sysctl(net, path, t->devinet_vars);
2573	if (!t->sysctl_header)
2574		goto free;
2575
2576	p->sysctl = t;
2577
2578	inet_netconf_notify_devconf(net, RTM_NEWNETCONF, NETCONFA_ALL,
2579				    ifindex, p);
2580	return 0;
2581
2582free:
2583	kfree(t);
2584out:
2585	return -ENOBUFS;
2586}
2587
2588static void __devinet_sysctl_unregister(struct net *net,
2589					struct ipv4_devconf *cnf, int ifindex)
2590{
2591	struct devinet_sysctl_table *t = cnf->sysctl;
2592
2593	if (t) {
2594		cnf->sysctl = NULL;
2595		unregister_net_sysctl_table(t->sysctl_header);
2596		kfree(t);
2597	}
2598
2599	inet_netconf_notify_devconf(net, RTM_DELNETCONF, 0, ifindex, NULL);
 
 
2600}
2601
2602static int devinet_sysctl_register(struct in_device *idev)
2603{
2604	int err;
2605
2606	if (!sysctl_dev_name_is_allowed(idev->dev->name))
2607		return -EINVAL;
2608
2609	err = neigh_sysctl_register(idev->dev, idev->arp_parms, NULL);
2610	if (err)
2611		return err;
2612	err = __devinet_sysctl_register(dev_net(idev->dev), idev->dev->name,
2613					idev->dev->ifindex, &idev->cnf);
2614	if (err)
2615		neigh_sysctl_unregister(idev->arp_parms);
2616	return err;
2617}
2618
2619static void devinet_sysctl_unregister(struct in_device *idev)
2620{
2621	struct net *net = dev_net(idev->dev);
2622
2623	__devinet_sysctl_unregister(net, &idev->cnf, idev->dev->ifindex);
2624	neigh_sysctl_unregister(idev->arp_parms);
2625}
2626
2627static struct ctl_table ctl_forward_entry[] = {
2628	{
2629		.procname	= "ip_forward",
2630		.data		= &ipv4_devconf.data[
2631					IPV4_DEVCONF_FORWARDING - 1],
2632		.maxlen		= sizeof(int),
2633		.mode		= 0644,
2634		.proc_handler	= devinet_sysctl_forward,
2635		.extra1		= &ipv4_devconf,
2636		.extra2		= &init_net,
2637	},
2638	{ },
2639};
2640#endif
2641
2642static __net_init int devinet_init_net(struct net *net)
2643{
2644	int err;
2645	struct ipv4_devconf *all, *dflt;
2646#ifdef CONFIG_SYSCTL
2647	struct ctl_table *tbl;
2648	struct ctl_table_header *forw_hdr;
2649#endif
2650
2651	err = -ENOMEM;
2652	all = kmemdup(&ipv4_devconf, sizeof(ipv4_devconf), GFP_KERNEL);
2653	if (!all)
2654		goto err_alloc_all;
2655
2656	dflt = kmemdup(&ipv4_devconf_dflt, sizeof(ipv4_devconf_dflt), GFP_KERNEL);
2657	if (!dflt)
2658		goto err_alloc_dflt;
 
 
 
 
2659
2660#ifdef CONFIG_SYSCTL
2661	tbl = kmemdup(ctl_forward_entry, sizeof(ctl_forward_entry), GFP_KERNEL);
2662	if (!tbl)
2663		goto err_alloc_ctl;
2664
2665	tbl[0].data = &all->data[IPV4_DEVCONF_FORWARDING - 1];
2666	tbl[0].extra1 = all;
2667	tbl[0].extra2 = net;
2668#endif
2669
2670	if (!net_eq(net, &init_net)) {
2671		if (IS_ENABLED(CONFIG_SYSCTL) &&
2672		    sysctl_devconf_inherit_init_net == 3) {
2673			/* copy from the current netns */
2674			memcpy(all, current->nsproxy->net_ns->ipv4.devconf_all,
2675			       sizeof(ipv4_devconf));
2676			memcpy(dflt,
2677			       current->nsproxy->net_ns->ipv4.devconf_dflt,
2678			       sizeof(ipv4_devconf_dflt));
2679		} else if (!IS_ENABLED(CONFIG_SYSCTL) ||
2680			   sysctl_devconf_inherit_init_net != 2) {
2681			/* inherit == 0 or 1: copy from init_net */
2682			memcpy(all, init_net.ipv4.devconf_all,
2683			       sizeof(ipv4_devconf));
2684			memcpy(dflt, init_net.ipv4.devconf_dflt,
2685			       sizeof(ipv4_devconf_dflt));
2686		}
2687		/* else inherit == 2: use compiled values */
2688	}
2689
2690#ifdef CONFIG_SYSCTL
2691	err = __devinet_sysctl_register(net, "all", NETCONFA_IFINDEX_ALL, all);
2692	if (err < 0)
2693		goto err_reg_all;
2694
2695	err = __devinet_sysctl_register(net, "default",
2696					NETCONFA_IFINDEX_DEFAULT, dflt);
2697	if (err < 0)
2698		goto err_reg_dflt;
2699
2700	err = -ENOMEM;
2701	forw_hdr = register_net_sysctl(net, "net/ipv4", tbl);
2702	if (!forw_hdr)
2703		goto err_reg_ctl;
2704	net->ipv4.forw_hdr = forw_hdr;
2705#endif
2706
2707	net->ipv4.devconf_all = all;
2708	net->ipv4.devconf_dflt = dflt;
2709	return 0;
2710
2711#ifdef CONFIG_SYSCTL
2712err_reg_ctl:
2713	__devinet_sysctl_unregister(net, dflt, NETCONFA_IFINDEX_DEFAULT);
2714err_reg_dflt:
2715	__devinet_sysctl_unregister(net, all, NETCONFA_IFINDEX_ALL);
2716err_reg_all:
2717	kfree(tbl);
 
2718err_alloc_ctl:
2719#endif
2720	kfree(dflt);
 
2721err_alloc_dflt:
2722	kfree(all);
 
2723err_alloc_all:
2724	return err;
2725}
2726
2727static __net_exit void devinet_exit_net(struct net *net)
2728{
2729#ifdef CONFIG_SYSCTL
2730	struct ctl_table *tbl;
2731
2732	tbl = net->ipv4.forw_hdr->ctl_table_arg;
2733	unregister_net_sysctl_table(net->ipv4.forw_hdr);
2734	__devinet_sysctl_unregister(net, net->ipv4.devconf_dflt,
2735				    NETCONFA_IFINDEX_DEFAULT);
2736	__devinet_sysctl_unregister(net, net->ipv4.devconf_all,
2737				    NETCONFA_IFINDEX_ALL);
2738	kfree(tbl);
2739#endif
2740	kfree(net->ipv4.devconf_dflt);
2741	kfree(net->ipv4.devconf_all);
2742}
2743
2744static __net_initdata struct pernet_operations devinet_ops = {
2745	.init = devinet_init_net,
2746	.exit = devinet_exit_net,
2747};
2748
2749static struct rtnl_af_ops inet_af_ops __read_mostly = {
2750	.family		  = AF_INET,
2751	.fill_link_af	  = inet_fill_link_af,
2752	.get_link_af_size = inet_get_link_af_size,
2753	.validate_link_af = inet_validate_link_af,
2754	.set_link_af	  = inet_set_link_af,
2755};
2756
2757void __init devinet_init(void)
2758{
2759	int i;
2760
2761	for (i = 0; i < IN4_ADDR_HSIZE; i++)
2762		INIT_HLIST_HEAD(&inet_addr_lst[i]);
2763
2764	register_pernet_subsys(&devinet_ops);
2765
2766	register_gifconf(PF_INET, inet_gifconf);
2767	register_netdevice_notifier(&ip_netdev_notifier);
2768
2769	queue_delayed_work(system_power_efficient_wq, &check_lifetime_work, 0);
2770
2771	rtnl_af_register(&inet_af_ops);
2772
2773	rtnl_register(PF_INET, RTM_NEWADDR, inet_rtm_newaddr, NULL, 0);
2774	rtnl_register(PF_INET, RTM_DELADDR, inet_rtm_deladdr, NULL, 0);
2775	rtnl_register(PF_INET, RTM_GETADDR, NULL, inet_dump_ifaddr, 0);
2776	rtnl_register(PF_INET, RTM_GETNETCONF, inet_netconf_get_devconf,
2777		      inet_netconf_dump_devconf, 0);
2778}