Linux Audio

Check our new training course

Loading...
v4.17
   1/*
   2 * net/sched/cls_api.c	Packet classifier API.
   3 *
   4 *		This program is free software; you can redistribute it and/or
   5 *		modify it under the terms of the GNU General Public License
   6 *		as published by the Free Software Foundation; either version
   7 *		2 of the License, or (at your option) any later version.
   8 *
   9 * Authors:	Alexey Kuznetsov, <kuznet@ms2.inr.ac.ru>
  10 *
  11 * Changes:
  12 *
  13 * Eduardo J. Blanco <ejbs@netlabs.com.uy> :990222: kmod support
  14 *
  15 */
  16
  17#include <linux/module.h>
  18#include <linux/types.h>
  19#include <linux/kernel.h>
  20#include <linux/string.h>
  21#include <linux/errno.h>
  22#include <linux/err.h>
  23#include <linux/skbuff.h>
  24#include <linux/init.h>
  25#include <linux/kmod.h>
 
  26#include <linux/slab.h>
  27#include <linux/idr.h>
  28#include <net/net_namespace.h>
  29#include <net/sock.h>
  30#include <net/netlink.h>
  31#include <net/pkt_sched.h>
  32#include <net/pkt_cls.h>
  33
  34/* The list of all installed classifier types */
  35static LIST_HEAD(tcf_proto_base);
  36
  37/* Protects list of registered TC modules. It is pure SMP lock. */
  38static DEFINE_RWLOCK(cls_mod_lock);
  39
  40/* Find classifier type by string name */
  41
  42static const struct tcf_proto_ops *tcf_proto_lookup_ops(const char *kind)
  43{
  44	const struct tcf_proto_ops *t, *res = NULL;
  45
  46	if (kind) {
  47		read_lock(&cls_mod_lock);
  48		list_for_each_entry(t, &tcf_proto_base, head) {
  49			if (strcmp(kind, t->kind) == 0) {
  50				if (try_module_get(t->owner))
  51					res = t;
  52				break;
  53			}
  54		}
  55		read_unlock(&cls_mod_lock);
  56	}
  57	return res;
  58}
  59
  60/* Register(unregister) new classifier type */
  61
  62int register_tcf_proto_ops(struct tcf_proto_ops *ops)
  63{
  64	struct tcf_proto_ops *t;
  65	int rc = -EEXIST;
  66
  67	write_lock(&cls_mod_lock);
  68	list_for_each_entry(t, &tcf_proto_base, head)
  69		if (!strcmp(ops->kind, t->kind))
  70			goto out;
  71
  72	list_add_tail(&ops->head, &tcf_proto_base);
  73	rc = 0;
  74out:
  75	write_unlock(&cls_mod_lock);
  76	return rc;
  77}
  78EXPORT_SYMBOL(register_tcf_proto_ops);
  79
  80static struct workqueue_struct *tc_filter_wq;
  81
  82int unregister_tcf_proto_ops(struct tcf_proto_ops *ops)
  83{
  84	struct tcf_proto_ops *t;
  85	int rc = -ENOENT;
  86
  87	/* Wait for outstanding call_rcu()s, if any, from a
  88	 * tcf_proto_ops's destroy() handler.
  89	 */
  90	rcu_barrier();
  91	flush_workqueue(tc_filter_wq);
  92
  93	write_lock(&cls_mod_lock);
  94	list_for_each_entry(t, &tcf_proto_base, head) {
  95		if (t == ops) {
  96			list_del(&t->head);
  97			rc = 0;
  98			break;
  99		}
 100	}
 101	write_unlock(&cls_mod_lock);
 102	return rc;
 103}
 104EXPORT_SYMBOL(unregister_tcf_proto_ops);
 105
 106bool tcf_queue_work(struct work_struct *work)
 107{
 108	return queue_work(tc_filter_wq, work);
 109}
 110EXPORT_SYMBOL(tcf_queue_work);
 111
 112/* Select new prio value from the range, managed by kernel. */
 113
 114static inline u32 tcf_auto_prio(struct tcf_proto *tp)
 115{
 116	u32 first = TC_H_MAKE(0xC0000000U, 0U);
 117
 118	if (tp)
 119		first = tp->prio - 1;
 120
 121	return TC_H_MAJ(first);
 122}
 123
 124static struct tcf_proto *tcf_proto_create(const char *kind, u32 protocol,
 125					  u32 prio, struct tcf_chain *chain,
 126					  struct netlink_ext_ack *extack)
 127{
 128	struct tcf_proto *tp;
 129	int err;
 130
 131	tp = kzalloc(sizeof(*tp), GFP_KERNEL);
 132	if (!tp)
 133		return ERR_PTR(-ENOBUFS);
 134
 135	err = -ENOENT;
 136	tp->ops = tcf_proto_lookup_ops(kind);
 137	if (!tp->ops) {
 138#ifdef CONFIG_MODULES
 139		rtnl_unlock();
 140		request_module("cls_%s", kind);
 141		rtnl_lock();
 142		tp->ops = tcf_proto_lookup_ops(kind);
 143		/* We dropped the RTNL semaphore in order to perform
 144		 * the module load. So, even if we succeeded in loading
 145		 * the module we have to replay the request. We indicate
 146		 * this using -EAGAIN.
 147		 */
 148		if (tp->ops) {
 149			module_put(tp->ops->owner);
 150			err = -EAGAIN;
 151		} else {
 152			NL_SET_ERR_MSG(extack, "TC classifier not found");
 153			err = -ENOENT;
 154		}
 155#endif
 156		goto errout;
 157	}
 158	tp->classify = tp->ops->classify;
 159	tp->protocol = protocol;
 160	tp->prio = prio;
 161	tp->chain = chain;
 162
 163	err = tp->ops->init(tp);
 164	if (err) {
 165		module_put(tp->ops->owner);
 166		goto errout;
 167	}
 168	return tp;
 169
 170errout:
 171	kfree(tp);
 172	return ERR_PTR(err);
 173}
 174
 175static void tcf_proto_destroy(struct tcf_proto *tp,
 176			      struct netlink_ext_ack *extack)
 177{
 178	tp->ops->destroy(tp, extack);
 179	module_put(tp->ops->owner);
 180	kfree_rcu(tp, rcu);
 181}
 182
 183struct tcf_filter_chain_list_item {
 184	struct list_head list;
 185	tcf_chain_head_change_t *chain_head_change;
 186	void *chain_head_change_priv;
 187};
 188
 189static struct tcf_chain *tcf_chain_create(struct tcf_block *block,
 190					  u32 chain_index)
 191{
 192	struct tcf_chain *chain;
 193
 194	chain = kzalloc(sizeof(*chain), GFP_KERNEL);
 195	if (!chain)
 196		return NULL;
 197	INIT_LIST_HEAD(&chain->filter_chain_list);
 198	list_add_tail(&chain->list, &block->chain_list);
 199	chain->block = block;
 200	chain->index = chain_index;
 201	chain->refcnt = 1;
 202	return chain;
 203}
 204
 205static void tcf_chain_head_change_item(struct tcf_filter_chain_list_item *item,
 206				       struct tcf_proto *tp_head)
 207{
 208	if (item->chain_head_change)
 209		item->chain_head_change(tp_head, item->chain_head_change_priv);
 210}
 211static void tcf_chain_head_change(struct tcf_chain *chain,
 212				  struct tcf_proto *tp_head)
 213{
 214	struct tcf_filter_chain_list_item *item;
 215
 216	list_for_each_entry(item, &chain->filter_chain_list, list)
 217		tcf_chain_head_change_item(item, tp_head);
 218}
 219
 220static void tcf_chain_flush(struct tcf_chain *chain)
 221{
 222	struct tcf_proto *tp = rtnl_dereference(chain->filter_chain);
 223
 224	tcf_chain_head_change(chain, NULL);
 225	while (tp) {
 226		RCU_INIT_POINTER(chain->filter_chain, tp->next);
 227		tcf_proto_destroy(tp, NULL);
 228		tp = rtnl_dereference(chain->filter_chain);
 229		tcf_chain_put(chain);
 230	}
 231}
 232
 233static void tcf_chain_destroy(struct tcf_chain *chain)
 234{
 235	struct tcf_block *block = chain->block;
 236
 237	list_del(&chain->list);
 238	kfree(chain);
 239	if (list_empty(&block->chain_list))
 240		kfree(block);
 241}
 242
 243static void tcf_chain_hold(struct tcf_chain *chain)
 244{
 245	++chain->refcnt;
 246}
 247
 248struct tcf_chain *tcf_chain_get(struct tcf_block *block, u32 chain_index,
 249				bool create)
 250{
 251	struct tcf_chain *chain;
 252
 253	list_for_each_entry(chain, &block->chain_list, list) {
 254		if (chain->index == chain_index) {
 255			tcf_chain_hold(chain);
 256			return chain;
 257		}
 258	}
 259
 260	return create ? tcf_chain_create(block, chain_index) : NULL;
 261}
 262EXPORT_SYMBOL(tcf_chain_get);
 263
 264void tcf_chain_put(struct tcf_chain *chain)
 265{
 266	if (--chain->refcnt == 0)
 267		tcf_chain_destroy(chain);
 268}
 269EXPORT_SYMBOL(tcf_chain_put);
 270
 271static bool tcf_block_offload_in_use(struct tcf_block *block)
 272{
 273	return block->offloadcnt;
 274}
 275
 276static int tcf_block_offload_cmd(struct tcf_block *block,
 277				 struct net_device *dev,
 278				 struct tcf_block_ext_info *ei,
 279				 enum tc_block_command command)
 280{
 281	struct tc_block_offload bo = {};
 282
 283	bo.command = command;
 284	bo.binder_type = ei->binder_type;
 285	bo.block = block;
 286	return dev->netdev_ops->ndo_setup_tc(dev, TC_SETUP_BLOCK, &bo);
 287}
 288
 289static int tcf_block_offload_bind(struct tcf_block *block, struct Qdisc *q,
 290				  struct tcf_block_ext_info *ei)
 291{
 292	struct net_device *dev = q->dev_queue->dev;
 293	int err;
 294
 295	if (!dev->netdev_ops->ndo_setup_tc)
 296		goto no_offload_dev_inc;
 297
 298	/* If tc offload feature is disabled and the block we try to bind
 299	 * to already has some offloaded filters, forbid to bind.
 300	 */
 301	if (!tc_can_offload(dev) && tcf_block_offload_in_use(block))
 302		return -EOPNOTSUPP;
 303
 304	err = tcf_block_offload_cmd(block, dev, ei, TC_BLOCK_BIND);
 305	if (err == -EOPNOTSUPP)
 306		goto no_offload_dev_inc;
 307	return err;
 308
 309no_offload_dev_inc:
 310	if (tcf_block_offload_in_use(block))
 311		return -EOPNOTSUPP;
 312	block->nooffloaddevcnt++;
 313	return 0;
 314}
 315
 316static void tcf_block_offload_unbind(struct tcf_block *block, struct Qdisc *q,
 317				     struct tcf_block_ext_info *ei)
 318{
 319	struct net_device *dev = q->dev_queue->dev;
 320	int err;
 321
 322	if (!dev->netdev_ops->ndo_setup_tc)
 323		goto no_offload_dev_dec;
 324	err = tcf_block_offload_cmd(block, dev, ei, TC_BLOCK_UNBIND);
 325	if (err == -EOPNOTSUPP)
 326		goto no_offload_dev_dec;
 327	return;
 328
 329no_offload_dev_dec:
 330	WARN_ON(block->nooffloaddevcnt-- == 0);
 331}
 332
 333static int
 334tcf_chain_head_change_cb_add(struct tcf_chain *chain,
 335			     struct tcf_block_ext_info *ei,
 336			     struct netlink_ext_ack *extack)
 337{
 338	struct tcf_filter_chain_list_item *item;
 339
 340	item = kmalloc(sizeof(*item), GFP_KERNEL);
 341	if (!item) {
 342		NL_SET_ERR_MSG(extack, "Memory allocation for head change callback item failed");
 343		return -ENOMEM;
 344	}
 345	item->chain_head_change = ei->chain_head_change;
 346	item->chain_head_change_priv = ei->chain_head_change_priv;
 347	if (chain->filter_chain)
 348		tcf_chain_head_change_item(item, chain->filter_chain);
 349	list_add(&item->list, &chain->filter_chain_list);
 350	return 0;
 351}
 352
 353static void
 354tcf_chain_head_change_cb_del(struct tcf_chain *chain,
 355			     struct tcf_block_ext_info *ei)
 356{
 357	struct tcf_filter_chain_list_item *item;
 358
 359	list_for_each_entry(item, &chain->filter_chain_list, list) {
 360		if ((!ei->chain_head_change && !ei->chain_head_change_priv) ||
 361		    (item->chain_head_change == ei->chain_head_change &&
 362		     item->chain_head_change_priv == ei->chain_head_change_priv)) {
 363			tcf_chain_head_change_item(item, NULL);
 364			list_del(&item->list);
 365			kfree(item);
 366			return;
 367		}
 368	}
 369	WARN_ON(1);
 370}
 371
 372struct tcf_net {
 373	struct idr idr;
 374};
 375
 376static unsigned int tcf_net_id;
 377
 378static int tcf_block_insert(struct tcf_block *block, struct net *net,
 379			    struct netlink_ext_ack *extack)
 380{
 381	struct tcf_net *tn = net_generic(net, tcf_net_id);
 382
 383	return idr_alloc_u32(&tn->idr, block, &block->index, block->index,
 384			     GFP_KERNEL);
 385}
 386
 387static void tcf_block_remove(struct tcf_block *block, struct net *net)
 388{
 389	struct tcf_net *tn = net_generic(net, tcf_net_id);
 390
 391	idr_remove(&tn->idr, block->index);
 392}
 393
 394static struct tcf_block *tcf_block_create(struct net *net, struct Qdisc *q,
 395					  u32 block_index,
 396					  struct netlink_ext_ack *extack)
 397{
 398	struct tcf_block *block;
 399	struct tcf_chain *chain;
 400	int err;
 401
 402	block = kzalloc(sizeof(*block), GFP_KERNEL);
 403	if (!block) {
 404		NL_SET_ERR_MSG(extack, "Memory allocation for block failed");
 405		return ERR_PTR(-ENOMEM);
 406	}
 407	INIT_LIST_HEAD(&block->chain_list);
 408	INIT_LIST_HEAD(&block->cb_list);
 409	INIT_LIST_HEAD(&block->owner_list);
 410
 411	/* Create chain 0 by default, it has to be always present. */
 412	chain = tcf_chain_create(block, 0);
 413	if (!chain) {
 414		NL_SET_ERR_MSG(extack, "Failed to create new tcf chain");
 415		err = -ENOMEM;
 416		goto err_chain_create;
 417	}
 418	block->refcnt = 1;
 419	block->net = net;
 420	block->index = block_index;
 421
 422	/* Don't store q pointer for blocks which are shared */
 423	if (!tcf_block_shared(block))
 424		block->q = q;
 425	return block;
 426
 427err_chain_create:
 428	kfree(block);
 429	return ERR_PTR(err);
 430}
 431
 432static struct tcf_block *tcf_block_lookup(struct net *net, u32 block_index)
 433{
 434	struct tcf_net *tn = net_generic(net, tcf_net_id);
 435
 436	return idr_find(&tn->idr, block_index);
 437}
 438
 439static struct tcf_chain *tcf_block_chain_zero(struct tcf_block *block)
 440{
 441	return list_first_entry(&block->chain_list, struct tcf_chain, list);
 442}
 443
 444struct tcf_block_owner_item {
 445	struct list_head list;
 446	struct Qdisc *q;
 447	enum tcf_block_binder_type binder_type;
 448};
 449
 450static void
 451tcf_block_owner_netif_keep_dst(struct tcf_block *block,
 452			       struct Qdisc *q,
 453			       enum tcf_block_binder_type binder_type)
 454{
 455	if (block->keep_dst &&
 456	    binder_type != TCF_BLOCK_BINDER_TYPE_CLSACT_INGRESS &&
 457	    binder_type != TCF_BLOCK_BINDER_TYPE_CLSACT_EGRESS)
 458		netif_keep_dst(qdisc_dev(q));
 459}
 460
 461void tcf_block_netif_keep_dst(struct tcf_block *block)
 462{
 463	struct tcf_block_owner_item *item;
 464
 465	block->keep_dst = true;
 466	list_for_each_entry(item, &block->owner_list, list)
 467		tcf_block_owner_netif_keep_dst(block, item->q,
 468					       item->binder_type);
 469}
 470EXPORT_SYMBOL(tcf_block_netif_keep_dst);
 471
 472static int tcf_block_owner_add(struct tcf_block *block,
 473			       struct Qdisc *q,
 474			       enum tcf_block_binder_type binder_type)
 475{
 476	struct tcf_block_owner_item *item;
 477
 478	item = kmalloc(sizeof(*item), GFP_KERNEL);
 479	if (!item)
 480		return -ENOMEM;
 481	item->q = q;
 482	item->binder_type = binder_type;
 483	list_add(&item->list, &block->owner_list);
 484	return 0;
 485}
 486
 487static void tcf_block_owner_del(struct tcf_block *block,
 488				struct Qdisc *q,
 489				enum tcf_block_binder_type binder_type)
 490{
 491	struct tcf_block_owner_item *item;
 492
 493	list_for_each_entry(item, &block->owner_list, list) {
 494		if (item->q == q && item->binder_type == binder_type) {
 495			list_del(&item->list);
 496			kfree(item);
 497			return;
 498		}
 499	}
 500	WARN_ON(1);
 501}
 502
 503int tcf_block_get_ext(struct tcf_block **p_block, struct Qdisc *q,
 504		      struct tcf_block_ext_info *ei,
 505		      struct netlink_ext_ack *extack)
 506{
 507	struct net *net = qdisc_net(q);
 508	struct tcf_block *block = NULL;
 509	bool created = false;
 510	int err;
 511
 512	if (ei->block_index) {
 513		/* block_index not 0 means the shared block is requested */
 514		block = tcf_block_lookup(net, ei->block_index);
 515		if (block)
 516			block->refcnt++;
 517	}
 518
 519	if (!block) {
 520		block = tcf_block_create(net, q, ei->block_index, extack);
 521		if (IS_ERR(block))
 522			return PTR_ERR(block);
 523		created = true;
 524		if (tcf_block_shared(block)) {
 525			err = tcf_block_insert(block, net, extack);
 526			if (err)
 527				goto err_block_insert;
 528		}
 529	}
 530
 531	err = tcf_block_owner_add(block, q, ei->binder_type);
 532	if (err)
 533		goto err_block_owner_add;
 534
 535	tcf_block_owner_netif_keep_dst(block, q, ei->binder_type);
 536
 537	err = tcf_chain_head_change_cb_add(tcf_block_chain_zero(block),
 538					   ei, extack);
 539	if (err)
 540		goto err_chain_head_change_cb_add;
 541
 542	err = tcf_block_offload_bind(block, q, ei);
 543	if (err)
 544		goto err_block_offload_bind;
 545
 546	*p_block = block;
 547	return 0;
 548
 549err_block_offload_bind:
 550	tcf_chain_head_change_cb_del(tcf_block_chain_zero(block), ei);
 551err_chain_head_change_cb_add:
 552	tcf_block_owner_del(block, q, ei->binder_type);
 553err_block_owner_add:
 554	if (created) {
 555		if (tcf_block_shared(block))
 556			tcf_block_remove(block, net);
 557err_block_insert:
 558		kfree(tcf_block_chain_zero(block));
 559		kfree(block);
 560	} else {
 561		block->refcnt--;
 562	}
 563	return err;
 564}
 565EXPORT_SYMBOL(tcf_block_get_ext);
 566
 567static void tcf_chain_head_change_dflt(struct tcf_proto *tp_head, void *priv)
 568{
 569	struct tcf_proto __rcu **p_filter_chain = priv;
 570
 571	rcu_assign_pointer(*p_filter_chain, tp_head);
 572}
 573
 574int tcf_block_get(struct tcf_block **p_block,
 575		  struct tcf_proto __rcu **p_filter_chain, struct Qdisc *q,
 576		  struct netlink_ext_ack *extack)
 577{
 578	struct tcf_block_ext_info ei = {
 579		.chain_head_change = tcf_chain_head_change_dflt,
 580		.chain_head_change_priv = p_filter_chain,
 581	};
 582
 583	WARN_ON(!p_filter_chain);
 584	return tcf_block_get_ext(p_block, q, &ei, extack);
 585}
 586EXPORT_SYMBOL(tcf_block_get);
 587
 588/* XXX: Standalone actions are not allowed to jump to any chain, and bound
 589 * actions should be all removed after flushing.
 590 */
 591void tcf_block_put_ext(struct tcf_block *block, struct Qdisc *q,
 592		       struct tcf_block_ext_info *ei)
 593{
 594	struct tcf_chain *chain, *tmp;
 595
 596	if (!block)
 597		return;
 598	tcf_chain_head_change_cb_del(tcf_block_chain_zero(block), ei);
 599	tcf_block_owner_del(block, q, ei->binder_type);
 600
 601	if (--block->refcnt == 0) {
 602		if (tcf_block_shared(block))
 603			tcf_block_remove(block, block->net);
 604
 605		/* Hold a refcnt for all chains, so that they don't disappear
 606		 * while we are iterating.
 607		 */
 608		list_for_each_entry(chain, &block->chain_list, list)
 609			tcf_chain_hold(chain);
 610
 611		list_for_each_entry(chain, &block->chain_list, list)
 612			tcf_chain_flush(chain);
 613	}
 614
 615	tcf_block_offload_unbind(block, q, ei);
 616
 617	if (block->refcnt == 0) {
 618		/* At this point, all the chains should have refcnt >= 1. */
 619		list_for_each_entry_safe(chain, tmp, &block->chain_list, list)
 620			tcf_chain_put(chain);
 621
 622		/* Finally, put chain 0 and allow block to be freed. */
 623		tcf_chain_put(tcf_block_chain_zero(block));
 624	}
 625}
 626EXPORT_SYMBOL(tcf_block_put_ext);
 627
 628void tcf_block_put(struct tcf_block *block)
 629{
 630	struct tcf_block_ext_info ei = {0, };
 631
 632	if (!block)
 633		return;
 634	tcf_block_put_ext(block, block->q, &ei);
 635}
 636
 637EXPORT_SYMBOL(tcf_block_put);
 638
 639struct tcf_block_cb {
 640	struct list_head list;
 641	tc_setup_cb_t *cb;
 642	void *cb_ident;
 643	void *cb_priv;
 644	unsigned int refcnt;
 645};
 646
 647void *tcf_block_cb_priv(struct tcf_block_cb *block_cb)
 648{
 649	return block_cb->cb_priv;
 650}
 651EXPORT_SYMBOL(tcf_block_cb_priv);
 652
 653struct tcf_block_cb *tcf_block_cb_lookup(struct tcf_block *block,
 654					 tc_setup_cb_t *cb, void *cb_ident)
 655{	struct tcf_block_cb *block_cb;
 656
 657	list_for_each_entry(block_cb, &block->cb_list, list)
 658		if (block_cb->cb == cb && block_cb->cb_ident == cb_ident)
 659			return block_cb;
 660	return NULL;
 661}
 662EXPORT_SYMBOL(tcf_block_cb_lookup);
 663
 664void tcf_block_cb_incref(struct tcf_block_cb *block_cb)
 665{
 666	block_cb->refcnt++;
 667}
 668EXPORT_SYMBOL(tcf_block_cb_incref);
 669
 670unsigned int tcf_block_cb_decref(struct tcf_block_cb *block_cb)
 671{
 672	return --block_cb->refcnt;
 673}
 674EXPORT_SYMBOL(tcf_block_cb_decref);
 675
 676struct tcf_block_cb *__tcf_block_cb_register(struct tcf_block *block,
 677					     tc_setup_cb_t *cb, void *cb_ident,
 678					     void *cb_priv)
 679{
 680	struct tcf_block_cb *block_cb;
 681
 682	/* At this point, playback of previous block cb calls is not supported,
 683	 * so forbid to register to block which already has some offloaded
 684	 * filters present.
 685	 */
 686	if (tcf_block_offload_in_use(block))
 687		return ERR_PTR(-EOPNOTSUPP);
 688
 689	block_cb = kzalloc(sizeof(*block_cb), GFP_KERNEL);
 690	if (!block_cb)
 691		return ERR_PTR(-ENOMEM);
 692	block_cb->cb = cb;
 693	block_cb->cb_ident = cb_ident;
 694	block_cb->cb_priv = cb_priv;
 695	list_add(&block_cb->list, &block->cb_list);
 696	return block_cb;
 697}
 698EXPORT_SYMBOL(__tcf_block_cb_register);
 699
 700int tcf_block_cb_register(struct tcf_block *block,
 701			  tc_setup_cb_t *cb, void *cb_ident,
 702			  void *cb_priv)
 703{
 704	struct tcf_block_cb *block_cb;
 705
 706	block_cb = __tcf_block_cb_register(block, cb, cb_ident, cb_priv);
 707	return IS_ERR(block_cb) ? PTR_ERR(block_cb) : 0;
 708}
 709EXPORT_SYMBOL(tcf_block_cb_register);
 710
 711void __tcf_block_cb_unregister(struct tcf_block_cb *block_cb)
 712{
 713	list_del(&block_cb->list);
 714	kfree(block_cb);
 715}
 716EXPORT_SYMBOL(__tcf_block_cb_unregister);
 717
 718void tcf_block_cb_unregister(struct tcf_block *block,
 719			     tc_setup_cb_t *cb, void *cb_ident)
 720{
 721	struct tcf_block_cb *block_cb;
 722
 723	block_cb = tcf_block_cb_lookup(block, cb, cb_ident);
 724	if (!block_cb)
 725		return;
 726	__tcf_block_cb_unregister(block_cb);
 727}
 728EXPORT_SYMBOL(tcf_block_cb_unregister);
 729
 730static int tcf_block_cb_call(struct tcf_block *block, enum tc_setup_type type,
 731			     void *type_data, bool err_stop)
 732{
 733	struct tcf_block_cb *block_cb;
 734	int ok_count = 0;
 735	int err;
 736
 737	/* Make sure all netdevs sharing this block are offload-capable. */
 738	if (block->nooffloaddevcnt && err_stop)
 739		return -EOPNOTSUPP;
 740
 741	list_for_each_entry(block_cb, &block->cb_list, list) {
 742		err = block_cb->cb(type, type_data, block_cb->cb_priv);
 743		if (err) {
 744			if (err_stop)
 745				return err;
 746		} else {
 747			ok_count++;
 748		}
 749	}
 750	return ok_count;
 751}
 752
 753/* Main classifier routine: scans classifier chain attached
 754 * to this qdisc, (optionally) tests for protocol and asks
 755 * specific classifiers.
 756 */
 757int tcf_classify(struct sk_buff *skb, const struct tcf_proto *tp,
 758		 struct tcf_result *res, bool compat_mode)
 759{
 760	__be16 protocol = tc_skb_protocol(skb);
 761#ifdef CONFIG_NET_CLS_ACT
 762	const int max_reclassify_loop = 4;
 763	const struct tcf_proto *orig_tp = tp;
 764	const struct tcf_proto *first_tp;
 765	int limit = 0;
 766
 767reclassify:
 768#endif
 769	for (; tp; tp = rcu_dereference_bh(tp->next)) {
 770		int err;
 771
 772		if (tp->protocol != protocol &&
 773		    tp->protocol != htons(ETH_P_ALL))
 774			continue;
 775
 776		err = tp->classify(skb, tp, res);
 777#ifdef CONFIG_NET_CLS_ACT
 778		if (unlikely(err == TC_ACT_RECLASSIFY && !compat_mode)) {
 779			first_tp = orig_tp;
 780			goto reset;
 781		} else if (unlikely(TC_ACT_EXT_CMP(err, TC_ACT_GOTO_CHAIN))) {
 782			first_tp = res->goto_tp;
 783			goto reset;
 784		}
 785#endif
 786		if (err >= 0)
 787			return err;
 788	}
 789
 790	return TC_ACT_UNSPEC; /* signal: continue lookup */
 791#ifdef CONFIG_NET_CLS_ACT
 792reset:
 793	if (unlikely(limit++ >= max_reclassify_loop)) {
 794		net_notice_ratelimited("%u: reclassify loop, rule prio %u, protocol %02x\n",
 795				       tp->chain->block->index,
 796				       tp->prio & 0xffff,
 797				       ntohs(tp->protocol));
 798		return TC_ACT_SHOT;
 799	}
 800
 801	tp = first_tp;
 802	protocol = tc_skb_protocol(skb);
 803	goto reclassify;
 804#endif
 805}
 806EXPORT_SYMBOL(tcf_classify);
 807
 808struct tcf_chain_info {
 809	struct tcf_proto __rcu **pprev;
 810	struct tcf_proto __rcu *next;
 811};
 812
 813static struct tcf_proto *tcf_chain_tp_prev(struct tcf_chain_info *chain_info)
 814{
 815	return rtnl_dereference(*chain_info->pprev);
 816}
 817
 818static void tcf_chain_tp_insert(struct tcf_chain *chain,
 819				struct tcf_chain_info *chain_info,
 820				struct tcf_proto *tp)
 821{
 822	if (*chain_info->pprev == chain->filter_chain)
 823		tcf_chain_head_change(chain, tp);
 824	RCU_INIT_POINTER(tp->next, tcf_chain_tp_prev(chain_info));
 825	rcu_assign_pointer(*chain_info->pprev, tp);
 826	tcf_chain_hold(chain);
 827}
 828
 829static void tcf_chain_tp_remove(struct tcf_chain *chain,
 830				struct tcf_chain_info *chain_info,
 831				struct tcf_proto *tp)
 832{
 833	struct tcf_proto *next = rtnl_dereference(chain_info->next);
 834
 835	if (tp == chain->filter_chain)
 836		tcf_chain_head_change(chain, next);
 837	RCU_INIT_POINTER(*chain_info->pprev, next);
 838	tcf_chain_put(chain);
 839}
 840
 841static struct tcf_proto *tcf_chain_tp_find(struct tcf_chain *chain,
 842					   struct tcf_chain_info *chain_info,
 843					   u32 protocol, u32 prio,
 844					   bool prio_allocate)
 845{
 846	struct tcf_proto **pprev;
 847	struct tcf_proto *tp;
 848
 849	/* Check the chain for existence of proto-tcf with this priority */
 850	for (pprev = &chain->filter_chain;
 851	     (tp = rtnl_dereference(*pprev)); pprev = &tp->next) {
 852		if (tp->prio >= prio) {
 853			if (tp->prio == prio) {
 854				if (prio_allocate ||
 855				    (tp->protocol != protocol && protocol))
 856					return ERR_PTR(-EINVAL);
 857			} else {
 858				tp = NULL;
 859			}
 860			break;
 861		}
 862	}
 863	chain_info->pprev = pprev;
 864	chain_info->next = tp ? tp->next : NULL;
 865	return tp;
 866}
 867
 868static int tcf_fill_node(struct net *net, struct sk_buff *skb,
 869			 struct tcf_proto *tp, struct tcf_block *block,
 870			 struct Qdisc *q, u32 parent, void *fh,
 871			 u32 portid, u32 seq, u16 flags, int event)
 872{
 873	struct tcmsg *tcm;
 874	struct nlmsghdr  *nlh;
 875	unsigned char *b = skb_tail_pointer(skb);
 876
 877	nlh = nlmsg_put(skb, portid, seq, event, sizeof(*tcm), flags);
 878	if (!nlh)
 879		goto out_nlmsg_trim;
 880	tcm = nlmsg_data(nlh);
 881	tcm->tcm_family = AF_UNSPEC;
 882	tcm->tcm__pad1 = 0;
 883	tcm->tcm__pad2 = 0;
 884	if (q) {
 885		tcm->tcm_ifindex = qdisc_dev(q)->ifindex;
 886		tcm->tcm_parent = parent;
 887	} else {
 888		tcm->tcm_ifindex = TCM_IFINDEX_MAGIC_BLOCK;
 889		tcm->tcm_block_index = block->index;
 890	}
 891	tcm->tcm_info = TC_H_MAKE(tp->prio, tp->protocol);
 892	if (nla_put_string(skb, TCA_KIND, tp->ops->kind))
 893		goto nla_put_failure;
 894	if (nla_put_u32(skb, TCA_CHAIN, tp->chain->index))
 895		goto nla_put_failure;
 896	if (!fh) {
 897		tcm->tcm_handle = 0;
 898	} else {
 899		if (tp->ops->dump && tp->ops->dump(net, tp, fh, skb, tcm) < 0)
 900			goto nla_put_failure;
 901	}
 902	nlh->nlmsg_len = skb_tail_pointer(skb) - b;
 903	return skb->len;
 904
 905out_nlmsg_trim:
 906nla_put_failure:
 907	nlmsg_trim(skb, b);
 908	return -1;
 909}
 910
 911static int tfilter_notify(struct net *net, struct sk_buff *oskb,
 912			  struct nlmsghdr *n, struct tcf_proto *tp,
 913			  struct tcf_block *block, struct Qdisc *q,
 914			  u32 parent, void *fh, int event, bool unicast)
 915{
 916	struct sk_buff *skb;
 917	u32 portid = oskb ? NETLINK_CB(oskb).portid : 0;
 918
 919	skb = alloc_skb(NLMSG_GOODSIZE, GFP_KERNEL);
 920	if (!skb)
 921		return -ENOBUFS;
 922
 923	if (tcf_fill_node(net, skb, tp, block, q, parent, fh, portid,
 924			  n->nlmsg_seq, n->nlmsg_flags, event) <= 0) {
 925		kfree_skb(skb);
 926		return -EINVAL;
 927	}
 928
 929	if (unicast)
 930		return netlink_unicast(net->rtnl, skb, portid, MSG_DONTWAIT);
 931
 932	return rtnetlink_send(skb, net, portid, RTNLGRP_TC,
 933			      n->nlmsg_flags & NLM_F_ECHO);
 934}
 935
 936static int tfilter_del_notify(struct net *net, struct sk_buff *oskb,
 937			      struct nlmsghdr *n, struct tcf_proto *tp,
 938			      struct tcf_block *block, struct Qdisc *q,
 939			      u32 parent, void *fh, bool unicast, bool *last,
 940			      struct netlink_ext_ack *extack)
 941{
 942	struct sk_buff *skb;
 943	u32 portid = oskb ? NETLINK_CB(oskb).portid : 0;
 944	int err;
 945
 946	skb = alloc_skb(NLMSG_GOODSIZE, GFP_KERNEL);
 947	if (!skb)
 948		return -ENOBUFS;
 949
 950	if (tcf_fill_node(net, skb, tp, block, q, parent, fh, portid,
 951			  n->nlmsg_seq, n->nlmsg_flags, RTM_DELTFILTER) <= 0) {
 952		NL_SET_ERR_MSG(extack, "Failed to build del event notification");
 953		kfree_skb(skb);
 954		return -EINVAL;
 955	}
 956
 957	err = tp->ops->delete(tp, fh, last, extack);
 958	if (err) {
 959		kfree_skb(skb);
 960		return err;
 961	}
 962
 963	if (unicast)
 964		return netlink_unicast(net->rtnl, skb, portid, MSG_DONTWAIT);
 965
 966	err = rtnetlink_send(skb, net, portid, RTNLGRP_TC,
 967			     n->nlmsg_flags & NLM_F_ECHO);
 968	if (err < 0)
 969		NL_SET_ERR_MSG(extack, "Failed to send filter delete notification");
 970	return err;
 971}
 972
 973static void tfilter_notify_chain(struct net *net, struct sk_buff *oskb,
 974				 struct tcf_block *block, struct Qdisc *q,
 975				 u32 parent, struct nlmsghdr *n,
 976				 struct tcf_chain *chain, int event)
 977{
 978	struct tcf_proto *tp;
 979
 980	for (tp = rtnl_dereference(chain->filter_chain);
 981	     tp; tp = rtnl_dereference(tp->next))
 982		tfilter_notify(net, oskb, n, tp, block,
 983			       q, parent, 0, event, false);
 984}
 985
 986/* Add/change/delete/get a filter node */
 987
 988static int tc_ctl_tfilter(struct sk_buff *skb, struct nlmsghdr *n,
 989			  struct netlink_ext_ack *extack)
 990{
 991	struct net *net = sock_net(skb->sk);
 992	struct nlattr *tca[TCA_MAX + 1];
 
 993	struct tcmsg *t;
 994	u32 protocol;
 995	u32 prio;
 996	bool prio_allocate;
 997	u32 parent;
 998	u32 chain_index;
 999	struct Qdisc *q = NULL;
1000	struct tcf_chain_info chain_info;
1001	struct tcf_chain *chain = NULL;
1002	struct tcf_block *block;
1003	struct tcf_proto *tp;
 
 
1004	unsigned long cl;
1005	void *fh;
1006	int err;
1007	int tp_created;
1008
1009	if ((n->nlmsg_type != RTM_GETTFILTER) &&
1010	    !netlink_ns_capable(skb, net->user_ns, CAP_NET_ADMIN))
1011		return -EPERM;
1012
1013replay:
1014	tp_created = 0;
1015
1016	err = nlmsg_parse(n, sizeof(*t), tca, TCA_MAX, NULL, extack);
1017	if (err < 0)
1018		return err;
1019
1020	t = nlmsg_data(n);
1021	protocol = TC_H_MIN(t->tcm_info);
1022	prio = TC_H_MAJ(t->tcm_info);
1023	prio_allocate = false;
1024	parent = t->tcm_parent;
1025	cl = 0;
1026
1027	if (prio == 0) {
1028		switch (n->nlmsg_type) {
1029		case RTM_DELTFILTER:
1030			if (protocol || t->tcm_handle || tca[TCA_KIND]) {
1031				NL_SET_ERR_MSG(extack, "Cannot flush filters with protocol, handle or kind set");
1032				return -ENOENT;
1033			}
1034			break;
1035		case RTM_NEWTFILTER:
1036			/* If no priority is provided by the user,
1037			 * we allocate one.
1038			 */
1039			if (n->nlmsg_flags & NLM_F_CREATE) {
1040				prio = TC_H_MAKE(0x80000000U, 0U);
1041				prio_allocate = true;
1042				break;
1043			}
1044			/* fall-through */
1045		default:
1046			NL_SET_ERR_MSG(extack, "Invalid filter command with priority of zero");
1047			return -ENOENT;
1048		}
1049	}
1050
1051	/* Find head of filter chain. */
1052
1053	if (t->tcm_ifindex == TCM_IFINDEX_MAGIC_BLOCK) {
1054		block = tcf_block_lookup(net, t->tcm_block_index);
1055		if (!block) {
1056			NL_SET_ERR_MSG(extack, "Block of given index was not found");
1057			err = -EINVAL;
1058			goto errout;
1059		}
 
 
1060	} else {
1061		const struct Qdisc_class_ops *cops;
1062		struct net_device *dev;
1063
1064		/* Find link */
1065		dev = __dev_get_by_index(net, t->tcm_ifindex);
1066		if (!dev)
1067			return -ENODEV;
1068
1069		/* Find qdisc */
1070		if (!parent) {
1071			q = dev->qdisc;
1072			parent = q->handle;
1073		} else {
1074			q = qdisc_lookup(dev, TC_H_MAJ(t->tcm_parent));
1075			if (!q) {
1076				NL_SET_ERR_MSG(extack, "Parent Qdisc doesn't exists");
1077				return -EINVAL;
1078			}
1079		}
1080
1081		/* Is it classful? */
1082		cops = q->ops->cl_ops;
1083		if (!cops) {
1084			NL_SET_ERR_MSG(extack, "Qdisc not classful");
1085			return -EINVAL;
1086		}
1087
1088		if (!cops->tcf_block) {
1089			NL_SET_ERR_MSG(extack, "Class doesn't support blocks");
1090			return -EOPNOTSUPP;
1091		}
1092
1093		/* Do we search for filter, attached to class? */
1094		if (TC_H_MIN(parent)) {
1095			cl = cops->find(q, parent);
1096			if (cl == 0) {
1097				NL_SET_ERR_MSG(extack, "Specified class doesn't exist");
1098				return -ENOENT;
1099			}
1100		}
1101
1102		/* And the last stroke */
1103		block = cops->tcf_block(q, cl, extack);
1104		if (!block) {
1105			err = -EINVAL;
1106			goto errout;
1107		}
1108		if (tcf_block_shared(block)) {
1109			NL_SET_ERR_MSG(extack, "This filter block is shared. Please use the block index to manipulate the filters");
1110			err = -EOPNOTSUPP;
1111			goto errout;
1112		}
1113	}
1114
1115	chain_index = tca[TCA_CHAIN] ? nla_get_u32(tca[TCA_CHAIN]) : 0;
1116	if (chain_index > TC_ACT_EXT_VAL_MASK) {
1117		NL_SET_ERR_MSG(extack, "Specified chain index exceeds upper limit");
1118		err = -EINVAL;
1119		goto errout;
1120	}
1121	chain = tcf_chain_get(block, chain_index,
1122			      n->nlmsg_type == RTM_NEWTFILTER);
1123	if (!chain) {
1124		NL_SET_ERR_MSG(extack, "Cannot find specified filter chain");
1125		err = n->nlmsg_type == RTM_NEWTFILTER ? -ENOMEM : -EINVAL;
1126		goto errout;
1127	}
1128
1129	if (n->nlmsg_type == RTM_DELTFILTER && prio == 0) {
1130		tfilter_notify_chain(net, skb, block, q, parent, n,
1131				     chain, RTM_DELTFILTER);
1132		tcf_chain_flush(chain);
1133		err = 0;
1134		goto errout;
 
 
 
 
 
1135	}
1136
1137	tp = tcf_chain_tp_find(chain, &chain_info, protocol,
1138			       prio, prio_allocate);
1139	if (IS_ERR(tp)) {
1140		NL_SET_ERR_MSG(extack, "Filter with specified priority/protocol not found");
1141		err = PTR_ERR(tp);
1142		goto errout;
1143	}
1144
1145	if (tp == NULL) {
1146		/* Proto-tcf does not exist, create new one */
1147
1148		if (tca[TCA_KIND] == NULL || !protocol) {
1149			NL_SET_ERR_MSG(extack, "Filter kind and protocol must be specified");
1150			err = -EINVAL;
1151			goto errout;
1152		}
1153
 
1154		if (n->nlmsg_type != RTM_NEWTFILTER ||
1155		    !(n->nlmsg_flags & NLM_F_CREATE)) {
1156			NL_SET_ERR_MSG(extack, "Need both RTM_NEWTFILTER and NLM_F_CREATE to create a new filter");
1157			err = -ENOENT;
1158			goto errout;
1159		}
1160
1161		if (prio_allocate)
1162			prio = tcf_auto_prio(tcf_chain_tp_prev(&chain_info));
1163
1164		tp = tcf_proto_create(nla_data(tca[TCA_KIND]),
1165				      protocol, prio, chain, extack);
1166		if (IS_ERR(tp)) {
1167			err = PTR_ERR(tp);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1168			goto errout;
1169		}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1170		tp_created = 1;
1171	} else if (tca[TCA_KIND] && nla_strcmp(tca[TCA_KIND], tp->ops->kind)) {
1172		NL_SET_ERR_MSG(extack, "Specified filter kind does not match existing one");
1173		err = -EINVAL;
1174		goto errout;
1175	}
1176
1177	fh = tp->ops->get(tp, t->tcm_handle);
1178
1179	if (!fh) {
1180		if (n->nlmsg_type == RTM_DELTFILTER && t->tcm_handle == 0) {
1181			tcf_chain_tp_remove(chain, &chain_info, tp);
1182			tfilter_notify(net, skb, n, tp, block, q, parent, fh,
1183				       RTM_DELTFILTER, false);
1184			tcf_proto_destroy(tp, extack);
 
 
1185			err = 0;
1186			goto errout;
1187		}
1188
 
1189		if (n->nlmsg_type != RTM_NEWTFILTER ||
1190		    !(n->nlmsg_flags & NLM_F_CREATE)) {
1191			NL_SET_ERR_MSG(extack, "Need both RTM_NEWTFILTER and NLM_F_CREATE to create a new filter");
1192			err = -ENOENT;
1193			goto errout;
1194		}
1195	} else {
1196		bool last;
1197
1198		switch (n->nlmsg_type) {
1199		case RTM_NEWTFILTER:
 
1200			if (n->nlmsg_flags & NLM_F_EXCL) {
1201				if (tp_created)
1202					tcf_proto_destroy(tp, NULL);
1203				NL_SET_ERR_MSG(extack, "Filter already exists");
1204				err = -EEXIST;
1205				goto errout;
1206			}
1207			break;
1208		case RTM_DELTFILTER:
1209			err = tfilter_del_notify(net, skb, n, tp, block,
1210						 q, parent, fh, false, &last,
1211						 extack);
1212			if (err)
1213				goto errout;
1214			if (last) {
1215				tcf_chain_tp_remove(chain, &chain_info, tp);
1216				tcf_proto_destroy(tp, extack);
1217			}
1218			goto errout;
1219		case RTM_GETTFILTER:
1220			err = tfilter_notify(net, skb, n, tp, block, q, parent,
1221					     fh, RTM_NEWTFILTER, true);
1222			if (err < 0)
1223				NL_SET_ERR_MSG(extack, "Failed to send filter notify message");
1224			goto errout;
1225		default:
1226			NL_SET_ERR_MSG(extack, "Invalid netlink message type");
1227			err = -EINVAL;
1228			goto errout;
1229		}
1230	}
1231
1232	err = tp->ops->change(net, skb, tp, cl, t->tcm_handle, tca, &fh,
1233			      n->nlmsg_flags & NLM_F_CREATE ? TCA_ACT_NOREPLACE : TCA_ACT_REPLACE,
1234			      extack);
1235	if (err == 0) {
1236		if (tp_created)
1237			tcf_chain_tp_insert(chain, &chain_info, tp);
1238		tfilter_notify(net, skb, n, tp, block, q, parent, fh,
1239			       RTM_NEWTFILTER, false);
 
 
 
1240	} else {
1241		if (tp_created)
1242			tcf_proto_destroy(tp, NULL);
1243	}
1244
1245errout:
1246	if (chain)
1247		tcf_chain_put(chain);
1248	if (err == -EAGAIN)
1249		/* Replay the request. */
1250		goto replay;
1251	return err;
1252}
1253
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1254struct tcf_dump_args {
1255	struct tcf_walker w;
1256	struct sk_buff *skb;
1257	struct netlink_callback *cb;
1258	struct tcf_block *block;
1259	struct Qdisc *q;
1260	u32 parent;
1261};
1262
1263static int tcf_node_dump(struct tcf_proto *tp, void *n, struct tcf_walker *arg)
 
1264{
1265	struct tcf_dump_args *a = (void *)arg;
1266	struct net *net = sock_net(a->skb->sk);
1267
1268	return tcf_fill_node(net, a->skb, tp, a->block, a->q, a->parent,
1269			     n, NETLINK_CB(a->cb->skb).portid,
1270			     a->cb->nlh->nlmsg_seq, NLM_F_MULTI,
1271			     RTM_NEWTFILTER);
1272}
1273
1274static bool tcf_chain_dump(struct tcf_chain *chain, struct Qdisc *q, u32 parent,
1275			   struct sk_buff *skb, struct netlink_callback *cb,
1276			   long index_start, long *p_index)
1277{
1278	struct net *net = sock_net(skb->sk);
1279	struct tcf_block *block = chain->block;
 
 
 
 
1280	struct tcmsg *tcm = nlmsg_data(cb->nlh);
 
 
1281	struct tcf_dump_args arg;
1282	struct tcf_proto *tp;
1283
1284	for (tp = rtnl_dereference(chain->filter_chain);
1285	     tp; tp = rtnl_dereference(tp->next), (*p_index)++) {
1286		if (*p_index < index_start)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1287			continue;
1288		if (TC_H_MAJ(tcm->tcm_info) &&
1289		    TC_H_MAJ(tcm->tcm_info) != tp->prio)
1290			continue;
1291		if (TC_H_MIN(tcm->tcm_info) &&
1292		    TC_H_MIN(tcm->tcm_info) != tp->protocol)
1293			continue;
1294		if (*p_index > index_start)
1295			memset(&cb->args[1], 0,
1296			       sizeof(cb->args) - sizeof(cb->args[0]));
1297		if (cb->args[1] == 0) {
1298			if (tcf_fill_node(net, skb, tp, block, q, parent, 0,
1299					  NETLINK_CB(cb->skb).portid,
1300					  cb->nlh->nlmsg_seq, NLM_F_MULTI,
1301					  RTM_NEWTFILTER) <= 0)
1302				return false;
1303
1304			cb->args[1] = 1;
1305		}
1306		if (!tp->ops->walk)
1307			continue;
1308		arg.w.fn = tcf_node_dump;
1309		arg.skb = skb;
1310		arg.cb = cb;
1311		arg.block = block;
1312		arg.q = q;
1313		arg.parent = parent;
1314		arg.w.stop = 0;
1315		arg.w.skip = cb->args[1] - 1;
1316		arg.w.count = 0;
1317		tp->ops->walk(tp, &arg.w);
1318		cb->args[1] = arg.w.count + 1;
1319		if (arg.w.stop)
1320			return false;
1321	}
1322	return true;
1323}
1324
1325/* called with RTNL */
1326static int tc_dump_tfilter(struct sk_buff *skb, struct netlink_callback *cb)
1327{
1328	struct net *net = sock_net(skb->sk);
1329	struct nlattr *tca[TCA_MAX + 1];
1330	struct Qdisc *q = NULL;
1331	struct tcf_block *block;
1332	struct tcf_chain *chain;
1333	struct tcmsg *tcm = nlmsg_data(cb->nlh);
1334	long index_start;
1335	long index;
1336	u32 parent;
1337	int err;
1338
1339	if (nlmsg_len(cb->nlh) < sizeof(*tcm))
1340		return skb->len;
1341
1342	err = nlmsg_parse(cb->nlh, sizeof(*tcm), tca, TCA_MAX, NULL, NULL);
1343	if (err)
1344		return err;
1345
1346	if (tcm->tcm_ifindex == TCM_IFINDEX_MAGIC_BLOCK) {
1347		block = tcf_block_lookup(net, tcm->tcm_block_index);
1348		if (!block)
1349			goto out;
1350		/* If we work with block index, q is NULL and parent value
1351		 * will never be used in the following code. The check
1352		 * in tcf_fill_node prevents it. However, compiler does not
1353		 * see that far, so set parent to zero to silence the warning
1354		 * about parent being uninitialized.
1355		 */
1356		parent = 0;
1357	} else {
1358		const struct Qdisc_class_ops *cops;
1359		struct net_device *dev;
1360		unsigned long cl = 0;
1361
1362		dev = __dev_get_by_index(net, tcm->tcm_ifindex);
1363		if (!dev)
1364			return skb->len;
1365
1366		parent = tcm->tcm_parent;
1367		if (!parent) {
1368			q = dev->qdisc;
1369			parent = q->handle;
1370		} else {
1371			q = qdisc_lookup(dev, TC_H_MAJ(tcm->tcm_parent));
1372		}
1373		if (!q)
1374			goto out;
1375		cops = q->ops->cl_ops;
1376		if (!cops)
1377			goto out;
1378		if (!cops->tcf_block)
1379			goto out;
1380		if (TC_H_MIN(tcm->tcm_parent)) {
1381			cl = cops->find(q, tcm->tcm_parent);
1382			if (cl == 0)
1383				goto out;
1384		}
1385		block = cops->tcf_block(q, cl, NULL);
1386		if (!block)
1387			goto out;
1388		if (tcf_block_shared(block))
1389			q = NULL;
1390	}
1391
1392	index_start = cb->args[0];
1393	index = 0;
1394
1395	list_for_each_entry(chain, &block->chain_list, list) {
1396		if (tca[TCA_CHAIN] &&
1397		    nla_get_u32(tca[TCA_CHAIN]) != chain->index)
1398			continue;
1399		if (!tcf_chain_dump(chain, q, parent, skb, cb,
1400				    index_start, &index)) {
1401			err = -EMSGSIZE;
1402			break;
1403		}
1404	}
1405
1406	cb->args[0] = index;
1407
 
 
 
1408out:
1409	/* If we did no progress, the error (EMSGSIZE) is real */
1410	if (skb->len == 0 && err)
1411		return err;
1412	return skb->len;
1413}
1414
1415void tcf_exts_destroy(struct tcf_exts *exts)
1416{
1417#ifdef CONFIG_NET_CLS_ACT
1418	LIST_HEAD(actions);
1419
1420	ASSERT_RTNL();
1421	tcf_exts_to_list(exts, &actions);
1422	tcf_action_destroy(&actions, TCA_ACT_UNBIND);
1423	kfree(exts->actions);
1424	exts->nr_actions = 0;
1425#endif
1426}
1427EXPORT_SYMBOL(tcf_exts_destroy);
1428
1429int tcf_exts_validate(struct net *net, struct tcf_proto *tp, struct nlattr **tb,
1430		      struct nlattr *rate_tlv, struct tcf_exts *exts, bool ovr,
1431		      struct netlink_ext_ack *extack)
1432{
1433#ifdef CONFIG_NET_CLS_ACT
1434	{
1435		struct tc_action *act;
1436		size_t attr_size = 0;
1437
 
1438		if (exts->police && tb[exts->police]) {
1439			act = tcf_action_init_1(net, tp, tb[exts->police],
1440						rate_tlv, "police", ovr,
1441						TCA_ACT_BIND, extack);
1442			if (IS_ERR(act))
1443				return PTR_ERR(act);
1444
1445			act->type = exts->type = TCA_OLD_COMPAT;
1446			exts->actions[0] = act;
1447			exts->nr_actions = 1;
1448		} else if (exts->action && tb[exts->action]) {
1449			LIST_HEAD(actions);
1450			int err, i = 0;
1451
1452			err = tcf_action_init(net, tp, tb[exts->action],
1453					      rate_tlv, NULL, ovr, TCA_ACT_BIND,
1454					      &actions, &attr_size, extack);
1455			if (err)
1456				return err;
1457			list_for_each_entry(act, &actions, list)
1458				exts->actions[i++] = act;
1459			exts->nr_actions = i;
1460		}
1461		exts->net = net;
1462	}
1463#else
1464	if ((exts->action && tb[exts->action]) ||
1465	    (exts->police && tb[exts->police])) {
1466		NL_SET_ERR_MSG(extack, "Classifier actions are not supported per compile options (CONFIG_NET_CLS_ACT)");
1467		return -EOPNOTSUPP;
1468	}
1469#endif
1470
1471	return 0;
1472}
1473EXPORT_SYMBOL(tcf_exts_validate);
1474
1475void tcf_exts_change(struct tcf_exts *dst, struct tcf_exts *src)
 
1476{
1477#ifdef CONFIG_NET_CLS_ACT
1478	struct tcf_exts old = *dst;
1479
1480	*dst = *src;
1481	tcf_exts_destroy(&old);
 
 
 
 
1482#endif
1483}
1484EXPORT_SYMBOL(tcf_exts_change);
1485
1486#ifdef CONFIG_NET_CLS_ACT
1487static struct tc_action *tcf_exts_first_act(struct tcf_exts *exts)
1488{
1489	if (exts->nr_actions == 0)
1490		return NULL;
1491	else
1492		return exts->actions[0];
1493}
1494#endif
1495
1496int tcf_exts_dump(struct sk_buff *skb, struct tcf_exts *exts)
1497{
1498#ifdef CONFIG_NET_CLS_ACT
1499	struct nlattr *nest;
1500
1501	if (exts->action && tcf_exts_has_actions(exts)) {
1502		/*
1503		 * again for backward compatible mode - we want
1504		 * to work with both old and new modes of entering
1505		 * tc data even if iproute2  was newer - jhs
1506		 */
 
1507		if (exts->type != TCA_OLD_COMPAT) {
1508			LIST_HEAD(actions);
1509
1510			nest = nla_nest_start(skb, exts->action);
1511			if (nest == NULL)
1512				goto nla_put_failure;
1513
1514			tcf_exts_to_list(exts, &actions);
1515			if (tcf_action_dump(skb, &actions, 0, 0) < 0)
1516				goto nla_put_failure;
1517			nla_nest_end(skb, nest);
1518		} else if (exts->police) {
1519			struct tc_action *act = tcf_exts_first_act(exts);
1520			nest = nla_nest_start(skb, exts->police);
1521			if (nest == NULL || !act)
1522				goto nla_put_failure;
1523			if (tcf_action_dump_old(skb, act, 0, 0) < 0)
1524				goto nla_put_failure;
1525			nla_nest_end(skb, nest);
1526		}
1527	}
 
1528	return 0;
1529
1530nla_put_failure:
1531	nla_nest_cancel(skb, nest);
1532	return -1;
1533#else
1534	return 0;
1535#endif
1536}
1537EXPORT_SYMBOL(tcf_exts_dump);
1538
1539
1540int tcf_exts_dump_stats(struct sk_buff *skb, struct tcf_exts *exts)
1541{
1542#ifdef CONFIG_NET_CLS_ACT
1543	struct tc_action *a = tcf_exts_first_act(exts);
1544	if (a != NULL && tcf_action_copy_stats(skb, a, 1) < 0)
1545		return -1;
1546#endif
1547	return 0;
1548}
1549EXPORT_SYMBOL(tcf_exts_dump_stats);
1550
1551static int tc_exts_setup_cb_egdev_call(struct tcf_exts *exts,
1552				       enum tc_setup_type type,
1553				       void *type_data, bool err_stop)
1554{
1555	int ok_count = 0;
1556#ifdef CONFIG_NET_CLS_ACT
1557	const struct tc_action *a;
1558	struct net_device *dev;
1559	int i, ret;
1560
1561	if (!tcf_exts_has_actions(exts))
1562		return 0;
1563
1564	for (i = 0; i < exts->nr_actions; i++) {
1565		a = exts->actions[i];
1566		if (!a->ops->get_dev)
1567			continue;
1568		dev = a->ops->get_dev(a);
1569		if (!dev)
1570			continue;
1571		ret = tc_setup_cb_egdev_call(dev, type, type_data, err_stop);
1572		if (ret < 0)
1573			return ret;
1574		ok_count += ret;
1575	}
1576#endif
1577	return ok_count;
1578}
1579
1580int tc_setup_cb_call(struct tcf_block *block, struct tcf_exts *exts,
1581		     enum tc_setup_type type, void *type_data, bool err_stop)
1582{
1583	int ok_count;
1584	int ret;
1585
1586	ret = tcf_block_cb_call(block, type, type_data, err_stop);
1587	if (ret < 0)
1588		return ret;
1589	ok_count = ret;
1590
1591	if (!exts || ok_count)
1592		return ok_count;
1593	ret = tc_exts_setup_cb_egdev_call(exts, type, type_data, err_stop);
1594	if (ret < 0)
1595		return ret;
1596	ok_count += ret;
1597
1598	return ok_count;
1599}
1600EXPORT_SYMBOL(tc_setup_cb_call);
1601
1602static __net_init int tcf_net_init(struct net *net)
1603{
1604	struct tcf_net *tn = net_generic(net, tcf_net_id);
1605
1606	idr_init(&tn->idr);
1607	return 0;
1608}
1609
1610static void __net_exit tcf_net_exit(struct net *net)
1611{
1612	struct tcf_net *tn = net_generic(net, tcf_net_id);
1613
1614	idr_destroy(&tn->idr);
1615}
1616
1617static struct pernet_operations tcf_net_ops = {
1618	.init = tcf_net_init,
1619	.exit = tcf_net_exit,
1620	.id   = &tcf_net_id,
1621	.size = sizeof(struct tcf_net),
1622};
1623
1624static int __init tc_filter_init(void)
1625{
1626	int err;
1627
1628	tc_filter_wq = alloc_ordered_workqueue("tc_filter_workqueue", 0);
1629	if (!tc_filter_wq)
1630		return -ENOMEM;
1631
1632	err = register_pernet_subsys(&tcf_net_ops);
1633	if (err)
1634		goto err_register_pernet_subsys;
1635
1636	rtnl_register(PF_UNSPEC, RTM_NEWTFILTER, tc_ctl_tfilter, NULL, 0);
1637	rtnl_register(PF_UNSPEC, RTM_DELTFILTER, tc_ctl_tfilter, NULL, 0);
1638	rtnl_register(PF_UNSPEC, RTM_GETTFILTER, tc_ctl_tfilter,
1639		      tc_dump_tfilter, 0);
1640
1641	return 0;
1642
1643err_register_pernet_subsys:
1644	destroy_workqueue(tc_filter_wq);
1645	return err;
1646}
1647
1648subsys_initcall(tc_filter_init);
v3.15
  1/*
  2 * net/sched/cls_api.c	Packet classifier API.
  3 *
  4 *		This program is free software; you can redistribute it and/or
  5 *		modify it under the terms of the GNU General Public License
  6 *		as published by the Free Software Foundation; either version
  7 *		2 of the License, or (at your option) any later version.
  8 *
  9 * Authors:	Alexey Kuznetsov, <kuznet@ms2.inr.ac.ru>
 10 *
 11 * Changes:
 12 *
 13 * Eduardo J. Blanco <ejbs@netlabs.com.uy> :990222: kmod support
 14 *
 15 */
 16
 17#include <linux/module.h>
 18#include <linux/types.h>
 19#include <linux/kernel.h>
 20#include <linux/string.h>
 21#include <linux/errno.h>
 
 22#include <linux/skbuff.h>
 23#include <linux/init.h>
 24#include <linux/kmod.h>
 25#include <linux/err.h>
 26#include <linux/slab.h>
 
 27#include <net/net_namespace.h>
 28#include <net/sock.h>
 29#include <net/netlink.h>
 30#include <net/pkt_sched.h>
 31#include <net/pkt_cls.h>
 32
 33/* The list of all installed classifier types */
 34static LIST_HEAD(tcf_proto_base);
 35
 36/* Protects list of registered TC modules. It is pure SMP lock. */
 37static DEFINE_RWLOCK(cls_mod_lock);
 38
 39/* Find classifier type by string name */
 40
 41static const struct tcf_proto_ops *tcf_proto_lookup_ops(struct nlattr *kind)
 42{
 43	const struct tcf_proto_ops *t, *res = NULL;
 44
 45	if (kind) {
 46		read_lock(&cls_mod_lock);
 47		list_for_each_entry(t, &tcf_proto_base, head) {
 48			if (nla_strcmp(kind, t->kind) == 0) {
 49				if (try_module_get(t->owner))
 50					res = t;
 51				break;
 52			}
 53		}
 54		read_unlock(&cls_mod_lock);
 55	}
 56	return res;
 57}
 58
 59/* Register(unregister) new classifier type */
 60
 61int register_tcf_proto_ops(struct tcf_proto_ops *ops)
 62{
 63	struct tcf_proto_ops *t;
 64	int rc = -EEXIST;
 65
 66	write_lock(&cls_mod_lock);
 67	list_for_each_entry(t, &tcf_proto_base, head)
 68		if (!strcmp(ops->kind, t->kind))
 69			goto out;
 70
 71	list_add_tail(&ops->head, &tcf_proto_base);
 72	rc = 0;
 73out:
 74	write_unlock(&cls_mod_lock);
 75	return rc;
 76}
 77EXPORT_SYMBOL(register_tcf_proto_ops);
 78
 
 
 79int unregister_tcf_proto_ops(struct tcf_proto_ops *ops)
 80{
 81	struct tcf_proto_ops *t;
 82	int rc = -ENOENT;
 83
 
 
 
 
 
 
 84	write_lock(&cls_mod_lock);
 85	list_for_each_entry(t, &tcf_proto_base, head) {
 86		if (t == ops) {
 87			list_del(&t->head);
 88			rc = 0;
 89			break;
 90		}
 91	}
 92	write_unlock(&cls_mod_lock);
 93	return rc;
 94}
 95EXPORT_SYMBOL(unregister_tcf_proto_ops);
 96
 97static int tfilter_notify(struct net *net, struct sk_buff *oskb,
 98			  struct nlmsghdr *n, struct tcf_proto *tp,
 99			  unsigned long fh, int event);
100
 
101
102/* Select new prio value from the range, managed by kernel. */
103
104static inline u32 tcf_auto_prio(struct tcf_proto *tp)
105{
106	u32 first = TC_H_MAKE(0xC0000000U, 0U);
107
108	if (tp)
109		first = tp->prio - 1;
110
111	return first;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
112}
113
114/* Add/change/delete/get a filter node */
115
116static int tc_ctl_tfilter(struct sk_buff *skb, struct nlmsghdr *n)
 
117{
118	struct net *net = sock_net(skb->sk);
119	struct nlattr *tca[TCA_MAX + 1];
120	spinlock_t *root_lock;
121	struct tcmsg *t;
122	u32 protocol;
123	u32 prio;
124	u32 nprio;
125	u32 parent;
126	struct net_device *dev;
127	struct Qdisc  *q;
128	struct tcf_proto **back, **chain;
 
 
129	struct tcf_proto *tp;
130	const struct tcf_proto_ops *tp_ops;
131	const struct Qdisc_class_ops *cops;
132	unsigned long cl;
133	unsigned long fh;
134	int err;
135	int tp_created = 0;
136
137	if ((n->nlmsg_type != RTM_GETTFILTER) && !netlink_capable(skb, CAP_NET_ADMIN))
 
138		return -EPERM;
139
140replay:
141	err = nlmsg_parse(n, sizeof(*t), tca, TCA_MAX, NULL);
 
 
142	if (err < 0)
143		return err;
144
145	t = nlmsg_data(n);
146	protocol = TC_H_MIN(t->tcm_info);
147	prio = TC_H_MAJ(t->tcm_info);
148	nprio = prio;
149	parent = t->tcm_parent;
150	cl = 0;
151
152	if (prio == 0) {
153		/* If no priority is given, user wants we allocated it. */
154		if (n->nlmsg_type != RTM_NEWTFILTER ||
155		    !(n->nlmsg_flags & NLM_F_CREATE))
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
156			return -ENOENT;
157		prio = TC_H_MAKE(0x80000000U, 0U);
158	}
159
160	/* Find head of filter chain. */
161
162	/* Find link */
163	dev = __dev_get_by_index(net, t->tcm_ifindex);
164	if (dev == NULL)
165		return -ENODEV;
166
167	/* Find qdisc */
168	if (!parent) {
169		q = dev->qdisc;
170		parent = q->handle;
171	} else {
172		q = qdisc_lookup(dev, TC_H_MAJ(t->tcm_parent));
173		if (q == NULL)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
174			return -EINVAL;
175	}
176
177	/* Is it classful? */
178	cops = q->ops->cl_ops;
179	if (!cops)
180		return -EINVAL;
181
182	if (cops->tcf_chain == NULL)
183		return -EOPNOTSUPP;
 
 
 
 
 
 
184
185	/* Do we search for filter, attached to class? */
186	if (TC_H_MIN(parent)) {
187		cl = cops->get(q, parent);
188		if (cl == 0)
189			return -ENOENT;
 
 
 
 
 
 
190	}
191
192	/* And the last stroke */
193	chain = cops->tcf_chain(q, cl);
194	err = -EINVAL;
195	if (chain == NULL)
196		goto errout;
 
 
 
 
 
 
 
 
197
198	/* Check the chain for existence of proto-tcf with this priority */
199	for (back = chain; (tp = *back) != NULL; back = &tp->next) {
200		if (tp->prio >= prio) {
201			if (tp->prio == prio) {
202				if (!nprio ||
203				    (tp->protocol != protocol && protocol))
204					goto errout;
205			} else
206				tp = NULL;
207			break;
208		}
209	}
210
211	root_lock = qdisc_root_sleeping_lock(q);
 
 
 
 
 
 
212
213	if (tp == NULL) {
214		/* Proto-tcf does not exist, create new one */
215
216		if (tca[TCA_KIND] == NULL || !protocol)
 
 
217			goto errout;
 
218
219		err = -ENOENT;
220		if (n->nlmsg_type != RTM_NEWTFILTER ||
221		    !(n->nlmsg_flags & NLM_F_CREATE))
 
 
222			goto errout;
 
223
 
 
224
225		/* Create new proto tcf */
226
227		err = -ENOBUFS;
228		tp = kzalloc(sizeof(*tp), GFP_KERNEL);
229		if (tp == NULL)
230			goto errout;
231		err = -ENOENT;
232		tp_ops = tcf_proto_lookup_ops(tca[TCA_KIND]);
233		if (tp_ops == NULL) {
234#ifdef CONFIG_MODULES
235			struct nlattr *kind = tca[TCA_KIND];
236			char name[IFNAMSIZ];
237
238			if (kind != NULL &&
239			    nla_strlcpy(name, kind, IFNAMSIZ) < IFNAMSIZ) {
240				rtnl_unlock();
241				request_module("cls_%s", name);
242				rtnl_lock();
243				tp_ops = tcf_proto_lookup_ops(kind);
244				/* We dropped the RTNL semaphore in order to
245				 * perform the module load.  So, even if we
246				 * succeeded in loading the module we have to
247				 * replay the request.  We indicate this using
248				 * -EAGAIN.
249				 */
250				if (tp_ops != NULL) {
251					module_put(tp_ops->owner);
252					err = -EAGAIN;
253				}
254			}
255#endif
256			kfree(tp);
257			goto errout;
258		}
259		tp->ops = tp_ops;
260		tp->protocol = protocol;
261		tp->prio = nprio ? : TC_H_MAJ(tcf_auto_prio(*back));
262		tp->q = q;
263		tp->classify = tp_ops->classify;
264		tp->classid = parent;
265
266		err = tp_ops->init(tp);
267		if (err != 0) {
268			module_put(tp_ops->owner);
269			kfree(tp);
270			goto errout;
271		}
272
273		tp_created = 1;
274
275	} else if (tca[TCA_KIND] && nla_strcmp(tca[TCA_KIND], tp->ops->kind))
 
276		goto errout;
 
277
278	fh = tp->ops->get(tp, t->tcm_handle);
279
280	if (fh == 0) {
281		if (n->nlmsg_type == RTM_DELTFILTER && t->tcm_handle == 0) {
282			spin_lock_bh(root_lock);
283			*back = tp->next;
284			spin_unlock_bh(root_lock);
285
286			tfilter_notify(net, skb, n, tp, fh, RTM_DELTFILTER);
287			tcf_destroy(tp);
288			err = 0;
289			goto errout;
290		}
291
292		err = -ENOENT;
293		if (n->nlmsg_type != RTM_NEWTFILTER ||
294		    !(n->nlmsg_flags & NLM_F_CREATE))
 
 
295			goto errout;
 
296	} else {
 
 
297		switch (n->nlmsg_type) {
298		case RTM_NEWTFILTER:
299			err = -EEXIST;
300			if (n->nlmsg_flags & NLM_F_EXCL) {
301				if (tp_created)
302					tcf_destroy(tp);
 
 
303				goto errout;
304			}
305			break;
306		case RTM_DELTFILTER:
307			err = tp->ops->delete(tp, fh);
308			if (err == 0)
309				tfilter_notify(net, skb, n, tp, fh, RTM_DELTFILTER);
 
 
 
 
 
 
310			goto errout;
311		case RTM_GETTFILTER:
312			err = tfilter_notify(net, skb, n, tp, fh, RTM_NEWTFILTER);
 
 
 
313			goto errout;
314		default:
 
315			err = -EINVAL;
316			goto errout;
317		}
318	}
319
320	err = tp->ops->change(net, skb, tp, cl, t->tcm_handle, tca, &fh);
 
 
321	if (err == 0) {
322		if (tp_created) {
323			spin_lock_bh(root_lock);
324			tp->next = *back;
325			*back = tp;
326			spin_unlock_bh(root_lock);
327		}
328		tfilter_notify(net, skb, n, tp, fh, RTM_NEWTFILTER);
329	} else {
330		if (tp_created)
331			tcf_destroy(tp);
332	}
333
334errout:
335	if (cl)
336		cops->put(q, cl);
337	if (err == -EAGAIN)
338		/* Replay the request. */
339		goto replay;
340	return err;
341}
342
343static int tcf_fill_node(struct net *net, struct sk_buff *skb, struct tcf_proto *tp,
344			 unsigned long fh, u32 portid, u32 seq, u16 flags, int event)
345{
346	struct tcmsg *tcm;
347	struct nlmsghdr  *nlh;
348	unsigned char *b = skb_tail_pointer(skb);
349
350	nlh = nlmsg_put(skb, portid, seq, event, sizeof(*tcm), flags);
351	if (!nlh)
352		goto out_nlmsg_trim;
353	tcm = nlmsg_data(nlh);
354	tcm->tcm_family = AF_UNSPEC;
355	tcm->tcm__pad1 = 0;
356	tcm->tcm__pad2 = 0;
357	tcm->tcm_ifindex = qdisc_dev(tp->q)->ifindex;
358	tcm->tcm_parent = tp->classid;
359	tcm->tcm_info = TC_H_MAKE(tp->prio, tp->protocol);
360	if (nla_put_string(skb, TCA_KIND, tp->ops->kind))
361		goto nla_put_failure;
362	tcm->tcm_handle = fh;
363	if (RTM_DELTFILTER != event) {
364		tcm->tcm_handle = 0;
365		if (tp->ops->dump && tp->ops->dump(net, tp, fh, skb, tcm) < 0)
366			goto nla_put_failure;
367	}
368	nlh->nlmsg_len = skb_tail_pointer(skb) - b;
369	return skb->len;
370
371out_nlmsg_trim:
372nla_put_failure:
373	nlmsg_trim(skb, b);
374	return -1;
375}
376
377static int tfilter_notify(struct net *net, struct sk_buff *oskb,
378			  struct nlmsghdr *n, struct tcf_proto *tp,
379			  unsigned long fh, int event)
380{
381	struct sk_buff *skb;
382	u32 portid = oskb ? NETLINK_CB(oskb).portid : 0;
383
384	skb = alloc_skb(NLMSG_GOODSIZE, GFP_KERNEL);
385	if (!skb)
386		return -ENOBUFS;
387
388	if (tcf_fill_node(net, skb, tp, fh, portid, n->nlmsg_seq, 0, event) <= 0) {
389		kfree_skb(skb);
390		return -EINVAL;
391	}
392
393	return rtnetlink_send(skb, net, portid, RTNLGRP_TC,
394			      n->nlmsg_flags & NLM_F_ECHO);
395}
396
397struct tcf_dump_args {
398	struct tcf_walker w;
399	struct sk_buff *skb;
400	struct netlink_callback *cb;
 
 
 
401};
402
403static int tcf_node_dump(struct tcf_proto *tp, unsigned long n,
404			 struct tcf_walker *arg)
405{
406	struct tcf_dump_args *a = (void *)arg;
407	struct net *net = sock_net(a->skb->sk);
408
409	return tcf_fill_node(net, a->skb, tp, n, NETLINK_CB(a->cb->skb).portid,
410			     a->cb->nlh->nlmsg_seq, NLM_F_MULTI, RTM_NEWTFILTER);
 
 
411}
412
413/* called with RTNL */
414static int tc_dump_tfilter(struct sk_buff *skb, struct netlink_callback *cb)
 
415{
416	struct net *net = sock_net(skb->sk);
417	int t;
418	int s_t;
419	struct net_device *dev;
420	struct Qdisc *q;
421	struct tcf_proto *tp, **chain;
422	struct tcmsg *tcm = nlmsg_data(cb->nlh);
423	unsigned long cl = 0;
424	const struct Qdisc_class_ops *cops;
425	struct tcf_dump_args arg;
 
426
427	if (nlmsg_len(cb->nlh) < sizeof(*tcm))
428		return skb->len;
429	dev = __dev_get_by_index(net, tcm->tcm_ifindex);
430	if (!dev)
431		return skb->len;
432
433	if (!tcm->tcm_parent)
434		q = dev->qdisc;
435	else
436		q = qdisc_lookup(dev, TC_H_MAJ(tcm->tcm_parent));
437	if (!q)
438		goto out;
439	cops = q->ops->cl_ops;
440	if (!cops)
441		goto errout;
442	if (cops->tcf_chain == NULL)
443		goto errout;
444	if (TC_H_MIN(tcm->tcm_parent)) {
445		cl = cops->get(q, tcm->tcm_parent);
446		if (cl == 0)
447			goto errout;
448	}
449	chain = cops->tcf_chain(q, cl);
450	if (chain == NULL)
451		goto errout;
452
453	s_t = cb->args[0];
454
455	for (tp = *chain, t = 0; tp; tp = tp->next, t++) {
456		if (t < s_t)
457			continue;
458		if (TC_H_MAJ(tcm->tcm_info) &&
459		    TC_H_MAJ(tcm->tcm_info) != tp->prio)
460			continue;
461		if (TC_H_MIN(tcm->tcm_info) &&
462		    TC_H_MIN(tcm->tcm_info) != tp->protocol)
463			continue;
464		if (t > s_t)
465			memset(&cb->args[1], 0, sizeof(cb->args)-sizeof(cb->args[0]));
 
466		if (cb->args[1] == 0) {
467			if (tcf_fill_node(net, skb, tp, 0, NETLINK_CB(cb->skb).portid,
 
468					  cb->nlh->nlmsg_seq, NLM_F_MULTI,
469					  RTM_NEWTFILTER) <= 0)
470				break;
471
472			cb->args[1] = 1;
473		}
474		if (tp->ops->walk == NULL)
475			continue;
476		arg.w.fn = tcf_node_dump;
477		arg.skb = skb;
478		arg.cb = cb;
 
 
 
479		arg.w.stop = 0;
480		arg.w.skip = cb->args[1] - 1;
481		arg.w.count = 0;
482		tp->ops->walk(tp, &arg.w);
483		cb->args[1] = arg.w.count + 1;
484		if (arg.w.stop)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
485			break;
 
486	}
487
488	cb->args[0] = t;
489
490errout:
491	if (cl)
492		cops->put(q, cl);
493out:
 
 
 
494	return skb->len;
495}
496
497void tcf_exts_destroy(struct tcf_proto *tp, struct tcf_exts *exts)
498{
499#ifdef CONFIG_NET_CLS_ACT
500	tcf_action_destroy(&exts->actions, TCA_ACT_UNBIND);
501	INIT_LIST_HEAD(&exts->actions);
 
 
 
 
 
502#endif
503}
504EXPORT_SYMBOL(tcf_exts_destroy);
505
506int tcf_exts_validate(struct net *net, struct tcf_proto *tp, struct nlattr **tb,
507		  struct nlattr *rate_tlv, struct tcf_exts *exts)
 
508{
509#ifdef CONFIG_NET_CLS_ACT
510	{
511		struct tc_action *act;
 
512
513		INIT_LIST_HEAD(&exts->actions);
514		if (exts->police && tb[exts->police]) {
515			act = tcf_action_init_1(net, tb[exts->police], rate_tlv,
516						"police", TCA_ACT_NOREPLACE,
517						TCA_ACT_BIND);
518			if (IS_ERR(act))
519				return PTR_ERR(act);
520
521			act->type = exts->type = TCA_OLD_COMPAT;
522			list_add(&act->list, &exts->actions);
 
523		} else if (exts->action && tb[exts->action]) {
524			int err;
525			err = tcf_action_init(net, tb[exts->action], rate_tlv,
526					      NULL, TCA_ACT_NOREPLACE,
527					      TCA_ACT_BIND, &exts->actions);
 
 
528			if (err)
529				return err;
 
 
 
530		}
 
531	}
532#else
533	if ((exts->action && tb[exts->action]) ||
534	    (exts->police && tb[exts->police]))
 
535		return -EOPNOTSUPP;
 
536#endif
537
538	return 0;
539}
540EXPORT_SYMBOL(tcf_exts_validate);
541
542void tcf_exts_change(struct tcf_proto *tp, struct tcf_exts *dst,
543		     struct tcf_exts *src)
544{
545#ifdef CONFIG_NET_CLS_ACT
546	if (!list_empty(&src->actions)) {
547		LIST_HEAD(tmp);
548		tcf_tree_lock(tp);
549		list_splice_init(&dst->actions, &tmp);
550		list_splice(&src->actions, &dst->actions);
551		tcf_tree_unlock(tp);
552		tcf_action_destroy(&tmp, TCA_ACT_UNBIND);
553	}
554#endif
555}
556EXPORT_SYMBOL(tcf_exts_change);
557
558#define tcf_exts_first_act(ext) \
559		list_first_entry(&(exts)->actions, struct tc_action, list)
 
 
 
 
 
 
 
560
561int tcf_exts_dump(struct sk_buff *skb, struct tcf_exts *exts)
562{
563#ifdef CONFIG_NET_CLS_ACT
564	if (exts->action && !list_empty(&exts->actions)) {
 
 
565		/*
566		 * again for backward compatible mode - we want
567		 * to work with both old and new modes of entering
568		 * tc data even if iproute2  was newer - jhs
569		 */
570		struct nlattr *nest;
571		if (exts->type != TCA_OLD_COMPAT) {
 
 
572			nest = nla_nest_start(skb, exts->action);
573			if (nest == NULL)
574				goto nla_put_failure;
575			if (tcf_action_dump(skb, &exts->actions, 0, 0) < 0)
 
 
576				goto nla_put_failure;
577			nla_nest_end(skb, nest);
578		} else if (exts->police) {
579			struct tc_action *act = tcf_exts_first_act(exts);
580			nest = nla_nest_start(skb, exts->police);
581			if (nest == NULL || !act)
582				goto nla_put_failure;
583			if (tcf_action_dump_old(skb, act, 0, 0) < 0)
584				goto nla_put_failure;
585			nla_nest_end(skb, nest);
586		}
587	}
588#endif
589	return 0;
590nla_put_failure: __attribute__ ((unused))
 
 
591	return -1;
 
 
 
592}
593EXPORT_SYMBOL(tcf_exts_dump);
594
595
596int tcf_exts_dump_stats(struct sk_buff *skb, struct tcf_exts *exts)
597{
598#ifdef CONFIG_NET_CLS_ACT
599	struct tc_action *a = tcf_exts_first_act(exts);
600	if (tcf_action_copy_stats(skb, a, 1) < 0)
601		return -1;
602#endif
603	return 0;
604}
605EXPORT_SYMBOL(tcf_exts_dump_stats);
606
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
607static int __init tc_filter_init(void)
608{
609	rtnl_register(PF_UNSPEC, RTM_NEWTFILTER, tc_ctl_tfilter, NULL, NULL);
610	rtnl_register(PF_UNSPEC, RTM_DELTFILTER, tc_ctl_tfilter, NULL, NULL);
 
 
 
 
 
 
 
 
 
 
611	rtnl_register(PF_UNSPEC, RTM_GETTFILTER, tc_ctl_tfilter,
612		      tc_dump_tfilter, NULL);
613
614	return 0;
 
 
 
 
615}
616
617subsys_initcall(tc_filter_init);