2006-08-04 10:38:38 +00:00
|
|
|
/*
|
|
|
|
* net/core/fib_rules.c Generic Routing Rules
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU General Public License as
|
|
|
|
* published by the Free Software Foundation, version 2.
|
|
|
|
*
|
|
|
|
* Authors: Thomas Graf <tgraf@suug.ch>
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/types.h>
|
|
|
|
#include <linux/kernel.h>
|
include cleanup: Update gfp.h and slab.h includes to prepare for breaking implicit slab.h inclusion from percpu.h
percpu.h is included by sched.h and module.h and thus ends up being
included when building most .c files. percpu.h includes slab.h which
in turn includes gfp.h making everything defined by the two files
universally available and complicating inclusion dependencies.
percpu.h -> slab.h dependency is about to be removed. Prepare for
this change by updating users of gfp and slab facilities include those
headers directly instead of assuming availability. As this conversion
needs to touch large number of source files, the following script is
used as the basis of conversion.
http://userweb.kernel.org/~tj/misc/slabh-sweep.py
The script does the followings.
* Scan files for gfp and slab usages and update includes such that
only the necessary includes are there. ie. if only gfp is used,
gfp.h, if slab is used, slab.h.
* When the script inserts a new include, it looks at the include
blocks and try to put the new include such that its order conforms
to its surrounding. It's put in the include block which contains
core kernel includes, in the same order that the rest are ordered -
alphabetical, Christmas tree, rev-Xmas-tree or at the end if there
doesn't seem to be any matching order.
* If the script can't find a place to put a new include (mostly
because the file doesn't have fitting include block), it prints out
an error message indicating which .h file needs to be added to the
file.
The conversion was done in the following steps.
1. The initial automatic conversion of all .c files updated slightly
over 4000 files, deleting around 700 includes and adding ~480 gfp.h
and ~3000 slab.h inclusions. The script emitted errors for ~400
files.
2. Each error was manually checked. Some didn't need the inclusion,
some needed manual addition while adding it to implementation .h or
embedding .c file was more appropriate for others. This step added
inclusions to around 150 files.
3. The script was run again and the output was compared to the edits
from #2 to make sure no file was left behind.
4. Several build tests were done and a couple of problems were fixed.
e.g. lib/decompress_*.c used malloc/free() wrappers around slab
APIs requiring slab.h to be added manually.
5. The script was run on all .h files but without automatically
editing them as sprinkling gfp.h and slab.h inclusions around .h
files could easily lead to inclusion dependency hell. Most gfp.h
inclusion directives were ignored as stuff from gfp.h was usually
wildly available and often used in preprocessor macros. Each
slab.h inclusion directive was examined and added manually as
necessary.
6. percpu.h was updated not to include slab.h.
7. Build test were done on the following configurations and failures
were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my
distributed build env didn't work with gcov compiles) and a few
more options had to be turned off depending on archs to make things
build (like ipr on powerpc/64 which failed due to missing writeq).
* x86 and x86_64 UP and SMP allmodconfig and a custom test config.
* powerpc and powerpc64 SMP allmodconfig
* sparc and sparc64 SMP allmodconfig
* ia64 SMP allmodconfig
* s390 SMP allmodconfig
* alpha SMP allmodconfig
* um on x86_64 SMP allmodconfig
8. percpu.h modifications were reverted so that it could be applied as
a separate patch and serve as bisection point.
Given the fact that I had only a couple of failures from tests on step
6, I'm fairly confident about the coverage of this conversion patch.
If there is a breakage, it's likely to be something in one of the arch
headers which should be easily discoverable easily on most builds of
the specific arch.
Signed-off-by: Tejun Heo <tj@kernel.org>
Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
2010-03-24 08:04:11 +00:00
|
|
|
#include <linux/slab.h>
|
2006-08-04 10:38:38 +00:00
|
|
|
#include <linux/list.h>
|
2007-09-12 11:02:17 +00:00
|
|
|
#include <net/net_namespace.h>
|
2007-09-17 18:56:21 +00:00
|
|
|
#include <net/sock.h>
|
2006-08-04 10:38:38 +00:00
|
|
|
#include <net/fib_rules.h>
|
|
|
|
|
2007-11-11 06:12:03 +00:00
|
|
|
int fib_default_rule_add(struct fib_rules_ops *ops,
|
|
|
|
u32 pref, u32 table, u32 flags)
|
|
|
|
{
|
|
|
|
struct fib_rule *r;
|
|
|
|
|
|
|
|
r = kzalloc(ops->rule_size, GFP_KERNEL);
|
|
|
|
if (r == NULL)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
atomic_set(&r->refcnt, 1);
|
|
|
|
r->action = FR_ACT_TO_TBL;
|
|
|
|
r->pref = pref;
|
|
|
|
r->table = table;
|
|
|
|
r->flags = flags;
|
2008-04-16 09:01:56 +00:00
|
|
|
r->fr_net = hold_net(ops->fro_net);
|
2007-11-11 06:12:03 +00:00
|
|
|
|
|
|
|
/* The lock is not required here, the list in unreacheable
|
|
|
|
* at the moment this function is called */
|
|
|
|
list_add_tail(&r->list, &ops->rules_list);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(fib_default_rule_add);
|
|
|
|
|
2010-04-13 05:03:15 +00:00
|
|
|
u32 fib_default_rule_pref(struct fib_rules_ops *ops)
|
|
|
|
{
|
|
|
|
struct list_head *pos;
|
|
|
|
struct fib_rule *rule;
|
|
|
|
|
|
|
|
if (!list_empty(&ops->rules_list)) {
|
|
|
|
pos = ops->rules_list.next;
|
|
|
|
if (pos->next != &ops->rules_list) {
|
|
|
|
rule = list_entry(pos->next, struct fib_rule, list);
|
|
|
|
if (rule->pref)
|
|
|
|
return rule->pref - 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(fib_default_rule_pref);
|
|
|
|
|
2008-01-21 00:46:41 +00:00
|
|
|
static void notify_rule_change(int event, struct fib_rule *rule,
|
2006-08-15 07:32:48 +00:00
|
|
|
struct fib_rules_ops *ops, struct nlmsghdr *nlh,
|
|
|
|
u32 pid);
|
2006-08-04 10:38:38 +00:00
|
|
|
|
2008-01-10 11:20:28 +00:00
|
|
|
static struct fib_rules_ops *lookup_rules_ops(struct net *net, int family)
|
2006-08-04 10:38:38 +00:00
|
|
|
{
|
|
|
|
struct fib_rules_ops *ops;
|
|
|
|
|
|
|
|
rcu_read_lock();
|
2008-01-10 11:20:28 +00:00
|
|
|
list_for_each_entry_rcu(ops, &net->rules_ops, list) {
|
2006-08-04 10:38:38 +00:00
|
|
|
if (ops->family == family) {
|
|
|
|
if (!try_module_get(ops->owner))
|
|
|
|
ops = NULL;
|
|
|
|
rcu_read_unlock();
|
|
|
|
return ops;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
rcu_read_unlock();
|
|
|
|
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void rules_ops_put(struct fib_rules_ops *ops)
|
|
|
|
{
|
|
|
|
if (ops)
|
|
|
|
module_put(ops->owner);
|
|
|
|
}
|
|
|
|
|
2007-03-27 20:56:52 +00:00
|
|
|
static void flush_route_cache(struct fib_rules_ops *ops)
|
|
|
|
{
|
|
|
|
if (ops->flush_cache)
|
2008-07-06 02:01:28 +00:00
|
|
|
ops->flush_cache(ops);
|
2007-03-27 20:56:52 +00:00
|
|
|
}
|
|
|
|
|
2009-12-03 20:22:55 +00:00
|
|
|
static int __fib_rules_register(struct fib_rules_ops *ops)
|
2006-08-04 10:38:38 +00:00
|
|
|
{
|
|
|
|
int err = -EEXIST;
|
|
|
|
struct fib_rules_ops *o;
|
2008-01-21 00:46:41 +00:00
|
|
|
struct net *net;
|
|
|
|
|
|
|
|
net = ops->fro_net;
|
2006-08-04 10:38:38 +00:00
|
|
|
|
|
|
|
if (ops->rule_size < sizeof(struct fib_rule))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (ops->match == NULL || ops->configure == NULL ||
|
|
|
|
ops->compare == NULL || ops->fill == NULL ||
|
|
|
|
ops->action == NULL)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2008-01-10 11:20:28 +00:00
|
|
|
spin_lock(&net->rules_mod_lock);
|
|
|
|
list_for_each_entry(o, &net->rules_ops, list)
|
2006-08-04 10:38:38 +00:00
|
|
|
if (ops->family == o->family)
|
|
|
|
goto errout;
|
|
|
|
|
2008-01-10 11:20:28 +00:00
|
|
|
hold_net(net);
|
|
|
|
list_add_tail_rcu(&ops->list, &net->rules_ops);
|
2006-08-04 10:38:38 +00:00
|
|
|
err = 0;
|
|
|
|
errout:
|
2008-01-10 11:20:28 +00:00
|
|
|
spin_unlock(&net->rules_mod_lock);
|
2006-08-04 10:38:38 +00:00
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2009-12-03 20:22:55 +00:00
|
|
|
struct fib_rules_ops *
|
2010-04-26 14:02:04 +00:00
|
|
|
fib_rules_register(const struct fib_rules_ops *tmpl, struct net *net)
|
2009-12-03 20:22:55 +00:00
|
|
|
{
|
|
|
|
struct fib_rules_ops *ops;
|
|
|
|
int err;
|
|
|
|
|
2010-03-09 20:03:38 +00:00
|
|
|
ops = kmemdup(tmpl, sizeof(*ops), GFP_KERNEL);
|
2009-12-03 20:22:55 +00:00
|
|
|
if (ops == NULL)
|
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
|
|
|
|
INIT_LIST_HEAD(&ops->rules_list);
|
|
|
|
ops->fro_net = net;
|
|
|
|
|
|
|
|
err = __fib_rules_register(ops);
|
|
|
|
if (err) {
|
|
|
|
kfree(ops);
|
|
|
|
ops = ERR_PTR(err);
|
|
|
|
}
|
|
|
|
|
|
|
|
return ops;
|
|
|
|
}
|
2006-08-04 10:38:38 +00:00
|
|
|
EXPORT_SYMBOL_GPL(fib_rules_register);
|
|
|
|
|
2010-10-04 20:14:17 +00:00
|
|
|
static void fib_rules_cleanup_ops(struct fib_rules_ops *ops)
|
2006-08-04 10:38:38 +00:00
|
|
|
{
|
|
|
|
struct fib_rule *rule, *tmp;
|
|
|
|
|
2007-09-16 22:44:27 +00:00
|
|
|
list_for_each_entry_safe(rule, tmp, &ops->rules_list, list) {
|
2006-08-04 10:38:38 +00:00
|
|
|
list_del_rcu(&rule->list);
|
|
|
|
fib_rule_put(rule);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2009-12-03 20:22:55 +00:00
|
|
|
static void fib_rules_put_rcu(struct rcu_head *head)
|
|
|
|
{
|
|
|
|
struct fib_rules_ops *ops = container_of(head, struct fib_rules_ops, rcu);
|
|
|
|
struct net *net = ops->fro_net;
|
|
|
|
|
|
|
|
release_net(net);
|
|
|
|
kfree(ops);
|
|
|
|
}
|
|
|
|
|
2008-01-21 00:46:41 +00:00
|
|
|
void fib_rules_unregister(struct fib_rules_ops *ops)
|
2006-08-04 10:38:38 +00:00
|
|
|
{
|
2008-01-21 00:46:41 +00:00
|
|
|
struct net *net = ops->fro_net;
|
2006-08-04 10:38:38 +00:00
|
|
|
|
2008-01-10 11:20:28 +00:00
|
|
|
spin_lock(&net->rules_mod_lock);
|
2008-01-15 06:59:30 +00:00
|
|
|
list_del_rcu(&ops->list);
|
|
|
|
fib_rules_cleanup_ops(ops);
|
2008-01-10 11:20:28 +00:00
|
|
|
spin_unlock(&net->rules_mod_lock);
|
2006-08-04 10:38:38 +00:00
|
|
|
|
2009-12-03 20:22:55 +00:00
|
|
|
call_rcu(&ops->rcu, fib_rules_put_rcu);
|
2006-08-04 10:38:38 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(fib_rules_unregister);
|
|
|
|
|
2006-11-09 23:23:20 +00:00
|
|
|
static int fib_rule_match(struct fib_rule *rule, struct fib_rules_ops *ops,
|
|
|
|
struct flowi *fl, int flags)
|
|
|
|
{
|
|
|
|
int ret = 0;
|
|
|
|
|
2010-12-23 20:03:57 +00:00
|
|
|
if (rule->iifindex && (rule->iifindex != fl->iif))
|
2006-11-09 23:23:20 +00:00
|
|
|
goto out;
|
|
|
|
|
2009-12-03 01:25:56 +00:00
|
|
|
if (rule->oifindex && (rule->oifindex != fl->oif))
|
|
|
|
goto out;
|
|
|
|
|
2006-11-09 23:23:20 +00:00
|
|
|
if ((rule->mark ^ fl->mark) & rule->mark_mask)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
ret = ops->match(rule, fl, flags);
|
|
|
|
out:
|
|
|
|
return (rule->flags & FIB_RULE_INVERT) ? !ret : ret;
|
|
|
|
}
|
|
|
|
|
2006-08-04 10:38:38 +00:00
|
|
|
int fib_rules_lookup(struct fib_rules_ops *ops, struct flowi *fl,
|
|
|
|
int flags, struct fib_lookup_arg *arg)
|
|
|
|
{
|
|
|
|
struct fib_rule *rule;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
rcu_read_lock();
|
|
|
|
|
2007-09-16 22:44:27 +00:00
|
|
|
list_for_each_entry_rcu(rule, &ops->rules_list, list) {
|
2007-03-27 00:14:15 +00:00
|
|
|
jumped:
|
2006-11-09 23:23:20 +00:00
|
|
|
if (!fib_rule_match(rule, ops, fl, flags))
|
2006-08-04 10:38:38 +00:00
|
|
|
continue;
|
|
|
|
|
2007-03-27 00:14:15 +00:00
|
|
|
if (rule->action == FR_ACT_GOTO) {
|
|
|
|
struct fib_rule *target;
|
|
|
|
|
|
|
|
target = rcu_dereference(rule->ctarget);
|
|
|
|
if (target == NULL) {
|
|
|
|
continue;
|
|
|
|
} else {
|
|
|
|
rule = target;
|
|
|
|
goto jumped;
|
|
|
|
}
|
2007-03-27 00:38:53 +00:00
|
|
|
} else if (rule->action == FR_ACT_NOP)
|
|
|
|
continue;
|
|
|
|
else
|
2007-03-27 00:14:15 +00:00
|
|
|
err = ops->action(rule, fl, flags, arg);
|
|
|
|
|
2006-08-04 10:38:38 +00:00
|
|
|
if (err != -EAGAIN) {
|
2010-10-05 10:41:36 +00:00
|
|
|
if ((arg->flags & FIB_LOOKUP_NOREF) ||
|
|
|
|
likely(atomic_inc_not_zero(&rule->refcnt))) {
|
2010-09-27 04:18:27 +00:00
|
|
|
arg->rule = rule;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
break;
|
2006-08-04 10:38:38 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2007-03-30 20:34:27 +00:00
|
|
|
err = -ESRCH;
|
2006-08-04 10:38:38 +00:00
|
|
|
out:
|
|
|
|
rcu_read_unlock();
|
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(fib_rules_lookup);
|
|
|
|
|
2007-03-24 19:46:02 +00:00
|
|
|
static int validate_rulemsg(struct fib_rule_hdr *frh, struct nlattr **tb,
|
|
|
|
struct fib_rules_ops *ops)
|
|
|
|
{
|
|
|
|
int err = -EINVAL;
|
|
|
|
|
|
|
|
if (frh->src_len)
|
|
|
|
if (tb[FRA_SRC] == NULL ||
|
|
|
|
frh->src_len > (ops->addr_size * 8) ||
|
|
|
|
nla_len(tb[FRA_SRC]) != ops->addr_size)
|
|
|
|
goto errout;
|
|
|
|
|
|
|
|
if (frh->dst_len)
|
|
|
|
if (tb[FRA_DST] == NULL ||
|
|
|
|
frh->dst_len > (ops->addr_size * 8) ||
|
|
|
|
nla_len(tb[FRA_DST]) != ops->addr_size)
|
|
|
|
goto errout;
|
|
|
|
|
|
|
|
err = 0;
|
|
|
|
errout:
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2007-03-26 06:20:05 +00:00
|
|
|
static int fib_nl_newrule(struct sk_buff *skb, struct nlmsghdr* nlh, void *arg)
|
2006-08-04 10:38:38 +00:00
|
|
|
{
|
2008-03-25 17:26:21 +00:00
|
|
|
struct net *net = sock_net(skb->sk);
|
2006-08-04 10:38:38 +00:00
|
|
|
struct fib_rule_hdr *frh = nlmsg_data(nlh);
|
|
|
|
struct fib_rules_ops *ops = NULL;
|
|
|
|
struct fib_rule *rule, *r, *last = NULL;
|
|
|
|
struct nlattr *tb[FRA_MAX+1];
|
2007-03-27 00:14:15 +00:00
|
|
|
int err = -EINVAL, unresolved = 0;
|
2006-08-04 10:38:38 +00:00
|
|
|
|
|
|
|
if (nlh->nlmsg_len < nlmsg_msg_size(sizeof(*frh)))
|
|
|
|
goto errout;
|
|
|
|
|
2008-01-10 11:20:28 +00:00
|
|
|
ops = lookup_rules_ops(net, frh->family);
|
2006-08-04 10:38:38 +00:00
|
|
|
if (ops == NULL) {
|
2008-07-02 02:59:37 +00:00
|
|
|
err = -EAFNOSUPPORT;
|
2006-08-04 10:38:38 +00:00
|
|
|
goto errout;
|
|
|
|
}
|
|
|
|
|
|
|
|
err = nlmsg_parse(nlh, sizeof(*frh), tb, FRA_MAX, ops->policy);
|
|
|
|
if (err < 0)
|
|
|
|
goto errout;
|
|
|
|
|
2007-03-24 19:46:02 +00:00
|
|
|
err = validate_rulemsg(frh, tb, ops);
|
|
|
|
if (err < 0)
|
|
|
|
goto errout;
|
|
|
|
|
2006-08-04 10:38:38 +00:00
|
|
|
rule = kzalloc(ops->rule_size, GFP_KERNEL);
|
|
|
|
if (rule == NULL) {
|
|
|
|
err = -ENOMEM;
|
|
|
|
goto errout;
|
|
|
|
}
|
2008-04-16 09:01:56 +00:00
|
|
|
rule->fr_net = hold_net(net);
|
2006-08-04 10:38:38 +00:00
|
|
|
|
|
|
|
if (tb[FRA_PRIORITY])
|
|
|
|
rule->pref = nla_get_u32(tb[FRA_PRIORITY]);
|
|
|
|
|
2009-12-03 01:25:54 +00:00
|
|
|
if (tb[FRA_IIFNAME]) {
|
2006-08-04 10:38:38 +00:00
|
|
|
struct net_device *dev;
|
|
|
|
|
2009-12-03 01:25:54 +00:00
|
|
|
rule->iifindex = -1;
|
|
|
|
nla_strlcpy(rule->iifname, tb[FRA_IIFNAME], IFNAMSIZ);
|
|
|
|
dev = __dev_get_by_name(net, rule->iifname);
|
2006-08-04 10:38:38 +00:00
|
|
|
if (dev)
|
2009-12-03 01:25:54 +00:00
|
|
|
rule->iifindex = dev->ifindex;
|
2006-08-04 10:38:38 +00:00
|
|
|
}
|
|
|
|
|
2009-12-03 01:25:56 +00:00
|
|
|
if (tb[FRA_OIFNAME]) {
|
|
|
|
struct net_device *dev;
|
|
|
|
|
|
|
|
rule->oifindex = -1;
|
|
|
|
nla_strlcpy(rule->oifname, tb[FRA_OIFNAME], IFNAMSIZ);
|
|
|
|
dev = __dev_get_by_name(net, rule->oifname);
|
|
|
|
if (dev)
|
|
|
|
rule->oifindex = dev->ifindex;
|
|
|
|
}
|
|
|
|
|
2006-11-09 23:22:18 +00:00
|
|
|
if (tb[FRA_FWMARK]) {
|
|
|
|
rule->mark = nla_get_u32(tb[FRA_FWMARK]);
|
|
|
|
if (rule->mark)
|
|
|
|
/* compatibility: if the mark value is non-zero all bits
|
|
|
|
* are compared unless a mask is explicitly specified.
|
|
|
|
*/
|
|
|
|
rule->mark_mask = 0xFFFFFFFF;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (tb[FRA_FWMASK])
|
|
|
|
rule->mark_mask = nla_get_u32(tb[FRA_FWMASK]);
|
|
|
|
|
2006-08-04 10:38:38 +00:00
|
|
|
rule->action = frh->action;
|
|
|
|
rule->flags = frh->flags;
|
2006-08-11 06:09:48 +00:00
|
|
|
rule->table = frh_get_table(frh, tb);
|
2006-08-04 10:38:38 +00:00
|
|
|
|
2009-12-03 01:25:57 +00:00
|
|
|
if (!tb[FRA_PRIORITY] && ops->default_pref)
|
2008-01-10 11:18:25 +00:00
|
|
|
rule->pref = ops->default_pref(ops);
|
2006-08-04 10:38:38 +00:00
|
|
|
|
2007-03-27 00:14:15 +00:00
|
|
|
err = -EINVAL;
|
|
|
|
if (tb[FRA_GOTO]) {
|
|
|
|
if (rule->action != FR_ACT_GOTO)
|
|
|
|
goto errout_free;
|
|
|
|
|
|
|
|
rule->target = nla_get_u32(tb[FRA_GOTO]);
|
|
|
|
/* Backward jumps are prohibited to avoid endless loops */
|
|
|
|
if (rule->target <= rule->pref)
|
|
|
|
goto errout_free;
|
|
|
|
|
2007-09-16 22:44:27 +00:00
|
|
|
list_for_each_entry(r, &ops->rules_list, list) {
|
2007-03-27 00:14:15 +00:00
|
|
|
if (r->pref == rule->target) {
|
2010-10-26 09:24:55 +00:00
|
|
|
RCU_INIT_POINTER(rule->ctarget, r);
|
2007-03-27 00:14:15 +00:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2010-10-26 09:24:55 +00:00
|
|
|
if (rcu_dereference_protected(rule->ctarget, 1) == NULL)
|
2007-03-27 00:14:15 +00:00
|
|
|
unresolved = 1;
|
|
|
|
} else if (rule->action == FR_ACT_GOTO)
|
|
|
|
goto errout_free;
|
|
|
|
|
2009-05-11 05:52:49 +00:00
|
|
|
err = ops->configure(rule, skb, frh, tb);
|
2006-08-04 10:38:38 +00:00
|
|
|
if (err < 0)
|
|
|
|
goto errout_free;
|
|
|
|
|
2007-09-16 22:44:27 +00:00
|
|
|
list_for_each_entry(r, &ops->rules_list, list) {
|
2006-08-04 10:38:38 +00:00
|
|
|
if (r->pref > rule->pref)
|
|
|
|
break;
|
|
|
|
last = r;
|
|
|
|
}
|
|
|
|
|
|
|
|
fib_rule_get(rule);
|
|
|
|
|
2010-10-23 09:44:25 +00:00
|
|
|
if (last)
|
|
|
|
list_add_rcu(&rule->list, &last->list);
|
|
|
|
else
|
|
|
|
list_add_rcu(&rule->list, &ops->rules_list);
|
|
|
|
|
2007-03-27 00:14:15 +00:00
|
|
|
if (ops->unresolved_rules) {
|
|
|
|
/*
|
|
|
|
* There are unresolved goto rules in the list, check if
|
|
|
|
* any of them are pointing to this new rule.
|
|
|
|
*/
|
2007-09-16 22:44:27 +00:00
|
|
|
list_for_each_entry(r, &ops->rules_list, list) {
|
2007-03-27 00:14:15 +00:00
|
|
|
if (r->action == FR_ACT_GOTO &&
|
|
|
|
r->target == rule->pref) {
|
2010-10-26 09:24:55 +00:00
|
|
|
BUG_ON(rtnl_dereference(r->ctarget) != NULL);
|
2007-03-27 00:14:15 +00:00
|
|
|
rcu_assign_pointer(r->ctarget, rule);
|
|
|
|
if (--ops->unresolved_rules == 0)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (rule->action == FR_ACT_GOTO)
|
|
|
|
ops->nr_goto_rules++;
|
|
|
|
|
|
|
|
if (unresolved)
|
|
|
|
ops->unresolved_rules++;
|
|
|
|
|
2008-01-21 00:46:41 +00:00
|
|
|
notify_rule_change(RTM_NEWRULE, rule, ops, nlh, NETLINK_CB(skb).pid);
|
2007-03-27 20:56:52 +00:00
|
|
|
flush_route_cache(ops);
|
2006-08-04 10:38:38 +00:00
|
|
|
rules_ops_put(ops);
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
errout_free:
|
2008-04-16 09:01:56 +00:00
|
|
|
release_net(rule->fr_net);
|
2006-08-04 10:38:38 +00:00
|
|
|
kfree(rule);
|
|
|
|
errout:
|
|
|
|
rules_ops_put(ops);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2007-03-26 06:20:05 +00:00
|
|
|
static int fib_nl_delrule(struct sk_buff *skb, struct nlmsghdr* nlh, void *arg)
|
2006-08-04 10:38:38 +00:00
|
|
|
{
|
2008-03-25 17:26:21 +00:00
|
|
|
struct net *net = sock_net(skb->sk);
|
2006-08-04 10:38:38 +00:00
|
|
|
struct fib_rule_hdr *frh = nlmsg_data(nlh);
|
|
|
|
struct fib_rules_ops *ops = NULL;
|
2007-03-27 00:14:15 +00:00
|
|
|
struct fib_rule *rule, *tmp;
|
2006-08-04 10:38:38 +00:00
|
|
|
struct nlattr *tb[FRA_MAX+1];
|
|
|
|
int err = -EINVAL;
|
|
|
|
|
|
|
|
if (nlh->nlmsg_len < nlmsg_msg_size(sizeof(*frh)))
|
|
|
|
goto errout;
|
|
|
|
|
2008-01-10 11:20:28 +00:00
|
|
|
ops = lookup_rules_ops(net, frh->family);
|
2006-08-04 10:38:38 +00:00
|
|
|
if (ops == NULL) {
|
2008-07-02 02:59:37 +00:00
|
|
|
err = -EAFNOSUPPORT;
|
2006-08-04 10:38:38 +00:00
|
|
|
goto errout;
|
|
|
|
}
|
|
|
|
|
|
|
|
err = nlmsg_parse(nlh, sizeof(*frh), tb, FRA_MAX, ops->policy);
|
|
|
|
if (err < 0)
|
|
|
|
goto errout;
|
|
|
|
|
2007-03-24 19:46:02 +00:00
|
|
|
err = validate_rulemsg(frh, tb, ops);
|
|
|
|
if (err < 0)
|
|
|
|
goto errout;
|
|
|
|
|
2007-09-16 22:44:27 +00:00
|
|
|
list_for_each_entry(rule, &ops->rules_list, list) {
|
2006-08-04 10:38:38 +00:00
|
|
|
if (frh->action && (frh->action != rule->action))
|
|
|
|
continue;
|
|
|
|
|
2006-08-11 06:09:48 +00:00
|
|
|
if (frh->table && (frh_get_table(frh, tb) != rule->table))
|
2006-08-04 10:38:38 +00:00
|
|
|
continue;
|
|
|
|
|
|
|
|
if (tb[FRA_PRIORITY] &&
|
|
|
|
(rule->pref != nla_get_u32(tb[FRA_PRIORITY])))
|
|
|
|
continue;
|
|
|
|
|
2009-12-03 01:25:54 +00:00
|
|
|
if (tb[FRA_IIFNAME] &&
|
|
|
|
nla_strcmp(tb[FRA_IIFNAME], rule->iifname))
|
2006-08-04 10:38:38 +00:00
|
|
|
continue;
|
|
|
|
|
2009-12-03 01:25:56 +00:00
|
|
|
if (tb[FRA_OIFNAME] &&
|
|
|
|
nla_strcmp(tb[FRA_OIFNAME], rule->oifname))
|
|
|
|
continue;
|
|
|
|
|
2006-11-09 23:22:18 +00:00
|
|
|
if (tb[FRA_FWMARK] &&
|
|
|
|
(rule->mark != nla_get_u32(tb[FRA_FWMARK])))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (tb[FRA_FWMASK] &&
|
|
|
|
(rule->mark_mask != nla_get_u32(tb[FRA_FWMASK])))
|
|
|
|
continue;
|
|
|
|
|
2006-08-04 10:38:38 +00:00
|
|
|
if (!ops->compare(rule, frh, tb))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (rule->flags & FIB_RULE_PERMANENT) {
|
|
|
|
err = -EPERM;
|
|
|
|
goto errout;
|
|
|
|
}
|
|
|
|
|
|
|
|
list_del_rcu(&rule->list);
|
2007-03-27 00:14:15 +00:00
|
|
|
|
|
|
|
if (rule->action == FR_ACT_GOTO)
|
|
|
|
ops->nr_goto_rules--;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Check if this rule is a target to any of them. If so,
|
|
|
|
* disable them. As this operation is eventually very
|
|
|
|
* expensive, it is only performed if goto rules have
|
|
|
|
* actually been added.
|
|
|
|
*/
|
|
|
|
if (ops->nr_goto_rules > 0) {
|
2007-09-16 22:44:27 +00:00
|
|
|
list_for_each_entry(tmp, &ops->rules_list, list) {
|
2010-10-26 09:24:55 +00:00
|
|
|
if (rtnl_dereference(tmp->ctarget) == rule) {
|
2007-03-27 00:14:15 +00:00
|
|
|
rcu_assign_pointer(tmp->ctarget, NULL);
|
|
|
|
ops->unresolved_rules++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2008-01-21 00:46:41 +00:00
|
|
|
notify_rule_change(RTM_DELRULE, rule, ops, nlh,
|
2006-08-15 07:32:48 +00:00
|
|
|
NETLINK_CB(skb).pid);
|
2006-08-04 10:38:38 +00:00
|
|
|
fib_rule_put(rule);
|
2007-03-27 20:56:52 +00:00
|
|
|
flush_route_cache(ops);
|
2006-08-04 10:38:38 +00:00
|
|
|
rules_ops_put(ops);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
err = -ENOENT;
|
|
|
|
errout:
|
|
|
|
rules_ops_put(ops);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2006-11-10 22:10:15 +00:00
|
|
|
static inline size_t fib_rule_nlmsg_size(struct fib_rules_ops *ops,
|
|
|
|
struct fib_rule *rule)
|
|
|
|
{
|
|
|
|
size_t payload = NLMSG_ALIGN(sizeof(struct fib_rule_hdr))
|
2009-12-03 01:25:54 +00:00
|
|
|
+ nla_total_size(IFNAMSIZ) /* FRA_IIFNAME */
|
2009-12-03 01:25:56 +00:00
|
|
|
+ nla_total_size(IFNAMSIZ) /* FRA_OIFNAME */
|
2006-11-10 22:10:15 +00:00
|
|
|
+ nla_total_size(4) /* FRA_PRIORITY */
|
|
|
|
+ nla_total_size(4) /* FRA_TABLE */
|
|
|
|
+ nla_total_size(4) /* FRA_FWMARK */
|
|
|
|
+ nla_total_size(4); /* FRA_FWMASK */
|
|
|
|
|
|
|
|
if (ops->nlmsg_payload)
|
|
|
|
payload += ops->nlmsg_payload(rule);
|
|
|
|
|
|
|
|
return payload;
|
|
|
|
}
|
|
|
|
|
2006-08-04 10:38:38 +00:00
|
|
|
static int fib_nl_fill_rule(struct sk_buff *skb, struct fib_rule *rule,
|
|
|
|
u32 pid, u32 seq, int type, int flags,
|
|
|
|
struct fib_rules_ops *ops)
|
|
|
|
{
|
|
|
|
struct nlmsghdr *nlh;
|
|
|
|
struct fib_rule_hdr *frh;
|
|
|
|
|
|
|
|
nlh = nlmsg_put(skb, pid, seq, type, sizeof(*frh), flags);
|
|
|
|
if (nlh == NULL)
|
2007-02-01 07:16:40 +00:00
|
|
|
return -EMSGSIZE;
|
2006-08-04 10:38:38 +00:00
|
|
|
|
|
|
|
frh = nlmsg_data(nlh);
|
2010-04-13 05:03:16 +00:00
|
|
|
frh->family = ops->family;
|
2006-08-04 10:38:38 +00:00
|
|
|
frh->table = rule->table;
|
2006-08-11 06:09:48 +00:00
|
|
|
NLA_PUT_U32(skb, FRA_TABLE, rule->table);
|
2006-08-04 10:38:38 +00:00
|
|
|
frh->res1 = 0;
|
|
|
|
frh->res2 = 0;
|
|
|
|
frh->action = rule->action;
|
|
|
|
frh->flags = rule->flags;
|
|
|
|
|
2010-10-26 09:24:55 +00:00
|
|
|
if (rule->action == FR_ACT_GOTO &&
|
|
|
|
rcu_dereference_raw(rule->ctarget) == NULL)
|
2007-03-27 00:14:15 +00:00
|
|
|
frh->flags |= FIB_RULE_UNRESOLVED;
|
|
|
|
|
2009-12-03 01:25:54 +00:00
|
|
|
if (rule->iifname[0]) {
|
|
|
|
NLA_PUT_STRING(skb, FRA_IIFNAME, rule->iifname);
|
2006-08-04 10:38:38 +00:00
|
|
|
|
2009-12-03 01:25:54 +00:00
|
|
|
if (rule->iifindex == -1)
|
|
|
|
frh->flags |= FIB_RULE_IIF_DETACHED;
|
2007-03-27 00:37:59 +00:00
|
|
|
}
|
|
|
|
|
2009-12-03 01:25:56 +00:00
|
|
|
if (rule->oifname[0]) {
|
|
|
|
NLA_PUT_STRING(skb, FRA_OIFNAME, rule->oifname);
|
|
|
|
|
|
|
|
if (rule->oifindex == -1)
|
|
|
|
frh->flags |= FIB_RULE_OIF_DETACHED;
|
|
|
|
}
|
|
|
|
|
2006-08-04 10:38:38 +00:00
|
|
|
if (rule->pref)
|
|
|
|
NLA_PUT_U32(skb, FRA_PRIORITY, rule->pref);
|
|
|
|
|
2006-11-09 23:22:18 +00:00
|
|
|
if (rule->mark)
|
|
|
|
NLA_PUT_U32(skb, FRA_FWMARK, rule->mark);
|
|
|
|
|
|
|
|
if (rule->mark_mask || rule->mark)
|
|
|
|
NLA_PUT_U32(skb, FRA_FWMASK, rule->mark_mask);
|
|
|
|
|
2007-03-27 00:14:15 +00:00
|
|
|
if (rule->target)
|
|
|
|
NLA_PUT_U32(skb, FRA_GOTO, rule->target);
|
|
|
|
|
2009-05-21 00:26:23 +00:00
|
|
|
if (ops->fill(rule, skb, frh) < 0)
|
2006-08-04 10:38:38 +00:00
|
|
|
goto nla_put_failure;
|
|
|
|
|
|
|
|
return nlmsg_end(skb, nlh);
|
|
|
|
|
|
|
|
nla_put_failure:
|
2007-02-01 07:16:40 +00:00
|
|
|
nlmsg_cancel(skb, nlh);
|
|
|
|
return -EMSGSIZE;
|
2006-08-04 10:38:38 +00:00
|
|
|
}
|
|
|
|
|
2007-03-26 06:24:24 +00:00
|
|
|
static int dump_rules(struct sk_buff *skb, struct netlink_callback *cb,
|
|
|
|
struct fib_rules_ops *ops)
|
2006-08-04 10:38:38 +00:00
|
|
|
{
|
|
|
|
int idx = 0;
|
|
|
|
struct fib_rule *rule;
|
|
|
|
|
2007-09-16 22:44:27 +00:00
|
|
|
list_for_each_entry(rule, &ops->rules_list, list) {
|
2007-03-26 06:24:24 +00:00
|
|
|
if (idx < cb->args[1])
|
2006-08-04 10:38:38 +00:00
|
|
|
goto skip;
|
|
|
|
|
|
|
|
if (fib_nl_fill_rule(skb, rule, NETLINK_CB(cb->skb).pid,
|
|
|
|
cb->nlh->nlmsg_seq, RTM_NEWRULE,
|
|
|
|
NLM_F_MULTI, ops) < 0)
|
|
|
|
break;
|
|
|
|
skip:
|
|
|
|
idx++;
|
|
|
|
}
|
2007-03-26 06:24:24 +00:00
|
|
|
cb->args[1] = idx;
|
2006-08-04 10:38:38 +00:00
|
|
|
rules_ops_put(ops);
|
|
|
|
|
|
|
|
return skb->len;
|
|
|
|
}
|
|
|
|
|
2007-03-26 06:24:24 +00:00
|
|
|
static int fib_nl_dumprule(struct sk_buff *skb, struct netlink_callback *cb)
|
|
|
|
{
|
2008-03-25 17:26:21 +00:00
|
|
|
struct net *net = sock_net(skb->sk);
|
2007-03-26 06:24:24 +00:00
|
|
|
struct fib_rules_ops *ops;
|
|
|
|
int idx = 0, family;
|
|
|
|
|
|
|
|
family = rtnl_msg_family(cb->nlh);
|
|
|
|
if (family != AF_UNSPEC) {
|
|
|
|
/* Protocol specific dump request */
|
2008-01-10 11:20:28 +00:00
|
|
|
ops = lookup_rules_ops(net, family);
|
2007-03-26 06:24:24 +00:00
|
|
|
if (ops == NULL)
|
|
|
|
return -EAFNOSUPPORT;
|
|
|
|
|
|
|
|
return dump_rules(skb, cb, ops);
|
|
|
|
}
|
|
|
|
|
|
|
|
rcu_read_lock();
|
2008-01-10 11:20:28 +00:00
|
|
|
list_for_each_entry_rcu(ops, &net->rules_ops, list) {
|
2007-03-26 06:24:24 +00:00
|
|
|
if (idx < cb->args[0] || !try_module_get(ops->owner))
|
|
|
|
goto skip;
|
|
|
|
|
|
|
|
if (dump_rules(skb, cb, ops) < 0)
|
|
|
|
break;
|
|
|
|
|
|
|
|
cb->args[1] = 0;
|
2010-03-09 20:03:38 +00:00
|
|
|
skip:
|
2007-03-26 06:24:24 +00:00
|
|
|
idx++;
|
|
|
|
}
|
|
|
|
rcu_read_unlock();
|
|
|
|
cb->args[0] = idx;
|
|
|
|
|
|
|
|
return skb->len;
|
|
|
|
}
|
2006-08-04 10:38:38 +00:00
|
|
|
|
2008-01-21 00:46:41 +00:00
|
|
|
static void notify_rule_change(int event, struct fib_rule *rule,
|
2006-08-15 07:32:48 +00:00
|
|
|
struct fib_rules_ops *ops, struct nlmsghdr *nlh,
|
|
|
|
u32 pid)
|
2006-08-04 10:38:38 +00:00
|
|
|
{
|
2008-01-21 00:46:41 +00:00
|
|
|
struct net *net;
|
2006-08-15 07:32:48 +00:00
|
|
|
struct sk_buff *skb;
|
|
|
|
int err = -ENOBUFS;
|
2006-08-04 10:38:38 +00:00
|
|
|
|
2008-01-21 00:46:41 +00:00
|
|
|
net = ops->fro_net;
|
2006-11-10 22:10:15 +00:00
|
|
|
skb = nlmsg_new(fib_rule_nlmsg_size(ops, rule), GFP_KERNEL);
|
2006-08-04 10:38:38 +00:00
|
|
|
if (skb == NULL)
|
2006-08-15 07:32:48 +00:00
|
|
|
goto errout;
|
|
|
|
|
|
|
|
err = fib_nl_fill_rule(skb, rule, pid, nlh->nlmsg_seq, event, 0, ops);
|
2007-02-01 07:16:40 +00:00
|
|
|
if (err < 0) {
|
|
|
|
/* -EMSGSIZE implies BUG in fib_rule_nlmsg_size() */
|
|
|
|
WARN_ON(err == -EMSGSIZE);
|
|
|
|
kfree_skb(skb);
|
|
|
|
goto errout;
|
|
|
|
}
|
2008-01-21 00:46:41 +00:00
|
|
|
|
2009-02-25 07:18:28 +00:00
|
|
|
rtnl_notify(skb, net, pid, ops->nlgroup, nlh, GFP_KERNEL);
|
|
|
|
return;
|
2006-08-15 07:32:48 +00:00
|
|
|
errout:
|
|
|
|
if (err < 0)
|
2008-01-10 11:20:28 +00:00
|
|
|
rtnl_set_sk_err(net, ops->nlgroup, err);
|
2006-08-04 10:38:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void attach_rules(struct list_head *rules, struct net_device *dev)
|
|
|
|
{
|
|
|
|
struct fib_rule *rule;
|
|
|
|
|
|
|
|
list_for_each_entry(rule, rules, list) {
|
2009-12-03 01:25:54 +00:00
|
|
|
if (rule->iifindex == -1 &&
|
|
|
|
strcmp(dev->name, rule->iifname) == 0)
|
|
|
|
rule->iifindex = dev->ifindex;
|
2009-12-03 01:25:56 +00:00
|
|
|
if (rule->oifindex == -1 &&
|
|
|
|
strcmp(dev->name, rule->oifname) == 0)
|
|
|
|
rule->oifindex = dev->ifindex;
|
2006-08-04 10:38:38 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void detach_rules(struct list_head *rules, struct net_device *dev)
|
|
|
|
{
|
|
|
|
struct fib_rule *rule;
|
|
|
|
|
2009-12-03 01:25:56 +00:00
|
|
|
list_for_each_entry(rule, rules, list) {
|
2009-12-03 01:25:54 +00:00
|
|
|
if (rule->iifindex == dev->ifindex)
|
|
|
|
rule->iifindex = -1;
|
2009-12-03 01:25:56 +00:00
|
|
|
if (rule->oifindex == dev->ifindex)
|
|
|
|
rule->oifindex = -1;
|
|
|
|
}
|
2006-08-04 10:38:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static int fib_rules_event(struct notifier_block *this, unsigned long event,
|
|
|
|
void *ptr)
|
|
|
|
{
|
|
|
|
struct net_device *dev = ptr;
|
2008-03-25 12:47:49 +00:00
|
|
|
struct net *net = dev_net(dev);
|
2006-08-04 10:38:38 +00:00
|
|
|
struct fib_rules_ops *ops;
|
|
|
|
|
|
|
|
ASSERT_RTNL();
|
|
|
|
|
|
|
|
switch (event) {
|
|
|
|
case NETDEV_REGISTER:
|
2008-01-10 11:20:28 +00:00
|
|
|
list_for_each_entry(ops, &net->rules_ops, list)
|
2007-09-16 22:44:27 +00:00
|
|
|
attach_rules(&ops->rules_list, dev);
|
2006-08-04 10:38:38 +00:00
|
|
|
break;
|
|
|
|
|
|
|
|
case NETDEV_UNREGISTER:
|
2008-01-10 11:20:28 +00:00
|
|
|
list_for_each_entry(ops, &net->rules_ops, list)
|
2007-09-16 22:44:27 +00:00
|
|
|
detach_rules(&ops->rules_list, dev);
|
2006-08-04 10:38:38 +00:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
return NOTIFY_DONE;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct notifier_block fib_rules_notifier = {
|
|
|
|
.notifier_call = fib_rules_event,
|
|
|
|
};
|
|
|
|
|
2010-01-17 03:35:32 +00:00
|
|
|
static int __net_init fib_rules_net_init(struct net *net)
|
2008-01-10 11:20:28 +00:00
|
|
|
{
|
|
|
|
INIT_LIST_HEAD(&net->rules_ops);
|
|
|
|
spin_lock_init(&net->rules_mod_lock);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct pernet_operations fib_rules_net_ops = {
|
|
|
|
.init = fib_rules_net_init,
|
|
|
|
};
|
|
|
|
|
2006-08-04 10:38:38 +00:00
|
|
|
static int __init fib_rules_init(void)
|
|
|
|
{
|
2008-01-10 11:20:28 +00:00
|
|
|
int err;
|
2007-03-26 06:20:05 +00:00
|
|
|
rtnl_register(PF_UNSPEC, RTM_NEWRULE, fib_nl_newrule, NULL);
|
|
|
|
rtnl_register(PF_UNSPEC, RTM_DELRULE, fib_nl_delrule, NULL);
|
2007-03-26 06:24:24 +00:00
|
|
|
rtnl_register(PF_UNSPEC, RTM_GETRULE, NULL, fib_nl_dumprule);
|
2007-03-26 06:20:05 +00:00
|
|
|
|
2008-11-08 06:52:34 +00:00
|
|
|
err = register_pernet_subsys(&fib_rules_net_ops);
|
2008-01-10 11:20:28 +00:00
|
|
|
if (err < 0)
|
|
|
|
goto fail;
|
|
|
|
|
2008-11-08 06:52:34 +00:00
|
|
|
err = register_netdevice_notifier(&fib_rules_notifier);
|
2008-01-10 11:20:28 +00:00
|
|
|
if (err < 0)
|
|
|
|
goto fail_unregister;
|
2008-11-08 06:52:34 +00:00
|
|
|
|
2008-01-10 11:20:28 +00:00
|
|
|
return 0;
|
|
|
|
|
|
|
|
fail_unregister:
|
2008-11-08 06:52:34 +00:00
|
|
|
unregister_pernet_subsys(&fib_rules_net_ops);
|
2008-01-10 11:20:28 +00:00
|
|
|
fail:
|
|
|
|
rtnl_unregister(PF_UNSPEC, RTM_NEWRULE);
|
|
|
|
rtnl_unregister(PF_UNSPEC, RTM_DELRULE);
|
|
|
|
rtnl_unregister(PF_UNSPEC, RTM_GETRULE);
|
|
|
|
return err;
|
2006-08-04 10:38:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
subsys_initcall(fib_rules_init);
|