]>
Commit | Line | Data |
---|---|---|
1da177e4 LT |
1 | /* |
2 | * INET An implementation of the TCP/IP protocol suite for the LINUX | |
3 | * operating system. INET is implemented using the BSD Socket | |
4 | * interface as the means of communication with the user level. | |
5 | * | |
6 | * IPv4 Forwarding Information Base: semantics. | |
7 | * | |
8 | * Version: $Id: fib_semantics.c,v 1.19 2002/01/12 07:54:56 davem Exp $ | |
9 | * | |
10 | * Authors: Alexey Kuznetsov, <[email protected]> | |
11 | * | |
12 | * This program is free software; you can redistribute it and/or | |
13 | * modify it under the terms of the GNU General Public License | |
14 | * as published by the Free Software Foundation; either version | |
15 | * 2 of the License, or (at your option) any later version. | |
16 | */ | |
17 | ||
18 | #include <linux/config.h> | |
19 | #include <asm/uaccess.h> | |
20 | #include <asm/system.h> | |
21 | #include <linux/bitops.h> | |
22 | #include <linux/types.h> | |
23 | #include <linux/kernel.h> | |
24 | #include <linux/jiffies.h> | |
25 | #include <linux/mm.h> | |
26 | #include <linux/string.h> | |
27 | #include <linux/socket.h> | |
28 | #include <linux/sockios.h> | |
29 | #include <linux/errno.h> | |
30 | #include <linux/in.h> | |
31 | #include <linux/inet.h> | |
14c85021 | 32 | #include <linux/inetdevice.h> |
1da177e4 LT |
33 | #include <linux/netdevice.h> |
34 | #include <linux/if_arp.h> | |
35 | #include <linux/proc_fs.h> | |
36 | #include <linux/skbuff.h> | |
37 | #include <linux/netlink.h> | |
38 | #include <linux/init.h> | |
39 | ||
14c85021 | 40 | #include <net/arp.h> |
1da177e4 LT |
41 | #include <net/ip.h> |
42 | #include <net/protocol.h> | |
43 | #include <net/route.h> | |
44 | #include <net/tcp.h> | |
45 | #include <net/sock.h> | |
46 | #include <net/ip_fib.h> | |
47 | #include <net/ip_mp_alg.h> | |
48 | ||
49 | #include "fib_lookup.h" | |
50 | ||
51 | #define FSprintk(a...) | |
52 | ||
53 | static DEFINE_RWLOCK(fib_info_lock); | |
54 | static struct hlist_head *fib_info_hash; | |
55 | static struct hlist_head *fib_info_laddrhash; | |
56 | static unsigned int fib_hash_size; | |
57 | static unsigned int fib_info_cnt; | |
58 | ||
59 | #define DEVINDEX_HASHBITS 8 | |
60 | #define DEVINDEX_HASHSIZE (1U << DEVINDEX_HASHBITS) | |
61 | static struct hlist_head fib_info_devhash[DEVINDEX_HASHSIZE]; | |
62 | ||
63 | #ifdef CONFIG_IP_ROUTE_MULTIPATH | |
64 | ||
65 | static DEFINE_SPINLOCK(fib_multipath_lock); | |
66 | ||
67 | #define for_nexthops(fi) { int nhsel; const struct fib_nh * nh; \ | |
68 | for (nhsel=0, nh = (fi)->fib_nh; nhsel < (fi)->fib_nhs; nh++, nhsel++) | |
69 | ||
70 | #define change_nexthops(fi) { int nhsel; struct fib_nh * nh; \ | |
71 | for (nhsel=0, nh = (struct fib_nh*)((fi)->fib_nh); nhsel < (fi)->fib_nhs; nh++, nhsel++) | |
72 | ||
73 | #else /* CONFIG_IP_ROUTE_MULTIPATH */ | |
74 | ||
75 | /* Hope, that gcc will optimize it to get rid of dummy loop */ | |
76 | ||
77 | #define for_nexthops(fi) { int nhsel=0; const struct fib_nh * nh = (fi)->fib_nh; \ | |
78 | for (nhsel=0; nhsel < 1; nhsel++) | |
79 | ||
80 | #define change_nexthops(fi) { int nhsel=0; struct fib_nh * nh = (struct fib_nh*)((fi)->fib_nh); \ | |
81 | for (nhsel=0; nhsel < 1; nhsel++) | |
82 | ||
83 | #endif /* CONFIG_IP_ROUTE_MULTIPATH */ | |
84 | ||
85 | #define endfor_nexthops(fi) } | |
86 | ||
87 | ||
9b5b5cff | 88 | static const struct |
1da177e4 LT |
89 | { |
90 | int error; | |
91 | u8 scope; | |
92 | } fib_props[RTA_MAX + 1] = { | |
93 | { | |
94 | .error = 0, | |
95 | .scope = RT_SCOPE_NOWHERE, | |
96 | }, /* RTN_UNSPEC */ | |
97 | { | |
98 | .error = 0, | |
99 | .scope = RT_SCOPE_UNIVERSE, | |
100 | }, /* RTN_UNICAST */ | |
101 | { | |
102 | .error = 0, | |
103 | .scope = RT_SCOPE_HOST, | |
104 | }, /* RTN_LOCAL */ | |
105 | { | |
106 | .error = 0, | |
107 | .scope = RT_SCOPE_LINK, | |
108 | }, /* RTN_BROADCAST */ | |
109 | { | |
110 | .error = 0, | |
111 | .scope = RT_SCOPE_LINK, | |
112 | }, /* RTN_ANYCAST */ | |
113 | { | |
114 | .error = 0, | |
115 | .scope = RT_SCOPE_UNIVERSE, | |
116 | }, /* RTN_MULTICAST */ | |
117 | { | |
118 | .error = -EINVAL, | |
119 | .scope = RT_SCOPE_UNIVERSE, | |
120 | }, /* RTN_BLACKHOLE */ | |
121 | { | |
122 | .error = -EHOSTUNREACH, | |
123 | .scope = RT_SCOPE_UNIVERSE, | |
124 | }, /* RTN_UNREACHABLE */ | |
125 | { | |
126 | .error = -EACCES, | |
127 | .scope = RT_SCOPE_UNIVERSE, | |
128 | }, /* RTN_PROHIBIT */ | |
129 | { | |
130 | .error = -EAGAIN, | |
131 | .scope = RT_SCOPE_UNIVERSE, | |
132 | }, /* RTN_THROW */ | |
133 | { | |
134 | .error = -EINVAL, | |
135 | .scope = RT_SCOPE_NOWHERE, | |
136 | }, /* RTN_NAT */ | |
137 | { | |
138 | .error = -EINVAL, | |
139 | .scope = RT_SCOPE_NOWHERE, | |
140 | }, /* RTN_XRESOLVE */ | |
141 | }; | |
142 | ||
143 | ||
144 | /* Release a nexthop info record */ | |
145 | ||
146 | void free_fib_info(struct fib_info *fi) | |
147 | { | |
148 | if (fi->fib_dead == 0) { | |
149 | printk("Freeing alive fib_info %p\n", fi); | |
150 | return; | |
151 | } | |
152 | change_nexthops(fi) { | |
153 | if (nh->nh_dev) | |
154 | dev_put(nh->nh_dev); | |
155 | nh->nh_dev = NULL; | |
156 | } endfor_nexthops(fi); | |
157 | fib_info_cnt--; | |
158 | kfree(fi); | |
159 | } | |
160 | ||
161 | void fib_release_info(struct fib_info *fi) | |
162 | { | |
163 | write_lock(&fib_info_lock); | |
164 | if (fi && --fi->fib_treeref == 0) { | |
165 | hlist_del(&fi->fib_hash); | |
166 | if (fi->fib_prefsrc) | |
167 | hlist_del(&fi->fib_lhash); | |
168 | change_nexthops(fi) { | |
169 | if (!nh->nh_dev) | |
170 | continue; | |
171 | hlist_del(&nh->nh_hash); | |
172 | } endfor_nexthops(fi) | |
173 | fi->fib_dead = 1; | |
174 | fib_info_put(fi); | |
175 | } | |
176 | write_unlock(&fib_info_lock); | |
177 | } | |
178 | ||
179 | static __inline__ int nh_comp(const struct fib_info *fi, const struct fib_info *ofi) | |
180 | { | |
181 | const struct fib_nh *onh = ofi->fib_nh; | |
182 | ||
183 | for_nexthops(fi) { | |
184 | if (nh->nh_oif != onh->nh_oif || | |
185 | nh->nh_gw != onh->nh_gw || | |
186 | nh->nh_scope != onh->nh_scope || | |
187 | #ifdef CONFIG_IP_ROUTE_MULTIPATH | |
188 | nh->nh_weight != onh->nh_weight || | |
189 | #endif | |
190 | #ifdef CONFIG_NET_CLS_ROUTE | |
191 | nh->nh_tclassid != onh->nh_tclassid || | |
192 | #endif | |
193 | ((nh->nh_flags^onh->nh_flags)&~RTNH_F_DEAD)) | |
194 | return -1; | |
195 | onh++; | |
196 | } endfor_nexthops(fi); | |
197 | return 0; | |
198 | } | |
199 | ||
200 | static inline unsigned int fib_info_hashfn(const struct fib_info *fi) | |
201 | { | |
202 | unsigned int mask = (fib_hash_size - 1); | |
203 | unsigned int val = fi->fib_nhs; | |
204 | ||
205 | val ^= fi->fib_protocol; | |
206 | val ^= fi->fib_prefsrc; | |
207 | val ^= fi->fib_priority; | |
208 | ||
209 | return (val ^ (val >> 7) ^ (val >> 12)) & mask; | |
210 | } | |
211 | ||
212 | static struct fib_info *fib_find_info(const struct fib_info *nfi) | |
213 | { | |
214 | struct hlist_head *head; | |
215 | struct hlist_node *node; | |
216 | struct fib_info *fi; | |
217 | unsigned int hash; | |
218 | ||
219 | hash = fib_info_hashfn(nfi); | |
220 | head = &fib_info_hash[hash]; | |
221 | ||
222 | hlist_for_each_entry(fi, node, head, fib_hash) { | |
223 | if (fi->fib_nhs != nfi->fib_nhs) | |
224 | continue; | |
225 | if (nfi->fib_protocol == fi->fib_protocol && | |
226 | nfi->fib_prefsrc == fi->fib_prefsrc && | |
227 | nfi->fib_priority == fi->fib_priority && | |
228 | memcmp(nfi->fib_metrics, fi->fib_metrics, | |
229 | sizeof(fi->fib_metrics)) == 0 && | |
230 | ((nfi->fib_flags^fi->fib_flags)&~RTNH_F_DEAD) == 0 && | |
231 | (nfi->fib_nhs == 0 || nh_comp(fi, nfi) == 0)) | |
232 | return fi; | |
233 | } | |
234 | ||
235 | return NULL; | |
236 | } | |
237 | ||
238 | static inline unsigned int fib_devindex_hashfn(unsigned int val) | |
239 | { | |
240 | unsigned int mask = DEVINDEX_HASHSIZE - 1; | |
241 | ||
242 | return (val ^ | |
243 | (val >> DEVINDEX_HASHBITS) ^ | |
244 | (val >> (DEVINDEX_HASHBITS * 2))) & mask; | |
245 | } | |
246 | ||
247 | /* Check, that the gateway is already configured. | |
248 | Used only by redirect accept routine. | |
249 | */ | |
250 | ||
251 | int ip_fib_check_default(u32 gw, struct net_device *dev) | |
252 | { | |
253 | struct hlist_head *head; | |
254 | struct hlist_node *node; | |
255 | struct fib_nh *nh; | |
256 | unsigned int hash; | |
257 | ||
258 | read_lock(&fib_info_lock); | |
259 | ||
260 | hash = fib_devindex_hashfn(dev->ifindex); | |
261 | head = &fib_info_devhash[hash]; | |
262 | hlist_for_each_entry(nh, node, head, nh_hash) { | |
263 | if (nh->nh_dev == dev && | |
264 | nh->nh_gw == gw && | |
265 | !(nh->nh_flags&RTNH_F_DEAD)) { | |
266 | read_unlock(&fib_info_lock); | |
267 | return 0; | |
268 | } | |
269 | } | |
270 | ||
271 | read_unlock(&fib_info_lock); | |
272 | ||
273 | return -1; | |
274 | } | |
275 | ||
276 | void rtmsg_fib(int event, u32 key, struct fib_alias *fa, | |
277 | int z, int tb_id, | |
278 | struct nlmsghdr *n, struct netlink_skb_parms *req) | |
279 | { | |
280 | struct sk_buff *skb; | |
9ed19f33 | 281 | u32 pid = req ? req->pid : n->nlmsg_pid; |
1da177e4 LT |
282 | int size = NLMSG_SPACE(sizeof(struct rtmsg)+256); |
283 | ||
284 | skb = alloc_skb(size, GFP_KERNEL); | |
285 | if (!skb) | |
286 | return; | |
287 | ||
288 | if (fib_dump_info(skb, pid, n->nlmsg_seq, event, tb_id, | |
289 | fa->fa_type, fa->fa_scope, &key, z, | |
290 | fa->fa_tos, | |
b6544c0b | 291 | fa->fa_info, 0) < 0) { |
1da177e4 LT |
292 | kfree_skb(skb); |
293 | return; | |
294 | } | |
ac6d439d | 295 | NETLINK_CB(skb).dst_group = RTNLGRP_IPV4_ROUTE; |
1da177e4 LT |
296 | if (n->nlmsg_flags&NLM_F_ECHO) |
297 | atomic_inc(&skb->users); | |
ac6d439d | 298 | netlink_broadcast(rtnl, skb, pid, RTNLGRP_IPV4_ROUTE, GFP_KERNEL); |
1da177e4 LT |
299 | if (n->nlmsg_flags&NLM_F_ECHO) |
300 | netlink_unicast(rtnl, skb, pid, MSG_DONTWAIT); | |
301 | } | |
302 | ||
303 | /* Return the first fib alias matching TOS with | |
304 | * priority less than or equal to PRIO. | |
305 | */ | |
306 | struct fib_alias *fib_find_alias(struct list_head *fah, u8 tos, u32 prio) | |
307 | { | |
308 | if (fah) { | |
309 | struct fib_alias *fa; | |
310 | list_for_each_entry(fa, fah, fa_list) { | |
311 | if (fa->fa_tos > tos) | |
312 | continue; | |
313 | if (fa->fa_info->fib_priority >= prio || | |
314 | fa->fa_tos < tos) | |
315 | return fa; | |
316 | } | |
317 | } | |
318 | return NULL; | |
319 | } | |
320 | ||
321 | int fib_detect_death(struct fib_info *fi, int order, | |
322 | struct fib_info **last_resort, int *last_idx, int *dflt) | |
323 | { | |
324 | struct neighbour *n; | |
325 | int state = NUD_NONE; | |
326 | ||
327 | n = neigh_lookup(&arp_tbl, &fi->fib_nh[0].nh_gw, fi->fib_dev); | |
328 | if (n) { | |
329 | state = n->nud_state; | |
330 | neigh_release(n); | |
331 | } | |
332 | if (state==NUD_REACHABLE) | |
333 | return 0; | |
334 | if ((state&NUD_VALID) && order != *dflt) | |
335 | return 0; | |
336 | if ((state&NUD_VALID) || | |
337 | (*last_idx<0 && order > *dflt)) { | |
338 | *last_resort = fi; | |
339 | *last_idx = order; | |
340 | } | |
341 | return 1; | |
342 | } | |
343 | ||
344 | #ifdef CONFIG_IP_ROUTE_MULTIPATH | |
345 | ||
346 | static u32 fib_get_attr32(struct rtattr *attr, int attrlen, int type) | |
347 | { | |
348 | while (RTA_OK(attr,attrlen)) { | |
349 | if (attr->rta_type == type) | |
350 | return *(u32*)RTA_DATA(attr); | |
351 | attr = RTA_NEXT(attr, attrlen); | |
352 | } | |
353 | return 0; | |
354 | } | |
355 | ||
356 | static int | |
357 | fib_count_nexthops(struct rtattr *rta) | |
358 | { | |
359 | int nhs = 0; | |
360 | struct rtnexthop *nhp = RTA_DATA(rta); | |
361 | int nhlen = RTA_PAYLOAD(rta); | |
362 | ||
363 | while (nhlen >= (int)sizeof(struct rtnexthop)) { | |
364 | if ((nhlen -= nhp->rtnh_len) < 0) | |
365 | return 0; | |
366 | nhs++; | |
367 | nhp = RTNH_NEXT(nhp); | |
368 | }; | |
369 | return nhs; | |
370 | } | |
371 | ||
372 | static int | |
373 | fib_get_nhs(struct fib_info *fi, const struct rtattr *rta, const struct rtmsg *r) | |
374 | { | |
375 | struct rtnexthop *nhp = RTA_DATA(rta); | |
376 | int nhlen = RTA_PAYLOAD(rta); | |
377 | ||
378 | change_nexthops(fi) { | |
379 | int attrlen = nhlen - sizeof(struct rtnexthop); | |
380 | if (attrlen < 0 || (nhlen -= nhp->rtnh_len) < 0) | |
381 | return -EINVAL; | |
382 | nh->nh_flags = (r->rtm_flags&~0xFF) | nhp->rtnh_flags; | |
383 | nh->nh_oif = nhp->rtnh_ifindex; | |
384 | nh->nh_weight = nhp->rtnh_hops + 1; | |
385 | if (attrlen) { | |
386 | nh->nh_gw = fib_get_attr32(RTNH_DATA(nhp), attrlen, RTA_GATEWAY); | |
387 | #ifdef CONFIG_NET_CLS_ROUTE | |
388 | nh->nh_tclassid = fib_get_attr32(RTNH_DATA(nhp), attrlen, RTA_FLOW); | |
389 | #endif | |
390 | } | |
391 | nhp = RTNH_NEXT(nhp); | |
392 | } endfor_nexthops(fi); | |
393 | return 0; | |
394 | } | |
395 | ||
396 | #endif | |
397 | ||
398 | int fib_nh_match(struct rtmsg *r, struct nlmsghdr *nlh, struct kern_rta *rta, | |
399 | struct fib_info *fi) | |
400 | { | |
401 | #ifdef CONFIG_IP_ROUTE_MULTIPATH | |
402 | struct rtnexthop *nhp; | |
403 | int nhlen; | |
404 | #endif | |
405 | ||
406 | if (rta->rta_priority && | |
407 | *rta->rta_priority != fi->fib_priority) | |
408 | return 1; | |
409 | ||
410 | if (rta->rta_oif || rta->rta_gw) { | |
411 | if ((!rta->rta_oif || *rta->rta_oif == fi->fib_nh->nh_oif) && | |
412 | (!rta->rta_gw || memcmp(rta->rta_gw, &fi->fib_nh->nh_gw, 4) == 0)) | |
413 | return 0; | |
414 | return 1; | |
415 | } | |
416 | ||
417 | #ifdef CONFIG_IP_ROUTE_MULTIPATH | |
418 | if (rta->rta_mp == NULL) | |
419 | return 0; | |
420 | nhp = RTA_DATA(rta->rta_mp); | |
421 | nhlen = RTA_PAYLOAD(rta->rta_mp); | |
422 | ||
423 | for_nexthops(fi) { | |
424 | int attrlen = nhlen - sizeof(struct rtnexthop); | |
425 | u32 gw; | |
426 | ||
427 | if (attrlen < 0 || (nhlen -= nhp->rtnh_len) < 0) | |
428 | return -EINVAL; | |
429 | if (nhp->rtnh_ifindex && nhp->rtnh_ifindex != nh->nh_oif) | |
430 | return 1; | |
431 | if (attrlen) { | |
432 | gw = fib_get_attr32(RTNH_DATA(nhp), attrlen, RTA_GATEWAY); | |
433 | if (gw && gw != nh->nh_gw) | |
434 | return 1; | |
435 | #ifdef CONFIG_NET_CLS_ROUTE | |
436 | gw = fib_get_attr32(RTNH_DATA(nhp), attrlen, RTA_FLOW); | |
437 | if (gw && gw != nh->nh_tclassid) | |
438 | return 1; | |
439 | #endif | |
440 | } | |
441 | nhp = RTNH_NEXT(nhp); | |
442 | } endfor_nexthops(fi); | |
443 | #endif | |
444 | return 0; | |
445 | } | |
446 | ||
447 | ||
448 | /* | |
449 | Picture | |
450 | ------- | |
451 | ||
452 | Semantics of nexthop is very messy by historical reasons. | |
453 | We have to take into account, that: | |
454 | a) gateway can be actually local interface address, | |
455 | so that gatewayed route is direct. | |
456 | b) gateway must be on-link address, possibly | |
457 | described not by an ifaddr, but also by a direct route. | |
458 | c) If both gateway and interface are specified, they should not | |
459 | contradict. | |
460 | d) If we use tunnel routes, gateway could be not on-link. | |
461 | ||
462 | Attempt to reconcile all of these (alas, self-contradictory) conditions | |
463 | results in pretty ugly and hairy code with obscure logic. | |
464 | ||
465 | I chose to generalized it instead, so that the size | |
466 | of code does not increase practically, but it becomes | |
467 | much more general. | |
468 | Every prefix is assigned a "scope" value: "host" is local address, | |
469 | "link" is direct route, | |
470 | [ ... "site" ... "interior" ... ] | |
471 | and "universe" is true gateway route with global meaning. | |
472 | ||
473 | Every prefix refers to a set of "nexthop"s (gw, oif), | |
474 | where gw must have narrower scope. This recursion stops | |
475 | when gw has LOCAL scope or if "nexthop" is declared ONLINK, | |
476 | which means that gw is forced to be on link. | |
477 | ||
478 | Code is still hairy, but now it is apparently logically | |
479 | consistent and very flexible. F.e. as by-product it allows | |
480 | to co-exists in peace independent exterior and interior | |
481 | routing processes. | |
482 | ||
483 | Normally it looks as following. | |
484 | ||
485 | {universe prefix} -> (gw, oif) [scope link] | |
486 | | | |
487 | |-> {link prefix} -> (gw, oif) [scope local] | |
488 | | | |
489 | |-> {local prefix} (terminal node) | |
490 | */ | |
491 | ||
492 | static int fib_check_nh(const struct rtmsg *r, struct fib_info *fi, struct fib_nh *nh) | |
493 | { | |
494 | int err; | |
495 | ||
496 | if (nh->nh_gw) { | |
497 | struct fib_result res; | |
498 | ||
499 | #ifdef CONFIG_IP_ROUTE_PERVASIVE | |
500 | if (nh->nh_flags&RTNH_F_PERVASIVE) | |
501 | return 0; | |
502 | #endif | |
503 | if (nh->nh_flags&RTNH_F_ONLINK) { | |
504 | struct net_device *dev; | |
505 | ||
506 | if (r->rtm_scope >= RT_SCOPE_LINK) | |
507 | return -EINVAL; | |
508 | if (inet_addr_type(nh->nh_gw) != RTN_UNICAST) | |
509 | return -EINVAL; | |
510 | if ((dev = __dev_get_by_index(nh->nh_oif)) == NULL) | |
511 | return -ENODEV; | |
512 | if (!(dev->flags&IFF_UP)) | |
513 | return -ENETDOWN; | |
514 | nh->nh_dev = dev; | |
515 | dev_hold(dev); | |
516 | nh->nh_scope = RT_SCOPE_LINK; | |
517 | return 0; | |
518 | } | |
519 | { | |
520 | struct flowi fl = { .nl_u = { .ip4_u = | |
521 | { .daddr = nh->nh_gw, | |
522 | .scope = r->rtm_scope + 1 } }, | |
523 | .oif = nh->nh_oif }; | |
524 | ||
525 | /* It is not necessary, but requires a bit of thinking */ | |
526 | if (fl.fl4_scope < RT_SCOPE_LINK) | |
527 | fl.fl4_scope = RT_SCOPE_LINK; | |
528 | if ((err = fib_lookup(&fl, &res)) != 0) | |
529 | return err; | |
530 | } | |
531 | err = -EINVAL; | |
532 | if (res.type != RTN_UNICAST && res.type != RTN_LOCAL) | |
533 | goto out; | |
534 | nh->nh_scope = res.scope; | |
535 | nh->nh_oif = FIB_RES_OIF(res); | |
536 | if ((nh->nh_dev = FIB_RES_DEV(res)) == NULL) | |
537 | goto out; | |
538 | dev_hold(nh->nh_dev); | |
539 | err = -ENETDOWN; | |
540 | if (!(nh->nh_dev->flags & IFF_UP)) | |
541 | goto out; | |
542 | err = 0; | |
543 | out: | |
544 | fib_res_put(&res); | |
545 | return err; | |
546 | } else { | |
547 | struct in_device *in_dev; | |
548 | ||
549 | if (nh->nh_flags&(RTNH_F_PERVASIVE|RTNH_F_ONLINK)) | |
550 | return -EINVAL; | |
551 | ||
552 | in_dev = inetdev_by_index(nh->nh_oif); | |
553 | if (in_dev == NULL) | |
554 | return -ENODEV; | |
555 | if (!(in_dev->dev->flags&IFF_UP)) { | |
556 | in_dev_put(in_dev); | |
557 | return -ENETDOWN; | |
558 | } | |
559 | nh->nh_dev = in_dev->dev; | |
560 | dev_hold(nh->nh_dev); | |
561 | nh->nh_scope = RT_SCOPE_HOST; | |
562 | in_dev_put(in_dev); | |
563 | } | |
564 | return 0; | |
565 | } | |
566 | ||
567 | static inline unsigned int fib_laddr_hashfn(u32 val) | |
568 | { | |
569 | unsigned int mask = (fib_hash_size - 1); | |
570 | ||
571 | return (val ^ (val >> 7) ^ (val >> 14)) & mask; | |
572 | } | |
573 | ||
574 | static struct hlist_head *fib_hash_alloc(int bytes) | |
575 | { | |
576 | if (bytes <= PAGE_SIZE) | |
577 | return kmalloc(bytes, GFP_KERNEL); | |
578 | else | |
579 | return (struct hlist_head *) | |
580 | __get_free_pages(GFP_KERNEL, get_order(bytes)); | |
581 | } | |
582 | ||
583 | static void fib_hash_free(struct hlist_head *hash, int bytes) | |
584 | { | |
585 | if (!hash) | |
586 | return; | |
587 | ||
588 | if (bytes <= PAGE_SIZE) | |
589 | kfree(hash); | |
590 | else | |
591 | free_pages((unsigned long) hash, get_order(bytes)); | |
592 | } | |
593 | ||
594 | static void fib_hash_move(struct hlist_head *new_info_hash, | |
595 | struct hlist_head *new_laddrhash, | |
596 | unsigned int new_size) | |
597 | { | |
b7656e7f | 598 | struct hlist_head *old_info_hash, *old_laddrhash; |
1da177e4 | 599 | unsigned int old_size = fib_hash_size; |
b7656e7f | 600 | unsigned int i, bytes; |
1da177e4 LT |
601 | |
602 | write_lock(&fib_info_lock); | |
b7656e7f DM |
603 | old_info_hash = fib_info_hash; |
604 | old_laddrhash = fib_info_laddrhash; | |
1da177e4 LT |
605 | fib_hash_size = new_size; |
606 | ||
607 | for (i = 0; i < old_size; i++) { | |
608 | struct hlist_head *head = &fib_info_hash[i]; | |
609 | struct hlist_node *node, *n; | |
610 | struct fib_info *fi; | |
611 | ||
612 | hlist_for_each_entry_safe(fi, node, n, head, fib_hash) { | |
613 | struct hlist_head *dest; | |
614 | unsigned int new_hash; | |
615 | ||
616 | hlist_del(&fi->fib_hash); | |
617 | ||
618 | new_hash = fib_info_hashfn(fi); | |
619 | dest = &new_info_hash[new_hash]; | |
620 | hlist_add_head(&fi->fib_hash, dest); | |
621 | } | |
622 | } | |
623 | fib_info_hash = new_info_hash; | |
624 | ||
625 | for (i = 0; i < old_size; i++) { | |
626 | struct hlist_head *lhead = &fib_info_laddrhash[i]; | |
627 | struct hlist_node *node, *n; | |
628 | struct fib_info *fi; | |
629 | ||
630 | hlist_for_each_entry_safe(fi, node, n, lhead, fib_lhash) { | |
631 | struct hlist_head *ldest; | |
632 | unsigned int new_hash; | |
633 | ||
634 | hlist_del(&fi->fib_lhash); | |
635 | ||
636 | new_hash = fib_laddr_hashfn(fi->fib_prefsrc); | |
637 | ldest = &new_laddrhash[new_hash]; | |
638 | hlist_add_head(&fi->fib_lhash, ldest); | |
639 | } | |
640 | } | |
641 | fib_info_laddrhash = new_laddrhash; | |
642 | ||
643 | write_unlock(&fib_info_lock); | |
b7656e7f DM |
644 | |
645 | bytes = old_size * sizeof(struct hlist_head *); | |
646 | fib_hash_free(old_info_hash, bytes); | |
647 | fib_hash_free(old_laddrhash, bytes); | |
1da177e4 LT |
648 | } |
649 | ||
650 | struct fib_info * | |
651 | fib_create_info(const struct rtmsg *r, struct kern_rta *rta, | |
652 | const struct nlmsghdr *nlh, int *errp) | |
653 | { | |
654 | int err; | |
655 | struct fib_info *fi = NULL; | |
656 | struct fib_info *ofi; | |
657 | #ifdef CONFIG_IP_ROUTE_MULTIPATH | |
658 | int nhs = 1; | |
659 | #else | |
660 | const int nhs = 1; | |
661 | #endif | |
662 | #ifdef CONFIG_IP_ROUTE_MULTIPATH_CACHED | |
663 | u32 mp_alg = IP_MP_ALG_NONE; | |
664 | #endif | |
665 | ||
666 | /* Fast check to catch the most weird cases */ | |
667 | if (fib_props[r->rtm_type].scope > r->rtm_scope) | |
668 | goto err_inval; | |
669 | ||
670 | #ifdef CONFIG_IP_ROUTE_MULTIPATH | |
671 | if (rta->rta_mp) { | |
672 | nhs = fib_count_nexthops(rta->rta_mp); | |
673 | if (nhs == 0) | |
674 | goto err_inval; | |
675 | } | |
676 | #endif | |
677 | #ifdef CONFIG_IP_ROUTE_MULTIPATH_CACHED | |
678 | if (rta->rta_mp_alg) { | |
679 | mp_alg = *rta->rta_mp_alg; | |
680 | ||
681 | if (mp_alg < IP_MP_ALG_NONE || | |
682 | mp_alg > IP_MP_ALG_MAX) | |
683 | goto err_inval; | |
684 | } | |
685 | #endif | |
686 | ||
687 | err = -ENOBUFS; | |
688 | if (fib_info_cnt >= fib_hash_size) { | |
689 | unsigned int new_size = fib_hash_size << 1; | |
690 | struct hlist_head *new_info_hash; | |
691 | struct hlist_head *new_laddrhash; | |
692 | unsigned int bytes; | |
693 | ||
694 | if (!new_size) | |
695 | new_size = 1; | |
696 | bytes = new_size * sizeof(struct hlist_head *); | |
697 | new_info_hash = fib_hash_alloc(bytes); | |
698 | new_laddrhash = fib_hash_alloc(bytes); | |
699 | if (!new_info_hash || !new_laddrhash) { | |
700 | fib_hash_free(new_info_hash, bytes); | |
701 | fib_hash_free(new_laddrhash, bytes); | |
702 | } else { | |
703 | memset(new_info_hash, 0, bytes); | |
704 | memset(new_laddrhash, 0, bytes); | |
705 | ||
706 | fib_hash_move(new_info_hash, new_laddrhash, new_size); | |
707 | } | |
708 | ||
709 | if (!fib_hash_size) | |
710 | goto failure; | |
711 | } | |
712 | ||
713 | fi = kmalloc(sizeof(*fi)+nhs*sizeof(struct fib_nh), GFP_KERNEL); | |
714 | if (fi == NULL) | |
715 | goto failure; | |
716 | fib_info_cnt++; | |
717 | memset(fi, 0, sizeof(*fi)+nhs*sizeof(struct fib_nh)); | |
718 | ||
719 | fi->fib_protocol = r->rtm_protocol; | |
720 | ||
721 | fi->fib_nhs = nhs; | |
722 | change_nexthops(fi) { | |
723 | nh->nh_parent = fi; | |
724 | } endfor_nexthops(fi) | |
725 | ||
726 | fi->fib_flags = r->rtm_flags; | |
727 | if (rta->rta_priority) | |
728 | fi->fib_priority = *rta->rta_priority; | |
729 | if (rta->rta_mx) { | |
730 | int attrlen = RTA_PAYLOAD(rta->rta_mx); | |
731 | struct rtattr *attr = RTA_DATA(rta->rta_mx); | |
732 | ||
733 | while (RTA_OK(attr, attrlen)) { | |
734 | unsigned flavor = attr->rta_type; | |
735 | if (flavor) { | |
736 | if (flavor > RTAX_MAX) | |
737 | goto err_inval; | |
738 | fi->fib_metrics[flavor-1] = *(unsigned*)RTA_DATA(attr); | |
739 | } | |
740 | attr = RTA_NEXT(attr, attrlen); | |
741 | } | |
742 | } | |
743 | if (rta->rta_prefsrc) | |
744 | memcpy(&fi->fib_prefsrc, rta->rta_prefsrc, 4); | |
745 | ||
746 | if (rta->rta_mp) { | |
747 | #ifdef CONFIG_IP_ROUTE_MULTIPATH | |
748 | if ((err = fib_get_nhs(fi, rta->rta_mp, r)) != 0) | |
749 | goto failure; | |
750 | if (rta->rta_oif && fi->fib_nh->nh_oif != *rta->rta_oif) | |
751 | goto err_inval; | |
752 | if (rta->rta_gw && memcmp(&fi->fib_nh->nh_gw, rta->rta_gw, 4)) | |
753 | goto err_inval; | |
754 | #ifdef CONFIG_NET_CLS_ROUTE | |
755 | if (rta->rta_flow && memcmp(&fi->fib_nh->nh_tclassid, rta->rta_flow, 4)) | |
756 | goto err_inval; | |
757 | #endif | |
758 | #else | |
759 | goto err_inval; | |
760 | #endif | |
761 | } else { | |
762 | struct fib_nh *nh = fi->fib_nh; | |
763 | if (rta->rta_oif) | |
764 | nh->nh_oif = *rta->rta_oif; | |
765 | if (rta->rta_gw) | |
766 | memcpy(&nh->nh_gw, rta->rta_gw, 4); | |
767 | #ifdef CONFIG_NET_CLS_ROUTE | |
768 | if (rta->rta_flow) | |
769 | memcpy(&nh->nh_tclassid, rta->rta_flow, 4); | |
770 | #endif | |
771 | nh->nh_flags = r->rtm_flags; | |
772 | #ifdef CONFIG_IP_ROUTE_MULTIPATH | |
773 | nh->nh_weight = 1; | |
774 | #endif | |
775 | } | |
776 | ||
777 | #ifdef CONFIG_IP_ROUTE_MULTIPATH_CACHED | |
778 | fi->fib_mp_alg = mp_alg; | |
779 | #endif | |
780 | ||
781 | if (fib_props[r->rtm_type].error) { | |
782 | if (rta->rta_gw || rta->rta_oif || rta->rta_mp) | |
783 | goto err_inval; | |
784 | goto link_it; | |
785 | } | |
786 | ||
787 | if (r->rtm_scope > RT_SCOPE_HOST) | |
788 | goto err_inval; | |
789 | ||
790 | if (r->rtm_scope == RT_SCOPE_HOST) { | |
791 | struct fib_nh *nh = fi->fib_nh; | |
792 | ||
793 | /* Local address is added. */ | |
794 | if (nhs != 1 || nh->nh_gw) | |
795 | goto err_inval; | |
796 | nh->nh_scope = RT_SCOPE_NOWHERE; | |
797 | nh->nh_dev = dev_get_by_index(fi->fib_nh->nh_oif); | |
798 | err = -ENODEV; | |
799 | if (nh->nh_dev == NULL) | |
800 | goto failure; | |
801 | } else { | |
802 | change_nexthops(fi) { | |
803 | if ((err = fib_check_nh(r, fi, nh)) != 0) | |
804 | goto failure; | |
805 | } endfor_nexthops(fi) | |
806 | } | |
807 | ||
808 | if (fi->fib_prefsrc) { | |
809 | if (r->rtm_type != RTN_LOCAL || rta->rta_dst == NULL || | |
810 | memcmp(&fi->fib_prefsrc, rta->rta_dst, 4)) | |
811 | if (inet_addr_type(fi->fib_prefsrc) != RTN_LOCAL) | |
812 | goto err_inval; | |
813 | } | |
814 | ||
815 | link_it: | |
816 | if ((ofi = fib_find_info(fi)) != NULL) { | |
817 | fi->fib_dead = 1; | |
818 | free_fib_info(fi); | |
819 | ofi->fib_treeref++; | |
820 | return ofi; | |
821 | } | |
822 | ||
823 | fi->fib_treeref++; | |
824 | atomic_inc(&fi->fib_clntref); | |
825 | write_lock(&fib_info_lock); | |
826 | hlist_add_head(&fi->fib_hash, | |
827 | &fib_info_hash[fib_info_hashfn(fi)]); | |
828 | if (fi->fib_prefsrc) { | |
829 | struct hlist_head *head; | |
830 | ||
831 | head = &fib_info_laddrhash[fib_laddr_hashfn(fi->fib_prefsrc)]; | |
832 | hlist_add_head(&fi->fib_lhash, head); | |
833 | } | |
834 | change_nexthops(fi) { | |
835 | struct hlist_head *head; | |
836 | unsigned int hash; | |
837 | ||
838 | if (!nh->nh_dev) | |
839 | continue; | |
840 | hash = fib_devindex_hashfn(nh->nh_dev->ifindex); | |
841 | head = &fib_info_devhash[hash]; | |
842 | hlist_add_head(&nh->nh_hash, head); | |
843 | } endfor_nexthops(fi) | |
844 | write_unlock(&fib_info_lock); | |
845 | return fi; | |
846 | ||
847 | err_inval: | |
848 | err = -EINVAL; | |
849 | ||
850 | failure: | |
851 | *errp = err; | |
852 | if (fi) { | |
853 | fi->fib_dead = 1; | |
854 | free_fib_info(fi); | |
855 | } | |
856 | return NULL; | |
857 | } | |
858 | ||
e5b43760 | 859 | /* Note! fib_semantic_match intentionally uses RCU list functions. */ |
1da177e4 LT |
860 | int fib_semantic_match(struct list_head *head, const struct flowi *flp, |
861 | struct fib_result *res, __u32 zone, __u32 mask, | |
862 | int prefixlen) | |
863 | { | |
864 | struct fib_alias *fa; | |
865 | int nh_sel = 0; | |
866 | ||
e5b43760 | 867 | list_for_each_entry_rcu(fa, head, fa_list) { |
1da177e4 LT |
868 | int err; |
869 | ||
870 | if (fa->fa_tos && | |
871 | fa->fa_tos != flp->fl4_tos) | |
872 | continue; | |
873 | ||
874 | if (fa->fa_scope < flp->fl4_scope) | |
875 | continue; | |
876 | ||
877 | fa->fa_state |= FA_S_ACCESSED; | |
878 | ||
879 | err = fib_props[fa->fa_type].error; | |
880 | if (err == 0) { | |
881 | struct fib_info *fi = fa->fa_info; | |
882 | ||
883 | if (fi->fib_flags & RTNH_F_DEAD) | |
884 | continue; | |
885 | ||
886 | switch (fa->fa_type) { | |
887 | case RTN_UNICAST: | |
888 | case RTN_LOCAL: | |
889 | case RTN_BROADCAST: | |
890 | case RTN_ANYCAST: | |
891 | case RTN_MULTICAST: | |
892 | for_nexthops(fi) { | |
893 | if (nh->nh_flags&RTNH_F_DEAD) | |
894 | continue; | |
895 | if (!flp->oif || flp->oif == nh->nh_oif) | |
896 | break; | |
897 | } | |
898 | #ifdef CONFIG_IP_ROUTE_MULTIPATH | |
899 | if (nhsel < fi->fib_nhs) { | |
900 | nh_sel = nhsel; | |
901 | goto out_fill_res; | |
902 | } | |
903 | #else | |
904 | if (nhsel < 1) { | |
905 | goto out_fill_res; | |
906 | } | |
907 | #endif | |
908 | endfor_nexthops(fi); | |
909 | continue; | |
910 | ||
911 | default: | |
912 | printk(KERN_DEBUG "impossible 102\n"); | |
913 | return -EINVAL; | |
914 | }; | |
915 | } | |
916 | return err; | |
917 | } | |
918 | return 1; | |
919 | ||
920 | out_fill_res: | |
921 | res->prefixlen = prefixlen; | |
922 | res->nh_sel = nh_sel; | |
923 | res->type = fa->fa_type; | |
924 | res->scope = fa->fa_scope; | |
925 | res->fi = fa->fa_info; | |
926 | #ifdef CONFIG_IP_ROUTE_MULTIPATH_CACHED | |
927 | res->netmask = mask; | |
928 | res->network = zone & | |
929 | (0xFFFFFFFF >> (32 - prefixlen)); | |
930 | #endif | |
931 | atomic_inc(&res->fi->fib_clntref); | |
932 | return 0; | |
933 | } | |
934 | ||
935 | /* Find appropriate source address to this destination */ | |
936 | ||
937 | u32 __fib_res_prefsrc(struct fib_result *res) | |
938 | { | |
939 | return inet_select_addr(FIB_RES_DEV(*res), FIB_RES_GW(*res), res->scope); | |
940 | } | |
941 | ||
942 | int | |
943 | fib_dump_info(struct sk_buff *skb, u32 pid, u32 seq, int event, | |
944 | u8 tb_id, u8 type, u8 scope, void *dst, int dst_len, u8 tos, | |
b6544c0b | 945 | struct fib_info *fi, unsigned int flags) |
1da177e4 LT |
946 | { |
947 | struct rtmsg *rtm; | |
948 | struct nlmsghdr *nlh; | |
949 | unsigned char *b = skb->tail; | |
950 | ||
b6544c0b | 951 | nlh = NLMSG_NEW(skb, pid, seq, event, sizeof(*rtm), flags); |
1da177e4 LT |
952 | rtm = NLMSG_DATA(nlh); |
953 | rtm->rtm_family = AF_INET; | |
954 | rtm->rtm_dst_len = dst_len; | |
955 | rtm->rtm_src_len = 0; | |
956 | rtm->rtm_tos = tos; | |
957 | rtm->rtm_table = tb_id; | |
958 | rtm->rtm_type = type; | |
959 | rtm->rtm_flags = fi->fib_flags; | |
960 | rtm->rtm_scope = scope; | |
961 | if (rtm->rtm_dst_len) | |
962 | RTA_PUT(skb, RTA_DST, 4, dst); | |
963 | rtm->rtm_protocol = fi->fib_protocol; | |
964 | if (fi->fib_priority) | |
965 | RTA_PUT(skb, RTA_PRIORITY, 4, &fi->fib_priority); | |
966 | #ifdef CONFIG_NET_CLS_ROUTE | |
967 | if (fi->fib_nh[0].nh_tclassid) | |
968 | RTA_PUT(skb, RTA_FLOW, 4, &fi->fib_nh[0].nh_tclassid); | |
969 | #endif | |
970 | if (rtnetlink_put_metrics(skb, fi->fib_metrics) < 0) | |
971 | goto rtattr_failure; | |
972 | if (fi->fib_prefsrc) | |
973 | RTA_PUT(skb, RTA_PREFSRC, 4, &fi->fib_prefsrc); | |
974 | if (fi->fib_nhs == 1) { | |
975 | if (fi->fib_nh->nh_gw) | |
976 | RTA_PUT(skb, RTA_GATEWAY, 4, &fi->fib_nh->nh_gw); | |
977 | if (fi->fib_nh->nh_oif) | |
978 | RTA_PUT(skb, RTA_OIF, sizeof(int), &fi->fib_nh->nh_oif); | |
979 | } | |
980 | #ifdef CONFIG_IP_ROUTE_MULTIPATH | |
981 | if (fi->fib_nhs > 1) { | |
982 | struct rtnexthop *nhp; | |
983 | struct rtattr *mp_head; | |
984 | if (skb_tailroom(skb) <= RTA_SPACE(0)) | |
985 | goto rtattr_failure; | |
986 | mp_head = (struct rtattr*)skb_put(skb, RTA_SPACE(0)); | |
987 | ||
988 | for_nexthops(fi) { | |
989 | if (skb_tailroom(skb) < RTA_ALIGN(RTA_ALIGN(sizeof(*nhp)) + 4)) | |
990 | goto rtattr_failure; | |
991 | nhp = (struct rtnexthop*)skb_put(skb, RTA_ALIGN(sizeof(*nhp))); | |
992 | nhp->rtnh_flags = nh->nh_flags & 0xFF; | |
993 | nhp->rtnh_hops = nh->nh_weight-1; | |
994 | nhp->rtnh_ifindex = nh->nh_oif; | |
995 | if (nh->nh_gw) | |
996 | RTA_PUT(skb, RTA_GATEWAY, 4, &nh->nh_gw); | |
997 | nhp->rtnh_len = skb->tail - (unsigned char*)nhp; | |
998 | } endfor_nexthops(fi); | |
999 | mp_head->rta_type = RTA_MULTIPATH; | |
1000 | mp_head->rta_len = skb->tail - (u8*)mp_head; | |
1001 | } | |
1002 | #endif | |
1003 | nlh->nlmsg_len = skb->tail - b; | |
1004 | return skb->len; | |
1005 | ||
1006 | nlmsg_failure: | |
1007 | rtattr_failure: | |
1008 | skb_trim(skb, b - skb->data); | |
1009 | return -1; | |
1010 | } | |
1011 | ||
1012 | #ifndef CONFIG_IP_NOSIOCRT | |
1013 | ||
1014 | int | |
1015 | fib_convert_rtentry(int cmd, struct nlmsghdr *nl, struct rtmsg *rtm, | |
1016 | struct kern_rta *rta, struct rtentry *r) | |
1017 | { | |
1018 | int plen; | |
1019 | u32 *ptr; | |
1020 | ||
1021 | memset(rtm, 0, sizeof(*rtm)); | |
1022 | memset(rta, 0, sizeof(*rta)); | |
1023 | ||
1024 | if (r->rt_dst.sa_family != AF_INET) | |
1025 | return -EAFNOSUPPORT; | |
1026 | ||
1027 | /* Check mask for validity: | |
1028 | a) it must be contiguous. | |
1029 | b) destination must have all host bits clear. | |
1030 | c) if application forgot to set correct family (AF_INET), | |
1031 | reject request unless it is absolutely clear i.e. | |
1032 | both family and mask are zero. | |
1033 | */ | |
1034 | plen = 32; | |
1035 | ptr = &((struct sockaddr_in*)&r->rt_dst)->sin_addr.s_addr; | |
1036 | if (!(r->rt_flags&RTF_HOST)) { | |
1037 | u32 mask = ((struct sockaddr_in*)&r->rt_genmask)->sin_addr.s_addr; | |
1038 | if (r->rt_genmask.sa_family != AF_INET) { | |
1039 | if (mask || r->rt_genmask.sa_family) | |
1040 | return -EAFNOSUPPORT; | |
1041 | } | |
1042 | if (bad_mask(mask, *ptr)) | |
1043 | return -EINVAL; | |
1044 | plen = inet_mask_len(mask); | |
1045 | } | |
1046 | ||
1047 | nl->nlmsg_flags = NLM_F_REQUEST; | |
28633514 | 1048 | nl->nlmsg_pid = 0; |
1da177e4 LT |
1049 | nl->nlmsg_seq = 0; |
1050 | nl->nlmsg_len = NLMSG_LENGTH(sizeof(*rtm)); | |
1051 | if (cmd == SIOCDELRT) { | |
1052 | nl->nlmsg_type = RTM_DELROUTE; | |
1053 | nl->nlmsg_flags = 0; | |
1054 | } else { | |
1055 | nl->nlmsg_type = RTM_NEWROUTE; | |
1056 | nl->nlmsg_flags = NLM_F_REQUEST|NLM_F_CREATE; | |
1057 | rtm->rtm_protocol = RTPROT_BOOT; | |
1058 | } | |
1059 | ||
1060 | rtm->rtm_dst_len = plen; | |
1061 | rta->rta_dst = ptr; | |
1062 | ||
1063 | if (r->rt_metric) { | |
1064 | *(u32*)&r->rt_pad3 = r->rt_metric - 1; | |
1065 | rta->rta_priority = (u32*)&r->rt_pad3; | |
1066 | } | |
1067 | if (r->rt_flags&RTF_REJECT) { | |
1068 | rtm->rtm_scope = RT_SCOPE_HOST; | |
1069 | rtm->rtm_type = RTN_UNREACHABLE; | |
1070 | return 0; | |
1071 | } | |
1072 | rtm->rtm_scope = RT_SCOPE_NOWHERE; | |
1073 | rtm->rtm_type = RTN_UNICAST; | |
1074 | ||
1075 | if (r->rt_dev) { | |
1076 | char *colon; | |
1077 | struct net_device *dev; | |
1078 | char devname[IFNAMSIZ]; | |
1079 | ||
1080 | if (copy_from_user(devname, r->rt_dev, IFNAMSIZ-1)) | |
1081 | return -EFAULT; | |
1082 | devname[IFNAMSIZ-1] = 0; | |
1083 | colon = strchr(devname, ':'); | |
1084 | if (colon) | |
1085 | *colon = 0; | |
1086 | dev = __dev_get_by_name(devname); | |
1087 | if (!dev) | |
1088 | return -ENODEV; | |
1089 | rta->rta_oif = &dev->ifindex; | |
1090 | if (colon) { | |
1091 | struct in_ifaddr *ifa; | |
e5ed6399 | 1092 | struct in_device *in_dev = __in_dev_get_rtnl(dev); |
1da177e4 LT |
1093 | if (!in_dev) |
1094 | return -ENODEV; | |
1095 | *colon = ':'; | |
1096 | for (ifa = in_dev->ifa_list; ifa; ifa = ifa->ifa_next) | |
1097 | if (strcmp(ifa->ifa_label, devname) == 0) | |
1098 | break; | |
1099 | if (ifa == NULL) | |
1100 | return -ENODEV; | |
1101 | rta->rta_prefsrc = &ifa->ifa_local; | |
1102 | } | |
1103 | } | |
1104 | ||
1105 | ptr = &((struct sockaddr_in*)&r->rt_gateway)->sin_addr.s_addr; | |
1106 | if (r->rt_gateway.sa_family == AF_INET && *ptr) { | |
1107 | rta->rta_gw = ptr; | |
1108 | if (r->rt_flags&RTF_GATEWAY && inet_addr_type(*ptr) == RTN_UNICAST) | |
1109 | rtm->rtm_scope = RT_SCOPE_UNIVERSE; | |
1110 | } | |
1111 | ||
1112 | if (cmd == SIOCDELRT) | |
1113 | return 0; | |
1114 | ||
1115 | if (r->rt_flags&RTF_GATEWAY && rta->rta_gw == NULL) | |
1116 | return -EINVAL; | |
1117 | ||
1118 | if (rtm->rtm_scope == RT_SCOPE_NOWHERE) | |
1119 | rtm->rtm_scope = RT_SCOPE_LINK; | |
1120 | ||
1121 | if (r->rt_flags&(RTF_MTU|RTF_WINDOW|RTF_IRTT)) { | |
1122 | struct rtattr *rec; | |
1123 | struct rtattr *mx = kmalloc(RTA_LENGTH(3*RTA_LENGTH(4)), GFP_KERNEL); | |
1124 | if (mx == NULL) | |
1125 | return -ENOMEM; | |
1126 | rta->rta_mx = mx; | |
1127 | mx->rta_type = RTA_METRICS; | |
1128 | mx->rta_len = RTA_LENGTH(0); | |
1129 | if (r->rt_flags&RTF_MTU) { | |
1130 | rec = (void*)((char*)mx + RTA_ALIGN(mx->rta_len)); | |
1131 | rec->rta_type = RTAX_ADVMSS; | |
1132 | rec->rta_len = RTA_LENGTH(4); | |
1133 | mx->rta_len += RTA_LENGTH(4); | |
1134 | *(u32*)RTA_DATA(rec) = r->rt_mtu - 40; | |
1135 | } | |
1136 | if (r->rt_flags&RTF_WINDOW) { | |
1137 | rec = (void*)((char*)mx + RTA_ALIGN(mx->rta_len)); | |
1138 | rec->rta_type = RTAX_WINDOW; | |
1139 | rec->rta_len = RTA_LENGTH(4); | |
1140 | mx->rta_len += RTA_LENGTH(4); | |
1141 | *(u32*)RTA_DATA(rec) = r->rt_window; | |
1142 | } | |
1143 | if (r->rt_flags&RTF_IRTT) { | |
1144 | rec = (void*)((char*)mx + RTA_ALIGN(mx->rta_len)); | |
1145 | rec->rta_type = RTAX_RTT; | |
1146 | rec->rta_len = RTA_LENGTH(4); | |
1147 | mx->rta_len += RTA_LENGTH(4); | |
1148 | *(u32*)RTA_DATA(rec) = r->rt_irtt<<3; | |
1149 | } | |
1150 | } | |
1151 | return 0; | |
1152 | } | |
1153 | ||
1154 | #endif | |
1155 | ||
1156 | /* | |
1157 | Update FIB if: | |
1158 | - local address disappeared -> we must delete all the entries | |
1159 | referring to it. | |
1160 | - device went down -> we must shutdown all nexthops going via it. | |
1161 | */ | |
1162 | ||
1163 | int fib_sync_down(u32 local, struct net_device *dev, int force) | |
1164 | { | |
1165 | int ret = 0; | |
1166 | int scope = RT_SCOPE_NOWHERE; | |
1167 | ||
1168 | if (force) | |
1169 | scope = -1; | |
1170 | ||
1171 | if (local && fib_info_laddrhash) { | |
1172 | unsigned int hash = fib_laddr_hashfn(local); | |
1173 | struct hlist_head *head = &fib_info_laddrhash[hash]; | |
1174 | struct hlist_node *node; | |
1175 | struct fib_info *fi; | |
1176 | ||
1177 | hlist_for_each_entry(fi, node, head, fib_lhash) { | |
1178 | if (fi->fib_prefsrc == local) { | |
1179 | fi->fib_flags |= RTNH_F_DEAD; | |
1180 | ret++; | |
1181 | } | |
1182 | } | |
1183 | } | |
1184 | ||
1185 | if (dev) { | |
1186 | struct fib_info *prev_fi = NULL; | |
1187 | unsigned int hash = fib_devindex_hashfn(dev->ifindex); | |
1188 | struct hlist_head *head = &fib_info_devhash[hash]; | |
1189 | struct hlist_node *node; | |
1190 | struct fib_nh *nh; | |
1191 | ||
1192 | hlist_for_each_entry(nh, node, head, nh_hash) { | |
1193 | struct fib_info *fi = nh->nh_parent; | |
1194 | int dead; | |
1195 | ||
1196 | BUG_ON(!fi->fib_nhs); | |
1197 | if (nh->nh_dev != dev || fi == prev_fi) | |
1198 | continue; | |
1199 | prev_fi = fi; | |
1200 | dead = 0; | |
1201 | change_nexthops(fi) { | |
1202 | if (nh->nh_flags&RTNH_F_DEAD) | |
1203 | dead++; | |
1204 | else if (nh->nh_dev == dev && | |
1205 | nh->nh_scope != scope) { | |
1206 | nh->nh_flags |= RTNH_F_DEAD; | |
1207 | #ifdef CONFIG_IP_ROUTE_MULTIPATH | |
1208 | spin_lock_bh(&fib_multipath_lock); | |
1209 | fi->fib_power -= nh->nh_power; | |
1210 | nh->nh_power = 0; | |
1211 | spin_unlock_bh(&fib_multipath_lock); | |
1212 | #endif | |
1213 | dead++; | |
1214 | } | |
1215 | #ifdef CONFIG_IP_ROUTE_MULTIPATH | |
1216 | if (force > 1 && nh->nh_dev == dev) { | |
1217 | dead = fi->fib_nhs; | |
1218 | break; | |
1219 | } | |
1220 | #endif | |
1221 | } endfor_nexthops(fi) | |
1222 | if (dead == fi->fib_nhs) { | |
1223 | fi->fib_flags |= RTNH_F_DEAD; | |
1224 | ret++; | |
1225 | } | |
1226 | } | |
1227 | } | |
1228 | ||
1229 | return ret; | |
1230 | } | |
1231 | ||
1232 | #ifdef CONFIG_IP_ROUTE_MULTIPATH | |
1233 | ||
1234 | /* | |
1235 | Dead device goes up. We wake up dead nexthops. | |
1236 | It takes sense only on multipath routes. | |
1237 | */ | |
1238 | ||
1239 | int fib_sync_up(struct net_device *dev) | |
1240 | { | |
1241 | struct fib_info *prev_fi; | |
1242 | unsigned int hash; | |
1243 | struct hlist_head *head; | |
1244 | struct hlist_node *node; | |
1245 | struct fib_nh *nh; | |
1246 | int ret; | |
1247 | ||
1248 | if (!(dev->flags&IFF_UP)) | |
1249 | return 0; | |
1250 | ||
1251 | prev_fi = NULL; | |
1252 | hash = fib_devindex_hashfn(dev->ifindex); | |
1253 | head = &fib_info_devhash[hash]; | |
1254 | ret = 0; | |
1255 | ||
1256 | hlist_for_each_entry(nh, node, head, nh_hash) { | |
1257 | struct fib_info *fi = nh->nh_parent; | |
1258 | int alive; | |
1259 | ||
1260 | BUG_ON(!fi->fib_nhs); | |
1261 | if (nh->nh_dev != dev || fi == prev_fi) | |
1262 | continue; | |
1263 | ||
1264 | prev_fi = fi; | |
1265 | alive = 0; | |
1266 | change_nexthops(fi) { | |
1267 | if (!(nh->nh_flags&RTNH_F_DEAD)) { | |
1268 | alive++; | |
1269 | continue; | |
1270 | } | |
1271 | if (nh->nh_dev == NULL || !(nh->nh_dev->flags&IFF_UP)) | |
1272 | continue; | |
e5ed6399 | 1273 | if (nh->nh_dev != dev || !__in_dev_get_rtnl(dev)) |
1da177e4 LT |
1274 | continue; |
1275 | alive++; | |
1276 | spin_lock_bh(&fib_multipath_lock); | |
1277 | nh->nh_power = 0; | |
1278 | nh->nh_flags &= ~RTNH_F_DEAD; | |
1279 | spin_unlock_bh(&fib_multipath_lock); | |
1280 | } endfor_nexthops(fi) | |
1281 | ||
1282 | if (alive > 0) { | |
1283 | fi->fib_flags &= ~RTNH_F_DEAD; | |
1284 | ret++; | |
1285 | } | |
1286 | } | |
1287 | ||
1288 | return ret; | |
1289 | } | |
1290 | ||
1291 | /* | |
1292 | The algorithm is suboptimal, but it provides really | |
1293 | fair weighted route distribution. | |
1294 | */ | |
1295 | ||
1296 | void fib_select_multipath(const struct flowi *flp, struct fib_result *res) | |
1297 | { | |
1298 | struct fib_info *fi = res->fi; | |
1299 | int w; | |
1300 | ||
1301 | spin_lock_bh(&fib_multipath_lock); | |
1302 | if (fi->fib_power <= 0) { | |
1303 | int power = 0; | |
1304 | change_nexthops(fi) { | |
1305 | if (!(nh->nh_flags&RTNH_F_DEAD)) { | |
1306 | power += nh->nh_weight; | |
1307 | nh->nh_power = nh->nh_weight; | |
1308 | } | |
1309 | } endfor_nexthops(fi); | |
1310 | fi->fib_power = power; | |
1311 | if (power <= 0) { | |
1312 | spin_unlock_bh(&fib_multipath_lock); | |
1313 | /* Race condition: route has just become dead. */ | |
1314 | res->nh_sel = 0; | |
1315 | return; | |
1316 | } | |
1317 | } | |
1318 | ||
1319 | ||
1320 | /* w should be random number [0..fi->fib_power-1], | |
1321 | it is pretty bad approximation. | |
1322 | */ | |
1323 | ||
1324 | w = jiffies % fi->fib_power; | |
1325 | ||
1326 | change_nexthops(fi) { | |
1327 | if (!(nh->nh_flags&RTNH_F_DEAD) && nh->nh_power) { | |
1328 | if ((w -= nh->nh_power) <= 0) { | |
1329 | nh->nh_power--; | |
1330 | fi->fib_power--; | |
1331 | res->nh_sel = nhsel; | |
1332 | spin_unlock_bh(&fib_multipath_lock); | |
1333 | return; | |
1334 | } | |
1335 | } | |
1336 | } endfor_nexthops(fi); | |
1337 | ||
1338 | /* Race condition: route has just become dead. */ | |
1339 | res->nh_sel = 0; | |
1340 | spin_unlock_bh(&fib_multipath_lock); | |
1341 | } | |
1342 | #endif |