Loading...
1// SPDX-License-Identifier: GPL-2.0
2/* Copyright 2011-2014 Autronica Fire and Security AS
3 *
4 * Author(s):
5 * 2011-2014 Arvid Brodin, arvid.brodin@alten.se
6 *
7 * Routines for handling Netlink messages for HSR and PRP.
8 */
9
10#include "hsr_netlink.h"
11#include <linux/kernel.h>
12#include <net/rtnetlink.h>
13#include <net/genetlink.h>
14#include "hsr_main.h"
15#include "hsr_device.h"
16#include "hsr_framereg.h"
17
18static const struct nla_policy hsr_policy[IFLA_HSR_MAX + 1] = {
19 [IFLA_HSR_SLAVE1] = { .type = NLA_U32 },
20 [IFLA_HSR_SLAVE2] = { .type = NLA_U32 },
21 [IFLA_HSR_MULTICAST_SPEC] = { .type = NLA_U8 },
22 [IFLA_HSR_VERSION] = { .type = NLA_U8 },
23 [IFLA_HSR_SUPERVISION_ADDR] = { .len = ETH_ALEN },
24 [IFLA_HSR_SEQ_NR] = { .type = NLA_U16 },
25 [IFLA_HSR_PROTOCOL] = { .type = NLA_U8 },
26};
27
28/* Here, it seems a netdevice has already been allocated for us, and the
29 * hsr_dev_setup routine has been executed. Nice!
30 */
31static int hsr_newlink(struct net *src_net, struct net_device *dev,
32 struct nlattr *tb[], struct nlattr *data[],
33 struct netlink_ext_ack *extack)
34{
35 enum hsr_version proto_version;
36 unsigned char multicast_spec;
37 u8 proto = HSR_PROTOCOL_HSR;
38 struct net_device *link[2];
39
40 if (!data) {
41 NL_SET_ERR_MSG_MOD(extack, "No slave devices specified");
42 return -EINVAL;
43 }
44 if (!data[IFLA_HSR_SLAVE1]) {
45 NL_SET_ERR_MSG_MOD(extack, "Slave1 device not specified");
46 return -EINVAL;
47 }
48 link[0] = __dev_get_by_index(src_net,
49 nla_get_u32(data[IFLA_HSR_SLAVE1]));
50 if (!link[0]) {
51 NL_SET_ERR_MSG_MOD(extack, "Slave1 does not exist");
52 return -EINVAL;
53 }
54 if (!data[IFLA_HSR_SLAVE2]) {
55 NL_SET_ERR_MSG_MOD(extack, "Slave2 device not specified");
56 return -EINVAL;
57 }
58 link[1] = __dev_get_by_index(src_net,
59 nla_get_u32(data[IFLA_HSR_SLAVE2]));
60 if (!link[1]) {
61 NL_SET_ERR_MSG_MOD(extack, "Slave2 does not exist");
62 return -EINVAL;
63 }
64
65 if (link[0] == link[1]) {
66 NL_SET_ERR_MSG_MOD(extack, "Slave1 and Slave2 are same");
67 return -EINVAL;
68 }
69
70 if (!data[IFLA_HSR_MULTICAST_SPEC])
71 multicast_spec = 0;
72 else
73 multicast_spec = nla_get_u8(data[IFLA_HSR_MULTICAST_SPEC]);
74
75 if (data[IFLA_HSR_PROTOCOL])
76 proto = nla_get_u8(data[IFLA_HSR_PROTOCOL]);
77
78 if (proto >= HSR_PROTOCOL_MAX) {
79 NL_SET_ERR_MSG_MOD(extack, "Unsupported protocol");
80 return -EINVAL;
81 }
82
83 if (!data[IFLA_HSR_VERSION]) {
84 proto_version = HSR_V0;
85 } else {
86 if (proto == HSR_PROTOCOL_PRP) {
87 NL_SET_ERR_MSG_MOD(extack, "PRP version unsupported");
88 return -EINVAL;
89 }
90
91 proto_version = nla_get_u8(data[IFLA_HSR_VERSION]);
92 if (proto_version > HSR_V1) {
93 NL_SET_ERR_MSG_MOD(extack,
94 "Only HSR version 0/1 supported");
95 return -EINVAL;
96 }
97 }
98
99 if (proto == HSR_PROTOCOL_PRP)
100 proto_version = PRP_V1;
101
102 return hsr_dev_finalize(dev, link, multicast_spec, proto_version, extack);
103}
104
105static void hsr_dellink(struct net_device *dev, struct list_head *head)
106{
107 struct hsr_priv *hsr = netdev_priv(dev);
108
109 del_timer_sync(&hsr->prune_timer);
110 del_timer_sync(&hsr->announce_timer);
111
112 hsr_debugfs_term(hsr);
113 hsr_del_ports(hsr);
114
115 hsr_del_self_node(hsr);
116 hsr_del_nodes(&hsr->node_db);
117
118 unregister_netdevice_queue(dev, head);
119}
120
121static int hsr_fill_info(struct sk_buff *skb, const struct net_device *dev)
122{
123 struct hsr_priv *hsr = netdev_priv(dev);
124 u8 proto = HSR_PROTOCOL_HSR;
125 struct hsr_port *port;
126
127 port = hsr_port_get_hsr(hsr, HSR_PT_SLAVE_A);
128 if (port) {
129 if (nla_put_u32(skb, IFLA_HSR_SLAVE1, port->dev->ifindex))
130 goto nla_put_failure;
131 }
132
133 port = hsr_port_get_hsr(hsr, HSR_PT_SLAVE_B);
134 if (port) {
135 if (nla_put_u32(skb, IFLA_HSR_SLAVE2, port->dev->ifindex))
136 goto nla_put_failure;
137 }
138
139 if (nla_put(skb, IFLA_HSR_SUPERVISION_ADDR, ETH_ALEN,
140 hsr->sup_multicast_addr) ||
141 nla_put_u16(skb, IFLA_HSR_SEQ_NR, hsr->sequence_nr))
142 goto nla_put_failure;
143 if (hsr->prot_version == PRP_V1)
144 proto = HSR_PROTOCOL_PRP;
145 if (nla_put_u8(skb, IFLA_HSR_PROTOCOL, proto))
146 goto nla_put_failure;
147
148 return 0;
149
150nla_put_failure:
151 return -EMSGSIZE;
152}
153
154static struct rtnl_link_ops hsr_link_ops __read_mostly = {
155 .kind = "hsr",
156 .maxtype = IFLA_HSR_MAX,
157 .policy = hsr_policy,
158 .priv_size = sizeof(struct hsr_priv),
159 .setup = hsr_dev_setup,
160 .newlink = hsr_newlink,
161 .dellink = hsr_dellink,
162 .fill_info = hsr_fill_info,
163};
164
165/* attribute policy */
166static const struct nla_policy hsr_genl_policy[HSR_A_MAX + 1] = {
167 [HSR_A_NODE_ADDR] = { .len = ETH_ALEN },
168 [HSR_A_NODE_ADDR_B] = { .len = ETH_ALEN },
169 [HSR_A_IFINDEX] = { .type = NLA_U32 },
170 [HSR_A_IF1_AGE] = { .type = NLA_U32 },
171 [HSR_A_IF2_AGE] = { .type = NLA_U32 },
172 [HSR_A_IF1_SEQ] = { .type = NLA_U16 },
173 [HSR_A_IF2_SEQ] = { .type = NLA_U16 },
174};
175
176static struct genl_family hsr_genl_family;
177
178static const struct genl_multicast_group hsr_mcgrps[] = {
179 { .name = "hsr-network", },
180};
181
182/* This is called if for some node with MAC address addr, we only get frames
183 * over one of the slave interfaces. This would indicate an open network ring
184 * (i.e. a link has failed somewhere).
185 */
186void hsr_nl_ringerror(struct hsr_priv *hsr, unsigned char addr[ETH_ALEN],
187 struct hsr_port *port)
188{
189 struct sk_buff *skb;
190 void *msg_head;
191 struct hsr_port *master;
192 int res;
193
194 skb = genlmsg_new(NLMSG_GOODSIZE, GFP_ATOMIC);
195 if (!skb)
196 goto fail;
197
198 msg_head = genlmsg_put(skb, 0, 0, &hsr_genl_family, 0,
199 HSR_C_RING_ERROR);
200 if (!msg_head)
201 goto nla_put_failure;
202
203 res = nla_put(skb, HSR_A_NODE_ADDR, ETH_ALEN, addr);
204 if (res < 0)
205 goto nla_put_failure;
206
207 res = nla_put_u32(skb, HSR_A_IFINDEX, port->dev->ifindex);
208 if (res < 0)
209 goto nla_put_failure;
210
211 genlmsg_end(skb, msg_head);
212 genlmsg_multicast(&hsr_genl_family, skb, 0, 0, GFP_ATOMIC);
213
214 return;
215
216nla_put_failure:
217 kfree_skb(skb);
218
219fail:
220 rcu_read_lock();
221 master = hsr_port_get_hsr(hsr, HSR_PT_MASTER);
222 netdev_warn(master->dev, "Could not send HSR ring error message\n");
223 rcu_read_unlock();
224}
225
226/* This is called when we haven't heard from the node with MAC address addr for
227 * some time (just before the node is removed from the node table/list).
228 */
229void hsr_nl_nodedown(struct hsr_priv *hsr, unsigned char addr[ETH_ALEN])
230{
231 struct sk_buff *skb;
232 void *msg_head;
233 struct hsr_port *master;
234 int res;
235
236 skb = genlmsg_new(NLMSG_GOODSIZE, GFP_ATOMIC);
237 if (!skb)
238 goto fail;
239
240 msg_head = genlmsg_put(skb, 0, 0, &hsr_genl_family, 0, HSR_C_NODE_DOWN);
241 if (!msg_head)
242 goto nla_put_failure;
243
244 res = nla_put(skb, HSR_A_NODE_ADDR, ETH_ALEN, addr);
245 if (res < 0)
246 goto nla_put_failure;
247
248 genlmsg_end(skb, msg_head);
249 genlmsg_multicast(&hsr_genl_family, skb, 0, 0, GFP_ATOMIC);
250
251 return;
252
253nla_put_failure:
254 kfree_skb(skb);
255
256fail:
257 rcu_read_lock();
258 master = hsr_port_get_hsr(hsr, HSR_PT_MASTER);
259 netdev_warn(master->dev, "Could not send HSR node down\n");
260 rcu_read_unlock();
261}
262
263/* HSR_C_GET_NODE_STATUS lets userspace query the internal HSR node table
264 * about the status of a specific node in the network, defined by its MAC
265 * address.
266 *
267 * Input: hsr ifindex, node mac address
268 * Output: hsr ifindex, node mac address (copied from request),
269 * age of latest frame from node over slave 1, slave 2 [ms]
270 */
271static int hsr_get_node_status(struct sk_buff *skb_in, struct genl_info *info)
272{
273 /* For receiving */
274 struct nlattr *na;
275 struct net_device *hsr_dev;
276
277 /* For sending */
278 struct sk_buff *skb_out;
279 void *msg_head;
280 struct hsr_priv *hsr;
281 struct hsr_port *port;
282 unsigned char hsr_node_addr_b[ETH_ALEN];
283 int hsr_node_if1_age;
284 u16 hsr_node_if1_seq;
285 int hsr_node_if2_age;
286 u16 hsr_node_if2_seq;
287 int addr_b_ifindex;
288 int res;
289
290 if (!info)
291 goto invalid;
292
293 na = info->attrs[HSR_A_IFINDEX];
294 if (!na)
295 goto invalid;
296 na = info->attrs[HSR_A_NODE_ADDR];
297 if (!na)
298 goto invalid;
299
300 rcu_read_lock();
301 hsr_dev = dev_get_by_index_rcu(genl_info_net(info),
302 nla_get_u32(info->attrs[HSR_A_IFINDEX]));
303 if (!hsr_dev)
304 goto rcu_unlock;
305 if (!is_hsr_master(hsr_dev))
306 goto rcu_unlock;
307
308 /* Send reply */
309 skb_out = genlmsg_new(NLMSG_GOODSIZE, GFP_ATOMIC);
310 if (!skb_out) {
311 res = -ENOMEM;
312 goto fail;
313 }
314
315 msg_head = genlmsg_put(skb_out, NETLINK_CB(skb_in).portid,
316 info->snd_seq, &hsr_genl_family, 0,
317 HSR_C_SET_NODE_STATUS);
318 if (!msg_head) {
319 res = -ENOMEM;
320 goto nla_put_failure;
321 }
322
323 res = nla_put_u32(skb_out, HSR_A_IFINDEX, hsr_dev->ifindex);
324 if (res < 0)
325 goto nla_put_failure;
326
327 hsr = netdev_priv(hsr_dev);
328 res = hsr_get_node_data(hsr,
329 (unsigned char *)
330 nla_data(info->attrs[HSR_A_NODE_ADDR]),
331 hsr_node_addr_b,
332 &addr_b_ifindex,
333 &hsr_node_if1_age,
334 &hsr_node_if1_seq,
335 &hsr_node_if2_age,
336 &hsr_node_if2_seq);
337 if (res < 0)
338 goto nla_put_failure;
339
340 res = nla_put(skb_out, HSR_A_NODE_ADDR, ETH_ALEN,
341 nla_data(info->attrs[HSR_A_NODE_ADDR]));
342 if (res < 0)
343 goto nla_put_failure;
344
345 if (addr_b_ifindex > -1) {
346 res = nla_put(skb_out, HSR_A_NODE_ADDR_B, ETH_ALEN,
347 hsr_node_addr_b);
348 if (res < 0)
349 goto nla_put_failure;
350
351 res = nla_put_u32(skb_out, HSR_A_ADDR_B_IFINDEX,
352 addr_b_ifindex);
353 if (res < 0)
354 goto nla_put_failure;
355 }
356
357 res = nla_put_u32(skb_out, HSR_A_IF1_AGE, hsr_node_if1_age);
358 if (res < 0)
359 goto nla_put_failure;
360 res = nla_put_u16(skb_out, HSR_A_IF1_SEQ, hsr_node_if1_seq);
361 if (res < 0)
362 goto nla_put_failure;
363 port = hsr_port_get_hsr(hsr, HSR_PT_SLAVE_A);
364 if (port)
365 res = nla_put_u32(skb_out, HSR_A_IF1_IFINDEX,
366 port->dev->ifindex);
367 if (res < 0)
368 goto nla_put_failure;
369
370 res = nla_put_u32(skb_out, HSR_A_IF2_AGE, hsr_node_if2_age);
371 if (res < 0)
372 goto nla_put_failure;
373 res = nla_put_u16(skb_out, HSR_A_IF2_SEQ, hsr_node_if2_seq);
374 if (res < 0)
375 goto nla_put_failure;
376 port = hsr_port_get_hsr(hsr, HSR_PT_SLAVE_B);
377 if (port)
378 res = nla_put_u32(skb_out, HSR_A_IF2_IFINDEX,
379 port->dev->ifindex);
380 if (res < 0)
381 goto nla_put_failure;
382
383 rcu_read_unlock();
384
385 genlmsg_end(skb_out, msg_head);
386 genlmsg_unicast(genl_info_net(info), skb_out, info->snd_portid);
387
388 return 0;
389
390rcu_unlock:
391 rcu_read_unlock();
392invalid:
393 netlink_ack(skb_in, nlmsg_hdr(skb_in), -EINVAL, NULL);
394 return 0;
395
396nla_put_failure:
397 kfree_skb(skb_out);
398 /* Fall through */
399
400fail:
401 rcu_read_unlock();
402 return res;
403}
404
405/* Get a list of MacAddressA of all nodes known to this node (including self).
406 */
407static int hsr_get_node_list(struct sk_buff *skb_in, struct genl_info *info)
408{
409 unsigned char addr[ETH_ALEN];
410 struct net_device *hsr_dev;
411 struct sk_buff *skb_out;
412 struct hsr_priv *hsr;
413 bool restart = false;
414 struct nlattr *na;
415 void *pos = NULL;
416 void *msg_head;
417 int res;
418
419 if (!info)
420 goto invalid;
421
422 na = info->attrs[HSR_A_IFINDEX];
423 if (!na)
424 goto invalid;
425
426 rcu_read_lock();
427 hsr_dev = dev_get_by_index_rcu(genl_info_net(info),
428 nla_get_u32(info->attrs[HSR_A_IFINDEX]));
429 if (!hsr_dev)
430 goto rcu_unlock;
431 if (!is_hsr_master(hsr_dev))
432 goto rcu_unlock;
433
434restart:
435 /* Send reply */
436 skb_out = genlmsg_new(GENLMSG_DEFAULT_SIZE, GFP_ATOMIC);
437 if (!skb_out) {
438 res = -ENOMEM;
439 goto fail;
440 }
441
442 msg_head = genlmsg_put(skb_out, NETLINK_CB(skb_in).portid,
443 info->snd_seq, &hsr_genl_family, 0,
444 HSR_C_SET_NODE_LIST);
445 if (!msg_head) {
446 res = -ENOMEM;
447 goto nla_put_failure;
448 }
449
450 if (!restart) {
451 res = nla_put_u32(skb_out, HSR_A_IFINDEX, hsr_dev->ifindex);
452 if (res < 0)
453 goto nla_put_failure;
454 }
455
456 hsr = netdev_priv(hsr_dev);
457
458 if (!pos)
459 pos = hsr_get_next_node(hsr, NULL, addr);
460 while (pos) {
461 res = nla_put(skb_out, HSR_A_NODE_ADDR, ETH_ALEN, addr);
462 if (res < 0) {
463 if (res == -EMSGSIZE) {
464 genlmsg_end(skb_out, msg_head);
465 genlmsg_unicast(genl_info_net(info), skb_out,
466 info->snd_portid);
467 restart = true;
468 goto restart;
469 }
470 goto nla_put_failure;
471 }
472 pos = hsr_get_next_node(hsr, pos, addr);
473 }
474 rcu_read_unlock();
475
476 genlmsg_end(skb_out, msg_head);
477 genlmsg_unicast(genl_info_net(info), skb_out, info->snd_portid);
478
479 return 0;
480
481rcu_unlock:
482 rcu_read_unlock();
483invalid:
484 netlink_ack(skb_in, nlmsg_hdr(skb_in), -EINVAL, NULL);
485 return 0;
486
487nla_put_failure:
488 nlmsg_free(skb_out);
489 /* Fall through */
490
491fail:
492 rcu_read_unlock();
493 return res;
494}
495
496static const struct genl_small_ops hsr_ops[] = {
497 {
498 .cmd = HSR_C_GET_NODE_STATUS,
499 .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP,
500 .flags = 0,
501 .doit = hsr_get_node_status,
502 .dumpit = NULL,
503 },
504 {
505 .cmd = HSR_C_GET_NODE_LIST,
506 .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP,
507 .flags = 0,
508 .doit = hsr_get_node_list,
509 .dumpit = NULL,
510 },
511};
512
513static struct genl_family hsr_genl_family __ro_after_init = {
514 .hdrsize = 0,
515 .name = "HSR",
516 .version = 1,
517 .maxattr = HSR_A_MAX,
518 .policy = hsr_genl_policy,
519 .netnsok = true,
520 .module = THIS_MODULE,
521 .small_ops = hsr_ops,
522 .n_small_ops = ARRAY_SIZE(hsr_ops),
523 .mcgrps = hsr_mcgrps,
524 .n_mcgrps = ARRAY_SIZE(hsr_mcgrps),
525};
526
527int __init hsr_netlink_init(void)
528{
529 int rc;
530
531 rc = rtnl_link_register(&hsr_link_ops);
532 if (rc)
533 goto fail_rtnl_link_register;
534
535 rc = genl_register_family(&hsr_genl_family);
536 if (rc)
537 goto fail_genl_register_family;
538
539 hsr_debugfs_create_root();
540 return 0;
541
542fail_genl_register_family:
543 rtnl_link_unregister(&hsr_link_ops);
544fail_rtnl_link_register:
545
546 return rc;
547}
548
549void __exit hsr_netlink_exit(void)
550{
551 genl_unregister_family(&hsr_genl_family);
552 rtnl_link_unregister(&hsr_link_ops);
553}
554
555MODULE_ALIAS_RTNL_LINK("hsr");
1/* Copyright 2011-2013 Autronica Fire and Security AS
2 *
3 * This program is free software; you can redistribute it and/or modify it
4 * under the terms of the GNU General Public License as published by the Free
5 * Software Foundation; either version 2 of the License, or (at your option)
6 * any later version.
7 *
8 * Author(s):
9 * 2011-2013 Arvid Brodin, arvid.brodin@xdin.com
10 *
11 * Routines for handling Netlink messages for HSR.
12 */
13
14#include "hsr_netlink.h"
15#include <linux/kernel.h>
16#include <net/rtnetlink.h>
17#include <net/genetlink.h>
18#include "hsr_main.h"
19#include "hsr_device.h"
20#include "hsr_framereg.h"
21
22static const struct nla_policy hsr_policy[IFLA_HSR_MAX + 1] = {
23 [IFLA_HSR_SLAVE1] = { .type = NLA_U32 },
24 [IFLA_HSR_SLAVE2] = { .type = NLA_U32 },
25 [IFLA_HSR_MULTICAST_SPEC] = { .type = NLA_U8 },
26 [IFLA_HSR_SUPERVISION_ADDR] = { .type = NLA_BINARY, .len = ETH_ALEN },
27 [IFLA_HSR_SEQ_NR] = { .type = NLA_U16 },
28};
29
30
31/* Here, it seems a netdevice has already been allocated for us, and the
32 * hsr_dev_setup routine has been executed. Nice!
33 */
34static int hsr_newlink(struct net *src_net, struct net_device *dev,
35 struct nlattr *tb[], struct nlattr *data[])
36{
37 struct net_device *link[2];
38 unsigned char multicast_spec;
39
40 if (!data[IFLA_HSR_SLAVE1]) {
41 netdev_info(dev, "IFLA_HSR_SLAVE1 missing!\n");
42 return -EINVAL;
43 }
44 link[0] = __dev_get_by_index(src_net, nla_get_u32(data[IFLA_HSR_SLAVE1]));
45 if (!data[IFLA_HSR_SLAVE2]) {
46 netdev_info(dev, "IFLA_HSR_SLAVE2 missing!\n");
47 return -EINVAL;
48 }
49 link[1] = __dev_get_by_index(src_net, nla_get_u32(data[IFLA_HSR_SLAVE2]));
50
51 if (!link[0] || !link[1])
52 return -ENODEV;
53 if (link[0] == link[1])
54 return -EINVAL;
55
56 if (!data[IFLA_HSR_MULTICAST_SPEC])
57 multicast_spec = 0;
58 else
59 multicast_spec = nla_get_u8(data[IFLA_HSR_MULTICAST_SPEC]);
60
61 return hsr_dev_finalize(dev, link, multicast_spec);
62}
63
64static int hsr_fill_info(struct sk_buff *skb, const struct net_device *dev)
65{
66 struct hsr_priv *hsr_priv;
67
68 hsr_priv = netdev_priv(dev);
69
70 if (hsr_priv->slave[0])
71 if (nla_put_u32(skb, IFLA_HSR_SLAVE1, hsr_priv->slave[0]->ifindex))
72 goto nla_put_failure;
73
74 if (hsr_priv->slave[1])
75 if (nla_put_u32(skb, IFLA_HSR_SLAVE2, hsr_priv->slave[1]->ifindex))
76 goto nla_put_failure;
77
78 if (nla_put(skb, IFLA_HSR_SUPERVISION_ADDR, ETH_ALEN,
79 hsr_priv->sup_multicast_addr) ||
80 nla_put_u16(skb, IFLA_HSR_SEQ_NR, hsr_priv->sequence_nr))
81 goto nla_put_failure;
82
83 return 0;
84
85nla_put_failure:
86 return -EMSGSIZE;
87}
88
89static struct rtnl_link_ops hsr_link_ops __read_mostly = {
90 .kind = "hsr",
91 .maxtype = IFLA_HSR_MAX,
92 .policy = hsr_policy,
93 .priv_size = sizeof(struct hsr_priv),
94 .setup = hsr_dev_setup,
95 .newlink = hsr_newlink,
96 .fill_info = hsr_fill_info,
97};
98
99
100
101/* attribute policy */
102/* NLA_BINARY missing in libnl; use NLA_UNSPEC in userspace instead. */
103static const struct nla_policy hsr_genl_policy[HSR_A_MAX + 1] = {
104 [HSR_A_NODE_ADDR] = { .type = NLA_BINARY, .len = ETH_ALEN },
105 [HSR_A_NODE_ADDR_B] = { .type = NLA_BINARY, .len = ETH_ALEN },
106 [HSR_A_IFINDEX] = { .type = NLA_U32 },
107 [HSR_A_IF1_AGE] = { .type = NLA_U32 },
108 [HSR_A_IF2_AGE] = { .type = NLA_U32 },
109 [HSR_A_IF1_SEQ] = { .type = NLA_U16 },
110 [HSR_A_IF2_SEQ] = { .type = NLA_U16 },
111};
112
113static struct genl_family hsr_genl_family = {
114 .id = GENL_ID_GENERATE,
115 .hdrsize = 0,
116 .name = "HSR",
117 .version = 1,
118 .maxattr = HSR_A_MAX,
119};
120
121static const struct genl_multicast_group hsr_mcgrps[] = {
122 { .name = "hsr-network", },
123};
124
125
126
127/* This is called if for some node with MAC address addr, we only get frames
128 * over one of the slave interfaces. This would indicate an open network ring
129 * (i.e. a link has failed somewhere).
130 */
131void hsr_nl_ringerror(struct hsr_priv *hsr_priv, unsigned char addr[ETH_ALEN],
132 enum hsr_dev_idx dev_idx)
133{
134 struct sk_buff *skb;
135 void *msg_head;
136 int res;
137 int ifindex;
138
139 skb = genlmsg_new(NLMSG_GOODSIZE, GFP_ATOMIC);
140 if (!skb)
141 goto fail;
142
143 msg_head = genlmsg_put(skb, 0, 0, &hsr_genl_family, 0, HSR_C_RING_ERROR);
144 if (!msg_head)
145 goto nla_put_failure;
146
147 res = nla_put(skb, HSR_A_NODE_ADDR, ETH_ALEN, addr);
148 if (res < 0)
149 goto nla_put_failure;
150
151 if (hsr_priv->slave[dev_idx])
152 ifindex = hsr_priv->slave[dev_idx]->ifindex;
153 else
154 ifindex = -1;
155 res = nla_put_u32(skb, HSR_A_IFINDEX, ifindex);
156 if (res < 0)
157 goto nla_put_failure;
158
159 genlmsg_end(skb, msg_head);
160 genlmsg_multicast(&hsr_genl_family, skb, 0, 0, GFP_ATOMIC);
161
162 return;
163
164nla_put_failure:
165 kfree_skb(skb);
166
167fail:
168 netdev_warn(hsr_priv->dev, "Could not send HSR ring error message\n");
169}
170
171/* This is called when we haven't heard from the node with MAC address addr for
172 * some time (just before the node is removed from the node table/list).
173 */
174void hsr_nl_nodedown(struct hsr_priv *hsr_priv, unsigned char addr[ETH_ALEN])
175{
176 struct sk_buff *skb;
177 void *msg_head;
178 int res;
179
180 skb = genlmsg_new(NLMSG_GOODSIZE, GFP_ATOMIC);
181 if (!skb)
182 goto fail;
183
184 msg_head = genlmsg_put(skb, 0, 0, &hsr_genl_family, 0, HSR_C_NODE_DOWN);
185 if (!msg_head)
186 goto nla_put_failure;
187
188
189 res = nla_put(skb, HSR_A_NODE_ADDR, ETH_ALEN, addr);
190 if (res < 0)
191 goto nla_put_failure;
192
193 genlmsg_end(skb, msg_head);
194 genlmsg_multicast(&hsr_genl_family, skb, 0, 0, GFP_ATOMIC);
195
196 return;
197
198nla_put_failure:
199 kfree_skb(skb);
200
201fail:
202 netdev_warn(hsr_priv->dev, "Could not send HSR node down\n");
203}
204
205
206/* HSR_C_GET_NODE_STATUS lets userspace query the internal HSR node table
207 * about the status of a specific node in the network, defined by its MAC
208 * address.
209 *
210 * Input: hsr ifindex, node mac address
211 * Output: hsr ifindex, node mac address (copied from request),
212 * age of latest frame from node over slave 1, slave 2 [ms]
213 */
214static int hsr_get_node_status(struct sk_buff *skb_in, struct genl_info *info)
215{
216 /* For receiving */
217 struct nlattr *na;
218 struct net_device *hsr_dev;
219
220 /* For sending */
221 struct sk_buff *skb_out;
222 void *msg_head;
223 struct hsr_priv *hsr_priv;
224 unsigned char hsr_node_addr_b[ETH_ALEN];
225 int hsr_node_if1_age;
226 u16 hsr_node_if1_seq;
227 int hsr_node_if2_age;
228 u16 hsr_node_if2_seq;
229 int addr_b_ifindex;
230 int res;
231
232 if (!info)
233 goto invalid;
234
235 na = info->attrs[HSR_A_IFINDEX];
236 if (!na)
237 goto invalid;
238 na = info->attrs[HSR_A_NODE_ADDR];
239 if (!na)
240 goto invalid;
241
242 hsr_dev = __dev_get_by_index(genl_info_net(info),
243 nla_get_u32(info->attrs[HSR_A_IFINDEX]));
244 if (!hsr_dev)
245 goto invalid;
246 if (!is_hsr_master(hsr_dev))
247 goto invalid;
248
249
250 /* Send reply */
251
252 skb_out = genlmsg_new(NLMSG_GOODSIZE, GFP_KERNEL);
253 if (!skb_out) {
254 res = -ENOMEM;
255 goto fail;
256 }
257
258 msg_head = genlmsg_put(skb_out, NETLINK_CB(skb_in).portid,
259 info->snd_seq, &hsr_genl_family, 0,
260 HSR_C_SET_NODE_STATUS);
261 if (!msg_head) {
262 res = -ENOMEM;
263 goto nla_put_failure;
264 }
265
266 res = nla_put_u32(skb_out, HSR_A_IFINDEX, hsr_dev->ifindex);
267 if (res < 0)
268 goto nla_put_failure;
269
270 hsr_priv = netdev_priv(hsr_dev);
271 res = hsr_get_node_data(hsr_priv,
272 (unsigned char *) nla_data(info->attrs[HSR_A_NODE_ADDR]),
273 hsr_node_addr_b,
274 &addr_b_ifindex,
275 &hsr_node_if1_age,
276 &hsr_node_if1_seq,
277 &hsr_node_if2_age,
278 &hsr_node_if2_seq);
279 if (res < 0)
280 goto nla_put_failure;
281
282 res = nla_put(skb_out, HSR_A_NODE_ADDR, ETH_ALEN,
283 nla_data(info->attrs[HSR_A_NODE_ADDR]));
284 if (res < 0)
285 goto nla_put_failure;
286
287 if (addr_b_ifindex > -1) {
288 res = nla_put(skb_out, HSR_A_NODE_ADDR_B, ETH_ALEN,
289 hsr_node_addr_b);
290 if (res < 0)
291 goto nla_put_failure;
292
293 res = nla_put_u32(skb_out, HSR_A_ADDR_B_IFINDEX, addr_b_ifindex);
294 if (res < 0)
295 goto nla_put_failure;
296 }
297
298 res = nla_put_u32(skb_out, HSR_A_IF1_AGE, hsr_node_if1_age);
299 if (res < 0)
300 goto nla_put_failure;
301 res = nla_put_u16(skb_out, HSR_A_IF1_SEQ, hsr_node_if1_seq);
302 if (res < 0)
303 goto nla_put_failure;
304 if (hsr_priv->slave[0])
305 res = nla_put_u32(skb_out, HSR_A_IF1_IFINDEX,
306 hsr_priv->slave[0]->ifindex);
307 if (res < 0)
308 goto nla_put_failure;
309
310 res = nla_put_u32(skb_out, HSR_A_IF2_AGE, hsr_node_if2_age);
311 if (res < 0)
312 goto nla_put_failure;
313 res = nla_put_u16(skb_out, HSR_A_IF2_SEQ, hsr_node_if2_seq);
314 if (res < 0)
315 goto nla_put_failure;
316 if (hsr_priv->slave[1])
317 res = nla_put_u32(skb_out, HSR_A_IF2_IFINDEX,
318 hsr_priv->slave[1]->ifindex);
319
320 genlmsg_end(skb_out, msg_head);
321 genlmsg_unicast(genl_info_net(info), skb_out, info->snd_portid);
322
323 return 0;
324
325invalid:
326 netlink_ack(skb_in, nlmsg_hdr(skb_in), -EINVAL);
327 return 0;
328
329nla_put_failure:
330 kfree_skb(skb_out);
331 /* Fall through */
332
333fail:
334 return res;
335}
336
337/* Get a list of MacAddressA of all nodes known to this node (other than self).
338 */
339static int hsr_get_node_list(struct sk_buff *skb_in, struct genl_info *info)
340{
341 /* For receiving */
342 struct nlattr *na;
343 struct net_device *hsr_dev;
344
345 /* For sending */
346 struct sk_buff *skb_out;
347 void *msg_head;
348 struct hsr_priv *hsr_priv;
349 void *pos;
350 unsigned char addr[ETH_ALEN];
351 int res;
352
353 if (!info)
354 goto invalid;
355
356 na = info->attrs[HSR_A_IFINDEX];
357 if (!na)
358 goto invalid;
359
360 hsr_dev = __dev_get_by_index(genl_info_net(info),
361 nla_get_u32(info->attrs[HSR_A_IFINDEX]));
362 if (!hsr_dev)
363 goto invalid;
364 if (!is_hsr_master(hsr_dev))
365 goto invalid;
366
367
368 /* Send reply */
369
370 skb_out = genlmsg_new(NLMSG_GOODSIZE, GFP_KERNEL);
371 if (!skb_out) {
372 res = -ENOMEM;
373 goto fail;
374 }
375
376 msg_head = genlmsg_put(skb_out, NETLINK_CB(skb_in).portid,
377 info->snd_seq, &hsr_genl_family, 0,
378 HSR_C_SET_NODE_LIST);
379 if (!msg_head) {
380 res = -ENOMEM;
381 goto nla_put_failure;
382 }
383
384 res = nla_put_u32(skb_out, HSR_A_IFINDEX, hsr_dev->ifindex);
385 if (res < 0)
386 goto nla_put_failure;
387
388 hsr_priv = netdev_priv(hsr_dev);
389
390 rcu_read_lock();
391 pos = hsr_get_next_node(hsr_priv, NULL, addr);
392 while (pos) {
393 res = nla_put(skb_out, HSR_A_NODE_ADDR, ETH_ALEN, addr);
394 if (res < 0) {
395 rcu_read_unlock();
396 goto nla_put_failure;
397 }
398 pos = hsr_get_next_node(hsr_priv, pos, addr);
399 }
400 rcu_read_unlock();
401
402 genlmsg_end(skb_out, msg_head);
403 genlmsg_unicast(genl_info_net(info), skb_out, info->snd_portid);
404
405 return 0;
406
407invalid:
408 netlink_ack(skb_in, nlmsg_hdr(skb_in), -EINVAL);
409 return 0;
410
411nla_put_failure:
412 kfree_skb(skb_out);
413 /* Fall through */
414
415fail:
416 return res;
417}
418
419
420static const struct genl_ops hsr_ops[] = {
421 {
422 .cmd = HSR_C_GET_NODE_STATUS,
423 .flags = 0,
424 .policy = hsr_genl_policy,
425 .doit = hsr_get_node_status,
426 .dumpit = NULL,
427 },
428 {
429 .cmd = HSR_C_GET_NODE_LIST,
430 .flags = 0,
431 .policy = hsr_genl_policy,
432 .doit = hsr_get_node_list,
433 .dumpit = NULL,
434 },
435};
436
437int __init hsr_netlink_init(void)
438{
439 int rc;
440
441 rc = rtnl_link_register(&hsr_link_ops);
442 if (rc)
443 goto fail_rtnl_link_register;
444
445 rc = genl_register_family_with_ops_groups(&hsr_genl_family, hsr_ops,
446 hsr_mcgrps);
447 if (rc)
448 goto fail_genl_register_family;
449
450 return 0;
451
452fail_genl_register_family:
453 rtnl_link_unregister(&hsr_link_ops);
454fail_rtnl_link_register:
455
456 return rc;
457}
458
459void __exit hsr_netlink_exit(void)
460{
461 genl_unregister_family(&hsr_genl_family);
462 rtnl_link_unregister(&hsr_link_ops);
463}
464
465MODULE_ALIAS_RTNL_LINK("hsr");