1 // SPDX-License-Identifier: GPL-2.0
2 /* Copyright 2011-2014 Autronica Fire and Security AS
7 * Frame router for HSR and PRP.
10 #include "hsr_forward.h"
11 #include <linux/types.h>
12 #include <linux/skbuff.h>
13 #include <linux/etherdevice.h>
14 #include <linux/if_vlan.h>
16 #include "hsr_framereg.h"
20 /* The uses I can see for these HSR supervision frames are:
21 * 1) Use the frames that are sent after node initialization ("HSR_TLV.Type =
22 * 22") to reset any sequence_nr counters belonging to that node. Useful if
23 * the other node's counter has been reset for some reason.
25 * Or not - resetting the counter and bridging the frame would create a
26 * loop, unfortunately.
28 * 2) Use the LifeCheck frames to detect ring breaks. I.e. if no LifeCheck
29 * frame is received from a particular node, we know something is wrong.
30 * We just register these (as with normal frames) and throw them away.
32 * 3) Allow different MAC addresses for the two slave interfaces, using the
35 static bool is_supervision_frame(struct hsr_priv *hsr, struct sk_buff *skb)
37 struct ethhdr *eth_hdr;
38 struct hsr_sup_tag *hsr_sup_tag;
39 struct hsrv1_ethhdr_sp *hsr_V1_hdr;
40 struct hsr_sup_tlv *hsr_sup_tlv;
43 WARN_ON_ONCE(!skb_mac_header_was_set(skb));
44 eth_hdr = (struct ethhdr *)skb_mac_header(skb);
47 if (!ether_addr_equal(eth_hdr->h_dest,
48 hsr->sup_multicast_addr))
51 /* Correct ether type?. */
52 if (!(eth_hdr->h_proto == htons(ETH_P_PRP) ||
53 eth_hdr->h_proto == htons(ETH_P_HSR)))
56 /* Get the supervision header from correct location. */
57 if (eth_hdr->h_proto == htons(ETH_P_HSR)) { /* Okay HSRv1. */
58 total_length = sizeof(struct hsrv1_ethhdr_sp);
59 if (!pskb_may_pull(skb, total_length))
62 hsr_V1_hdr = (struct hsrv1_ethhdr_sp *)skb_mac_header(skb);
63 if (hsr_V1_hdr->hsr.encap_proto != htons(ETH_P_PRP))
66 hsr_sup_tag = &hsr_V1_hdr->hsr_sup;
68 total_length = sizeof(struct hsrv0_ethhdr_sp);
69 if (!pskb_may_pull(skb, total_length))
73 &((struct hsrv0_ethhdr_sp *)skb_mac_header(skb))->hsr_sup;
76 if (hsr_sup_tag->tlv.HSR_TLV_type != HSR_TLV_ANNOUNCE &&
77 hsr_sup_tag->tlv.HSR_TLV_type != HSR_TLV_LIFE_CHECK &&
78 hsr_sup_tag->tlv.HSR_TLV_type != PRP_TLV_LIFE_CHECK_DD &&
79 hsr_sup_tag->tlv.HSR_TLV_type != PRP_TLV_LIFE_CHECK_DA)
81 if (hsr_sup_tag->tlv.HSR_TLV_length != 12 &&
82 hsr_sup_tag->tlv.HSR_TLV_length != sizeof(struct hsr_sup_payload))
86 total_length += hsr_sup_tag->tlv.HSR_TLV_length;
87 if (!pskb_may_pull(skb, total_length))
89 skb_pull(skb, total_length);
90 hsr_sup_tlv = (struct hsr_sup_tlv *)skb->data;
91 skb_push(skb, total_length);
93 /* if this is a redbox supervision frame we need to verify
94 * that more data is available
96 if (hsr_sup_tlv->HSR_TLV_type == PRP_TLV_REDBOX_MAC) {
97 /* tlv length must be a length of a mac address */
98 if (hsr_sup_tlv->HSR_TLV_length != sizeof(struct hsr_sup_payload))
101 /* make sure another tlv follows */
102 total_length += sizeof(struct hsr_sup_tlv) + hsr_sup_tlv->HSR_TLV_length;
103 if (!pskb_may_pull(skb, total_length))
107 skb_pull(skb, total_length);
108 hsr_sup_tlv = (struct hsr_sup_tlv *)skb->data;
109 skb_push(skb, total_length);
112 /* end of tlvs must follow at the end */
113 if (hsr_sup_tlv->HSR_TLV_type == HSR_TLV_EOT &&
114 hsr_sup_tlv->HSR_TLV_length != 0)
120 static struct sk_buff *create_stripped_skb_hsr(struct sk_buff *skb_in,
121 struct hsr_frame_info *frame)
125 unsigned char *dst, *src;
127 skb_pull(skb_in, HSR_HLEN);
128 skb = __pskb_copy(skb_in, skb_headroom(skb_in) - HSR_HLEN, GFP_ATOMIC);
129 skb_push(skb_in, HSR_HLEN);
133 skb_reset_mac_header(skb);
135 if (skb->ip_summed == CHECKSUM_PARTIAL)
136 skb->csum_start -= HSR_HLEN;
138 copylen = 2 * ETH_ALEN;
140 copylen += VLAN_HLEN;
141 src = skb_mac_header(skb_in);
142 dst = skb_mac_header(skb);
143 memcpy(dst, src, copylen);
145 skb->protocol = eth_hdr(skb)->h_proto;
149 struct sk_buff *hsr_get_untagged_frame(struct hsr_frame_info *frame,
150 struct hsr_port *port)
152 if (!frame->skb_std) {
155 create_stripped_skb_hsr(frame->skb_hsr, frame);
157 netdev_warn_once(port->dev,
158 "Unexpected frame received in hsr_get_untagged_frame()\n");
164 return skb_clone(frame->skb_std, GFP_ATOMIC);
167 struct sk_buff *prp_get_untagged_frame(struct hsr_frame_info *frame,
168 struct hsr_port *port)
170 if (!frame->skb_std) {
171 if (frame->skb_prp) {
172 /* trim the skb by len - HSR_HLEN to exclude RCT */
173 skb_trim(frame->skb_prp,
174 frame->skb_prp->len - HSR_HLEN);
176 __pskb_copy(frame->skb_prp,
177 skb_headroom(frame->skb_prp),
181 WARN_ONCE(1, "%s:%d: Unexpected frame received (port_src %s)\n",
182 __FILE__, __LINE__, port->dev->name);
187 return skb_clone(frame->skb_std, GFP_ATOMIC);
190 static void prp_set_lan_id(struct prp_rct *trailer,
191 struct hsr_port *port)
195 if (port->type == HSR_PT_SLAVE_A)
200 /* Add net_id in the upper 3 bits of lane_id */
201 lane_id |= port->hsr->net_id;
202 set_prp_lan_id(trailer, lane_id);
205 /* Tailroom for PRP rct should have been created before calling this */
206 static struct sk_buff *prp_fill_rct(struct sk_buff *skb,
207 struct hsr_frame_info *frame,
208 struct hsr_port *port)
210 struct prp_rct *trailer;
211 int min_size = ETH_ZLEN;
218 min_size = VLAN_ETH_ZLEN;
220 if (skb_put_padto(skb, min_size))
223 trailer = (struct prp_rct *)skb_put(skb, HSR_HLEN);
224 lsdu_size = skb->len - 14;
227 prp_set_lan_id(trailer, port);
228 set_prp_LSDU_size(trailer, lsdu_size);
229 trailer->sequence_nr = htons(frame->sequence_nr);
230 trailer->PRP_suffix = htons(ETH_P_PRP);
231 skb->protocol = eth_hdr(skb)->h_proto;
236 static void hsr_set_path_id(struct hsr_ethhdr *hsr_ethhdr,
237 struct hsr_port *port)
241 if (port->type == HSR_PT_SLAVE_A)
246 set_hsr_tag_path(&hsr_ethhdr->hsr_tag, path_id);
249 static struct sk_buff *hsr_fill_tag(struct sk_buff *skb,
250 struct hsr_frame_info *frame,
251 struct hsr_port *port, u8 proto_version)
253 struct hsr_ethhdr *hsr_ethhdr;
256 /* pad to minimum packet size which is 60 + 6 (HSR tag) */
257 if (skb_put_padto(skb, ETH_ZLEN + HSR_HLEN))
260 lsdu_size = skb->len - 14;
264 hsr_ethhdr = (struct hsr_ethhdr *)skb_mac_header(skb);
266 hsr_set_path_id(hsr_ethhdr, port);
267 set_hsr_tag_LSDU_size(&hsr_ethhdr->hsr_tag, lsdu_size);
268 hsr_ethhdr->hsr_tag.sequence_nr = htons(frame->sequence_nr);
269 hsr_ethhdr->hsr_tag.encap_proto = hsr_ethhdr->ethhdr.h_proto;
270 hsr_ethhdr->ethhdr.h_proto = htons(proto_version ?
271 ETH_P_HSR : ETH_P_PRP);
272 skb->protocol = hsr_ethhdr->ethhdr.h_proto;
277 /* If the original frame was an HSR tagged frame, just clone it to be sent
278 * unchanged. Otherwise, create a private frame especially tagged for 'port'.
280 struct sk_buff *hsr_create_tagged_frame(struct hsr_frame_info *frame,
281 struct hsr_port *port)
283 unsigned char *dst, *src;
287 if (frame->skb_hsr) {
288 struct hsr_ethhdr *hsr_ethhdr =
289 (struct hsr_ethhdr *)skb_mac_header(frame->skb_hsr);
291 /* set the lane id properly */
292 hsr_set_path_id(hsr_ethhdr, port);
293 return skb_clone(frame->skb_hsr, GFP_ATOMIC);
294 } else if (port->dev->features & NETIF_F_HW_HSR_TAG_INS) {
295 return skb_clone(frame->skb_std, GFP_ATOMIC);
298 /* Create the new skb with enough headroom to fit the HSR tag */
299 skb = __pskb_copy(frame->skb_std,
300 skb_headroom(frame->skb_std) + HSR_HLEN, GFP_ATOMIC);
303 skb_reset_mac_header(skb);
305 if (skb->ip_summed == CHECKSUM_PARTIAL)
306 skb->csum_start += HSR_HLEN;
310 movelen += VLAN_HLEN;
312 src = skb_mac_header(skb);
313 dst = skb_push(skb, HSR_HLEN);
314 memmove(dst, src, movelen);
315 skb_reset_mac_header(skb);
317 /* skb_put_padto free skb on error and hsr_fill_tag returns NULL in
320 return hsr_fill_tag(skb, frame, port, port->hsr->prot_version);
323 struct sk_buff *prp_create_tagged_frame(struct hsr_frame_info *frame,
324 struct hsr_port *port)
328 if (frame->skb_prp) {
329 struct prp_rct *trailer = skb_get_PRP_rct(frame->skb_prp);
332 prp_set_lan_id(trailer, port);
334 WARN_ONCE(!trailer, "errored PRP skb");
337 return skb_clone(frame->skb_prp, GFP_ATOMIC);
338 } else if (port->dev->features & NETIF_F_HW_HSR_TAG_INS) {
339 return skb_clone(frame->skb_std, GFP_ATOMIC);
342 skb = skb_copy_expand(frame->skb_std, 0,
343 skb_tailroom(frame->skb_std) + HSR_HLEN,
345 return prp_fill_rct(skb, frame, port);
348 static void hsr_deliver_master(struct sk_buff *skb, struct net_device *dev,
349 struct hsr_node *node_src)
351 bool was_multicast_frame;
354 was_multicast_frame = (skb->pkt_type == PACKET_MULTICAST);
355 hsr_addr_subst_source(node_src, skb);
356 skb_pull(skb, ETH_HLEN);
359 if (res == NET_RX_DROP) {
360 dev->stats.rx_dropped++;
362 dev->stats.rx_packets++;
363 dev->stats.rx_bytes += recv_len;
364 if (was_multicast_frame)
365 dev->stats.multicast++;
369 static int hsr_xmit(struct sk_buff *skb, struct hsr_port *port,
370 struct hsr_frame_info *frame)
372 if (frame->port_rcv->type == HSR_PT_MASTER) {
373 hsr_addr_subst_dest(frame->node_src, skb, port);
375 /* Address substitution (IEC62439-3 pp 26, 50): replace mac
376 * address of outgoing frame with that of the outgoing slave's.
378 ether_addr_copy(eth_hdr(skb)->h_source, port->dev->dev_addr);
381 /* When HSR node is used as RedBox - the frame received from HSR ring
382 * requires source MAC address (SA) replacement to one which can be
383 * recognized by SAN devices (otherwise, frames are dropped by switch)
385 if (port->type == HSR_PT_INTERLINK)
386 ether_addr_copy(eth_hdr(skb)->h_source,
387 port->hsr->macaddress_redbox);
389 return dev_queue_xmit(skb);
392 bool prp_drop_frame(struct hsr_frame_info *frame, struct hsr_port *port)
394 return ((frame->port_rcv->type == HSR_PT_SLAVE_A &&
395 port->type == HSR_PT_SLAVE_B) ||
396 (frame->port_rcv->type == HSR_PT_SLAVE_B &&
397 port->type == HSR_PT_SLAVE_A));
400 bool hsr_drop_frame(struct hsr_frame_info *frame, struct hsr_port *port)
404 if (port->dev->features & NETIF_F_HW_HSR_FWD)
405 return prp_drop_frame(frame, port);
407 /* RedBox specific frames dropping policies
409 * Do not send HSR supervisory frames to SAN devices
411 if (frame->is_supervision && port->type == HSR_PT_INTERLINK)
414 /* Do not forward to other HSR port (A or B) unicast frames which
415 * are addressed to interlink port (and are in the ProxyNodeTable).
417 skb = frame->skb_hsr;
418 if (skb && prp_drop_frame(frame, port) &&
419 is_unicast_ether_addr(eth_hdr(skb)->h_dest) &&
420 hsr_is_node_in_db(&port->hsr->proxy_node_db,
421 eth_hdr(skb)->h_dest)) {
425 /* Do not forward to port C (Interlink) frames from nodes A and B
426 * if DA is in NodeTable.
428 if ((frame->port_rcv->type == HSR_PT_SLAVE_A ||
429 frame->port_rcv->type == HSR_PT_SLAVE_B) &&
430 port->type == HSR_PT_INTERLINK) {
431 skb = frame->skb_hsr;
432 if (skb && is_unicast_ether_addr(eth_hdr(skb)->h_dest) &&
433 hsr_is_node_in_db(&port->hsr->node_db,
434 eth_hdr(skb)->h_dest)) {
439 /* Do not forward to port A and B unicast frames received on the
440 * interlink port if it is addressed to one of nodes registered in
441 * the ProxyNodeTable.
443 if ((port->type == HSR_PT_SLAVE_A || port->type == HSR_PT_SLAVE_B) &&
444 frame->port_rcv->type == HSR_PT_INTERLINK) {
445 skb = frame->skb_std;
446 if (skb && is_unicast_ether_addr(eth_hdr(skb)->h_dest) &&
447 hsr_is_node_in_db(&port->hsr->proxy_node_db,
448 eth_hdr(skb)->h_dest)) {
456 /* Forward the frame through all devices except:
457 * - Back through the receiving device
458 * - If it's a HSR frame: through a device where it has passed before
459 * - if it's a PRP frame: through another PRP slave device (no bridge)
460 * - To the local HSR master only if the frame is directly addressed to it, or
461 * a non-supervision multicast or broadcast frame.
463 * HSR slave devices should insert a HSR tag into the frame, or forward the
464 * frame unchanged if it's already tagged. Interlink devices should strip HSR
465 * tags if they're of the non-HSR type (but only after duplicate discard). The
466 * master device always strips HSR tags.
468 static void hsr_forward_do(struct hsr_frame_info *frame)
470 struct hsr_port *port;
474 hsr_for_each_port(frame->port_rcv->hsr, port) {
475 struct hsr_priv *hsr = port->hsr;
476 /* Don't send frame back the way it came */
477 if (port == frame->port_rcv)
480 /* Don't deliver locally unless we should */
481 if (port->type == HSR_PT_MASTER && !frame->is_local_dest)
484 /* Deliver frames directly addressed to us to master only */
485 if (port->type != HSR_PT_MASTER && frame->is_local_exclusive)
488 /* If hardware duplicate generation is enabled, only send out
491 if ((port->dev->features & NETIF_F_HW_HSR_DUP) && sent)
494 /* Don't send frame over port where it has been sent before.
495 * Also for SAN, this shouldn't be done.
497 if (!frame->is_from_san &&
498 hsr_register_frame_out(port, frame->node_src,
502 if (frame->is_supervision && port->type == HSR_PT_MASTER) {
503 hsr_handle_sup_frame(frame);
507 /* Check if frame is to be dropped. Eg. for PRP no forward
508 * between ports, or sending HSR supervision to RedBox.
510 if (hsr->proto_ops->drop_frame &&
511 hsr->proto_ops->drop_frame(frame, port))
514 if (port->type == HSR_PT_SLAVE_A ||
515 port->type == HSR_PT_SLAVE_B)
516 skb = hsr->proto_ops->create_tagged_frame(frame, port);
518 skb = hsr->proto_ops->get_untagged_frame(frame, port);
521 frame->port_rcv->dev->stats.rx_dropped++;
525 skb->dev = port->dev;
526 if (port->type == HSR_PT_MASTER) {
527 hsr_deliver_master(skb, port->dev, frame->node_src);
529 if (!hsr_xmit(skb, port, frame))
530 if (port->type == HSR_PT_SLAVE_A ||
531 port->type == HSR_PT_SLAVE_B)
537 static void check_local_dest(struct hsr_priv *hsr, struct sk_buff *skb,
538 struct hsr_frame_info *frame)
540 if (hsr_addr_is_self(hsr, eth_hdr(skb)->h_dest)) {
541 frame->is_local_exclusive = true;
542 skb->pkt_type = PACKET_HOST;
544 frame->is_local_exclusive = false;
547 if (skb->pkt_type == PACKET_HOST ||
548 skb->pkt_type == PACKET_MULTICAST ||
549 skb->pkt_type == PACKET_BROADCAST) {
550 frame->is_local_dest = true;
552 frame->is_local_dest = false;
556 static void handle_std_frame(struct sk_buff *skb,
557 struct hsr_frame_info *frame)
559 struct hsr_port *port = frame->port_rcv;
560 struct hsr_priv *hsr = port->hsr;
562 frame->skb_hsr = NULL;
563 frame->skb_prp = NULL;
564 frame->skb_std = skb;
566 if (port->type != HSR_PT_MASTER)
567 frame->is_from_san = true;
569 if (port->type == HSR_PT_MASTER ||
570 port->type == HSR_PT_INTERLINK) {
571 /* Sequence nr for the master/interlink node */
572 lockdep_assert_held(&hsr->seqnr_lock);
573 frame->sequence_nr = hsr->sequence_nr;
578 int hsr_fill_frame_info(__be16 proto, struct sk_buff *skb,
579 struct hsr_frame_info *frame)
581 struct hsr_port *port = frame->port_rcv;
582 struct hsr_priv *hsr = port->hsr;
584 /* HSRv0 supervisory frames double as a tag so treat them as tagged. */
585 if ((!hsr->prot_version && proto == htons(ETH_P_PRP)) ||
586 proto == htons(ETH_P_HSR)) {
587 /* Check if skb contains hsr_ethhdr */
588 if (skb->mac_len < sizeof(struct hsr_ethhdr))
591 /* HSR tagged frame :- Data or Supervision */
592 frame->skb_std = NULL;
593 frame->skb_prp = NULL;
594 frame->skb_hsr = skb;
595 frame->sequence_nr = hsr_get_skb_sequence_nr(skb);
599 /* Standard frame or PRP from master port */
600 handle_std_frame(skb, frame);
605 int prp_fill_frame_info(__be16 proto, struct sk_buff *skb,
606 struct hsr_frame_info *frame)
608 /* Supervision frame */
609 struct prp_rct *rct = skb_get_PRP_rct(skb);
612 prp_check_lsdu_size(skb, rct, frame->is_supervision)) {
613 frame->skb_hsr = NULL;
614 frame->skb_std = NULL;
615 frame->skb_prp = skb;
616 frame->sequence_nr = prp_get_skb_sequence_nr(rct);
619 handle_std_frame(skb, frame);
624 static int fill_frame_info(struct hsr_frame_info *frame,
625 struct sk_buff *skb, struct hsr_port *port)
627 struct hsr_priv *hsr = port->hsr;
628 struct hsr_vlan_ethhdr *vlan_hdr;
629 struct list_head *n_db;
630 struct ethhdr *ethhdr;
634 /* Check if skb contains ethhdr */
635 if (skb->mac_len < sizeof(struct ethhdr))
638 memset(frame, 0, sizeof(*frame));
639 frame->is_supervision = is_supervision_frame(port->hsr, skb);
641 n_db = &hsr->node_db;
642 if (port->type == HSR_PT_INTERLINK)
643 n_db = &hsr->proxy_node_db;
645 frame->node_src = hsr_get_node(port, n_db, skb,
646 frame->is_supervision, port->type);
647 if (!frame->node_src)
648 return -1; /* Unknown node and !is_supervision, or no mem */
650 ethhdr = (struct ethhdr *)skb_mac_header(skb);
651 frame->is_vlan = false;
652 proto = ethhdr->h_proto;
654 if (proto == htons(ETH_P_8021Q))
655 frame->is_vlan = true;
657 if (frame->is_vlan) {
658 vlan_hdr = (struct hsr_vlan_ethhdr *)ethhdr;
659 proto = vlan_hdr->vlanhdr.h_vlan_encapsulated_proto;
661 netdev_warn_once(skb->dev, "VLAN not yet supported");
665 frame->is_from_san = false;
666 frame->port_rcv = port;
667 ret = hsr->proto_ops->fill_frame_info(proto, skb, frame);
671 check_local_dest(port->hsr, skb, frame);
676 /* Must be called holding rcu read lock (because of the port parameter) */
677 void hsr_forward_skb(struct sk_buff *skb, struct hsr_port *port)
679 struct hsr_frame_info frame;
682 if (fill_frame_info(&frame, skb, port) < 0)
685 hsr_register_frame_in(frame.node_src, port, frame.sequence_nr);
686 hsr_forward_do(&frame);
688 /* Gets called for ingress frames as well as egress from master port.
689 * So check and increment stats for master port only here.
691 if (port->type == HSR_PT_MASTER) {
692 port->dev->stats.tx_packets++;
693 port->dev->stats.tx_bytes += skb->len;
696 kfree_skb(frame.skb_hsr);
697 kfree_skb(frame.skb_prp);
698 kfree_skb(frame.skb_std);
703 port->dev->stats.tx_dropped++;