1 // SPDX-License-Identifier: GPL-2.0
2 /* Copyright 2011-2014 Autronica Fire and Security AS
3 *
4 * Author(s):
5 * 2011-2014 Arvid Brodin, arvid.brodin@alten.se
6 *
7 * Frame router for HSR and PRP.
8 */
9
10 #include "hsr_forward.h"
11 #include <linux/types.h>
12 #include <linux/skbuff.h>
13 #include <linux/etherdevice.h>
14 #include <linux/if_vlan.h>
15 #include "hsr_main.h"
16 #include "hsr_framereg.h"
17
18 struct hsr_node;
19
20 /* The uses I can see for these HSR supervision frames are:
21 * 1) Use the frames that are sent after node initialization ("HSR_TLV.Type =
22 * 22") to reset any sequence_nr counters belonging to that node. Useful if
23 * the other node's counter has been reset for some reason.
24 * --
25 * Or not - resetting the counter and bridging the frame would create a
26 * loop, unfortunately.
27 *
28 * 2) Use the LifeCheck frames to detect ring breaks. I.e. if no LifeCheck
29 * frame is received from a particular node, we know something is wrong.
30 * We just register these (as with normal frames) and throw them away.
31 *
32 * 3) Allow different MAC addresses for the two slave interfaces, using the
33 * MacAddressA field.
34 */
is_supervision_frame(struct hsr_priv * hsr,struct sk_buff * skb)35 static bool is_supervision_frame(struct hsr_priv *hsr, struct sk_buff *skb)
36 {
37 struct ethhdr *eth_hdr;
38 struct hsr_sup_tag *hsr_sup_tag;
39 struct hsrv1_ethhdr_sp *hsr_V1_hdr;
40 struct hsr_sup_tlv *hsr_sup_tlv;
41 u16 total_length = 0;
42
43 WARN_ON_ONCE(!skb_mac_header_was_set(skb));
44 eth_hdr = (struct ethhdr *)skb_mac_header(skb);
45
46 /* Correct addr? */
47 if (!ether_addr_equal(eth_hdr->h_dest,
48 hsr->sup_multicast_addr))
49 return false;
50
51 /* Correct ether type?. */
52 if (!(eth_hdr->h_proto == htons(ETH_P_PRP) ||
53 eth_hdr->h_proto == htons(ETH_P_HSR)))
54 return false;
55
56 /* Get the supervision header from correct location. */
57 if (eth_hdr->h_proto == htons(ETH_P_HSR)) { /* Okay HSRv1. */
58 total_length = sizeof(struct hsrv1_ethhdr_sp);
59 if (!pskb_may_pull(skb, total_length))
60 return false;
61
62 hsr_V1_hdr = (struct hsrv1_ethhdr_sp *)skb_mac_header(skb);
63 if (hsr_V1_hdr->hsr.encap_proto != htons(ETH_P_PRP))
64 return false;
65
66 hsr_sup_tag = &hsr_V1_hdr->hsr_sup;
67 } else {
68 total_length = sizeof(struct hsrv0_ethhdr_sp);
69 if (!pskb_may_pull(skb, total_length))
70 return false;
71
72 hsr_sup_tag =
73 &((struct hsrv0_ethhdr_sp *)skb_mac_header(skb))->hsr_sup;
74 }
75
76 if (hsr_sup_tag->tlv.HSR_TLV_type != HSR_TLV_ANNOUNCE &&
77 hsr_sup_tag->tlv.HSR_TLV_type != HSR_TLV_LIFE_CHECK &&
78 hsr_sup_tag->tlv.HSR_TLV_type != PRP_TLV_LIFE_CHECK_DD &&
79 hsr_sup_tag->tlv.HSR_TLV_type != PRP_TLV_LIFE_CHECK_DA)
80 return false;
81 if (hsr_sup_tag->tlv.HSR_TLV_length != 12 &&
82 hsr_sup_tag->tlv.HSR_TLV_length != sizeof(struct hsr_sup_payload))
83 return false;
84
85 /* Get next tlv */
86 total_length += hsr_sup_tag->tlv.HSR_TLV_length;
87 if (!pskb_may_pull(skb, total_length))
88 return false;
89 skb_pull(skb, total_length);
90 hsr_sup_tlv = (struct hsr_sup_tlv *)skb->data;
91 skb_push(skb, total_length);
92
93 /* if this is a redbox supervision frame we need to verify
94 * that more data is available
95 */
96 if (hsr_sup_tlv->HSR_TLV_type == PRP_TLV_REDBOX_MAC) {
97 /* tlv length must be a length of a mac address */
98 if (hsr_sup_tlv->HSR_TLV_length != sizeof(struct hsr_sup_payload))
99 return false;
100
101 /* make sure another tlv follows */
102 total_length += sizeof(struct hsr_sup_tlv) + hsr_sup_tlv->HSR_TLV_length;
103 if (!pskb_may_pull(skb, total_length))
104 return false;
105
106 /* get next tlv */
107 skb_pull(skb, total_length);
108 hsr_sup_tlv = (struct hsr_sup_tlv *)skb->data;
109 skb_push(skb, total_length);
110 }
111
112 /* end of tlvs must follow at the end */
113 if (hsr_sup_tlv->HSR_TLV_type == HSR_TLV_EOT &&
114 hsr_sup_tlv->HSR_TLV_length != 0)
115 return false;
116
117 return true;
118 }
119
is_proxy_supervision_frame(struct hsr_priv * hsr,struct sk_buff * skb)120 static bool is_proxy_supervision_frame(struct hsr_priv *hsr,
121 struct sk_buff *skb)
122 {
123 struct hsr_sup_payload *payload;
124 struct ethhdr *eth_hdr;
125 u16 total_length = 0;
126
127 eth_hdr = (struct ethhdr *)skb_mac_header(skb);
128
129 /* Get the HSR protocol revision. */
130 if (eth_hdr->h_proto == htons(ETH_P_HSR))
131 total_length = sizeof(struct hsrv1_ethhdr_sp);
132 else
133 total_length = sizeof(struct hsrv0_ethhdr_sp);
134
135 if (!pskb_may_pull(skb, total_length + sizeof(struct hsr_sup_payload)))
136 return false;
137
138 skb_pull(skb, total_length);
139 payload = (struct hsr_sup_payload *)skb->data;
140 skb_push(skb, total_length);
141
142 /* For RedBox (HSR-SAN) check if we have received the supervision
143 * frame with MAC addresses from own ProxyNodeTable.
144 */
145 return hsr_is_node_in_db(&hsr->proxy_node_db,
146 payload->macaddress_A);
147 }
148
create_stripped_skb_hsr(struct sk_buff * skb_in,struct hsr_frame_info * frame)149 static struct sk_buff *create_stripped_skb_hsr(struct sk_buff *skb_in,
150 struct hsr_frame_info *frame)
151 {
152 struct sk_buff *skb;
153 int copylen;
154 unsigned char *dst, *src;
155
156 skb_pull(skb_in, HSR_HLEN);
157 skb = __pskb_copy(skb_in, skb_headroom(skb_in) - HSR_HLEN, GFP_ATOMIC);
158 skb_push(skb_in, HSR_HLEN);
159 if (!skb)
160 return NULL;
161
162 skb_reset_mac_header(skb);
163
164 if (skb->ip_summed == CHECKSUM_PARTIAL)
165 skb->csum_start -= HSR_HLEN;
166
167 copylen = 2 * ETH_ALEN;
168 if (frame->is_vlan)
169 copylen += VLAN_HLEN;
170 src = skb_mac_header(skb_in);
171 dst = skb_mac_header(skb);
172 memcpy(dst, src, copylen);
173
174 skb->protocol = eth_hdr(skb)->h_proto;
175 return skb;
176 }
177
hsr_get_untagged_frame(struct hsr_frame_info * frame,struct hsr_port * port)178 struct sk_buff *hsr_get_untagged_frame(struct hsr_frame_info *frame,
179 struct hsr_port *port)
180 {
181 if (!frame->skb_std) {
182 if (frame->skb_hsr)
183 frame->skb_std =
184 create_stripped_skb_hsr(frame->skb_hsr, frame);
185 else
186 netdev_warn_once(port->dev,
187 "Unexpected frame received in hsr_get_untagged_frame()\n");
188
189 if (!frame->skb_std)
190 return NULL;
191 }
192
193 return skb_clone(frame->skb_std, GFP_ATOMIC);
194 }
195
prp_get_untagged_frame(struct hsr_frame_info * frame,struct hsr_port * port)196 struct sk_buff *prp_get_untagged_frame(struct hsr_frame_info *frame,
197 struct hsr_port *port)
198 {
199 if (!frame->skb_std) {
200 if (frame->skb_prp) {
201 /* trim the skb by len - HSR_HLEN to exclude RCT */
202 skb_trim(frame->skb_prp,
203 frame->skb_prp->len - HSR_HLEN);
204 frame->skb_std =
205 __pskb_copy(frame->skb_prp,
206 skb_headroom(frame->skb_prp),
207 GFP_ATOMIC);
208 } else {
209 /* Unexpected */
210 WARN_ONCE(1, "%s:%d: Unexpected frame received (port_src %s)\n",
211 __FILE__, __LINE__, port->dev->name);
212 return NULL;
213 }
214 }
215
216 return skb_clone(frame->skb_std, GFP_ATOMIC);
217 }
218
prp_set_lan_id(struct prp_rct * trailer,struct hsr_port * port)219 static void prp_set_lan_id(struct prp_rct *trailer,
220 struct hsr_port *port)
221 {
222 int lane_id;
223
224 if (port->type == HSR_PT_SLAVE_A)
225 lane_id = 0;
226 else
227 lane_id = 1;
228
229 /* Add net_id in the upper 3 bits of lane_id */
230 lane_id |= port->hsr->net_id;
231 set_prp_lan_id(trailer, lane_id);
232 }
233
234 /* Tailroom for PRP rct should have been created before calling this */
prp_fill_rct(struct sk_buff * skb,struct hsr_frame_info * frame,struct hsr_port * port)235 static struct sk_buff *prp_fill_rct(struct sk_buff *skb,
236 struct hsr_frame_info *frame,
237 struct hsr_port *port)
238 {
239 struct prp_rct *trailer;
240 int min_size = ETH_ZLEN;
241 int lsdu_size;
242
243 if (!skb)
244 return skb;
245
246 if (frame->is_vlan)
247 min_size = VLAN_ETH_ZLEN;
248
249 if (skb_put_padto(skb, min_size))
250 return NULL;
251
252 trailer = (struct prp_rct *)skb_put(skb, HSR_HLEN);
253 lsdu_size = skb->len - 14;
254 if (frame->is_vlan)
255 lsdu_size -= 4;
256 prp_set_lan_id(trailer, port);
257 set_prp_LSDU_size(trailer, lsdu_size);
258 trailer->sequence_nr = htons(frame->sequence_nr);
259 trailer->PRP_suffix = htons(ETH_P_PRP);
260 skb->protocol = eth_hdr(skb)->h_proto;
261
262 return skb;
263 }
264
hsr_set_path_id(struct hsr_frame_info * frame,struct hsr_ethhdr * hsr_ethhdr,struct hsr_port * port)265 static void hsr_set_path_id(struct hsr_frame_info *frame,
266 struct hsr_ethhdr *hsr_ethhdr,
267 struct hsr_port *port)
268 {
269 int path_id;
270
271 if (port->hsr->prot_version) {
272 if (port->type == HSR_PT_SLAVE_A)
273 path_id = 0;
274 else
275 path_id = 1;
276 } else {
277 if (frame->is_supervision)
278 path_id = 0xf;
279 else
280 path_id = 1;
281 }
282
283 set_hsr_tag_path(&hsr_ethhdr->hsr_tag, path_id);
284 }
285
hsr_fill_tag(struct sk_buff * skb,struct hsr_frame_info * frame,struct hsr_port * port,u8 proto_version)286 static struct sk_buff *hsr_fill_tag(struct sk_buff *skb,
287 struct hsr_frame_info *frame,
288 struct hsr_port *port, u8 proto_version)
289 {
290 struct hsr_ethhdr *hsr_ethhdr;
291 unsigned char *pc;
292 int lsdu_size;
293
294 /* pad to minimum packet size which is 60 + 6 (HSR tag) */
295 if (skb_put_padto(skb, ETH_ZLEN + HSR_HLEN))
296 return NULL;
297
298 lsdu_size = skb->len - 14;
299 if (frame->is_vlan)
300 lsdu_size -= 4;
301
302 pc = skb_mac_header(skb);
303 if (frame->is_vlan)
304 /* This 4-byte shift (size of a vlan tag) does not
305 * mean that the ethhdr starts there. But rather it
306 * provides the proper environment for accessing
307 * the fields, such as hsr_tag etc., just like
308 * when the vlan tag is not there. This is because
309 * the hsr tag is after the vlan tag.
310 */
311 hsr_ethhdr = (struct hsr_ethhdr *)(pc + VLAN_HLEN);
312 else
313 hsr_ethhdr = (struct hsr_ethhdr *)pc;
314
315 hsr_set_path_id(frame, hsr_ethhdr, port);
316 set_hsr_tag_LSDU_size(&hsr_ethhdr->hsr_tag, lsdu_size);
317 hsr_ethhdr->hsr_tag.sequence_nr = htons(frame->sequence_nr);
318 hsr_ethhdr->hsr_tag.encap_proto = hsr_ethhdr->ethhdr.h_proto;
319 hsr_ethhdr->ethhdr.h_proto = htons(proto_version ?
320 ETH_P_HSR : ETH_P_PRP);
321 skb->protocol = hsr_ethhdr->ethhdr.h_proto;
322
323 return skb;
324 }
325
326 /* If the original frame was an HSR tagged frame, just clone it to be sent
327 * unchanged. Otherwise, create a private frame especially tagged for 'port'.
328 */
hsr_create_tagged_frame(struct hsr_frame_info * frame,struct hsr_port * port)329 struct sk_buff *hsr_create_tagged_frame(struct hsr_frame_info *frame,
330 struct hsr_port *port)
331 {
332 unsigned char *dst, *src;
333 struct sk_buff *skb;
334 int movelen;
335
336 if (frame->skb_hsr) {
337 struct hsr_ethhdr *hsr_ethhdr =
338 (struct hsr_ethhdr *)skb_mac_header(frame->skb_hsr);
339
340 /* set the lane id properly */
341 hsr_set_path_id(frame, hsr_ethhdr, port);
342 return skb_clone(frame->skb_hsr, GFP_ATOMIC);
343 } else if (port->dev->features & NETIF_F_HW_HSR_TAG_INS) {
344 return skb_clone(frame->skb_std, GFP_ATOMIC);
345 }
346
347 /* Create the new skb with enough headroom to fit the HSR tag */
348 skb = __pskb_copy(frame->skb_std,
349 skb_headroom(frame->skb_std) + HSR_HLEN, GFP_ATOMIC);
350 if (!skb)
351 return NULL;
352 skb_reset_mac_header(skb);
353
354 if (skb->ip_summed == CHECKSUM_PARTIAL)
355 skb->csum_start += HSR_HLEN;
356
357 movelen = ETH_HLEN;
358 if (frame->is_vlan)
359 movelen += VLAN_HLEN;
360
361 src = skb_mac_header(skb);
362 dst = skb_push(skb, HSR_HLEN);
363 memmove(dst, src, movelen);
364 skb_reset_mac_header(skb);
365
366 /* skb_put_padto free skb on error and hsr_fill_tag returns NULL in
367 * that case
368 */
369 return hsr_fill_tag(skb, frame, port, port->hsr->prot_version);
370 }
371
prp_create_tagged_frame(struct hsr_frame_info * frame,struct hsr_port * port)372 struct sk_buff *prp_create_tagged_frame(struct hsr_frame_info *frame,
373 struct hsr_port *port)
374 {
375 struct sk_buff *skb;
376
377 if (frame->skb_prp) {
378 struct prp_rct *trailer = skb_get_PRP_rct(frame->skb_prp);
379
380 if (trailer) {
381 prp_set_lan_id(trailer, port);
382 } else {
383 WARN_ONCE(!trailer, "errored PRP skb");
384 return NULL;
385 }
386 return skb_clone(frame->skb_prp, GFP_ATOMIC);
387 } else if (port->dev->features & NETIF_F_HW_HSR_TAG_INS) {
388 return skb_clone(frame->skb_std, GFP_ATOMIC);
389 }
390
391 skb = skb_copy_expand(frame->skb_std, skb_headroom(frame->skb_std),
392 skb_tailroom(frame->skb_std) + HSR_HLEN,
393 GFP_ATOMIC);
394 return prp_fill_rct(skb, frame, port);
395 }
396
hsr_deliver_master(struct sk_buff * skb,struct net_device * dev,struct hsr_node * node_src)397 static void hsr_deliver_master(struct sk_buff *skb, struct net_device *dev,
398 struct hsr_node *node_src)
399 {
400 bool was_multicast_frame;
401 int res, recv_len;
402
403 was_multicast_frame = (skb->pkt_type == PACKET_MULTICAST);
404 hsr_addr_subst_source(node_src, skb);
405 skb_pull(skb, ETH_HLEN);
406 recv_len = skb->len;
407 res = netif_rx(skb);
408 if (res == NET_RX_DROP) {
409 dev->stats.rx_dropped++;
410 } else {
411 dev->stats.rx_packets++;
412 dev->stats.rx_bytes += recv_len;
413 if (was_multicast_frame)
414 dev->stats.multicast++;
415 }
416 }
417
hsr_xmit(struct sk_buff * skb,struct hsr_port * port,struct hsr_frame_info * frame)418 static int hsr_xmit(struct sk_buff *skb, struct hsr_port *port,
419 struct hsr_frame_info *frame)
420 {
421 if (frame->port_rcv->type == HSR_PT_MASTER) {
422 hsr_addr_subst_dest(frame->node_src, skb, port);
423
424 /* Address substitution (IEC62439-3 pp 26, 50): replace mac
425 * address of outgoing frame with that of the outgoing slave's.
426 */
427 ether_addr_copy(eth_hdr(skb)->h_source, port->dev->dev_addr);
428 }
429
430 /* When HSR node is used as RedBox - the frame received from HSR ring
431 * requires source MAC address (SA) replacement to one which can be
432 * recognized by SAN devices (otherwise, frames are dropped by switch)
433 */
434 if (port->type == HSR_PT_INTERLINK)
435 ether_addr_copy(eth_hdr(skb)->h_source,
436 port->hsr->macaddress_redbox);
437
438 return dev_queue_xmit(skb);
439 }
440
prp_drop_frame(struct hsr_frame_info * frame,struct hsr_port * port)441 bool prp_drop_frame(struct hsr_frame_info *frame, struct hsr_port *port)
442 {
443 return ((frame->port_rcv->type == HSR_PT_SLAVE_A &&
444 port->type == HSR_PT_SLAVE_B) ||
445 (frame->port_rcv->type == HSR_PT_SLAVE_B &&
446 port->type == HSR_PT_SLAVE_A));
447 }
448
hsr_drop_frame(struct hsr_frame_info * frame,struct hsr_port * port)449 bool hsr_drop_frame(struct hsr_frame_info *frame, struct hsr_port *port)
450 {
451 struct sk_buff *skb;
452
453 if (port->dev->features & NETIF_F_HW_HSR_FWD)
454 return prp_drop_frame(frame, port);
455
456 /* RedBox specific frames dropping policies
457 *
458 * Do not send HSR supervisory frames to SAN devices
459 */
460 if (frame->is_supervision && port->type == HSR_PT_INTERLINK)
461 return true;
462
463 /* Do not forward to other HSR port (A or B) unicast frames which
464 * are addressed to interlink port (and are in the ProxyNodeTable).
465 */
466 skb = frame->skb_hsr;
467 if (skb && prp_drop_frame(frame, port) &&
468 is_unicast_ether_addr(eth_hdr(skb)->h_dest) &&
469 hsr_is_node_in_db(&port->hsr->proxy_node_db,
470 eth_hdr(skb)->h_dest)) {
471 return true;
472 }
473
474 /* Do not forward to port C (Interlink) frames from nodes A and B
475 * if DA is in NodeTable.
476 */
477 if ((frame->port_rcv->type == HSR_PT_SLAVE_A ||
478 frame->port_rcv->type == HSR_PT_SLAVE_B) &&
479 port->type == HSR_PT_INTERLINK) {
480 skb = frame->skb_hsr;
481 if (skb && is_unicast_ether_addr(eth_hdr(skb)->h_dest) &&
482 hsr_is_node_in_db(&port->hsr->node_db,
483 eth_hdr(skb)->h_dest)) {
484 return true;
485 }
486 }
487
488 /* Do not forward to port A and B unicast frames received on the
489 * interlink port if it is addressed to one of nodes registered in
490 * the ProxyNodeTable.
491 */
492 if ((port->type == HSR_PT_SLAVE_A || port->type == HSR_PT_SLAVE_B) &&
493 frame->port_rcv->type == HSR_PT_INTERLINK) {
494 skb = frame->skb_std;
495 if (skb && is_unicast_ether_addr(eth_hdr(skb)->h_dest) &&
496 hsr_is_node_in_db(&port->hsr->proxy_node_db,
497 eth_hdr(skb)->h_dest)) {
498 return true;
499 }
500 }
501
502 return false;
503 }
504
505 /* Forward the frame through all devices except:
506 * - Back through the receiving device
507 * - If it's a HSR frame: through a device where it has passed before
508 * - if it's a PRP frame: through another PRP slave device (no bridge)
509 * - To the local HSR master only if the frame is directly addressed to it, or
510 * a non-supervision multicast or broadcast frame.
511 *
512 * HSR slave devices should insert a HSR tag into the frame, or forward the
513 * frame unchanged if it's already tagged. Interlink devices should strip HSR
514 * tags if they're of the non-HSR type (but only after duplicate discard). The
515 * master device always strips HSR tags.
516 */
hsr_forward_do(struct hsr_frame_info * frame)517 static void hsr_forward_do(struct hsr_frame_info *frame)
518 {
519 struct hsr_port *port;
520 struct sk_buff *skb;
521 bool sent = false;
522
523 hsr_for_each_port(frame->port_rcv->hsr, port) {
524 struct hsr_priv *hsr = port->hsr;
525 /* Don't send frame back the way it came */
526 if (port == frame->port_rcv)
527 continue;
528
529 /* Don't deliver locally unless we should */
530 if (port->type == HSR_PT_MASTER && !frame->is_local_dest)
531 continue;
532
533 /* Deliver frames directly addressed to us to master only */
534 if (port->type != HSR_PT_MASTER && frame->is_local_exclusive)
535 continue;
536
537 /* If hardware duplicate generation is enabled, only send out
538 * one port.
539 */
540 if ((port->dev->features & NETIF_F_HW_HSR_DUP) && sent)
541 continue;
542
543 /* Don't send frame over port where it has been sent before.
544 * Also for SAN, this shouldn't be done.
545 */
546 if (!frame->is_from_san &&
547 hsr->proto_ops->register_frame_out &&
548 hsr->proto_ops->register_frame_out(port, frame))
549 continue;
550
551 if (frame->is_supervision && port->type == HSR_PT_MASTER &&
552 !frame->is_proxy_supervision) {
553 hsr_handle_sup_frame(frame);
554 continue;
555 }
556
557 /* Check if frame is to be dropped. Eg. for PRP no forward
558 * between ports, or sending HSR supervision to RedBox.
559 */
560 if (hsr->proto_ops->drop_frame &&
561 hsr->proto_ops->drop_frame(frame, port))
562 continue;
563
564 if (port->type == HSR_PT_SLAVE_A ||
565 port->type == HSR_PT_SLAVE_B)
566 skb = hsr->proto_ops->create_tagged_frame(frame, port);
567 else
568 skb = hsr->proto_ops->get_untagged_frame(frame, port);
569
570 if (!skb) {
571 frame->port_rcv->dev->stats.rx_dropped++;
572 continue;
573 }
574
575 skb->dev = port->dev;
576 if (port->type == HSR_PT_MASTER) {
577 hsr_deliver_master(skb, port->dev, frame->node_src);
578 } else {
579 if (!hsr_xmit(skb, port, frame))
580 if (port->type == HSR_PT_SLAVE_A ||
581 port->type == HSR_PT_SLAVE_B)
582 sent = true;
583 }
584 }
585 }
586
check_local_dest(struct hsr_priv * hsr,struct sk_buff * skb,struct hsr_frame_info * frame)587 static void check_local_dest(struct hsr_priv *hsr, struct sk_buff *skb,
588 struct hsr_frame_info *frame)
589 {
590 if (hsr_addr_is_self(hsr, eth_hdr(skb)->h_dest)) {
591 frame->is_local_exclusive = true;
592 skb->pkt_type = PACKET_HOST;
593 } else {
594 frame->is_local_exclusive = false;
595 }
596
597 if (skb->pkt_type == PACKET_HOST ||
598 skb->pkt_type == PACKET_MULTICAST ||
599 skb->pkt_type == PACKET_BROADCAST) {
600 frame->is_local_dest = true;
601 } else {
602 frame->is_local_dest = false;
603 }
604 }
605
handle_std_frame(struct sk_buff * skb,struct hsr_frame_info * frame)606 static void handle_std_frame(struct sk_buff *skb,
607 struct hsr_frame_info *frame)
608 {
609 struct hsr_port *port = frame->port_rcv;
610 struct hsr_priv *hsr = port->hsr;
611
612 frame->skb_hsr = NULL;
613 frame->skb_prp = NULL;
614 frame->skb_std = skb;
615
616 if (port->type != HSR_PT_MASTER)
617 frame->is_from_san = true;
618
619 if (port->type == HSR_PT_MASTER ||
620 port->type == HSR_PT_INTERLINK) {
621 /* Sequence nr for the master/interlink node */
622 lockdep_assert_held(&hsr->seqnr_lock);
623 frame->sequence_nr = hsr->sequence_nr;
624 hsr->sequence_nr++;
625 }
626 }
627
hsr_fill_frame_info(__be16 proto,struct sk_buff * skb,struct hsr_frame_info * frame)628 int hsr_fill_frame_info(__be16 proto, struct sk_buff *skb,
629 struct hsr_frame_info *frame)
630 {
631 struct hsr_port *port = frame->port_rcv;
632 struct hsr_priv *hsr = port->hsr;
633
634 /* HSRv0 supervisory frames double as a tag so treat them as tagged. */
635 if ((!hsr->prot_version && proto == htons(ETH_P_PRP)) ||
636 proto == htons(ETH_P_HSR)) {
637 /* Check if skb contains hsr_ethhdr */
638 if (skb->mac_len < sizeof(struct hsr_ethhdr))
639 return -EINVAL;
640
641 /* HSR tagged frame :- Data or Supervision */
642 frame->skb_std = NULL;
643 frame->skb_prp = NULL;
644 frame->skb_hsr = skb;
645 frame->sequence_nr = hsr_get_skb_sequence_nr(skb);
646 return 0;
647 }
648
649 /* Standard frame or PRP from master port */
650 handle_std_frame(skb, frame);
651
652 return 0;
653 }
654
prp_fill_frame_info(__be16 proto,struct sk_buff * skb,struct hsr_frame_info * frame)655 int prp_fill_frame_info(__be16 proto, struct sk_buff *skb,
656 struct hsr_frame_info *frame)
657 {
658 /* Supervision frame */
659 struct prp_rct *rct = skb_get_PRP_rct(skb);
660
661 if (rct &&
662 prp_check_lsdu_size(skb, rct, frame->is_supervision)) {
663 frame->skb_hsr = NULL;
664 frame->skb_std = NULL;
665 frame->skb_prp = skb;
666 frame->sequence_nr = prp_get_skb_sequence_nr(rct);
667 return 0;
668 }
669 handle_std_frame(skb, frame);
670
671 return 0;
672 }
673
fill_frame_info(struct hsr_frame_info * frame,struct sk_buff * skb,struct hsr_port * port)674 static int fill_frame_info(struct hsr_frame_info *frame,
675 struct sk_buff *skb, struct hsr_port *port)
676 {
677 struct hsr_priv *hsr = port->hsr;
678 struct hsr_vlan_ethhdr *vlan_hdr;
679 struct list_head *n_db;
680 struct ethhdr *ethhdr;
681 __be16 proto;
682 int ret;
683
684 /* Check if skb contains ethhdr */
685 if (skb->mac_len < sizeof(struct ethhdr))
686 return -EINVAL;
687
688 memset(frame, 0, sizeof(*frame));
689 frame->is_supervision = is_supervision_frame(port->hsr, skb);
690 if (frame->is_supervision && hsr->redbox)
691 frame->is_proxy_supervision =
692 is_proxy_supervision_frame(port->hsr, skb);
693
694 n_db = &hsr->node_db;
695 if (port->type == HSR_PT_INTERLINK)
696 n_db = &hsr->proxy_node_db;
697
698 frame->node_src = hsr_get_node(port, n_db, skb,
699 frame->is_supervision, port->type);
700 if (!frame->node_src)
701 return -1; /* Unknown node and !is_supervision, or no mem */
702
703 ethhdr = (struct ethhdr *)skb_mac_header(skb);
704 frame->is_vlan = false;
705 proto = ethhdr->h_proto;
706
707 if (proto == htons(ETH_P_8021Q))
708 frame->is_vlan = true;
709
710 if (frame->is_vlan) {
711 /* Note: skb->mac_len might be wrong here. */
712 if (!pskb_may_pull(skb,
713 skb_mac_offset(skb) +
714 offsetofend(struct hsr_vlan_ethhdr, vlanhdr)))
715 return -EINVAL;
716 vlan_hdr = (struct hsr_vlan_ethhdr *)skb_mac_header(skb);
717 proto = vlan_hdr->vlanhdr.h_vlan_encapsulated_proto;
718 }
719
720 frame->is_from_san = false;
721 frame->port_rcv = port;
722 ret = hsr->proto_ops->fill_frame_info(proto, skb, frame);
723 if (ret)
724 return ret;
725
726 check_local_dest(port->hsr, skb, frame);
727
728 return 0;
729 }
730
731 /* Must be called holding rcu read lock (because of the port parameter) */
hsr_forward_skb(struct sk_buff * skb,struct hsr_port * port)732 void hsr_forward_skb(struct sk_buff *skb, struct hsr_port *port)
733 {
734 struct hsr_frame_info frame;
735
736 rcu_read_lock();
737 if (fill_frame_info(&frame, skb, port) < 0)
738 goto out_drop;
739
740 hsr_register_frame_in(frame.node_src, port, frame.sequence_nr);
741 hsr_forward_do(&frame);
742 rcu_read_unlock();
743 /* Gets called for ingress frames as well as egress from master port.
744 * So check and increment stats for master port only here.
745 */
746 if (port->type == HSR_PT_MASTER || port->type == HSR_PT_INTERLINK) {
747 port->dev->stats.tx_packets++;
748 port->dev->stats.tx_bytes += skb->len;
749 }
750
751 kfree_skb(frame.skb_hsr);
752 kfree_skb(frame.skb_prp);
753 kfree_skb(frame.skb_std);
754 return;
755
756 out_drop:
757 rcu_read_unlock();
758 port->dev->stats.tx_dropped++;
759 kfree_skb(skb);
760 }
761