1 // SPDX-License-Identifier: GPL-2.0 2 /* Copyright 2011-2014 Autronica Fire and Security AS 3 * 4 * Author(s): 5 * 2011-2014 Arvid Brodin, arvid.brodin@alten.se 6 * 7 * Frame router for HSR and PRP. 8 */ 9 10 #include "hsr_forward.h" 11 #include <linux/types.h> 12 #include <linux/skbuff.h> 13 #include <linux/etherdevice.h> 14 #include <linux/if_vlan.h> 15 #include "hsr_main.h" 16 #include "hsr_framereg.h" 17 18 struct hsr_node; 19 20 /* The uses I can see for these HSR supervision frames are: 21 * 1) Use the frames that are sent after node initialization ("HSR_TLV.Type = 22 * 22") to reset any sequence_nr counters belonging to that node. Useful if 23 * the other node's counter has been reset for some reason. 24 * -- 25 * Or not - resetting the counter and bridging the frame would create a 26 * loop, unfortunately. 27 * 28 * 2) Use the LifeCheck frames to detect ring breaks. I.e. if no LifeCheck 29 * frame is received from a particular node, we know something is wrong. 30 * We just register these (as with normal frames) and throw them away. 31 * 32 * 3) Allow different MAC addresses for the two slave interfaces, using the 33 * MacAddressA field. 34 */ 35 static bool is_supervision_frame(struct hsr_priv *hsr, struct sk_buff *skb) 36 { 37 struct ethhdr *eth_hdr; 38 struct hsr_sup_tag *hsr_sup_tag; 39 struct hsrv1_ethhdr_sp *hsr_V1_hdr; 40 struct hsr_sup_tlv *hsr_sup_tlv; 41 u16 total_length = 0; 42 43 WARN_ON_ONCE(!skb_mac_header_was_set(skb)); 44 eth_hdr = (struct ethhdr *)skb_mac_header(skb); 45 46 /* Correct addr? */ 47 if (!ether_addr_equal(eth_hdr->h_dest, 48 hsr->sup_multicast_addr)) 49 return false; 50 51 /* Correct ether type?. */ 52 if (!(eth_hdr->h_proto == htons(ETH_P_PRP) || 53 eth_hdr->h_proto == htons(ETH_P_HSR))) 54 return false; 55 56 /* Get the supervision header from correct location. */ 57 if (eth_hdr->h_proto == htons(ETH_P_HSR)) { /* Okay HSRv1. */ 58 total_length = sizeof(struct hsrv1_ethhdr_sp); 59 if (!pskb_may_pull(skb, total_length)) 60 return false; 61 62 hsr_V1_hdr = (struct hsrv1_ethhdr_sp *)skb_mac_header(skb); 63 if (hsr_V1_hdr->hsr.encap_proto != htons(ETH_P_PRP)) 64 return false; 65 66 hsr_sup_tag = &hsr_V1_hdr->hsr_sup; 67 } else { 68 total_length = sizeof(struct hsrv0_ethhdr_sp); 69 if (!pskb_may_pull(skb, total_length)) 70 return false; 71 72 hsr_sup_tag = 73 &((struct hsrv0_ethhdr_sp *)skb_mac_header(skb))->hsr_sup; 74 } 75 76 if (hsr_sup_tag->tlv.HSR_TLV_type != HSR_TLV_ANNOUNCE && 77 hsr_sup_tag->tlv.HSR_TLV_type != HSR_TLV_LIFE_CHECK && 78 hsr_sup_tag->tlv.HSR_TLV_type != PRP_TLV_LIFE_CHECK_DD && 79 hsr_sup_tag->tlv.HSR_TLV_type != PRP_TLV_LIFE_CHECK_DA) 80 return false; 81 if (hsr_sup_tag->tlv.HSR_TLV_length != 12 && 82 hsr_sup_tag->tlv.HSR_TLV_length != sizeof(struct hsr_sup_payload)) 83 return false; 84 85 /* Get next tlv */ 86 total_length += hsr_sup_tag->tlv.HSR_TLV_length; 87 if (!pskb_may_pull(skb, total_length)) 88 return false; 89 skb_pull(skb, total_length); 90 hsr_sup_tlv = (struct hsr_sup_tlv *)skb->data; 91 skb_push(skb, total_length); 92 93 /* if this is a redbox supervision frame we need to verify 94 * that more data is available 95 */ 96 if (hsr_sup_tlv->HSR_TLV_type == PRP_TLV_REDBOX_MAC) { 97 /* tlv length must be a length of a mac address */ 98 if (hsr_sup_tlv->HSR_TLV_length != sizeof(struct hsr_sup_payload)) 99 return false; 100 101 /* make sure another tlv follows */ 102 total_length += sizeof(struct hsr_sup_tlv) + hsr_sup_tlv->HSR_TLV_length; 103 if (!pskb_may_pull(skb, total_length)) 104 return false; 105 106 /* get next tlv */ 107 skb_pull(skb, total_length); 108 hsr_sup_tlv = (struct hsr_sup_tlv *)skb->data; 109 skb_push(skb, total_length); 110 } 111 112 /* end of tlvs must follow at the end */ 113 if (hsr_sup_tlv->HSR_TLV_type == HSR_TLV_EOT && 114 hsr_sup_tlv->HSR_TLV_length != 0) 115 return false; 116 117 return true; 118 } 119 120 static bool is_proxy_supervision_frame(struct hsr_priv *hsr, 121 struct sk_buff *skb) 122 { 123 struct hsr_sup_payload *payload; 124 struct ethhdr *eth_hdr; 125 u16 total_length = 0; 126 127 eth_hdr = (struct ethhdr *)skb_mac_header(skb); 128 129 /* Get the HSR protocol revision. */ 130 if (eth_hdr->h_proto == htons(ETH_P_HSR)) 131 total_length = sizeof(struct hsrv1_ethhdr_sp); 132 else 133 total_length = sizeof(struct hsrv0_ethhdr_sp); 134 135 if (!pskb_may_pull(skb, total_length + sizeof(struct hsr_sup_payload))) 136 return false; 137 138 skb_pull(skb, total_length); 139 payload = (struct hsr_sup_payload *)skb->data; 140 skb_push(skb, total_length); 141 142 /* For RedBox (HSR-SAN) check if we have received the supervision 143 * frame with MAC addresses from own ProxyNodeTable. 144 */ 145 return hsr_is_node_in_db(&hsr->proxy_node_db, 146 payload->macaddress_A); 147 } 148 149 static struct sk_buff *create_stripped_skb_hsr(struct sk_buff *skb_in, 150 struct hsr_frame_info *frame) 151 { 152 struct sk_buff *skb; 153 int copylen; 154 unsigned char *dst, *src; 155 156 skb_pull(skb_in, HSR_HLEN); 157 skb = __pskb_copy(skb_in, skb_headroom(skb_in) - HSR_HLEN, GFP_ATOMIC); 158 skb_push(skb_in, HSR_HLEN); 159 if (!skb) 160 return NULL; 161 162 skb_reset_mac_header(skb); 163 164 if (skb->ip_summed == CHECKSUM_PARTIAL) 165 skb->csum_start -= HSR_HLEN; 166 167 copylen = 2 * ETH_ALEN; 168 if (frame->is_vlan) 169 copylen += VLAN_HLEN; 170 src = skb_mac_header(skb_in); 171 dst = skb_mac_header(skb); 172 memcpy(dst, src, copylen); 173 174 skb->protocol = eth_hdr(skb)->h_proto; 175 return skb; 176 } 177 178 struct sk_buff *hsr_get_untagged_frame(struct hsr_frame_info *frame, 179 struct hsr_port *port) 180 { 181 if (!frame->skb_std) { 182 if (frame->skb_hsr) 183 frame->skb_std = 184 create_stripped_skb_hsr(frame->skb_hsr, frame); 185 else 186 netdev_warn_once(port->dev, 187 "Unexpected frame received in hsr_get_untagged_frame()\n"); 188 189 if (!frame->skb_std) 190 return NULL; 191 } 192 193 return skb_clone(frame->skb_std, GFP_ATOMIC); 194 } 195 196 struct sk_buff *prp_get_untagged_frame(struct hsr_frame_info *frame, 197 struct hsr_port *port) 198 { 199 if (!frame->skb_std) { 200 if (frame->skb_prp) { 201 /* trim the skb by len - HSR_HLEN to exclude RCT */ 202 skb_trim(frame->skb_prp, 203 frame->skb_prp->len - HSR_HLEN); 204 frame->skb_std = 205 __pskb_copy(frame->skb_prp, 206 skb_headroom(frame->skb_prp), 207 GFP_ATOMIC); 208 } else { 209 /* Unexpected */ 210 WARN_ONCE(1, "%s:%d: Unexpected frame received (port_src %s)\n", 211 __FILE__, __LINE__, port->dev->name); 212 return NULL; 213 } 214 } 215 216 return skb_clone(frame->skb_std, GFP_ATOMIC); 217 } 218 219 static void prp_set_lan_id(struct prp_rct *trailer, 220 struct hsr_port *port) 221 { 222 int lane_id; 223 224 if (port->type == HSR_PT_SLAVE_A) 225 lane_id = 0; 226 else 227 lane_id = 1; 228 229 /* Add net_id in the upper 3 bits of lane_id */ 230 lane_id |= port->hsr->net_id; 231 set_prp_lan_id(trailer, lane_id); 232 } 233 234 /* Tailroom for PRP rct should have been created before calling this */ 235 static struct sk_buff *prp_fill_rct(struct sk_buff *skb, 236 struct hsr_frame_info *frame, 237 struct hsr_port *port) 238 { 239 struct prp_rct *trailer; 240 int min_size = ETH_ZLEN; 241 int lsdu_size; 242 243 if (!skb) 244 return skb; 245 246 if (frame->is_vlan) 247 min_size = VLAN_ETH_ZLEN; 248 249 if (skb_put_padto(skb, min_size)) 250 return NULL; 251 252 trailer = (struct prp_rct *)skb_put(skb, HSR_HLEN); 253 lsdu_size = skb->len - 14; 254 if (frame->is_vlan) 255 lsdu_size -= 4; 256 prp_set_lan_id(trailer, port); 257 set_prp_LSDU_size(trailer, lsdu_size); 258 trailer->sequence_nr = htons(frame->sequence_nr); 259 trailer->PRP_suffix = htons(ETH_P_PRP); 260 skb->protocol = eth_hdr(skb)->h_proto; 261 262 return skb; 263 } 264 265 static void hsr_set_path_id(struct hsr_frame_info *frame, 266 struct hsr_ethhdr *hsr_ethhdr, 267 struct hsr_port *port) 268 { 269 int path_id; 270 271 if (port->hsr->prot_version) { 272 if (port->type == HSR_PT_SLAVE_A) 273 path_id = 0; 274 else 275 path_id = 1; 276 } else { 277 if (frame->is_supervision) 278 path_id = 0xf; 279 else 280 path_id = 1; 281 } 282 283 set_hsr_tag_path(&hsr_ethhdr->hsr_tag, path_id); 284 } 285 286 static struct sk_buff *hsr_fill_tag(struct sk_buff *skb, 287 struct hsr_frame_info *frame, 288 struct hsr_port *port, u8 proto_version) 289 { 290 struct hsr_ethhdr *hsr_ethhdr; 291 unsigned char *pc; 292 int lsdu_size; 293 294 /* pad to minimum packet size which is 60 + 6 (HSR tag) */ 295 if (skb_put_padto(skb, ETH_ZLEN + HSR_HLEN)) 296 return NULL; 297 298 lsdu_size = skb->len - 14; 299 if (frame->is_vlan) 300 lsdu_size -= 4; 301 302 pc = skb_mac_header(skb); 303 if (frame->is_vlan) 304 /* This 4-byte shift (size of a vlan tag) does not 305 * mean that the ethhdr starts there. But rather it 306 * provides the proper environment for accessing 307 * the fields, such as hsr_tag etc., just like 308 * when the vlan tag is not there. This is because 309 * the hsr tag is after the vlan tag. 310 */ 311 hsr_ethhdr = (struct hsr_ethhdr *)(pc + VLAN_HLEN); 312 else 313 hsr_ethhdr = (struct hsr_ethhdr *)pc; 314 315 hsr_set_path_id(frame, hsr_ethhdr, port); 316 set_hsr_tag_LSDU_size(&hsr_ethhdr->hsr_tag, lsdu_size); 317 hsr_ethhdr->hsr_tag.sequence_nr = htons(frame->sequence_nr); 318 hsr_ethhdr->hsr_tag.encap_proto = hsr_ethhdr->ethhdr.h_proto; 319 hsr_ethhdr->ethhdr.h_proto = htons(proto_version ? 320 ETH_P_HSR : ETH_P_PRP); 321 skb->protocol = hsr_ethhdr->ethhdr.h_proto; 322 323 return skb; 324 } 325 326 /* If the original frame was an HSR tagged frame, just clone it to be sent 327 * unchanged. Otherwise, create a private frame especially tagged for 'port'. 328 */ 329 struct sk_buff *hsr_create_tagged_frame(struct hsr_frame_info *frame, 330 struct hsr_port *port) 331 { 332 unsigned char *dst, *src; 333 struct sk_buff *skb; 334 int movelen; 335 336 if (frame->skb_hsr) { 337 struct hsr_ethhdr *hsr_ethhdr = 338 (struct hsr_ethhdr *)skb_mac_header(frame->skb_hsr); 339 340 /* set the lane id properly */ 341 hsr_set_path_id(frame, hsr_ethhdr, port); 342 return skb_clone(frame->skb_hsr, GFP_ATOMIC); 343 } else if (port->dev->features & NETIF_F_HW_HSR_TAG_INS) { 344 return skb_clone(frame->skb_std, GFP_ATOMIC); 345 } 346 347 /* Create the new skb with enough headroom to fit the HSR tag */ 348 skb = __pskb_copy(frame->skb_std, 349 skb_headroom(frame->skb_std) + HSR_HLEN, GFP_ATOMIC); 350 if (!skb) 351 return NULL; 352 skb_reset_mac_header(skb); 353 354 if (skb->ip_summed == CHECKSUM_PARTIAL) 355 skb->csum_start += HSR_HLEN; 356 357 movelen = ETH_HLEN; 358 if (frame->is_vlan) 359 movelen += VLAN_HLEN; 360 361 src = skb_mac_header(skb); 362 dst = skb_push(skb, HSR_HLEN); 363 memmove(dst, src, movelen); 364 skb_reset_mac_header(skb); 365 366 /* skb_put_padto free skb on error and hsr_fill_tag returns NULL in 367 * that case 368 */ 369 return hsr_fill_tag(skb, frame, port, port->hsr->prot_version); 370 } 371 372 struct sk_buff *prp_create_tagged_frame(struct hsr_frame_info *frame, 373 struct hsr_port *port) 374 { 375 struct sk_buff *skb; 376 377 if (frame->skb_prp) { 378 struct prp_rct *trailer = skb_get_PRP_rct(frame->skb_prp); 379 380 if (trailer) { 381 prp_set_lan_id(trailer, port); 382 } else { 383 WARN_ONCE(!trailer, "errored PRP skb"); 384 return NULL; 385 } 386 return skb_clone(frame->skb_prp, GFP_ATOMIC); 387 } else if (port->dev->features & NETIF_F_HW_HSR_TAG_INS) { 388 return skb_clone(frame->skb_std, GFP_ATOMIC); 389 } 390 391 skb = skb_copy_expand(frame->skb_std, skb_headroom(frame->skb_std), 392 skb_tailroom(frame->skb_std) + HSR_HLEN, 393 GFP_ATOMIC); 394 return prp_fill_rct(skb, frame, port); 395 } 396 397 static void hsr_deliver_master(struct sk_buff *skb, struct net_device *dev, 398 struct hsr_node *node_src) 399 { 400 bool was_multicast_frame; 401 int res, recv_len; 402 403 was_multicast_frame = (skb->pkt_type == PACKET_MULTICAST); 404 hsr_addr_subst_source(node_src, skb); 405 skb_pull(skb, ETH_HLEN); 406 recv_len = skb->len; 407 res = netif_rx(skb); 408 if (res == NET_RX_DROP) { 409 dev->stats.rx_dropped++; 410 } else { 411 dev->stats.rx_packets++; 412 dev->stats.rx_bytes += recv_len; 413 if (was_multicast_frame) 414 dev->stats.multicast++; 415 } 416 } 417 418 static int hsr_xmit(struct sk_buff *skb, struct hsr_port *port, 419 struct hsr_frame_info *frame) 420 { 421 if (frame->port_rcv->type == HSR_PT_MASTER) { 422 hsr_addr_subst_dest(frame->node_src, skb, port); 423 424 /* Address substitution (IEC62439-3 pp 26, 50): replace mac 425 * address of outgoing frame with that of the outgoing slave's. 426 */ 427 ether_addr_copy(eth_hdr(skb)->h_source, port->dev->dev_addr); 428 } 429 430 /* When HSR node is used as RedBox - the frame received from HSR ring 431 * requires source MAC address (SA) replacement to one which can be 432 * recognized by SAN devices (otherwise, frames are dropped by switch) 433 */ 434 if (port->type == HSR_PT_INTERLINK) 435 ether_addr_copy(eth_hdr(skb)->h_source, 436 port->hsr->macaddress_redbox); 437 438 return dev_queue_xmit(skb); 439 } 440 441 bool prp_drop_frame(struct hsr_frame_info *frame, struct hsr_port *port) 442 { 443 return ((frame->port_rcv->type == HSR_PT_SLAVE_A && 444 port->type == HSR_PT_SLAVE_B) || 445 (frame->port_rcv->type == HSR_PT_SLAVE_B && 446 port->type == HSR_PT_SLAVE_A)); 447 } 448 449 bool hsr_drop_frame(struct hsr_frame_info *frame, struct hsr_port *port) 450 { 451 struct sk_buff *skb; 452 453 if (port->dev->features & NETIF_F_HW_HSR_FWD) 454 return prp_drop_frame(frame, port); 455 456 /* RedBox specific frames dropping policies 457 * 458 * Do not send HSR supervisory frames to SAN devices 459 */ 460 if (frame->is_supervision && port->type == HSR_PT_INTERLINK) 461 return true; 462 463 /* Do not forward to other HSR port (A or B) unicast frames which 464 * are addressed to interlink port (and are in the ProxyNodeTable). 465 */ 466 skb = frame->skb_hsr; 467 if (skb && prp_drop_frame(frame, port) && 468 is_unicast_ether_addr(eth_hdr(skb)->h_dest) && 469 hsr_is_node_in_db(&port->hsr->proxy_node_db, 470 eth_hdr(skb)->h_dest)) { 471 return true; 472 } 473 474 /* Do not forward to port C (Interlink) frames from nodes A and B 475 * if DA is in NodeTable. 476 */ 477 if ((frame->port_rcv->type == HSR_PT_SLAVE_A || 478 frame->port_rcv->type == HSR_PT_SLAVE_B) && 479 port->type == HSR_PT_INTERLINK) { 480 skb = frame->skb_hsr; 481 if (skb && is_unicast_ether_addr(eth_hdr(skb)->h_dest) && 482 hsr_is_node_in_db(&port->hsr->node_db, 483 eth_hdr(skb)->h_dest)) { 484 return true; 485 } 486 } 487 488 /* Do not forward to port A and B unicast frames received on the 489 * interlink port if it is addressed to one of nodes registered in 490 * the ProxyNodeTable. 491 */ 492 if ((port->type == HSR_PT_SLAVE_A || port->type == HSR_PT_SLAVE_B) && 493 frame->port_rcv->type == HSR_PT_INTERLINK) { 494 skb = frame->skb_std; 495 if (skb && is_unicast_ether_addr(eth_hdr(skb)->h_dest) && 496 hsr_is_node_in_db(&port->hsr->proxy_node_db, 497 eth_hdr(skb)->h_dest)) { 498 return true; 499 } 500 } 501 502 return false; 503 } 504 505 /* Forward the frame through all devices except: 506 * - Back through the receiving device 507 * - If it's a HSR frame: through a device where it has passed before 508 * - if it's a PRP frame: through another PRP slave device (no bridge) 509 * - To the local HSR master only if the frame is directly addressed to it, or 510 * a non-supervision multicast or broadcast frame. 511 * 512 * HSR slave devices should insert a HSR tag into the frame, or forward the 513 * frame unchanged if it's already tagged. Interlink devices should strip HSR 514 * tags if they're of the non-HSR type (but only after duplicate discard). The 515 * master device always strips HSR tags. 516 */ 517 static void hsr_forward_do(struct hsr_frame_info *frame) 518 { 519 struct hsr_port *port; 520 struct sk_buff *skb; 521 bool sent = false; 522 523 hsr_for_each_port(frame->port_rcv->hsr, port) { 524 struct hsr_priv *hsr = port->hsr; 525 /* Don't send frame back the way it came */ 526 if (port == frame->port_rcv) 527 continue; 528 529 /* Don't deliver locally unless we should */ 530 if (port->type == HSR_PT_MASTER && !frame->is_local_dest) 531 continue; 532 533 /* Deliver frames directly addressed to us to master only */ 534 if (port->type != HSR_PT_MASTER && frame->is_local_exclusive) 535 continue; 536 537 /* If hardware duplicate generation is enabled, only send out 538 * one port. 539 */ 540 if ((port->dev->features & NETIF_F_HW_HSR_DUP) && sent) 541 continue; 542 543 /* Don't send frame over port where it has been sent before. 544 * Also for SAN, this shouldn't be done. 545 */ 546 if (!frame->is_from_san && 547 hsr->proto_ops->register_frame_out && 548 hsr->proto_ops->register_frame_out(port, frame)) 549 continue; 550 551 if (frame->is_supervision && port->type == HSR_PT_MASTER && 552 !frame->is_proxy_supervision) { 553 hsr_handle_sup_frame(frame); 554 continue; 555 } 556 557 /* Check if frame is to be dropped. Eg. for PRP no forward 558 * between ports, or sending HSR supervision to RedBox. 559 */ 560 if (hsr->proto_ops->drop_frame && 561 hsr->proto_ops->drop_frame(frame, port)) 562 continue; 563 564 if (port->type == HSR_PT_SLAVE_A || 565 port->type == HSR_PT_SLAVE_B) 566 skb = hsr->proto_ops->create_tagged_frame(frame, port); 567 else 568 skb = hsr->proto_ops->get_untagged_frame(frame, port); 569 570 if (!skb) { 571 frame->port_rcv->dev->stats.rx_dropped++; 572 continue; 573 } 574 575 skb->dev = port->dev; 576 if (port->type == HSR_PT_MASTER) { 577 hsr_deliver_master(skb, port->dev, frame->node_src); 578 } else { 579 if (!hsr_xmit(skb, port, frame)) 580 if (port->type == HSR_PT_SLAVE_A || 581 port->type == HSR_PT_SLAVE_B) 582 sent = true; 583 } 584 } 585 } 586 587 static void check_local_dest(struct hsr_priv *hsr, struct sk_buff *skb, 588 struct hsr_frame_info *frame) 589 { 590 if (hsr_addr_is_self(hsr, eth_hdr(skb)->h_dest)) { 591 frame->is_local_exclusive = true; 592 skb->pkt_type = PACKET_HOST; 593 } else { 594 frame->is_local_exclusive = false; 595 } 596 597 if (skb->pkt_type == PACKET_HOST || 598 skb->pkt_type == PACKET_MULTICAST || 599 skb->pkt_type == PACKET_BROADCAST) { 600 frame->is_local_dest = true; 601 } else { 602 frame->is_local_dest = false; 603 } 604 } 605 606 static void handle_std_frame(struct sk_buff *skb, 607 struct hsr_frame_info *frame) 608 { 609 struct hsr_port *port = frame->port_rcv; 610 struct hsr_priv *hsr = port->hsr; 611 612 frame->skb_hsr = NULL; 613 frame->skb_prp = NULL; 614 frame->skb_std = skb; 615 616 if (port->type != HSR_PT_MASTER) 617 frame->is_from_san = true; 618 619 if (port->type == HSR_PT_MASTER || 620 port->type == HSR_PT_INTERLINK) { 621 /* Sequence nr for the master/interlink node */ 622 lockdep_assert_held(&hsr->seqnr_lock); 623 frame->sequence_nr = hsr->sequence_nr; 624 hsr->sequence_nr++; 625 } 626 } 627 628 int hsr_fill_frame_info(__be16 proto, struct sk_buff *skb, 629 struct hsr_frame_info *frame) 630 { 631 struct hsr_port *port = frame->port_rcv; 632 struct hsr_priv *hsr = port->hsr; 633 634 /* HSRv0 supervisory frames double as a tag so treat them as tagged. */ 635 if ((!hsr->prot_version && proto == htons(ETH_P_PRP)) || 636 proto == htons(ETH_P_HSR)) { 637 /* Check if skb contains hsr_ethhdr */ 638 if (skb->mac_len < sizeof(struct hsr_ethhdr)) 639 return -EINVAL; 640 641 /* HSR tagged frame :- Data or Supervision */ 642 frame->skb_std = NULL; 643 frame->skb_prp = NULL; 644 frame->skb_hsr = skb; 645 frame->sequence_nr = hsr_get_skb_sequence_nr(skb); 646 return 0; 647 } 648 649 /* Standard frame or PRP from master port */ 650 handle_std_frame(skb, frame); 651 652 return 0; 653 } 654 655 int prp_fill_frame_info(__be16 proto, struct sk_buff *skb, 656 struct hsr_frame_info *frame) 657 { 658 /* Supervision frame */ 659 struct prp_rct *rct = skb_get_PRP_rct(skb); 660 661 if (rct && 662 prp_check_lsdu_size(skb, rct, frame->is_supervision)) { 663 frame->skb_hsr = NULL; 664 frame->skb_std = NULL; 665 frame->skb_prp = skb; 666 frame->sequence_nr = prp_get_skb_sequence_nr(rct); 667 return 0; 668 } 669 handle_std_frame(skb, frame); 670 671 return 0; 672 } 673 674 static int fill_frame_info(struct hsr_frame_info *frame, 675 struct sk_buff *skb, struct hsr_port *port) 676 { 677 struct hsr_priv *hsr = port->hsr; 678 struct hsr_vlan_ethhdr *vlan_hdr; 679 struct list_head *n_db; 680 struct ethhdr *ethhdr; 681 __be16 proto; 682 int ret; 683 684 /* Check if skb contains ethhdr */ 685 if (skb->mac_len < sizeof(struct ethhdr)) 686 return -EINVAL; 687 688 memset(frame, 0, sizeof(*frame)); 689 frame->is_supervision = is_supervision_frame(port->hsr, skb); 690 if (frame->is_supervision && hsr->redbox) 691 frame->is_proxy_supervision = 692 is_proxy_supervision_frame(port->hsr, skb); 693 694 n_db = &hsr->node_db; 695 if (port->type == HSR_PT_INTERLINK) 696 n_db = &hsr->proxy_node_db; 697 698 frame->node_src = hsr_get_node(port, n_db, skb, 699 frame->is_supervision, port->type); 700 if (!frame->node_src) 701 return -1; /* Unknown node and !is_supervision, or no mem */ 702 703 ethhdr = (struct ethhdr *)skb_mac_header(skb); 704 frame->is_vlan = false; 705 proto = ethhdr->h_proto; 706 707 if (proto == htons(ETH_P_8021Q)) 708 frame->is_vlan = true; 709 710 if (frame->is_vlan) { 711 /* Note: skb->mac_len might be wrong here. */ 712 if (!pskb_may_pull(skb, 713 skb_mac_offset(skb) + 714 offsetofend(struct hsr_vlan_ethhdr, vlanhdr))) 715 return -EINVAL; 716 vlan_hdr = (struct hsr_vlan_ethhdr *)skb_mac_header(skb); 717 proto = vlan_hdr->vlanhdr.h_vlan_encapsulated_proto; 718 } 719 720 frame->is_from_san = false; 721 frame->port_rcv = port; 722 ret = hsr->proto_ops->fill_frame_info(proto, skb, frame); 723 if (ret) 724 return ret; 725 726 check_local_dest(port->hsr, skb, frame); 727 728 return 0; 729 } 730 731 /* Must be called holding rcu read lock (because of the port parameter) */ 732 void hsr_forward_skb(struct sk_buff *skb, struct hsr_port *port) 733 { 734 struct hsr_frame_info frame; 735 736 rcu_read_lock(); 737 if (fill_frame_info(&frame, skb, port) < 0) 738 goto out_drop; 739 740 hsr_register_frame_in(frame.node_src, port, frame.sequence_nr); 741 hsr_forward_do(&frame); 742 rcu_read_unlock(); 743 /* Gets called for ingress frames as well as egress from master port. 744 * So check and increment stats for master port only here. 745 */ 746 if (port->type == HSR_PT_MASTER || port->type == HSR_PT_INTERLINK) { 747 port->dev->stats.tx_packets++; 748 port->dev->stats.tx_bytes += skb->len; 749 } 750 751 kfree_skb(frame.skb_hsr); 752 kfree_skb(frame.skb_prp); 753 kfree_skb(frame.skb_std); 754 return; 755 756 out_drop: 757 rcu_read_unlock(); 758 port->dev->stats.tx_dropped++; 759 kfree_skb(skb); 760 } 761