1 /* 2 * Copyright (C) 2017 Netronome Systems, Inc. 3 * 4 * This software is licensed under the GNU General License Version 2, 5 * June 1991 as shown in the file COPYING in the top-level directory of this 6 * source tree. 7 * 8 * THE COPYRIGHT HOLDERS AND/OR OTHER PARTIES PROVIDE THE PROGRAM "AS IS" 9 * WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED OR IMPLIED, INCLUDING, 10 * BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS 11 * FOR A PARTICULAR PURPOSE. THE ENTIRE RISK AS TO THE QUALITY AND PERFORMANCE 12 * OF THE PROGRAM IS WITH YOU. SHOULD THE PROGRAM PROVE DEFECTIVE, YOU ASSUME 13 * THE COST OF ALL NECESSARY SERVICING, REPAIR OR CORRECTION. 14 */ 15 16 #include <linux/debugfs.h> 17 #include <linux/etherdevice.h> 18 #include <linux/ethtool_netlink.h> 19 #include <linux/kernel.h> 20 #include <linux/module.h> 21 #include <linux/netdevice.h> 22 #include <linux/slab.h> 23 #include <net/netdev_queues.h> 24 #include <net/netdev_rx_queue.h> 25 #include <net/page_pool/helpers.h> 26 #include <net/netlink.h> 27 #include <net/net_shaper.h> 28 #include <net/netdev_lock.h> 29 #include <net/pkt_cls.h> 30 #include <net/rtnetlink.h> 31 #include <net/udp_tunnel.h> 32 #include <net/busy_poll.h> 33 34 #include "netdevsim.h" 35 36 MODULE_IMPORT_NS("NETDEV_INTERNAL"); 37 38 #define NSIM_RING_SIZE 256 39 40 static void nsim_start_peer_tx_queue(struct net_device *dev, struct nsim_rq *rq) 41 { 42 struct netdevsim *ns = netdev_priv(dev); 43 struct net_device *peer_dev; 44 struct netdevsim *peer_ns; 45 struct netdev_queue *txq; 46 u16 idx; 47 48 idx = rq->napi.index; 49 rcu_read_lock(); 50 peer_ns = rcu_dereference(ns->peer); 51 if (!peer_ns) 52 goto out; 53 54 /* TX device */ 55 peer_dev = peer_ns->netdev; 56 if (dev->real_num_tx_queues != peer_dev->num_rx_queues) 57 goto out; 58 59 txq = netdev_get_tx_queue(peer_dev, idx); 60 if (!netif_tx_queue_stopped(txq)) 61 goto out; 62 63 netif_tx_wake_queue(txq); 64 out: 65 rcu_read_unlock(); 66 } 67 68 static void nsim_stop_tx_queue(struct net_device *tx_dev, 69 struct net_device *rx_dev, 70 struct nsim_rq *rq, 71 u16 idx) 72 { 73 /* If different queues size, do not stop, since it is not 74 * easy to find which TX queue is mapped here 75 */ 76 if (rx_dev->real_num_tx_queues != tx_dev->num_rx_queues) 77 return; 78 79 /* rq is the queue on the receive side */ 80 netif_subqueue_try_stop(tx_dev, idx, 81 NSIM_RING_SIZE - skb_queue_len(&rq->skb_queue), 82 NSIM_RING_SIZE / 2); 83 } 84 85 static int nsim_napi_rx(struct net_device *tx_dev, struct net_device *rx_dev, 86 struct nsim_rq *rq, struct sk_buff *skb) 87 { 88 if (skb_queue_len(&rq->skb_queue) > NSIM_RING_SIZE) { 89 dev_kfree_skb_any(skb); 90 return NET_RX_DROP; 91 } 92 93 skb_queue_tail(&rq->skb_queue, skb); 94 95 /* Stop the peer TX queue avoiding dropping packets later */ 96 if (skb_queue_len(&rq->skb_queue) >= NSIM_RING_SIZE) 97 nsim_stop_tx_queue(tx_dev, rx_dev, rq, 98 skb_get_queue_mapping(skb)); 99 100 return NET_RX_SUCCESS; 101 } 102 103 static int nsim_forward_skb(struct net_device *tx_dev, 104 struct net_device *rx_dev, 105 struct sk_buff *skb, 106 struct nsim_rq *rq) 107 { 108 return __dev_forward_skb(rx_dev, skb) ?: 109 nsim_napi_rx(tx_dev, rx_dev, rq, skb); 110 } 111 112 static netdev_tx_t nsim_start_xmit(struct sk_buff *skb, struct net_device *dev) 113 { 114 struct netdevsim *ns = netdev_priv(dev); 115 struct net_device *peer_dev; 116 unsigned int len = skb->len; 117 struct netdevsim *peer_ns; 118 struct netdev_config *cfg; 119 struct nsim_rq *rq; 120 int rxq; 121 122 rcu_read_lock(); 123 if (!nsim_ipsec_tx(ns, skb)) 124 goto out_drop_free; 125 126 peer_ns = rcu_dereference(ns->peer); 127 if (!peer_ns) 128 goto out_drop_free; 129 130 peer_dev = peer_ns->netdev; 131 rxq = skb_get_queue_mapping(skb); 132 if (rxq >= peer_dev->num_rx_queues) 133 rxq = rxq % peer_dev->num_rx_queues; 134 rq = peer_ns->rq[rxq]; 135 136 cfg = peer_dev->cfg; 137 if (skb_is_nonlinear(skb) && 138 (cfg->hds_config != ETHTOOL_TCP_DATA_SPLIT_ENABLED || 139 (cfg->hds_config == ETHTOOL_TCP_DATA_SPLIT_ENABLED && 140 cfg->hds_thresh > len))) 141 skb_linearize(skb); 142 143 skb_tx_timestamp(skb); 144 if (unlikely(nsim_forward_skb(dev, peer_dev, skb, rq) == NET_RX_DROP)) 145 goto out_drop_cnt; 146 147 if (!hrtimer_active(&rq->napi_timer)) 148 hrtimer_start(&rq->napi_timer, us_to_ktime(5), HRTIMER_MODE_REL); 149 150 rcu_read_unlock(); 151 dev_dstats_tx_add(dev, len); 152 return NETDEV_TX_OK; 153 154 out_drop_free: 155 dev_kfree_skb(skb); 156 out_drop_cnt: 157 rcu_read_unlock(); 158 dev_dstats_tx_dropped(dev); 159 return NETDEV_TX_OK; 160 } 161 162 static void nsim_set_rx_mode(struct net_device *dev) 163 { 164 } 165 166 static int nsim_change_mtu(struct net_device *dev, int new_mtu) 167 { 168 struct netdevsim *ns = netdev_priv(dev); 169 170 if (ns->xdp.prog && !ns->xdp.prog->aux->xdp_has_frags && 171 new_mtu > NSIM_XDP_MAX_MTU) 172 return -EBUSY; 173 174 WRITE_ONCE(dev->mtu, new_mtu); 175 176 return 0; 177 } 178 179 static int 180 nsim_setup_tc_block_cb(enum tc_setup_type type, void *type_data, void *cb_priv) 181 { 182 return nsim_bpf_setup_tc_block_cb(type, type_data, cb_priv); 183 } 184 185 static int nsim_set_vf_mac(struct net_device *dev, int vf, u8 *mac) 186 { 187 struct netdevsim *ns = netdev_priv(dev); 188 struct nsim_dev *nsim_dev = ns->nsim_dev; 189 190 /* Only refuse multicast addresses, zero address can mean unset/any. */ 191 if (vf >= nsim_dev_get_vfs(nsim_dev) || is_multicast_ether_addr(mac)) 192 return -EINVAL; 193 memcpy(nsim_dev->vfconfigs[vf].vf_mac, mac, ETH_ALEN); 194 195 return 0; 196 } 197 198 static int nsim_set_vf_vlan(struct net_device *dev, int vf, 199 u16 vlan, u8 qos, __be16 vlan_proto) 200 { 201 struct netdevsim *ns = netdev_priv(dev); 202 struct nsim_dev *nsim_dev = ns->nsim_dev; 203 204 if (vf >= nsim_dev_get_vfs(nsim_dev) || vlan > 4095 || qos > 7) 205 return -EINVAL; 206 207 nsim_dev->vfconfigs[vf].vlan = vlan; 208 nsim_dev->vfconfigs[vf].qos = qos; 209 nsim_dev->vfconfigs[vf].vlan_proto = vlan_proto; 210 211 return 0; 212 } 213 214 static int nsim_set_vf_rate(struct net_device *dev, int vf, int min, int max) 215 { 216 struct netdevsim *ns = netdev_priv(dev); 217 struct nsim_dev *nsim_dev = ns->nsim_dev; 218 219 if (nsim_esw_mode_is_switchdev(ns->nsim_dev)) { 220 pr_err("Not supported in switchdev mode. Please use devlink API.\n"); 221 return -EOPNOTSUPP; 222 } 223 224 if (vf >= nsim_dev_get_vfs(nsim_dev)) 225 return -EINVAL; 226 227 nsim_dev->vfconfigs[vf].min_tx_rate = min; 228 nsim_dev->vfconfigs[vf].max_tx_rate = max; 229 230 return 0; 231 } 232 233 static int nsim_set_vf_spoofchk(struct net_device *dev, int vf, bool val) 234 { 235 struct netdevsim *ns = netdev_priv(dev); 236 struct nsim_dev *nsim_dev = ns->nsim_dev; 237 238 if (vf >= nsim_dev_get_vfs(nsim_dev)) 239 return -EINVAL; 240 nsim_dev->vfconfigs[vf].spoofchk_enabled = val; 241 242 return 0; 243 } 244 245 static int nsim_set_vf_rss_query_en(struct net_device *dev, int vf, bool val) 246 { 247 struct netdevsim *ns = netdev_priv(dev); 248 struct nsim_dev *nsim_dev = ns->nsim_dev; 249 250 if (vf >= nsim_dev_get_vfs(nsim_dev)) 251 return -EINVAL; 252 nsim_dev->vfconfigs[vf].rss_query_enabled = val; 253 254 return 0; 255 } 256 257 static int nsim_set_vf_trust(struct net_device *dev, int vf, bool val) 258 { 259 struct netdevsim *ns = netdev_priv(dev); 260 struct nsim_dev *nsim_dev = ns->nsim_dev; 261 262 if (vf >= nsim_dev_get_vfs(nsim_dev)) 263 return -EINVAL; 264 nsim_dev->vfconfigs[vf].trusted = val; 265 266 return 0; 267 } 268 269 static int 270 nsim_get_vf_config(struct net_device *dev, int vf, struct ifla_vf_info *ivi) 271 { 272 struct netdevsim *ns = netdev_priv(dev); 273 struct nsim_dev *nsim_dev = ns->nsim_dev; 274 275 if (vf >= nsim_dev_get_vfs(nsim_dev)) 276 return -EINVAL; 277 278 ivi->vf = vf; 279 ivi->linkstate = nsim_dev->vfconfigs[vf].link_state; 280 ivi->min_tx_rate = nsim_dev->vfconfigs[vf].min_tx_rate; 281 ivi->max_tx_rate = nsim_dev->vfconfigs[vf].max_tx_rate; 282 ivi->vlan = nsim_dev->vfconfigs[vf].vlan; 283 ivi->vlan_proto = nsim_dev->vfconfigs[vf].vlan_proto; 284 ivi->qos = nsim_dev->vfconfigs[vf].qos; 285 memcpy(&ivi->mac, nsim_dev->vfconfigs[vf].vf_mac, ETH_ALEN); 286 ivi->spoofchk = nsim_dev->vfconfigs[vf].spoofchk_enabled; 287 ivi->trusted = nsim_dev->vfconfigs[vf].trusted; 288 ivi->rss_query_en = nsim_dev->vfconfigs[vf].rss_query_enabled; 289 290 return 0; 291 } 292 293 static int nsim_set_vf_link_state(struct net_device *dev, int vf, int state) 294 { 295 struct netdevsim *ns = netdev_priv(dev); 296 struct nsim_dev *nsim_dev = ns->nsim_dev; 297 298 if (vf >= nsim_dev_get_vfs(nsim_dev)) 299 return -EINVAL; 300 301 switch (state) { 302 case IFLA_VF_LINK_STATE_AUTO: 303 case IFLA_VF_LINK_STATE_ENABLE: 304 case IFLA_VF_LINK_STATE_DISABLE: 305 break; 306 default: 307 return -EINVAL; 308 } 309 310 nsim_dev->vfconfigs[vf].link_state = state; 311 312 return 0; 313 } 314 315 static void nsim_taprio_stats(struct tc_taprio_qopt_stats *stats) 316 { 317 stats->window_drops = 0; 318 stats->tx_overruns = 0; 319 } 320 321 static int nsim_setup_tc_taprio(struct net_device *dev, 322 struct tc_taprio_qopt_offload *offload) 323 { 324 int err = 0; 325 326 switch (offload->cmd) { 327 case TAPRIO_CMD_REPLACE: 328 case TAPRIO_CMD_DESTROY: 329 break; 330 case TAPRIO_CMD_STATS: 331 nsim_taprio_stats(&offload->stats); 332 break; 333 default: 334 err = -EOPNOTSUPP; 335 } 336 337 return err; 338 } 339 340 static LIST_HEAD(nsim_block_cb_list); 341 342 static int 343 nsim_setup_tc(struct net_device *dev, enum tc_setup_type type, void *type_data) 344 { 345 struct netdevsim *ns = netdev_priv(dev); 346 347 switch (type) { 348 case TC_SETUP_QDISC_TAPRIO: 349 return nsim_setup_tc_taprio(dev, type_data); 350 case TC_SETUP_BLOCK: 351 return flow_block_cb_setup_simple(type_data, 352 &nsim_block_cb_list, 353 nsim_setup_tc_block_cb, 354 ns, ns, true); 355 default: 356 return -EOPNOTSUPP; 357 } 358 } 359 360 static int 361 nsim_set_features(struct net_device *dev, netdev_features_t features) 362 { 363 struct netdevsim *ns = netdev_priv(dev); 364 365 if ((dev->features & NETIF_F_HW_TC) > (features & NETIF_F_HW_TC)) 366 return nsim_bpf_disable_tc(ns); 367 368 return 0; 369 } 370 371 static int nsim_get_iflink(const struct net_device *dev) 372 { 373 struct netdevsim *nsim, *peer; 374 int iflink; 375 376 nsim = netdev_priv(dev); 377 378 rcu_read_lock(); 379 peer = rcu_dereference(nsim->peer); 380 iflink = peer ? READ_ONCE(peer->netdev->ifindex) : 381 READ_ONCE(dev->ifindex); 382 rcu_read_unlock(); 383 384 return iflink; 385 } 386 387 static int nsim_rcv(struct nsim_rq *rq, int budget) 388 { 389 struct net_device *dev = rq->napi.dev; 390 struct bpf_prog *xdp_prog; 391 struct netdevsim *ns; 392 struct sk_buff *skb; 393 unsigned int skblen; 394 int i, ret; 395 396 ns = netdev_priv(dev); 397 xdp_prog = READ_ONCE(ns->xdp.prog); 398 399 for (i = 0; i < budget; i++) { 400 if (skb_queue_empty(&rq->skb_queue)) 401 break; 402 403 skb = skb_dequeue(&rq->skb_queue); 404 405 if (xdp_prog) { 406 /* skb might be freed directly by XDP, save the len */ 407 skblen = skb->len; 408 409 if (skb->ip_summed == CHECKSUM_PARTIAL) 410 skb_checksum_help(skb); 411 ret = do_xdp_generic(xdp_prog, &skb); 412 if (ret != XDP_PASS) { 413 dev_dstats_rx_add(dev, skblen); 414 continue; 415 } 416 } 417 418 /* skb might be discard at netif_receive_skb, save the len */ 419 skblen = skb->len; 420 skb_mark_napi_id(skb, &rq->napi); 421 ret = netif_receive_skb(skb); 422 if (ret == NET_RX_SUCCESS) 423 dev_dstats_rx_add(dev, skblen); 424 else 425 dev_dstats_rx_dropped(dev); 426 } 427 428 nsim_start_peer_tx_queue(dev, rq); 429 return i; 430 } 431 432 static int nsim_poll(struct napi_struct *napi, int budget) 433 { 434 struct nsim_rq *rq = container_of(napi, struct nsim_rq, napi); 435 int done; 436 437 done = nsim_rcv(rq, budget); 438 if (done < budget) 439 napi_complete_done(napi, done); 440 441 return done; 442 } 443 444 static int nsim_create_page_pool(struct page_pool **p, struct napi_struct *napi) 445 { 446 struct page_pool_params params = { 447 .order = 0, 448 .pool_size = NSIM_RING_SIZE, 449 .nid = NUMA_NO_NODE, 450 .dev = &napi->dev->dev, 451 .napi = napi, 452 .dma_dir = DMA_BIDIRECTIONAL, 453 .netdev = napi->dev, 454 }; 455 struct page_pool *pool; 456 457 pool = page_pool_create(¶ms); 458 if (IS_ERR(pool)) 459 return PTR_ERR(pool); 460 461 *p = pool; 462 return 0; 463 } 464 465 static int nsim_init_napi(struct netdevsim *ns) 466 { 467 struct net_device *dev = ns->netdev; 468 struct nsim_rq *rq; 469 int err, i; 470 471 for (i = 0; i < dev->num_rx_queues; i++) { 472 rq = ns->rq[i]; 473 474 netif_napi_add_config_locked(dev, &rq->napi, nsim_poll, i); 475 } 476 477 for (i = 0; i < dev->num_rx_queues; i++) { 478 rq = ns->rq[i]; 479 480 err = nsim_create_page_pool(&rq->page_pool, &rq->napi); 481 if (err) 482 goto err_pp_destroy; 483 } 484 485 return 0; 486 487 err_pp_destroy: 488 while (i--) { 489 page_pool_destroy(ns->rq[i]->page_pool); 490 ns->rq[i]->page_pool = NULL; 491 } 492 493 for (i = 0; i < dev->num_rx_queues; i++) 494 __netif_napi_del_locked(&ns->rq[i]->napi); 495 496 return err; 497 } 498 499 static enum hrtimer_restart nsim_napi_schedule(struct hrtimer *timer) 500 { 501 struct nsim_rq *rq; 502 503 rq = container_of(timer, struct nsim_rq, napi_timer); 504 napi_schedule(&rq->napi); 505 506 return HRTIMER_NORESTART; 507 } 508 509 static void nsim_rq_timer_init(struct nsim_rq *rq) 510 { 511 hrtimer_setup(&rq->napi_timer, nsim_napi_schedule, CLOCK_MONOTONIC, 512 HRTIMER_MODE_REL); 513 } 514 515 static void nsim_enable_napi(struct netdevsim *ns) 516 { 517 struct net_device *dev = ns->netdev; 518 int i; 519 520 for (i = 0; i < dev->num_rx_queues; i++) { 521 struct nsim_rq *rq = ns->rq[i]; 522 523 netif_queue_set_napi(dev, i, NETDEV_QUEUE_TYPE_RX, &rq->napi); 524 napi_enable_locked(&rq->napi); 525 } 526 } 527 528 static int nsim_open(struct net_device *dev) 529 { 530 struct netdevsim *ns = netdev_priv(dev); 531 int err; 532 533 netdev_assert_locked(dev); 534 535 err = nsim_init_napi(ns); 536 if (err) 537 return err; 538 539 nsim_enable_napi(ns); 540 541 return 0; 542 } 543 544 static void nsim_del_napi(struct netdevsim *ns) 545 { 546 struct net_device *dev = ns->netdev; 547 int i; 548 549 for (i = 0; i < dev->num_rx_queues; i++) { 550 struct nsim_rq *rq = ns->rq[i]; 551 552 napi_disable_locked(&rq->napi); 553 __netif_napi_del_locked(&rq->napi); 554 } 555 synchronize_net(); 556 557 for (i = 0; i < dev->num_rx_queues; i++) { 558 page_pool_destroy(ns->rq[i]->page_pool); 559 ns->rq[i]->page_pool = NULL; 560 } 561 } 562 563 static int nsim_stop(struct net_device *dev) 564 { 565 struct netdevsim *ns = netdev_priv(dev); 566 struct netdevsim *peer; 567 568 netdev_assert_locked(dev); 569 570 netif_carrier_off(dev); 571 peer = rtnl_dereference(ns->peer); 572 if (peer) 573 netif_carrier_off(peer->netdev); 574 575 nsim_del_napi(ns); 576 577 return 0; 578 } 579 580 static int nsim_shaper_set(struct net_shaper_binding *binding, 581 const struct net_shaper *shaper, 582 struct netlink_ext_ack *extack) 583 { 584 return 0; 585 } 586 587 static int nsim_shaper_del(struct net_shaper_binding *binding, 588 const struct net_shaper_handle *handle, 589 struct netlink_ext_ack *extack) 590 { 591 return 0; 592 } 593 594 static int nsim_shaper_group(struct net_shaper_binding *binding, 595 int leaves_count, 596 const struct net_shaper *leaves, 597 const struct net_shaper *root, 598 struct netlink_ext_ack *extack) 599 { 600 return 0; 601 } 602 603 static void nsim_shaper_cap(struct net_shaper_binding *binding, 604 enum net_shaper_scope scope, 605 unsigned long *flags) 606 { 607 *flags = ULONG_MAX; 608 } 609 610 static const struct net_shaper_ops nsim_shaper_ops = { 611 .set = nsim_shaper_set, 612 .delete = nsim_shaper_del, 613 .group = nsim_shaper_group, 614 .capabilities = nsim_shaper_cap, 615 }; 616 617 static const struct net_device_ops nsim_netdev_ops = { 618 .ndo_start_xmit = nsim_start_xmit, 619 .ndo_set_rx_mode = nsim_set_rx_mode, 620 .ndo_set_mac_address = eth_mac_addr, 621 .ndo_validate_addr = eth_validate_addr, 622 .ndo_change_mtu = nsim_change_mtu, 623 .ndo_set_vf_mac = nsim_set_vf_mac, 624 .ndo_set_vf_vlan = nsim_set_vf_vlan, 625 .ndo_set_vf_rate = nsim_set_vf_rate, 626 .ndo_set_vf_spoofchk = nsim_set_vf_spoofchk, 627 .ndo_set_vf_trust = nsim_set_vf_trust, 628 .ndo_get_vf_config = nsim_get_vf_config, 629 .ndo_set_vf_link_state = nsim_set_vf_link_state, 630 .ndo_set_vf_rss_query_en = nsim_set_vf_rss_query_en, 631 .ndo_setup_tc = nsim_setup_tc, 632 .ndo_set_features = nsim_set_features, 633 .ndo_get_iflink = nsim_get_iflink, 634 .ndo_bpf = nsim_bpf, 635 .ndo_open = nsim_open, 636 .ndo_stop = nsim_stop, 637 .net_shaper_ops = &nsim_shaper_ops, 638 }; 639 640 static const struct net_device_ops nsim_vf_netdev_ops = { 641 .ndo_start_xmit = nsim_start_xmit, 642 .ndo_set_rx_mode = nsim_set_rx_mode, 643 .ndo_set_mac_address = eth_mac_addr, 644 .ndo_validate_addr = eth_validate_addr, 645 .ndo_change_mtu = nsim_change_mtu, 646 .ndo_setup_tc = nsim_setup_tc, 647 .ndo_set_features = nsim_set_features, 648 }; 649 650 /* We don't have true per-queue stats, yet, so do some random fakery here. 651 * Only report stuff for queue 0. 652 */ 653 static void nsim_get_queue_stats_rx(struct net_device *dev, int idx, 654 struct netdev_queue_stats_rx *stats) 655 { 656 struct rtnl_link_stats64 rtstats = {}; 657 658 if (!idx) 659 dev_get_stats(dev, &rtstats); 660 661 stats->packets = rtstats.rx_packets - !!rtstats.rx_packets; 662 stats->bytes = rtstats.rx_bytes; 663 } 664 665 static void nsim_get_queue_stats_tx(struct net_device *dev, int idx, 666 struct netdev_queue_stats_tx *stats) 667 { 668 struct rtnl_link_stats64 rtstats = {}; 669 670 if (!idx) 671 dev_get_stats(dev, &rtstats); 672 673 stats->packets = rtstats.tx_packets - !!rtstats.tx_packets; 674 stats->bytes = rtstats.tx_bytes; 675 } 676 677 static void nsim_get_base_stats(struct net_device *dev, 678 struct netdev_queue_stats_rx *rx, 679 struct netdev_queue_stats_tx *tx) 680 { 681 struct rtnl_link_stats64 rtstats = {}; 682 683 dev_get_stats(dev, &rtstats); 684 685 rx->packets = !!rtstats.rx_packets; 686 rx->bytes = 0; 687 tx->packets = !!rtstats.tx_packets; 688 tx->bytes = 0; 689 } 690 691 static const struct netdev_stat_ops nsim_stat_ops = { 692 .get_queue_stats_tx = nsim_get_queue_stats_tx, 693 .get_queue_stats_rx = nsim_get_queue_stats_rx, 694 .get_base_stats = nsim_get_base_stats, 695 }; 696 697 static struct nsim_rq *nsim_queue_alloc(void) 698 { 699 struct nsim_rq *rq; 700 701 rq = kzalloc(sizeof(*rq), GFP_KERNEL_ACCOUNT); 702 if (!rq) 703 return NULL; 704 705 skb_queue_head_init(&rq->skb_queue); 706 nsim_rq_timer_init(rq); 707 return rq; 708 } 709 710 static void nsim_queue_free(struct net_device *dev, struct nsim_rq *rq) 711 { 712 hrtimer_cancel(&rq->napi_timer); 713 714 if (rq->skb_queue.qlen) { 715 local_bh_disable(); 716 dev_dstats_rx_dropped_add(dev, rq->skb_queue.qlen); 717 local_bh_enable(); 718 } 719 720 skb_queue_purge_reason(&rq->skb_queue, SKB_DROP_REASON_QUEUE_PURGE); 721 kfree(rq); 722 } 723 724 /* Queue reset mode is controlled by ns->rq_reset_mode. 725 * - normal - new NAPI new pool (old NAPI enabled when new added) 726 * - mode 1 - allocate new pool (NAPI is only disabled / enabled) 727 * - mode 2 - new NAPI new pool (old NAPI removed before new added) 728 * - mode 3 - new NAPI new pool (old NAPI disabled when new added) 729 */ 730 struct nsim_queue_mem { 731 struct nsim_rq *rq; 732 struct page_pool *pp; 733 }; 734 735 static int 736 nsim_queue_mem_alloc(struct net_device *dev, void *per_queue_mem, int idx) 737 { 738 struct nsim_queue_mem *qmem = per_queue_mem; 739 struct netdevsim *ns = netdev_priv(dev); 740 int err; 741 742 if (ns->rq_reset_mode > 3) 743 return -EINVAL; 744 745 if (ns->rq_reset_mode == 1) { 746 if (!netif_running(ns->netdev)) 747 return -ENETDOWN; 748 return nsim_create_page_pool(&qmem->pp, &ns->rq[idx]->napi); 749 } 750 751 qmem->rq = nsim_queue_alloc(); 752 if (!qmem->rq) 753 return -ENOMEM; 754 755 err = nsim_create_page_pool(&qmem->rq->page_pool, &qmem->rq->napi); 756 if (err) 757 goto err_free; 758 759 if (!ns->rq_reset_mode) 760 netif_napi_add_config_locked(dev, &qmem->rq->napi, nsim_poll, 761 idx); 762 763 return 0; 764 765 err_free: 766 nsim_queue_free(dev, qmem->rq); 767 return err; 768 } 769 770 static void nsim_queue_mem_free(struct net_device *dev, void *per_queue_mem) 771 { 772 struct nsim_queue_mem *qmem = per_queue_mem; 773 struct netdevsim *ns = netdev_priv(dev); 774 775 page_pool_destroy(qmem->pp); 776 if (qmem->rq) { 777 if (!ns->rq_reset_mode) 778 netif_napi_del_locked(&qmem->rq->napi); 779 page_pool_destroy(qmem->rq->page_pool); 780 nsim_queue_free(dev, qmem->rq); 781 } 782 } 783 784 static int 785 nsim_queue_start(struct net_device *dev, void *per_queue_mem, int idx) 786 { 787 struct nsim_queue_mem *qmem = per_queue_mem; 788 struct netdevsim *ns = netdev_priv(dev); 789 790 netdev_assert_locked(dev); 791 792 if (ns->rq_reset_mode == 1) { 793 ns->rq[idx]->page_pool = qmem->pp; 794 napi_enable_locked(&ns->rq[idx]->napi); 795 return 0; 796 } 797 798 /* netif_napi_add()/_del() should normally be called from alloc/free, 799 * here we want to test various call orders. 800 */ 801 if (ns->rq_reset_mode == 2) { 802 netif_napi_del_locked(&ns->rq[idx]->napi); 803 netif_napi_add_config_locked(dev, &qmem->rq->napi, nsim_poll, 804 idx); 805 } else if (ns->rq_reset_mode == 3) { 806 netif_napi_add_config_locked(dev, &qmem->rq->napi, nsim_poll, 807 idx); 808 netif_napi_del_locked(&ns->rq[idx]->napi); 809 } 810 811 ns->rq[idx] = qmem->rq; 812 napi_enable_locked(&ns->rq[idx]->napi); 813 814 return 0; 815 } 816 817 static int nsim_queue_stop(struct net_device *dev, void *per_queue_mem, int idx) 818 { 819 struct nsim_queue_mem *qmem = per_queue_mem; 820 struct netdevsim *ns = netdev_priv(dev); 821 822 netdev_assert_locked(dev); 823 824 napi_disable_locked(&ns->rq[idx]->napi); 825 826 if (ns->rq_reset_mode == 1) { 827 qmem->pp = ns->rq[idx]->page_pool; 828 page_pool_disable_direct_recycling(qmem->pp); 829 } else { 830 qmem->rq = ns->rq[idx]; 831 } 832 833 return 0; 834 } 835 836 static const struct netdev_queue_mgmt_ops nsim_queue_mgmt_ops = { 837 .ndo_queue_mem_size = sizeof(struct nsim_queue_mem), 838 .ndo_queue_mem_alloc = nsim_queue_mem_alloc, 839 .ndo_queue_mem_free = nsim_queue_mem_free, 840 .ndo_queue_start = nsim_queue_start, 841 .ndo_queue_stop = nsim_queue_stop, 842 }; 843 844 static ssize_t 845 nsim_qreset_write(struct file *file, const char __user *data, 846 size_t count, loff_t *ppos) 847 { 848 struct netdevsim *ns = file->private_data; 849 unsigned int queue, mode; 850 char buf[32]; 851 ssize_t ret; 852 853 if (count >= sizeof(buf)) 854 return -EINVAL; 855 if (copy_from_user(buf, data, count)) 856 return -EFAULT; 857 buf[count] = '\0'; 858 859 ret = sscanf(buf, "%u %u", &queue, &mode); 860 if (ret != 2) 861 return -EINVAL; 862 863 netdev_lock(ns->netdev); 864 if (queue >= ns->netdev->real_num_rx_queues) { 865 ret = -EINVAL; 866 goto exit_unlock; 867 } 868 869 ns->rq_reset_mode = mode; 870 ret = netdev_rx_queue_restart(ns->netdev, queue); 871 ns->rq_reset_mode = 0; 872 if (ret) 873 goto exit_unlock; 874 875 ret = count; 876 exit_unlock: 877 netdev_unlock(ns->netdev); 878 return ret; 879 } 880 881 static const struct file_operations nsim_qreset_fops = { 882 .open = simple_open, 883 .write = nsim_qreset_write, 884 .owner = THIS_MODULE, 885 }; 886 887 static ssize_t 888 nsim_pp_hold_read(struct file *file, char __user *data, 889 size_t count, loff_t *ppos) 890 { 891 struct netdevsim *ns = file->private_data; 892 char buf[3] = "n\n"; 893 894 if (ns->page) 895 buf[0] = 'y'; 896 897 return simple_read_from_buffer(data, count, ppos, buf, 2); 898 } 899 900 static ssize_t 901 nsim_pp_hold_write(struct file *file, const char __user *data, 902 size_t count, loff_t *ppos) 903 { 904 struct netdevsim *ns = file->private_data; 905 ssize_t ret; 906 bool val; 907 908 ret = kstrtobool_from_user(data, count, &val); 909 if (ret) 910 return ret; 911 912 rtnl_lock(); 913 ret = count; 914 if (val == !!ns->page) 915 goto exit; 916 917 if (!netif_running(ns->netdev) && val) { 918 ret = -ENETDOWN; 919 } else if (val) { 920 ns->page = page_pool_dev_alloc_pages(ns->rq[0]->page_pool); 921 if (!ns->page) 922 ret = -ENOMEM; 923 } else { 924 page_pool_put_full_page(pp_page_to_nmdesc(ns->page)->pp, 925 ns->page, false); 926 ns->page = NULL; 927 } 928 929 exit: 930 rtnl_unlock(); 931 return ret; 932 } 933 934 static const struct file_operations nsim_pp_hold_fops = { 935 .open = simple_open, 936 .read = nsim_pp_hold_read, 937 .write = nsim_pp_hold_write, 938 .llseek = generic_file_llseek, 939 .owner = THIS_MODULE, 940 }; 941 942 static void nsim_setup(struct net_device *dev) 943 { 944 ether_setup(dev); 945 eth_hw_addr_random(dev); 946 947 dev->flags &= ~IFF_MULTICAST; 948 dev->priv_flags |= IFF_LIVE_ADDR_CHANGE; 949 dev->features |= NETIF_F_HIGHDMA | 950 NETIF_F_SG | 951 NETIF_F_FRAGLIST | 952 NETIF_F_HW_CSUM | 953 NETIF_F_LRO | 954 NETIF_F_TSO; 955 dev->hw_features |= NETIF_F_HW_TC | 956 NETIF_F_SG | 957 NETIF_F_FRAGLIST | 958 NETIF_F_HW_CSUM | 959 NETIF_F_LRO | 960 NETIF_F_TSO; 961 dev->pcpu_stat_type = NETDEV_PCPU_STAT_DSTATS; 962 dev->max_mtu = ETH_MAX_MTU; 963 dev->xdp_features = NETDEV_XDP_ACT_BASIC | NETDEV_XDP_ACT_HW_OFFLOAD; 964 } 965 966 static int nsim_queue_init(struct netdevsim *ns) 967 { 968 struct net_device *dev = ns->netdev; 969 int i; 970 971 ns->rq = kcalloc(dev->num_rx_queues, sizeof(*ns->rq), 972 GFP_KERNEL_ACCOUNT); 973 if (!ns->rq) 974 return -ENOMEM; 975 976 for (i = 0; i < dev->num_rx_queues; i++) { 977 ns->rq[i] = nsim_queue_alloc(); 978 if (!ns->rq[i]) 979 goto err_free_prev; 980 } 981 982 return 0; 983 984 err_free_prev: 985 while (i--) 986 kfree(ns->rq[i]); 987 kfree(ns->rq); 988 return -ENOMEM; 989 } 990 991 static void nsim_queue_uninit(struct netdevsim *ns) 992 { 993 struct net_device *dev = ns->netdev; 994 int i; 995 996 for (i = 0; i < dev->num_rx_queues; i++) 997 nsim_queue_free(dev, ns->rq[i]); 998 999 kfree(ns->rq); 1000 ns->rq = NULL; 1001 } 1002 1003 static int nsim_init_netdevsim(struct netdevsim *ns) 1004 { 1005 struct mock_phc *phc; 1006 int err; 1007 1008 phc = mock_phc_create(&ns->nsim_bus_dev->dev); 1009 if (IS_ERR(phc)) 1010 return PTR_ERR(phc); 1011 1012 ns->phc = phc; 1013 ns->netdev->netdev_ops = &nsim_netdev_ops; 1014 ns->netdev->stat_ops = &nsim_stat_ops; 1015 ns->netdev->queue_mgmt_ops = &nsim_queue_mgmt_ops; 1016 netdev_lockdep_set_classes(ns->netdev); 1017 1018 err = nsim_udp_tunnels_info_create(ns->nsim_dev, ns->netdev); 1019 if (err) 1020 goto err_phc_destroy; 1021 1022 rtnl_lock(); 1023 err = nsim_queue_init(ns); 1024 if (err) 1025 goto err_utn_destroy; 1026 1027 err = nsim_bpf_init(ns); 1028 if (err) 1029 goto err_rq_destroy; 1030 1031 nsim_macsec_init(ns); 1032 nsim_ipsec_init(ns); 1033 1034 err = register_netdevice(ns->netdev); 1035 if (err) 1036 goto err_ipsec_teardown; 1037 rtnl_unlock(); 1038 1039 if (IS_ENABLED(CONFIG_DEBUG_NET)) { 1040 ns->nb.notifier_call = netdev_debug_event; 1041 if (register_netdevice_notifier_dev_net(ns->netdev, &ns->nb, 1042 &ns->nn)) 1043 ns->nb.notifier_call = NULL; 1044 } 1045 1046 return 0; 1047 1048 err_ipsec_teardown: 1049 nsim_ipsec_teardown(ns); 1050 nsim_macsec_teardown(ns); 1051 nsim_bpf_uninit(ns); 1052 err_rq_destroy: 1053 nsim_queue_uninit(ns); 1054 err_utn_destroy: 1055 rtnl_unlock(); 1056 nsim_udp_tunnels_info_destroy(ns->netdev); 1057 err_phc_destroy: 1058 mock_phc_destroy(ns->phc); 1059 return err; 1060 } 1061 1062 static int nsim_init_netdevsim_vf(struct netdevsim *ns) 1063 { 1064 int err; 1065 1066 ns->netdev->netdev_ops = &nsim_vf_netdev_ops; 1067 rtnl_lock(); 1068 err = register_netdevice(ns->netdev); 1069 rtnl_unlock(); 1070 return err; 1071 } 1072 1073 static void nsim_exit_netdevsim(struct netdevsim *ns) 1074 { 1075 nsim_udp_tunnels_info_destroy(ns->netdev); 1076 mock_phc_destroy(ns->phc); 1077 } 1078 1079 struct netdevsim *nsim_create(struct nsim_dev *nsim_dev, 1080 struct nsim_dev_port *nsim_dev_port, 1081 u8 perm_addr[ETH_ALEN]) 1082 { 1083 struct net_device *dev; 1084 struct netdevsim *ns; 1085 int err; 1086 1087 dev = alloc_netdev_mq(sizeof(*ns), "eth%d", NET_NAME_UNKNOWN, nsim_setup, 1088 nsim_dev->nsim_bus_dev->num_queues); 1089 if (!dev) 1090 return ERR_PTR(-ENOMEM); 1091 1092 if (perm_addr) 1093 memcpy(dev->perm_addr, perm_addr, ETH_ALEN); 1094 1095 dev_net_set(dev, nsim_dev_net(nsim_dev)); 1096 ns = netdev_priv(dev); 1097 ns->netdev = dev; 1098 ns->nsim_dev = nsim_dev; 1099 ns->nsim_dev_port = nsim_dev_port; 1100 ns->nsim_bus_dev = nsim_dev->nsim_bus_dev; 1101 SET_NETDEV_DEV(dev, &ns->nsim_bus_dev->dev); 1102 SET_NETDEV_DEVLINK_PORT(dev, &nsim_dev_port->devlink_port); 1103 nsim_ethtool_init(ns); 1104 if (nsim_dev_port_is_pf(nsim_dev_port)) 1105 err = nsim_init_netdevsim(ns); 1106 else 1107 err = nsim_init_netdevsim_vf(ns); 1108 if (err) 1109 goto err_free_netdev; 1110 1111 ns->pp_dfs = debugfs_create_file("pp_hold", 0600, nsim_dev_port->ddir, 1112 ns, &nsim_pp_hold_fops); 1113 ns->qr_dfs = debugfs_create_file("queue_reset", 0200, 1114 nsim_dev_port->ddir, ns, 1115 &nsim_qreset_fops); 1116 return ns; 1117 1118 err_free_netdev: 1119 free_netdev(dev); 1120 return ERR_PTR(err); 1121 } 1122 1123 void nsim_destroy(struct netdevsim *ns) 1124 { 1125 struct net_device *dev = ns->netdev; 1126 struct netdevsim *peer; 1127 1128 debugfs_remove(ns->qr_dfs); 1129 debugfs_remove(ns->pp_dfs); 1130 1131 if (ns->nb.notifier_call) 1132 unregister_netdevice_notifier_dev_net(ns->netdev, &ns->nb, 1133 &ns->nn); 1134 1135 rtnl_lock(); 1136 peer = rtnl_dereference(ns->peer); 1137 if (peer) 1138 RCU_INIT_POINTER(peer->peer, NULL); 1139 RCU_INIT_POINTER(ns->peer, NULL); 1140 unregister_netdevice(dev); 1141 if (nsim_dev_port_is_pf(ns->nsim_dev_port)) { 1142 nsim_macsec_teardown(ns); 1143 nsim_ipsec_teardown(ns); 1144 nsim_bpf_uninit(ns); 1145 nsim_queue_uninit(ns); 1146 } 1147 rtnl_unlock(); 1148 if (nsim_dev_port_is_pf(ns->nsim_dev_port)) 1149 nsim_exit_netdevsim(ns); 1150 1151 /* Put this intentionally late to exercise the orphaning path */ 1152 if (ns->page) { 1153 page_pool_put_full_page(pp_page_to_nmdesc(ns->page)->pp, 1154 ns->page, false); 1155 ns->page = NULL; 1156 } 1157 1158 free_netdev(dev); 1159 } 1160 1161 bool netdev_is_nsim(struct net_device *dev) 1162 { 1163 return dev->netdev_ops == &nsim_netdev_ops; 1164 } 1165 1166 static int nsim_validate(struct nlattr *tb[], struct nlattr *data[], 1167 struct netlink_ext_ack *extack) 1168 { 1169 NL_SET_ERR_MSG_MOD(extack, 1170 "Please use: echo \"[ID] [PORT_COUNT] [NUM_QUEUES]\" > /sys/bus/netdevsim/new_device"); 1171 return -EOPNOTSUPP; 1172 } 1173 1174 static struct rtnl_link_ops nsim_link_ops __read_mostly = { 1175 .kind = DRV_NAME, 1176 .validate = nsim_validate, 1177 }; 1178 1179 static int __init nsim_module_init(void) 1180 { 1181 int err; 1182 1183 err = nsim_dev_init(); 1184 if (err) 1185 return err; 1186 1187 err = nsim_bus_init(); 1188 if (err) 1189 goto err_dev_exit; 1190 1191 err = rtnl_link_register(&nsim_link_ops); 1192 if (err) 1193 goto err_bus_exit; 1194 1195 return 0; 1196 1197 err_bus_exit: 1198 nsim_bus_exit(); 1199 err_dev_exit: 1200 nsim_dev_exit(); 1201 return err; 1202 } 1203 1204 static void __exit nsim_module_exit(void) 1205 { 1206 rtnl_link_unregister(&nsim_link_ops); 1207 nsim_bus_exit(); 1208 nsim_dev_exit(); 1209 } 1210 1211 module_init(nsim_module_init); 1212 module_exit(nsim_module_exit); 1213 MODULE_LICENSE("GPL"); 1214 MODULE_DESCRIPTION("Simulated networking device for testing"); 1215 MODULE_ALIAS_RTNL_LINK(DRV_NAME); 1216