1 // SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB 2 /* 3 * Copyright 2015-2020 Amazon.com, Inc. or its affiliates. All rights reserved. 4 */ 5 6 #include <linux/ethtool.h> 7 #include <linux/pci.h> 8 9 #include "ena_netdev.h" 10 11 struct ena_stats { 12 char name[ETH_GSTRING_LEN]; 13 int stat_offset; 14 }; 15 16 #define ENA_STAT_ENA_COM_ENTRY(stat) { \ 17 .name = #stat, \ 18 .stat_offset = offsetof(struct ena_com_stats_admin, stat) / sizeof(u64) \ 19 } 20 21 #define ENA_STAT_ENTRY(stat, stat_type) { \ 22 .name = #stat, \ 23 .stat_offset = offsetof(struct ena_stats_##stat_type, stat) / sizeof(u64) \ 24 } 25 26 #define ENA_STAT_HW_ENTRY(stat, stat_type) { \ 27 .name = #stat, \ 28 .stat_offset = offsetof(struct ena_admin_##stat_type, stat) / sizeof(u64) \ 29 } 30 31 #define ENA_STAT_RX_ENTRY(stat) \ 32 ENA_STAT_ENTRY(stat, rx) 33 34 #define ENA_STAT_TX_ENTRY(stat) \ 35 ENA_STAT_ENTRY(stat, tx) 36 37 #define ENA_STAT_GLOBAL_ENTRY(stat) \ 38 ENA_STAT_ENTRY(stat, dev) 39 40 #define ENA_STAT_ENI_ENTRY(stat) \ 41 ENA_STAT_HW_ENTRY(stat, eni_stats) 42 43 static const struct ena_stats ena_stats_global_strings[] = { 44 ENA_STAT_GLOBAL_ENTRY(tx_timeout), 45 ENA_STAT_GLOBAL_ENTRY(suspend), 46 ENA_STAT_GLOBAL_ENTRY(resume), 47 ENA_STAT_GLOBAL_ENTRY(wd_expired), 48 ENA_STAT_GLOBAL_ENTRY(interface_up), 49 ENA_STAT_GLOBAL_ENTRY(interface_down), 50 ENA_STAT_GLOBAL_ENTRY(admin_q_pause), 51 }; 52 53 static const struct ena_stats ena_stats_eni_strings[] = { 54 ENA_STAT_ENI_ENTRY(bw_in_allowance_exceeded), 55 ENA_STAT_ENI_ENTRY(bw_out_allowance_exceeded), 56 ENA_STAT_ENI_ENTRY(pps_allowance_exceeded), 57 ENA_STAT_ENI_ENTRY(conntrack_allowance_exceeded), 58 ENA_STAT_ENI_ENTRY(linklocal_allowance_exceeded), 59 }; 60 61 static const struct ena_stats ena_stats_tx_strings[] = { 62 ENA_STAT_TX_ENTRY(cnt), 63 ENA_STAT_TX_ENTRY(bytes), 64 ENA_STAT_TX_ENTRY(queue_stop), 65 ENA_STAT_TX_ENTRY(queue_wakeup), 66 ENA_STAT_TX_ENTRY(dma_mapping_err), 67 ENA_STAT_TX_ENTRY(linearize), 68 ENA_STAT_TX_ENTRY(linearize_failed), 69 ENA_STAT_TX_ENTRY(napi_comp), 70 ENA_STAT_TX_ENTRY(tx_poll), 71 ENA_STAT_TX_ENTRY(doorbells), 72 ENA_STAT_TX_ENTRY(prepare_ctx_err), 73 ENA_STAT_TX_ENTRY(bad_req_id), 74 ENA_STAT_TX_ENTRY(llq_buffer_copy), 75 ENA_STAT_TX_ENTRY(missed_tx), 76 ENA_STAT_TX_ENTRY(unmask_interrupt), 77 }; 78 79 static const struct ena_stats ena_stats_rx_strings[] = { 80 ENA_STAT_RX_ENTRY(cnt), 81 ENA_STAT_RX_ENTRY(bytes), 82 ENA_STAT_RX_ENTRY(rx_copybreak_pkt), 83 ENA_STAT_RX_ENTRY(csum_good), 84 ENA_STAT_RX_ENTRY(refil_partial), 85 ENA_STAT_RX_ENTRY(bad_csum), 86 ENA_STAT_RX_ENTRY(page_alloc_fail), 87 ENA_STAT_RX_ENTRY(skb_alloc_fail), 88 ENA_STAT_RX_ENTRY(dma_mapping_err), 89 ENA_STAT_RX_ENTRY(bad_desc_num), 90 ENA_STAT_RX_ENTRY(bad_req_id), 91 ENA_STAT_RX_ENTRY(empty_rx_ring), 92 ENA_STAT_RX_ENTRY(csum_unchecked), 93 ENA_STAT_RX_ENTRY(xdp_aborted), 94 ENA_STAT_RX_ENTRY(xdp_drop), 95 ENA_STAT_RX_ENTRY(xdp_pass), 96 ENA_STAT_RX_ENTRY(xdp_tx), 97 ENA_STAT_RX_ENTRY(xdp_invalid), 98 }; 99 100 static const struct ena_stats ena_stats_ena_com_strings[] = { 101 ENA_STAT_ENA_COM_ENTRY(aborted_cmd), 102 ENA_STAT_ENA_COM_ENTRY(submitted_cmd), 103 ENA_STAT_ENA_COM_ENTRY(completed_cmd), 104 ENA_STAT_ENA_COM_ENTRY(out_of_space), 105 ENA_STAT_ENA_COM_ENTRY(no_completion), 106 }; 107 108 #define ENA_STATS_ARRAY_GLOBAL ARRAY_SIZE(ena_stats_global_strings) 109 #define ENA_STATS_ARRAY_TX ARRAY_SIZE(ena_stats_tx_strings) 110 #define ENA_STATS_ARRAY_RX ARRAY_SIZE(ena_stats_rx_strings) 111 #define ENA_STATS_ARRAY_ENA_COM ARRAY_SIZE(ena_stats_ena_com_strings) 112 #define ENA_STATS_ARRAY_ENI(adapter) \ 113 (ARRAY_SIZE(ena_stats_eni_strings) * (adapter)->eni_stats_supported) 114 115 static void ena_safe_update_stat(u64 *src, u64 *dst, 116 struct u64_stats_sync *syncp) 117 { 118 unsigned int start; 119 120 do { 121 start = u64_stats_fetch_begin_irq(syncp); 122 *(dst) = *src; 123 } while (u64_stats_fetch_retry_irq(syncp, start)); 124 } 125 126 static void ena_queue_stats(struct ena_adapter *adapter, u64 **data) 127 { 128 const struct ena_stats *ena_stats; 129 struct ena_ring *ring; 130 131 u64 *ptr; 132 int i, j; 133 134 for (i = 0; i < adapter->num_io_queues + adapter->xdp_num_queues; i++) { 135 /* Tx stats */ 136 ring = &adapter->tx_ring[i]; 137 138 for (j = 0; j < ENA_STATS_ARRAY_TX; j++) { 139 ena_stats = &ena_stats_tx_strings[j]; 140 141 ptr = (u64 *)&ring->tx_stats + ena_stats->stat_offset; 142 143 ena_safe_update_stat(ptr, (*data)++, &ring->syncp); 144 } 145 /* XDP TX queues don't have a RX queue counterpart */ 146 if (!ENA_IS_XDP_INDEX(adapter, i)) { 147 /* Rx stats */ 148 ring = &adapter->rx_ring[i]; 149 150 for (j = 0; j < ENA_STATS_ARRAY_RX; j++) { 151 ena_stats = &ena_stats_rx_strings[j]; 152 153 ptr = (u64 *)&ring->rx_stats + 154 ena_stats->stat_offset; 155 156 ena_safe_update_stat(ptr, (*data)++, &ring->syncp); 157 } 158 } 159 } 160 } 161 162 static void ena_dev_admin_queue_stats(struct ena_adapter *adapter, u64 **data) 163 { 164 const struct ena_stats *ena_stats; 165 u64 *ptr; 166 int i; 167 168 for (i = 0; i < ENA_STATS_ARRAY_ENA_COM; i++) { 169 ena_stats = &ena_stats_ena_com_strings[i]; 170 171 ptr = (u64 *)&adapter->ena_dev->admin_queue.stats + 172 ena_stats->stat_offset; 173 174 *(*data)++ = *ptr; 175 } 176 } 177 178 static void ena_get_stats(struct ena_adapter *adapter, 179 u64 *data, 180 bool eni_stats_needed) 181 { 182 const struct ena_stats *ena_stats; 183 u64 *ptr; 184 int i; 185 186 for (i = 0; i < ENA_STATS_ARRAY_GLOBAL; i++) { 187 ena_stats = &ena_stats_global_strings[i]; 188 189 ptr = (u64 *)&adapter->dev_stats + ena_stats->stat_offset; 190 191 ena_safe_update_stat(ptr, data++, &adapter->syncp); 192 } 193 194 if (eni_stats_needed) { 195 ena_update_hw_stats(adapter); 196 for (i = 0; i < ENA_STATS_ARRAY_ENI(adapter); i++) { 197 ena_stats = &ena_stats_eni_strings[i]; 198 199 ptr = (u64 *)&adapter->eni_stats + 200 ena_stats->stat_offset; 201 202 ena_safe_update_stat(ptr, data++, &adapter->syncp); 203 } 204 } 205 206 ena_queue_stats(adapter, &data); 207 ena_dev_admin_queue_stats(adapter, &data); 208 } 209 210 static void ena_get_ethtool_stats(struct net_device *netdev, 211 struct ethtool_stats *stats, 212 u64 *data) 213 { 214 struct ena_adapter *adapter = netdev_priv(netdev); 215 216 ena_get_stats(adapter, data, adapter->eni_stats_supported); 217 } 218 219 static int ena_get_sw_stats_count(struct ena_adapter *adapter) 220 { 221 return adapter->num_io_queues * (ENA_STATS_ARRAY_TX + ENA_STATS_ARRAY_RX) 222 + adapter->xdp_num_queues * ENA_STATS_ARRAY_TX 223 + ENA_STATS_ARRAY_GLOBAL + ENA_STATS_ARRAY_ENA_COM; 224 } 225 226 static int ena_get_hw_stats_count(struct ena_adapter *adapter) 227 { 228 return ENA_STATS_ARRAY_ENI(adapter); 229 } 230 231 int ena_get_sset_count(struct net_device *netdev, int sset) 232 { 233 struct ena_adapter *adapter = netdev_priv(netdev); 234 235 if (sset != ETH_SS_STATS) 236 return -EOPNOTSUPP; 237 238 return ena_get_sw_stats_count(adapter) + ena_get_hw_stats_count(adapter); 239 } 240 241 static void ena_queue_strings(struct ena_adapter *adapter, u8 **data) 242 { 243 const struct ena_stats *ena_stats; 244 bool is_xdp; 245 int i, j; 246 247 for (i = 0; i < adapter->num_io_queues + adapter->xdp_num_queues; i++) { 248 is_xdp = ENA_IS_XDP_INDEX(adapter, i); 249 /* Tx stats */ 250 for (j = 0; j < ENA_STATS_ARRAY_TX; j++) { 251 ena_stats = &ena_stats_tx_strings[j]; 252 253 snprintf(*data, ETH_GSTRING_LEN, 254 "queue_%u_%s_%s", i, 255 is_xdp ? "xdp_tx" : "tx", ena_stats->name); 256 (*data) += ETH_GSTRING_LEN; 257 } 258 259 if (!is_xdp) { 260 /* RX stats, in XDP there isn't a RX queue 261 * counterpart 262 */ 263 for (j = 0; j < ENA_STATS_ARRAY_RX; j++) { 264 ena_stats = &ena_stats_rx_strings[j]; 265 266 snprintf(*data, ETH_GSTRING_LEN, 267 "queue_%u_rx_%s", i, ena_stats->name); 268 (*data) += ETH_GSTRING_LEN; 269 } 270 } 271 } 272 } 273 274 static void ena_com_dev_strings(u8 **data) 275 { 276 const struct ena_stats *ena_stats; 277 int i; 278 279 for (i = 0; i < ENA_STATS_ARRAY_ENA_COM; i++) { 280 ena_stats = &ena_stats_ena_com_strings[i]; 281 282 snprintf(*data, ETH_GSTRING_LEN, 283 "ena_admin_q_%s", ena_stats->name); 284 (*data) += ETH_GSTRING_LEN; 285 } 286 } 287 288 static void ena_get_strings(struct ena_adapter *adapter, 289 u8 *data, 290 bool eni_stats_needed) 291 { 292 const struct ena_stats *ena_stats; 293 int i; 294 295 for (i = 0; i < ENA_STATS_ARRAY_GLOBAL; i++) { 296 ena_stats = &ena_stats_global_strings[i]; 297 memcpy(data, ena_stats->name, ETH_GSTRING_LEN); 298 data += ETH_GSTRING_LEN; 299 } 300 301 if (eni_stats_needed) { 302 for (i = 0; i < ENA_STATS_ARRAY_ENI(adapter); i++) { 303 ena_stats = &ena_stats_eni_strings[i]; 304 memcpy(data, ena_stats->name, ETH_GSTRING_LEN); 305 data += ETH_GSTRING_LEN; 306 } 307 } 308 309 ena_queue_strings(adapter, &data); 310 ena_com_dev_strings(&data); 311 } 312 313 static void ena_get_ethtool_strings(struct net_device *netdev, 314 u32 sset, 315 u8 *data) 316 { 317 struct ena_adapter *adapter = netdev_priv(netdev); 318 319 if (sset != ETH_SS_STATS) 320 return; 321 322 ena_get_strings(adapter, data, adapter->eni_stats_supported); 323 } 324 325 static int ena_get_link_ksettings(struct net_device *netdev, 326 struct ethtool_link_ksettings *link_ksettings) 327 { 328 struct ena_adapter *adapter = netdev_priv(netdev); 329 struct ena_com_dev *ena_dev = adapter->ena_dev; 330 struct ena_admin_get_feature_link_desc *link; 331 struct ena_admin_get_feat_resp feat_resp; 332 int rc; 333 334 rc = ena_com_get_link_params(ena_dev, &feat_resp); 335 if (rc) 336 return rc; 337 338 link = &feat_resp.u.link; 339 link_ksettings->base.speed = link->speed; 340 341 if (link->flags & ENA_ADMIN_GET_FEATURE_LINK_DESC_AUTONEG_MASK) { 342 ethtool_link_ksettings_add_link_mode(link_ksettings, 343 supported, Autoneg); 344 ethtool_link_ksettings_add_link_mode(link_ksettings, 345 supported, Autoneg); 346 } 347 348 link_ksettings->base.autoneg = 349 (link->flags & ENA_ADMIN_GET_FEATURE_LINK_DESC_AUTONEG_MASK) ? 350 AUTONEG_ENABLE : AUTONEG_DISABLE; 351 352 link_ksettings->base.duplex = DUPLEX_FULL; 353 354 return 0; 355 } 356 357 static int ena_get_coalesce(struct net_device *net_dev, 358 struct ethtool_coalesce *coalesce) 359 { 360 struct ena_adapter *adapter = netdev_priv(net_dev); 361 struct ena_com_dev *ena_dev = adapter->ena_dev; 362 363 if (!ena_com_interrupt_moderation_supported(ena_dev)) 364 return -EOPNOTSUPP; 365 366 coalesce->tx_coalesce_usecs = 367 ena_com_get_nonadaptive_moderation_interval_tx(ena_dev) * 368 ena_dev->intr_delay_resolution; 369 370 coalesce->rx_coalesce_usecs = 371 ena_com_get_nonadaptive_moderation_interval_rx(ena_dev) 372 * ena_dev->intr_delay_resolution; 373 374 coalesce->use_adaptive_rx_coalesce = 375 ena_com_get_adaptive_moderation_enabled(ena_dev); 376 377 return 0; 378 } 379 380 static void ena_update_tx_rings_nonadaptive_intr_moderation(struct ena_adapter *adapter) 381 { 382 unsigned int val; 383 int i; 384 385 val = ena_com_get_nonadaptive_moderation_interval_tx(adapter->ena_dev); 386 387 for (i = 0; i < adapter->num_io_queues; i++) 388 adapter->tx_ring[i].smoothed_interval = val; 389 } 390 391 static void ena_update_rx_rings_nonadaptive_intr_moderation(struct ena_adapter *adapter) 392 { 393 unsigned int val; 394 int i; 395 396 val = ena_com_get_nonadaptive_moderation_interval_rx(adapter->ena_dev); 397 398 for (i = 0; i < adapter->num_io_queues; i++) 399 adapter->rx_ring[i].smoothed_interval = val; 400 } 401 402 static int ena_set_coalesce(struct net_device *net_dev, 403 struct ethtool_coalesce *coalesce) 404 { 405 struct ena_adapter *adapter = netdev_priv(net_dev); 406 struct ena_com_dev *ena_dev = adapter->ena_dev; 407 int rc; 408 409 if (!ena_com_interrupt_moderation_supported(ena_dev)) 410 return -EOPNOTSUPP; 411 412 rc = ena_com_update_nonadaptive_moderation_interval_tx(ena_dev, 413 coalesce->tx_coalesce_usecs); 414 if (rc) 415 return rc; 416 417 ena_update_tx_rings_nonadaptive_intr_moderation(adapter); 418 419 rc = ena_com_update_nonadaptive_moderation_interval_rx(ena_dev, 420 coalesce->rx_coalesce_usecs); 421 if (rc) 422 return rc; 423 424 ena_update_rx_rings_nonadaptive_intr_moderation(adapter); 425 426 if (coalesce->use_adaptive_rx_coalesce && 427 !ena_com_get_adaptive_moderation_enabled(ena_dev)) 428 ena_com_enable_adaptive_moderation(ena_dev); 429 430 if (!coalesce->use_adaptive_rx_coalesce && 431 ena_com_get_adaptive_moderation_enabled(ena_dev)) 432 ena_com_disable_adaptive_moderation(ena_dev); 433 434 return 0; 435 } 436 437 static u32 ena_get_msglevel(struct net_device *netdev) 438 { 439 struct ena_adapter *adapter = netdev_priv(netdev); 440 441 return adapter->msg_enable; 442 } 443 444 static void ena_set_msglevel(struct net_device *netdev, u32 value) 445 { 446 struct ena_adapter *adapter = netdev_priv(netdev); 447 448 adapter->msg_enable = value; 449 } 450 451 static void ena_get_drvinfo(struct net_device *dev, 452 struct ethtool_drvinfo *info) 453 { 454 struct ena_adapter *adapter = netdev_priv(dev); 455 456 strlcpy(info->driver, DRV_MODULE_NAME, sizeof(info->driver)); 457 strlcpy(info->bus_info, pci_name(adapter->pdev), 458 sizeof(info->bus_info)); 459 } 460 461 static void ena_get_ringparam(struct net_device *netdev, 462 struct ethtool_ringparam *ring) 463 { 464 struct ena_adapter *adapter = netdev_priv(netdev); 465 466 ring->tx_max_pending = adapter->max_tx_ring_size; 467 ring->rx_max_pending = adapter->max_rx_ring_size; 468 ring->tx_pending = adapter->tx_ring[0].ring_size; 469 ring->rx_pending = adapter->rx_ring[0].ring_size; 470 } 471 472 static int ena_set_ringparam(struct net_device *netdev, 473 struct ethtool_ringparam *ring) 474 { 475 struct ena_adapter *adapter = netdev_priv(netdev); 476 u32 new_tx_size, new_rx_size; 477 478 new_tx_size = ring->tx_pending < ENA_MIN_RING_SIZE ? 479 ENA_MIN_RING_SIZE : ring->tx_pending; 480 new_tx_size = rounddown_pow_of_two(new_tx_size); 481 482 new_rx_size = ring->rx_pending < ENA_MIN_RING_SIZE ? 483 ENA_MIN_RING_SIZE : ring->rx_pending; 484 new_rx_size = rounddown_pow_of_two(new_rx_size); 485 486 if (new_tx_size == adapter->requested_tx_ring_size && 487 new_rx_size == adapter->requested_rx_ring_size) 488 return 0; 489 490 return ena_update_queue_sizes(adapter, new_tx_size, new_rx_size); 491 } 492 493 static u32 ena_flow_hash_to_flow_type(u16 hash_fields) 494 { 495 u32 data = 0; 496 497 if (hash_fields & ENA_ADMIN_RSS_L2_DA) 498 data |= RXH_L2DA; 499 500 if (hash_fields & ENA_ADMIN_RSS_L3_DA) 501 data |= RXH_IP_DST; 502 503 if (hash_fields & ENA_ADMIN_RSS_L3_SA) 504 data |= RXH_IP_SRC; 505 506 if (hash_fields & ENA_ADMIN_RSS_L4_DP) 507 data |= RXH_L4_B_2_3; 508 509 if (hash_fields & ENA_ADMIN_RSS_L4_SP) 510 data |= RXH_L4_B_0_1; 511 512 return data; 513 } 514 515 static u16 ena_flow_data_to_flow_hash(u32 hash_fields) 516 { 517 u16 data = 0; 518 519 if (hash_fields & RXH_L2DA) 520 data |= ENA_ADMIN_RSS_L2_DA; 521 522 if (hash_fields & RXH_IP_DST) 523 data |= ENA_ADMIN_RSS_L3_DA; 524 525 if (hash_fields & RXH_IP_SRC) 526 data |= ENA_ADMIN_RSS_L3_SA; 527 528 if (hash_fields & RXH_L4_B_2_3) 529 data |= ENA_ADMIN_RSS_L4_DP; 530 531 if (hash_fields & RXH_L4_B_0_1) 532 data |= ENA_ADMIN_RSS_L4_SP; 533 534 return data; 535 } 536 537 static int ena_get_rss_hash(struct ena_com_dev *ena_dev, 538 struct ethtool_rxnfc *cmd) 539 { 540 enum ena_admin_flow_hash_proto proto; 541 u16 hash_fields; 542 int rc; 543 544 cmd->data = 0; 545 546 switch (cmd->flow_type) { 547 case TCP_V4_FLOW: 548 proto = ENA_ADMIN_RSS_TCP4; 549 break; 550 case UDP_V4_FLOW: 551 proto = ENA_ADMIN_RSS_UDP4; 552 break; 553 case TCP_V6_FLOW: 554 proto = ENA_ADMIN_RSS_TCP6; 555 break; 556 case UDP_V6_FLOW: 557 proto = ENA_ADMIN_RSS_UDP6; 558 break; 559 case IPV4_FLOW: 560 proto = ENA_ADMIN_RSS_IP4; 561 break; 562 case IPV6_FLOW: 563 proto = ENA_ADMIN_RSS_IP6; 564 break; 565 case ETHER_FLOW: 566 proto = ENA_ADMIN_RSS_NOT_IP; 567 break; 568 case AH_V4_FLOW: 569 case ESP_V4_FLOW: 570 case AH_V6_FLOW: 571 case ESP_V6_FLOW: 572 case SCTP_V4_FLOW: 573 case AH_ESP_V4_FLOW: 574 return -EOPNOTSUPP; 575 default: 576 return -EINVAL; 577 } 578 579 rc = ena_com_get_hash_ctrl(ena_dev, proto, &hash_fields); 580 if (rc) 581 return rc; 582 583 cmd->data = ena_flow_hash_to_flow_type(hash_fields); 584 585 return 0; 586 } 587 588 static int ena_set_rss_hash(struct ena_com_dev *ena_dev, 589 struct ethtool_rxnfc *cmd) 590 { 591 enum ena_admin_flow_hash_proto proto; 592 u16 hash_fields; 593 594 switch (cmd->flow_type) { 595 case TCP_V4_FLOW: 596 proto = ENA_ADMIN_RSS_TCP4; 597 break; 598 case UDP_V4_FLOW: 599 proto = ENA_ADMIN_RSS_UDP4; 600 break; 601 case TCP_V6_FLOW: 602 proto = ENA_ADMIN_RSS_TCP6; 603 break; 604 case UDP_V6_FLOW: 605 proto = ENA_ADMIN_RSS_UDP6; 606 break; 607 case IPV4_FLOW: 608 proto = ENA_ADMIN_RSS_IP4; 609 break; 610 case IPV6_FLOW: 611 proto = ENA_ADMIN_RSS_IP6; 612 break; 613 case ETHER_FLOW: 614 proto = ENA_ADMIN_RSS_NOT_IP; 615 break; 616 case AH_V4_FLOW: 617 case ESP_V4_FLOW: 618 case AH_V6_FLOW: 619 case ESP_V6_FLOW: 620 case SCTP_V4_FLOW: 621 case AH_ESP_V4_FLOW: 622 return -EOPNOTSUPP; 623 default: 624 return -EINVAL; 625 } 626 627 hash_fields = ena_flow_data_to_flow_hash(cmd->data); 628 629 return ena_com_fill_hash_ctrl(ena_dev, proto, hash_fields); 630 } 631 632 static int ena_set_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *info) 633 { 634 struct ena_adapter *adapter = netdev_priv(netdev); 635 int rc = 0; 636 637 switch (info->cmd) { 638 case ETHTOOL_SRXFH: 639 rc = ena_set_rss_hash(adapter->ena_dev, info); 640 break; 641 case ETHTOOL_SRXCLSRLDEL: 642 case ETHTOOL_SRXCLSRLINS: 643 default: 644 netif_err(adapter, drv, netdev, 645 "Command parameter %d is not supported\n", info->cmd); 646 rc = -EOPNOTSUPP; 647 } 648 649 return rc; 650 } 651 652 static int ena_get_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *info, 653 u32 *rules) 654 { 655 struct ena_adapter *adapter = netdev_priv(netdev); 656 int rc = 0; 657 658 switch (info->cmd) { 659 case ETHTOOL_GRXRINGS: 660 info->data = adapter->num_io_queues; 661 rc = 0; 662 break; 663 case ETHTOOL_GRXFH: 664 rc = ena_get_rss_hash(adapter->ena_dev, info); 665 break; 666 case ETHTOOL_GRXCLSRLCNT: 667 case ETHTOOL_GRXCLSRULE: 668 case ETHTOOL_GRXCLSRLALL: 669 default: 670 netif_err(adapter, drv, netdev, 671 "Command parameter %d is not supported\n", info->cmd); 672 rc = -EOPNOTSUPP; 673 } 674 675 return rc; 676 } 677 678 static u32 ena_get_rxfh_indir_size(struct net_device *netdev) 679 { 680 return ENA_RX_RSS_TABLE_SIZE; 681 } 682 683 static u32 ena_get_rxfh_key_size(struct net_device *netdev) 684 { 685 return ENA_HASH_KEY_SIZE; 686 } 687 688 static int ena_indirection_table_set(struct ena_adapter *adapter, 689 const u32 *indir) 690 { 691 struct ena_com_dev *ena_dev = adapter->ena_dev; 692 int i, rc; 693 694 for (i = 0; i < ENA_RX_RSS_TABLE_SIZE; i++) { 695 rc = ena_com_indirect_table_fill_entry(ena_dev, 696 i, 697 ENA_IO_RXQ_IDX(indir[i])); 698 if (unlikely(rc)) { 699 netif_err(adapter, drv, adapter->netdev, 700 "Cannot fill indirect table (index is too large)\n"); 701 return rc; 702 } 703 } 704 705 rc = ena_com_indirect_table_set(ena_dev); 706 if (rc) { 707 netif_err(adapter, drv, adapter->netdev, 708 "Cannot set indirect table\n"); 709 return rc == -EPERM ? -EOPNOTSUPP : rc; 710 } 711 return rc; 712 } 713 714 static int ena_indirection_table_get(struct ena_adapter *adapter, u32 *indir) 715 { 716 struct ena_com_dev *ena_dev = adapter->ena_dev; 717 int i, rc; 718 719 if (!indir) 720 return 0; 721 722 rc = ena_com_indirect_table_get(ena_dev, indir); 723 if (rc) 724 return rc; 725 726 /* Our internal representation of the indices is: even indices 727 * for Tx and uneven indices for Rx. We need to convert the Rx 728 * indices to be consecutive 729 */ 730 for (i = 0; i < ENA_RX_RSS_TABLE_SIZE; i++) 731 indir[i] = ENA_IO_RXQ_IDX_TO_COMBINED_IDX(indir[i]); 732 733 return rc; 734 } 735 736 static int ena_get_rxfh(struct net_device *netdev, u32 *indir, u8 *key, 737 u8 *hfunc) 738 { 739 struct ena_adapter *adapter = netdev_priv(netdev); 740 enum ena_admin_hash_functions ena_func; 741 u8 func; 742 int rc; 743 744 rc = ena_indirection_table_get(adapter, indir); 745 if (rc) 746 return rc; 747 748 /* We call this function in order to check if the device 749 * supports getting/setting the hash function. 750 */ 751 rc = ena_com_get_hash_function(adapter->ena_dev, &ena_func); 752 if (rc) { 753 if (rc == -EOPNOTSUPP) 754 rc = 0; 755 756 return rc; 757 } 758 759 rc = ena_com_get_hash_key(adapter->ena_dev, key); 760 if (rc) 761 return rc; 762 763 switch (ena_func) { 764 case ENA_ADMIN_TOEPLITZ: 765 func = ETH_RSS_HASH_TOP; 766 break; 767 case ENA_ADMIN_CRC32: 768 func = ETH_RSS_HASH_CRC32; 769 break; 770 default: 771 netif_err(adapter, drv, netdev, 772 "Command parameter is not supported\n"); 773 return -EOPNOTSUPP; 774 } 775 776 if (hfunc) 777 *hfunc = func; 778 779 return 0; 780 } 781 782 static int ena_set_rxfh(struct net_device *netdev, const u32 *indir, 783 const u8 *key, const u8 hfunc) 784 { 785 struct ena_adapter *adapter = netdev_priv(netdev); 786 struct ena_com_dev *ena_dev = adapter->ena_dev; 787 enum ena_admin_hash_functions func = 0; 788 int rc; 789 790 if (indir) { 791 rc = ena_indirection_table_set(adapter, indir); 792 if (rc) 793 return rc; 794 } 795 796 switch (hfunc) { 797 case ETH_RSS_HASH_NO_CHANGE: 798 func = ena_com_get_current_hash_function(ena_dev); 799 break; 800 case ETH_RSS_HASH_TOP: 801 func = ENA_ADMIN_TOEPLITZ; 802 break; 803 case ETH_RSS_HASH_CRC32: 804 func = ENA_ADMIN_CRC32; 805 break; 806 default: 807 netif_err(adapter, drv, netdev, "Unsupported hfunc %d\n", 808 hfunc); 809 return -EOPNOTSUPP; 810 } 811 812 if (key || func) { 813 rc = ena_com_fill_hash_function(ena_dev, func, key, 814 ENA_HASH_KEY_SIZE, 815 0xFFFFFFFF); 816 if (unlikely(rc)) { 817 netif_err(adapter, drv, netdev, "Cannot fill key\n"); 818 return rc == -EPERM ? -EOPNOTSUPP : rc; 819 } 820 } 821 822 return 0; 823 } 824 825 static void ena_get_channels(struct net_device *netdev, 826 struct ethtool_channels *channels) 827 { 828 struct ena_adapter *adapter = netdev_priv(netdev); 829 830 channels->max_combined = adapter->max_num_io_queues; 831 channels->combined_count = adapter->num_io_queues; 832 } 833 834 static int ena_set_channels(struct net_device *netdev, 835 struct ethtool_channels *channels) 836 { 837 struct ena_adapter *adapter = netdev_priv(netdev); 838 u32 count = channels->combined_count; 839 /* The check for max value is already done in ethtool */ 840 if (count < ENA_MIN_NUM_IO_QUEUES || 841 (ena_xdp_present(adapter) && 842 !ena_xdp_legal_queue_count(adapter, channels->combined_count))) 843 return -EINVAL; 844 845 return ena_update_queue_count(adapter, count); 846 } 847 848 static int ena_get_tunable(struct net_device *netdev, 849 const struct ethtool_tunable *tuna, void *data) 850 { 851 struct ena_adapter *adapter = netdev_priv(netdev); 852 int ret = 0; 853 854 switch (tuna->id) { 855 case ETHTOOL_RX_COPYBREAK: 856 *(u32 *)data = adapter->rx_copybreak; 857 break; 858 default: 859 ret = -EINVAL; 860 break; 861 } 862 863 return ret; 864 } 865 866 static int ena_set_tunable(struct net_device *netdev, 867 const struct ethtool_tunable *tuna, 868 const void *data) 869 { 870 struct ena_adapter *adapter = netdev_priv(netdev); 871 int ret = 0; 872 u32 len; 873 874 switch (tuna->id) { 875 case ETHTOOL_RX_COPYBREAK: 876 len = *(u32 *)data; 877 if (len > adapter->netdev->mtu) { 878 ret = -EINVAL; 879 break; 880 } 881 adapter->rx_copybreak = len; 882 break; 883 default: 884 ret = -EINVAL; 885 break; 886 } 887 888 return ret; 889 } 890 891 static const struct ethtool_ops ena_ethtool_ops = { 892 .supported_coalesce_params = ETHTOOL_COALESCE_USECS | 893 ETHTOOL_COALESCE_USE_ADAPTIVE_RX, 894 .get_link_ksettings = ena_get_link_ksettings, 895 .get_drvinfo = ena_get_drvinfo, 896 .get_msglevel = ena_get_msglevel, 897 .set_msglevel = ena_set_msglevel, 898 .get_link = ethtool_op_get_link, 899 .get_coalesce = ena_get_coalesce, 900 .set_coalesce = ena_set_coalesce, 901 .get_ringparam = ena_get_ringparam, 902 .set_ringparam = ena_set_ringparam, 903 .get_sset_count = ena_get_sset_count, 904 .get_strings = ena_get_ethtool_strings, 905 .get_ethtool_stats = ena_get_ethtool_stats, 906 .get_rxnfc = ena_get_rxnfc, 907 .set_rxnfc = ena_set_rxnfc, 908 .get_rxfh_indir_size = ena_get_rxfh_indir_size, 909 .get_rxfh_key_size = ena_get_rxfh_key_size, 910 .get_rxfh = ena_get_rxfh, 911 .set_rxfh = ena_set_rxfh, 912 .get_channels = ena_get_channels, 913 .set_channels = ena_set_channels, 914 .get_tunable = ena_get_tunable, 915 .set_tunable = ena_set_tunable, 916 .get_ts_info = ethtool_op_get_ts_info, 917 }; 918 919 void ena_set_ethtool_ops(struct net_device *netdev) 920 { 921 netdev->ethtool_ops = &ena_ethtool_ops; 922 } 923 924 static void ena_dump_stats_ex(struct ena_adapter *adapter, u8 *buf) 925 { 926 struct net_device *netdev = adapter->netdev; 927 u8 *strings_buf; 928 u64 *data_buf; 929 int strings_num; 930 int i, rc; 931 932 strings_num = ena_get_sw_stats_count(adapter); 933 if (strings_num <= 0) { 934 netif_err(adapter, drv, netdev, "Can't get stats num\n"); 935 return; 936 } 937 938 strings_buf = devm_kcalloc(&adapter->pdev->dev, 939 ETH_GSTRING_LEN, strings_num, 940 GFP_ATOMIC); 941 if (!strings_buf) { 942 netif_err(adapter, drv, netdev, 943 "Failed to allocate strings_buf\n"); 944 return; 945 } 946 947 data_buf = devm_kcalloc(&adapter->pdev->dev, 948 strings_num, sizeof(u64), 949 GFP_ATOMIC); 950 if (!data_buf) { 951 netif_err(adapter, drv, netdev, 952 "Failed to allocate data buf\n"); 953 devm_kfree(&adapter->pdev->dev, strings_buf); 954 return; 955 } 956 957 ena_get_strings(adapter, strings_buf, false); 958 ena_get_stats(adapter, data_buf, false); 959 960 /* If there is a buffer, dump stats, otherwise print them to dmesg */ 961 if (buf) 962 for (i = 0; i < strings_num; i++) { 963 rc = snprintf(buf, ETH_GSTRING_LEN + sizeof(u64), 964 "%s %llu\n", 965 strings_buf + i * ETH_GSTRING_LEN, 966 data_buf[i]); 967 buf += rc; 968 } 969 else 970 for (i = 0; i < strings_num; i++) 971 netif_err(adapter, drv, netdev, "%s: %llu\n", 972 strings_buf + i * ETH_GSTRING_LEN, 973 data_buf[i]); 974 975 devm_kfree(&adapter->pdev->dev, strings_buf); 976 devm_kfree(&adapter->pdev->dev, data_buf); 977 } 978 979 void ena_dump_stats_to_buf(struct ena_adapter *adapter, u8 *buf) 980 { 981 if (!buf) 982 return; 983 984 ena_dump_stats_ex(adapter, buf); 985 } 986 987 void ena_dump_stats_to_dmesg(struct ena_adapter *adapter) 988 { 989 ena_dump_stats_ex(adapter, NULL); 990 } 991