1 /* Framework for configuring and reading PHY devices 2 * Based on code in sungem_phy.c and gianfar_phy.c 3 * 4 * Author: Andy Fleming 5 * 6 * Copyright (c) 2004 Freescale Semiconductor, Inc. 7 * Copyright (c) 2006, 2007 Maciej W. Rozycki 8 * 9 * This program is free software; you can redistribute it and/or modify it 10 * under the terms of the GNU General Public License as published by the 11 * Free Software Foundation; either version 2 of the License, or (at your 12 * option) any later version. 13 * 14 */ 15 16 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 17 18 #include <linux/kernel.h> 19 #include <linux/string.h> 20 #include <linux/errno.h> 21 #include <linux/unistd.h> 22 #include <linux/interrupt.h> 23 #include <linux/delay.h> 24 #include <linux/netdevice.h> 25 #include <linux/etherdevice.h> 26 #include <linux/skbuff.h> 27 #include <linux/mm.h> 28 #include <linux/module.h> 29 #include <linux/mii.h> 30 #include <linux/ethtool.h> 31 #include <linux/phy.h> 32 #include <linux/phy_led_triggers.h> 33 #include <linux/timer.h> 34 #include <linux/workqueue.h> 35 #include <linux/mdio.h> 36 #include <linux/io.h> 37 #include <linux/uaccess.h> 38 #include <linux/atomic.h> 39 40 #include <asm/irq.h> 41 42 static const char *phy_speed_to_str(int speed) 43 { 44 switch (speed) { 45 case SPEED_10: 46 return "10Mbps"; 47 case SPEED_100: 48 return "100Mbps"; 49 case SPEED_1000: 50 return "1Gbps"; 51 case SPEED_2500: 52 return "2.5Gbps"; 53 case SPEED_10000: 54 return "10Gbps"; 55 case SPEED_UNKNOWN: 56 return "Unknown"; 57 default: 58 return "Unsupported (update phy.c)"; 59 } 60 } 61 62 #define PHY_STATE_STR(_state) \ 63 case PHY_##_state: \ 64 return __stringify(_state); \ 65 66 static const char *phy_state_to_str(enum phy_state st) 67 { 68 switch (st) { 69 PHY_STATE_STR(DOWN) 70 PHY_STATE_STR(STARTING) 71 PHY_STATE_STR(READY) 72 PHY_STATE_STR(PENDING) 73 PHY_STATE_STR(UP) 74 PHY_STATE_STR(AN) 75 PHY_STATE_STR(RUNNING) 76 PHY_STATE_STR(NOLINK) 77 PHY_STATE_STR(FORCING) 78 PHY_STATE_STR(CHANGELINK) 79 PHY_STATE_STR(HALTED) 80 PHY_STATE_STR(RESUMING) 81 } 82 83 return NULL; 84 } 85 86 87 /** 88 * phy_print_status - Convenience function to print out the current phy status 89 * @phydev: the phy_device struct 90 */ 91 void phy_print_status(struct phy_device *phydev) 92 { 93 if (phydev->link) { 94 netdev_info(phydev->attached_dev, 95 "Link is Up - %s/%s - flow control %s\n", 96 phy_speed_to_str(phydev->speed), 97 DUPLEX_FULL == phydev->duplex ? "Full" : "Half", 98 phydev->pause ? "rx/tx" : "off"); 99 } else { 100 netdev_info(phydev->attached_dev, "Link is Down\n"); 101 } 102 } 103 EXPORT_SYMBOL(phy_print_status); 104 105 /** 106 * phy_clear_interrupt - Ack the phy device's interrupt 107 * @phydev: the phy_device struct 108 * 109 * If the @phydev driver has an ack_interrupt function, call it to 110 * ack and clear the phy device's interrupt. 111 * 112 * Returns 0 on success or < 0 on error. 113 */ 114 static int phy_clear_interrupt(struct phy_device *phydev) 115 { 116 if (phydev->drv->ack_interrupt) 117 return phydev->drv->ack_interrupt(phydev); 118 119 return 0; 120 } 121 122 /** 123 * phy_config_interrupt - configure the PHY device for the requested interrupts 124 * @phydev: the phy_device struct 125 * @interrupts: interrupt flags to configure for this @phydev 126 * 127 * Returns 0 on success or < 0 on error. 128 */ 129 static int phy_config_interrupt(struct phy_device *phydev, u32 interrupts) 130 { 131 phydev->interrupts = interrupts; 132 if (phydev->drv->config_intr) 133 return phydev->drv->config_intr(phydev); 134 135 return 0; 136 } 137 138 139 /** 140 * phy_aneg_done - return auto-negotiation status 141 * @phydev: target phy_device struct 142 * 143 * Description: Return the auto-negotiation status from this @phydev 144 * Returns > 0 on success or < 0 on error. 0 means that auto-negotiation 145 * is still pending. 146 */ 147 int phy_aneg_done(struct phy_device *phydev) 148 { 149 if (phydev->drv && phydev->drv->aneg_done) 150 return phydev->drv->aneg_done(phydev); 151 152 return genphy_aneg_done(phydev); 153 } 154 EXPORT_SYMBOL(phy_aneg_done); 155 156 /* A structure for mapping a particular speed and duplex 157 * combination to a particular SUPPORTED and ADVERTISED value 158 */ 159 struct phy_setting { 160 int speed; 161 int duplex; 162 u32 setting; 163 }; 164 165 /* A mapping of all SUPPORTED settings to speed/duplex */ 166 static const struct phy_setting settings[] = { 167 { 168 .speed = SPEED_10000, 169 .duplex = DUPLEX_FULL, 170 .setting = SUPPORTED_10000baseKR_Full, 171 }, 172 { 173 .speed = SPEED_10000, 174 .duplex = DUPLEX_FULL, 175 .setting = SUPPORTED_10000baseKX4_Full, 176 }, 177 { 178 .speed = SPEED_10000, 179 .duplex = DUPLEX_FULL, 180 .setting = SUPPORTED_10000baseT_Full, 181 }, 182 { 183 .speed = SPEED_2500, 184 .duplex = DUPLEX_FULL, 185 .setting = SUPPORTED_2500baseX_Full, 186 }, 187 { 188 .speed = SPEED_1000, 189 .duplex = DUPLEX_FULL, 190 .setting = SUPPORTED_1000baseKX_Full, 191 }, 192 { 193 .speed = SPEED_1000, 194 .duplex = DUPLEX_FULL, 195 .setting = SUPPORTED_1000baseT_Full, 196 }, 197 { 198 .speed = SPEED_1000, 199 .duplex = DUPLEX_HALF, 200 .setting = SUPPORTED_1000baseT_Half, 201 }, 202 { 203 .speed = SPEED_100, 204 .duplex = DUPLEX_FULL, 205 .setting = SUPPORTED_100baseT_Full, 206 }, 207 { 208 .speed = SPEED_100, 209 .duplex = DUPLEX_HALF, 210 .setting = SUPPORTED_100baseT_Half, 211 }, 212 { 213 .speed = SPEED_10, 214 .duplex = DUPLEX_FULL, 215 .setting = SUPPORTED_10baseT_Full, 216 }, 217 { 218 .speed = SPEED_10, 219 .duplex = DUPLEX_HALF, 220 .setting = SUPPORTED_10baseT_Half, 221 }, 222 }; 223 224 #define MAX_NUM_SETTINGS ARRAY_SIZE(settings) 225 226 /** 227 * phy_find_setting - find a PHY settings array entry that matches speed & duplex 228 * @speed: speed to match 229 * @duplex: duplex to match 230 * 231 * Description: Searches the settings array for the setting which 232 * matches the desired speed and duplex, and returns the index 233 * of that setting. Returns the index of the last setting if 234 * none of the others match. 235 */ 236 static inline unsigned int phy_find_setting(int speed, int duplex) 237 { 238 unsigned int idx = 0; 239 240 while (idx < ARRAY_SIZE(settings) && 241 (settings[idx].speed != speed || settings[idx].duplex != duplex)) 242 idx++; 243 244 return idx < MAX_NUM_SETTINGS ? idx : MAX_NUM_SETTINGS - 1; 245 } 246 247 /** 248 * phy_find_valid - find a PHY setting that matches the requested features mask 249 * @idx: The first index in settings[] to search 250 * @features: A mask of the valid settings 251 * 252 * Description: Returns the index of the first valid setting less 253 * than or equal to the one pointed to by idx, as determined by 254 * the mask in features. Returns the index of the last setting 255 * if nothing else matches. 256 */ 257 static inline unsigned int phy_find_valid(unsigned int idx, u32 features) 258 { 259 while (idx < MAX_NUM_SETTINGS && !(settings[idx].setting & features)) 260 idx++; 261 262 return idx < MAX_NUM_SETTINGS ? idx : MAX_NUM_SETTINGS - 1; 263 } 264 265 /** 266 * phy_supported_speeds - return all speeds currently supported by a phy device 267 * @phy: The phy device to return supported speeds of. 268 * @speeds: buffer to store supported speeds in. 269 * @size: size of speeds buffer. 270 * 271 * Description: Returns the number of supported speeds, and fills the speeds 272 * buffer with the supported speeds. If speeds buffer is too small to contain 273 * all currently supported speeds, will return as many speeds as can fit. 274 */ 275 unsigned int phy_supported_speeds(struct phy_device *phy, 276 unsigned int *speeds, 277 unsigned int size) 278 { 279 unsigned int count = 0; 280 unsigned int idx = 0; 281 282 while (idx < MAX_NUM_SETTINGS && count < size) { 283 idx = phy_find_valid(idx, phy->supported); 284 285 if (!(settings[idx].setting & phy->supported)) 286 break; 287 288 /* Assumes settings are grouped by speed */ 289 if ((count == 0) || 290 (speeds[count - 1] != settings[idx].speed)) { 291 speeds[count] = settings[idx].speed; 292 count++; 293 } 294 idx++; 295 } 296 297 return count; 298 } 299 300 /** 301 * phy_check_valid - check if there is a valid PHY setting which matches 302 * speed, duplex, and feature mask 303 * @speed: speed to match 304 * @duplex: duplex to match 305 * @features: A mask of the valid settings 306 * 307 * Description: Returns true if there is a valid setting, false otherwise. 308 */ 309 static inline bool phy_check_valid(int speed, int duplex, u32 features) 310 { 311 unsigned int idx; 312 313 idx = phy_find_valid(phy_find_setting(speed, duplex), features); 314 315 return settings[idx].speed == speed && settings[idx].duplex == duplex && 316 (settings[idx].setting & features); 317 } 318 319 /** 320 * phy_sanitize_settings - make sure the PHY is set to supported speed and duplex 321 * @phydev: the target phy_device struct 322 * 323 * Description: Make sure the PHY is set to supported speeds and 324 * duplexes. Drop down by one in this order: 1000/FULL, 325 * 1000/HALF, 100/FULL, 100/HALF, 10/FULL, 10/HALF. 326 */ 327 static void phy_sanitize_settings(struct phy_device *phydev) 328 { 329 u32 features = phydev->supported; 330 unsigned int idx; 331 332 /* Sanitize settings based on PHY capabilities */ 333 if ((features & SUPPORTED_Autoneg) == 0) 334 phydev->autoneg = AUTONEG_DISABLE; 335 336 idx = phy_find_valid(phy_find_setting(phydev->speed, phydev->duplex), 337 features); 338 339 phydev->speed = settings[idx].speed; 340 phydev->duplex = settings[idx].duplex; 341 } 342 343 /** 344 * phy_ethtool_sset - generic ethtool sset function, handles all the details 345 * @phydev: target phy_device struct 346 * @cmd: ethtool_cmd 347 * 348 * A few notes about parameter checking: 349 * - We don't set port or transceiver, so we don't care what they 350 * were set to. 351 * - phy_start_aneg() will make sure forced settings are sane, and 352 * choose the next best ones from the ones selected, so we don't 353 * care if ethtool tries to give us bad values. 354 */ 355 int phy_ethtool_sset(struct phy_device *phydev, struct ethtool_cmd *cmd) 356 { 357 u32 speed = ethtool_cmd_speed(cmd); 358 359 if (cmd->phy_address != phydev->mdio.addr) 360 return -EINVAL; 361 362 /* We make sure that we don't pass unsupported values in to the PHY */ 363 cmd->advertising &= phydev->supported; 364 365 /* Verify the settings we care about. */ 366 if (cmd->autoneg != AUTONEG_ENABLE && cmd->autoneg != AUTONEG_DISABLE) 367 return -EINVAL; 368 369 if (cmd->autoneg == AUTONEG_ENABLE && cmd->advertising == 0) 370 return -EINVAL; 371 372 if (cmd->autoneg == AUTONEG_DISABLE && 373 ((speed != SPEED_1000 && 374 speed != SPEED_100 && 375 speed != SPEED_10) || 376 (cmd->duplex != DUPLEX_HALF && 377 cmd->duplex != DUPLEX_FULL))) 378 return -EINVAL; 379 380 phydev->autoneg = cmd->autoneg; 381 382 phydev->speed = speed; 383 384 phydev->advertising = cmd->advertising; 385 386 if (AUTONEG_ENABLE == cmd->autoneg) 387 phydev->advertising |= ADVERTISED_Autoneg; 388 else 389 phydev->advertising &= ~ADVERTISED_Autoneg; 390 391 phydev->duplex = cmd->duplex; 392 393 phydev->mdix_ctrl = cmd->eth_tp_mdix_ctrl; 394 395 /* Restart the PHY */ 396 phy_start_aneg(phydev); 397 398 return 0; 399 } 400 EXPORT_SYMBOL(phy_ethtool_sset); 401 402 int phy_ethtool_ksettings_set(struct phy_device *phydev, 403 const struct ethtool_link_ksettings *cmd) 404 { 405 u8 autoneg = cmd->base.autoneg; 406 u8 duplex = cmd->base.duplex; 407 u32 speed = cmd->base.speed; 408 u32 advertising; 409 410 if (cmd->base.phy_address != phydev->mdio.addr) 411 return -EINVAL; 412 413 ethtool_convert_link_mode_to_legacy_u32(&advertising, 414 cmd->link_modes.advertising); 415 416 /* We make sure that we don't pass unsupported values in to the PHY */ 417 advertising &= phydev->supported; 418 419 /* Verify the settings we care about. */ 420 if (autoneg != AUTONEG_ENABLE && autoneg != AUTONEG_DISABLE) 421 return -EINVAL; 422 423 if (autoneg == AUTONEG_ENABLE && advertising == 0) 424 return -EINVAL; 425 426 if (autoneg == AUTONEG_DISABLE && 427 ((speed != SPEED_1000 && 428 speed != SPEED_100 && 429 speed != SPEED_10) || 430 (duplex != DUPLEX_HALF && 431 duplex != DUPLEX_FULL))) 432 return -EINVAL; 433 434 phydev->autoneg = autoneg; 435 436 phydev->speed = speed; 437 438 phydev->advertising = advertising; 439 440 if (autoneg == AUTONEG_ENABLE) 441 phydev->advertising |= ADVERTISED_Autoneg; 442 else 443 phydev->advertising &= ~ADVERTISED_Autoneg; 444 445 phydev->duplex = duplex; 446 447 phydev->mdix_ctrl = cmd->base.eth_tp_mdix_ctrl; 448 449 /* Restart the PHY */ 450 phy_start_aneg(phydev); 451 452 return 0; 453 } 454 EXPORT_SYMBOL(phy_ethtool_ksettings_set); 455 456 int phy_ethtool_gset(struct phy_device *phydev, struct ethtool_cmd *cmd) 457 { 458 cmd->supported = phydev->supported; 459 460 cmd->advertising = phydev->advertising; 461 cmd->lp_advertising = phydev->lp_advertising; 462 463 ethtool_cmd_speed_set(cmd, phydev->speed); 464 cmd->duplex = phydev->duplex; 465 if (phydev->interface == PHY_INTERFACE_MODE_MOCA) 466 cmd->port = PORT_BNC; 467 else 468 cmd->port = PORT_MII; 469 cmd->phy_address = phydev->mdio.addr; 470 cmd->transceiver = phy_is_internal(phydev) ? 471 XCVR_INTERNAL : XCVR_EXTERNAL; 472 cmd->autoneg = phydev->autoneg; 473 cmd->eth_tp_mdix_ctrl = phydev->mdix_ctrl; 474 cmd->eth_tp_mdix = phydev->mdix; 475 476 return 0; 477 } 478 EXPORT_SYMBOL(phy_ethtool_gset); 479 480 int phy_ethtool_ksettings_get(struct phy_device *phydev, 481 struct ethtool_link_ksettings *cmd) 482 { 483 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.supported, 484 phydev->supported); 485 486 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.advertising, 487 phydev->advertising); 488 489 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.lp_advertising, 490 phydev->lp_advertising); 491 492 cmd->base.speed = phydev->speed; 493 cmd->base.duplex = phydev->duplex; 494 if (phydev->interface == PHY_INTERFACE_MODE_MOCA) 495 cmd->base.port = PORT_BNC; 496 else 497 cmd->base.port = PORT_MII; 498 499 cmd->base.phy_address = phydev->mdio.addr; 500 cmd->base.autoneg = phydev->autoneg; 501 cmd->base.eth_tp_mdix_ctrl = phydev->mdix_ctrl; 502 cmd->base.eth_tp_mdix = phydev->mdix; 503 504 return 0; 505 } 506 EXPORT_SYMBOL(phy_ethtool_ksettings_get); 507 508 /** 509 * phy_mii_ioctl - generic PHY MII ioctl interface 510 * @phydev: the phy_device struct 511 * @ifr: &struct ifreq for socket ioctl's 512 * @cmd: ioctl cmd to execute 513 * 514 * Note that this function is currently incompatible with the 515 * PHYCONTROL layer. It changes registers without regard to 516 * current state. Use at own risk. 517 */ 518 int phy_mii_ioctl(struct phy_device *phydev, struct ifreq *ifr, int cmd) 519 { 520 struct mii_ioctl_data *mii_data = if_mii(ifr); 521 u16 val = mii_data->val_in; 522 bool change_autoneg = false; 523 524 switch (cmd) { 525 case SIOCGMIIPHY: 526 mii_data->phy_id = phydev->mdio.addr; 527 /* fall through */ 528 529 case SIOCGMIIREG: 530 mii_data->val_out = mdiobus_read(phydev->mdio.bus, 531 mii_data->phy_id, 532 mii_data->reg_num); 533 return 0; 534 535 case SIOCSMIIREG: 536 if (mii_data->phy_id == phydev->mdio.addr) { 537 switch (mii_data->reg_num) { 538 case MII_BMCR: 539 if ((val & (BMCR_RESET | BMCR_ANENABLE)) == 0) { 540 if (phydev->autoneg == AUTONEG_ENABLE) 541 change_autoneg = true; 542 phydev->autoneg = AUTONEG_DISABLE; 543 if (val & BMCR_FULLDPLX) 544 phydev->duplex = DUPLEX_FULL; 545 else 546 phydev->duplex = DUPLEX_HALF; 547 if (val & BMCR_SPEED1000) 548 phydev->speed = SPEED_1000; 549 else if (val & BMCR_SPEED100) 550 phydev->speed = SPEED_100; 551 else phydev->speed = SPEED_10; 552 } 553 else { 554 if (phydev->autoneg == AUTONEG_DISABLE) 555 change_autoneg = true; 556 phydev->autoneg = AUTONEG_ENABLE; 557 } 558 break; 559 case MII_ADVERTISE: 560 phydev->advertising = mii_adv_to_ethtool_adv_t(val); 561 change_autoneg = true; 562 break; 563 default: 564 /* do nothing */ 565 break; 566 } 567 } 568 569 mdiobus_write(phydev->mdio.bus, mii_data->phy_id, 570 mii_data->reg_num, val); 571 572 if (mii_data->phy_id == phydev->mdio.addr && 573 mii_data->reg_num == MII_BMCR && 574 val & BMCR_RESET) 575 return phy_init_hw(phydev); 576 577 if (change_autoneg) 578 return phy_start_aneg(phydev); 579 580 return 0; 581 582 case SIOCSHWTSTAMP: 583 if (phydev->drv && phydev->drv->hwtstamp) 584 return phydev->drv->hwtstamp(phydev, ifr); 585 /* fall through */ 586 587 default: 588 return -EOPNOTSUPP; 589 } 590 } 591 EXPORT_SYMBOL(phy_mii_ioctl); 592 593 /** 594 * phy_start_aneg_priv - start auto-negotiation for this PHY device 595 * @phydev: the phy_device struct 596 * @sync: indicate whether we should wait for the workqueue cancelation 597 * 598 * Description: Sanitizes the settings (if we're not autonegotiating 599 * them), and then calls the driver's config_aneg function. 600 * If the PHYCONTROL Layer is operating, we change the state to 601 * reflect the beginning of Auto-negotiation or forcing. 602 */ 603 static int phy_start_aneg_priv(struct phy_device *phydev, bool sync) 604 { 605 bool trigger = 0; 606 int err; 607 608 if (!phydev->drv) 609 return -EIO; 610 611 mutex_lock(&phydev->lock); 612 613 if (AUTONEG_DISABLE == phydev->autoneg) 614 phy_sanitize_settings(phydev); 615 616 /* Invalidate LP advertising flags */ 617 phydev->lp_advertising = 0; 618 619 err = phydev->drv->config_aneg(phydev); 620 if (err < 0) 621 goto out_unlock; 622 623 if (phydev->state != PHY_HALTED) { 624 if (AUTONEG_ENABLE == phydev->autoneg) { 625 phydev->state = PHY_AN; 626 phydev->link_timeout = PHY_AN_TIMEOUT; 627 } else { 628 phydev->state = PHY_FORCING; 629 phydev->link_timeout = PHY_FORCE_TIMEOUT; 630 } 631 } 632 633 /* Re-schedule a PHY state machine to check PHY status because 634 * negotiation may already be done and aneg interrupt may not be 635 * generated. 636 */ 637 if (phy_interrupt_is_valid(phydev) && (phydev->state == PHY_AN)) { 638 err = phy_aneg_done(phydev); 639 if (err > 0) { 640 trigger = true; 641 err = 0; 642 } 643 } 644 645 out_unlock: 646 mutex_unlock(&phydev->lock); 647 648 if (trigger) 649 phy_trigger_machine(phydev, sync); 650 651 return err; 652 } 653 654 /** 655 * phy_start_aneg - start auto-negotiation for this PHY device 656 * @phydev: the phy_device struct 657 * 658 * Description: Sanitizes the settings (if we're not autonegotiating 659 * them), and then calls the driver's config_aneg function. 660 * If the PHYCONTROL Layer is operating, we change the state to 661 * reflect the beginning of Auto-negotiation or forcing. 662 */ 663 int phy_start_aneg(struct phy_device *phydev) 664 { 665 return phy_start_aneg_priv(phydev, true); 666 } 667 EXPORT_SYMBOL(phy_start_aneg); 668 669 /** 670 * phy_start_machine - start PHY state machine tracking 671 * @phydev: the phy_device struct 672 * 673 * Description: The PHY infrastructure can run a state machine 674 * which tracks whether the PHY is starting up, negotiating, 675 * etc. This function starts the timer which tracks the state 676 * of the PHY. If you want to maintain your own state machine, 677 * do not call this function. 678 */ 679 void phy_start_machine(struct phy_device *phydev) 680 { 681 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, HZ); 682 } 683 684 /** 685 * phy_trigger_machine - trigger the state machine to run 686 * 687 * @phydev: the phy_device struct 688 * @sync: indicate whether we should wait for the workqueue cancelation 689 * 690 * Description: There has been a change in state which requires that the 691 * state machine runs. 692 */ 693 694 void phy_trigger_machine(struct phy_device *phydev, bool sync) 695 { 696 if (sync) 697 cancel_delayed_work_sync(&phydev->state_queue); 698 else 699 cancel_delayed_work(&phydev->state_queue); 700 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, 0); 701 } 702 703 /** 704 * phy_stop_machine - stop the PHY state machine tracking 705 * @phydev: target phy_device struct 706 * 707 * Description: Stops the state machine timer, sets the state to UP 708 * (unless it wasn't up yet). This function must be called BEFORE 709 * phy_detach. 710 */ 711 void phy_stop_machine(struct phy_device *phydev) 712 { 713 cancel_delayed_work_sync(&phydev->state_queue); 714 715 mutex_lock(&phydev->lock); 716 if (phydev->state > PHY_UP && phydev->state != PHY_HALTED) 717 phydev->state = PHY_UP; 718 mutex_unlock(&phydev->lock); 719 } 720 721 /** 722 * phy_error - enter HALTED state for this PHY device 723 * @phydev: target phy_device struct 724 * 725 * Moves the PHY to the HALTED state in response to a read 726 * or write error, and tells the controller the link is down. 727 * Must not be called from interrupt context, or while the 728 * phydev->lock is held. 729 */ 730 static void phy_error(struct phy_device *phydev) 731 { 732 mutex_lock(&phydev->lock); 733 phydev->state = PHY_HALTED; 734 mutex_unlock(&phydev->lock); 735 736 phy_trigger_machine(phydev, false); 737 } 738 739 /** 740 * phy_interrupt - PHY interrupt handler 741 * @irq: interrupt line 742 * @phy_dat: phy_device pointer 743 * 744 * Description: When a PHY interrupt occurs, the handler disables 745 * interrupts, and uses phy_change to handle the interrupt. 746 */ 747 static irqreturn_t phy_interrupt(int irq, void *phy_dat) 748 { 749 struct phy_device *phydev = phy_dat; 750 751 if (PHY_HALTED == phydev->state) 752 return IRQ_NONE; /* It can't be ours. */ 753 754 disable_irq_nosync(irq); 755 atomic_inc(&phydev->irq_disable); 756 757 phy_change(phydev); 758 759 return IRQ_HANDLED; 760 } 761 762 /** 763 * phy_enable_interrupts - Enable the interrupts from the PHY side 764 * @phydev: target phy_device struct 765 */ 766 static int phy_enable_interrupts(struct phy_device *phydev) 767 { 768 int err = phy_clear_interrupt(phydev); 769 770 if (err < 0) 771 return err; 772 773 return phy_config_interrupt(phydev, PHY_INTERRUPT_ENABLED); 774 } 775 776 /** 777 * phy_disable_interrupts - Disable the PHY interrupts from the PHY side 778 * @phydev: target phy_device struct 779 */ 780 static int phy_disable_interrupts(struct phy_device *phydev) 781 { 782 int err; 783 784 /* Disable PHY interrupts */ 785 err = phy_config_interrupt(phydev, PHY_INTERRUPT_DISABLED); 786 if (err) 787 goto phy_err; 788 789 /* Clear the interrupt */ 790 err = phy_clear_interrupt(phydev); 791 if (err) 792 goto phy_err; 793 794 return 0; 795 796 phy_err: 797 phy_error(phydev); 798 799 return err; 800 } 801 802 /** 803 * phy_start_interrupts - request and enable interrupts for a PHY device 804 * @phydev: target phy_device struct 805 * 806 * Description: Request the interrupt for the given PHY. 807 * If this fails, then we set irq to PHY_POLL. 808 * Otherwise, we enable the interrupts in the PHY. 809 * This should only be called with a valid IRQ number. 810 * Returns 0 on success or < 0 on error. 811 */ 812 int phy_start_interrupts(struct phy_device *phydev) 813 { 814 atomic_set(&phydev->irq_disable, 0); 815 if (request_threaded_irq(phydev->irq, NULL, phy_interrupt, 816 IRQF_ONESHOT | IRQF_SHARED, 817 phydev_name(phydev), phydev) < 0) { 818 pr_warn("%s: Can't get IRQ %d (PHY)\n", 819 phydev->mdio.bus->name, phydev->irq); 820 phydev->irq = PHY_POLL; 821 return 0; 822 } 823 824 return phy_enable_interrupts(phydev); 825 } 826 EXPORT_SYMBOL(phy_start_interrupts); 827 828 /** 829 * phy_stop_interrupts - disable interrupts from a PHY device 830 * @phydev: target phy_device struct 831 */ 832 int phy_stop_interrupts(struct phy_device *phydev) 833 { 834 int err = phy_disable_interrupts(phydev); 835 836 if (err) 837 phy_error(phydev); 838 839 free_irq(phydev->irq, phydev); 840 841 /* If work indeed has been cancelled, disable_irq() will have 842 * been left unbalanced from phy_interrupt() and enable_irq() 843 * has to be called so that other devices on the line work. 844 */ 845 while (atomic_dec_return(&phydev->irq_disable) >= 0) 846 enable_irq(phydev->irq); 847 848 return err; 849 } 850 EXPORT_SYMBOL(phy_stop_interrupts); 851 852 /** 853 * phy_change - Called by the phy_interrupt to handle PHY changes 854 * @phydev: phy_device struct that interrupted 855 */ 856 void phy_change(struct phy_device *phydev) 857 { 858 if (phy_interrupt_is_valid(phydev)) { 859 if (phydev->drv->did_interrupt && 860 !phydev->drv->did_interrupt(phydev)) 861 goto ignore; 862 863 if (phy_disable_interrupts(phydev)) 864 goto phy_err; 865 } 866 867 mutex_lock(&phydev->lock); 868 if ((PHY_RUNNING == phydev->state) || (PHY_NOLINK == phydev->state)) 869 phydev->state = PHY_CHANGELINK; 870 mutex_unlock(&phydev->lock); 871 872 if (phy_interrupt_is_valid(phydev)) { 873 atomic_dec(&phydev->irq_disable); 874 enable_irq(phydev->irq); 875 876 /* Reenable interrupts */ 877 if (PHY_HALTED != phydev->state && 878 phy_config_interrupt(phydev, PHY_INTERRUPT_ENABLED)) 879 goto irq_enable_err; 880 } 881 882 /* reschedule state queue work to run as soon as possible */ 883 phy_trigger_machine(phydev, true); 884 return; 885 886 ignore: 887 atomic_dec(&phydev->irq_disable); 888 enable_irq(phydev->irq); 889 return; 890 891 irq_enable_err: 892 disable_irq(phydev->irq); 893 atomic_inc(&phydev->irq_disable); 894 phy_err: 895 phy_error(phydev); 896 } 897 898 /** 899 * phy_change_work - Scheduled by the phy_mac_interrupt to handle PHY changes 900 * @work: work_struct that describes the work to be done 901 */ 902 void phy_change_work(struct work_struct *work) 903 { 904 struct phy_device *phydev = 905 container_of(work, struct phy_device, phy_queue); 906 907 phy_change(phydev); 908 } 909 910 /** 911 * phy_stop - Bring down the PHY link, and stop checking the status 912 * @phydev: target phy_device struct 913 */ 914 void phy_stop(struct phy_device *phydev) 915 { 916 mutex_lock(&phydev->lock); 917 918 if (PHY_HALTED == phydev->state) 919 goto out_unlock; 920 921 if (phy_interrupt_is_valid(phydev)) { 922 /* Disable PHY Interrupts */ 923 phy_config_interrupt(phydev, PHY_INTERRUPT_DISABLED); 924 925 /* Clear any pending interrupts */ 926 phy_clear_interrupt(phydev); 927 } 928 929 phydev->state = PHY_HALTED; 930 931 out_unlock: 932 mutex_unlock(&phydev->lock); 933 934 /* Cannot call flush_scheduled_work() here as desired because 935 * of rtnl_lock(), but PHY_HALTED shall guarantee phy_change() 936 * will not reenable interrupts. 937 */ 938 } 939 EXPORT_SYMBOL(phy_stop); 940 941 /** 942 * phy_start - start or restart a PHY device 943 * @phydev: target phy_device struct 944 * 945 * Description: Indicates the attached device's readiness to 946 * handle PHY-related work. Used during startup to start the 947 * PHY, and after a call to phy_stop() to resume operation. 948 * Also used to indicate the MDIO bus has cleared an error 949 * condition. 950 */ 951 void phy_start(struct phy_device *phydev) 952 { 953 bool do_resume = false; 954 int err = 0; 955 956 mutex_lock(&phydev->lock); 957 958 switch (phydev->state) { 959 case PHY_STARTING: 960 phydev->state = PHY_PENDING; 961 break; 962 case PHY_READY: 963 phydev->state = PHY_UP; 964 break; 965 case PHY_HALTED: 966 /* make sure interrupts are re-enabled for the PHY */ 967 if (phydev->irq != PHY_POLL) { 968 err = phy_enable_interrupts(phydev); 969 if (err < 0) 970 break; 971 } 972 973 phydev->state = PHY_RESUMING; 974 do_resume = true; 975 break; 976 default: 977 break; 978 } 979 mutex_unlock(&phydev->lock); 980 981 /* if phy was suspended, bring the physical link up again */ 982 if (do_resume) 983 phy_resume(phydev); 984 985 phy_trigger_machine(phydev, true); 986 } 987 EXPORT_SYMBOL(phy_start); 988 989 static void phy_adjust_link(struct phy_device *phydev) 990 { 991 phydev->adjust_link(phydev->attached_dev); 992 phy_led_trigger_change_speed(phydev); 993 } 994 995 /** 996 * phy_state_machine - Handle the state machine 997 * @work: work_struct that describes the work to be done 998 */ 999 void phy_state_machine(struct work_struct *work) 1000 { 1001 struct delayed_work *dwork = to_delayed_work(work); 1002 struct phy_device *phydev = 1003 container_of(dwork, struct phy_device, state_queue); 1004 bool needs_aneg = false, do_suspend = false; 1005 enum phy_state old_state; 1006 int err = 0; 1007 int old_link; 1008 1009 mutex_lock(&phydev->lock); 1010 1011 old_state = phydev->state; 1012 1013 if (phydev->drv && phydev->drv->link_change_notify) 1014 phydev->drv->link_change_notify(phydev); 1015 1016 switch (phydev->state) { 1017 case PHY_DOWN: 1018 case PHY_STARTING: 1019 case PHY_READY: 1020 case PHY_PENDING: 1021 break; 1022 case PHY_UP: 1023 needs_aneg = true; 1024 1025 phydev->link_timeout = PHY_AN_TIMEOUT; 1026 1027 break; 1028 case PHY_AN: 1029 err = phy_read_status(phydev); 1030 if (err < 0) 1031 break; 1032 1033 /* If the link is down, give up on negotiation for now */ 1034 if (!phydev->link) { 1035 phydev->state = PHY_NOLINK; 1036 netif_carrier_off(phydev->attached_dev); 1037 phy_adjust_link(phydev); 1038 break; 1039 } 1040 1041 /* Check if negotiation is done. Break if there's an error */ 1042 err = phy_aneg_done(phydev); 1043 if (err < 0) 1044 break; 1045 1046 /* If AN is done, we're running */ 1047 if (err > 0) { 1048 phydev->state = PHY_RUNNING; 1049 netif_carrier_on(phydev->attached_dev); 1050 phy_adjust_link(phydev); 1051 1052 } else if (0 == phydev->link_timeout--) 1053 needs_aneg = true; 1054 break; 1055 case PHY_NOLINK: 1056 if (phy_interrupt_is_valid(phydev)) 1057 break; 1058 1059 err = phy_read_status(phydev); 1060 if (err) 1061 break; 1062 1063 if (phydev->link) { 1064 if (AUTONEG_ENABLE == phydev->autoneg) { 1065 err = phy_aneg_done(phydev); 1066 if (err < 0) 1067 break; 1068 1069 if (!err) { 1070 phydev->state = PHY_AN; 1071 phydev->link_timeout = PHY_AN_TIMEOUT; 1072 break; 1073 } 1074 } 1075 phydev->state = PHY_RUNNING; 1076 netif_carrier_on(phydev->attached_dev); 1077 phy_adjust_link(phydev); 1078 } 1079 break; 1080 case PHY_FORCING: 1081 err = genphy_update_link(phydev); 1082 if (err) 1083 break; 1084 1085 if (phydev->link) { 1086 phydev->state = PHY_RUNNING; 1087 netif_carrier_on(phydev->attached_dev); 1088 } else { 1089 if (0 == phydev->link_timeout--) 1090 needs_aneg = true; 1091 } 1092 1093 phy_adjust_link(phydev); 1094 break; 1095 case PHY_RUNNING: 1096 /* Only register a CHANGE if we are polling and link changed 1097 * since latest checking. 1098 */ 1099 if (phydev->irq == PHY_POLL) { 1100 old_link = phydev->link; 1101 err = phy_read_status(phydev); 1102 if (err) 1103 break; 1104 1105 if (old_link != phydev->link) 1106 phydev->state = PHY_CHANGELINK; 1107 } 1108 /* 1109 * Failsafe: check that nobody set phydev->link=0 between two 1110 * poll cycles, otherwise we won't leave RUNNING state as long 1111 * as link remains down. 1112 */ 1113 if (!phydev->link && phydev->state == PHY_RUNNING) { 1114 phydev->state = PHY_CHANGELINK; 1115 phydev_err(phydev, "no link in PHY_RUNNING\n"); 1116 } 1117 break; 1118 case PHY_CHANGELINK: 1119 err = phy_read_status(phydev); 1120 if (err) 1121 break; 1122 1123 if (phydev->link) { 1124 phydev->state = PHY_RUNNING; 1125 netif_carrier_on(phydev->attached_dev); 1126 } else { 1127 phydev->state = PHY_NOLINK; 1128 netif_carrier_off(phydev->attached_dev); 1129 } 1130 1131 phy_adjust_link(phydev); 1132 1133 if (phy_interrupt_is_valid(phydev)) 1134 err = phy_config_interrupt(phydev, 1135 PHY_INTERRUPT_ENABLED); 1136 break; 1137 case PHY_HALTED: 1138 if (phydev->link) { 1139 phydev->link = 0; 1140 netif_carrier_off(phydev->attached_dev); 1141 phy_adjust_link(phydev); 1142 do_suspend = true; 1143 } 1144 break; 1145 case PHY_RESUMING: 1146 if (AUTONEG_ENABLE == phydev->autoneg) { 1147 err = phy_aneg_done(phydev); 1148 if (err < 0) 1149 break; 1150 1151 /* err > 0 if AN is done. 1152 * Otherwise, it's 0, and we're still waiting for AN 1153 */ 1154 if (err > 0) { 1155 err = phy_read_status(phydev); 1156 if (err) 1157 break; 1158 1159 if (phydev->link) { 1160 phydev->state = PHY_RUNNING; 1161 netif_carrier_on(phydev->attached_dev); 1162 } else { 1163 phydev->state = PHY_NOLINK; 1164 } 1165 phy_adjust_link(phydev); 1166 } else { 1167 phydev->state = PHY_AN; 1168 phydev->link_timeout = PHY_AN_TIMEOUT; 1169 } 1170 } else { 1171 err = phy_read_status(phydev); 1172 if (err) 1173 break; 1174 1175 if (phydev->link) { 1176 phydev->state = PHY_RUNNING; 1177 netif_carrier_on(phydev->attached_dev); 1178 } else { 1179 phydev->state = PHY_NOLINK; 1180 } 1181 phy_adjust_link(phydev); 1182 } 1183 break; 1184 } 1185 1186 mutex_unlock(&phydev->lock); 1187 1188 if (needs_aneg) 1189 err = phy_start_aneg_priv(phydev, false); 1190 else if (do_suspend) 1191 phy_suspend(phydev); 1192 1193 if (err < 0) 1194 phy_error(phydev); 1195 1196 phydev_dbg(phydev, "PHY state change %s -> %s\n", 1197 phy_state_to_str(old_state), 1198 phy_state_to_str(phydev->state)); 1199 1200 /* Only re-schedule a PHY state machine change if we are polling the 1201 * PHY, if PHY_IGNORE_INTERRUPT is set, then we will be moving 1202 * between states from phy_mac_interrupt() 1203 */ 1204 if (phydev->irq == PHY_POLL) 1205 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, 1206 PHY_STATE_TIME * HZ); 1207 } 1208 1209 /** 1210 * phy_mac_interrupt - MAC says the link has changed 1211 * @phydev: phy_device struct with changed link 1212 * @new_link: Link is Up/Down. 1213 * 1214 * Description: The MAC layer is able indicate there has been a change 1215 * in the PHY link status. Set the new link status, and trigger the 1216 * state machine, work a work queue. 1217 */ 1218 void phy_mac_interrupt(struct phy_device *phydev, int new_link) 1219 { 1220 phydev->link = new_link; 1221 1222 /* Trigger a state machine change */ 1223 queue_work(system_power_efficient_wq, &phydev->phy_queue); 1224 } 1225 EXPORT_SYMBOL(phy_mac_interrupt); 1226 1227 static inline void mmd_phy_indirect(struct mii_bus *bus, int prtad, int devad, 1228 int addr) 1229 { 1230 /* Write the desired MMD Devad */ 1231 bus->write(bus, addr, MII_MMD_CTRL, devad); 1232 1233 /* Write the desired MMD register address */ 1234 bus->write(bus, addr, MII_MMD_DATA, prtad); 1235 1236 /* Select the Function : DATA with no post increment */ 1237 bus->write(bus, addr, MII_MMD_CTRL, (devad | MII_MMD_CTRL_NOINCR)); 1238 } 1239 1240 /** 1241 * phy_read_mmd_indirect - reads data from the MMD registers 1242 * @phydev: The PHY device bus 1243 * @prtad: MMD Address 1244 * @devad: MMD DEVAD 1245 * 1246 * Description: it reads data from the MMD registers (clause 22 to access to 1247 * clause 45) of the specified phy address. 1248 * To read these register we have: 1249 * 1) Write reg 13 // DEVAD 1250 * 2) Write reg 14 // MMD Address 1251 * 3) Write reg 13 // MMD Data Command for MMD DEVAD 1252 * 3) Read reg 14 // Read MMD data 1253 */ 1254 int phy_read_mmd_indirect(struct phy_device *phydev, int prtad, int devad) 1255 { 1256 struct phy_driver *phydrv = phydev->drv; 1257 int addr = phydev->mdio.addr; 1258 int value = -1; 1259 1260 if (!phydrv->read_mmd_indirect) { 1261 struct mii_bus *bus = phydev->mdio.bus; 1262 1263 mutex_lock(&bus->mdio_lock); 1264 mmd_phy_indirect(bus, prtad, devad, addr); 1265 1266 /* Read the content of the MMD's selected register */ 1267 value = bus->read(bus, addr, MII_MMD_DATA); 1268 mutex_unlock(&bus->mdio_lock); 1269 } else { 1270 value = phydrv->read_mmd_indirect(phydev, prtad, devad, addr); 1271 } 1272 return value; 1273 } 1274 EXPORT_SYMBOL(phy_read_mmd_indirect); 1275 1276 /** 1277 * phy_write_mmd_indirect - writes data to the MMD registers 1278 * @phydev: The PHY device 1279 * @prtad: MMD Address 1280 * @devad: MMD DEVAD 1281 * @data: data to write in the MMD register 1282 * 1283 * Description: Write data from the MMD registers of the specified 1284 * phy address. 1285 * To write these register we have: 1286 * 1) Write reg 13 // DEVAD 1287 * 2) Write reg 14 // MMD Address 1288 * 3) Write reg 13 // MMD Data Command for MMD DEVAD 1289 * 3) Write reg 14 // Write MMD data 1290 */ 1291 void phy_write_mmd_indirect(struct phy_device *phydev, int prtad, 1292 int devad, u32 data) 1293 { 1294 struct phy_driver *phydrv = phydev->drv; 1295 int addr = phydev->mdio.addr; 1296 1297 if (!phydrv->write_mmd_indirect) { 1298 struct mii_bus *bus = phydev->mdio.bus; 1299 1300 mutex_lock(&bus->mdio_lock); 1301 mmd_phy_indirect(bus, prtad, devad, addr); 1302 1303 /* Write the data into MMD's selected register */ 1304 bus->write(bus, addr, MII_MMD_DATA, data); 1305 mutex_unlock(&bus->mdio_lock); 1306 } else { 1307 phydrv->write_mmd_indirect(phydev, prtad, devad, addr, data); 1308 } 1309 } 1310 EXPORT_SYMBOL(phy_write_mmd_indirect); 1311 1312 /** 1313 * phy_init_eee - init and check the EEE feature 1314 * @phydev: target phy_device struct 1315 * @clk_stop_enable: PHY may stop the clock during LPI 1316 * 1317 * Description: it checks if the Energy-Efficient Ethernet (EEE) 1318 * is supported by looking at the MMD registers 3.20 and 7.60/61 1319 * and it programs the MMD register 3.0 setting the "Clock stop enable" 1320 * bit if required. 1321 */ 1322 int phy_init_eee(struct phy_device *phydev, bool clk_stop_enable) 1323 { 1324 if (!phydev->drv) 1325 return -EIO; 1326 1327 /* According to 802.3az,the EEE is supported only in full duplex-mode. 1328 * Also EEE feature is active when core is operating with MII, GMII 1329 * or RGMII (all kinds). Internal PHYs are also allowed to proceed and 1330 * should return an error if they do not support EEE. 1331 */ 1332 if ((phydev->duplex == DUPLEX_FULL) && 1333 ((phydev->interface == PHY_INTERFACE_MODE_MII) || 1334 (phydev->interface == PHY_INTERFACE_MODE_GMII) || 1335 phy_interface_is_rgmii(phydev) || 1336 phy_is_internal(phydev))) { 1337 int eee_lp, eee_cap, eee_adv; 1338 u32 lp, cap, adv; 1339 int status; 1340 1341 /* Read phy status to properly get the right settings */ 1342 status = phy_read_status(phydev); 1343 if (status) 1344 return status; 1345 1346 /* First check if the EEE ability is supported */ 1347 eee_cap = phy_read_mmd_indirect(phydev, MDIO_PCS_EEE_ABLE, 1348 MDIO_MMD_PCS); 1349 if (eee_cap <= 0) 1350 goto eee_exit_err; 1351 1352 cap = mmd_eee_cap_to_ethtool_sup_t(eee_cap); 1353 if (!cap) 1354 goto eee_exit_err; 1355 1356 /* Check which link settings negotiated and verify it in 1357 * the EEE advertising registers. 1358 */ 1359 eee_lp = phy_read_mmd_indirect(phydev, MDIO_AN_EEE_LPABLE, 1360 MDIO_MMD_AN); 1361 if (eee_lp <= 0) 1362 goto eee_exit_err; 1363 1364 eee_adv = phy_read_mmd_indirect(phydev, MDIO_AN_EEE_ADV, 1365 MDIO_MMD_AN); 1366 if (eee_adv <= 0) 1367 goto eee_exit_err; 1368 1369 adv = mmd_eee_adv_to_ethtool_adv_t(eee_adv); 1370 lp = mmd_eee_adv_to_ethtool_adv_t(eee_lp); 1371 if (!phy_check_valid(phydev->speed, phydev->duplex, lp & adv)) 1372 goto eee_exit_err; 1373 1374 if (clk_stop_enable) { 1375 /* Configure the PHY to stop receiving xMII 1376 * clock while it is signaling LPI. 1377 */ 1378 int val = phy_read_mmd_indirect(phydev, MDIO_CTRL1, 1379 MDIO_MMD_PCS); 1380 if (val < 0) 1381 return val; 1382 1383 val |= MDIO_PCS_CTRL1_CLKSTOP_EN; 1384 phy_write_mmd_indirect(phydev, MDIO_CTRL1, 1385 MDIO_MMD_PCS, val); 1386 } 1387 1388 return 0; /* EEE supported */ 1389 } 1390 eee_exit_err: 1391 return -EPROTONOSUPPORT; 1392 } 1393 EXPORT_SYMBOL(phy_init_eee); 1394 1395 /** 1396 * phy_get_eee_err - report the EEE wake error count 1397 * @phydev: target phy_device struct 1398 * 1399 * Description: it is to report the number of time where the PHY 1400 * failed to complete its normal wake sequence. 1401 */ 1402 int phy_get_eee_err(struct phy_device *phydev) 1403 { 1404 if (!phydev->drv) 1405 return -EIO; 1406 1407 return phy_read_mmd_indirect(phydev, MDIO_PCS_EEE_WK_ERR, MDIO_MMD_PCS); 1408 } 1409 EXPORT_SYMBOL(phy_get_eee_err); 1410 1411 /** 1412 * phy_ethtool_get_eee - get EEE supported and status 1413 * @phydev: target phy_device struct 1414 * @data: ethtool_eee data 1415 * 1416 * Description: it reportes the Supported/Advertisement/LP Advertisement 1417 * capabilities. 1418 */ 1419 int phy_ethtool_get_eee(struct phy_device *phydev, struct ethtool_eee *data) 1420 { 1421 int val; 1422 1423 if (!phydev->drv) 1424 return -EIO; 1425 1426 /* Get Supported EEE */ 1427 val = phy_read_mmd_indirect(phydev, MDIO_PCS_EEE_ABLE, MDIO_MMD_PCS); 1428 if (val < 0) 1429 return val; 1430 data->supported = mmd_eee_cap_to_ethtool_sup_t(val); 1431 1432 /* Get advertisement EEE */ 1433 val = phy_read_mmd_indirect(phydev, MDIO_AN_EEE_ADV, MDIO_MMD_AN); 1434 if (val < 0) 1435 return val; 1436 data->advertised = mmd_eee_adv_to_ethtool_adv_t(val); 1437 1438 /* Get LP advertisement EEE */ 1439 val = phy_read_mmd_indirect(phydev, MDIO_AN_EEE_LPABLE, MDIO_MMD_AN); 1440 if (val < 0) 1441 return val; 1442 data->lp_advertised = mmd_eee_adv_to_ethtool_adv_t(val); 1443 1444 return 0; 1445 } 1446 EXPORT_SYMBOL(phy_ethtool_get_eee); 1447 1448 /** 1449 * phy_ethtool_set_eee - set EEE supported and status 1450 * @phydev: target phy_device struct 1451 * @data: ethtool_eee data 1452 * 1453 * Description: it is to program the Advertisement EEE register. 1454 */ 1455 int phy_ethtool_set_eee(struct phy_device *phydev, struct ethtool_eee *data) 1456 { 1457 int val = ethtool_adv_to_mmd_eee_adv_t(data->advertised); 1458 1459 if (!phydev->drv) 1460 return -EIO; 1461 1462 /* Mask prohibited EEE modes */ 1463 val &= ~phydev->eee_broken_modes; 1464 1465 phy_write_mmd_indirect(phydev, MDIO_AN_EEE_ADV, MDIO_MMD_AN, val); 1466 1467 return 0; 1468 } 1469 EXPORT_SYMBOL(phy_ethtool_set_eee); 1470 1471 int phy_ethtool_set_wol(struct phy_device *phydev, struct ethtool_wolinfo *wol) 1472 { 1473 if (phydev->drv && phydev->drv->set_wol) 1474 return phydev->drv->set_wol(phydev, wol); 1475 1476 return -EOPNOTSUPP; 1477 } 1478 EXPORT_SYMBOL(phy_ethtool_set_wol); 1479 1480 void phy_ethtool_get_wol(struct phy_device *phydev, struct ethtool_wolinfo *wol) 1481 { 1482 if (phydev->drv && phydev->drv->get_wol) 1483 phydev->drv->get_wol(phydev, wol); 1484 } 1485 EXPORT_SYMBOL(phy_ethtool_get_wol); 1486 1487 int phy_ethtool_get_link_ksettings(struct net_device *ndev, 1488 struct ethtool_link_ksettings *cmd) 1489 { 1490 struct phy_device *phydev = ndev->phydev; 1491 1492 if (!phydev) 1493 return -ENODEV; 1494 1495 return phy_ethtool_ksettings_get(phydev, cmd); 1496 } 1497 EXPORT_SYMBOL(phy_ethtool_get_link_ksettings); 1498 1499 int phy_ethtool_set_link_ksettings(struct net_device *ndev, 1500 const struct ethtool_link_ksettings *cmd) 1501 { 1502 struct phy_device *phydev = ndev->phydev; 1503 1504 if (!phydev) 1505 return -ENODEV; 1506 1507 return phy_ethtool_ksettings_set(phydev, cmd); 1508 } 1509 EXPORT_SYMBOL(phy_ethtool_set_link_ksettings); 1510 1511 int phy_ethtool_nway_reset(struct net_device *ndev) 1512 { 1513 struct phy_device *phydev = ndev->phydev; 1514 1515 if (!phydev) 1516 return -ENODEV; 1517 1518 if (!phydev->drv) 1519 return -EIO; 1520 1521 return genphy_restart_aneg(phydev); 1522 } 1523 EXPORT_SYMBOL(phy_ethtool_nway_reset); 1524