1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Microchip KSZ9477 switch driver main logic 4 * 5 * Copyright (C) 2017-2019 Microchip Technology Inc. 6 */ 7 8 #include <linux/kernel.h> 9 #include <linux/module.h> 10 #include <linux/iopoll.h> 11 #include <linux/platform_data/microchip-ksz.h> 12 #include <linux/phy.h> 13 #include <linux/if_bridge.h> 14 #include <linux/if_vlan.h> 15 #include <net/dsa.h> 16 #include <net/switchdev.h> 17 18 #include "ksz9477_reg.h" 19 #include "ksz_common.h" 20 #include "ksz9477.h" 21 22 static void ksz_cfg(struct ksz_device *dev, u32 addr, u8 bits, bool set) 23 { 24 regmap_update_bits(dev->regmap[0], addr, bits, set ? bits : 0); 25 } 26 27 static void ksz_port_cfg(struct ksz_device *dev, int port, int offset, u8 bits, 28 bool set) 29 { 30 regmap_update_bits(dev->regmap[0], PORT_CTRL_ADDR(port, offset), 31 bits, set ? bits : 0); 32 } 33 34 static void ksz9477_cfg32(struct ksz_device *dev, u32 addr, u32 bits, bool set) 35 { 36 regmap_update_bits(dev->regmap[2], addr, bits, set ? bits : 0); 37 } 38 39 static void ksz9477_port_cfg32(struct ksz_device *dev, int port, int offset, 40 u32 bits, bool set) 41 { 42 regmap_update_bits(dev->regmap[2], PORT_CTRL_ADDR(port, offset), 43 bits, set ? bits : 0); 44 } 45 46 int ksz9477_change_mtu(struct ksz_device *dev, int port, int mtu) 47 { 48 u16 frame_size; 49 50 if (!dsa_is_cpu_port(dev->ds, port)) 51 return 0; 52 53 frame_size = mtu + VLAN_ETH_HLEN + ETH_FCS_LEN; 54 55 return regmap_update_bits(dev->regmap[1], REG_SW_MTU__2, 56 REG_SW_MTU_MASK, frame_size); 57 } 58 59 static int ksz9477_wait_vlan_ctrl_ready(struct ksz_device *dev) 60 { 61 unsigned int val; 62 63 return regmap_read_poll_timeout(dev->regmap[0], REG_SW_VLAN_CTRL, 64 val, !(val & VLAN_START), 10, 1000); 65 } 66 67 static int ksz9477_get_vlan_table(struct ksz_device *dev, u16 vid, 68 u32 *vlan_table) 69 { 70 int ret; 71 72 mutex_lock(&dev->vlan_mutex); 73 74 ksz_write16(dev, REG_SW_VLAN_ENTRY_INDEX__2, vid & VLAN_INDEX_M); 75 ksz_write8(dev, REG_SW_VLAN_CTRL, VLAN_READ | VLAN_START); 76 77 /* wait to be cleared */ 78 ret = ksz9477_wait_vlan_ctrl_ready(dev); 79 if (ret) { 80 dev_dbg(dev->dev, "Failed to read vlan table\n"); 81 goto exit; 82 } 83 84 ksz_read32(dev, REG_SW_VLAN_ENTRY__4, &vlan_table[0]); 85 ksz_read32(dev, REG_SW_VLAN_ENTRY_UNTAG__4, &vlan_table[1]); 86 ksz_read32(dev, REG_SW_VLAN_ENTRY_PORTS__4, &vlan_table[2]); 87 88 ksz_write8(dev, REG_SW_VLAN_CTRL, 0); 89 90 exit: 91 mutex_unlock(&dev->vlan_mutex); 92 93 return ret; 94 } 95 96 static int ksz9477_set_vlan_table(struct ksz_device *dev, u16 vid, 97 u32 *vlan_table) 98 { 99 int ret; 100 101 mutex_lock(&dev->vlan_mutex); 102 103 ksz_write32(dev, REG_SW_VLAN_ENTRY__4, vlan_table[0]); 104 ksz_write32(dev, REG_SW_VLAN_ENTRY_UNTAG__4, vlan_table[1]); 105 ksz_write32(dev, REG_SW_VLAN_ENTRY_PORTS__4, vlan_table[2]); 106 107 ksz_write16(dev, REG_SW_VLAN_ENTRY_INDEX__2, vid & VLAN_INDEX_M); 108 ksz_write8(dev, REG_SW_VLAN_CTRL, VLAN_START | VLAN_WRITE); 109 110 /* wait to be cleared */ 111 ret = ksz9477_wait_vlan_ctrl_ready(dev); 112 if (ret) { 113 dev_dbg(dev->dev, "Failed to write vlan table\n"); 114 goto exit; 115 } 116 117 ksz_write8(dev, REG_SW_VLAN_CTRL, 0); 118 119 /* update vlan cache table */ 120 dev->vlan_cache[vid].table[0] = vlan_table[0]; 121 dev->vlan_cache[vid].table[1] = vlan_table[1]; 122 dev->vlan_cache[vid].table[2] = vlan_table[2]; 123 124 exit: 125 mutex_unlock(&dev->vlan_mutex); 126 127 return ret; 128 } 129 130 static void ksz9477_read_table(struct ksz_device *dev, u32 *table) 131 { 132 ksz_read32(dev, REG_SW_ALU_VAL_A, &table[0]); 133 ksz_read32(dev, REG_SW_ALU_VAL_B, &table[1]); 134 ksz_read32(dev, REG_SW_ALU_VAL_C, &table[2]); 135 ksz_read32(dev, REG_SW_ALU_VAL_D, &table[3]); 136 } 137 138 static void ksz9477_write_table(struct ksz_device *dev, u32 *table) 139 { 140 ksz_write32(dev, REG_SW_ALU_VAL_A, table[0]); 141 ksz_write32(dev, REG_SW_ALU_VAL_B, table[1]); 142 ksz_write32(dev, REG_SW_ALU_VAL_C, table[2]); 143 ksz_write32(dev, REG_SW_ALU_VAL_D, table[3]); 144 } 145 146 static int ksz9477_wait_alu_ready(struct ksz_device *dev) 147 { 148 unsigned int val; 149 150 return regmap_read_poll_timeout(dev->regmap[2], REG_SW_ALU_CTRL__4, 151 val, !(val & ALU_START), 10, 1000); 152 } 153 154 static int ksz9477_wait_alu_sta_ready(struct ksz_device *dev) 155 { 156 unsigned int val; 157 158 return regmap_read_poll_timeout(dev->regmap[2], 159 REG_SW_ALU_STAT_CTRL__4, 160 val, !(val & ALU_STAT_START), 161 10, 1000); 162 } 163 164 int ksz9477_reset_switch(struct ksz_device *dev) 165 { 166 u8 data8; 167 u32 data32; 168 169 /* reset switch */ 170 ksz_cfg(dev, REG_SW_OPERATION, SW_RESET, true); 171 172 /* turn off SPI DO Edge select */ 173 regmap_update_bits(dev->regmap[0], REG_SW_GLOBAL_SERIAL_CTRL_0, 174 SPI_AUTO_EDGE_DETECTION, 0); 175 176 /* default configuration */ 177 ksz_read8(dev, REG_SW_LUE_CTRL_1, &data8); 178 data8 = SW_AGING_ENABLE | SW_LINK_AUTO_AGING | 179 SW_SRC_ADDR_FILTER | SW_FLUSH_STP_TABLE | SW_FLUSH_MSTP_TABLE; 180 ksz_write8(dev, REG_SW_LUE_CTRL_1, data8); 181 182 /* disable interrupts */ 183 ksz_write32(dev, REG_SW_INT_MASK__4, SWITCH_INT_MASK); 184 ksz_write32(dev, REG_SW_PORT_INT_MASK__4, 0x7F); 185 ksz_read32(dev, REG_SW_PORT_INT_STATUS__4, &data32); 186 187 /* KSZ9893 compatible chips do not support refclk configuration */ 188 if (dev->chip_id == KSZ9893_CHIP_ID || 189 dev->chip_id == KSZ8563_CHIP_ID || 190 dev->chip_id == KSZ9563_CHIP_ID) 191 return 0; 192 193 data8 = SW_ENABLE_REFCLKO; 194 if (dev->synclko_disable) 195 data8 = 0; 196 else if (dev->synclko_125) 197 data8 = SW_ENABLE_REFCLKO | SW_REFCLKO_IS_125MHZ; 198 ksz_write8(dev, REG_SW_GLOBAL_OUTPUT_CTRL__1, data8); 199 200 return 0; 201 } 202 203 void ksz9477_r_mib_cnt(struct ksz_device *dev, int port, u16 addr, u64 *cnt) 204 { 205 struct ksz_port *p = &dev->ports[port]; 206 unsigned int val; 207 u32 data; 208 int ret; 209 210 /* retain the flush/freeze bit */ 211 data = p->freeze ? MIB_COUNTER_FLUSH_FREEZE : 0; 212 data |= MIB_COUNTER_READ; 213 data |= (addr << MIB_COUNTER_INDEX_S); 214 ksz_pwrite32(dev, port, REG_PORT_MIB_CTRL_STAT__4, data); 215 216 ret = regmap_read_poll_timeout(dev->regmap[2], 217 PORT_CTRL_ADDR(port, REG_PORT_MIB_CTRL_STAT__4), 218 val, !(val & MIB_COUNTER_READ), 10, 1000); 219 /* failed to read MIB. get out of loop */ 220 if (ret) { 221 dev_dbg(dev->dev, "Failed to get MIB\n"); 222 return; 223 } 224 225 /* count resets upon read */ 226 ksz_pread32(dev, port, REG_PORT_MIB_DATA, &data); 227 *cnt += data; 228 } 229 230 void ksz9477_r_mib_pkt(struct ksz_device *dev, int port, u16 addr, 231 u64 *dropped, u64 *cnt) 232 { 233 addr = dev->info->mib_names[addr].index; 234 ksz9477_r_mib_cnt(dev, port, addr, cnt); 235 } 236 237 void ksz9477_freeze_mib(struct ksz_device *dev, int port, bool freeze) 238 { 239 u32 val = freeze ? MIB_COUNTER_FLUSH_FREEZE : 0; 240 struct ksz_port *p = &dev->ports[port]; 241 242 /* enable/disable the port for flush/freeze function */ 243 mutex_lock(&p->mib.cnt_mutex); 244 ksz_pwrite32(dev, port, REG_PORT_MIB_CTRL_STAT__4, val); 245 246 /* used by MIB counter reading code to know freeze is enabled */ 247 p->freeze = freeze; 248 mutex_unlock(&p->mib.cnt_mutex); 249 } 250 251 void ksz9477_port_init_cnt(struct ksz_device *dev, int port) 252 { 253 struct ksz_port_mib *mib = &dev->ports[port].mib; 254 255 /* flush all enabled port MIB counters */ 256 mutex_lock(&mib->cnt_mutex); 257 ksz_pwrite32(dev, port, REG_PORT_MIB_CTRL_STAT__4, 258 MIB_COUNTER_FLUSH_FREEZE); 259 ksz_write8(dev, REG_SW_MAC_CTRL_6, SW_MIB_COUNTER_FLUSH); 260 ksz_pwrite32(dev, port, REG_PORT_MIB_CTRL_STAT__4, 0); 261 mutex_unlock(&mib->cnt_mutex); 262 } 263 264 static void ksz9477_r_phy_quirks(struct ksz_device *dev, u16 addr, u16 reg, 265 u16 *data) 266 { 267 /* KSZ8563R do not have extended registers but BMSR_ESTATEN and 268 * BMSR_ERCAP bits are set. 269 */ 270 if (dev->chip_id == KSZ8563_CHIP_ID && reg == MII_BMSR) 271 *data &= ~(BMSR_ESTATEN | BMSR_ERCAP); 272 } 273 274 int ksz9477_r_phy(struct ksz_device *dev, u16 addr, u16 reg, u16 *data) 275 { 276 u16 val = 0xffff; 277 int ret; 278 279 /* No real PHY after this. Simulate the PHY. 280 * A fixed PHY can be setup in the device tree, but this function is 281 * still called for that port during initialization. 282 * For RGMII PHY there is no way to access it so the fixed PHY should 283 * be used. For SGMII PHY the supporting code will be added later. 284 */ 285 if (!dev->info->internal_phy[addr]) { 286 struct ksz_port *p = &dev->ports[addr]; 287 288 switch (reg) { 289 case MII_BMCR: 290 val = 0x1140; 291 break; 292 case MII_BMSR: 293 val = 0x796d; 294 break; 295 case MII_PHYSID1: 296 val = 0x0022; 297 break; 298 case MII_PHYSID2: 299 val = 0x1631; 300 break; 301 case MII_ADVERTISE: 302 val = 0x05e1; 303 break; 304 case MII_LPA: 305 val = 0xc5e1; 306 break; 307 case MII_CTRL1000: 308 val = 0x0700; 309 break; 310 case MII_STAT1000: 311 if (p->phydev.speed == SPEED_1000) 312 val = 0x3800; 313 else 314 val = 0; 315 break; 316 } 317 } else { 318 ret = ksz_pread16(dev, addr, 0x100 + (reg << 1), &val); 319 if (ret) 320 return ret; 321 322 ksz9477_r_phy_quirks(dev, addr, reg, &val); 323 } 324 325 *data = val; 326 327 return 0; 328 } 329 330 int ksz9477_w_phy(struct ksz_device *dev, u16 addr, u16 reg, u16 val) 331 { 332 /* No real PHY after this. */ 333 if (!dev->info->internal_phy[addr]) 334 return 0; 335 336 return ksz_pwrite16(dev, addr, 0x100 + (reg << 1), val); 337 } 338 339 void ksz9477_cfg_port_member(struct ksz_device *dev, int port, u8 member) 340 { 341 ksz_pwrite32(dev, port, REG_PORT_VLAN_MEMBERSHIP__4, member); 342 } 343 344 void ksz9477_flush_dyn_mac_table(struct ksz_device *dev, int port) 345 { 346 const u16 *regs = dev->info->regs; 347 u8 data; 348 349 regmap_update_bits(dev->regmap[0], REG_SW_LUE_CTRL_2, 350 SW_FLUSH_OPTION_M << SW_FLUSH_OPTION_S, 351 SW_FLUSH_OPTION_DYN_MAC << SW_FLUSH_OPTION_S); 352 353 if (port < dev->info->port_cnt) { 354 /* flush individual port */ 355 ksz_pread8(dev, port, regs[P_STP_CTRL], &data); 356 if (!(data & PORT_LEARN_DISABLE)) 357 ksz_pwrite8(dev, port, regs[P_STP_CTRL], 358 data | PORT_LEARN_DISABLE); 359 ksz_cfg(dev, S_FLUSH_TABLE_CTRL, SW_FLUSH_DYN_MAC_TABLE, true); 360 ksz_pwrite8(dev, port, regs[P_STP_CTRL], data); 361 } else { 362 /* flush all */ 363 ksz_cfg(dev, S_FLUSH_TABLE_CTRL, SW_FLUSH_STP_TABLE, true); 364 } 365 } 366 367 int ksz9477_port_vlan_filtering(struct ksz_device *dev, int port, 368 bool flag, struct netlink_ext_ack *extack) 369 { 370 if (flag) { 371 ksz_port_cfg(dev, port, REG_PORT_LUE_CTRL, 372 PORT_VLAN_LOOKUP_VID_0, true); 373 ksz_cfg(dev, REG_SW_LUE_CTRL_0, SW_VLAN_ENABLE, true); 374 } else { 375 ksz_cfg(dev, REG_SW_LUE_CTRL_0, SW_VLAN_ENABLE, false); 376 ksz_port_cfg(dev, port, REG_PORT_LUE_CTRL, 377 PORT_VLAN_LOOKUP_VID_0, false); 378 } 379 380 return 0; 381 } 382 383 int ksz9477_port_vlan_add(struct ksz_device *dev, int port, 384 const struct switchdev_obj_port_vlan *vlan, 385 struct netlink_ext_ack *extack) 386 { 387 u32 vlan_table[3]; 388 bool untagged = vlan->flags & BRIDGE_VLAN_INFO_UNTAGGED; 389 int err; 390 391 err = ksz9477_get_vlan_table(dev, vlan->vid, vlan_table); 392 if (err) { 393 NL_SET_ERR_MSG_MOD(extack, "Failed to get vlan table"); 394 return err; 395 } 396 397 vlan_table[0] = VLAN_VALID | (vlan->vid & VLAN_FID_M); 398 if (untagged) 399 vlan_table[1] |= BIT(port); 400 else 401 vlan_table[1] &= ~BIT(port); 402 vlan_table[1] &= ~(BIT(dev->cpu_port)); 403 404 vlan_table[2] |= BIT(port) | BIT(dev->cpu_port); 405 406 err = ksz9477_set_vlan_table(dev, vlan->vid, vlan_table); 407 if (err) { 408 NL_SET_ERR_MSG_MOD(extack, "Failed to set vlan table"); 409 return err; 410 } 411 412 /* change PVID */ 413 if (vlan->flags & BRIDGE_VLAN_INFO_PVID) 414 ksz_pwrite16(dev, port, REG_PORT_DEFAULT_VID, vlan->vid); 415 416 return 0; 417 } 418 419 int ksz9477_port_vlan_del(struct ksz_device *dev, int port, 420 const struct switchdev_obj_port_vlan *vlan) 421 { 422 bool untagged = vlan->flags & BRIDGE_VLAN_INFO_UNTAGGED; 423 u32 vlan_table[3]; 424 u16 pvid; 425 426 ksz_pread16(dev, port, REG_PORT_DEFAULT_VID, &pvid); 427 pvid = pvid & 0xFFF; 428 429 if (ksz9477_get_vlan_table(dev, vlan->vid, vlan_table)) { 430 dev_dbg(dev->dev, "Failed to get vlan table\n"); 431 return -ETIMEDOUT; 432 } 433 434 vlan_table[2] &= ~BIT(port); 435 436 if (pvid == vlan->vid) 437 pvid = 1; 438 439 if (untagged) 440 vlan_table[1] &= ~BIT(port); 441 442 if (ksz9477_set_vlan_table(dev, vlan->vid, vlan_table)) { 443 dev_dbg(dev->dev, "Failed to set vlan table\n"); 444 return -ETIMEDOUT; 445 } 446 447 ksz_pwrite16(dev, port, REG_PORT_DEFAULT_VID, pvid); 448 449 return 0; 450 } 451 452 int ksz9477_fdb_add(struct ksz_device *dev, int port, 453 const unsigned char *addr, u16 vid, struct dsa_db db) 454 { 455 u32 alu_table[4]; 456 u32 data; 457 int ret = 0; 458 459 mutex_lock(&dev->alu_mutex); 460 461 /* find any entry with mac & vid */ 462 data = vid << ALU_FID_INDEX_S; 463 data |= ((addr[0] << 8) | addr[1]); 464 ksz_write32(dev, REG_SW_ALU_INDEX_0, data); 465 466 data = ((addr[2] << 24) | (addr[3] << 16)); 467 data |= ((addr[4] << 8) | addr[5]); 468 ksz_write32(dev, REG_SW_ALU_INDEX_1, data); 469 470 /* start read operation */ 471 ksz_write32(dev, REG_SW_ALU_CTRL__4, ALU_READ | ALU_START); 472 473 /* wait to be finished */ 474 ret = ksz9477_wait_alu_ready(dev); 475 if (ret) { 476 dev_dbg(dev->dev, "Failed to read ALU\n"); 477 goto exit; 478 } 479 480 /* read ALU entry */ 481 ksz9477_read_table(dev, alu_table); 482 483 /* update ALU entry */ 484 alu_table[0] = ALU_V_STATIC_VALID; 485 alu_table[1] |= BIT(port); 486 if (vid) 487 alu_table[1] |= ALU_V_USE_FID; 488 alu_table[2] = (vid << ALU_V_FID_S); 489 alu_table[2] |= ((addr[0] << 8) | addr[1]); 490 alu_table[3] = ((addr[2] << 24) | (addr[3] << 16)); 491 alu_table[3] |= ((addr[4] << 8) | addr[5]); 492 493 ksz9477_write_table(dev, alu_table); 494 495 ksz_write32(dev, REG_SW_ALU_CTRL__4, ALU_WRITE | ALU_START); 496 497 /* wait to be finished */ 498 ret = ksz9477_wait_alu_ready(dev); 499 if (ret) 500 dev_dbg(dev->dev, "Failed to write ALU\n"); 501 502 exit: 503 mutex_unlock(&dev->alu_mutex); 504 505 return ret; 506 } 507 508 int ksz9477_fdb_del(struct ksz_device *dev, int port, 509 const unsigned char *addr, u16 vid, struct dsa_db db) 510 { 511 u32 alu_table[4]; 512 u32 data; 513 int ret = 0; 514 515 mutex_lock(&dev->alu_mutex); 516 517 /* read any entry with mac & vid */ 518 data = vid << ALU_FID_INDEX_S; 519 data |= ((addr[0] << 8) | addr[1]); 520 ksz_write32(dev, REG_SW_ALU_INDEX_0, data); 521 522 data = ((addr[2] << 24) | (addr[3] << 16)); 523 data |= ((addr[4] << 8) | addr[5]); 524 ksz_write32(dev, REG_SW_ALU_INDEX_1, data); 525 526 /* start read operation */ 527 ksz_write32(dev, REG_SW_ALU_CTRL__4, ALU_READ | ALU_START); 528 529 /* wait to be finished */ 530 ret = ksz9477_wait_alu_ready(dev); 531 if (ret) { 532 dev_dbg(dev->dev, "Failed to read ALU\n"); 533 goto exit; 534 } 535 536 ksz_read32(dev, REG_SW_ALU_VAL_A, &alu_table[0]); 537 if (alu_table[0] & ALU_V_STATIC_VALID) { 538 ksz_read32(dev, REG_SW_ALU_VAL_B, &alu_table[1]); 539 ksz_read32(dev, REG_SW_ALU_VAL_C, &alu_table[2]); 540 ksz_read32(dev, REG_SW_ALU_VAL_D, &alu_table[3]); 541 542 /* clear forwarding port */ 543 alu_table[1] &= ~BIT(port); 544 545 /* if there is no port to forward, clear table */ 546 if ((alu_table[1] & ALU_V_PORT_MAP) == 0) { 547 alu_table[0] = 0; 548 alu_table[1] = 0; 549 alu_table[2] = 0; 550 alu_table[3] = 0; 551 } 552 } else { 553 alu_table[0] = 0; 554 alu_table[1] = 0; 555 alu_table[2] = 0; 556 alu_table[3] = 0; 557 } 558 559 ksz9477_write_table(dev, alu_table); 560 561 ksz_write32(dev, REG_SW_ALU_CTRL__4, ALU_WRITE | ALU_START); 562 563 /* wait to be finished */ 564 ret = ksz9477_wait_alu_ready(dev); 565 if (ret) 566 dev_dbg(dev->dev, "Failed to write ALU\n"); 567 568 exit: 569 mutex_unlock(&dev->alu_mutex); 570 571 return ret; 572 } 573 574 static void ksz9477_convert_alu(struct alu_struct *alu, u32 *alu_table) 575 { 576 alu->is_static = !!(alu_table[0] & ALU_V_STATIC_VALID); 577 alu->is_src_filter = !!(alu_table[0] & ALU_V_SRC_FILTER); 578 alu->is_dst_filter = !!(alu_table[0] & ALU_V_DST_FILTER); 579 alu->prio_age = (alu_table[0] >> ALU_V_PRIO_AGE_CNT_S) & 580 ALU_V_PRIO_AGE_CNT_M; 581 alu->mstp = alu_table[0] & ALU_V_MSTP_M; 582 583 alu->is_override = !!(alu_table[1] & ALU_V_OVERRIDE); 584 alu->is_use_fid = !!(alu_table[1] & ALU_V_USE_FID); 585 alu->port_forward = alu_table[1] & ALU_V_PORT_MAP; 586 587 alu->fid = (alu_table[2] >> ALU_V_FID_S) & ALU_V_FID_M; 588 589 alu->mac[0] = (alu_table[2] >> 8) & 0xFF; 590 alu->mac[1] = alu_table[2] & 0xFF; 591 alu->mac[2] = (alu_table[3] >> 24) & 0xFF; 592 alu->mac[3] = (alu_table[3] >> 16) & 0xFF; 593 alu->mac[4] = (alu_table[3] >> 8) & 0xFF; 594 alu->mac[5] = alu_table[3] & 0xFF; 595 } 596 597 int ksz9477_fdb_dump(struct ksz_device *dev, int port, 598 dsa_fdb_dump_cb_t *cb, void *data) 599 { 600 int ret = 0; 601 u32 ksz_data; 602 u32 alu_table[4]; 603 struct alu_struct alu; 604 int timeout; 605 606 mutex_lock(&dev->alu_mutex); 607 608 /* start ALU search */ 609 ksz_write32(dev, REG_SW_ALU_CTRL__4, ALU_START | ALU_SEARCH); 610 611 do { 612 timeout = 1000; 613 do { 614 ksz_read32(dev, REG_SW_ALU_CTRL__4, &ksz_data); 615 if ((ksz_data & ALU_VALID) || !(ksz_data & ALU_START)) 616 break; 617 usleep_range(1, 10); 618 } while (timeout-- > 0); 619 620 if (!timeout) { 621 dev_dbg(dev->dev, "Failed to search ALU\n"); 622 ret = -ETIMEDOUT; 623 goto exit; 624 } 625 626 if (!(ksz_data & ALU_VALID)) 627 continue; 628 629 /* read ALU table */ 630 ksz9477_read_table(dev, alu_table); 631 632 ksz9477_convert_alu(&alu, alu_table); 633 634 if (alu.port_forward & BIT(port)) { 635 ret = cb(alu.mac, alu.fid, alu.is_static, data); 636 if (ret) 637 goto exit; 638 } 639 } while (ksz_data & ALU_START); 640 641 exit: 642 643 /* stop ALU search */ 644 ksz_write32(dev, REG_SW_ALU_CTRL__4, 0); 645 646 mutex_unlock(&dev->alu_mutex); 647 648 return ret; 649 } 650 651 int ksz9477_mdb_add(struct ksz_device *dev, int port, 652 const struct switchdev_obj_port_mdb *mdb, struct dsa_db db) 653 { 654 u32 static_table[4]; 655 const u8 *shifts; 656 const u32 *masks; 657 u32 data; 658 int index; 659 u32 mac_hi, mac_lo; 660 int err = 0; 661 662 shifts = dev->info->shifts; 663 masks = dev->info->masks; 664 665 mac_hi = ((mdb->addr[0] << 8) | mdb->addr[1]); 666 mac_lo = ((mdb->addr[2] << 24) | (mdb->addr[3] << 16)); 667 mac_lo |= ((mdb->addr[4] << 8) | mdb->addr[5]); 668 669 mutex_lock(&dev->alu_mutex); 670 671 for (index = 0; index < dev->info->num_statics; index++) { 672 /* find empty slot first */ 673 data = (index << shifts[ALU_STAT_INDEX]) | 674 masks[ALU_STAT_READ] | ALU_STAT_START; 675 ksz_write32(dev, REG_SW_ALU_STAT_CTRL__4, data); 676 677 /* wait to be finished */ 678 err = ksz9477_wait_alu_sta_ready(dev); 679 if (err) { 680 dev_dbg(dev->dev, "Failed to read ALU STATIC\n"); 681 goto exit; 682 } 683 684 /* read ALU static table */ 685 ksz9477_read_table(dev, static_table); 686 687 if (static_table[0] & ALU_V_STATIC_VALID) { 688 /* check this has same vid & mac address */ 689 if (((static_table[2] >> ALU_V_FID_S) == mdb->vid) && 690 ((static_table[2] & ALU_V_MAC_ADDR_HI) == mac_hi) && 691 static_table[3] == mac_lo) { 692 /* found matching one */ 693 break; 694 } 695 } else { 696 /* found empty one */ 697 break; 698 } 699 } 700 701 /* no available entry */ 702 if (index == dev->info->num_statics) { 703 err = -ENOSPC; 704 goto exit; 705 } 706 707 /* add entry */ 708 static_table[0] = ALU_V_STATIC_VALID; 709 static_table[1] |= BIT(port); 710 if (mdb->vid) 711 static_table[1] |= ALU_V_USE_FID; 712 static_table[2] = (mdb->vid << ALU_V_FID_S); 713 static_table[2] |= mac_hi; 714 static_table[3] = mac_lo; 715 716 ksz9477_write_table(dev, static_table); 717 718 data = (index << shifts[ALU_STAT_INDEX]) | ALU_STAT_START; 719 ksz_write32(dev, REG_SW_ALU_STAT_CTRL__4, data); 720 721 /* wait to be finished */ 722 if (ksz9477_wait_alu_sta_ready(dev)) 723 dev_dbg(dev->dev, "Failed to read ALU STATIC\n"); 724 725 exit: 726 mutex_unlock(&dev->alu_mutex); 727 return err; 728 } 729 730 int ksz9477_mdb_del(struct ksz_device *dev, int port, 731 const struct switchdev_obj_port_mdb *mdb, struct dsa_db db) 732 { 733 u32 static_table[4]; 734 const u8 *shifts; 735 const u32 *masks; 736 u32 data; 737 int index; 738 int ret = 0; 739 u32 mac_hi, mac_lo; 740 741 shifts = dev->info->shifts; 742 masks = dev->info->masks; 743 744 mac_hi = ((mdb->addr[0] << 8) | mdb->addr[1]); 745 mac_lo = ((mdb->addr[2] << 24) | (mdb->addr[3] << 16)); 746 mac_lo |= ((mdb->addr[4] << 8) | mdb->addr[5]); 747 748 mutex_lock(&dev->alu_mutex); 749 750 for (index = 0; index < dev->info->num_statics; index++) { 751 /* find empty slot first */ 752 data = (index << shifts[ALU_STAT_INDEX]) | 753 masks[ALU_STAT_READ] | ALU_STAT_START; 754 ksz_write32(dev, REG_SW_ALU_STAT_CTRL__4, data); 755 756 /* wait to be finished */ 757 ret = ksz9477_wait_alu_sta_ready(dev); 758 if (ret) { 759 dev_dbg(dev->dev, "Failed to read ALU STATIC\n"); 760 goto exit; 761 } 762 763 /* read ALU static table */ 764 ksz9477_read_table(dev, static_table); 765 766 if (static_table[0] & ALU_V_STATIC_VALID) { 767 /* check this has same vid & mac address */ 768 769 if (((static_table[2] >> ALU_V_FID_S) == mdb->vid) && 770 ((static_table[2] & ALU_V_MAC_ADDR_HI) == mac_hi) && 771 static_table[3] == mac_lo) { 772 /* found matching one */ 773 break; 774 } 775 } 776 } 777 778 /* no available entry */ 779 if (index == dev->info->num_statics) 780 goto exit; 781 782 /* clear port */ 783 static_table[1] &= ~BIT(port); 784 785 if ((static_table[1] & ALU_V_PORT_MAP) == 0) { 786 /* delete entry */ 787 static_table[0] = 0; 788 static_table[1] = 0; 789 static_table[2] = 0; 790 static_table[3] = 0; 791 } 792 793 ksz9477_write_table(dev, static_table); 794 795 data = (index << shifts[ALU_STAT_INDEX]) | ALU_STAT_START; 796 ksz_write32(dev, REG_SW_ALU_STAT_CTRL__4, data); 797 798 /* wait to be finished */ 799 ret = ksz9477_wait_alu_sta_ready(dev); 800 if (ret) 801 dev_dbg(dev->dev, "Failed to read ALU STATIC\n"); 802 803 exit: 804 mutex_unlock(&dev->alu_mutex); 805 806 return ret; 807 } 808 809 int ksz9477_port_mirror_add(struct ksz_device *dev, int port, 810 struct dsa_mall_mirror_tc_entry *mirror, 811 bool ingress, struct netlink_ext_ack *extack) 812 { 813 u8 data; 814 int p; 815 816 /* Limit to one sniffer port 817 * Check if any of the port is already set for sniffing 818 * If yes, instruct the user to remove the previous entry & exit 819 */ 820 for (p = 0; p < dev->info->port_cnt; p++) { 821 /* Skip the current sniffing port */ 822 if (p == mirror->to_local_port) 823 continue; 824 825 ksz_pread8(dev, p, P_MIRROR_CTRL, &data); 826 827 if (data & PORT_MIRROR_SNIFFER) { 828 NL_SET_ERR_MSG_MOD(extack, 829 "Sniffer port is already configured, delete existing rules & retry"); 830 return -EBUSY; 831 } 832 } 833 834 if (ingress) 835 ksz_port_cfg(dev, port, P_MIRROR_CTRL, PORT_MIRROR_RX, true); 836 else 837 ksz_port_cfg(dev, port, P_MIRROR_CTRL, PORT_MIRROR_TX, true); 838 839 /* configure mirror port */ 840 ksz_port_cfg(dev, mirror->to_local_port, P_MIRROR_CTRL, 841 PORT_MIRROR_SNIFFER, true); 842 843 ksz_cfg(dev, S_MIRROR_CTRL, SW_MIRROR_RX_TX, false); 844 845 return 0; 846 } 847 848 void ksz9477_port_mirror_del(struct ksz_device *dev, int port, 849 struct dsa_mall_mirror_tc_entry *mirror) 850 { 851 bool in_use = false; 852 u8 data; 853 int p; 854 855 if (mirror->ingress) 856 ksz_port_cfg(dev, port, P_MIRROR_CTRL, PORT_MIRROR_RX, false); 857 else 858 ksz_port_cfg(dev, port, P_MIRROR_CTRL, PORT_MIRROR_TX, false); 859 860 861 /* Check if any of the port is still referring to sniffer port */ 862 for (p = 0; p < dev->info->port_cnt; p++) { 863 ksz_pread8(dev, p, P_MIRROR_CTRL, &data); 864 865 if ((data & (PORT_MIRROR_RX | PORT_MIRROR_TX))) { 866 in_use = true; 867 break; 868 } 869 } 870 871 /* delete sniffing if there are no other mirroring rules */ 872 if (!in_use) 873 ksz_port_cfg(dev, mirror->to_local_port, P_MIRROR_CTRL, 874 PORT_MIRROR_SNIFFER, false); 875 } 876 877 static phy_interface_t ksz9477_get_interface(struct ksz_device *dev, int port) 878 { 879 phy_interface_t interface; 880 bool gbit; 881 882 if (dev->info->internal_phy[port]) 883 return PHY_INTERFACE_MODE_NA; 884 885 gbit = ksz_get_gbit(dev, port); 886 887 interface = ksz_get_xmii(dev, port, gbit); 888 889 return interface; 890 } 891 892 static void ksz9477_port_mmd_write(struct ksz_device *dev, int port, 893 u8 dev_addr, u16 reg_addr, u16 val) 894 { 895 ksz_pwrite16(dev, port, REG_PORT_PHY_MMD_SETUP, 896 MMD_SETUP(PORT_MMD_OP_INDEX, dev_addr)); 897 ksz_pwrite16(dev, port, REG_PORT_PHY_MMD_INDEX_DATA, reg_addr); 898 ksz_pwrite16(dev, port, REG_PORT_PHY_MMD_SETUP, 899 MMD_SETUP(PORT_MMD_OP_DATA_NO_INCR, dev_addr)); 900 ksz_pwrite16(dev, port, REG_PORT_PHY_MMD_INDEX_DATA, val); 901 } 902 903 static void ksz9477_phy_errata_setup(struct ksz_device *dev, int port) 904 { 905 /* Apply PHY settings to address errata listed in 906 * KSZ9477, KSZ9897, KSZ9896, KSZ9567, KSZ8565 907 * Silicon Errata and Data Sheet Clarification documents: 908 * 909 * Register settings are needed to improve PHY receive performance 910 */ 911 ksz9477_port_mmd_write(dev, port, 0x01, 0x6f, 0xdd0b); 912 ksz9477_port_mmd_write(dev, port, 0x01, 0x8f, 0x6032); 913 ksz9477_port_mmd_write(dev, port, 0x01, 0x9d, 0x248c); 914 ksz9477_port_mmd_write(dev, port, 0x01, 0x75, 0x0060); 915 ksz9477_port_mmd_write(dev, port, 0x01, 0xd3, 0x7777); 916 ksz9477_port_mmd_write(dev, port, 0x1c, 0x06, 0x3008); 917 ksz9477_port_mmd_write(dev, port, 0x1c, 0x08, 0x2001); 918 919 /* Transmit waveform amplitude can be improved 920 * (1000BASE-T, 100BASE-TX, 10BASE-Te) 921 */ 922 ksz9477_port_mmd_write(dev, port, 0x1c, 0x04, 0x00d0); 923 924 /* Energy Efficient Ethernet (EEE) feature select must 925 * be manually disabled (except on KSZ8565 which is 100Mbit) 926 */ 927 if (dev->info->gbit_capable[port]) 928 ksz9477_port_mmd_write(dev, port, 0x07, 0x3c, 0x0000); 929 930 /* Register settings are required to meet data sheet 931 * supply current specifications 932 */ 933 ksz9477_port_mmd_write(dev, port, 0x1c, 0x13, 0x6eff); 934 ksz9477_port_mmd_write(dev, port, 0x1c, 0x14, 0xe6ff); 935 ksz9477_port_mmd_write(dev, port, 0x1c, 0x15, 0x6eff); 936 ksz9477_port_mmd_write(dev, port, 0x1c, 0x16, 0xe6ff); 937 ksz9477_port_mmd_write(dev, port, 0x1c, 0x17, 0x00ff); 938 ksz9477_port_mmd_write(dev, port, 0x1c, 0x18, 0x43ff); 939 ksz9477_port_mmd_write(dev, port, 0x1c, 0x19, 0xc3ff); 940 ksz9477_port_mmd_write(dev, port, 0x1c, 0x1a, 0x6fff); 941 ksz9477_port_mmd_write(dev, port, 0x1c, 0x1b, 0x07ff); 942 ksz9477_port_mmd_write(dev, port, 0x1c, 0x1c, 0x0fff); 943 ksz9477_port_mmd_write(dev, port, 0x1c, 0x1d, 0xe7ff); 944 ksz9477_port_mmd_write(dev, port, 0x1c, 0x1e, 0xefff); 945 ksz9477_port_mmd_write(dev, port, 0x1c, 0x20, 0xeeee); 946 } 947 948 void ksz9477_get_caps(struct ksz_device *dev, int port, 949 struct phylink_config *config) 950 { 951 config->mac_capabilities = MAC_10 | MAC_100 | MAC_ASYM_PAUSE | 952 MAC_SYM_PAUSE; 953 954 if (dev->info->gbit_capable[port]) 955 config->mac_capabilities |= MAC_1000FD; 956 } 957 958 int ksz9477_set_ageing_time(struct ksz_device *dev, unsigned int msecs) 959 { 960 u32 secs = msecs / 1000; 961 u8 value; 962 u8 data; 963 int ret; 964 965 value = FIELD_GET(SW_AGE_PERIOD_7_0_M, secs); 966 967 ret = ksz_write8(dev, REG_SW_LUE_CTRL_3, value); 968 if (ret < 0) 969 return ret; 970 971 data = FIELD_GET(SW_AGE_PERIOD_10_8_M, secs); 972 973 ret = ksz_read8(dev, REG_SW_LUE_CTRL_0, &value); 974 if (ret < 0) 975 return ret; 976 977 value &= ~SW_AGE_CNT_M; 978 value |= FIELD_PREP(SW_AGE_CNT_M, data); 979 980 return ksz_write8(dev, REG_SW_LUE_CTRL_0, value); 981 } 982 983 void ksz9477_port_queue_split(struct ksz_device *dev, int port) 984 { 985 u8 data; 986 987 if (dev->info->num_tx_queues == 8) 988 data = PORT_EIGHT_QUEUE; 989 else if (dev->info->num_tx_queues == 4) 990 data = PORT_FOUR_QUEUE; 991 else if (dev->info->num_tx_queues == 2) 992 data = PORT_TWO_QUEUE; 993 else 994 data = PORT_SINGLE_QUEUE; 995 996 ksz_prmw8(dev, port, REG_PORT_CTRL_0, PORT_QUEUE_SPLIT_MASK, data); 997 } 998 999 void ksz9477_port_setup(struct ksz_device *dev, int port, bool cpu_port) 1000 { 1001 struct dsa_switch *ds = dev->ds; 1002 u16 data16; 1003 u8 member; 1004 1005 /* enable tag tail for host port */ 1006 if (cpu_port) 1007 ksz_port_cfg(dev, port, REG_PORT_CTRL_0, PORT_TAIL_TAG_ENABLE, 1008 true); 1009 1010 ksz9477_port_queue_split(dev, port); 1011 1012 ksz_port_cfg(dev, port, REG_PORT_CTRL_0, PORT_MAC_LOOPBACK, false); 1013 1014 /* set back pressure */ 1015 ksz_port_cfg(dev, port, REG_PORT_MAC_CTRL_1, PORT_BACK_PRESSURE, true); 1016 1017 /* enable broadcast storm limit */ 1018 ksz_port_cfg(dev, port, P_BCAST_STORM_CTRL, PORT_BROADCAST_STORM, true); 1019 1020 /* disable DiffServ priority */ 1021 ksz_port_cfg(dev, port, P_PRIO_CTRL, PORT_DIFFSERV_PRIO_ENABLE, false); 1022 1023 /* replace priority */ 1024 ksz_port_cfg(dev, port, REG_PORT_MRI_MAC_CTRL, PORT_USER_PRIO_CEILING, 1025 false); 1026 ksz9477_port_cfg32(dev, port, REG_PORT_MTI_QUEUE_CTRL_0__4, 1027 MTI_PVID_REPLACE, false); 1028 1029 /* enable 802.1p priority */ 1030 ksz_port_cfg(dev, port, P_PRIO_CTRL, PORT_802_1P_PRIO_ENABLE, true); 1031 1032 if (dev->info->internal_phy[port]) { 1033 /* do not force flow control */ 1034 ksz_port_cfg(dev, port, REG_PORT_CTRL_0, 1035 PORT_FORCE_TX_FLOW_CTRL | PORT_FORCE_RX_FLOW_CTRL, 1036 false); 1037 1038 if (dev->info->phy_errata_9477) 1039 ksz9477_phy_errata_setup(dev, port); 1040 } else { 1041 /* force flow control */ 1042 ksz_port_cfg(dev, port, REG_PORT_CTRL_0, 1043 PORT_FORCE_TX_FLOW_CTRL | PORT_FORCE_RX_FLOW_CTRL, 1044 true); 1045 } 1046 1047 if (cpu_port) 1048 member = dsa_user_ports(ds); 1049 else 1050 member = BIT(dsa_upstream_port(ds, port)); 1051 1052 ksz9477_cfg_port_member(dev, port, member); 1053 1054 /* clear pending interrupts */ 1055 if (dev->info->internal_phy[port]) 1056 ksz_pread16(dev, port, REG_PORT_PHY_INT_ENABLE, &data16); 1057 } 1058 1059 void ksz9477_config_cpu_port(struct dsa_switch *ds) 1060 { 1061 struct ksz_device *dev = ds->priv; 1062 struct ksz_port *p; 1063 int i; 1064 1065 for (i = 0; i < dev->info->port_cnt; i++) { 1066 if (dsa_is_cpu_port(ds, i) && 1067 (dev->info->cpu_ports & (1 << i))) { 1068 phy_interface_t interface; 1069 const char *prev_msg; 1070 const char *prev_mode; 1071 1072 dev->cpu_port = i; 1073 p = &dev->ports[i]; 1074 1075 /* Read from XMII register to determine host port 1076 * interface. If set specifically in device tree 1077 * note the difference to help debugging. 1078 */ 1079 interface = ksz9477_get_interface(dev, i); 1080 if (!p->interface) { 1081 if (dev->compat_interface) { 1082 dev_warn(dev->dev, 1083 "Using legacy switch \"phy-mode\" property, because it is missing on port %d node. " 1084 "Please update your device tree.\n", 1085 i); 1086 p->interface = dev->compat_interface; 1087 } else { 1088 p->interface = interface; 1089 } 1090 } 1091 if (interface && interface != p->interface) { 1092 prev_msg = " instead of "; 1093 prev_mode = phy_modes(interface); 1094 } else { 1095 prev_msg = ""; 1096 prev_mode = ""; 1097 } 1098 dev_info(dev->dev, 1099 "Port%d: using phy mode %s%s%s\n", 1100 i, 1101 phy_modes(p->interface), 1102 prev_msg, 1103 prev_mode); 1104 1105 /* enable cpu port */ 1106 ksz9477_port_setup(dev, i, true); 1107 } 1108 } 1109 1110 for (i = 0; i < dev->info->port_cnt; i++) { 1111 if (i == dev->cpu_port) 1112 continue; 1113 ksz_port_stp_state_set(ds, i, BR_STATE_DISABLED); 1114 } 1115 } 1116 1117 int ksz9477_enable_stp_addr(struct ksz_device *dev) 1118 { 1119 const u32 *masks; 1120 u32 data; 1121 int ret; 1122 1123 masks = dev->info->masks; 1124 1125 /* Enable Reserved multicast table */ 1126 ksz_cfg(dev, REG_SW_LUE_CTRL_0, SW_RESV_MCAST_ENABLE, true); 1127 1128 /* Set the Override bit for forwarding BPDU packet to CPU */ 1129 ret = ksz_write32(dev, REG_SW_ALU_VAL_B, 1130 ALU_V_OVERRIDE | BIT(dev->cpu_port)); 1131 if (ret < 0) 1132 return ret; 1133 1134 data = ALU_STAT_START | ALU_RESV_MCAST_ADDR | masks[ALU_STAT_WRITE]; 1135 1136 ret = ksz_write32(dev, REG_SW_ALU_STAT_CTRL__4, data); 1137 if (ret < 0) 1138 return ret; 1139 1140 /* wait to be finished */ 1141 ret = ksz9477_wait_alu_sta_ready(dev); 1142 if (ret < 0) { 1143 dev_err(dev->dev, "Failed to update Reserved Multicast table\n"); 1144 return ret; 1145 } 1146 1147 return 0; 1148 } 1149 1150 int ksz9477_setup(struct dsa_switch *ds) 1151 { 1152 struct ksz_device *dev = ds->priv; 1153 int ret = 0; 1154 1155 ds->mtu_enforcement_ingress = true; 1156 1157 /* Required for port partitioning. */ 1158 ksz9477_cfg32(dev, REG_SW_QM_CTRL__4, UNICAST_VLAN_BOUNDARY, 1159 true); 1160 1161 /* Do not work correctly with tail tagging. */ 1162 ksz_cfg(dev, REG_SW_MAC_CTRL_0, SW_CHECK_LENGTH, false); 1163 1164 /* Enable REG_SW_MTU__2 reg by setting SW_JUMBO_PACKET */ 1165 ksz_cfg(dev, REG_SW_MAC_CTRL_1, SW_JUMBO_PACKET, true); 1166 1167 /* Now we can configure default MTU value */ 1168 ret = regmap_update_bits(dev->regmap[1], REG_SW_MTU__2, REG_SW_MTU_MASK, 1169 VLAN_ETH_FRAME_LEN + ETH_FCS_LEN); 1170 if (ret) 1171 return ret; 1172 1173 /* queue based egress rate limit */ 1174 ksz_cfg(dev, REG_SW_MAC_CTRL_5, SW_OUT_RATE_LIMIT_QUEUE_BASED, true); 1175 1176 /* enable global MIB counter freeze function */ 1177 ksz_cfg(dev, REG_SW_MAC_CTRL_6, SW_MIB_COUNTER_FREEZE, true); 1178 1179 return 0; 1180 } 1181 1182 u32 ksz9477_get_port_addr(int port, int offset) 1183 { 1184 return PORT_CTRL_ADDR(port, offset); 1185 } 1186 1187 int ksz9477_tc_cbs_set_cinc(struct ksz_device *dev, int port, u32 val) 1188 { 1189 val = val >> 8; 1190 1191 return ksz_pwrite16(dev, port, REG_PORT_MTI_CREDIT_INCREMENT, val); 1192 } 1193 1194 int ksz9477_switch_init(struct ksz_device *dev) 1195 { 1196 u8 data8; 1197 int ret; 1198 1199 dev->port_mask = (1 << dev->info->port_cnt) - 1; 1200 1201 /* turn off SPI DO Edge select */ 1202 ret = ksz_read8(dev, REG_SW_GLOBAL_SERIAL_CTRL_0, &data8); 1203 if (ret) 1204 return ret; 1205 1206 data8 &= ~SPI_AUTO_EDGE_DETECTION; 1207 ret = ksz_write8(dev, REG_SW_GLOBAL_SERIAL_CTRL_0, data8); 1208 if (ret) 1209 return ret; 1210 1211 return 0; 1212 } 1213 1214 void ksz9477_switch_exit(struct ksz_device *dev) 1215 { 1216 ksz9477_reset_switch(dev); 1217 } 1218 1219 MODULE_AUTHOR("Woojung Huh <Woojung.Huh@microchip.com>"); 1220 MODULE_DESCRIPTION("Microchip KSZ9477 Series Switch DSA Driver"); 1221 MODULE_LICENSE("GPL"); 1222