1 // SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause 2 /* 3 * Copyright (C) 2024-2025 Intel Corporation 4 */ 5 #include <linux/rtnetlink.h> 6 #include <net/mac80211.h> 7 8 #include "fw/api/rx.h" 9 #include "fw/api/datapath.h" 10 #include "fw/api/commands.h" 11 #include "fw/api/offload.h" 12 #include "fw/api/coex.h" 13 #include "fw/dbg.h" 14 #include "fw/uefi.h" 15 16 #include "mld.h" 17 #include "mlo.h" 18 #include "mac80211.h" 19 #include "led.h" 20 #include "scan.h" 21 #include "tx.h" 22 #include "sta.h" 23 #include "regulatory.h" 24 #include "thermal.h" 25 #include "low_latency.h" 26 #include "hcmd.h" 27 #include "fw/api/location.h" 28 29 #include "iwl-nvm-parse.h" 30 31 #define DRV_DESCRIPTION "Intel(R) MLD wireless driver for Linux" 32 MODULE_DESCRIPTION(DRV_DESCRIPTION); 33 MODULE_LICENSE("GPL"); 34 MODULE_IMPORT_NS("IWLWIFI"); 35 36 static const struct iwl_op_mode_ops iwl_mld_ops; 37 38 static int __init iwl_mld_init(void) 39 { 40 int ret = iwl_opmode_register("iwlmld", &iwl_mld_ops); 41 42 if (ret) 43 pr_err("Unable to register MLD op_mode: %d\n", ret); 44 45 return ret; 46 } 47 module_init(iwl_mld_init); 48 49 static void __exit iwl_mld_exit(void) 50 { 51 iwl_opmode_deregister("iwlmld"); 52 } 53 module_exit(iwl_mld_exit); 54 55 static void iwl_mld_hw_set_regulatory(struct iwl_mld *mld) 56 { 57 struct wiphy *wiphy = mld->wiphy; 58 59 wiphy->regulatory_flags |= REGULATORY_WIPHY_SELF_MANAGED; 60 wiphy->regulatory_flags |= REGULATORY_ENABLE_RELAX_NO_IR; 61 } 62 63 VISIBLE_IF_IWLWIFI_KUNIT 64 void iwl_construct_mld(struct iwl_mld *mld, struct iwl_trans *trans, 65 const struct iwl_cfg *cfg, const struct iwl_fw *fw, 66 struct ieee80211_hw *hw, struct dentry *dbgfs_dir) 67 { 68 mld->dev = trans->dev; 69 mld->trans = trans; 70 mld->cfg = cfg; 71 mld->fw = fw; 72 mld->hw = hw; 73 mld->wiphy = hw->wiphy; 74 mld->debugfs_dir = dbgfs_dir; 75 76 iwl_notification_wait_init(&mld->notif_wait); 77 78 /* Setup async RX handling */ 79 spin_lock_init(&mld->async_handlers_lock); 80 wiphy_work_init(&mld->async_handlers_wk, 81 iwl_mld_async_handlers_wk); 82 83 /* Dynamic Queue Allocation */ 84 spin_lock_init(&mld->add_txqs_lock); 85 INIT_LIST_HEAD(&mld->txqs_to_add); 86 wiphy_work_init(&mld->add_txqs_wk, iwl_mld_add_txqs_wk); 87 88 /* Setup RX queues sync wait queue */ 89 init_waitqueue_head(&mld->rxq_sync.waitq); 90 } 91 EXPORT_SYMBOL_IF_IWLWIFI_KUNIT(iwl_construct_mld); 92 93 static void __acquires(&mld->wiphy->mtx) 94 iwl_mld_fwrt_dump_start(void *ctx) 95 { 96 struct iwl_mld *mld = ctx; 97 98 wiphy_lock(mld->wiphy); 99 } 100 101 static void __releases(&mld->wiphy->mtx) 102 iwl_mld_fwrt_dump_end(void *ctx) 103 { 104 struct iwl_mld *mld = ctx; 105 106 wiphy_unlock(mld->wiphy); 107 } 108 109 static bool iwl_mld_d3_debug_enable(void *ctx) 110 { 111 return IWL_MLD_D3_DEBUG; 112 } 113 114 static int iwl_mld_fwrt_send_hcmd(void *ctx, struct iwl_host_cmd *host_cmd) 115 { 116 struct iwl_mld *mld = (struct iwl_mld *)ctx; 117 int ret; 118 119 wiphy_lock(mld->wiphy); 120 ret = iwl_mld_send_cmd(mld, host_cmd); 121 wiphy_unlock(mld->wiphy); 122 123 return ret; 124 } 125 126 static const struct iwl_fw_runtime_ops iwl_mld_fwrt_ops = { 127 .dump_start = iwl_mld_fwrt_dump_start, 128 .dump_end = iwl_mld_fwrt_dump_end, 129 .send_hcmd = iwl_mld_fwrt_send_hcmd, 130 .d3_debug_enable = iwl_mld_d3_debug_enable, 131 }; 132 133 static void 134 iwl_mld_construct_fw_runtime(struct iwl_mld *mld, struct iwl_trans *trans, 135 const struct iwl_fw *fw, 136 struct dentry *debugfs_dir) 137 { 138 iwl_fw_runtime_init(&mld->fwrt, trans, fw, &iwl_mld_fwrt_ops, mld, 139 NULL, NULL, debugfs_dir); 140 141 iwl_fw_set_current_image(&mld->fwrt, IWL_UCODE_REGULAR); 142 } 143 144 /* Please keep this array *SORTED* by hex value. 145 * Access is done through binary search 146 */ 147 static const struct iwl_hcmd_names iwl_mld_legacy_names[] = { 148 HCMD_NAME(UCODE_ALIVE_NTFY), 149 HCMD_NAME(INIT_COMPLETE_NOTIF), 150 HCMD_NAME(PHY_CONTEXT_CMD), 151 HCMD_NAME(SCAN_CFG_CMD), 152 HCMD_NAME(SCAN_REQ_UMAC), 153 HCMD_NAME(SCAN_ABORT_UMAC), 154 HCMD_NAME(SCAN_COMPLETE_UMAC), 155 HCMD_NAME(TX_CMD), 156 HCMD_NAME(TXPATH_FLUSH), 157 HCMD_NAME(LEDS_CMD), 158 HCMD_NAME(WNM_80211V_TIMING_MEASUREMENT_NOTIFICATION), 159 HCMD_NAME(WNM_80211V_TIMING_MEASUREMENT_CONFIRM_NOTIFICATION), 160 HCMD_NAME(SCAN_OFFLOAD_UPDATE_PROFILES_CMD), 161 HCMD_NAME(POWER_TABLE_CMD), 162 HCMD_NAME(PSM_UAPSD_AP_MISBEHAVING_NOTIFICATION), 163 HCMD_NAME(BEACON_NOTIFICATION), 164 HCMD_NAME(BEACON_TEMPLATE_CMD), 165 HCMD_NAME(TX_ANT_CONFIGURATION_CMD), 166 HCMD_NAME(REDUCE_TX_POWER_CMD), 167 HCMD_NAME(MISSED_BEACONS_NOTIFICATION), 168 HCMD_NAME(MAC_PM_POWER_TABLE), 169 HCMD_NAME(MFUART_LOAD_NOTIFICATION), 170 HCMD_NAME(RSS_CONFIG_CMD), 171 HCMD_NAME(SCAN_ITERATION_COMPLETE_UMAC), 172 HCMD_NAME(REPLY_RX_MPDU_CMD), 173 HCMD_NAME(BA_NOTIF), 174 HCMD_NAME(MCC_UPDATE_CMD), 175 HCMD_NAME(MCC_CHUB_UPDATE_CMD), 176 HCMD_NAME(MCAST_FILTER_CMD), 177 HCMD_NAME(REPLY_BEACON_FILTERING_CMD), 178 HCMD_NAME(PROT_OFFLOAD_CONFIG_CMD), 179 HCMD_NAME(MATCH_FOUND_NOTIFICATION), 180 HCMD_NAME(WOWLAN_PATTERNS), 181 HCMD_NAME(WOWLAN_CONFIGURATION), 182 HCMD_NAME(WOWLAN_TSC_RSC_PARAM), 183 HCMD_NAME(WOWLAN_KEK_KCK_MATERIAL), 184 HCMD_NAME(DEBUG_HOST_COMMAND), 185 HCMD_NAME(LDBG_CONFIG_CMD), 186 }; 187 188 /* Please keep this array *SORTED* by hex value. 189 * Access is done through binary search 190 */ 191 static const struct iwl_hcmd_names iwl_mld_system_names[] = { 192 HCMD_NAME(SHARED_MEM_CFG_CMD), 193 HCMD_NAME(SOC_CONFIGURATION_CMD), 194 HCMD_NAME(INIT_EXTENDED_CFG_CMD), 195 HCMD_NAME(FW_ERROR_RECOVERY_CMD), 196 HCMD_NAME(RFI_GET_FREQ_TABLE_CMD), 197 HCMD_NAME(SYSTEM_STATISTICS_CMD), 198 HCMD_NAME(SYSTEM_STATISTICS_END_NOTIF), 199 }; 200 201 /* Please keep this array *SORTED* by hex value. 202 * Access is done through binary search 203 */ 204 static const struct iwl_hcmd_names iwl_mld_reg_and_nvm_names[] = { 205 HCMD_NAME(LARI_CONFIG_CHANGE), 206 HCMD_NAME(NVM_GET_INFO), 207 HCMD_NAME(TAS_CONFIG), 208 HCMD_NAME(SAR_OFFSET_MAPPING_TABLE_CMD), 209 HCMD_NAME(MCC_ALLOWED_AP_TYPE_CMD), 210 }; 211 212 /* Please keep this array *SORTED* by hex value. 213 * Access is done through binary search 214 */ 215 static const struct iwl_hcmd_names iwl_mld_debug_names[] = { 216 HCMD_NAME(HOST_EVENT_CFG), 217 HCMD_NAME(DBGC_SUSPEND_RESUME), 218 }; 219 220 /* Please keep this array *SORTED* by hex value. 221 * Access is done through binary search 222 */ 223 static const struct iwl_hcmd_names iwl_mld_mac_conf_names[] = { 224 HCMD_NAME(LOW_LATENCY_CMD), 225 HCMD_NAME(SESSION_PROTECTION_CMD), 226 HCMD_NAME(MAC_CONFIG_CMD), 227 HCMD_NAME(LINK_CONFIG_CMD), 228 HCMD_NAME(STA_CONFIG_CMD), 229 HCMD_NAME(AUX_STA_CMD), 230 HCMD_NAME(STA_REMOVE_CMD), 231 HCMD_NAME(ROC_CMD), 232 HCMD_NAME(MISSED_BEACONS_NOTIF), 233 HCMD_NAME(EMLSR_TRANS_FAIL_NOTIF), 234 HCMD_NAME(ROC_NOTIF), 235 HCMD_NAME(CHANNEL_SWITCH_ERROR_NOTIF), 236 HCMD_NAME(SESSION_PROTECTION_NOTIF), 237 HCMD_NAME(PROBE_RESPONSE_DATA_NOTIF), 238 HCMD_NAME(CHANNEL_SWITCH_START_NOTIF), 239 }; 240 241 /* Please keep this array *SORTED* by hex value. 242 * Access is done through binary search 243 */ 244 static const struct iwl_hcmd_names iwl_mld_data_path_names[] = { 245 HCMD_NAME(TRIGGER_RX_QUEUES_NOTIF_CMD), 246 HCMD_NAME(WNM_PLATFORM_PTM_REQUEST_CMD), 247 HCMD_NAME(WNM_80211V_TIMING_MEASUREMENT_CONFIG_CMD), 248 HCMD_NAME(RFH_QUEUE_CONFIG_CMD), 249 HCMD_NAME(TLC_MNG_CONFIG_CMD), 250 HCMD_NAME(RX_BAID_ALLOCATION_CONFIG_CMD), 251 HCMD_NAME(SCD_QUEUE_CONFIG_CMD), 252 HCMD_NAME(OMI_SEND_STATUS_NOTIF), 253 HCMD_NAME(ESR_MODE_NOTIF), 254 HCMD_NAME(MONITOR_NOTIF), 255 HCMD_NAME(TLC_MNG_UPDATE_NOTIF), 256 HCMD_NAME(MU_GROUP_MGMT_NOTIF), 257 }; 258 259 /* Please keep this array *SORTED* by hex value. 260 * Access is done through binary search 261 */ 262 static const struct iwl_hcmd_names iwl_mld_location_names[] = { 263 HCMD_NAME(TOF_RANGE_REQ_CMD), 264 HCMD_NAME(TOF_RANGE_RESPONSE_NOTIF), 265 }; 266 267 /* Please keep this array *SORTED* by hex value. 268 * Access is done through binary search 269 */ 270 static const struct iwl_hcmd_names iwl_mld_phy_names[] = { 271 HCMD_NAME(CMD_DTS_MEASUREMENT_TRIGGER_WIDE), 272 HCMD_NAME(CTDP_CONFIG_CMD), 273 HCMD_NAME(TEMP_REPORTING_THRESHOLDS_CMD), 274 HCMD_NAME(PER_CHAIN_LIMIT_OFFSET_CMD), 275 HCMD_NAME(CT_KILL_NOTIFICATION), 276 HCMD_NAME(DTS_MEASUREMENT_NOTIF_WIDE), 277 }; 278 279 /* Please keep this array *SORTED* by hex value. 280 * Access is done through binary search 281 */ 282 static const struct iwl_hcmd_names iwl_mld_statistics_names[] = { 283 HCMD_NAME(STATISTICS_OPER_NOTIF), 284 HCMD_NAME(STATISTICS_OPER_PART1_NOTIF), 285 }; 286 287 /* Please keep this array *SORTED* by hex value. 288 * Access is done through binary search 289 */ 290 static const struct iwl_hcmd_names iwl_mld_prot_offload_names[] = { 291 HCMD_NAME(WOWLAN_WAKE_PKT_NOTIFICATION), 292 HCMD_NAME(WOWLAN_INFO_NOTIFICATION), 293 HCMD_NAME(D3_END_NOTIFICATION), 294 }; 295 296 /* Please keep this array *SORTED* by hex value. 297 * Access is done through binary search 298 */ 299 static const struct iwl_hcmd_names iwl_mld_coex_names[] = { 300 HCMD_NAME(PROFILE_NOTIF), 301 }; 302 303 VISIBLE_IF_IWLWIFI_KUNIT 304 const struct iwl_hcmd_arr iwl_mld_groups[] = { 305 [LEGACY_GROUP] = HCMD_ARR(iwl_mld_legacy_names), 306 [LONG_GROUP] = HCMD_ARR(iwl_mld_legacy_names), 307 [SYSTEM_GROUP] = HCMD_ARR(iwl_mld_system_names), 308 [MAC_CONF_GROUP] = HCMD_ARR(iwl_mld_mac_conf_names), 309 [DATA_PATH_GROUP] = HCMD_ARR(iwl_mld_data_path_names), 310 [LOCATION_GROUP] = HCMD_ARR(iwl_mld_location_names), 311 [REGULATORY_AND_NVM_GROUP] = HCMD_ARR(iwl_mld_reg_and_nvm_names), 312 [DEBUG_GROUP] = HCMD_ARR(iwl_mld_debug_names), 313 [PHY_OPS_GROUP] = HCMD_ARR(iwl_mld_phy_names), 314 [STATISTICS_GROUP] = HCMD_ARR(iwl_mld_statistics_names), 315 [PROT_OFFLOAD_GROUP] = HCMD_ARR(iwl_mld_prot_offload_names), 316 [BT_COEX_GROUP] = HCMD_ARR(iwl_mld_coex_names), 317 }; 318 EXPORT_SYMBOL_IF_IWLWIFI_KUNIT(iwl_mld_groups); 319 320 #if IS_ENABLED(CONFIG_IWLWIFI_KUNIT_TESTS) 321 const unsigned int global_iwl_mld_goups_size = ARRAY_SIZE(iwl_mld_groups); 322 EXPORT_SYMBOL_IF_IWLWIFI_KUNIT(global_iwl_mld_goups_size); 323 #endif 324 325 static void 326 iwl_mld_configure_trans(struct iwl_op_mode *op_mode) 327 { 328 const struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 329 static const u8 no_reclaim_cmds[] = {TX_CMD}; 330 struct iwl_trans_config trans_cfg = { 331 .op_mode = op_mode, 332 /* Rx is not supported yet, but add it to avoid warnings */ 333 .rx_buf_size = iwl_amsdu_size_to_rxb_size(), 334 .command_groups = iwl_mld_groups, 335 .command_groups_size = ARRAY_SIZE(iwl_mld_groups), 336 .fw_reset_handshake = true, 337 .queue_alloc_cmd_ver = 338 iwl_fw_lookup_cmd_ver(mld->fw, 339 WIDE_ID(DATA_PATH_GROUP, 340 SCD_QUEUE_CONFIG_CMD), 341 0), 342 .no_reclaim_cmds = no_reclaim_cmds, 343 .n_no_reclaim_cmds = ARRAY_SIZE(no_reclaim_cmds), 344 .cb_data_offs = offsetof(struct ieee80211_tx_info, 345 driver_data[2]), 346 }; 347 struct iwl_trans *trans = mld->trans; 348 349 trans->rx_mpdu_cmd = REPLY_RX_MPDU_CMD; 350 trans->rx_mpdu_cmd_hdr_size = sizeof(struct iwl_rx_mpdu_res_start); 351 trans->iml = mld->fw->iml; 352 trans->iml_len = mld->fw->iml_len; 353 trans->wide_cmd_header = true; 354 355 iwl_trans_configure(trans, &trans_cfg); 356 } 357 358 /* 359 ***************************************************** 360 * op mode ops functions 361 ***************************************************** 362 */ 363 364 #define NUM_FW_LOAD_RETRIES 3 365 static struct iwl_op_mode * 366 iwl_op_mode_mld_start(struct iwl_trans *trans, const struct iwl_cfg *cfg, 367 const struct iwl_fw *fw, struct dentry *dbgfs_dir) 368 { 369 struct ieee80211_hw *hw; 370 struct iwl_op_mode *op_mode; 371 struct iwl_mld *mld; 372 u32 eckv_value; 373 int ret; 374 375 /* Allocate and initialize a new hardware device */ 376 hw = ieee80211_alloc_hw(sizeof(struct iwl_op_mode) + 377 sizeof(struct iwl_mld), 378 &iwl_mld_hw_ops); 379 if (!hw) 380 return ERR_PTR(-ENOMEM); 381 382 op_mode = hw->priv; 383 384 op_mode->ops = &iwl_mld_ops; 385 386 mld = IWL_OP_MODE_GET_MLD(op_mode); 387 388 iwl_construct_mld(mld, trans, cfg, fw, hw, dbgfs_dir); 389 390 iwl_mld_construct_fw_runtime(mld, trans, fw, dbgfs_dir); 391 392 iwl_mld_get_bios_tables(mld); 393 iwl_uefi_get_sgom_table(trans, &mld->fwrt); 394 iwl_uefi_get_step_table(trans); 395 if (iwl_bios_get_eckv(&mld->fwrt, &eckv_value)) 396 IWL_DEBUG_RADIO(mld, "ECKV table doesn't exist in BIOS\n"); 397 else 398 trans->ext_32khz_clock_valid = !!eckv_value; 399 iwl_bios_setup_step(trans, &mld->fwrt); 400 mld->bios_enable_puncturing = iwl_uefi_get_puncturing(&mld->fwrt); 401 402 iwl_mld_hw_set_regulatory(mld); 403 404 /* Configure transport layer with the opmode specific params */ 405 iwl_mld_configure_trans(op_mode); 406 407 /* needed for regulatory init */ 408 rtnl_lock(); 409 /* Needed for sending commands */ 410 wiphy_lock(mld->wiphy); 411 412 for (int i = 0; i < NUM_FW_LOAD_RETRIES; i++) { 413 ret = iwl_mld_load_fw(mld); 414 if (!ret) 415 break; 416 } 417 418 if (!ret) { 419 mld->nvm_data = iwl_get_nvm(mld->trans, mld->fw, 0, 0); 420 if (IS_ERR(mld->nvm_data)) { 421 IWL_ERR(mld, "Failed to read NVM: %d\n", ret); 422 ret = PTR_ERR(mld->nvm_data); 423 } 424 } 425 426 if (ret) { 427 wiphy_unlock(mld->wiphy); 428 rtnl_unlock(); 429 goto err; 430 } 431 432 /* We are about to stop the FW. Notifications may require an 433 * operational FW, so handle them all here before we stop. 434 */ 435 wiphy_work_flush(mld->wiphy, &mld->async_handlers_wk); 436 437 iwl_mld_stop_fw(mld); 438 439 wiphy_unlock(mld->wiphy); 440 rtnl_unlock(); 441 442 ret = iwl_mld_leds_init(mld); 443 if (ret) 444 goto free_nvm; 445 446 ret = iwl_mld_alloc_scan_cmd(mld); 447 if (ret) 448 goto leds_exit; 449 450 ret = iwl_mld_low_latency_init(mld); 451 if (ret) 452 goto free_scan_cmd; 453 454 ret = iwl_mld_register_hw(mld); 455 if (ret) 456 goto low_latency_free; 457 458 iwl_mld_toggle_tx_ant(mld, &mld->mgmt_tx_ant); 459 460 iwl_mld_add_debugfs_files(mld, dbgfs_dir); 461 iwl_mld_thermal_initialize(mld); 462 463 iwl_mld_ptp_init(mld); 464 465 return op_mode; 466 467 low_latency_free: 468 iwl_mld_low_latency_free(mld); 469 free_scan_cmd: 470 kfree(mld->scan.cmd); 471 leds_exit: 472 iwl_mld_leds_exit(mld); 473 free_nvm: 474 kfree(mld->nvm_data); 475 err: 476 iwl_trans_op_mode_leave(mld->trans); 477 ieee80211_free_hw(mld->hw); 478 return ERR_PTR(ret); 479 } 480 481 static void 482 iwl_op_mode_mld_stop(struct iwl_op_mode *op_mode) 483 { 484 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 485 486 iwl_mld_ptp_remove(mld); 487 iwl_mld_leds_exit(mld); 488 489 wiphy_lock(mld->wiphy); 490 iwl_mld_thermal_exit(mld); 491 iwl_mld_low_latency_stop(mld); 492 iwl_mld_deinit_time_sync(mld); 493 wiphy_unlock(mld->wiphy); 494 495 ieee80211_unregister_hw(mld->hw); 496 497 iwl_fw_runtime_free(&mld->fwrt); 498 iwl_mld_low_latency_free(mld); 499 500 iwl_trans_op_mode_leave(mld->trans); 501 502 kfree(mld->nvm_data); 503 kfree(mld->scan.cmd); 504 kfree(mld->error_recovery_buf); 505 kfree(mld->mcast_filter_cmd); 506 507 ieee80211_free_hw(mld->hw); 508 } 509 510 static void iwl_mld_queue_state_change(struct iwl_op_mode *op_mode, 511 int hw_queue, bool queue_full) 512 { 513 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 514 struct ieee80211_txq *txq; 515 struct iwl_mld_sta *mld_sta; 516 struct iwl_mld_txq *mld_txq; 517 518 rcu_read_lock(); 519 520 txq = rcu_dereference(mld->fw_id_to_txq[hw_queue]); 521 if (!txq) { 522 rcu_read_unlock(); 523 524 if (queue_full) { 525 /* An internal queue is not expected to become full */ 526 IWL_WARN(mld, 527 "Internal hw_queue %d is full! stopping all queues\n", 528 hw_queue); 529 /* Stop all queues, as an internal queue is not 530 * mapped to a mac80211 one 531 */ 532 ieee80211_stop_queues(mld->hw); 533 } else { 534 ieee80211_wake_queues(mld->hw); 535 } 536 537 return; 538 } 539 540 mld_txq = iwl_mld_txq_from_mac80211(txq); 541 mld_sta = txq->sta ? iwl_mld_sta_from_mac80211(txq->sta) : NULL; 542 543 mld_txq->status.stop_full = queue_full; 544 545 if (!queue_full && mld_sta && 546 mld_sta->sta_state != IEEE80211_STA_NOTEXIST) { 547 local_bh_disable(); 548 iwl_mld_tx_from_txq(mld, txq); 549 local_bh_enable(); 550 } 551 552 rcu_read_unlock(); 553 } 554 555 static void 556 iwl_mld_queue_full(struct iwl_op_mode *op_mode, int hw_queue) 557 { 558 iwl_mld_queue_state_change(op_mode, hw_queue, true); 559 } 560 561 static void 562 iwl_mld_queue_not_full(struct iwl_op_mode *op_mode, int hw_queue) 563 { 564 iwl_mld_queue_state_change(op_mode, hw_queue, false); 565 } 566 567 static bool 568 iwl_mld_set_hw_rfkill_state(struct iwl_op_mode *op_mode, bool state) 569 { 570 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 571 572 iwl_mld_set_hwkill(mld, state); 573 574 return false; 575 } 576 577 static void 578 iwl_mld_free_skb(struct iwl_op_mode *op_mode, struct sk_buff *skb) 579 { 580 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 581 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 582 583 iwl_trans_free_tx_cmd(mld->trans, info->driver_data[1]); 584 ieee80211_free_txskb(mld->hw, skb); 585 } 586 587 static void iwl_mld_read_error_recovery_buffer(struct iwl_mld *mld) 588 { 589 u32 src_size = mld->fw->ucode_capa.error_log_size; 590 u32 src_addr = mld->fw->ucode_capa.error_log_addr; 591 u8 *recovery_buf; 592 int ret; 593 594 /* no recovery buffer size defined in a TLV */ 595 if (!src_size) 596 return; 597 598 recovery_buf = kzalloc(src_size, GFP_ATOMIC); 599 if (!recovery_buf) 600 return; 601 602 ret = iwl_trans_read_mem_bytes(mld->trans, src_addr, 603 recovery_buf, src_size); 604 if (ret) { 605 IWL_ERR(mld, "Failed to read error recovery buffer (%d)\n", 606 ret); 607 kfree(recovery_buf); 608 return; 609 } 610 611 mld->error_recovery_buf = recovery_buf; 612 } 613 614 static void iwl_mld_restart_nic(struct iwl_mld *mld) 615 { 616 iwl_mld_read_error_recovery_buffer(mld); 617 618 mld->fwrt.trans->dbg.restart_required = false; 619 620 ieee80211_restart_hw(mld->hw); 621 } 622 623 static void 624 iwl_mld_nic_error(struct iwl_op_mode *op_mode, 625 enum iwl_fw_error_type type) 626 { 627 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 628 bool trans_dead = test_bit(STATUS_TRANS_DEAD, &mld->trans->status); 629 630 if (type == IWL_ERR_TYPE_CMD_QUEUE_FULL) 631 IWL_ERR(mld, "Command queue full!\n"); 632 else if (!trans_dead && !mld->fw_status.do_not_dump_once) 633 iwl_fwrt_dump_error_logs(&mld->fwrt); 634 635 mld->fw_status.do_not_dump_once = false; 636 637 /* It is necessary to abort any os scan here because mac80211 requires 638 * having the scan cleared before restarting. 639 * We'll reset the scan_status to NONE in restart cleanup in 640 * the next drv_start() call from mac80211. If ieee80211_hw_restart 641 * isn't called scan status will stay busy. 642 */ 643 iwl_mld_report_scan_aborted(mld); 644 645 /* 646 * This should be first thing before trying to collect any 647 * data to avoid endless loops if any HW error happens while 648 * collecting debug data. 649 * It might not actually be true that we'll restart, but the 650 * setting doesn't matter if we're going to be unbound either. 651 */ 652 if (type != IWL_ERR_TYPE_RESET_HS_TIMEOUT) 653 mld->fw_status.in_hw_restart = true; 654 } 655 656 static void iwl_mld_dump_error(struct iwl_op_mode *op_mode, 657 struct iwl_fw_error_dump_mode *mode) 658 { 659 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 660 661 /* if we come in from opmode we have the mutex held */ 662 if (mode->context == IWL_ERR_CONTEXT_FROM_OPMODE) { 663 lockdep_assert_wiphy(mld->wiphy); 664 iwl_fw_error_collect(&mld->fwrt); 665 } else { 666 wiphy_lock(mld->wiphy); 667 if (mode->context != IWL_ERR_CONTEXT_ABORT) 668 iwl_fw_error_collect(&mld->fwrt); 669 wiphy_unlock(mld->wiphy); 670 } 671 } 672 673 static bool iwl_mld_sw_reset(struct iwl_op_mode *op_mode, 674 enum iwl_fw_error_type type) 675 { 676 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 677 678 /* SW reset can happen for TOP error w/o NIC error, so 679 * also abort scan here and set in_hw_restart, when we 680 * had a NIC error both were already done. 681 */ 682 iwl_mld_report_scan_aborted(mld); 683 mld->fw_status.in_hw_restart = true; 684 685 /* Do restart only in the following conditions are met: 686 * - we consider the FW as running 687 * - The trigger that brought us here is defined as one that requires 688 * a restart (in the debug TLVs) 689 */ 690 if (!mld->fw_status.running || !mld->fwrt.trans->dbg.restart_required) 691 return false; 692 693 iwl_mld_restart_nic(mld); 694 return true; 695 } 696 697 static void 698 iwl_mld_time_point(struct iwl_op_mode *op_mode, 699 enum iwl_fw_ini_time_point tp_id, 700 union iwl_dbg_tlv_tp_data *tp_data) 701 { 702 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 703 704 iwl_dbg_tlv_time_point(&mld->fwrt, tp_id, tp_data); 705 } 706 707 #ifdef CONFIG_PM_SLEEP 708 static void iwl_mld_device_powered_off(struct iwl_op_mode *op_mode) 709 { 710 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 711 712 wiphy_lock(mld->wiphy); 713 mld->trans->system_pm_mode = IWL_PLAT_PM_MODE_DISABLED; 714 iwl_mld_stop_fw(mld); 715 mld->fw_status.in_d3 = false; 716 wiphy_unlock(mld->wiphy); 717 } 718 #else 719 static void iwl_mld_device_powered_off(struct iwl_op_mode *op_mode) 720 {} 721 #endif 722 723 static const struct iwl_op_mode_ops iwl_mld_ops = { 724 .start = iwl_op_mode_mld_start, 725 .stop = iwl_op_mode_mld_stop, 726 .rx = iwl_mld_rx, 727 .rx_rss = iwl_mld_rx_rss, 728 .queue_full = iwl_mld_queue_full, 729 .queue_not_full = iwl_mld_queue_not_full, 730 .hw_rf_kill = iwl_mld_set_hw_rfkill_state, 731 .free_skb = iwl_mld_free_skb, 732 .nic_error = iwl_mld_nic_error, 733 .dump_error = iwl_mld_dump_error, 734 .sw_reset = iwl_mld_sw_reset, 735 .time_point = iwl_mld_time_point, 736 .device_powered_off = pm_sleep_ptr(iwl_mld_device_powered_off), 737 }; 738 739 struct iwl_mld_mod_params iwlmld_mod_params = { 740 .power_scheme = IWL_POWER_SCHEME_BPS, 741 }; 742 743 module_param_named(power_scheme, iwlmld_mod_params.power_scheme, int, 0444); 744 MODULE_PARM_DESC(power_scheme, 745 "power management scheme: 1-active, 2-balanced, default: 2"); 746