1 // SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause 2 /* 3 * Copyright (C) 2024-2025 Intel Corporation 4 */ 5 #include <linux/rtnetlink.h> 6 #include <net/mac80211.h> 7 8 #include "fw/api/rx.h" 9 #include "fw/api/datapath.h" 10 #include "fw/api/commands.h" 11 #include "fw/api/offload.h" 12 #include "fw/api/coex.h" 13 #include "fw/dbg.h" 14 #include "fw/uefi.h" 15 16 #include "mld.h" 17 #include "mlo.h" 18 #include "mac80211.h" 19 #include "led.h" 20 #include "scan.h" 21 #include "tx.h" 22 #include "sta.h" 23 #include "regulatory.h" 24 #include "thermal.h" 25 #include "low_latency.h" 26 #include "hcmd.h" 27 #include "fw/api/location.h" 28 29 #include "iwl-nvm-parse.h" 30 31 #define DRV_DESCRIPTION "Intel(R) MLD wireless driver for Linux" 32 MODULE_DESCRIPTION(DRV_DESCRIPTION); 33 MODULE_LICENSE("GPL"); 34 MODULE_IMPORT_NS("IWLWIFI"); 35 36 static const struct iwl_op_mode_ops iwl_mld_ops; 37 38 static int __init iwl_mld_init(void) 39 { 40 int ret = iwl_opmode_register("iwlmld", &iwl_mld_ops); 41 42 if (ret) 43 pr_err("Unable to register MLD op_mode: %d\n", ret); 44 45 return ret; 46 } 47 module_init(iwl_mld_init); 48 49 static void __exit iwl_mld_exit(void) 50 { 51 iwl_opmode_deregister("iwlmld"); 52 } 53 module_exit(iwl_mld_exit); 54 55 static void iwl_mld_hw_set_regulatory(struct iwl_mld *mld) 56 { 57 struct wiphy *wiphy = mld->wiphy; 58 59 wiphy->regulatory_flags |= REGULATORY_WIPHY_SELF_MANAGED; 60 wiphy->regulatory_flags |= REGULATORY_ENABLE_RELAX_NO_IR; 61 } 62 63 VISIBLE_IF_IWLWIFI_KUNIT 64 void iwl_construct_mld(struct iwl_mld *mld, struct iwl_trans *trans, 65 const struct iwl_cfg *cfg, const struct iwl_fw *fw, 66 struct ieee80211_hw *hw, struct dentry *dbgfs_dir) 67 { 68 mld->dev = trans->dev; 69 mld->trans = trans; 70 mld->cfg = cfg; 71 mld->fw = fw; 72 mld->hw = hw; 73 mld->wiphy = hw->wiphy; 74 mld->debugfs_dir = dbgfs_dir; 75 76 iwl_notification_wait_init(&mld->notif_wait); 77 78 /* Setup async RX handling */ 79 spin_lock_init(&mld->async_handlers_lock); 80 wiphy_work_init(&mld->async_handlers_wk, 81 iwl_mld_async_handlers_wk); 82 83 /* Dynamic Queue Allocation */ 84 spin_lock_init(&mld->add_txqs_lock); 85 INIT_LIST_HEAD(&mld->txqs_to_add); 86 wiphy_work_init(&mld->add_txqs_wk, iwl_mld_add_txqs_wk); 87 88 /* Setup RX queues sync wait queue */ 89 init_waitqueue_head(&mld->rxq_sync.waitq); 90 } 91 EXPORT_SYMBOL_IF_IWLWIFI_KUNIT(iwl_construct_mld); 92 93 static void __acquires(&mld->wiphy->mtx) 94 iwl_mld_fwrt_dump_start(void *ctx) 95 { 96 struct iwl_mld *mld = ctx; 97 98 wiphy_lock(mld->wiphy); 99 } 100 101 static void __releases(&mld->wiphy->mtx) 102 iwl_mld_fwrt_dump_end(void *ctx) 103 { 104 struct iwl_mld *mld = ctx; 105 106 wiphy_unlock(mld->wiphy); 107 } 108 109 static bool iwl_mld_d3_debug_enable(void *ctx) 110 { 111 return IWL_MLD_D3_DEBUG; 112 } 113 114 static int iwl_mld_fwrt_send_hcmd(void *ctx, struct iwl_host_cmd *host_cmd) 115 { 116 struct iwl_mld *mld = (struct iwl_mld *)ctx; 117 int ret; 118 119 wiphy_lock(mld->wiphy); 120 ret = iwl_mld_send_cmd(mld, host_cmd); 121 wiphy_unlock(mld->wiphy); 122 123 return ret; 124 } 125 126 static const struct iwl_fw_runtime_ops iwl_mld_fwrt_ops = { 127 .dump_start = iwl_mld_fwrt_dump_start, 128 .dump_end = iwl_mld_fwrt_dump_end, 129 .send_hcmd = iwl_mld_fwrt_send_hcmd, 130 .d3_debug_enable = iwl_mld_d3_debug_enable, 131 }; 132 133 static void 134 iwl_mld_construct_fw_runtime(struct iwl_mld *mld, struct iwl_trans *trans, 135 const struct iwl_fw *fw, 136 struct dentry *debugfs_dir) 137 { 138 iwl_fw_runtime_init(&mld->fwrt, trans, fw, &iwl_mld_fwrt_ops, mld, 139 NULL, NULL, debugfs_dir); 140 141 iwl_fw_set_current_image(&mld->fwrt, IWL_UCODE_REGULAR); 142 } 143 144 /* Please keep this array *SORTED* by hex value. 145 * Access is done through binary search 146 */ 147 static const struct iwl_hcmd_names iwl_mld_legacy_names[] = { 148 HCMD_NAME(UCODE_ALIVE_NTFY), 149 HCMD_NAME(INIT_COMPLETE_NOTIF), 150 HCMD_NAME(PHY_CONTEXT_CMD), 151 HCMD_NAME(SCAN_CFG_CMD), 152 HCMD_NAME(SCAN_REQ_UMAC), 153 HCMD_NAME(SCAN_ABORT_UMAC), 154 HCMD_NAME(SCAN_COMPLETE_UMAC), 155 HCMD_NAME(TX_CMD), 156 HCMD_NAME(TXPATH_FLUSH), 157 HCMD_NAME(LEDS_CMD), 158 HCMD_NAME(WNM_80211V_TIMING_MEASUREMENT_NOTIFICATION), 159 HCMD_NAME(WNM_80211V_TIMING_MEASUREMENT_CONFIRM_NOTIFICATION), 160 HCMD_NAME(SCAN_OFFLOAD_UPDATE_PROFILES_CMD), 161 HCMD_NAME(POWER_TABLE_CMD), 162 HCMD_NAME(PSM_UAPSD_AP_MISBEHAVING_NOTIFICATION), 163 HCMD_NAME(BEACON_NOTIFICATION), 164 HCMD_NAME(BEACON_TEMPLATE_CMD), 165 HCMD_NAME(TX_ANT_CONFIGURATION_CMD), 166 HCMD_NAME(REDUCE_TX_POWER_CMD), 167 HCMD_NAME(MISSED_BEACONS_NOTIFICATION), 168 HCMD_NAME(MAC_PM_POWER_TABLE), 169 HCMD_NAME(MFUART_LOAD_NOTIFICATION), 170 HCMD_NAME(RSS_CONFIG_CMD), 171 HCMD_NAME(SCAN_ITERATION_COMPLETE_UMAC), 172 HCMD_NAME(REPLY_RX_MPDU_CMD), 173 HCMD_NAME(BA_NOTIF), 174 HCMD_NAME(MCC_UPDATE_CMD), 175 HCMD_NAME(MCC_CHUB_UPDATE_CMD), 176 HCMD_NAME(MCAST_FILTER_CMD), 177 HCMD_NAME(REPLY_BEACON_FILTERING_CMD), 178 HCMD_NAME(PROT_OFFLOAD_CONFIG_CMD), 179 HCMD_NAME(MATCH_FOUND_NOTIFICATION), 180 HCMD_NAME(WOWLAN_PATTERNS), 181 HCMD_NAME(WOWLAN_CONFIGURATION), 182 HCMD_NAME(WOWLAN_TSC_RSC_PARAM), 183 HCMD_NAME(WOWLAN_KEK_KCK_MATERIAL), 184 HCMD_NAME(DEBUG_HOST_COMMAND), 185 HCMD_NAME(LDBG_CONFIG_CMD), 186 }; 187 188 /* Please keep this array *SORTED* by hex value. 189 * Access is done through binary search 190 */ 191 static const struct iwl_hcmd_names iwl_mld_system_names[] = { 192 HCMD_NAME(SHARED_MEM_CFG_CMD), 193 HCMD_NAME(SOC_CONFIGURATION_CMD), 194 HCMD_NAME(INIT_EXTENDED_CFG_CMD), 195 HCMD_NAME(FW_ERROR_RECOVERY_CMD), 196 HCMD_NAME(RFI_GET_FREQ_TABLE_CMD), 197 HCMD_NAME(SYSTEM_STATISTICS_CMD), 198 HCMD_NAME(SYSTEM_STATISTICS_END_NOTIF), 199 }; 200 201 /* Please keep this array *SORTED* by hex value. 202 * Access is done through binary search 203 */ 204 static const struct iwl_hcmd_names iwl_mld_reg_and_nvm_names[] = { 205 HCMD_NAME(LARI_CONFIG_CHANGE), 206 HCMD_NAME(NVM_GET_INFO), 207 HCMD_NAME(TAS_CONFIG), 208 HCMD_NAME(SAR_OFFSET_MAPPING_TABLE_CMD), 209 HCMD_NAME(MCC_ALLOWED_AP_TYPE_CMD), 210 }; 211 212 /* Please keep this array *SORTED* by hex value. 213 * Access is done through binary search 214 */ 215 static const struct iwl_hcmd_names iwl_mld_debug_names[] = { 216 HCMD_NAME(HOST_EVENT_CFG), 217 HCMD_NAME(DBGC_SUSPEND_RESUME), 218 }; 219 220 /* Please keep this array *SORTED* by hex value. 221 * Access is done through binary search 222 */ 223 static const struct iwl_hcmd_names iwl_mld_mac_conf_names[] = { 224 HCMD_NAME(LOW_LATENCY_CMD), 225 HCMD_NAME(SESSION_PROTECTION_CMD), 226 HCMD_NAME(MAC_CONFIG_CMD), 227 HCMD_NAME(LINK_CONFIG_CMD), 228 HCMD_NAME(STA_CONFIG_CMD), 229 HCMD_NAME(AUX_STA_CMD), 230 HCMD_NAME(STA_REMOVE_CMD), 231 HCMD_NAME(ROC_CMD), 232 HCMD_NAME(MISSED_BEACONS_NOTIF), 233 HCMD_NAME(EMLSR_TRANS_FAIL_NOTIF), 234 HCMD_NAME(ROC_NOTIF), 235 HCMD_NAME(CHANNEL_SWITCH_ERROR_NOTIF), 236 HCMD_NAME(SESSION_PROTECTION_NOTIF), 237 HCMD_NAME(PROBE_RESPONSE_DATA_NOTIF), 238 HCMD_NAME(CHANNEL_SWITCH_START_NOTIF), 239 }; 240 241 /* Please keep this array *SORTED* by hex value. 242 * Access is done through binary search 243 */ 244 static const struct iwl_hcmd_names iwl_mld_data_path_names[] = { 245 HCMD_NAME(TRIGGER_RX_QUEUES_NOTIF_CMD), 246 HCMD_NAME(WNM_PLATFORM_PTM_REQUEST_CMD), 247 HCMD_NAME(WNM_80211V_TIMING_MEASUREMENT_CONFIG_CMD), 248 HCMD_NAME(RFH_QUEUE_CONFIG_CMD), 249 HCMD_NAME(TLC_MNG_CONFIG_CMD), 250 HCMD_NAME(RX_BAID_ALLOCATION_CONFIG_CMD), 251 HCMD_NAME(SCD_QUEUE_CONFIG_CMD), 252 HCMD_NAME(OMI_SEND_STATUS_NOTIF), 253 HCMD_NAME(ESR_MODE_NOTIF), 254 HCMD_NAME(MONITOR_NOTIF), 255 HCMD_NAME(TLC_MNG_UPDATE_NOTIF), 256 HCMD_NAME(MU_GROUP_MGMT_NOTIF), 257 }; 258 259 /* Please keep this array *SORTED* by hex value. 260 * Access is done through binary search 261 */ 262 static const struct iwl_hcmd_names iwl_mld_location_names[] = { 263 HCMD_NAME(TOF_RANGE_REQ_CMD), 264 HCMD_NAME(TOF_RANGE_RESPONSE_NOTIF), 265 }; 266 267 /* Please keep this array *SORTED* by hex value. 268 * Access is done through binary search 269 */ 270 static const struct iwl_hcmd_names iwl_mld_phy_names[] = { 271 HCMD_NAME(CMD_DTS_MEASUREMENT_TRIGGER_WIDE), 272 HCMD_NAME(CTDP_CONFIG_CMD), 273 HCMD_NAME(TEMP_REPORTING_THRESHOLDS_CMD), 274 HCMD_NAME(PER_CHAIN_LIMIT_OFFSET_CMD), 275 HCMD_NAME(CT_KILL_NOTIFICATION), 276 HCMD_NAME(DTS_MEASUREMENT_NOTIF_WIDE), 277 }; 278 279 /* Please keep this array *SORTED* by hex value. 280 * Access is done through binary search 281 */ 282 static const struct iwl_hcmd_names iwl_mld_statistics_names[] = { 283 HCMD_NAME(STATISTICS_OPER_NOTIF), 284 HCMD_NAME(STATISTICS_OPER_PART1_NOTIF), 285 }; 286 287 /* Please keep this array *SORTED* by hex value. 288 * Access is done through binary search 289 */ 290 static const struct iwl_hcmd_names iwl_mld_prot_offload_names[] = { 291 HCMD_NAME(WOWLAN_WAKE_PKT_NOTIFICATION), 292 HCMD_NAME(WOWLAN_INFO_NOTIFICATION), 293 HCMD_NAME(D3_END_NOTIFICATION), 294 }; 295 296 /* Please keep this array *SORTED* by hex value. 297 * Access is done through binary search 298 */ 299 static const struct iwl_hcmd_names iwl_mld_coex_names[] = { 300 HCMD_NAME(PROFILE_NOTIF), 301 }; 302 303 VISIBLE_IF_IWLWIFI_KUNIT 304 const struct iwl_hcmd_arr iwl_mld_groups[] = { 305 [LEGACY_GROUP] = HCMD_ARR(iwl_mld_legacy_names), 306 [LONG_GROUP] = HCMD_ARR(iwl_mld_legacy_names), 307 [SYSTEM_GROUP] = HCMD_ARR(iwl_mld_system_names), 308 [MAC_CONF_GROUP] = HCMD_ARR(iwl_mld_mac_conf_names), 309 [DATA_PATH_GROUP] = HCMD_ARR(iwl_mld_data_path_names), 310 [LOCATION_GROUP] = HCMD_ARR(iwl_mld_location_names), 311 [REGULATORY_AND_NVM_GROUP] = HCMD_ARR(iwl_mld_reg_and_nvm_names), 312 [DEBUG_GROUP] = HCMD_ARR(iwl_mld_debug_names), 313 [PHY_OPS_GROUP] = HCMD_ARR(iwl_mld_phy_names), 314 [STATISTICS_GROUP] = HCMD_ARR(iwl_mld_statistics_names), 315 [PROT_OFFLOAD_GROUP] = HCMD_ARR(iwl_mld_prot_offload_names), 316 [BT_COEX_GROUP] = HCMD_ARR(iwl_mld_coex_names), 317 }; 318 EXPORT_SYMBOL_IF_IWLWIFI_KUNIT(iwl_mld_groups); 319 320 #if IS_ENABLED(CONFIG_IWLWIFI_KUNIT_TESTS) 321 const unsigned int global_iwl_mld_goups_size = ARRAY_SIZE(iwl_mld_groups); 322 EXPORT_SYMBOL_IF_IWLWIFI_KUNIT(global_iwl_mld_goups_size); 323 #endif 324 325 static void 326 iwl_mld_configure_trans(struct iwl_op_mode *op_mode) 327 { 328 const struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 329 static const u8 no_reclaim_cmds[] = {TX_CMD}; 330 struct iwl_trans *trans = mld->trans; 331 332 trans->conf.rx_buf_size = iwl_amsdu_size_to_rxb_size(); 333 trans->conf.command_groups = iwl_mld_groups; 334 trans->conf.command_groups_size = ARRAY_SIZE(iwl_mld_groups); 335 trans->conf.fw_reset_handshake = true; 336 trans->conf.queue_alloc_cmd_ver = 337 iwl_fw_lookup_cmd_ver(mld->fw, WIDE_ID(DATA_PATH_GROUP, 338 SCD_QUEUE_CONFIG_CMD), 339 0); 340 trans->conf.cb_data_offs = offsetof(struct ieee80211_tx_info, 341 driver_data[2]); 342 BUILD_BUG_ON(sizeof(no_reclaim_cmds) > 343 sizeof(trans->conf.no_reclaim_cmds)); 344 memcpy(trans->conf.no_reclaim_cmds, no_reclaim_cmds, 345 sizeof(no_reclaim_cmds)); 346 trans->conf.n_no_reclaim_cmds = ARRAY_SIZE(no_reclaim_cmds); 347 348 trans->conf.rx_mpdu_cmd = REPLY_RX_MPDU_CMD; 349 trans->conf.rx_mpdu_cmd_hdr_size = sizeof(struct iwl_rx_mpdu_res_start); 350 trans->conf.wide_cmd_header = true; 351 352 iwl_trans_op_mode_enter(trans, op_mode); 353 } 354 355 /* 356 ***************************************************** 357 * op mode ops functions 358 ***************************************************** 359 */ 360 361 #define NUM_FW_LOAD_RETRIES 3 362 static struct iwl_op_mode * 363 iwl_op_mode_mld_start(struct iwl_trans *trans, const struct iwl_cfg *cfg, 364 const struct iwl_fw *fw, struct dentry *dbgfs_dir) 365 { 366 struct ieee80211_hw *hw; 367 struct iwl_op_mode *op_mode; 368 struct iwl_mld *mld; 369 u32 eckv_value; 370 int ret; 371 372 /* Allocate and initialize a new hardware device */ 373 hw = ieee80211_alloc_hw(sizeof(struct iwl_op_mode) + 374 sizeof(struct iwl_mld), 375 &iwl_mld_hw_ops); 376 if (!hw) 377 return ERR_PTR(-ENOMEM); 378 379 op_mode = hw->priv; 380 381 op_mode->ops = &iwl_mld_ops; 382 383 mld = IWL_OP_MODE_GET_MLD(op_mode); 384 385 iwl_construct_mld(mld, trans, cfg, fw, hw, dbgfs_dir); 386 387 iwl_mld_construct_fw_runtime(mld, trans, fw, dbgfs_dir); 388 389 iwl_mld_get_bios_tables(mld); 390 iwl_uefi_get_sgom_table(trans, &mld->fwrt); 391 iwl_uefi_get_step_table(trans); 392 if (iwl_bios_get_eckv(&mld->fwrt, &eckv_value)) 393 IWL_DEBUG_RADIO(mld, "ECKV table doesn't exist in BIOS\n"); 394 else 395 trans->ext_32khz_clock_valid = !!eckv_value; 396 iwl_bios_setup_step(trans, &mld->fwrt); 397 mld->bios_enable_puncturing = iwl_uefi_get_puncturing(&mld->fwrt); 398 399 iwl_mld_hw_set_regulatory(mld); 400 401 /* Configure transport layer with the opmode specific params */ 402 iwl_mld_configure_trans(op_mode); 403 404 /* needed for regulatory init */ 405 rtnl_lock(); 406 /* Needed for sending commands */ 407 wiphy_lock(mld->wiphy); 408 409 for (int i = 0; i < NUM_FW_LOAD_RETRIES; i++) { 410 ret = iwl_mld_load_fw(mld); 411 if (!ret) 412 break; 413 } 414 415 if (!ret) { 416 mld->nvm_data = iwl_get_nvm(mld->trans, mld->fw, 0, 0); 417 if (IS_ERR(mld->nvm_data)) { 418 IWL_ERR(mld, "Failed to read NVM: %d\n", ret); 419 ret = PTR_ERR(mld->nvm_data); 420 } 421 } 422 423 if (ret) { 424 wiphy_unlock(mld->wiphy); 425 rtnl_unlock(); 426 goto err; 427 } 428 429 /* We are about to stop the FW. Notifications may require an 430 * operational FW, so handle them all here before we stop. 431 */ 432 wiphy_work_flush(mld->wiphy, &mld->async_handlers_wk); 433 434 iwl_mld_stop_fw(mld); 435 436 wiphy_unlock(mld->wiphy); 437 rtnl_unlock(); 438 439 ret = iwl_mld_leds_init(mld); 440 if (ret) 441 goto free_nvm; 442 443 ret = iwl_mld_alloc_scan_cmd(mld); 444 if (ret) 445 goto leds_exit; 446 447 ret = iwl_mld_low_latency_init(mld); 448 if (ret) 449 goto free_scan_cmd; 450 451 ret = iwl_mld_register_hw(mld); 452 if (ret) 453 goto low_latency_free; 454 455 iwl_mld_toggle_tx_ant(mld, &mld->mgmt_tx_ant); 456 457 iwl_mld_add_debugfs_files(mld, dbgfs_dir); 458 iwl_mld_thermal_initialize(mld); 459 460 iwl_mld_ptp_init(mld); 461 462 return op_mode; 463 464 low_latency_free: 465 iwl_mld_low_latency_free(mld); 466 free_scan_cmd: 467 kfree(mld->scan.cmd); 468 leds_exit: 469 iwl_mld_leds_exit(mld); 470 free_nvm: 471 kfree(mld->nvm_data); 472 err: 473 iwl_trans_op_mode_leave(mld->trans); 474 ieee80211_free_hw(mld->hw); 475 return ERR_PTR(ret); 476 } 477 478 static void 479 iwl_op_mode_mld_stop(struct iwl_op_mode *op_mode) 480 { 481 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 482 483 iwl_mld_ptp_remove(mld); 484 iwl_mld_leds_exit(mld); 485 486 wiphy_lock(mld->wiphy); 487 iwl_mld_thermal_exit(mld); 488 iwl_mld_low_latency_stop(mld); 489 iwl_mld_deinit_time_sync(mld); 490 wiphy_unlock(mld->wiphy); 491 492 ieee80211_unregister_hw(mld->hw); 493 494 iwl_fw_runtime_free(&mld->fwrt); 495 iwl_mld_low_latency_free(mld); 496 497 iwl_trans_op_mode_leave(mld->trans); 498 499 kfree(mld->nvm_data); 500 kfree(mld->scan.cmd); 501 kfree(mld->error_recovery_buf); 502 kfree(mld->mcast_filter_cmd); 503 504 ieee80211_free_hw(mld->hw); 505 } 506 507 static void iwl_mld_queue_state_change(struct iwl_op_mode *op_mode, 508 int hw_queue, bool queue_full) 509 { 510 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 511 struct ieee80211_txq *txq; 512 struct iwl_mld_sta *mld_sta; 513 struct iwl_mld_txq *mld_txq; 514 515 rcu_read_lock(); 516 517 txq = rcu_dereference(mld->fw_id_to_txq[hw_queue]); 518 if (!txq) { 519 rcu_read_unlock(); 520 521 if (queue_full) { 522 /* An internal queue is not expected to become full */ 523 IWL_WARN(mld, 524 "Internal hw_queue %d is full! stopping all queues\n", 525 hw_queue); 526 /* Stop all queues, as an internal queue is not 527 * mapped to a mac80211 one 528 */ 529 ieee80211_stop_queues(mld->hw); 530 } else { 531 ieee80211_wake_queues(mld->hw); 532 } 533 534 return; 535 } 536 537 mld_txq = iwl_mld_txq_from_mac80211(txq); 538 mld_sta = txq->sta ? iwl_mld_sta_from_mac80211(txq->sta) : NULL; 539 540 mld_txq->status.stop_full = queue_full; 541 542 if (!queue_full && mld_sta && 543 mld_sta->sta_state != IEEE80211_STA_NOTEXIST) { 544 local_bh_disable(); 545 iwl_mld_tx_from_txq(mld, txq); 546 local_bh_enable(); 547 } 548 549 rcu_read_unlock(); 550 } 551 552 static void 553 iwl_mld_queue_full(struct iwl_op_mode *op_mode, int hw_queue) 554 { 555 iwl_mld_queue_state_change(op_mode, hw_queue, true); 556 } 557 558 static void 559 iwl_mld_queue_not_full(struct iwl_op_mode *op_mode, int hw_queue) 560 { 561 iwl_mld_queue_state_change(op_mode, hw_queue, false); 562 } 563 564 static bool 565 iwl_mld_set_hw_rfkill_state(struct iwl_op_mode *op_mode, bool state) 566 { 567 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 568 569 iwl_mld_set_hwkill(mld, state); 570 571 return false; 572 } 573 574 static void 575 iwl_mld_free_skb(struct iwl_op_mode *op_mode, struct sk_buff *skb) 576 { 577 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 578 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 579 580 iwl_trans_free_tx_cmd(mld->trans, info->driver_data[1]); 581 ieee80211_free_txskb(mld->hw, skb); 582 } 583 584 static void iwl_mld_read_error_recovery_buffer(struct iwl_mld *mld) 585 { 586 u32 src_size = mld->fw->ucode_capa.error_log_size; 587 u32 src_addr = mld->fw->ucode_capa.error_log_addr; 588 u8 *recovery_buf; 589 int ret; 590 591 /* no recovery buffer size defined in a TLV */ 592 if (!src_size) 593 return; 594 595 recovery_buf = kzalloc(src_size, GFP_ATOMIC); 596 if (!recovery_buf) 597 return; 598 599 ret = iwl_trans_read_mem_bytes(mld->trans, src_addr, 600 recovery_buf, src_size); 601 if (ret) { 602 IWL_ERR(mld, "Failed to read error recovery buffer (%d)\n", 603 ret); 604 kfree(recovery_buf); 605 return; 606 } 607 608 mld->error_recovery_buf = recovery_buf; 609 } 610 611 static void iwl_mld_restart_nic(struct iwl_mld *mld) 612 { 613 iwl_mld_read_error_recovery_buffer(mld); 614 615 mld->fwrt.trans->dbg.restart_required = false; 616 617 ieee80211_restart_hw(mld->hw); 618 } 619 620 static void 621 iwl_mld_nic_error(struct iwl_op_mode *op_mode, 622 enum iwl_fw_error_type type) 623 { 624 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 625 bool trans_dead = test_bit(STATUS_TRANS_DEAD, &mld->trans->status); 626 627 if (type == IWL_ERR_TYPE_CMD_QUEUE_FULL) 628 IWL_ERR(mld, "Command queue full!\n"); 629 else if (!trans_dead && !mld->fw_status.do_not_dump_once) 630 iwl_fwrt_dump_error_logs(&mld->fwrt); 631 632 mld->fw_status.do_not_dump_once = false; 633 634 /* It is necessary to abort any os scan here because mac80211 requires 635 * having the scan cleared before restarting. 636 * We'll reset the scan_status to NONE in restart cleanup in 637 * the next drv_start() call from mac80211. If ieee80211_hw_restart 638 * isn't called scan status will stay busy. 639 */ 640 iwl_mld_report_scan_aborted(mld); 641 642 /* 643 * This should be first thing before trying to collect any 644 * data to avoid endless loops if any HW error happens while 645 * collecting debug data. 646 * It might not actually be true that we'll restart, but the 647 * setting doesn't matter if we're going to be unbound either. 648 */ 649 if (type != IWL_ERR_TYPE_RESET_HS_TIMEOUT) 650 mld->fw_status.in_hw_restart = true; 651 } 652 653 static void iwl_mld_dump_error(struct iwl_op_mode *op_mode, 654 struct iwl_fw_error_dump_mode *mode) 655 { 656 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 657 658 /* if we come in from opmode we have the mutex held */ 659 if (mode->context == IWL_ERR_CONTEXT_FROM_OPMODE) { 660 lockdep_assert_wiphy(mld->wiphy); 661 iwl_fw_error_collect(&mld->fwrt); 662 } else { 663 wiphy_lock(mld->wiphy); 664 if (mode->context != IWL_ERR_CONTEXT_ABORT) 665 iwl_fw_error_collect(&mld->fwrt); 666 wiphy_unlock(mld->wiphy); 667 } 668 } 669 670 static bool iwl_mld_sw_reset(struct iwl_op_mode *op_mode, 671 enum iwl_fw_error_type type) 672 { 673 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 674 675 /* SW reset can happen for TOP error w/o NIC error, so 676 * also abort scan here and set in_hw_restart, when we 677 * had a NIC error both were already done. 678 */ 679 iwl_mld_report_scan_aborted(mld); 680 mld->fw_status.in_hw_restart = true; 681 682 /* Do restart only in the following conditions are met: 683 * - we consider the FW as running 684 * - The trigger that brought us here is defined as one that requires 685 * a restart (in the debug TLVs) 686 */ 687 if (!mld->fw_status.running || !mld->fwrt.trans->dbg.restart_required) 688 return false; 689 690 iwl_mld_restart_nic(mld); 691 return true; 692 } 693 694 static void 695 iwl_mld_time_point(struct iwl_op_mode *op_mode, 696 enum iwl_fw_ini_time_point tp_id, 697 union iwl_dbg_tlv_tp_data *tp_data) 698 { 699 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 700 701 iwl_dbg_tlv_time_point(&mld->fwrt, tp_id, tp_data); 702 } 703 704 #ifdef CONFIG_PM_SLEEP 705 static void iwl_mld_device_powered_off(struct iwl_op_mode *op_mode) 706 { 707 struct iwl_mld *mld = IWL_OP_MODE_GET_MLD(op_mode); 708 709 wiphy_lock(mld->wiphy); 710 iwl_mld_stop_fw(mld); 711 mld->fw_status.in_d3 = false; 712 wiphy_unlock(mld->wiphy); 713 } 714 #else 715 static void iwl_mld_device_powered_off(struct iwl_op_mode *op_mode) 716 {} 717 #endif 718 719 static const struct iwl_op_mode_ops iwl_mld_ops = { 720 .start = iwl_op_mode_mld_start, 721 .stop = iwl_op_mode_mld_stop, 722 .rx = iwl_mld_rx, 723 .rx_rss = iwl_mld_rx_rss, 724 .queue_full = iwl_mld_queue_full, 725 .queue_not_full = iwl_mld_queue_not_full, 726 .hw_rf_kill = iwl_mld_set_hw_rfkill_state, 727 .free_skb = iwl_mld_free_skb, 728 .nic_error = iwl_mld_nic_error, 729 .dump_error = iwl_mld_dump_error, 730 .sw_reset = iwl_mld_sw_reset, 731 .time_point = iwl_mld_time_point, 732 .device_powered_off = pm_sleep_ptr(iwl_mld_device_powered_off), 733 }; 734 735 struct iwl_mld_mod_params iwlmld_mod_params = { 736 .power_scheme = IWL_POWER_SCHEME_BPS, 737 }; 738 739 module_param_named(power_scheme, iwlmld_mod_params.power_scheme, int, 0444); 740 MODULE_PARM_DESC(power_scheme, 741 "power management scheme: 1-active, 2-balanced, default: 2"); 742