1 // SPDX-License-Identifier: GPL-2.0-only 2 /****************************************************************************** 3 * 4 * Copyright(c) 2008 - 2014, 2022 Intel Corporation. All rights reserved. 5 *****************************************************************************/ 6 #include <linux/etherdevice.h> 7 #include <linux/kernel.h> 8 #include <linux/module.h> 9 #include <linux/sched.h> 10 #include <net/mac80211.h> 11 12 #include "iwl-io.h" 13 #include "iwl-agn-hw.h" 14 #include "iwl-trans.h" 15 #include "iwl-modparams.h" 16 17 #include "dev.h" 18 #include "agn.h" 19 20 int iwlagn_hw_valid_rtc_data_addr(u32 addr) 21 { 22 return (addr >= IWLAGN_RTC_DATA_LOWER_BOUND) && 23 (addr < IWLAGN_RTC_DATA_UPPER_BOUND); 24 } 25 26 int iwlagn_send_tx_power(struct iwl_priv *priv) 27 { 28 struct iwlagn_tx_power_dbm_cmd tx_power_cmd; 29 u8 tx_ant_cfg_cmd; 30 31 if (WARN_ONCE(test_bit(STATUS_SCAN_HW, &priv->status), 32 "TX Power requested while scanning!\n")) 33 return -EAGAIN; 34 35 /* half dBm need to multiply */ 36 tx_power_cmd.global_lmt = (s8)(2 * priv->tx_power_user_lmt); 37 38 if (tx_power_cmd.global_lmt > priv->nvm_data->max_tx_pwr_half_dbm) { 39 /* 40 * For the newer devices which using enhanced/extend tx power 41 * table in EEPROM, the format is in half dBm. driver need to 42 * convert to dBm format before report to mac80211. 43 * By doing so, there is a possibility of 1/2 dBm resolution 44 * lost. driver will perform "round-up" operation before 45 * reporting, but it will cause 1/2 dBm tx power over the 46 * regulatory limit. Perform the checking here, if the 47 * "tx_power_user_lmt" is higher than EEPROM value (in 48 * half-dBm format), lower the tx power based on EEPROM 49 */ 50 tx_power_cmd.global_lmt = 51 priv->nvm_data->max_tx_pwr_half_dbm; 52 } 53 tx_power_cmd.flags = IWLAGN_TX_POWER_NO_CLOSED; 54 tx_power_cmd.srv_chan_lmt = IWLAGN_TX_POWER_AUTO; 55 56 if (IWL_UCODE_API(priv->fw->ucode_ver) == 1) 57 tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD_V1; 58 else 59 tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD; 60 61 return iwl_dvm_send_cmd_pdu(priv, tx_ant_cfg_cmd, 0, 62 sizeof(tx_power_cmd), &tx_power_cmd); 63 } 64 65 void iwlagn_temperature(struct iwl_priv *priv) 66 { 67 lockdep_assert_held(&priv->statistics.lock); 68 69 /* store temperature from correct statistics (in Celsius) */ 70 priv->temperature = le32_to_cpu(priv->statistics.common.temperature); 71 iwl_tt_handler(priv); 72 } 73 74 int iwlagn_hwrate_to_mac80211_idx(u32 rate_n_flags, enum nl80211_band band) 75 { 76 int idx = 0; 77 int band_offset = 0; 78 79 /* HT rate format: mac80211 wants an MCS number, which is just LSB */ 80 if (rate_n_flags & RATE_MCS_HT_MSK) { 81 idx = (rate_n_flags & 0xff); 82 return idx; 83 /* Legacy rate format, search for match in table */ 84 } else { 85 if (band == NL80211_BAND_5GHZ) 86 band_offset = IWL_FIRST_OFDM_RATE; 87 for (idx = band_offset; idx < IWL_RATE_COUNT_LEGACY; idx++) 88 if (iwl_rates[idx].plcp == (rate_n_flags & 0xFF)) 89 return idx - band_offset; 90 } 91 92 return -1; 93 } 94 95 int iwlagn_manage_ibss_station(struct iwl_priv *priv, 96 struct ieee80211_vif *vif, bool add) 97 { 98 struct iwl_vif_priv *vif_priv = (void *)vif->drv_priv; 99 100 if (add) 101 return iwlagn_add_bssid_station(priv, vif_priv->ctx, 102 vif->bss_conf.bssid, 103 &vif_priv->ibss_bssid_sta_id); 104 return iwl_remove_station(priv, vif_priv->ibss_bssid_sta_id, 105 vif->bss_conf.bssid); 106 } 107 108 /* 109 * iwlagn_txfifo_flush: send REPLY_TXFIFO_FLUSH command to uCode 110 * 111 * pre-requirements: 112 * 1. acquire mutex before calling 113 * 2. make sure rf is on and not in exit state 114 */ 115 int iwlagn_txfifo_flush(struct iwl_priv *priv, u32 scd_q_msk) 116 { 117 struct iwl_txfifo_flush_cmd_v3 flush_cmd_v3 = { 118 .flush_control = cpu_to_le16(IWL_DROP_ALL), 119 }; 120 struct iwl_txfifo_flush_cmd_v2 flush_cmd_v2 = { 121 .flush_control = cpu_to_le16(IWL_DROP_ALL), 122 }; 123 124 u32 queue_control = IWL_SCD_VO_MSK | IWL_SCD_VI_MSK | 125 IWL_SCD_BE_MSK | IWL_SCD_BK_MSK | IWL_SCD_MGMT_MSK; 126 127 if ((priv->valid_contexts != BIT(IWL_RXON_CTX_BSS))) 128 queue_control |= IWL_PAN_SCD_VO_MSK | IWL_PAN_SCD_VI_MSK | 129 IWL_PAN_SCD_BE_MSK | IWL_PAN_SCD_BK_MSK | 130 IWL_PAN_SCD_MGMT_MSK | 131 IWL_PAN_SCD_MULTICAST_MSK; 132 133 if (priv->nvm_data->sku_cap_11n_enable) 134 queue_control |= IWL_AGG_TX_QUEUE_MSK; 135 136 if (scd_q_msk) 137 queue_control = scd_q_msk; 138 139 IWL_DEBUG_INFO(priv, "queue control: 0x%x\n", queue_control); 140 flush_cmd_v3.queue_control = cpu_to_le32(queue_control); 141 flush_cmd_v2.queue_control = cpu_to_le16((u16)queue_control); 142 143 if (IWL_UCODE_API(priv->fw->ucode_ver) > 2) 144 return iwl_dvm_send_cmd_pdu(priv, REPLY_TXFIFO_FLUSH, 0, 145 sizeof(flush_cmd_v3), 146 &flush_cmd_v3); 147 return iwl_dvm_send_cmd_pdu(priv, REPLY_TXFIFO_FLUSH, 0, 148 sizeof(flush_cmd_v2), &flush_cmd_v2); 149 } 150 151 void iwlagn_dev_txfifo_flush(struct iwl_priv *priv) 152 { 153 mutex_lock(&priv->mutex); 154 ieee80211_stop_queues(priv->hw); 155 if (iwlagn_txfifo_flush(priv, 0)) { 156 IWL_ERR(priv, "flush request fail\n"); 157 goto done; 158 } 159 IWL_DEBUG_INFO(priv, "wait transmit/flush all frames\n"); 160 iwl_trans_wait_tx_queues_empty(priv->trans, 0xffffffff); 161 done: 162 ieee80211_wake_queues(priv->hw); 163 mutex_unlock(&priv->mutex); 164 } 165 166 /* 167 * BT coex 168 */ 169 /* Notmal TDM */ 170 static const __le32 iwlagn_def_3w_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = { 171 cpu_to_le32(0xaaaaaaaa), 172 cpu_to_le32(0xaaaaaaaa), 173 cpu_to_le32(0xaeaaaaaa), 174 cpu_to_le32(0xaaaaaaaa), 175 cpu_to_le32(0xcc00ff28), 176 cpu_to_le32(0x0000aaaa), 177 cpu_to_le32(0xcc00aaaa), 178 cpu_to_le32(0x0000aaaa), 179 cpu_to_le32(0xc0004000), 180 cpu_to_le32(0x00004000), 181 cpu_to_le32(0xf0005000), 182 cpu_to_le32(0xf0005000), 183 }; 184 185 /* Full concurrency */ 186 static const __le32 iwlagn_concurrent_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = { 187 cpu_to_le32(0xaaaaaaaa), 188 cpu_to_le32(0xaaaaaaaa), 189 cpu_to_le32(0xaaaaaaaa), 190 cpu_to_le32(0xaaaaaaaa), 191 cpu_to_le32(0xaaaaaaaa), 192 cpu_to_le32(0xaaaaaaaa), 193 cpu_to_le32(0xaaaaaaaa), 194 cpu_to_le32(0xaaaaaaaa), 195 cpu_to_le32(0x00000000), 196 cpu_to_le32(0x00000000), 197 cpu_to_le32(0x00000000), 198 cpu_to_le32(0x00000000), 199 }; 200 201 void iwlagn_send_advance_bt_config(struct iwl_priv *priv) 202 { 203 struct iwl_basic_bt_cmd basic = { 204 .max_kill = IWLAGN_BT_MAX_KILL_DEFAULT, 205 .bt3_timer_t7_value = IWLAGN_BT3_T7_DEFAULT, 206 .bt3_prio_sample_time = IWLAGN_BT3_PRIO_SAMPLE_DEFAULT, 207 .bt3_timer_t2_value = IWLAGN_BT3_T2_DEFAULT, 208 }; 209 struct iwl_bt_cmd_v1 bt_cmd_v1; 210 struct iwl_bt_cmd_v2 bt_cmd_v2; 211 int ret; 212 213 BUILD_BUG_ON(sizeof(iwlagn_def_3w_lookup) != 214 sizeof(basic.bt3_lookup_table)); 215 216 if (priv->lib->bt_params) { 217 /* 218 * newer generation of devices (2000 series and newer) 219 * use the version 2 of the bt command 220 * we need to make sure sending the host command 221 * with correct data structure to avoid uCode assert 222 */ 223 if (priv->lib->bt_params->bt_session_2) { 224 bt_cmd_v2.prio_boost = cpu_to_le32( 225 priv->lib->bt_params->bt_prio_boost); 226 bt_cmd_v2.tx_prio_boost = 0; 227 bt_cmd_v2.rx_prio_boost = 0; 228 } else { 229 /* older version only has 8 bits */ 230 WARN_ON(priv->lib->bt_params->bt_prio_boost & ~0xFF); 231 bt_cmd_v1.prio_boost = 232 priv->lib->bt_params->bt_prio_boost; 233 bt_cmd_v1.tx_prio_boost = 0; 234 bt_cmd_v1.rx_prio_boost = 0; 235 } 236 } else { 237 IWL_ERR(priv, "failed to construct BT Coex Config\n"); 238 return; 239 } 240 241 /* 242 * Possible situations when BT needs to take over for receive, 243 * at the same time where STA needs to response to AP's frame(s), 244 * reduce the tx power of the required response frames, by that, 245 * allow the concurrent BT receive & WiFi transmit 246 * (BT - ANT A, WiFi -ANT B), without interference to one another 247 * 248 * Reduced tx power apply to control frames only (ACK/Back/CTS) 249 * when indicated by the BT config command 250 */ 251 basic.kill_ack_mask = priv->kill_ack_mask; 252 basic.kill_cts_mask = priv->kill_cts_mask; 253 if (priv->reduced_txpower) 254 basic.reduce_txpower = IWLAGN_BT_REDUCED_TX_PWR; 255 basic.valid = priv->bt_valid; 256 257 /* 258 * Configure BT coex mode to "no coexistence" when the 259 * user disabled BT coexistence, we have no interface 260 * (might be in monitor mode), or the interface is in 261 * IBSS mode (no proper uCode support for coex then). 262 */ 263 if (!iwlwifi_mod_params.bt_coex_active || 264 priv->iw_mode == NL80211_IFTYPE_ADHOC) { 265 basic.flags = IWLAGN_BT_FLAG_COEX_MODE_DISABLED; 266 } else { 267 basic.flags = IWLAGN_BT_FLAG_COEX_MODE_3W << 268 IWLAGN_BT_FLAG_COEX_MODE_SHIFT; 269 270 if (!priv->bt_enable_pspoll) 271 basic.flags |= IWLAGN_BT_FLAG_SYNC_2_BT_DISABLE; 272 else 273 basic.flags &= ~IWLAGN_BT_FLAG_SYNC_2_BT_DISABLE; 274 275 if (priv->bt_ch_announce) 276 basic.flags |= IWLAGN_BT_FLAG_CHANNEL_INHIBITION; 277 IWL_DEBUG_COEX(priv, "BT coex flag: 0X%x\n", basic.flags); 278 } 279 priv->bt_enable_flag = basic.flags; 280 if (priv->bt_full_concurrent) 281 memcpy(basic.bt3_lookup_table, iwlagn_concurrent_lookup, 282 sizeof(iwlagn_concurrent_lookup)); 283 else 284 memcpy(basic.bt3_lookup_table, iwlagn_def_3w_lookup, 285 sizeof(iwlagn_def_3w_lookup)); 286 287 IWL_DEBUG_COEX(priv, "BT coex %s in %s mode\n", 288 basic.flags ? "active" : "disabled", 289 priv->bt_full_concurrent ? 290 "full concurrency" : "3-wire"); 291 292 if (priv->lib->bt_params->bt_session_2) { 293 memcpy(&bt_cmd_v2.basic, &basic, 294 sizeof(basic)); 295 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_BT_CONFIG, 296 0, sizeof(bt_cmd_v2), &bt_cmd_v2); 297 } else { 298 memcpy(&bt_cmd_v1.basic, &basic, 299 sizeof(basic)); 300 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_BT_CONFIG, 301 0, sizeof(bt_cmd_v1), &bt_cmd_v1); 302 } 303 if (ret) 304 IWL_ERR(priv, "failed to send BT Coex Config\n"); 305 306 } 307 308 void iwlagn_bt_adjust_rssi_monitor(struct iwl_priv *priv, bool rssi_ena) 309 { 310 struct iwl_rxon_context *ctx, *found_ctx = NULL; 311 bool found_ap = false; 312 313 lockdep_assert_held(&priv->mutex); 314 315 /* Check whether AP or GO mode is active. */ 316 if (rssi_ena) { 317 for_each_context(priv, ctx) { 318 if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_AP && 319 iwl_is_associated_ctx(ctx)) { 320 found_ap = true; 321 break; 322 } 323 } 324 } 325 326 /* 327 * If disable was received or If GO/AP mode, disable RSSI 328 * measurements. 329 */ 330 if (!rssi_ena || found_ap) { 331 if (priv->cur_rssi_ctx) { 332 ctx = priv->cur_rssi_ctx; 333 ieee80211_disable_rssi_reports(ctx->vif); 334 priv->cur_rssi_ctx = NULL; 335 } 336 return; 337 } 338 339 /* 340 * If rssi measurements need to be enabled, consider all cases now. 341 * Figure out how many contexts are active. 342 */ 343 for_each_context(priv, ctx) { 344 if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION && 345 iwl_is_associated_ctx(ctx)) { 346 found_ctx = ctx; 347 break; 348 } 349 } 350 351 /* 352 * rssi monitor already enabled for the correct interface...nothing 353 * to do. 354 */ 355 if (found_ctx == priv->cur_rssi_ctx) 356 return; 357 358 /* 359 * Figure out if rssi monitor is currently enabled, and needs 360 * to be changed. If rssi monitor is already enabled, disable 361 * it first else just enable rssi measurements on the 362 * interface found above. 363 */ 364 if (priv->cur_rssi_ctx) { 365 ctx = priv->cur_rssi_ctx; 366 if (ctx->vif) 367 ieee80211_disable_rssi_reports(ctx->vif); 368 } 369 370 priv->cur_rssi_ctx = found_ctx; 371 372 if (!found_ctx) 373 return; 374 375 ieee80211_enable_rssi_reports(found_ctx->vif, 376 IWLAGN_BT_PSP_MIN_RSSI_THRESHOLD, 377 IWLAGN_BT_PSP_MAX_RSSI_THRESHOLD); 378 } 379 380 static bool iwlagn_bt_traffic_is_sco(struct iwl_bt_uart_msg *uart_msg) 381 { 382 return (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3) >> 383 BT_UART_MSG_FRAME3SCOESCO_POS; 384 } 385 386 static void iwlagn_bt_traffic_change_work(struct work_struct *work) 387 { 388 struct iwl_priv *priv = 389 container_of(work, struct iwl_priv, bt_traffic_change_work); 390 struct iwl_rxon_context *ctx; 391 int smps_request = -1; 392 393 if (priv->bt_enable_flag == IWLAGN_BT_FLAG_COEX_MODE_DISABLED) { 394 /* bt coex disabled */ 395 return; 396 } 397 398 /* 399 * Note: bt_traffic_load can be overridden by scan complete and 400 * coex profile notifications. Ignore that since only bad consequence 401 * can be not matching debug print with actual state. 402 */ 403 IWL_DEBUG_COEX(priv, "BT traffic load changes: %d\n", 404 priv->bt_traffic_load); 405 406 switch (priv->bt_traffic_load) { 407 case IWL_BT_COEX_TRAFFIC_LOAD_NONE: 408 if (priv->bt_status) 409 smps_request = IEEE80211_SMPS_DYNAMIC; 410 else 411 smps_request = IEEE80211_SMPS_AUTOMATIC; 412 break; 413 case IWL_BT_COEX_TRAFFIC_LOAD_LOW: 414 smps_request = IEEE80211_SMPS_DYNAMIC; 415 break; 416 case IWL_BT_COEX_TRAFFIC_LOAD_HIGH: 417 case IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS: 418 smps_request = IEEE80211_SMPS_STATIC; 419 break; 420 default: 421 IWL_ERR(priv, "Invalid BT traffic load: %d\n", 422 priv->bt_traffic_load); 423 break; 424 } 425 426 mutex_lock(&priv->mutex); 427 428 /* 429 * We can not send command to firmware while scanning. When the scan 430 * complete we will schedule this work again. We do check with mutex 431 * locked to prevent new scan request to arrive. We do not check 432 * STATUS_SCANNING to avoid race when queue_work two times from 433 * different notifications, but quit and not perform any work at all. 434 */ 435 if (test_bit(STATUS_SCAN_HW, &priv->status)) 436 goto out; 437 438 iwl_update_chain_flags(priv); 439 440 if (smps_request != -1) { 441 priv->current_ht_config.smps = smps_request; 442 for_each_context(priv, ctx) { 443 if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION) 444 ieee80211_request_smps(ctx->vif, 0, smps_request); 445 } 446 } 447 448 /* 449 * Dynamic PS poll related functionality. Adjust RSSI measurements if 450 * necessary. 451 */ 452 iwlagn_bt_coex_rssi_monitor(priv); 453 out: 454 mutex_unlock(&priv->mutex); 455 } 456 457 /* 458 * If BT sco traffic, and RSSI monitor is enabled, move measurements to the 459 * correct interface or disable it if this is the last interface to be 460 * removed. 461 */ 462 void iwlagn_bt_coex_rssi_monitor(struct iwl_priv *priv) 463 { 464 if (priv->bt_is_sco && 465 priv->bt_traffic_load == IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS) 466 iwlagn_bt_adjust_rssi_monitor(priv, true); 467 else 468 iwlagn_bt_adjust_rssi_monitor(priv, false); 469 } 470 471 static void iwlagn_print_uartmsg(struct iwl_priv *priv, 472 struct iwl_bt_uart_msg *uart_msg) 473 { 474 IWL_DEBUG_COEX(priv, "Message Type = 0x%X, SSN = 0x%X, " 475 "Update Req = 0x%X\n", 476 (BT_UART_MSG_FRAME1MSGTYPE_MSK & uart_msg->frame1) >> 477 BT_UART_MSG_FRAME1MSGTYPE_POS, 478 (BT_UART_MSG_FRAME1SSN_MSK & uart_msg->frame1) >> 479 BT_UART_MSG_FRAME1SSN_POS, 480 (BT_UART_MSG_FRAME1UPDATEREQ_MSK & uart_msg->frame1) >> 481 BT_UART_MSG_FRAME1UPDATEREQ_POS); 482 483 IWL_DEBUG_COEX(priv, "Open connections = 0x%X, Traffic load = 0x%X, " 484 "Chl_SeqN = 0x%X, In band = 0x%X\n", 485 (BT_UART_MSG_FRAME2OPENCONNECTIONS_MSK & uart_msg->frame2) >> 486 BT_UART_MSG_FRAME2OPENCONNECTIONS_POS, 487 (BT_UART_MSG_FRAME2TRAFFICLOAD_MSK & uart_msg->frame2) >> 488 BT_UART_MSG_FRAME2TRAFFICLOAD_POS, 489 (BT_UART_MSG_FRAME2CHLSEQN_MSK & uart_msg->frame2) >> 490 BT_UART_MSG_FRAME2CHLSEQN_POS, 491 (BT_UART_MSG_FRAME2INBAND_MSK & uart_msg->frame2) >> 492 BT_UART_MSG_FRAME2INBAND_POS); 493 494 IWL_DEBUG_COEX(priv, "SCO/eSCO = 0x%X, Sniff = 0x%X, A2DP = 0x%X, " 495 "ACL = 0x%X, Master = 0x%X, OBEX = 0x%X\n", 496 (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3) >> 497 BT_UART_MSG_FRAME3SCOESCO_POS, 498 (BT_UART_MSG_FRAME3SNIFF_MSK & uart_msg->frame3) >> 499 BT_UART_MSG_FRAME3SNIFF_POS, 500 (BT_UART_MSG_FRAME3A2DP_MSK & uart_msg->frame3) >> 501 BT_UART_MSG_FRAME3A2DP_POS, 502 (BT_UART_MSG_FRAME3ACL_MSK & uart_msg->frame3) >> 503 BT_UART_MSG_FRAME3ACL_POS, 504 (BT_UART_MSG_FRAME3MASTER_MSK & uart_msg->frame3) >> 505 BT_UART_MSG_FRAME3MASTER_POS, 506 (BT_UART_MSG_FRAME3OBEX_MSK & uart_msg->frame3) >> 507 BT_UART_MSG_FRAME3OBEX_POS); 508 509 IWL_DEBUG_COEX(priv, "Idle duration = 0x%X\n", 510 (BT_UART_MSG_FRAME4IDLEDURATION_MSK & uart_msg->frame4) >> 511 BT_UART_MSG_FRAME4IDLEDURATION_POS); 512 513 IWL_DEBUG_COEX(priv, "Tx Activity = 0x%X, Rx Activity = 0x%X, " 514 "eSCO Retransmissions = 0x%X\n", 515 (BT_UART_MSG_FRAME5TXACTIVITY_MSK & uart_msg->frame5) >> 516 BT_UART_MSG_FRAME5TXACTIVITY_POS, 517 (BT_UART_MSG_FRAME5RXACTIVITY_MSK & uart_msg->frame5) >> 518 BT_UART_MSG_FRAME5RXACTIVITY_POS, 519 (BT_UART_MSG_FRAME5ESCORETRANSMIT_MSK & uart_msg->frame5) >> 520 BT_UART_MSG_FRAME5ESCORETRANSMIT_POS); 521 522 IWL_DEBUG_COEX(priv, "Sniff Interval = 0x%X, Discoverable = 0x%X\n", 523 (BT_UART_MSG_FRAME6SNIFFINTERVAL_MSK & uart_msg->frame6) >> 524 BT_UART_MSG_FRAME6SNIFFINTERVAL_POS, 525 (BT_UART_MSG_FRAME6DISCOVERABLE_MSK & uart_msg->frame6) >> 526 BT_UART_MSG_FRAME6DISCOVERABLE_POS); 527 528 IWL_DEBUG_COEX(priv, "Sniff Activity = 0x%X, Page = " 529 "0x%X, Inquiry = 0x%X, Connectable = 0x%X\n", 530 (BT_UART_MSG_FRAME7SNIFFACTIVITY_MSK & uart_msg->frame7) >> 531 BT_UART_MSG_FRAME7SNIFFACTIVITY_POS, 532 (BT_UART_MSG_FRAME7PAGE_MSK & uart_msg->frame7) >> 533 BT_UART_MSG_FRAME7PAGE_POS, 534 (BT_UART_MSG_FRAME7INQUIRY_MSK & uart_msg->frame7) >> 535 BT_UART_MSG_FRAME7INQUIRY_POS, 536 (BT_UART_MSG_FRAME7CONNECTABLE_MSK & uart_msg->frame7) >> 537 BT_UART_MSG_FRAME7CONNECTABLE_POS); 538 } 539 540 static bool iwlagn_set_kill_msk(struct iwl_priv *priv, 541 struct iwl_bt_uart_msg *uart_msg) 542 { 543 bool need_update = false; 544 u8 kill_msk = IWL_BT_KILL_REDUCE; 545 static const __le32 bt_kill_ack_msg[3] = { 546 IWLAGN_BT_KILL_ACK_MASK_DEFAULT, 547 IWLAGN_BT_KILL_ACK_CTS_MASK_SCO, 548 IWLAGN_BT_KILL_ACK_CTS_MASK_REDUCE}; 549 static const __le32 bt_kill_cts_msg[3] = { 550 IWLAGN_BT_KILL_CTS_MASK_DEFAULT, 551 IWLAGN_BT_KILL_ACK_CTS_MASK_SCO, 552 IWLAGN_BT_KILL_ACK_CTS_MASK_REDUCE}; 553 554 if (!priv->reduced_txpower) 555 kill_msk = (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3) 556 ? IWL_BT_KILL_OVERRIDE : IWL_BT_KILL_DEFAULT; 557 if (priv->kill_ack_mask != bt_kill_ack_msg[kill_msk] || 558 priv->kill_cts_mask != bt_kill_cts_msg[kill_msk]) { 559 priv->bt_valid |= IWLAGN_BT_VALID_KILL_ACK_MASK; 560 priv->kill_ack_mask = bt_kill_ack_msg[kill_msk]; 561 priv->bt_valid |= IWLAGN_BT_VALID_KILL_CTS_MASK; 562 priv->kill_cts_mask = bt_kill_cts_msg[kill_msk]; 563 need_update = true; 564 } 565 return need_update; 566 } 567 568 /* 569 * Upon RSSI changes, sends a bt config command with following changes 570 * 1. enable/disable "reduced control frames tx power 571 * 2. update the "kill)ack_mask" and "kill_cts_mask" 572 * 573 * If "reduced tx power" is enabled, uCode shall 574 * 1. ACK/Back/CTS rate shall reduced to 6Mbps 575 * 2. not use duplciate 20/40MHz mode 576 */ 577 static bool iwlagn_fill_txpower_mode(struct iwl_priv *priv, 578 struct iwl_bt_uart_msg *uart_msg) 579 { 580 bool need_update = false; 581 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS]; 582 int ave_rssi; 583 584 if (!ctx->vif || (ctx->vif->type != NL80211_IFTYPE_STATION)) { 585 IWL_DEBUG_INFO(priv, "BSS ctx not active or not in sta mode\n"); 586 return false; 587 } 588 589 ave_rssi = ieee80211_ave_rssi(ctx->vif); 590 if (!ave_rssi) { 591 /* no rssi data, no changes to reduce tx power */ 592 IWL_DEBUG_COEX(priv, "no rssi data available\n"); 593 return need_update; 594 } 595 if (!priv->reduced_txpower && 596 !iwl_is_associated(priv, IWL_RXON_CTX_PAN) && 597 (ave_rssi > BT_ENABLE_REDUCED_TXPOWER_THRESHOLD) && 598 (uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK | 599 BT_UART_MSG_FRAME3OBEX_MSK)) && 600 !(uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK | 601 BT_UART_MSG_FRAME3SNIFF_MSK | BT_UART_MSG_FRAME3A2DP_MSK))) { 602 /* enabling reduced tx power */ 603 priv->reduced_txpower = true; 604 priv->bt_valid |= IWLAGN_BT_VALID_REDUCED_TX_PWR; 605 need_update = true; 606 } else if (priv->reduced_txpower && 607 (iwl_is_associated(priv, IWL_RXON_CTX_PAN) || 608 (ave_rssi < BT_DISABLE_REDUCED_TXPOWER_THRESHOLD) || 609 (uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK | 610 BT_UART_MSG_FRAME3SNIFF_MSK | BT_UART_MSG_FRAME3A2DP_MSK)) || 611 !(uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK | 612 BT_UART_MSG_FRAME3OBEX_MSK)))) { 613 /* disable reduced tx power */ 614 priv->reduced_txpower = false; 615 priv->bt_valid |= IWLAGN_BT_VALID_REDUCED_TX_PWR; 616 need_update = true; 617 } 618 619 return need_update; 620 } 621 622 static void iwlagn_bt_coex_profile_notif(struct iwl_priv *priv, 623 struct iwl_rx_cmd_buffer *rxb) 624 { 625 struct iwl_rx_packet *pkt = rxb_addr(rxb); 626 struct iwl_bt_coex_profile_notif *coex = (void *)pkt->data; 627 struct iwl_bt_uart_msg *uart_msg = &coex->last_bt_uart_msg; 628 629 if (priv->bt_enable_flag == IWLAGN_BT_FLAG_COEX_MODE_DISABLED) { 630 /* bt coex disabled */ 631 return; 632 } 633 634 IWL_DEBUG_COEX(priv, "BT Coex notification:\n"); 635 IWL_DEBUG_COEX(priv, " status: %d\n", coex->bt_status); 636 IWL_DEBUG_COEX(priv, " traffic load: %d\n", coex->bt_traffic_load); 637 IWL_DEBUG_COEX(priv, " CI compliance: %d\n", 638 coex->bt_ci_compliance); 639 iwlagn_print_uartmsg(priv, uart_msg); 640 641 priv->last_bt_traffic_load = priv->bt_traffic_load; 642 priv->bt_is_sco = iwlagn_bt_traffic_is_sco(uart_msg); 643 644 if (priv->iw_mode != NL80211_IFTYPE_ADHOC) { 645 if (priv->bt_status != coex->bt_status || 646 priv->last_bt_traffic_load != coex->bt_traffic_load) { 647 if (coex->bt_status) { 648 /* BT on */ 649 if (!priv->bt_ch_announce) 650 priv->bt_traffic_load = 651 IWL_BT_COEX_TRAFFIC_LOAD_HIGH; 652 else 653 priv->bt_traffic_load = 654 coex->bt_traffic_load; 655 } else { 656 /* BT off */ 657 priv->bt_traffic_load = 658 IWL_BT_COEX_TRAFFIC_LOAD_NONE; 659 } 660 priv->bt_status = coex->bt_status; 661 queue_work(priv->workqueue, 662 &priv->bt_traffic_change_work); 663 } 664 } 665 666 /* schedule to send runtime bt_config */ 667 /* check reduce power before change ack/cts kill mask */ 668 if (iwlagn_fill_txpower_mode(priv, uart_msg) || 669 iwlagn_set_kill_msk(priv, uart_msg)) 670 queue_work(priv->workqueue, &priv->bt_runtime_config); 671 672 673 /* FIXME: based on notification, adjust the prio_boost */ 674 675 priv->bt_ci_compliance = coex->bt_ci_compliance; 676 } 677 678 void iwlagn_bt_rx_handler_setup(struct iwl_priv *priv) 679 { 680 priv->rx_handlers[REPLY_BT_COEX_PROFILE_NOTIF] = 681 iwlagn_bt_coex_profile_notif; 682 } 683 684 void iwlagn_bt_setup_deferred_work(struct iwl_priv *priv) 685 { 686 INIT_WORK(&priv->bt_traffic_change_work, 687 iwlagn_bt_traffic_change_work); 688 } 689 690 void iwlagn_bt_cancel_deferred_work(struct iwl_priv *priv) 691 { 692 cancel_work_sync(&priv->bt_traffic_change_work); 693 } 694 695 static bool is_single_rx_stream(struct iwl_priv *priv) 696 { 697 return priv->current_ht_config.smps == IEEE80211_SMPS_STATIC || 698 priv->current_ht_config.single_chain_sufficient; 699 } 700 701 #define IWL_NUM_RX_CHAINS_MULTIPLE 3 702 #define IWL_NUM_RX_CHAINS_SINGLE 2 703 #define IWL_NUM_IDLE_CHAINS_DUAL 2 704 #define IWL_NUM_IDLE_CHAINS_SINGLE 1 705 706 /* 707 * Determine how many receiver/antenna chains to use. 708 * 709 * More provides better reception via diversity. Fewer saves power 710 * at the expense of throughput, but only when not in powersave to 711 * start with. 712 * 713 * MIMO (dual stream) requires at least 2, but works better with 3. 714 * This does not determine *which* chains to use, just how many. 715 */ 716 static int iwl_get_active_rx_chain_count(struct iwl_priv *priv) 717 { 718 if (priv->lib->bt_params && 719 priv->lib->bt_params->advanced_bt_coexist && 720 (priv->bt_full_concurrent || 721 priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)) { 722 /* 723 * only use chain 'A' in bt high traffic load or 724 * full concurrency mode 725 */ 726 return IWL_NUM_RX_CHAINS_SINGLE; 727 } 728 /* # of Rx chains to use when expecting MIMO. */ 729 if (is_single_rx_stream(priv)) 730 return IWL_NUM_RX_CHAINS_SINGLE; 731 else 732 return IWL_NUM_RX_CHAINS_MULTIPLE; 733 } 734 735 /* 736 * When we are in power saving mode, unless device support spatial 737 * multiplexing power save, use the active count for rx chain count. 738 */ 739 static int iwl_get_idle_rx_chain_count(struct iwl_priv *priv, int active_cnt) 740 { 741 /* # Rx chains when idling, depending on SMPS mode */ 742 switch (priv->current_ht_config.smps) { 743 case IEEE80211_SMPS_STATIC: 744 case IEEE80211_SMPS_DYNAMIC: 745 return IWL_NUM_IDLE_CHAINS_SINGLE; 746 case IEEE80211_SMPS_AUTOMATIC: 747 case IEEE80211_SMPS_OFF: 748 return active_cnt; 749 default: 750 WARN(1, "invalid SMPS mode %d", 751 priv->current_ht_config.smps); 752 return active_cnt; 753 } 754 } 755 756 /* up to 4 chains */ 757 static u8 iwl_count_chain_bitmap(u32 chain_bitmap) 758 { 759 u8 res; 760 res = (chain_bitmap & BIT(0)) >> 0; 761 res += (chain_bitmap & BIT(1)) >> 1; 762 res += (chain_bitmap & BIT(2)) >> 2; 763 res += (chain_bitmap & BIT(3)) >> 3; 764 return res; 765 } 766 767 /* 768 * iwlagn_set_rxon_chain - Set up Rx chain usage in "staging" RXON image 769 * 770 * Selects how many and which Rx receivers/antennas/chains to use. 771 * This should not be used for scan command ... it puts data in wrong place. 772 */ 773 void iwlagn_set_rxon_chain(struct iwl_priv *priv, struct iwl_rxon_context *ctx) 774 { 775 bool is_single = is_single_rx_stream(priv); 776 bool is_cam = !test_bit(STATUS_POWER_PMI, &priv->status); 777 u8 idle_rx_cnt, active_rx_cnt, valid_rx_cnt; 778 u32 active_chains; 779 u16 rx_chain; 780 781 /* Tell uCode which antennas are actually connected. 782 * Before first association, we assume all antennas are connected. 783 * Just after first association, iwl_chain_noise_calibration() 784 * checks which antennas actually *are* connected. */ 785 if (priv->chain_noise_data.active_chains) 786 active_chains = priv->chain_noise_data.active_chains; 787 else 788 active_chains = priv->nvm_data->valid_rx_ant; 789 790 if (priv->lib->bt_params && 791 priv->lib->bt_params->advanced_bt_coexist && 792 (priv->bt_full_concurrent || 793 priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)) { 794 /* 795 * only use chain 'A' in bt high traffic load or 796 * full concurrency mode 797 */ 798 active_chains = first_antenna(active_chains); 799 } 800 801 rx_chain = active_chains << RXON_RX_CHAIN_VALID_POS; 802 803 /* How many receivers should we use? */ 804 active_rx_cnt = iwl_get_active_rx_chain_count(priv); 805 idle_rx_cnt = iwl_get_idle_rx_chain_count(priv, active_rx_cnt); 806 807 808 /* correct rx chain count according hw settings 809 * and chain noise calibration 810 */ 811 valid_rx_cnt = iwl_count_chain_bitmap(active_chains); 812 if (valid_rx_cnt < active_rx_cnt) 813 active_rx_cnt = valid_rx_cnt; 814 815 if (valid_rx_cnt < idle_rx_cnt) 816 idle_rx_cnt = valid_rx_cnt; 817 818 rx_chain |= active_rx_cnt << RXON_RX_CHAIN_MIMO_CNT_POS; 819 rx_chain |= idle_rx_cnt << RXON_RX_CHAIN_CNT_POS; 820 821 ctx->staging.rx_chain = cpu_to_le16(rx_chain); 822 823 if (!is_single && (active_rx_cnt >= IWL_NUM_RX_CHAINS_SINGLE) && is_cam) 824 ctx->staging.rx_chain |= RXON_RX_CHAIN_MIMO_FORCE_MSK; 825 else 826 ctx->staging.rx_chain &= ~RXON_RX_CHAIN_MIMO_FORCE_MSK; 827 828 IWL_DEBUG_ASSOC(priv, "rx_chain=0x%X active=%d idle=%d\n", 829 ctx->staging.rx_chain, 830 active_rx_cnt, idle_rx_cnt); 831 832 WARN_ON(active_rx_cnt == 0 || idle_rx_cnt == 0 || 833 active_rx_cnt < idle_rx_cnt); 834 } 835 836 u8 iwl_toggle_tx_ant(struct iwl_priv *priv, u8 ant, u8 valid) 837 { 838 int i; 839 u8 ind = ant; 840 841 if (priv->band == NL80211_BAND_2GHZ && 842 priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH) 843 return 0; 844 845 for (i = 0; i < RATE_ANT_NUM - 1; i++) { 846 ind = (ind + 1) < RATE_ANT_NUM ? ind + 1 : 0; 847 if (valid & BIT(ind)) 848 return ind; 849 } 850 return ant; 851 } 852 853 #ifdef CONFIG_PM_SLEEP 854 static void iwlagn_convert_p1k(u16 *p1k, __le16 *out) 855 { 856 int i; 857 858 for (i = 0; i < IWLAGN_P1K_SIZE; i++) 859 out[i] = cpu_to_le16(p1k[i]); 860 } 861 862 struct wowlan_key_data { 863 struct iwl_rxon_context *ctx; 864 struct iwlagn_wowlan_rsc_tsc_params_cmd *rsc_tsc; 865 struct iwlagn_wowlan_tkip_params_cmd *tkip; 866 const u8 *bssid; 867 bool error, use_rsc_tsc, use_tkip; 868 }; 869 870 871 static void iwlagn_wowlan_program_keys(struct ieee80211_hw *hw, 872 struct ieee80211_vif *vif, 873 struct ieee80211_sta *sta, 874 struct ieee80211_key_conf *key, 875 void *_data) 876 { 877 struct iwl_priv *priv = IWL_MAC80211_GET_DVM(hw); 878 struct wowlan_key_data *data = _data; 879 struct iwl_rxon_context *ctx = data->ctx; 880 struct aes_sc *aes_sc, *aes_tx_sc = NULL; 881 struct tkip_sc *tkip_sc, *tkip_tx_sc = NULL; 882 struct iwlagn_p1k_cache *rx_p1ks; 883 u8 *rx_mic_key; 884 struct ieee80211_key_seq seq; 885 u32 cur_rx_iv32 = 0; 886 u16 p1k[IWLAGN_P1K_SIZE]; 887 int ret, i; 888 889 mutex_lock(&priv->mutex); 890 891 if ((key->cipher == WLAN_CIPHER_SUITE_WEP40 || 892 key->cipher == WLAN_CIPHER_SUITE_WEP104) && 893 !sta && !ctx->key_mapping_keys) 894 ret = iwl_set_default_wep_key(priv, ctx, key); 895 else 896 ret = iwl_set_dynamic_key(priv, ctx, key, sta); 897 898 if (ret) { 899 IWL_ERR(priv, "Error setting key during suspend!\n"); 900 data->error = true; 901 } 902 903 switch (key->cipher) { 904 case WLAN_CIPHER_SUITE_TKIP: 905 if (sta) { 906 u64 pn64; 907 908 tkip_sc = data->rsc_tsc->all_tsc_rsc.tkip.unicast_rsc; 909 tkip_tx_sc = &data->rsc_tsc->all_tsc_rsc.tkip.tsc; 910 911 rx_p1ks = data->tkip->rx_uni; 912 913 pn64 = atomic64_read(&key->tx_pn); 914 tkip_tx_sc->iv16 = cpu_to_le16(TKIP_PN_TO_IV16(pn64)); 915 tkip_tx_sc->iv32 = cpu_to_le32(TKIP_PN_TO_IV32(pn64)); 916 917 ieee80211_get_tkip_p1k_iv(key, seq.tkip.iv32, p1k); 918 iwlagn_convert_p1k(p1k, data->tkip->tx.p1k); 919 920 memcpy(data->tkip->mic_keys.tx, 921 &key->key[NL80211_TKIP_DATA_OFFSET_TX_MIC_KEY], 922 IWLAGN_MIC_KEY_SIZE); 923 924 rx_mic_key = data->tkip->mic_keys.rx_unicast; 925 } else { 926 tkip_sc = 927 data->rsc_tsc->all_tsc_rsc.tkip.multicast_rsc; 928 rx_p1ks = data->tkip->rx_multi; 929 rx_mic_key = data->tkip->mic_keys.rx_mcast; 930 } 931 932 /* 933 * For non-QoS this relies on the fact that both the uCode and 934 * mac80211 use TID 0 (as they need to to avoid replay attacks) 935 * for checking the IV in the frames. 936 */ 937 for (i = 0; i < IWLAGN_NUM_RSC; i++) { 938 ieee80211_get_key_rx_seq(key, i, &seq); 939 tkip_sc[i].iv16 = cpu_to_le16(seq.tkip.iv16); 940 tkip_sc[i].iv32 = cpu_to_le32(seq.tkip.iv32); 941 /* wrapping isn't allowed, AP must rekey */ 942 if (seq.tkip.iv32 > cur_rx_iv32) 943 cur_rx_iv32 = seq.tkip.iv32; 944 } 945 946 ieee80211_get_tkip_rx_p1k(key, data->bssid, cur_rx_iv32, p1k); 947 iwlagn_convert_p1k(p1k, rx_p1ks[0].p1k); 948 ieee80211_get_tkip_rx_p1k(key, data->bssid, 949 cur_rx_iv32 + 1, p1k); 950 iwlagn_convert_p1k(p1k, rx_p1ks[1].p1k); 951 952 memcpy(rx_mic_key, 953 &key->key[NL80211_TKIP_DATA_OFFSET_RX_MIC_KEY], 954 IWLAGN_MIC_KEY_SIZE); 955 956 data->use_tkip = true; 957 data->use_rsc_tsc = true; 958 break; 959 case WLAN_CIPHER_SUITE_CCMP: 960 if (sta) { 961 u64 pn64; 962 963 aes_sc = data->rsc_tsc->all_tsc_rsc.aes.unicast_rsc; 964 aes_tx_sc = &data->rsc_tsc->all_tsc_rsc.aes.tsc; 965 966 pn64 = atomic64_read(&key->tx_pn); 967 aes_tx_sc->pn = cpu_to_le64(pn64); 968 } else 969 aes_sc = data->rsc_tsc->all_tsc_rsc.aes.multicast_rsc; 970 971 /* 972 * For non-QoS this relies on the fact that both the uCode and 973 * mac80211 use TID 0 for checking the IV in the frames. 974 */ 975 for (i = 0; i < IWLAGN_NUM_RSC; i++) { 976 u8 *pn = seq.ccmp.pn; 977 978 ieee80211_get_key_rx_seq(key, i, &seq); 979 aes_sc[i].pn = cpu_to_le64( 980 (u64)pn[5] | 981 ((u64)pn[4] << 8) | 982 ((u64)pn[3] << 16) | 983 ((u64)pn[2] << 24) | 984 ((u64)pn[1] << 32) | 985 ((u64)pn[0] << 40)); 986 } 987 data->use_rsc_tsc = true; 988 break; 989 } 990 991 mutex_unlock(&priv->mutex); 992 } 993 994 int iwlagn_send_patterns(struct iwl_priv *priv, 995 struct cfg80211_wowlan *wowlan) 996 { 997 struct iwlagn_wowlan_patterns_cmd *pattern_cmd; 998 struct iwl_host_cmd cmd = { 999 .id = REPLY_WOWLAN_PATTERNS, 1000 .dataflags[0] = IWL_HCMD_DFL_NOCOPY, 1001 }; 1002 int i, err; 1003 1004 if (!wowlan->n_patterns) 1005 return 0; 1006 1007 cmd.len[0] = struct_size(pattern_cmd, patterns, wowlan->n_patterns); 1008 1009 pattern_cmd = kmalloc(cmd.len[0], GFP_KERNEL); 1010 if (!pattern_cmd) 1011 return -ENOMEM; 1012 1013 pattern_cmd->n_patterns = cpu_to_le32(wowlan->n_patterns); 1014 1015 for (i = 0; i < wowlan->n_patterns; i++) { 1016 int mask_len = DIV_ROUND_UP(wowlan->patterns[i].pattern_len, 8); 1017 1018 memcpy(&pattern_cmd->patterns[i].mask, 1019 wowlan->patterns[i].mask, mask_len); 1020 memcpy(&pattern_cmd->patterns[i].pattern, 1021 wowlan->patterns[i].pattern, 1022 wowlan->patterns[i].pattern_len); 1023 pattern_cmd->patterns[i].mask_size = mask_len; 1024 pattern_cmd->patterns[i].pattern_size = 1025 wowlan->patterns[i].pattern_len; 1026 } 1027 1028 cmd.data[0] = pattern_cmd; 1029 err = iwl_dvm_send_cmd(priv, &cmd); 1030 kfree(pattern_cmd); 1031 return err; 1032 } 1033 1034 int iwlagn_suspend(struct iwl_priv *priv, struct cfg80211_wowlan *wowlan) 1035 { 1036 struct iwlagn_wowlan_wakeup_filter_cmd wakeup_filter_cmd; 1037 struct iwl_rxon_cmd rxon; 1038 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS]; 1039 struct iwlagn_wowlan_kek_kck_material_cmd kek_kck_cmd; 1040 struct iwlagn_wowlan_tkip_params_cmd tkip_cmd = {}; 1041 struct iwlagn_d3_config_cmd d3_cfg_cmd = { 1042 /* 1043 * Program the minimum sleep time to 10 seconds, as many 1044 * platforms have issues processing a wakeup signal while 1045 * still being in the process of suspending. 1046 */ 1047 .min_sleep_time = cpu_to_le32(10 * 1000 * 1000), 1048 }; 1049 struct wowlan_key_data key_data = { 1050 .ctx = ctx, 1051 .bssid = ctx->active.bssid_addr, 1052 .use_rsc_tsc = false, 1053 .tkip = &tkip_cmd, 1054 .use_tkip = false, 1055 }; 1056 int ret, i; 1057 u16 seq; 1058 1059 key_data.rsc_tsc = kzalloc(sizeof(*key_data.rsc_tsc), GFP_KERNEL); 1060 if (!key_data.rsc_tsc) 1061 return -ENOMEM; 1062 1063 memset(&wakeup_filter_cmd, 0, sizeof(wakeup_filter_cmd)); 1064 1065 /* 1066 * We know the last used seqno, and the uCode expects to know that 1067 * one, it will increment before TX. 1068 */ 1069 seq = le16_to_cpu(priv->last_seq_ctl) & IEEE80211_SCTL_SEQ; 1070 wakeup_filter_cmd.non_qos_seq = cpu_to_le16(seq); 1071 1072 /* 1073 * For QoS counters, we store the one to use next, so subtract 0x10 1074 * since the uCode will add 0x10 before using the value. 1075 */ 1076 for (i = 0; i < IWL_MAX_TID_COUNT; i++) { 1077 seq = priv->tid_data[IWL_AP_ID][i].seq_number; 1078 seq -= 0x10; 1079 wakeup_filter_cmd.qos_seq[i] = cpu_to_le16(seq); 1080 } 1081 1082 if (wowlan->disconnect) 1083 wakeup_filter_cmd.enabled |= 1084 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_BEACON_MISS | 1085 IWLAGN_WOWLAN_WAKEUP_LINK_CHANGE); 1086 if (wowlan->magic_pkt) 1087 wakeup_filter_cmd.enabled |= 1088 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_MAGIC_PACKET); 1089 if (wowlan->gtk_rekey_failure) 1090 wakeup_filter_cmd.enabled |= 1091 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_GTK_REKEY_FAIL); 1092 if (wowlan->eap_identity_req) 1093 wakeup_filter_cmd.enabled |= 1094 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_EAP_IDENT_REQ); 1095 if (wowlan->four_way_handshake) 1096 wakeup_filter_cmd.enabled |= 1097 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_4WAY_HANDSHAKE); 1098 if (wowlan->n_patterns) 1099 wakeup_filter_cmd.enabled |= 1100 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_PATTERN_MATCH); 1101 1102 if (wowlan->rfkill_release) 1103 d3_cfg_cmd.wakeup_flags |= 1104 cpu_to_le32(IWLAGN_D3_WAKEUP_RFKILL); 1105 1106 iwl_scan_cancel_timeout(priv, 200); 1107 1108 memcpy(&rxon, &ctx->active, sizeof(rxon)); 1109 1110 priv->ucode_loaded = false; 1111 iwl_trans_stop_device(priv->trans); 1112 ret = iwl_trans_start_hw(priv->trans); 1113 if (ret) 1114 goto out; 1115 1116 priv->wowlan = true; 1117 1118 ret = iwl_load_ucode_wait_alive(priv, IWL_UCODE_WOWLAN); 1119 if (ret) 1120 goto out; 1121 1122 /* now configure WoWLAN ucode */ 1123 ret = iwl_alive_start(priv); 1124 if (ret) 1125 goto out; 1126 1127 memcpy(&ctx->staging, &rxon, sizeof(rxon)); 1128 ret = iwlagn_commit_rxon(priv, ctx); 1129 if (ret) 1130 goto out; 1131 1132 ret = iwl_power_update_mode(priv, true); 1133 if (ret) 1134 goto out; 1135 1136 if (!iwlwifi_mod_params.swcrypto) { 1137 /* mark all keys clear */ 1138 priv->ucode_key_table = 0; 1139 ctx->key_mapping_keys = 0; 1140 1141 /* 1142 * This needs to be unlocked due to lock ordering 1143 * constraints. Since we're in the suspend path 1144 * that isn't really a problem though. 1145 */ 1146 mutex_unlock(&priv->mutex); 1147 ieee80211_iter_keys(priv->hw, ctx->vif, 1148 iwlagn_wowlan_program_keys, 1149 &key_data); 1150 mutex_lock(&priv->mutex); 1151 if (key_data.error) { 1152 ret = -EIO; 1153 goto out; 1154 } 1155 1156 if (key_data.use_rsc_tsc) { 1157 struct iwl_host_cmd rsc_tsc_cmd = { 1158 .id = REPLY_WOWLAN_TSC_RSC_PARAMS, 1159 .data[0] = key_data.rsc_tsc, 1160 .dataflags[0] = IWL_HCMD_DFL_NOCOPY, 1161 .len[0] = sizeof(*key_data.rsc_tsc), 1162 }; 1163 1164 ret = iwl_dvm_send_cmd(priv, &rsc_tsc_cmd); 1165 if (ret) 1166 goto out; 1167 } 1168 1169 if (key_data.use_tkip) { 1170 ret = iwl_dvm_send_cmd_pdu(priv, 1171 REPLY_WOWLAN_TKIP_PARAMS, 1172 0, sizeof(tkip_cmd), 1173 &tkip_cmd); 1174 if (ret) 1175 goto out; 1176 } 1177 1178 if (priv->have_rekey_data) { 1179 memset(&kek_kck_cmd, 0, sizeof(kek_kck_cmd)); 1180 memcpy(kek_kck_cmd.kck, priv->kck, NL80211_KCK_LEN); 1181 kek_kck_cmd.kck_len = cpu_to_le16(NL80211_KCK_LEN); 1182 memcpy(kek_kck_cmd.kek, priv->kek, NL80211_KEK_LEN); 1183 kek_kck_cmd.kek_len = cpu_to_le16(NL80211_KEK_LEN); 1184 kek_kck_cmd.replay_ctr = priv->replay_ctr; 1185 1186 ret = iwl_dvm_send_cmd_pdu(priv, 1187 REPLY_WOWLAN_KEK_KCK_MATERIAL, 1188 0, sizeof(kek_kck_cmd), 1189 &kek_kck_cmd); 1190 if (ret) 1191 goto out; 1192 } 1193 } 1194 1195 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_D3_CONFIG, 0, 1196 sizeof(d3_cfg_cmd), &d3_cfg_cmd); 1197 if (ret) 1198 goto out; 1199 1200 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_WOWLAN_WAKEUP_FILTER, 1201 0, sizeof(wakeup_filter_cmd), 1202 &wakeup_filter_cmd); 1203 if (ret) 1204 goto out; 1205 1206 ret = iwlagn_send_patterns(priv, wowlan); 1207 out: 1208 kfree(key_data.rsc_tsc); 1209 return ret; 1210 } 1211 #endif 1212 1213 int iwl_dvm_send_cmd(struct iwl_priv *priv, struct iwl_host_cmd *cmd) 1214 { 1215 if (iwl_is_rfkill(priv) || iwl_is_ctkill(priv)) { 1216 IWL_WARN(priv, "Not sending command - %s KILL\n", 1217 iwl_is_rfkill(priv) ? "RF" : "CT"); 1218 return -EIO; 1219 } 1220 1221 if (test_bit(STATUS_FW_ERROR, &priv->status)) { 1222 IWL_ERR(priv, "Command %s failed: FW Error\n", 1223 iwl_get_cmd_string(priv->trans, cmd->id)); 1224 return -EIO; 1225 } 1226 1227 /* 1228 * This can happen upon FW ASSERT: we clear the STATUS_FW_ERROR flag 1229 * in iwl_down but cancel the workers only later. 1230 */ 1231 if (!priv->ucode_loaded) { 1232 IWL_ERR(priv, "Fw not loaded - dropping CMD: %x\n", cmd->id); 1233 return -EIO; 1234 } 1235 1236 /* 1237 * Synchronous commands from this op-mode must hold 1238 * the mutex, this ensures we don't try to send two 1239 * (or more) synchronous commands at a time. 1240 */ 1241 if (!(cmd->flags & CMD_ASYNC)) 1242 lockdep_assert_held(&priv->mutex); 1243 1244 return iwl_trans_send_cmd(priv->trans, cmd); 1245 } 1246 1247 int iwl_dvm_send_cmd_pdu(struct iwl_priv *priv, u8 id, 1248 u32 flags, u16 len, const void *data) 1249 { 1250 struct iwl_host_cmd cmd = { 1251 .id = id, 1252 .len = { len, }, 1253 .data = { data, }, 1254 .flags = flags, 1255 }; 1256 1257 return iwl_dvm_send_cmd(priv, &cmd); 1258 } 1259