1 /****************************************************************************** 2 * 3 * This file is provided under a dual BSD/GPLv2 license. When using or 4 * redistributing this file, you may do so under either license. 5 * 6 * GPL LICENSE SUMMARY 7 * 8 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved. 9 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH 10 * Copyright(c) 2016 Intel Deutschland GmbH 11 * 12 * This program is free software; you can redistribute it and/or modify 13 * it under the terms of version 2 of the GNU General Public License as 14 * published by the Free Software Foundation. 15 * 16 * This program is distributed in the hope that it will be useful, but 17 * WITHOUT ANY WARRANTY; without even the implied warranty of 18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 19 * General Public License for more details. 20 * 21 * You should have received a copy of the GNU General Public License 22 * along with this program; if not, write to the Free Software 23 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110, 24 * USA 25 * 26 * The full GNU General Public License is included in this distribution 27 * in the file called COPYING. 28 * 29 * Contact Information: 30 * Intel Linux Wireless <linuxwifi@intel.com> 31 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497 32 * 33 * BSD LICENSE 34 * 35 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved. 36 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH 37 * All rights reserved. 38 * 39 * Redistribution and use in source and binary forms, with or without 40 * modification, are permitted provided that the following conditions 41 * are met: 42 * 43 * * Redistributions of source code must retain the above copyright 44 * notice, this list of conditions and the following disclaimer. 45 * * Redistributions in binary form must reproduce the above copyright 46 * notice, this list of conditions and the following disclaimer in 47 * the documentation and/or other materials provided with the 48 * distribution. 49 * * Neither the name Intel Corporation nor the names of its 50 * contributors may be used to endorse or promote products derived 51 * from this software without specific prior written permission. 52 * 53 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 54 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 55 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR 56 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT 57 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, 58 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT 59 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 60 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 61 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 62 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE 63 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 64 * 65 *****************************************************************************/ 66 #include <linux/ieee80211.h> 67 #include <linux/etherdevice.h> 68 #include <linux/tcp.h> 69 #include <net/ip.h> 70 #include <net/ipv6.h> 71 72 #include "iwl-trans.h" 73 #include "iwl-eeprom-parse.h" 74 #include "mvm.h" 75 #include "sta.h" 76 #include "fw-dbg.h" 77 78 static void 79 iwl_mvm_bar_check_trigger(struct iwl_mvm *mvm, const u8 *addr, 80 u16 tid, u16 ssn) 81 { 82 struct iwl_fw_dbg_trigger_tlv *trig; 83 struct iwl_fw_dbg_trigger_ba *ba_trig; 84 85 if (!iwl_fw_dbg_trigger_enabled(mvm->fw, FW_DBG_TRIGGER_BA)) 86 return; 87 88 trig = iwl_fw_dbg_get_trigger(mvm->fw, FW_DBG_TRIGGER_BA); 89 ba_trig = (void *)trig->data; 90 91 if (!iwl_fw_dbg_trigger_check_stop(mvm, NULL, trig)) 92 return; 93 94 if (!(le16_to_cpu(ba_trig->tx_bar) & BIT(tid))) 95 return; 96 97 iwl_mvm_fw_dbg_collect_trig(mvm, trig, 98 "BAR sent to %pM, tid %d, ssn %d", 99 addr, tid, ssn); 100 } 101 102 #define OPT_HDR(type, skb, off) \ 103 (type *)(skb_network_header(skb) + (off)) 104 105 static void iwl_mvm_tx_csum(struct iwl_mvm *mvm, struct sk_buff *skb, 106 struct ieee80211_hdr *hdr, 107 struct ieee80211_tx_info *info, 108 struct iwl_tx_cmd *tx_cmd) 109 { 110 #if IS_ENABLED(CONFIG_INET) 111 u16 mh_len = ieee80211_hdrlen(hdr->frame_control); 112 u16 offload_assist = le16_to_cpu(tx_cmd->offload_assist); 113 u8 protocol = 0; 114 115 /* 116 * Do not compute checksum if already computed or if transport will 117 * compute it 118 */ 119 if (skb->ip_summed != CHECKSUM_PARTIAL || IWL_MVM_SW_TX_CSUM_OFFLOAD) 120 return; 121 122 /* We do not expect to be requested to csum stuff we do not support */ 123 if (WARN_ONCE(!(mvm->hw->netdev_features & IWL_TX_CSUM_NETIF_FLAGS) || 124 (skb->protocol != htons(ETH_P_IP) && 125 skb->protocol != htons(ETH_P_IPV6)), 126 "No support for requested checksum\n")) { 127 skb_checksum_help(skb); 128 return; 129 } 130 131 if (skb->protocol == htons(ETH_P_IP)) { 132 protocol = ip_hdr(skb)->protocol; 133 } else { 134 #if IS_ENABLED(CONFIG_IPV6) 135 struct ipv6hdr *ipv6h = 136 (struct ipv6hdr *)skb_network_header(skb); 137 unsigned int off = sizeof(*ipv6h); 138 139 protocol = ipv6h->nexthdr; 140 while (protocol != NEXTHDR_NONE && ipv6_ext_hdr(protocol)) { 141 /* only supported extension headers */ 142 if (protocol != NEXTHDR_ROUTING && 143 protocol != NEXTHDR_HOP && 144 protocol != NEXTHDR_DEST && 145 protocol != NEXTHDR_FRAGMENT) { 146 skb_checksum_help(skb); 147 return; 148 } 149 150 if (protocol == NEXTHDR_FRAGMENT) { 151 struct frag_hdr *hp = 152 OPT_HDR(struct frag_hdr, skb, off); 153 154 protocol = hp->nexthdr; 155 off += sizeof(struct frag_hdr); 156 } else { 157 struct ipv6_opt_hdr *hp = 158 OPT_HDR(struct ipv6_opt_hdr, skb, off); 159 160 protocol = hp->nexthdr; 161 off += ipv6_optlen(hp); 162 } 163 } 164 /* if we get here - protocol now should be TCP/UDP */ 165 #endif 166 } 167 168 if (protocol != IPPROTO_TCP && protocol != IPPROTO_UDP) { 169 WARN_ON_ONCE(1); 170 skb_checksum_help(skb); 171 return; 172 } 173 174 /* enable L4 csum */ 175 offload_assist |= BIT(TX_CMD_OFFLD_L4_EN); 176 177 /* 178 * Set offset to IP header (snap). 179 * We don't support tunneling so no need to take care of inner header. 180 * Size is in words. 181 */ 182 offload_assist |= (4 << TX_CMD_OFFLD_IP_HDR); 183 184 /* Do IPv4 csum for AMSDU only (no IP csum for Ipv6) */ 185 if (skb->protocol == htons(ETH_P_IP) && 186 (offload_assist & BIT(TX_CMD_OFFLD_AMSDU))) { 187 ip_hdr(skb)->check = 0; 188 offload_assist |= BIT(TX_CMD_OFFLD_L3_EN); 189 } 190 191 /* reset UDP/TCP header csum */ 192 if (protocol == IPPROTO_TCP) 193 tcp_hdr(skb)->check = 0; 194 else 195 udp_hdr(skb)->check = 0; 196 197 /* mac header len should include IV, size is in words */ 198 if (info->control.hw_key) 199 mh_len += info->control.hw_key->iv_len; 200 mh_len /= 2; 201 offload_assist |= mh_len << TX_CMD_OFFLD_MH_SIZE; 202 203 tx_cmd->offload_assist = cpu_to_le16(offload_assist); 204 #endif 205 } 206 207 /* 208 * Sets most of the Tx cmd's fields 209 */ 210 void iwl_mvm_set_tx_cmd(struct iwl_mvm *mvm, struct sk_buff *skb, 211 struct iwl_tx_cmd *tx_cmd, 212 struct ieee80211_tx_info *info, u8 sta_id) 213 { 214 struct ieee80211_tx_info *skb_info = IEEE80211_SKB_CB(skb); 215 struct ieee80211_hdr *hdr = (void *)skb->data; 216 __le16 fc = hdr->frame_control; 217 u32 tx_flags = le32_to_cpu(tx_cmd->tx_flags); 218 u32 len = skb->len + FCS_LEN; 219 u8 ac; 220 221 if (!(info->flags & IEEE80211_TX_CTL_NO_ACK)) 222 tx_flags |= TX_CMD_FLG_ACK; 223 else 224 tx_flags &= ~TX_CMD_FLG_ACK; 225 226 if (ieee80211_is_probe_resp(fc)) 227 tx_flags |= TX_CMD_FLG_TSF; 228 229 if (ieee80211_has_morefrags(fc)) 230 tx_flags |= TX_CMD_FLG_MORE_FRAG; 231 232 if (ieee80211_is_data_qos(fc)) { 233 u8 *qc = ieee80211_get_qos_ctl(hdr); 234 tx_cmd->tid_tspec = qc[0] & 0xf; 235 tx_flags &= ~TX_CMD_FLG_SEQ_CTL; 236 if (*qc & IEEE80211_QOS_CTL_A_MSDU_PRESENT) 237 tx_cmd->offload_assist |= 238 cpu_to_le16(BIT(TX_CMD_OFFLD_AMSDU)); 239 } else if (ieee80211_is_back_req(fc)) { 240 struct ieee80211_bar *bar = (void *)skb->data; 241 u16 control = le16_to_cpu(bar->control); 242 u16 ssn = le16_to_cpu(bar->start_seq_num); 243 244 tx_flags |= TX_CMD_FLG_ACK | TX_CMD_FLG_BAR; 245 tx_cmd->tid_tspec = (control & 246 IEEE80211_BAR_CTRL_TID_INFO_MASK) >> 247 IEEE80211_BAR_CTRL_TID_INFO_SHIFT; 248 WARN_ON_ONCE(tx_cmd->tid_tspec >= IWL_MAX_TID_COUNT); 249 iwl_mvm_bar_check_trigger(mvm, bar->ra, tx_cmd->tid_tspec, 250 ssn); 251 } else { 252 tx_cmd->tid_tspec = IWL_TID_NON_QOS; 253 if (info->flags & IEEE80211_TX_CTL_ASSIGN_SEQ) 254 tx_flags |= TX_CMD_FLG_SEQ_CTL; 255 else 256 tx_flags &= ~TX_CMD_FLG_SEQ_CTL; 257 } 258 259 /* Default to 0 (BE) when tid_spec is set to IWL_TID_NON_QOS */ 260 if (tx_cmd->tid_tspec < IWL_MAX_TID_COUNT) 261 ac = tid_to_mac80211_ac[tx_cmd->tid_tspec]; 262 else 263 ac = tid_to_mac80211_ac[0]; 264 265 tx_flags |= iwl_mvm_bt_coex_tx_prio(mvm, hdr, info, ac) << 266 TX_CMD_FLG_BT_PRIO_POS; 267 268 if (ieee80211_is_mgmt(fc)) { 269 if (ieee80211_is_assoc_req(fc) || ieee80211_is_reassoc_req(fc)) 270 tx_cmd->pm_frame_timeout = cpu_to_le16(PM_FRAME_ASSOC); 271 else if (ieee80211_is_action(fc)) 272 tx_cmd->pm_frame_timeout = cpu_to_le16(PM_FRAME_NONE); 273 else 274 tx_cmd->pm_frame_timeout = cpu_to_le16(PM_FRAME_MGMT); 275 276 /* The spec allows Action frames in A-MPDU, we don't support 277 * it 278 */ 279 WARN_ON_ONCE(info->flags & IEEE80211_TX_CTL_AMPDU); 280 } else if (info->control.flags & IEEE80211_TX_CTRL_PORT_CTRL_PROTO) { 281 tx_cmd->pm_frame_timeout = cpu_to_le16(PM_FRAME_MGMT); 282 } else { 283 tx_cmd->pm_frame_timeout = cpu_to_le16(PM_FRAME_NONE); 284 } 285 286 if (ieee80211_is_data(fc) && len > mvm->rts_threshold && 287 !is_multicast_ether_addr(ieee80211_get_DA(hdr))) 288 tx_flags |= TX_CMD_FLG_PROT_REQUIRE; 289 290 if (fw_has_capa(&mvm->fw->ucode_capa, 291 IWL_UCODE_TLV_CAPA_TXPOWER_INSERTION_SUPPORT) && 292 ieee80211_action_contains_tpc(skb)) 293 tx_flags |= TX_CMD_FLG_WRITE_TX_POWER; 294 295 tx_cmd->tx_flags = cpu_to_le32(tx_flags); 296 /* Total # bytes to be transmitted */ 297 tx_cmd->len = cpu_to_le16((u16)skb->len + 298 (uintptr_t)skb_info->driver_data[0]); 299 tx_cmd->life_time = cpu_to_le32(TX_CMD_LIFE_TIME_INFINITE); 300 tx_cmd->sta_id = sta_id; 301 302 /* padding is inserted later in transport */ 303 if (ieee80211_hdrlen(fc) % 4 && 304 !(tx_cmd->offload_assist & cpu_to_le16(BIT(TX_CMD_OFFLD_AMSDU)))) 305 tx_cmd->offload_assist |= cpu_to_le16(BIT(TX_CMD_OFFLD_PAD)); 306 307 iwl_mvm_tx_csum(mvm, skb, hdr, info, tx_cmd); 308 } 309 310 /* 311 * Sets the fields in the Tx cmd that are rate related 312 */ 313 void iwl_mvm_set_tx_cmd_rate(struct iwl_mvm *mvm, struct iwl_tx_cmd *tx_cmd, 314 struct ieee80211_tx_info *info, 315 struct ieee80211_sta *sta, __le16 fc) 316 { 317 u32 rate_flags; 318 int rate_idx; 319 u8 rate_plcp; 320 321 /* Set retry limit on RTS packets */ 322 tx_cmd->rts_retry_limit = IWL_RTS_DFAULT_RETRY_LIMIT; 323 324 /* Set retry limit on DATA packets and Probe Responses*/ 325 if (ieee80211_is_probe_resp(fc)) { 326 tx_cmd->data_retry_limit = IWL_MGMT_DFAULT_RETRY_LIMIT; 327 tx_cmd->rts_retry_limit = 328 min(tx_cmd->data_retry_limit, tx_cmd->rts_retry_limit); 329 } else if (ieee80211_is_back_req(fc)) { 330 tx_cmd->data_retry_limit = IWL_BAR_DFAULT_RETRY_LIMIT; 331 } else { 332 tx_cmd->data_retry_limit = IWL_DEFAULT_TX_RETRY; 333 } 334 335 /* 336 * for data packets, rate info comes from the table inside the fw. This 337 * table is controlled by LINK_QUALITY commands 338 */ 339 340 if (ieee80211_is_data(fc) && sta) { 341 tx_cmd->initial_rate_index = 0; 342 tx_cmd->tx_flags |= cpu_to_le32(TX_CMD_FLG_STA_RATE); 343 return; 344 } else if (ieee80211_is_back_req(fc)) { 345 tx_cmd->tx_flags |= 346 cpu_to_le32(TX_CMD_FLG_ACK | TX_CMD_FLG_BAR); 347 } 348 349 /* HT rate doesn't make sense for a non data frame */ 350 WARN_ONCE(info->control.rates[0].flags & IEEE80211_TX_RC_MCS, 351 "Got an HT rate (flags:0x%x/mcs:%d) for a non data frame (fc:0x%x)\n", 352 info->control.rates[0].flags, 353 info->control.rates[0].idx, 354 le16_to_cpu(fc)); 355 356 rate_idx = info->control.rates[0].idx; 357 /* if the rate isn't a well known legacy rate, take the lowest one */ 358 if (rate_idx < 0 || rate_idx > IWL_RATE_COUNT_LEGACY) 359 rate_idx = rate_lowest_index( 360 &mvm->nvm_data->bands[info->band], sta); 361 362 /* For 5 GHZ band, remap mac80211 rate indices into driver indices */ 363 if (info->band == NL80211_BAND_5GHZ) 364 rate_idx += IWL_FIRST_OFDM_RATE; 365 366 /* For 2.4 GHZ band, check that there is no need to remap */ 367 BUILD_BUG_ON(IWL_FIRST_CCK_RATE != 0); 368 369 /* Get PLCP rate for tx_cmd->rate_n_flags */ 370 rate_plcp = iwl_mvm_mac80211_idx_to_hwrate(rate_idx); 371 372 mvm->mgmt_last_antenna_idx = 373 iwl_mvm_next_antenna(mvm, iwl_mvm_get_valid_tx_ant(mvm), 374 mvm->mgmt_last_antenna_idx); 375 376 if (info->band == NL80211_BAND_2GHZ && 377 !iwl_mvm_bt_coex_is_shared_ant_avail(mvm)) 378 rate_flags = mvm->cfg->non_shared_ant << RATE_MCS_ANT_POS; 379 else 380 rate_flags = 381 BIT(mvm->mgmt_last_antenna_idx) << RATE_MCS_ANT_POS; 382 383 /* Set CCK flag as needed */ 384 if ((rate_idx >= IWL_FIRST_CCK_RATE) && (rate_idx <= IWL_LAST_CCK_RATE)) 385 rate_flags |= RATE_MCS_CCK_MSK; 386 387 /* Set the rate in the TX cmd */ 388 tx_cmd->rate_n_flags = cpu_to_le32((u32)rate_plcp | rate_flags); 389 } 390 391 /* 392 * Sets the fields in the Tx cmd that are crypto related 393 */ 394 static void iwl_mvm_set_tx_cmd_crypto(struct iwl_mvm *mvm, 395 struct ieee80211_tx_info *info, 396 struct iwl_tx_cmd *tx_cmd, 397 struct sk_buff *skb_frag, 398 int hdrlen) 399 { 400 struct ieee80211_key_conf *keyconf = info->control.hw_key; 401 u8 *crypto_hdr = skb_frag->data + hdrlen; 402 u64 pn; 403 404 switch (keyconf->cipher) { 405 case WLAN_CIPHER_SUITE_CCMP: 406 case WLAN_CIPHER_SUITE_CCMP_256: 407 iwl_mvm_set_tx_cmd_ccmp(info, tx_cmd); 408 pn = atomic64_inc_return(&keyconf->tx_pn); 409 crypto_hdr[0] = pn; 410 crypto_hdr[2] = 0; 411 crypto_hdr[3] = 0x20 | (keyconf->keyidx << 6); 412 crypto_hdr[1] = pn >> 8; 413 crypto_hdr[4] = pn >> 16; 414 crypto_hdr[5] = pn >> 24; 415 crypto_hdr[6] = pn >> 32; 416 crypto_hdr[7] = pn >> 40; 417 break; 418 419 case WLAN_CIPHER_SUITE_TKIP: 420 tx_cmd->sec_ctl = TX_CMD_SEC_TKIP; 421 pn = atomic64_inc_return(&keyconf->tx_pn); 422 ieee80211_tkip_add_iv(crypto_hdr, keyconf, pn); 423 ieee80211_get_tkip_p2k(keyconf, skb_frag, tx_cmd->key); 424 break; 425 426 case WLAN_CIPHER_SUITE_WEP104: 427 tx_cmd->sec_ctl |= TX_CMD_SEC_KEY128; 428 /* fall through */ 429 case WLAN_CIPHER_SUITE_WEP40: 430 tx_cmd->sec_ctl |= TX_CMD_SEC_WEP | 431 ((keyconf->keyidx << TX_CMD_SEC_WEP_KEY_IDX_POS) & 432 TX_CMD_SEC_WEP_KEY_IDX_MSK); 433 434 memcpy(&tx_cmd->key[3], keyconf->key, keyconf->keylen); 435 break; 436 default: 437 tx_cmd->sec_ctl |= TX_CMD_SEC_EXT; 438 } 439 } 440 441 /* 442 * Allocates and sets the Tx cmd the driver data pointers in the skb 443 */ 444 static struct iwl_device_cmd * 445 iwl_mvm_set_tx_params(struct iwl_mvm *mvm, struct sk_buff *skb, 446 struct ieee80211_tx_info *info, int hdrlen, 447 struct ieee80211_sta *sta, u8 sta_id) 448 { 449 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)skb->data; 450 struct ieee80211_tx_info *skb_info = IEEE80211_SKB_CB(skb); 451 struct iwl_device_cmd *dev_cmd; 452 struct iwl_tx_cmd *tx_cmd; 453 454 dev_cmd = iwl_trans_alloc_tx_cmd(mvm->trans); 455 456 if (unlikely(!dev_cmd)) 457 return NULL; 458 459 memset(dev_cmd, 0, sizeof(*dev_cmd)); 460 dev_cmd->hdr.cmd = TX_CMD; 461 tx_cmd = (struct iwl_tx_cmd *)dev_cmd->payload; 462 463 if (info->control.hw_key) 464 iwl_mvm_set_tx_cmd_crypto(mvm, info, tx_cmd, skb, hdrlen); 465 466 iwl_mvm_set_tx_cmd(mvm, skb, tx_cmd, info, sta_id); 467 468 iwl_mvm_set_tx_cmd_rate(mvm, tx_cmd, info, sta, hdr->frame_control); 469 470 memset(&skb_info->status, 0, sizeof(skb_info->status)); 471 memset(skb_info->driver_data, 0, sizeof(skb_info->driver_data)); 472 473 skb_info->driver_data[1] = dev_cmd; 474 475 return dev_cmd; 476 } 477 478 static int iwl_mvm_get_ctrl_vif_queue(struct iwl_mvm *mvm, 479 struct ieee80211_tx_info *info, __le16 fc) 480 { 481 if (iwl_mvm_is_dqa_supported(mvm)) { 482 if (info->control.vif->type == NL80211_IFTYPE_AP && 483 ieee80211_is_probe_resp(fc)) 484 return IWL_MVM_DQA_AP_PROBE_RESP_QUEUE; 485 else if (ieee80211_is_mgmt(fc) && 486 info->control.vif->type == NL80211_IFTYPE_P2P_DEVICE) 487 return IWL_MVM_DQA_P2P_DEVICE_QUEUE; 488 } 489 490 return info->hw_queue; 491 } 492 493 int iwl_mvm_tx_skb_non_sta(struct iwl_mvm *mvm, struct sk_buff *skb) 494 { 495 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)skb->data; 496 struct ieee80211_tx_info *skb_info = IEEE80211_SKB_CB(skb); 497 struct ieee80211_tx_info info; 498 struct iwl_device_cmd *dev_cmd; 499 struct iwl_tx_cmd *tx_cmd; 500 u8 sta_id; 501 int hdrlen = ieee80211_hdrlen(hdr->frame_control); 502 int queue; 503 504 memcpy(&info, skb->cb, sizeof(info)); 505 506 if (WARN_ON_ONCE(info.flags & IEEE80211_TX_CTL_AMPDU)) 507 return -1; 508 509 if (WARN_ON_ONCE(info.flags & IEEE80211_TX_CTL_SEND_AFTER_DTIM && 510 (!info.control.vif || 511 info.hw_queue != info.control.vif->cab_queue))) 512 return -1; 513 514 /* This holds the amsdu headers length */ 515 skb_info->driver_data[0] = (void *)(uintptr_t)0; 516 517 /* 518 * IWL_MVM_OFFCHANNEL_QUEUE is used for ROC packets that can be used 519 * in 2 different types of vifs, P2P & STATION. P2P uses the offchannel 520 * queue. STATION (HS2.0) uses the auxiliary context of the FW, 521 * and hence needs to be sent on the aux queue 522 */ 523 if (IEEE80211_SKB_CB(skb)->hw_queue == IWL_MVM_OFFCHANNEL_QUEUE && 524 info.control.vif->type == NL80211_IFTYPE_STATION) 525 IEEE80211_SKB_CB(skb)->hw_queue = mvm->aux_queue; 526 527 queue = info.hw_queue; 528 529 /* 530 * If the interface on which the frame is sent is the P2P_DEVICE 531 * or an AP/GO interface use the broadcast station associated 532 * with it; otherwise if the interface is a managed interface 533 * use the AP station associated with it for multicast traffic 534 * (this is not possible for unicast packets as a TLDS discovery 535 * response are sent without a station entry); otherwise use the 536 * AUX station. 537 */ 538 sta_id = mvm->aux_sta.sta_id; 539 if (info.control.vif) { 540 struct iwl_mvm_vif *mvmvif = 541 iwl_mvm_vif_from_mac80211(info.control.vif); 542 543 if (info.control.vif->type == NL80211_IFTYPE_P2P_DEVICE || 544 info.control.vif->type == NL80211_IFTYPE_AP) { 545 sta_id = mvmvif->bcast_sta.sta_id; 546 queue = iwl_mvm_get_ctrl_vif_queue(mvm, &info, 547 hdr->frame_control); 548 } else if (info.control.vif->type == NL80211_IFTYPE_STATION && 549 is_multicast_ether_addr(hdr->addr1)) { 550 u8 ap_sta_id = ACCESS_ONCE(mvmvif->ap_sta_id); 551 552 if (ap_sta_id != IWL_MVM_STATION_COUNT) 553 sta_id = ap_sta_id; 554 } 555 } 556 557 IWL_DEBUG_TX(mvm, "station Id %d, queue=%d\n", sta_id, queue); 558 559 dev_cmd = iwl_mvm_set_tx_params(mvm, skb, &info, hdrlen, NULL, sta_id); 560 if (!dev_cmd) 561 return -1; 562 563 tx_cmd = (struct iwl_tx_cmd *)dev_cmd->payload; 564 565 /* Copy MAC header from skb into command buffer */ 566 memcpy(tx_cmd->hdr, hdr, hdrlen); 567 568 if (iwl_trans_tx(mvm->trans, skb, dev_cmd, queue)) { 569 iwl_trans_free_tx_cmd(mvm->trans, dev_cmd); 570 return -1; 571 } 572 573 /* 574 * Increase the pending frames counter, so that later when a reply comes 575 * in and the counter is decreased - we don't start getting negative 576 * values. 577 * Note that we don't need to make sure it isn't agg'd, since we're 578 * TXing non-sta 579 */ 580 atomic_inc(&mvm->pending_frames[sta_id]); 581 582 return 0; 583 } 584 585 #ifdef CONFIG_INET 586 static int iwl_mvm_tx_tso(struct iwl_mvm *mvm, struct sk_buff *skb, 587 struct ieee80211_tx_info *info, 588 struct ieee80211_sta *sta, 589 struct sk_buff_head *mpdus_skb) 590 { 591 struct iwl_mvm_sta *mvmsta = iwl_mvm_sta_from_mac80211(sta); 592 struct ieee80211_hdr *hdr = (void *)skb->data; 593 unsigned int mss = skb_shinfo(skb)->gso_size; 594 struct sk_buff *tmp, *next; 595 char cb[sizeof(skb->cb)]; 596 unsigned int num_subframes, tcp_payload_len, subf_len, max_amsdu_len; 597 bool ipv4 = (skb->protocol == htons(ETH_P_IP)); 598 u16 ip_base_id = ipv4 ? ntohs(ip_hdr(skb)->id) : 0; 599 u16 amsdu_add, snap_ip_tcp, pad, i = 0; 600 unsigned int dbg_max_amsdu_len; 601 netdev_features_t netdev_features = NETIF_F_CSUM_MASK | NETIF_F_SG; 602 u8 *qc, tid, txf; 603 604 snap_ip_tcp = 8 + skb_transport_header(skb) - skb_network_header(skb) + 605 tcp_hdrlen(skb); 606 607 qc = ieee80211_get_qos_ctl(hdr); 608 tid = *qc & IEEE80211_QOS_CTL_TID_MASK; 609 if (WARN_ON_ONCE(tid >= IWL_MAX_TID_COUNT)) 610 return -EINVAL; 611 612 dbg_max_amsdu_len = ACCESS_ONCE(mvm->max_amsdu_len); 613 614 if (!sta->max_amsdu_len || 615 !ieee80211_is_data_qos(hdr->frame_control) || 616 (!mvmsta->tlc_amsdu && !dbg_max_amsdu_len)) { 617 num_subframes = 1; 618 pad = 0; 619 goto segment; 620 } 621 622 /* 623 * Do not build AMSDU for IPv6 with extension headers. 624 * ask stack to segment and checkum the generated MPDUs for us. 625 */ 626 if (skb->protocol == htons(ETH_P_IPV6) && 627 ((struct ipv6hdr *)skb_network_header(skb))->nexthdr != 628 IPPROTO_TCP) { 629 num_subframes = 1; 630 pad = 0; 631 netdev_features &= ~NETIF_F_CSUM_MASK; 632 goto segment; 633 } 634 635 /* 636 * No need to lock amsdu_in_ampdu_allowed since it can't be modified 637 * during an BA session. 638 */ 639 if (info->flags & IEEE80211_TX_CTL_AMPDU && 640 !mvmsta->tid_data[tid].amsdu_in_ampdu_allowed) { 641 num_subframes = 1; 642 pad = 0; 643 goto segment; 644 } 645 646 max_amsdu_len = sta->max_amsdu_len; 647 648 /* the Tx FIFO to which this A-MSDU will be routed */ 649 txf = iwl_mvm_ac_to_tx_fifo[tid_to_mac80211_ac[tid]]; 650 651 /* 652 * Don't send an AMSDU that will be longer than the TXF. 653 * Add a security margin of 256 for the TX command + headers. 654 * We also want to have the start of the next packet inside the 655 * fifo to be able to send bursts. 656 */ 657 max_amsdu_len = min_t(unsigned int, max_amsdu_len, 658 mvm->shared_mem_cfg.txfifo_size[txf] - 256); 659 660 if (unlikely(dbg_max_amsdu_len)) 661 max_amsdu_len = min_t(unsigned int, max_amsdu_len, 662 dbg_max_amsdu_len); 663 664 /* 665 * Limit A-MSDU in A-MPDU to 4095 bytes when VHT is not 666 * supported. This is a spec requirement (IEEE 802.11-2015 667 * section 8.7.3 NOTE 3). 668 */ 669 if (info->flags & IEEE80211_TX_CTL_AMPDU && 670 !sta->vht_cap.vht_supported) 671 max_amsdu_len = min_t(unsigned int, max_amsdu_len, 4095); 672 673 /* Sub frame header + SNAP + IP header + TCP header + MSS */ 674 subf_len = sizeof(struct ethhdr) + snap_ip_tcp + mss; 675 pad = (4 - subf_len) & 0x3; 676 677 /* 678 * If we have N subframes in the A-MSDU, then the A-MSDU's size is 679 * N * subf_len + (N - 1) * pad. 680 */ 681 num_subframes = (max_amsdu_len + pad) / (subf_len + pad); 682 if (num_subframes > 1) 683 *qc |= IEEE80211_QOS_CTL_A_MSDU_PRESENT; 684 685 tcp_payload_len = skb_tail_pointer(skb) - skb_transport_header(skb) - 686 tcp_hdrlen(skb) + skb->data_len; 687 688 /* 689 * Make sure we have enough TBs for the A-MSDU: 690 * 2 for each subframe 691 * 1 more for each fragment 692 * 1 more for the potential data in the header 693 */ 694 num_subframes = 695 min_t(unsigned int, num_subframes, 696 (mvm->trans->max_skb_frags - 1 - 697 skb_shinfo(skb)->nr_frags) / 2); 698 699 /* This skb fits in one single A-MSDU */ 700 if (num_subframes * mss >= tcp_payload_len) { 701 struct ieee80211_tx_info *skb_info = IEEE80211_SKB_CB(skb); 702 703 /* 704 * Compute the length of all the data added for the A-MSDU. 705 * This will be used to compute the length to write in the TX 706 * command. We have: SNAP + IP + TCP for n -1 subframes and 707 * ETH header for n subframes. Note that the original skb 708 * already had one set of SNAP / IP / TCP headers. 709 */ 710 num_subframes = DIV_ROUND_UP(tcp_payload_len, mss); 711 amsdu_add = num_subframes * sizeof(struct ethhdr) + 712 (num_subframes - 1) * (snap_ip_tcp + pad); 713 /* This holds the amsdu headers length */ 714 skb_info->driver_data[0] = (void *)(uintptr_t)amsdu_add; 715 716 __skb_queue_tail(mpdus_skb, skb); 717 return 0; 718 } 719 720 /* 721 * Trick the segmentation function to make it 722 * create SKBs that can fit into one A-MSDU. 723 */ 724 segment: 725 skb_shinfo(skb)->gso_size = num_subframes * mss; 726 memcpy(cb, skb->cb, sizeof(cb)); 727 728 next = skb_gso_segment(skb, netdev_features); 729 skb_shinfo(skb)->gso_size = mss; 730 if (WARN_ON_ONCE(IS_ERR(next))) 731 return -EINVAL; 732 else if (next) 733 consume_skb(skb); 734 735 while (next) { 736 tmp = next; 737 next = tmp->next; 738 739 memcpy(tmp->cb, cb, sizeof(tmp->cb)); 740 /* 741 * Compute the length of all the data added for the A-MSDU. 742 * This will be used to compute the length to write in the TX 743 * command. We have: SNAP + IP + TCP for n -1 subframes and 744 * ETH header for n subframes. 745 */ 746 tcp_payload_len = skb_tail_pointer(tmp) - 747 skb_transport_header(tmp) - 748 tcp_hdrlen(tmp) + tmp->data_len; 749 750 if (ipv4) 751 ip_hdr(tmp)->id = htons(ip_base_id + i * num_subframes); 752 753 if (tcp_payload_len > mss) { 754 struct ieee80211_tx_info *skb_info = 755 IEEE80211_SKB_CB(tmp); 756 757 num_subframes = DIV_ROUND_UP(tcp_payload_len, mss); 758 amsdu_add = num_subframes * sizeof(struct ethhdr) + 759 (num_subframes - 1) * (snap_ip_tcp + pad); 760 skb_info->driver_data[0] = 761 (void *)(uintptr_t)amsdu_add; 762 skb_shinfo(tmp)->gso_size = mss; 763 } else { 764 qc = ieee80211_get_qos_ctl((void *)tmp->data); 765 766 if (ipv4) 767 ip_send_check(ip_hdr(tmp)); 768 *qc &= ~IEEE80211_QOS_CTL_A_MSDU_PRESENT; 769 skb_shinfo(tmp)->gso_size = 0; 770 } 771 772 tmp->prev = NULL; 773 tmp->next = NULL; 774 775 __skb_queue_tail(mpdus_skb, tmp); 776 i++; 777 } 778 779 return 0; 780 } 781 #else /* CONFIG_INET */ 782 static int iwl_mvm_tx_tso(struct iwl_mvm *mvm, struct sk_buff *skb, 783 struct ieee80211_tx_info *info, 784 struct ieee80211_sta *sta, 785 struct sk_buff_head *mpdus_skb) 786 { 787 /* Impossible to get TSO with CONFIG_INET */ 788 WARN_ON(1); 789 790 return -1; 791 } 792 #endif 793 794 static void iwl_mvm_tx_add_stream(struct iwl_mvm *mvm, 795 struct iwl_mvm_sta *mvm_sta, u8 tid, 796 struct sk_buff *skb) 797 { 798 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 799 u8 mac_queue = info->hw_queue; 800 struct sk_buff_head *deferred_tx_frames; 801 802 lockdep_assert_held(&mvm_sta->lock); 803 804 mvm_sta->deferred_traffic_tid_map |= BIT(tid); 805 set_bit(mvm_sta->sta_id, mvm->sta_deferred_frames); 806 807 deferred_tx_frames = &mvm_sta->tid_data[tid].deferred_tx_frames; 808 809 skb_queue_tail(deferred_tx_frames, skb); 810 811 /* 812 * The first deferred frame should've stopped the MAC queues, so we 813 * should never get a second deferred frame for the RA/TID. 814 */ 815 if (!WARN(skb_queue_len(deferred_tx_frames) != 1, 816 "RATID %d/%d has %d deferred frames\n", mvm_sta->sta_id, tid, 817 skb_queue_len(deferred_tx_frames))) { 818 iwl_mvm_stop_mac_queues(mvm, BIT(mac_queue)); 819 schedule_work(&mvm->add_stream_wk); 820 } 821 } 822 823 /* 824 * Sets the fields in the Tx cmd that are crypto related 825 */ 826 static int iwl_mvm_tx_mpdu(struct iwl_mvm *mvm, struct sk_buff *skb, 827 struct ieee80211_tx_info *info, 828 struct ieee80211_sta *sta) 829 { 830 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)skb->data; 831 struct iwl_mvm_sta *mvmsta; 832 struct iwl_device_cmd *dev_cmd; 833 struct iwl_tx_cmd *tx_cmd; 834 __le16 fc; 835 u16 seq_number = 0; 836 u8 tid = IWL_MAX_TID_COUNT; 837 u8 txq_id = info->hw_queue; 838 bool is_ampdu = false; 839 int hdrlen; 840 841 mvmsta = iwl_mvm_sta_from_mac80211(sta); 842 fc = hdr->frame_control; 843 hdrlen = ieee80211_hdrlen(fc); 844 845 if (WARN_ON_ONCE(!mvmsta)) 846 return -1; 847 848 if (WARN_ON_ONCE(mvmsta->sta_id == IWL_MVM_STATION_COUNT)) 849 return -1; 850 851 dev_cmd = iwl_mvm_set_tx_params(mvm, skb, info, hdrlen, 852 sta, mvmsta->sta_id); 853 if (!dev_cmd) 854 goto drop; 855 856 tx_cmd = (struct iwl_tx_cmd *)dev_cmd->payload; 857 /* From now on, we cannot access info->control */ 858 859 /* 860 * we handle that entirely ourselves -- for uAPSD the firmware 861 * will always send a notification, and for PS-Poll responses 862 * we'll notify mac80211 when getting frame status 863 */ 864 info->flags &= ~IEEE80211_TX_STATUS_EOSP; 865 866 spin_lock(&mvmsta->lock); 867 868 if (ieee80211_is_data_qos(fc) && !ieee80211_is_qos_nullfunc(fc)) { 869 u8 *qc = NULL; 870 qc = ieee80211_get_qos_ctl(hdr); 871 tid = qc[0] & IEEE80211_QOS_CTL_TID_MASK; 872 if (WARN_ON_ONCE(tid >= IWL_MAX_TID_COUNT)) 873 goto drop_unlock_sta; 874 875 seq_number = mvmsta->tid_data[tid].seq_number; 876 seq_number &= IEEE80211_SCTL_SEQ; 877 hdr->seq_ctrl &= cpu_to_le16(IEEE80211_SCTL_FRAG); 878 hdr->seq_ctrl |= cpu_to_le16(seq_number); 879 is_ampdu = info->flags & IEEE80211_TX_CTL_AMPDU; 880 } else if (iwl_mvm_is_dqa_supported(mvm) && 881 (ieee80211_is_qos_nullfunc(fc) || 882 ieee80211_is_nullfunc(fc))) { 883 /* 884 * nullfunc frames should go to the MGMT queue regardless of QOS 885 */ 886 tid = IWL_MAX_TID_COUNT; 887 txq_id = mvmsta->tid_data[tid].txq_id; 888 } 889 890 /* Copy MAC header from skb into command buffer */ 891 memcpy(tx_cmd->hdr, hdr, hdrlen); 892 893 WARN_ON_ONCE(info->flags & IEEE80211_TX_CTL_SEND_AFTER_DTIM); 894 895 if (sta->tdls) { 896 /* default to TID 0 for non-QoS packets */ 897 u8 tdls_tid = tid == IWL_MAX_TID_COUNT ? 0 : tid; 898 899 txq_id = mvmsta->hw_queue[tid_to_mac80211_ac[tdls_tid]]; 900 } 901 902 if (is_ampdu) { 903 if (WARN_ON_ONCE(mvmsta->tid_data[tid].state != IWL_AGG_ON)) 904 goto drop_unlock_sta; 905 txq_id = mvmsta->tid_data[tid].txq_id; 906 } 907 908 if (iwl_mvm_is_dqa_supported(mvm)) { 909 if (unlikely(mvmsta->tid_data[tid].txq_id == 910 IEEE80211_INVAL_HW_QUEUE)) { 911 iwl_mvm_tx_add_stream(mvm, mvmsta, tid, skb); 912 913 /* 914 * The frame is now deferred, and the worker scheduled 915 * will re-allocate it, so we can free it for now. 916 */ 917 iwl_trans_free_tx_cmd(mvm->trans, dev_cmd); 918 spin_unlock(&mvmsta->lock); 919 return 0; 920 } 921 922 txq_id = mvmsta->tid_data[tid].txq_id; 923 } 924 925 IWL_DEBUG_TX(mvm, "TX to [%d|%d] Q:%d - seq: 0x%x\n", mvmsta->sta_id, 926 tid, txq_id, IEEE80211_SEQ_TO_SN(seq_number)); 927 928 if (iwl_trans_tx(mvm->trans, skb, dev_cmd, txq_id)) 929 goto drop_unlock_sta; 930 931 if (tid < IWL_MAX_TID_COUNT && !ieee80211_has_morefrags(fc)) 932 mvmsta->tid_data[tid].seq_number = seq_number + 0x10; 933 934 spin_unlock(&mvmsta->lock); 935 936 /* Increase pending frames count if this isn't AMPDU */ 937 if (!is_ampdu) 938 atomic_inc(&mvm->pending_frames[mvmsta->sta_id]); 939 940 return 0; 941 942 drop_unlock_sta: 943 iwl_trans_free_tx_cmd(mvm->trans, dev_cmd); 944 spin_unlock(&mvmsta->lock); 945 drop: 946 return -1; 947 } 948 949 int iwl_mvm_tx_skb(struct iwl_mvm *mvm, struct sk_buff *skb, 950 struct ieee80211_sta *sta) 951 { 952 struct iwl_mvm_sta *mvmsta = iwl_mvm_sta_from_mac80211(sta); 953 struct ieee80211_tx_info *skb_info = IEEE80211_SKB_CB(skb); 954 struct ieee80211_tx_info info; 955 struct sk_buff_head mpdus_skbs; 956 unsigned int payload_len; 957 int ret; 958 959 if (WARN_ON_ONCE(!mvmsta)) 960 return -1; 961 962 if (WARN_ON_ONCE(mvmsta->sta_id == IWL_MVM_STATION_COUNT)) 963 return -1; 964 965 memcpy(&info, skb->cb, sizeof(info)); 966 967 /* This holds the amsdu headers length */ 968 skb_info->driver_data[0] = (void *)(uintptr_t)0; 969 970 if (!skb_is_gso(skb)) 971 return iwl_mvm_tx_mpdu(mvm, skb, &info, sta); 972 973 payload_len = skb_tail_pointer(skb) - skb_transport_header(skb) - 974 tcp_hdrlen(skb) + skb->data_len; 975 976 if (payload_len <= skb_shinfo(skb)->gso_size) 977 return iwl_mvm_tx_mpdu(mvm, skb, &info, sta); 978 979 __skb_queue_head_init(&mpdus_skbs); 980 981 ret = iwl_mvm_tx_tso(mvm, skb, &info, sta, &mpdus_skbs); 982 if (ret) 983 return ret; 984 985 if (WARN_ON(skb_queue_empty(&mpdus_skbs))) 986 return ret; 987 988 while (!skb_queue_empty(&mpdus_skbs)) { 989 skb = __skb_dequeue(&mpdus_skbs); 990 991 ret = iwl_mvm_tx_mpdu(mvm, skb, &info, sta); 992 if (ret) { 993 __skb_queue_purge(&mpdus_skbs); 994 return ret; 995 } 996 } 997 998 return 0; 999 } 1000 1001 static void iwl_mvm_check_ratid_empty(struct iwl_mvm *mvm, 1002 struct ieee80211_sta *sta, u8 tid) 1003 { 1004 struct iwl_mvm_sta *mvmsta = iwl_mvm_sta_from_mac80211(sta); 1005 struct iwl_mvm_tid_data *tid_data = &mvmsta->tid_data[tid]; 1006 struct ieee80211_vif *vif = mvmsta->vif; 1007 1008 lockdep_assert_held(&mvmsta->lock); 1009 1010 if ((tid_data->state == IWL_AGG_ON || 1011 tid_data->state == IWL_EMPTYING_HW_QUEUE_DELBA) && 1012 iwl_mvm_tid_queued(tid_data) == 0) { 1013 /* 1014 * Now that this aggregation queue is empty tell mac80211 so it 1015 * knows we no longer have frames buffered for the station on 1016 * this TID (for the TIM bitmap calculation.) 1017 */ 1018 ieee80211_sta_set_buffered(sta, tid, false); 1019 } 1020 1021 if (tid_data->ssn != tid_data->next_reclaimed) 1022 return; 1023 1024 switch (tid_data->state) { 1025 case IWL_EMPTYING_HW_QUEUE_ADDBA: 1026 IWL_DEBUG_TX_QUEUES(mvm, 1027 "Can continue addBA flow ssn = next_recl = %d\n", 1028 tid_data->next_reclaimed); 1029 tid_data->state = IWL_AGG_STARTING; 1030 ieee80211_start_tx_ba_cb_irqsafe(vif, sta->addr, tid); 1031 break; 1032 1033 case IWL_EMPTYING_HW_QUEUE_DELBA: 1034 IWL_DEBUG_TX_QUEUES(mvm, 1035 "Can continue DELBA flow ssn = next_recl = %d\n", 1036 tid_data->next_reclaimed); 1037 iwl_mvm_disable_txq(mvm, tid_data->txq_id, 1038 vif->hw_queue[tid_to_mac80211_ac[tid]], tid, 1039 CMD_ASYNC); 1040 tid_data->state = IWL_AGG_OFF; 1041 ieee80211_stop_tx_ba_cb_irqsafe(vif, sta->addr, tid); 1042 break; 1043 1044 default: 1045 break; 1046 } 1047 } 1048 1049 #ifdef CONFIG_IWLWIFI_DEBUG 1050 const char *iwl_mvm_get_tx_fail_reason(u32 status) 1051 { 1052 #define TX_STATUS_FAIL(x) case TX_STATUS_FAIL_ ## x: return #x 1053 #define TX_STATUS_POSTPONE(x) case TX_STATUS_POSTPONE_ ## x: return #x 1054 1055 switch (status & TX_STATUS_MSK) { 1056 case TX_STATUS_SUCCESS: 1057 return "SUCCESS"; 1058 TX_STATUS_POSTPONE(DELAY); 1059 TX_STATUS_POSTPONE(FEW_BYTES); 1060 TX_STATUS_POSTPONE(BT_PRIO); 1061 TX_STATUS_POSTPONE(QUIET_PERIOD); 1062 TX_STATUS_POSTPONE(CALC_TTAK); 1063 TX_STATUS_FAIL(INTERNAL_CROSSED_RETRY); 1064 TX_STATUS_FAIL(SHORT_LIMIT); 1065 TX_STATUS_FAIL(LONG_LIMIT); 1066 TX_STATUS_FAIL(UNDERRUN); 1067 TX_STATUS_FAIL(DRAIN_FLOW); 1068 TX_STATUS_FAIL(RFKILL_FLUSH); 1069 TX_STATUS_FAIL(LIFE_EXPIRE); 1070 TX_STATUS_FAIL(DEST_PS); 1071 TX_STATUS_FAIL(HOST_ABORTED); 1072 TX_STATUS_FAIL(BT_RETRY); 1073 TX_STATUS_FAIL(STA_INVALID); 1074 TX_STATUS_FAIL(FRAG_DROPPED); 1075 TX_STATUS_FAIL(TID_DISABLE); 1076 TX_STATUS_FAIL(FIFO_FLUSHED); 1077 TX_STATUS_FAIL(SMALL_CF_POLL); 1078 TX_STATUS_FAIL(FW_DROP); 1079 TX_STATUS_FAIL(STA_COLOR_MISMATCH); 1080 } 1081 1082 return "UNKNOWN"; 1083 1084 #undef TX_STATUS_FAIL 1085 #undef TX_STATUS_POSTPONE 1086 } 1087 #endif /* CONFIG_IWLWIFI_DEBUG */ 1088 1089 void iwl_mvm_hwrate_to_tx_rate(u32 rate_n_flags, 1090 enum nl80211_band band, 1091 struct ieee80211_tx_rate *r) 1092 { 1093 if (rate_n_flags & RATE_HT_MCS_GF_MSK) 1094 r->flags |= IEEE80211_TX_RC_GREEN_FIELD; 1095 switch (rate_n_flags & RATE_MCS_CHAN_WIDTH_MSK) { 1096 case RATE_MCS_CHAN_WIDTH_20: 1097 break; 1098 case RATE_MCS_CHAN_WIDTH_40: 1099 r->flags |= IEEE80211_TX_RC_40_MHZ_WIDTH; 1100 break; 1101 case RATE_MCS_CHAN_WIDTH_80: 1102 r->flags |= IEEE80211_TX_RC_80_MHZ_WIDTH; 1103 break; 1104 case RATE_MCS_CHAN_WIDTH_160: 1105 r->flags |= IEEE80211_TX_RC_160_MHZ_WIDTH; 1106 break; 1107 } 1108 if (rate_n_flags & RATE_MCS_SGI_MSK) 1109 r->flags |= IEEE80211_TX_RC_SHORT_GI; 1110 if (rate_n_flags & RATE_MCS_HT_MSK) { 1111 r->flags |= IEEE80211_TX_RC_MCS; 1112 r->idx = rate_n_flags & RATE_HT_MCS_INDEX_MSK; 1113 } else if (rate_n_flags & RATE_MCS_VHT_MSK) { 1114 ieee80211_rate_set_vht( 1115 r, rate_n_flags & RATE_VHT_MCS_RATE_CODE_MSK, 1116 ((rate_n_flags & RATE_VHT_MCS_NSS_MSK) >> 1117 RATE_VHT_MCS_NSS_POS) + 1); 1118 r->flags |= IEEE80211_TX_RC_VHT_MCS; 1119 } else { 1120 r->idx = iwl_mvm_legacy_rate_to_mac80211_idx(rate_n_flags, 1121 band); 1122 } 1123 } 1124 1125 /** 1126 * translate ucode response to mac80211 tx status control values 1127 */ 1128 static void iwl_mvm_hwrate_to_tx_status(u32 rate_n_flags, 1129 struct ieee80211_tx_info *info) 1130 { 1131 struct ieee80211_tx_rate *r = &info->status.rates[0]; 1132 1133 info->status.antenna = 1134 ((rate_n_flags & RATE_MCS_ANT_ABC_MSK) >> RATE_MCS_ANT_POS); 1135 iwl_mvm_hwrate_to_tx_rate(rate_n_flags, info->band, r); 1136 } 1137 1138 static void iwl_mvm_tx_status_check_trigger(struct iwl_mvm *mvm, 1139 u32 status) 1140 { 1141 struct iwl_fw_dbg_trigger_tlv *trig; 1142 struct iwl_fw_dbg_trigger_tx_status *status_trig; 1143 int i; 1144 1145 if (!iwl_fw_dbg_trigger_enabled(mvm->fw, FW_DBG_TRIGGER_TX_STATUS)) 1146 return; 1147 1148 trig = iwl_fw_dbg_get_trigger(mvm->fw, FW_DBG_TRIGGER_TX_STATUS); 1149 status_trig = (void *)trig->data; 1150 1151 if (!iwl_fw_dbg_trigger_check_stop(mvm, NULL, trig)) 1152 return; 1153 1154 for (i = 0; i < ARRAY_SIZE(status_trig->statuses); i++) { 1155 /* don't collect on status 0 */ 1156 if (!status_trig->statuses[i].status) 1157 break; 1158 1159 if (status_trig->statuses[i].status != (status & TX_STATUS_MSK)) 1160 continue; 1161 1162 iwl_mvm_fw_dbg_collect_trig(mvm, trig, 1163 "Tx status %d was received", 1164 status & TX_STATUS_MSK); 1165 break; 1166 } 1167 } 1168 1169 static void iwl_mvm_rx_tx_cmd_single(struct iwl_mvm *mvm, 1170 struct iwl_rx_packet *pkt) 1171 { 1172 struct ieee80211_sta *sta; 1173 u16 sequence = le16_to_cpu(pkt->hdr.sequence); 1174 int txq_id = SEQ_TO_QUEUE(sequence); 1175 struct iwl_mvm_tx_resp *tx_resp = (void *)pkt->data; 1176 int sta_id = IWL_MVM_TX_RES_GET_RA(tx_resp->ra_tid); 1177 int tid = IWL_MVM_TX_RES_GET_TID(tx_resp->ra_tid); 1178 u32 status = le16_to_cpu(tx_resp->status.status); 1179 u16 ssn = iwl_mvm_get_scd_ssn(tx_resp); 1180 struct iwl_mvm_sta *mvmsta; 1181 struct sk_buff_head skbs; 1182 u8 skb_freed = 0; 1183 u16 next_reclaimed, seq_ctl; 1184 bool is_ndp = false; 1185 bool txq_agg = false; /* Is this TXQ aggregated */ 1186 1187 __skb_queue_head_init(&skbs); 1188 1189 seq_ctl = le16_to_cpu(tx_resp->seq_ctl); 1190 1191 /* we can free until ssn % q.n_bd not inclusive */ 1192 iwl_trans_reclaim(mvm->trans, txq_id, ssn, &skbs); 1193 1194 while (!skb_queue_empty(&skbs)) { 1195 struct sk_buff *skb = __skb_dequeue(&skbs); 1196 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 1197 1198 skb_freed++; 1199 1200 iwl_trans_free_tx_cmd(mvm->trans, info->driver_data[1]); 1201 1202 memset(&info->status, 0, sizeof(info->status)); 1203 1204 info->flags &= ~IEEE80211_TX_CTL_AMPDU; 1205 1206 /* inform mac80211 about what happened with the frame */ 1207 switch (status & TX_STATUS_MSK) { 1208 case TX_STATUS_SUCCESS: 1209 case TX_STATUS_DIRECT_DONE: 1210 info->flags |= IEEE80211_TX_STAT_ACK; 1211 break; 1212 case TX_STATUS_FAIL_DEST_PS: 1213 info->flags |= IEEE80211_TX_STAT_TX_FILTERED; 1214 break; 1215 default: 1216 break; 1217 } 1218 1219 iwl_mvm_tx_status_check_trigger(mvm, status); 1220 1221 info->status.rates[0].count = tx_resp->failure_frame + 1; 1222 iwl_mvm_hwrate_to_tx_status(le32_to_cpu(tx_resp->initial_rate), 1223 info); 1224 info->status.status_driver_data[1] = 1225 (void *)(uintptr_t)le32_to_cpu(tx_resp->initial_rate); 1226 1227 /* Single frame failure in an AMPDU queue => send BAR */ 1228 if (txq_id >= mvm->first_agg_queue && 1229 !(info->flags & IEEE80211_TX_STAT_ACK) && 1230 !(info->flags & IEEE80211_TX_STAT_TX_FILTERED)) 1231 info->flags |= IEEE80211_TX_STAT_AMPDU_NO_BACK; 1232 1233 /* W/A FW bug: seq_ctl is wrong when the status isn't success */ 1234 if (status != TX_STATUS_SUCCESS) { 1235 struct ieee80211_hdr *hdr = (void *)skb->data; 1236 seq_ctl = le16_to_cpu(hdr->seq_ctrl); 1237 } 1238 1239 if (unlikely(!seq_ctl)) { 1240 struct ieee80211_hdr *hdr = (void *)skb->data; 1241 1242 /* 1243 * If it is an NDP, we can't update next_reclaim since 1244 * its sequence control is 0. Note that for that same 1245 * reason, NDPs are never sent to A-MPDU'able queues 1246 * so that we can never have more than one freed frame 1247 * for a single Tx resonse (see WARN_ON below). 1248 */ 1249 if (ieee80211_is_qos_nullfunc(hdr->frame_control)) 1250 is_ndp = true; 1251 } 1252 1253 /* 1254 * TODO: this is not accurate if we are freeing more than one 1255 * packet. 1256 */ 1257 info->status.tx_time = 1258 le16_to_cpu(tx_resp->wireless_media_time); 1259 BUILD_BUG_ON(ARRAY_SIZE(info->status.status_driver_data) < 1); 1260 info->status.status_driver_data[0] = 1261 (void *)(uintptr_t)tx_resp->reduced_tpc; 1262 1263 ieee80211_tx_status(mvm->hw, skb); 1264 } 1265 1266 if (txq_id >= mvm->first_agg_queue) { 1267 /* If this is an aggregation queue, we use the ssn since: 1268 * ssn = wifi seq_num % 256. 1269 * The seq_ctl is the sequence control of the packet to which 1270 * this Tx response relates. But if there is a hole in the 1271 * bitmap of the BA we received, this Tx response may allow to 1272 * reclaim the hole and all the subsequent packets that were 1273 * already acked. In that case, seq_ctl != ssn, and the next 1274 * packet to be reclaimed will be ssn and not seq_ctl. In that 1275 * case, several packets will be reclaimed even if 1276 * frame_count = 1. 1277 * 1278 * The ssn is the index (% 256) of the latest packet that has 1279 * treated (acked / dropped) + 1. 1280 */ 1281 next_reclaimed = ssn; 1282 } else { 1283 /* The next packet to be reclaimed is the one after this one */ 1284 next_reclaimed = IEEE80211_SEQ_TO_SN(seq_ctl + 0x10); 1285 } 1286 1287 IWL_DEBUG_TX_REPLY(mvm, 1288 "TXQ %d status %s (0x%08x)\n", 1289 txq_id, iwl_mvm_get_tx_fail_reason(status), status); 1290 1291 IWL_DEBUG_TX_REPLY(mvm, 1292 "\t\t\t\tinitial_rate 0x%x retries %d, idx=%d ssn=%d next_reclaimed=0x%x seq_ctl=0x%x\n", 1293 le32_to_cpu(tx_resp->initial_rate), 1294 tx_resp->failure_frame, SEQ_TO_INDEX(sequence), 1295 ssn, next_reclaimed, seq_ctl); 1296 1297 rcu_read_lock(); 1298 1299 sta = rcu_dereference(mvm->fw_id_to_mac_id[sta_id]); 1300 /* 1301 * sta can't be NULL otherwise it'd mean that the sta has been freed in 1302 * the firmware while we still have packets for it in the Tx queues. 1303 */ 1304 if (WARN_ON_ONCE(!sta)) 1305 goto out; 1306 1307 if (!IS_ERR(sta)) { 1308 mvmsta = iwl_mvm_sta_from_mac80211(sta); 1309 1310 if (tid != IWL_TID_NON_QOS) { 1311 struct iwl_mvm_tid_data *tid_data = 1312 &mvmsta->tid_data[tid]; 1313 bool send_eosp_ndp = false; 1314 1315 spin_lock_bh(&mvmsta->lock); 1316 txq_agg = (mvmsta->tid_data[tid].state == IWL_AGG_ON); 1317 1318 if (!is_ndp) { 1319 tid_data->next_reclaimed = next_reclaimed; 1320 IWL_DEBUG_TX_REPLY(mvm, 1321 "Next reclaimed packet:%d\n", 1322 next_reclaimed); 1323 } else { 1324 IWL_DEBUG_TX_REPLY(mvm, 1325 "NDP - don't update next_reclaimed\n"); 1326 } 1327 1328 iwl_mvm_check_ratid_empty(mvm, sta, tid); 1329 1330 if (mvmsta->sleep_tx_count) { 1331 mvmsta->sleep_tx_count--; 1332 if (mvmsta->sleep_tx_count && 1333 !iwl_mvm_tid_queued(tid_data)) { 1334 /* 1335 * The number of frames in the queue 1336 * dropped to 0 even if we sent less 1337 * frames than we thought we had on the 1338 * Tx queue. 1339 * This means we had holes in the BA 1340 * window that we just filled, ask 1341 * mac80211 to send EOSP since the 1342 * firmware won't know how to do that. 1343 * Send NDP and the firmware will send 1344 * EOSP notification that will trigger 1345 * a call to ieee80211_sta_eosp(). 1346 */ 1347 send_eosp_ndp = true; 1348 } 1349 } 1350 1351 spin_unlock_bh(&mvmsta->lock); 1352 if (send_eosp_ndp) { 1353 iwl_mvm_sta_modify_sleep_tx_count(mvm, sta, 1354 IEEE80211_FRAME_RELEASE_UAPSD, 1355 1, tid, false, false); 1356 mvmsta->sleep_tx_count = 0; 1357 ieee80211_send_eosp_nullfunc(sta, tid); 1358 } 1359 } 1360 1361 if (mvmsta->next_status_eosp) { 1362 mvmsta->next_status_eosp = false; 1363 ieee80211_sta_eosp(sta); 1364 } 1365 } else { 1366 mvmsta = NULL; 1367 } 1368 1369 /* 1370 * If the txq is not an AMPDU queue, there is no chance we freed 1371 * several skbs. Check that out... 1372 */ 1373 if (txq_agg) 1374 goto out; 1375 1376 /* We can't free more than one frame at once on a shared queue */ 1377 WARN_ON(!iwl_mvm_is_dqa_supported(mvm) && (skb_freed > 1)); 1378 1379 /* If we have still frames for this STA nothing to do here */ 1380 if (!atomic_sub_and_test(skb_freed, &mvm->pending_frames[sta_id])) 1381 goto out; 1382 1383 if (mvmsta && mvmsta->vif->type == NL80211_IFTYPE_AP) { 1384 1385 /* 1386 * If there are no pending frames for this STA and 1387 * the tx to this station is not disabled, notify 1388 * mac80211 that this station can now wake up in its 1389 * STA table. 1390 * If mvmsta is not NULL, sta is valid. 1391 */ 1392 1393 spin_lock_bh(&mvmsta->lock); 1394 1395 if (!mvmsta->disable_tx) 1396 ieee80211_sta_block_awake(mvm->hw, sta, false); 1397 1398 spin_unlock_bh(&mvmsta->lock); 1399 } 1400 1401 if (PTR_ERR(sta) == -EBUSY || PTR_ERR(sta) == -ENOENT) { 1402 /* 1403 * We are draining and this was the last packet - pre_rcu_remove 1404 * has been called already. We might be after the 1405 * synchronize_net already. 1406 * Don't rely on iwl_mvm_rm_sta to see the empty Tx queues. 1407 */ 1408 set_bit(sta_id, mvm->sta_drained); 1409 schedule_work(&mvm->sta_drained_wk); 1410 } 1411 1412 out: 1413 rcu_read_unlock(); 1414 } 1415 1416 #ifdef CONFIG_IWLWIFI_DEBUG 1417 #define AGG_TX_STATE_(x) case AGG_TX_STATE_ ## x: return #x 1418 static const char *iwl_get_agg_tx_status(u16 status) 1419 { 1420 switch (status & AGG_TX_STATE_STATUS_MSK) { 1421 AGG_TX_STATE_(TRANSMITTED); 1422 AGG_TX_STATE_(UNDERRUN); 1423 AGG_TX_STATE_(BT_PRIO); 1424 AGG_TX_STATE_(FEW_BYTES); 1425 AGG_TX_STATE_(ABORT); 1426 AGG_TX_STATE_(LAST_SENT_TTL); 1427 AGG_TX_STATE_(LAST_SENT_TRY_CNT); 1428 AGG_TX_STATE_(LAST_SENT_BT_KILL); 1429 AGG_TX_STATE_(SCD_QUERY); 1430 AGG_TX_STATE_(TEST_BAD_CRC32); 1431 AGG_TX_STATE_(RESPONSE); 1432 AGG_TX_STATE_(DUMP_TX); 1433 AGG_TX_STATE_(DELAY_TX); 1434 } 1435 1436 return "UNKNOWN"; 1437 } 1438 1439 static void iwl_mvm_rx_tx_cmd_agg_dbg(struct iwl_mvm *mvm, 1440 struct iwl_rx_packet *pkt) 1441 { 1442 struct iwl_mvm_tx_resp *tx_resp = (void *)pkt->data; 1443 struct agg_tx_status *frame_status = &tx_resp->status; 1444 int i; 1445 1446 for (i = 0; i < tx_resp->frame_count; i++) { 1447 u16 fstatus = le16_to_cpu(frame_status[i].status); 1448 1449 IWL_DEBUG_TX_REPLY(mvm, 1450 "status %s (0x%04x), try-count (%d) seq (0x%x)\n", 1451 iwl_get_agg_tx_status(fstatus), 1452 fstatus & AGG_TX_STATE_STATUS_MSK, 1453 (fstatus & AGG_TX_STATE_TRY_CNT_MSK) >> 1454 AGG_TX_STATE_TRY_CNT_POS, 1455 le16_to_cpu(frame_status[i].sequence)); 1456 } 1457 } 1458 #else 1459 static void iwl_mvm_rx_tx_cmd_agg_dbg(struct iwl_mvm *mvm, 1460 struct iwl_rx_packet *pkt) 1461 {} 1462 #endif /* CONFIG_IWLWIFI_DEBUG */ 1463 1464 static void iwl_mvm_rx_tx_cmd_agg(struct iwl_mvm *mvm, 1465 struct iwl_rx_packet *pkt) 1466 { 1467 struct iwl_mvm_tx_resp *tx_resp = (void *)pkt->data; 1468 int sta_id = IWL_MVM_TX_RES_GET_RA(tx_resp->ra_tid); 1469 int tid = IWL_MVM_TX_RES_GET_TID(tx_resp->ra_tid); 1470 u16 sequence = le16_to_cpu(pkt->hdr.sequence); 1471 struct iwl_mvm_sta *mvmsta; 1472 int queue = SEQ_TO_QUEUE(sequence); 1473 1474 if (WARN_ON_ONCE(queue < mvm->first_agg_queue && 1475 (!iwl_mvm_is_dqa_supported(mvm) || 1476 (queue != IWL_MVM_DQA_BSS_CLIENT_QUEUE)))) 1477 return; 1478 1479 if (WARN_ON_ONCE(tid == IWL_TID_NON_QOS)) 1480 return; 1481 1482 iwl_mvm_rx_tx_cmd_agg_dbg(mvm, pkt); 1483 1484 rcu_read_lock(); 1485 1486 mvmsta = iwl_mvm_sta_from_staid_rcu(mvm, sta_id); 1487 1488 if (!WARN_ON_ONCE(!mvmsta)) { 1489 mvmsta->tid_data[tid].rate_n_flags = 1490 le32_to_cpu(tx_resp->initial_rate); 1491 mvmsta->tid_data[tid].tx_time = 1492 le16_to_cpu(tx_resp->wireless_media_time); 1493 } 1494 1495 rcu_read_unlock(); 1496 } 1497 1498 void iwl_mvm_rx_tx_cmd(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb) 1499 { 1500 struct iwl_rx_packet *pkt = rxb_addr(rxb); 1501 struct iwl_mvm_tx_resp *tx_resp = (void *)pkt->data; 1502 1503 if (tx_resp->frame_count == 1) 1504 iwl_mvm_rx_tx_cmd_single(mvm, pkt); 1505 else 1506 iwl_mvm_rx_tx_cmd_agg(mvm, pkt); 1507 } 1508 1509 static void iwl_mvm_tx_info_from_ba_notif(struct ieee80211_tx_info *info, 1510 struct iwl_mvm_ba_notif *ba_notif, 1511 struct iwl_mvm_tid_data *tid_data) 1512 { 1513 info->flags |= IEEE80211_TX_STAT_AMPDU; 1514 info->status.ampdu_ack_len = ba_notif->txed_2_done; 1515 info->status.ampdu_len = ba_notif->txed; 1516 iwl_mvm_hwrate_to_tx_status(tid_data->rate_n_flags, 1517 info); 1518 /* TODO: not accounted if the whole A-MPDU failed */ 1519 info->status.tx_time = tid_data->tx_time; 1520 info->status.status_driver_data[0] = 1521 (void *)(uintptr_t)ba_notif->reduced_txp; 1522 info->status.status_driver_data[1] = 1523 (void *)(uintptr_t)tid_data->rate_n_flags; 1524 } 1525 1526 void iwl_mvm_rx_ba_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb) 1527 { 1528 struct iwl_rx_packet *pkt = rxb_addr(rxb); 1529 struct iwl_mvm_ba_notif *ba_notif = (void *)pkt->data; 1530 struct sk_buff_head reclaimed_skbs; 1531 struct iwl_mvm_tid_data *tid_data; 1532 struct ieee80211_sta *sta; 1533 struct iwl_mvm_sta *mvmsta; 1534 struct sk_buff *skb; 1535 int sta_id, tid, freed; 1536 /* "flow" corresponds to Tx queue */ 1537 u16 scd_flow = le16_to_cpu(ba_notif->scd_flow); 1538 /* "ssn" is start of block-ack Tx window, corresponds to index 1539 * (in Tx queue's circular buffer) of first TFD/frame in window */ 1540 u16 ba_resp_scd_ssn = le16_to_cpu(ba_notif->scd_ssn); 1541 1542 sta_id = ba_notif->sta_id; 1543 tid = ba_notif->tid; 1544 1545 if (WARN_ONCE(sta_id >= IWL_MVM_STATION_COUNT || 1546 tid >= IWL_MAX_TID_COUNT, 1547 "sta_id %d tid %d", sta_id, tid)) 1548 return; 1549 1550 rcu_read_lock(); 1551 1552 sta = rcu_dereference(mvm->fw_id_to_mac_id[sta_id]); 1553 1554 /* Reclaiming frames for a station that has been deleted ? */ 1555 if (WARN_ON_ONCE(IS_ERR_OR_NULL(sta))) { 1556 rcu_read_unlock(); 1557 return; 1558 } 1559 1560 mvmsta = iwl_mvm_sta_from_mac80211(sta); 1561 tid_data = &mvmsta->tid_data[tid]; 1562 1563 if (tid_data->txq_id != scd_flow) { 1564 IWL_ERR(mvm, 1565 "invalid BA notification: Q %d, tid %d, flow %d\n", 1566 tid_data->txq_id, tid, scd_flow); 1567 rcu_read_unlock(); 1568 return; 1569 } 1570 1571 spin_lock_bh(&mvmsta->lock); 1572 1573 __skb_queue_head_init(&reclaimed_skbs); 1574 1575 /* 1576 * Release all TFDs before the SSN, i.e. all TFDs in front of 1577 * block-ack window (we assume that they've been successfully 1578 * transmitted ... if not, it's too late anyway). 1579 */ 1580 iwl_trans_reclaim(mvm->trans, scd_flow, ba_resp_scd_ssn, 1581 &reclaimed_skbs); 1582 1583 IWL_DEBUG_TX_REPLY(mvm, 1584 "BA_NOTIFICATION Received from %pM, sta_id = %d\n", 1585 (u8 *)&ba_notif->sta_addr_lo32, 1586 ba_notif->sta_id); 1587 IWL_DEBUG_TX_REPLY(mvm, 1588 "TID = %d, SeqCtl = %d, bitmap = 0x%llx, scd_flow = %d, scd_ssn = %d sent:%d, acked:%d\n", 1589 ba_notif->tid, le16_to_cpu(ba_notif->seq_ctl), 1590 (unsigned long long)le64_to_cpu(ba_notif->bitmap), 1591 scd_flow, ba_resp_scd_ssn, ba_notif->txed, 1592 ba_notif->txed_2_done); 1593 1594 IWL_DEBUG_TX_REPLY(mvm, "reduced txp from ba notif %d\n", 1595 ba_notif->reduced_txp); 1596 tid_data->next_reclaimed = ba_resp_scd_ssn; 1597 1598 iwl_mvm_check_ratid_empty(mvm, sta, tid); 1599 1600 freed = 0; 1601 1602 skb_queue_walk(&reclaimed_skbs, skb) { 1603 struct ieee80211_hdr *hdr = (void *)skb->data; 1604 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 1605 1606 if (ieee80211_is_data_qos(hdr->frame_control)) 1607 freed++; 1608 else 1609 WARN_ON_ONCE(1); 1610 1611 iwl_trans_free_tx_cmd(mvm->trans, info->driver_data[1]); 1612 1613 memset(&info->status, 0, sizeof(info->status)); 1614 /* Packet was transmitted successfully, failures come as single 1615 * frames because before failing a frame the firmware transmits 1616 * it without aggregation at least once. 1617 */ 1618 info->flags |= IEEE80211_TX_STAT_ACK; 1619 1620 /* this is the first skb we deliver in this batch */ 1621 /* put the rate scaling data there */ 1622 if (freed == 1) 1623 iwl_mvm_tx_info_from_ba_notif(info, ba_notif, tid_data); 1624 } 1625 1626 spin_unlock_bh(&mvmsta->lock); 1627 1628 /* We got a BA notif with 0 acked or scd_ssn didn't progress which is 1629 * possible (i.e. first MPDU in the aggregation wasn't acked) 1630 * Still it's important to update RS about sent vs. acked. 1631 */ 1632 if (skb_queue_empty(&reclaimed_skbs)) { 1633 struct ieee80211_tx_info ba_info = {}; 1634 struct ieee80211_chanctx_conf *chanctx_conf = NULL; 1635 1636 if (mvmsta->vif) 1637 chanctx_conf = 1638 rcu_dereference(mvmsta->vif->chanctx_conf); 1639 1640 if (WARN_ON_ONCE(!chanctx_conf)) 1641 goto out; 1642 1643 ba_info.band = chanctx_conf->def.chan->band; 1644 iwl_mvm_tx_info_from_ba_notif(&ba_info, ba_notif, tid_data); 1645 1646 IWL_DEBUG_TX_REPLY(mvm, "No reclaim. Update rs directly\n"); 1647 iwl_mvm_rs_tx_status(mvm, sta, tid, &ba_info); 1648 } 1649 1650 out: 1651 rcu_read_unlock(); 1652 1653 while (!skb_queue_empty(&reclaimed_skbs)) { 1654 skb = __skb_dequeue(&reclaimed_skbs); 1655 ieee80211_tx_status(mvm->hw, skb); 1656 } 1657 } 1658 1659 /* 1660 * Note that there are transports that buffer frames before they reach 1661 * the firmware. This means that after flush_tx_path is called, the 1662 * queue might not be empty. The race-free way to handle this is to: 1663 * 1) set the station as draining 1664 * 2) flush the Tx path 1665 * 3) wait for the transport queues to be empty 1666 */ 1667 int iwl_mvm_flush_tx_path(struct iwl_mvm *mvm, u32 tfd_msk, u32 flags) 1668 { 1669 int ret; 1670 struct iwl_tx_path_flush_cmd flush_cmd = { 1671 .queues_ctl = cpu_to_le32(tfd_msk), 1672 .flush_ctl = cpu_to_le16(DUMP_TX_FIFO_FLUSH), 1673 }; 1674 1675 ret = iwl_mvm_send_cmd_pdu(mvm, TXPATH_FLUSH, flags, 1676 sizeof(flush_cmd), &flush_cmd); 1677 if (ret) 1678 IWL_ERR(mvm, "Failed to send flush command (%d)\n", ret); 1679 return ret; 1680 } 1681