1 // SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause 2 /* 3 * Copyright (C) 2003-2014, 2018-2021 Intel Corporation 4 * Copyright (C) 2013-2015 Intel Mobile Communications GmbH 5 * Copyright (C) 2016-2017 Intel Deutschland GmbH 6 */ 7 #include <linux/etherdevice.h> 8 #include <linux/ieee80211.h> 9 #include <linux/slab.h> 10 #include <linux/sched.h> 11 #ifdef CONFIG_INET 12 #include <net/ip6_checksum.h> 13 #include <net/tso.h> 14 #endif 15 #if defined(__FreeBSD__) 16 #include <net/mac80211.h> 17 #endif 18 19 #include "iwl-debug.h" 20 #include "iwl-csr.h" 21 #include "iwl-prph.h" 22 #include "iwl-io.h" 23 #include "iwl-scd.h" 24 #include "iwl-op-mode.h" 25 #include "internal.h" 26 #include "fw/api/tx.h" 27 28 /*************** DMA-QUEUE-GENERAL-FUNCTIONS ***** 29 * DMA services 30 * 31 * Theory of operation 32 * 33 * A Tx or Rx queue resides in host DRAM, and is comprised of a circular buffer 34 * of buffer descriptors, each of which points to one or more data buffers for 35 * the device to read from or fill. Driver and device exchange status of each 36 * queue via "read" and "write" pointers. Driver keeps minimum of 2 empty 37 * entries in each circular buffer, to protect against confusing empty and full 38 * queue states. 39 * 40 * The device reads or writes the data in the queues via the device's several 41 * DMA/FIFO channels. Each queue is mapped to a single DMA channel. 42 * 43 * For Tx queue, there are low mark and high mark limits. If, after queuing 44 * the packet for Tx, free space become < low mark, Tx queue stopped. When 45 * reclaiming packets (on 'tx done IRQ), if free space become > high mark, 46 * Tx queue resumed. 47 * 48 ***************************************************/ 49 50 51 int iwl_pcie_alloc_dma_ptr(struct iwl_trans *trans, 52 struct iwl_dma_ptr *ptr, size_t size) 53 { 54 if (WARN_ON(ptr->addr)) 55 return -EINVAL; 56 57 ptr->addr = dma_alloc_coherent(trans->dev, size, 58 &ptr->dma, GFP_KERNEL); 59 if (!ptr->addr) 60 return -ENOMEM; 61 ptr->size = size; 62 return 0; 63 } 64 65 void iwl_pcie_free_dma_ptr(struct iwl_trans *trans, struct iwl_dma_ptr *ptr) 66 { 67 if (unlikely(!ptr->addr)) 68 return; 69 70 dma_free_coherent(trans->dev, ptr->size, ptr->addr, ptr->dma); 71 memset(ptr, 0, sizeof(*ptr)); 72 } 73 74 /* 75 * iwl_pcie_txq_inc_wr_ptr - Send new write index to hardware 76 */ 77 static void iwl_pcie_txq_inc_wr_ptr(struct iwl_trans *trans, 78 struct iwl_txq *txq) 79 { 80 u32 reg = 0; 81 int txq_id = txq->id; 82 83 lockdep_assert_held(&txq->lock); 84 85 /* 86 * explicitly wake up the NIC if: 87 * 1. shadow registers aren't enabled 88 * 2. NIC is woken up for CMD regardless of shadow outside this function 89 * 3. there is a chance that the NIC is asleep 90 */ 91 if (!trans->trans_cfg->base_params->shadow_reg_enable && 92 txq_id != trans->txqs.cmd.q_id && 93 test_bit(STATUS_TPOWER_PMI, &trans->status)) { 94 /* 95 * wake up nic if it's powered down ... 96 * uCode will wake up, and interrupt us again, so next 97 * time we'll skip this part. 98 */ 99 reg = iwl_read32(trans, CSR_UCODE_DRV_GP1); 100 101 if (reg & CSR_UCODE_DRV_GP1_BIT_MAC_SLEEP) { 102 IWL_DEBUG_INFO(trans, "Tx queue %d requesting wakeup, GP1 = 0x%x\n", 103 txq_id, reg); 104 iwl_set_bit(trans, CSR_GP_CNTRL, 105 CSR_GP_CNTRL_REG_FLAG_MAC_ACCESS_REQ); 106 txq->need_update = true; 107 return; 108 } 109 } 110 111 /* 112 * if not in power-save mode, uCode will never sleep when we're 113 * trying to tx (during RFKILL, we're not trying to tx). 114 */ 115 IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq_id, txq->write_ptr); 116 if (!txq->block) 117 iwl_write32(trans, HBUS_TARG_WRPTR, 118 txq->write_ptr | (txq_id << 8)); 119 } 120 121 void iwl_pcie_txq_check_wrptrs(struct iwl_trans *trans) 122 { 123 int i; 124 125 for (i = 0; i < trans->trans_cfg->base_params->num_of_queues; i++) { 126 struct iwl_txq *txq = trans->txqs.txq[i]; 127 128 if (!test_bit(i, trans->txqs.queue_used)) 129 continue; 130 131 spin_lock_bh(&txq->lock); 132 if (txq->need_update) { 133 iwl_pcie_txq_inc_wr_ptr(trans, txq); 134 txq->need_update = false; 135 } 136 spin_unlock_bh(&txq->lock); 137 } 138 } 139 140 static inline void iwl_pcie_tfd_set_tb(struct iwl_trans *trans, void *tfd, 141 u8 idx, dma_addr_t addr, u16 len) 142 { 143 struct iwl_tfd *tfd_fh = (void *)tfd; 144 struct iwl_tfd_tb *tb = &tfd_fh->tbs[idx]; 145 146 u16 hi_n_len = len << 4; 147 148 put_unaligned_le32(addr, &tb->lo); 149 hi_n_len |= iwl_get_dma_hi_addr(addr); 150 151 tb->hi_n_len = cpu_to_le16(hi_n_len); 152 153 tfd_fh->num_tbs = idx + 1; 154 } 155 156 static int iwl_pcie_txq_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq, 157 dma_addr_t addr, u16 len, bool reset) 158 { 159 void *tfd; 160 u32 num_tbs; 161 162 tfd = (u8 *)txq->tfds + trans->txqs.tfd.size * txq->write_ptr; 163 164 if (reset) 165 memset(tfd, 0, trans->txqs.tfd.size); 166 167 num_tbs = iwl_txq_gen1_tfd_get_num_tbs(trans, tfd); 168 169 /* Each TFD can point to a maximum max_tbs Tx buffers */ 170 if (num_tbs >= trans->txqs.tfd.max_tbs) { 171 IWL_ERR(trans, "Error can not send more than %d chunks\n", 172 trans->txqs.tfd.max_tbs); 173 return -EINVAL; 174 } 175 176 if (WARN(addr & ~IWL_TX_DMA_MASK, 177 "Unaligned address = %llx\n", (unsigned long long)addr)) 178 return -EINVAL; 179 180 iwl_pcie_tfd_set_tb(trans, tfd, num_tbs, addr, len); 181 182 return num_tbs; 183 } 184 185 static void iwl_pcie_clear_cmd_in_flight(struct iwl_trans *trans) 186 { 187 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans); 188 189 if (!trans->trans_cfg->base_params->apmg_wake_up_wa) 190 return; 191 192 spin_lock(&trans_pcie->reg_lock); 193 194 if (WARN_ON(!trans_pcie->cmd_hold_nic_awake)) { 195 spin_unlock(&trans_pcie->reg_lock); 196 return; 197 } 198 199 trans_pcie->cmd_hold_nic_awake = false; 200 __iwl_trans_pcie_clear_bit(trans, CSR_GP_CNTRL, 201 CSR_GP_CNTRL_REG_FLAG_MAC_ACCESS_REQ); 202 spin_unlock(&trans_pcie->reg_lock); 203 } 204 205 /* 206 * iwl_pcie_txq_unmap - Unmap any remaining DMA mappings and free skb's 207 */ 208 static void iwl_pcie_txq_unmap(struct iwl_trans *trans, int txq_id) 209 { 210 struct iwl_txq *txq = trans->txqs.txq[txq_id]; 211 212 if (!txq) { 213 IWL_ERR(trans, "Trying to free a queue that wasn't allocated?\n"); 214 return; 215 } 216 217 spin_lock_bh(&txq->lock); 218 while (txq->write_ptr != txq->read_ptr) { 219 IWL_DEBUG_TX_REPLY(trans, "Q %d Free %d\n", 220 txq_id, txq->read_ptr); 221 222 if (txq_id != trans->txqs.cmd.q_id) { 223 struct sk_buff *skb = txq->entries[txq->read_ptr].skb; 224 225 if (WARN_ON_ONCE(!skb)) 226 continue; 227 228 iwl_txq_free_tso_page(trans, skb); 229 } 230 iwl_txq_free_tfd(trans, txq); 231 txq->read_ptr = iwl_txq_inc_wrap(trans, txq->read_ptr); 232 233 if (txq->read_ptr == txq->write_ptr && 234 txq_id == trans->txqs.cmd.q_id) 235 iwl_pcie_clear_cmd_in_flight(trans); 236 } 237 238 while (!skb_queue_empty(&txq->overflow_q)) { 239 struct sk_buff *skb = __skb_dequeue(&txq->overflow_q); 240 241 iwl_op_mode_free_skb(trans->op_mode, skb); 242 } 243 244 spin_unlock_bh(&txq->lock); 245 246 /* just in case - this queue may have been stopped */ 247 iwl_wake_queue(trans, txq); 248 } 249 250 /* 251 * iwl_pcie_txq_free - Deallocate DMA queue. 252 * @txq: Transmit queue to deallocate. 253 * 254 * Empty queue by removing and destroying all BD's. 255 * Free all buffers. 256 * 0-fill, but do not free "txq" descriptor structure. 257 */ 258 static void iwl_pcie_txq_free(struct iwl_trans *trans, int txq_id) 259 { 260 struct iwl_txq *txq = trans->txqs.txq[txq_id]; 261 struct device *dev = trans->dev; 262 int i; 263 264 if (WARN_ON(!txq)) 265 return; 266 267 iwl_pcie_txq_unmap(trans, txq_id); 268 269 /* De-alloc array of command/tx buffers */ 270 if (txq_id == trans->txqs.cmd.q_id) 271 for (i = 0; i < txq->n_window; i++) { 272 kfree_sensitive(txq->entries[i].cmd); 273 kfree_sensitive(txq->entries[i].free_buf); 274 } 275 276 /* De-alloc circular buffer of TFDs */ 277 if (txq->tfds) { 278 dma_free_coherent(dev, 279 trans->txqs.tfd.size * 280 trans->trans_cfg->base_params->max_tfd_queue_size, 281 txq->tfds, txq->dma_addr); 282 txq->dma_addr = 0; 283 txq->tfds = NULL; 284 285 dma_free_coherent(dev, 286 sizeof(*txq->first_tb_bufs) * txq->n_window, 287 txq->first_tb_bufs, txq->first_tb_dma); 288 } 289 290 kfree(txq->entries); 291 txq->entries = NULL; 292 293 del_timer_sync(&txq->stuck_timer); 294 295 /* 0-fill queue descriptor structure */ 296 memset(txq, 0, sizeof(*txq)); 297 } 298 299 void iwl_pcie_tx_start(struct iwl_trans *trans, u32 scd_base_addr) 300 { 301 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans); 302 int nq = trans->trans_cfg->base_params->num_of_queues; 303 int chan; 304 u32 reg_val; 305 int clear_dwords = (SCD_TRANS_TBL_OFFSET_QUEUE(nq) - 306 SCD_CONTEXT_MEM_LOWER_BOUND) / sizeof(u32); 307 308 /* make sure all queue are not stopped/used */ 309 memset(trans->txqs.queue_stopped, 0, 310 sizeof(trans->txqs.queue_stopped)); 311 memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used)); 312 313 trans_pcie->scd_base_addr = 314 iwl_read_prph(trans, SCD_SRAM_BASE_ADDR); 315 316 WARN_ON(scd_base_addr != 0 && 317 scd_base_addr != trans_pcie->scd_base_addr); 318 319 /* reset context data, TX status and translation data */ 320 iwl_trans_write_mem(trans, trans_pcie->scd_base_addr + 321 SCD_CONTEXT_MEM_LOWER_BOUND, 322 NULL, clear_dwords); 323 324 iwl_write_prph(trans, SCD_DRAM_BASE_ADDR, 325 trans->txqs.scd_bc_tbls.dma >> 10); 326 327 /* The chain extension of the SCD doesn't work well. This feature is 328 * enabled by default by the HW, so we need to disable it manually. 329 */ 330 if (trans->trans_cfg->base_params->scd_chain_ext_wa) 331 iwl_write_prph(trans, SCD_CHAINEXT_EN, 0); 332 333 iwl_trans_ac_txq_enable(trans, trans->txqs.cmd.q_id, 334 trans->txqs.cmd.fifo, 335 trans->txqs.cmd.wdg_timeout); 336 337 /* Activate all Tx DMA/FIFO channels */ 338 iwl_scd_activate_fifos(trans); 339 340 /* Enable DMA channel */ 341 for (chan = 0; chan < FH_TCSR_CHNL_NUM; chan++) 342 iwl_write_direct32(trans, FH_TCSR_CHNL_TX_CONFIG_REG(chan), 343 FH_TCSR_TX_CONFIG_REG_VAL_DMA_CHNL_ENABLE | 344 FH_TCSR_TX_CONFIG_REG_VAL_DMA_CREDIT_ENABLE); 345 346 /* Update FH chicken bits */ 347 reg_val = iwl_read_direct32(trans, FH_TX_CHICKEN_BITS_REG); 348 iwl_write_direct32(trans, FH_TX_CHICKEN_BITS_REG, 349 reg_val | FH_TX_CHICKEN_BITS_SCD_AUTO_RETRY_EN); 350 351 /* Enable L1-Active */ 352 if (trans->trans_cfg->device_family < IWL_DEVICE_FAMILY_8000) 353 iwl_clear_bits_prph(trans, APMG_PCIDEV_STT_REG, 354 APMG_PCIDEV_STT_VAL_L1_ACT_DIS); 355 } 356 357 void iwl_trans_pcie_tx_reset(struct iwl_trans *trans) 358 { 359 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans); 360 int txq_id; 361 362 /* 363 * we should never get here in gen2 trans mode return early to avoid 364 * having invalid accesses 365 */ 366 if (WARN_ON_ONCE(trans->trans_cfg->gen2)) 367 return; 368 369 for (txq_id = 0; txq_id < trans->trans_cfg->base_params->num_of_queues; 370 txq_id++) { 371 struct iwl_txq *txq = trans->txqs.txq[txq_id]; 372 if (trans->trans_cfg->use_tfh) 373 iwl_write_direct64(trans, 374 FH_MEM_CBBC_QUEUE(trans, txq_id), 375 txq->dma_addr); 376 else 377 iwl_write_direct32(trans, 378 FH_MEM_CBBC_QUEUE(trans, txq_id), 379 txq->dma_addr >> 8); 380 iwl_pcie_txq_unmap(trans, txq_id); 381 txq->read_ptr = 0; 382 txq->write_ptr = 0; 383 } 384 385 /* Tell NIC where to find the "keep warm" buffer */ 386 iwl_write_direct32(trans, FH_KW_MEM_ADDR_REG, 387 trans_pcie->kw.dma >> 4); 388 389 /* 390 * Send 0 as the scd_base_addr since the device may have be reset 391 * while we were in WoWLAN in which case SCD_SRAM_BASE_ADDR will 392 * contain garbage. 393 */ 394 iwl_pcie_tx_start(trans, 0); 395 } 396 397 static void iwl_pcie_tx_stop_fh(struct iwl_trans *trans) 398 { 399 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans); 400 int ch, ret; 401 u32 mask = 0; 402 403 spin_lock_bh(&trans_pcie->irq_lock); 404 405 if (!iwl_trans_grab_nic_access(trans)) 406 goto out; 407 408 /* Stop each Tx DMA channel */ 409 for (ch = 0; ch < FH_TCSR_CHNL_NUM; ch++) { 410 iwl_write32(trans, FH_TCSR_CHNL_TX_CONFIG_REG(ch), 0x0); 411 mask |= FH_TSSR_TX_STATUS_REG_MSK_CHNL_IDLE(ch); 412 } 413 414 /* Wait for DMA channels to be idle */ 415 ret = iwl_poll_bit(trans, FH_TSSR_TX_STATUS_REG, mask, mask, 5000); 416 if (ret < 0) 417 IWL_ERR(trans, 418 "Failing on timeout while stopping DMA channel %d [0x%08x]\n", 419 ch, iwl_read32(trans, FH_TSSR_TX_STATUS_REG)); 420 421 iwl_trans_release_nic_access(trans); 422 423 out: 424 spin_unlock_bh(&trans_pcie->irq_lock); 425 } 426 427 /* 428 * iwl_pcie_tx_stop - Stop all Tx DMA channels 429 */ 430 int iwl_pcie_tx_stop(struct iwl_trans *trans) 431 { 432 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans); 433 int txq_id; 434 435 /* Turn off all Tx DMA fifos */ 436 iwl_scd_deactivate_fifos(trans); 437 438 /* Turn off all Tx DMA channels */ 439 iwl_pcie_tx_stop_fh(trans); 440 441 /* 442 * This function can be called before the op_mode disabled the 443 * queues. This happens when we have an rfkill interrupt. 444 * Since we stop Tx altogether - mark the queues as stopped. 445 */ 446 memset(trans->txqs.queue_stopped, 0, 447 sizeof(trans->txqs.queue_stopped)); 448 memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used)); 449 450 /* This can happen: start_hw, stop_device */ 451 if (!trans_pcie->txq_memory) 452 return 0; 453 454 /* Unmap DMA from host system and free skb's */ 455 for (txq_id = 0; txq_id < trans->trans_cfg->base_params->num_of_queues; 456 txq_id++) 457 iwl_pcie_txq_unmap(trans, txq_id); 458 459 return 0; 460 } 461 462 /* 463 * iwl_trans_tx_free - Free TXQ Context 464 * 465 * Destroy all TX DMA queues and structures 466 */ 467 void iwl_pcie_tx_free(struct iwl_trans *trans) 468 { 469 int txq_id; 470 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans); 471 472 memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used)); 473 474 /* Tx queues */ 475 if (trans_pcie->txq_memory) { 476 for (txq_id = 0; 477 txq_id < trans->trans_cfg->base_params->num_of_queues; 478 txq_id++) { 479 iwl_pcie_txq_free(trans, txq_id); 480 trans->txqs.txq[txq_id] = NULL; 481 } 482 } 483 484 kfree(trans_pcie->txq_memory); 485 trans_pcie->txq_memory = NULL; 486 487 iwl_pcie_free_dma_ptr(trans, &trans_pcie->kw); 488 489 iwl_pcie_free_dma_ptr(trans, &trans->txqs.scd_bc_tbls); 490 } 491 492 /* 493 * iwl_pcie_tx_alloc - allocate TX context 494 * Allocate all Tx DMA structures and initialize them 495 */ 496 static int iwl_pcie_tx_alloc(struct iwl_trans *trans) 497 { 498 int ret; 499 int txq_id, slots_num; 500 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans); 501 u16 bc_tbls_size = trans->trans_cfg->base_params->num_of_queues; 502 503 if (WARN_ON(trans->trans_cfg->device_family >= IWL_DEVICE_FAMILY_AX210)) 504 return -EINVAL; 505 506 bc_tbls_size *= sizeof(struct iwlagn_scd_bc_tbl); 507 508 /*It is not allowed to alloc twice, so warn when this happens. 509 * We cannot rely on the previous allocation, so free and fail */ 510 if (WARN_ON(trans_pcie->txq_memory)) { 511 ret = -EINVAL; 512 goto error; 513 } 514 515 ret = iwl_pcie_alloc_dma_ptr(trans, &trans->txqs.scd_bc_tbls, 516 bc_tbls_size); 517 if (ret) { 518 IWL_ERR(trans, "Scheduler BC Table allocation failed\n"); 519 goto error; 520 } 521 522 /* Alloc keep-warm buffer */ 523 ret = iwl_pcie_alloc_dma_ptr(trans, &trans_pcie->kw, IWL_KW_SIZE); 524 if (ret) { 525 IWL_ERR(trans, "Keep Warm allocation failed\n"); 526 goto error; 527 } 528 529 trans_pcie->txq_memory = 530 kcalloc(trans->trans_cfg->base_params->num_of_queues, 531 sizeof(struct iwl_txq), GFP_KERNEL); 532 if (!trans_pcie->txq_memory) { 533 IWL_ERR(trans, "Not enough memory for txq\n"); 534 ret = -ENOMEM; 535 goto error; 536 } 537 538 /* Alloc and init all Tx queues, including the command queue (#4/#9) */ 539 for (txq_id = 0; txq_id < trans->trans_cfg->base_params->num_of_queues; 540 txq_id++) { 541 bool cmd_queue = (txq_id == trans->txqs.cmd.q_id); 542 543 if (cmd_queue) 544 slots_num = max_t(u32, IWL_CMD_QUEUE_SIZE, 545 trans->cfg->min_txq_size); 546 else 547 slots_num = max_t(u32, IWL_DEFAULT_QUEUE_SIZE, 548 trans->cfg->min_ba_txq_size); 549 trans->txqs.txq[txq_id] = &trans_pcie->txq_memory[txq_id]; 550 ret = iwl_txq_alloc(trans, trans->txqs.txq[txq_id], slots_num, 551 cmd_queue); 552 if (ret) { 553 IWL_ERR(trans, "Tx %d queue alloc failed\n", txq_id); 554 goto error; 555 } 556 trans->txqs.txq[txq_id]->id = txq_id; 557 } 558 559 return 0; 560 561 error: 562 iwl_pcie_tx_free(trans); 563 564 return ret; 565 } 566 567 int iwl_pcie_tx_init(struct iwl_trans *trans) 568 { 569 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans); 570 int ret; 571 int txq_id, slots_num; 572 bool alloc = false; 573 574 if (!trans_pcie->txq_memory) { 575 ret = iwl_pcie_tx_alloc(trans); 576 if (ret) 577 goto error; 578 alloc = true; 579 } 580 581 spin_lock_bh(&trans_pcie->irq_lock); 582 583 /* Turn off all Tx DMA fifos */ 584 iwl_scd_deactivate_fifos(trans); 585 586 /* Tell NIC where to find the "keep warm" buffer */ 587 iwl_write_direct32(trans, FH_KW_MEM_ADDR_REG, 588 trans_pcie->kw.dma >> 4); 589 590 spin_unlock_bh(&trans_pcie->irq_lock); 591 592 /* Alloc and init all Tx queues, including the command queue (#4/#9) */ 593 for (txq_id = 0; txq_id < trans->trans_cfg->base_params->num_of_queues; 594 txq_id++) { 595 bool cmd_queue = (txq_id == trans->txqs.cmd.q_id); 596 597 if (cmd_queue) 598 slots_num = max_t(u32, IWL_CMD_QUEUE_SIZE, 599 trans->cfg->min_txq_size); 600 else 601 slots_num = max_t(u32, IWL_DEFAULT_QUEUE_SIZE, 602 trans->cfg->min_ba_txq_size); 603 ret = iwl_txq_init(trans, trans->txqs.txq[txq_id], slots_num, 604 cmd_queue); 605 if (ret) { 606 IWL_ERR(trans, "Tx %d queue init failed\n", txq_id); 607 goto error; 608 } 609 610 /* 611 * Tell nic where to find circular buffer of TFDs for a 612 * given Tx queue, and enable the DMA channel used for that 613 * queue. 614 * Circular buffer (TFD queue in DRAM) physical base address 615 */ 616 iwl_write_direct32(trans, FH_MEM_CBBC_QUEUE(trans, txq_id), 617 trans->txqs.txq[txq_id]->dma_addr >> 8); 618 } 619 620 iwl_set_bits_prph(trans, SCD_GP_CTRL, SCD_GP_CTRL_AUTO_ACTIVE_MODE); 621 if (trans->trans_cfg->base_params->num_of_queues > 20) 622 iwl_set_bits_prph(trans, SCD_GP_CTRL, 623 SCD_GP_CTRL_ENABLE_31_QUEUES); 624 625 return 0; 626 error: 627 /*Upon error, free only if we allocated something */ 628 if (alloc) 629 iwl_pcie_tx_free(trans); 630 return ret; 631 } 632 633 static int iwl_pcie_set_cmd_in_flight(struct iwl_trans *trans, 634 const struct iwl_host_cmd *cmd) 635 { 636 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans); 637 638 /* Make sure the NIC is still alive in the bus */ 639 if (test_bit(STATUS_TRANS_DEAD, &trans->status)) 640 return -ENODEV; 641 642 if (!trans->trans_cfg->base_params->apmg_wake_up_wa) 643 return 0; 644 645 /* 646 * wake up the NIC to make sure that the firmware will see the host 647 * command - we will let the NIC sleep once all the host commands 648 * returned. This needs to be done only on NICs that have 649 * apmg_wake_up_wa set (see above.) 650 */ 651 if (!_iwl_trans_pcie_grab_nic_access(trans)) 652 return -EIO; 653 654 /* 655 * In iwl_trans_grab_nic_access(), we've acquired the reg_lock. 656 * There, we also returned immediately if cmd_hold_nic_awake is 657 * already true, so it's OK to unconditionally set it to true. 658 */ 659 trans_pcie->cmd_hold_nic_awake = true; 660 spin_unlock(&trans_pcie->reg_lock); 661 662 return 0; 663 } 664 665 /* 666 * iwl_pcie_cmdq_reclaim - Reclaim TX command queue entries already Tx'd 667 * 668 * When FW advances 'R' index, all entries between old and new 'R' index 669 * need to be reclaimed. As result, some free space forms. If there is 670 * enough free space (> low mark), wake the stack that feeds us. 671 */ 672 static void iwl_pcie_cmdq_reclaim(struct iwl_trans *trans, int txq_id, int idx) 673 { 674 struct iwl_txq *txq = trans->txqs.txq[txq_id]; 675 int nfreed = 0; 676 u16 r; 677 678 lockdep_assert_held(&txq->lock); 679 680 idx = iwl_txq_get_cmd_index(txq, idx); 681 r = iwl_txq_get_cmd_index(txq, txq->read_ptr); 682 683 if (idx >= trans->trans_cfg->base_params->max_tfd_queue_size || 684 (!iwl_txq_used(txq, idx))) { 685 WARN_ONCE(test_bit(txq_id, trans->txqs.queue_used), 686 "%s: Read index for DMA queue txq id (%d), index %d is out of range [0-%d] %d %d.\n", 687 __func__, txq_id, idx, 688 trans->trans_cfg->base_params->max_tfd_queue_size, 689 txq->write_ptr, txq->read_ptr); 690 return; 691 } 692 693 for (idx = iwl_txq_inc_wrap(trans, idx); r != idx; 694 r = iwl_txq_inc_wrap(trans, r)) { 695 txq->read_ptr = iwl_txq_inc_wrap(trans, txq->read_ptr); 696 697 if (nfreed++ > 0) { 698 IWL_ERR(trans, "HCMD skipped: index (%d) %d %d\n", 699 idx, txq->write_ptr, r); 700 iwl_force_nmi(trans); 701 } 702 } 703 704 if (txq->read_ptr == txq->write_ptr) 705 iwl_pcie_clear_cmd_in_flight(trans); 706 707 iwl_txq_progress(txq); 708 } 709 710 static int iwl_pcie_txq_set_ratid_map(struct iwl_trans *trans, u16 ra_tid, 711 u16 txq_id) 712 { 713 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans); 714 u32 tbl_dw_addr; 715 u32 tbl_dw; 716 u16 scd_q2ratid; 717 718 scd_q2ratid = ra_tid & SCD_QUEUE_RA_TID_MAP_RATID_MSK; 719 720 tbl_dw_addr = trans_pcie->scd_base_addr + 721 SCD_TRANS_TBL_OFFSET_QUEUE(txq_id); 722 723 tbl_dw = iwl_trans_read_mem32(trans, tbl_dw_addr); 724 725 if (txq_id & 0x1) 726 tbl_dw = (scd_q2ratid << 16) | (tbl_dw & 0x0000FFFF); 727 else 728 tbl_dw = scd_q2ratid | (tbl_dw & 0xFFFF0000); 729 730 iwl_trans_write_mem32(trans, tbl_dw_addr, tbl_dw); 731 732 return 0; 733 } 734 735 /* Receiver address (actually, Rx station's index into station table), 736 * combined with Traffic ID (QOS priority), in format used by Tx Scheduler */ 737 #define BUILD_RAxTID(sta_id, tid) (((sta_id) << 4) + (tid)) 738 739 bool iwl_trans_pcie_txq_enable(struct iwl_trans *trans, int txq_id, u16 ssn, 740 const struct iwl_trans_txq_scd_cfg *cfg, 741 unsigned int wdg_timeout) 742 { 743 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans); 744 struct iwl_txq *txq = trans->txqs.txq[txq_id]; 745 int fifo = -1; 746 bool scd_bug = false; 747 748 if (test_and_set_bit(txq_id, trans->txqs.queue_used)) 749 WARN_ONCE(1, "queue %d already used - expect issues", txq_id); 750 751 txq->wd_timeout = msecs_to_jiffies(wdg_timeout); 752 753 if (cfg) { 754 fifo = cfg->fifo; 755 756 /* Disable the scheduler prior configuring the cmd queue */ 757 if (txq_id == trans->txqs.cmd.q_id && 758 trans_pcie->scd_set_active) 759 iwl_scd_enable_set_active(trans, 0); 760 761 /* Stop this Tx queue before configuring it */ 762 iwl_scd_txq_set_inactive(trans, txq_id); 763 764 /* Set this queue as a chain-building queue unless it is CMD */ 765 if (txq_id != trans->txqs.cmd.q_id) 766 iwl_scd_txq_set_chain(trans, txq_id); 767 768 if (cfg->aggregate) { 769 u16 ra_tid = BUILD_RAxTID(cfg->sta_id, cfg->tid); 770 771 /* Map receiver-address / traffic-ID to this queue */ 772 iwl_pcie_txq_set_ratid_map(trans, ra_tid, txq_id); 773 774 /* enable aggregations for the queue */ 775 iwl_scd_txq_enable_agg(trans, txq_id); 776 txq->ampdu = true; 777 } else { 778 /* 779 * disable aggregations for the queue, this will also 780 * make the ra_tid mapping configuration irrelevant 781 * since it is now a non-AGG queue. 782 */ 783 iwl_scd_txq_disable_agg(trans, txq_id); 784 785 ssn = txq->read_ptr; 786 } 787 } else { 788 /* 789 * If we need to move the SCD write pointer by steps of 790 * 0x40, 0x80 or 0xc0, it gets stuck. Avoids this and let 791 * the op_mode know by returning true later. 792 * Do this only in case cfg is NULL since this trick can 793 * be done only if we have DQA enabled which is true for mvm 794 * only. And mvm never sets a cfg pointer. 795 * This is really ugly, but this is the easiest way out for 796 * this sad hardware issue. 797 * This bug has been fixed on devices 9000 and up. 798 */ 799 scd_bug = !trans->trans_cfg->mq_rx_supported && 800 !((ssn - txq->write_ptr) & 0x3f) && 801 (ssn != txq->write_ptr); 802 if (scd_bug) 803 ssn++; 804 } 805 806 /* Place first TFD at index corresponding to start sequence number. 807 * Assumes that ssn_idx is valid (!= 0xFFF) */ 808 txq->read_ptr = (ssn & 0xff); 809 txq->write_ptr = (ssn & 0xff); 810 iwl_write_direct32(trans, HBUS_TARG_WRPTR, 811 (ssn & 0xff) | (txq_id << 8)); 812 813 if (cfg) { 814 u8 frame_limit = cfg->frame_limit; 815 816 iwl_write_prph(trans, SCD_QUEUE_RDPTR(txq_id), ssn); 817 818 /* Set up Tx window size and frame limit for this queue */ 819 iwl_trans_write_mem32(trans, trans_pcie->scd_base_addr + 820 SCD_CONTEXT_QUEUE_OFFSET(txq_id), 0); 821 iwl_trans_write_mem32(trans, 822 trans_pcie->scd_base_addr + 823 SCD_CONTEXT_QUEUE_OFFSET(txq_id) + sizeof(u32), 824 SCD_QUEUE_CTX_REG2_VAL(WIN_SIZE, frame_limit) | 825 SCD_QUEUE_CTX_REG2_VAL(FRAME_LIMIT, frame_limit)); 826 827 /* Set up status area in SRAM, map to Tx DMA/FIFO, activate */ 828 iwl_write_prph(trans, SCD_QUEUE_STATUS_BITS(txq_id), 829 (1 << SCD_QUEUE_STTS_REG_POS_ACTIVE) | 830 (cfg->fifo << SCD_QUEUE_STTS_REG_POS_TXF) | 831 (1 << SCD_QUEUE_STTS_REG_POS_WSL) | 832 SCD_QUEUE_STTS_REG_MSK); 833 834 /* enable the scheduler for this queue (only) */ 835 if (txq_id == trans->txqs.cmd.q_id && 836 trans_pcie->scd_set_active) 837 iwl_scd_enable_set_active(trans, BIT(txq_id)); 838 839 IWL_DEBUG_TX_QUEUES(trans, 840 "Activate queue %d on FIFO %d WrPtr: %d\n", 841 txq_id, fifo, ssn & 0xff); 842 } else { 843 IWL_DEBUG_TX_QUEUES(trans, 844 "Activate queue %d WrPtr: %d\n", 845 txq_id, ssn & 0xff); 846 } 847 848 return scd_bug; 849 } 850 851 void iwl_trans_pcie_txq_set_shared_mode(struct iwl_trans *trans, u32 txq_id, 852 bool shared_mode) 853 { 854 struct iwl_txq *txq = trans->txqs.txq[txq_id]; 855 856 txq->ampdu = !shared_mode; 857 } 858 859 void iwl_trans_pcie_txq_disable(struct iwl_trans *trans, int txq_id, 860 bool configure_scd) 861 { 862 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans); 863 u32 stts_addr = trans_pcie->scd_base_addr + 864 SCD_TX_STTS_QUEUE_OFFSET(txq_id); 865 static const u32 zero_val[4] = {}; 866 867 trans->txqs.txq[txq_id]->frozen_expiry_remainder = 0; 868 trans->txqs.txq[txq_id]->frozen = false; 869 870 /* 871 * Upon HW Rfkill - we stop the device, and then stop the queues 872 * in the op_mode. Just for the sake of the simplicity of the op_mode, 873 * allow the op_mode to call txq_disable after it already called 874 * stop_device. 875 */ 876 if (!test_and_clear_bit(txq_id, trans->txqs.queue_used)) { 877 WARN_ONCE(test_bit(STATUS_DEVICE_ENABLED, &trans->status), 878 "queue %d not used", txq_id); 879 return; 880 } 881 882 if (configure_scd) { 883 iwl_scd_txq_set_inactive(trans, txq_id); 884 885 iwl_trans_write_mem(trans, stts_addr, (const void *)zero_val, 886 ARRAY_SIZE(zero_val)); 887 } 888 889 iwl_pcie_txq_unmap(trans, txq_id); 890 trans->txqs.txq[txq_id]->ampdu = false; 891 892 IWL_DEBUG_TX_QUEUES(trans, "Deactivate queue %d\n", txq_id); 893 } 894 895 /*************** HOST COMMAND QUEUE FUNCTIONS *****/ 896 897 /* 898 * iwl_pcie_enqueue_hcmd - enqueue a uCode command 899 * @priv: device private data point 900 * @cmd: a pointer to the ucode command structure 901 * 902 * The function returns < 0 values to indicate the operation 903 * failed. On success, it returns the index (>= 0) of command in the 904 * command queue. 905 */ 906 int iwl_pcie_enqueue_hcmd(struct iwl_trans *trans, 907 struct iwl_host_cmd *cmd) 908 { 909 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; 910 struct iwl_device_cmd *out_cmd; 911 struct iwl_cmd_meta *out_meta; 912 void *dup_buf = NULL; 913 dma_addr_t phys_addr; 914 int idx; 915 u16 copy_size, cmd_size, tb0_size; 916 bool had_nocopy = false; 917 u8 group_id = iwl_cmd_groupid(cmd->id); 918 int i, ret; 919 u32 cmd_pos; 920 const u8 *cmddata[IWL_MAX_CMD_TBS_PER_TFD]; 921 u16 cmdlen[IWL_MAX_CMD_TBS_PER_TFD]; 922 unsigned long flags; 923 924 if (WARN(!trans->wide_cmd_header && 925 group_id > IWL_ALWAYS_LONG_GROUP, 926 "unsupported wide command %#x\n", cmd->id)) 927 return -EINVAL; 928 929 if (group_id != 0) { 930 copy_size = sizeof(struct iwl_cmd_header_wide); 931 cmd_size = sizeof(struct iwl_cmd_header_wide); 932 } else { 933 copy_size = sizeof(struct iwl_cmd_header); 934 cmd_size = sizeof(struct iwl_cmd_header); 935 } 936 937 /* need one for the header if the first is NOCOPY */ 938 BUILD_BUG_ON(IWL_MAX_CMD_TBS_PER_TFD > IWL_NUM_OF_TBS - 1); 939 940 for (i = 0; i < IWL_MAX_CMD_TBS_PER_TFD; i++) { 941 cmddata[i] = cmd->data[i]; 942 cmdlen[i] = cmd->len[i]; 943 944 if (!cmd->len[i]) 945 continue; 946 947 /* need at least IWL_FIRST_TB_SIZE copied */ 948 if (copy_size < IWL_FIRST_TB_SIZE) { 949 int copy = IWL_FIRST_TB_SIZE - copy_size; 950 951 if (copy > cmdlen[i]) 952 copy = cmdlen[i]; 953 cmdlen[i] -= copy; 954 cmddata[i] += copy; 955 copy_size += copy; 956 } 957 958 if (cmd->dataflags[i] & IWL_HCMD_DFL_NOCOPY) { 959 had_nocopy = true; 960 if (WARN_ON(cmd->dataflags[i] & IWL_HCMD_DFL_DUP)) { 961 idx = -EINVAL; 962 goto free_dup_buf; 963 } 964 } else if (cmd->dataflags[i] & IWL_HCMD_DFL_DUP) { 965 /* 966 * This is also a chunk that isn't copied 967 * to the static buffer so set had_nocopy. 968 */ 969 had_nocopy = true; 970 971 /* only allowed once */ 972 if (WARN_ON(dup_buf)) { 973 idx = -EINVAL; 974 goto free_dup_buf; 975 } 976 977 dup_buf = kmemdup(cmddata[i], cmdlen[i], 978 GFP_ATOMIC); 979 if (!dup_buf) 980 return -ENOMEM; 981 } else { 982 /* NOCOPY must not be followed by normal! */ 983 if (WARN_ON(had_nocopy)) { 984 idx = -EINVAL; 985 goto free_dup_buf; 986 } 987 copy_size += cmdlen[i]; 988 } 989 cmd_size += cmd->len[i]; 990 } 991 992 /* 993 * If any of the command structures end up being larger than 994 * the TFD_MAX_PAYLOAD_SIZE and they aren't dynamically 995 * allocated into separate TFDs, then we will need to 996 * increase the size of the buffers. 997 */ 998 if (WARN(copy_size > TFD_MAX_PAYLOAD_SIZE, 999 "Command %s (%#x) is too large (%d bytes)\n", 1000 iwl_get_cmd_string(trans, cmd->id), 1001 cmd->id, copy_size)) { 1002 idx = -EINVAL; 1003 goto free_dup_buf; 1004 } 1005 1006 spin_lock_irqsave(&txq->lock, flags); 1007 1008 if (iwl_txq_space(trans, txq) < ((cmd->flags & CMD_ASYNC) ? 2 : 1)) { 1009 spin_unlock_irqrestore(&txq->lock, flags); 1010 1011 IWL_ERR(trans, "No space in command queue\n"); 1012 iwl_op_mode_cmd_queue_full(trans->op_mode); 1013 idx = -ENOSPC; 1014 goto free_dup_buf; 1015 } 1016 1017 idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); 1018 out_cmd = txq->entries[idx].cmd; 1019 out_meta = &txq->entries[idx].meta; 1020 1021 memset(out_meta, 0, sizeof(*out_meta)); /* re-initialize to NULL */ 1022 if (cmd->flags & CMD_WANT_SKB) 1023 out_meta->source = cmd; 1024 1025 /* set up the header */ 1026 if (group_id != 0) { 1027 out_cmd->hdr_wide.cmd = iwl_cmd_opcode(cmd->id); 1028 out_cmd->hdr_wide.group_id = group_id; 1029 out_cmd->hdr_wide.version = iwl_cmd_version(cmd->id); 1030 out_cmd->hdr_wide.length = 1031 cpu_to_le16(cmd_size - 1032 sizeof(struct iwl_cmd_header_wide)); 1033 out_cmd->hdr_wide.reserved = 0; 1034 out_cmd->hdr_wide.sequence = 1035 cpu_to_le16(QUEUE_TO_SEQ(trans->txqs.cmd.q_id) | 1036 INDEX_TO_SEQ(txq->write_ptr)); 1037 1038 cmd_pos = sizeof(struct iwl_cmd_header_wide); 1039 copy_size = sizeof(struct iwl_cmd_header_wide); 1040 } else { 1041 out_cmd->hdr.cmd = iwl_cmd_opcode(cmd->id); 1042 out_cmd->hdr.sequence = 1043 cpu_to_le16(QUEUE_TO_SEQ(trans->txqs.cmd.q_id) | 1044 INDEX_TO_SEQ(txq->write_ptr)); 1045 out_cmd->hdr.group_id = 0; 1046 1047 cmd_pos = sizeof(struct iwl_cmd_header); 1048 copy_size = sizeof(struct iwl_cmd_header); 1049 } 1050 1051 /* and copy the data that needs to be copied */ 1052 for (i = 0; i < IWL_MAX_CMD_TBS_PER_TFD; i++) { 1053 int copy; 1054 1055 if (!cmd->len[i]) 1056 continue; 1057 1058 /* copy everything if not nocopy/dup */ 1059 if (!(cmd->dataflags[i] & (IWL_HCMD_DFL_NOCOPY | 1060 IWL_HCMD_DFL_DUP))) { 1061 copy = cmd->len[i]; 1062 1063 memcpy((u8 *)out_cmd + cmd_pos, cmd->data[i], copy); 1064 cmd_pos += copy; 1065 copy_size += copy; 1066 continue; 1067 } 1068 1069 /* 1070 * Otherwise we need at least IWL_FIRST_TB_SIZE copied 1071 * in total (for bi-directional DMA), but copy up to what 1072 * we can fit into the payload for debug dump purposes. 1073 */ 1074 copy = min_t(int, TFD_MAX_PAYLOAD_SIZE - cmd_pos, cmd->len[i]); 1075 1076 memcpy((u8 *)out_cmd + cmd_pos, cmd->data[i], copy); 1077 cmd_pos += copy; 1078 1079 /* However, treat copy_size the proper way, we need it below */ 1080 if (copy_size < IWL_FIRST_TB_SIZE) { 1081 copy = IWL_FIRST_TB_SIZE - copy_size; 1082 1083 if (copy > cmd->len[i]) 1084 copy = cmd->len[i]; 1085 copy_size += copy; 1086 } 1087 } 1088 1089 IWL_DEBUG_HC(trans, 1090 "Sending command %s (%.2x.%.2x), seq: 0x%04X, %d bytes at %d[%d]:%d\n", 1091 iwl_get_cmd_string(trans, cmd->id), 1092 group_id, out_cmd->hdr.cmd, 1093 le16_to_cpu(out_cmd->hdr.sequence), 1094 cmd_size, txq->write_ptr, idx, trans->txqs.cmd.q_id); 1095 1096 /* start the TFD with the minimum copy bytes */ 1097 tb0_size = min_t(int, copy_size, IWL_FIRST_TB_SIZE); 1098 memcpy(&txq->first_tb_bufs[idx], &out_cmd->hdr, tb0_size); 1099 iwl_pcie_txq_build_tfd(trans, txq, 1100 iwl_txq_get_first_tb_dma(txq, idx), 1101 tb0_size, true); 1102 1103 /* map first command fragment, if any remains */ 1104 if (copy_size > tb0_size) { 1105 phys_addr = dma_map_single(trans->dev, 1106 ((u8 *)&out_cmd->hdr) + tb0_size, 1107 copy_size - tb0_size, 1108 DMA_TO_DEVICE); 1109 if (dma_mapping_error(trans->dev, phys_addr)) { 1110 iwl_txq_gen1_tfd_unmap(trans, out_meta, txq, 1111 txq->write_ptr); 1112 idx = -ENOMEM; 1113 goto out; 1114 } 1115 1116 iwl_pcie_txq_build_tfd(trans, txq, phys_addr, 1117 copy_size - tb0_size, false); 1118 } 1119 1120 /* map the remaining (adjusted) nocopy/dup fragments */ 1121 for (i = 0; i < IWL_MAX_CMD_TBS_PER_TFD; i++) { 1122 void *data = (void *)(uintptr_t)cmddata[i]; 1123 1124 if (!cmdlen[i]) 1125 continue; 1126 if (!(cmd->dataflags[i] & (IWL_HCMD_DFL_NOCOPY | 1127 IWL_HCMD_DFL_DUP))) 1128 continue; 1129 if (cmd->dataflags[i] & IWL_HCMD_DFL_DUP) 1130 data = dup_buf; 1131 phys_addr = dma_map_single(trans->dev, data, 1132 cmdlen[i], DMA_TO_DEVICE); 1133 if (dma_mapping_error(trans->dev, phys_addr)) { 1134 iwl_txq_gen1_tfd_unmap(trans, out_meta, txq, 1135 txq->write_ptr); 1136 idx = -ENOMEM; 1137 goto out; 1138 } 1139 1140 iwl_pcie_txq_build_tfd(trans, txq, phys_addr, cmdlen[i], false); 1141 } 1142 1143 BUILD_BUG_ON(IWL_TFH_NUM_TBS > sizeof(out_meta->tbs) * BITS_PER_BYTE); 1144 out_meta->flags = cmd->flags; 1145 if (WARN_ON_ONCE(txq->entries[idx].free_buf)) 1146 kfree_sensitive(txq->entries[idx].free_buf); 1147 txq->entries[idx].free_buf = dup_buf; 1148 1149 trace_iwlwifi_dev_hcmd(trans->dev, cmd, cmd_size, &out_cmd->hdr_wide); 1150 1151 /* start timer if queue currently empty */ 1152 if (txq->read_ptr == txq->write_ptr && txq->wd_timeout) 1153 mod_timer(&txq->stuck_timer, jiffies + txq->wd_timeout); 1154 1155 ret = iwl_pcie_set_cmd_in_flight(trans, cmd); 1156 if (ret < 0) { 1157 idx = ret; 1158 goto out; 1159 } 1160 1161 /* Increment and update queue's write index */ 1162 txq->write_ptr = iwl_txq_inc_wrap(trans, txq->write_ptr); 1163 iwl_pcie_txq_inc_wr_ptr(trans, txq); 1164 1165 out: 1166 spin_unlock_irqrestore(&txq->lock, flags); 1167 free_dup_buf: 1168 if (idx < 0) 1169 kfree(dup_buf); 1170 return idx; 1171 } 1172 1173 /* 1174 * iwl_pcie_hcmd_complete - Pull unused buffers off the queue and reclaim them 1175 * @rxb: Rx buffer to reclaim 1176 */ 1177 void iwl_pcie_hcmd_complete(struct iwl_trans *trans, 1178 struct iwl_rx_cmd_buffer *rxb) 1179 { 1180 struct iwl_rx_packet *pkt = rxb_addr(rxb); 1181 u16 sequence = le16_to_cpu(pkt->hdr.sequence); 1182 u8 group_id; 1183 u32 cmd_id; 1184 int txq_id = SEQ_TO_QUEUE(sequence); 1185 int index = SEQ_TO_INDEX(sequence); 1186 int cmd_index; 1187 struct iwl_device_cmd *cmd; 1188 struct iwl_cmd_meta *meta; 1189 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans); 1190 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; 1191 1192 /* If a Tx command is being handled and it isn't in the actual 1193 * command queue then there a command routing bug has been introduced 1194 * in the queue management code. */ 1195 if (WARN(txq_id != trans->txqs.cmd.q_id, 1196 "wrong command queue %d (should be %d), sequence 0x%X readp=%d writep=%d\n", 1197 txq_id, trans->txqs.cmd.q_id, sequence, txq->read_ptr, 1198 txq->write_ptr)) { 1199 iwl_print_hex_error(trans, pkt, 32); 1200 return; 1201 } 1202 1203 spin_lock_bh(&txq->lock); 1204 1205 cmd_index = iwl_txq_get_cmd_index(txq, index); 1206 cmd = txq->entries[cmd_index].cmd; 1207 meta = &txq->entries[cmd_index].meta; 1208 group_id = cmd->hdr.group_id; 1209 cmd_id = WIDE_ID(group_id, cmd->hdr.cmd); 1210 1211 iwl_txq_gen1_tfd_unmap(trans, meta, txq, index); 1212 1213 /* Input error checking is done when commands are added to queue. */ 1214 if (meta->flags & CMD_WANT_SKB) { 1215 struct page *p = rxb_steal_page(rxb); 1216 1217 meta->source->resp_pkt = pkt; 1218 #if defined(__linux__) 1219 meta->source->_rx_page_addr = (unsigned long)page_address(p); 1220 #elif defined(__FreeBSD__) 1221 meta->source->_page = p; 1222 #endif 1223 meta->source->_rx_page_order = trans_pcie->rx_page_order; 1224 } 1225 1226 if (meta->flags & CMD_WANT_ASYNC_CALLBACK) 1227 iwl_op_mode_async_cb(trans->op_mode, cmd); 1228 1229 iwl_pcie_cmdq_reclaim(trans, txq_id, index); 1230 1231 if (!(meta->flags & CMD_ASYNC)) { 1232 if (!test_bit(STATUS_SYNC_HCMD_ACTIVE, &trans->status)) { 1233 IWL_WARN(trans, 1234 "HCMD_ACTIVE already clear for command %s\n", 1235 iwl_get_cmd_string(trans, cmd_id)); 1236 } 1237 clear_bit(STATUS_SYNC_HCMD_ACTIVE, &trans->status); 1238 IWL_DEBUG_INFO(trans, "Clearing HCMD_ACTIVE for command %s\n", 1239 iwl_get_cmd_string(trans, cmd_id)); 1240 wake_up(&trans->wait_command_queue); 1241 } 1242 1243 meta->flags = 0; 1244 1245 spin_unlock_bh(&txq->lock); 1246 } 1247 1248 static int iwl_fill_data_tbs(struct iwl_trans *trans, struct sk_buff *skb, 1249 struct iwl_txq *txq, u8 hdr_len, 1250 struct iwl_cmd_meta *out_meta) 1251 { 1252 u16 head_tb_len; 1253 int i; 1254 1255 /* 1256 * Set up TFD's third entry to point directly to remainder 1257 * of skb's head, if any 1258 */ 1259 head_tb_len = skb_headlen(skb) - hdr_len; 1260 1261 if (head_tb_len > 0) { 1262 dma_addr_t tb_phys = dma_map_single(trans->dev, 1263 skb->data + hdr_len, 1264 head_tb_len, DMA_TO_DEVICE); 1265 if (unlikely(dma_mapping_error(trans->dev, tb_phys))) 1266 return -EINVAL; 1267 trace_iwlwifi_dev_tx_tb(trans->dev, skb, skb->data + hdr_len, 1268 tb_phys, head_tb_len); 1269 iwl_pcie_txq_build_tfd(trans, txq, tb_phys, head_tb_len, false); 1270 } 1271 1272 /* set up the remaining entries to point to the data */ 1273 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { 1274 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; 1275 dma_addr_t tb_phys; 1276 int tb_idx; 1277 1278 if (!skb_frag_size(frag)) 1279 continue; 1280 1281 tb_phys = skb_frag_dma_map(trans->dev, frag, 0, 1282 skb_frag_size(frag), DMA_TO_DEVICE); 1283 1284 if (unlikely(dma_mapping_error(trans->dev, tb_phys))) 1285 return -EINVAL; 1286 trace_iwlwifi_dev_tx_tb(trans->dev, skb, skb_frag_address(frag), 1287 tb_phys, skb_frag_size(frag)); 1288 tb_idx = iwl_pcie_txq_build_tfd(trans, txq, tb_phys, 1289 skb_frag_size(frag), false); 1290 if (tb_idx < 0) 1291 return tb_idx; 1292 1293 out_meta->tbs |= BIT(tb_idx); 1294 } 1295 1296 return 0; 1297 } 1298 1299 #ifdef CONFIG_INET 1300 static int iwl_fill_data_tbs_amsdu(struct iwl_trans *trans, struct sk_buff *skb, 1301 struct iwl_txq *txq, u8 hdr_len, 1302 struct iwl_cmd_meta *out_meta, 1303 struct iwl_device_tx_cmd *dev_cmd, 1304 u16 tb1_len) 1305 { 1306 struct iwl_tx_cmd *tx_cmd = (void *)dev_cmd->payload; 1307 struct ieee80211_hdr *hdr = (void *)skb->data; 1308 unsigned int snap_ip_tcp_hdrlen, ip_hdrlen, total_len, hdr_room; 1309 unsigned int mss = skb_shinfo(skb)->gso_size; 1310 u16 length, iv_len, amsdu_pad; 1311 u8 *start_hdr; 1312 struct iwl_tso_hdr_page *hdr_page; 1313 struct tso_t tso; 1314 1315 /* if the packet is protected, then it must be CCMP or GCMP */ 1316 BUILD_BUG_ON(IEEE80211_CCMP_HDR_LEN != IEEE80211_GCMP_HDR_LEN); 1317 iv_len = ieee80211_has_protected(hdr->frame_control) ? 1318 IEEE80211_CCMP_HDR_LEN : 0; 1319 1320 trace_iwlwifi_dev_tx(trans->dev, skb, 1321 iwl_txq_get_tfd(trans, txq, txq->write_ptr), 1322 trans->txqs.tfd.size, 1323 &dev_cmd->hdr, IWL_FIRST_TB_SIZE + tb1_len, 0); 1324 1325 ip_hdrlen = skb_transport_header(skb) - skb_network_header(skb); 1326 snap_ip_tcp_hdrlen = 8 + ip_hdrlen + tcp_hdrlen(skb); 1327 total_len = skb->len - snap_ip_tcp_hdrlen - hdr_len - iv_len; 1328 amsdu_pad = 0; 1329 1330 /* total amount of header we may need for this A-MSDU */ 1331 hdr_room = DIV_ROUND_UP(total_len, mss) * 1332 (3 + snap_ip_tcp_hdrlen + sizeof(struct ethhdr)) + iv_len; 1333 1334 /* Our device supports 9 segments at most, it will fit in 1 page */ 1335 hdr_page = get_page_hdr(trans, hdr_room, skb); 1336 if (!hdr_page) 1337 return -ENOMEM; 1338 1339 start_hdr = hdr_page->pos; 1340 memcpy(hdr_page->pos, skb->data + hdr_len, iv_len); 1341 hdr_page->pos += iv_len; 1342 1343 /* 1344 * Pull the ieee80211 header + IV to be able to use TSO core, 1345 * we will restore it for the tx_status flow. 1346 */ 1347 skb_pull(skb, hdr_len + iv_len); 1348 1349 /* 1350 * Remove the length of all the headers that we don't actually 1351 * have in the MPDU by themselves, but that we duplicate into 1352 * all the different MSDUs inside the A-MSDU. 1353 */ 1354 le16_add_cpu(&tx_cmd->len, -snap_ip_tcp_hdrlen); 1355 1356 tso_start(skb, &tso); 1357 1358 while (total_len) { 1359 /* this is the data left for this subframe */ 1360 unsigned int data_left = 1361 min_t(unsigned int, mss, total_len); 1362 unsigned int hdr_tb_len; 1363 dma_addr_t hdr_tb_phys; 1364 u8 *subf_hdrs_start = hdr_page->pos; 1365 1366 total_len -= data_left; 1367 1368 memset(hdr_page->pos, 0, amsdu_pad); 1369 hdr_page->pos += amsdu_pad; 1370 amsdu_pad = (4 - (sizeof(struct ethhdr) + snap_ip_tcp_hdrlen + 1371 data_left)) & 0x3; 1372 ether_addr_copy(hdr_page->pos, ieee80211_get_DA(hdr)); 1373 hdr_page->pos += ETH_ALEN; 1374 ether_addr_copy(hdr_page->pos, ieee80211_get_SA(hdr)); 1375 hdr_page->pos += ETH_ALEN; 1376 1377 length = snap_ip_tcp_hdrlen + data_left; 1378 *((__be16 *)hdr_page->pos) = cpu_to_be16(length); 1379 hdr_page->pos += sizeof(length); 1380 1381 /* 1382 * This will copy the SNAP as well which will be considered 1383 * as MAC header. 1384 */ 1385 tso_build_hdr(skb, hdr_page->pos, &tso, data_left, !total_len); 1386 1387 hdr_page->pos += snap_ip_tcp_hdrlen; 1388 1389 hdr_tb_len = hdr_page->pos - start_hdr; 1390 hdr_tb_phys = dma_map_single(trans->dev, start_hdr, 1391 hdr_tb_len, DMA_TO_DEVICE); 1392 if (unlikely(dma_mapping_error(trans->dev, hdr_tb_phys))) 1393 return -EINVAL; 1394 iwl_pcie_txq_build_tfd(trans, txq, hdr_tb_phys, 1395 hdr_tb_len, false); 1396 trace_iwlwifi_dev_tx_tb(trans->dev, skb, start_hdr, 1397 hdr_tb_phys, hdr_tb_len); 1398 /* add this subframe's headers' length to the tx_cmd */ 1399 le16_add_cpu(&tx_cmd->len, hdr_page->pos - subf_hdrs_start); 1400 1401 /* prepare the start_hdr for the next subframe */ 1402 start_hdr = hdr_page->pos; 1403 1404 /* put the payload */ 1405 while (data_left) { 1406 unsigned int size = min_t(unsigned int, tso.size, 1407 data_left); 1408 dma_addr_t tb_phys; 1409 1410 tb_phys = dma_map_single(trans->dev, tso.data, 1411 size, DMA_TO_DEVICE); 1412 if (unlikely(dma_mapping_error(trans->dev, tb_phys))) 1413 return -EINVAL; 1414 1415 iwl_pcie_txq_build_tfd(trans, txq, tb_phys, 1416 size, false); 1417 trace_iwlwifi_dev_tx_tb(trans->dev, skb, tso.data, 1418 tb_phys, size); 1419 1420 data_left -= size; 1421 tso_build_data(skb, &tso, size); 1422 } 1423 } 1424 1425 /* re -add the WiFi header and IV */ 1426 skb_push(skb, hdr_len + iv_len); 1427 1428 return 0; 1429 } 1430 #else /* CONFIG_INET */ 1431 static int iwl_fill_data_tbs_amsdu(struct iwl_trans *trans, struct sk_buff *skb, 1432 struct iwl_txq *txq, u8 hdr_len, 1433 struct iwl_cmd_meta *out_meta, 1434 struct iwl_device_tx_cmd *dev_cmd, 1435 u16 tb1_len) 1436 { 1437 /* No A-MSDU without CONFIG_INET */ 1438 WARN_ON(1); 1439 1440 return -1; 1441 } 1442 #endif /* CONFIG_INET */ 1443 1444 int iwl_trans_pcie_tx(struct iwl_trans *trans, struct sk_buff *skb, 1445 struct iwl_device_tx_cmd *dev_cmd, int txq_id) 1446 { 1447 struct ieee80211_hdr *hdr; 1448 struct iwl_tx_cmd *tx_cmd = (struct iwl_tx_cmd *)dev_cmd->payload; 1449 struct iwl_cmd_meta *out_meta; 1450 struct iwl_txq *txq; 1451 dma_addr_t tb0_phys, tb1_phys, scratch_phys; 1452 void *tb1_addr; 1453 void *tfd; 1454 u16 len, tb1_len; 1455 bool wait_write_ptr; 1456 __le16 fc; 1457 u8 hdr_len; 1458 u16 wifi_seq; 1459 bool amsdu; 1460 1461 txq = trans->txqs.txq[txq_id]; 1462 1463 if (WARN_ONCE(!test_bit(txq_id, trans->txqs.queue_used), 1464 "TX on unused queue %d\n", txq_id)) 1465 return -EINVAL; 1466 1467 if (skb_is_nonlinear(skb) && 1468 skb_shinfo(skb)->nr_frags > IWL_TRANS_MAX_FRAGS(trans) && 1469 __skb_linearize(skb)) 1470 return -ENOMEM; 1471 1472 /* mac80211 always puts the full header into the SKB's head, 1473 * so there's no need to check if it's readable there 1474 */ 1475 hdr = (struct ieee80211_hdr *)skb->data; 1476 fc = hdr->frame_control; 1477 hdr_len = ieee80211_hdrlen(fc); 1478 1479 spin_lock(&txq->lock); 1480 1481 if (iwl_txq_space(trans, txq) < txq->high_mark) { 1482 iwl_txq_stop(trans, txq); 1483 1484 /* don't put the packet on the ring, if there is no room */ 1485 if (unlikely(iwl_txq_space(trans, txq) < 3)) { 1486 struct iwl_device_tx_cmd **dev_cmd_ptr; 1487 1488 dev_cmd_ptr = (void *)((u8 *)skb->cb + 1489 trans->txqs.dev_cmd_offs); 1490 1491 *dev_cmd_ptr = dev_cmd; 1492 __skb_queue_tail(&txq->overflow_q, skb); 1493 1494 spin_unlock(&txq->lock); 1495 return 0; 1496 } 1497 } 1498 1499 /* In AGG mode, the index in the ring must correspond to the WiFi 1500 * sequence number. This is a HW requirements to help the SCD to parse 1501 * the BA. 1502 * Check here that the packets are in the right place on the ring. 1503 */ 1504 wifi_seq = IEEE80211_SEQ_TO_SN(le16_to_cpu(hdr->seq_ctrl)); 1505 WARN_ONCE(txq->ampdu && 1506 (wifi_seq & 0xff) != txq->write_ptr, 1507 "Q: %d WiFi Seq %d tfdNum %d", 1508 txq_id, wifi_seq, txq->write_ptr); 1509 1510 /* Set up driver data for this TFD */ 1511 txq->entries[txq->write_ptr].skb = skb; 1512 txq->entries[txq->write_ptr].cmd = dev_cmd; 1513 1514 dev_cmd->hdr.sequence = 1515 cpu_to_le16((u16)(QUEUE_TO_SEQ(txq_id) | 1516 INDEX_TO_SEQ(txq->write_ptr))); 1517 1518 tb0_phys = iwl_txq_get_first_tb_dma(txq, txq->write_ptr); 1519 scratch_phys = tb0_phys + sizeof(struct iwl_cmd_header) + 1520 offsetof(struct iwl_tx_cmd, scratch); 1521 1522 tx_cmd->dram_lsb_ptr = cpu_to_le32(scratch_phys); 1523 tx_cmd->dram_msb_ptr = iwl_get_dma_hi_addr(scratch_phys); 1524 1525 /* Set up first empty entry in queue's array of Tx/cmd buffers */ 1526 out_meta = &txq->entries[txq->write_ptr].meta; 1527 out_meta->flags = 0; 1528 1529 /* 1530 * The second TB (tb1) points to the remainder of the TX command 1531 * and the 802.11 header - dword aligned size 1532 * (This calculation modifies the TX command, so do it before the 1533 * setup of the first TB) 1534 */ 1535 len = sizeof(struct iwl_tx_cmd) + sizeof(struct iwl_cmd_header) + 1536 hdr_len - IWL_FIRST_TB_SIZE; 1537 /* do not align A-MSDU to dword as the subframe header aligns it */ 1538 amsdu = ieee80211_is_data_qos(fc) && 1539 (*ieee80211_get_qos_ctl(hdr) & 1540 IEEE80211_QOS_CTL_A_MSDU_PRESENT); 1541 if (!amsdu) { 1542 tb1_len = ALIGN(len, 4); 1543 /* Tell NIC about any 2-byte padding after MAC header */ 1544 if (tb1_len != len) 1545 tx_cmd->tx_flags |= cpu_to_le32(TX_CMD_FLG_MH_PAD); 1546 } else { 1547 tb1_len = len; 1548 } 1549 1550 /* 1551 * The first TB points to bi-directional DMA data, we'll 1552 * memcpy the data into it later. 1553 */ 1554 iwl_pcie_txq_build_tfd(trans, txq, tb0_phys, 1555 IWL_FIRST_TB_SIZE, true); 1556 1557 /* there must be data left over for TB1 or this code must be changed */ 1558 BUILD_BUG_ON(sizeof(struct iwl_tx_cmd) < IWL_FIRST_TB_SIZE); 1559 1560 /* map the data for TB1 */ 1561 tb1_addr = ((u8 *)&dev_cmd->hdr) + IWL_FIRST_TB_SIZE; 1562 tb1_phys = dma_map_single(trans->dev, tb1_addr, tb1_len, DMA_TO_DEVICE); 1563 if (unlikely(dma_mapping_error(trans->dev, tb1_phys))) 1564 goto out_err; 1565 iwl_pcie_txq_build_tfd(trans, txq, tb1_phys, tb1_len, false); 1566 1567 trace_iwlwifi_dev_tx(trans->dev, skb, 1568 iwl_txq_get_tfd(trans, txq, txq->write_ptr), 1569 trans->txqs.tfd.size, 1570 &dev_cmd->hdr, IWL_FIRST_TB_SIZE + tb1_len, 1571 hdr_len); 1572 1573 /* 1574 * If gso_size wasn't set, don't give the frame "amsdu treatment" 1575 * (adding subframes, etc.). 1576 * This can happen in some testing flows when the amsdu was already 1577 * pre-built, and we just need to send the resulting skb. 1578 */ 1579 if (amsdu && skb_shinfo(skb)->gso_size) { 1580 if (unlikely(iwl_fill_data_tbs_amsdu(trans, skb, txq, hdr_len, 1581 out_meta, dev_cmd, 1582 tb1_len))) 1583 goto out_err; 1584 } else { 1585 struct sk_buff *frag; 1586 1587 if (unlikely(iwl_fill_data_tbs(trans, skb, txq, hdr_len, 1588 out_meta))) 1589 goto out_err; 1590 1591 skb_walk_frags(skb, frag) { 1592 if (unlikely(iwl_fill_data_tbs(trans, frag, txq, 0, 1593 out_meta))) 1594 goto out_err; 1595 } 1596 } 1597 1598 /* building the A-MSDU might have changed this data, so memcpy it now */ 1599 memcpy(&txq->first_tb_bufs[txq->write_ptr], dev_cmd, IWL_FIRST_TB_SIZE); 1600 1601 tfd = iwl_txq_get_tfd(trans, txq, txq->write_ptr); 1602 /* Set up entry for this TFD in Tx byte-count array */ 1603 iwl_txq_gen1_update_byte_cnt_tbl(trans, txq, le16_to_cpu(tx_cmd->len), 1604 iwl_txq_gen1_tfd_get_num_tbs(trans, 1605 tfd)); 1606 1607 wait_write_ptr = ieee80211_has_morefrags(fc); 1608 1609 /* start timer if queue currently empty */ 1610 if (txq->read_ptr == txq->write_ptr && txq->wd_timeout) { 1611 /* 1612 * If the TXQ is active, then set the timer, if not, 1613 * set the timer in remainder so that the timer will 1614 * be armed with the right value when the station will 1615 * wake up. 1616 */ 1617 if (!txq->frozen) 1618 mod_timer(&txq->stuck_timer, 1619 jiffies + txq->wd_timeout); 1620 else 1621 txq->frozen_expiry_remainder = txq->wd_timeout; 1622 } 1623 1624 /* Tell device the write index *just past* this latest filled TFD */ 1625 txq->write_ptr = iwl_txq_inc_wrap(trans, txq->write_ptr); 1626 if (!wait_write_ptr) 1627 iwl_pcie_txq_inc_wr_ptr(trans, txq); 1628 1629 /* 1630 * At this point the frame is "transmitted" successfully 1631 * and we will get a TX status notification eventually. 1632 */ 1633 spin_unlock(&txq->lock); 1634 return 0; 1635 out_err: 1636 iwl_txq_gen1_tfd_unmap(trans, out_meta, txq, txq->write_ptr); 1637 spin_unlock(&txq->lock); 1638 return -1; 1639 } 1640