1 // SPDX-License-Identifier: BSD-3-Clause-Clear 2 /* 3 * Copyright (c) 2018-2019 The Linux Foundation. All rights reserved. 4 * Copyright (c) 2022-2023 Qualcomm Innovation Center, Inc. All rights reserved. 5 */ 6 7 #include <linux/module.h> 8 #include <linux/platform_device.h> 9 #include <linux/of_device.h> 10 #include <linux/of.h> 11 #include <linux/dma-mapping.h> 12 #include <linux/of_address.h> 13 #include <linux/iommu.h> 14 #include "ahb.h" 15 #include "debug.h" 16 #include "hif.h" 17 #include "qmi.h" 18 #include <linux/remoteproc.h> 19 #include "pcic.h" 20 #include <linux/soc/qcom/smem.h> 21 #include <linux/soc/qcom/smem_state.h> 22 23 static const struct of_device_id ath11k_ahb_of_match[] = { 24 /* TODO: Should we change the compatible string to something similar 25 * to one that ath10k uses? 26 */ 27 { .compatible = "qcom,ipq8074-wifi", 28 .data = (void *)ATH11K_HW_IPQ8074, 29 }, 30 { .compatible = "qcom,ipq6018-wifi", 31 .data = (void *)ATH11K_HW_IPQ6018_HW10, 32 }, 33 { .compatible = "qcom,wcn6750-wifi", 34 .data = (void *)ATH11K_HW_WCN6750_HW10, 35 }, 36 { .compatible = "qcom,ipq5018-wifi", 37 .data = (void *)ATH11K_HW_IPQ5018_HW10, 38 }, 39 { } 40 }; 41 42 MODULE_DEVICE_TABLE(of, ath11k_ahb_of_match); 43 44 #define ATH11K_IRQ_CE0_OFFSET 4 45 46 static const char *irq_name[ATH11K_IRQ_NUM_MAX] = { 47 "misc-pulse1", 48 "misc-latch", 49 "sw-exception", 50 "watchdog", 51 "ce0", 52 "ce1", 53 "ce2", 54 "ce3", 55 "ce4", 56 "ce5", 57 "ce6", 58 "ce7", 59 "ce8", 60 "ce9", 61 "ce10", 62 "ce11", 63 "host2wbm-desc-feed", 64 "host2reo-re-injection", 65 "host2reo-command", 66 "host2rxdma-monitor-ring3", 67 "host2rxdma-monitor-ring2", 68 "host2rxdma-monitor-ring1", 69 "reo2ost-exception", 70 "wbm2host-rx-release", 71 "reo2host-status", 72 "reo2host-destination-ring4", 73 "reo2host-destination-ring3", 74 "reo2host-destination-ring2", 75 "reo2host-destination-ring1", 76 "rxdma2host-monitor-destination-mac3", 77 "rxdma2host-monitor-destination-mac2", 78 "rxdma2host-monitor-destination-mac1", 79 "ppdu-end-interrupts-mac3", 80 "ppdu-end-interrupts-mac2", 81 "ppdu-end-interrupts-mac1", 82 "rxdma2host-monitor-status-ring-mac3", 83 "rxdma2host-monitor-status-ring-mac2", 84 "rxdma2host-monitor-status-ring-mac1", 85 "host2rxdma-host-buf-ring-mac3", 86 "host2rxdma-host-buf-ring-mac2", 87 "host2rxdma-host-buf-ring-mac1", 88 "rxdma2host-destination-ring-mac3", 89 "rxdma2host-destination-ring-mac2", 90 "rxdma2host-destination-ring-mac1", 91 "host2tcl-input-ring4", 92 "host2tcl-input-ring3", 93 "host2tcl-input-ring2", 94 "host2tcl-input-ring1", 95 "wbm2host-tx-completions-ring3", 96 "wbm2host-tx-completions-ring2", 97 "wbm2host-tx-completions-ring1", 98 "tcl2host-status-ring", 99 }; 100 101 /* enum ext_irq_num - irq numbers that can be used by external modules 102 * like datapath 103 */ 104 enum ext_irq_num { 105 host2wbm_desc_feed = 16, 106 host2reo_re_injection, 107 host2reo_command, 108 host2rxdma_monitor_ring3, 109 host2rxdma_monitor_ring2, 110 host2rxdma_monitor_ring1, 111 reo2host_exception, 112 wbm2host_rx_release, 113 reo2host_status, 114 reo2host_destination_ring4, 115 reo2host_destination_ring3, 116 reo2host_destination_ring2, 117 reo2host_destination_ring1, 118 rxdma2host_monitor_destination_mac3, 119 rxdma2host_monitor_destination_mac2, 120 rxdma2host_monitor_destination_mac1, 121 ppdu_end_interrupts_mac3, 122 ppdu_end_interrupts_mac2, 123 ppdu_end_interrupts_mac1, 124 rxdma2host_monitor_status_ring_mac3, 125 rxdma2host_monitor_status_ring_mac2, 126 rxdma2host_monitor_status_ring_mac1, 127 host2rxdma_host_buf_ring_mac3, 128 host2rxdma_host_buf_ring_mac2, 129 host2rxdma_host_buf_ring_mac1, 130 rxdma2host_destination_ring_mac3, 131 rxdma2host_destination_ring_mac2, 132 rxdma2host_destination_ring_mac1, 133 host2tcl_input_ring4, 134 host2tcl_input_ring3, 135 host2tcl_input_ring2, 136 host2tcl_input_ring1, 137 wbm2host_tx_completions_ring3, 138 wbm2host_tx_completions_ring2, 139 wbm2host_tx_completions_ring1, 140 tcl2host_status_ring, 141 }; 142 143 static int 144 ath11k_ahb_get_msi_irq_wcn6750(struct ath11k_base *ab, unsigned int vector) 145 { 146 return ab->pci.msi.irqs[vector]; 147 } 148 149 static inline u32 150 ath11k_ahb_get_window_start_wcn6750(struct ath11k_base *ab, u32 offset) 151 { 152 u32 window_start = 0; 153 154 /* If offset lies within DP register range, use 1st window */ 155 if ((offset ^ HAL_SEQ_WCSS_UMAC_OFFSET) < ATH11K_PCI_WINDOW_RANGE_MASK) 156 window_start = ATH11K_PCI_WINDOW_START; 157 /* If offset lies within CE register range, use 2nd window */ 158 else if ((offset ^ HAL_SEQ_WCSS_UMAC_CE0_SRC_REG(ab)) < 159 ATH11K_PCI_WINDOW_RANGE_MASK) 160 window_start = 2 * ATH11K_PCI_WINDOW_START; 161 162 return window_start; 163 } 164 165 static void 166 ath11k_ahb_window_write32_wcn6750(struct ath11k_base *ab, u32 offset, u32 value) 167 { 168 u32 window_start; 169 170 /* WCN6750 uses static window based register access*/ 171 window_start = ath11k_ahb_get_window_start_wcn6750(ab, offset); 172 173 iowrite32(value, ab->mem + window_start + 174 (offset & ATH11K_PCI_WINDOW_RANGE_MASK)); 175 } 176 177 static u32 ath11k_ahb_window_read32_wcn6750(struct ath11k_base *ab, u32 offset) 178 { 179 u32 window_start; 180 u32 val; 181 182 /* WCN6750 uses static window based register access */ 183 window_start = ath11k_ahb_get_window_start_wcn6750(ab, offset); 184 185 val = ioread32(ab->mem + window_start + 186 (offset & ATH11K_PCI_WINDOW_RANGE_MASK)); 187 return val; 188 } 189 190 static const struct ath11k_pci_ops ath11k_ahb_pci_ops_wcn6750 = { 191 .wakeup = NULL, 192 .release = NULL, 193 .get_msi_irq = ath11k_ahb_get_msi_irq_wcn6750, 194 .window_write32 = ath11k_ahb_window_write32_wcn6750, 195 .window_read32 = ath11k_ahb_window_read32_wcn6750, 196 }; 197 198 static inline u32 ath11k_ahb_read32(struct ath11k_base *ab, u32 offset) 199 { 200 return ioread32(ab->mem + offset); 201 } 202 203 static inline void ath11k_ahb_write32(struct ath11k_base *ab, u32 offset, u32 value) 204 { 205 iowrite32(value, ab->mem + offset); 206 } 207 208 static void ath11k_ahb_kill_tasklets(struct ath11k_base *ab) 209 { 210 int i; 211 212 for (i = 0; i < ab->hw_params.ce_count; i++) { 213 struct ath11k_ce_pipe *ce_pipe = &ab->ce.ce_pipe[i]; 214 215 if (ath11k_ce_get_attr_flags(ab, i) & CE_ATTR_DIS_INTR) 216 continue; 217 218 tasklet_kill(&ce_pipe->intr_tq); 219 } 220 } 221 222 static void ath11k_ahb_ext_grp_disable(struct ath11k_ext_irq_grp *irq_grp) 223 { 224 int i; 225 226 for (i = 0; i < irq_grp->num_irq; i++) 227 disable_irq_nosync(irq_grp->ab->irq_num[irq_grp->irqs[i]]); 228 } 229 230 static void __ath11k_ahb_ext_irq_disable(struct ath11k_base *ab) 231 { 232 int i; 233 234 for (i = 0; i < ATH11K_EXT_IRQ_GRP_NUM_MAX; i++) { 235 struct ath11k_ext_irq_grp *irq_grp = &ab->ext_irq_grp[i]; 236 237 ath11k_ahb_ext_grp_disable(irq_grp); 238 239 if (irq_grp->napi_enabled) { 240 napi_synchronize(&irq_grp->napi); 241 napi_disable(&irq_grp->napi); 242 irq_grp->napi_enabled = false; 243 } 244 } 245 } 246 247 static void ath11k_ahb_ext_grp_enable(struct ath11k_ext_irq_grp *irq_grp) 248 { 249 int i; 250 251 for (i = 0; i < irq_grp->num_irq; i++) 252 enable_irq(irq_grp->ab->irq_num[irq_grp->irqs[i]]); 253 } 254 255 static void ath11k_ahb_setbit32(struct ath11k_base *ab, u8 bit, u32 offset) 256 { 257 u32 val; 258 259 val = ath11k_ahb_read32(ab, offset); 260 ath11k_ahb_write32(ab, offset, val | BIT(bit)); 261 } 262 263 static void ath11k_ahb_clearbit32(struct ath11k_base *ab, u8 bit, u32 offset) 264 { 265 u32 val; 266 267 val = ath11k_ahb_read32(ab, offset); 268 ath11k_ahb_write32(ab, offset, val & ~BIT(bit)); 269 } 270 271 static void ath11k_ahb_ce_irq_enable(struct ath11k_base *ab, u16 ce_id) 272 { 273 const struct ce_attr *ce_attr; 274 const struct ce_ie_addr *ce_ie_addr = ab->hw_params.ce_ie_addr; 275 u32 ie1_reg_addr, ie2_reg_addr, ie3_reg_addr; 276 277 ie1_reg_addr = ce_ie_addr->ie1_reg_addr + ATH11K_CE_OFFSET(ab); 278 ie2_reg_addr = ce_ie_addr->ie2_reg_addr + ATH11K_CE_OFFSET(ab); 279 ie3_reg_addr = ce_ie_addr->ie3_reg_addr + ATH11K_CE_OFFSET(ab); 280 281 ce_attr = &ab->hw_params.host_ce_config[ce_id]; 282 if (ce_attr->src_nentries) 283 ath11k_ahb_setbit32(ab, ce_id, ie1_reg_addr); 284 285 if (ce_attr->dest_nentries) { 286 ath11k_ahb_setbit32(ab, ce_id, ie2_reg_addr); 287 ath11k_ahb_setbit32(ab, ce_id + CE_HOST_IE_3_SHIFT, 288 ie3_reg_addr); 289 } 290 } 291 292 static void ath11k_ahb_ce_irq_disable(struct ath11k_base *ab, u16 ce_id) 293 { 294 const struct ce_attr *ce_attr; 295 const struct ce_ie_addr *ce_ie_addr = ab->hw_params.ce_ie_addr; 296 u32 ie1_reg_addr, ie2_reg_addr, ie3_reg_addr; 297 298 ie1_reg_addr = ce_ie_addr->ie1_reg_addr + ATH11K_CE_OFFSET(ab); 299 ie2_reg_addr = ce_ie_addr->ie2_reg_addr + ATH11K_CE_OFFSET(ab); 300 ie3_reg_addr = ce_ie_addr->ie3_reg_addr + ATH11K_CE_OFFSET(ab); 301 302 ce_attr = &ab->hw_params.host_ce_config[ce_id]; 303 if (ce_attr->src_nentries) 304 ath11k_ahb_clearbit32(ab, ce_id, ie1_reg_addr); 305 306 if (ce_attr->dest_nentries) { 307 ath11k_ahb_clearbit32(ab, ce_id, ie2_reg_addr); 308 ath11k_ahb_clearbit32(ab, ce_id + CE_HOST_IE_3_SHIFT, 309 ie3_reg_addr); 310 } 311 } 312 313 static void ath11k_ahb_sync_ce_irqs(struct ath11k_base *ab) 314 { 315 int i; 316 int irq_idx; 317 318 for (i = 0; i < ab->hw_params.ce_count; i++) { 319 if (ath11k_ce_get_attr_flags(ab, i) & CE_ATTR_DIS_INTR) 320 continue; 321 322 irq_idx = ATH11K_IRQ_CE0_OFFSET + i; 323 synchronize_irq(ab->irq_num[irq_idx]); 324 } 325 } 326 327 static void ath11k_ahb_sync_ext_irqs(struct ath11k_base *ab) 328 { 329 int i, j; 330 int irq_idx; 331 332 for (i = 0; i < ATH11K_EXT_IRQ_GRP_NUM_MAX; i++) { 333 struct ath11k_ext_irq_grp *irq_grp = &ab->ext_irq_grp[i]; 334 335 for (j = 0; j < irq_grp->num_irq; j++) { 336 irq_idx = irq_grp->irqs[j]; 337 synchronize_irq(ab->irq_num[irq_idx]); 338 } 339 } 340 } 341 342 static void ath11k_ahb_ce_irqs_enable(struct ath11k_base *ab) 343 { 344 int i; 345 346 for (i = 0; i < ab->hw_params.ce_count; i++) { 347 if (ath11k_ce_get_attr_flags(ab, i) & CE_ATTR_DIS_INTR) 348 continue; 349 ath11k_ahb_ce_irq_enable(ab, i); 350 } 351 } 352 353 static void ath11k_ahb_ce_irqs_disable(struct ath11k_base *ab) 354 { 355 int i; 356 357 for (i = 0; i < ab->hw_params.ce_count; i++) { 358 if (ath11k_ce_get_attr_flags(ab, i) & CE_ATTR_DIS_INTR) 359 continue; 360 ath11k_ahb_ce_irq_disable(ab, i); 361 } 362 } 363 364 static int ath11k_ahb_start(struct ath11k_base *ab) 365 { 366 ath11k_ahb_ce_irqs_enable(ab); 367 ath11k_ce_rx_post_buf(ab); 368 369 return 0; 370 } 371 372 static void ath11k_ahb_ext_irq_enable(struct ath11k_base *ab) 373 { 374 int i; 375 376 for (i = 0; i < ATH11K_EXT_IRQ_GRP_NUM_MAX; i++) { 377 struct ath11k_ext_irq_grp *irq_grp = &ab->ext_irq_grp[i]; 378 379 if (!irq_grp->napi_enabled) { 380 napi_enable(&irq_grp->napi); 381 irq_grp->napi_enabled = true; 382 } 383 ath11k_ahb_ext_grp_enable(irq_grp); 384 } 385 } 386 387 static void ath11k_ahb_ext_irq_disable(struct ath11k_base *ab) 388 { 389 __ath11k_ahb_ext_irq_disable(ab); 390 ath11k_ahb_sync_ext_irqs(ab); 391 } 392 393 static void ath11k_ahb_stop(struct ath11k_base *ab) 394 { 395 if (!test_bit(ATH11K_FLAG_CRASH_FLUSH, &ab->dev_flags)) 396 ath11k_ahb_ce_irqs_disable(ab); 397 ath11k_ahb_sync_ce_irqs(ab); 398 ath11k_ahb_kill_tasklets(ab); 399 del_timer_sync(&ab->rx_replenish_retry); 400 ath11k_ce_cleanup_pipes(ab); 401 } 402 403 static int ath11k_ahb_power_up(struct ath11k_base *ab) 404 { 405 struct ath11k_ahb *ab_ahb = ath11k_ahb_priv(ab); 406 int ret; 407 408 ret = rproc_boot(ab_ahb->tgt_rproc); 409 if (ret) 410 ath11k_err(ab, "failed to boot the remote processor Q6\n"); 411 412 return ret; 413 } 414 415 static void ath11k_ahb_power_down(struct ath11k_base *ab) 416 { 417 struct ath11k_ahb *ab_ahb = ath11k_ahb_priv(ab); 418 419 rproc_shutdown(ab_ahb->tgt_rproc); 420 } 421 422 static void ath11k_ahb_init_qmi_ce_config(struct ath11k_base *ab) 423 { 424 struct ath11k_qmi_ce_cfg *cfg = &ab->qmi.ce_cfg; 425 426 cfg->tgt_ce_len = ab->hw_params.target_ce_count; 427 cfg->tgt_ce = ab->hw_params.target_ce_config; 428 cfg->svc_to_ce_map_len = ab->hw_params.svc_to_ce_map_len; 429 cfg->svc_to_ce_map = ab->hw_params.svc_to_ce_map; 430 ab->qmi.service_ins_id = ab->hw_params.qmi_service_ins_id; 431 } 432 433 static void ath11k_ahb_free_ext_irq(struct ath11k_base *ab) 434 { 435 int i, j; 436 437 for (i = 0; i < ATH11K_EXT_IRQ_GRP_NUM_MAX; i++) { 438 struct ath11k_ext_irq_grp *irq_grp = &ab->ext_irq_grp[i]; 439 440 for (j = 0; j < irq_grp->num_irq; j++) 441 free_irq(ab->irq_num[irq_grp->irqs[j]], irq_grp); 442 443 netif_napi_del(&irq_grp->napi); 444 } 445 } 446 447 static void ath11k_ahb_free_irq(struct ath11k_base *ab) 448 { 449 int irq_idx; 450 int i; 451 452 if (ab->hw_params.hybrid_bus_type) 453 return ath11k_pcic_free_irq(ab); 454 455 for (i = 0; i < ab->hw_params.ce_count; i++) { 456 if (ath11k_ce_get_attr_flags(ab, i) & CE_ATTR_DIS_INTR) 457 continue; 458 irq_idx = ATH11K_IRQ_CE0_OFFSET + i; 459 free_irq(ab->irq_num[irq_idx], &ab->ce.ce_pipe[i]); 460 } 461 462 ath11k_ahb_free_ext_irq(ab); 463 } 464 465 static void ath11k_ahb_ce_tasklet(struct tasklet_struct *t) 466 { 467 struct ath11k_ce_pipe *ce_pipe = from_tasklet(ce_pipe, t, intr_tq); 468 469 ath11k_ce_per_engine_service(ce_pipe->ab, ce_pipe->pipe_num); 470 471 ath11k_ahb_ce_irq_enable(ce_pipe->ab, ce_pipe->pipe_num); 472 } 473 474 static irqreturn_t ath11k_ahb_ce_interrupt_handler(int irq, void *arg) 475 { 476 struct ath11k_ce_pipe *ce_pipe = arg; 477 478 /* last interrupt received for this CE */ 479 ce_pipe->timestamp = jiffies; 480 481 ath11k_ahb_ce_irq_disable(ce_pipe->ab, ce_pipe->pipe_num); 482 483 tasklet_schedule(&ce_pipe->intr_tq); 484 485 return IRQ_HANDLED; 486 } 487 488 static int ath11k_ahb_ext_grp_napi_poll(struct napi_struct *napi, int budget) 489 { 490 struct ath11k_ext_irq_grp *irq_grp = container_of(napi, 491 struct ath11k_ext_irq_grp, 492 napi); 493 struct ath11k_base *ab = irq_grp->ab; 494 int work_done; 495 496 work_done = ath11k_dp_service_srng(ab, irq_grp, budget); 497 if (work_done < budget) { 498 napi_complete_done(napi, work_done); 499 ath11k_ahb_ext_grp_enable(irq_grp); 500 } 501 502 if (work_done > budget) 503 work_done = budget; 504 505 return work_done; 506 } 507 508 static irqreturn_t ath11k_ahb_ext_interrupt_handler(int irq, void *arg) 509 { 510 struct ath11k_ext_irq_grp *irq_grp = arg; 511 512 /* last interrupt received for this group */ 513 irq_grp->timestamp = jiffies; 514 515 ath11k_ahb_ext_grp_disable(irq_grp); 516 517 napi_schedule(&irq_grp->napi); 518 519 return IRQ_HANDLED; 520 } 521 522 static int ath11k_ahb_config_ext_irq(struct ath11k_base *ab) 523 { 524 struct ath11k_hw_params *hw = &ab->hw_params; 525 int i, j; 526 int irq; 527 int ret; 528 529 for (i = 0; i < ATH11K_EXT_IRQ_GRP_NUM_MAX; i++) { 530 struct ath11k_ext_irq_grp *irq_grp = &ab->ext_irq_grp[i]; 531 u32 num_irq = 0; 532 533 irq_grp->ab = ab; 534 irq_grp->grp_id = i; 535 init_dummy_netdev(&irq_grp->napi_ndev); 536 netif_napi_add(&irq_grp->napi_ndev, &irq_grp->napi, 537 ath11k_ahb_ext_grp_napi_poll); 538 539 for (j = 0; j < ATH11K_EXT_IRQ_NUM_MAX; j++) { 540 if (ab->hw_params.ring_mask->tx[i] & BIT(j)) { 541 irq_grp->irqs[num_irq++] = 542 wbm2host_tx_completions_ring1 - j; 543 } 544 545 if (ab->hw_params.ring_mask->rx[i] & BIT(j)) { 546 irq_grp->irqs[num_irq++] = 547 reo2host_destination_ring1 - j; 548 } 549 550 if (ab->hw_params.ring_mask->rx_err[i] & BIT(j)) 551 irq_grp->irqs[num_irq++] = reo2host_exception; 552 553 if (ab->hw_params.ring_mask->rx_wbm_rel[i] & BIT(j)) 554 irq_grp->irqs[num_irq++] = wbm2host_rx_release; 555 556 if (ab->hw_params.ring_mask->reo_status[i] & BIT(j)) 557 irq_grp->irqs[num_irq++] = reo2host_status; 558 559 if (j < ab->hw_params.max_radios) { 560 if (ab->hw_params.ring_mask->rxdma2host[i] & BIT(j)) { 561 irq_grp->irqs[num_irq++] = 562 rxdma2host_destination_ring_mac1 - 563 ath11k_hw_get_mac_from_pdev_id(hw, j); 564 } 565 566 if (ab->hw_params.ring_mask->host2rxdma[i] & BIT(j)) { 567 irq_grp->irqs[num_irq++] = 568 host2rxdma_host_buf_ring_mac1 - 569 ath11k_hw_get_mac_from_pdev_id(hw, j); 570 } 571 572 if (ab->hw_params.ring_mask->rx_mon_status[i] & BIT(j)) { 573 irq_grp->irqs[num_irq++] = 574 ppdu_end_interrupts_mac1 - 575 ath11k_hw_get_mac_from_pdev_id(hw, j); 576 irq_grp->irqs[num_irq++] = 577 rxdma2host_monitor_status_ring_mac1 - 578 ath11k_hw_get_mac_from_pdev_id(hw, j); 579 } 580 } 581 } 582 irq_grp->num_irq = num_irq; 583 584 for (j = 0; j < irq_grp->num_irq; j++) { 585 int irq_idx = irq_grp->irqs[j]; 586 587 irq = platform_get_irq_byname(ab->pdev, 588 irq_name[irq_idx]); 589 ab->irq_num[irq_idx] = irq; 590 irq_set_status_flags(irq, IRQ_NOAUTOEN | IRQ_DISABLE_UNLAZY); 591 ret = request_irq(irq, ath11k_ahb_ext_interrupt_handler, 592 IRQF_TRIGGER_RISING, 593 irq_name[irq_idx], irq_grp); 594 if (ret) { 595 ath11k_err(ab, "failed request_irq for %d\n", 596 irq); 597 } 598 } 599 } 600 601 return 0; 602 } 603 604 static int ath11k_ahb_config_irq(struct ath11k_base *ab) 605 { 606 int irq, irq_idx, i; 607 int ret; 608 609 if (ab->hw_params.hybrid_bus_type) 610 return ath11k_pcic_config_irq(ab); 611 612 /* Configure CE irqs */ 613 for (i = 0; i < ab->hw_params.ce_count; i++) { 614 struct ath11k_ce_pipe *ce_pipe = &ab->ce.ce_pipe[i]; 615 616 if (ath11k_ce_get_attr_flags(ab, i) & CE_ATTR_DIS_INTR) 617 continue; 618 619 irq_idx = ATH11K_IRQ_CE0_OFFSET + i; 620 621 tasklet_setup(&ce_pipe->intr_tq, ath11k_ahb_ce_tasklet); 622 irq = platform_get_irq_byname(ab->pdev, irq_name[irq_idx]); 623 ret = request_irq(irq, ath11k_ahb_ce_interrupt_handler, 624 IRQF_TRIGGER_RISING, irq_name[irq_idx], 625 ce_pipe); 626 if (ret) 627 return ret; 628 629 ab->irq_num[irq_idx] = irq; 630 } 631 632 /* Configure external interrupts */ 633 ret = ath11k_ahb_config_ext_irq(ab); 634 635 return ret; 636 } 637 638 static int ath11k_ahb_map_service_to_pipe(struct ath11k_base *ab, u16 service_id, 639 u8 *ul_pipe, u8 *dl_pipe) 640 { 641 const struct service_to_pipe *entry; 642 bool ul_set = false, dl_set = false; 643 int i; 644 645 for (i = 0; i < ab->hw_params.svc_to_ce_map_len; i++) { 646 entry = &ab->hw_params.svc_to_ce_map[i]; 647 648 if (__le32_to_cpu(entry->service_id) != service_id) 649 continue; 650 651 switch (__le32_to_cpu(entry->pipedir)) { 652 case PIPEDIR_NONE: 653 break; 654 case PIPEDIR_IN: 655 WARN_ON(dl_set); 656 *dl_pipe = __le32_to_cpu(entry->pipenum); 657 dl_set = true; 658 break; 659 case PIPEDIR_OUT: 660 WARN_ON(ul_set); 661 *ul_pipe = __le32_to_cpu(entry->pipenum); 662 ul_set = true; 663 break; 664 case PIPEDIR_INOUT: 665 WARN_ON(dl_set); 666 WARN_ON(ul_set); 667 *dl_pipe = __le32_to_cpu(entry->pipenum); 668 *ul_pipe = __le32_to_cpu(entry->pipenum); 669 dl_set = true; 670 ul_set = true; 671 break; 672 } 673 } 674 675 if (WARN_ON(!ul_set || !dl_set)) 676 return -ENOENT; 677 678 return 0; 679 } 680 681 static int ath11k_ahb_hif_suspend(struct ath11k_base *ab) 682 { 683 struct ath11k_ahb *ab_ahb = ath11k_ahb_priv(ab); 684 u32 wake_irq; 685 u32 value = 0; 686 int ret; 687 688 if (!device_may_wakeup(ab->dev)) 689 return -EPERM; 690 691 wake_irq = ab->irq_num[ATH11K_PCI_IRQ_CE0_OFFSET + ATH11K_PCI_CE_WAKE_IRQ]; 692 693 ret = enable_irq_wake(wake_irq); 694 if (ret) { 695 ath11k_err(ab, "failed to enable wakeup irq :%d\n", ret); 696 return ret; 697 } 698 699 value = u32_encode_bits(ab_ahb->smp2p_info.seq_no++, 700 ATH11K_AHB_SMP2P_SMEM_SEQ_NO); 701 value |= u32_encode_bits(ATH11K_AHB_POWER_SAVE_ENTER, 702 ATH11K_AHB_SMP2P_SMEM_MSG); 703 704 ret = qcom_smem_state_update_bits(ab_ahb->smp2p_info.smem_state, 705 ATH11K_AHB_SMP2P_SMEM_VALUE_MASK, value); 706 if (ret) { 707 ath11k_err(ab, "failed to send smp2p power save enter cmd :%d\n", ret); 708 return ret; 709 } 710 711 ath11k_dbg(ab, ATH11K_DBG_AHB, "device suspended\n"); 712 713 return ret; 714 } 715 716 static int ath11k_ahb_hif_resume(struct ath11k_base *ab) 717 { 718 struct ath11k_ahb *ab_ahb = ath11k_ahb_priv(ab); 719 u32 wake_irq; 720 u32 value = 0; 721 int ret; 722 723 if (!device_may_wakeup(ab->dev)) 724 return -EPERM; 725 726 wake_irq = ab->irq_num[ATH11K_PCI_IRQ_CE0_OFFSET + ATH11K_PCI_CE_WAKE_IRQ]; 727 728 ret = disable_irq_wake(wake_irq); 729 if (ret) { 730 ath11k_err(ab, "failed to disable wakeup irq: %d\n", ret); 731 return ret; 732 } 733 734 reinit_completion(&ab->wow.wakeup_completed); 735 736 value = u32_encode_bits(ab_ahb->smp2p_info.seq_no++, 737 ATH11K_AHB_SMP2P_SMEM_SEQ_NO); 738 value |= u32_encode_bits(ATH11K_AHB_POWER_SAVE_EXIT, 739 ATH11K_AHB_SMP2P_SMEM_MSG); 740 741 ret = qcom_smem_state_update_bits(ab_ahb->smp2p_info.smem_state, 742 ATH11K_AHB_SMP2P_SMEM_VALUE_MASK, value); 743 if (ret) { 744 ath11k_err(ab, "failed to send smp2p power save enter cmd :%d\n", ret); 745 return ret; 746 } 747 748 ret = wait_for_completion_timeout(&ab->wow.wakeup_completed, 3 * HZ); 749 if (ret == 0) { 750 ath11k_warn(ab, "timed out while waiting for wow wakeup completion\n"); 751 return -ETIMEDOUT; 752 } 753 754 ath11k_dbg(ab, ATH11K_DBG_AHB, "device resumed\n"); 755 756 return 0; 757 } 758 759 static const struct ath11k_hif_ops ath11k_ahb_hif_ops_ipq8074 = { 760 .start = ath11k_ahb_start, 761 .stop = ath11k_ahb_stop, 762 .read32 = ath11k_ahb_read32, 763 .write32 = ath11k_ahb_write32, 764 .read = NULL, 765 .irq_enable = ath11k_ahb_ext_irq_enable, 766 .irq_disable = ath11k_ahb_ext_irq_disable, 767 .map_service_to_pipe = ath11k_ahb_map_service_to_pipe, 768 .power_down = ath11k_ahb_power_down, 769 .power_up = ath11k_ahb_power_up, 770 }; 771 772 static const struct ath11k_hif_ops ath11k_ahb_hif_ops_wcn6750 = { 773 .start = ath11k_pcic_start, 774 .stop = ath11k_pcic_stop, 775 .read32 = ath11k_pcic_read32, 776 .write32 = ath11k_pcic_write32, 777 .read = NULL, 778 .irq_enable = ath11k_pcic_ext_irq_enable, 779 .irq_disable = ath11k_pcic_ext_irq_disable, 780 .get_msi_address = ath11k_pcic_get_msi_address, 781 .get_user_msi_vector = ath11k_pcic_get_user_msi_assignment, 782 .map_service_to_pipe = ath11k_pcic_map_service_to_pipe, 783 .power_down = ath11k_ahb_power_down, 784 .power_up = ath11k_ahb_power_up, 785 .suspend = ath11k_ahb_hif_suspend, 786 .resume = ath11k_ahb_hif_resume, 787 .ce_irq_enable = ath11k_pci_enable_ce_irqs_except_wake_irq, 788 .ce_irq_disable = ath11k_pci_disable_ce_irqs_except_wake_irq, 789 }; 790 791 static int ath11k_core_get_rproc(struct ath11k_base *ab) 792 { 793 struct ath11k_ahb *ab_ahb = ath11k_ahb_priv(ab); 794 struct device *dev = ab->dev; 795 struct rproc *prproc; 796 phandle rproc_phandle; 797 798 if (of_property_read_u32(dev->of_node, "qcom,rproc", &rproc_phandle)) { 799 ath11k_err(ab, "failed to get q6_rproc handle\n"); 800 return -ENOENT; 801 } 802 803 prproc = rproc_get_by_phandle(rproc_phandle); 804 if (!prproc) { 805 ath11k_err(ab, "failed to get rproc\n"); 806 return -EINVAL; 807 } 808 ab_ahb->tgt_rproc = prproc; 809 810 return 0; 811 } 812 813 static int ath11k_ahb_setup_msi_resources(struct ath11k_base *ab) 814 { 815 struct platform_device *pdev = ab->pdev; 816 phys_addr_t msi_addr_pa; 817 dma_addr_t msi_addr_iova; 818 struct resource *res; 819 int int_prop; 820 int ret; 821 int i; 822 823 ret = ath11k_pcic_init_msi_config(ab); 824 if (ret) { 825 ath11k_err(ab, "failed to init msi config: %d\n", ret); 826 return ret; 827 } 828 829 res = platform_get_resource(pdev, IORESOURCE_MEM, 0); 830 if (!res) { 831 ath11k_err(ab, "failed to fetch msi_addr\n"); 832 return -ENOENT; 833 } 834 835 msi_addr_pa = res->start; 836 msi_addr_iova = dma_map_resource(ab->dev, msi_addr_pa, PAGE_SIZE, 837 DMA_FROM_DEVICE, 0); 838 if (dma_mapping_error(ab->dev, msi_addr_iova)) 839 return -ENOMEM; 840 841 ab->pci.msi.addr_lo = lower_32_bits(msi_addr_iova); 842 ab->pci.msi.addr_hi = upper_32_bits(msi_addr_iova); 843 844 ret = of_property_read_u32_index(ab->dev->of_node, "interrupts", 1, &int_prop); 845 if (ret) 846 return ret; 847 848 ab->pci.msi.ep_base_data = int_prop + 32; 849 850 for (i = 0; i < ab->pci.msi.config->total_vectors; i++) { 851 ret = platform_get_irq(pdev, i); 852 if (ret < 0) 853 return ret; 854 855 ab->pci.msi.irqs[i] = ret; 856 } 857 858 set_bit(ATH11K_FLAG_MULTI_MSI_VECTORS, &ab->dev_flags); 859 860 return 0; 861 } 862 863 static int ath11k_ahb_setup_smp2p_handle(struct ath11k_base *ab) 864 { 865 struct ath11k_ahb *ab_ahb = ath11k_ahb_priv(ab); 866 867 if (!ab->hw_params.smp2p_wow_exit) 868 return 0; 869 870 ab_ahb->smp2p_info.smem_state = qcom_smem_state_get(ab->dev, "wlan-smp2p-out", 871 &ab_ahb->smp2p_info.smem_bit); 872 if (IS_ERR(ab_ahb->smp2p_info.smem_state)) { 873 ath11k_err(ab, "failed to fetch smem state: %ld\n", 874 PTR_ERR(ab_ahb->smp2p_info.smem_state)); 875 return PTR_ERR(ab_ahb->smp2p_info.smem_state); 876 } 877 878 return 0; 879 } 880 881 static void ath11k_ahb_release_smp2p_handle(struct ath11k_base *ab) 882 { 883 struct ath11k_ahb *ab_ahb = ath11k_ahb_priv(ab); 884 885 if (!ab->hw_params.smp2p_wow_exit) 886 return; 887 888 qcom_smem_state_put(ab_ahb->smp2p_info.smem_state); 889 } 890 891 static int ath11k_ahb_setup_resources(struct ath11k_base *ab) 892 { 893 struct platform_device *pdev = ab->pdev; 894 struct resource *mem_res; 895 void __iomem *mem; 896 897 if (ab->hw_params.hybrid_bus_type) 898 return ath11k_ahb_setup_msi_resources(ab); 899 900 mem = devm_platform_get_and_ioremap_resource(pdev, 0, &mem_res); 901 if (IS_ERR(mem)) { 902 dev_err(&pdev->dev, "ioremap error\n"); 903 return PTR_ERR(mem); 904 } 905 906 ab->mem = mem; 907 ab->mem_len = resource_size(mem_res); 908 909 return 0; 910 } 911 912 static int ath11k_ahb_setup_msa_resources(struct ath11k_base *ab) 913 { 914 struct ath11k_ahb *ab_ahb = ath11k_ahb_priv(ab); 915 struct device *dev = ab->dev; 916 struct device_node *node; 917 struct resource r; 918 int ret; 919 920 node = of_parse_phandle(dev->of_node, "memory-region", 0); 921 if (!node) 922 return -ENOENT; 923 924 ret = of_address_to_resource(node, 0, &r); 925 of_node_put(node); 926 if (ret) { 927 dev_err(dev, "failed to resolve msa fixed region\n"); 928 return ret; 929 } 930 931 ab_ahb->fw.msa_paddr = r.start; 932 ab_ahb->fw.msa_size = resource_size(&r); 933 934 node = of_parse_phandle(dev->of_node, "memory-region", 1); 935 if (!node) 936 return -ENOENT; 937 938 ret = of_address_to_resource(node, 0, &r); 939 of_node_put(node); 940 if (ret) { 941 dev_err(dev, "failed to resolve ce fixed region\n"); 942 return ret; 943 } 944 945 ab_ahb->fw.ce_paddr = r.start; 946 ab_ahb->fw.ce_size = resource_size(&r); 947 948 return 0; 949 } 950 951 static int ath11k_ahb_fw_resources_init(struct ath11k_base *ab) 952 { 953 struct ath11k_ahb *ab_ahb = ath11k_ahb_priv(ab); 954 struct device *host_dev = ab->dev; 955 struct platform_device_info info = {0}; 956 struct iommu_domain *iommu_dom; 957 struct platform_device *pdev; 958 struct device_node *node; 959 int ret; 960 961 /* Chipsets not requiring MSA need not initialize 962 * MSA resources, return success in such cases. 963 */ 964 if (!ab->hw_params.fixed_fw_mem) 965 return 0; 966 967 ret = ath11k_ahb_setup_msa_resources(ab); 968 if (ret) { 969 ath11k_err(ab, "failed to setup msa resources\n"); 970 return ret; 971 } 972 973 node = of_get_child_by_name(host_dev->of_node, "wifi-firmware"); 974 if (!node) { 975 ab_ahb->fw.use_tz = true; 976 return 0; 977 } 978 979 info.fwnode = &node->fwnode; 980 info.parent = host_dev; 981 info.name = node->name; 982 info.dma_mask = DMA_BIT_MASK(32); 983 984 pdev = platform_device_register_full(&info); 985 if (IS_ERR(pdev)) { 986 of_node_put(node); 987 return PTR_ERR(pdev); 988 } 989 990 ret = of_dma_configure(&pdev->dev, node, true); 991 if (ret) { 992 ath11k_err(ab, "dma configure fail: %d\n", ret); 993 goto err_unregister; 994 } 995 996 ab_ahb->fw.dev = &pdev->dev; 997 998 iommu_dom = iommu_domain_alloc(&platform_bus_type); 999 if (!iommu_dom) { 1000 ath11k_err(ab, "failed to allocate iommu domain\n"); 1001 ret = -ENOMEM; 1002 goto err_unregister; 1003 } 1004 1005 ret = iommu_attach_device(iommu_dom, ab_ahb->fw.dev); 1006 if (ret) { 1007 ath11k_err(ab, "could not attach device: %d\n", ret); 1008 goto err_iommu_free; 1009 } 1010 1011 ret = iommu_map(iommu_dom, ab_ahb->fw.msa_paddr, 1012 ab_ahb->fw.msa_paddr, ab_ahb->fw.msa_size, 1013 IOMMU_READ | IOMMU_WRITE, GFP_KERNEL); 1014 if (ret) { 1015 ath11k_err(ab, "failed to map firmware region: %d\n", ret); 1016 goto err_iommu_detach; 1017 } 1018 1019 ret = iommu_map(iommu_dom, ab_ahb->fw.ce_paddr, 1020 ab_ahb->fw.ce_paddr, ab_ahb->fw.ce_size, 1021 IOMMU_READ | IOMMU_WRITE, GFP_KERNEL); 1022 if (ret) { 1023 ath11k_err(ab, "failed to map firmware CE region: %d\n", ret); 1024 goto err_iommu_unmap; 1025 } 1026 1027 ab_ahb->fw.use_tz = false; 1028 ab_ahb->fw.iommu_domain = iommu_dom; 1029 of_node_put(node); 1030 1031 return 0; 1032 1033 err_iommu_unmap: 1034 iommu_unmap(iommu_dom, ab_ahb->fw.msa_paddr, ab_ahb->fw.msa_size); 1035 1036 err_iommu_detach: 1037 iommu_detach_device(iommu_dom, ab_ahb->fw.dev); 1038 1039 err_iommu_free: 1040 iommu_domain_free(iommu_dom); 1041 1042 err_unregister: 1043 platform_device_unregister(pdev); 1044 of_node_put(node); 1045 1046 return ret; 1047 } 1048 1049 static int ath11k_ahb_fw_resource_deinit(struct ath11k_base *ab) 1050 { 1051 struct ath11k_ahb *ab_ahb = ath11k_ahb_priv(ab); 1052 struct iommu_domain *iommu; 1053 size_t unmapped_size; 1054 1055 /* Chipsets not requiring MSA would have not initialized 1056 * MSA resources, return success in such cases. 1057 */ 1058 if (!ab->hw_params.fixed_fw_mem) 1059 return 0; 1060 1061 if (ab_ahb->fw.use_tz) 1062 return 0; 1063 1064 iommu = ab_ahb->fw.iommu_domain; 1065 1066 unmapped_size = iommu_unmap(iommu, ab_ahb->fw.msa_paddr, ab_ahb->fw.msa_size); 1067 if (unmapped_size != ab_ahb->fw.msa_size) 1068 ath11k_err(ab, "failed to unmap firmware: %zu\n", 1069 unmapped_size); 1070 1071 unmapped_size = iommu_unmap(iommu, ab_ahb->fw.ce_paddr, ab_ahb->fw.ce_size); 1072 if (unmapped_size != ab_ahb->fw.ce_size) 1073 ath11k_err(ab, "failed to unmap firmware CE memory: %zu\n", 1074 unmapped_size); 1075 1076 iommu_detach_device(iommu, ab_ahb->fw.dev); 1077 iommu_domain_free(iommu); 1078 1079 platform_device_unregister(to_platform_device(ab_ahb->fw.dev)); 1080 1081 return 0; 1082 } 1083 1084 static int ath11k_ahb_probe(struct platform_device *pdev) 1085 { 1086 struct ath11k_base *ab; 1087 const struct of_device_id *of_id; 1088 const struct ath11k_hif_ops *hif_ops; 1089 const struct ath11k_pci_ops *pci_ops; 1090 enum ath11k_hw_rev hw_rev; 1091 int ret; 1092 1093 of_id = of_match_device(ath11k_ahb_of_match, &pdev->dev); 1094 if (!of_id) { 1095 dev_err(&pdev->dev, "failed to find matching device tree id\n"); 1096 return -EINVAL; 1097 } 1098 1099 hw_rev = (enum ath11k_hw_rev)of_id->data; 1100 1101 switch (hw_rev) { 1102 case ATH11K_HW_IPQ8074: 1103 case ATH11K_HW_IPQ6018_HW10: 1104 case ATH11K_HW_IPQ5018_HW10: 1105 hif_ops = &ath11k_ahb_hif_ops_ipq8074; 1106 pci_ops = NULL; 1107 break; 1108 case ATH11K_HW_WCN6750_HW10: 1109 hif_ops = &ath11k_ahb_hif_ops_wcn6750; 1110 pci_ops = &ath11k_ahb_pci_ops_wcn6750; 1111 break; 1112 default: 1113 dev_err(&pdev->dev, "unsupported device type %d\n", hw_rev); 1114 return -EOPNOTSUPP; 1115 } 1116 1117 ret = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(32)); 1118 if (ret) { 1119 dev_err(&pdev->dev, "failed to set 32-bit consistent dma\n"); 1120 return ret; 1121 } 1122 1123 ab = ath11k_core_alloc(&pdev->dev, sizeof(struct ath11k_ahb), 1124 ATH11K_BUS_AHB); 1125 if (!ab) { 1126 dev_err(&pdev->dev, "failed to allocate ath11k base\n"); 1127 return -ENOMEM; 1128 } 1129 1130 ab->hif.ops = hif_ops; 1131 ab->pdev = pdev; 1132 ab->hw_rev = hw_rev; 1133 ab->fw_mode = ATH11K_FIRMWARE_MODE_NORMAL; 1134 platform_set_drvdata(pdev, ab); 1135 1136 ret = ath11k_pcic_register_pci_ops(ab, pci_ops); 1137 if (ret) { 1138 ath11k_err(ab, "failed to register PCI ops: %d\n", ret); 1139 goto err_core_free; 1140 } 1141 1142 ret = ath11k_core_pre_init(ab); 1143 if (ret) 1144 goto err_core_free; 1145 1146 ret = ath11k_ahb_setup_resources(ab); 1147 if (ret) 1148 goto err_core_free; 1149 1150 ab->mem_ce = ab->mem; 1151 1152 if (ab->hw_params.ce_remap) { 1153 const struct ce_remap *ce_remap = ab->hw_params.ce_remap; 1154 /* ce register space is moved out of wcss unlike ipq8074 or ipq6018 1155 * and the space is not contiguous, hence remapping the CE registers 1156 * to a new space for accessing them. 1157 */ 1158 ab->mem_ce = ioremap(ce_remap->base, ce_remap->size); 1159 if (!ab->mem_ce) { 1160 dev_err(&pdev->dev, "ce ioremap error\n"); 1161 ret = -ENOMEM; 1162 goto err_core_free; 1163 } 1164 } 1165 1166 ret = ath11k_ahb_fw_resources_init(ab); 1167 if (ret) 1168 goto err_core_free; 1169 1170 ret = ath11k_ahb_setup_smp2p_handle(ab); 1171 if (ret) 1172 goto err_fw_deinit; 1173 1174 ret = ath11k_hal_srng_init(ab); 1175 if (ret) 1176 goto err_release_smp2p_handle; 1177 1178 ret = ath11k_ce_alloc_pipes(ab); 1179 if (ret) { 1180 ath11k_err(ab, "failed to allocate ce pipes: %d\n", ret); 1181 goto err_hal_srng_deinit; 1182 } 1183 1184 ath11k_ahb_init_qmi_ce_config(ab); 1185 1186 ret = ath11k_core_get_rproc(ab); 1187 if (ret) { 1188 ath11k_err(ab, "failed to get rproc: %d\n", ret); 1189 goto err_ce_free; 1190 } 1191 1192 ret = ath11k_core_init(ab); 1193 if (ret) { 1194 ath11k_err(ab, "failed to init core: %d\n", ret); 1195 goto err_ce_free; 1196 } 1197 1198 ret = ath11k_ahb_config_irq(ab); 1199 if (ret) { 1200 ath11k_err(ab, "failed to configure irq: %d\n", ret); 1201 goto err_ce_free; 1202 } 1203 1204 ath11k_qmi_fwreset_from_cold_boot(ab); 1205 1206 return 0; 1207 1208 err_ce_free: 1209 ath11k_ce_free_pipes(ab); 1210 1211 err_hal_srng_deinit: 1212 ath11k_hal_srng_deinit(ab); 1213 1214 err_release_smp2p_handle: 1215 ath11k_ahb_release_smp2p_handle(ab); 1216 1217 err_fw_deinit: 1218 ath11k_ahb_fw_resource_deinit(ab); 1219 1220 err_core_free: 1221 ath11k_core_free(ab); 1222 platform_set_drvdata(pdev, NULL); 1223 1224 return ret; 1225 } 1226 1227 static void ath11k_ahb_remove_prepare(struct ath11k_base *ab) 1228 { 1229 unsigned long left; 1230 1231 if (test_bit(ATH11K_FLAG_RECOVERY, &ab->dev_flags)) { 1232 left = wait_for_completion_timeout(&ab->driver_recovery, 1233 ATH11K_AHB_RECOVERY_TIMEOUT); 1234 if (!left) 1235 ath11k_warn(ab, "failed to receive recovery response completion\n"); 1236 } 1237 1238 set_bit(ATH11K_FLAG_UNREGISTERING, &ab->dev_flags); 1239 cancel_work_sync(&ab->restart_work); 1240 cancel_work_sync(&ab->qmi.event_work); 1241 } 1242 1243 static void ath11k_ahb_free_resources(struct ath11k_base *ab) 1244 { 1245 struct platform_device *pdev = ab->pdev; 1246 1247 ath11k_ahb_free_irq(ab); 1248 ath11k_hal_srng_deinit(ab); 1249 ath11k_ahb_release_smp2p_handle(ab); 1250 ath11k_ahb_fw_resource_deinit(ab); 1251 ath11k_ce_free_pipes(ab); 1252 1253 if (ab->hw_params.ce_remap) 1254 iounmap(ab->mem_ce); 1255 1256 ath11k_core_free(ab); 1257 platform_set_drvdata(pdev, NULL); 1258 } 1259 1260 static int ath11k_ahb_remove(struct platform_device *pdev) 1261 { 1262 struct ath11k_base *ab = platform_get_drvdata(pdev); 1263 1264 if (test_bit(ATH11K_FLAG_QMI_FAIL, &ab->dev_flags)) { 1265 ath11k_ahb_power_down(ab); 1266 ath11k_debugfs_soc_destroy(ab); 1267 ath11k_qmi_deinit_service(ab); 1268 goto qmi_fail; 1269 } 1270 1271 ath11k_ahb_remove_prepare(ab); 1272 ath11k_core_deinit(ab); 1273 1274 qmi_fail: 1275 ath11k_ahb_free_resources(ab); 1276 1277 return 0; 1278 } 1279 1280 static void ath11k_ahb_shutdown(struct platform_device *pdev) 1281 { 1282 struct ath11k_base *ab = platform_get_drvdata(pdev); 1283 1284 /* platform shutdown() & remove() are mutually exclusive. 1285 * remove() is invoked during rmmod & shutdown() during 1286 * system reboot/shutdown. 1287 */ 1288 ath11k_ahb_remove_prepare(ab); 1289 1290 if (!(test_bit(ATH11K_FLAG_REGISTERED, &ab->dev_flags))) 1291 goto free_resources; 1292 1293 ath11k_core_deinit(ab); 1294 1295 free_resources: 1296 ath11k_ahb_free_resources(ab); 1297 } 1298 1299 static struct platform_driver ath11k_ahb_driver = { 1300 .driver = { 1301 .name = "ath11k", 1302 .of_match_table = ath11k_ahb_of_match, 1303 }, 1304 .probe = ath11k_ahb_probe, 1305 .remove = ath11k_ahb_remove, 1306 .shutdown = ath11k_ahb_shutdown, 1307 }; 1308 1309 static int ath11k_ahb_init(void) 1310 { 1311 return platform_driver_register(&ath11k_ahb_driver); 1312 } 1313 module_init(ath11k_ahb_init); 1314 1315 static void ath11k_ahb_exit(void) 1316 { 1317 platform_driver_unregister(&ath11k_ahb_driver); 1318 } 1319 module_exit(ath11k_ahb_exit); 1320 1321 MODULE_DESCRIPTION("Driver support for Qualcomm Technologies 802.11ax WLAN AHB devices"); 1322 MODULE_LICENSE("Dual BSD/GPL"); 1323