1 /*- 2 * SPDX-License-Identifier: BSD-2-Clause 3 * 4 * Copyright (c) 2007-2009 Sam Leffler, Errno Consulting 5 * Copyright (c) 2007-2008 Marvell Semiconductor, Inc. 6 * All rights reserved. 7 * 8 * Redistribution and use in source and binary forms, with or without 9 * modification, are permitted provided that the following conditions 10 * are met: 11 * 1. Redistributions of source code must retain the above copyright 12 * notice, this list of conditions and the following disclaimer, 13 * without modification. 14 * 2. Redistributions in binary form must reproduce at minimum a disclaimer 15 * similar to the "NO WARRANTY" disclaimer below ("Disclaimer") and any 16 * redistribution must be conditioned upon including a substantially 17 * similar Disclaimer requirement for further binary redistribution. 18 * 19 * NO WARRANTY 20 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 21 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 22 * LIMITED TO, THE IMPLIED WARRANTIES OF NONINFRINGEMENT, MERCHANTIBILITY 23 * AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL 24 * THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY, 25 * OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 26 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 27 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER 28 * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 29 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF 30 * THE POSSIBILITY OF SUCH DAMAGES. 31 */ 32 33 #include <sys/cdefs.h> 34 /* 35 * Driver for the Marvell 88W8363 Wireless LAN controller. 36 */ 37 38 #include "opt_inet.h" 39 #include "opt_mwl.h" 40 #include "opt_wlan.h" 41 42 #include <sys/param.h> 43 #include <sys/systm.h> 44 #include <sys/sysctl.h> 45 #include <sys/mbuf.h> 46 #include <sys/malloc.h> 47 #include <sys/lock.h> 48 #include <sys/mutex.h> 49 #include <sys/kernel.h> 50 #include <sys/socket.h> 51 #include <sys/sockio.h> 52 #include <sys/errno.h> 53 #include <sys/callout.h> 54 #include <sys/bus.h> 55 #include <sys/endian.h> 56 #include <sys/kthread.h> 57 #include <sys/taskqueue.h> 58 59 #include <machine/bus.h> 60 61 #include <net/if.h> 62 #include <net/if_var.h> 63 #include <net/if_dl.h> 64 #include <net/if_media.h> 65 #include <net/if_types.h> 66 #include <net/if_arp.h> 67 #include <net/ethernet.h> 68 #include <net/if_llc.h> 69 70 #include <net/bpf.h> 71 72 #include <net80211/ieee80211_var.h> 73 #include <net80211/ieee80211_input.h> 74 #include <net80211/ieee80211_regdomain.h> 75 76 #ifdef INET 77 #include <netinet/in.h> 78 #include <netinet/if_ether.h> 79 #endif /* INET */ 80 81 #include <dev/mwl/if_mwlvar.h> 82 #include <dev/mwl/mwldiag.h> 83 84 static struct ieee80211vap *mwl_vap_create(struct ieee80211com *, 85 const char [IFNAMSIZ], int, enum ieee80211_opmode, int, 86 const uint8_t [IEEE80211_ADDR_LEN], 87 const uint8_t [IEEE80211_ADDR_LEN]); 88 static void mwl_vap_delete(struct ieee80211vap *); 89 static int mwl_setupdma(struct mwl_softc *); 90 static int mwl_hal_reset(struct mwl_softc *sc); 91 static int mwl_init(struct mwl_softc *); 92 static void mwl_parent(struct ieee80211com *); 93 static int mwl_reset(struct ieee80211vap *, u_long); 94 static void mwl_stop(struct mwl_softc *); 95 static void mwl_start(struct mwl_softc *); 96 static int mwl_transmit(struct ieee80211com *, struct mbuf *); 97 static int mwl_raw_xmit(struct ieee80211_node *, struct mbuf *, 98 const struct ieee80211_bpf_params *); 99 static int mwl_media_change(if_t); 100 static void mwl_watchdog(void *); 101 static int mwl_ioctl(struct ieee80211com *, u_long, void *); 102 static void mwl_radar_proc(void *, int); 103 static void mwl_chanswitch_proc(void *, int); 104 static void mwl_bawatchdog_proc(void *, int); 105 static int mwl_key_alloc(struct ieee80211vap *, 106 struct ieee80211_key *, 107 ieee80211_keyix *, ieee80211_keyix *); 108 static int mwl_key_delete(struct ieee80211vap *, 109 const struct ieee80211_key *); 110 static int mwl_key_set(struct ieee80211vap *, 111 const struct ieee80211_key *); 112 static int _mwl_key_set(struct ieee80211vap *, 113 const struct ieee80211_key *, 114 const uint8_t mac[IEEE80211_ADDR_LEN]); 115 static int mwl_mode_init(struct mwl_softc *); 116 static void mwl_update_mcast(struct ieee80211com *); 117 static void mwl_update_promisc(struct ieee80211com *); 118 static void mwl_updateslot(struct ieee80211com *); 119 static int mwl_beacon_setup(struct ieee80211vap *); 120 static void mwl_beacon_update(struct ieee80211vap *, int); 121 #ifdef MWL_HOST_PS_SUPPORT 122 static void mwl_update_ps(struct ieee80211vap *, int); 123 static int mwl_set_tim(struct ieee80211_node *, int); 124 #endif 125 static int mwl_dma_setup(struct mwl_softc *); 126 static void mwl_dma_cleanup(struct mwl_softc *); 127 static struct ieee80211_node *mwl_node_alloc(struct ieee80211vap *, 128 const uint8_t [IEEE80211_ADDR_LEN]); 129 static void mwl_node_cleanup(struct ieee80211_node *); 130 static void mwl_node_drain(struct ieee80211_node *); 131 static void mwl_node_getsignal(const struct ieee80211_node *, 132 int8_t *, int8_t *); 133 static void mwl_node_getmimoinfo(const struct ieee80211_node *, 134 struct ieee80211_mimo_info *); 135 static int mwl_rxbuf_init(struct mwl_softc *, struct mwl_rxbuf *); 136 static void mwl_rx_proc(void *, int); 137 static void mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *, int); 138 static int mwl_tx_setup(struct mwl_softc *, int, int); 139 static int mwl_wme_update(struct ieee80211com *); 140 static void mwl_tx_cleanupq(struct mwl_softc *, struct mwl_txq *); 141 static void mwl_tx_cleanup(struct mwl_softc *); 142 static uint16_t mwl_calcformat(uint8_t rate, const struct ieee80211_node *); 143 static int mwl_tx_start(struct mwl_softc *, struct ieee80211_node *, 144 struct mwl_txbuf *, struct mbuf *); 145 static void mwl_tx_proc(void *, int); 146 static int mwl_chan_set(struct mwl_softc *, struct ieee80211_channel *); 147 static void mwl_draintxq(struct mwl_softc *); 148 static void mwl_cleartxq(struct mwl_softc *, struct ieee80211vap *); 149 static int mwl_recv_action(struct ieee80211_node *, 150 const struct ieee80211_frame *, 151 const uint8_t *, const uint8_t *); 152 static int mwl_addba_request(struct ieee80211_node *, 153 struct ieee80211_tx_ampdu *, int dialogtoken, 154 int baparamset, int batimeout); 155 static int mwl_addba_response(struct ieee80211_node *, 156 struct ieee80211_tx_ampdu *, int status, 157 int baparamset, int batimeout); 158 static void mwl_addba_stop(struct ieee80211_node *, 159 struct ieee80211_tx_ampdu *); 160 static int mwl_startrecv(struct mwl_softc *); 161 static MWL_HAL_APMODE mwl_getapmode(const struct ieee80211vap *, 162 struct ieee80211_channel *); 163 static int mwl_setapmode(struct ieee80211vap *, struct ieee80211_channel*); 164 static void mwl_scan_start(struct ieee80211com *); 165 static void mwl_scan_end(struct ieee80211com *); 166 static void mwl_set_channel(struct ieee80211com *); 167 static int mwl_peerstadb(struct ieee80211_node *, 168 int aid, int staid, MWL_HAL_PEERINFO *pi); 169 static int mwl_localstadb(struct ieee80211vap *); 170 static int mwl_newstate(struct ieee80211vap *, enum ieee80211_state, int); 171 static int allocstaid(struct mwl_softc *sc, int aid); 172 static void delstaid(struct mwl_softc *sc, int staid); 173 static void mwl_newassoc(struct ieee80211_node *, int); 174 static void mwl_agestations(void *); 175 static int mwl_setregdomain(struct ieee80211com *, 176 struct ieee80211_regdomain *, int, 177 struct ieee80211_channel []); 178 static void mwl_getradiocaps(struct ieee80211com *, int, int *, 179 struct ieee80211_channel []); 180 static int mwl_getchannels(struct mwl_softc *); 181 182 static void mwl_sysctlattach(struct mwl_softc *); 183 static void mwl_announce(struct mwl_softc *); 184 185 SYSCTL_NODE(_hw, OID_AUTO, mwl, CTLFLAG_RD | CTLFLAG_MPSAFE, 0, 186 "Marvell driver parameters"); 187 188 static int mwl_rxdesc = MWL_RXDESC; /* # rx desc's to allocate */ 189 SYSCTL_INT(_hw_mwl, OID_AUTO, rxdesc, CTLFLAG_RW, &mwl_rxdesc, 190 0, "rx descriptors allocated"); 191 static int mwl_rxbuf = MWL_RXBUF; /* # rx buffers to allocate */ 192 SYSCTL_INT(_hw_mwl, OID_AUTO, rxbuf, CTLFLAG_RWTUN, &mwl_rxbuf, 193 0, "rx buffers allocated"); 194 static int mwl_txbuf = MWL_TXBUF; /* # tx buffers to allocate */ 195 SYSCTL_INT(_hw_mwl, OID_AUTO, txbuf, CTLFLAG_RWTUN, &mwl_txbuf, 196 0, "tx buffers allocated"); 197 static int mwl_txcoalesce = 8; /* # tx packets to q before poking f/w*/ 198 SYSCTL_INT(_hw_mwl, OID_AUTO, txcoalesce, CTLFLAG_RWTUN, &mwl_txcoalesce, 199 0, "tx buffers to send at once"); 200 static int mwl_rxquota = MWL_RXBUF; /* # max buffers to process */ 201 SYSCTL_INT(_hw_mwl, OID_AUTO, rxquota, CTLFLAG_RWTUN, &mwl_rxquota, 202 0, "max rx buffers to process per interrupt"); 203 static int mwl_rxdmalow = 3; /* # min buffers for wakeup */ 204 SYSCTL_INT(_hw_mwl, OID_AUTO, rxdmalow, CTLFLAG_RWTUN, &mwl_rxdmalow, 205 0, "min free rx buffers before restarting traffic"); 206 207 #ifdef MWL_DEBUG 208 static int mwl_debug = 0; 209 SYSCTL_INT(_hw_mwl, OID_AUTO, debug, CTLFLAG_RWTUN, &mwl_debug, 210 0, "control debugging printfs"); 211 enum { 212 MWL_DEBUG_XMIT = 0x00000001, /* basic xmit operation */ 213 MWL_DEBUG_XMIT_DESC = 0x00000002, /* xmit descriptors */ 214 MWL_DEBUG_RECV = 0x00000004, /* basic recv operation */ 215 MWL_DEBUG_RECV_DESC = 0x00000008, /* recv descriptors */ 216 MWL_DEBUG_RESET = 0x00000010, /* reset processing */ 217 MWL_DEBUG_BEACON = 0x00000020, /* beacon handling */ 218 MWL_DEBUG_INTR = 0x00000040, /* ISR */ 219 MWL_DEBUG_TX_PROC = 0x00000080, /* tx ISR proc */ 220 MWL_DEBUG_RX_PROC = 0x00000100, /* rx ISR proc */ 221 MWL_DEBUG_KEYCACHE = 0x00000200, /* key cache management */ 222 MWL_DEBUG_STATE = 0x00000400, /* 802.11 state transitions */ 223 MWL_DEBUG_NODE = 0x00000800, /* node management */ 224 MWL_DEBUG_RECV_ALL = 0x00001000, /* trace all frames (beacons) */ 225 MWL_DEBUG_TSO = 0x00002000, /* TSO processing */ 226 MWL_DEBUG_AMPDU = 0x00004000, /* BA stream handling */ 227 MWL_DEBUG_ANY = 0xffffffff 228 }; 229 #define IS_BEACON(wh) \ 230 ((wh->i_fc[0] & (IEEE80211_FC0_TYPE_MASK|IEEE80211_FC0_SUBTYPE_MASK)) == \ 231 (IEEE80211_FC0_TYPE_MGT|IEEE80211_FC0_SUBTYPE_BEACON)) 232 #define IFF_DUMPPKTS_RECV(sc, wh) \ 233 ((sc->sc_debug & MWL_DEBUG_RECV) && \ 234 ((sc->sc_debug & MWL_DEBUG_RECV_ALL) || !IS_BEACON(wh))) 235 #define IFF_DUMPPKTS_XMIT(sc) \ 236 (sc->sc_debug & MWL_DEBUG_XMIT) 237 238 #define DPRINTF(sc, m, fmt, ...) do { \ 239 if (sc->sc_debug & (m)) \ 240 printf(fmt, __VA_ARGS__); \ 241 } while (0) 242 #define KEYPRINTF(sc, hk, mac) do { \ 243 if (sc->sc_debug & MWL_DEBUG_KEYCACHE) \ 244 mwl_keyprint(sc, __func__, hk, mac); \ 245 } while (0) 246 static void mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix); 247 static void mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix); 248 #else 249 #define IFF_DUMPPKTS_RECV(sc, wh) 0 250 #define IFF_DUMPPKTS_XMIT(sc) 0 251 #define DPRINTF(sc, m, fmt, ...) do { (void )sc; } while (0) 252 #define KEYPRINTF(sc, k, mac) do { (void )sc; } while (0) 253 #endif 254 255 static MALLOC_DEFINE(M_MWLDEV, "mwldev", "mwl driver dma buffers"); 256 257 /* 258 * Each packet has fixed front matter: a 2-byte length 259 * of the payload, followed by a 4-address 802.11 header 260 * (regardless of the actual header and always w/o any 261 * QoS header). The payload then follows. 262 */ 263 struct mwltxrec { 264 uint16_t fwlen; 265 struct ieee80211_frame_addr4 wh; 266 } __packed; 267 268 /* 269 * Read/Write shorthands for accesses to BAR 0. Note 270 * that all BAR 1 operations are done in the "hal" and 271 * there should be no reference to them here. 272 */ 273 #ifdef MWL_DEBUG 274 static __inline uint32_t 275 RD4(struct mwl_softc *sc, bus_size_t off) 276 { 277 return bus_space_read_4(sc->sc_io0t, sc->sc_io0h, off); 278 } 279 #endif 280 281 static __inline void 282 WR4(struct mwl_softc *sc, bus_size_t off, uint32_t val) 283 { 284 bus_space_write_4(sc->sc_io0t, sc->sc_io0h, off, val); 285 } 286 287 int 288 mwl_attach(uint16_t devid, struct mwl_softc *sc) 289 { 290 struct ieee80211com *ic = &sc->sc_ic; 291 struct mwl_hal *mh; 292 int error = 0; 293 294 DPRINTF(sc, MWL_DEBUG_ANY, "%s: devid 0x%x\n", __func__, devid); 295 296 /* 297 * Setup the RX free list lock early, so it can be consistently 298 * removed. 299 */ 300 MWL_RXFREE_INIT(sc); 301 302 mh = mwl_hal_attach(sc->sc_dev, devid, 303 sc->sc_io1h, sc->sc_io1t, sc->sc_dmat); 304 if (mh == NULL) { 305 device_printf(sc->sc_dev, "unable to attach HAL\n"); 306 error = EIO; 307 goto bad; 308 } 309 sc->sc_mh = mh; 310 /* 311 * Load firmware so we can get setup. We arbitrarily 312 * pick station firmware; we'll re-load firmware as 313 * needed so setting up the wrong mode isn't a big deal. 314 */ 315 if (mwl_hal_fwload(mh, NULL) != 0) { 316 device_printf(sc->sc_dev, "unable to setup builtin firmware\n"); 317 error = EIO; 318 goto bad1; 319 } 320 if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) { 321 device_printf(sc->sc_dev, "unable to fetch h/w specs\n"); 322 error = EIO; 323 goto bad1; 324 } 325 error = mwl_getchannels(sc); 326 if (error != 0) 327 goto bad1; 328 329 sc->sc_txantenna = 0; /* h/w default */ 330 sc->sc_rxantenna = 0; /* h/w default */ 331 sc->sc_invalid = 0; /* ready to go, enable int handling */ 332 sc->sc_ageinterval = MWL_AGEINTERVAL; 333 334 /* 335 * Allocate tx+rx descriptors and populate the lists. 336 * We immediately push the information to the firmware 337 * as otherwise it gets upset. 338 */ 339 error = mwl_dma_setup(sc); 340 if (error != 0) { 341 device_printf(sc->sc_dev, "failed to setup descriptors: %d\n", 342 error); 343 goto bad1; 344 } 345 error = mwl_setupdma(sc); /* push to firmware */ 346 if (error != 0) /* NB: mwl_setupdma prints msg */ 347 goto bad1; 348 349 callout_init(&sc->sc_timer, 1); 350 callout_init_mtx(&sc->sc_watchdog, &sc->sc_mtx, 0); 351 mbufq_init(&sc->sc_snd, ifqmaxlen); 352 353 sc->sc_tq = taskqueue_create("mwl_taskq", M_NOWAIT, 354 taskqueue_thread_enqueue, &sc->sc_tq); 355 taskqueue_start_threads(&sc->sc_tq, 1, PI_NET, 356 "%s taskq", device_get_nameunit(sc->sc_dev)); 357 358 NET_TASK_INIT(&sc->sc_rxtask, 0, mwl_rx_proc, sc); 359 TASK_INIT(&sc->sc_radartask, 0, mwl_radar_proc, sc); 360 TASK_INIT(&sc->sc_chanswitchtask, 0, mwl_chanswitch_proc, sc); 361 TASK_INIT(&sc->sc_bawatchdogtask, 0, mwl_bawatchdog_proc, sc); 362 363 /* NB: insure BK queue is the lowest priority h/w queue */ 364 if (!mwl_tx_setup(sc, WME_AC_BK, MWL_WME_AC_BK)) { 365 device_printf(sc->sc_dev, 366 "unable to setup xmit queue for %s traffic!\n", 367 ieee80211_wme_acnames[WME_AC_BK]); 368 error = EIO; 369 goto bad2; 370 } 371 if (!mwl_tx_setup(sc, WME_AC_BE, MWL_WME_AC_BE) || 372 !mwl_tx_setup(sc, WME_AC_VI, MWL_WME_AC_VI) || 373 !mwl_tx_setup(sc, WME_AC_VO, MWL_WME_AC_VO)) { 374 /* 375 * Not enough hardware tx queues to properly do WME; 376 * just punt and assign them all to the same h/w queue. 377 * We could do a better job of this if, for example, 378 * we allocate queues when we switch from station to 379 * AP mode. 380 */ 381 if (sc->sc_ac2q[WME_AC_VI] != NULL) 382 mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_VI]); 383 if (sc->sc_ac2q[WME_AC_BE] != NULL) 384 mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_BE]); 385 sc->sc_ac2q[WME_AC_BE] = sc->sc_ac2q[WME_AC_BK]; 386 sc->sc_ac2q[WME_AC_VI] = sc->sc_ac2q[WME_AC_BK]; 387 sc->sc_ac2q[WME_AC_VO] = sc->sc_ac2q[WME_AC_BK]; 388 } 389 TASK_INIT(&sc->sc_txtask, 0, mwl_tx_proc, sc); 390 391 ic->ic_softc = sc; 392 ic->ic_name = device_get_nameunit(sc->sc_dev); 393 /* XXX not right but it's not used anywhere important */ 394 ic->ic_phytype = IEEE80211_T_OFDM; 395 ic->ic_opmode = IEEE80211_M_STA; 396 ic->ic_caps = 397 IEEE80211_C_STA /* station mode supported */ 398 | IEEE80211_C_HOSTAP /* hostap mode */ 399 | IEEE80211_C_MONITOR /* monitor mode */ 400 #if 0 401 | IEEE80211_C_IBSS /* ibss, nee adhoc, mode */ 402 | IEEE80211_C_AHDEMO /* adhoc demo mode */ 403 #endif 404 | IEEE80211_C_MBSS /* mesh point link mode */ 405 | IEEE80211_C_WDS /* WDS supported */ 406 | IEEE80211_C_SHPREAMBLE /* short preamble supported */ 407 | IEEE80211_C_SHSLOT /* short slot time supported */ 408 | IEEE80211_C_WME /* WME/WMM supported */ 409 | IEEE80211_C_BURST /* xmit bursting supported */ 410 | IEEE80211_C_WPA /* capable of WPA1+WPA2 */ 411 | IEEE80211_C_BGSCAN /* capable of bg scanning */ 412 | IEEE80211_C_TXFRAG /* handle tx frags */ 413 | IEEE80211_C_TXPMGT /* capable of txpow mgt */ 414 | IEEE80211_C_DFS /* DFS supported */ 415 ; 416 417 ic->ic_htcaps = 418 IEEE80211_HTCAP_SMPS_ENA /* SM PS mode enabled */ 419 | IEEE80211_HTCAP_CHWIDTH40 /* 40MHz channel width */ 420 | IEEE80211_HTCAP_SHORTGI20 /* short GI in 20MHz */ 421 | IEEE80211_HTCAP_SHORTGI40 /* short GI in 40MHz */ 422 | IEEE80211_HTCAP_RXSTBC_2STREAM/* 1-2 spatial streams */ 423 #if MWL_AGGR_SIZE == 7935 424 | IEEE80211_HTCAP_MAXAMSDU_7935 /* max A-MSDU length */ 425 #else 426 | IEEE80211_HTCAP_MAXAMSDU_3839 /* max A-MSDU length */ 427 #endif 428 #if 0 429 | IEEE80211_HTCAP_PSMP /* PSMP supported */ 430 | IEEE80211_HTCAP_40INTOLERANT /* 40MHz intolerant */ 431 #endif 432 /* s/w capabilities */ 433 | IEEE80211_HTC_HT /* HT operation */ 434 | IEEE80211_HTC_AMPDU /* tx A-MPDU */ 435 | IEEE80211_HTC_AMSDU /* tx A-MSDU */ 436 | IEEE80211_HTC_SMPS /* SMPS available */ 437 ; 438 439 /* 440 * Mark h/w crypto support. 441 * XXX no way to query h/w support. 442 */ 443 ic->ic_cryptocaps |= IEEE80211_CRYPTO_WEP 444 | IEEE80211_CRYPTO_AES_CCM 445 | IEEE80211_CRYPTO_TKIP 446 | IEEE80211_CRYPTO_TKIPMIC 447 ; 448 /* 449 * Transmit requires space in the packet for a special 450 * format transmit record and optional padding between 451 * this record and the payload. Ask the net80211 layer 452 * to arrange this when encapsulating packets so we can 453 * add it efficiently. 454 */ 455 ic->ic_headroom = sizeof(struct mwltxrec) - 456 sizeof(struct ieee80211_frame); 457 458 IEEE80211_ADDR_COPY(ic->ic_macaddr, sc->sc_hwspecs.macAddr); 459 460 /* call MI attach routine. */ 461 ieee80211_ifattach(ic); 462 ic->ic_setregdomain = mwl_setregdomain; 463 ic->ic_getradiocaps = mwl_getradiocaps; 464 /* override default methods */ 465 ic->ic_raw_xmit = mwl_raw_xmit; 466 ic->ic_newassoc = mwl_newassoc; 467 ic->ic_updateslot = mwl_updateslot; 468 ic->ic_update_mcast = mwl_update_mcast; 469 ic->ic_update_promisc = mwl_update_promisc; 470 ic->ic_wme.wme_update = mwl_wme_update; 471 ic->ic_transmit = mwl_transmit; 472 ic->ic_ioctl = mwl_ioctl; 473 ic->ic_parent = mwl_parent; 474 475 ic->ic_node_alloc = mwl_node_alloc; 476 sc->sc_node_cleanup = ic->ic_node_cleanup; 477 ic->ic_node_cleanup = mwl_node_cleanup; 478 sc->sc_node_drain = ic->ic_node_drain; 479 ic->ic_node_drain = mwl_node_drain; 480 ic->ic_node_getsignal = mwl_node_getsignal; 481 ic->ic_node_getmimoinfo = mwl_node_getmimoinfo; 482 483 ic->ic_scan_start = mwl_scan_start; 484 ic->ic_scan_end = mwl_scan_end; 485 ic->ic_set_channel = mwl_set_channel; 486 487 sc->sc_recv_action = ic->ic_recv_action; 488 ic->ic_recv_action = mwl_recv_action; 489 sc->sc_addba_request = ic->ic_addba_request; 490 ic->ic_addba_request = mwl_addba_request; 491 sc->sc_addba_response = ic->ic_addba_response; 492 ic->ic_addba_response = mwl_addba_response; 493 sc->sc_addba_stop = ic->ic_addba_stop; 494 ic->ic_addba_stop = mwl_addba_stop; 495 496 ic->ic_vap_create = mwl_vap_create; 497 ic->ic_vap_delete = mwl_vap_delete; 498 499 ieee80211_radiotap_attach(ic, 500 &sc->sc_tx_th.wt_ihdr, sizeof(sc->sc_tx_th), 501 MWL_TX_RADIOTAP_PRESENT, 502 &sc->sc_rx_th.wr_ihdr, sizeof(sc->sc_rx_th), 503 MWL_RX_RADIOTAP_PRESENT); 504 /* 505 * Setup dynamic sysctl's now that country code and 506 * regdomain are available from the hal. 507 */ 508 mwl_sysctlattach(sc); 509 510 if (bootverbose) 511 ieee80211_announce(ic); 512 mwl_announce(sc); 513 return 0; 514 bad2: 515 mwl_dma_cleanup(sc); 516 bad1: 517 mwl_hal_detach(mh); 518 bad: 519 MWL_RXFREE_DESTROY(sc); 520 sc->sc_invalid = 1; 521 return error; 522 } 523 524 int 525 mwl_detach(struct mwl_softc *sc) 526 { 527 struct ieee80211com *ic = &sc->sc_ic; 528 529 MWL_LOCK(sc); 530 mwl_stop(sc); 531 MWL_UNLOCK(sc); 532 /* 533 * NB: the order of these is important: 534 * o call the 802.11 layer before detaching the hal to 535 * insure callbacks into the driver to delete global 536 * key cache entries can be handled 537 * o reclaim the tx queue data structures after calling 538 * the 802.11 layer as we'll get called back to reclaim 539 * node state and potentially want to use them 540 * o to cleanup the tx queues the hal is called, so detach 541 * it last 542 * Other than that, it's straightforward... 543 */ 544 ieee80211_ifdetach(ic); 545 callout_drain(&sc->sc_watchdog); 546 mwl_dma_cleanup(sc); 547 MWL_RXFREE_DESTROY(sc); 548 mwl_tx_cleanup(sc); 549 mwl_hal_detach(sc->sc_mh); 550 mbufq_drain(&sc->sc_snd); 551 552 return 0; 553 } 554 555 /* 556 * MAC address handling for multiple BSS on the same radio. 557 * The first vap uses the MAC address from the EEPROM. For 558 * subsequent vap's we set the U/L bit (bit 1) in the MAC 559 * address and use the next six bits as an index. 560 */ 561 static void 562 assign_address(struct mwl_softc *sc, uint8_t mac[IEEE80211_ADDR_LEN], int clone) 563 { 564 int i; 565 566 if (clone && mwl_hal_ismbsscapable(sc->sc_mh)) { 567 /* NB: we only do this if h/w supports multiple bssid */ 568 for (i = 0; i < 32; i++) 569 if ((sc->sc_bssidmask & (1<<i)) == 0) 570 break; 571 if (i != 0) 572 mac[0] |= (i << 2)|0x2; 573 } else 574 i = 0; 575 sc->sc_bssidmask |= 1<<i; 576 if (i == 0) 577 sc->sc_nbssid0++; 578 } 579 580 static void 581 reclaim_address(struct mwl_softc *sc, const uint8_t mac[IEEE80211_ADDR_LEN]) 582 { 583 int i = mac[0] >> 2; 584 if (i != 0 || --sc->sc_nbssid0 == 0) 585 sc->sc_bssidmask &= ~(1<<i); 586 } 587 588 static struct ieee80211vap * 589 mwl_vap_create(struct ieee80211com *ic, const char name[IFNAMSIZ], int unit, 590 enum ieee80211_opmode opmode, int flags, 591 const uint8_t bssid[IEEE80211_ADDR_LEN], 592 const uint8_t mac0[IEEE80211_ADDR_LEN]) 593 { 594 struct mwl_softc *sc = ic->ic_softc; 595 struct mwl_hal *mh = sc->sc_mh; 596 struct ieee80211vap *vap, *apvap; 597 struct mwl_hal_vap *hvap; 598 struct mwl_vap *mvp; 599 uint8_t mac[IEEE80211_ADDR_LEN]; 600 601 IEEE80211_ADDR_COPY(mac, mac0); 602 switch (opmode) { 603 case IEEE80211_M_HOSTAP: 604 case IEEE80211_M_MBSS: 605 if ((flags & IEEE80211_CLONE_MACADDR) == 0) 606 assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID); 607 hvap = mwl_hal_newvap(mh, MWL_HAL_AP, mac); 608 if (hvap == NULL) { 609 if ((flags & IEEE80211_CLONE_MACADDR) == 0) 610 reclaim_address(sc, mac); 611 return NULL; 612 } 613 break; 614 case IEEE80211_M_STA: 615 if ((flags & IEEE80211_CLONE_MACADDR) == 0) 616 assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID); 617 hvap = mwl_hal_newvap(mh, MWL_HAL_STA, mac); 618 if (hvap == NULL) { 619 if ((flags & IEEE80211_CLONE_MACADDR) == 0) 620 reclaim_address(sc, mac); 621 return NULL; 622 } 623 /* no h/w beacon miss support; always use s/w */ 624 flags |= IEEE80211_CLONE_NOBEACONS; 625 break; 626 case IEEE80211_M_WDS: 627 hvap = NULL; /* NB: we use associated AP vap */ 628 if (sc->sc_napvaps == 0) 629 return NULL; /* no existing AP vap */ 630 break; 631 case IEEE80211_M_MONITOR: 632 hvap = NULL; 633 break; 634 case IEEE80211_M_IBSS: 635 case IEEE80211_M_AHDEMO: 636 default: 637 return NULL; 638 } 639 640 mvp = malloc(sizeof(struct mwl_vap), M_80211_VAP, M_WAITOK | M_ZERO); 641 mvp->mv_hvap = hvap; 642 if (opmode == IEEE80211_M_WDS) { 643 /* 644 * WDS vaps must have an associated AP vap; find one. 645 * XXX not right. 646 */ 647 TAILQ_FOREACH(apvap, &ic->ic_vaps, iv_next) 648 if (apvap->iv_opmode == IEEE80211_M_HOSTAP) { 649 mvp->mv_ap_hvap = MWL_VAP(apvap)->mv_hvap; 650 break; 651 } 652 KASSERT(mvp->mv_ap_hvap != NULL, ("no ap vap")); 653 } 654 vap = &mvp->mv_vap; 655 ieee80211_vap_setup(ic, vap, name, unit, opmode, flags, bssid); 656 /* override with driver methods */ 657 mvp->mv_newstate = vap->iv_newstate; 658 vap->iv_newstate = mwl_newstate; 659 vap->iv_max_keyix = 0; /* XXX */ 660 vap->iv_key_alloc = mwl_key_alloc; 661 vap->iv_key_delete = mwl_key_delete; 662 vap->iv_key_set = mwl_key_set; 663 #ifdef MWL_HOST_PS_SUPPORT 664 if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS) { 665 vap->iv_update_ps = mwl_update_ps; 666 mvp->mv_set_tim = vap->iv_set_tim; 667 vap->iv_set_tim = mwl_set_tim; 668 } 669 #endif 670 vap->iv_reset = mwl_reset; 671 vap->iv_update_beacon = mwl_beacon_update; 672 673 /* override max aid so sta's cannot assoc when we're out of sta id's */ 674 vap->iv_max_aid = MWL_MAXSTAID; 675 /* override default A-MPDU rx parameters */ 676 vap->iv_ampdu_rxmax = IEEE80211_HTCAP_MAXRXAMPDU_64K; 677 vap->iv_ampdu_density = IEEE80211_HTCAP_MPDUDENSITY_4; 678 679 /* complete setup */ 680 ieee80211_vap_attach(vap, mwl_media_change, ieee80211_media_status, 681 mac); 682 683 switch (vap->iv_opmode) { 684 case IEEE80211_M_HOSTAP: 685 case IEEE80211_M_MBSS: 686 case IEEE80211_M_STA: 687 /* 688 * Setup sta db entry for local address. 689 */ 690 mwl_localstadb(vap); 691 if (vap->iv_opmode == IEEE80211_M_HOSTAP || 692 vap->iv_opmode == IEEE80211_M_MBSS) 693 sc->sc_napvaps++; 694 else 695 sc->sc_nstavaps++; 696 break; 697 case IEEE80211_M_WDS: 698 sc->sc_nwdsvaps++; 699 break; 700 default: 701 break; 702 } 703 /* 704 * Setup overall operating mode. 705 */ 706 if (sc->sc_napvaps) 707 ic->ic_opmode = IEEE80211_M_HOSTAP; 708 else if (sc->sc_nstavaps) 709 ic->ic_opmode = IEEE80211_M_STA; 710 else 711 ic->ic_opmode = opmode; 712 713 return vap; 714 } 715 716 static void 717 mwl_vap_delete(struct ieee80211vap *vap) 718 { 719 struct mwl_vap *mvp = MWL_VAP(vap); 720 struct mwl_softc *sc = vap->iv_ic->ic_softc; 721 struct mwl_hal *mh = sc->sc_mh; 722 struct mwl_hal_vap *hvap = mvp->mv_hvap; 723 enum ieee80211_opmode opmode = vap->iv_opmode; 724 725 /* XXX disallow ap vap delete if WDS still present */ 726 if (sc->sc_running) { 727 /* quiesce h/w while we remove the vap */ 728 mwl_hal_intrset(mh, 0); /* disable interrupts */ 729 } 730 ieee80211_vap_detach(vap); 731 switch (opmode) { 732 case IEEE80211_M_HOSTAP: 733 case IEEE80211_M_MBSS: 734 case IEEE80211_M_STA: 735 KASSERT(hvap != NULL, ("no hal vap handle")); 736 (void) mwl_hal_delstation(hvap, vap->iv_myaddr); 737 mwl_hal_delvap(hvap); 738 if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS) 739 sc->sc_napvaps--; 740 else 741 sc->sc_nstavaps--; 742 /* XXX don't do it for IEEE80211_CLONE_MACADDR */ 743 reclaim_address(sc, vap->iv_myaddr); 744 break; 745 case IEEE80211_M_WDS: 746 sc->sc_nwdsvaps--; 747 break; 748 default: 749 break; 750 } 751 mwl_cleartxq(sc, vap); 752 free(mvp, M_80211_VAP); 753 if (sc->sc_running) 754 mwl_hal_intrset(mh, sc->sc_imask); 755 } 756 757 void 758 mwl_suspend(struct mwl_softc *sc) 759 { 760 761 MWL_LOCK(sc); 762 mwl_stop(sc); 763 MWL_UNLOCK(sc); 764 } 765 766 void 767 mwl_resume(struct mwl_softc *sc) 768 { 769 int error = EDOOFUS; 770 771 MWL_LOCK(sc); 772 if (sc->sc_ic.ic_nrunning > 0) 773 error = mwl_init(sc); 774 MWL_UNLOCK(sc); 775 776 if (error == 0) 777 ieee80211_start_all(&sc->sc_ic); /* start all vap's */ 778 } 779 780 void 781 mwl_shutdown(void *arg) 782 { 783 struct mwl_softc *sc = arg; 784 785 MWL_LOCK(sc); 786 mwl_stop(sc); 787 MWL_UNLOCK(sc); 788 } 789 790 /* 791 * Interrupt handler. Most of the actual processing is deferred. 792 */ 793 void 794 mwl_intr(void *arg) 795 { 796 struct mwl_softc *sc = arg; 797 struct mwl_hal *mh = sc->sc_mh; 798 uint32_t status; 799 800 if (sc->sc_invalid) { 801 /* 802 * The hardware is not ready/present, don't touch anything. 803 * Note this can happen early on if the IRQ is shared. 804 */ 805 DPRINTF(sc, MWL_DEBUG_ANY, "%s: invalid; ignored\n", __func__); 806 return; 807 } 808 /* 809 * Figure out the reason(s) for the interrupt. 810 */ 811 mwl_hal_getisr(mh, &status); /* NB: clears ISR too */ 812 if (status == 0) /* must be a shared irq */ 813 return; 814 815 DPRINTF(sc, MWL_DEBUG_INTR, "%s: status 0x%x imask 0x%x\n", 816 __func__, status, sc->sc_imask); 817 if (status & MACREG_A2HRIC_BIT_RX_RDY) 818 taskqueue_enqueue(sc->sc_tq, &sc->sc_rxtask); 819 if (status & MACREG_A2HRIC_BIT_TX_DONE) 820 taskqueue_enqueue(sc->sc_tq, &sc->sc_txtask); 821 if (status & MACREG_A2HRIC_BIT_BA_WATCHDOG) 822 taskqueue_enqueue(sc->sc_tq, &sc->sc_bawatchdogtask); 823 if (status & MACREG_A2HRIC_BIT_OPC_DONE) 824 mwl_hal_cmddone(mh); 825 if (status & MACREG_A2HRIC_BIT_MAC_EVENT) { 826 ; 827 } 828 if (status & MACREG_A2HRIC_BIT_ICV_ERROR) { 829 /* TKIP ICV error */ 830 sc->sc_stats.mst_rx_badtkipicv++; 831 } 832 if (status & MACREG_A2HRIC_BIT_QUEUE_EMPTY) { 833 /* 11n aggregation queue is empty, re-fill */ 834 ; 835 } 836 if (status & MACREG_A2HRIC_BIT_QUEUE_FULL) { 837 ; 838 } 839 if (status & MACREG_A2HRIC_BIT_RADAR_DETECT) { 840 /* radar detected, process event */ 841 taskqueue_enqueue(sc->sc_tq, &sc->sc_radartask); 842 } 843 if (status & MACREG_A2HRIC_BIT_CHAN_SWITCH) { 844 /* DFS channel switch */ 845 taskqueue_enqueue(sc->sc_tq, &sc->sc_chanswitchtask); 846 } 847 } 848 849 static void 850 mwl_radar_proc(void *arg, int pending) 851 { 852 struct mwl_softc *sc = arg; 853 struct ieee80211com *ic = &sc->sc_ic; 854 855 DPRINTF(sc, MWL_DEBUG_ANY, "%s: radar detected, pending %u\n", 856 __func__, pending); 857 858 sc->sc_stats.mst_radardetect++; 859 /* XXX stop h/w BA streams? */ 860 861 IEEE80211_LOCK(ic); 862 ieee80211_dfs_notify_radar(ic, ic->ic_curchan); 863 IEEE80211_UNLOCK(ic); 864 } 865 866 static void 867 mwl_chanswitch_proc(void *arg, int pending) 868 { 869 struct mwl_softc *sc = arg; 870 struct ieee80211com *ic = &sc->sc_ic; 871 872 DPRINTF(sc, MWL_DEBUG_ANY, "%s: channel switch notice, pending %u\n", 873 __func__, pending); 874 875 IEEE80211_LOCK(ic); 876 sc->sc_csapending = 0; 877 ieee80211_csa_completeswitch(ic); 878 IEEE80211_UNLOCK(ic); 879 } 880 881 static void 882 mwl_bawatchdog(const MWL_HAL_BASTREAM *sp) 883 { 884 struct ieee80211_node *ni = sp->data[0]; 885 886 /* send DELBA and drop the stream */ 887 ieee80211_ampdu_stop(ni, sp->data[1], IEEE80211_REASON_UNSPECIFIED); 888 } 889 890 static void 891 mwl_bawatchdog_proc(void *arg, int pending) 892 { 893 struct mwl_softc *sc = arg; 894 struct mwl_hal *mh = sc->sc_mh; 895 const MWL_HAL_BASTREAM *sp; 896 uint8_t bitmap, n; 897 898 sc->sc_stats.mst_bawatchdog++; 899 900 if (mwl_hal_getwatchdogbitmap(mh, &bitmap) != 0) { 901 DPRINTF(sc, MWL_DEBUG_AMPDU, 902 "%s: could not get bitmap\n", __func__); 903 sc->sc_stats.mst_bawatchdog_failed++; 904 return; 905 } 906 DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: bitmap 0x%x\n", __func__, bitmap); 907 if (bitmap == 0xff) { 908 n = 0; 909 /* disable all ba streams */ 910 for (bitmap = 0; bitmap < 8; bitmap++) { 911 sp = mwl_hal_bastream_lookup(mh, bitmap); 912 if (sp != NULL) { 913 mwl_bawatchdog(sp); 914 n++; 915 } 916 } 917 if (n == 0) { 918 DPRINTF(sc, MWL_DEBUG_AMPDU, 919 "%s: no BA streams found\n", __func__); 920 sc->sc_stats.mst_bawatchdog_empty++; 921 } 922 } else if (bitmap != 0xaa) { 923 /* disable a single ba stream */ 924 sp = mwl_hal_bastream_lookup(mh, bitmap); 925 if (sp != NULL) { 926 mwl_bawatchdog(sp); 927 } else { 928 DPRINTF(sc, MWL_DEBUG_AMPDU, 929 "%s: no BA stream %d\n", __func__, bitmap); 930 sc->sc_stats.mst_bawatchdog_notfound++; 931 } 932 } 933 } 934 935 /* 936 * Convert net80211 channel to a HAL channel. 937 */ 938 static void 939 mwl_mapchan(MWL_HAL_CHANNEL *hc, const struct ieee80211_channel *chan) 940 { 941 hc->channel = chan->ic_ieee; 942 943 *(uint32_t *)&hc->channelFlags = 0; 944 if (IEEE80211_IS_CHAN_2GHZ(chan)) 945 hc->channelFlags.FreqBand = MWL_FREQ_BAND_2DOT4GHZ; 946 else if (IEEE80211_IS_CHAN_5GHZ(chan)) 947 hc->channelFlags.FreqBand = MWL_FREQ_BAND_5GHZ; 948 if (IEEE80211_IS_CHAN_HT40(chan)) { 949 hc->channelFlags.ChnlWidth = MWL_CH_40_MHz_WIDTH; 950 if (IEEE80211_IS_CHAN_HT40U(chan)) 951 hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_ABOVE_CTRL_CH; 952 else 953 hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_BELOW_CTRL_CH; 954 } else 955 hc->channelFlags.ChnlWidth = MWL_CH_20_MHz_WIDTH; 956 /* XXX 10MHz channels */ 957 } 958 959 /* 960 * Inform firmware of our tx/rx dma setup. The BAR 0 961 * writes below are for compatibility with older firmware. 962 * For current firmware we send this information with a 963 * cmd block via mwl_hal_sethwdma. 964 */ 965 static int 966 mwl_setupdma(struct mwl_softc *sc) 967 { 968 int error, i; 969 970 sc->sc_hwdma.rxDescRead = sc->sc_rxdma.dd_desc_paddr; 971 WR4(sc, sc->sc_hwspecs.rxDescRead, sc->sc_hwdma.rxDescRead); 972 WR4(sc, sc->sc_hwspecs.rxDescWrite, sc->sc_hwdma.rxDescRead); 973 974 for (i = 0; i < MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES; i++) { 975 struct mwl_txq *txq = &sc->sc_txq[i]; 976 sc->sc_hwdma.wcbBase[i] = txq->dma.dd_desc_paddr; 977 WR4(sc, sc->sc_hwspecs.wcbBase[i], sc->sc_hwdma.wcbBase[i]); 978 } 979 sc->sc_hwdma.maxNumTxWcb = mwl_txbuf; 980 sc->sc_hwdma.maxNumWCB = MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES; 981 982 error = mwl_hal_sethwdma(sc->sc_mh, &sc->sc_hwdma); 983 if (error != 0) { 984 device_printf(sc->sc_dev, 985 "unable to setup tx/rx dma; hal status %u\n", error); 986 /* XXX */ 987 } 988 return error; 989 } 990 991 /* 992 * Inform firmware of tx rate parameters. 993 * Called after a channel change. 994 */ 995 static int 996 mwl_setcurchanrates(struct mwl_softc *sc) 997 { 998 struct ieee80211com *ic = &sc->sc_ic; 999 const struct ieee80211_rateset *rs; 1000 MWL_HAL_TXRATE rates; 1001 1002 memset(&rates, 0, sizeof(rates)); 1003 rs = ieee80211_get_suprates(ic, ic->ic_curchan); 1004 /* rate used to send management frames */ 1005 rates.MgtRate = rs->rs_rates[0] & IEEE80211_RATE_VAL; 1006 /* rate used to send multicast frames */ 1007 rates.McastRate = rates.MgtRate; 1008 1009 return mwl_hal_settxrate_auto(sc->sc_mh, &rates); 1010 } 1011 1012 /* 1013 * Inform firmware of tx rate parameters. Called whenever 1014 * user-settable params change and after a channel change. 1015 */ 1016 static int 1017 mwl_setrates(struct ieee80211vap *vap) 1018 { 1019 struct mwl_vap *mvp = MWL_VAP(vap); 1020 struct ieee80211_node *ni = vap->iv_bss; 1021 const struct ieee80211_txparam *tp = ni->ni_txparms; 1022 MWL_HAL_TXRATE rates; 1023 1024 KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state)); 1025 1026 /* 1027 * Update the h/w rate map. 1028 * NB: 0x80 for MCS is passed through unchanged 1029 */ 1030 memset(&rates, 0, sizeof(rates)); 1031 /* rate used to send management frames */ 1032 rates.MgtRate = tp->mgmtrate; 1033 /* rate used to send multicast frames */ 1034 rates.McastRate = tp->mcastrate; 1035 1036 /* while here calculate EAPOL fixed rate cookie */ 1037 mvp->mv_eapolformat = htole16(mwl_calcformat(rates.MgtRate, ni)); 1038 1039 return mwl_hal_settxrate(mvp->mv_hvap, 1040 tp->ucastrate != IEEE80211_FIXED_RATE_NONE ? 1041 RATE_FIXED : RATE_AUTO, &rates); 1042 } 1043 1044 /* 1045 * Setup a fixed xmit rate cookie for EAPOL frames. 1046 */ 1047 static void 1048 mwl_seteapolformat(struct ieee80211vap *vap) 1049 { 1050 struct mwl_vap *mvp = MWL_VAP(vap); 1051 struct ieee80211_node *ni = vap->iv_bss; 1052 enum ieee80211_phymode mode; 1053 uint8_t rate; 1054 1055 KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state)); 1056 1057 mode = ieee80211_chan2mode(ni->ni_chan); 1058 /* 1059 * Use legacy rates when operating a mixed HT+non-HT bss. 1060 * NB: this may violate POLA for sta and wds vap's. 1061 */ 1062 if (mode == IEEE80211_MODE_11NA && 1063 (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0) 1064 rate = vap->iv_txparms[IEEE80211_MODE_11A].mgmtrate; 1065 else if (mode == IEEE80211_MODE_11NG && 1066 (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0) 1067 rate = vap->iv_txparms[IEEE80211_MODE_11G].mgmtrate; 1068 else 1069 rate = vap->iv_txparms[mode].mgmtrate; 1070 1071 mvp->mv_eapolformat = htole16(mwl_calcformat(rate, ni)); 1072 } 1073 1074 /* 1075 * Map SKU+country code to region code for radar bin'ing. 1076 */ 1077 static int 1078 mwl_map2regioncode(const struct ieee80211_regdomain *rd) 1079 { 1080 switch (rd->regdomain) { 1081 case SKU_FCC: 1082 case SKU_FCC3: 1083 return DOMAIN_CODE_FCC; 1084 case SKU_CA: 1085 return DOMAIN_CODE_IC; 1086 case SKU_ETSI: 1087 case SKU_ETSI2: 1088 case SKU_ETSI3: 1089 if (rd->country == CTRY_SPAIN) 1090 return DOMAIN_CODE_SPAIN; 1091 if (rd->country == CTRY_FRANCE || rd->country == CTRY_FRANCE2) 1092 return DOMAIN_CODE_FRANCE; 1093 /* XXX force 1.3.1 radar type */ 1094 return DOMAIN_CODE_ETSI_131; 1095 case SKU_JAPAN: 1096 return DOMAIN_CODE_MKK; 1097 case SKU_ROW: 1098 return DOMAIN_CODE_DGT; /* Taiwan */ 1099 case SKU_APAC: 1100 case SKU_APAC2: 1101 case SKU_APAC3: 1102 return DOMAIN_CODE_AUS; /* Australia */ 1103 } 1104 /* XXX KOREA? */ 1105 return DOMAIN_CODE_FCC; /* XXX? */ 1106 } 1107 1108 static int 1109 mwl_hal_reset(struct mwl_softc *sc) 1110 { 1111 struct ieee80211com *ic = &sc->sc_ic; 1112 struct mwl_hal *mh = sc->sc_mh; 1113 1114 mwl_hal_setantenna(mh, WL_ANTENNATYPE_RX, sc->sc_rxantenna); 1115 mwl_hal_setantenna(mh, WL_ANTENNATYPE_TX, sc->sc_txantenna); 1116 mwl_hal_setradio(mh, 1, WL_AUTO_PREAMBLE); 1117 mwl_hal_setwmm(sc->sc_mh, (ic->ic_flags & IEEE80211_F_WME) != 0); 1118 mwl_chan_set(sc, ic->ic_curchan); 1119 /* NB: RF/RA performance tuned for indoor mode */ 1120 mwl_hal_setrateadaptmode(mh, 0); 1121 mwl_hal_setoptimizationlevel(mh, 1122 (ic->ic_flags & IEEE80211_F_BURST) != 0); 1123 1124 mwl_hal_setregioncode(mh, mwl_map2regioncode(&ic->ic_regdomain)); 1125 1126 mwl_hal_setaggampduratemode(mh, 1, 80); /* XXX */ 1127 mwl_hal_setcfend(mh, 0); /* XXX */ 1128 1129 return 1; 1130 } 1131 1132 static int 1133 mwl_init(struct mwl_softc *sc) 1134 { 1135 struct mwl_hal *mh = sc->sc_mh; 1136 int error = 0; 1137 1138 MWL_LOCK_ASSERT(sc); 1139 1140 /* 1141 * Stop anything previously setup. This is safe 1142 * whether this is the first time through or not. 1143 */ 1144 mwl_stop(sc); 1145 1146 /* 1147 * Push vap-independent state to the firmware. 1148 */ 1149 if (!mwl_hal_reset(sc)) { 1150 device_printf(sc->sc_dev, "unable to reset hardware\n"); 1151 return EIO; 1152 } 1153 1154 /* 1155 * Setup recv (once); transmit is already good to go. 1156 */ 1157 error = mwl_startrecv(sc); 1158 if (error != 0) { 1159 device_printf(sc->sc_dev, "unable to start recv logic\n"); 1160 return error; 1161 } 1162 1163 /* 1164 * Enable interrupts. 1165 */ 1166 sc->sc_imask = MACREG_A2HRIC_BIT_RX_RDY 1167 | MACREG_A2HRIC_BIT_TX_DONE 1168 | MACREG_A2HRIC_BIT_OPC_DONE 1169 #if 0 1170 | MACREG_A2HRIC_BIT_MAC_EVENT 1171 #endif 1172 | MACREG_A2HRIC_BIT_ICV_ERROR 1173 | MACREG_A2HRIC_BIT_RADAR_DETECT 1174 | MACREG_A2HRIC_BIT_CHAN_SWITCH 1175 #if 0 1176 | MACREG_A2HRIC_BIT_QUEUE_EMPTY 1177 #endif 1178 | MACREG_A2HRIC_BIT_BA_WATCHDOG 1179 | MACREQ_A2HRIC_BIT_TX_ACK 1180 ; 1181 1182 sc->sc_running = 1; 1183 mwl_hal_intrset(mh, sc->sc_imask); 1184 callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc); 1185 1186 return 0; 1187 } 1188 1189 static void 1190 mwl_stop(struct mwl_softc *sc) 1191 { 1192 1193 MWL_LOCK_ASSERT(sc); 1194 if (sc->sc_running) { 1195 /* 1196 * Shutdown the hardware and driver. 1197 */ 1198 sc->sc_running = 0; 1199 callout_stop(&sc->sc_watchdog); 1200 sc->sc_tx_timer = 0; 1201 mwl_draintxq(sc); 1202 } 1203 } 1204 1205 static int 1206 mwl_reset_vap(struct ieee80211vap *vap, int state) 1207 { 1208 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap; 1209 struct ieee80211com *ic = vap->iv_ic; 1210 1211 if (state == IEEE80211_S_RUN) 1212 mwl_setrates(vap); 1213 /* XXX off by 1? */ 1214 mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold); 1215 /* XXX auto? 20/40 split? */ 1216 mwl_hal_sethtgi(hvap, (vap->iv_flags_ht & 1217 (IEEE80211_FHT_SHORTGI20|IEEE80211_FHT_SHORTGI40)) ? 1 : 0); 1218 mwl_hal_setnprot(hvap, ic->ic_htprotmode == IEEE80211_PROT_NONE ? 1219 HTPROTECT_NONE : HTPROTECT_AUTO); 1220 /* XXX txpower cap */ 1221 1222 /* re-setup beacons */ 1223 if (state == IEEE80211_S_RUN && 1224 (vap->iv_opmode == IEEE80211_M_HOSTAP || 1225 vap->iv_opmode == IEEE80211_M_MBSS || 1226 vap->iv_opmode == IEEE80211_M_IBSS)) { 1227 mwl_setapmode(vap, vap->iv_bss->ni_chan); 1228 mwl_hal_setnprotmode(hvap, _IEEE80211_MASKSHIFT( 1229 ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE)); 1230 return mwl_beacon_setup(vap); 1231 } 1232 return 0; 1233 } 1234 1235 /* 1236 * Reset the hardware w/o losing operational state. 1237 * Used to reset or reload hardware state for a vap. 1238 */ 1239 static int 1240 mwl_reset(struct ieee80211vap *vap, u_long cmd) 1241 { 1242 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap; 1243 int error = 0; 1244 1245 if (hvap != NULL) { /* WDS, MONITOR, etc. */ 1246 struct ieee80211com *ic = vap->iv_ic; 1247 struct mwl_softc *sc = ic->ic_softc; 1248 struct mwl_hal *mh = sc->sc_mh; 1249 1250 /* XXX handle DWDS sta vap change */ 1251 /* XXX do we need to disable interrupts? */ 1252 mwl_hal_intrset(mh, 0); /* disable interrupts */ 1253 error = mwl_reset_vap(vap, vap->iv_state); 1254 mwl_hal_intrset(mh, sc->sc_imask); 1255 } 1256 return error; 1257 } 1258 1259 /* 1260 * Allocate a tx buffer for sending a frame. The 1261 * packet is assumed to have the WME AC stored so 1262 * we can use it to select the appropriate h/w queue. 1263 */ 1264 static struct mwl_txbuf * 1265 mwl_gettxbuf(struct mwl_softc *sc, struct mwl_txq *txq) 1266 { 1267 struct mwl_txbuf *bf; 1268 1269 /* 1270 * Grab a TX buffer and associated resources. 1271 */ 1272 MWL_TXQ_LOCK(txq); 1273 bf = STAILQ_FIRST(&txq->free); 1274 if (bf != NULL) { 1275 STAILQ_REMOVE_HEAD(&txq->free, bf_list); 1276 txq->nfree--; 1277 } 1278 MWL_TXQ_UNLOCK(txq); 1279 if (bf == NULL) 1280 DPRINTF(sc, MWL_DEBUG_XMIT, 1281 "%s: out of xmit buffers on q %d\n", __func__, txq->qnum); 1282 return bf; 1283 } 1284 1285 /* 1286 * Return a tx buffer to the queue it came from. Note there 1287 * are two cases because we must preserve the order of buffers 1288 * as it reflects the fixed order of descriptors in memory 1289 * (the firmware pre-fetches descriptors so we cannot reorder). 1290 */ 1291 static void 1292 mwl_puttxbuf_head(struct mwl_txq *txq, struct mwl_txbuf *bf) 1293 { 1294 bf->bf_m = NULL; 1295 bf->bf_node = NULL; 1296 MWL_TXQ_LOCK(txq); 1297 STAILQ_INSERT_HEAD(&txq->free, bf, bf_list); 1298 txq->nfree++; 1299 MWL_TXQ_UNLOCK(txq); 1300 } 1301 1302 static void 1303 mwl_puttxbuf_tail(struct mwl_txq *txq, struct mwl_txbuf *bf) 1304 { 1305 bf->bf_m = NULL; 1306 bf->bf_node = NULL; 1307 MWL_TXQ_LOCK(txq); 1308 STAILQ_INSERT_TAIL(&txq->free, bf, bf_list); 1309 txq->nfree++; 1310 MWL_TXQ_UNLOCK(txq); 1311 } 1312 1313 static int 1314 mwl_transmit(struct ieee80211com *ic, struct mbuf *m) 1315 { 1316 struct mwl_softc *sc = ic->ic_softc; 1317 int error; 1318 1319 MWL_LOCK(sc); 1320 if (!sc->sc_running) { 1321 MWL_UNLOCK(sc); 1322 return (ENXIO); 1323 } 1324 error = mbufq_enqueue(&sc->sc_snd, m); 1325 if (error) { 1326 MWL_UNLOCK(sc); 1327 return (error); 1328 } 1329 mwl_start(sc); 1330 MWL_UNLOCK(sc); 1331 return (0); 1332 } 1333 1334 static void 1335 mwl_start(struct mwl_softc *sc) 1336 { 1337 struct ieee80211_node *ni; 1338 struct mwl_txbuf *bf; 1339 struct mbuf *m; 1340 struct mwl_txq *txq = NULL; /* XXX silence gcc */ 1341 int nqueued; 1342 1343 MWL_LOCK_ASSERT(sc); 1344 if (!sc->sc_running || sc->sc_invalid) 1345 return; 1346 nqueued = 0; 1347 while ((m = mbufq_dequeue(&sc->sc_snd)) != NULL) { 1348 /* 1349 * Grab the node for the destination. 1350 */ 1351 ni = (struct ieee80211_node *) m->m_pkthdr.rcvif; 1352 KASSERT(ni != NULL, ("no node")); 1353 m->m_pkthdr.rcvif = NULL; /* committed, clear ref */ 1354 /* 1355 * Grab a TX buffer and associated resources. 1356 * We honor the classification by the 802.11 layer. 1357 */ 1358 txq = sc->sc_ac2q[M_WME_GETAC(m)]; 1359 bf = mwl_gettxbuf(sc, txq); 1360 if (bf == NULL) { 1361 m_freem(m); 1362 ieee80211_free_node(ni); 1363 #ifdef MWL_TX_NODROP 1364 sc->sc_stats.mst_tx_qstop++; 1365 break; 1366 #else 1367 DPRINTF(sc, MWL_DEBUG_XMIT, 1368 "%s: tail drop on q %d\n", __func__, txq->qnum); 1369 sc->sc_stats.mst_tx_qdrop++; 1370 continue; 1371 #endif /* MWL_TX_NODROP */ 1372 } 1373 1374 /* 1375 * Pass the frame to the h/w for transmission. 1376 */ 1377 if (mwl_tx_start(sc, ni, bf, m)) { 1378 if_inc_counter(ni->ni_vap->iv_ifp, 1379 IFCOUNTER_OERRORS, 1); 1380 mwl_puttxbuf_head(txq, bf); 1381 ieee80211_free_node(ni); 1382 continue; 1383 } 1384 nqueued++; 1385 if (nqueued >= mwl_txcoalesce) { 1386 /* 1387 * Poke the firmware to process queued frames; 1388 * see below about (lack of) locking. 1389 */ 1390 nqueued = 0; 1391 mwl_hal_txstart(sc->sc_mh, 0/*XXX*/); 1392 } 1393 } 1394 if (nqueued) { 1395 /* 1396 * NB: We don't need to lock against tx done because 1397 * this just prods the firmware to check the transmit 1398 * descriptors. The firmware will also start fetching 1399 * descriptors by itself if it notices new ones are 1400 * present when it goes to deliver a tx done interrupt 1401 * to the host. So if we race with tx done processing 1402 * it's ok. Delivering the kick here rather than in 1403 * mwl_tx_start is an optimization to avoid poking the 1404 * firmware for each packet. 1405 * 1406 * NB: the queue id isn't used so 0 is ok. 1407 */ 1408 mwl_hal_txstart(sc->sc_mh, 0/*XXX*/); 1409 } 1410 } 1411 1412 static int 1413 mwl_raw_xmit(struct ieee80211_node *ni, struct mbuf *m, 1414 const struct ieee80211_bpf_params *params) 1415 { 1416 struct ieee80211com *ic = ni->ni_ic; 1417 struct mwl_softc *sc = ic->ic_softc; 1418 struct mwl_txbuf *bf; 1419 struct mwl_txq *txq; 1420 1421 if (!sc->sc_running || sc->sc_invalid) { 1422 m_freem(m); 1423 return ENETDOWN; 1424 } 1425 /* 1426 * Grab a TX buffer and associated resources. 1427 * Note that we depend on the classification 1428 * by the 802.11 layer to get to the right h/w 1429 * queue. Management frames must ALWAYS go on 1430 * queue 1 but we cannot just force that here 1431 * because we may receive non-mgt frames. 1432 */ 1433 txq = sc->sc_ac2q[M_WME_GETAC(m)]; 1434 bf = mwl_gettxbuf(sc, txq); 1435 if (bf == NULL) { 1436 sc->sc_stats.mst_tx_qstop++; 1437 m_freem(m); 1438 return ENOBUFS; 1439 } 1440 /* 1441 * Pass the frame to the h/w for transmission. 1442 */ 1443 if (mwl_tx_start(sc, ni, bf, m)) { 1444 mwl_puttxbuf_head(txq, bf); 1445 1446 return EIO; /* XXX */ 1447 } 1448 /* 1449 * NB: We don't need to lock against tx done because 1450 * this just prods the firmware to check the transmit 1451 * descriptors. The firmware will also start fetching 1452 * descriptors by itself if it notices new ones are 1453 * present when it goes to deliver a tx done interrupt 1454 * to the host. So if we race with tx done processing 1455 * it's ok. Delivering the kick here rather than in 1456 * mwl_tx_start is an optimization to avoid poking the 1457 * firmware for each packet. 1458 * 1459 * NB: the queue id isn't used so 0 is ok. 1460 */ 1461 mwl_hal_txstart(sc->sc_mh, 0/*XXX*/); 1462 return 0; 1463 } 1464 1465 static int 1466 mwl_media_change(if_t ifp) 1467 { 1468 struct ieee80211vap *vap; 1469 int error; 1470 1471 /* NB: only the fixed rate can change and that doesn't need a reset */ 1472 error = ieee80211_media_change(ifp); 1473 if (error != 0) 1474 return (error); 1475 1476 vap = if_getsoftc(ifp); 1477 mwl_setrates(vap); 1478 return (0); 1479 } 1480 1481 #ifdef MWL_DEBUG 1482 static void 1483 mwl_keyprint(struct mwl_softc *sc, const char *tag, 1484 const MWL_HAL_KEYVAL *hk, const uint8_t mac[IEEE80211_ADDR_LEN]) 1485 { 1486 static const char *ciphers[] = { 1487 "WEP", 1488 "TKIP", 1489 "AES-CCM", 1490 }; 1491 int i, n; 1492 1493 printf("%s: [%u] %-7s", tag, hk->keyIndex, ciphers[hk->keyTypeId]); 1494 for (i = 0, n = hk->keyLen; i < n; i++) 1495 printf(" %02x", hk->key.aes[i]); 1496 printf(" mac %s", ether_sprintf(mac)); 1497 if (hk->keyTypeId == KEY_TYPE_ID_TKIP) { 1498 printf(" %s", "rxmic"); 1499 for (i = 0; i < sizeof(hk->key.tkip.rxMic); i++) 1500 printf(" %02x", hk->key.tkip.rxMic[i]); 1501 printf(" txmic"); 1502 for (i = 0; i < sizeof(hk->key.tkip.txMic); i++) 1503 printf(" %02x", hk->key.tkip.txMic[i]); 1504 } 1505 printf(" flags 0x%x\n", hk->keyFlags); 1506 } 1507 #endif 1508 1509 /* 1510 * Allocate a key cache slot for a unicast key. The 1511 * firmware handles key allocation and every station is 1512 * guaranteed key space so we are always successful. 1513 */ 1514 static int 1515 mwl_key_alloc(struct ieee80211vap *vap, struct ieee80211_key *k, 1516 ieee80211_keyix *keyix, ieee80211_keyix *rxkeyix) 1517 { 1518 struct mwl_softc *sc = vap->iv_ic->ic_softc; 1519 1520 if (k->wk_keyix != IEEE80211_KEYIX_NONE || 1521 (k->wk_flags & IEEE80211_KEY_GROUP)) { 1522 if (!ieee80211_is_key_global(vap, k)) { 1523 /* should not happen */ 1524 DPRINTF(sc, MWL_DEBUG_KEYCACHE, 1525 "%s: bogus group key\n", __func__); 1526 return 0; 1527 } 1528 /* give the caller what they requested */ 1529 *keyix = *rxkeyix = ieee80211_crypto_get_key_wepidx(vap, k); 1530 } else { 1531 /* 1532 * Firmware handles key allocation. 1533 */ 1534 *keyix = *rxkeyix = 0; 1535 } 1536 return 1; 1537 } 1538 1539 /* 1540 * Delete a key entry allocated by mwl_key_alloc. 1541 */ 1542 static int 1543 mwl_key_delete(struct ieee80211vap *vap, const struct ieee80211_key *k) 1544 { 1545 struct mwl_softc *sc = vap->iv_ic->ic_softc; 1546 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap; 1547 MWL_HAL_KEYVAL hk; 1548 const uint8_t bcastaddr[IEEE80211_ADDR_LEN] = 1549 { 0xff, 0xff, 0xff, 0xff, 0xff, 0xff }; 1550 1551 if (hvap == NULL) { 1552 if (vap->iv_opmode != IEEE80211_M_WDS) { 1553 /* XXX monitor mode? */ 1554 DPRINTF(sc, MWL_DEBUG_KEYCACHE, 1555 "%s: no hvap for opmode %d\n", __func__, 1556 vap->iv_opmode); 1557 return 0; 1558 } 1559 hvap = MWL_VAP(vap)->mv_ap_hvap; 1560 } 1561 1562 DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: delete key %u\n", 1563 __func__, k->wk_keyix); 1564 1565 memset(&hk, 0, sizeof(hk)); 1566 hk.keyIndex = k->wk_keyix; 1567 switch (k->wk_cipher->ic_cipher) { 1568 case IEEE80211_CIPHER_WEP: 1569 hk.keyTypeId = KEY_TYPE_ID_WEP; 1570 break; 1571 case IEEE80211_CIPHER_TKIP: 1572 hk.keyTypeId = KEY_TYPE_ID_TKIP; 1573 break; 1574 case IEEE80211_CIPHER_AES_CCM: 1575 hk.keyTypeId = KEY_TYPE_ID_AES; 1576 break; 1577 default: 1578 /* XXX should not happen */ 1579 DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n", 1580 __func__, k->wk_cipher->ic_cipher); 1581 return 0; 1582 } 1583 return (mwl_hal_keyreset(hvap, &hk, bcastaddr) == 0); /*XXX*/ 1584 } 1585 1586 static __inline int 1587 addgroupflags(MWL_HAL_KEYVAL *hk, const struct ieee80211_key *k) 1588 { 1589 if (k->wk_flags & IEEE80211_KEY_GROUP) { 1590 if (k->wk_flags & IEEE80211_KEY_XMIT) 1591 hk->keyFlags |= KEY_FLAG_TXGROUPKEY; 1592 if (k->wk_flags & IEEE80211_KEY_RECV) 1593 hk->keyFlags |= KEY_FLAG_RXGROUPKEY; 1594 return 1; 1595 } else 1596 return 0; 1597 } 1598 1599 /* 1600 * Set the key cache contents for the specified key. Key cache 1601 * slot(s) must already have been allocated by mwl_key_alloc. 1602 */ 1603 static int 1604 mwl_key_set(struct ieee80211vap *vap, const struct ieee80211_key *k) 1605 { 1606 return (_mwl_key_set(vap, k, k->wk_macaddr)); 1607 } 1608 1609 static int 1610 _mwl_key_set(struct ieee80211vap *vap, const struct ieee80211_key *k, 1611 const uint8_t mac[IEEE80211_ADDR_LEN]) 1612 { 1613 #define GRPXMIT (IEEE80211_KEY_XMIT | IEEE80211_KEY_GROUP) 1614 /* NB: static wep keys are marked GROUP+tx/rx; GTK will be tx or rx */ 1615 #define IEEE80211_IS_STATICKEY(k) \ 1616 (((k)->wk_flags & (GRPXMIT|IEEE80211_KEY_RECV)) == \ 1617 (GRPXMIT|IEEE80211_KEY_RECV)) 1618 struct mwl_softc *sc = vap->iv_ic->ic_softc; 1619 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap; 1620 const struct ieee80211_cipher *cip = k->wk_cipher; 1621 const uint8_t *macaddr; 1622 MWL_HAL_KEYVAL hk; 1623 1624 KASSERT((k->wk_flags & IEEE80211_KEY_SWCRYPT) == 0, 1625 ("s/w crypto set?")); 1626 1627 if (hvap == NULL) { 1628 if (vap->iv_opmode != IEEE80211_M_WDS) { 1629 /* XXX monitor mode? */ 1630 DPRINTF(sc, MWL_DEBUG_KEYCACHE, 1631 "%s: no hvap for opmode %d\n", __func__, 1632 vap->iv_opmode); 1633 return 0; 1634 } 1635 hvap = MWL_VAP(vap)->mv_ap_hvap; 1636 } 1637 memset(&hk, 0, sizeof(hk)); 1638 hk.keyIndex = k->wk_keyix; 1639 switch (cip->ic_cipher) { 1640 case IEEE80211_CIPHER_WEP: 1641 hk.keyTypeId = KEY_TYPE_ID_WEP; 1642 hk.keyLen = k->wk_keylen; 1643 if (k->wk_keyix == vap->iv_def_txkey) 1644 hk.keyFlags = KEY_FLAG_WEP_TXKEY; 1645 if (!IEEE80211_IS_STATICKEY(k)) { 1646 /* NB: WEP is never used for the PTK */ 1647 (void) addgroupflags(&hk, k); 1648 } 1649 break; 1650 case IEEE80211_CIPHER_TKIP: 1651 hk.keyTypeId = KEY_TYPE_ID_TKIP; 1652 hk.key.tkip.tsc.high = (uint32_t)(k->wk_keytsc >> 16); 1653 hk.key.tkip.tsc.low = (uint16_t)k->wk_keytsc; 1654 hk.keyFlags = KEY_FLAG_TSC_VALID | KEY_FLAG_MICKEY_VALID; 1655 hk.keyLen = k->wk_keylen + IEEE80211_MICBUF_SIZE; 1656 if (!addgroupflags(&hk, k)) 1657 hk.keyFlags |= KEY_FLAG_PAIRWISE; 1658 break; 1659 case IEEE80211_CIPHER_AES_CCM: 1660 hk.keyTypeId = KEY_TYPE_ID_AES; 1661 hk.keyLen = k->wk_keylen; 1662 if (!addgroupflags(&hk, k)) 1663 hk.keyFlags |= KEY_FLAG_PAIRWISE; 1664 break; 1665 default: 1666 /* XXX should not happen */ 1667 DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n", 1668 __func__, k->wk_cipher->ic_cipher); 1669 return 0; 1670 } 1671 /* 1672 * NB: tkip mic keys get copied here too; the layout 1673 * just happens to match that in ieee80211_key. 1674 */ 1675 memcpy(hk.key.aes, k->wk_key, hk.keyLen); 1676 1677 /* 1678 * Locate address of sta db entry for writing key; 1679 * the convention unfortunately is somewhat different 1680 * than how net80211, hostapd, and wpa_supplicant think. 1681 */ 1682 if (vap->iv_opmode == IEEE80211_M_STA) { 1683 /* 1684 * NB: keys plumbed before the sta reaches AUTH state 1685 * will be discarded or written to the wrong sta db 1686 * entry because iv_bss is meaningless. This is ok 1687 * (right now) because we handle deferred plumbing of 1688 * WEP keys when the sta reaches AUTH state. 1689 */ 1690 macaddr = vap->iv_bss->ni_bssid; 1691 if ((k->wk_flags & IEEE80211_KEY_GROUP) == 0) { 1692 /* XXX plumb to local sta db too for static key wep */ 1693 mwl_hal_keyset(hvap, &hk, vap->iv_myaddr); 1694 } 1695 } else if (vap->iv_opmode == IEEE80211_M_WDS && 1696 vap->iv_state != IEEE80211_S_RUN) { 1697 /* 1698 * Prior to RUN state a WDS vap will not it's BSS node 1699 * setup so we will plumb the key to the wrong mac 1700 * address (it'll be our local address). Workaround 1701 * this for the moment by grabbing the correct address. 1702 */ 1703 macaddr = vap->iv_des_bssid; 1704 } else if ((k->wk_flags & GRPXMIT) == GRPXMIT) 1705 macaddr = vap->iv_myaddr; 1706 else 1707 macaddr = mac; 1708 KEYPRINTF(sc, &hk, macaddr); 1709 return (mwl_hal_keyset(hvap, &hk, macaddr) == 0); 1710 #undef IEEE80211_IS_STATICKEY 1711 #undef GRPXMIT 1712 } 1713 1714 /* 1715 * Set the multicast filter contents into the hardware. 1716 * XXX f/w has no support; just defer to the os. 1717 */ 1718 static void 1719 mwl_setmcastfilter(struct mwl_softc *sc) 1720 { 1721 #if 0 1722 struct ether_multi *enm; 1723 struct ether_multistep estep; 1724 uint8_t macs[IEEE80211_ADDR_LEN*MWL_HAL_MCAST_MAX];/* XXX stack use */ 1725 uint8_t *mp; 1726 int nmc; 1727 1728 mp = macs; 1729 nmc = 0; 1730 ETHER_FIRST_MULTI(estep, &sc->sc_ec, enm); 1731 while (enm != NULL) { 1732 /* XXX Punt on ranges. */ 1733 if (nmc == MWL_HAL_MCAST_MAX || 1734 !IEEE80211_ADDR_EQ(enm->enm_addrlo, enm->enm_addrhi)) { 1735 if_setflagsbit(ifp, IFF_ALLMULTI, 0); 1736 return; 1737 } 1738 IEEE80211_ADDR_COPY(mp, enm->enm_addrlo); 1739 mp += IEEE80211_ADDR_LEN, nmc++; 1740 ETHER_NEXT_MULTI(estep, enm); 1741 } 1742 if_setflagsbit(ifp, 0, IFF_ALLMULTI); 1743 mwl_hal_setmcast(sc->sc_mh, nmc, macs); 1744 #endif 1745 } 1746 1747 static int 1748 mwl_mode_init(struct mwl_softc *sc) 1749 { 1750 struct ieee80211com *ic = &sc->sc_ic; 1751 struct mwl_hal *mh = sc->sc_mh; 1752 1753 mwl_hal_setpromisc(mh, ic->ic_promisc > 0); 1754 mwl_setmcastfilter(sc); 1755 1756 return 0; 1757 } 1758 1759 /* 1760 * Callback from the 802.11 layer after a multicast state change. 1761 */ 1762 static void 1763 mwl_update_mcast(struct ieee80211com *ic) 1764 { 1765 struct mwl_softc *sc = ic->ic_softc; 1766 1767 mwl_setmcastfilter(sc); 1768 } 1769 1770 /* 1771 * Callback from the 802.11 layer after a promiscuous mode change. 1772 * Note this interface does not check the operating mode as this 1773 * is an internal callback and we are expected to honor the current 1774 * state (e.g. this is used for setting the interface in promiscuous 1775 * mode when operating in hostap mode to do ACS). 1776 */ 1777 static void 1778 mwl_update_promisc(struct ieee80211com *ic) 1779 { 1780 struct mwl_softc *sc = ic->ic_softc; 1781 1782 mwl_hal_setpromisc(sc->sc_mh, ic->ic_promisc > 0); 1783 } 1784 1785 /* 1786 * Callback from the 802.11 layer to update the slot time 1787 * based on the current setting. We use it to notify the 1788 * firmware of ERP changes and the f/w takes care of things 1789 * like slot time and preamble. 1790 */ 1791 static void 1792 mwl_updateslot(struct ieee80211com *ic) 1793 { 1794 struct mwl_softc *sc = ic->ic_softc; 1795 struct mwl_hal *mh = sc->sc_mh; 1796 int prot; 1797 1798 /* NB: can be called early; suppress needless cmds */ 1799 if (!sc->sc_running) 1800 return; 1801 1802 /* 1803 * Calculate the ERP flags. The firwmare will use 1804 * this to carry out the appropriate measures. 1805 */ 1806 prot = 0; 1807 if (IEEE80211_IS_CHAN_ANYG(ic->ic_curchan)) { 1808 if ((ic->ic_flags & IEEE80211_F_SHSLOT) == 0) 1809 prot |= IEEE80211_ERP_NON_ERP_PRESENT; 1810 if (ic->ic_flags & IEEE80211_F_USEPROT) 1811 prot |= IEEE80211_ERP_USE_PROTECTION; 1812 if (ic->ic_flags & IEEE80211_F_USEBARKER) 1813 prot |= IEEE80211_ERP_LONG_PREAMBLE; 1814 } 1815 1816 DPRINTF(sc, MWL_DEBUG_RESET, 1817 "%s: chan %u MHz/flags 0x%x %s slot, (prot 0x%x ic_flags 0x%x)\n", 1818 __func__, ic->ic_curchan->ic_freq, ic->ic_curchan->ic_flags, 1819 ic->ic_flags & IEEE80211_F_SHSLOT ? "short" : "long", prot, 1820 ic->ic_flags); 1821 1822 mwl_hal_setgprot(mh, prot); 1823 } 1824 1825 /* 1826 * Setup the beacon frame. 1827 */ 1828 static int 1829 mwl_beacon_setup(struct ieee80211vap *vap) 1830 { 1831 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap; 1832 struct ieee80211_node *ni = vap->iv_bss; 1833 struct mbuf *m; 1834 1835 m = ieee80211_beacon_alloc(ni); 1836 if (m == NULL) 1837 return ENOBUFS; 1838 mwl_hal_setbeacon(hvap, mtod(m, const void *), m->m_len); 1839 m_free(m); 1840 1841 return 0; 1842 } 1843 1844 /* 1845 * Update the beacon frame in response to a change. 1846 */ 1847 static void 1848 mwl_beacon_update(struct ieee80211vap *vap, int item) 1849 { 1850 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap; 1851 struct ieee80211com *ic = vap->iv_ic; 1852 1853 KASSERT(hvap != NULL, ("no beacon")); 1854 switch (item) { 1855 case IEEE80211_BEACON_ERP: 1856 mwl_updateslot(ic); 1857 break; 1858 case IEEE80211_BEACON_HTINFO: 1859 mwl_hal_setnprotmode(hvap, _IEEE80211_MASKSHIFT( 1860 ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE)); 1861 break; 1862 case IEEE80211_BEACON_CAPS: 1863 case IEEE80211_BEACON_WME: 1864 case IEEE80211_BEACON_APPIE: 1865 case IEEE80211_BEACON_CSA: 1866 break; 1867 case IEEE80211_BEACON_TIM: 1868 /* NB: firmware always forms TIM */ 1869 return; 1870 } 1871 /* XXX retain beacon frame and update */ 1872 mwl_beacon_setup(vap); 1873 } 1874 1875 static void 1876 mwl_load_cb(void *arg, bus_dma_segment_t *segs, int nsegs, int error) 1877 { 1878 bus_addr_t *paddr = (bus_addr_t*) arg; 1879 KASSERT(error == 0, ("error %u on bus_dma callback", error)); 1880 *paddr = segs->ds_addr; 1881 } 1882 1883 #ifdef MWL_HOST_PS_SUPPORT 1884 /* 1885 * Handle power save station occupancy changes. 1886 */ 1887 static void 1888 mwl_update_ps(struct ieee80211vap *vap, int nsta) 1889 { 1890 struct mwl_vap *mvp = MWL_VAP(vap); 1891 1892 if (nsta == 0 || mvp->mv_last_ps_sta == 0) 1893 mwl_hal_setpowersave_bss(mvp->mv_hvap, nsta); 1894 mvp->mv_last_ps_sta = nsta; 1895 } 1896 1897 /* 1898 * Handle associated station power save state changes. 1899 */ 1900 static int 1901 mwl_set_tim(struct ieee80211_node *ni, int set) 1902 { 1903 struct ieee80211vap *vap = ni->ni_vap; 1904 struct mwl_vap *mvp = MWL_VAP(vap); 1905 1906 if (mvp->mv_set_tim(ni, set)) { /* NB: state change */ 1907 mwl_hal_setpowersave_sta(mvp->mv_hvap, 1908 IEEE80211_AID(ni->ni_associd), set); 1909 return 1; 1910 } else 1911 return 0; 1912 } 1913 #endif /* MWL_HOST_PS_SUPPORT */ 1914 1915 static int 1916 mwl_desc_setup(struct mwl_softc *sc, const char *name, 1917 struct mwl_descdma *dd, 1918 int nbuf, size_t bufsize, int ndesc, size_t descsize) 1919 { 1920 uint8_t *ds; 1921 int error; 1922 1923 DPRINTF(sc, MWL_DEBUG_RESET, 1924 "%s: %s DMA: %u bufs (%ju) %u desc/buf (%ju)\n", 1925 __func__, name, nbuf, (uintmax_t) bufsize, 1926 ndesc, (uintmax_t) descsize); 1927 1928 dd->dd_name = name; 1929 dd->dd_desc_len = nbuf * ndesc * descsize; 1930 1931 /* 1932 * Setup DMA descriptor area. 1933 */ 1934 error = bus_dma_tag_create(bus_get_dma_tag(sc->sc_dev), /* parent */ 1935 PAGE_SIZE, 0, /* alignment, bounds */ 1936 BUS_SPACE_MAXADDR_32BIT, /* lowaddr */ 1937 BUS_SPACE_MAXADDR, /* highaddr */ 1938 NULL, NULL, /* filter, filterarg */ 1939 dd->dd_desc_len, /* maxsize */ 1940 1, /* nsegments */ 1941 dd->dd_desc_len, /* maxsegsize */ 1942 BUS_DMA_ALLOCNOW, /* flags */ 1943 NULL, /* lockfunc */ 1944 NULL, /* lockarg */ 1945 &dd->dd_dmat); 1946 if (error != 0) { 1947 device_printf(sc->sc_dev, "cannot allocate %s DMA tag\n", dd->dd_name); 1948 return error; 1949 } 1950 1951 /* allocate descriptors */ 1952 error = bus_dmamem_alloc(dd->dd_dmat, (void**) &dd->dd_desc, 1953 BUS_DMA_NOWAIT | BUS_DMA_COHERENT, 1954 &dd->dd_dmamap); 1955 if (error != 0) { 1956 device_printf(sc->sc_dev, "unable to alloc memory for %u %s descriptors, " 1957 "error %u\n", nbuf * ndesc, dd->dd_name, error); 1958 goto fail1; 1959 } 1960 1961 error = bus_dmamap_load(dd->dd_dmat, dd->dd_dmamap, 1962 dd->dd_desc, dd->dd_desc_len, 1963 mwl_load_cb, &dd->dd_desc_paddr, 1964 BUS_DMA_NOWAIT); 1965 if (error != 0) { 1966 device_printf(sc->sc_dev, "unable to map %s descriptors, error %u\n", 1967 dd->dd_name, error); 1968 goto fail2; 1969 } 1970 1971 ds = dd->dd_desc; 1972 memset(ds, 0, dd->dd_desc_len); 1973 DPRINTF(sc, MWL_DEBUG_RESET, 1974 "%s: %s DMA map: %p (%lu) -> 0x%jx (%lu)\n", 1975 __func__, dd->dd_name, ds, (u_long) dd->dd_desc_len, 1976 (uintmax_t) dd->dd_desc_paddr, /*XXX*/ (u_long) dd->dd_desc_len); 1977 1978 return 0; 1979 fail2: 1980 bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap); 1981 fail1: 1982 bus_dma_tag_destroy(dd->dd_dmat); 1983 memset(dd, 0, sizeof(*dd)); 1984 return error; 1985 #undef DS2PHYS 1986 } 1987 1988 static void 1989 mwl_desc_cleanup(struct mwl_softc *sc, struct mwl_descdma *dd) 1990 { 1991 bus_dmamap_unload(dd->dd_dmat, dd->dd_dmamap); 1992 bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap); 1993 bus_dma_tag_destroy(dd->dd_dmat); 1994 1995 memset(dd, 0, sizeof(*dd)); 1996 } 1997 1998 /* 1999 * Construct a tx q's free list. The order of entries on 2000 * the list must reflect the physical layout of tx descriptors 2001 * because the firmware pre-fetches descriptors. 2002 * 2003 * XXX might be better to use indices into the buffer array. 2004 */ 2005 static void 2006 mwl_txq_reset(struct mwl_softc *sc, struct mwl_txq *txq) 2007 { 2008 struct mwl_txbuf *bf; 2009 int i; 2010 2011 bf = txq->dma.dd_bufptr; 2012 STAILQ_INIT(&txq->free); 2013 for (i = 0; i < mwl_txbuf; i++, bf++) 2014 STAILQ_INSERT_TAIL(&txq->free, bf, bf_list); 2015 txq->nfree = i; 2016 } 2017 2018 #define DS2PHYS(_dd, _ds) \ 2019 ((_dd)->dd_desc_paddr + ((caddr_t)(_ds) - (caddr_t)(_dd)->dd_desc)) 2020 2021 static int 2022 mwl_txdma_setup(struct mwl_softc *sc, struct mwl_txq *txq) 2023 { 2024 int error, bsize, i; 2025 struct mwl_txbuf *bf; 2026 struct mwl_txdesc *ds; 2027 2028 error = mwl_desc_setup(sc, "tx", &txq->dma, 2029 mwl_txbuf, sizeof(struct mwl_txbuf), 2030 MWL_TXDESC, sizeof(struct mwl_txdesc)); 2031 if (error != 0) 2032 return error; 2033 2034 /* allocate and setup tx buffers */ 2035 bsize = mwl_txbuf * sizeof(struct mwl_txbuf); 2036 bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO); 2037 if (bf == NULL) { 2038 device_printf(sc->sc_dev, "malloc of %u tx buffers failed\n", 2039 mwl_txbuf); 2040 return ENOMEM; 2041 } 2042 txq->dma.dd_bufptr = bf; 2043 2044 ds = txq->dma.dd_desc; 2045 for (i = 0; i < mwl_txbuf; i++, bf++, ds += MWL_TXDESC) { 2046 bf->bf_desc = ds; 2047 bf->bf_daddr = DS2PHYS(&txq->dma, ds); 2048 error = bus_dmamap_create(sc->sc_dmat, BUS_DMA_NOWAIT, 2049 &bf->bf_dmamap); 2050 if (error != 0) { 2051 device_printf(sc->sc_dev, "unable to create dmamap for tx " 2052 "buffer %u, error %u\n", i, error); 2053 return error; 2054 } 2055 } 2056 mwl_txq_reset(sc, txq); 2057 return 0; 2058 } 2059 2060 static void 2061 mwl_txdma_cleanup(struct mwl_softc *sc, struct mwl_txq *txq) 2062 { 2063 struct mwl_txbuf *bf; 2064 int i; 2065 2066 bf = txq->dma.dd_bufptr; 2067 for (i = 0; i < mwl_txbuf; i++, bf++) { 2068 KASSERT(bf->bf_m == NULL, ("mbuf on free list")); 2069 KASSERT(bf->bf_node == NULL, ("node on free list")); 2070 if (bf->bf_dmamap != NULL) 2071 bus_dmamap_destroy(sc->sc_dmat, bf->bf_dmamap); 2072 } 2073 STAILQ_INIT(&txq->free); 2074 txq->nfree = 0; 2075 if (txq->dma.dd_bufptr != NULL) { 2076 free(txq->dma.dd_bufptr, M_MWLDEV); 2077 txq->dma.dd_bufptr = NULL; 2078 } 2079 if (txq->dma.dd_desc_len != 0) 2080 mwl_desc_cleanup(sc, &txq->dma); 2081 } 2082 2083 static int 2084 mwl_rxdma_setup(struct mwl_softc *sc) 2085 { 2086 int error, jumbosize, bsize, i; 2087 struct mwl_rxbuf *bf; 2088 struct mwl_jumbo *rbuf; 2089 struct mwl_rxdesc *ds; 2090 caddr_t data; 2091 2092 error = mwl_desc_setup(sc, "rx", &sc->sc_rxdma, 2093 mwl_rxdesc, sizeof(struct mwl_rxbuf), 2094 1, sizeof(struct mwl_rxdesc)); 2095 if (error != 0) 2096 return error; 2097 2098 /* 2099 * Receive is done to a private pool of jumbo buffers. 2100 * This allows us to attach to mbuf's and avoid re-mapping 2101 * memory on each rx we post. We allocate a large chunk 2102 * of memory and manage it in the driver. The mbuf free 2103 * callback method is used to reclaim frames after sending 2104 * them up the stack. By default we allocate 2x the number of 2105 * rx descriptors configured so we have some slop to hold 2106 * us while frames are processed. 2107 */ 2108 if (mwl_rxbuf < 2*mwl_rxdesc) { 2109 device_printf(sc->sc_dev, 2110 "too few rx dma buffers (%d); increasing to %d\n", 2111 mwl_rxbuf, 2*mwl_rxdesc); 2112 mwl_rxbuf = 2*mwl_rxdesc; 2113 } 2114 jumbosize = roundup(MWL_AGGR_SIZE, PAGE_SIZE); 2115 sc->sc_rxmemsize = mwl_rxbuf*jumbosize; 2116 2117 error = bus_dma_tag_create(sc->sc_dmat, /* parent */ 2118 PAGE_SIZE, 0, /* alignment, bounds */ 2119 BUS_SPACE_MAXADDR_32BIT, /* lowaddr */ 2120 BUS_SPACE_MAXADDR, /* highaddr */ 2121 NULL, NULL, /* filter, filterarg */ 2122 sc->sc_rxmemsize, /* maxsize */ 2123 1, /* nsegments */ 2124 sc->sc_rxmemsize, /* maxsegsize */ 2125 BUS_DMA_ALLOCNOW, /* flags */ 2126 NULL, /* lockfunc */ 2127 NULL, /* lockarg */ 2128 &sc->sc_rxdmat); 2129 if (error != 0) { 2130 device_printf(sc->sc_dev, "could not create rx DMA tag\n"); 2131 return error; 2132 } 2133 2134 error = bus_dmamem_alloc(sc->sc_rxdmat, (void**) &sc->sc_rxmem, 2135 BUS_DMA_NOWAIT | BUS_DMA_COHERENT, 2136 &sc->sc_rxmap); 2137 if (error != 0) { 2138 device_printf(sc->sc_dev, "could not alloc %ju bytes of rx DMA memory\n", 2139 (uintmax_t) sc->sc_rxmemsize); 2140 return error; 2141 } 2142 2143 error = bus_dmamap_load(sc->sc_rxdmat, sc->sc_rxmap, 2144 sc->sc_rxmem, sc->sc_rxmemsize, 2145 mwl_load_cb, &sc->sc_rxmem_paddr, 2146 BUS_DMA_NOWAIT); 2147 if (error != 0) { 2148 device_printf(sc->sc_dev, "could not load rx DMA map\n"); 2149 return error; 2150 } 2151 2152 /* 2153 * Allocate rx buffers and set them up. 2154 */ 2155 bsize = mwl_rxdesc * sizeof(struct mwl_rxbuf); 2156 bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO); 2157 if (bf == NULL) { 2158 device_printf(sc->sc_dev, "malloc of %u rx buffers failed\n", bsize); 2159 return error; 2160 } 2161 sc->sc_rxdma.dd_bufptr = bf; 2162 2163 STAILQ_INIT(&sc->sc_rxbuf); 2164 ds = sc->sc_rxdma.dd_desc; 2165 for (i = 0; i < mwl_rxdesc; i++, bf++, ds++) { 2166 bf->bf_desc = ds; 2167 bf->bf_daddr = DS2PHYS(&sc->sc_rxdma, ds); 2168 /* pre-assign dma buffer */ 2169 bf->bf_data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize); 2170 /* NB: tail is intentional to preserve descriptor order */ 2171 STAILQ_INSERT_TAIL(&sc->sc_rxbuf, bf, bf_list); 2172 } 2173 2174 /* 2175 * Place remainder of dma memory buffers on the free list. 2176 */ 2177 SLIST_INIT(&sc->sc_rxfree); 2178 for (; i < mwl_rxbuf; i++) { 2179 data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize); 2180 rbuf = MWL_JUMBO_DATA2BUF(data); 2181 SLIST_INSERT_HEAD(&sc->sc_rxfree, rbuf, next); 2182 sc->sc_nrxfree++; 2183 } 2184 return 0; 2185 } 2186 #undef DS2PHYS 2187 2188 static void 2189 mwl_rxdma_cleanup(struct mwl_softc *sc) 2190 { 2191 if (sc->sc_rxmem_paddr != 0) { 2192 bus_dmamap_unload(sc->sc_rxdmat, sc->sc_rxmap); 2193 sc->sc_rxmem_paddr = 0; 2194 } 2195 if (sc->sc_rxmem != NULL) { 2196 bus_dmamem_free(sc->sc_rxdmat, sc->sc_rxmem, sc->sc_rxmap); 2197 sc->sc_rxmem = NULL; 2198 } 2199 if (sc->sc_rxdma.dd_bufptr != NULL) { 2200 free(sc->sc_rxdma.dd_bufptr, M_MWLDEV); 2201 sc->sc_rxdma.dd_bufptr = NULL; 2202 } 2203 if (sc->sc_rxdma.dd_desc_len != 0) 2204 mwl_desc_cleanup(sc, &sc->sc_rxdma); 2205 } 2206 2207 static int 2208 mwl_dma_setup(struct mwl_softc *sc) 2209 { 2210 int error, i; 2211 2212 error = mwl_rxdma_setup(sc); 2213 if (error != 0) { 2214 mwl_rxdma_cleanup(sc); 2215 return error; 2216 } 2217 2218 for (i = 0; i < MWL_NUM_TX_QUEUES; i++) { 2219 error = mwl_txdma_setup(sc, &sc->sc_txq[i]); 2220 if (error != 0) { 2221 mwl_dma_cleanup(sc); 2222 return error; 2223 } 2224 } 2225 return 0; 2226 } 2227 2228 static void 2229 mwl_dma_cleanup(struct mwl_softc *sc) 2230 { 2231 int i; 2232 2233 for (i = 0; i < MWL_NUM_TX_QUEUES; i++) 2234 mwl_txdma_cleanup(sc, &sc->sc_txq[i]); 2235 mwl_rxdma_cleanup(sc); 2236 } 2237 2238 static struct ieee80211_node * 2239 mwl_node_alloc(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN]) 2240 { 2241 struct ieee80211com *ic = vap->iv_ic; 2242 struct mwl_softc *sc = ic->ic_softc; 2243 const size_t space = sizeof(struct mwl_node); 2244 struct mwl_node *mn; 2245 2246 mn = malloc(space, M_80211_NODE, M_NOWAIT|M_ZERO); 2247 if (mn == NULL) { 2248 /* XXX stat+msg */ 2249 return NULL; 2250 } 2251 DPRINTF(sc, MWL_DEBUG_NODE, "%s: mn %p\n", __func__, mn); 2252 return &mn->mn_node; 2253 } 2254 2255 static void 2256 mwl_node_cleanup(struct ieee80211_node *ni) 2257 { 2258 struct ieee80211com *ic = ni->ni_ic; 2259 struct mwl_softc *sc = ic->ic_softc; 2260 struct mwl_node *mn = MWL_NODE(ni); 2261 2262 DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p ic %p staid %d\n", 2263 __func__, ni, ni->ni_ic, mn->mn_staid); 2264 2265 if (mn->mn_staid != 0) { 2266 struct ieee80211vap *vap = ni->ni_vap; 2267 2268 if (mn->mn_hvap != NULL) { 2269 if (vap->iv_opmode == IEEE80211_M_STA) 2270 mwl_hal_delstation(mn->mn_hvap, vap->iv_myaddr); 2271 else 2272 mwl_hal_delstation(mn->mn_hvap, ni->ni_macaddr); 2273 } 2274 /* 2275 * NB: legacy WDS peer sta db entry is installed using 2276 * the associate ap's hvap; use it again to delete it. 2277 * XXX can vap be NULL? 2278 */ 2279 else if (vap->iv_opmode == IEEE80211_M_WDS && 2280 MWL_VAP(vap)->mv_ap_hvap != NULL) 2281 mwl_hal_delstation(MWL_VAP(vap)->mv_ap_hvap, 2282 ni->ni_macaddr); 2283 delstaid(sc, mn->mn_staid); 2284 mn->mn_staid = 0; 2285 } 2286 sc->sc_node_cleanup(ni); 2287 } 2288 2289 /* 2290 * Reclaim rx dma buffers from packets sitting on the ampdu 2291 * reorder queue for a station. We replace buffers with a 2292 * system cluster (if available). 2293 */ 2294 static void 2295 mwl_ampdu_rxdma_reclaim(struct ieee80211_rx_ampdu *rap) 2296 { 2297 #if 0 2298 int i, n, off; 2299 struct mbuf *m; 2300 void *cl; 2301 2302 n = rap->rxa_qframes; 2303 for (i = 0; i < rap->rxa_wnd && n > 0; i++) { 2304 m = rap->rxa_m[i]; 2305 if (m == NULL) 2306 continue; 2307 n--; 2308 /* our dma buffers have a well-known free routine */ 2309 if ((m->m_flags & M_EXT) == 0 || 2310 m->m_ext.ext_free != mwl_ext_free) 2311 continue; 2312 /* 2313 * Try to allocate a cluster and move the data. 2314 */ 2315 off = m->m_data - m->m_ext.ext_buf; 2316 if (off + m->m_pkthdr.len > MCLBYTES) { 2317 /* XXX no AMSDU for now */ 2318 continue; 2319 } 2320 cl = pool_cache_get_paddr(&mclpool_cache, 0, 2321 &m->m_ext.ext_paddr); 2322 if (cl != NULL) { 2323 /* 2324 * Copy the existing data to the cluster, remove 2325 * the rx dma buffer, and attach the cluster in 2326 * its place. Note we preserve the offset to the 2327 * data so frames being bridged can still prepend 2328 * their headers without adding another mbuf. 2329 */ 2330 memcpy((caddr_t) cl + off, m->m_data, m->m_pkthdr.len); 2331 MEXTREMOVE(m); 2332 MEXTADD(m, cl, MCLBYTES, 0, NULL, &mclpool_cache); 2333 /* setup mbuf like _MCLGET does */ 2334 m->m_flags |= M_CLUSTER | M_EXT_RW; 2335 _MOWNERREF(m, M_EXT | M_CLUSTER); 2336 /* NB: m_data is clobbered by MEXTADDR, adjust */ 2337 m->m_data += off; 2338 } 2339 } 2340 #endif 2341 } 2342 2343 /* 2344 * Callback to reclaim resources. We first let the 2345 * net80211 layer do it's thing, then if we are still 2346 * blocked by a lack of rx dma buffers we walk the ampdu 2347 * reorder q's to reclaim buffers by copying to a system 2348 * cluster. 2349 */ 2350 static void 2351 mwl_node_drain(struct ieee80211_node *ni) 2352 { 2353 struct ieee80211com *ic = ni->ni_ic; 2354 struct mwl_softc *sc = ic->ic_softc; 2355 struct mwl_node *mn = MWL_NODE(ni); 2356 2357 DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p vap %p staid %d\n", 2358 __func__, ni, ni->ni_vap, mn->mn_staid); 2359 2360 /* NB: call up first to age out ampdu q's */ 2361 sc->sc_node_drain(ni); 2362 2363 /* XXX better to not check low water mark? */ 2364 if (sc->sc_rxblocked && mn->mn_staid != 0 && 2365 (ni->ni_flags & IEEE80211_NODE_HT)) { 2366 uint8_t tid; 2367 /* 2368 * Walk the reorder q and reclaim rx dma buffers by copying 2369 * the packet contents into clusters. 2370 */ 2371 for (tid = 0; tid < WME_NUM_TID; tid++) { 2372 struct ieee80211_rx_ampdu *rap; 2373 2374 rap = &ni->ni_rx_ampdu[tid]; 2375 if ((rap->rxa_flags & IEEE80211_AGGR_XCHGPEND) == 0) 2376 continue; 2377 if (rap->rxa_qframes) 2378 mwl_ampdu_rxdma_reclaim(rap); 2379 } 2380 } 2381 } 2382 2383 static void 2384 mwl_node_getsignal(const struct ieee80211_node *ni, int8_t *rssi, int8_t *noise) 2385 { 2386 *rssi = ni->ni_ic->ic_node_getrssi(ni); 2387 #ifdef MWL_ANT_INFO_SUPPORT 2388 #if 0 2389 /* XXX need to smooth data */ 2390 *noise = -MWL_NODE_CONST(ni)->mn_ai.nf; 2391 #else 2392 *noise = -95; /* XXX */ 2393 #endif 2394 #else 2395 *noise = -95; /* XXX */ 2396 #endif 2397 } 2398 2399 /* 2400 * Convert Hardware per-antenna rssi info to common format: 2401 * Let a1, a2, a3 represent the amplitudes per chain 2402 * Let amax represent max[a1, a2, a3] 2403 * Rssi1_dBm = RSSI_dBm + 20*log10(a1/amax) 2404 * Rssi1_dBm = RSSI_dBm + 20*log10(a1) - 20*log10(amax) 2405 * We store a table that is 4*20*log10(idx) - the extra 4 is to store or 2406 * maintain some extra precision. 2407 * 2408 * Values are stored in .5 db format capped at 127. 2409 */ 2410 static void 2411 mwl_node_getmimoinfo(const struct ieee80211_node *ni, 2412 struct ieee80211_mimo_info *mi) 2413 { 2414 #define CVT(_dst, _src) do { \ 2415 (_dst) = rssi + ((logdbtbl[_src] - logdbtbl[rssi_max]) >> 2); \ 2416 (_dst) = (_dst) > 64 ? 127 : ((_dst) << 1); \ 2417 } while (0) 2418 static const int8_t logdbtbl[32] = { 2419 0, 0, 24, 38, 48, 56, 62, 68, 2420 72, 76, 80, 83, 86, 89, 92, 94, 2421 96, 98, 100, 102, 104, 106, 107, 109, 2422 110, 112, 113, 115, 116, 117, 118, 119 2423 }; 2424 const struct mwl_node *mn = MWL_NODE_CONST(ni); 2425 uint8_t rssi = mn->mn_ai.rsvd1/2; /* XXX */ 2426 uint32_t rssi_max; 2427 2428 rssi_max = mn->mn_ai.rssi_a; 2429 if (mn->mn_ai.rssi_b > rssi_max) 2430 rssi_max = mn->mn_ai.rssi_b; 2431 if (mn->mn_ai.rssi_c > rssi_max) 2432 rssi_max = mn->mn_ai.rssi_c; 2433 2434 CVT(mi->ch[0].rssi[0], mn->mn_ai.rssi_a); 2435 CVT(mi->ch[1].rssi[0], mn->mn_ai.rssi_b); 2436 CVT(mi->ch[2].rssi[0], mn->mn_ai.rssi_c); 2437 2438 mi->ch[0].noise[0] = mn->mn_ai.nf_a; 2439 mi->ch[1].noise[0] = mn->mn_ai.nf_b; 2440 mi->ch[2].noise[0] = mn->mn_ai.nf_c; 2441 #undef CVT 2442 } 2443 2444 static __inline void * 2445 mwl_getrxdma(struct mwl_softc *sc) 2446 { 2447 struct mwl_jumbo *buf; 2448 void *data; 2449 2450 /* 2451 * Allocate from jumbo pool. 2452 */ 2453 MWL_RXFREE_LOCK(sc); 2454 buf = SLIST_FIRST(&sc->sc_rxfree); 2455 if (buf == NULL) { 2456 DPRINTF(sc, MWL_DEBUG_ANY, 2457 "%s: out of rx dma buffers\n", __func__); 2458 sc->sc_stats.mst_rx_nodmabuf++; 2459 data = NULL; 2460 } else { 2461 SLIST_REMOVE_HEAD(&sc->sc_rxfree, next); 2462 sc->sc_nrxfree--; 2463 data = MWL_JUMBO_BUF2DATA(buf); 2464 } 2465 MWL_RXFREE_UNLOCK(sc); 2466 return data; 2467 } 2468 2469 static __inline void 2470 mwl_putrxdma(struct mwl_softc *sc, void *data) 2471 { 2472 struct mwl_jumbo *buf; 2473 2474 /* XXX bounds check data */ 2475 MWL_RXFREE_LOCK(sc); 2476 buf = MWL_JUMBO_DATA2BUF(data); 2477 SLIST_INSERT_HEAD(&sc->sc_rxfree, buf, next); 2478 sc->sc_nrxfree++; 2479 MWL_RXFREE_UNLOCK(sc); 2480 } 2481 2482 static int 2483 mwl_rxbuf_init(struct mwl_softc *sc, struct mwl_rxbuf *bf) 2484 { 2485 struct mwl_rxdesc *ds; 2486 2487 ds = bf->bf_desc; 2488 if (bf->bf_data == NULL) { 2489 bf->bf_data = mwl_getrxdma(sc); 2490 if (bf->bf_data == NULL) { 2491 /* mark descriptor to be skipped */ 2492 ds->RxControl = EAGLE_RXD_CTRL_OS_OWN; 2493 /* NB: don't need PREREAD */ 2494 MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREWRITE); 2495 sc->sc_stats.mst_rxbuf_failed++; 2496 return ENOMEM; 2497 } 2498 } 2499 /* 2500 * NB: DMA buffer contents is known to be unmodified 2501 * so there's no need to flush the data cache. 2502 */ 2503 2504 /* 2505 * Setup descriptor. 2506 */ 2507 ds->QosCtrl = 0; 2508 ds->RSSI = 0; 2509 ds->Status = EAGLE_RXD_STATUS_IDLE; 2510 ds->Channel = 0; 2511 ds->PktLen = htole16(MWL_AGGR_SIZE); 2512 ds->SQ2 = 0; 2513 ds->pPhysBuffData = htole32(MWL_JUMBO_DMA_ADDR(sc, bf->bf_data)); 2514 /* NB: don't touch pPhysNext, set once */ 2515 ds->RxControl = EAGLE_RXD_CTRL_DRIVER_OWN; 2516 MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE); 2517 2518 return 0; 2519 } 2520 2521 static void 2522 mwl_ext_free(struct mbuf *m) 2523 { 2524 struct mwl_softc *sc = m->m_ext.ext_arg1; 2525 2526 /* XXX bounds check data */ 2527 mwl_putrxdma(sc, m->m_ext.ext_buf); 2528 /* 2529 * If we were previously blocked by a lack of rx dma buffers 2530 * check if we now have enough to restart rx interrupt handling. 2531 */ 2532 if (sc->sc_rxblocked && sc->sc_nrxfree > mwl_rxdmalow) { 2533 sc->sc_rxblocked = 0; 2534 mwl_hal_intrset(sc->sc_mh, sc->sc_imask); 2535 } 2536 } 2537 2538 struct mwl_frame_bar { 2539 u_int8_t i_fc[2]; 2540 u_int8_t i_dur[2]; 2541 u_int8_t i_ra[IEEE80211_ADDR_LEN]; 2542 u_int8_t i_ta[IEEE80211_ADDR_LEN]; 2543 /* ctl, seq, FCS */ 2544 } __packed; 2545 2546 /* 2547 * Like ieee80211_anyhdrsize, but handles BAR frames 2548 * specially so the logic below to piece the 802.11 2549 * header together works. 2550 */ 2551 static __inline int 2552 mwl_anyhdrsize(const void *data) 2553 { 2554 const struct ieee80211_frame *wh = data; 2555 2556 if ((wh->i_fc[0]&IEEE80211_FC0_TYPE_MASK) == IEEE80211_FC0_TYPE_CTL) { 2557 switch (wh->i_fc[0] & IEEE80211_FC0_SUBTYPE_MASK) { 2558 case IEEE80211_FC0_SUBTYPE_CTS: 2559 case IEEE80211_FC0_SUBTYPE_ACK: 2560 return sizeof(struct ieee80211_frame_ack); 2561 case IEEE80211_FC0_SUBTYPE_BAR: 2562 return sizeof(struct mwl_frame_bar); 2563 } 2564 return sizeof(struct ieee80211_frame_min); 2565 } else 2566 return ieee80211_hdrsize(data); 2567 } 2568 2569 static void 2570 mwl_handlemicerror(struct ieee80211com *ic, const uint8_t *data) 2571 { 2572 const struct ieee80211_frame *wh; 2573 struct ieee80211_node *ni; 2574 2575 wh = (const struct ieee80211_frame *)(data + sizeof(uint16_t)); 2576 ni = ieee80211_find_rxnode(ic, (const struct ieee80211_frame_min *) wh); 2577 if (ni != NULL) { 2578 ieee80211_notify_michael_failure(ni->ni_vap, wh, 0); 2579 ieee80211_free_node(ni); 2580 } 2581 } 2582 2583 /* 2584 * Convert hardware signal strength to rssi. The value 2585 * provided by the device has the noise floor added in; 2586 * we need to compensate for this but we don't have that 2587 * so we use a fixed value. 2588 * 2589 * The offset of 8 is good for both 2.4 and 5GHz. The LNA 2590 * offset is already set as part of the initial gain. This 2591 * will give at least +/- 3dB for 2.4GHz and +/- 5dB for 5GHz. 2592 */ 2593 static __inline int 2594 cvtrssi(uint8_t ssi) 2595 { 2596 int rssi = (int) ssi + 8; 2597 /* XXX hack guess until we have a real noise floor */ 2598 rssi = 2*(87 - rssi); /* NB: .5 dBm units */ 2599 return (rssi < 0 ? 0 : rssi > 127 ? 127 : rssi); 2600 } 2601 2602 static void 2603 mwl_rx_proc(void *arg, int npending) 2604 { 2605 struct mwl_softc *sc = arg; 2606 struct ieee80211com *ic = &sc->sc_ic; 2607 struct mwl_rxbuf *bf; 2608 struct mwl_rxdesc *ds; 2609 struct mbuf *m; 2610 struct ieee80211_qosframe *wh; 2611 struct ieee80211_node *ni; 2612 struct mwl_node *mn; 2613 int off, len, hdrlen, pktlen, rssi, ntodo; 2614 uint8_t *data, status; 2615 void *newdata; 2616 int16_t nf; 2617 2618 DPRINTF(sc, MWL_DEBUG_RX_PROC, "%s: pending %u rdptr 0x%x wrptr 0x%x\n", 2619 __func__, npending, RD4(sc, sc->sc_hwspecs.rxDescRead), 2620 RD4(sc, sc->sc_hwspecs.rxDescWrite)); 2621 nf = -96; /* XXX */ 2622 bf = sc->sc_rxnext; 2623 for (ntodo = mwl_rxquota; ntodo > 0; ntodo--) { 2624 if (bf == NULL) 2625 bf = STAILQ_FIRST(&sc->sc_rxbuf); 2626 ds = bf->bf_desc; 2627 data = bf->bf_data; 2628 if (data == NULL) { 2629 /* 2630 * If data allocation failed previously there 2631 * will be no buffer; try again to re-populate it. 2632 * Note the firmware will not advance to the next 2633 * descriptor with a dma buffer so we must mimic 2634 * this or we'll get out of sync. 2635 */ 2636 DPRINTF(sc, MWL_DEBUG_ANY, 2637 "%s: rx buf w/o dma memory\n", __func__); 2638 (void) mwl_rxbuf_init(sc, bf); 2639 sc->sc_stats.mst_rx_dmabufmissing++; 2640 break; 2641 } 2642 MWL_RXDESC_SYNC(sc, ds, 2643 BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE); 2644 if (ds->RxControl != EAGLE_RXD_CTRL_DMA_OWN) 2645 break; 2646 #ifdef MWL_DEBUG 2647 if (sc->sc_debug & MWL_DEBUG_RECV_DESC) 2648 mwl_printrxbuf(bf, 0); 2649 #endif 2650 status = ds->Status; 2651 if (status & EAGLE_RXD_STATUS_DECRYPT_ERR_MASK) { 2652 counter_u64_add(ic->ic_ierrors, 1); 2653 sc->sc_stats.mst_rx_crypto++; 2654 /* 2655 * NB: Check EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR 2656 * for backwards compatibility. 2657 */ 2658 if (status != EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR && 2659 (status & EAGLE_RXD_STATUS_TKIP_MIC_DECRYPT_ERR)) { 2660 /* 2661 * MIC error, notify upper layers. 2662 */ 2663 bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap, 2664 BUS_DMASYNC_POSTREAD); 2665 mwl_handlemicerror(ic, data); 2666 sc->sc_stats.mst_rx_tkipmic++; 2667 } 2668 /* XXX too painful to tap packets */ 2669 goto rx_next; 2670 } 2671 /* 2672 * Sync the data buffer. 2673 */ 2674 len = le16toh(ds->PktLen); 2675 bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap, BUS_DMASYNC_POSTREAD); 2676 /* 2677 * The 802.11 header is provided all or in part at the front; 2678 * use it to calculate the true size of the header that we'll 2679 * construct below. We use this to figure out where to copy 2680 * payload prior to constructing the header. 2681 */ 2682 hdrlen = mwl_anyhdrsize(data + sizeof(uint16_t)); 2683 off = sizeof(uint16_t) + sizeof(struct ieee80211_frame_addr4); 2684 2685 /* calculate rssi early so we can re-use for each aggregate */ 2686 rssi = cvtrssi(ds->RSSI); 2687 2688 pktlen = hdrlen + (len - off); 2689 /* 2690 * NB: we know our frame is at least as large as 2691 * IEEE80211_MIN_LEN because there is a 4-address 2692 * frame at the front. Hence there's no need to 2693 * vet the packet length. If the frame in fact 2694 * is too small it should be discarded at the 2695 * net80211 layer. 2696 */ 2697 2698 /* 2699 * Attach dma buffer to an mbuf. We tried 2700 * doing this based on the packet size (i.e. 2701 * copying small packets) but it turns out to 2702 * be a net loss. The tradeoff might be system 2703 * dependent (cache architecture is important). 2704 */ 2705 MGETHDR(m, M_NOWAIT, MT_DATA); 2706 if (m == NULL) { 2707 DPRINTF(sc, MWL_DEBUG_ANY, 2708 "%s: no rx mbuf\n", __func__); 2709 sc->sc_stats.mst_rx_nombuf++; 2710 goto rx_next; 2711 } 2712 /* 2713 * Acquire the replacement dma buffer before 2714 * processing the frame. If we're out of dma 2715 * buffers we disable rx interrupts and wait 2716 * for the free pool to reach mlw_rxdmalow buffers 2717 * before starting to do work again. If the firmware 2718 * runs out of descriptors then it will toss frames 2719 * which is better than our doing it as that can 2720 * starve our processing. It is also important that 2721 * we always process rx'd frames in case they are 2722 * A-MPDU as otherwise the host's view of the BA 2723 * window may get out of sync with the firmware. 2724 */ 2725 newdata = mwl_getrxdma(sc); 2726 if (newdata == NULL) { 2727 /* NB: stat+msg in mwl_getrxdma */ 2728 m_free(m); 2729 /* disable RX interrupt and mark state */ 2730 mwl_hal_intrset(sc->sc_mh, 2731 sc->sc_imask &~ MACREG_A2HRIC_BIT_RX_RDY); 2732 sc->sc_rxblocked = 1; 2733 ieee80211_drain(ic); 2734 /* XXX check rxblocked and immediately start again? */ 2735 goto rx_stop; 2736 } 2737 bf->bf_data = newdata; 2738 /* 2739 * Attach the dma buffer to the mbuf; 2740 * mwl_rxbuf_init will re-setup the rx 2741 * descriptor using the replacement dma 2742 * buffer we just installed above. 2743 */ 2744 m_extadd(m, data, MWL_AGGR_SIZE, mwl_ext_free, sc, NULL, 0, 2745 EXT_NET_DRV); 2746 m->m_data += off - hdrlen; 2747 m->m_pkthdr.len = m->m_len = pktlen; 2748 /* NB: dma buffer assumed read-only */ 2749 2750 /* 2751 * Piece 802.11 header together. 2752 */ 2753 wh = mtod(m, struct ieee80211_qosframe *); 2754 /* NB: don't need to do this sometimes but ... */ 2755 /* XXX special case so we can memcpy after m_devget? */ 2756 ovbcopy(data + sizeof(uint16_t), wh, hdrlen); 2757 if (IEEE80211_QOS_HAS_SEQ(wh)) 2758 *(uint16_t *)ieee80211_getqos(wh) = ds->QosCtrl; 2759 /* 2760 * The f/w strips WEP header but doesn't clear 2761 * the WEP bit; mark the packet with M_WEP so 2762 * net80211 will treat the data as decrypted. 2763 * While here also clear the PWR_MGT bit since 2764 * power save is handled by the firmware and 2765 * passing this up will potentially cause the 2766 * upper layer to put a station in power save 2767 * (except when configured with MWL_HOST_PS_SUPPORT). 2768 */ 2769 if (wh->i_fc[1] & IEEE80211_FC1_PROTECTED) 2770 m->m_flags |= M_WEP; 2771 #ifdef MWL_HOST_PS_SUPPORT 2772 wh->i_fc[1] &= ~IEEE80211_FC1_PROTECTED; 2773 #else 2774 wh->i_fc[1] &= ~(IEEE80211_FC1_PROTECTED | 2775 IEEE80211_FC1_PWR_MGT); 2776 #endif 2777 2778 if (ieee80211_radiotap_active(ic)) { 2779 struct mwl_rx_radiotap_header *tap = &sc->sc_rx_th; 2780 2781 tap->wr_flags = 0; 2782 tap->wr_rate = ds->Rate; 2783 tap->wr_antsignal = rssi + nf; 2784 tap->wr_antnoise = nf; 2785 } 2786 if (IFF_DUMPPKTS_RECV(sc, wh)) { 2787 ieee80211_dump_pkt(ic, mtod(m, caddr_t), 2788 len, ds->Rate, rssi); 2789 } 2790 /* dispatch */ 2791 ni = ieee80211_find_rxnode(ic, 2792 (const struct ieee80211_frame_min *) wh); 2793 if (ni != NULL) { 2794 mn = MWL_NODE(ni); 2795 #ifdef MWL_ANT_INFO_SUPPORT 2796 mn->mn_ai.rssi_a = ds->ai.rssi_a; 2797 mn->mn_ai.rssi_b = ds->ai.rssi_b; 2798 mn->mn_ai.rssi_c = ds->ai.rssi_c; 2799 mn->mn_ai.rsvd1 = rssi; 2800 #endif 2801 /* tag AMPDU aggregates for reorder processing */ 2802 if (ni->ni_flags & IEEE80211_NODE_HT) 2803 m->m_flags |= M_AMPDU; 2804 (void) ieee80211_input(ni, m, rssi, nf); 2805 ieee80211_free_node(ni); 2806 } else 2807 (void) ieee80211_input_all(ic, m, rssi, nf); 2808 rx_next: 2809 /* NB: ignore ENOMEM so we process more descriptors */ 2810 (void) mwl_rxbuf_init(sc, bf); 2811 bf = STAILQ_NEXT(bf, bf_list); 2812 } 2813 rx_stop: 2814 sc->sc_rxnext = bf; 2815 2816 if (mbufq_first(&sc->sc_snd) != NULL) { 2817 /* NB: kick fw; the tx thread may have been preempted */ 2818 mwl_hal_txstart(sc->sc_mh, 0); 2819 mwl_start(sc); 2820 } 2821 } 2822 2823 static void 2824 mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *txq, int qnum) 2825 { 2826 struct mwl_txbuf *bf, *bn; 2827 struct mwl_txdesc *ds; 2828 2829 MWL_TXQ_LOCK_INIT(sc, txq); 2830 txq->qnum = qnum; 2831 txq->txpri = 0; /* XXX */ 2832 #if 0 2833 /* NB: q setup by mwl_txdma_setup XXX */ 2834 STAILQ_INIT(&txq->free); 2835 #endif 2836 STAILQ_FOREACH(bf, &txq->free, bf_list) { 2837 bf->bf_txq = txq; 2838 2839 ds = bf->bf_desc; 2840 bn = STAILQ_NEXT(bf, bf_list); 2841 if (bn == NULL) 2842 bn = STAILQ_FIRST(&txq->free); 2843 ds->pPhysNext = htole32(bn->bf_daddr); 2844 } 2845 STAILQ_INIT(&txq->active); 2846 } 2847 2848 /* 2849 * Setup a hardware data transmit queue for the specified 2850 * access control. We record the mapping from ac's 2851 * to h/w queues for use by mwl_tx_start. 2852 */ 2853 static int 2854 mwl_tx_setup(struct mwl_softc *sc, int ac, int mvtype) 2855 { 2856 struct mwl_txq *txq; 2857 2858 if (ac >= nitems(sc->sc_ac2q)) { 2859 device_printf(sc->sc_dev, "AC %u out of range, max %zu!\n", 2860 ac, nitems(sc->sc_ac2q)); 2861 return 0; 2862 } 2863 if (mvtype >= MWL_NUM_TX_QUEUES) { 2864 device_printf(sc->sc_dev, "mvtype %u out of range, max %u!\n", 2865 mvtype, MWL_NUM_TX_QUEUES); 2866 return 0; 2867 } 2868 txq = &sc->sc_txq[mvtype]; 2869 mwl_txq_init(sc, txq, mvtype); 2870 sc->sc_ac2q[ac] = txq; 2871 return 1; 2872 } 2873 2874 /* 2875 * Update WME parameters for a transmit queue. 2876 */ 2877 static int 2878 mwl_txq_update(struct mwl_softc *sc, int ac) 2879 { 2880 #define MWL_EXPONENT_TO_VALUE(v) ((1<<v)-1) 2881 struct ieee80211com *ic = &sc->sc_ic; 2882 struct chanAccParams chp; 2883 struct mwl_txq *txq = sc->sc_ac2q[ac]; 2884 struct wmeParams *wmep; 2885 struct mwl_hal *mh = sc->sc_mh; 2886 int aifs, cwmin, cwmax, txoplim; 2887 2888 ieee80211_wme_ic_getparams(ic, &chp); 2889 wmep = &chp.cap_wmeParams[ac]; 2890 2891 aifs = wmep->wmep_aifsn; 2892 /* XXX in sta mode need to pass log values for cwmin/max */ 2893 cwmin = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmin); 2894 cwmax = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmax); 2895 txoplim = wmep->wmep_txopLimit; /* NB: units of 32us */ 2896 2897 if (mwl_hal_setedcaparams(mh, txq->qnum, cwmin, cwmax, aifs, txoplim)) { 2898 device_printf(sc->sc_dev, "unable to update hardware queue " 2899 "parameters for %s traffic!\n", 2900 ieee80211_wme_acnames[ac]); 2901 return 0; 2902 } 2903 return 1; 2904 #undef MWL_EXPONENT_TO_VALUE 2905 } 2906 2907 /* 2908 * Callback from the 802.11 layer to update WME parameters. 2909 */ 2910 static int 2911 mwl_wme_update(struct ieee80211com *ic) 2912 { 2913 struct mwl_softc *sc = ic->ic_softc; 2914 2915 return !mwl_txq_update(sc, WME_AC_BE) || 2916 !mwl_txq_update(sc, WME_AC_BK) || 2917 !mwl_txq_update(sc, WME_AC_VI) || 2918 !mwl_txq_update(sc, WME_AC_VO) ? EIO : 0; 2919 } 2920 2921 /* 2922 * Reclaim resources for a setup queue. 2923 */ 2924 static void 2925 mwl_tx_cleanupq(struct mwl_softc *sc, struct mwl_txq *txq) 2926 { 2927 /* XXX hal work? */ 2928 MWL_TXQ_LOCK_DESTROY(txq); 2929 } 2930 2931 /* 2932 * Reclaim all tx queue resources. 2933 */ 2934 static void 2935 mwl_tx_cleanup(struct mwl_softc *sc) 2936 { 2937 int i; 2938 2939 for (i = 0; i < MWL_NUM_TX_QUEUES; i++) 2940 mwl_tx_cleanupq(sc, &sc->sc_txq[i]); 2941 } 2942 2943 static int 2944 mwl_tx_dmasetup(struct mwl_softc *sc, struct mwl_txbuf *bf, struct mbuf *m0) 2945 { 2946 struct mbuf *m; 2947 int error; 2948 2949 /* 2950 * Load the DMA map so any coalescing is done. This 2951 * also calculates the number of descriptors we need. 2952 */ 2953 error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0, 2954 bf->bf_segs, &bf->bf_nseg, 2955 BUS_DMA_NOWAIT); 2956 if (error == EFBIG) { 2957 /* XXX packet requires too many descriptors */ 2958 bf->bf_nseg = MWL_TXDESC+1; 2959 } else if (error != 0) { 2960 sc->sc_stats.mst_tx_busdma++; 2961 m_freem(m0); 2962 return error; 2963 } 2964 /* 2965 * Discard null packets and check for packets that 2966 * require too many TX descriptors. We try to convert 2967 * the latter to a cluster. 2968 */ 2969 if (error == EFBIG) { /* too many desc's, linearize */ 2970 sc->sc_stats.mst_tx_linear++; 2971 #if MWL_TXDESC > 1 2972 m = m_collapse(m0, M_NOWAIT, MWL_TXDESC); 2973 #else 2974 m = m_defrag(m0, M_NOWAIT); 2975 #endif 2976 if (m == NULL) { 2977 m_freem(m0); 2978 sc->sc_stats.mst_tx_nombuf++; 2979 return ENOMEM; 2980 } 2981 m0 = m; 2982 error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0, 2983 bf->bf_segs, &bf->bf_nseg, 2984 BUS_DMA_NOWAIT); 2985 if (error != 0) { 2986 sc->sc_stats.mst_tx_busdma++; 2987 m_freem(m0); 2988 return error; 2989 } 2990 KASSERT(bf->bf_nseg <= MWL_TXDESC, 2991 ("too many segments after defrag; nseg %u", bf->bf_nseg)); 2992 } else if (bf->bf_nseg == 0) { /* null packet, discard */ 2993 sc->sc_stats.mst_tx_nodata++; 2994 m_freem(m0); 2995 return EIO; 2996 } 2997 DPRINTF(sc, MWL_DEBUG_XMIT, "%s: m %p len %u\n", 2998 __func__, m0, m0->m_pkthdr.len); 2999 bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap, BUS_DMASYNC_PREWRITE); 3000 bf->bf_m = m0; 3001 3002 return 0; 3003 } 3004 3005 static __inline int 3006 mwl_cvtlegacyrate(int rate) 3007 { 3008 switch (rate) { 3009 case 2: return 0; 3010 case 4: return 1; 3011 case 11: return 2; 3012 case 22: return 3; 3013 case 44: return 4; 3014 case 12: return 5; 3015 case 18: return 6; 3016 case 24: return 7; 3017 case 36: return 8; 3018 case 48: return 9; 3019 case 72: return 10; 3020 case 96: return 11; 3021 case 108:return 12; 3022 } 3023 return 0; 3024 } 3025 3026 /* 3027 * Calculate fixed tx rate information per client state; 3028 * this value is suitable for writing to the Format field 3029 * of a tx descriptor. 3030 */ 3031 static uint16_t 3032 mwl_calcformat(uint8_t rate, const struct ieee80211_node *ni) 3033 { 3034 uint16_t fmt; 3035 3036 fmt = _IEEE80211_SHIFTMASK(3, EAGLE_TXD_ANTENNA) 3037 | (IEEE80211_IS_CHAN_HT40D(ni->ni_chan) ? 3038 EAGLE_TXD_EXTCHAN_LO : EAGLE_TXD_EXTCHAN_HI); 3039 if (rate & IEEE80211_RATE_MCS) { /* HT MCS */ 3040 fmt |= EAGLE_TXD_FORMAT_HT 3041 /* NB: 0x80 implicitly stripped from ucastrate */ 3042 | _IEEE80211_SHIFTMASK(rate, EAGLE_TXD_RATE); 3043 /* XXX short/long GI may be wrong; re-check */ 3044 if (IEEE80211_IS_CHAN_HT40(ni->ni_chan)) { 3045 fmt |= EAGLE_TXD_CHW_40 3046 | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI40 ? 3047 EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG); 3048 } else { 3049 fmt |= EAGLE_TXD_CHW_20 3050 | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI20 ? 3051 EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG); 3052 } 3053 } else { /* legacy rate */ 3054 fmt |= EAGLE_TXD_FORMAT_LEGACY 3055 | _IEEE80211_SHIFTMASK(mwl_cvtlegacyrate(rate), 3056 EAGLE_TXD_RATE) 3057 | EAGLE_TXD_CHW_20 3058 /* XXX iv_flags & IEEE80211_F_SHPREAMBLE? */ 3059 | (ni->ni_capinfo & IEEE80211_CAPINFO_SHORT_PREAMBLE ? 3060 EAGLE_TXD_PREAMBLE_SHORT : EAGLE_TXD_PREAMBLE_LONG); 3061 } 3062 return fmt; 3063 } 3064 3065 static int 3066 mwl_tx_start(struct mwl_softc *sc, struct ieee80211_node *ni, struct mwl_txbuf *bf, 3067 struct mbuf *m0) 3068 { 3069 struct ieee80211com *ic = &sc->sc_ic; 3070 struct ieee80211vap *vap = ni->ni_vap; 3071 int error, iswep, ismcast; 3072 int hdrlen, pktlen; 3073 struct mwl_txdesc *ds; 3074 struct mwl_txq *txq; 3075 struct ieee80211_frame *wh; 3076 struct mwltxrec *tr; 3077 struct mwl_node *mn; 3078 uint16_t qos; 3079 #if MWL_TXDESC > 1 3080 int i; 3081 #endif 3082 3083 wh = mtod(m0, struct ieee80211_frame *); 3084 iswep = wh->i_fc[1] & IEEE80211_FC1_PROTECTED; 3085 ismcast = IEEE80211_IS_MULTICAST(wh->i_addr1); 3086 hdrlen = ieee80211_anyhdrsize(wh); 3087 pktlen = m0->m_pkthdr.len; 3088 if (IEEE80211_QOS_HAS_SEQ(wh)) { 3089 qos = *(uint16_t *)ieee80211_getqos(wh); 3090 } else 3091 qos = 0; 3092 3093 if (iswep) { 3094 const struct ieee80211_cipher *cip; 3095 struct ieee80211_key *k; 3096 3097 /* 3098 * Construct the 802.11 header+trailer for an encrypted 3099 * frame. The only reason this can fail is because of an 3100 * unknown or unsupported cipher/key type. 3101 * 3102 * NB: we do this even though the firmware will ignore 3103 * what we've done for WEP and TKIP as we need the 3104 * ExtIV filled in for CCMP and this also adjusts 3105 * the headers which simplifies our work below. 3106 */ 3107 k = ieee80211_crypto_encap(ni, m0); 3108 if (k == NULL) { 3109 /* 3110 * This can happen when the key is yanked after the 3111 * frame was queued. Just discard the frame; the 3112 * 802.11 layer counts failures and provides 3113 * debugging/diagnostics. 3114 */ 3115 m_freem(m0); 3116 return EIO; 3117 } 3118 /* 3119 * Adjust the packet length for the crypto additions 3120 * done during encap and any other bits that the f/w 3121 * will add later on. 3122 */ 3123 cip = k->wk_cipher; 3124 pktlen += cip->ic_header + cip->ic_miclen + cip->ic_trailer; 3125 3126 /* packet header may have moved, reset our local pointer */ 3127 wh = mtod(m0, struct ieee80211_frame *); 3128 } 3129 3130 if (ieee80211_radiotap_active_vap(vap)) { 3131 sc->sc_tx_th.wt_flags = 0; /* XXX */ 3132 if (iswep) 3133 sc->sc_tx_th.wt_flags |= IEEE80211_RADIOTAP_F_WEP; 3134 #if 0 3135 sc->sc_tx_th.wt_rate = ds->DataRate; 3136 #endif 3137 sc->sc_tx_th.wt_txpower = ni->ni_txpower; 3138 sc->sc_tx_th.wt_antenna = sc->sc_txantenna; 3139 3140 ieee80211_radiotap_tx(vap, m0); 3141 } 3142 /* 3143 * Copy up/down the 802.11 header; the firmware requires 3144 * we present a 2-byte payload length followed by a 3145 * 4-address header (w/o QoS), followed (optionally) by 3146 * any WEP/ExtIV header (but only filled in for CCMP). 3147 * We are assured the mbuf has sufficient headroom to 3148 * prepend in-place by the setup of ic_headroom in 3149 * mwl_attach. 3150 */ 3151 if (hdrlen < sizeof(struct mwltxrec)) { 3152 const int space = sizeof(struct mwltxrec) - hdrlen; 3153 if (M_LEADINGSPACE(m0) < space) { 3154 /* NB: should never happen */ 3155 device_printf(sc->sc_dev, 3156 "not enough headroom, need %d found %zd, " 3157 "m_flags 0x%x m_len %d\n", 3158 space, M_LEADINGSPACE(m0), m0->m_flags, m0->m_len); 3159 ieee80211_dump_pkt(ic, 3160 mtod(m0, const uint8_t *), m0->m_len, 0, -1); 3161 m_freem(m0); 3162 sc->sc_stats.mst_tx_noheadroom++; 3163 return EIO; 3164 } 3165 M_PREPEND(m0, space, M_NOWAIT); 3166 } 3167 tr = mtod(m0, struct mwltxrec *); 3168 if (wh != (struct ieee80211_frame *) &tr->wh) 3169 ovbcopy(wh, &tr->wh, hdrlen); 3170 /* 3171 * Note: the "firmware length" is actually the length 3172 * of the fully formed "802.11 payload". That is, it's 3173 * everything except for the 802.11 header. In particular 3174 * this includes all crypto material including the MIC! 3175 */ 3176 tr->fwlen = htole16(pktlen - hdrlen); 3177 3178 /* 3179 * Load the DMA map so any coalescing is done. This 3180 * also calculates the number of descriptors we need. 3181 */ 3182 error = mwl_tx_dmasetup(sc, bf, m0); 3183 if (error != 0) { 3184 /* NB: stat collected in mwl_tx_dmasetup */ 3185 DPRINTF(sc, MWL_DEBUG_XMIT, 3186 "%s: unable to setup dma\n", __func__); 3187 return error; 3188 } 3189 bf->bf_node = ni; /* NB: held reference */ 3190 m0 = bf->bf_m; /* NB: may have changed */ 3191 tr = mtod(m0, struct mwltxrec *); 3192 wh = (struct ieee80211_frame *)&tr->wh; 3193 3194 /* 3195 * Formulate tx descriptor. 3196 */ 3197 ds = bf->bf_desc; 3198 txq = bf->bf_txq; 3199 3200 ds->QosCtrl = qos; /* NB: already little-endian */ 3201 #if MWL_TXDESC == 1 3202 /* 3203 * NB: multiframes should be zero because the descriptors 3204 * are initialized to zero. This should handle the case 3205 * where the driver is built with MWL_TXDESC=1 but we are 3206 * using firmware with multi-segment support. 3207 */ 3208 ds->PktPtr = htole32(bf->bf_segs[0].ds_addr); 3209 ds->PktLen = htole16(bf->bf_segs[0].ds_len); 3210 #else 3211 ds->multiframes = htole32(bf->bf_nseg); 3212 ds->PktLen = htole16(m0->m_pkthdr.len); 3213 for (i = 0; i < bf->bf_nseg; i++) { 3214 ds->PktPtrArray[i] = htole32(bf->bf_segs[i].ds_addr); 3215 ds->PktLenArray[i] = htole16(bf->bf_segs[i].ds_len); 3216 } 3217 #endif 3218 /* NB: pPhysNext, DataRate, and SapPktInfo setup once, don't touch */ 3219 ds->Format = 0; 3220 ds->pad = 0; 3221 ds->ack_wcb_addr = 0; 3222 3223 mn = MWL_NODE(ni); 3224 /* 3225 * Select transmit rate. 3226 */ 3227 switch (wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK) { 3228 case IEEE80211_FC0_TYPE_MGT: 3229 sc->sc_stats.mst_tx_mgmt++; 3230 /* fall thru... */ 3231 case IEEE80211_FC0_TYPE_CTL: 3232 /* NB: assign to BE q to avoid bursting */ 3233 ds->TxPriority = MWL_WME_AC_BE; 3234 break; 3235 case IEEE80211_FC0_TYPE_DATA: 3236 if (!ismcast) { 3237 const struct ieee80211_txparam *tp = ni->ni_txparms; 3238 /* 3239 * EAPOL frames get forced to a fixed rate and w/o 3240 * aggregation; otherwise check for any fixed rate 3241 * for the client (may depend on association state). 3242 */ 3243 if (m0->m_flags & M_EAPOL) { 3244 const struct mwl_vap *mvp = MWL_VAP_CONST(vap); 3245 ds->Format = mvp->mv_eapolformat; 3246 ds->pad = htole16( 3247 EAGLE_TXD_FIXED_RATE | EAGLE_TXD_DONT_AGGR); 3248 } else if (tp->ucastrate != IEEE80211_FIXED_RATE_NONE) { 3249 /* XXX pre-calculate per node */ 3250 ds->Format = htole16( 3251 mwl_calcformat(tp->ucastrate, ni)); 3252 ds->pad = htole16(EAGLE_TXD_FIXED_RATE); 3253 } 3254 /* NB: EAPOL frames will never have qos set */ 3255 if (qos == 0) 3256 ds->TxPriority = txq->qnum; 3257 #if MWL_MAXBA > 3 3258 else if (mwl_bastream_match(&mn->mn_ba[3], qos)) 3259 ds->TxPriority = mn->mn_ba[3].txq; 3260 #endif 3261 #if MWL_MAXBA > 2 3262 else if (mwl_bastream_match(&mn->mn_ba[2], qos)) 3263 ds->TxPriority = mn->mn_ba[2].txq; 3264 #endif 3265 #if MWL_MAXBA > 1 3266 else if (mwl_bastream_match(&mn->mn_ba[1], qos)) 3267 ds->TxPriority = mn->mn_ba[1].txq; 3268 #endif 3269 #if MWL_MAXBA > 0 3270 else if (mwl_bastream_match(&mn->mn_ba[0], qos)) 3271 ds->TxPriority = mn->mn_ba[0].txq; 3272 #endif 3273 else 3274 ds->TxPriority = txq->qnum; 3275 } else 3276 ds->TxPriority = txq->qnum; 3277 break; 3278 default: 3279 device_printf(sc->sc_dev, "bogus frame type 0x%x (%s)\n", 3280 wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK, __func__); 3281 sc->sc_stats.mst_tx_badframetype++; 3282 m_freem(m0); 3283 return EIO; 3284 } 3285 3286 if (IFF_DUMPPKTS_XMIT(sc)) 3287 ieee80211_dump_pkt(ic, 3288 mtod(m0, const uint8_t *)+sizeof(uint16_t), 3289 m0->m_len - sizeof(uint16_t), ds->DataRate, -1); 3290 3291 MWL_TXQ_LOCK(txq); 3292 ds->Status = htole32(EAGLE_TXD_STATUS_FW_OWNED); 3293 STAILQ_INSERT_TAIL(&txq->active, bf, bf_list); 3294 MWL_TXDESC_SYNC(txq, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE); 3295 3296 sc->sc_tx_timer = 5; 3297 MWL_TXQ_UNLOCK(txq); 3298 3299 return 0; 3300 } 3301 3302 static __inline int 3303 mwl_cvtlegacyrix(int rix) 3304 { 3305 static const int ieeerates[] = 3306 { 2, 4, 11, 22, 44, 12, 18, 24, 36, 48, 72, 96, 108 }; 3307 return (rix < nitems(ieeerates) ? ieeerates[rix] : 0); 3308 } 3309 3310 /* 3311 * Process completed xmit descriptors from the specified queue. 3312 */ 3313 static int 3314 mwl_tx_processq(struct mwl_softc *sc, struct mwl_txq *txq) 3315 { 3316 #define EAGLE_TXD_STATUS_MCAST \ 3317 (EAGLE_TXD_STATUS_MULTICAST_TX | EAGLE_TXD_STATUS_BROADCAST_TX) 3318 struct ieee80211com *ic = &sc->sc_ic; 3319 struct mwl_txbuf *bf; 3320 struct mwl_txdesc *ds; 3321 struct ieee80211_node *ni; 3322 int nreaped; 3323 uint32_t status; 3324 3325 DPRINTF(sc, MWL_DEBUG_TX_PROC, "%s: tx queue %u\n", __func__, txq->qnum); 3326 for (nreaped = 0;; nreaped++) { 3327 MWL_TXQ_LOCK(txq); 3328 bf = STAILQ_FIRST(&txq->active); 3329 if (bf == NULL) { 3330 MWL_TXQ_UNLOCK(txq); 3331 break; 3332 } 3333 ds = bf->bf_desc; 3334 MWL_TXDESC_SYNC(txq, ds, 3335 BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE); 3336 if (ds->Status & htole32(EAGLE_TXD_STATUS_FW_OWNED)) { 3337 MWL_TXQ_UNLOCK(txq); 3338 break; 3339 } 3340 STAILQ_REMOVE_HEAD(&txq->active, bf_list); 3341 MWL_TXQ_UNLOCK(txq); 3342 3343 #ifdef MWL_DEBUG 3344 if (sc->sc_debug & MWL_DEBUG_XMIT_DESC) 3345 mwl_printtxbuf(bf, txq->qnum, nreaped); 3346 #endif 3347 ni = bf->bf_node; 3348 if (ni != NULL) { 3349 status = le32toh(ds->Status); 3350 if (status & EAGLE_TXD_STATUS_OK) { 3351 uint16_t Format = le16toh(ds->Format); 3352 uint8_t txant = _IEEE80211_MASKSHIFT(Format, 3353 EAGLE_TXD_ANTENNA); 3354 3355 sc->sc_stats.mst_ant_tx[txant]++; 3356 if (status & EAGLE_TXD_STATUS_OK_RETRY) 3357 sc->sc_stats.mst_tx_retries++; 3358 if (status & EAGLE_TXD_STATUS_OK_MORE_RETRY) 3359 sc->sc_stats.mst_tx_mretries++; 3360 if (txq->qnum >= MWL_WME_AC_VO) 3361 ic->ic_wme.wme_hipri_traffic++; 3362 ni->ni_txrate = _IEEE80211_MASKSHIFT(Format, 3363 EAGLE_TXD_RATE); 3364 if ((Format & EAGLE_TXD_FORMAT_HT) == 0) { 3365 ni->ni_txrate = mwl_cvtlegacyrix( 3366 ni->ni_txrate); 3367 } else 3368 ni->ni_txrate |= IEEE80211_RATE_MCS; 3369 sc->sc_stats.mst_tx_rate = ni->ni_txrate; 3370 } else { 3371 if (status & EAGLE_TXD_STATUS_FAILED_LINK_ERROR) 3372 sc->sc_stats.mst_tx_linkerror++; 3373 if (status & EAGLE_TXD_STATUS_FAILED_XRETRY) 3374 sc->sc_stats.mst_tx_xretries++; 3375 if (status & EAGLE_TXD_STATUS_FAILED_AGING) 3376 sc->sc_stats.mst_tx_aging++; 3377 if (bf->bf_m->m_flags & M_FF) 3378 sc->sc_stats.mst_ff_txerr++; 3379 } 3380 if (bf->bf_m->m_flags & M_TXCB) 3381 /* XXX strip fw len in case header inspected */ 3382 m_adj(bf->bf_m, sizeof(uint16_t)); 3383 ieee80211_tx_complete(ni, bf->bf_m, 3384 (status & EAGLE_TXD_STATUS_OK) == 0); 3385 } else 3386 m_freem(bf->bf_m); 3387 ds->Status = htole32(EAGLE_TXD_STATUS_IDLE); 3388 3389 bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap, 3390 BUS_DMASYNC_POSTWRITE); 3391 bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap); 3392 3393 mwl_puttxbuf_tail(txq, bf); 3394 } 3395 return nreaped; 3396 #undef EAGLE_TXD_STATUS_MCAST 3397 } 3398 3399 /* 3400 * Deferred processing of transmit interrupt; special-cased 3401 * for four hardware queues, 0-3. 3402 */ 3403 static void 3404 mwl_tx_proc(void *arg, int npending) 3405 { 3406 struct mwl_softc *sc = arg; 3407 int nreaped; 3408 3409 /* 3410 * Process each active queue. 3411 */ 3412 nreaped = 0; 3413 if (!STAILQ_EMPTY(&sc->sc_txq[0].active)) 3414 nreaped += mwl_tx_processq(sc, &sc->sc_txq[0]); 3415 if (!STAILQ_EMPTY(&sc->sc_txq[1].active)) 3416 nreaped += mwl_tx_processq(sc, &sc->sc_txq[1]); 3417 if (!STAILQ_EMPTY(&sc->sc_txq[2].active)) 3418 nreaped += mwl_tx_processq(sc, &sc->sc_txq[2]); 3419 if (!STAILQ_EMPTY(&sc->sc_txq[3].active)) 3420 nreaped += mwl_tx_processq(sc, &sc->sc_txq[3]); 3421 3422 if (nreaped != 0) { 3423 sc->sc_tx_timer = 0; 3424 if (mbufq_first(&sc->sc_snd) != NULL) { 3425 /* NB: kick fw; the tx thread may have been preempted */ 3426 mwl_hal_txstart(sc->sc_mh, 0); 3427 mwl_start(sc); 3428 } 3429 } 3430 } 3431 3432 static void 3433 mwl_tx_draintxq(struct mwl_softc *sc, struct mwl_txq *txq) 3434 { 3435 struct ieee80211_node *ni; 3436 struct mwl_txbuf *bf; 3437 u_int ix __unused; 3438 3439 /* 3440 * NB: this assumes output has been stopped and 3441 * we do not need to block mwl_tx_tasklet 3442 */ 3443 for (ix = 0;; ix++) { 3444 MWL_TXQ_LOCK(txq); 3445 bf = STAILQ_FIRST(&txq->active); 3446 if (bf == NULL) { 3447 MWL_TXQ_UNLOCK(txq); 3448 break; 3449 } 3450 STAILQ_REMOVE_HEAD(&txq->active, bf_list); 3451 MWL_TXQ_UNLOCK(txq); 3452 #ifdef MWL_DEBUG 3453 if (sc->sc_debug & MWL_DEBUG_RESET) { 3454 struct ieee80211com *ic = &sc->sc_ic; 3455 const struct mwltxrec *tr = 3456 mtod(bf->bf_m, const struct mwltxrec *); 3457 mwl_printtxbuf(bf, txq->qnum, ix); 3458 ieee80211_dump_pkt(ic, (const uint8_t *)&tr->wh, 3459 bf->bf_m->m_len - sizeof(tr->fwlen), 0, -1); 3460 } 3461 #endif /* MWL_DEBUG */ 3462 bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap); 3463 ni = bf->bf_node; 3464 if (ni != NULL) { 3465 /* 3466 * Reclaim node reference. 3467 */ 3468 ieee80211_free_node(ni); 3469 } 3470 m_freem(bf->bf_m); 3471 3472 mwl_puttxbuf_tail(txq, bf); 3473 } 3474 } 3475 3476 /* 3477 * Drain the transmit queues and reclaim resources. 3478 */ 3479 static void 3480 mwl_draintxq(struct mwl_softc *sc) 3481 { 3482 int i; 3483 3484 for (i = 0; i < MWL_NUM_TX_QUEUES; i++) 3485 mwl_tx_draintxq(sc, &sc->sc_txq[i]); 3486 sc->sc_tx_timer = 0; 3487 } 3488 3489 #ifdef MWL_DIAGAPI 3490 /* 3491 * Reset the transmit queues to a pristine state after a fw download. 3492 */ 3493 static void 3494 mwl_resettxq(struct mwl_softc *sc) 3495 { 3496 int i; 3497 3498 for (i = 0; i < MWL_NUM_TX_QUEUES; i++) 3499 mwl_txq_reset(sc, &sc->sc_txq[i]); 3500 } 3501 #endif /* MWL_DIAGAPI */ 3502 3503 /* 3504 * Clear the transmit queues of any frames submitted for the 3505 * specified vap. This is done when the vap is deleted so we 3506 * don't potentially reference the vap after it is gone. 3507 * Note we cannot remove the frames; we only reclaim the node 3508 * reference. 3509 */ 3510 static void 3511 mwl_cleartxq(struct mwl_softc *sc, struct ieee80211vap *vap) 3512 { 3513 struct mwl_txq *txq; 3514 struct mwl_txbuf *bf; 3515 int i; 3516 3517 for (i = 0; i < MWL_NUM_TX_QUEUES; i++) { 3518 txq = &sc->sc_txq[i]; 3519 MWL_TXQ_LOCK(txq); 3520 STAILQ_FOREACH(bf, &txq->active, bf_list) { 3521 struct ieee80211_node *ni = bf->bf_node; 3522 if (ni != NULL && ni->ni_vap == vap) { 3523 bf->bf_node = NULL; 3524 ieee80211_free_node(ni); 3525 } 3526 } 3527 MWL_TXQ_UNLOCK(txq); 3528 } 3529 } 3530 3531 static int 3532 mwl_recv_action(struct ieee80211_node *ni, const struct ieee80211_frame *wh, 3533 const uint8_t *frm, const uint8_t *efrm) 3534 { 3535 struct mwl_softc *sc = ni->ni_ic->ic_softc; 3536 const struct ieee80211_action *ia; 3537 3538 ia = (const struct ieee80211_action *) frm; 3539 if (ia->ia_category == IEEE80211_ACTION_CAT_HT && 3540 ia->ia_action == IEEE80211_ACTION_HT_MIMOPWRSAVE) { 3541 const struct ieee80211_action_ht_mimopowersave *mps = 3542 (const struct ieee80211_action_ht_mimopowersave *) ia; 3543 3544 mwl_hal_setmimops(sc->sc_mh, ni->ni_macaddr, 3545 mps->am_control & IEEE80211_A_HT_MIMOPWRSAVE_ENA, 3546 _IEEE80211_MASKSHIFT(mps->am_control, 3547 IEEE80211_A_HT_MIMOPWRSAVE_MODE)); 3548 return 0; 3549 } else 3550 return sc->sc_recv_action(ni, wh, frm, efrm); 3551 } 3552 3553 static int 3554 mwl_addba_request(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap, 3555 int dialogtoken, int baparamset, int batimeout) 3556 { 3557 struct mwl_softc *sc = ni->ni_ic->ic_softc; 3558 struct ieee80211vap *vap = ni->ni_vap; 3559 struct mwl_node *mn = MWL_NODE(ni); 3560 struct mwl_bastate *bas; 3561 3562 bas = tap->txa_private; 3563 if (bas == NULL) { 3564 const MWL_HAL_BASTREAM *sp; 3565 /* 3566 * Check for a free BA stream slot. 3567 */ 3568 #if MWL_MAXBA > 3 3569 if (mn->mn_ba[3].bastream == NULL) 3570 bas = &mn->mn_ba[3]; 3571 else 3572 #endif 3573 #if MWL_MAXBA > 2 3574 if (mn->mn_ba[2].bastream == NULL) 3575 bas = &mn->mn_ba[2]; 3576 else 3577 #endif 3578 #if MWL_MAXBA > 1 3579 if (mn->mn_ba[1].bastream == NULL) 3580 bas = &mn->mn_ba[1]; 3581 else 3582 #endif 3583 #if MWL_MAXBA > 0 3584 if (mn->mn_ba[0].bastream == NULL) 3585 bas = &mn->mn_ba[0]; 3586 else 3587 #endif 3588 { 3589 /* sta already has max BA streams */ 3590 /* XXX assign BA stream to highest priority tid */ 3591 DPRINTF(sc, MWL_DEBUG_AMPDU, 3592 "%s: already has max bastreams\n", __func__); 3593 sc->sc_stats.mst_ampdu_reject++; 3594 return 0; 3595 } 3596 /* NB: no held reference to ni */ 3597 sp = mwl_hal_bastream_alloc(MWL_VAP(vap)->mv_hvap, 3598 (baparamset & IEEE80211_BAPS_POLICY_IMMEDIATE) != 0, 3599 ni->ni_macaddr, tap->txa_tid, ni->ni_htparam, 3600 ni, tap); 3601 if (sp == NULL) { 3602 /* 3603 * No available stream, return 0 so no 3604 * a-mpdu aggregation will be done. 3605 */ 3606 DPRINTF(sc, MWL_DEBUG_AMPDU, 3607 "%s: no bastream available\n", __func__); 3608 sc->sc_stats.mst_ampdu_nostream++; 3609 return 0; 3610 } 3611 DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: alloc bastream %p\n", 3612 __func__, sp); 3613 /* NB: qos is left zero so we won't match in mwl_tx_start */ 3614 bas->bastream = sp; 3615 tap->txa_private = bas; 3616 } 3617 /* fetch current seq# from the firmware; if available */ 3618 if (mwl_hal_bastream_get_seqno(sc->sc_mh, bas->bastream, 3619 vap->iv_opmode == IEEE80211_M_STA ? vap->iv_myaddr : ni->ni_macaddr, 3620 &tap->txa_start) != 0) 3621 tap->txa_start = 0; 3622 return sc->sc_addba_request(ni, tap, dialogtoken, baparamset, batimeout); 3623 } 3624 3625 static int 3626 mwl_addba_response(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap, 3627 int code, int baparamset, int batimeout) 3628 { 3629 struct mwl_softc *sc = ni->ni_ic->ic_softc; 3630 struct mwl_bastate *bas; 3631 3632 bas = tap->txa_private; 3633 if (bas == NULL) { 3634 /* XXX should not happen */ 3635 DPRINTF(sc, MWL_DEBUG_AMPDU, 3636 "%s: no BA stream allocated, TID %d\n", 3637 __func__, tap->txa_tid); 3638 sc->sc_stats.mst_addba_nostream++; 3639 return 0; 3640 } 3641 if (code == IEEE80211_STATUS_SUCCESS) { 3642 struct ieee80211vap *vap = ni->ni_vap; 3643 int bufsiz, error; 3644 3645 /* 3646 * Tell the firmware to setup the BA stream; 3647 * we know resources are available because we 3648 * pre-allocated one before forming the request. 3649 */ 3650 bufsiz = _IEEE80211_MASKSHIFT(baparamset, IEEE80211_BAPS_BUFSIZ); 3651 if (bufsiz == 0) 3652 bufsiz = IEEE80211_AGGR_BAWMAX; 3653 error = mwl_hal_bastream_create(MWL_VAP(vap)->mv_hvap, 3654 bas->bastream, bufsiz, bufsiz, tap->txa_start); 3655 if (error != 0) { 3656 /* 3657 * Setup failed, return immediately so no a-mpdu 3658 * aggregation will be done. 3659 */ 3660 mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream); 3661 mwl_bastream_free(bas); 3662 tap->txa_private = NULL; 3663 3664 DPRINTF(sc, MWL_DEBUG_AMPDU, 3665 "%s: create failed, error %d, bufsiz %d TID %d " 3666 "htparam 0x%x\n", __func__, error, bufsiz, 3667 tap->txa_tid, ni->ni_htparam); 3668 sc->sc_stats.mst_bacreate_failed++; 3669 return 0; 3670 } 3671 /* NB: cache txq to avoid ptr indirect */ 3672 mwl_bastream_setup(bas, tap->txa_tid, bas->bastream->txq); 3673 DPRINTF(sc, MWL_DEBUG_AMPDU, 3674 "%s: bastream %p assigned to txq %d TID %d bufsiz %d " 3675 "htparam 0x%x\n", __func__, bas->bastream, 3676 bas->txq, tap->txa_tid, bufsiz, ni->ni_htparam); 3677 } else { 3678 /* 3679 * Other side NAK'd us; return the resources. 3680 */ 3681 DPRINTF(sc, MWL_DEBUG_AMPDU, 3682 "%s: request failed with code %d, destroy bastream %p\n", 3683 __func__, code, bas->bastream); 3684 mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream); 3685 mwl_bastream_free(bas); 3686 tap->txa_private = NULL; 3687 } 3688 /* NB: firmware sends BAR so we don't need to */ 3689 return sc->sc_addba_response(ni, tap, code, baparamset, batimeout); 3690 } 3691 3692 static void 3693 mwl_addba_stop(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap) 3694 { 3695 struct mwl_softc *sc = ni->ni_ic->ic_softc; 3696 struct mwl_bastate *bas; 3697 3698 bas = tap->txa_private; 3699 if (bas != NULL) { 3700 DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: destroy bastream %p\n", 3701 __func__, bas->bastream); 3702 mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream); 3703 mwl_bastream_free(bas); 3704 tap->txa_private = NULL; 3705 } 3706 sc->sc_addba_stop(ni, tap); 3707 } 3708 3709 /* 3710 * Setup the rx data structures. This should only be 3711 * done once or we may get out of sync with the firmware. 3712 */ 3713 static int 3714 mwl_startrecv(struct mwl_softc *sc) 3715 { 3716 if (!sc->sc_recvsetup) { 3717 struct mwl_rxbuf *bf, *prev; 3718 struct mwl_rxdesc *ds; 3719 3720 prev = NULL; 3721 STAILQ_FOREACH(bf, &sc->sc_rxbuf, bf_list) { 3722 int error = mwl_rxbuf_init(sc, bf); 3723 if (error != 0) { 3724 DPRINTF(sc, MWL_DEBUG_RECV, 3725 "%s: mwl_rxbuf_init failed %d\n", 3726 __func__, error); 3727 return error; 3728 } 3729 if (prev != NULL) { 3730 ds = prev->bf_desc; 3731 ds->pPhysNext = htole32(bf->bf_daddr); 3732 } 3733 prev = bf; 3734 } 3735 if (prev != NULL) { 3736 ds = prev->bf_desc; 3737 ds->pPhysNext = 3738 htole32(STAILQ_FIRST(&sc->sc_rxbuf)->bf_daddr); 3739 } 3740 sc->sc_recvsetup = 1; 3741 } 3742 mwl_mode_init(sc); /* set filters, etc. */ 3743 return 0; 3744 } 3745 3746 static MWL_HAL_APMODE 3747 mwl_getapmode(const struct ieee80211vap *vap, struct ieee80211_channel *chan) 3748 { 3749 MWL_HAL_APMODE mode; 3750 3751 if (IEEE80211_IS_CHAN_HT(chan)) { 3752 if (vap->iv_flags_ht & IEEE80211_FHT_PUREN) 3753 mode = AP_MODE_N_ONLY; 3754 else if (IEEE80211_IS_CHAN_5GHZ(chan)) 3755 mode = AP_MODE_AandN; 3756 else if (vap->iv_flags & IEEE80211_F_PUREG) 3757 mode = AP_MODE_GandN; 3758 else 3759 mode = AP_MODE_BandGandN; 3760 } else if (IEEE80211_IS_CHAN_ANYG(chan)) { 3761 if (vap->iv_flags & IEEE80211_F_PUREG) 3762 mode = AP_MODE_G_ONLY; 3763 else 3764 mode = AP_MODE_MIXED; 3765 } else if (IEEE80211_IS_CHAN_B(chan)) 3766 mode = AP_MODE_B_ONLY; 3767 else if (IEEE80211_IS_CHAN_A(chan)) 3768 mode = AP_MODE_A_ONLY; 3769 else 3770 mode = AP_MODE_MIXED; /* XXX should not happen? */ 3771 return mode; 3772 } 3773 3774 static int 3775 mwl_setapmode(struct ieee80211vap *vap, struct ieee80211_channel *chan) 3776 { 3777 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap; 3778 return mwl_hal_setapmode(hvap, mwl_getapmode(vap, chan)); 3779 } 3780 3781 /* 3782 * Set/change channels. 3783 */ 3784 static int 3785 mwl_chan_set(struct mwl_softc *sc, struct ieee80211_channel *chan) 3786 { 3787 struct mwl_hal *mh = sc->sc_mh; 3788 struct ieee80211com *ic = &sc->sc_ic; 3789 MWL_HAL_CHANNEL hchan; 3790 int maxtxpow; 3791 3792 DPRINTF(sc, MWL_DEBUG_RESET, "%s: chan %u MHz/flags 0x%x\n", 3793 __func__, chan->ic_freq, chan->ic_flags); 3794 3795 /* 3796 * Convert to a HAL channel description with 3797 * the flags constrained to reflect the current 3798 * operating mode. 3799 */ 3800 mwl_mapchan(&hchan, chan); 3801 mwl_hal_intrset(mh, 0); /* disable interrupts */ 3802 #if 0 3803 mwl_draintxq(sc); /* clear pending tx frames */ 3804 #endif 3805 mwl_hal_setchannel(mh, &hchan); 3806 /* 3807 * Tx power is cap'd by the regulatory setting and 3808 * possibly a user-set limit. We pass the min of 3809 * these to the hal to apply them to the cal data 3810 * for this channel. 3811 * XXX min bound? 3812 */ 3813 maxtxpow = 2*chan->ic_maxregpower; 3814 if (maxtxpow > ic->ic_txpowlimit) 3815 maxtxpow = ic->ic_txpowlimit; 3816 mwl_hal_settxpower(mh, &hchan, maxtxpow / 2); 3817 /* NB: potentially change mcast/mgt rates */ 3818 mwl_setcurchanrates(sc); 3819 3820 /* 3821 * Update internal state. 3822 */ 3823 sc->sc_tx_th.wt_chan_freq = htole16(chan->ic_freq); 3824 sc->sc_rx_th.wr_chan_freq = htole16(chan->ic_freq); 3825 if (IEEE80211_IS_CHAN_A(chan)) { 3826 sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_A); 3827 sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_A); 3828 } else if (IEEE80211_IS_CHAN_ANYG(chan)) { 3829 sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_G); 3830 sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_G); 3831 } else { 3832 sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_B); 3833 sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_B); 3834 } 3835 sc->sc_curchan = hchan; 3836 mwl_hal_intrset(mh, sc->sc_imask); 3837 3838 return 0; 3839 } 3840 3841 static void 3842 mwl_scan_start(struct ieee80211com *ic) 3843 { 3844 struct mwl_softc *sc = ic->ic_softc; 3845 3846 DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__); 3847 } 3848 3849 static void 3850 mwl_scan_end(struct ieee80211com *ic) 3851 { 3852 struct mwl_softc *sc = ic->ic_softc; 3853 3854 DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__); 3855 } 3856 3857 static void 3858 mwl_set_channel(struct ieee80211com *ic) 3859 { 3860 struct mwl_softc *sc = ic->ic_softc; 3861 3862 (void) mwl_chan_set(sc, ic->ic_curchan); 3863 } 3864 3865 /* 3866 * Handle a channel switch request. We inform the firmware 3867 * and mark the global state to suppress various actions. 3868 * NB: we issue only one request to the fw; we may be called 3869 * multiple times if there are multiple vap's. 3870 */ 3871 static void 3872 mwl_startcsa(struct ieee80211vap *vap) 3873 { 3874 struct ieee80211com *ic = vap->iv_ic; 3875 struct mwl_softc *sc = ic->ic_softc; 3876 MWL_HAL_CHANNEL hchan; 3877 3878 if (sc->sc_csapending) 3879 return; 3880 3881 mwl_mapchan(&hchan, ic->ic_csa_newchan); 3882 /* 1 =>'s quiet channel */ 3883 mwl_hal_setchannelswitchie(sc->sc_mh, &hchan, 1, ic->ic_csa_count); 3884 sc->sc_csapending = 1; 3885 } 3886 3887 /* 3888 * Plumb any static WEP key for the station. This is 3889 * necessary as we must propagate the key from the 3890 * global key table of the vap to each sta db entry. 3891 */ 3892 static void 3893 mwl_setanywepkey(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN]) 3894 { 3895 if ((vap->iv_flags & (IEEE80211_F_PRIVACY|IEEE80211_F_WPA)) == 3896 IEEE80211_F_PRIVACY && 3897 vap->iv_def_txkey != IEEE80211_KEYIX_NONE && 3898 vap->iv_nw_keys[vap->iv_def_txkey].wk_keyix != IEEE80211_KEYIX_NONE) 3899 (void) _mwl_key_set(vap, &vap->iv_nw_keys[vap->iv_def_txkey], 3900 mac); 3901 } 3902 3903 static int 3904 mwl_peerstadb(struct ieee80211_node *ni, int aid, int staid, MWL_HAL_PEERINFO *pi) 3905 { 3906 #define WME(ie) ((const struct ieee80211_wme_info *) ie) 3907 struct ieee80211vap *vap = ni->ni_vap; 3908 struct mwl_hal_vap *hvap; 3909 int error; 3910 3911 if (vap->iv_opmode == IEEE80211_M_WDS) { 3912 /* 3913 * WDS vap's do not have a f/w vap; instead they piggyback 3914 * on an AP vap and we must install the sta db entry and 3915 * crypto state using that AP's handle (the WDS vap has none). 3916 */ 3917 hvap = MWL_VAP(vap)->mv_ap_hvap; 3918 } else 3919 hvap = MWL_VAP(vap)->mv_hvap; 3920 error = mwl_hal_newstation(hvap, ni->ni_macaddr, 3921 aid, staid, pi, 3922 ni->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT), 3923 ni->ni_ies.wme_ie != NULL ? WME(ni->ni_ies.wme_ie)->wme_info : 0); 3924 if (error == 0) { 3925 /* 3926 * Setup security for this station. For sta mode this is 3927 * needed even though do the same thing on transition to 3928 * AUTH state because the call to mwl_hal_newstation 3929 * clobbers the crypto state we setup. 3930 */ 3931 mwl_setanywepkey(vap, ni->ni_macaddr); 3932 } 3933 return error; 3934 #undef WME 3935 } 3936 3937 static void 3938 mwl_setglobalkeys(struct ieee80211vap *vap) 3939 { 3940 struct ieee80211_key *wk; 3941 3942 wk = &vap->iv_nw_keys[0]; 3943 for (; wk < &vap->iv_nw_keys[IEEE80211_WEP_NKID]; wk++) 3944 if (wk->wk_keyix != IEEE80211_KEYIX_NONE) 3945 (void) _mwl_key_set(vap, wk, vap->iv_myaddr); 3946 } 3947 3948 /* 3949 * Convert a legacy rate set to a firmware bitmask. 3950 */ 3951 static uint32_t 3952 get_rate_bitmap(const struct ieee80211_rateset *rs) 3953 { 3954 uint32_t rates; 3955 int i; 3956 3957 rates = 0; 3958 for (i = 0; i < rs->rs_nrates; i++) 3959 switch (rs->rs_rates[i] & IEEE80211_RATE_VAL) { 3960 case 2: rates |= 0x001; break; 3961 case 4: rates |= 0x002; break; 3962 case 11: rates |= 0x004; break; 3963 case 22: rates |= 0x008; break; 3964 case 44: rates |= 0x010; break; 3965 case 12: rates |= 0x020; break; 3966 case 18: rates |= 0x040; break; 3967 case 24: rates |= 0x080; break; 3968 case 36: rates |= 0x100; break; 3969 case 48: rates |= 0x200; break; 3970 case 72: rates |= 0x400; break; 3971 case 96: rates |= 0x800; break; 3972 case 108: rates |= 0x1000; break; 3973 } 3974 return rates; 3975 } 3976 3977 /* 3978 * Construct an HT firmware bitmask from an HT rate set. 3979 */ 3980 static uint32_t 3981 get_htrate_bitmap(const struct ieee80211_htrateset *rs) 3982 { 3983 uint32_t rates; 3984 int i; 3985 3986 rates = 0; 3987 for (i = 0; i < rs->rs_nrates; i++) { 3988 if (rs->rs_rates[i] < 16) 3989 rates |= 1<<rs->rs_rates[i]; 3990 } 3991 return rates; 3992 } 3993 3994 /* 3995 * Craft station database entry for station. 3996 * NB: use host byte order here, the hal handles byte swapping. 3997 */ 3998 static MWL_HAL_PEERINFO * 3999 mkpeerinfo(MWL_HAL_PEERINFO *pi, const struct ieee80211_node *ni) 4000 { 4001 const struct ieee80211vap *vap = ni->ni_vap; 4002 4003 memset(pi, 0, sizeof(*pi)); 4004 pi->LegacyRateBitMap = get_rate_bitmap(&ni->ni_rates); 4005 pi->CapInfo = ni->ni_capinfo; 4006 if (ni->ni_flags & IEEE80211_NODE_HT) { 4007 /* HT capabilities, etc */ 4008 pi->HTCapabilitiesInfo = ni->ni_htcap; 4009 /* XXX pi.HTCapabilitiesInfo */ 4010 pi->MacHTParamInfo = ni->ni_htparam; 4011 pi->HTRateBitMap = get_htrate_bitmap(&ni->ni_htrates); 4012 pi->AddHtInfo.ControlChan = ni->ni_htctlchan; 4013 pi->AddHtInfo.AddChan = ni->ni_ht2ndchan; 4014 pi->AddHtInfo.OpMode = ni->ni_htopmode; 4015 pi->AddHtInfo.stbc = ni->ni_htstbc; 4016 4017 /* constrain according to local configuration */ 4018 if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI40) == 0) 4019 pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI40; 4020 if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI20) == 0) 4021 pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI20; 4022 if (ni->ni_chw != 40) 4023 pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_CHWIDTH40; 4024 } 4025 return pi; 4026 } 4027 4028 /* 4029 * Re-create the local sta db entry for a vap to ensure 4030 * up to date WME state is pushed to the firmware. Because 4031 * this resets crypto state this must be followed by a 4032 * reload of any keys in the global key table. 4033 */ 4034 static int 4035 mwl_localstadb(struct ieee80211vap *vap) 4036 { 4037 #define WME(ie) ((const struct ieee80211_wme_info *) ie) 4038 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap; 4039 struct ieee80211_node *bss; 4040 MWL_HAL_PEERINFO pi; 4041 int error; 4042 4043 switch (vap->iv_opmode) { 4044 case IEEE80211_M_STA: 4045 bss = vap->iv_bss; 4046 error = mwl_hal_newstation(hvap, vap->iv_myaddr, 0, 0, 4047 vap->iv_state == IEEE80211_S_RUN ? 4048 mkpeerinfo(&pi, bss) : NULL, 4049 (bss->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT)), 4050 bss->ni_ies.wme_ie != NULL ? 4051 WME(bss->ni_ies.wme_ie)->wme_info : 0); 4052 if (error == 0) 4053 mwl_setglobalkeys(vap); 4054 break; 4055 case IEEE80211_M_HOSTAP: 4056 case IEEE80211_M_MBSS: 4057 error = mwl_hal_newstation(hvap, vap->iv_myaddr, 4058 0, 0, NULL, vap->iv_flags & IEEE80211_F_WME, 0); 4059 if (error == 0) 4060 mwl_setglobalkeys(vap); 4061 break; 4062 default: 4063 error = 0; 4064 break; 4065 } 4066 return error; 4067 #undef WME 4068 } 4069 4070 static int 4071 mwl_newstate(struct ieee80211vap *vap, enum ieee80211_state nstate, int arg) 4072 { 4073 struct mwl_vap *mvp = MWL_VAP(vap); 4074 struct mwl_hal_vap *hvap = mvp->mv_hvap; 4075 struct ieee80211com *ic = vap->iv_ic; 4076 struct ieee80211_node *ni = NULL; 4077 struct mwl_softc *sc = ic->ic_softc; 4078 struct mwl_hal *mh = sc->sc_mh; 4079 enum ieee80211_state ostate = vap->iv_state; 4080 int error; 4081 4082 DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: %s -> %s\n", 4083 if_name(vap->iv_ifp), __func__, 4084 ieee80211_state_name[ostate], ieee80211_state_name[nstate]); 4085 4086 callout_stop(&sc->sc_timer); 4087 /* 4088 * Clear current radar detection state. 4089 */ 4090 if (ostate == IEEE80211_S_CAC) { 4091 /* stop quiet mode radar detection */ 4092 mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_STOP); 4093 } else if (sc->sc_radarena) { 4094 /* stop in-service radar detection */ 4095 mwl_hal_setradardetection(mh, DR_DFS_DISABLE); 4096 sc->sc_radarena = 0; 4097 } 4098 /* 4099 * Carry out per-state actions before doing net80211 work. 4100 */ 4101 if (nstate == IEEE80211_S_INIT) { 4102 /* NB: only ap+sta vap's have a fw entity */ 4103 if (hvap != NULL) 4104 mwl_hal_stop(hvap); 4105 } else if (nstate == IEEE80211_S_SCAN) { 4106 mwl_hal_start(hvap); 4107 /* NB: this disables beacon frames */ 4108 mwl_hal_setinframode(hvap); 4109 } else if (nstate == IEEE80211_S_AUTH) { 4110 /* 4111 * Must create a sta db entry in case a WEP key needs to 4112 * be plumbed. This entry will be overwritten if we 4113 * associate; otherwise it will be reclaimed on node free. 4114 */ 4115 ni = vap->iv_bss; 4116 MWL_NODE(ni)->mn_hvap = hvap; 4117 (void) mwl_peerstadb(ni, 0, 0, NULL); 4118 } else if (nstate == IEEE80211_S_CSA) { 4119 /* XXX move to below? */ 4120 if (vap->iv_opmode == IEEE80211_M_HOSTAP || 4121 vap->iv_opmode == IEEE80211_M_MBSS) 4122 mwl_startcsa(vap); 4123 } else if (nstate == IEEE80211_S_CAC) { 4124 /* XXX move to below? */ 4125 /* stop ap xmit and enable quiet mode radar detection */ 4126 mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_START); 4127 } 4128 4129 /* 4130 * Invoke the parent method to do net80211 work. 4131 */ 4132 error = mvp->mv_newstate(vap, nstate, arg); 4133 4134 /* 4135 * Carry out work that must be done after net80211 runs; 4136 * this work requires up to date state (e.g. iv_bss). 4137 */ 4138 if (error == 0 && nstate == IEEE80211_S_RUN) { 4139 /* NB: collect bss node again, it may have changed */ 4140 ni = vap->iv_bss; 4141 4142 DPRINTF(sc, MWL_DEBUG_STATE, 4143 "%s: %s(RUN): iv_flags 0x%08x bintvl %d bssid %s " 4144 "capinfo 0x%04x chan %d\n", 4145 if_name(vap->iv_ifp), __func__, vap->iv_flags, 4146 ni->ni_intval, ether_sprintf(ni->ni_bssid), ni->ni_capinfo, 4147 ieee80211_chan2ieee(ic, ic->ic_curchan)); 4148 4149 /* 4150 * Recreate local sta db entry to update WME/HT state. 4151 */ 4152 mwl_localstadb(vap); 4153 switch (vap->iv_opmode) { 4154 case IEEE80211_M_HOSTAP: 4155 case IEEE80211_M_MBSS: 4156 if (ostate == IEEE80211_S_CAC) { 4157 /* enable in-service radar detection */ 4158 mwl_hal_setradardetection(mh, 4159 DR_IN_SERVICE_MONITOR_START); 4160 sc->sc_radarena = 1; 4161 } 4162 /* 4163 * Allocate and setup the beacon frame 4164 * (and related state). 4165 */ 4166 error = mwl_reset_vap(vap, IEEE80211_S_RUN); 4167 if (error != 0) { 4168 DPRINTF(sc, MWL_DEBUG_STATE, 4169 "%s: beacon setup failed, error %d\n", 4170 __func__, error); 4171 goto bad; 4172 } 4173 /* NB: must be after setting up beacon */ 4174 mwl_hal_start(hvap); 4175 break; 4176 case IEEE80211_M_STA: 4177 DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: aid 0x%x\n", 4178 if_name(vap->iv_ifp), __func__, ni->ni_associd); 4179 /* 4180 * Set state now that we're associated. 4181 */ 4182 mwl_hal_setassocid(hvap, ni->ni_bssid, ni->ni_associd); 4183 mwl_setrates(vap); 4184 mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold); 4185 if ((vap->iv_flags & IEEE80211_F_DWDS) && 4186 sc->sc_ndwdsvaps++ == 0) 4187 mwl_hal_setdwds(mh, 1); 4188 break; 4189 case IEEE80211_M_WDS: 4190 DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: bssid %s\n", 4191 if_name(vap->iv_ifp), __func__, 4192 ether_sprintf(ni->ni_bssid)); 4193 mwl_seteapolformat(vap); 4194 break; 4195 default: 4196 break; 4197 } 4198 /* 4199 * Set CS mode according to operating channel; 4200 * this mostly an optimization for 5GHz. 4201 * 4202 * NB: must follow mwl_hal_start which resets csmode 4203 */ 4204 if (IEEE80211_IS_CHAN_5GHZ(ic->ic_bsschan)) 4205 mwl_hal_setcsmode(mh, CSMODE_AGGRESSIVE); 4206 else 4207 mwl_hal_setcsmode(mh, CSMODE_AUTO_ENA); 4208 /* 4209 * Start timer to prod firmware. 4210 */ 4211 if (sc->sc_ageinterval != 0) 4212 callout_reset(&sc->sc_timer, sc->sc_ageinterval*hz, 4213 mwl_agestations, sc); 4214 } else if (nstate == IEEE80211_S_SLEEP) { 4215 /* XXX set chip in power save */ 4216 } else if ((vap->iv_flags & IEEE80211_F_DWDS) && 4217 --sc->sc_ndwdsvaps == 0) 4218 mwl_hal_setdwds(mh, 0); 4219 bad: 4220 return error; 4221 } 4222 4223 /* 4224 * Manage station id's; these are separate from AID's 4225 * as AID's may have values out of the range of possible 4226 * station id's acceptable to the firmware. 4227 */ 4228 static int 4229 allocstaid(struct mwl_softc *sc, int aid) 4230 { 4231 int staid; 4232 4233 if (!(0 < aid && aid < MWL_MAXSTAID) || isset(sc->sc_staid, aid)) { 4234 /* NB: don't use 0 */ 4235 for (staid = 1; staid < MWL_MAXSTAID; staid++) 4236 if (isclr(sc->sc_staid, staid)) 4237 break; 4238 } else 4239 staid = aid; 4240 setbit(sc->sc_staid, staid); 4241 return staid; 4242 } 4243 4244 static void 4245 delstaid(struct mwl_softc *sc, int staid) 4246 { 4247 clrbit(sc->sc_staid, staid); 4248 } 4249 4250 /* 4251 * Setup driver-specific state for a newly associated node. 4252 * Note that we're called also on a re-associate, the isnew 4253 * param tells us if this is the first time or not. 4254 */ 4255 static void 4256 mwl_newassoc(struct ieee80211_node *ni, int isnew) 4257 { 4258 struct ieee80211vap *vap = ni->ni_vap; 4259 struct mwl_softc *sc = vap->iv_ic->ic_softc; 4260 struct mwl_node *mn = MWL_NODE(ni); 4261 MWL_HAL_PEERINFO pi; 4262 uint16_t aid; 4263 int error; 4264 4265 aid = IEEE80211_AID(ni->ni_associd); 4266 if (isnew) { 4267 mn->mn_staid = allocstaid(sc, aid); 4268 mn->mn_hvap = MWL_VAP(vap)->mv_hvap; 4269 } else { 4270 mn = MWL_NODE(ni); 4271 /* XXX reset BA stream? */ 4272 } 4273 DPRINTF(sc, MWL_DEBUG_NODE, "%s: mac %s isnew %d aid %d staid %d\n", 4274 __func__, ether_sprintf(ni->ni_macaddr), isnew, aid, mn->mn_staid); 4275 error = mwl_peerstadb(ni, aid, mn->mn_staid, mkpeerinfo(&pi, ni)); 4276 if (error != 0) { 4277 DPRINTF(sc, MWL_DEBUG_NODE, 4278 "%s: error %d creating sta db entry\n", 4279 __func__, error); 4280 /* XXX how to deal with error? */ 4281 } 4282 } 4283 4284 /* 4285 * Periodically poke the firmware to age out station state 4286 * (power save queues, pending tx aggregates). 4287 */ 4288 static void 4289 mwl_agestations(void *arg) 4290 { 4291 struct mwl_softc *sc = arg; 4292 4293 mwl_hal_setkeepalive(sc->sc_mh); 4294 if (sc->sc_ageinterval != 0) /* NB: catch dynamic changes */ 4295 callout_schedule(&sc->sc_timer, sc->sc_ageinterval*hz); 4296 } 4297 4298 static const struct mwl_hal_channel * 4299 findhalchannel(const MWL_HAL_CHANNELINFO *ci, int ieee) 4300 { 4301 int i; 4302 4303 for (i = 0; i < ci->nchannels; i++) { 4304 const struct mwl_hal_channel *hc = &ci->channels[i]; 4305 if (hc->ieee == ieee) 4306 return hc; 4307 } 4308 return NULL; 4309 } 4310 4311 static int 4312 mwl_setregdomain(struct ieee80211com *ic, struct ieee80211_regdomain *rd, 4313 int nchan, struct ieee80211_channel chans[]) 4314 { 4315 struct mwl_softc *sc = ic->ic_softc; 4316 struct mwl_hal *mh = sc->sc_mh; 4317 const MWL_HAL_CHANNELINFO *ci; 4318 int i; 4319 4320 for (i = 0; i < nchan; i++) { 4321 struct ieee80211_channel *c = &chans[i]; 4322 const struct mwl_hal_channel *hc; 4323 4324 if (IEEE80211_IS_CHAN_2GHZ(c)) { 4325 mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_2DOT4GHZ, 4326 IEEE80211_IS_CHAN_HT40(c) ? 4327 MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci); 4328 } else if (IEEE80211_IS_CHAN_5GHZ(c)) { 4329 mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_5GHZ, 4330 IEEE80211_IS_CHAN_HT40(c) ? 4331 MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci); 4332 } else { 4333 device_printf(sc->sc_dev, 4334 "%s: channel %u freq %u/0x%x not 2.4/5GHz\n", 4335 __func__, c->ic_ieee, c->ic_freq, c->ic_flags); 4336 return EINVAL; 4337 } 4338 /* 4339 * Verify channel has cal data and cap tx power. 4340 */ 4341 hc = findhalchannel(ci, c->ic_ieee); 4342 if (hc != NULL) { 4343 if (c->ic_maxpower > 2*hc->maxTxPow) 4344 c->ic_maxpower = 2*hc->maxTxPow; 4345 goto next; 4346 } 4347 if (IEEE80211_IS_CHAN_HT40(c)) { 4348 /* 4349 * Look for the extension channel since the 4350 * hal table only has the primary channel. 4351 */ 4352 hc = findhalchannel(ci, c->ic_extieee); 4353 if (hc != NULL) { 4354 if (c->ic_maxpower > 2*hc->maxTxPow) 4355 c->ic_maxpower = 2*hc->maxTxPow; 4356 goto next; 4357 } 4358 } 4359 device_printf(sc->sc_dev, 4360 "%s: no cal data for channel %u ext %u freq %u/0x%x\n", 4361 __func__, c->ic_ieee, c->ic_extieee, 4362 c->ic_freq, c->ic_flags); 4363 return EINVAL; 4364 next: 4365 ; 4366 } 4367 return 0; 4368 } 4369 4370 #define IEEE80211_CHAN_HTG (IEEE80211_CHAN_HT|IEEE80211_CHAN_G) 4371 #define IEEE80211_CHAN_HTA (IEEE80211_CHAN_HT|IEEE80211_CHAN_A) 4372 4373 static void 4374 addht40channels(struct ieee80211_channel chans[], int maxchans, int *nchans, 4375 const MWL_HAL_CHANNELINFO *ci, int flags) 4376 { 4377 int i, error; 4378 4379 for (i = 0; i < ci->nchannels; i++) { 4380 const struct mwl_hal_channel *hc = &ci->channels[i]; 4381 4382 error = ieee80211_add_channel_ht40(chans, maxchans, nchans, 4383 hc->ieee, hc->maxTxPow, flags); 4384 if (error != 0 && error != ENOENT) 4385 break; 4386 } 4387 } 4388 4389 static void 4390 addchannels(struct ieee80211_channel chans[], int maxchans, int *nchans, 4391 const MWL_HAL_CHANNELINFO *ci, const uint8_t bands[]) 4392 { 4393 int i, error; 4394 4395 error = 0; 4396 for (i = 0; i < ci->nchannels && error == 0; i++) { 4397 const struct mwl_hal_channel *hc = &ci->channels[i]; 4398 4399 error = ieee80211_add_channel(chans, maxchans, nchans, 4400 hc->ieee, hc->freq, hc->maxTxPow, 0, bands); 4401 } 4402 } 4403 4404 static void 4405 getchannels(struct mwl_softc *sc, int maxchans, int *nchans, 4406 struct ieee80211_channel chans[]) 4407 { 4408 const MWL_HAL_CHANNELINFO *ci; 4409 uint8_t bands[IEEE80211_MODE_BYTES]; 4410 4411 /* 4412 * Use the channel info from the hal to craft the 4413 * channel list. Note that we pass back an unsorted 4414 * list; the caller is required to sort it for us 4415 * (if desired). 4416 */ 4417 *nchans = 0; 4418 if (mwl_hal_getchannelinfo(sc->sc_mh, 4419 MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0) { 4420 memset(bands, 0, sizeof(bands)); 4421 setbit(bands, IEEE80211_MODE_11B); 4422 setbit(bands, IEEE80211_MODE_11G); 4423 setbit(bands, IEEE80211_MODE_11NG); 4424 addchannels(chans, maxchans, nchans, ci, bands); 4425 } 4426 if (mwl_hal_getchannelinfo(sc->sc_mh, 4427 MWL_FREQ_BAND_5GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0) { 4428 memset(bands, 0, sizeof(bands)); 4429 setbit(bands, IEEE80211_MODE_11A); 4430 setbit(bands, IEEE80211_MODE_11NA); 4431 addchannels(chans, maxchans, nchans, ci, bands); 4432 } 4433 if (mwl_hal_getchannelinfo(sc->sc_mh, 4434 MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0) 4435 addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTG); 4436 if (mwl_hal_getchannelinfo(sc->sc_mh, 4437 MWL_FREQ_BAND_5GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0) 4438 addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTA); 4439 } 4440 4441 static void 4442 mwl_getradiocaps(struct ieee80211com *ic, 4443 int maxchans, int *nchans, struct ieee80211_channel chans[]) 4444 { 4445 struct mwl_softc *sc = ic->ic_softc; 4446 4447 getchannels(sc, maxchans, nchans, chans); 4448 } 4449 4450 static int 4451 mwl_getchannels(struct mwl_softc *sc) 4452 { 4453 struct ieee80211com *ic = &sc->sc_ic; 4454 4455 /* 4456 * Use the channel info from the hal to craft the 4457 * channel list for net80211. Note that we pass up 4458 * an unsorted list; net80211 will sort it for us. 4459 */ 4460 memset(ic->ic_channels, 0, sizeof(ic->ic_channels)); 4461 ic->ic_nchans = 0; 4462 getchannels(sc, IEEE80211_CHAN_MAX, &ic->ic_nchans, ic->ic_channels); 4463 4464 ic->ic_regdomain.regdomain = SKU_DEBUG; 4465 ic->ic_regdomain.country = CTRY_DEFAULT; 4466 ic->ic_regdomain.location = 'I'; 4467 ic->ic_regdomain.isocc[0] = ' '; /* XXX? */ 4468 ic->ic_regdomain.isocc[1] = ' '; 4469 return (ic->ic_nchans == 0 ? EIO : 0); 4470 } 4471 #undef IEEE80211_CHAN_HTA 4472 #undef IEEE80211_CHAN_HTG 4473 4474 #ifdef MWL_DEBUG 4475 static void 4476 mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix) 4477 { 4478 const struct mwl_rxdesc *ds = bf->bf_desc; 4479 uint32_t status = le32toh(ds->Status); 4480 4481 printf("R[%2u] (DS.V:%p DS.P:0x%jx) NEXT:%08x DATA:%08x RC:%02x%s\n" 4482 " STAT:%02x LEN:%04x RSSI:%02x CHAN:%02x RATE:%02x QOS:%04x HT:%04x\n", 4483 ix, ds, (uintmax_t)bf->bf_daddr, le32toh(ds->pPhysNext), 4484 le32toh(ds->pPhysBuffData), ds->RxControl, 4485 ds->RxControl != EAGLE_RXD_CTRL_DRIVER_OWN ? 4486 "" : (status & EAGLE_RXD_STATUS_OK) ? " *" : " !", 4487 ds->Status, le16toh(ds->PktLen), ds->RSSI, ds->Channel, 4488 ds->Rate, le16toh(ds->QosCtrl), le16toh(ds->HtSig2)); 4489 } 4490 4491 static void 4492 mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix) 4493 { 4494 const struct mwl_txdesc *ds = bf->bf_desc; 4495 uint32_t status = le32toh(ds->Status); 4496 4497 printf("Q%u[%3u]", qnum, ix); 4498 printf(" (DS.V:%p DS.P:0x%jx)\n", ds, (uintmax_t)bf->bf_daddr); 4499 printf(" NEXT:%08x DATA:%08x LEN:%04x STAT:%08x%s\n", 4500 le32toh(ds->pPhysNext), 4501 le32toh(ds->PktPtr), le16toh(ds->PktLen), status, 4502 status & EAGLE_TXD_STATUS_USED ? 4503 "" : (status & 3) != 0 ? " *" : " !"); 4504 printf(" RATE:%02x PRI:%x QOS:%04x SAP:%08x FORMAT:%04x\n", 4505 ds->DataRate, ds->TxPriority, le16toh(ds->QosCtrl), 4506 le32toh(ds->SapPktInfo), le16toh(ds->Format)); 4507 #if MWL_TXDESC > 1 4508 printf(" MULTIFRAMES:%u LEN:%04x %04x %04x %04x %04x %04x\n" 4509 , le32toh(ds->multiframes) 4510 , le16toh(ds->PktLenArray[0]), le16toh(ds->PktLenArray[1]) 4511 , le16toh(ds->PktLenArray[2]), le16toh(ds->PktLenArray[3]) 4512 , le16toh(ds->PktLenArray[4]), le16toh(ds->PktLenArray[5]) 4513 ); 4514 printf(" DATA:%08x %08x %08x %08x %08x %08x\n" 4515 , le32toh(ds->PktPtrArray[0]), le32toh(ds->PktPtrArray[1]) 4516 , le32toh(ds->PktPtrArray[2]), le32toh(ds->PktPtrArray[3]) 4517 , le32toh(ds->PktPtrArray[4]), le32toh(ds->PktPtrArray[5]) 4518 ); 4519 #endif 4520 #if 0 4521 { const uint8_t *cp = (const uint8_t *) ds; 4522 int i; 4523 for (i = 0; i < sizeof(struct mwl_txdesc); i++) { 4524 printf("%02x ", cp[i]); 4525 if (((i+1) % 16) == 0) 4526 printf("\n"); 4527 } 4528 printf("\n"); 4529 } 4530 #endif 4531 } 4532 #endif /* MWL_DEBUG */ 4533 4534 #if 0 4535 static void 4536 mwl_txq_dump(struct mwl_txq *txq) 4537 { 4538 struct mwl_txbuf *bf; 4539 int i = 0; 4540 4541 MWL_TXQ_LOCK(txq); 4542 STAILQ_FOREACH(bf, &txq->active, bf_list) { 4543 struct mwl_txdesc *ds = bf->bf_desc; 4544 MWL_TXDESC_SYNC(txq, ds, 4545 BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE); 4546 #ifdef MWL_DEBUG 4547 mwl_printtxbuf(bf, txq->qnum, i); 4548 #endif 4549 i++; 4550 } 4551 MWL_TXQ_UNLOCK(txq); 4552 } 4553 #endif 4554 4555 static void 4556 mwl_watchdog(void *arg) 4557 { 4558 struct mwl_softc *sc = arg; 4559 4560 callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc); 4561 if (sc->sc_tx_timer == 0 || --sc->sc_tx_timer > 0) 4562 return; 4563 4564 if (sc->sc_running && !sc->sc_invalid) { 4565 if (mwl_hal_setkeepalive(sc->sc_mh)) 4566 device_printf(sc->sc_dev, 4567 "transmit timeout (firmware hung?)\n"); 4568 else 4569 device_printf(sc->sc_dev, 4570 "transmit timeout\n"); 4571 #if 0 4572 mwl_reset(sc); 4573 mwl_txq_dump(&sc->sc_txq[0]);/*XXX*/ 4574 #endif 4575 counter_u64_add(sc->sc_ic.ic_oerrors, 1); 4576 sc->sc_stats.mst_watchdog++; 4577 } 4578 } 4579 4580 #ifdef MWL_DIAGAPI 4581 /* 4582 * Diagnostic interface to the HAL. This is used by various 4583 * tools to do things like retrieve register contents for 4584 * debugging. The mechanism is intentionally opaque so that 4585 * it can change frequently w/o concern for compatibility. 4586 */ 4587 static int 4588 mwl_ioctl_diag(struct mwl_softc *sc, struct mwl_diag *md) 4589 { 4590 struct mwl_hal *mh = sc->sc_mh; 4591 u_int id = md->md_id & MWL_DIAG_ID; 4592 void *indata = NULL; 4593 void *outdata = NULL; 4594 u_int32_t insize = md->md_in_size; 4595 u_int32_t outsize = md->md_out_size; 4596 int error = 0; 4597 4598 if (md->md_id & MWL_DIAG_IN) { 4599 /* 4600 * Copy in data. 4601 */ 4602 indata = malloc(insize, M_TEMP, M_NOWAIT); 4603 if (indata == NULL) { 4604 error = ENOMEM; 4605 goto bad; 4606 } 4607 error = copyin(md->md_in_data, indata, insize); 4608 if (error) 4609 goto bad; 4610 } 4611 if (md->md_id & MWL_DIAG_DYN) { 4612 /* 4613 * Allocate a buffer for the results (otherwise the HAL 4614 * returns a pointer to a buffer where we can read the 4615 * results). Note that we depend on the HAL leaving this 4616 * pointer for us to use below in reclaiming the buffer; 4617 * may want to be more defensive. 4618 */ 4619 outdata = malloc(outsize, M_TEMP, M_NOWAIT); 4620 if (outdata == NULL) { 4621 error = ENOMEM; 4622 goto bad; 4623 } 4624 } 4625 if (mwl_hal_getdiagstate(mh, id, indata, insize, &outdata, &outsize)) { 4626 if (outsize < md->md_out_size) 4627 md->md_out_size = outsize; 4628 if (outdata != NULL) 4629 error = copyout(outdata, md->md_out_data, 4630 md->md_out_size); 4631 } else { 4632 error = EINVAL; 4633 } 4634 bad: 4635 if ((md->md_id & MWL_DIAG_IN) && indata != NULL) 4636 free(indata, M_TEMP); 4637 if ((md->md_id & MWL_DIAG_DYN) && outdata != NULL) 4638 free(outdata, M_TEMP); 4639 return error; 4640 } 4641 4642 static int 4643 mwl_ioctl_reset(struct mwl_softc *sc, struct mwl_diag *md) 4644 { 4645 struct mwl_hal *mh = sc->sc_mh; 4646 int error; 4647 4648 MWL_LOCK_ASSERT(sc); 4649 4650 if (md->md_id == 0 && mwl_hal_fwload(mh, NULL) != 0) { 4651 device_printf(sc->sc_dev, "unable to load firmware\n"); 4652 return EIO; 4653 } 4654 if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) { 4655 device_printf(sc->sc_dev, "unable to fetch h/w specs\n"); 4656 return EIO; 4657 } 4658 error = mwl_setupdma(sc); 4659 if (error != 0) { 4660 /* NB: mwl_setupdma prints a msg */ 4661 return error; 4662 } 4663 /* 4664 * Reset tx/rx data structures; after reload we must 4665 * re-start the driver's notion of the next xmit/recv. 4666 */ 4667 mwl_draintxq(sc); /* clear pending frames */ 4668 mwl_resettxq(sc); /* rebuild tx q lists */ 4669 sc->sc_rxnext = NULL; /* force rx to start at the list head */ 4670 return 0; 4671 } 4672 #endif /* MWL_DIAGAPI */ 4673 4674 static void 4675 mwl_parent(struct ieee80211com *ic) 4676 { 4677 struct mwl_softc *sc = ic->ic_softc; 4678 int startall = 0; 4679 4680 MWL_LOCK(sc); 4681 if (ic->ic_nrunning > 0) { 4682 if (sc->sc_running) { 4683 /* 4684 * To avoid rescanning another access point, 4685 * do not call mwl_init() here. Instead, 4686 * only reflect promisc mode settings. 4687 */ 4688 mwl_mode_init(sc); 4689 } else { 4690 /* 4691 * Beware of being called during attach/detach 4692 * to reset promiscuous mode. In that case we 4693 * will still be marked UP but not RUNNING. 4694 * However trying to re-init the interface 4695 * is the wrong thing to do as we've already 4696 * torn down much of our state. There's 4697 * probably a better way to deal with this. 4698 */ 4699 if (!sc->sc_invalid) { 4700 mwl_init(sc); /* XXX lose error */ 4701 startall = 1; 4702 } 4703 } 4704 } else 4705 mwl_stop(sc); 4706 MWL_UNLOCK(sc); 4707 if (startall) 4708 ieee80211_start_all(ic); 4709 } 4710 4711 static int 4712 mwl_ioctl(struct ieee80211com *ic, u_long cmd, void *data) 4713 { 4714 struct mwl_softc *sc = ic->ic_softc; 4715 struct ifreq *ifr = data; 4716 int error = 0; 4717 4718 switch (cmd) { 4719 case SIOCGMVSTATS: 4720 mwl_hal_gethwstats(sc->sc_mh, &sc->sc_stats.hw_stats); 4721 #if 0 4722 /* NB: embed these numbers to get a consistent view */ 4723 sc->sc_stats.mst_tx_packets = 4724 if_get_counter(ifp, IFCOUNTER_OPACKETS); 4725 sc->sc_stats.mst_rx_packets = 4726 if_get_counter(ifp, IFCOUNTER_IPACKETS); 4727 #endif 4728 /* 4729 * NB: Drop the softc lock in case of a page fault; 4730 * we'll accept any potential inconsisentcy in the 4731 * statistics. The alternative is to copy the data 4732 * to a local structure. 4733 */ 4734 return (copyout(&sc->sc_stats, ifr_data_get_ptr(ifr), 4735 sizeof (sc->sc_stats))); 4736 #ifdef MWL_DIAGAPI 4737 case SIOCGMVDIAG: 4738 /* XXX check privs */ 4739 return mwl_ioctl_diag(sc, (struct mwl_diag *) ifr); 4740 case SIOCGMVRESET: 4741 /* XXX check privs */ 4742 MWL_LOCK(sc); 4743 error = mwl_ioctl_reset(sc,(struct mwl_diag *) ifr); 4744 MWL_UNLOCK(sc); 4745 break; 4746 #endif /* MWL_DIAGAPI */ 4747 default: 4748 error = ENOTTY; 4749 break; 4750 } 4751 return (error); 4752 } 4753 4754 #ifdef MWL_DEBUG 4755 static int 4756 mwl_sysctl_debug(SYSCTL_HANDLER_ARGS) 4757 { 4758 struct mwl_softc *sc = arg1; 4759 int debug, error; 4760 4761 debug = sc->sc_debug | (mwl_hal_getdebug(sc->sc_mh) << 24); 4762 error = sysctl_handle_int(oidp, &debug, 0, req); 4763 if (error || !req->newptr) 4764 return error; 4765 mwl_hal_setdebug(sc->sc_mh, debug >> 24); 4766 sc->sc_debug = debug & 0x00ffffff; 4767 return 0; 4768 } 4769 #endif /* MWL_DEBUG */ 4770 4771 static void 4772 mwl_sysctlattach(struct mwl_softc *sc) 4773 { 4774 #ifdef MWL_DEBUG 4775 struct sysctl_ctx_list *ctx = device_get_sysctl_ctx(sc->sc_dev); 4776 struct sysctl_oid *tree = device_get_sysctl_tree(sc->sc_dev); 4777 4778 sc->sc_debug = mwl_debug; 4779 SYSCTL_ADD_PROC(ctx, SYSCTL_CHILDREN(tree), OID_AUTO, "debug", 4780 CTLTYPE_INT | CTLFLAG_RW | CTLFLAG_NEEDGIANT, sc, 0, 4781 mwl_sysctl_debug, "I", "control debugging printfs"); 4782 #endif 4783 } 4784 4785 /* 4786 * Announce various information on device/driver attach. 4787 */ 4788 static void 4789 mwl_announce(struct mwl_softc *sc) 4790 { 4791 4792 device_printf(sc->sc_dev, "Rev A%d hardware, v%d.%d.%d.%d firmware (regioncode %d)\n", 4793 sc->sc_hwspecs.hwVersion, 4794 (sc->sc_hwspecs.fwReleaseNumber>>24) & 0xff, 4795 (sc->sc_hwspecs.fwReleaseNumber>>16) & 0xff, 4796 (sc->sc_hwspecs.fwReleaseNumber>>8) & 0xff, 4797 (sc->sc_hwspecs.fwReleaseNumber>>0) & 0xff, 4798 sc->sc_hwspecs.regionCode); 4799 sc->sc_fwrelease = sc->sc_hwspecs.fwReleaseNumber; 4800 4801 if (bootverbose) { 4802 int i; 4803 for (i = 0; i <= WME_AC_VO; i++) { 4804 struct mwl_txq *txq = sc->sc_ac2q[i]; 4805 device_printf(sc->sc_dev, "Use hw queue %u for %s traffic\n", 4806 txq->qnum, ieee80211_wme_acnames[i]); 4807 } 4808 } 4809 if (bootverbose || mwl_rxdesc != MWL_RXDESC) 4810 device_printf(sc->sc_dev, "using %u rx descriptors\n", mwl_rxdesc); 4811 if (bootverbose || mwl_rxbuf != MWL_RXBUF) 4812 device_printf(sc->sc_dev, "using %u rx buffers\n", mwl_rxbuf); 4813 if (bootverbose || mwl_txbuf != MWL_TXBUF) 4814 device_printf(sc->sc_dev, "using %u tx buffers\n", mwl_txbuf); 4815 if (bootverbose && mwl_hal_ismbsscapable(sc->sc_mh)) 4816 device_printf(sc->sc_dev, "multi-bss support\n"); 4817 #ifdef MWL_TX_NODROP 4818 if (bootverbose) 4819 device_printf(sc->sc_dev, "no tx drop\n"); 4820 #endif 4821 } 4822