1 /*- 2 * Copyright (c) 2007-2009 Sam Leffler, Errno Consulting 3 * Copyright (c) 2007-2008 Marvell Semiconductor, Inc. 4 * All rights reserved. 5 * 6 * Redistribution and use in source and binary forms, with or without 7 * modification, are permitted provided that the following conditions 8 * are met: 9 * 1. Redistributions of source code must retain the above copyright 10 * notice, this list of conditions and the following disclaimer, 11 * without modification. 12 * 2. Redistributions in binary form must reproduce at minimum a disclaimer 13 * similar to the "NO WARRANTY" disclaimer below ("Disclaimer") and any 14 * redistribution must be conditioned upon including a substantially 15 * similar Disclaimer requirement for further binary redistribution. 16 * 17 * NO WARRANTY 18 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 19 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 20 * LIMITED TO, THE IMPLIED WARRANTIES OF NONINFRINGEMENT, MERCHANTIBILITY 21 * AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL 22 * THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY, 23 * OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 24 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 25 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER 26 * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 27 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF 28 * THE POSSIBILITY OF SUCH DAMAGES. 29 */ 30 31 #include <sys/cdefs.h> 32 __FBSDID("$FreeBSD$"); 33 34 /* 35 * Driver for the Marvell 88W8363 Wireless LAN controller. 36 */ 37 38 #include "opt_inet.h" 39 #include "opt_mwl.h" 40 #include "opt_wlan.h" 41 42 #include <sys/param.h> 43 #include <sys/systm.h> 44 #include <sys/sysctl.h> 45 #include <sys/mbuf.h> 46 #include <sys/malloc.h> 47 #include <sys/lock.h> 48 #include <sys/mutex.h> 49 #include <sys/kernel.h> 50 #include <sys/socket.h> 51 #include <sys/sockio.h> 52 #include <sys/errno.h> 53 #include <sys/callout.h> 54 #include <sys/bus.h> 55 #include <sys/endian.h> 56 #include <sys/kthread.h> 57 #include <sys/taskqueue.h> 58 59 #include <machine/bus.h> 60 61 #include <net/if.h> 62 #include <net/if_var.h> 63 #include <net/if_dl.h> 64 #include <net/if_media.h> 65 #include <net/if_types.h> 66 #include <net/if_arp.h> 67 #include <net/ethernet.h> 68 #include <net/if_llc.h> 69 70 #include <net/bpf.h> 71 72 #include <net80211/ieee80211_var.h> 73 #include <net80211/ieee80211_regdomain.h> 74 75 #ifdef INET 76 #include <netinet/in.h> 77 #include <netinet/if_ether.h> 78 #endif /* INET */ 79 80 #include <dev/mwl/if_mwlvar.h> 81 #include <dev/mwl/mwldiag.h> 82 83 /* idiomatic shorthands: MS = mask+shift, SM = shift+mask */ 84 #define MS(v,x) (((v) & x) >> x##_S) 85 #define SM(v,x) (((v) << x##_S) & x) 86 87 static struct ieee80211vap *mwl_vap_create(struct ieee80211com *, 88 const char [IFNAMSIZ], int, enum ieee80211_opmode, int, 89 const uint8_t [IEEE80211_ADDR_LEN], 90 const uint8_t [IEEE80211_ADDR_LEN]); 91 static void mwl_vap_delete(struct ieee80211vap *); 92 static int mwl_setupdma(struct mwl_softc *); 93 static int mwl_hal_reset(struct mwl_softc *sc); 94 static int mwl_init_locked(struct mwl_softc *); 95 static void mwl_init(void *); 96 static void mwl_stop_locked(struct ifnet *, int); 97 static int mwl_reset(struct ieee80211vap *, u_long); 98 static void mwl_stop(struct ifnet *, int); 99 static void mwl_start(struct ifnet *); 100 static int mwl_raw_xmit(struct ieee80211_node *, struct mbuf *, 101 const struct ieee80211_bpf_params *); 102 static int mwl_media_change(struct ifnet *); 103 static void mwl_watchdog(void *); 104 static int mwl_ioctl(struct ifnet *, u_long, caddr_t); 105 static void mwl_radar_proc(void *, int); 106 static void mwl_chanswitch_proc(void *, int); 107 static void mwl_bawatchdog_proc(void *, int); 108 static int mwl_key_alloc(struct ieee80211vap *, 109 struct ieee80211_key *, 110 ieee80211_keyix *, ieee80211_keyix *); 111 static int mwl_key_delete(struct ieee80211vap *, 112 const struct ieee80211_key *); 113 static int mwl_key_set(struct ieee80211vap *, const struct ieee80211_key *, 114 const uint8_t mac[IEEE80211_ADDR_LEN]); 115 static int mwl_mode_init(struct mwl_softc *); 116 static void mwl_update_mcast(struct ifnet *); 117 static void mwl_update_promisc(struct ifnet *); 118 static void mwl_updateslot(struct ifnet *); 119 static int mwl_beacon_setup(struct ieee80211vap *); 120 static void mwl_beacon_update(struct ieee80211vap *, int); 121 #ifdef MWL_HOST_PS_SUPPORT 122 static void mwl_update_ps(struct ieee80211vap *, int); 123 static int mwl_set_tim(struct ieee80211_node *, int); 124 #endif 125 static int mwl_dma_setup(struct mwl_softc *); 126 static void mwl_dma_cleanup(struct mwl_softc *); 127 static struct ieee80211_node *mwl_node_alloc(struct ieee80211vap *, 128 const uint8_t [IEEE80211_ADDR_LEN]); 129 static void mwl_node_cleanup(struct ieee80211_node *); 130 static void mwl_node_drain(struct ieee80211_node *); 131 static void mwl_node_getsignal(const struct ieee80211_node *, 132 int8_t *, int8_t *); 133 static void mwl_node_getmimoinfo(const struct ieee80211_node *, 134 struct ieee80211_mimo_info *); 135 static int mwl_rxbuf_init(struct mwl_softc *, struct mwl_rxbuf *); 136 static void mwl_rx_proc(void *, int); 137 static void mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *, int); 138 static int mwl_tx_setup(struct mwl_softc *, int, int); 139 static int mwl_wme_update(struct ieee80211com *); 140 static void mwl_tx_cleanupq(struct mwl_softc *, struct mwl_txq *); 141 static void mwl_tx_cleanup(struct mwl_softc *); 142 static uint16_t mwl_calcformat(uint8_t rate, const struct ieee80211_node *); 143 static int mwl_tx_start(struct mwl_softc *, struct ieee80211_node *, 144 struct mwl_txbuf *, struct mbuf *); 145 static void mwl_tx_proc(void *, int); 146 static int mwl_chan_set(struct mwl_softc *, struct ieee80211_channel *); 147 static void mwl_draintxq(struct mwl_softc *); 148 static void mwl_cleartxq(struct mwl_softc *, struct ieee80211vap *); 149 static int mwl_recv_action(struct ieee80211_node *, 150 const struct ieee80211_frame *, 151 const uint8_t *, const uint8_t *); 152 static int mwl_addba_request(struct ieee80211_node *, 153 struct ieee80211_tx_ampdu *, int dialogtoken, 154 int baparamset, int batimeout); 155 static int mwl_addba_response(struct ieee80211_node *, 156 struct ieee80211_tx_ampdu *, int status, 157 int baparamset, int batimeout); 158 static void mwl_addba_stop(struct ieee80211_node *, 159 struct ieee80211_tx_ampdu *); 160 static int mwl_startrecv(struct mwl_softc *); 161 static MWL_HAL_APMODE mwl_getapmode(const struct ieee80211vap *, 162 struct ieee80211_channel *); 163 static int mwl_setapmode(struct ieee80211vap *, struct ieee80211_channel*); 164 static void mwl_scan_start(struct ieee80211com *); 165 static void mwl_scan_end(struct ieee80211com *); 166 static void mwl_set_channel(struct ieee80211com *); 167 static int mwl_peerstadb(struct ieee80211_node *, 168 int aid, int staid, MWL_HAL_PEERINFO *pi); 169 static int mwl_localstadb(struct ieee80211vap *); 170 static int mwl_newstate(struct ieee80211vap *, enum ieee80211_state, int); 171 static int allocstaid(struct mwl_softc *sc, int aid); 172 static void delstaid(struct mwl_softc *sc, int staid); 173 static void mwl_newassoc(struct ieee80211_node *, int); 174 static void mwl_agestations(void *); 175 static int mwl_setregdomain(struct ieee80211com *, 176 struct ieee80211_regdomain *, int, 177 struct ieee80211_channel []); 178 static void mwl_getradiocaps(struct ieee80211com *, int, int *, 179 struct ieee80211_channel []); 180 static int mwl_getchannels(struct mwl_softc *); 181 182 static void mwl_sysctlattach(struct mwl_softc *); 183 static void mwl_announce(struct mwl_softc *); 184 185 SYSCTL_NODE(_hw, OID_AUTO, mwl, CTLFLAG_RD, 0, "Marvell driver parameters"); 186 187 static int mwl_rxdesc = MWL_RXDESC; /* # rx desc's to allocate */ 188 SYSCTL_INT(_hw_mwl, OID_AUTO, rxdesc, CTLFLAG_RW, &mwl_rxdesc, 189 0, "rx descriptors allocated"); 190 static int mwl_rxbuf = MWL_RXBUF; /* # rx buffers to allocate */ 191 SYSCTL_INT(_hw_mwl, OID_AUTO, rxbuf, CTLFLAG_RW, &mwl_rxbuf, 192 0, "rx buffers allocated"); 193 TUNABLE_INT("hw.mwl.rxbuf", &mwl_rxbuf); 194 static int mwl_txbuf = MWL_TXBUF; /* # tx buffers to allocate */ 195 SYSCTL_INT(_hw_mwl, OID_AUTO, txbuf, CTLFLAG_RW, &mwl_txbuf, 196 0, "tx buffers allocated"); 197 TUNABLE_INT("hw.mwl.txbuf", &mwl_txbuf); 198 static int mwl_txcoalesce = 8; /* # tx packets to q before poking f/w*/ 199 SYSCTL_INT(_hw_mwl, OID_AUTO, txcoalesce, CTLFLAG_RW, &mwl_txcoalesce, 200 0, "tx buffers to send at once"); 201 TUNABLE_INT("hw.mwl.txcoalesce", &mwl_txcoalesce); 202 static int mwl_rxquota = MWL_RXBUF; /* # max buffers to process */ 203 SYSCTL_INT(_hw_mwl, OID_AUTO, rxquota, CTLFLAG_RW, &mwl_rxquota, 204 0, "max rx buffers to process per interrupt"); 205 TUNABLE_INT("hw.mwl.rxquota", &mwl_rxquota); 206 static int mwl_rxdmalow = 3; /* # min buffers for wakeup */ 207 SYSCTL_INT(_hw_mwl, OID_AUTO, rxdmalow, CTLFLAG_RW, &mwl_rxdmalow, 208 0, "min free rx buffers before restarting traffic"); 209 TUNABLE_INT("hw.mwl.rxdmalow", &mwl_rxdmalow); 210 211 #ifdef MWL_DEBUG 212 static int mwl_debug = 0; 213 SYSCTL_INT(_hw_mwl, OID_AUTO, debug, CTLFLAG_RW, &mwl_debug, 214 0, "control debugging printfs"); 215 TUNABLE_INT("hw.mwl.debug", &mwl_debug); 216 enum { 217 MWL_DEBUG_XMIT = 0x00000001, /* basic xmit operation */ 218 MWL_DEBUG_XMIT_DESC = 0x00000002, /* xmit descriptors */ 219 MWL_DEBUG_RECV = 0x00000004, /* basic recv operation */ 220 MWL_DEBUG_RECV_DESC = 0x00000008, /* recv descriptors */ 221 MWL_DEBUG_RESET = 0x00000010, /* reset processing */ 222 MWL_DEBUG_BEACON = 0x00000020, /* beacon handling */ 223 MWL_DEBUG_INTR = 0x00000040, /* ISR */ 224 MWL_DEBUG_TX_PROC = 0x00000080, /* tx ISR proc */ 225 MWL_DEBUG_RX_PROC = 0x00000100, /* rx ISR proc */ 226 MWL_DEBUG_KEYCACHE = 0x00000200, /* key cache management */ 227 MWL_DEBUG_STATE = 0x00000400, /* 802.11 state transitions */ 228 MWL_DEBUG_NODE = 0x00000800, /* node management */ 229 MWL_DEBUG_RECV_ALL = 0x00001000, /* trace all frames (beacons) */ 230 MWL_DEBUG_TSO = 0x00002000, /* TSO processing */ 231 MWL_DEBUG_AMPDU = 0x00004000, /* BA stream handling */ 232 MWL_DEBUG_ANY = 0xffffffff 233 }; 234 #define IS_BEACON(wh) \ 235 ((wh->i_fc[0] & (IEEE80211_FC0_TYPE_MASK|IEEE80211_FC0_SUBTYPE_MASK)) == \ 236 (IEEE80211_FC0_TYPE_MGT|IEEE80211_FC0_SUBTYPE_BEACON)) 237 #define IFF_DUMPPKTS_RECV(sc, wh) \ 238 (((sc->sc_debug & MWL_DEBUG_RECV) && \ 239 ((sc->sc_debug & MWL_DEBUG_RECV_ALL) || !IS_BEACON(wh))) || \ 240 (sc->sc_ifp->if_flags & (IFF_DEBUG|IFF_LINK2)) == (IFF_DEBUG|IFF_LINK2)) 241 #define IFF_DUMPPKTS_XMIT(sc) \ 242 ((sc->sc_debug & MWL_DEBUG_XMIT) || \ 243 (sc->sc_ifp->if_flags & (IFF_DEBUG|IFF_LINK2)) == (IFF_DEBUG|IFF_LINK2)) 244 #define DPRINTF(sc, m, fmt, ...) do { \ 245 if (sc->sc_debug & (m)) \ 246 printf(fmt, __VA_ARGS__); \ 247 } while (0) 248 #define KEYPRINTF(sc, hk, mac) do { \ 249 if (sc->sc_debug & MWL_DEBUG_KEYCACHE) \ 250 mwl_keyprint(sc, __func__, hk, mac); \ 251 } while (0) 252 static void mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix); 253 static void mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix); 254 #else 255 #define IFF_DUMPPKTS_RECV(sc, wh) \ 256 ((sc->sc_ifp->if_flags & (IFF_DEBUG|IFF_LINK2)) == (IFF_DEBUG|IFF_LINK2)) 257 #define IFF_DUMPPKTS_XMIT(sc) \ 258 ((sc->sc_ifp->if_flags & (IFF_DEBUG|IFF_LINK2)) == (IFF_DEBUG|IFF_LINK2)) 259 #define DPRINTF(sc, m, fmt, ...) do { \ 260 (void) sc; \ 261 } while (0) 262 #define KEYPRINTF(sc, k, mac) do { \ 263 (void) sc; \ 264 } while (0) 265 #endif 266 267 static MALLOC_DEFINE(M_MWLDEV, "mwldev", "mwl driver dma buffers"); 268 269 /* 270 * Each packet has fixed front matter: a 2-byte length 271 * of the payload, followed by a 4-address 802.11 header 272 * (regardless of the actual header and always w/o any 273 * QoS header). The payload then follows. 274 */ 275 struct mwltxrec { 276 uint16_t fwlen; 277 struct ieee80211_frame_addr4 wh; 278 } __packed; 279 280 /* 281 * Read/Write shorthands for accesses to BAR 0. Note 282 * that all BAR 1 operations are done in the "hal" and 283 * there should be no reference to them here. 284 */ 285 #ifdef MWL_DEBUG 286 static __inline uint32_t 287 RD4(struct mwl_softc *sc, bus_size_t off) 288 { 289 return bus_space_read_4(sc->sc_io0t, sc->sc_io0h, off); 290 } 291 #endif 292 293 static __inline void 294 WR4(struct mwl_softc *sc, bus_size_t off, uint32_t val) 295 { 296 bus_space_write_4(sc->sc_io0t, sc->sc_io0h, off, val); 297 } 298 299 int 300 mwl_attach(uint16_t devid, struct mwl_softc *sc) 301 { 302 struct ifnet *ifp; 303 struct ieee80211com *ic; 304 struct mwl_hal *mh; 305 int error = 0; 306 307 DPRINTF(sc, MWL_DEBUG_ANY, "%s: devid 0x%x\n", __func__, devid); 308 309 ifp = sc->sc_ifp = if_alloc(IFT_IEEE80211); 310 if (ifp == NULL) { 311 device_printf(sc->sc_dev, "cannot if_alloc()\n"); 312 return ENOSPC; 313 } 314 ic = ifp->if_l2com; 315 316 /* 317 * Setup the RX free list lock early, so it can be consistently 318 * removed. 319 */ 320 MWL_RXFREE_INIT(sc); 321 322 /* set these up early for if_printf use */ 323 if_initname(ifp, device_get_name(sc->sc_dev), 324 device_get_unit(sc->sc_dev)); 325 326 mh = mwl_hal_attach(sc->sc_dev, devid, 327 sc->sc_io1h, sc->sc_io1t, sc->sc_dmat); 328 if (mh == NULL) { 329 if_printf(ifp, "unable to attach HAL\n"); 330 error = EIO; 331 goto bad; 332 } 333 sc->sc_mh = mh; 334 /* 335 * Load firmware so we can get setup. We arbitrarily 336 * pick station firmware; we'll re-load firmware as 337 * needed so setting up the wrong mode isn't a big deal. 338 */ 339 if (mwl_hal_fwload(mh, NULL) != 0) { 340 if_printf(ifp, "unable to setup builtin firmware\n"); 341 error = EIO; 342 goto bad1; 343 } 344 if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) { 345 if_printf(ifp, "unable to fetch h/w specs\n"); 346 error = EIO; 347 goto bad1; 348 } 349 error = mwl_getchannels(sc); 350 if (error != 0) 351 goto bad1; 352 353 sc->sc_txantenna = 0; /* h/w default */ 354 sc->sc_rxantenna = 0; /* h/w default */ 355 sc->sc_invalid = 0; /* ready to go, enable int handling */ 356 sc->sc_ageinterval = MWL_AGEINTERVAL; 357 358 /* 359 * Allocate tx+rx descriptors and populate the lists. 360 * We immediately push the information to the firmware 361 * as otherwise it gets upset. 362 */ 363 error = mwl_dma_setup(sc); 364 if (error != 0) { 365 if_printf(ifp, "failed to setup descriptors: %d\n", error); 366 goto bad1; 367 } 368 error = mwl_setupdma(sc); /* push to firmware */ 369 if (error != 0) /* NB: mwl_setupdma prints msg */ 370 goto bad1; 371 372 callout_init(&sc->sc_timer, CALLOUT_MPSAFE); 373 callout_init_mtx(&sc->sc_watchdog, &sc->sc_mtx, 0); 374 375 sc->sc_tq = taskqueue_create("mwl_taskq", M_NOWAIT, 376 taskqueue_thread_enqueue, &sc->sc_tq); 377 taskqueue_start_threads(&sc->sc_tq, 1, PI_NET, 378 "%s taskq", ifp->if_xname); 379 380 TASK_INIT(&sc->sc_rxtask, 0, mwl_rx_proc, sc); 381 TASK_INIT(&sc->sc_radartask, 0, mwl_radar_proc, sc); 382 TASK_INIT(&sc->sc_chanswitchtask, 0, mwl_chanswitch_proc, sc); 383 TASK_INIT(&sc->sc_bawatchdogtask, 0, mwl_bawatchdog_proc, sc); 384 385 /* NB: insure BK queue is the lowest priority h/w queue */ 386 if (!mwl_tx_setup(sc, WME_AC_BK, MWL_WME_AC_BK)) { 387 if_printf(ifp, "unable to setup xmit queue for %s traffic!\n", 388 ieee80211_wme_acnames[WME_AC_BK]); 389 error = EIO; 390 goto bad2; 391 } 392 if (!mwl_tx_setup(sc, WME_AC_BE, MWL_WME_AC_BE) || 393 !mwl_tx_setup(sc, WME_AC_VI, MWL_WME_AC_VI) || 394 !mwl_tx_setup(sc, WME_AC_VO, MWL_WME_AC_VO)) { 395 /* 396 * Not enough hardware tx queues to properly do WME; 397 * just punt and assign them all to the same h/w queue. 398 * We could do a better job of this if, for example, 399 * we allocate queues when we switch from station to 400 * AP mode. 401 */ 402 if (sc->sc_ac2q[WME_AC_VI] != NULL) 403 mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_VI]); 404 if (sc->sc_ac2q[WME_AC_BE] != NULL) 405 mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_BE]); 406 sc->sc_ac2q[WME_AC_BE] = sc->sc_ac2q[WME_AC_BK]; 407 sc->sc_ac2q[WME_AC_VI] = sc->sc_ac2q[WME_AC_BK]; 408 sc->sc_ac2q[WME_AC_VO] = sc->sc_ac2q[WME_AC_BK]; 409 } 410 TASK_INIT(&sc->sc_txtask, 0, mwl_tx_proc, sc); 411 412 ifp->if_softc = sc; 413 ifp->if_flags = IFF_SIMPLEX | IFF_BROADCAST | IFF_MULTICAST; 414 ifp->if_start = mwl_start; 415 ifp->if_ioctl = mwl_ioctl; 416 ifp->if_init = mwl_init; 417 IFQ_SET_MAXLEN(&ifp->if_snd, ifqmaxlen); 418 ifp->if_snd.ifq_drv_maxlen = ifqmaxlen; 419 IFQ_SET_READY(&ifp->if_snd); 420 421 ic->ic_ifp = ifp; 422 /* XXX not right but it's not used anywhere important */ 423 ic->ic_phytype = IEEE80211_T_OFDM; 424 ic->ic_opmode = IEEE80211_M_STA; 425 ic->ic_caps = 426 IEEE80211_C_STA /* station mode supported */ 427 | IEEE80211_C_HOSTAP /* hostap mode */ 428 | IEEE80211_C_MONITOR /* monitor mode */ 429 #if 0 430 | IEEE80211_C_IBSS /* ibss, nee adhoc, mode */ 431 | IEEE80211_C_AHDEMO /* adhoc demo mode */ 432 #endif 433 | IEEE80211_C_MBSS /* mesh point link mode */ 434 | IEEE80211_C_WDS /* WDS supported */ 435 | IEEE80211_C_SHPREAMBLE /* short preamble supported */ 436 | IEEE80211_C_SHSLOT /* short slot time supported */ 437 | IEEE80211_C_WME /* WME/WMM supported */ 438 | IEEE80211_C_BURST /* xmit bursting supported */ 439 | IEEE80211_C_WPA /* capable of WPA1+WPA2 */ 440 | IEEE80211_C_BGSCAN /* capable of bg scanning */ 441 | IEEE80211_C_TXFRAG /* handle tx frags */ 442 | IEEE80211_C_TXPMGT /* capable of txpow mgt */ 443 | IEEE80211_C_DFS /* DFS supported */ 444 ; 445 446 ic->ic_htcaps = 447 IEEE80211_HTCAP_SMPS_ENA /* SM PS mode enabled */ 448 | IEEE80211_HTCAP_CHWIDTH40 /* 40MHz channel width */ 449 | IEEE80211_HTCAP_SHORTGI20 /* short GI in 20MHz */ 450 | IEEE80211_HTCAP_SHORTGI40 /* short GI in 40MHz */ 451 | IEEE80211_HTCAP_RXSTBC_2STREAM/* 1-2 spatial streams */ 452 #if MWL_AGGR_SIZE == 7935 453 | IEEE80211_HTCAP_MAXAMSDU_7935 /* max A-MSDU length */ 454 #else 455 | IEEE80211_HTCAP_MAXAMSDU_3839 /* max A-MSDU length */ 456 #endif 457 #if 0 458 | IEEE80211_HTCAP_PSMP /* PSMP supported */ 459 | IEEE80211_HTCAP_40INTOLERANT /* 40MHz intolerant */ 460 #endif 461 /* s/w capabilities */ 462 | IEEE80211_HTC_HT /* HT operation */ 463 | IEEE80211_HTC_AMPDU /* tx A-MPDU */ 464 | IEEE80211_HTC_AMSDU /* tx A-MSDU */ 465 | IEEE80211_HTC_SMPS /* SMPS available */ 466 ; 467 468 /* 469 * Mark h/w crypto support. 470 * XXX no way to query h/w support. 471 */ 472 ic->ic_cryptocaps |= IEEE80211_CRYPTO_WEP 473 | IEEE80211_CRYPTO_AES_CCM 474 | IEEE80211_CRYPTO_TKIP 475 | IEEE80211_CRYPTO_TKIPMIC 476 ; 477 /* 478 * Transmit requires space in the packet for a special 479 * format transmit record and optional padding between 480 * this record and the payload. Ask the net80211 layer 481 * to arrange this when encapsulating packets so we can 482 * add it efficiently. 483 */ 484 ic->ic_headroom = sizeof(struct mwltxrec) - 485 sizeof(struct ieee80211_frame); 486 487 /* call MI attach routine. */ 488 ieee80211_ifattach(ic, sc->sc_hwspecs.macAddr); 489 ic->ic_setregdomain = mwl_setregdomain; 490 ic->ic_getradiocaps = mwl_getradiocaps; 491 /* override default methods */ 492 ic->ic_raw_xmit = mwl_raw_xmit; 493 ic->ic_newassoc = mwl_newassoc; 494 ic->ic_updateslot = mwl_updateslot; 495 ic->ic_update_mcast = mwl_update_mcast; 496 ic->ic_update_promisc = mwl_update_promisc; 497 ic->ic_wme.wme_update = mwl_wme_update; 498 499 ic->ic_node_alloc = mwl_node_alloc; 500 sc->sc_node_cleanup = ic->ic_node_cleanup; 501 ic->ic_node_cleanup = mwl_node_cleanup; 502 sc->sc_node_drain = ic->ic_node_drain; 503 ic->ic_node_drain = mwl_node_drain; 504 ic->ic_node_getsignal = mwl_node_getsignal; 505 ic->ic_node_getmimoinfo = mwl_node_getmimoinfo; 506 507 ic->ic_scan_start = mwl_scan_start; 508 ic->ic_scan_end = mwl_scan_end; 509 ic->ic_set_channel = mwl_set_channel; 510 511 sc->sc_recv_action = ic->ic_recv_action; 512 ic->ic_recv_action = mwl_recv_action; 513 sc->sc_addba_request = ic->ic_addba_request; 514 ic->ic_addba_request = mwl_addba_request; 515 sc->sc_addba_response = ic->ic_addba_response; 516 ic->ic_addba_response = mwl_addba_response; 517 sc->sc_addba_stop = ic->ic_addba_stop; 518 ic->ic_addba_stop = mwl_addba_stop; 519 520 ic->ic_vap_create = mwl_vap_create; 521 ic->ic_vap_delete = mwl_vap_delete; 522 523 ieee80211_radiotap_attach(ic, 524 &sc->sc_tx_th.wt_ihdr, sizeof(sc->sc_tx_th), 525 MWL_TX_RADIOTAP_PRESENT, 526 &sc->sc_rx_th.wr_ihdr, sizeof(sc->sc_rx_th), 527 MWL_RX_RADIOTAP_PRESENT); 528 /* 529 * Setup dynamic sysctl's now that country code and 530 * regdomain are available from the hal. 531 */ 532 mwl_sysctlattach(sc); 533 534 if (bootverbose) 535 ieee80211_announce(ic); 536 mwl_announce(sc); 537 return 0; 538 bad2: 539 mwl_dma_cleanup(sc); 540 bad1: 541 mwl_hal_detach(mh); 542 bad: 543 MWL_RXFREE_DESTROY(sc); 544 if_free(ifp); 545 sc->sc_invalid = 1; 546 return error; 547 } 548 549 int 550 mwl_detach(struct mwl_softc *sc) 551 { 552 struct ifnet *ifp = sc->sc_ifp; 553 struct ieee80211com *ic = ifp->if_l2com; 554 555 DPRINTF(sc, MWL_DEBUG_ANY, "%s: if_flags %x\n", 556 __func__, ifp->if_flags); 557 558 mwl_stop(ifp, 1); 559 /* 560 * NB: the order of these is important: 561 * o call the 802.11 layer before detaching the hal to 562 * insure callbacks into the driver to delete global 563 * key cache entries can be handled 564 * o reclaim the tx queue data structures after calling 565 * the 802.11 layer as we'll get called back to reclaim 566 * node state and potentially want to use them 567 * o to cleanup the tx queues the hal is called, so detach 568 * it last 569 * Other than that, it's straightforward... 570 */ 571 ieee80211_ifdetach(ic); 572 callout_drain(&sc->sc_watchdog); 573 mwl_dma_cleanup(sc); 574 MWL_RXFREE_DESTROY(sc); 575 mwl_tx_cleanup(sc); 576 mwl_hal_detach(sc->sc_mh); 577 if_free(ifp); 578 579 return 0; 580 } 581 582 /* 583 * MAC address handling for multiple BSS on the same radio. 584 * The first vap uses the MAC address from the EEPROM. For 585 * subsequent vap's we set the U/L bit (bit 1) in the MAC 586 * address and use the next six bits as an index. 587 */ 588 static void 589 assign_address(struct mwl_softc *sc, uint8_t mac[IEEE80211_ADDR_LEN], int clone) 590 { 591 int i; 592 593 if (clone && mwl_hal_ismbsscapable(sc->sc_mh)) { 594 /* NB: we only do this if h/w supports multiple bssid */ 595 for (i = 0; i < 32; i++) 596 if ((sc->sc_bssidmask & (1<<i)) == 0) 597 break; 598 if (i != 0) 599 mac[0] |= (i << 2)|0x2; 600 } else 601 i = 0; 602 sc->sc_bssidmask |= 1<<i; 603 if (i == 0) 604 sc->sc_nbssid0++; 605 } 606 607 static void 608 reclaim_address(struct mwl_softc *sc, uint8_t mac[IEEE80211_ADDR_LEN]) 609 { 610 int i = mac[0] >> 2; 611 if (i != 0 || --sc->sc_nbssid0 == 0) 612 sc->sc_bssidmask &= ~(1<<i); 613 } 614 615 static struct ieee80211vap * 616 mwl_vap_create(struct ieee80211com *ic, const char name[IFNAMSIZ], int unit, 617 enum ieee80211_opmode opmode, int flags, 618 const uint8_t bssid[IEEE80211_ADDR_LEN], 619 const uint8_t mac0[IEEE80211_ADDR_LEN]) 620 { 621 struct ifnet *ifp = ic->ic_ifp; 622 struct mwl_softc *sc = ifp->if_softc; 623 struct mwl_hal *mh = sc->sc_mh; 624 struct ieee80211vap *vap, *apvap; 625 struct mwl_hal_vap *hvap; 626 struct mwl_vap *mvp; 627 uint8_t mac[IEEE80211_ADDR_LEN]; 628 629 IEEE80211_ADDR_COPY(mac, mac0); 630 switch (opmode) { 631 case IEEE80211_M_HOSTAP: 632 case IEEE80211_M_MBSS: 633 if ((flags & IEEE80211_CLONE_MACADDR) == 0) 634 assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID); 635 hvap = mwl_hal_newvap(mh, MWL_HAL_AP, mac); 636 if (hvap == NULL) { 637 if ((flags & IEEE80211_CLONE_MACADDR) == 0) 638 reclaim_address(sc, mac); 639 return NULL; 640 } 641 break; 642 case IEEE80211_M_STA: 643 if ((flags & IEEE80211_CLONE_MACADDR) == 0) 644 assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID); 645 hvap = mwl_hal_newvap(mh, MWL_HAL_STA, mac); 646 if (hvap == NULL) { 647 if ((flags & IEEE80211_CLONE_MACADDR) == 0) 648 reclaim_address(sc, mac); 649 return NULL; 650 } 651 /* no h/w beacon miss support; always use s/w */ 652 flags |= IEEE80211_CLONE_NOBEACONS; 653 break; 654 case IEEE80211_M_WDS: 655 hvap = NULL; /* NB: we use associated AP vap */ 656 if (sc->sc_napvaps == 0) 657 return NULL; /* no existing AP vap */ 658 break; 659 case IEEE80211_M_MONITOR: 660 hvap = NULL; 661 break; 662 case IEEE80211_M_IBSS: 663 case IEEE80211_M_AHDEMO: 664 default: 665 return NULL; 666 } 667 668 mvp = (struct mwl_vap *) malloc(sizeof(struct mwl_vap), 669 M_80211_VAP, M_NOWAIT | M_ZERO); 670 if (mvp == NULL) { 671 if (hvap != NULL) { 672 mwl_hal_delvap(hvap); 673 if ((flags & IEEE80211_CLONE_MACADDR) == 0) 674 reclaim_address(sc, mac); 675 } 676 /* XXX msg */ 677 return NULL; 678 } 679 mvp->mv_hvap = hvap; 680 if (opmode == IEEE80211_M_WDS) { 681 /* 682 * WDS vaps must have an associated AP vap; find one. 683 * XXX not right. 684 */ 685 TAILQ_FOREACH(apvap, &ic->ic_vaps, iv_next) 686 if (apvap->iv_opmode == IEEE80211_M_HOSTAP) { 687 mvp->mv_ap_hvap = MWL_VAP(apvap)->mv_hvap; 688 break; 689 } 690 KASSERT(mvp->mv_ap_hvap != NULL, ("no ap vap")); 691 } 692 vap = &mvp->mv_vap; 693 ieee80211_vap_setup(ic, vap, name, unit, opmode, flags, bssid, mac); 694 if (hvap != NULL) 695 IEEE80211_ADDR_COPY(vap->iv_myaddr, mac); 696 /* override with driver methods */ 697 mvp->mv_newstate = vap->iv_newstate; 698 vap->iv_newstate = mwl_newstate; 699 vap->iv_max_keyix = 0; /* XXX */ 700 vap->iv_key_alloc = mwl_key_alloc; 701 vap->iv_key_delete = mwl_key_delete; 702 vap->iv_key_set = mwl_key_set; 703 #ifdef MWL_HOST_PS_SUPPORT 704 if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS) { 705 vap->iv_update_ps = mwl_update_ps; 706 mvp->mv_set_tim = vap->iv_set_tim; 707 vap->iv_set_tim = mwl_set_tim; 708 } 709 #endif 710 vap->iv_reset = mwl_reset; 711 vap->iv_update_beacon = mwl_beacon_update; 712 713 /* override max aid so sta's cannot assoc when we're out of sta id's */ 714 vap->iv_max_aid = MWL_MAXSTAID; 715 /* override default A-MPDU rx parameters */ 716 vap->iv_ampdu_rxmax = IEEE80211_HTCAP_MAXRXAMPDU_64K; 717 vap->iv_ampdu_density = IEEE80211_HTCAP_MPDUDENSITY_4; 718 719 /* complete setup */ 720 ieee80211_vap_attach(vap, mwl_media_change, ieee80211_media_status); 721 722 switch (vap->iv_opmode) { 723 case IEEE80211_M_HOSTAP: 724 case IEEE80211_M_MBSS: 725 case IEEE80211_M_STA: 726 /* 727 * Setup sta db entry for local address. 728 */ 729 mwl_localstadb(vap); 730 if (vap->iv_opmode == IEEE80211_M_HOSTAP || 731 vap->iv_opmode == IEEE80211_M_MBSS) 732 sc->sc_napvaps++; 733 else 734 sc->sc_nstavaps++; 735 break; 736 case IEEE80211_M_WDS: 737 sc->sc_nwdsvaps++; 738 break; 739 default: 740 break; 741 } 742 /* 743 * Setup overall operating mode. 744 */ 745 if (sc->sc_napvaps) 746 ic->ic_opmode = IEEE80211_M_HOSTAP; 747 else if (sc->sc_nstavaps) 748 ic->ic_opmode = IEEE80211_M_STA; 749 else 750 ic->ic_opmode = opmode; 751 752 return vap; 753 } 754 755 static void 756 mwl_vap_delete(struct ieee80211vap *vap) 757 { 758 struct mwl_vap *mvp = MWL_VAP(vap); 759 struct ifnet *parent = vap->iv_ic->ic_ifp; 760 struct mwl_softc *sc = parent->if_softc; 761 struct mwl_hal *mh = sc->sc_mh; 762 struct mwl_hal_vap *hvap = mvp->mv_hvap; 763 enum ieee80211_opmode opmode = vap->iv_opmode; 764 765 /* XXX disallow ap vap delete if WDS still present */ 766 if (parent->if_drv_flags & IFF_DRV_RUNNING) { 767 /* quiesce h/w while we remove the vap */ 768 mwl_hal_intrset(mh, 0); /* disable interrupts */ 769 } 770 ieee80211_vap_detach(vap); 771 switch (opmode) { 772 case IEEE80211_M_HOSTAP: 773 case IEEE80211_M_MBSS: 774 case IEEE80211_M_STA: 775 KASSERT(hvap != NULL, ("no hal vap handle")); 776 (void) mwl_hal_delstation(hvap, vap->iv_myaddr); 777 mwl_hal_delvap(hvap); 778 if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS) 779 sc->sc_napvaps--; 780 else 781 sc->sc_nstavaps--; 782 /* XXX don't do it for IEEE80211_CLONE_MACADDR */ 783 reclaim_address(sc, vap->iv_myaddr); 784 break; 785 case IEEE80211_M_WDS: 786 sc->sc_nwdsvaps--; 787 break; 788 default: 789 break; 790 } 791 mwl_cleartxq(sc, vap); 792 free(mvp, M_80211_VAP); 793 if (parent->if_drv_flags & IFF_DRV_RUNNING) 794 mwl_hal_intrset(mh, sc->sc_imask); 795 } 796 797 void 798 mwl_suspend(struct mwl_softc *sc) 799 { 800 struct ifnet *ifp = sc->sc_ifp; 801 802 DPRINTF(sc, MWL_DEBUG_ANY, "%s: if_flags %x\n", 803 __func__, ifp->if_flags); 804 805 mwl_stop(ifp, 1); 806 } 807 808 void 809 mwl_resume(struct mwl_softc *sc) 810 { 811 struct ifnet *ifp = sc->sc_ifp; 812 813 DPRINTF(sc, MWL_DEBUG_ANY, "%s: if_flags %x\n", 814 __func__, ifp->if_flags); 815 816 if (ifp->if_flags & IFF_UP) 817 mwl_init(sc); 818 } 819 820 void 821 mwl_shutdown(void *arg) 822 { 823 struct mwl_softc *sc = arg; 824 825 mwl_stop(sc->sc_ifp, 1); 826 } 827 828 /* 829 * Interrupt handler. Most of the actual processing is deferred. 830 */ 831 void 832 mwl_intr(void *arg) 833 { 834 struct mwl_softc *sc = arg; 835 struct mwl_hal *mh = sc->sc_mh; 836 uint32_t status; 837 838 if (sc->sc_invalid) { 839 /* 840 * The hardware is not ready/present, don't touch anything. 841 * Note this can happen early on if the IRQ is shared. 842 */ 843 DPRINTF(sc, MWL_DEBUG_ANY, "%s: invalid; ignored\n", __func__); 844 return; 845 } 846 /* 847 * Figure out the reason(s) for the interrupt. 848 */ 849 mwl_hal_getisr(mh, &status); /* NB: clears ISR too */ 850 if (status == 0) /* must be a shared irq */ 851 return; 852 853 DPRINTF(sc, MWL_DEBUG_INTR, "%s: status 0x%x imask 0x%x\n", 854 __func__, status, sc->sc_imask); 855 if (status & MACREG_A2HRIC_BIT_RX_RDY) 856 taskqueue_enqueue(sc->sc_tq, &sc->sc_rxtask); 857 if (status & MACREG_A2HRIC_BIT_TX_DONE) 858 taskqueue_enqueue(sc->sc_tq, &sc->sc_txtask); 859 if (status & MACREG_A2HRIC_BIT_BA_WATCHDOG) 860 taskqueue_enqueue(sc->sc_tq, &sc->sc_bawatchdogtask); 861 if (status & MACREG_A2HRIC_BIT_OPC_DONE) 862 mwl_hal_cmddone(mh); 863 if (status & MACREG_A2HRIC_BIT_MAC_EVENT) { 864 ; 865 } 866 if (status & MACREG_A2HRIC_BIT_ICV_ERROR) { 867 /* TKIP ICV error */ 868 sc->sc_stats.mst_rx_badtkipicv++; 869 } 870 if (status & MACREG_A2HRIC_BIT_QUEUE_EMPTY) { 871 /* 11n aggregation queue is empty, re-fill */ 872 ; 873 } 874 if (status & MACREG_A2HRIC_BIT_QUEUE_FULL) { 875 ; 876 } 877 if (status & MACREG_A2HRIC_BIT_RADAR_DETECT) { 878 /* radar detected, process event */ 879 taskqueue_enqueue(sc->sc_tq, &sc->sc_radartask); 880 } 881 if (status & MACREG_A2HRIC_BIT_CHAN_SWITCH) { 882 /* DFS channel switch */ 883 taskqueue_enqueue(sc->sc_tq, &sc->sc_chanswitchtask); 884 } 885 } 886 887 static void 888 mwl_radar_proc(void *arg, int pending) 889 { 890 struct mwl_softc *sc = arg; 891 struct ifnet *ifp = sc->sc_ifp; 892 struct ieee80211com *ic = ifp->if_l2com; 893 894 DPRINTF(sc, MWL_DEBUG_ANY, "%s: radar detected, pending %u\n", 895 __func__, pending); 896 897 sc->sc_stats.mst_radardetect++; 898 /* XXX stop h/w BA streams? */ 899 900 IEEE80211_LOCK(ic); 901 ieee80211_dfs_notify_radar(ic, ic->ic_curchan); 902 IEEE80211_UNLOCK(ic); 903 } 904 905 static void 906 mwl_chanswitch_proc(void *arg, int pending) 907 { 908 struct mwl_softc *sc = arg; 909 struct ifnet *ifp = sc->sc_ifp; 910 struct ieee80211com *ic = ifp->if_l2com; 911 912 DPRINTF(sc, MWL_DEBUG_ANY, "%s: channel switch notice, pending %u\n", 913 __func__, pending); 914 915 IEEE80211_LOCK(ic); 916 sc->sc_csapending = 0; 917 ieee80211_csa_completeswitch(ic); 918 IEEE80211_UNLOCK(ic); 919 } 920 921 static void 922 mwl_bawatchdog(const MWL_HAL_BASTREAM *sp) 923 { 924 struct ieee80211_node *ni = sp->data[0]; 925 926 /* send DELBA and drop the stream */ 927 ieee80211_ampdu_stop(ni, sp->data[1], IEEE80211_REASON_UNSPECIFIED); 928 } 929 930 static void 931 mwl_bawatchdog_proc(void *arg, int pending) 932 { 933 struct mwl_softc *sc = arg; 934 struct mwl_hal *mh = sc->sc_mh; 935 const MWL_HAL_BASTREAM *sp; 936 uint8_t bitmap, n; 937 938 sc->sc_stats.mst_bawatchdog++; 939 940 if (mwl_hal_getwatchdogbitmap(mh, &bitmap) != 0) { 941 DPRINTF(sc, MWL_DEBUG_AMPDU, 942 "%s: could not get bitmap\n", __func__); 943 sc->sc_stats.mst_bawatchdog_failed++; 944 return; 945 } 946 DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: bitmap 0x%x\n", __func__, bitmap); 947 if (bitmap == 0xff) { 948 n = 0; 949 /* disable all ba streams */ 950 for (bitmap = 0; bitmap < 8; bitmap++) { 951 sp = mwl_hal_bastream_lookup(mh, bitmap); 952 if (sp != NULL) { 953 mwl_bawatchdog(sp); 954 n++; 955 } 956 } 957 if (n == 0) { 958 DPRINTF(sc, MWL_DEBUG_AMPDU, 959 "%s: no BA streams found\n", __func__); 960 sc->sc_stats.mst_bawatchdog_empty++; 961 } 962 } else if (bitmap != 0xaa) { 963 /* disable a single ba stream */ 964 sp = mwl_hal_bastream_lookup(mh, bitmap); 965 if (sp != NULL) { 966 mwl_bawatchdog(sp); 967 } else { 968 DPRINTF(sc, MWL_DEBUG_AMPDU, 969 "%s: no BA stream %d\n", __func__, bitmap); 970 sc->sc_stats.mst_bawatchdog_notfound++; 971 } 972 } 973 } 974 975 /* 976 * Convert net80211 channel to a HAL channel. 977 */ 978 static void 979 mwl_mapchan(MWL_HAL_CHANNEL *hc, const struct ieee80211_channel *chan) 980 { 981 hc->channel = chan->ic_ieee; 982 983 *(uint32_t *)&hc->channelFlags = 0; 984 if (IEEE80211_IS_CHAN_2GHZ(chan)) 985 hc->channelFlags.FreqBand = MWL_FREQ_BAND_2DOT4GHZ; 986 else if (IEEE80211_IS_CHAN_5GHZ(chan)) 987 hc->channelFlags.FreqBand = MWL_FREQ_BAND_5GHZ; 988 if (IEEE80211_IS_CHAN_HT40(chan)) { 989 hc->channelFlags.ChnlWidth = MWL_CH_40_MHz_WIDTH; 990 if (IEEE80211_IS_CHAN_HT40U(chan)) 991 hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_ABOVE_CTRL_CH; 992 else 993 hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_BELOW_CTRL_CH; 994 } else 995 hc->channelFlags.ChnlWidth = MWL_CH_20_MHz_WIDTH; 996 /* XXX 10MHz channels */ 997 } 998 999 /* 1000 * Inform firmware of our tx/rx dma setup. The BAR 0 1001 * writes below are for compatibility with older firmware. 1002 * For current firmware we send this information with a 1003 * cmd block via mwl_hal_sethwdma. 1004 */ 1005 static int 1006 mwl_setupdma(struct mwl_softc *sc) 1007 { 1008 int error, i; 1009 1010 sc->sc_hwdma.rxDescRead = sc->sc_rxdma.dd_desc_paddr; 1011 WR4(sc, sc->sc_hwspecs.rxDescRead, sc->sc_hwdma.rxDescRead); 1012 WR4(sc, sc->sc_hwspecs.rxDescWrite, sc->sc_hwdma.rxDescRead); 1013 1014 for (i = 0; i < MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES; i++) { 1015 struct mwl_txq *txq = &sc->sc_txq[i]; 1016 sc->sc_hwdma.wcbBase[i] = txq->dma.dd_desc_paddr; 1017 WR4(sc, sc->sc_hwspecs.wcbBase[i], sc->sc_hwdma.wcbBase[i]); 1018 } 1019 sc->sc_hwdma.maxNumTxWcb = mwl_txbuf; 1020 sc->sc_hwdma.maxNumWCB = MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES; 1021 1022 error = mwl_hal_sethwdma(sc->sc_mh, &sc->sc_hwdma); 1023 if (error != 0) { 1024 device_printf(sc->sc_dev, 1025 "unable to setup tx/rx dma; hal status %u\n", error); 1026 /* XXX */ 1027 } 1028 return error; 1029 } 1030 1031 /* 1032 * Inform firmware of tx rate parameters. 1033 * Called after a channel change. 1034 */ 1035 static int 1036 mwl_setcurchanrates(struct mwl_softc *sc) 1037 { 1038 struct ifnet *ifp = sc->sc_ifp; 1039 struct ieee80211com *ic = ifp->if_l2com; 1040 const struct ieee80211_rateset *rs; 1041 MWL_HAL_TXRATE rates; 1042 1043 memset(&rates, 0, sizeof(rates)); 1044 rs = ieee80211_get_suprates(ic, ic->ic_curchan); 1045 /* rate used to send management frames */ 1046 rates.MgtRate = rs->rs_rates[0] & IEEE80211_RATE_VAL; 1047 /* rate used to send multicast frames */ 1048 rates.McastRate = rates.MgtRate; 1049 1050 return mwl_hal_settxrate_auto(sc->sc_mh, &rates); 1051 } 1052 1053 /* 1054 * Inform firmware of tx rate parameters. Called whenever 1055 * user-settable params change and after a channel change. 1056 */ 1057 static int 1058 mwl_setrates(struct ieee80211vap *vap) 1059 { 1060 struct mwl_vap *mvp = MWL_VAP(vap); 1061 struct ieee80211_node *ni = vap->iv_bss; 1062 const struct ieee80211_txparam *tp = ni->ni_txparms; 1063 MWL_HAL_TXRATE rates; 1064 1065 KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state)); 1066 1067 /* 1068 * Update the h/w rate map. 1069 * NB: 0x80 for MCS is passed through unchanged 1070 */ 1071 memset(&rates, 0, sizeof(rates)); 1072 /* rate used to send management frames */ 1073 rates.MgtRate = tp->mgmtrate; 1074 /* rate used to send multicast frames */ 1075 rates.McastRate = tp->mcastrate; 1076 1077 /* while here calculate EAPOL fixed rate cookie */ 1078 mvp->mv_eapolformat = htole16(mwl_calcformat(rates.MgtRate, ni)); 1079 1080 return mwl_hal_settxrate(mvp->mv_hvap, 1081 tp->ucastrate != IEEE80211_FIXED_RATE_NONE ? 1082 RATE_FIXED : RATE_AUTO, &rates); 1083 } 1084 1085 /* 1086 * Setup a fixed xmit rate cookie for EAPOL frames. 1087 */ 1088 static void 1089 mwl_seteapolformat(struct ieee80211vap *vap) 1090 { 1091 struct mwl_vap *mvp = MWL_VAP(vap); 1092 struct ieee80211_node *ni = vap->iv_bss; 1093 enum ieee80211_phymode mode; 1094 uint8_t rate; 1095 1096 KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state)); 1097 1098 mode = ieee80211_chan2mode(ni->ni_chan); 1099 /* 1100 * Use legacy rates when operating a mixed HT+non-HT bss. 1101 * NB: this may violate POLA for sta and wds vap's. 1102 */ 1103 if (mode == IEEE80211_MODE_11NA && 1104 (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0) 1105 rate = vap->iv_txparms[IEEE80211_MODE_11A].mgmtrate; 1106 else if (mode == IEEE80211_MODE_11NG && 1107 (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0) 1108 rate = vap->iv_txparms[IEEE80211_MODE_11G].mgmtrate; 1109 else 1110 rate = vap->iv_txparms[mode].mgmtrate; 1111 1112 mvp->mv_eapolformat = htole16(mwl_calcformat(rate, ni)); 1113 } 1114 1115 /* 1116 * Map SKU+country code to region code for radar bin'ing. 1117 */ 1118 static int 1119 mwl_map2regioncode(const struct ieee80211_regdomain *rd) 1120 { 1121 switch (rd->regdomain) { 1122 case SKU_FCC: 1123 case SKU_FCC3: 1124 return DOMAIN_CODE_FCC; 1125 case SKU_CA: 1126 return DOMAIN_CODE_IC; 1127 case SKU_ETSI: 1128 case SKU_ETSI2: 1129 case SKU_ETSI3: 1130 if (rd->country == CTRY_SPAIN) 1131 return DOMAIN_CODE_SPAIN; 1132 if (rd->country == CTRY_FRANCE || rd->country == CTRY_FRANCE2) 1133 return DOMAIN_CODE_FRANCE; 1134 /* XXX force 1.3.1 radar type */ 1135 return DOMAIN_CODE_ETSI_131; 1136 case SKU_JAPAN: 1137 return DOMAIN_CODE_MKK; 1138 case SKU_ROW: 1139 return DOMAIN_CODE_DGT; /* Taiwan */ 1140 case SKU_APAC: 1141 case SKU_APAC2: 1142 case SKU_APAC3: 1143 return DOMAIN_CODE_AUS; /* Australia */ 1144 } 1145 /* XXX KOREA? */ 1146 return DOMAIN_CODE_FCC; /* XXX? */ 1147 } 1148 1149 static int 1150 mwl_hal_reset(struct mwl_softc *sc) 1151 { 1152 struct ifnet *ifp = sc->sc_ifp; 1153 struct ieee80211com *ic = ifp->if_l2com; 1154 struct mwl_hal *mh = sc->sc_mh; 1155 1156 mwl_hal_setantenna(mh, WL_ANTENNATYPE_RX, sc->sc_rxantenna); 1157 mwl_hal_setantenna(mh, WL_ANTENNATYPE_TX, sc->sc_txantenna); 1158 mwl_hal_setradio(mh, 1, WL_AUTO_PREAMBLE); 1159 mwl_hal_setwmm(sc->sc_mh, (ic->ic_flags & IEEE80211_F_WME) != 0); 1160 mwl_chan_set(sc, ic->ic_curchan); 1161 /* NB: RF/RA performance tuned for indoor mode */ 1162 mwl_hal_setrateadaptmode(mh, 0); 1163 mwl_hal_setoptimizationlevel(mh, 1164 (ic->ic_flags & IEEE80211_F_BURST) != 0); 1165 1166 mwl_hal_setregioncode(mh, mwl_map2regioncode(&ic->ic_regdomain)); 1167 1168 mwl_hal_setaggampduratemode(mh, 1, 80); /* XXX */ 1169 mwl_hal_setcfend(mh, 0); /* XXX */ 1170 1171 return 1; 1172 } 1173 1174 static int 1175 mwl_init_locked(struct mwl_softc *sc) 1176 { 1177 struct ifnet *ifp = sc->sc_ifp; 1178 struct mwl_hal *mh = sc->sc_mh; 1179 int error = 0; 1180 1181 DPRINTF(sc, MWL_DEBUG_ANY, "%s: if_flags 0x%x\n", 1182 __func__, ifp->if_flags); 1183 1184 MWL_LOCK_ASSERT(sc); 1185 1186 /* 1187 * Stop anything previously setup. This is safe 1188 * whether this is the first time through or not. 1189 */ 1190 mwl_stop_locked(ifp, 0); 1191 1192 /* 1193 * Push vap-independent state to the firmware. 1194 */ 1195 if (!mwl_hal_reset(sc)) { 1196 if_printf(ifp, "unable to reset hardware\n"); 1197 return EIO; 1198 } 1199 1200 /* 1201 * Setup recv (once); transmit is already good to go. 1202 */ 1203 error = mwl_startrecv(sc); 1204 if (error != 0) { 1205 if_printf(ifp, "unable to start recv logic\n"); 1206 return error; 1207 } 1208 1209 /* 1210 * Enable interrupts. 1211 */ 1212 sc->sc_imask = MACREG_A2HRIC_BIT_RX_RDY 1213 | MACREG_A2HRIC_BIT_TX_DONE 1214 | MACREG_A2HRIC_BIT_OPC_DONE 1215 #if 0 1216 | MACREG_A2HRIC_BIT_MAC_EVENT 1217 #endif 1218 | MACREG_A2HRIC_BIT_ICV_ERROR 1219 | MACREG_A2HRIC_BIT_RADAR_DETECT 1220 | MACREG_A2HRIC_BIT_CHAN_SWITCH 1221 #if 0 1222 | MACREG_A2HRIC_BIT_QUEUE_EMPTY 1223 #endif 1224 | MACREG_A2HRIC_BIT_BA_WATCHDOG 1225 | MACREQ_A2HRIC_BIT_TX_ACK 1226 ; 1227 1228 ifp->if_drv_flags |= IFF_DRV_RUNNING; 1229 mwl_hal_intrset(mh, sc->sc_imask); 1230 callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc); 1231 1232 return 0; 1233 } 1234 1235 static void 1236 mwl_init(void *arg) 1237 { 1238 struct mwl_softc *sc = arg; 1239 struct ifnet *ifp = sc->sc_ifp; 1240 struct ieee80211com *ic = ifp->if_l2com; 1241 int error = 0; 1242 1243 DPRINTF(sc, MWL_DEBUG_ANY, "%s: if_flags 0x%x\n", 1244 __func__, ifp->if_flags); 1245 1246 MWL_LOCK(sc); 1247 error = mwl_init_locked(sc); 1248 MWL_UNLOCK(sc); 1249 1250 if (error == 0) 1251 ieee80211_start_all(ic); /* start all vap's */ 1252 } 1253 1254 static void 1255 mwl_stop_locked(struct ifnet *ifp, int disable) 1256 { 1257 struct mwl_softc *sc = ifp->if_softc; 1258 1259 DPRINTF(sc, MWL_DEBUG_ANY, "%s: invalid %u if_flags 0x%x\n", 1260 __func__, sc->sc_invalid, ifp->if_flags); 1261 1262 MWL_LOCK_ASSERT(sc); 1263 if (ifp->if_drv_flags & IFF_DRV_RUNNING) { 1264 /* 1265 * Shutdown the hardware and driver. 1266 */ 1267 ifp->if_drv_flags &= ~IFF_DRV_RUNNING; 1268 callout_stop(&sc->sc_watchdog); 1269 sc->sc_tx_timer = 0; 1270 mwl_draintxq(sc); 1271 } 1272 } 1273 1274 static void 1275 mwl_stop(struct ifnet *ifp, int disable) 1276 { 1277 struct mwl_softc *sc = ifp->if_softc; 1278 1279 MWL_LOCK(sc); 1280 mwl_stop_locked(ifp, disable); 1281 MWL_UNLOCK(sc); 1282 } 1283 1284 static int 1285 mwl_reset_vap(struct ieee80211vap *vap, int state) 1286 { 1287 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap; 1288 struct ieee80211com *ic = vap->iv_ic; 1289 1290 if (state == IEEE80211_S_RUN) 1291 mwl_setrates(vap); 1292 /* XXX off by 1? */ 1293 mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold); 1294 /* XXX auto? 20/40 split? */ 1295 mwl_hal_sethtgi(hvap, (vap->iv_flags_ht & 1296 (IEEE80211_FHT_SHORTGI20|IEEE80211_FHT_SHORTGI40)) ? 1 : 0); 1297 mwl_hal_setnprot(hvap, ic->ic_htprotmode == IEEE80211_PROT_NONE ? 1298 HTPROTECT_NONE : HTPROTECT_AUTO); 1299 /* XXX txpower cap */ 1300 1301 /* re-setup beacons */ 1302 if (state == IEEE80211_S_RUN && 1303 (vap->iv_opmode == IEEE80211_M_HOSTAP || 1304 vap->iv_opmode == IEEE80211_M_MBSS || 1305 vap->iv_opmode == IEEE80211_M_IBSS)) { 1306 mwl_setapmode(vap, vap->iv_bss->ni_chan); 1307 mwl_hal_setnprotmode(hvap, 1308 MS(ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE)); 1309 return mwl_beacon_setup(vap); 1310 } 1311 return 0; 1312 } 1313 1314 /* 1315 * Reset the hardware w/o losing operational state. 1316 * Used to to reset or reload hardware state for a vap. 1317 */ 1318 static int 1319 mwl_reset(struct ieee80211vap *vap, u_long cmd) 1320 { 1321 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap; 1322 int error = 0; 1323 1324 if (hvap != NULL) { /* WDS, MONITOR, etc. */ 1325 struct ieee80211com *ic = vap->iv_ic; 1326 struct ifnet *ifp = ic->ic_ifp; 1327 struct mwl_softc *sc = ifp->if_softc; 1328 struct mwl_hal *mh = sc->sc_mh; 1329 1330 /* XXX handle DWDS sta vap change */ 1331 /* XXX do we need to disable interrupts? */ 1332 mwl_hal_intrset(mh, 0); /* disable interrupts */ 1333 error = mwl_reset_vap(vap, vap->iv_state); 1334 mwl_hal_intrset(mh, sc->sc_imask); 1335 } 1336 return error; 1337 } 1338 1339 /* 1340 * Allocate a tx buffer for sending a frame. The 1341 * packet is assumed to have the WME AC stored so 1342 * we can use it to select the appropriate h/w queue. 1343 */ 1344 static struct mwl_txbuf * 1345 mwl_gettxbuf(struct mwl_softc *sc, struct mwl_txq *txq) 1346 { 1347 struct mwl_txbuf *bf; 1348 1349 /* 1350 * Grab a TX buffer and associated resources. 1351 */ 1352 MWL_TXQ_LOCK(txq); 1353 bf = STAILQ_FIRST(&txq->free); 1354 if (bf != NULL) { 1355 STAILQ_REMOVE_HEAD(&txq->free, bf_list); 1356 txq->nfree--; 1357 } 1358 MWL_TXQ_UNLOCK(txq); 1359 if (bf == NULL) 1360 DPRINTF(sc, MWL_DEBUG_XMIT, 1361 "%s: out of xmit buffers on q %d\n", __func__, txq->qnum); 1362 return bf; 1363 } 1364 1365 /* 1366 * Return a tx buffer to the queue it came from. Note there 1367 * are two cases because we must preserve the order of buffers 1368 * as it reflects the fixed order of descriptors in memory 1369 * (the firmware pre-fetches descriptors so we cannot reorder). 1370 */ 1371 static void 1372 mwl_puttxbuf_head(struct mwl_txq *txq, struct mwl_txbuf *bf) 1373 { 1374 bf->bf_m = NULL; 1375 bf->bf_node = NULL; 1376 MWL_TXQ_LOCK(txq); 1377 STAILQ_INSERT_HEAD(&txq->free, bf, bf_list); 1378 txq->nfree++; 1379 MWL_TXQ_UNLOCK(txq); 1380 } 1381 1382 static void 1383 mwl_puttxbuf_tail(struct mwl_txq *txq, struct mwl_txbuf *bf) 1384 { 1385 bf->bf_m = NULL; 1386 bf->bf_node = NULL; 1387 MWL_TXQ_LOCK(txq); 1388 STAILQ_INSERT_TAIL(&txq->free, bf, bf_list); 1389 txq->nfree++; 1390 MWL_TXQ_UNLOCK(txq); 1391 } 1392 1393 static void 1394 mwl_start(struct ifnet *ifp) 1395 { 1396 struct mwl_softc *sc = ifp->if_softc; 1397 struct ieee80211_node *ni; 1398 struct mwl_txbuf *bf; 1399 struct mbuf *m; 1400 struct mwl_txq *txq = NULL; /* XXX silence gcc */ 1401 int nqueued; 1402 1403 if ((ifp->if_drv_flags & IFF_DRV_RUNNING) == 0 || sc->sc_invalid) 1404 return; 1405 nqueued = 0; 1406 for (;;) { 1407 bf = NULL; 1408 IFQ_DEQUEUE(&ifp->if_snd, m); 1409 if (m == NULL) 1410 break; 1411 /* 1412 * Grab the node for the destination. 1413 */ 1414 ni = (struct ieee80211_node *) m->m_pkthdr.rcvif; 1415 KASSERT(ni != NULL, ("no node")); 1416 m->m_pkthdr.rcvif = NULL; /* committed, clear ref */ 1417 /* 1418 * Grab a TX buffer and associated resources. 1419 * We honor the classification by the 802.11 layer. 1420 */ 1421 txq = sc->sc_ac2q[M_WME_GETAC(m)]; 1422 bf = mwl_gettxbuf(sc, txq); 1423 if (bf == NULL) { 1424 m_freem(m); 1425 ieee80211_free_node(ni); 1426 #ifdef MWL_TX_NODROP 1427 sc->sc_stats.mst_tx_qstop++; 1428 /* XXX blocks other traffic */ 1429 ifp->if_drv_flags |= IFF_DRV_OACTIVE; 1430 break; 1431 #else 1432 DPRINTF(sc, MWL_DEBUG_XMIT, 1433 "%s: tail drop on q %d\n", __func__, txq->qnum); 1434 sc->sc_stats.mst_tx_qdrop++; 1435 continue; 1436 #endif /* MWL_TX_NODROP */ 1437 } 1438 1439 /* 1440 * Pass the frame to the h/w for transmission. 1441 */ 1442 if (mwl_tx_start(sc, ni, bf, m)) { 1443 ifp->if_oerrors++; 1444 mwl_puttxbuf_head(txq, bf); 1445 ieee80211_free_node(ni); 1446 continue; 1447 } 1448 nqueued++; 1449 if (nqueued >= mwl_txcoalesce) { 1450 /* 1451 * Poke the firmware to process queued frames; 1452 * see below about (lack of) locking. 1453 */ 1454 nqueued = 0; 1455 mwl_hal_txstart(sc->sc_mh, 0/*XXX*/); 1456 } 1457 } 1458 if (nqueued) { 1459 /* 1460 * NB: We don't need to lock against tx done because 1461 * this just prods the firmware to check the transmit 1462 * descriptors. The firmware will also start fetching 1463 * descriptors by itself if it notices new ones are 1464 * present when it goes to deliver a tx done interrupt 1465 * to the host. So if we race with tx done processing 1466 * it's ok. Delivering the kick here rather than in 1467 * mwl_tx_start is an optimization to avoid poking the 1468 * firmware for each packet. 1469 * 1470 * NB: the queue id isn't used so 0 is ok. 1471 */ 1472 mwl_hal_txstart(sc->sc_mh, 0/*XXX*/); 1473 } 1474 } 1475 1476 static int 1477 mwl_raw_xmit(struct ieee80211_node *ni, struct mbuf *m, 1478 const struct ieee80211_bpf_params *params) 1479 { 1480 struct ieee80211com *ic = ni->ni_ic; 1481 struct ifnet *ifp = ic->ic_ifp; 1482 struct mwl_softc *sc = ifp->if_softc; 1483 struct mwl_txbuf *bf; 1484 struct mwl_txq *txq; 1485 1486 if ((ifp->if_drv_flags & IFF_DRV_RUNNING) == 0 || sc->sc_invalid) { 1487 ieee80211_free_node(ni); 1488 m_freem(m); 1489 return ENETDOWN; 1490 } 1491 /* 1492 * Grab a TX buffer and associated resources. 1493 * Note that we depend on the classification 1494 * by the 802.11 layer to get to the right h/w 1495 * queue. Management frames must ALWAYS go on 1496 * queue 1 but we cannot just force that here 1497 * because we may receive non-mgt frames. 1498 */ 1499 txq = sc->sc_ac2q[M_WME_GETAC(m)]; 1500 bf = mwl_gettxbuf(sc, txq); 1501 if (bf == NULL) { 1502 sc->sc_stats.mst_tx_qstop++; 1503 /* XXX blocks other traffic */ 1504 ifp->if_drv_flags |= IFF_DRV_OACTIVE; 1505 ieee80211_free_node(ni); 1506 m_freem(m); 1507 return ENOBUFS; 1508 } 1509 /* 1510 * Pass the frame to the h/w for transmission. 1511 */ 1512 if (mwl_tx_start(sc, ni, bf, m)) { 1513 ifp->if_oerrors++; 1514 mwl_puttxbuf_head(txq, bf); 1515 1516 ieee80211_free_node(ni); 1517 return EIO; /* XXX */ 1518 } 1519 /* 1520 * NB: We don't need to lock against tx done because 1521 * this just prods the firmware to check the transmit 1522 * descriptors. The firmware will also start fetching 1523 * descriptors by itself if it notices new ones are 1524 * present when it goes to deliver a tx done interrupt 1525 * to the host. So if we race with tx done processing 1526 * it's ok. Delivering the kick here rather than in 1527 * mwl_tx_start is an optimization to avoid poking the 1528 * firmware for each packet. 1529 * 1530 * NB: the queue id isn't used so 0 is ok. 1531 */ 1532 mwl_hal_txstart(sc->sc_mh, 0/*XXX*/); 1533 return 0; 1534 } 1535 1536 static int 1537 mwl_media_change(struct ifnet *ifp) 1538 { 1539 struct ieee80211vap *vap = ifp->if_softc; 1540 int error; 1541 1542 error = ieee80211_media_change(ifp); 1543 /* NB: only the fixed rate can change and that doesn't need a reset */ 1544 if (error == ENETRESET) { 1545 mwl_setrates(vap); 1546 error = 0; 1547 } 1548 return error; 1549 } 1550 1551 #ifdef MWL_DEBUG 1552 static void 1553 mwl_keyprint(struct mwl_softc *sc, const char *tag, 1554 const MWL_HAL_KEYVAL *hk, const uint8_t mac[IEEE80211_ADDR_LEN]) 1555 { 1556 static const char *ciphers[] = { 1557 "WEP", 1558 "TKIP", 1559 "AES-CCM", 1560 }; 1561 int i, n; 1562 1563 printf("%s: [%u] %-7s", tag, hk->keyIndex, ciphers[hk->keyTypeId]); 1564 for (i = 0, n = hk->keyLen; i < n; i++) 1565 printf(" %02x", hk->key.aes[i]); 1566 printf(" mac %s", ether_sprintf(mac)); 1567 if (hk->keyTypeId == KEY_TYPE_ID_TKIP) { 1568 printf(" %s", "rxmic"); 1569 for (i = 0; i < sizeof(hk->key.tkip.rxMic); i++) 1570 printf(" %02x", hk->key.tkip.rxMic[i]); 1571 printf(" txmic"); 1572 for (i = 0; i < sizeof(hk->key.tkip.txMic); i++) 1573 printf(" %02x", hk->key.tkip.txMic[i]); 1574 } 1575 printf(" flags 0x%x\n", hk->keyFlags); 1576 } 1577 #endif 1578 1579 /* 1580 * Allocate a key cache slot for a unicast key. The 1581 * firmware handles key allocation and every station is 1582 * guaranteed key space so we are always successful. 1583 */ 1584 static int 1585 mwl_key_alloc(struct ieee80211vap *vap, struct ieee80211_key *k, 1586 ieee80211_keyix *keyix, ieee80211_keyix *rxkeyix) 1587 { 1588 struct mwl_softc *sc = vap->iv_ic->ic_ifp->if_softc; 1589 1590 if (k->wk_keyix != IEEE80211_KEYIX_NONE || 1591 (k->wk_flags & IEEE80211_KEY_GROUP)) { 1592 if (!(&vap->iv_nw_keys[0] <= k && 1593 k < &vap->iv_nw_keys[IEEE80211_WEP_NKID])) { 1594 /* should not happen */ 1595 DPRINTF(sc, MWL_DEBUG_KEYCACHE, 1596 "%s: bogus group key\n", __func__); 1597 return 0; 1598 } 1599 /* give the caller what they requested */ 1600 *keyix = *rxkeyix = k - vap->iv_nw_keys; 1601 } else { 1602 /* 1603 * Firmware handles key allocation. 1604 */ 1605 *keyix = *rxkeyix = 0; 1606 } 1607 return 1; 1608 } 1609 1610 /* 1611 * Delete a key entry allocated by mwl_key_alloc. 1612 */ 1613 static int 1614 mwl_key_delete(struct ieee80211vap *vap, const struct ieee80211_key *k) 1615 { 1616 struct mwl_softc *sc = vap->iv_ic->ic_ifp->if_softc; 1617 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap; 1618 MWL_HAL_KEYVAL hk; 1619 const uint8_t bcastaddr[IEEE80211_ADDR_LEN] = 1620 { 0xff, 0xff, 0xff, 0xff, 0xff, 0xff }; 1621 1622 if (hvap == NULL) { 1623 if (vap->iv_opmode != IEEE80211_M_WDS) { 1624 /* XXX monitor mode? */ 1625 DPRINTF(sc, MWL_DEBUG_KEYCACHE, 1626 "%s: no hvap for opmode %d\n", __func__, 1627 vap->iv_opmode); 1628 return 0; 1629 } 1630 hvap = MWL_VAP(vap)->mv_ap_hvap; 1631 } 1632 1633 DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: delete key %u\n", 1634 __func__, k->wk_keyix); 1635 1636 memset(&hk, 0, sizeof(hk)); 1637 hk.keyIndex = k->wk_keyix; 1638 switch (k->wk_cipher->ic_cipher) { 1639 case IEEE80211_CIPHER_WEP: 1640 hk.keyTypeId = KEY_TYPE_ID_WEP; 1641 break; 1642 case IEEE80211_CIPHER_TKIP: 1643 hk.keyTypeId = KEY_TYPE_ID_TKIP; 1644 break; 1645 case IEEE80211_CIPHER_AES_CCM: 1646 hk.keyTypeId = KEY_TYPE_ID_AES; 1647 break; 1648 default: 1649 /* XXX should not happen */ 1650 DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n", 1651 __func__, k->wk_cipher->ic_cipher); 1652 return 0; 1653 } 1654 return (mwl_hal_keyreset(hvap, &hk, bcastaddr) == 0); /*XXX*/ 1655 } 1656 1657 static __inline int 1658 addgroupflags(MWL_HAL_KEYVAL *hk, const struct ieee80211_key *k) 1659 { 1660 if (k->wk_flags & IEEE80211_KEY_GROUP) { 1661 if (k->wk_flags & IEEE80211_KEY_XMIT) 1662 hk->keyFlags |= KEY_FLAG_TXGROUPKEY; 1663 if (k->wk_flags & IEEE80211_KEY_RECV) 1664 hk->keyFlags |= KEY_FLAG_RXGROUPKEY; 1665 return 1; 1666 } else 1667 return 0; 1668 } 1669 1670 /* 1671 * Set the key cache contents for the specified key. Key cache 1672 * slot(s) must already have been allocated by mwl_key_alloc. 1673 */ 1674 static int 1675 mwl_key_set(struct ieee80211vap *vap, const struct ieee80211_key *k, 1676 const uint8_t mac[IEEE80211_ADDR_LEN]) 1677 { 1678 #define GRPXMIT (IEEE80211_KEY_XMIT | IEEE80211_KEY_GROUP) 1679 /* NB: static wep keys are marked GROUP+tx/rx; GTK will be tx or rx */ 1680 #define IEEE80211_IS_STATICKEY(k) \ 1681 (((k)->wk_flags & (GRPXMIT|IEEE80211_KEY_RECV)) == \ 1682 (GRPXMIT|IEEE80211_KEY_RECV)) 1683 struct mwl_softc *sc = vap->iv_ic->ic_ifp->if_softc; 1684 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap; 1685 const struct ieee80211_cipher *cip = k->wk_cipher; 1686 const uint8_t *macaddr; 1687 MWL_HAL_KEYVAL hk; 1688 1689 KASSERT((k->wk_flags & IEEE80211_KEY_SWCRYPT) == 0, 1690 ("s/w crypto set?")); 1691 1692 if (hvap == NULL) { 1693 if (vap->iv_opmode != IEEE80211_M_WDS) { 1694 /* XXX monitor mode? */ 1695 DPRINTF(sc, MWL_DEBUG_KEYCACHE, 1696 "%s: no hvap for opmode %d\n", __func__, 1697 vap->iv_opmode); 1698 return 0; 1699 } 1700 hvap = MWL_VAP(vap)->mv_ap_hvap; 1701 } 1702 memset(&hk, 0, sizeof(hk)); 1703 hk.keyIndex = k->wk_keyix; 1704 switch (cip->ic_cipher) { 1705 case IEEE80211_CIPHER_WEP: 1706 hk.keyTypeId = KEY_TYPE_ID_WEP; 1707 hk.keyLen = k->wk_keylen; 1708 if (k->wk_keyix == vap->iv_def_txkey) 1709 hk.keyFlags = KEY_FLAG_WEP_TXKEY; 1710 if (!IEEE80211_IS_STATICKEY(k)) { 1711 /* NB: WEP is never used for the PTK */ 1712 (void) addgroupflags(&hk, k); 1713 } 1714 break; 1715 case IEEE80211_CIPHER_TKIP: 1716 hk.keyTypeId = KEY_TYPE_ID_TKIP; 1717 hk.key.tkip.tsc.high = (uint32_t)(k->wk_keytsc >> 16); 1718 hk.key.tkip.tsc.low = (uint16_t)k->wk_keytsc; 1719 hk.keyFlags = KEY_FLAG_TSC_VALID | KEY_FLAG_MICKEY_VALID; 1720 hk.keyLen = k->wk_keylen + IEEE80211_MICBUF_SIZE; 1721 if (!addgroupflags(&hk, k)) 1722 hk.keyFlags |= KEY_FLAG_PAIRWISE; 1723 break; 1724 case IEEE80211_CIPHER_AES_CCM: 1725 hk.keyTypeId = KEY_TYPE_ID_AES; 1726 hk.keyLen = k->wk_keylen; 1727 if (!addgroupflags(&hk, k)) 1728 hk.keyFlags |= KEY_FLAG_PAIRWISE; 1729 break; 1730 default: 1731 /* XXX should not happen */ 1732 DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n", 1733 __func__, k->wk_cipher->ic_cipher); 1734 return 0; 1735 } 1736 /* 1737 * NB: tkip mic keys get copied here too; the layout 1738 * just happens to match that in ieee80211_key. 1739 */ 1740 memcpy(hk.key.aes, k->wk_key, hk.keyLen); 1741 1742 /* 1743 * Locate address of sta db entry for writing key; 1744 * the convention unfortunately is somewhat different 1745 * than how net80211, hostapd, and wpa_supplicant think. 1746 */ 1747 if (vap->iv_opmode == IEEE80211_M_STA) { 1748 /* 1749 * NB: keys plumbed before the sta reaches AUTH state 1750 * will be discarded or written to the wrong sta db 1751 * entry because iv_bss is meaningless. This is ok 1752 * (right now) because we handle deferred plumbing of 1753 * WEP keys when the sta reaches AUTH state. 1754 */ 1755 macaddr = vap->iv_bss->ni_bssid; 1756 if ((k->wk_flags & IEEE80211_KEY_GROUP) == 0) { 1757 /* XXX plumb to local sta db too for static key wep */ 1758 mwl_hal_keyset(hvap, &hk, vap->iv_myaddr); 1759 } 1760 } else if (vap->iv_opmode == IEEE80211_M_WDS && 1761 vap->iv_state != IEEE80211_S_RUN) { 1762 /* 1763 * Prior to RUN state a WDS vap will not it's BSS node 1764 * setup so we will plumb the key to the wrong mac 1765 * address (it'll be our local address). Workaround 1766 * this for the moment by grabbing the correct address. 1767 */ 1768 macaddr = vap->iv_des_bssid; 1769 } else if ((k->wk_flags & GRPXMIT) == GRPXMIT) 1770 macaddr = vap->iv_myaddr; 1771 else 1772 macaddr = mac; 1773 KEYPRINTF(sc, &hk, macaddr); 1774 return (mwl_hal_keyset(hvap, &hk, macaddr) == 0); 1775 #undef IEEE80211_IS_STATICKEY 1776 #undef GRPXMIT 1777 } 1778 1779 /* unaligned little endian access */ 1780 #define LE_READ_2(p) \ 1781 ((uint16_t) \ 1782 ((((const uint8_t *)(p))[0] ) | \ 1783 (((const uint8_t *)(p))[1] << 8))) 1784 #define LE_READ_4(p) \ 1785 ((uint32_t) \ 1786 ((((const uint8_t *)(p))[0] ) | \ 1787 (((const uint8_t *)(p))[1] << 8) | \ 1788 (((const uint8_t *)(p))[2] << 16) | \ 1789 (((const uint8_t *)(p))[3] << 24))) 1790 1791 /* 1792 * Set the multicast filter contents into the hardware. 1793 * XXX f/w has no support; just defer to the os. 1794 */ 1795 static void 1796 mwl_setmcastfilter(struct mwl_softc *sc) 1797 { 1798 struct ifnet *ifp = sc->sc_ifp; 1799 #if 0 1800 struct ether_multi *enm; 1801 struct ether_multistep estep; 1802 uint8_t macs[IEEE80211_ADDR_LEN*MWL_HAL_MCAST_MAX];/* XXX stack use */ 1803 uint8_t *mp; 1804 int nmc; 1805 1806 mp = macs; 1807 nmc = 0; 1808 ETHER_FIRST_MULTI(estep, &sc->sc_ec, enm); 1809 while (enm != NULL) { 1810 /* XXX Punt on ranges. */ 1811 if (nmc == MWL_HAL_MCAST_MAX || 1812 !IEEE80211_ADDR_EQ(enm->enm_addrlo, enm->enm_addrhi)) { 1813 ifp->if_flags |= IFF_ALLMULTI; 1814 return; 1815 } 1816 IEEE80211_ADDR_COPY(mp, enm->enm_addrlo); 1817 mp += IEEE80211_ADDR_LEN, nmc++; 1818 ETHER_NEXT_MULTI(estep, enm); 1819 } 1820 ifp->if_flags &= ~IFF_ALLMULTI; 1821 mwl_hal_setmcast(sc->sc_mh, nmc, macs); 1822 #else 1823 /* XXX no mcast filter support; we get everything */ 1824 ifp->if_flags |= IFF_ALLMULTI; 1825 #endif 1826 } 1827 1828 static int 1829 mwl_mode_init(struct mwl_softc *sc) 1830 { 1831 struct ifnet *ifp = sc->sc_ifp; 1832 struct ieee80211com *ic = ifp->if_l2com; 1833 struct mwl_hal *mh = sc->sc_mh; 1834 1835 /* 1836 * NB: Ignore promisc in hostap mode; it's set by the 1837 * bridge. This is wrong but we have no way to 1838 * identify internal requests (from the bridge) 1839 * versus external requests such as for tcpdump. 1840 */ 1841 mwl_hal_setpromisc(mh, (ifp->if_flags & IFF_PROMISC) && 1842 ic->ic_opmode != IEEE80211_M_HOSTAP); 1843 mwl_setmcastfilter(sc); 1844 1845 return 0; 1846 } 1847 1848 /* 1849 * Callback from the 802.11 layer after a multicast state change. 1850 */ 1851 static void 1852 mwl_update_mcast(struct ifnet *ifp) 1853 { 1854 struct mwl_softc *sc = ifp->if_softc; 1855 1856 mwl_setmcastfilter(sc); 1857 } 1858 1859 /* 1860 * Callback from the 802.11 layer after a promiscuous mode change. 1861 * Note this interface does not check the operating mode as this 1862 * is an internal callback and we are expected to honor the current 1863 * state (e.g. this is used for setting the interface in promiscuous 1864 * mode when operating in hostap mode to do ACS). 1865 */ 1866 static void 1867 mwl_update_promisc(struct ifnet *ifp) 1868 { 1869 struct mwl_softc *sc = ifp->if_softc; 1870 1871 mwl_hal_setpromisc(sc->sc_mh, (ifp->if_flags & IFF_PROMISC) != 0); 1872 } 1873 1874 /* 1875 * Callback from the 802.11 layer to update the slot time 1876 * based on the current setting. We use it to notify the 1877 * firmware of ERP changes and the f/w takes care of things 1878 * like slot time and preamble. 1879 */ 1880 static void 1881 mwl_updateslot(struct ifnet *ifp) 1882 { 1883 struct mwl_softc *sc = ifp->if_softc; 1884 struct ieee80211com *ic = ifp->if_l2com; 1885 struct mwl_hal *mh = sc->sc_mh; 1886 int prot; 1887 1888 /* NB: can be called early; suppress needless cmds */ 1889 if ((ifp->if_drv_flags & IFF_DRV_RUNNING) == 0) 1890 return; 1891 1892 /* 1893 * Calculate the ERP flags. The firwmare will use 1894 * this to carry out the appropriate measures. 1895 */ 1896 prot = 0; 1897 if (IEEE80211_IS_CHAN_ANYG(ic->ic_curchan)) { 1898 if ((ic->ic_flags & IEEE80211_F_SHSLOT) == 0) 1899 prot |= IEEE80211_ERP_NON_ERP_PRESENT; 1900 if (ic->ic_flags & IEEE80211_F_USEPROT) 1901 prot |= IEEE80211_ERP_USE_PROTECTION; 1902 if (ic->ic_flags & IEEE80211_F_USEBARKER) 1903 prot |= IEEE80211_ERP_LONG_PREAMBLE; 1904 } 1905 1906 DPRINTF(sc, MWL_DEBUG_RESET, 1907 "%s: chan %u MHz/flags 0x%x %s slot, (prot 0x%x ic_flags 0x%x)\n", 1908 __func__, ic->ic_curchan->ic_freq, ic->ic_curchan->ic_flags, 1909 ic->ic_flags & IEEE80211_F_SHSLOT ? "short" : "long", prot, 1910 ic->ic_flags); 1911 1912 mwl_hal_setgprot(mh, prot); 1913 } 1914 1915 /* 1916 * Setup the beacon frame. 1917 */ 1918 static int 1919 mwl_beacon_setup(struct ieee80211vap *vap) 1920 { 1921 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap; 1922 struct ieee80211_node *ni = vap->iv_bss; 1923 struct ieee80211_beacon_offsets bo; 1924 struct mbuf *m; 1925 1926 m = ieee80211_beacon_alloc(ni, &bo); 1927 if (m == NULL) 1928 return ENOBUFS; 1929 mwl_hal_setbeacon(hvap, mtod(m, const void *), m->m_len); 1930 m_free(m); 1931 1932 return 0; 1933 } 1934 1935 /* 1936 * Update the beacon frame in response to a change. 1937 */ 1938 static void 1939 mwl_beacon_update(struct ieee80211vap *vap, int item) 1940 { 1941 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap; 1942 struct ieee80211com *ic = vap->iv_ic; 1943 1944 KASSERT(hvap != NULL, ("no beacon")); 1945 switch (item) { 1946 case IEEE80211_BEACON_ERP: 1947 mwl_updateslot(ic->ic_ifp); 1948 break; 1949 case IEEE80211_BEACON_HTINFO: 1950 mwl_hal_setnprotmode(hvap, 1951 MS(ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE)); 1952 break; 1953 case IEEE80211_BEACON_CAPS: 1954 case IEEE80211_BEACON_WME: 1955 case IEEE80211_BEACON_APPIE: 1956 case IEEE80211_BEACON_CSA: 1957 break; 1958 case IEEE80211_BEACON_TIM: 1959 /* NB: firmware always forms TIM */ 1960 return; 1961 } 1962 /* XXX retain beacon frame and update */ 1963 mwl_beacon_setup(vap); 1964 } 1965 1966 static void 1967 mwl_load_cb(void *arg, bus_dma_segment_t *segs, int nsegs, int error) 1968 { 1969 bus_addr_t *paddr = (bus_addr_t*) arg; 1970 KASSERT(error == 0, ("error %u on bus_dma callback", error)); 1971 *paddr = segs->ds_addr; 1972 } 1973 1974 #ifdef MWL_HOST_PS_SUPPORT 1975 /* 1976 * Handle power save station occupancy changes. 1977 */ 1978 static void 1979 mwl_update_ps(struct ieee80211vap *vap, int nsta) 1980 { 1981 struct mwl_vap *mvp = MWL_VAP(vap); 1982 1983 if (nsta == 0 || mvp->mv_last_ps_sta == 0) 1984 mwl_hal_setpowersave_bss(mvp->mv_hvap, nsta); 1985 mvp->mv_last_ps_sta = nsta; 1986 } 1987 1988 /* 1989 * Handle associated station power save state changes. 1990 */ 1991 static int 1992 mwl_set_tim(struct ieee80211_node *ni, int set) 1993 { 1994 struct ieee80211vap *vap = ni->ni_vap; 1995 struct mwl_vap *mvp = MWL_VAP(vap); 1996 1997 if (mvp->mv_set_tim(ni, set)) { /* NB: state change */ 1998 mwl_hal_setpowersave_sta(mvp->mv_hvap, 1999 IEEE80211_AID(ni->ni_associd), set); 2000 return 1; 2001 } else 2002 return 0; 2003 } 2004 #endif /* MWL_HOST_PS_SUPPORT */ 2005 2006 static int 2007 mwl_desc_setup(struct mwl_softc *sc, const char *name, 2008 struct mwl_descdma *dd, 2009 int nbuf, size_t bufsize, int ndesc, size_t descsize) 2010 { 2011 struct ifnet *ifp = sc->sc_ifp; 2012 uint8_t *ds; 2013 int error; 2014 2015 DPRINTF(sc, MWL_DEBUG_RESET, 2016 "%s: %s DMA: %u bufs (%ju) %u desc/buf (%ju)\n", 2017 __func__, name, nbuf, (uintmax_t) bufsize, 2018 ndesc, (uintmax_t) descsize); 2019 2020 dd->dd_name = name; 2021 dd->dd_desc_len = nbuf * ndesc * descsize; 2022 2023 /* 2024 * Setup DMA descriptor area. 2025 */ 2026 error = bus_dma_tag_create(bus_get_dma_tag(sc->sc_dev), /* parent */ 2027 PAGE_SIZE, 0, /* alignment, bounds */ 2028 BUS_SPACE_MAXADDR_32BIT, /* lowaddr */ 2029 BUS_SPACE_MAXADDR, /* highaddr */ 2030 NULL, NULL, /* filter, filterarg */ 2031 dd->dd_desc_len, /* maxsize */ 2032 1, /* nsegments */ 2033 dd->dd_desc_len, /* maxsegsize */ 2034 BUS_DMA_ALLOCNOW, /* flags */ 2035 NULL, /* lockfunc */ 2036 NULL, /* lockarg */ 2037 &dd->dd_dmat); 2038 if (error != 0) { 2039 if_printf(ifp, "cannot allocate %s DMA tag\n", dd->dd_name); 2040 return error; 2041 } 2042 2043 /* allocate descriptors */ 2044 error = bus_dmamap_create(dd->dd_dmat, BUS_DMA_NOWAIT, &dd->dd_dmamap); 2045 if (error != 0) { 2046 if_printf(ifp, "unable to create dmamap for %s descriptors, " 2047 "error %u\n", dd->dd_name, error); 2048 goto fail0; 2049 } 2050 2051 error = bus_dmamem_alloc(dd->dd_dmat, (void**) &dd->dd_desc, 2052 BUS_DMA_NOWAIT | BUS_DMA_COHERENT, 2053 &dd->dd_dmamap); 2054 if (error != 0) { 2055 if_printf(ifp, "unable to alloc memory for %u %s descriptors, " 2056 "error %u\n", nbuf * ndesc, dd->dd_name, error); 2057 goto fail1; 2058 } 2059 2060 error = bus_dmamap_load(dd->dd_dmat, dd->dd_dmamap, 2061 dd->dd_desc, dd->dd_desc_len, 2062 mwl_load_cb, &dd->dd_desc_paddr, 2063 BUS_DMA_NOWAIT); 2064 if (error != 0) { 2065 if_printf(ifp, "unable to map %s descriptors, error %u\n", 2066 dd->dd_name, error); 2067 goto fail2; 2068 } 2069 2070 ds = dd->dd_desc; 2071 memset(ds, 0, dd->dd_desc_len); 2072 DPRINTF(sc, MWL_DEBUG_RESET, "%s: %s DMA map: %p (%lu) -> %p (%lu)\n", 2073 __func__, dd->dd_name, ds, (u_long) dd->dd_desc_len, 2074 (caddr_t) dd->dd_desc_paddr, /*XXX*/ (u_long) dd->dd_desc_len); 2075 2076 return 0; 2077 fail2: 2078 bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap); 2079 fail1: 2080 bus_dmamap_destroy(dd->dd_dmat, dd->dd_dmamap); 2081 fail0: 2082 bus_dma_tag_destroy(dd->dd_dmat); 2083 memset(dd, 0, sizeof(*dd)); 2084 return error; 2085 #undef DS2PHYS 2086 } 2087 2088 static void 2089 mwl_desc_cleanup(struct mwl_softc *sc, struct mwl_descdma *dd) 2090 { 2091 bus_dmamap_unload(dd->dd_dmat, dd->dd_dmamap); 2092 bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap); 2093 bus_dmamap_destroy(dd->dd_dmat, dd->dd_dmamap); 2094 bus_dma_tag_destroy(dd->dd_dmat); 2095 2096 memset(dd, 0, sizeof(*dd)); 2097 } 2098 2099 /* 2100 * Construct a tx q's free list. The order of entries on 2101 * the list must reflect the physical layout of tx descriptors 2102 * because the firmware pre-fetches descriptors. 2103 * 2104 * XXX might be better to use indices into the buffer array. 2105 */ 2106 static void 2107 mwl_txq_reset(struct mwl_softc *sc, struct mwl_txq *txq) 2108 { 2109 struct mwl_txbuf *bf; 2110 int i; 2111 2112 bf = txq->dma.dd_bufptr; 2113 STAILQ_INIT(&txq->free); 2114 for (i = 0; i < mwl_txbuf; i++, bf++) 2115 STAILQ_INSERT_TAIL(&txq->free, bf, bf_list); 2116 txq->nfree = i; 2117 } 2118 2119 #define DS2PHYS(_dd, _ds) \ 2120 ((_dd)->dd_desc_paddr + ((caddr_t)(_ds) - (caddr_t)(_dd)->dd_desc)) 2121 2122 static int 2123 mwl_txdma_setup(struct mwl_softc *sc, struct mwl_txq *txq) 2124 { 2125 struct ifnet *ifp = sc->sc_ifp; 2126 int error, bsize, i; 2127 struct mwl_txbuf *bf; 2128 struct mwl_txdesc *ds; 2129 2130 error = mwl_desc_setup(sc, "tx", &txq->dma, 2131 mwl_txbuf, sizeof(struct mwl_txbuf), 2132 MWL_TXDESC, sizeof(struct mwl_txdesc)); 2133 if (error != 0) 2134 return error; 2135 2136 /* allocate and setup tx buffers */ 2137 bsize = mwl_txbuf * sizeof(struct mwl_txbuf); 2138 bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO); 2139 if (bf == NULL) { 2140 if_printf(ifp, "malloc of %u tx buffers failed\n", 2141 mwl_txbuf); 2142 return ENOMEM; 2143 } 2144 txq->dma.dd_bufptr = bf; 2145 2146 ds = txq->dma.dd_desc; 2147 for (i = 0; i < mwl_txbuf; i++, bf++, ds += MWL_TXDESC) { 2148 bf->bf_desc = ds; 2149 bf->bf_daddr = DS2PHYS(&txq->dma, ds); 2150 error = bus_dmamap_create(sc->sc_dmat, BUS_DMA_NOWAIT, 2151 &bf->bf_dmamap); 2152 if (error != 0) { 2153 if_printf(ifp, "unable to create dmamap for tx " 2154 "buffer %u, error %u\n", i, error); 2155 return error; 2156 } 2157 } 2158 mwl_txq_reset(sc, txq); 2159 return 0; 2160 } 2161 2162 static void 2163 mwl_txdma_cleanup(struct mwl_softc *sc, struct mwl_txq *txq) 2164 { 2165 struct mwl_txbuf *bf; 2166 int i; 2167 2168 bf = txq->dma.dd_bufptr; 2169 for (i = 0; i < mwl_txbuf; i++, bf++) { 2170 KASSERT(bf->bf_m == NULL, ("mbuf on free list")); 2171 KASSERT(bf->bf_node == NULL, ("node on free list")); 2172 if (bf->bf_dmamap != NULL) 2173 bus_dmamap_destroy(sc->sc_dmat, bf->bf_dmamap); 2174 } 2175 STAILQ_INIT(&txq->free); 2176 txq->nfree = 0; 2177 if (txq->dma.dd_bufptr != NULL) { 2178 free(txq->dma.dd_bufptr, M_MWLDEV); 2179 txq->dma.dd_bufptr = NULL; 2180 } 2181 if (txq->dma.dd_desc_len != 0) 2182 mwl_desc_cleanup(sc, &txq->dma); 2183 } 2184 2185 static int 2186 mwl_rxdma_setup(struct mwl_softc *sc) 2187 { 2188 struct ifnet *ifp = sc->sc_ifp; 2189 int error, jumbosize, bsize, i; 2190 struct mwl_rxbuf *bf; 2191 struct mwl_jumbo *rbuf; 2192 struct mwl_rxdesc *ds; 2193 caddr_t data; 2194 2195 error = mwl_desc_setup(sc, "rx", &sc->sc_rxdma, 2196 mwl_rxdesc, sizeof(struct mwl_rxbuf), 2197 1, sizeof(struct mwl_rxdesc)); 2198 if (error != 0) 2199 return error; 2200 2201 /* 2202 * Receive is done to a private pool of jumbo buffers. 2203 * This allows us to attach to mbuf's and avoid re-mapping 2204 * memory on each rx we post. We allocate a large chunk 2205 * of memory and manage it in the driver. The mbuf free 2206 * callback method is used to reclaim frames after sending 2207 * them up the stack. By default we allocate 2x the number of 2208 * rx descriptors configured so we have some slop to hold 2209 * us while frames are processed. 2210 */ 2211 if (mwl_rxbuf < 2*mwl_rxdesc) { 2212 if_printf(ifp, 2213 "too few rx dma buffers (%d); increasing to %d\n", 2214 mwl_rxbuf, 2*mwl_rxdesc); 2215 mwl_rxbuf = 2*mwl_rxdesc; 2216 } 2217 jumbosize = roundup(MWL_AGGR_SIZE, PAGE_SIZE); 2218 sc->sc_rxmemsize = mwl_rxbuf*jumbosize; 2219 2220 error = bus_dma_tag_create(sc->sc_dmat, /* parent */ 2221 PAGE_SIZE, 0, /* alignment, bounds */ 2222 BUS_SPACE_MAXADDR_32BIT, /* lowaddr */ 2223 BUS_SPACE_MAXADDR, /* highaddr */ 2224 NULL, NULL, /* filter, filterarg */ 2225 sc->sc_rxmemsize, /* maxsize */ 2226 1, /* nsegments */ 2227 sc->sc_rxmemsize, /* maxsegsize */ 2228 BUS_DMA_ALLOCNOW, /* flags */ 2229 NULL, /* lockfunc */ 2230 NULL, /* lockarg */ 2231 &sc->sc_rxdmat); 2232 error = bus_dmamap_create(sc->sc_rxdmat, BUS_DMA_NOWAIT, &sc->sc_rxmap); 2233 if (error != 0) { 2234 if_printf(ifp, "could not create rx DMA map\n"); 2235 return error; 2236 } 2237 2238 error = bus_dmamem_alloc(sc->sc_rxdmat, (void**) &sc->sc_rxmem, 2239 BUS_DMA_NOWAIT | BUS_DMA_COHERENT, 2240 &sc->sc_rxmap); 2241 if (error != 0) { 2242 if_printf(ifp, "could not alloc %ju bytes of rx DMA memory\n", 2243 (uintmax_t) sc->sc_rxmemsize); 2244 return error; 2245 } 2246 2247 error = bus_dmamap_load(sc->sc_rxdmat, sc->sc_rxmap, 2248 sc->sc_rxmem, sc->sc_rxmemsize, 2249 mwl_load_cb, &sc->sc_rxmem_paddr, 2250 BUS_DMA_NOWAIT); 2251 if (error != 0) { 2252 if_printf(ifp, "could not load rx DMA map\n"); 2253 return error; 2254 } 2255 2256 /* 2257 * Allocate rx buffers and set them up. 2258 */ 2259 bsize = mwl_rxdesc * sizeof(struct mwl_rxbuf); 2260 bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO); 2261 if (bf == NULL) { 2262 if_printf(ifp, "malloc of %u rx buffers failed\n", bsize); 2263 return error; 2264 } 2265 sc->sc_rxdma.dd_bufptr = bf; 2266 2267 STAILQ_INIT(&sc->sc_rxbuf); 2268 ds = sc->sc_rxdma.dd_desc; 2269 for (i = 0; i < mwl_rxdesc; i++, bf++, ds++) { 2270 bf->bf_desc = ds; 2271 bf->bf_daddr = DS2PHYS(&sc->sc_rxdma, ds); 2272 /* pre-assign dma buffer */ 2273 bf->bf_data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize); 2274 /* NB: tail is intentional to preserve descriptor order */ 2275 STAILQ_INSERT_TAIL(&sc->sc_rxbuf, bf, bf_list); 2276 } 2277 2278 /* 2279 * Place remainder of dma memory buffers on the free list. 2280 */ 2281 SLIST_INIT(&sc->sc_rxfree); 2282 for (; i < mwl_rxbuf; i++) { 2283 data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize); 2284 rbuf = MWL_JUMBO_DATA2BUF(data); 2285 SLIST_INSERT_HEAD(&sc->sc_rxfree, rbuf, next); 2286 sc->sc_nrxfree++; 2287 } 2288 return 0; 2289 } 2290 #undef DS2PHYS 2291 2292 static void 2293 mwl_rxdma_cleanup(struct mwl_softc *sc) 2294 { 2295 if (sc->sc_rxmap != NULL) 2296 bus_dmamap_unload(sc->sc_rxdmat, sc->sc_rxmap); 2297 if (sc->sc_rxmem != NULL) { 2298 bus_dmamem_free(sc->sc_rxdmat, sc->sc_rxmem, sc->sc_rxmap); 2299 sc->sc_rxmem = NULL; 2300 } 2301 if (sc->sc_rxmap != NULL) { 2302 bus_dmamap_destroy(sc->sc_rxdmat, sc->sc_rxmap); 2303 sc->sc_rxmap = NULL; 2304 } 2305 if (sc->sc_rxdma.dd_bufptr != NULL) { 2306 free(sc->sc_rxdma.dd_bufptr, M_MWLDEV); 2307 sc->sc_rxdma.dd_bufptr = NULL; 2308 } 2309 if (sc->sc_rxdma.dd_desc_len != 0) 2310 mwl_desc_cleanup(sc, &sc->sc_rxdma); 2311 } 2312 2313 static int 2314 mwl_dma_setup(struct mwl_softc *sc) 2315 { 2316 int error, i; 2317 2318 error = mwl_rxdma_setup(sc); 2319 if (error != 0) { 2320 mwl_rxdma_cleanup(sc); 2321 return error; 2322 } 2323 2324 for (i = 0; i < MWL_NUM_TX_QUEUES; i++) { 2325 error = mwl_txdma_setup(sc, &sc->sc_txq[i]); 2326 if (error != 0) { 2327 mwl_dma_cleanup(sc); 2328 return error; 2329 } 2330 } 2331 return 0; 2332 } 2333 2334 static void 2335 mwl_dma_cleanup(struct mwl_softc *sc) 2336 { 2337 int i; 2338 2339 for (i = 0; i < MWL_NUM_TX_QUEUES; i++) 2340 mwl_txdma_cleanup(sc, &sc->sc_txq[i]); 2341 mwl_rxdma_cleanup(sc); 2342 } 2343 2344 static struct ieee80211_node * 2345 mwl_node_alloc(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN]) 2346 { 2347 struct ieee80211com *ic = vap->iv_ic; 2348 struct mwl_softc *sc = ic->ic_ifp->if_softc; 2349 const size_t space = sizeof(struct mwl_node); 2350 struct mwl_node *mn; 2351 2352 mn = malloc(space, M_80211_NODE, M_NOWAIT|M_ZERO); 2353 if (mn == NULL) { 2354 /* XXX stat+msg */ 2355 return NULL; 2356 } 2357 DPRINTF(sc, MWL_DEBUG_NODE, "%s: mn %p\n", __func__, mn); 2358 return &mn->mn_node; 2359 } 2360 2361 static void 2362 mwl_node_cleanup(struct ieee80211_node *ni) 2363 { 2364 struct ieee80211com *ic = ni->ni_ic; 2365 struct mwl_softc *sc = ic->ic_ifp->if_softc; 2366 struct mwl_node *mn = MWL_NODE(ni); 2367 2368 DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p ic %p staid %d\n", 2369 __func__, ni, ni->ni_ic, mn->mn_staid); 2370 2371 if (mn->mn_staid != 0) { 2372 struct ieee80211vap *vap = ni->ni_vap; 2373 2374 if (mn->mn_hvap != NULL) { 2375 if (vap->iv_opmode == IEEE80211_M_STA) 2376 mwl_hal_delstation(mn->mn_hvap, vap->iv_myaddr); 2377 else 2378 mwl_hal_delstation(mn->mn_hvap, ni->ni_macaddr); 2379 } 2380 /* 2381 * NB: legacy WDS peer sta db entry is installed using 2382 * the associate ap's hvap; use it again to delete it. 2383 * XXX can vap be NULL? 2384 */ 2385 else if (vap->iv_opmode == IEEE80211_M_WDS && 2386 MWL_VAP(vap)->mv_ap_hvap != NULL) 2387 mwl_hal_delstation(MWL_VAP(vap)->mv_ap_hvap, 2388 ni->ni_macaddr); 2389 delstaid(sc, mn->mn_staid); 2390 mn->mn_staid = 0; 2391 } 2392 sc->sc_node_cleanup(ni); 2393 } 2394 2395 /* 2396 * Reclaim rx dma buffers from packets sitting on the ampdu 2397 * reorder queue for a station. We replace buffers with a 2398 * system cluster (if available). 2399 */ 2400 static void 2401 mwl_ampdu_rxdma_reclaim(struct ieee80211_rx_ampdu *rap) 2402 { 2403 #if 0 2404 int i, n, off; 2405 struct mbuf *m; 2406 void *cl; 2407 2408 n = rap->rxa_qframes; 2409 for (i = 0; i < rap->rxa_wnd && n > 0; i++) { 2410 m = rap->rxa_m[i]; 2411 if (m == NULL) 2412 continue; 2413 n--; 2414 /* our dma buffers have a well-known free routine */ 2415 if ((m->m_flags & M_EXT) == 0 || 2416 m->m_ext.ext_free != mwl_ext_free) 2417 continue; 2418 /* 2419 * Try to allocate a cluster and move the data. 2420 */ 2421 off = m->m_data - m->m_ext.ext_buf; 2422 if (off + m->m_pkthdr.len > MCLBYTES) { 2423 /* XXX no AMSDU for now */ 2424 continue; 2425 } 2426 cl = pool_cache_get_paddr(&mclpool_cache, 0, 2427 &m->m_ext.ext_paddr); 2428 if (cl != NULL) { 2429 /* 2430 * Copy the existing data to the cluster, remove 2431 * the rx dma buffer, and attach the cluster in 2432 * its place. Note we preserve the offset to the 2433 * data so frames being bridged can still prepend 2434 * their headers without adding another mbuf. 2435 */ 2436 memcpy((caddr_t) cl + off, m->m_data, m->m_pkthdr.len); 2437 MEXTREMOVE(m); 2438 MEXTADD(m, cl, MCLBYTES, 0, NULL, &mclpool_cache); 2439 /* setup mbuf like _MCLGET does */ 2440 m->m_flags |= M_CLUSTER | M_EXT_RW; 2441 _MOWNERREF(m, M_EXT | M_CLUSTER); 2442 /* NB: m_data is clobbered by MEXTADDR, adjust */ 2443 m->m_data += off; 2444 } 2445 } 2446 #endif 2447 } 2448 2449 /* 2450 * Callback to reclaim resources. We first let the 2451 * net80211 layer do it's thing, then if we are still 2452 * blocked by a lack of rx dma buffers we walk the ampdu 2453 * reorder q's to reclaim buffers by copying to a system 2454 * cluster. 2455 */ 2456 static void 2457 mwl_node_drain(struct ieee80211_node *ni) 2458 { 2459 struct ieee80211com *ic = ni->ni_ic; 2460 struct mwl_softc *sc = ic->ic_ifp->if_softc; 2461 struct mwl_node *mn = MWL_NODE(ni); 2462 2463 DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p vap %p staid %d\n", 2464 __func__, ni, ni->ni_vap, mn->mn_staid); 2465 2466 /* NB: call up first to age out ampdu q's */ 2467 sc->sc_node_drain(ni); 2468 2469 /* XXX better to not check low water mark? */ 2470 if (sc->sc_rxblocked && mn->mn_staid != 0 && 2471 (ni->ni_flags & IEEE80211_NODE_HT)) { 2472 uint8_t tid; 2473 /* 2474 * Walk the reorder q and reclaim rx dma buffers by copying 2475 * the packet contents into clusters. 2476 */ 2477 for (tid = 0; tid < WME_NUM_TID; tid++) { 2478 struct ieee80211_rx_ampdu *rap; 2479 2480 rap = &ni->ni_rx_ampdu[tid]; 2481 if ((rap->rxa_flags & IEEE80211_AGGR_XCHGPEND) == 0) 2482 continue; 2483 if (rap->rxa_qframes) 2484 mwl_ampdu_rxdma_reclaim(rap); 2485 } 2486 } 2487 } 2488 2489 static void 2490 mwl_node_getsignal(const struct ieee80211_node *ni, int8_t *rssi, int8_t *noise) 2491 { 2492 *rssi = ni->ni_ic->ic_node_getrssi(ni); 2493 #ifdef MWL_ANT_INFO_SUPPORT 2494 #if 0 2495 /* XXX need to smooth data */ 2496 *noise = -MWL_NODE_CONST(ni)->mn_ai.nf; 2497 #else 2498 *noise = -95; /* XXX */ 2499 #endif 2500 #else 2501 *noise = -95; /* XXX */ 2502 #endif 2503 } 2504 2505 /* 2506 * Convert Hardware per-antenna rssi info to common format: 2507 * Let a1, a2, a3 represent the amplitudes per chain 2508 * Let amax represent max[a1, a2, a3] 2509 * Rssi1_dBm = RSSI_dBm + 20*log10(a1/amax) 2510 * Rssi1_dBm = RSSI_dBm + 20*log10(a1) - 20*log10(amax) 2511 * We store a table that is 4*20*log10(idx) - the extra 4 is to store or 2512 * maintain some extra precision. 2513 * 2514 * Values are stored in .5 db format capped at 127. 2515 */ 2516 static void 2517 mwl_node_getmimoinfo(const struct ieee80211_node *ni, 2518 struct ieee80211_mimo_info *mi) 2519 { 2520 #define CVT(_dst, _src) do { \ 2521 (_dst) = rssi + ((logdbtbl[_src] - logdbtbl[rssi_max]) >> 2); \ 2522 (_dst) = (_dst) > 64 ? 127 : ((_dst) << 1); \ 2523 } while (0) 2524 static const int8_t logdbtbl[32] = { 2525 0, 0, 24, 38, 48, 56, 62, 68, 2526 72, 76, 80, 83, 86, 89, 92, 94, 2527 96, 98, 100, 102, 104, 106, 107, 109, 2528 110, 112, 113, 115, 116, 117, 118, 119 2529 }; 2530 const struct mwl_node *mn = MWL_NODE_CONST(ni); 2531 uint8_t rssi = mn->mn_ai.rsvd1/2; /* XXX */ 2532 uint32_t rssi_max; 2533 2534 rssi_max = mn->mn_ai.rssi_a; 2535 if (mn->mn_ai.rssi_b > rssi_max) 2536 rssi_max = mn->mn_ai.rssi_b; 2537 if (mn->mn_ai.rssi_c > rssi_max) 2538 rssi_max = mn->mn_ai.rssi_c; 2539 2540 CVT(mi->rssi[0], mn->mn_ai.rssi_a); 2541 CVT(mi->rssi[1], mn->mn_ai.rssi_b); 2542 CVT(mi->rssi[2], mn->mn_ai.rssi_c); 2543 2544 mi->noise[0] = mn->mn_ai.nf_a; 2545 mi->noise[1] = mn->mn_ai.nf_b; 2546 mi->noise[2] = mn->mn_ai.nf_c; 2547 #undef CVT 2548 } 2549 2550 static __inline void * 2551 mwl_getrxdma(struct mwl_softc *sc) 2552 { 2553 struct mwl_jumbo *buf; 2554 void *data; 2555 2556 /* 2557 * Allocate from jumbo pool. 2558 */ 2559 MWL_RXFREE_LOCK(sc); 2560 buf = SLIST_FIRST(&sc->sc_rxfree); 2561 if (buf == NULL) { 2562 DPRINTF(sc, MWL_DEBUG_ANY, 2563 "%s: out of rx dma buffers\n", __func__); 2564 sc->sc_stats.mst_rx_nodmabuf++; 2565 data = NULL; 2566 } else { 2567 SLIST_REMOVE_HEAD(&sc->sc_rxfree, next); 2568 sc->sc_nrxfree--; 2569 data = MWL_JUMBO_BUF2DATA(buf); 2570 } 2571 MWL_RXFREE_UNLOCK(sc); 2572 return data; 2573 } 2574 2575 static __inline void 2576 mwl_putrxdma(struct mwl_softc *sc, void *data) 2577 { 2578 struct mwl_jumbo *buf; 2579 2580 /* XXX bounds check data */ 2581 MWL_RXFREE_LOCK(sc); 2582 buf = MWL_JUMBO_DATA2BUF(data); 2583 SLIST_INSERT_HEAD(&sc->sc_rxfree, buf, next); 2584 sc->sc_nrxfree++; 2585 MWL_RXFREE_UNLOCK(sc); 2586 } 2587 2588 static int 2589 mwl_rxbuf_init(struct mwl_softc *sc, struct mwl_rxbuf *bf) 2590 { 2591 struct mwl_rxdesc *ds; 2592 2593 ds = bf->bf_desc; 2594 if (bf->bf_data == NULL) { 2595 bf->bf_data = mwl_getrxdma(sc); 2596 if (bf->bf_data == NULL) { 2597 /* mark descriptor to be skipped */ 2598 ds->RxControl = EAGLE_RXD_CTRL_OS_OWN; 2599 /* NB: don't need PREREAD */ 2600 MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREWRITE); 2601 sc->sc_stats.mst_rxbuf_failed++; 2602 return ENOMEM; 2603 } 2604 } 2605 /* 2606 * NB: DMA buffer contents is known to be unmodified 2607 * so there's no need to flush the data cache. 2608 */ 2609 2610 /* 2611 * Setup descriptor. 2612 */ 2613 ds->QosCtrl = 0; 2614 ds->RSSI = 0; 2615 ds->Status = EAGLE_RXD_STATUS_IDLE; 2616 ds->Channel = 0; 2617 ds->PktLen = htole16(MWL_AGGR_SIZE); 2618 ds->SQ2 = 0; 2619 ds->pPhysBuffData = htole32(MWL_JUMBO_DMA_ADDR(sc, bf->bf_data)); 2620 /* NB: don't touch pPhysNext, set once */ 2621 ds->RxControl = EAGLE_RXD_CTRL_DRIVER_OWN; 2622 MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE); 2623 2624 return 0; 2625 } 2626 2627 static int 2628 mwl_ext_free(struct mbuf *m, void *data, void *arg) 2629 { 2630 struct mwl_softc *sc = arg; 2631 2632 /* XXX bounds check data */ 2633 mwl_putrxdma(sc, data); 2634 /* 2635 * If we were previously blocked by a lack of rx dma buffers 2636 * check if we now have enough to restart rx interrupt handling. 2637 * NB: we know we are called at splvm which is above splnet. 2638 */ 2639 if (sc->sc_rxblocked && sc->sc_nrxfree > mwl_rxdmalow) { 2640 sc->sc_rxblocked = 0; 2641 mwl_hal_intrset(sc->sc_mh, sc->sc_imask); 2642 } 2643 return (EXT_FREE_OK); 2644 } 2645 2646 struct mwl_frame_bar { 2647 u_int8_t i_fc[2]; 2648 u_int8_t i_dur[2]; 2649 u_int8_t i_ra[IEEE80211_ADDR_LEN]; 2650 u_int8_t i_ta[IEEE80211_ADDR_LEN]; 2651 /* ctl, seq, FCS */ 2652 } __packed; 2653 2654 /* 2655 * Like ieee80211_anyhdrsize, but handles BAR frames 2656 * specially so the logic below to piece the 802.11 2657 * header together works. 2658 */ 2659 static __inline int 2660 mwl_anyhdrsize(const void *data) 2661 { 2662 const struct ieee80211_frame *wh = data; 2663 2664 if ((wh->i_fc[0]&IEEE80211_FC0_TYPE_MASK) == IEEE80211_FC0_TYPE_CTL) { 2665 switch (wh->i_fc[0] & IEEE80211_FC0_SUBTYPE_MASK) { 2666 case IEEE80211_FC0_SUBTYPE_CTS: 2667 case IEEE80211_FC0_SUBTYPE_ACK: 2668 return sizeof(struct ieee80211_frame_ack); 2669 case IEEE80211_FC0_SUBTYPE_BAR: 2670 return sizeof(struct mwl_frame_bar); 2671 } 2672 return sizeof(struct ieee80211_frame_min); 2673 } else 2674 return ieee80211_hdrsize(data); 2675 } 2676 2677 static void 2678 mwl_handlemicerror(struct ieee80211com *ic, const uint8_t *data) 2679 { 2680 const struct ieee80211_frame *wh; 2681 struct ieee80211_node *ni; 2682 2683 wh = (const struct ieee80211_frame *)(data + sizeof(uint16_t)); 2684 ni = ieee80211_find_rxnode(ic, (const struct ieee80211_frame_min *) wh); 2685 if (ni != NULL) { 2686 ieee80211_notify_michael_failure(ni->ni_vap, wh, 0); 2687 ieee80211_free_node(ni); 2688 } 2689 } 2690 2691 /* 2692 * Convert hardware signal strength to rssi. The value 2693 * provided by the device has the noise floor added in; 2694 * we need to compensate for this but we don't have that 2695 * so we use a fixed value. 2696 * 2697 * The offset of 8 is good for both 2.4 and 5GHz. The LNA 2698 * offset is already set as part of the initial gain. This 2699 * will give at least +/- 3dB for 2.4GHz and +/- 5dB for 5GHz. 2700 */ 2701 static __inline int 2702 cvtrssi(uint8_t ssi) 2703 { 2704 int rssi = (int) ssi + 8; 2705 /* XXX hack guess until we have a real noise floor */ 2706 rssi = 2*(87 - rssi); /* NB: .5 dBm units */ 2707 return (rssi < 0 ? 0 : rssi > 127 ? 127 : rssi); 2708 } 2709 2710 static void 2711 mwl_rx_proc(void *arg, int npending) 2712 { 2713 #define IEEE80211_DIR_DSTODS(wh) \ 2714 ((((const struct ieee80211_frame *)wh)->i_fc[1] & IEEE80211_FC1_DIR_MASK) == IEEE80211_FC1_DIR_DSTODS) 2715 struct mwl_softc *sc = arg; 2716 struct ifnet *ifp = sc->sc_ifp; 2717 struct ieee80211com *ic = ifp->if_l2com; 2718 struct mwl_rxbuf *bf; 2719 struct mwl_rxdesc *ds; 2720 struct mbuf *m; 2721 struct ieee80211_qosframe *wh; 2722 struct ieee80211_qosframe_addr4 *wh4; 2723 struct ieee80211_node *ni; 2724 struct mwl_node *mn; 2725 int off, len, hdrlen, pktlen, rssi, ntodo; 2726 uint8_t *data, status; 2727 void *newdata; 2728 int16_t nf; 2729 2730 DPRINTF(sc, MWL_DEBUG_RX_PROC, "%s: pending %u rdptr 0x%x wrptr 0x%x\n", 2731 __func__, npending, RD4(sc, sc->sc_hwspecs.rxDescRead), 2732 RD4(sc, sc->sc_hwspecs.rxDescWrite)); 2733 nf = -96; /* XXX */ 2734 bf = sc->sc_rxnext; 2735 for (ntodo = mwl_rxquota; ntodo > 0; ntodo--) { 2736 if (bf == NULL) 2737 bf = STAILQ_FIRST(&sc->sc_rxbuf); 2738 ds = bf->bf_desc; 2739 data = bf->bf_data; 2740 if (data == NULL) { 2741 /* 2742 * If data allocation failed previously there 2743 * will be no buffer; try again to re-populate it. 2744 * Note the firmware will not advance to the next 2745 * descriptor with a dma buffer so we must mimic 2746 * this or we'll get out of sync. 2747 */ 2748 DPRINTF(sc, MWL_DEBUG_ANY, 2749 "%s: rx buf w/o dma memory\n", __func__); 2750 (void) mwl_rxbuf_init(sc, bf); 2751 sc->sc_stats.mst_rx_dmabufmissing++; 2752 break; 2753 } 2754 MWL_RXDESC_SYNC(sc, ds, 2755 BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE); 2756 if (ds->RxControl != EAGLE_RXD_CTRL_DMA_OWN) 2757 break; 2758 #ifdef MWL_DEBUG 2759 if (sc->sc_debug & MWL_DEBUG_RECV_DESC) 2760 mwl_printrxbuf(bf, 0); 2761 #endif 2762 status = ds->Status; 2763 if (status & EAGLE_RXD_STATUS_DECRYPT_ERR_MASK) { 2764 ifp->if_ierrors++; 2765 sc->sc_stats.mst_rx_crypto++; 2766 /* 2767 * NB: Check EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR 2768 * for backwards compatibility. 2769 */ 2770 if (status != EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR && 2771 (status & EAGLE_RXD_STATUS_TKIP_MIC_DECRYPT_ERR)) { 2772 /* 2773 * MIC error, notify upper layers. 2774 */ 2775 bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap, 2776 BUS_DMASYNC_POSTREAD); 2777 mwl_handlemicerror(ic, data); 2778 sc->sc_stats.mst_rx_tkipmic++; 2779 } 2780 /* XXX too painful to tap packets */ 2781 goto rx_next; 2782 } 2783 /* 2784 * Sync the data buffer. 2785 */ 2786 len = le16toh(ds->PktLen); 2787 bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap, BUS_DMASYNC_POSTREAD); 2788 /* 2789 * The 802.11 header is provided all or in part at the front; 2790 * use it to calculate the true size of the header that we'll 2791 * construct below. We use this to figure out where to copy 2792 * payload prior to constructing the header. 2793 */ 2794 hdrlen = mwl_anyhdrsize(data + sizeof(uint16_t)); 2795 off = sizeof(uint16_t) + sizeof(struct ieee80211_frame_addr4); 2796 2797 /* calculate rssi early so we can re-use for each aggregate */ 2798 rssi = cvtrssi(ds->RSSI); 2799 2800 pktlen = hdrlen + (len - off); 2801 /* 2802 * NB: we know our frame is at least as large as 2803 * IEEE80211_MIN_LEN because there is a 4-address 2804 * frame at the front. Hence there's no need to 2805 * vet the packet length. If the frame in fact 2806 * is too small it should be discarded at the 2807 * net80211 layer. 2808 */ 2809 2810 /* 2811 * Attach dma buffer to an mbuf. We tried 2812 * doing this based on the packet size (i.e. 2813 * copying small packets) but it turns out to 2814 * be a net loss. The tradeoff might be system 2815 * dependent (cache architecture is important). 2816 */ 2817 MGETHDR(m, M_NOWAIT, MT_DATA); 2818 if (m == NULL) { 2819 DPRINTF(sc, MWL_DEBUG_ANY, 2820 "%s: no rx mbuf\n", __func__); 2821 sc->sc_stats.mst_rx_nombuf++; 2822 goto rx_next; 2823 } 2824 /* 2825 * Acquire the replacement dma buffer before 2826 * processing the frame. If we're out of dma 2827 * buffers we disable rx interrupts and wait 2828 * for the free pool to reach mlw_rxdmalow buffers 2829 * before starting to do work again. If the firmware 2830 * runs out of descriptors then it will toss frames 2831 * which is better than our doing it as that can 2832 * starve our processing. It is also important that 2833 * we always process rx'd frames in case they are 2834 * A-MPDU as otherwise the host's view of the BA 2835 * window may get out of sync with the firmware. 2836 */ 2837 newdata = mwl_getrxdma(sc); 2838 if (newdata == NULL) { 2839 /* NB: stat+msg in mwl_getrxdma */ 2840 m_free(m); 2841 /* disable RX interrupt and mark state */ 2842 mwl_hal_intrset(sc->sc_mh, 2843 sc->sc_imask &~ MACREG_A2HRIC_BIT_RX_RDY); 2844 sc->sc_rxblocked = 1; 2845 ieee80211_drain(ic); 2846 /* XXX check rxblocked and immediately start again? */ 2847 goto rx_stop; 2848 } 2849 bf->bf_data = newdata; 2850 /* 2851 * Attach the dma buffer to the mbuf; 2852 * mwl_rxbuf_init will re-setup the rx 2853 * descriptor using the replacement dma 2854 * buffer we just installed above. 2855 */ 2856 MEXTADD(m, data, MWL_AGGR_SIZE, mwl_ext_free, 2857 data, sc, 0, EXT_NET_DRV); 2858 m->m_data += off - hdrlen; 2859 m->m_pkthdr.len = m->m_len = pktlen; 2860 m->m_pkthdr.rcvif = ifp; 2861 /* NB: dma buffer assumed read-only */ 2862 2863 /* 2864 * Piece 802.11 header together. 2865 */ 2866 wh = mtod(m, struct ieee80211_qosframe *); 2867 /* NB: don't need to do this sometimes but ... */ 2868 /* XXX special case so we can memcpy after m_devget? */ 2869 ovbcopy(data + sizeof(uint16_t), wh, hdrlen); 2870 if (IEEE80211_QOS_HAS_SEQ(wh)) { 2871 if (IEEE80211_DIR_DSTODS(wh)) { 2872 wh4 = mtod(m, 2873 struct ieee80211_qosframe_addr4*); 2874 *(uint16_t *)wh4->i_qos = ds->QosCtrl; 2875 } else { 2876 *(uint16_t *)wh->i_qos = ds->QosCtrl; 2877 } 2878 } 2879 /* 2880 * The f/w strips WEP header but doesn't clear 2881 * the WEP bit; mark the packet with M_WEP so 2882 * net80211 will treat the data as decrypted. 2883 * While here also clear the PWR_MGT bit since 2884 * power save is handled by the firmware and 2885 * passing this up will potentially cause the 2886 * upper layer to put a station in power save 2887 * (except when configured with MWL_HOST_PS_SUPPORT). 2888 */ 2889 if (wh->i_fc[1] & IEEE80211_FC1_PROTECTED) 2890 m->m_flags |= M_WEP; 2891 #ifdef MWL_HOST_PS_SUPPORT 2892 wh->i_fc[1] &= ~IEEE80211_FC1_PROTECTED; 2893 #else 2894 wh->i_fc[1] &= ~(IEEE80211_FC1_PROTECTED | 2895 IEEE80211_FC1_PWR_MGT); 2896 #endif 2897 2898 if (ieee80211_radiotap_active(ic)) { 2899 struct mwl_rx_radiotap_header *tap = &sc->sc_rx_th; 2900 2901 tap->wr_flags = 0; 2902 tap->wr_rate = ds->Rate; 2903 tap->wr_antsignal = rssi + nf; 2904 tap->wr_antnoise = nf; 2905 } 2906 if (IFF_DUMPPKTS_RECV(sc, wh)) { 2907 ieee80211_dump_pkt(ic, mtod(m, caddr_t), 2908 len, ds->Rate, rssi); 2909 } 2910 ifp->if_ipackets++; 2911 2912 /* dispatch */ 2913 ni = ieee80211_find_rxnode(ic, 2914 (const struct ieee80211_frame_min *) wh); 2915 if (ni != NULL) { 2916 mn = MWL_NODE(ni); 2917 #ifdef MWL_ANT_INFO_SUPPORT 2918 mn->mn_ai.rssi_a = ds->ai.rssi_a; 2919 mn->mn_ai.rssi_b = ds->ai.rssi_b; 2920 mn->mn_ai.rssi_c = ds->ai.rssi_c; 2921 mn->mn_ai.rsvd1 = rssi; 2922 #endif 2923 /* tag AMPDU aggregates for reorder processing */ 2924 if (ni->ni_flags & IEEE80211_NODE_HT) 2925 m->m_flags |= M_AMPDU; 2926 (void) ieee80211_input(ni, m, rssi, nf); 2927 ieee80211_free_node(ni); 2928 } else 2929 (void) ieee80211_input_all(ic, m, rssi, nf); 2930 rx_next: 2931 /* NB: ignore ENOMEM so we process more descriptors */ 2932 (void) mwl_rxbuf_init(sc, bf); 2933 bf = STAILQ_NEXT(bf, bf_list); 2934 } 2935 rx_stop: 2936 sc->sc_rxnext = bf; 2937 2938 if ((ifp->if_drv_flags & IFF_DRV_OACTIVE) == 0 && 2939 !IFQ_IS_EMPTY(&ifp->if_snd)) { 2940 /* NB: kick fw; the tx thread may have been preempted */ 2941 mwl_hal_txstart(sc->sc_mh, 0); 2942 mwl_start(ifp); 2943 } 2944 #undef IEEE80211_DIR_DSTODS 2945 } 2946 2947 static void 2948 mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *txq, int qnum) 2949 { 2950 struct mwl_txbuf *bf, *bn; 2951 struct mwl_txdesc *ds; 2952 2953 MWL_TXQ_LOCK_INIT(sc, txq); 2954 txq->qnum = qnum; 2955 txq->txpri = 0; /* XXX */ 2956 #if 0 2957 /* NB: q setup by mwl_txdma_setup XXX */ 2958 STAILQ_INIT(&txq->free); 2959 #endif 2960 STAILQ_FOREACH(bf, &txq->free, bf_list) { 2961 bf->bf_txq = txq; 2962 2963 ds = bf->bf_desc; 2964 bn = STAILQ_NEXT(bf, bf_list); 2965 if (bn == NULL) 2966 bn = STAILQ_FIRST(&txq->free); 2967 ds->pPhysNext = htole32(bn->bf_daddr); 2968 } 2969 STAILQ_INIT(&txq->active); 2970 } 2971 2972 /* 2973 * Setup a hardware data transmit queue for the specified 2974 * access control. We record the mapping from ac's 2975 * to h/w queues for use by mwl_tx_start. 2976 */ 2977 static int 2978 mwl_tx_setup(struct mwl_softc *sc, int ac, int mvtype) 2979 { 2980 #define N(a) (sizeof(a)/sizeof(a[0])) 2981 struct mwl_txq *txq; 2982 2983 if (ac >= N(sc->sc_ac2q)) { 2984 device_printf(sc->sc_dev, "AC %u out of range, max %zu!\n", 2985 ac, N(sc->sc_ac2q)); 2986 return 0; 2987 } 2988 if (mvtype >= MWL_NUM_TX_QUEUES) { 2989 device_printf(sc->sc_dev, "mvtype %u out of range, max %u!\n", 2990 mvtype, MWL_NUM_TX_QUEUES); 2991 return 0; 2992 } 2993 txq = &sc->sc_txq[mvtype]; 2994 mwl_txq_init(sc, txq, mvtype); 2995 sc->sc_ac2q[ac] = txq; 2996 return 1; 2997 #undef N 2998 } 2999 3000 /* 3001 * Update WME parameters for a transmit queue. 3002 */ 3003 static int 3004 mwl_txq_update(struct mwl_softc *sc, int ac) 3005 { 3006 #define MWL_EXPONENT_TO_VALUE(v) ((1<<v)-1) 3007 struct ifnet *ifp = sc->sc_ifp; 3008 struct ieee80211com *ic = ifp->if_l2com; 3009 struct mwl_txq *txq = sc->sc_ac2q[ac]; 3010 struct wmeParams *wmep = &ic->ic_wme.wme_chanParams.cap_wmeParams[ac]; 3011 struct mwl_hal *mh = sc->sc_mh; 3012 int aifs, cwmin, cwmax, txoplim; 3013 3014 aifs = wmep->wmep_aifsn; 3015 /* XXX in sta mode need to pass log values for cwmin/max */ 3016 cwmin = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmin); 3017 cwmax = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmax); 3018 txoplim = wmep->wmep_txopLimit; /* NB: units of 32us */ 3019 3020 if (mwl_hal_setedcaparams(mh, txq->qnum, cwmin, cwmax, aifs, txoplim)) { 3021 device_printf(sc->sc_dev, "unable to update hardware queue " 3022 "parameters for %s traffic!\n", 3023 ieee80211_wme_acnames[ac]); 3024 return 0; 3025 } 3026 return 1; 3027 #undef MWL_EXPONENT_TO_VALUE 3028 } 3029 3030 /* 3031 * Callback from the 802.11 layer to update WME parameters. 3032 */ 3033 static int 3034 mwl_wme_update(struct ieee80211com *ic) 3035 { 3036 struct mwl_softc *sc = ic->ic_ifp->if_softc; 3037 3038 return !mwl_txq_update(sc, WME_AC_BE) || 3039 !mwl_txq_update(sc, WME_AC_BK) || 3040 !mwl_txq_update(sc, WME_AC_VI) || 3041 !mwl_txq_update(sc, WME_AC_VO) ? EIO : 0; 3042 } 3043 3044 /* 3045 * Reclaim resources for a setup queue. 3046 */ 3047 static void 3048 mwl_tx_cleanupq(struct mwl_softc *sc, struct mwl_txq *txq) 3049 { 3050 /* XXX hal work? */ 3051 MWL_TXQ_LOCK_DESTROY(txq); 3052 } 3053 3054 /* 3055 * Reclaim all tx queue resources. 3056 */ 3057 static void 3058 mwl_tx_cleanup(struct mwl_softc *sc) 3059 { 3060 int i; 3061 3062 for (i = 0; i < MWL_NUM_TX_QUEUES; i++) 3063 mwl_tx_cleanupq(sc, &sc->sc_txq[i]); 3064 } 3065 3066 static int 3067 mwl_tx_dmasetup(struct mwl_softc *sc, struct mwl_txbuf *bf, struct mbuf *m0) 3068 { 3069 struct mbuf *m; 3070 int error; 3071 3072 /* 3073 * Load the DMA map so any coalescing is done. This 3074 * also calculates the number of descriptors we need. 3075 */ 3076 error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0, 3077 bf->bf_segs, &bf->bf_nseg, 3078 BUS_DMA_NOWAIT); 3079 if (error == EFBIG) { 3080 /* XXX packet requires too many descriptors */ 3081 bf->bf_nseg = MWL_TXDESC+1; 3082 } else if (error != 0) { 3083 sc->sc_stats.mst_tx_busdma++; 3084 m_freem(m0); 3085 return error; 3086 } 3087 /* 3088 * Discard null packets and check for packets that 3089 * require too many TX descriptors. We try to convert 3090 * the latter to a cluster. 3091 */ 3092 if (error == EFBIG) { /* too many desc's, linearize */ 3093 sc->sc_stats.mst_tx_linear++; 3094 #if MWL_TXDESC > 1 3095 m = m_collapse(m0, M_NOWAIT, MWL_TXDESC); 3096 #else 3097 m = m_defrag(m0, M_NOWAIT); 3098 #endif 3099 if (m == NULL) { 3100 m_freem(m0); 3101 sc->sc_stats.mst_tx_nombuf++; 3102 return ENOMEM; 3103 } 3104 m0 = m; 3105 error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0, 3106 bf->bf_segs, &bf->bf_nseg, 3107 BUS_DMA_NOWAIT); 3108 if (error != 0) { 3109 sc->sc_stats.mst_tx_busdma++; 3110 m_freem(m0); 3111 return error; 3112 } 3113 KASSERT(bf->bf_nseg <= MWL_TXDESC, 3114 ("too many segments after defrag; nseg %u", bf->bf_nseg)); 3115 } else if (bf->bf_nseg == 0) { /* null packet, discard */ 3116 sc->sc_stats.mst_tx_nodata++; 3117 m_freem(m0); 3118 return EIO; 3119 } 3120 DPRINTF(sc, MWL_DEBUG_XMIT, "%s: m %p len %u\n", 3121 __func__, m0, m0->m_pkthdr.len); 3122 bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap, BUS_DMASYNC_PREWRITE); 3123 bf->bf_m = m0; 3124 3125 return 0; 3126 } 3127 3128 static __inline int 3129 mwl_cvtlegacyrate(int rate) 3130 { 3131 switch (rate) { 3132 case 2: return 0; 3133 case 4: return 1; 3134 case 11: return 2; 3135 case 22: return 3; 3136 case 44: return 4; 3137 case 12: return 5; 3138 case 18: return 6; 3139 case 24: return 7; 3140 case 36: return 8; 3141 case 48: return 9; 3142 case 72: return 10; 3143 case 96: return 11; 3144 case 108:return 12; 3145 } 3146 return 0; 3147 } 3148 3149 /* 3150 * Calculate fixed tx rate information per client state; 3151 * this value is suitable for writing to the Format field 3152 * of a tx descriptor. 3153 */ 3154 static uint16_t 3155 mwl_calcformat(uint8_t rate, const struct ieee80211_node *ni) 3156 { 3157 uint16_t fmt; 3158 3159 fmt = SM(3, EAGLE_TXD_ANTENNA) 3160 | (IEEE80211_IS_CHAN_HT40D(ni->ni_chan) ? 3161 EAGLE_TXD_EXTCHAN_LO : EAGLE_TXD_EXTCHAN_HI); 3162 if (rate & IEEE80211_RATE_MCS) { /* HT MCS */ 3163 fmt |= EAGLE_TXD_FORMAT_HT 3164 /* NB: 0x80 implicitly stripped from ucastrate */ 3165 | SM(rate, EAGLE_TXD_RATE); 3166 /* XXX short/long GI may be wrong; re-check */ 3167 if (IEEE80211_IS_CHAN_HT40(ni->ni_chan)) { 3168 fmt |= EAGLE_TXD_CHW_40 3169 | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI40 ? 3170 EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG); 3171 } else { 3172 fmt |= EAGLE_TXD_CHW_20 3173 | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI20 ? 3174 EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG); 3175 } 3176 } else { /* legacy rate */ 3177 fmt |= EAGLE_TXD_FORMAT_LEGACY 3178 | SM(mwl_cvtlegacyrate(rate), EAGLE_TXD_RATE) 3179 | EAGLE_TXD_CHW_20 3180 /* XXX iv_flags & IEEE80211_F_SHPREAMBLE? */ 3181 | (ni->ni_capinfo & IEEE80211_CAPINFO_SHORT_PREAMBLE ? 3182 EAGLE_TXD_PREAMBLE_SHORT : EAGLE_TXD_PREAMBLE_LONG); 3183 } 3184 return fmt; 3185 } 3186 3187 static int 3188 mwl_tx_start(struct mwl_softc *sc, struct ieee80211_node *ni, struct mwl_txbuf *bf, 3189 struct mbuf *m0) 3190 { 3191 #define IEEE80211_DIR_DSTODS(wh) \ 3192 ((wh->i_fc[1] & IEEE80211_FC1_DIR_MASK) == IEEE80211_FC1_DIR_DSTODS) 3193 struct ifnet *ifp = sc->sc_ifp; 3194 struct ieee80211com *ic = ifp->if_l2com; 3195 struct ieee80211vap *vap = ni->ni_vap; 3196 int error, iswep, ismcast; 3197 int hdrlen, copyhdrlen, pktlen; 3198 struct mwl_txdesc *ds; 3199 struct mwl_txq *txq; 3200 struct ieee80211_frame *wh; 3201 struct mwltxrec *tr; 3202 struct mwl_node *mn; 3203 uint16_t qos; 3204 #if MWL_TXDESC > 1 3205 int i; 3206 #endif 3207 3208 wh = mtod(m0, struct ieee80211_frame *); 3209 iswep = wh->i_fc[1] & IEEE80211_FC1_PROTECTED; 3210 ismcast = IEEE80211_IS_MULTICAST(wh->i_addr1); 3211 hdrlen = ieee80211_anyhdrsize(wh); 3212 copyhdrlen = hdrlen; 3213 pktlen = m0->m_pkthdr.len; 3214 if (IEEE80211_QOS_HAS_SEQ(wh)) { 3215 if (IEEE80211_DIR_DSTODS(wh)) { 3216 qos = *(uint16_t *) 3217 (((struct ieee80211_qosframe_addr4 *) wh)->i_qos); 3218 copyhdrlen -= sizeof(qos); 3219 } else 3220 qos = *(uint16_t *) 3221 (((struct ieee80211_qosframe *) wh)->i_qos); 3222 } else 3223 qos = 0; 3224 3225 if (iswep) { 3226 const struct ieee80211_cipher *cip; 3227 struct ieee80211_key *k; 3228 3229 /* 3230 * Construct the 802.11 header+trailer for an encrypted 3231 * frame. The only reason this can fail is because of an 3232 * unknown or unsupported cipher/key type. 3233 * 3234 * NB: we do this even though the firmware will ignore 3235 * what we've done for WEP and TKIP as we need the 3236 * ExtIV filled in for CCMP and this also adjusts 3237 * the headers which simplifies our work below. 3238 */ 3239 k = ieee80211_crypto_encap(ni, m0); 3240 if (k == NULL) { 3241 /* 3242 * This can happen when the key is yanked after the 3243 * frame was queued. Just discard the frame; the 3244 * 802.11 layer counts failures and provides 3245 * debugging/diagnostics. 3246 */ 3247 m_freem(m0); 3248 return EIO; 3249 } 3250 /* 3251 * Adjust the packet length for the crypto additions 3252 * done during encap and any other bits that the f/w 3253 * will add later on. 3254 */ 3255 cip = k->wk_cipher; 3256 pktlen += cip->ic_header + cip->ic_miclen + cip->ic_trailer; 3257 3258 /* packet header may have moved, reset our local pointer */ 3259 wh = mtod(m0, struct ieee80211_frame *); 3260 } 3261 3262 if (ieee80211_radiotap_active_vap(vap)) { 3263 sc->sc_tx_th.wt_flags = 0; /* XXX */ 3264 if (iswep) 3265 sc->sc_tx_th.wt_flags |= IEEE80211_RADIOTAP_F_WEP; 3266 #if 0 3267 sc->sc_tx_th.wt_rate = ds->DataRate; 3268 #endif 3269 sc->sc_tx_th.wt_txpower = ni->ni_txpower; 3270 sc->sc_tx_th.wt_antenna = sc->sc_txantenna; 3271 3272 ieee80211_radiotap_tx(vap, m0); 3273 } 3274 /* 3275 * Copy up/down the 802.11 header; the firmware requires 3276 * we present a 2-byte payload length followed by a 3277 * 4-address header (w/o QoS), followed (optionally) by 3278 * any WEP/ExtIV header (but only filled in for CCMP). 3279 * We are assured the mbuf has sufficient headroom to 3280 * prepend in-place by the setup of ic_headroom in 3281 * mwl_attach. 3282 */ 3283 if (hdrlen < sizeof(struct mwltxrec)) { 3284 const int space = sizeof(struct mwltxrec) - hdrlen; 3285 if (M_LEADINGSPACE(m0) < space) { 3286 /* NB: should never happen */ 3287 device_printf(sc->sc_dev, 3288 "not enough headroom, need %d found %zd, " 3289 "m_flags 0x%x m_len %d\n", 3290 space, M_LEADINGSPACE(m0), m0->m_flags, m0->m_len); 3291 ieee80211_dump_pkt(ic, 3292 mtod(m0, const uint8_t *), m0->m_len, 0, -1); 3293 m_freem(m0); 3294 sc->sc_stats.mst_tx_noheadroom++; 3295 return EIO; 3296 } 3297 M_PREPEND(m0, space, M_NOWAIT); 3298 } 3299 tr = mtod(m0, struct mwltxrec *); 3300 if (wh != (struct ieee80211_frame *) &tr->wh) 3301 ovbcopy(wh, &tr->wh, hdrlen); 3302 /* 3303 * Note: the "firmware length" is actually the length 3304 * of the fully formed "802.11 payload". That is, it's 3305 * everything except for the 802.11 header. In particular 3306 * this includes all crypto material including the MIC! 3307 */ 3308 tr->fwlen = htole16(pktlen - hdrlen); 3309 3310 /* 3311 * Load the DMA map so any coalescing is done. This 3312 * also calculates the number of descriptors we need. 3313 */ 3314 error = mwl_tx_dmasetup(sc, bf, m0); 3315 if (error != 0) { 3316 /* NB: stat collected in mwl_tx_dmasetup */ 3317 DPRINTF(sc, MWL_DEBUG_XMIT, 3318 "%s: unable to setup dma\n", __func__); 3319 return error; 3320 } 3321 bf->bf_node = ni; /* NB: held reference */ 3322 m0 = bf->bf_m; /* NB: may have changed */ 3323 tr = mtod(m0, struct mwltxrec *); 3324 wh = (struct ieee80211_frame *)&tr->wh; 3325 3326 /* 3327 * Formulate tx descriptor. 3328 */ 3329 ds = bf->bf_desc; 3330 txq = bf->bf_txq; 3331 3332 ds->QosCtrl = qos; /* NB: already little-endian */ 3333 #if MWL_TXDESC == 1 3334 /* 3335 * NB: multiframes should be zero because the descriptors 3336 * are initialized to zero. This should handle the case 3337 * where the driver is built with MWL_TXDESC=1 but we are 3338 * using firmware with multi-segment support. 3339 */ 3340 ds->PktPtr = htole32(bf->bf_segs[0].ds_addr); 3341 ds->PktLen = htole16(bf->bf_segs[0].ds_len); 3342 #else 3343 ds->multiframes = htole32(bf->bf_nseg); 3344 ds->PktLen = htole16(m0->m_pkthdr.len); 3345 for (i = 0; i < bf->bf_nseg; i++) { 3346 ds->PktPtrArray[i] = htole32(bf->bf_segs[i].ds_addr); 3347 ds->PktLenArray[i] = htole16(bf->bf_segs[i].ds_len); 3348 } 3349 #endif 3350 /* NB: pPhysNext, DataRate, and SapPktInfo setup once, don't touch */ 3351 ds->Format = 0; 3352 ds->pad = 0; 3353 ds->ack_wcb_addr = 0; 3354 3355 mn = MWL_NODE(ni); 3356 /* 3357 * Select transmit rate. 3358 */ 3359 switch (wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK) { 3360 case IEEE80211_FC0_TYPE_MGT: 3361 sc->sc_stats.mst_tx_mgmt++; 3362 /* fall thru... */ 3363 case IEEE80211_FC0_TYPE_CTL: 3364 /* NB: assign to BE q to avoid bursting */ 3365 ds->TxPriority = MWL_WME_AC_BE; 3366 break; 3367 case IEEE80211_FC0_TYPE_DATA: 3368 if (!ismcast) { 3369 const struct ieee80211_txparam *tp = ni->ni_txparms; 3370 /* 3371 * EAPOL frames get forced to a fixed rate and w/o 3372 * aggregation; otherwise check for any fixed rate 3373 * for the client (may depend on association state). 3374 */ 3375 if (m0->m_flags & M_EAPOL) { 3376 const struct mwl_vap *mvp = MWL_VAP_CONST(vap); 3377 ds->Format = mvp->mv_eapolformat; 3378 ds->pad = htole16( 3379 EAGLE_TXD_FIXED_RATE | EAGLE_TXD_DONT_AGGR); 3380 } else if (tp->ucastrate != IEEE80211_FIXED_RATE_NONE) { 3381 /* XXX pre-calculate per node */ 3382 ds->Format = htole16( 3383 mwl_calcformat(tp->ucastrate, ni)); 3384 ds->pad = htole16(EAGLE_TXD_FIXED_RATE); 3385 } 3386 /* NB: EAPOL frames will never have qos set */ 3387 if (qos == 0) 3388 ds->TxPriority = txq->qnum; 3389 #if MWL_MAXBA > 3 3390 else if (mwl_bastream_match(&mn->mn_ba[3], qos)) 3391 ds->TxPriority = mn->mn_ba[3].txq; 3392 #endif 3393 #if MWL_MAXBA > 2 3394 else if (mwl_bastream_match(&mn->mn_ba[2], qos)) 3395 ds->TxPriority = mn->mn_ba[2].txq; 3396 #endif 3397 #if MWL_MAXBA > 1 3398 else if (mwl_bastream_match(&mn->mn_ba[1], qos)) 3399 ds->TxPriority = mn->mn_ba[1].txq; 3400 #endif 3401 #if MWL_MAXBA > 0 3402 else if (mwl_bastream_match(&mn->mn_ba[0], qos)) 3403 ds->TxPriority = mn->mn_ba[0].txq; 3404 #endif 3405 else 3406 ds->TxPriority = txq->qnum; 3407 } else 3408 ds->TxPriority = txq->qnum; 3409 break; 3410 default: 3411 if_printf(ifp, "bogus frame type 0x%x (%s)\n", 3412 wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK, __func__); 3413 sc->sc_stats.mst_tx_badframetype++; 3414 m_freem(m0); 3415 return EIO; 3416 } 3417 3418 if (IFF_DUMPPKTS_XMIT(sc)) 3419 ieee80211_dump_pkt(ic, 3420 mtod(m0, const uint8_t *)+sizeof(uint16_t), 3421 m0->m_len - sizeof(uint16_t), ds->DataRate, -1); 3422 3423 MWL_TXQ_LOCK(txq); 3424 ds->Status = htole32(EAGLE_TXD_STATUS_FW_OWNED); 3425 STAILQ_INSERT_TAIL(&txq->active, bf, bf_list); 3426 MWL_TXDESC_SYNC(txq, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE); 3427 3428 ifp->if_opackets++; 3429 sc->sc_tx_timer = 5; 3430 MWL_TXQ_UNLOCK(txq); 3431 3432 return 0; 3433 #undef IEEE80211_DIR_DSTODS 3434 } 3435 3436 static __inline int 3437 mwl_cvtlegacyrix(int rix) 3438 { 3439 #define N(x) (sizeof(x)/sizeof(x[0])) 3440 static const int ieeerates[] = 3441 { 2, 4, 11, 22, 44, 12, 18, 24, 36, 48, 72, 96, 108 }; 3442 return (rix < N(ieeerates) ? ieeerates[rix] : 0); 3443 #undef N 3444 } 3445 3446 /* 3447 * Process completed xmit descriptors from the specified queue. 3448 */ 3449 static int 3450 mwl_tx_processq(struct mwl_softc *sc, struct mwl_txq *txq) 3451 { 3452 #define EAGLE_TXD_STATUS_MCAST \ 3453 (EAGLE_TXD_STATUS_MULTICAST_TX | EAGLE_TXD_STATUS_BROADCAST_TX) 3454 struct ifnet *ifp = sc->sc_ifp; 3455 struct ieee80211com *ic = ifp->if_l2com; 3456 struct mwl_txbuf *bf; 3457 struct mwl_txdesc *ds; 3458 struct ieee80211_node *ni; 3459 struct mwl_node *an; 3460 int nreaped; 3461 uint32_t status; 3462 3463 DPRINTF(sc, MWL_DEBUG_TX_PROC, "%s: tx queue %u\n", __func__, txq->qnum); 3464 for (nreaped = 0;; nreaped++) { 3465 MWL_TXQ_LOCK(txq); 3466 bf = STAILQ_FIRST(&txq->active); 3467 if (bf == NULL) { 3468 MWL_TXQ_UNLOCK(txq); 3469 break; 3470 } 3471 ds = bf->bf_desc; 3472 MWL_TXDESC_SYNC(txq, ds, 3473 BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE); 3474 if (ds->Status & htole32(EAGLE_TXD_STATUS_FW_OWNED)) { 3475 MWL_TXQ_UNLOCK(txq); 3476 break; 3477 } 3478 STAILQ_REMOVE_HEAD(&txq->active, bf_list); 3479 MWL_TXQ_UNLOCK(txq); 3480 3481 #ifdef MWL_DEBUG 3482 if (sc->sc_debug & MWL_DEBUG_XMIT_DESC) 3483 mwl_printtxbuf(bf, txq->qnum, nreaped); 3484 #endif 3485 ni = bf->bf_node; 3486 if (ni != NULL) { 3487 an = MWL_NODE(ni); 3488 status = le32toh(ds->Status); 3489 if (status & EAGLE_TXD_STATUS_OK) { 3490 uint16_t Format = le16toh(ds->Format); 3491 uint8_t txant = MS(Format, EAGLE_TXD_ANTENNA); 3492 3493 sc->sc_stats.mst_ant_tx[txant]++; 3494 if (status & EAGLE_TXD_STATUS_OK_RETRY) 3495 sc->sc_stats.mst_tx_retries++; 3496 if (status & EAGLE_TXD_STATUS_OK_MORE_RETRY) 3497 sc->sc_stats.mst_tx_mretries++; 3498 if (txq->qnum >= MWL_WME_AC_VO) 3499 ic->ic_wme.wme_hipri_traffic++; 3500 ni->ni_txrate = MS(Format, EAGLE_TXD_RATE); 3501 if ((Format & EAGLE_TXD_FORMAT_HT) == 0) { 3502 ni->ni_txrate = mwl_cvtlegacyrix( 3503 ni->ni_txrate); 3504 } else 3505 ni->ni_txrate |= IEEE80211_RATE_MCS; 3506 sc->sc_stats.mst_tx_rate = ni->ni_txrate; 3507 } else { 3508 if (status & EAGLE_TXD_STATUS_FAILED_LINK_ERROR) 3509 sc->sc_stats.mst_tx_linkerror++; 3510 if (status & EAGLE_TXD_STATUS_FAILED_XRETRY) 3511 sc->sc_stats.mst_tx_xretries++; 3512 if (status & EAGLE_TXD_STATUS_FAILED_AGING) 3513 sc->sc_stats.mst_tx_aging++; 3514 if (bf->bf_m->m_flags & M_FF) 3515 sc->sc_stats.mst_ff_txerr++; 3516 } 3517 /* 3518 * Do any tx complete callback. Note this must 3519 * be done before releasing the node reference. 3520 * XXX no way to figure out if frame was ACK'd 3521 */ 3522 if (bf->bf_m->m_flags & M_TXCB) { 3523 /* XXX strip fw len in case header inspected */ 3524 m_adj(bf->bf_m, sizeof(uint16_t)); 3525 ieee80211_process_callback(ni, bf->bf_m, 3526 (status & EAGLE_TXD_STATUS_OK) == 0); 3527 } 3528 /* 3529 * Reclaim reference to node. 3530 * 3531 * NB: the node may be reclaimed here if, for example 3532 * this is a DEAUTH message that was sent and the 3533 * node was timed out due to inactivity. 3534 */ 3535 ieee80211_free_node(ni); 3536 } 3537 ds->Status = htole32(EAGLE_TXD_STATUS_IDLE); 3538 3539 bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap, 3540 BUS_DMASYNC_POSTWRITE); 3541 bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap); 3542 m_freem(bf->bf_m); 3543 3544 mwl_puttxbuf_tail(txq, bf); 3545 } 3546 return nreaped; 3547 #undef EAGLE_TXD_STATUS_MCAST 3548 } 3549 3550 /* 3551 * Deferred processing of transmit interrupt; special-cased 3552 * for four hardware queues, 0-3. 3553 */ 3554 static void 3555 mwl_tx_proc(void *arg, int npending) 3556 { 3557 struct mwl_softc *sc = arg; 3558 struct ifnet *ifp = sc->sc_ifp; 3559 int nreaped; 3560 3561 /* 3562 * Process each active queue. 3563 */ 3564 nreaped = 0; 3565 if (!STAILQ_EMPTY(&sc->sc_txq[0].active)) 3566 nreaped += mwl_tx_processq(sc, &sc->sc_txq[0]); 3567 if (!STAILQ_EMPTY(&sc->sc_txq[1].active)) 3568 nreaped += mwl_tx_processq(sc, &sc->sc_txq[1]); 3569 if (!STAILQ_EMPTY(&sc->sc_txq[2].active)) 3570 nreaped += mwl_tx_processq(sc, &sc->sc_txq[2]); 3571 if (!STAILQ_EMPTY(&sc->sc_txq[3].active)) 3572 nreaped += mwl_tx_processq(sc, &sc->sc_txq[3]); 3573 3574 if (nreaped != 0) { 3575 ifp->if_drv_flags &= ~IFF_DRV_OACTIVE; 3576 sc->sc_tx_timer = 0; 3577 if (!IFQ_IS_EMPTY(&ifp->if_snd)) { 3578 /* NB: kick fw; the tx thread may have been preempted */ 3579 mwl_hal_txstart(sc->sc_mh, 0); 3580 mwl_start(ifp); 3581 } 3582 } 3583 } 3584 3585 static void 3586 mwl_tx_draintxq(struct mwl_softc *sc, struct mwl_txq *txq) 3587 { 3588 struct ieee80211_node *ni; 3589 struct mwl_txbuf *bf; 3590 u_int ix; 3591 3592 /* 3593 * NB: this assumes output has been stopped and 3594 * we do not need to block mwl_tx_tasklet 3595 */ 3596 for (ix = 0;; ix++) { 3597 MWL_TXQ_LOCK(txq); 3598 bf = STAILQ_FIRST(&txq->active); 3599 if (bf == NULL) { 3600 MWL_TXQ_UNLOCK(txq); 3601 break; 3602 } 3603 STAILQ_REMOVE_HEAD(&txq->active, bf_list); 3604 MWL_TXQ_UNLOCK(txq); 3605 #ifdef MWL_DEBUG 3606 if (sc->sc_debug & MWL_DEBUG_RESET) { 3607 struct ifnet *ifp = sc->sc_ifp; 3608 struct ieee80211com *ic = ifp->if_l2com; 3609 const struct mwltxrec *tr = 3610 mtod(bf->bf_m, const struct mwltxrec *); 3611 mwl_printtxbuf(bf, txq->qnum, ix); 3612 ieee80211_dump_pkt(ic, (const uint8_t *)&tr->wh, 3613 bf->bf_m->m_len - sizeof(tr->fwlen), 0, -1); 3614 } 3615 #endif /* MWL_DEBUG */ 3616 bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap); 3617 ni = bf->bf_node; 3618 if (ni != NULL) { 3619 /* 3620 * Reclaim node reference. 3621 */ 3622 ieee80211_free_node(ni); 3623 } 3624 m_freem(bf->bf_m); 3625 3626 mwl_puttxbuf_tail(txq, bf); 3627 } 3628 } 3629 3630 /* 3631 * Drain the transmit queues and reclaim resources. 3632 */ 3633 static void 3634 mwl_draintxq(struct mwl_softc *sc) 3635 { 3636 struct ifnet *ifp = sc->sc_ifp; 3637 int i; 3638 3639 for (i = 0; i < MWL_NUM_TX_QUEUES; i++) 3640 mwl_tx_draintxq(sc, &sc->sc_txq[i]); 3641 ifp->if_drv_flags &= ~IFF_DRV_OACTIVE; 3642 sc->sc_tx_timer = 0; 3643 } 3644 3645 #ifdef MWL_DIAGAPI 3646 /* 3647 * Reset the transmit queues to a pristine state after a fw download. 3648 */ 3649 static void 3650 mwl_resettxq(struct mwl_softc *sc) 3651 { 3652 int i; 3653 3654 for (i = 0; i < MWL_NUM_TX_QUEUES; i++) 3655 mwl_txq_reset(sc, &sc->sc_txq[i]); 3656 } 3657 #endif /* MWL_DIAGAPI */ 3658 3659 /* 3660 * Clear the transmit queues of any frames submitted for the 3661 * specified vap. This is done when the vap is deleted so we 3662 * don't potentially reference the vap after it is gone. 3663 * Note we cannot remove the frames; we only reclaim the node 3664 * reference. 3665 */ 3666 static void 3667 mwl_cleartxq(struct mwl_softc *sc, struct ieee80211vap *vap) 3668 { 3669 struct mwl_txq *txq; 3670 struct mwl_txbuf *bf; 3671 int i; 3672 3673 for (i = 0; i < MWL_NUM_TX_QUEUES; i++) { 3674 txq = &sc->sc_txq[i]; 3675 MWL_TXQ_LOCK(txq); 3676 STAILQ_FOREACH(bf, &txq->active, bf_list) { 3677 struct ieee80211_node *ni = bf->bf_node; 3678 if (ni != NULL && ni->ni_vap == vap) { 3679 bf->bf_node = NULL; 3680 ieee80211_free_node(ni); 3681 } 3682 } 3683 MWL_TXQ_UNLOCK(txq); 3684 } 3685 } 3686 3687 static int 3688 mwl_recv_action(struct ieee80211_node *ni, const struct ieee80211_frame *wh, 3689 const uint8_t *frm, const uint8_t *efrm) 3690 { 3691 struct mwl_softc *sc = ni->ni_ic->ic_ifp->if_softc; 3692 const struct ieee80211_action *ia; 3693 3694 ia = (const struct ieee80211_action *) frm; 3695 if (ia->ia_category == IEEE80211_ACTION_CAT_HT && 3696 ia->ia_action == IEEE80211_ACTION_HT_MIMOPWRSAVE) { 3697 const struct ieee80211_action_ht_mimopowersave *mps = 3698 (const struct ieee80211_action_ht_mimopowersave *) ia; 3699 3700 mwl_hal_setmimops(sc->sc_mh, ni->ni_macaddr, 3701 mps->am_control & IEEE80211_A_HT_MIMOPWRSAVE_ENA, 3702 MS(mps->am_control, IEEE80211_A_HT_MIMOPWRSAVE_MODE)); 3703 return 0; 3704 } else 3705 return sc->sc_recv_action(ni, wh, frm, efrm); 3706 } 3707 3708 static int 3709 mwl_addba_request(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap, 3710 int dialogtoken, int baparamset, int batimeout) 3711 { 3712 struct mwl_softc *sc = ni->ni_ic->ic_ifp->if_softc; 3713 struct ieee80211vap *vap = ni->ni_vap; 3714 struct mwl_node *mn = MWL_NODE(ni); 3715 struct mwl_bastate *bas; 3716 3717 bas = tap->txa_private; 3718 if (bas == NULL) { 3719 const MWL_HAL_BASTREAM *sp; 3720 /* 3721 * Check for a free BA stream slot. 3722 */ 3723 #if MWL_MAXBA > 3 3724 if (mn->mn_ba[3].bastream == NULL) 3725 bas = &mn->mn_ba[3]; 3726 else 3727 #endif 3728 #if MWL_MAXBA > 2 3729 if (mn->mn_ba[2].bastream == NULL) 3730 bas = &mn->mn_ba[2]; 3731 else 3732 #endif 3733 #if MWL_MAXBA > 1 3734 if (mn->mn_ba[1].bastream == NULL) 3735 bas = &mn->mn_ba[1]; 3736 else 3737 #endif 3738 #if MWL_MAXBA > 0 3739 if (mn->mn_ba[0].bastream == NULL) 3740 bas = &mn->mn_ba[0]; 3741 else 3742 #endif 3743 { 3744 /* sta already has max BA streams */ 3745 /* XXX assign BA stream to highest priority tid */ 3746 DPRINTF(sc, MWL_DEBUG_AMPDU, 3747 "%s: already has max bastreams\n", __func__); 3748 sc->sc_stats.mst_ampdu_reject++; 3749 return 0; 3750 } 3751 /* NB: no held reference to ni */ 3752 sp = mwl_hal_bastream_alloc(MWL_VAP(vap)->mv_hvap, 3753 (baparamset & IEEE80211_BAPS_POLICY_IMMEDIATE) != 0, 3754 ni->ni_macaddr, tap->txa_tid, ni->ni_htparam, 3755 ni, tap); 3756 if (sp == NULL) { 3757 /* 3758 * No available stream, return 0 so no 3759 * a-mpdu aggregation will be done. 3760 */ 3761 DPRINTF(sc, MWL_DEBUG_AMPDU, 3762 "%s: no bastream available\n", __func__); 3763 sc->sc_stats.mst_ampdu_nostream++; 3764 return 0; 3765 } 3766 DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: alloc bastream %p\n", 3767 __func__, sp); 3768 /* NB: qos is left zero so we won't match in mwl_tx_start */ 3769 bas->bastream = sp; 3770 tap->txa_private = bas; 3771 } 3772 /* fetch current seq# from the firmware; if available */ 3773 if (mwl_hal_bastream_get_seqno(sc->sc_mh, bas->bastream, 3774 vap->iv_opmode == IEEE80211_M_STA ? vap->iv_myaddr : ni->ni_macaddr, 3775 &tap->txa_start) != 0) 3776 tap->txa_start = 0; 3777 return sc->sc_addba_request(ni, tap, dialogtoken, baparamset, batimeout); 3778 } 3779 3780 static int 3781 mwl_addba_response(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap, 3782 int code, int baparamset, int batimeout) 3783 { 3784 struct mwl_softc *sc = ni->ni_ic->ic_ifp->if_softc; 3785 struct mwl_bastate *bas; 3786 3787 bas = tap->txa_private; 3788 if (bas == NULL) { 3789 /* XXX should not happen */ 3790 DPRINTF(sc, MWL_DEBUG_AMPDU, 3791 "%s: no BA stream allocated, TID %d\n", 3792 __func__, tap->txa_tid); 3793 sc->sc_stats.mst_addba_nostream++; 3794 return 0; 3795 } 3796 if (code == IEEE80211_STATUS_SUCCESS) { 3797 struct ieee80211vap *vap = ni->ni_vap; 3798 int bufsiz, error; 3799 3800 /* 3801 * Tell the firmware to setup the BA stream; 3802 * we know resources are available because we 3803 * pre-allocated one before forming the request. 3804 */ 3805 bufsiz = MS(baparamset, IEEE80211_BAPS_BUFSIZ); 3806 if (bufsiz == 0) 3807 bufsiz = IEEE80211_AGGR_BAWMAX; 3808 error = mwl_hal_bastream_create(MWL_VAP(vap)->mv_hvap, 3809 bas->bastream, bufsiz, bufsiz, tap->txa_start); 3810 if (error != 0) { 3811 /* 3812 * Setup failed, return immediately so no a-mpdu 3813 * aggregation will be done. 3814 */ 3815 mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream); 3816 mwl_bastream_free(bas); 3817 tap->txa_private = NULL; 3818 3819 DPRINTF(sc, MWL_DEBUG_AMPDU, 3820 "%s: create failed, error %d, bufsiz %d TID %d " 3821 "htparam 0x%x\n", __func__, error, bufsiz, 3822 tap->txa_tid, ni->ni_htparam); 3823 sc->sc_stats.mst_bacreate_failed++; 3824 return 0; 3825 } 3826 /* NB: cache txq to avoid ptr indirect */ 3827 mwl_bastream_setup(bas, tap->txa_tid, bas->bastream->txq); 3828 DPRINTF(sc, MWL_DEBUG_AMPDU, 3829 "%s: bastream %p assigned to txq %d TID %d bufsiz %d " 3830 "htparam 0x%x\n", __func__, bas->bastream, 3831 bas->txq, tap->txa_tid, bufsiz, ni->ni_htparam); 3832 } else { 3833 /* 3834 * Other side NAK'd us; return the resources. 3835 */ 3836 DPRINTF(sc, MWL_DEBUG_AMPDU, 3837 "%s: request failed with code %d, destroy bastream %p\n", 3838 __func__, code, bas->bastream); 3839 mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream); 3840 mwl_bastream_free(bas); 3841 tap->txa_private = NULL; 3842 } 3843 /* NB: firmware sends BAR so we don't need to */ 3844 return sc->sc_addba_response(ni, tap, code, baparamset, batimeout); 3845 } 3846 3847 static void 3848 mwl_addba_stop(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap) 3849 { 3850 struct mwl_softc *sc = ni->ni_ic->ic_ifp->if_softc; 3851 struct mwl_bastate *bas; 3852 3853 bas = tap->txa_private; 3854 if (bas != NULL) { 3855 DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: destroy bastream %p\n", 3856 __func__, bas->bastream); 3857 mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream); 3858 mwl_bastream_free(bas); 3859 tap->txa_private = NULL; 3860 } 3861 sc->sc_addba_stop(ni, tap); 3862 } 3863 3864 /* 3865 * Setup the rx data structures. This should only be 3866 * done once or we may get out of sync with the firmware. 3867 */ 3868 static int 3869 mwl_startrecv(struct mwl_softc *sc) 3870 { 3871 if (!sc->sc_recvsetup) { 3872 struct mwl_rxbuf *bf, *prev; 3873 struct mwl_rxdesc *ds; 3874 3875 prev = NULL; 3876 STAILQ_FOREACH(bf, &sc->sc_rxbuf, bf_list) { 3877 int error = mwl_rxbuf_init(sc, bf); 3878 if (error != 0) { 3879 DPRINTF(sc, MWL_DEBUG_RECV, 3880 "%s: mwl_rxbuf_init failed %d\n", 3881 __func__, error); 3882 return error; 3883 } 3884 if (prev != NULL) { 3885 ds = prev->bf_desc; 3886 ds->pPhysNext = htole32(bf->bf_daddr); 3887 } 3888 prev = bf; 3889 } 3890 if (prev != NULL) { 3891 ds = prev->bf_desc; 3892 ds->pPhysNext = 3893 htole32(STAILQ_FIRST(&sc->sc_rxbuf)->bf_daddr); 3894 } 3895 sc->sc_recvsetup = 1; 3896 } 3897 mwl_mode_init(sc); /* set filters, etc. */ 3898 return 0; 3899 } 3900 3901 static MWL_HAL_APMODE 3902 mwl_getapmode(const struct ieee80211vap *vap, struct ieee80211_channel *chan) 3903 { 3904 MWL_HAL_APMODE mode; 3905 3906 if (IEEE80211_IS_CHAN_HT(chan)) { 3907 if (vap->iv_flags_ht & IEEE80211_FHT_PUREN) 3908 mode = AP_MODE_N_ONLY; 3909 else if (IEEE80211_IS_CHAN_5GHZ(chan)) 3910 mode = AP_MODE_AandN; 3911 else if (vap->iv_flags & IEEE80211_F_PUREG) 3912 mode = AP_MODE_GandN; 3913 else 3914 mode = AP_MODE_BandGandN; 3915 } else if (IEEE80211_IS_CHAN_ANYG(chan)) { 3916 if (vap->iv_flags & IEEE80211_F_PUREG) 3917 mode = AP_MODE_G_ONLY; 3918 else 3919 mode = AP_MODE_MIXED; 3920 } else if (IEEE80211_IS_CHAN_B(chan)) 3921 mode = AP_MODE_B_ONLY; 3922 else if (IEEE80211_IS_CHAN_A(chan)) 3923 mode = AP_MODE_A_ONLY; 3924 else 3925 mode = AP_MODE_MIXED; /* XXX should not happen? */ 3926 return mode; 3927 } 3928 3929 static int 3930 mwl_setapmode(struct ieee80211vap *vap, struct ieee80211_channel *chan) 3931 { 3932 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap; 3933 return mwl_hal_setapmode(hvap, mwl_getapmode(vap, chan)); 3934 } 3935 3936 /* 3937 * Set/change channels. 3938 */ 3939 static int 3940 mwl_chan_set(struct mwl_softc *sc, struct ieee80211_channel *chan) 3941 { 3942 struct mwl_hal *mh = sc->sc_mh; 3943 struct ifnet *ifp = sc->sc_ifp; 3944 struct ieee80211com *ic = ifp->if_l2com; 3945 MWL_HAL_CHANNEL hchan; 3946 int maxtxpow; 3947 3948 DPRINTF(sc, MWL_DEBUG_RESET, "%s: chan %u MHz/flags 0x%x\n", 3949 __func__, chan->ic_freq, chan->ic_flags); 3950 3951 /* 3952 * Convert to a HAL channel description with 3953 * the flags constrained to reflect the current 3954 * operating mode. 3955 */ 3956 mwl_mapchan(&hchan, chan); 3957 mwl_hal_intrset(mh, 0); /* disable interrupts */ 3958 #if 0 3959 mwl_draintxq(sc); /* clear pending tx frames */ 3960 #endif 3961 mwl_hal_setchannel(mh, &hchan); 3962 /* 3963 * Tx power is cap'd by the regulatory setting and 3964 * possibly a user-set limit. We pass the min of 3965 * these to the hal to apply them to the cal data 3966 * for this channel. 3967 * XXX min bound? 3968 */ 3969 maxtxpow = 2*chan->ic_maxregpower; 3970 if (maxtxpow > ic->ic_txpowlimit) 3971 maxtxpow = ic->ic_txpowlimit; 3972 mwl_hal_settxpower(mh, &hchan, maxtxpow / 2); 3973 /* NB: potentially change mcast/mgt rates */ 3974 mwl_setcurchanrates(sc); 3975 3976 /* 3977 * Update internal state. 3978 */ 3979 sc->sc_tx_th.wt_chan_freq = htole16(chan->ic_freq); 3980 sc->sc_rx_th.wr_chan_freq = htole16(chan->ic_freq); 3981 if (IEEE80211_IS_CHAN_A(chan)) { 3982 sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_A); 3983 sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_A); 3984 } else if (IEEE80211_IS_CHAN_ANYG(chan)) { 3985 sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_G); 3986 sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_G); 3987 } else { 3988 sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_B); 3989 sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_B); 3990 } 3991 sc->sc_curchan = hchan; 3992 mwl_hal_intrset(mh, sc->sc_imask); 3993 3994 return 0; 3995 } 3996 3997 static void 3998 mwl_scan_start(struct ieee80211com *ic) 3999 { 4000 struct ifnet *ifp = ic->ic_ifp; 4001 struct mwl_softc *sc = ifp->if_softc; 4002 4003 DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__); 4004 } 4005 4006 static void 4007 mwl_scan_end(struct ieee80211com *ic) 4008 { 4009 struct ifnet *ifp = ic->ic_ifp; 4010 struct mwl_softc *sc = ifp->if_softc; 4011 4012 DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__); 4013 } 4014 4015 static void 4016 mwl_set_channel(struct ieee80211com *ic) 4017 { 4018 struct ifnet *ifp = ic->ic_ifp; 4019 struct mwl_softc *sc = ifp->if_softc; 4020 4021 (void) mwl_chan_set(sc, ic->ic_curchan); 4022 } 4023 4024 /* 4025 * Handle a channel switch request. We inform the firmware 4026 * and mark the global state to suppress various actions. 4027 * NB: we issue only one request to the fw; we may be called 4028 * multiple times if there are multiple vap's. 4029 */ 4030 static void 4031 mwl_startcsa(struct ieee80211vap *vap) 4032 { 4033 struct ieee80211com *ic = vap->iv_ic; 4034 struct mwl_softc *sc = ic->ic_ifp->if_softc; 4035 MWL_HAL_CHANNEL hchan; 4036 4037 if (sc->sc_csapending) 4038 return; 4039 4040 mwl_mapchan(&hchan, ic->ic_csa_newchan); 4041 /* 1 =>'s quiet channel */ 4042 mwl_hal_setchannelswitchie(sc->sc_mh, &hchan, 1, ic->ic_csa_count); 4043 sc->sc_csapending = 1; 4044 } 4045 4046 /* 4047 * Plumb any static WEP key for the station. This is 4048 * necessary as we must propagate the key from the 4049 * global key table of the vap to each sta db entry. 4050 */ 4051 static void 4052 mwl_setanywepkey(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN]) 4053 { 4054 if ((vap->iv_flags & (IEEE80211_F_PRIVACY|IEEE80211_F_WPA)) == 4055 IEEE80211_F_PRIVACY && 4056 vap->iv_def_txkey != IEEE80211_KEYIX_NONE && 4057 vap->iv_nw_keys[vap->iv_def_txkey].wk_keyix != IEEE80211_KEYIX_NONE) 4058 (void) mwl_key_set(vap, &vap->iv_nw_keys[vap->iv_def_txkey], mac); 4059 } 4060 4061 static int 4062 mwl_peerstadb(struct ieee80211_node *ni, int aid, int staid, MWL_HAL_PEERINFO *pi) 4063 { 4064 #define WME(ie) ((const struct ieee80211_wme_info *) ie) 4065 struct ieee80211vap *vap = ni->ni_vap; 4066 struct mwl_hal_vap *hvap; 4067 int error; 4068 4069 if (vap->iv_opmode == IEEE80211_M_WDS) { 4070 /* 4071 * WDS vap's do not have a f/w vap; instead they piggyback 4072 * on an AP vap and we must install the sta db entry and 4073 * crypto state using that AP's handle (the WDS vap has none). 4074 */ 4075 hvap = MWL_VAP(vap)->mv_ap_hvap; 4076 } else 4077 hvap = MWL_VAP(vap)->mv_hvap; 4078 error = mwl_hal_newstation(hvap, ni->ni_macaddr, 4079 aid, staid, pi, 4080 ni->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT), 4081 ni->ni_ies.wme_ie != NULL ? WME(ni->ni_ies.wme_ie)->wme_info : 0); 4082 if (error == 0) { 4083 /* 4084 * Setup security for this station. For sta mode this is 4085 * needed even though do the same thing on transition to 4086 * AUTH state because the call to mwl_hal_newstation 4087 * clobbers the crypto state we setup. 4088 */ 4089 mwl_setanywepkey(vap, ni->ni_macaddr); 4090 } 4091 return error; 4092 #undef WME 4093 } 4094 4095 static void 4096 mwl_setglobalkeys(struct ieee80211vap *vap) 4097 { 4098 struct ieee80211_key *wk; 4099 4100 wk = &vap->iv_nw_keys[0]; 4101 for (; wk < &vap->iv_nw_keys[IEEE80211_WEP_NKID]; wk++) 4102 if (wk->wk_keyix != IEEE80211_KEYIX_NONE) 4103 (void) mwl_key_set(vap, wk, vap->iv_myaddr); 4104 } 4105 4106 /* 4107 * Convert a legacy rate set to a firmware bitmask. 4108 */ 4109 static uint32_t 4110 get_rate_bitmap(const struct ieee80211_rateset *rs) 4111 { 4112 uint32_t rates; 4113 int i; 4114 4115 rates = 0; 4116 for (i = 0; i < rs->rs_nrates; i++) 4117 switch (rs->rs_rates[i] & IEEE80211_RATE_VAL) { 4118 case 2: rates |= 0x001; break; 4119 case 4: rates |= 0x002; break; 4120 case 11: rates |= 0x004; break; 4121 case 22: rates |= 0x008; break; 4122 case 44: rates |= 0x010; break; 4123 case 12: rates |= 0x020; break; 4124 case 18: rates |= 0x040; break; 4125 case 24: rates |= 0x080; break; 4126 case 36: rates |= 0x100; break; 4127 case 48: rates |= 0x200; break; 4128 case 72: rates |= 0x400; break; 4129 case 96: rates |= 0x800; break; 4130 case 108: rates |= 0x1000; break; 4131 } 4132 return rates; 4133 } 4134 4135 /* 4136 * Construct an HT firmware bitmask from an HT rate set. 4137 */ 4138 static uint32_t 4139 get_htrate_bitmap(const struct ieee80211_htrateset *rs) 4140 { 4141 uint32_t rates; 4142 int i; 4143 4144 rates = 0; 4145 for (i = 0; i < rs->rs_nrates; i++) { 4146 if (rs->rs_rates[i] < 16) 4147 rates |= 1<<rs->rs_rates[i]; 4148 } 4149 return rates; 4150 } 4151 4152 /* 4153 * Craft station database entry for station. 4154 * NB: use host byte order here, the hal handles byte swapping. 4155 */ 4156 static MWL_HAL_PEERINFO * 4157 mkpeerinfo(MWL_HAL_PEERINFO *pi, const struct ieee80211_node *ni) 4158 { 4159 const struct ieee80211vap *vap = ni->ni_vap; 4160 4161 memset(pi, 0, sizeof(*pi)); 4162 pi->LegacyRateBitMap = get_rate_bitmap(&ni->ni_rates); 4163 pi->CapInfo = ni->ni_capinfo; 4164 if (ni->ni_flags & IEEE80211_NODE_HT) { 4165 /* HT capabilities, etc */ 4166 pi->HTCapabilitiesInfo = ni->ni_htcap; 4167 /* XXX pi.HTCapabilitiesInfo */ 4168 pi->MacHTParamInfo = ni->ni_htparam; 4169 pi->HTRateBitMap = get_htrate_bitmap(&ni->ni_htrates); 4170 pi->AddHtInfo.ControlChan = ni->ni_htctlchan; 4171 pi->AddHtInfo.AddChan = ni->ni_ht2ndchan; 4172 pi->AddHtInfo.OpMode = ni->ni_htopmode; 4173 pi->AddHtInfo.stbc = ni->ni_htstbc; 4174 4175 /* constrain according to local configuration */ 4176 if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI40) == 0) 4177 pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI40; 4178 if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI20) == 0) 4179 pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI20; 4180 if (ni->ni_chw != 40) 4181 pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_CHWIDTH40; 4182 } 4183 return pi; 4184 } 4185 4186 /* 4187 * Re-create the local sta db entry for a vap to ensure 4188 * up to date WME state is pushed to the firmware. Because 4189 * this resets crypto state this must be followed by a 4190 * reload of any keys in the global key table. 4191 */ 4192 static int 4193 mwl_localstadb(struct ieee80211vap *vap) 4194 { 4195 #define WME(ie) ((const struct ieee80211_wme_info *) ie) 4196 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap; 4197 struct ieee80211_node *bss; 4198 MWL_HAL_PEERINFO pi; 4199 int error; 4200 4201 switch (vap->iv_opmode) { 4202 case IEEE80211_M_STA: 4203 bss = vap->iv_bss; 4204 error = mwl_hal_newstation(hvap, vap->iv_myaddr, 0, 0, 4205 vap->iv_state == IEEE80211_S_RUN ? 4206 mkpeerinfo(&pi, bss) : NULL, 4207 (bss->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT)), 4208 bss->ni_ies.wme_ie != NULL ? 4209 WME(bss->ni_ies.wme_ie)->wme_info : 0); 4210 if (error == 0) 4211 mwl_setglobalkeys(vap); 4212 break; 4213 case IEEE80211_M_HOSTAP: 4214 case IEEE80211_M_MBSS: 4215 error = mwl_hal_newstation(hvap, vap->iv_myaddr, 4216 0, 0, NULL, vap->iv_flags & IEEE80211_F_WME, 0); 4217 if (error == 0) 4218 mwl_setglobalkeys(vap); 4219 break; 4220 default: 4221 error = 0; 4222 break; 4223 } 4224 return error; 4225 #undef WME 4226 } 4227 4228 static int 4229 mwl_newstate(struct ieee80211vap *vap, enum ieee80211_state nstate, int arg) 4230 { 4231 struct mwl_vap *mvp = MWL_VAP(vap); 4232 struct mwl_hal_vap *hvap = mvp->mv_hvap; 4233 struct ieee80211com *ic = vap->iv_ic; 4234 struct ieee80211_node *ni = NULL; 4235 struct ifnet *ifp = ic->ic_ifp; 4236 struct mwl_softc *sc = ifp->if_softc; 4237 struct mwl_hal *mh = sc->sc_mh; 4238 enum ieee80211_state ostate = vap->iv_state; 4239 int error; 4240 4241 DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: %s -> %s\n", 4242 vap->iv_ifp->if_xname, __func__, 4243 ieee80211_state_name[ostate], ieee80211_state_name[nstate]); 4244 4245 callout_stop(&sc->sc_timer); 4246 /* 4247 * Clear current radar detection state. 4248 */ 4249 if (ostate == IEEE80211_S_CAC) { 4250 /* stop quiet mode radar detection */ 4251 mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_STOP); 4252 } else if (sc->sc_radarena) { 4253 /* stop in-service radar detection */ 4254 mwl_hal_setradardetection(mh, DR_DFS_DISABLE); 4255 sc->sc_radarena = 0; 4256 } 4257 /* 4258 * Carry out per-state actions before doing net80211 work. 4259 */ 4260 if (nstate == IEEE80211_S_INIT) { 4261 /* NB: only ap+sta vap's have a fw entity */ 4262 if (hvap != NULL) 4263 mwl_hal_stop(hvap); 4264 } else if (nstate == IEEE80211_S_SCAN) { 4265 mwl_hal_start(hvap); 4266 /* NB: this disables beacon frames */ 4267 mwl_hal_setinframode(hvap); 4268 } else if (nstate == IEEE80211_S_AUTH) { 4269 /* 4270 * Must create a sta db entry in case a WEP key needs to 4271 * be plumbed. This entry will be overwritten if we 4272 * associate; otherwise it will be reclaimed on node free. 4273 */ 4274 ni = vap->iv_bss; 4275 MWL_NODE(ni)->mn_hvap = hvap; 4276 (void) mwl_peerstadb(ni, 0, 0, NULL); 4277 } else if (nstate == IEEE80211_S_CSA) { 4278 /* XXX move to below? */ 4279 if (vap->iv_opmode == IEEE80211_M_HOSTAP || 4280 vap->iv_opmode == IEEE80211_M_MBSS) 4281 mwl_startcsa(vap); 4282 } else if (nstate == IEEE80211_S_CAC) { 4283 /* XXX move to below? */ 4284 /* stop ap xmit and enable quiet mode radar detection */ 4285 mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_START); 4286 } 4287 4288 /* 4289 * Invoke the parent method to do net80211 work. 4290 */ 4291 error = mvp->mv_newstate(vap, nstate, arg); 4292 4293 /* 4294 * Carry out work that must be done after net80211 runs; 4295 * this work requires up to date state (e.g. iv_bss). 4296 */ 4297 if (error == 0 && nstate == IEEE80211_S_RUN) { 4298 /* NB: collect bss node again, it may have changed */ 4299 ni = vap->iv_bss; 4300 4301 DPRINTF(sc, MWL_DEBUG_STATE, 4302 "%s: %s(RUN): iv_flags 0x%08x bintvl %d bssid %s " 4303 "capinfo 0x%04x chan %d\n", 4304 vap->iv_ifp->if_xname, __func__, vap->iv_flags, 4305 ni->ni_intval, ether_sprintf(ni->ni_bssid), ni->ni_capinfo, 4306 ieee80211_chan2ieee(ic, ic->ic_curchan)); 4307 4308 /* 4309 * Recreate local sta db entry to update WME/HT state. 4310 */ 4311 mwl_localstadb(vap); 4312 switch (vap->iv_opmode) { 4313 case IEEE80211_M_HOSTAP: 4314 case IEEE80211_M_MBSS: 4315 if (ostate == IEEE80211_S_CAC) { 4316 /* enable in-service radar detection */ 4317 mwl_hal_setradardetection(mh, 4318 DR_IN_SERVICE_MONITOR_START); 4319 sc->sc_radarena = 1; 4320 } 4321 /* 4322 * Allocate and setup the beacon frame 4323 * (and related state). 4324 */ 4325 error = mwl_reset_vap(vap, IEEE80211_S_RUN); 4326 if (error != 0) { 4327 DPRINTF(sc, MWL_DEBUG_STATE, 4328 "%s: beacon setup failed, error %d\n", 4329 __func__, error); 4330 goto bad; 4331 } 4332 /* NB: must be after setting up beacon */ 4333 mwl_hal_start(hvap); 4334 break; 4335 case IEEE80211_M_STA: 4336 DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: aid 0x%x\n", 4337 vap->iv_ifp->if_xname, __func__, ni->ni_associd); 4338 /* 4339 * Set state now that we're associated. 4340 */ 4341 mwl_hal_setassocid(hvap, ni->ni_bssid, ni->ni_associd); 4342 mwl_setrates(vap); 4343 mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold); 4344 if ((vap->iv_flags & IEEE80211_F_DWDS) && 4345 sc->sc_ndwdsvaps++ == 0) 4346 mwl_hal_setdwds(mh, 1); 4347 break; 4348 case IEEE80211_M_WDS: 4349 DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: bssid %s\n", 4350 vap->iv_ifp->if_xname, __func__, 4351 ether_sprintf(ni->ni_bssid)); 4352 mwl_seteapolformat(vap); 4353 break; 4354 default: 4355 break; 4356 } 4357 /* 4358 * Set CS mode according to operating channel; 4359 * this mostly an optimization for 5GHz. 4360 * 4361 * NB: must follow mwl_hal_start which resets csmode 4362 */ 4363 if (IEEE80211_IS_CHAN_5GHZ(ic->ic_bsschan)) 4364 mwl_hal_setcsmode(mh, CSMODE_AGGRESSIVE); 4365 else 4366 mwl_hal_setcsmode(mh, CSMODE_AUTO_ENA); 4367 /* 4368 * Start timer to prod firmware. 4369 */ 4370 if (sc->sc_ageinterval != 0) 4371 callout_reset(&sc->sc_timer, sc->sc_ageinterval*hz, 4372 mwl_agestations, sc); 4373 } else if (nstate == IEEE80211_S_SLEEP) { 4374 /* XXX set chip in power save */ 4375 } else if ((vap->iv_flags & IEEE80211_F_DWDS) && 4376 --sc->sc_ndwdsvaps == 0) 4377 mwl_hal_setdwds(mh, 0); 4378 bad: 4379 return error; 4380 } 4381 4382 /* 4383 * Manage station id's; these are separate from AID's 4384 * as AID's may have values out of the range of possible 4385 * station id's acceptable to the firmware. 4386 */ 4387 static int 4388 allocstaid(struct mwl_softc *sc, int aid) 4389 { 4390 int staid; 4391 4392 if (!(0 < aid && aid < MWL_MAXSTAID) || isset(sc->sc_staid, aid)) { 4393 /* NB: don't use 0 */ 4394 for (staid = 1; staid < MWL_MAXSTAID; staid++) 4395 if (isclr(sc->sc_staid, staid)) 4396 break; 4397 } else 4398 staid = aid; 4399 setbit(sc->sc_staid, staid); 4400 return staid; 4401 } 4402 4403 static void 4404 delstaid(struct mwl_softc *sc, int staid) 4405 { 4406 clrbit(sc->sc_staid, staid); 4407 } 4408 4409 /* 4410 * Setup driver-specific state for a newly associated node. 4411 * Note that we're called also on a re-associate, the isnew 4412 * param tells us if this is the first time or not. 4413 */ 4414 static void 4415 mwl_newassoc(struct ieee80211_node *ni, int isnew) 4416 { 4417 struct ieee80211vap *vap = ni->ni_vap; 4418 struct mwl_softc *sc = vap->iv_ic->ic_ifp->if_softc; 4419 struct mwl_node *mn = MWL_NODE(ni); 4420 MWL_HAL_PEERINFO pi; 4421 uint16_t aid; 4422 int error; 4423 4424 aid = IEEE80211_AID(ni->ni_associd); 4425 if (isnew) { 4426 mn->mn_staid = allocstaid(sc, aid); 4427 mn->mn_hvap = MWL_VAP(vap)->mv_hvap; 4428 } else { 4429 mn = MWL_NODE(ni); 4430 /* XXX reset BA stream? */ 4431 } 4432 DPRINTF(sc, MWL_DEBUG_NODE, "%s: mac %s isnew %d aid %d staid %d\n", 4433 __func__, ether_sprintf(ni->ni_macaddr), isnew, aid, mn->mn_staid); 4434 error = mwl_peerstadb(ni, aid, mn->mn_staid, mkpeerinfo(&pi, ni)); 4435 if (error != 0) { 4436 DPRINTF(sc, MWL_DEBUG_NODE, 4437 "%s: error %d creating sta db entry\n", 4438 __func__, error); 4439 /* XXX how to deal with error? */ 4440 } 4441 } 4442 4443 /* 4444 * Periodically poke the firmware to age out station state 4445 * (power save queues, pending tx aggregates). 4446 */ 4447 static void 4448 mwl_agestations(void *arg) 4449 { 4450 struct mwl_softc *sc = arg; 4451 4452 mwl_hal_setkeepalive(sc->sc_mh); 4453 if (sc->sc_ageinterval != 0) /* NB: catch dynamic changes */ 4454 callout_schedule(&sc->sc_timer, sc->sc_ageinterval*hz); 4455 } 4456 4457 static const struct mwl_hal_channel * 4458 findhalchannel(const MWL_HAL_CHANNELINFO *ci, int ieee) 4459 { 4460 int i; 4461 4462 for (i = 0; i < ci->nchannels; i++) { 4463 const struct mwl_hal_channel *hc = &ci->channels[i]; 4464 if (hc->ieee == ieee) 4465 return hc; 4466 } 4467 return NULL; 4468 } 4469 4470 static int 4471 mwl_setregdomain(struct ieee80211com *ic, struct ieee80211_regdomain *rd, 4472 int nchan, struct ieee80211_channel chans[]) 4473 { 4474 struct mwl_softc *sc = ic->ic_ifp->if_softc; 4475 struct mwl_hal *mh = sc->sc_mh; 4476 const MWL_HAL_CHANNELINFO *ci; 4477 int i; 4478 4479 for (i = 0; i < nchan; i++) { 4480 struct ieee80211_channel *c = &chans[i]; 4481 const struct mwl_hal_channel *hc; 4482 4483 if (IEEE80211_IS_CHAN_2GHZ(c)) { 4484 mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_2DOT4GHZ, 4485 IEEE80211_IS_CHAN_HT40(c) ? 4486 MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci); 4487 } else if (IEEE80211_IS_CHAN_5GHZ(c)) { 4488 mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_5GHZ, 4489 IEEE80211_IS_CHAN_HT40(c) ? 4490 MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci); 4491 } else { 4492 if_printf(ic->ic_ifp, 4493 "%s: channel %u freq %u/0x%x not 2.4/5GHz\n", 4494 __func__, c->ic_ieee, c->ic_freq, c->ic_flags); 4495 return EINVAL; 4496 } 4497 /* 4498 * Verify channel has cal data and cap tx power. 4499 */ 4500 hc = findhalchannel(ci, c->ic_ieee); 4501 if (hc != NULL) { 4502 if (c->ic_maxpower > 2*hc->maxTxPow) 4503 c->ic_maxpower = 2*hc->maxTxPow; 4504 goto next; 4505 } 4506 if (IEEE80211_IS_CHAN_HT40(c)) { 4507 /* 4508 * Look for the extension channel since the 4509 * hal table only has the primary channel. 4510 */ 4511 hc = findhalchannel(ci, c->ic_extieee); 4512 if (hc != NULL) { 4513 if (c->ic_maxpower > 2*hc->maxTxPow) 4514 c->ic_maxpower = 2*hc->maxTxPow; 4515 goto next; 4516 } 4517 } 4518 if_printf(ic->ic_ifp, 4519 "%s: no cal data for channel %u ext %u freq %u/0x%x\n", 4520 __func__, c->ic_ieee, c->ic_extieee, 4521 c->ic_freq, c->ic_flags); 4522 return EINVAL; 4523 next: 4524 ; 4525 } 4526 return 0; 4527 } 4528 4529 #define IEEE80211_CHAN_HTG (IEEE80211_CHAN_HT|IEEE80211_CHAN_G) 4530 #define IEEE80211_CHAN_HTA (IEEE80211_CHAN_HT|IEEE80211_CHAN_A) 4531 4532 static void 4533 addchan(struct ieee80211_channel *c, int freq, int flags, int ieee, int txpow) 4534 { 4535 c->ic_freq = freq; 4536 c->ic_flags = flags; 4537 c->ic_ieee = ieee; 4538 c->ic_minpower = 0; 4539 c->ic_maxpower = 2*txpow; 4540 c->ic_maxregpower = txpow; 4541 } 4542 4543 static const struct ieee80211_channel * 4544 findchannel(const struct ieee80211_channel chans[], int nchans, 4545 int freq, int flags) 4546 { 4547 const struct ieee80211_channel *c; 4548 int i; 4549 4550 for (i = 0; i < nchans; i++) { 4551 c = &chans[i]; 4552 if (c->ic_freq == freq && c->ic_flags == flags) 4553 return c; 4554 } 4555 return NULL; 4556 } 4557 4558 static void 4559 addht40channels(struct ieee80211_channel chans[], int maxchans, int *nchans, 4560 const MWL_HAL_CHANNELINFO *ci, int flags) 4561 { 4562 struct ieee80211_channel *c; 4563 const struct ieee80211_channel *extc; 4564 const struct mwl_hal_channel *hc; 4565 int i; 4566 4567 c = &chans[*nchans]; 4568 4569 flags &= ~IEEE80211_CHAN_HT; 4570 for (i = 0; i < ci->nchannels; i++) { 4571 /* 4572 * Each entry defines an HT40 channel pair; find the 4573 * extension channel above and the insert the pair. 4574 */ 4575 hc = &ci->channels[i]; 4576 extc = findchannel(chans, *nchans, hc->freq+20, 4577 flags | IEEE80211_CHAN_HT20); 4578 if (extc != NULL) { 4579 if (*nchans >= maxchans) 4580 break; 4581 addchan(c, hc->freq, flags | IEEE80211_CHAN_HT40U, 4582 hc->ieee, hc->maxTxPow); 4583 c->ic_extieee = extc->ic_ieee; 4584 c++, (*nchans)++; 4585 if (*nchans >= maxchans) 4586 break; 4587 addchan(c, extc->ic_freq, flags | IEEE80211_CHAN_HT40D, 4588 extc->ic_ieee, hc->maxTxPow); 4589 c->ic_extieee = hc->ieee; 4590 c++, (*nchans)++; 4591 } 4592 } 4593 } 4594 4595 static void 4596 addchannels(struct ieee80211_channel chans[], int maxchans, int *nchans, 4597 const MWL_HAL_CHANNELINFO *ci, int flags) 4598 { 4599 struct ieee80211_channel *c; 4600 int i; 4601 4602 c = &chans[*nchans]; 4603 4604 for (i = 0; i < ci->nchannels; i++) { 4605 const struct mwl_hal_channel *hc; 4606 4607 hc = &ci->channels[i]; 4608 if (*nchans >= maxchans) 4609 break; 4610 addchan(c, hc->freq, flags, hc->ieee, hc->maxTxPow); 4611 c++, (*nchans)++; 4612 if (flags == IEEE80211_CHAN_G || flags == IEEE80211_CHAN_HTG) { 4613 /* g channel have a separate b-only entry */ 4614 if (*nchans >= maxchans) 4615 break; 4616 c[0] = c[-1]; 4617 c[-1].ic_flags = IEEE80211_CHAN_B; 4618 c++, (*nchans)++; 4619 } 4620 if (flags == IEEE80211_CHAN_HTG) { 4621 /* HT g channel have a separate g-only entry */ 4622 if (*nchans >= maxchans) 4623 break; 4624 c[-1].ic_flags = IEEE80211_CHAN_G; 4625 c[0] = c[-1]; 4626 c[0].ic_flags &= ~IEEE80211_CHAN_HT; 4627 c[0].ic_flags |= IEEE80211_CHAN_HT20; /* HT20 */ 4628 c++, (*nchans)++; 4629 } 4630 if (flags == IEEE80211_CHAN_HTA) { 4631 /* HT a channel have a separate a-only entry */ 4632 if (*nchans >= maxchans) 4633 break; 4634 c[-1].ic_flags = IEEE80211_CHAN_A; 4635 c[0] = c[-1]; 4636 c[0].ic_flags &= ~IEEE80211_CHAN_HT; 4637 c[0].ic_flags |= IEEE80211_CHAN_HT20; /* HT20 */ 4638 c++, (*nchans)++; 4639 } 4640 } 4641 } 4642 4643 static void 4644 getchannels(struct mwl_softc *sc, int maxchans, int *nchans, 4645 struct ieee80211_channel chans[]) 4646 { 4647 const MWL_HAL_CHANNELINFO *ci; 4648 4649 /* 4650 * Use the channel info from the hal to craft the 4651 * channel list. Note that we pass back an unsorted 4652 * list; the caller is required to sort it for us 4653 * (if desired). 4654 */ 4655 *nchans = 0; 4656 if (mwl_hal_getchannelinfo(sc->sc_mh, 4657 MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0) 4658 addchannels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTG); 4659 if (mwl_hal_getchannelinfo(sc->sc_mh, 4660 MWL_FREQ_BAND_5GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0) 4661 addchannels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTA); 4662 if (mwl_hal_getchannelinfo(sc->sc_mh, 4663 MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0) 4664 addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTG); 4665 if (mwl_hal_getchannelinfo(sc->sc_mh, 4666 MWL_FREQ_BAND_5GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0) 4667 addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTA); 4668 } 4669 4670 static void 4671 mwl_getradiocaps(struct ieee80211com *ic, 4672 int maxchans, int *nchans, struct ieee80211_channel chans[]) 4673 { 4674 struct mwl_softc *sc = ic->ic_ifp->if_softc; 4675 4676 getchannels(sc, maxchans, nchans, chans); 4677 } 4678 4679 static int 4680 mwl_getchannels(struct mwl_softc *sc) 4681 { 4682 struct ifnet *ifp = sc->sc_ifp; 4683 struct ieee80211com *ic = ifp->if_l2com; 4684 4685 /* 4686 * Use the channel info from the hal to craft the 4687 * channel list for net80211. Note that we pass up 4688 * an unsorted list; net80211 will sort it for us. 4689 */ 4690 memset(ic->ic_channels, 0, sizeof(ic->ic_channels)); 4691 ic->ic_nchans = 0; 4692 getchannels(sc, IEEE80211_CHAN_MAX, &ic->ic_nchans, ic->ic_channels); 4693 4694 ic->ic_regdomain.regdomain = SKU_DEBUG; 4695 ic->ic_regdomain.country = CTRY_DEFAULT; 4696 ic->ic_regdomain.location = 'I'; 4697 ic->ic_regdomain.isocc[0] = ' '; /* XXX? */ 4698 ic->ic_regdomain.isocc[1] = ' '; 4699 return (ic->ic_nchans == 0 ? EIO : 0); 4700 } 4701 #undef IEEE80211_CHAN_HTA 4702 #undef IEEE80211_CHAN_HTG 4703 4704 #ifdef MWL_DEBUG 4705 static void 4706 mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix) 4707 { 4708 const struct mwl_rxdesc *ds = bf->bf_desc; 4709 uint32_t status = le32toh(ds->Status); 4710 4711 printf("R[%2u] (DS.V:%p DS.P:%p) NEXT:%08x DATA:%08x RC:%02x%s\n" 4712 " STAT:%02x LEN:%04x RSSI:%02x CHAN:%02x RATE:%02x QOS:%04x HT:%04x\n", 4713 ix, ds, (const struct mwl_desc *)bf->bf_daddr, 4714 le32toh(ds->pPhysNext), le32toh(ds->pPhysBuffData), 4715 ds->RxControl, 4716 ds->RxControl != EAGLE_RXD_CTRL_DRIVER_OWN ? 4717 "" : (status & EAGLE_RXD_STATUS_OK) ? " *" : " !", 4718 ds->Status, le16toh(ds->PktLen), ds->RSSI, ds->Channel, 4719 ds->Rate, le16toh(ds->QosCtrl), le16toh(ds->HtSig2)); 4720 } 4721 4722 static void 4723 mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix) 4724 { 4725 const struct mwl_txdesc *ds = bf->bf_desc; 4726 uint32_t status = le32toh(ds->Status); 4727 4728 printf("Q%u[%3u]", qnum, ix); 4729 printf(" (DS.V:%p DS.P:%p)\n", 4730 ds, (const struct mwl_txdesc *)bf->bf_daddr); 4731 printf(" NEXT:%08x DATA:%08x LEN:%04x STAT:%08x%s\n", 4732 le32toh(ds->pPhysNext), 4733 le32toh(ds->PktPtr), le16toh(ds->PktLen), status, 4734 status & EAGLE_TXD_STATUS_USED ? 4735 "" : (status & 3) != 0 ? " *" : " !"); 4736 printf(" RATE:%02x PRI:%x QOS:%04x SAP:%08x FORMAT:%04x\n", 4737 ds->DataRate, ds->TxPriority, le16toh(ds->QosCtrl), 4738 le32toh(ds->SapPktInfo), le16toh(ds->Format)); 4739 #if MWL_TXDESC > 1 4740 printf(" MULTIFRAMES:%u LEN:%04x %04x %04x %04x %04x %04x\n" 4741 , le32toh(ds->multiframes) 4742 , le16toh(ds->PktLenArray[0]), le16toh(ds->PktLenArray[1]) 4743 , le16toh(ds->PktLenArray[2]), le16toh(ds->PktLenArray[3]) 4744 , le16toh(ds->PktLenArray[4]), le16toh(ds->PktLenArray[5]) 4745 ); 4746 printf(" DATA:%08x %08x %08x %08x %08x %08x\n" 4747 , le32toh(ds->PktPtrArray[0]), le32toh(ds->PktPtrArray[1]) 4748 , le32toh(ds->PktPtrArray[2]), le32toh(ds->PktPtrArray[3]) 4749 , le32toh(ds->PktPtrArray[4]), le32toh(ds->PktPtrArray[5]) 4750 ); 4751 #endif 4752 #if 0 4753 { const uint8_t *cp = (const uint8_t *) ds; 4754 int i; 4755 for (i = 0; i < sizeof(struct mwl_txdesc); i++) { 4756 printf("%02x ", cp[i]); 4757 if (((i+1) % 16) == 0) 4758 printf("\n"); 4759 } 4760 printf("\n"); 4761 } 4762 #endif 4763 } 4764 #endif /* MWL_DEBUG */ 4765 4766 #if 0 4767 static void 4768 mwl_txq_dump(struct mwl_txq *txq) 4769 { 4770 struct mwl_txbuf *bf; 4771 int i = 0; 4772 4773 MWL_TXQ_LOCK(txq); 4774 STAILQ_FOREACH(bf, &txq->active, bf_list) { 4775 struct mwl_txdesc *ds = bf->bf_desc; 4776 MWL_TXDESC_SYNC(txq, ds, 4777 BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE); 4778 #ifdef MWL_DEBUG 4779 mwl_printtxbuf(bf, txq->qnum, i); 4780 #endif 4781 i++; 4782 } 4783 MWL_TXQ_UNLOCK(txq); 4784 } 4785 #endif 4786 4787 static void 4788 mwl_watchdog(void *arg) 4789 { 4790 struct mwl_softc *sc; 4791 struct ifnet *ifp; 4792 4793 sc = arg; 4794 callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc); 4795 if (sc->sc_tx_timer == 0 || --sc->sc_tx_timer > 0) 4796 return; 4797 4798 ifp = sc->sc_ifp; 4799 if ((ifp->if_drv_flags & IFF_DRV_RUNNING) && !sc->sc_invalid) { 4800 if (mwl_hal_setkeepalive(sc->sc_mh)) 4801 if_printf(ifp, "transmit timeout (firmware hung?)\n"); 4802 else 4803 if_printf(ifp, "transmit timeout\n"); 4804 #if 0 4805 mwl_reset(ifp); 4806 mwl_txq_dump(&sc->sc_txq[0]);/*XXX*/ 4807 #endif 4808 ifp->if_oerrors++; 4809 sc->sc_stats.mst_watchdog++; 4810 } 4811 } 4812 4813 #ifdef MWL_DIAGAPI 4814 /* 4815 * Diagnostic interface to the HAL. This is used by various 4816 * tools to do things like retrieve register contents for 4817 * debugging. The mechanism is intentionally opaque so that 4818 * it can change frequently w/o concern for compatiblity. 4819 */ 4820 static int 4821 mwl_ioctl_diag(struct mwl_softc *sc, struct mwl_diag *md) 4822 { 4823 struct mwl_hal *mh = sc->sc_mh; 4824 u_int id = md->md_id & MWL_DIAG_ID; 4825 void *indata = NULL; 4826 void *outdata = NULL; 4827 u_int32_t insize = md->md_in_size; 4828 u_int32_t outsize = md->md_out_size; 4829 int error = 0; 4830 4831 if (md->md_id & MWL_DIAG_IN) { 4832 /* 4833 * Copy in data. 4834 */ 4835 indata = malloc(insize, M_TEMP, M_NOWAIT); 4836 if (indata == NULL) { 4837 error = ENOMEM; 4838 goto bad; 4839 } 4840 error = copyin(md->md_in_data, indata, insize); 4841 if (error) 4842 goto bad; 4843 } 4844 if (md->md_id & MWL_DIAG_DYN) { 4845 /* 4846 * Allocate a buffer for the results (otherwise the HAL 4847 * returns a pointer to a buffer where we can read the 4848 * results). Note that we depend on the HAL leaving this 4849 * pointer for us to use below in reclaiming the buffer; 4850 * may want to be more defensive. 4851 */ 4852 outdata = malloc(outsize, M_TEMP, M_NOWAIT); 4853 if (outdata == NULL) { 4854 error = ENOMEM; 4855 goto bad; 4856 } 4857 } 4858 if (mwl_hal_getdiagstate(mh, id, indata, insize, &outdata, &outsize)) { 4859 if (outsize < md->md_out_size) 4860 md->md_out_size = outsize; 4861 if (outdata != NULL) 4862 error = copyout(outdata, md->md_out_data, 4863 md->md_out_size); 4864 } else { 4865 error = EINVAL; 4866 } 4867 bad: 4868 if ((md->md_id & MWL_DIAG_IN) && indata != NULL) 4869 free(indata, M_TEMP); 4870 if ((md->md_id & MWL_DIAG_DYN) && outdata != NULL) 4871 free(outdata, M_TEMP); 4872 return error; 4873 } 4874 4875 static int 4876 mwl_ioctl_reset(struct mwl_softc *sc, struct mwl_diag *md) 4877 { 4878 struct mwl_hal *mh = sc->sc_mh; 4879 int error; 4880 4881 MWL_LOCK_ASSERT(sc); 4882 4883 if (md->md_id == 0 && mwl_hal_fwload(mh, NULL) != 0) { 4884 device_printf(sc->sc_dev, "unable to load firmware\n"); 4885 return EIO; 4886 } 4887 if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) { 4888 device_printf(sc->sc_dev, "unable to fetch h/w specs\n"); 4889 return EIO; 4890 } 4891 error = mwl_setupdma(sc); 4892 if (error != 0) { 4893 /* NB: mwl_setupdma prints a msg */ 4894 return error; 4895 } 4896 /* 4897 * Reset tx/rx data structures; after reload we must 4898 * re-start the driver's notion of the next xmit/recv. 4899 */ 4900 mwl_draintxq(sc); /* clear pending frames */ 4901 mwl_resettxq(sc); /* rebuild tx q lists */ 4902 sc->sc_rxnext = NULL; /* force rx to start at the list head */ 4903 return 0; 4904 } 4905 #endif /* MWL_DIAGAPI */ 4906 4907 static int 4908 mwl_ioctl(struct ifnet *ifp, u_long cmd, caddr_t data) 4909 { 4910 #define IS_RUNNING(ifp) \ 4911 ((ifp->if_flags & IFF_UP) && (ifp->if_drv_flags & IFF_DRV_RUNNING)) 4912 struct mwl_softc *sc = ifp->if_softc; 4913 struct ieee80211com *ic = ifp->if_l2com; 4914 struct ifreq *ifr = (struct ifreq *)data; 4915 int error = 0, startall; 4916 4917 switch (cmd) { 4918 case SIOCSIFFLAGS: 4919 MWL_LOCK(sc); 4920 startall = 0; 4921 if (IS_RUNNING(ifp)) { 4922 /* 4923 * To avoid rescanning another access point, 4924 * do not call mwl_init() here. Instead, 4925 * only reflect promisc mode settings. 4926 */ 4927 mwl_mode_init(sc); 4928 } else if (ifp->if_flags & IFF_UP) { 4929 /* 4930 * Beware of being called during attach/detach 4931 * to reset promiscuous mode. In that case we 4932 * will still be marked UP but not RUNNING. 4933 * However trying to re-init the interface 4934 * is the wrong thing to do as we've already 4935 * torn down much of our state. There's 4936 * probably a better way to deal with this. 4937 */ 4938 if (!sc->sc_invalid) { 4939 mwl_init_locked(sc); /* XXX lose error */ 4940 startall = 1; 4941 } 4942 } else 4943 mwl_stop_locked(ifp, 1); 4944 MWL_UNLOCK(sc); 4945 if (startall) 4946 ieee80211_start_all(ic); 4947 break; 4948 case SIOCGMVSTATS: 4949 mwl_hal_gethwstats(sc->sc_mh, &sc->sc_stats.hw_stats); 4950 /* NB: embed these numbers to get a consistent view */ 4951 sc->sc_stats.mst_tx_packets = ifp->if_opackets; 4952 sc->sc_stats.mst_rx_packets = ifp->if_ipackets; 4953 /* 4954 * NB: Drop the softc lock in case of a page fault; 4955 * we'll accept any potential inconsisentcy in the 4956 * statistics. The alternative is to copy the data 4957 * to a local structure. 4958 */ 4959 return copyout(&sc->sc_stats, 4960 ifr->ifr_data, sizeof (sc->sc_stats)); 4961 #ifdef MWL_DIAGAPI 4962 case SIOCGMVDIAG: 4963 /* XXX check privs */ 4964 return mwl_ioctl_diag(sc, (struct mwl_diag *) ifr); 4965 case SIOCGMVRESET: 4966 /* XXX check privs */ 4967 MWL_LOCK(sc); 4968 error = mwl_ioctl_reset(sc,(struct mwl_diag *) ifr); 4969 MWL_UNLOCK(sc); 4970 break; 4971 #endif /* MWL_DIAGAPI */ 4972 case SIOCGIFMEDIA: 4973 error = ifmedia_ioctl(ifp, ifr, &ic->ic_media, cmd); 4974 break; 4975 case SIOCGIFADDR: 4976 error = ether_ioctl(ifp, cmd, data); 4977 break; 4978 default: 4979 error = EINVAL; 4980 break; 4981 } 4982 return error; 4983 #undef IS_RUNNING 4984 } 4985 4986 #ifdef MWL_DEBUG 4987 static int 4988 mwl_sysctl_debug(SYSCTL_HANDLER_ARGS) 4989 { 4990 struct mwl_softc *sc = arg1; 4991 int debug, error; 4992 4993 debug = sc->sc_debug | (mwl_hal_getdebug(sc->sc_mh) << 24); 4994 error = sysctl_handle_int(oidp, &debug, 0, req); 4995 if (error || !req->newptr) 4996 return error; 4997 mwl_hal_setdebug(sc->sc_mh, debug >> 24); 4998 sc->sc_debug = debug & 0x00ffffff; 4999 return 0; 5000 } 5001 #endif /* MWL_DEBUG */ 5002 5003 static void 5004 mwl_sysctlattach(struct mwl_softc *sc) 5005 { 5006 #ifdef MWL_DEBUG 5007 struct sysctl_ctx_list *ctx = device_get_sysctl_ctx(sc->sc_dev); 5008 struct sysctl_oid *tree = device_get_sysctl_tree(sc->sc_dev); 5009 5010 sc->sc_debug = mwl_debug; 5011 SYSCTL_ADD_PROC(ctx, SYSCTL_CHILDREN(tree), OID_AUTO, 5012 "debug", CTLTYPE_INT | CTLFLAG_RW, sc, 0, 5013 mwl_sysctl_debug, "I", "control debugging printfs"); 5014 #endif 5015 } 5016 5017 /* 5018 * Announce various information on device/driver attach. 5019 */ 5020 static void 5021 mwl_announce(struct mwl_softc *sc) 5022 { 5023 struct ifnet *ifp = sc->sc_ifp; 5024 5025 if_printf(ifp, "Rev A%d hardware, v%d.%d.%d.%d firmware (regioncode %d)\n", 5026 sc->sc_hwspecs.hwVersion, 5027 (sc->sc_hwspecs.fwReleaseNumber>>24) & 0xff, 5028 (sc->sc_hwspecs.fwReleaseNumber>>16) & 0xff, 5029 (sc->sc_hwspecs.fwReleaseNumber>>8) & 0xff, 5030 (sc->sc_hwspecs.fwReleaseNumber>>0) & 0xff, 5031 sc->sc_hwspecs.regionCode); 5032 sc->sc_fwrelease = sc->sc_hwspecs.fwReleaseNumber; 5033 5034 if (bootverbose) { 5035 int i; 5036 for (i = 0; i <= WME_AC_VO; i++) { 5037 struct mwl_txq *txq = sc->sc_ac2q[i]; 5038 if_printf(ifp, "Use hw queue %u for %s traffic\n", 5039 txq->qnum, ieee80211_wme_acnames[i]); 5040 } 5041 } 5042 if (bootverbose || mwl_rxdesc != MWL_RXDESC) 5043 if_printf(ifp, "using %u rx descriptors\n", mwl_rxdesc); 5044 if (bootverbose || mwl_rxbuf != MWL_RXBUF) 5045 if_printf(ifp, "using %u rx buffers\n", mwl_rxbuf); 5046 if (bootverbose || mwl_txbuf != MWL_TXBUF) 5047 if_printf(ifp, "using %u tx buffers\n", mwl_txbuf); 5048 if (bootverbose && mwl_hal_ismbsscapable(sc->sc_mh)) 5049 if_printf(ifp, "multi-bss support\n"); 5050 #ifdef MWL_TX_NODROP 5051 if (bootverbose) 5052 if_printf(ifp, "no tx drop\n"); 5053 #endif 5054 } 5055