xref: /freebsd/sys/dev/mwl/if_mwl.c (revision fe01740653ed066369f3c892b7aa18a710c608c5)
1 /*-
2  * Copyright (c) 2007-2009 Sam Leffler, Errno Consulting
3  * Copyright (c) 2007-2008 Marvell Semiconductor, Inc.
4  * All rights reserved.
5  *
6  * Redistribution and use in source and binary forms, with or without
7  * modification, are permitted provided that the following conditions
8  * are met:
9  * 1. Redistributions of source code must retain the above copyright
10  *    notice, this list of conditions and the following disclaimer,
11  *    without modification.
12  * 2. Redistributions in binary form must reproduce at minimum a disclaimer
13  *    similar to the "NO WARRANTY" disclaimer below ("Disclaimer") and any
14  *    redistribution must be conditioned upon including a substantially
15  *    similar Disclaimer requirement for further binary redistribution.
16  *
17  * NO WARRANTY
18  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
19  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
20  * LIMITED TO, THE IMPLIED WARRANTIES OF NONINFRINGEMENT, MERCHANTIBILITY
21  * AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL
22  * THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY,
23  * OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
24  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
25  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER
26  * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
27  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
28  * THE POSSIBILITY OF SUCH DAMAGES.
29  */
30 
31 #include <sys/cdefs.h>
32 __FBSDID("$FreeBSD$");
33 
34 /*
35  * Driver for the Marvell 88W8363 Wireless LAN controller.
36  */
37 
38 #include "opt_inet.h"
39 #include "opt_mwl.h"
40 #include "opt_wlan.h"
41 
42 #include <sys/param.h>
43 #include <sys/systm.h>
44 #include <sys/sysctl.h>
45 #include <sys/mbuf.h>
46 #include <sys/malloc.h>
47 #include <sys/lock.h>
48 #include <sys/mutex.h>
49 #include <sys/kernel.h>
50 #include <sys/socket.h>
51 #include <sys/sockio.h>
52 #include <sys/errno.h>
53 #include <sys/callout.h>
54 #include <sys/bus.h>
55 #include <sys/endian.h>
56 #include <sys/kthread.h>
57 #include <sys/taskqueue.h>
58 
59 #include <machine/bus.h>
60 
61 #include <net/if.h>
62 #include <net/if_var.h>
63 #include <net/if_dl.h>
64 #include <net/if_media.h>
65 #include <net/if_types.h>
66 #include <net/if_arp.h>
67 #include <net/ethernet.h>
68 #include <net/if_llc.h>
69 
70 #include <net/bpf.h>
71 
72 #include <net80211/ieee80211_var.h>
73 #include <net80211/ieee80211_regdomain.h>
74 
75 #ifdef INET
76 #include <netinet/in.h>
77 #include <netinet/if_ether.h>
78 #endif /* INET */
79 
80 #include <dev/mwl/if_mwlvar.h>
81 #include <dev/mwl/mwldiag.h>
82 
83 /* idiomatic shorthands: MS = mask+shift, SM = shift+mask */
84 #define	MS(v,x)	(((v) & x) >> x##_S)
85 #define	SM(v,x)	(((v) << x##_S) & x)
86 
87 static struct ieee80211vap *mwl_vap_create(struct ieee80211com *,
88 		    const char [IFNAMSIZ], int, enum ieee80211_opmode, int,
89 		    const uint8_t [IEEE80211_ADDR_LEN],
90 		    const uint8_t [IEEE80211_ADDR_LEN]);
91 static void	mwl_vap_delete(struct ieee80211vap *);
92 static int	mwl_setupdma(struct mwl_softc *);
93 static int	mwl_hal_reset(struct mwl_softc *sc);
94 static int	mwl_init_locked(struct mwl_softc *);
95 static void	mwl_init(void *);
96 static void	mwl_stop_locked(struct ifnet *, int);
97 static int	mwl_reset(struct ieee80211vap *, u_long);
98 static void	mwl_stop(struct ifnet *, int);
99 static void	mwl_start(struct ifnet *);
100 static int	mwl_raw_xmit(struct ieee80211_node *, struct mbuf *,
101 			const struct ieee80211_bpf_params *);
102 static int	mwl_media_change(struct ifnet *);
103 static void	mwl_watchdog(void *);
104 static int	mwl_ioctl(struct ifnet *, u_long, caddr_t);
105 static void	mwl_radar_proc(void *, int);
106 static void	mwl_chanswitch_proc(void *, int);
107 static void	mwl_bawatchdog_proc(void *, int);
108 static int	mwl_key_alloc(struct ieee80211vap *,
109 			struct ieee80211_key *,
110 			ieee80211_keyix *, ieee80211_keyix *);
111 static int	mwl_key_delete(struct ieee80211vap *,
112 			const struct ieee80211_key *);
113 static int	mwl_key_set(struct ieee80211vap *, const struct ieee80211_key *,
114 			const uint8_t mac[IEEE80211_ADDR_LEN]);
115 static int	mwl_mode_init(struct mwl_softc *);
116 static void	mwl_update_mcast(struct ifnet *);
117 static void	mwl_update_promisc(struct ifnet *);
118 static void	mwl_updateslot(struct ifnet *);
119 static int	mwl_beacon_setup(struct ieee80211vap *);
120 static void	mwl_beacon_update(struct ieee80211vap *, int);
121 #ifdef MWL_HOST_PS_SUPPORT
122 static void	mwl_update_ps(struct ieee80211vap *, int);
123 static int	mwl_set_tim(struct ieee80211_node *, int);
124 #endif
125 static int	mwl_dma_setup(struct mwl_softc *);
126 static void	mwl_dma_cleanup(struct mwl_softc *);
127 static struct ieee80211_node *mwl_node_alloc(struct ieee80211vap *,
128 		    const uint8_t [IEEE80211_ADDR_LEN]);
129 static void	mwl_node_cleanup(struct ieee80211_node *);
130 static void	mwl_node_drain(struct ieee80211_node *);
131 static void	mwl_node_getsignal(const struct ieee80211_node *,
132 			int8_t *, int8_t *);
133 static void	mwl_node_getmimoinfo(const struct ieee80211_node *,
134 			struct ieee80211_mimo_info *);
135 static int	mwl_rxbuf_init(struct mwl_softc *, struct mwl_rxbuf *);
136 static void	mwl_rx_proc(void *, int);
137 static void	mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *, int);
138 static int	mwl_tx_setup(struct mwl_softc *, int, int);
139 static int	mwl_wme_update(struct ieee80211com *);
140 static void	mwl_tx_cleanupq(struct mwl_softc *, struct mwl_txq *);
141 static void	mwl_tx_cleanup(struct mwl_softc *);
142 static uint16_t	mwl_calcformat(uint8_t rate, const struct ieee80211_node *);
143 static int	mwl_tx_start(struct mwl_softc *, struct ieee80211_node *,
144 			     struct mwl_txbuf *, struct mbuf *);
145 static void	mwl_tx_proc(void *, int);
146 static int	mwl_chan_set(struct mwl_softc *, struct ieee80211_channel *);
147 static void	mwl_draintxq(struct mwl_softc *);
148 static void	mwl_cleartxq(struct mwl_softc *, struct ieee80211vap *);
149 static int	mwl_recv_action(struct ieee80211_node *,
150 			const struct ieee80211_frame *,
151 			const uint8_t *, const uint8_t *);
152 static int	mwl_addba_request(struct ieee80211_node *,
153 			struct ieee80211_tx_ampdu *, int dialogtoken,
154 			int baparamset, int batimeout);
155 static int	mwl_addba_response(struct ieee80211_node *,
156 			struct ieee80211_tx_ampdu *, int status,
157 			int baparamset, int batimeout);
158 static void	mwl_addba_stop(struct ieee80211_node *,
159 			struct ieee80211_tx_ampdu *);
160 static int	mwl_startrecv(struct mwl_softc *);
161 static MWL_HAL_APMODE mwl_getapmode(const struct ieee80211vap *,
162 			struct ieee80211_channel *);
163 static int	mwl_setapmode(struct ieee80211vap *, struct ieee80211_channel*);
164 static void	mwl_scan_start(struct ieee80211com *);
165 static void	mwl_scan_end(struct ieee80211com *);
166 static void	mwl_set_channel(struct ieee80211com *);
167 static int	mwl_peerstadb(struct ieee80211_node *,
168 			int aid, int staid, MWL_HAL_PEERINFO *pi);
169 static int	mwl_localstadb(struct ieee80211vap *);
170 static int	mwl_newstate(struct ieee80211vap *, enum ieee80211_state, int);
171 static int	allocstaid(struct mwl_softc *sc, int aid);
172 static void	delstaid(struct mwl_softc *sc, int staid);
173 static void	mwl_newassoc(struct ieee80211_node *, int);
174 static void	mwl_agestations(void *);
175 static int	mwl_setregdomain(struct ieee80211com *,
176 			struct ieee80211_regdomain *, int,
177 			struct ieee80211_channel []);
178 static void	mwl_getradiocaps(struct ieee80211com *, int, int *,
179 			struct ieee80211_channel []);
180 static int	mwl_getchannels(struct mwl_softc *);
181 
182 static void	mwl_sysctlattach(struct mwl_softc *);
183 static void	mwl_announce(struct mwl_softc *);
184 
185 SYSCTL_NODE(_hw, OID_AUTO, mwl, CTLFLAG_RD, 0, "Marvell driver parameters");
186 
187 static	int mwl_rxdesc = MWL_RXDESC;		/* # rx desc's to allocate */
188 SYSCTL_INT(_hw_mwl, OID_AUTO, rxdesc, CTLFLAG_RW, &mwl_rxdesc,
189 	    0, "rx descriptors allocated");
190 static	int mwl_rxbuf = MWL_RXBUF;		/* # rx buffers to allocate */
191 SYSCTL_INT(_hw_mwl, OID_AUTO, rxbuf, CTLFLAG_RW, &mwl_rxbuf,
192 	    0, "rx buffers allocated");
193 TUNABLE_INT("hw.mwl.rxbuf", &mwl_rxbuf);
194 static	int mwl_txbuf = MWL_TXBUF;		/* # tx buffers to allocate */
195 SYSCTL_INT(_hw_mwl, OID_AUTO, txbuf, CTLFLAG_RW, &mwl_txbuf,
196 	    0, "tx buffers allocated");
197 TUNABLE_INT("hw.mwl.txbuf", &mwl_txbuf);
198 static	int mwl_txcoalesce = 8;		/* # tx packets to q before poking f/w*/
199 SYSCTL_INT(_hw_mwl, OID_AUTO, txcoalesce, CTLFLAG_RW, &mwl_txcoalesce,
200 	    0, "tx buffers to send at once");
201 TUNABLE_INT("hw.mwl.txcoalesce", &mwl_txcoalesce);
202 static	int mwl_rxquota = MWL_RXBUF;		/* # max buffers to process */
203 SYSCTL_INT(_hw_mwl, OID_AUTO, rxquota, CTLFLAG_RW, &mwl_rxquota,
204 	    0, "max rx buffers to process per interrupt");
205 TUNABLE_INT("hw.mwl.rxquota", &mwl_rxquota);
206 static	int mwl_rxdmalow = 3;			/* # min buffers for wakeup */
207 SYSCTL_INT(_hw_mwl, OID_AUTO, rxdmalow, CTLFLAG_RW, &mwl_rxdmalow,
208 	    0, "min free rx buffers before restarting traffic");
209 TUNABLE_INT("hw.mwl.rxdmalow", &mwl_rxdmalow);
210 
211 #ifdef MWL_DEBUG
212 static	int mwl_debug = 0;
213 SYSCTL_INT(_hw_mwl, OID_AUTO, debug, CTLFLAG_RW, &mwl_debug,
214 	    0, "control debugging printfs");
215 TUNABLE_INT("hw.mwl.debug", &mwl_debug);
216 enum {
217 	MWL_DEBUG_XMIT		= 0x00000001,	/* basic xmit operation */
218 	MWL_DEBUG_XMIT_DESC	= 0x00000002,	/* xmit descriptors */
219 	MWL_DEBUG_RECV		= 0x00000004,	/* basic recv operation */
220 	MWL_DEBUG_RECV_DESC	= 0x00000008,	/* recv descriptors */
221 	MWL_DEBUG_RESET		= 0x00000010,	/* reset processing */
222 	MWL_DEBUG_BEACON 	= 0x00000020,	/* beacon handling */
223 	MWL_DEBUG_INTR		= 0x00000040,	/* ISR */
224 	MWL_DEBUG_TX_PROC	= 0x00000080,	/* tx ISR proc */
225 	MWL_DEBUG_RX_PROC	= 0x00000100,	/* rx ISR proc */
226 	MWL_DEBUG_KEYCACHE	= 0x00000200,	/* key cache management */
227 	MWL_DEBUG_STATE		= 0x00000400,	/* 802.11 state transitions */
228 	MWL_DEBUG_NODE		= 0x00000800,	/* node management */
229 	MWL_DEBUG_RECV_ALL	= 0x00001000,	/* trace all frames (beacons) */
230 	MWL_DEBUG_TSO		= 0x00002000,	/* TSO processing */
231 	MWL_DEBUG_AMPDU		= 0x00004000,	/* BA stream handling */
232 	MWL_DEBUG_ANY		= 0xffffffff
233 };
234 #define	IS_BEACON(wh) \
235     ((wh->i_fc[0] & (IEEE80211_FC0_TYPE_MASK|IEEE80211_FC0_SUBTYPE_MASK)) == \
236 	 (IEEE80211_FC0_TYPE_MGT|IEEE80211_FC0_SUBTYPE_BEACON))
237 #define	IFF_DUMPPKTS_RECV(sc, wh) \
238     (((sc->sc_debug & MWL_DEBUG_RECV) && \
239       ((sc->sc_debug & MWL_DEBUG_RECV_ALL) || !IS_BEACON(wh))) || \
240      (sc->sc_ifp->if_flags & (IFF_DEBUG|IFF_LINK2)) == (IFF_DEBUG|IFF_LINK2))
241 #define	IFF_DUMPPKTS_XMIT(sc) \
242 	((sc->sc_debug & MWL_DEBUG_XMIT) || \
243 	 (sc->sc_ifp->if_flags & (IFF_DEBUG|IFF_LINK2)) == (IFF_DEBUG|IFF_LINK2))
244 #define	DPRINTF(sc, m, fmt, ...) do {				\
245 	if (sc->sc_debug & (m))					\
246 		printf(fmt, __VA_ARGS__);			\
247 } while (0)
248 #define	KEYPRINTF(sc, hk, mac) do {				\
249 	if (sc->sc_debug & MWL_DEBUG_KEYCACHE)			\
250 		mwl_keyprint(sc, __func__, hk, mac);		\
251 } while (0)
252 static	void mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix);
253 static	void mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix);
254 #else
255 #define	IFF_DUMPPKTS_RECV(sc, wh) \
256 	((sc->sc_ifp->if_flags & (IFF_DEBUG|IFF_LINK2)) == (IFF_DEBUG|IFF_LINK2))
257 #define	IFF_DUMPPKTS_XMIT(sc) \
258 	((sc->sc_ifp->if_flags & (IFF_DEBUG|IFF_LINK2)) == (IFF_DEBUG|IFF_LINK2))
259 #define	DPRINTF(sc, m, fmt, ...) do {				\
260 	(void) sc;						\
261 } while (0)
262 #define	KEYPRINTF(sc, k, mac) do {				\
263 	(void) sc;						\
264 } while (0)
265 #endif
266 
267 static MALLOC_DEFINE(M_MWLDEV, "mwldev", "mwl driver dma buffers");
268 
269 /*
270  * Each packet has fixed front matter: a 2-byte length
271  * of the payload, followed by a 4-address 802.11 header
272  * (regardless of the actual header and always w/o any
273  * QoS header).  The payload then follows.
274  */
275 struct mwltxrec {
276 	uint16_t fwlen;
277 	struct ieee80211_frame_addr4 wh;
278 } __packed;
279 
280 /*
281  * Read/Write shorthands for accesses to BAR 0.  Note
282  * that all BAR 1 operations are done in the "hal" and
283  * there should be no reference to them here.
284  */
285 #ifdef MWL_DEBUG
286 static __inline uint32_t
287 RD4(struct mwl_softc *sc, bus_size_t off)
288 {
289 	return bus_space_read_4(sc->sc_io0t, sc->sc_io0h, off);
290 }
291 #endif
292 
293 static __inline void
294 WR4(struct mwl_softc *sc, bus_size_t off, uint32_t val)
295 {
296 	bus_space_write_4(sc->sc_io0t, sc->sc_io0h, off, val);
297 }
298 
299 int
300 mwl_attach(uint16_t devid, struct mwl_softc *sc)
301 {
302 	struct ifnet *ifp;
303 	struct ieee80211com *ic;
304 	struct mwl_hal *mh;
305 	int error = 0;
306 
307 	DPRINTF(sc, MWL_DEBUG_ANY, "%s: devid 0x%x\n", __func__, devid);
308 
309 	ifp = sc->sc_ifp = if_alloc(IFT_IEEE80211);
310 	if (ifp == NULL) {
311 		device_printf(sc->sc_dev, "cannot if_alloc()\n");
312 		return ENOSPC;
313 	}
314 	ic = ifp->if_l2com;
315 
316 	/*
317 	 * Setup the RX free list lock early, so it can be consistently
318 	 * removed.
319 	 */
320 	MWL_RXFREE_INIT(sc);
321 
322 	/* set these up early for if_printf use */
323 	if_initname(ifp, device_get_name(sc->sc_dev),
324 		device_get_unit(sc->sc_dev));
325 
326 	mh = mwl_hal_attach(sc->sc_dev, devid,
327 	    sc->sc_io1h, sc->sc_io1t, sc->sc_dmat);
328 	if (mh == NULL) {
329 		if_printf(ifp, "unable to attach HAL\n");
330 		error = EIO;
331 		goto bad;
332 	}
333 	sc->sc_mh = mh;
334 	/*
335 	 * Load firmware so we can get setup.  We arbitrarily
336 	 * pick station firmware; we'll re-load firmware as
337 	 * needed so setting up the wrong mode isn't a big deal.
338 	 */
339 	if (mwl_hal_fwload(mh, NULL) != 0) {
340 		if_printf(ifp, "unable to setup builtin firmware\n");
341 		error = EIO;
342 		goto bad1;
343 	}
344 	if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) {
345 		if_printf(ifp, "unable to fetch h/w specs\n");
346 		error = EIO;
347 		goto bad1;
348 	}
349 	error = mwl_getchannels(sc);
350 	if (error != 0)
351 		goto bad1;
352 
353 	sc->sc_txantenna = 0;		/* h/w default */
354 	sc->sc_rxantenna = 0;		/* h/w default */
355 	sc->sc_invalid = 0;		/* ready to go, enable int handling */
356 	sc->sc_ageinterval = MWL_AGEINTERVAL;
357 
358 	/*
359 	 * Allocate tx+rx descriptors and populate the lists.
360 	 * We immediately push the information to the firmware
361 	 * as otherwise it gets upset.
362 	 */
363 	error = mwl_dma_setup(sc);
364 	if (error != 0) {
365 		if_printf(ifp, "failed to setup descriptors: %d\n", error);
366 		goto bad1;
367 	}
368 	error = mwl_setupdma(sc);	/* push to firmware */
369 	if (error != 0)			/* NB: mwl_setupdma prints msg */
370 		goto bad1;
371 
372 	callout_init(&sc->sc_timer, CALLOUT_MPSAFE);
373 	callout_init_mtx(&sc->sc_watchdog, &sc->sc_mtx, 0);
374 
375 	sc->sc_tq = taskqueue_create("mwl_taskq", M_NOWAIT,
376 		taskqueue_thread_enqueue, &sc->sc_tq);
377 	taskqueue_start_threads(&sc->sc_tq, 1, PI_NET,
378 		"%s taskq", ifp->if_xname);
379 
380 	TASK_INIT(&sc->sc_rxtask, 0, mwl_rx_proc, sc);
381 	TASK_INIT(&sc->sc_radartask, 0, mwl_radar_proc, sc);
382 	TASK_INIT(&sc->sc_chanswitchtask, 0, mwl_chanswitch_proc, sc);
383 	TASK_INIT(&sc->sc_bawatchdogtask, 0, mwl_bawatchdog_proc, sc);
384 
385 	/* NB: insure BK queue is the lowest priority h/w queue */
386 	if (!mwl_tx_setup(sc, WME_AC_BK, MWL_WME_AC_BK)) {
387 		if_printf(ifp, "unable to setup xmit queue for %s traffic!\n",
388 			ieee80211_wme_acnames[WME_AC_BK]);
389 		error = EIO;
390 		goto bad2;
391 	}
392 	if (!mwl_tx_setup(sc, WME_AC_BE, MWL_WME_AC_BE) ||
393 	    !mwl_tx_setup(sc, WME_AC_VI, MWL_WME_AC_VI) ||
394 	    !mwl_tx_setup(sc, WME_AC_VO, MWL_WME_AC_VO)) {
395 		/*
396 		 * Not enough hardware tx queues to properly do WME;
397 		 * just punt and assign them all to the same h/w queue.
398 		 * We could do a better job of this if, for example,
399 		 * we allocate queues when we switch from station to
400 		 * AP mode.
401 		 */
402 		if (sc->sc_ac2q[WME_AC_VI] != NULL)
403 			mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_VI]);
404 		if (sc->sc_ac2q[WME_AC_BE] != NULL)
405 			mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_BE]);
406 		sc->sc_ac2q[WME_AC_BE] = sc->sc_ac2q[WME_AC_BK];
407 		sc->sc_ac2q[WME_AC_VI] = sc->sc_ac2q[WME_AC_BK];
408 		sc->sc_ac2q[WME_AC_VO] = sc->sc_ac2q[WME_AC_BK];
409 	}
410 	TASK_INIT(&sc->sc_txtask, 0, mwl_tx_proc, sc);
411 
412 	ifp->if_softc = sc;
413 	ifp->if_flags = IFF_SIMPLEX | IFF_BROADCAST | IFF_MULTICAST;
414 	ifp->if_start = mwl_start;
415 	ifp->if_ioctl = mwl_ioctl;
416 	ifp->if_init = mwl_init;
417 	IFQ_SET_MAXLEN(&ifp->if_snd, ifqmaxlen);
418 	ifp->if_snd.ifq_drv_maxlen = ifqmaxlen;
419 	IFQ_SET_READY(&ifp->if_snd);
420 
421 	ic->ic_ifp = ifp;
422 	/* XXX not right but it's not used anywhere important */
423 	ic->ic_phytype = IEEE80211_T_OFDM;
424 	ic->ic_opmode = IEEE80211_M_STA;
425 	ic->ic_caps =
426 		  IEEE80211_C_STA		/* station mode supported */
427 		| IEEE80211_C_HOSTAP		/* hostap mode */
428 		| IEEE80211_C_MONITOR		/* monitor mode */
429 #if 0
430 		| IEEE80211_C_IBSS		/* ibss, nee adhoc, mode */
431 		| IEEE80211_C_AHDEMO		/* adhoc demo mode */
432 #endif
433 		| IEEE80211_C_MBSS		/* mesh point link mode */
434 		| IEEE80211_C_WDS		/* WDS supported */
435 		| IEEE80211_C_SHPREAMBLE	/* short preamble supported */
436 		| IEEE80211_C_SHSLOT		/* short slot time supported */
437 		| IEEE80211_C_WME		/* WME/WMM supported */
438 		| IEEE80211_C_BURST		/* xmit bursting supported */
439 		| IEEE80211_C_WPA		/* capable of WPA1+WPA2 */
440 		| IEEE80211_C_BGSCAN		/* capable of bg scanning */
441 		| IEEE80211_C_TXFRAG		/* handle tx frags */
442 		| IEEE80211_C_TXPMGT		/* capable of txpow mgt */
443 		| IEEE80211_C_DFS		/* DFS supported */
444 		;
445 
446 	ic->ic_htcaps =
447 		  IEEE80211_HTCAP_SMPS_ENA	/* SM PS mode enabled */
448 		| IEEE80211_HTCAP_CHWIDTH40	/* 40MHz channel width */
449 		| IEEE80211_HTCAP_SHORTGI20	/* short GI in 20MHz */
450 		| IEEE80211_HTCAP_SHORTGI40	/* short GI in 40MHz */
451 		| IEEE80211_HTCAP_RXSTBC_2STREAM/* 1-2 spatial streams */
452 #if MWL_AGGR_SIZE == 7935
453 		| IEEE80211_HTCAP_MAXAMSDU_7935	/* max A-MSDU length */
454 #else
455 		| IEEE80211_HTCAP_MAXAMSDU_3839	/* max A-MSDU length */
456 #endif
457 #if 0
458 		| IEEE80211_HTCAP_PSMP		/* PSMP supported */
459 		| IEEE80211_HTCAP_40INTOLERANT	/* 40MHz intolerant */
460 #endif
461 		/* s/w capabilities */
462 		| IEEE80211_HTC_HT		/* HT operation */
463 		| IEEE80211_HTC_AMPDU		/* tx A-MPDU */
464 		| IEEE80211_HTC_AMSDU		/* tx A-MSDU */
465 		| IEEE80211_HTC_SMPS		/* SMPS available */
466 		;
467 
468 	/*
469 	 * Mark h/w crypto support.
470 	 * XXX no way to query h/w support.
471 	 */
472 	ic->ic_cryptocaps |= IEEE80211_CRYPTO_WEP
473 			  |  IEEE80211_CRYPTO_AES_CCM
474 			  |  IEEE80211_CRYPTO_TKIP
475 			  |  IEEE80211_CRYPTO_TKIPMIC
476 			  ;
477 	/*
478 	 * Transmit requires space in the packet for a special
479 	 * format transmit record and optional padding between
480 	 * this record and the payload.  Ask the net80211 layer
481 	 * to arrange this when encapsulating packets so we can
482 	 * add it efficiently.
483 	 */
484 	ic->ic_headroom = sizeof(struct mwltxrec) -
485 		sizeof(struct ieee80211_frame);
486 
487 	/* call MI attach routine. */
488 	ieee80211_ifattach(ic, sc->sc_hwspecs.macAddr);
489 	ic->ic_setregdomain = mwl_setregdomain;
490 	ic->ic_getradiocaps = mwl_getradiocaps;
491 	/* override default methods */
492 	ic->ic_raw_xmit = mwl_raw_xmit;
493 	ic->ic_newassoc = mwl_newassoc;
494 	ic->ic_updateslot = mwl_updateslot;
495 	ic->ic_update_mcast = mwl_update_mcast;
496 	ic->ic_update_promisc = mwl_update_promisc;
497 	ic->ic_wme.wme_update = mwl_wme_update;
498 
499 	ic->ic_node_alloc = mwl_node_alloc;
500 	sc->sc_node_cleanup = ic->ic_node_cleanup;
501 	ic->ic_node_cleanup = mwl_node_cleanup;
502 	sc->sc_node_drain = ic->ic_node_drain;
503 	ic->ic_node_drain = mwl_node_drain;
504 	ic->ic_node_getsignal = mwl_node_getsignal;
505 	ic->ic_node_getmimoinfo = mwl_node_getmimoinfo;
506 
507 	ic->ic_scan_start = mwl_scan_start;
508 	ic->ic_scan_end = mwl_scan_end;
509 	ic->ic_set_channel = mwl_set_channel;
510 
511 	sc->sc_recv_action = ic->ic_recv_action;
512 	ic->ic_recv_action = mwl_recv_action;
513 	sc->sc_addba_request = ic->ic_addba_request;
514 	ic->ic_addba_request = mwl_addba_request;
515 	sc->sc_addba_response = ic->ic_addba_response;
516 	ic->ic_addba_response = mwl_addba_response;
517 	sc->sc_addba_stop = ic->ic_addba_stop;
518 	ic->ic_addba_stop = mwl_addba_stop;
519 
520 	ic->ic_vap_create = mwl_vap_create;
521 	ic->ic_vap_delete = mwl_vap_delete;
522 
523 	ieee80211_radiotap_attach(ic,
524 	    &sc->sc_tx_th.wt_ihdr, sizeof(sc->sc_tx_th),
525 		MWL_TX_RADIOTAP_PRESENT,
526 	    &sc->sc_rx_th.wr_ihdr, sizeof(sc->sc_rx_th),
527 		MWL_RX_RADIOTAP_PRESENT);
528 	/*
529 	 * Setup dynamic sysctl's now that country code and
530 	 * regdomain are available from the hal.
531 	 */
532 	mwl_sysctlattach(sc);
533 
534 	if (bootverbose)
535 		ieee80211_announce(ic);
536 	mwl_announce(sc);
537 	return 0;
538 bad2:
539 	mwl_dma_cleanup(sc);
540 bad1:
541 	mwl_hal_detach(mh);
542 bad:
543 	MWL_RXFREE_DESTROY(sc);
544 	if_free(ifp);
545 	sc->sc_invalid = 1;
546 	return error;
547 }
548 
549 int
550 mwl_detach(struct mwl_softc *sc)
551 {
552 	struct ifnet *ifp = sc->sc_ifp;
553 	struct ieee80211com *ic = ifp->if_l2com;
554 
555 	DPRINTF(sc, MWL_DEBUG_ANY, "%s: if_flags %x\n",
556 		__func__, ifp->if_flags);
557 
558 	mwl_stop(ifp, 1);
559 	/*
560 	 * NB: the order of these is important:
561 	 * o call the 802.11 layer before detaching the hal to
562 	 *   insure callbacks into the driver to delete global
563 	 *   key cache entries can be handled
564 	 * o reclaim the tx queue data structures after calling
565 	 *   the 802.11 layer as we'll get called back to reclaim
566 	 *   node state and potentially want to use them
567 	 * o to cleanup the tx queues the hal is called, so detach
568 	 *   it last
569 	 * Other than that, it's straightforward...
570 	 */
571 	ieee80211_ifdetach(ic);
572 	callout_drain(&sc->sc_watchdog);
573 	mwl_dma_cleanup(sc);
574 	MWL_RXFREE_DESTROY(sc);
575 	mwl_tx_cleanup(sc);
576 	mwl_hal_detach(sc->sc_mh);
577 	if_free(ifp);
578 
579 	return 0;
580 }
581 
582 /*
583  * MAC address handling for multiple BSS on the same radio.
584  * The first vap uses the MAC address from the EEPROM.  For
585  * subsequent vap's we set the U/L bit (bit 1) in the MAC
586  * address and use the next six bits as an index.
587  */
588 static void
589 assign_address(struct mwl_softc *sc, uint8_t mac[IEEE80211_ADDR_LEN], int clone)
590 {
591 	int i;
592 
593 	if (clone && mwl_hal_ismbsscapable(sc->sc_mh)) {
594 		/* NB: we only do this if h/w supports multiple bssid */
595 		for (i = 0; i < 32; i++)
596 			if ((sc->sc_bssidmask & (1<<i)) == 0)
597 				break;
598 		if (i != 0)
599 			mac[0] |= (i << 2)|0x2;
600 	} else
601 		i = 0;
602 	sc->sc_bssidmask |= 1<<i;
603 	if (i == 0)
604 		sc->sc_nbssid0++;
605 }
606 
607 static void
608 reclaim_address(struct mwl_softc *sc, uint8_t mac[IEEE80211_ADDR_LEN])
609 {
610 	int i = mac[0] >> 2;
611 	if (i != 0 || --sc->sc_nbssid0 == 0)
612 		sc->sc_bssidmask &= ~(1<<i);
613 }
614 
615 static struct ieee80211vap *
616 mwl_vap_create(struct ieee80211com *ic, const char name[IFNAMSIZ], int unit,
617     enum ieee80211_opmode opmode, int flags,
618     const uint8_t bssid[IEEE80211_ADDR_LEN],
619     const uint8_t mac0[IEEE80211_ADDR_LEN])
620 {
621 	struct ifnet *ifp = ic->ic_ifp;
622 	struct mwl_softc *sc = ifp->if_softc;
623 	struct mwl_hal *mh = sc->sc_mh;
624 	struct ieee80211vap *vap, *apvap;
625 	struct mwl_hal_vap *hvap;
626 	struct mwl_vap *mvp;
627 	uint8_t mac[IEEE80211_ADDR_LEN];
628 
629 	IEEE80211_ADDR_COPY(mac, mac0);
630 	switch (opmode) {
631 	case IEEE80211_M_HOSTAP:
632 	case IEEE80211_M_MBSS:
633 		if ((flags & IEEE80211_CLONE_MACADDR) == 0)
634 			assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID);
635 		hvap = mwl_hal_newvap(mh, MWL_HAL_AP, mac);
636 		if (hvap == NULL) {
637 			if ((flags & IEEE80211_CLONE_MACADDR) == 0)
638 				reclaim_address(sc, mac);
639 			return NULL;
640 		}
641 		break;
642 	case IEEE80211_M_STA:
643 		if ((flags & IEEE80211_CLONE_MACADDR) == 0)
644 			assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID);
645 		hvap = mwl_hal_newvap(mh, MWL_HAL_STA, mac);
646 		if (hvap == NULL) {
647 			if ((flags & IEEE80211_CLONE_MACADDR) == 0)
648 				reclaim_address(sc, mac);
649 			return NULL;
650 		}
651 		/* no h/w beacon miss support; always use s/w */
652 		flags |= IEEE80211_CLONE_NOBEACONS;
653 		break;
654 	case IEEE80211_M_WDS:
655 		hvap = NULL;		/* NB: we use associated AP vap */
656 		if (sc->sc_napvaps == 0)
657 			return NULL;	/* no existing AP vap */
658 		break;
659 	case IEEE80211_M_MONITOR:
660 		hvap = NULL;
661 		break;
662 	case IEEE80211_M_IBSS:
663 	case IEEE80211_M_AHDEMO:
664 	default:
665 		return NULL;
666 	}
667 
668 	mvp = (struct mwl_vap *) malloc(sizeof(struct mwl_vap),
669 	    M_80211_VAP, M_NOWAIT | M_ZERO);
670 	if (mvp == NULL) {
671 		if (hvap != NULL) {
672 			mwl_hal_delvap(hvap);
673 			if ((flags & IEEE80211_CLONE_MACADDR) == 0)
674 				reclaim_address(sc, mac);
675 		}
676 		/* XXX msg */
677 		return NULL;
678 	}
679 	mvp->mv_hvap = hvap;
680 	if (opmode == IEEE80211_M_WDS) {
681 		/*
682 		 * WDS vaps must have an associated AP vap; find one.
683 		 * XXX not right.
684 		 */
685 		TAILQ_FOREACH(apvap, &ic->ic_vaps, iv_next)
686 			if (apvap->iv_opmode == IEEE80211_M_HOSTAP) {
687 				mvp->mv_ap_hvap = MWL_VAP(apvap)->mv_hvap;
688 				break;
689 			}
690 		KASSERT(mvp->mv_ap_hvap != NULL, ("no ap vap"));
691 	}
692 	vap = &mvp->mv_vap;
693 	ieee80211_vap_setup(ic, vap, name, unit, opmode, flags, bssid, mac);
694 	if (hvap != NULL)
695 		IEEE80211_ADDR_COPY(vap->iv_myaddr, mac);
696 	/* override with driver methods */
697 	mvp->mv_newstate = vap->iv_newstate;
698 	vap->iv_newstate = mwl_newstate;
699 	vap->iv_max_keyix = 0;	/* XXX */
700 	vap->iv_key_alloc = mwl_key_alloc;
701 	vap->iv_key_delete = mwl_key_delete;
702 	vap->iv_key_set = mwl_key_set;
703 #ifdef MWL_HOST_PS_SUPPORT
704 	if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS) {
705 		vap->iv_update_ps = mwl_update_ps;
706 		mvp->mv_set_tim = vap->iv_set_tim;
707 		vap->iv_set_tim = mwl_set_tim;
708 	}
709 #endif
710 	vap->iv_reset = mwl_reset;
711 	vap->iv_update_beacon = mwl_beacon_update;
712 
713 	/* override max aid so sta's cannot assoc when we're out of sta id's */
714 	vap->iv_max_aid = MWL_MAXSTAID;
715 	/* override default A-MPDU rx parameters */
716 	vap->iv_ampdu_rxmax = IEEE80211_HTCAP_MAXRXAMPDU_64K;
717 	vap->iv_ampdu_density = IEEE80211_HTCAP_MPDUDENSITY_4;
718 
719 	/* complete setup */
720 	ieee80211_vap_attach(vap, mwl_media_change, ieee80211_media_status);
721 
722 	switch (vap->iv_opmode) {
723 	case IEEE80211_M_HOSTAP:
724 	case IEEE80211_M_MBSS:
725 	case IEEE80211_M_STA:
726 		/*
727 		 * Setup sta db entry for local address.
728 		 */
729 		mwl_localstadb(vap);
730 		if (vap->iv_opmode == IEEE80211_M_HOSTAP ||
731 		    vap->iv_opmode == IEEE80211_M_MBSS)
732 			sc->sc_napvaps++;
733 		else
734 			sc->sc_nstavaps++;
735 		break;
736 	case IEEE80211_M_WDS:
737 		sc->sc_nwdsvaps++;
738 		break;
739 	default:
740 		break;
741 	}
742 	/*
743 	 * Setup overall operating mode.
744 	 */
745 	if (sc->sc_napvaps)
746 		ic->ic_opmode = IEEE80211_M_HOSTAP;
747 	else if (sc->sc_nstavaps)
748 		ic->ic_opmode = IEEE80211_M_STA;
749 	else
750 		ic->ic_opmode = opmode;
751 
752 	return vap;
753 }
754 
755 static void
756 mwl_vap_delete(struct ieee80211vap *vap)
757 {
758 	struct mwl_vap *mvp = MWL_VAP(vap);
759 	struct ifnet *parent = vap->iv_ic->ic_ifp;
760 	struct mwl_softc *sc = parent->if_softc;
761 	struct mwl_hal *mh = sc->sc_mh;
762 	struct mwl_hal_vap *hvap = mvp->mv_hvap;
763 	enum ieee80211_opmode opmode = vap->iv_opmode;
764 
765 	/* XXX disallow ap vap delete if WDS still present */
766 	if (parent->if_drv_flags & IFF_DRV_RUNNING) {
767 		/* quiesce h/w while we remove the vap */
768 		mwl_hal_intrset(mh, 0);		/* disable interrupts */
769 	}
770 	ieee80211_vap_detach(vap);
771 	switch (opmode) {
772 	case IEEE80211_M_HOSTAP:
773 	case IEEE80211_M_MBSS:
774 	case IEEE80211_M_STA:
775 		KASSERT(hvap != NULL, ("no hal vap handle"));
776 		(void) mwl_hal_delstation(hvap, vap->iv_myaddr);
777 		mwl_hal_delvap(hvap);
778 		if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS)
779 			sc->sc_napvaps--;
780 		else
781 			sc->sc_nstavaps--;
782 		/* XXX don't do it for IEEE80211_CLONE_MACADDR */
783 		reclaim_address(sc, vap->iv_myaddr);
784 		break;
785 	case IEEE80211_M_WDS:
786 		sc->sc_nwdsvaps--;
787 		break;
788 	default:
789 		break;
790 	}
791 	mwl_cleartxq(sc, vap);
792 	free(mvp, M_80211_VAP);
793 	if (parent->if_drv_flags & IFF_DRV_RUNNING)
794 		mwl_hal_intrset(mh, sc->sc_imask);
795 }
796 
797 void
798 mwl_suspend(struct mwl_softc *sc)
799 {
800 	struct ifnet *ifp = sc->sc_ifp;
801 
802 	DPRINTF(sc, MWL_DEBUG_ANY, "%s: if_flags %x\n",
803 		__func__, ifp->if_flags);
804 
805 	mwl_stop(ifp, 1);
806 }
807 
808 void
809 mwl_resume(struct mwl_softc *sc)
810 {
811 	struct ifnet *ifp = sc->sc_ifp;
812 
813 	DPRINTF(sc, MWL_DEBUG_ANY, "%s: if_flags %x\n",
814 		__func__, ifp->if_flags);
815 
816 	if (ifp->if_flags & IFF_UP)
817 		mwl_init(sc);
818 }
819 
820 void
821 mwl_shutdown(void *arg)
822 {
823 	struct mwl_softc *sc = arg;
824 
825 	mwl_stop(sc->sc_ifp, 1);
826 }
827 
828 /*
829  * Interrupt handler.  Most of the actual processing is deferred.
830  */
831 void
832 mwl_intr(void *arg)
833 {
834 	struct mwl_softc *sc = arg;
835 	struct mwl_hal *mh = sc->sc_mh;
836 	uint32_t status;
837 
838 	if (sc->sc_invalid) {
839 		/*
840 		 * The hardware is not ready/present, don't touch anything.
841 		 * Note this can happen early on if the IRQ is shared.
842 		 */
843 		DPRINTF(sc, MWL_DEBUG_ANY, "%s: invalid; ignored\n", __func__);
844 		return;
845 	}
846 	/*
847 	 * Figure out the reason(s) for the interrupt.
848 	 */
849 	mwl_hal_getisr(mh, &status);		/* NB: clears ISR too */
850 	if (status == 0)			/* must be a shared irq */
851 		return;
852 
853 	DPRINTF(sc, MWL_DEBUG_INTR, "%s: status 0x%x imask 0x%x\n",
854 	    __func__, status, sc->sc_imask);
855 	if (status & MACREG_A2HRIC_BIT_RX_RDY)
856 		taskqueue_enqueue(sc->sc_tq, &sc->sc_rxtask);
857 	if (status & MACREG_A2HRIC_BIT_TX_DONE)
858 		taskqueue_enqueue(sc->sc_tq, &sc->sc_txtask);
859 	if (status & MACREG_A2HRIC_BIT_BA_WATCHDOG)
860 		taskqueue_enqueue(sc->sc_tq, &sc->sc_bawatchdogtask);
861 	if (status & MACREG_A2HRIC_BIT_OPC_DONE)
862 		mwl_hal_cmddone(mh);
863 	if (status & MACREG_A2HRIC_BIT_MAC_EVENT) {
864 		;
865 	}
866 	if (status & MACREG_A2HRIC_BIT_ICV_ERROR) {
867 		/* TKIP ICV error */
868 		sc->sc_stats.mst_rx_badtkipicv++;
869 	}
870 	if (status & MACREG_A2HRIC_BIT_QUEUE_EMPTY) {
871 		/* 11n aggregation queue is empty, re-fill */
872 		;
873 	}
874 	if (status & MACREG_A2HRIC_BIT_QUEUE_FULL) {
875 		;
876 	}
877 	if (status & MACREG_A2HRIC_BIT_RADAR_DETECT) {
878 		/* radar detected, process event */
879 		taskqueue_enqueue(sc->sc_tq, &sc->sc_radartask);
880 	}
881 	if (status & MACREG_A2HRIC_BIT_CHAN_SWITCH) {
882 		/* DFS channel switch */
883 		taskqueue_enqueue(sc->sc_tq, &sc->sc_chanswitchtask);
884 	}
885 }
886 
887 static void
888 mwl_radar_proc(void *arg, int pending)
889 {
890 	struct mwl_softc *sc = arg;
891 	struct ifnet *ifp = sc->sc_ifp;
892 	struct ieee80211com *ic = ifp->if_l2com;
893 
894 	DPRINTF(sc, MWL_DEBUG_ANY, "%s: radar detected, pending %u\n",
895 	    __func__, pending);
896 
897 	sc->sc_stats.mst_radardetect++;
898 	/* XXX stop h/w BA streams? */
899 
900 	IEEE80211_LOCK(ic);
901 	ieee80211_dfs_notify_radar(ic, ic->ic_curchan);
902 	IEEE80211_UNLOCK(ic);
903 }
904 
905 static void
906 mwl_chanswitch_proc(void *arg, int pending)
907 {
908 	struct mwl_softc *sc = arg;
909 	struct ifnet *ifp = sc->sc_ifp;
910 	struct ieee80211com *ic = ifp->if_l2com;
911 
912 	DPRINTF(sc, MWL_DEBUG_ANY, "%s: channel switch notice, pending %u\n",
913 	    __func__, pending);
914 
915 	IEEE80211_LOCK(ic);
916 	sc->sc_csapending = 0;
917 	ieee80211_csa_completeswitch(ic);
918 	IEEE80211_UNLOCK(ic);
919 }
920 
921 static void
922 mwl_bawatchdog(const MWL_HAL_BASTREAM *sp)
923 {
924 	struct ieee80211_node *ni = sp->data[0];
925 
926 	/* send DELBA and drop the stream */
927 	ieee80211_ampdu_stop(ni, sp->data[1], IEEE80211_REASON_UNSPECIFIED);
928 }
929 
930 static void
931 mwl_bawatchdog_proc(void *arg, int pending)
932 {
933 	struct mwl_softc *sc = arg;
934 	struct mwl_hal *mh = sc->sc_mh;
935 	const MWL_HAL_BASTREAM *sp;
936 	uint8_t bitmap, n;
937 
938 	sc->sc_stats.mst_bawatchdog++;
939 
940 	if (mwl_hal_getwatchdogbitmap(mh, &bitmap) != 0) {
941 		DPRINTF(sc, MWL_DEBUG_AMPDU,
942 		    "%s: could not get bitmap\n", __func__);
943 		sc->sc_stats.mst_bawatchdog_failed++;
944 		return;
945 	}
946 	DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: bitmap 0x%x\n", __func__, bitmap);
947 	if (bitmap == 0xff) {
948 		n = 0;
949 		/* disable all ba streams */
950 		for (bitmap = 0; bitmap < 8; bitmap++) {
951 			sp = mwl_hal_bastream_lookup(mh, bitmap);
952 			if (sp != NULL) {
953 				mwl_bawatchdog(sp);
954 				n++;
955 			}
956 		}
957 		if (n == 0) {
958 			DPRINTF(sc, MWL_DEBUG_AMPDU,
959 			    "%s: no BA streams found\n", __func__);
960 			sc->sc_stats.mst_bawatchdog_empty++;
961 		}
962 	} else if (bitmap != 0xaa) {
963 		/* disable a single ba stream */
964 		sp = mwl_hal_bastream_lookup(mh, bitmap);
965 		if (sp != NULL) {
966 			mwl_bawatchdog(sp);
967 		} else {
968 			DPRINTF(sc, MWL_DEBUG_AMPDU,
969 			    "%s: no BA stream %d\n", __func__, bitmap);
970 			sc->sc_stats.mst_bawatchdog_notfound++;
971 		}
972 	}
973 }
974 
975 /*
976  * Convert net80211 channel to a HAL channel.
977  */
978 static void
979 mwl_mapchan(MWL_HAL_CHANNEL *hc, const struct ieee80211_channel *chan)
980 {
981 	hc->channel = chan->ic_ieee;
982 
983 	*(uint32_t *)&hc->channelFlags = 0;
984 	if (IEEE80211_IS_CHAN_2GHZ(chan))
985 		hc->channelFlags.FreqBand = MWL_FREQ_BAND_2DOT4GHZ;
986 	else if (IEEE80211_IS_CHAN_5GHZ(chan))
987 		hc->channelFlags.FreqBand = MWL_FREQ_BAND_5GHZ;
988 	if (IEEE80211_IS_CHAN_HT40(chan)) {
989 		hc->channelFlags.ChnlWidth = MWL_CH_40_MHz_WIDTH;
990 		if (IEEE80211_IS_CHAN_HT40U(chan))
991 			hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_ABOVE_CTRL_CH;
992 		else
993 			hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_BELOW_CTRL_CH;
994 	} else
995 		hc->channelFlags.ChnlWidth = MWL_CH_20_MHz_WIDTH;
996 	/* XXX 10MHz channels */
997 }
998 
999 /*
1000  * Inform firmware of our tx/rx dma setup.  The BAR 0
1001  * writes below are for compatibility with older firmware.
1002  * For current firmware we send this information with a
1003  * cmd block via mwl_hal_sethwdma.
1004  */
1005 static int
1006 mwl_setupdma(struct mwl_softc *sc)
1007 {
1008 	int error, i;
1009 
1010 	sc->sc_hwdma.rxDescRead = sc->sc_rxdma.dd_desc_paddr;
1011 	WR4(sc, sc->sc_hwspecs.rxDescRead, sc->sc_hwdma.rxDescRead);
1012 	WR4(sc, sc->sc_hwspecs.rxDescWrite, sc->sc_hwdma.rxDescRead);
1013 
1014 	for (i = 0; i < MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES; i++) {
1015 		struct mwl_txq *txq = &sc->sc_txq[i];
1016 		sc->sc_hwdma.wcbBase[i] = txq->dma.dd_desc_paddr;
1017 		WR4(sc, sc->sc_hwspecs.wcbBase[i], sc->sc_hwdma.wcbBase[i]);
1018 	}
1019 	sc->sc_hwdma.maxNumTxWcb = mwl_txbuf;
1020 	sc->sc_hwdma.maxNumWCB = MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES;
1021 
1022 	error = mwl_hal_sethwdma(sc->sc_mh, &sc->sc_hwdma);
1023 	if (error != 0) {
1024 		device_printf(sc->sc_dev,
1025 		    "unable to setup tx/rx dma; hal status %u\n", error);
1026 		/* XXX */
1027 	}
1028 	return error;
1029 }
1030 
1031 /*
1032  * Inform firmware of tx rate parameters.
1033  * Called after a channel change.
1034  */
1035 static int
1036 mwl_setcurchanrates(struct mwl_softc *sc)
1037 {
1038 	struct ifnet *ifp = sc->sc_ifp;
1039 	struct ieee80211com *ic = ifp->if_l2com;
1040 	const struct ieee80211_rateset *rs;
1041 	MWL_HAL_TXRATE rates;
1042 
1043 	memset(&rates, 0, sizeof(rates));
1044 	rs = ieee80211_get_suprates(ic, ic->ic_curchan);
1045 	/* rate used to send management frames */
1046 	rates.MgtRate = rs->rs_rates[0] & IEEE80211_RATE_VAL;
1047 	/* rate used to send multicast frames */
1048 	rates.McastRate = rates.MgtRate;
1049 
1050 	return mwl_hal_settxrate_auto(sc->sc_mh, &rates);
1051 }
1052 
1053 /*
1054  * Inform firmware of tx rate parameters.  Called whenever
1055  * user-settable params change and after a channel change.
1056  */
1057 static int
1058 mwl_setrates(struct ieee80211vap *vap)
1059 {
1060 	struct mwl_vap *mvp = MWL_VAP(vap);
1061 	struct ieee80211_node *ni = vap->iv_bss;
1062 	const struct ieee80211_txparam *tp = ni->ni_txparms;
1063 	MWL_HAL_TXRATE rates;
1064 
1065 	KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state));
1066 
1067 	/*
1068 	 * Update the h/w rate map.
1069 	 * NB: 0x80 for MCS is passed through unchanged
1070 	 */
1071 	memset(&rates, 0, sizeof(rates));
1072 	/* rate used to send management frames */
1073 	rates.MgtRate = tp->mgmtrate;
1074 	/* rate used to send multicast frames */
1075 	rates.McastRate = tp->mcastrate;
1076 
1077 	/* while here calculate EAPOL fixed rate cookie */
1078 	mvp->mv_eapolformat = htole16(mwl_calcformat(rates.MgtRate, ni));
1079 
1080 	return mwl_hal_settxrate(mvp->mv_hvap,
1081 	    tp->ucastrate != IEEE80211_FIXED_RATE_NONE ?
1082 		RATE_FIXED : RATE_AUTO, &rates);
1083 }
1084 
1085 /*
1086  * Setup a fixed xmit rate cookie for EAPOL frames.
1087  */
1088 static void
1089 mwl_seteapolformat(struct ieee80211vap *vap)
1090 {
1091 	struct mwl_vap *mvp = MWL_VAP(vap);
1092 	struct ieee80211_node *ni = vap->iv_bss;
1093 	enum ieee80211_phymode mode;
1094 	uint8_t rate;
1095 
1096 	KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state));
1097 
1098 	mode = ieee80211_chan2mode(ni->ni_chan);
1099 	/*
1100 	 * Use legacy rates when operating a mixed HT+non-HT bss.
1101 	 * NB: this may violate POLA for sta and wds vap's.
1102 	 */
1103 	if (mode == IEEE80211_MODE_11NA &&
1104 	    (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0)
1105 		rate = vap->iv_txparms[IEEE80211_MODE_11A].mgmtrate;
1106 	else if (mode == IEEE80211_MODE_11NG &&
1107 	    (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0)
1108 		rate = vap->iv_txparms[IEEE80211_MODE_11G].mgmtrate;
1109 	else
1110 		rate = vap->iv_txparms[mode].mgmtrate;
1111 
1112 	mvp->mv_eapolformat = htole16(mwl_calcformat(rate, ni));
1113 }
1114 
1115 /*
1116  * Map SKU+country code to region code for radar bin'ing.
1117  */
1118 static int
1119 mwl_map2regioncode(const struct ieee80211_regdomain *rd)
1120 {
1121 	switch (rd->regdomain) {
1122 	case SKU_FCC:
1123 	case SKU_FCC3:
1124 		return DOMAIN_CODE_FCC;
1125 	case SKU_CA:
1126 		return DOMAIN_CODE_IC;
1127 	case SKU_ETSI:
1128 	case SKU_ETSI2:
1129 	case SKU_ETSI3:
1130 		if (rd->country == CTRY_SPAIN)
1131 			return DOMAIN_CODE_SPAIN;
1132 		if (rd->country == CTRY_FRANCE || rd->country == CTRY_FRANCE2)
1133 			return DOMAIN_CODE_FRANCE;
1134 		/* XXX force 1.3.1 radar type */
1135 		return DOMAIN_CODE_ETSI_131;
1136 	case SKU_JAPAN:
1137 		return DOMAIN_CODE_MKK;
1138 	case SKU_ROW:
1139 		return DOMAIN_CODE_DGT;	/* Taiwan */
1140 	case SKU_APAC:
1141 	case SKU_APAC2:
1142 	case SKU_APAC3:
1143 		return DOMAIN_CODE_AUS;	/* Australia */
1144 	}
1145 	/* XXX KOREA? */
1146 	return DOMAIN_CODE_FCC;			/* XXX? */
1147 }
1148 
1149 static int
1150 mwl_hal_reset(struct mwl_softc *sc)
1151 {
1152 	struct ifnet *ifp = sc->sc_ifp;
1153 	struct ieee80211com *ic = ifp->if_l2com;
1154 	struct mwl_hal *mh = sc->sc_mh;
1155 
1156 	mwl_hal_setantenna(mh, WL_ANTENNATYPE_RX, sc->sc_rxantenna);
1157 	mwl_hal_setantenna(mh, WL_ANTENNATYPE_TX, sc->sc_txantenna);
1158 	mwl_hal_setradio(mh, 1, WL_AUTO_PREAMBLE);
1159 	mwl_hal_setwmm(sc->sc_mh, (ic->ic_flags & IEEE80211_F_WME) != 0);
1160 	mwl_chan_set(sc, ic->ic_curchan);
1161 	/* NB: RF/RA performance tuned for indoor mode */
1162 	mwl_hal_setrateadaptmode(mh, 0);
1163 	mwl_hal_setoptimizationlevel(mh,
1164 	    (ic->ic_flags & IEEE80211_F_BURST) != 0);
1165 
1166 	mwl_hal_setregioncode(mh, mwl_map2regioncode(&ic->ic_regdomain));
1167 
1168 	mwl_hal_setaggampduratemode(mh, 1, 80);		/* XXX */
1169 	mwl_hal_setcfend(mh, 0);			/* XXX */
1170 
1171 	return 1;
1172 }
1173 
1174 static int
1175 mwl_init_locked(struct mwl_softc *sc)
1176 {
1177 	struct ifnet *ifp = sc->sc_ifp;
1178 	struct mwl_hal *mh = sc->sc_mh;
1179 	int error = 0;
1180 
1181 	DPRINTF(sc, MWL_DEBUG_ANY, "%s: if_flags 0x%x\n",
1182 		__func__, ifp->if_flags);
1183 
1184 	MWL_LOCK_ASSERT(sc);
1185 
1186 	/*
1187 	 * Stop anything previously setup.  This is safe
1188 	 * whether this is the first time through or not.
1189 	 */
1190 	mwl_stop_locked(ifp, 0);
1191 
1192 	/*
1193 	 * Push vap-independent state to the firmware.
1194 	 */
1195 	if (!mwl_hal_reset(sc)) {
1196 		if_printf(ifp, "unable to reset hardware\n");
1197 		return EIO;
1198 	}
1199 
1200 	/*
1201 	 * Setup recv (once); transmit is already good to go.
1202 	 */
1203 	error = mwl_startrecv(sc);
1204 	if (error != 0) {
1205 		if_printf(ifp, "unable to start recv logic\n");
1206 		return error;
1207 	}
1208 
1209 	/*
1210 	 * Enable interrupts.
1211 	 */
1212 	sc->sc_imask = MACREG_A2HRIC_BIT_RX_RDY
1213 		     | MACREG_A2HRIC_BIT_TX_DONE
1214 		     | MACREG_A2HRIC_BIT_OPC_DONE
1215 #if 0
1216 		     | MACREG_A2HRIC_BIT_MAC_EVENT
1217 #endif
1218 		     | MACREG_A2HRIC_BIT_ICV_ERROR
1219 		     | MACREG_A2HRIC_BIT_RADAR_DETECT
1220 		     | MACREG_A2HRIC_BIT_CHAN_SWITCH
1221 #if 0
1222 		     | MACREG_A2HRIC_BIT_QUEUE_EMPTY
1223 #endif
1224 		     | MACREG_A2HRIC_BIT_BA_WATCHDOG
1225 		     | MACREQ_A2HRIC_BIT_TX_ACK
1226 		     ;
1227 
1228 	ifp->if_drv_flags |= IFF_DRV_RUNNING;
1229 	mwl_hal_intrset(mh, sc->sc_imask);
1230 	callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc);
1231 
1232 	return 0;
1233 }
1234 
1235 static void
1236 mwl_init(void *arg)
1237 {
1238 	struct mwl_softc *sc = arg;
1239 	struct ifnet *ifp = sc->sc_ifp;
1240 	struct ieee80211com *ic = ifp->if_l2com;
1241 	int error = 0;
1242 
1243 	DPRINTF(sc, MWL_DEBUG_ANY, "%s: if_flags 0x%x\n",
1244 		__func__, ifp->if_flags);
1245 
1246 	MWL_LOCK(sc);
1247 	error = mwl_init_locked(sc);
1248 	MWL_UNLOCK(sc);
1249 
1250 	if (error == 0)
1251 		ieee80211_start_all(ic);	/* start all vap's */
1252 }
1253 
1254 static void
1255 mwl_stop_locked(struct ifnet *ifp, int disable)
1256 {
1257 	struct mwl_softc *sc = ifp->if_softc;
1258 
1259 	DPRINTF(sc, MWL_DEBUG_ANY, "%s: invalid %u if_flags 0x%x\n",
1260 		__func__, sc->sc_invalid, ifp->if_flags);
1261 
1262 	MWL_LOCK_ASSERT(sc);
1263 	if (ifp->if_drv_flags & IFF_DRV_RUNNING) {
1264 		/*
1265 		 * Shutdown the hardware and driver.
1266 		 */
1267 		ifp->if_drv_flags &= ~IFF_DRV_RUNNING;
1268 		callout_stop(&sc->sc_watchdog);
1269 		sc->sc_tx_timer = 0;
1270 		mwl_draintxq(sc);
1271 	}
1272 }
1273 
1274 static void
1275 mwl_stop(struct ifnet *ifp, int disable)
1276 {
1277 	struct mwl_softc *sc = ifp->if_softc;
1278 
1279 	MWL_LOCK(sc);
1280 	mwl_stop_locked(ifp, disable);
1281 	MWL_UNLOCK(sc);
1282 }
1283 
1284 static int
1285 mwl_reset_vap(struct ieee80211vap *vap, int state)
1286 {
1287 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1288 	struct ieee80211com *ic = vap->iv_ic;
1289 
1290 	if (state == IEEE80211_S_RUN)
1291 		mwl_setrates(vap);
1292 	/* XXX off by 1? */
1293 	mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold);
1294 	/* XXX auto? 20/40 split? */
1295 	mwl_hal_sethtgi(hvap, (vap->iv_flags_ht &
1296 	    (IEEE80211_FHT_SHORTGI20|IEEE80211_FHT_SHORTGI40)) ? 1 : 0);
1297 	mwl_hal_setnprot(hvap, ic->ic_htprotmode == IEEE80211_PROT_NONE ?
1298 	    HTPROTECT_NONE : HTPROTECT_AUTO);
1299 	/* XXX txpower cap */
1300 
1301 	/* re-setup beacons */
1302 	if (state == IEEE80211_S_RUN &&
1303 	    (vap->iv_opmode == IEEE80211_M_HOSTAP ||
1304 	     vap->iv_opmode == IEEE80211_M_MBSS ||
1305 	     vap->iv_opmode == IEEE80211_M_IBSS)) {
1306 		mwl_setapmode(vap, vap->iv_bss->ni_chan);
1307 		mwl_hal_setnprotmode(hvap,
1308 		    MS(ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE));
1309 		return mwl_beacon_setup(vap);
1310 	}
1311 	return 0;
1312 }
1313 
1314 /*
1315  * Reset the hardware w/o losing operational state.
1316  * Used to to reset or reload hardware state for a vap.
1317  */
1318 static int
1319 mwl_reset(struct ieee80211vap *vap, u_long cmd)
1320 {
1321 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1322 	int error = 0;
1323 
1324 	if (hvap != NULL) {			/* WDS, MONITOR, etc. */
1325 		struct ieee80211com *ic = vap->iv_ic;
1326 		struct ifnet *ifp = ic->ic_ifp;
1327 		struct mwl_softc *sc = ifp->if_softc;
1328 		struct mwl_hal *mh = sc->sc_mh;
1329 
1330 		/* XXX handle DWDS sta vap change */
1331 		/* XXX do we need to disable interrupts? */
1332 		mwl_hal_intrset(mh, 0);		/* disable interrupts */
1333 		error = mwl_reset_vap(vap, vap->iv_state);
1334 		mwl_hal_intrset(mh, sc->sc_imask);
1335 	}
1336 	return error;
1337 }
1338 
1339 /*
1340  * Allocate a tx buffer for sending a frame.  The
1341  * packet is assumed to have the WME AC stored so
1342  * we can use it to select the appropriate h/w queue.
1343  */
1344 static struct mwl_txbuf *
1345 mwl_gettxbuf(struct mwl_softc *sc, struct mwl_txq *txq)
1346 {
1347 	struct mwl_txbuf *bf;
1348 
1349 	/*
1350 	 * Grab a TX buffer and associated resources.
1351 	 */
1352 	MWL_TXQ_LOCK(txq);
1353 	bf = STAILQ_FIRST(&txq->free);
1354 	if (bf != NULL) {
1355 		STAILQ_REMOVE_HEAD(&txq->free, bf_list);
1356 		txq->nfree--;
1357 	}
1358 	MWL_TXQ_UNLOCK(txq);
1359 	if (bf == NULL)
1360 		DPRINTF(sc, MWL_DEBUG_XMIT,
1361 		    "%s: out of xmit buffers on q %d\n", __func__, txq->qnum);
1362 	return bf;
1363 }
1364 
1365 /*
1366  * Return a tx buffer to the queue it came from.  Note there
1367  * are two cases because we must preserve the order of buffers
1368  * as it reflects the fixed order of descriptors in memory
1369  * (the firmware pre-fetches descriptors so we cannot reorder).
1370  */
1371 static void
1372 mwl_puttxbuf_head(struct mwl_txq *txq, struct mwl_txbuf *bf)
1373 {
1374 	bf->bf_m = NULL;
1375 	bf->bf_node = NULL;
1376 	MWL_TXQ_LOCK(txq);
1377 	STAILQ_INSERT_HEAD(&txq->free, bf, bf_list);
1378 	txq->nfree++;
1379 	MWL_TXQ_UNLOCK(txq);
1380 }
1381 
1382 static void
1383 mwl_puttxbuf_tail(struct mwl_txq *txq, struct mwl_txbuf *bf)
1384 {
1385 	bf->bf_m = NULL;
1386 	bf->bf_node = NULL;
1387 	MWL_TXQ_LOCK(txq);
1388 	STAILQ_INSERT_TAIL(&txq->free, bf, bf_list);
1389 	txq->nfree++;
1390 	MWL_TXQ_UNLOCK(txq);
1391 }
1392 
1393 static void
1394 mwl_start(struct ifnet *ifp)
1395 {
1396 	struct mwl_softc *sc = ifp->if_softc;
1397 	struct ieee80211_node *ni;
1398 	struct mwl_txbuf *bf;
1399 	struct mbuf *m;
1400 	struct mwl_txq *txq = NULL;	/* XXX silence gcc */
1401 	int nqueued;
1402 
1403 	if ((ifp->if_drv_flags & IFF_DRV_RUNNING) == 0 || sc->sc_invalid)
1404 		return;
1405 	nqueued = 0;
1406 	for (;;) {
1407 		bf = NULL;
1408 		IFQ_DEQUEUE(&ifp->if_snd, m);
1409 		if (m == NULL)
1410 			break;
1411 		/*
1412 		 * Grab the node for the destination.
1413 		 */
1414 		ni = (struct ieee80211_node *) m->m_pkthdr.rcvif;
1415 		KASSERT(ni != NULL, ("no node"));
1416 		m->m_pkthdr.rcvif = NULL;	/* committed, clear ref */
1417 		/*
1418 		 * Grab a TX buffer and associated resources.
1419 		 * We honor the classification by the 802.11 layer.
1420 		 */
1421 		txq = sc->sc_ac2q[M_WME_GETAC(m)];
1422 		bf = mwl_gettxbuf(sc, txq);
1423 		if (bf == NULL) {
1424 			m_freem(m);
1425 			ieee80211_free_node(ni);
1426 #ifdef MWL_TX_NODROP
1427 			sc->sc_stats.mst_tx_qstop++;
1428 			/* XXX blocks other traffic */
1429 			ifp->if_drv_flags |= IFF_DRV_OACTIVE;
1430 			break;
1431 #else
1432 			DPRINTF(sc, MWL_DEBUG_XMIT,
1433 			    "%s: tail drop on q %d\n", __func__, txq->qnum);
1434 			sc->sc_stats.mst_tx_qdrop++;
1435 			continue;
1436 #endif /* MWL_TX_NODROP */
1437 		}
1438 
1439 		/*
1440 		 * Pass the frame to the h/w for transmission.
1441 		 */
1442 		if (mwl_tx_start(sc, ni, bf, m)) {
1443 			ifp->if_oerrors++;
1444 			mwl_puttxbuf_head(txq, bf);
1445 			ieee80211_free_node(ni);
1446 			continue;
1447 		}
1448 		nqueued++;
1449 		if (nqueued >= mwl_txcoalesce) {
1450 			/*
1451 			 * Poke the firmware to process queued frames;
1452 			 * see below about (lack of) locking.
1453 			 */
1454 			nqueued = 0;
1455 			mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1456 		}
1457 	}
1458 	if (nqueued) {
1459 		/*
1460 		 * NB: We don't need to lock against tx done because
1461 		 * this just prods the firmware to check the transmit
1462 		 * descriptors.  The firmware will also start fetching
1463 		 * descriptors by itself if it notices new ones are
1464 		 * present when it goes to deliver a tx done interrupt
1465 		 * to the host. So if we race with tx done processing
1466 		 * it's ok.  Delivering the kick here rather than in
1467 		 * mwl_tx_start is an optimization to avoid poking the
1468 		 * firmware for each packet.
1469 		 *
1470 		 * NB: the queue id isn't used so 0 is ok.
1471 		 */
1472 		mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1473 	}
1474 }
1475 
1476 static int
1477 mwl_raw_xmit(struct ieee80211_node *ni, struct mbuf *m,
1478 	const struct ieee80211_bpf_params *params)
1479 {
1480 	struct ieee80211com *ic = ni->ni_ic;
1481 	struct ifnet *ifp = ic->ic_ifp;
1482 	struct mwl_softc *sc = ifp->if_softc;
1483 	struct mwl_txbuf *bf;
1484 	struct mwl_txq *txq;
1485 
1486 	if ((ifp->if_drv_flags & IFF_DRV_RUNNING) == 0 || sc->sc_invalid) {
1487 		ieee80211_free_node(ni);
1488 		m_freem(m);
1489 		return ENETDOWN;
1490 	}
1491 	/*
1492 	 * Grab a TX buffer and associated resources.
1493 	 * Note that we depend on the classification
1494 	 * by the 802.11 layer to get to the right h/w
1495 	 * queue.  Management frames must ALWAYS go on
1496 	 * queue 1 but we cannot just force that here
1497 	 * because we may receive non-mgt frames.
1498 	 */
1499 	txq = sc->sc_ac2q[M_WME_GETAC(m)];
1500 	bf = mwl_gettxbuf(sc, txq);
1501 	if (bf == NULL) {
1502 		sc->sc_stats.mst_tx_qstop++;
1503 		/* XXX blocks other traffic */
1504 		ifp->if_drv_flags |= IFF_DRV_OACTIVE;
1505 		ieee80211_free_node(ni);
1506 		m_freem(m);
1507 		return ENOBUFS;
1508 	}
1509 	/*
1510 	 * Pass the frame to the h/w for transmission.
1511 	 */
1512 	if (mwl_tx_start(sc, ni, bf, m)) {
1513 		ifp->if_oerrors++;
1514 		mwl_puttxbuf_head(txq, bf);
1515 
1516 		ieee80211_free_node(ni);
1517 		return EIO;		/* XXX */
1518 	}
1519 	/*
1520 	 * NB: We don't need to lock against tx done because
1521 	 * this just prods the firmware to check the transmit
1522 	 * descriptors.  The firmware will also start fetching
1523 	 * descriptors by itself if it notices new ones are
1524 	 * present when it goes to deliver a tx done interrupt
1525 	 * to the host. So if we race with tx done processing
1526 	 * it's ok.  Delivering the kick here rather than in
1527 	 * mwl_tx_start is an optimization to avoid poking the
1528 	 * firmware for each packet.
1529 	 *
1530 	 * NB: the queue id isn't used so 0 is ok.
1531 	 */
1532 	mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1533 	return 0;
1534 }
1535 
1536 static int
1537 mwl_media_change(struct ifnet *ifp)
1538 {
1539 	struct ieee80211vap *vap = ifp->if_softc;
1540 	int error;
1541 
1542 	error = ieee80211_media_change(ifp);
1543 	/* NB: only the fixed rate can change and that doesn't need a reset */
1544 	if (error == ENETRESET) {
1545 		mwl_setrates(vap);
1546 		error = 0;
1547 	}
1548 	return error;
1549 }
1550 
1551 #ifdef MWL_DEBUG
1552 static void
1553 mwl_keyprint(struct mwl_softc *sc, const char *tag,
1554 	const MWL_HAL_KEYVAL *hk, const uint8_t mac[IEEE80211_ADDR_LEN])
1555 {
1556 	static const char *ciphers[] = {
1557 		"WEP",
1558 		"TKIP",
1559 		"AES-CCM",
1560 	};
1561 	int i, n;
1562 
1563 	printf("%s: [%u] %-7s", tag, hk->keyIndex, ciphers[hk->keyTypeId]);
1564 	for (i = 0, n = hk->keyLen; i < n; i++)
1565 		printf(" %02x", hk->key.aes[i]);
1566 	printf(" mac %s", ether_sprintf(mac));
1567 	if (hk->keyTypeId == KEY_TYPE_ID_TKIP) {
1568 		printf(" %s", "rxmic");
1569 		for (i = 0; i < sizeof(hk->key.tkip.rxMic); i++)
1570 			printf(" %02x", hk->key.tkip.rxMic[i]);
1571 		printf(" txmic");
1572 		for (i = 0; i < sizeof(hk->key.tkip.txMic); i++)
1573 			printf(" %02x", hk->key.tkip.txMic[i]);
1574 	}
1575 	printf(" flags 0x%x\n", hk->keyFlags);
1576 }
1577 #endif
1578 
1579 /*
1580  * Allocate a key cache slot for a unicast key.  The
1581  * firmware handles key allocation and every station is
1582  * guaranteed key space so we are always successful.
1583  */
1584 static int
1585 mwl_key_alloc(struct ieee80211vap *vap, struct ieee80211_key *k,
1586 	ieee80211_keyix *keyix, ieee80211_keyix *rxkeyix)
1587 {
1588 	struct mwl_softc *sc = vap->iv_ic->ic_ifp->if_softc;
1589 
1590 	if (k->wk_keyix != IEEE80211_KEYIX_NONE ||
1591 	    (k->wk_flags & IEEE80211_KEY_GROUP)) {
1592 		if (!(&vap->iv_nw_keys[0] <= k &&
1593 		      k < &vap->iv_nw_keys[IEEE80211_WEP_NKID])) {
1594 			/* should not happen */
1595 			DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1596 				"%s: bogus group key\n", __func__);
1597 			return 0;
1598 		}
1599 		/* give the caller what they requested */
1600 		*keyix = *rxkeyix = k - vap->iv_nw_keys;
1601 	} else {
1602 		/*
1603 		 * Firmware handles key allocation.
1604 		 */
1605 		*keyix = *rxkeyix = 0;
1606 	}
1607 	return 1;
1608 }
1609 
1610 /*
1611  * Delete a key entry allocated by mwl_key_alloc.
1612  */
1613 static int
1614 mwl_key_delete(struct ieee80211vap *vap, const struct ieee80211_key *k)
1615 {
1616 	struct mwl_softc *sc = vap->iv_ic->ic_ifp->if_softc;
1617 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1618 	MWL_HAL_KEYVAL hk;
1619 	const uint8_t bcastaddr[IEEE80211_ADDR_LEN] =
1620 	    { 0xff, 0xff, 0xff, 0xff, 0xff, 0xff };
1621 
1622 	if (hvap == NULL) {
1623 		if (vap->iv_opmode != IEEE80211_M_WDS) {
1624 			/* XXX monitor mode? */
1625 			DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1626 			    "%s: no hvap for opmode %d\n", __func__,
1627 			    vap->iv_opmode);
1628 			return 0;
1629 		}
1630 		hvap = MWL_VAP(vap)->mv_ap_hvap;
1631 	}
1632 
1633 	DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: delete key %u\n",
1634 	    __func__, k->wk_keyix);
1635 
1636 	memset(&hk, 0, sizeof(hk));
1637 	hk.keyIndex = k->wk_keyix;
1638 	switch (k->wk_cipher->ic_cipher) {
1639 	case IEEE80211_CIPHER_WEP:
1640 		hk.keyTypeId = KEY_TYPE_ID_WEP;
1641 		break;
1642 	case IEEE80211_CIPHER_TKIP:
1643 		hk.keyTypeId = KEY_TYPE_ID_TKIP;
1644 		break;
1645 	case IEEE80211_CIPHER_AES_CCM:
1646 		hk.keyTypeId = KEY_TYPE_ID_AES;
1647 		break;
1648 	default:
1649 		/* XXX should not happen */
1650 		DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n",
1651 		    __func__, k->wk_cipher->ic_cipher);
1652 		return 0;
1653 	}
1654 	return (mwl_hal_keyreset(hvap, &hk, bcastaddr) == 0);	/*XXX*/
1655 }
1656 
1657 static __inline int
1658 addgroupflags(MWL_HAL_KEYVAL *hk, const struct ieee80211_key *k)
1659 {
1660 	if (k->wk_flags & IEEE80211_KEY_GROUP) {
1661 		if (k->wk_flags & IEEE80211_KEY_XMIT)
1662 			hk->keyFlags |= KEY_FLAG_TXGROUPKEY;
1663 		if (k->wk_flags & IEEE80211_KEY_RECV)
1664 			hk->keyFlags |= KEY_FLAG_RXGROUPKEY;
1665 		return 1;
1666 	} else
1667 		return 0;
1668 }
1669 
1670 /*
1671  * Set the key cache contents for the specified key.  Key cache
1672  * slot(s) must already have been allocated by mwl_key_alloc.
1673  */
1674 static int
1675 mwl_key_set(struct ieee80211vap *vap, const struct ieee80211_key *k,
1676 	const uint8_t mac[IEEE80211_ADDR_LEN])
1677 {
1678 #define	GRPXMIT	(IEEE80211_KEY_XMIT | IEEE80211_KEY_GROUP)
1679 /* NB: static wep keys are marked GROUP+tx/rx; GTK will be tx or rx */
1680 #define	IEEE80211_IS_STATICKEY(k) \
1681 	(((k)->wk_flags & (GRPXMIT|IEEE80211_KEY_RECV)) == \
1682 	 (GRPXMIT|IEEE80211_KEY_RECV))
1683 	struct mwl_softc *sc = vap->iv_ic->ic_ifp->if_softc;
1684 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1685 	const struct ieee80211_cipher *cip = k->wk_cipher;
1686 	const uint8_t *macaddr;
1687 	MWL_HAL_KEYVAL hk;
1688 
1689 	KASSERT((k->wk_flags & IEEE80211_KEY_SWCRYPT) == 0,
1690 		("s/w crypto set?"));
1691 
1692 	if (hvap == NULL) {
1693 		if (vap->iv_opmode != IEEE80211_M_WDS) {
1694 			/* XXX monitor mode? */
1695 			DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1696 			    "%s: no hvap for opmode %d\n", __func__,
1697 			    vap->iv_opmode);
1698 			return 0;
1699 		}
1700 		hvap = MWL_VAP(vap)->mv_ap_hvap;
1701 	}
1702 	memset(&hk, 0, sizeof(hk));
1703 	hk.keyIndex = k->wk_keyix;
1704 	switch (cip->ic_cipher) {
1705 	case IEEE80211_CIPHER_WEP:
1706 		hk.keyTypeId = KEY_TYPE_ID_WEP;
1707 		hk.keyLen = k->wk_keylen;
1708 		if (k->wk_keyix == vap->iv_def_txkey)
1709 			hk.keyFlags = KEY_FLAG_WEP_TXKEY;
1710 		if (!IEEE80211_IS_STATICKEY(k)) {
1711 			/* NB: WEP is never used for the PTK */
1712 			(void) addgroupflags(&hk, k);
1713 		}
1714 		break;
1715 	case IEEE80211_CIPHER_TKIP:
1716 		hk.keyTypeId = KEY_TYPE_ID_TKIP;
1717 		hk.key.tkip.tsc.high = (uint32_t)(k->wk_keytsc >> 16);
1718 		hk.key.tkip.tsc.low = (uint16_t)k->wk_keytsc;
1719 		hk.keyFlags = KEY_FLAG_TSC_VALID | KEY_FLAG_MICKEY_VALID;
1720 		hk.keyLen = k->wk_keylen + IEEE80211_MICBUF_SIZE;
1721 		if (!addgroupflags(&hk, k))
1722 			hk.keyFlags |= KEY_FLAG_PAIRWISE;
1723 		break;
1724 	case IEEE80211_CIPHER_AES_CCM:
1725 		hk.keyTypeId = KEY_TYPE_ID_AES;
1726 		hk.keyLen = k->wk_keylen;
1727 		if (!addgroupflags(&hk, k))
1728 			hk.keyFlags |= KEY_FLAG_PAIRWISE;
1729 		break;
1730 	default:
1731 		/* XXX should not happen */
1732 		DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n",
1733 		    __func__, k->wk_cipher->ic_cipher);
1734 		return 0;
1735 	}
1736 	/*
1737 	 * NB: tkip mic keys get copied here too; the layout
1738 	 *     just happens to match that in ieee80211_key.
1739 	 */
1740 	memcpy(hk.key.aes, k->wk_key, hk.keyLen);
1741 
1742 	/*
1743 	 * Locate address of sta db entry for writing key;
1744 	 * the convention unfortunately is somewhat different
1745 	 * than how net80211, hostapd, and wpa_supplicant think.
1746 	 */
1747 	if (vap->iv_opmode == IEEE80211_M_STA) {
1748 		/*
1749 		 * NB: keys plumbed before the sta reaches AUTH state
1750 		 * will be discarded or written to the wrong sta db
1751 		 * entry because iv_bss is meaningless.  This is ok
1752 		 * (right now) because we handle deferred plumbing of
1753 		 * WEP keys when the sta reaches AUTH state.
1754 		 */
1755 		macaddr = vap->iv_bss->ni_bssid;
1756 		if ((k->wk_flags & IEEE80211_KEY_GROUP) == 0) {
1757 			/* XXX plumb to local sta db too for static key wep */
1758 			mwl_hal_keyset(hvap, &hk, vap->iv_myaddr);
1759 		}
1760 	} else if (vap->iv_opmode == IEEE80211_M_WDS &&
1761 	    vap->iv_state != IEEE80211_S_RUN) {
1762 		/*
1763 		 * Prior to RUN state a WDS vap will not it's BSS node
1764 		 * setup so we will plumb the key to the wrong mac
1765 		 * address (it'll be our local address).  Workaround
1766 		 * this for the moment by grabbing the correct address.
1767 		 */
1768 		macaddr = vap->iv_des_bssid;
1769 	} else if ((k->wk_flags & GRPXMIT) == GRPXMIT)
1770 		macaddr = vap->iv_myaddr;
1771 	else
1772 		macaddr = mac;
1773 	KEYPRINTF(sc, &hk, macaddr);
1774 	return (mwl_hal_keyset(hvap, &hk, macaddr) == 0);
1775 #undef IEEE80211_IS_STATICKEY
1776 #undef GRPXMIT
1777 }
1778 
1779 /* unaligned little endian access */
1780 #define LE_READ_2(p)				\
1781 	((uint16_t)				\
1782 	 ((((const uint8_t *)(p))[0]      ) |	\
1783 	  (((const uint8_t *)(p))[1] <<  8)))
1784 #define LE_READ_4(p)				\
1785 	((uint32_t)				\
1786 	 ((((const uint8_t *)(p))[0]      ) |	\
1787 	  (((const uint8_t *)(p))[1] <<  8) |	\
1788 	  (((const uint8_t *)(p))[2] << 16) |	\
1789 	  (((const uint8_t *)(p))[3] << 24)))
1790 
1791 /*
1792  * Set the multicast filter contents into the hardware.
1793  * XXX f/w has no support; just defer to the os.
1794  */
1795 static void
1796 mwl_setmcastfilter(struct mwl_softc *sc)
1797 {
1798 	struct ifnet *ifp = sc->sc_ifp;
1799 #if 0
1800 	struct ether_multi *enm;
1801 	struct ether_multistep estep;
1802 	uint8_t macs[IEEE80211_ADDR_LEN*MWL_HAL_MCAST_MAX];/* XXX stack use */
1803 	uint8_t *mp;
1804 	int nmc;
1805 
1806 	mp = macs;
1807 	nmc = 0;
1808 	ETHER_FIRST_MULTI(estep, &sc->sc_ec, enm);
1809 	while (enm != NULL) {
1810 		/* XXX Punt on ranges. */
1811 		if (nmc == MWL_HAL_MCAST_MAX ||
1812 		    !IEEE80211_ADDR_EQ(enm->enm_addrlo, enm->enm_addrhi)) {
1813 			ifp->if_flags |= IFF_ALLMULTI;
1814 			return;
1815 		}
1816 		IEEE80211_ADDR_COPY(mp, enm->enm_addrlo);
1817 		mp += IEEE80211_ADDR_LEN, nmc++;
1818 		ETHER_NEXT_MULTI(estep, enm);
1819 	}
1820 	ifp->if_flags &= ~IFF_ALLMULTI;
1821 	mwl_hal_setmcast(sc->sc_mh, nmc, macs);
1822 #else
1823 	/* XXX no mcast filter support; we get everything */
1824 	ifp->if_flags |= IFF_ALLMULTI;
1825 #endif
1826 }
1827 
1828 static int
1829 mwl_mode_init(struct mwl_softc *sc)
1830 {
1831 	struct ifnet *ifp = sc->sc_ifp;
1832 	struct ieee80211com *ic = ifp->if_l2com;
1833 	struct mwl_hal *mh = sc->sc_mh;
1834 
1835 	/*
1836 	 * NB: Ignore promisc in hostap mode; it's set by the
1837 	 * bridge.  This is wrong but we have no way to
1838 	 * identify internal requests (from the bridge)
1839 	 * versus external requests such as for tcpdump.
1840 	 */
1841 	mwl_hal_setpromisc(mh, (ifp->if_flags & IFF_PROMISC) &&
1842 	    ic->ic_opmode != IEEE80211_M_HOSTAP);
1843 	mwl_setmcastfilter(sc);
1844 
1845 	return 0;
1846 }
1847 
1848 /*
1849  * Callback from the 802.11 layer after a multicast state change.
1850  */
1851 static void
1852 mwl_update_mcast(struct ifnet *ifp)
1853 {
1854 	struct mwl_softc *sc = ifp->if_softc;
1855 
1856 	mwl_setmcastfilter(sc);
1857 }
1858 
1859 /*
1860  * Callback from the 802.11 layer after a promiscuous mode change.
1861  * Note this interface does not check the operating mode as this
1862  * is an internal callback and we are expected to honor the current
1863  * state (e.g. this is used for setting the interface in promiscuous
1864  * mode when operating in hostap mode to do ACS).
1865  */
1866 static void
1867 mwl_update_promisc(struct ifnet *ifp)
1868 {
1869 	struct mwl_softc *sc = ifp->if_softc;
1870 
1871 	mwl_hal_setpromisc(sc->sc_mh, (ifp->if_flags & IFF_PROMISC) != 0);
1872 }
1873 
1874 /*
1875  * Callback from the 802.11 layer to update the slot time
1876  * based on the current setting.  We use it to notify the
1877  * firmware of ERP changes and the f/w takes care of things
1878  * like slot time and preamble.
1879  */
1880 static void
1881 mwl_updateslot(struct ifnet *ifp)
1882 {
1883 	struct mwl_softc *sc = ifp->if_softc;
1884 	struct ieee80211com *ic = ifp->if_l2com;
1885 	struct mwl_hal *mh = sc->sc_mh;
1886 	int prot;
1887 
1888 	/* NB: can be called early; suppress needless cmds */
1889 	if ((ifp->if_drv_flags & IFF_DRV_RUNNING) == 0)
1890 		return;
1891 
1892 	/*
1893 	 * Calculate the ERP flags.  The firwmare will use
1894 	 * this to carry out the appropriate measures.
1895 	 */
1896 	prot = 0;
1897 	if (IEEE80211_IS_CHAN_ANYG(ic->ic_curchan)) {
1898 		if ((ic->ic_flags & IEEE80211_F_SHSLOT) == 0)
1899 			prot |= IEEE80211_ERP_NON_ERP_PRESENT;
1900 		if (ic->ic_flags & IEEE80211_F_USEPROT)
1901 			prot |= IEEE80211_ERP_USE_PROTECTION;
1902 		if (ic->ic_flags & IEEE80211_F_USEBARKER)
1903 			prot |= IEEE80211_ERP_LONG_PREAMBLE;
1904 	}
1905 
1906 	DPRINTF(sc, MWL_DEBUG_RESET,
1907 	    "%s: chan %u MHz/flags 0x%x %s slot, (prot 0x%x ic_flags 0x%x)\n",
1908 	    __func__, ic->ic_curchan->ic_freq, ic->ic_curchan->ic_flags,
1909 	    ic->ic_flags & IEEE80211_F_SHSLOT ? "short" : "long", prot,
1910 	    ic->ic_flags);
1911 
1912 	mwl_hal_setgprot(mh, prot);
1913 }
1914 
1915 /*
1916  * Setup the beacon frame.
1917  */
1918 static int
1919 mwl_beacon_setup(struct ieee80211vap *vap)
1920 {
1921 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1922 	struct ieee80211_node *ni = vap->iv_bss;
1923 	struct ieee80211_beacon_offsets bo;
1924 	struct mbuf *m;
1925 
1926 	m = ieee80211_beacon_alloc(ni, &bo);
1927 	if (m == NULL)
1928 		return ENOBUFS;
1929 	mwl_hal_setbeacon(hvap, mtod(m, const void *), m->m_len);
1930 	m_free(m);
1931 
1932 	return 0;
1933 }
1934 
1935 /*
1936  * Update the beacon frame in response to a change.
1937  */
1938 static void
1939 mwl_beacon_update(struct ieee80211vap *vap, int item)
1940 {
1941 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1942 	struct ieee80211com *ic = vap->iv_ic;
1943 
1944 	KASSERT(hvap != NULL, ("no beacon"));
1945 	switch (item) {
1946 	case IEEE80211_BEACON_ERP:
1947 		mwl_updateslot(ic->ic_ifp);
1948 		break;
1949 	case IEEE80211_BEACON_HTINFO:
1950 		mwl_hal_setnprotmode(hvap,
1951 		    MS(ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE));
1952 		break;
1953 	case IEEE80211_BEACON_CAPS:
1954 	case IEEE80211_BEACON_WME:
1955 	case IEEE80211_BEACON_APPIE:
1956 	case IEEE80211_BEACON_CSA:
1957 		break;
1958 	case IEEE80211_BEACON_TIM:
1959 		/* NB: firmware always forms TIM */
1960 		return;
1961 	}
1962 	/* XXX retain beacon frame and update */
1963 	mwl_beacon_setup(vap);
1964 }
1965 
1966 static void
1967 mwl_load_cb(void *arg, bus_dma_segment_t *segs, int nsegs, int error)
1968 {
1969 	bus_addr_t *paddr = (bus_addr_t*) arg;
1970 	KASSERT(error == 0, ("error %u on bus_dma callback", error));
1971 	*paddr = segs->ds_addr;
1972 }
1973 
1974 #ifdef MWL_HOST_PS_SUPPORT
1975 /*
1976  * Handle power save station occupancy changes.
1977  */
1978 static void
1979 mwl_update_ps(struct ieee80211vap *vap, int nsta)
1980 {
1981 	struct mwl_vap *mvp = MWL_VAP(vap);
1982 
1983 	if (nsta == 0 || mvp->mv_last_ps_sta == 0)
1984 		mwl_hal_setpowersave_bss(mvp->mv_hvap, nsta);
1985 	mvp->mv_last_ps_sta = nsta;
1986 }
1987 
1988 /*
1989  * Handle associated station power save state changes.
1990  */
1991 static int
1992 mwl_set_tim(struct ieee80211_node *ni, int set)
1993 {
1994 	struct ieee80211vap *vap = ni->ni_vap;
1995 	struct mwl_vap *mvp = MWL_VAP(vap);
1996 
1997 	if (mvp->mv_set_tim(ni, set)) {		/* NB: state change */
1998 		mwl_hal_setpowersave_sta(mvp->mv_hvap,
1999 		    IEEE80211_AID(ni->ni_associd), set);
2000 		return 1;
2001 	} else
2002 		return 0;
2003 }
2004 #endif /* MWL_HOST_PS_SUPPORT */
2005 
2006 static int
2007 mwl_desc_setup(struct mwl_softc *sc, const char *name,
2008 	struct mwl_descdma *dd,
2009 	int nbuf, size_t bufsize, int ndesc, size_t descsize)
2010 {
2011 	struct ifnet *ifp = sc->sc_ifp;
2012 	uint8_t *ds;
2013 	int error;
2014 
2015 	DPRINTF(sc, MWL_DEBUG_RESET,
2016 	    "%s: %s DMA: %u bufs (%ju) %u desc/buf (%ju)\n",
2017 	    __func__, name, nbuf, (uintmax_t) bufsize,
2018 	    ndesc, (uintmax_t) descsize);
2019 
2020 	dd->dd_name = name;
2021 	dd->dd_desc_len = nbuf * ndesc * descsize;
2022 
2023 	/*
2024 	 * Setup DMA descriptor area.
2025 	 */
2026 	error = bus_dma_tag_create(bus_get_dma_tag(sc->sc_dev),	/* parent */
2027 		       PAGE_SIZE, 0,		/* alignment, bounds */
2028 		       BUS_SPACE_MAXADDR_32BIT,	/* lowaddr */
2029 		       BUS_SPACE_MAXADDR,	/* highaddr */
2030 		       NULL, NULL,		/* filter, filterarg */
2031 		       dd->dd_desc_len,		/* maxsize */
2032 		       1,			/* nsegments */
2033 		       dd->dd_desc_len,		/* maxsegsize */
2034 		       BUS_DMA_ALLOCNOW,	/* flags */
2035 		       NULL,			/* lockfunc */
2036 		       NULL,			/* lockarg */
2037 		       &dd->dd_dmat);
2038 	if (error != 0) {
2039 		if_printf(ifp, "cannot allocate %s DMA tag\n", dd->dd_name);
2040 		return error;
2041 	}
2042 
2043 	/* allocate descriptors */
2044 	error = bus_dmamem_alloc(dd->dd_dmat, (void**) &dd->dd_desc,
2045 				 BUS_DMA_NOWAIT | BUS_DMA_COHERENT,
2046 				 &dd->dd_dmamap);
2047 	if (error != 0) {
2048 		if_printf(ifp, "unable to alloc memory for %u %s descriptors, "
2049 			"error %u\n", nbuf * ndesc, dd->dd_name, error);
2050 		goto fail1;
2051 	}
2052 
2053 	error = bus_dmamap_load(dd->dd_dmat, dd->dd_dmamap,
2054 				dd->dd_desc, dd->dd_desc_len,
2055 				mwl_load_cb, &dd->dd_desc_paddr,
2056 				BUS_DMA_NOWAIT);
2057 	if (error != 0) {
2058 		if_printf(ifp, "unable to map %s descriptors, error %u\n",
2059 			dd->dd_name, error);
2060 		goto fail2;
2061 	}
2062 
2063 	ds = dd->dd_desc;
2064 	memset(ds, 0, dd->dd_desc_len);
2065 	DPRINTF(sc, MWL_DEBUG_RESET, "%s: %s DMA map: %p (%lu) -> %p (%lu)\n",
2066 	    __func__, dd->dd_name, ds, (u_long) dd->dd_desc_len,
2067 	    (caddr_t) dd->dd_desc_paddr, /*XXX*/ (u_long) dd->dd_desc_len);
2068 
2069 	return 0;
2070 fail2:
2071 	bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap);
2072 fail1:
2073 	bus_dma_tag_destroy(dd->dd_dmat);
2074 	memset(dd, 0, sizeof(*dd));
2075 	return error;
2076 #undef DS2PHYS
2077 }
2078 
2079 static void
2080 mwl_desc_cleanup(struct mwl_softc *sc, struct mwl_descdma *dd)
2081 {
2082 	bus_dmamap_unload(dd->dd_dmat, dd->dd_dmamap);
2083 	bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap);
2084 	bus_dma_tag_destroy(dd->dd_dmat);
2085 
2086 	memset(dd, 0, sizeof(*dd));
2087 }
2088 
2089 /*
2090  * Construct a tx q's free list.  The order of entries on
2091  * the list must reflect the physical layout of tx descriptors
2092  * because the firmware pre-fetches descriptors.
2093  *
2094  * XXX might be better to use indices into the buffer array.
2095  */
2096 static void
2097 mwl_txq_reset(struct mwl_softc *sc, struct mwl_txq *txq)
2098 {
2099 	struct mwl_txbuf *bf;
2100 	int i;
2101 
2102 	bf = txq->dma.dd_bufptr;
2103 	STAILQ_INIT(&txq->free);
2104 	for (i = 0; i < mwl_txbuf; i++, bf++)
2105 		STAILQ_INSERT_TAIL(&txq->free, bf, bf_list);
2106 	txq->nfree = i;
2107 }
2108 
2109 #define	DS2PHYS(_dd, _ds) \
2110 	((_dd)->dd_desc_paddr + ((caddr_t)(_ds) - (caddr_t)(_dd)->dd_desc))
2111 
2112 static int
2113 mwl_txdma_setup(struct mwl_softc *sc, struct mwl_txq *txq)
2114 {
2115 	struct ifnet *ifp = sc->sc_ifp;
2116 	int error, bsize, i;
2117 	struct mwl_txbuf *bf;
2118 	struct mwl_txdesc *ds;
2119 
2120 	error = mwl_desc_setup(sc, "tx", &txq->dma,
2121 			mwl_txbuf, sizeof(struct mwl_txbuf),
2122 			MWL_TXDESC, sizeof(struct mwl_txdesc));
2123 	if (error != 0)
2124 		return error;
2125 
2126 	/* allocate and setup tx buffers */
2127 	bsize = mwl_txbuf * sizeof(struct mwl_txbuf);
2128 	bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO);
2129 	if (bf == NULL) {
2130 		if_printf(ifp, "malloc of %u tx buffers failed\n",
2131 			mwl_txbuf);
2132 		return ENOMEM;
2133 	}
2134 	txq->dma.dd_bufptr = bf;
2135 
2136 	ds = txq->dma.dd_desc;
2137 	for (i = 0; i < mwl_txbuf; i++, bf++, ds += MWL_TXDESC) {
2138 		bf->bf_desc = ds;
2139 		bf->bf_daddr = DS2PHYS(&txq->dma, ds);
2140 		error = bus_dmamap_create(sc->sc_dmat, BUS_DMA_NOWAIT,
2141 				&bf->bf_dmamap);
2142 		if (error != 0) {
2143 			if_printf(ifp, "unable to create dmamap for tx "
2144 				"buffer %u, error %u\n", i, error);
2145 			return error;
2146 		}
2147 	}
2148 	mwl_txq_reset(sc, txq);
2149 	return 0;
2150 }
2151 
2152 static void
2153 mwl_txdma_cleanup(struct mwl_softc *sc, struct mwl_txq *txq)
2154 {
2155 	struct mwl_txbuf *bf;
2156 	int i;
2157 
2158 	bf = txq->dma.dd_bufptr;
2159 	for (i = 0; i < mwl_txbuf; i++, bf++) {
2160 		KASSERT(bf->bf_m == NULL, ("mbuf on free list"));
2161 		KASSERT(bf->bf_node == NULL, ("node on free list"));
2162 		if (bf->bf_dmamap != NULL)
2163 			bus_dmamap_destroy(sc->sc_dmat, bf->bf_dmamap);
2164 	}
2165 	STAILQ_INIT(&txq->free);
2166 	txq->nfree = 0;
2167 	if (txq->dma.dd_bufptr != NULL) {
2168 		free(txq->dma.dd_bufptr, M_MWLDEV);
2169 		txq->dma.dd_bufptr = NULL;
2170 	}
2171 	if (txq->dma.dd_desc_len != 0)
2172 		mwl_desc_cleanup(sc, &txq->dma);
2173 }
2174 
2175 static int
2176 mwl_rxdma_setup(struct mwl_softc *sc)
2177 {
2178 	struct ifnet *ifp = sc->sc_ifp;
2179 	int error, jumbosize, bsize, i;
2180 	struct mwl_rxbuf *bf;
2181 	struct mwl_jumbo *rbuf;
2182 	struct mwl_rxdesc *ds;
2183 	caddr_t data;
2184 
2185 	error = mwl_desc_setup(sc, "rx", &sc->sc_rxdma,
2186 			mwl_rxdesc, sizeof(struct mwl_rxbuf),
2187 			1, sizeof(struct mwl_rxdesc));
2188 	if (error != 0)
2189 		return error;
2190 
2191 	/*
2192 	 * Receive is done to a private pool of jumbo buffers.
2193 	 * This allows us to attach to mbuf's and avoid re-mapping
2194 	 * memory on each rx we post.  We allocate a large chunk
2195 	 * of memory and manage it in the driver.  The mbuf free
2196 	 * callback method is used to reclaim frames after sending
2197 	 * them up the stack.  By default we allocate 2x the number of
2198 	 * rx descriptors configured so we have some slop to hold
2199 	 * us while frames are processed.
2200 	 */
2201 	if (mwl_rxbuf < 2*mwl_rxdesc) {
2202 		if_printf(ifp,
2203 		    "too few rx dma buffers (%d); increasing to %d\n",
2204 		    mwl_rxbuf, 2*mwl_rxdesc);
2205 		mwl_rxbuf = 2*mwl_rxdesc;
2206 	}
2207 	jumbosize = roundup(MWL_AGGR_SIZE, PAGE_SIZE);
2208 	sc->sc_rxmemsize = mwl_rxbuf*jumbosize;
2209 
2210 	error = bus_dma_tag_create(sc->sc_dmat,	/* parent */
2211 		       PAGE_SIZE, 0,		/* alignment, bounds */
2212 		       BUS_SPACE_MAXADDR_32BIT,	/* lowaddr */
2213 		       BUS_SPACE_MAXADDR,	/* highaddr */
2214 		       NULL, NULL,		/* filter, filterarg */
2215 		       sc->sc_rxmemsize,	/* maxsize */
2216 		       1,			/* nsegments */
2217 		       sc->sc_rxmemsize,	/* maxsegsize */
2218 		       BUS_DMA_ALLOCNOW,	/* flags */
2219 		       NULL,			/* lockfunc */
2220 		       NULL,			/* lockarg */
2221 		       &sc->sc_rxdmat);
2222 	if (error != 0) {
2223 		if_printf(ifp, "could not create rx DMA tag\n");
2224 		return error;
2225 	}
2226 
2227 	error = bus_dmamem_alloc(sc->sc_rxdmat, (void**) &sc->sc_rxmem,
2228 				 BUS_DMA_NOWAIT | BUS_DMA_COHERENT,
2229 				 &sc->sc_rxmap);
2230 	if (error != 0) {
2231 		if_printf(ifp, "could not alloc %ju bytes of rx DMA memory\n",
2232 		    (uintmax_t) sc->sc_rxmemsize);
2233 		return error;
2234 	}
2235 
2236 	error = bus_dmamap_load(sc->sc_rxdmat, sc->sc_rxmap,
2237 				sc->sc_rxmem, sc->sc_rxmemsize,
2238 				mwl_load_cb, &sc->sc_rxmem_paddr,
2239 				BUS_DMA_NOWAIT);
2240 	if (error != 0) {
2241 		if_printf(ifp, "could not load rx DMA map\n");
2242 		return error;
2243 	}
2244 
2245 	/*
2246 	 * Allocate rx buffers and set them up.
2247 	 */
2248 	bsize = mwl_rxdesc * sizeof(struct mwl_rxbuf);
2249 	bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO);
2250 	if (bf == NULL) {
2251 		if_printf(ifp, "malloc of %u rx buffers failed\n", bsize);
2252 		return error;
2253 	}
2254 	sc->sc_rxdma.dd_bufptr = bf;
2255 
2256 	STAILQ_INIT(&sc->sc_rxbuf);
2257 	ds = sc->sc_rxdma.dd_desc;
2258 	for (i = 0; i < mwl_rxdesc; i++, bf++, ds++) {
2259 		bf->bf_desc = ds;
2260 		bf->bf_daddr = DS2PHYS(&sc->sc_rxdma, ds);
2261 		/* pre-assign dma buffer */
2262 		bf->bf_data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize);
2263 		/* NB: tail is intentional to preserve descriptor order */
2264 		STAILQ_INSERT_TAIL(&sc->sc_rxbuf, bf, bf_list);
2265 	}
2266 
2267 	/*
2268 	 * Place remainder of dma memory buffers on the free list.
2269 	 */
2270 	SLIST_INIT(&sc->sc_rxfree);
2271 	for (; i < mwl_rxbuf; i++) {
2272 		data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize);
2273 		rbuf = MWL_JUMBO_DATA2BUF(data);
2274 		SLIST_INSERT_HEAD(&sc->sc_rxfree, rbuf, next);
2275 		sc->sc_nrxfree++;
2276 	}
2277 	return 0;
2278 }
2279 #undef DS2PHYS
2280 
2281 static void
2282 mwl_rxdma_cleanup(struct mwl_softc *sc)
2283 {
2284 	if (sc->sc_rxmem_paddr != 0) {
2285 		bus_dmamap_unload(sc->sc_rxdmat, sc->sc_rxmap);
2286 		sc->sc_rxmem_paddr = 0;
2287 	}
2288 	if (sc->sc_rxmem != NULL) {
2289 		bus_dmamem_free(sc->sc_rxdmat, sc->sc_rxmem, sc->sc_rxmap);
2290 		sc->sc_rxmem = NULL;
2291 	}
2292 	if (sc->sc_rxdma.dd_bufptr != NULL) {
2293 		free(sc->sc_rxdma.dd_bufptr, M_MWLDEV);
2294 		sc->sc_rxdma.dd_bufptr = NULL;
2295 	}
2296 	if (sc->sc_rxdma.dd_desc_len != 0)
2297 		mwl_desc_cleanup(sc, &sc->sc_rxdma);
2298 }
2299 
2300 static int
2301 mwl_dma_setup(struct mwl_softc *sc)
2302 {
2303 	int error, i;
2304 
2305 	error = mwl_rxdma_setup(sc);
2306 	if (error != 0) {
2307 		mwl_rxdma_cleanup(sc);
2308 		return error;
2309 	}
2310 
2311 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++) {
2312 		error = mwl_txdma_setup(sc, &sc->sc_txq[i]);
2313 		if (error != 0) {
2314 			mwl_dma_cleanup(sc);
2315 			return error;
2316 		}
2317 	}
2318 	return 0;
2319 }
2320 
2321 static void
2322 mwl_dma_cleanup(struct mwl_softc *sc)
2323 {
2324 	int i;
2325 
2326 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
2327 		mwl_txdma_cleanup(sc, &sc->sc_txq[i]);
2328 	mwl_rxdma_cleanup(sc);
2329 }
2330 
2331 static struct ieee80211_node *
2332 mwl_node_alloc(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN])
2333 {
2334 	struct ieee80211com *ic = vap->iv_ic;
2335 	struct mwl_softc *sc = ic->ic_ifp->if_softc;
2336 	const size_t space = sizeof(struct mwl_node);
2337 	struct mwl_node *mn;
2338 
2339 	mn = malloc(space, M_80211_NODE, M_NOWAIT|M_ZERO);
2340 	if (mn == NULL) {
2341 		/* XXX stat+msg */
2342 		return NULL;
2343 	}
2344 	DPRINTF(sc, MWL_DEBUG_NODE, "%s: mn %p\n", __func__, mn);
2345 	return &mn->mn_node;
2346 }
2347 
2348 static void
2349 mwl_node_cleanup(struct ieee80211_node *ni)
2350 {
2351 	struct ieee80211com *ic = ni->ni_ic;
2352         struct mwl_softc *sc = ic->ic_ifp->if_softc;
2353 	struct mwl_node *mn = MWL_NODE(ni);
2354 
2355 	DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p ic %p staid %d\n",
2356 	    __func__, ni, ni->ni_ic, mn->mn_staid);
2357 
2358 	if (mn->mn_staid != 0) {
2359 		struct ieee80211vap *vap = ni->ni_vap;
2360 
2361 		if (mn->mn_hvap != NULL) {
2362 			if (vap->iv_opmode == IEEE80211_M_STA)
2363 				mwl_hal_delstation(mn->mn_hvap, vap->iv_myaddr);
2364 			else
2365 				mwl_hal_delstation(mn->mn_hvap, ni->ni_macaddr);
2366 		}
2367 		/*
2368 		 * NB: legacy WDS peer sta db entry is installed using
2369 		 * the associate ap's hvap; use it again to delete it.
2370 		 * XXX can vap be NULL?
2371 		 */
2372 		else if (vap->iv_opmode == IEEE80211_M_WDS &&
2373 		    MWL_VAP(vap)->mv_ap_hvap != NULL)
2374 			mwl_hal_delstation(MWL_VAP(vap)->mv_ap_hvap,
2375 			    ni->ni_macaddr);
2376 		delstaid(sc, mn->mn_staid);
2377 		mn->mn_staid = 0;
2378 	}
2379 	sc->sc_node_cleanup(ni);
2380 }
2381 
2382 /*
2383  * Reclaim rx dma buffers from packets sitting on the ampdu
2384  * reorder queue for a station.  We replace buffers with a
2385  * system cluster (if available).
2386  */
2387 static void
2388 mwl_ampdu_rxdma_reclaim(struct ieee80211_rx_ampdu *rap)
2389 {
2390 #if 0
2391 	int i, n, off;
2392 	struct mbuf *m;
2393 	void *cl;
2394 
2395 	n = rap->rxa_qframes;
2396 	for (i = 0; i < rap->rxa_wnd && n > 0; i++) {
2397 		m = rap->rxa_m[i];
2398 		if (m == NULL)
2399 			continue;
2400 		n--;
2401 		/* our dma buffers have a well-known free routine */
2402 		if ((m->m_flags & M_EXT) == 0 ||
2403 		    m->m_ext.ext_free != mwl_ext_free)
2404 			continue;
2405 		/*
2406 		 * Try to allocate a cluster and move the data.
2407 		 */
2408 		off = m->m_data - m->m_ext.ext_buf;
2409 		if (off + m->m_pkthdr.len > MCLBYTES) {
2410 			/* XXX no AMSDU for now */
2411 			continue;
2412 		}
2413 		cl = pool_cache_get_paddr(&mclpool_cache, 0,
2414 		    &m->m_ext.ext_paddr);
2415 		if (cl != NULL) {
2416 			/*
2417 			 * Copy the existing data to the cluster, remove
2418 			 * the rx dma buffer, and attach the cluster in
2419 			 * its place.  Note we preserve the offset to the
2420 			 * data so frames being bridged can still prepend
2421 			 * their headers without adding another mbuf.
2422 			 */
2423 			memcpy((caddr_t) cl + off, m->m_data, m->m_pkthdr.len);
2424 			MEXTREMOVE(m);
2425 			MEXTADD(m, cl, MCLBYTES, 0, NULL, &mclpool_cache);
2426 			/* setup mbuf like _MCLGET does */
2427 			m->m_flags |= M_CLUSTER | M_EXT_RW;
2428 			_MOWNERREF(m, M_EXT | M_CLUSTER);
2429 			/* NB: m_data is clobbered by MEXTADDR, adjust */
2430 			m->m_data += off;
2431 		}
2432 	}
2433 #endif
2434 }
2435 
2436 /*
2437  * Callback to reclaim resources.  We first let the
2438  * net80211 layer do it's thing, then if we are still
2439  * blocked by a lack of rx dma buffers we walk the ampdu
2440  * reorder q's to reclaim buffers by copying to a system
2441  * cluster.
2442  */
2443 static void
2444 mwl_node_drain(struct ieee80211_node *ni)
2445 {
2446 	struct ieee80211com *ic = ni->ni_ic;
2447         struct mwl_softc *sc = ic->ic_ifp->if_softc;
2448 	struct mwl_node *mn = MWL_NODE(ni);
2449 
2450 	DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p vap %p staid %d\n",
2451 	    __func__, ni, ni->ni_vap, mn->mn_staid);
2452 
2453 	/* NB: call up first to age out ampdu q's */
2454 	sc->sc_node_drain(ni);
2455 
2456 	/* XXX better to not check low water mark? */
2457 	if (sc->sc_rxblocked && mn->mn_staid != 0 &&
2458 	    (ni->ni_flags & IEEE80211_NODE_HT)) {
2459 		uint8_t tid;
2460 		/*
2461 		 * Walk the reorder q and reclaim rx dma buffers by copying
2462 		 * the packet contents into clusters.
2463 		 */
2464 		for (tid = 0; tid < WME_NUM_TID; tid++) {
2465 			struct ieee80211_rx_ampdu *rap;
2466 
2467 			rap = &ni->ni_rx_ampdu[tid];
2468 			if ((rap->rxa_flags & IEEE80211_AGGR_XCHGPEND) == 0)
2469 				continue;
2470 			if (rap->rxa_qframes)
2471 				mwl_ampdu_rxdma_reclaim(rap);
2472 		}
2473 	}
2474 }
2475 
2476 static void
2477 mwl_node_getsignal(const struct ieee80211_node *ni, int8_t *rssi, int8_t *noise)
2478 {
2479 	*rssi = ni->ni_ic->ic_node_getrssi(ni);
2480 #ifdef MWL_ANT_INFO_SUPPORT
2481 #if 0
2482 	/* XXX need to smooth data */
2483 	*noise = -MWL_NODE_CONST(ni)->mn_ai.nf;
2484 #else
2485 	*noise = -95;		/* XXX */
2486 #endif
2487 #else
2488 	*noise = -95;		/* XXX */
2489 #endif
2490 }
2491 
2492 /*
2493  * Convert Hardware per-antenna rssi info to common format:
2494  * Let a1, a2, a3 represent the amplitudes per chain
2495  * Let amax represent max[a1, a2, a3]
2496  * Rssi1_dBm = RSSI_dBm + 20*log10(a1/amax)
2497  * Rssi1_dBm = RSSI_dBm + 20*log10(a1) - 20*log10(amax)
2498  * We store a table that is 4*20*log10(idx) - the extra 4 is to store or
2499  * maintain some extra precision.
2500  *
2501  * Values are stored in .5 db format capped at 127.
2502  */
2503 static void
2504 mwl_node_getmimoinfo(const struct ieee80211_node *ni,
2505 	struct ieee80211_mimo_info *mi)
2506 {
2507 #define	CVT(_dst, _src) do {						\
2508 	(_dst) = rssi + ((logdbtbl[_src] - logdbtbl[rssi_max]) >> 2);	\
2509 	(_dst) = (_dst) > 64 ? 127 : ((_dst) << 1);			\
2510 } while (0)
2511 	static const int8_t logdbtbl[32] = {
2512 	       0,   0,  24,  38,  48,  56,  62,  68,
2513 	      72,  76,  80,  83,  86,  89,  92,  94,
2514 	      96,  98, 100, 102, 104, 106, 107, 109,
2515 	     110, 112, 113, 115, 116, 117, 118, 119
2516 	};
2517 	const struct mwl_node *mn = MWL_NODE_CONST(ni);
2518 	uint8_t rssi = mn->mn_ai.rsvd1/2;		/* XXX */
2519 	uint32_t rssi_max;
2520 
2521 	rssi_max = mn->mn_ai.rssi_a;
2522 	if (mn->mn_ai.rssi_b > rssi_max)
2523 		rssi_max = mn->mn_ai.rssi_b;
2524 	if (mn->mn_ai.rssi_c > rssi_max)
2525 		rssi_max = mn->mn_ai.rssi_c;
2526 
2527 	CVT(mi->rssi[0], mn->mn_ai.rssi_a);
2528 	CVT(mi->rssi[1], mn->mn_ai.rssi_b);
2529 	CVT(mi->rssi[2], mn->mn_ai.rssi_c);
2530 
2531 	mi->noise[0] = mn->mn_ai.nf_a;
2532 	mi->noise[1] = mn->mn_ai.nf_b;
2533 	mi->noise[2] = mn->mn_ai.nf_c;
2534 #undef CVT
2535 }
2536 
2537 static __inline void *
2538 mwl_getrxdma(struct mwl_softc *sc)
2539 {
2540 	struct mwl_jumbo *buf;
2541 	void *data;
2542 
2543 	/*
2544 	 * Allocate from jumbo pool.
2545 	 */
2546 	MWL_RXFREE_LOCK(sc);
2547 	buf = SLIST_FIRST(&sc->sc_rxfree);
2548 	if (buf == NULL) {
2549 		DPRINTF(sc, MWL_DEBUG_ANY,
2550 		    "%s: out of rx dma buffers\n", __func__);
2551 		sc->sc_stats.mst_rx_nodmabuf++;
2552 		data = NULL;
2553 	} else {
2554 		SLIST_REMOVE_HEAD(&sc->sc_rxfree, next);
2555 		sc->sc_nrxfree--;
2556 		data = MWL_JUMBO_BUF2DATA(buf);
2557 	}
2558 	MWL_RXFREE_UNLOCK(sc);
2559 	return data;
2560 }
2561 
2562 static __inline void
2563 mwl_putrxdma(struct mwl_softc *sc, void *data)
2564 {
2565 	struct mwl_jumbo *buf;
2566 
2567 	/* XXX bounds check data */
2568 	MWL_RXFREE_LOCK(sc);
2569 	buf = MWL_JUMBO_DATA2BUF(data);
2570 	SLIST_INSERT_HEAD(&sc->sc_rxfree, buf, next);
2571 	sc->sc_nrxfree++;
2572 	MWL_RXFREE_UNLOCK(sc);
2573 }
2574 
2575 static int
2576 mwl_rxbuf_init(struct mwl_softc *sc, struct mwl_rxbuf *bf)
2577 {
2578 	struct mwl_rxdesc *ds;
2579 
2580 	ds = bf->bf_desc;
2581 	if (bf->bf_data == NULL) {
2582 		bf->bf_data = mwl_getrxdma(sc);
2583 		if (bf->bf_data == NULL) {
2584 			/* mark descriptor to be skipped */
2585 			ds->RxControl = EAGLE_RXD_CTRL_OS_OWN;
2586 			/* NB: don't need PREREAD */
2587 			MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREWRITE);
2588 			sc->sc_stats.mst_rxbuf_failed++;
2589 			return ENOMEM;
2590 		}
2591 	}
2592 	/*
2593 	 * NB: DMA buffer contents is known to be unmodified
2594 	 *     so there's no need to flush the data cache.
2595 	 */
2596 
2597 	/*
2598 	 * Setup descriptor.
2599 	 */
2600 	ds->QosCtrl = 0;
2601 	ds->RSSI = 0;
2602 	ds->Status = EAGLE_RXD_STATUS_IDLE;
2603 	ds->Channel = 0;
2604 	ds->PktLen = htole16(MWL_AGGR_SIZE);
2605 	ds->SQ2 = 0;
2606 	ds->pPhysBuffData = htole32(MWL_JUMBO_DMA_ADDR(sc, bf->bf_data));
2607 	/* NB: don't touch pPhysNext, set once */
2608 	ds->RxControl = EAGLE_RXD_CTRL_DRIVER_OWN;
2609 	MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
2610 
2611 	return 0;
2612 }
2613 
2614 static int
2615 mwl_ext_free(struct mbuf *m, void *data, void *arg)
2616 {
2617 	struct mwl_softc *sc = arg;
2618 
2619 	/* XXX bounds check data */
2620 	mwl_putrxdma(sc, data);
2621 	/*
2622 	 * If we were previously blocked by a lack of rx dma buffers
2623 	 * check if we now have enough to restart rx interrupt handling.
2624 	 * NB: we know we are called at splvm which is above splnet.
2625 	 */
2626 	if (sc->sc_rxblocked && sc->sc_nrxfree > mwl_rxdmalow) {
2627 		sc->sc_rxblocked = 0;
2628 		mwl_hal_intrset(sc->sc_mh, sc->sc_imask);
2629 	}
2630 	return (EXT_FREE_OK);
2631 }
2632 
2633 struct mwl_frame_bar {
2634 	u_int8_t	i_fc[2];
2635 	u_int8_t	i_dur[2];
2636 	u_int8_t	i_ra[IEEE80211_ADDR_LEN];
2637 	u_int8_t	i_ta[IEEE80211_ADDR_LEN];
2638 	/* ctl, seq, FCS */
2639 } __packed;
2640 
2641 /*
2642  * Like ieee80211_anyhdrsize, but handles BAR frames
2643  * specially so the logic below to piece the 802.11
2644  * header together works.
2645  */
2646 static __inline int
2647 mwl_anyhdrsize(const void *data)
2648 {
2649 	const struct ieee80211_frame *wh = data;
2650 
2651 	if ((wh->i_fc[0]&IEEE80211_FC0_TYPE_MASK) == IEEE80211_FC0_TYPE_CTL) {
2652 		switch (wh->i_fc[0] & IEEE80211_FC0_SUBTYPE_MASK) {
2653 		case IEEE80211_FC0_SUBTYPE_CTS:
2654 		case IEEE80211_FC0_SUBTYPE_ACK:
2655 			return sizeof(struct ieee80211_frame_ack);
2656 		case IEEE80211_FC0_SUBTYPE_BAR:
2657 			return sizeof(struct mwl_frame_bar);
2658 		}
2659 		return sizeof(struct ieee80211_frame_min);
2660 	} else
2661 		return ieee80211_hdrsize(data);
2662 }
2663 
2664 static void
2665 mwl_handlemicerror(struct ieee80211com *ic, const uint8_t *data)
2666 {
2667 	const struct ieee80211_frame *wh;
2668 	struct ieee80211_node *ni;
2669 
2670 	wh = (const struct ieee80211_frame *)(data + sizeof(uint16_t));
2671 	ni = ieee80211_find_rxnode(ic, (const struct ieee80211_frame_min *) wh);
2672 	if (ni != NULL) {
2673 		ieee80211_notify_michael_failure(ni->ni_vap, wh, 0);
2674 		ieee80211_free_node(ni);
2675 	}
2676 }
2677 
2678 /*
2679  * Convert hardware signal strength to rssi.  The value
2680  * provided by the device has the noise floor added in;
2681  * we need to compensate for this but we don't have that
2682  * so we use a fixed value.
2683  *
2684  * The offset of 8 is good for both 2.4 and 5GHz.  The LNA
2685  * offset is already set as part of the initial gain.  This
2686  * will give at least +/- 3dB for 2.4GHz and +/- 5dB for 5GHz.
2687  */
2688 static __inline int
2689 cvtrssi(uint8_t ssi)
2690 {
2691 	int rssi = (int) ssi + 8;
2692 	/* XXX hack guess until we have a real noise floor */
2693 	rssi = 2*(87 - rssi);	/* NB: .5 dBm units */
2694 	return (rssi < 0 ? 0 : rssi > 127 ? 127 : rssi);
2695 }
2696 
2697 static void
2698 mwl_rx_proc(void *arg, int npending)
2699 {
2700 #define	IEEE80211_DIR_DSTODS(wh) \
2701 	((((const struct ieee80211_frame *)wh)->i_fc[1] & IEEE80211_FC1_DIR_MASK) == IEEE80211_FC1_DIR_DSTODS)
2702 	struct mwl_softc *sc = arg;
2703 	struct ifnet *ifp = sc->sc_ifp;
2704 	struct ieee80211com *ic = ifp->if_l2com;
2705 	struct mwl_rxbuf *bf;
2706 	struct mwl_rxdesc *ds;
2707 	struct mbuf *m;
2708 	struct ieee80211_qosframe *wh;
2709 	struct ieee80211_qosframe_addr4 *wh4;
2710 	struct ieee80211_node *ni;
2711 	struct mwl_node *mn;
2712 	int off, len, hdrlen, pktlen, rssi, ntodo;
2713 	uint8_t *data, status;
2714 	void *newdata;
2715 	int16_t nf;
2716 
2717 	DPRINTF(sc, MWL_DEBUG_RX_PROC, "%s: pending %u rdptr 0x%x wrptr 0x%x\n",
2718 	    __func__, npending, RD4(sc, sc->sc_hwspecs.rxDescRead),
2719 	    RD4(sc, sc->sc_hwspecs.rxDescWrite));
2720 	nf = -96;			/* XXX */
2721 	bf = sc->sc_rxnext;
2722 	for (ntodo = mwl_rxquota; ntodo > 0; ntodo--) {
2723 		if (bf == NULL)
2724 			bf = STAILQ_FIRST(&sc->sc_rxbuf);
2725 		ds = bf->bf_desc;
2726 		data = bf->bf_data;
2727 		if (data == NULL) {
2728 			/*
2729 			 * If data allocation failed previously there
2730 			 * will be no buffer; try again to re-populate it.
2731 			 * Note the firmware will not advance to the next
2732 			 * descriptor with a dma buffer so we must mimic
2733 			 * this or we'll get out of sync.
2734 			 */
2735 			DPRINTF(sc, MWL_DEBUG_ANY,
2736 			    "%s: rx buf w/o dma memory\n", __func__);
2737 			(void) mwl_rxbuf_init(sc, bf);
2738 			sc->sc_stats.mst_rx_dmabufmissing++;
2739 			break;
2740 		}
2741 		MWL_RXDESC_SYNC(sc, ds,
2742 		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
2743 		if (ds->RxControl != EAGLE_RXD_CTRL_DMA_OWN)
2744 			break;
2745 #ifdef MWL_DEBUG
2746 		if (sc->sc_debug & MWL_DEBUG_RECV_DESC)
2747 			mwl_printrxbuf(bf, 0);
2748 #endif
2749 		status = ds->Status;
2750 		if (status & EAGLE_RXD_STATUS_DECRYPT_ERR_MASK) {
2751 			ifp->if_ierrors++;
2752 			sc->sc_stats.mst_rx_crypto++;
2753 			/*
2754 			 * NB: Check EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR
2755 			 *     for backwards compatibility.
2756 			 */
2757 			if (status != EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR &&
2758 			    (status & EAGLE_RXD_STATUS_TKIP_MIC_DECRYPT_ERR)) {
2759 				/*
2760 				 * MIC error, notify upper layers.
2761 				 */
2762 				bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap,
2763 				    BUS_DMASYNC_POSTREAD);
2764 				mwl_handlemicerror(ic, data);
2765 				sc->sc_stats.mst_rx_tkipmic++;
2766 			}
2767 			/* XXX too painful to tap packets */
2768 			goto rx_next;
2769 		}
2770 		/*
2771 		 * Sync the data buffer.
2772 		 */
2773 		len = le16toh(ds->PktLen);
2774 		bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap, BUS_DMASYNC_POSTREAD);
2775 		/*
2776 		 * The 802.11 header is provided all or in part at the front;
2777 		 * use it to calculate the true size of the header that we'll
2778 		 * construct below.  We use this to figure out where to copy
2779 		 * payload prior to constructing the header.
2780 		 */
2781 		hdrlen = mwl_anyhdrsize(data + sizeof(uint16_t));
2782 		off = sizeof(uint16_t) + sizeof(struct ieee80211_frame_addr4);
2783 
2784 		/* calculate rssi early so we can re-use for each aggregate */
2785 		rssi = cvtrssi(ds->RSSI);
2786 
2787 		pktlen = hdrlen + (len - off);
2788 		/*
2789 		 * NB: we know our frame is at least as large as
2790 		 * IEEE80211_MIN_LEN because there is a 4-address
2791 		 * frame at the front.  Hence there's no need to
2792 		 * vet the packet length.  If the frame in fact
2793 		 * is too small it should be discarded at the
2794 		 * net80211 layer.
2795 		 */
2796 
2797 		/*
2798 		 * Attach dma buffer to an mbuf.  We tried
2799 		 * doing this based on the packet size (i.e.
2800 		 * copying small packets) but it turns out to
2801 		 * be a net loss.  The tradeoff might be system
2802 		 * dependent (cache architecture is important).
2803 		 */
2804 		MGETHDR(m, M_NOWAIT, MT_DATA);
2805 		if (m == NULL) {
2806 			DPRINTF(sc, MWL_DEBUG_ANY,
2807 			    "%s: no rx mbuf\n", __func__);
2808 			sc->sc_stats.mst_rx_nombuf++;
2809 			goto rx_next;
2810 		}
2811 		/*
2812 		 * Acquire the replacement dma buffer before
2813 		 * processing the frame.  If we're out of dma
2814 		 * buffers we disable rx interrupts and wait
2815 		 * for the free pool to reach mlw_rxdmalow buffers
2816 		 * before starting to do work again.  If the firmware
2817 		 * runs out of descriptors then it will toss frames
2818 		 * which is better than our doing it as that can
2819 		 * starve our processing.  It is also important that
2820 		 * we always process rx'd frames in case they are
2821 		 * A-MPDU as otherwise the host's view of the BA
2822 		 * window may get out of sync with the firmware.
2823 		 */
2824 		newdata = mwl_getrxdma(sc);
2825 		if (newdata == NULL) {
2826 			/* NB: stat+msg in mwl_getrxdma */
2827 			m_free(m);
2828 			/* disable RX interrupt and mark state */
2829 			mwl_hal_intrset(sc->sc_mh,
2830 			    sc->sc_imask &~ MACREG_A2HRIC_BIT_RX_RDY);
2831 			sc->sc_rxblocked = 1;
2832 			ieee80211_drain(ic);
2833 			/* XXX check rxblocked and immediately start again? */
2834 			goto rx_stop;
2835 		}
2836 		bf->bf_data = newdata;
2837 		/*
2838 		 * Attach the dma buffer to the mbuf;
2839 		 * mwl_rxbuf_init will re-setup the rx
2840 		 * descriptor using the replacement dma
2841 		 * buffer we just installed above.
2842 		 */
2843 		MEXTADD(m, data, MWL_AGGR_SIZE, mwl_ext_free,
2844 		    data, sc, 0, EXT_NET_DRV);
2845 		m->m_data += off - hdrlen;
2846 		m->m_pkthdr.len = m->m_len = pktlen;
2847 		m->m_pkthdr.rcvif = ifp;
2848 		/* NB: dma buffer assumed read-only */
2849 
2850 		/*
2851 		 * Piece 802.11 header together.
2852 		 */
2853 		wh = mtod(m, struct ieee80211_qosframe *);
2854 		/* NB: don't need to do this sometimes but ... */
2855 		/* XXX special case so we can memcpy after m_devget? */
2856 		ovbcopy(data + sizeof(uint16_t), wh, hdrlen);
2857 		if (IEEE80211_QOS_HAS_SEQ(wh)) {
2858 			if (IEEE80211_DIR_DSTODS(wh)) {
2859 				wh4 = mtod(m,
2860 				    struct ieee80211_qosframe_addr4*);
2861 				*(uint16_t *)wh4->i_qos = ds->QosCtrl;
2862 			} else {
2863 				*(uint16_t *)wh->i_qos = ds->QosCtrl;
2864 			}
2865 		}
2866 		/*
2867 		 * The f/w strips WEP header but doesn't clear
2868 		 * the WEP bit; mark the packet with M_WEP so
2869 		 * net80211 will treat the data as decrypted.
2870 		 * While here also clear the PWR_MGT bit since
2871 		 * power save is handled by the firmware and
2872 		 * passing this up will potentially cause the
2873 		 * upper layer to put a station in power save
2874 		 * (except when configured with MWL_HOST_PS_SUPPORT).
2875 		 */
2876 		if (wh->i_fc[1] & IEEE80211_FC1_PROTECTED)
2877 			m->m_flags |= M_WEP;
2878 #ifdef MWL_HOST_PS_SUPPORT
2879 		wh->i_fc[1] &= ~IEEE80211_FC1_PROTECTED;
2880 #else
2881 		wh->i_fc[1] &= ~(IEEE80211_FC1_PROTECTED |
2882 		    IEEE80211_FC1_PWR_MGT);
2883 #endif
2884 
2885 		if (ieee80211_radiotap_active(ic)) {
2886 			struct mwl_rx_radiotap_header *tap = &sc->sc_rx_th;
2887 
2888 			tap->wr_flags = 0;
2889 			tap->wr_rate = ds->Rate;
2890 			tap->wr_antsignal = rssi + nf;
2891 			tap->wr_antnoise = nf;
2892 		}
2893 		if (IFF_DUMPPKTS_RECV(sc, wh)) {
2894 			ieee80211_dump_pkt(ic, mtod(m, caddr_t),
2895 			    len, ds->Rate, rssi);
2896 		}
2897 		ifp->if_ipackets++;
2898 
2899 		/* dispatch */
2900 		ni = ieee80211_find_rxnode(ic,
2901 		    (const struct ieee80211_frame_min *) wh);
2902 		if (ni != NULL) {
2903 			mn = MWL_NODE(ni);
2904 #ifdef MWL_ANT_INFO_SUPPORT
2905 			mn->mn_ai.rssi_a = ds->ai.rssi_a;
2906 			mn->mn_ai.rssi_b = ds->ai.rssi_b;
2907 			mn->mn_ai.rssi_c = ds->ai.rssi_c;
2908 			mn->mn_ai.rsvd1 = rssi;
2909 #endif
2910 			/* tag AMPDU aggregates for reorder processing */
2911 			if (ni->ni_flags & IEEE80211_NODE_HT)
2912 				m->m_flags |= M_AMPDU;
2913 			(void) ieee80211_input(ni, m, rssi, nf);
2914 			ieee80211_free_node(ni);
2915 		} else
2916 			(void) ieee80211_input_all(ic, m, rssi, nf);
2917 rx_next:
2918 		/* NB: ignore ENOMEM so we process more descriptors */
2919 		(void) mwl_rxbuf_init(sc, bf);
2920 		bf = STAILQ_NEXT(bf, bf_list);
2921 	}
2922 rx_stop:
2923 	sc->sc_rxnext = bf;
2924 
2925 	if ((ifp->if_drv_flags & IFF_DRV_OACTIVE) == 0 &&
2926 	    !IFQ_IS_EMPTY(&ifp->if_snd)) {
2927 		/* NB: kick fw; the tx thread may have been preempted */
2928 		mwl_hal_txstart(sc->sc_mh, 0);
2929 		mwl_start(ifp);
2930 	}
2931 #undef IEEE80211_DIR_DSTODS
2932 }
2933 
2934 static void
2935 mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *txq, int qnum)
2936 {
2937 	struct mwl_txbuf *bf, *bn;
2938 	struct mwl_txdesc *ds;
2939 
2940 	MWL_TXQ_LOCK_INIT(sc, txq);
2941 	txq->qnum = qnum;
2942 	txq->txpri = 0;	/* XXX */
2943 #if 0
2944 	/* NB: q setup by mwl_txdma_setup XXX */
2945 	STAILQ_INIT(&txq->free);
2946 #endif
2947 	STAILQ_FOREACH(bf, &txq->free, bf_list) {
2948 		bf->bf_txq = txq;
2949 
2950 		ds = bf->bf_desc;
2951 		bn = STAILQ_NEXT(bf, bf_list);
2952 		if (bn == NULL)
2953 			bn = STAILQ_FIRST(&txq->free);
2954 		ds->pPhysNext = htole32(bn->bf_daddr);
2955 	}
2956 	STAILQ_INIT(&txq->active);
2957 }
2958 
2959 /*
2960  * Setup a hardware data transmit queue for the specified
2961  * access control.  We record the mapping from ac's
2962  * to h/w queues for use by mwl_tx_start.
2963  */
2964 static int
2965 mwl_tx_setup(struct mwl_softc *sc, int ac, int mvtype)
2966 {
2967 #define	N(a)	(sizeof(a)/sizeof(a[0]))
2968 	struct mwl_txq *txq;
2969 
2970 	if (ac >= N(sc->sc_ac2q)) {
2971 		device_printf(sc->sc_dev, "AC %u out of range, max %zu!\n",
2972 			ac, N(sc->sc_ac2q));
2973 		return 0;
2974 	}
2975 	if (mvtype >= MWL_NUM_TX_QUEUES) {
2976 		device_printf(sc->sc_dev, "mvtype %u out of range, max %u!\n",
2977 			mvtype, MWL_NUM_TX_QUEUES);
2978 		return 0;
2979 	}
2980 	txq = &sc->sc_txq[mvtype];
2981 	mwl_txq_init(sc, txq, mvtype);
2982 	sc->sc_ac2q[ac] = txq;
2983 	return 1;
2984 #undef N
2985 }
2986 
2987 /*
2988  * Update WME parameters for a transmit queue.
2989  */
2990 static int
2991 mwl_txq_update(struct mwl_softc *sc, int ac)
2992 {
2993 #define	MWL_EXPONENT_TO_VALUE(v)	((1<<v)-1)
2994 	struct ifnet *ifp = sc->sc_ifp;
2995 	struct ieee80211com *ic = ifp->if_l2com;
2996 	struct mwl_txq *txq = sc->sc_ac2q[ac];
2997 	struct wmeParams *wmep = &ic->ic_wme.wme_chanParams.cap_wmeParams[ac];
2998 	struct mwl_hal *mh = sc->sc_mh;
2999 	int aifs, cwmin, cwmax, txoplim;
3000 
3001 	aifs = wmep->wmep_aifsn;
3002 	/* XXX in sta mode need to pass log values for cwmin/max */
3003 	cwmin = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmin);
3004 	cwmax = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmax);
3005 	txoplim = wmep->wmep_txopLimit;		/* NB: units of 32us */
3006 
3007 	if (mwl_hal_setedcaparams(mh, txq->qnum, cwmin, cwmax, aifs, txoplim)) {
3008 		device_printf(sc->sc_dev, "unable to update hardware queue "
3009 			"parameters for %s traffic!\n",
3010 			ieee80211_wme_acnames[ac]);
3011 		return 0;
3012 	}
3013 	return 1;
3014 #undef MWL_EXPONENT_TO_VALUE
3015 }
3016 
3017 /*
3018  * Callback from the 802.11 layer to update WME parameters.
3019  */
3020 static int
3021 mwl_wme_update(struct ieee80211com *ic)
3022 {
3023 	struct mwl_softc *sc = ic->ic_ifp->if_softc;
3024 
3025 	return !mwl_txq_update(sc, WME_AC_BE) ||
3026 	    !mwl_txq_update(sc, WME_AC_BK) ||
3027 	    !mwl_txq_update(sc, WME_AC_VI) ||
3028 	    !mwl_txq_update(sc, WME_AC_VO) ? EIO : 0;
3029 }
3030 
3031 /*
3032  * Reclaim resources for a setup queue.
3033  */
3034 static void
3035 mwl_tx_cleanupq(struct mwl_softc *sc, struct mwl_txq *txq)
3036 {
3037 	/* XXX hal work? */
3038 	MWL_TXQ_LOCK_DESTROY(txq);
3039 }
3040 
3041 /*
3042  * Reclaim all tx queue resources.
3043  */
3044 static void
3045 mwl_tx_cleanup(struct mwl_softc *sc)
3046 {
3047 	int i;
3048 
3049 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
3050 		mwl_tx_cleanupq(sc, &sc->sc_txq[i]);
3051 }
3052 
3053 static int
3054 mwl_tx_dmasetup(struct mwl_softc *sc, struct mwl_txbuf *bf, struct mbuf *m0)
3055 {
3056 	struct mbuf *m;
3057 	int error;
3058 
3059 	/*
3060 	 * Load the DMA map so any coalescing is done.  This
3061 	 * also calculates the number of descriptors we need.
3062 	 */
3063 	error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0,
3064 				     bf->bf_segs, &bf->bf_nseg,
3065 				     BUS_DMA_NOWAIT);
3066 	if (error == EFBIG) {
3067 		/* XXX packet requires too many descriptors */
3068 		bf->bf_nseg = MWL_TXDESC+1;
3069 	} else if (error != 0) {
3070 		sc->sc_stats.mst_tx_busdma++;
3071 		m_freem(m0);
3072 		return error;
3073 	}
3074 	/*
3075 	 * Discard null packets and check for packets that
3076 	 * require too many TX descriptors.  We try to convert
3077 	 * the latter to a cluster.
3078 	 */
3079 	if (error == EFBIG) {		/* too many desc's, linearize */
3080 		sc->sc_stats.mst_tx_linear++;
3081 #if MWL_TXDESC > 1
3082 		m = m_collapse(m0, M_NOWAIT, MWL_TXDESC);
3083 #else
3084 		m = m_defrag(m0, M_NOWAIT);
3085 #endif
3086 		if (m == NULL) {
3087 			m_freem(m0);
3088 			sc->sc_stats.mst_tx_nombuf++;
3089 			return ENOMEM;
3090 		}
3091 		m0 = m;
3092 		error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0,
3093 					     bf->bf_segs, &bf->bf_nseg,
3094 					     BUS_DMA_NOWAIT);
3095 		if (error != 0) {
3096 			sc->sc_stats.mst_tx_busdma++;
3097 			m_freem(m0);
3098 			return error;
3099 		}
3100 		KASSERT(bf->bf_nseg <= MWL_TXDESC,
3101 		    ("too many segments after defrag; nseg %u", bf->bf_nseg));
3102 	} else if (bf->bf_nseg == 0) {		/* null packet, discard */
3103 		sc->sc_stats.mst_tx_nodata++;
3104 		m_freem(m0);
3105 		return EIO;
3106 	}
3107 	DPRINTF(sc, MWL_DEBUG_XMIT, "%s: m %p len %u\n",
3108 		__func__, m0, m0->m_pkthdr.len);
3109 	bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap, BUS_DMASYNC_PREWRITE);
3110 	bf->bf_m = m0;
3111 
3112 	return 0;
3113 }
3114 
3115 static __inline int
3116 mwl_cvtlegacyrate(int rate)
3117 {
3118 	switch (rate) {
3119 	case 2:	 return 0;
3120 	case 4:	 return 1;
3121 	case 11: return 2;
3122 	case 22: return 3;
3123 	case 44: return 4;
3124 	case 12: return 5;
3125 	case 18: return 6;
3126 	case 24: return 7;
3127 	case 36: return 8;
3128 	case 48: return 9;
3129 	case 72: return 10;
3130 	case 96: return 11;
3131 	case 108:return 12;
3132 	}
3133 	return 0;
3134 }
3135 
3136 /*
3137  * Calculate fixed tx rate information per client state;
3138  * this value is suitable for writing to the Format field
3139  * of a tx descriptor.
3140  */
3141 static uint16_t
3142 mwl_calcformat(uint8_t rate, const struct ieee80211_node *ni)
3143 {
3144 	uint16_t fmt;
3145 
3146 	fmt = SM(3, EAGLE_TXD_ANTENNA)
3147 	    | (IEEE80211_IS_CHAN_HT40D(ni->ni_chan) ?
3148 		EAGLE_TXD_EXTCHAN_LO : EAGLE_TXD_EXTCHAN_HI);
3149 	if (rate & IEEE80211_RATE_MCS) {	/* HT MCS */
3150 		fmt |= EAGLE_TXD_FORMAT_HT
3151 		    /* NB: 0x80 implicitly stripped from ucastrate */
3152 		    | SM(rate, EAGLE_TXD_RATE);
3153 		/* XXX short/long GI may be wrong; re-check */
3154 		if (IEEE80211_IS_CHAN_HT40(ni->ni_chan)) {
3155 			fmt |= EAGLE_TXD_CHW_40
3156 			    | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI40 ?
3157 			        EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG);
3158 		} else {
3159 			fmt |= EAGLE_TXD_CHW_20
3160 			    | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI20 ?
3161 			        EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG);
3162 		}
3163 	} else {			/* legacy rate */
3164 		fmt |= EAGLE_TXD_FORMAT_LEGACY
3165 		    | SM(mwl_cvtlegacyrate(rate), EAGLE_TXD_RATE)
3166 		    | EAGLE_TXD_CHW_20
3167 		    /* XXX iv_flags & IEEE80211_F_SHPREAMBLE? */
3168 		    | (ni->ni_capinfo & IEEE80211_CAPINFO_SHORT_PREAMBLE ?
3169 			EAGLE_TXD_PREAMBLE_SHORT : EAGLE_TXD_PREAMBLE_LONG);
3170 	}
3171 	return fmt;
3172 }
3173 
3174 static int
3175 mwl_tx_start(struct mwl_softc *sc, struct ieee80211_node *ni, struct mwl_txbuf *bf,
3176     struct mbuf *m0)
3177 {
3178 #define	IEEE80211_DIR_DSTODS(wh) \
3179 	((wh->i_fc[1] & IEEE80211_FC1_DIR_MASK) == IEEE80211_FC1_DIR_DSTODS)
3180 	struct ifnet *ifp = sc->sc_ifp;
3181 	struct ieee80211com *ic = ifp->if_l2com;
3182 	struct ieee80211vap *vap = ni->ni_vap;
3183 	int error, iswep, ismcast;
3184 	int hdrlen, copyhdrlen, pktlen;
3185 	struct mwl_txdesc *ds;
3186 	struct mwl_txq *txq;
3187 	struct ieee80211_frame *wh;
3188 	struct mwltxrec *tr;
3189 	struct mwl_node *mn;
3190 	uint16_t qos;
3191 #if MWL_TXDESC > 1
3192 	int i;
3193 #endif
3194 
3195 	wh = mtod(m0, struct ieee80211_frame *);
3196 	iswep = wh->i_fc[1] & IEEE80211_FC1_PROTECTED;
3197 	ismcast = IEEE80211_IS_MULTICAST(wh->i_addr1);
3198 	hdrlen = ieee80211_anyhdrsize(wh);
3199 	copyhdrlen = hdrlen;
3200 	pktlen = m0->m_pkthdr.len;
3201 	if (IEEE80211_QOS_HAS_SEQ(wh)) {
3202 		if (IEEE80211_DIR_DSTODS(wh)) {
3203 			qos = *(uint16_t *)
3204 			    (((struct ieee80211_qosframe_addr4 *) wh)->i_qos);
3205 			copyhdrlen -= sizeof(qos);
3206 		} else
3207 			qos = *(uint16_t *)
3208 			    (((struct ieee80211_qosframe *) wh)->i_qos);
3209 	} else
3210 		qos = 0;
3211 
3212 	if (iswep) {
3213 		const struct ieee80211_cipher *cip;
3214 		struct ieee80211_key *k;
3215 
3216 		/*
3217 		 * Construct the 802.11 header+trailer for an encrypted
3218 		 * frame. The only reason this can fail is because of an
3219 		 * unknown or unsupported cipher/key type.
3220 		 *
3221 		 * NB: we do this even though the firmware will ignore
3222 		 *     what we've done for WEP and TKIP as we need the
3223 		 *     ExtIV filled in for CCMP and this also adjusts
3224 		 *     the headers which simplifies our work below.
3225 		 */
3226 		k = ieee80211_crypto_encap(ni, m0);
3227 		if (k == NULL) {
3228 			/*
3229 			 * This can happen when the key is yanked after the
3230 			 * frame was queued.  Just discard the frame; the
3231 			 * 802.11 layer counts failures and provides
3232 			 * debugging/diagnostics.
3233 			 */
3234 			m_freem(m0);
3235 			return EIO;
3236 		}
3237 		/*
3238 		 * Adjust the packet length for the crypto additions
3239 		 * done during encap and any other bits that the f/w
3240 		 * will add later on.
3241 		 */
3242 		cip = k->wk_cipher;
3243 		pktlen += cip->ic_header + cip->ic_miclen + cip->ic_trailer;
3244 
3245 		/* packet header may have moved, reset our local pointer */
3246 		wh = mtod(m0, struct ieee80211_frame *);
3247 	}
3248 
3249 	if (ieee80211_radiotap_active_vap(vap)) {
3250 		sc->sc_tx_th.wt_flags = 0;	/* XXX */
3251 		if (iswep)
3252 			sc->sc_tx_th.wt_flags |= IEEE80211_RADIOTAP_F_WEP;
3253 #if 0
3254 		sc->sc_tx_th.wt_rate = ds->DataRate;
3255 #endif
3256 		sc->sc_tx_th.wt_txpower = ni->ni_txpower;
3257 		sc->sc_tx_th.wt_antenna = sc->sc_txantenna;
3258 
3259 		ieee80211_radiotap_tx(vap, m0);
3260 	}
3261 	/*
3262 	 * Copy up/down the 802.11 header; the firmware requires
3263 	 * we present a 2-byte payload length followed by a
3264 	 * 4-address header (w/o QoS), followed (optionally) by
3265 	 * any WEP/ExtIV header (but only filled in for CCMP).
3266 	 * We are assured the mbuf has sufficient headroom to
3267 	 * prepend in-place by the setup of ic_headroom in
3268 	 * mwl_attach.
3269 	 */
3270 	if (hdrlen < sizeof(struct mwltxrec)) {
3271 		const int space = sizeof(struct mwltxrec) - hdrlen;
3272 		if (M_LEADINGSPACE(m0) < space) {
3273 			/* NB: should never happen */
3274 			device_printf(sc->sc_dev,
3275 			    "not enough headroom, need %d found %zd, "
3276 			    "m_flags 0x%x m_len %d\n",
3277 			    space, M_LEADINGSPACE(m0), m0->m_flags, m0->m_len);
3278 			ieee80211_dump_pkt(ic,
3279 			    mtod(m0, const uint8_t *), m0->m_len, 0, -1);
3280 			m_freem(m0);
3281 			sc->sc_stats.mst_tx_noheadroom++;
3282 			return EIO;
3283 		}
3284 		M_PREPEND(m0, space, M_NOWAIT);
3285 	}
3286 	tr = mtod(m0, struct mwltxrec *);
3287 	if (wh != (struct ieee80211_frame *) &tr->wh)
3288 		ovbcopy(wh, &tr->wh, hdrlen);
3289 	/*
3290 	 * Note: the "firmware length" is actually the length
3291 	 * of the fully formed "802.11 payload".  That is, it's
3292 	 * everything except for the 802.11 header.  In particular
3293 	 * this includes all crypto material including the MIC!
3294 	 */
3295 	tr->fwlen = htole16(pktlen - hdrlen);
3296 
3297 	/*
3298 	 * Load the DMA map so any coalescing is done.  This
3299 	 * also calculates the number of descriptors we need.
3300 	 */
3301 	error = mwl_tx_dmasetup(sc, bf, m0);
3302 	if (error != 0) {
3303 		/* NB: stat collected in mwl_tx_dmasetup */
3304 		DPRINTF(sc, MWL_DEBUG_XMIT,
3305 		    "%s: unable to setup dma\n", __func__);
3306 		return error;
3307 	}
3308 	bf->bf_node = ni;			/* NB: held reference */
3309 	m0 = bf->bf_m;				/* NB: may have changed */
3310 	tr = mtod(m0, struct mwltxrec *);
3311 	wh = (struct ieee80211_frame *)&tr->wh;
3312 
3313 	/*
3314 	 * Formulate tx descriptor.
3315 	 */
3316 	ds = bf->bf_desc;
3317 	txq = bf->bf_txq;
3318 
3319 	ds->QosCtrl = qos;			/* NB: already little-endian */
3320 #if MWL_TXDESC == 1
3321 	/*
3322 	 * NB: multiframes should be zero because the descriptors
3323 	 *     are initialized to zero.  This should handle the case
3324 	 *     where the driver is built with MWL_TXDESC=1 but we are
3325 	 *     using firmware with multi-segment support.
3326 	 */
3327 	ds->PktPtr = htole32(bf->bf_segs[0].ds_addr);
3328 	ds->PktLen = htole16(bf->bf_segs[0].ds_len);
3329 #else
3330 	ds->multiframes = htole32(bf->bf_nseg);
3331 	ds->PktLen = htole16(m0->m_pkthdr.len);
3332 	for (i = 0; i < bf->bf_nseg; i++) {
3333 		ds->PktPtrArray[i] = htole32(bf->bf_segs[i].ds_addr);
3334 		ds->PktLenArray[i] = htole16(bf->bf_segs[i].ds_len);
3335 	}
3336 #endif
3337 	/* NB: pPhysNext, DataRate, and SapPktInfo setup once, don't touch */
3338 	ds->Format = 0;
3339 	ds->pad = 0;
3340 	ds->ack_wcb_addr = 0;
3341 
3342 	mn = MWL_NODE(ni);
3343 	/*
3344 	 * Select transmit rate.
3345 	 */
3346 	switch (wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK) {
3347 	case IEEE80211_FC0_TYPE_MGT:
3348 		sc->sc_stats.mst_tx_mgmt++;
3349 		/* fall thru... */
3350 	case IEEE80211_FC0_TYPE_CTL:
3351 		/* NB: assign to BE q to avoid bursting */
3352 		ds->TxPriority = MWL_WME_AC_BE;
3353 		break;
3354 	case IEEE80211_FC0_TYPE_DATA:
3355 		if (!ismcast) {
3356 			const struct ieee80211_txparam *tp = ni->ni_txparms;
3357 			/*
3358 			 * EAPOL frames get forced to a fixed rate and w/o
3359 			 * aggregation; otherwise check for any fixed rate
3360 			 * for the client (may depend on association state).
3361 			 */
3362 			if (m0->m_flags & M_EAPOL) {
3363 				const struct mwl_vap *mvp = MWL_VAP_CONST(vap);
3364 				ds->Format = mvp->mv_eapolformat;
3365 				ds->pad = htole16(
3366 				    EAGLE_TXD_FIXED_RATE | EAGLE_TXD_DONT_AGGR);
3367 			} else if (tp->ucastrate != IEEE80211_FIXED_RATE_NONE) {
3368 				/* XXX pre-calculate per node */
3369 				ds->Format = htole16(
3370 				    mwl_calcformat(tp->ucastrate, ni));
3371 				ds->pad = htole16(EAGLE_TXD_FIXED_RATE);
3372 			}
3373 			/* NB: EAPOL frames will never have qos set */
3374 			if (qos == 0)
3375 				ds->TxPriority = txq->qnum;
3376 #if MWL_MAXBA > 3
3377 			else if (mwl_bastream_match(&mn->mn_ba[3], qos))
3378 				ds->TxPriority = mn->mn_ba[3].txq;
3379 #endif
3380 #if MWL_MAXBA > 2
3381 			else if (mwl_bastream_match(&mn->mn_ba[2], qos))
3382 				ds->TxPriority = mn->mn_ba[2].txq;
3383 #endif
3384 #if MWL_MAXBA > 1
3385 			else if (mwl_bastream_match(&mn->mn_ba[1], qos))
3386 				ds->TxPriority = mn->mn_ba[1].txq;
3387 #endif
3388 #if MWL_MAXBA > 0
3389 			else if (mwl_bastream_match(&mn->mn_ba[0], qos))
3390 				ds->TxPriority = mn->mn_ba[0].txq;
3391 #endif
3392 			else
3393 				ds->TxPriority = txq->qnum;
3394 		} else
3395 			ds->TxPriority = txq->qnum;
3396 		break;
3397 	default:
3398 		if_printf(ifp, "bogus frame type 0x%x (%s)\n",
3399 			wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK, __func__);
3400 		sc->sc_stats.mst_tx_badframetype++;
3401 		m_freem(m0);
3402 		return EIO;
3403 	}
3404 
3405 	if (IFF_DUMPPKTS_XMIT(sc))
3406 		ieee80211_dump_pkt(ic,
3407 		    mtod(m0, const uint8_t *)+sizeof(uint16_t),
3408 		    m0->m_len - sizeof(uint16_t), ds->DataRate, -1);
3409 
3410 	MWL_TXQ_LOCK(txq);
3411 	ds->Status = htole32(EAGLE_TXD_STATUS_FW_OWNED);
3412 	STAILQ_INSERT_TAIL(&txq->active, bf, bf_list);
3413 	MWL_TXDESC_SYNC(txq, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
3414 
3415 	ifp->if_opackets++;
3416 	sc->sc_tx_timer = 5;
3417 	MWL_TXQ_UNLOCK(txq);
3418 
3419 	return 0;
3420 #undef	IEEE80211_DIR_DSTODS
3421 }
3422 
3423 static __inline int
3424 mwl_cvtlegacyrix(int rix)
3425 {
3426 #define	N(x)	(sizeof(x)/sizeof(x[0]))
3427 	static const int ieeerates[] =
3428 	    { 2, 4, 11, 22, 44, 12, 18, 24, 36, 48, 72, 96, 108 };
3429 	return (rix < N(ieeerates) ? ieeerates[rix] : 0);
3430 #undef N
3431 }
3432 
3433 /*
3434  * Process completed xmit descriptors from the specified queue.
3435  */
3436 static int
3437 mwl_tx_processq(struct mwl_softc *sc, struct mwl_txq *txq)
3438 {
3439 #define	EAGLE_TXD_STATUS_MCAST \
3440 	(EAGLE_TXD_STATUS_MULTICAST_TX | EAGLE_TXD_STATUS_BROADCAST_TX)
3441 	struct ifnet *ifp = sc->sc_ifp;
3442 	struct ieee80211com *ic = ifp->if_l2com;
3443 	struct mwl_txbuf *bf;
3444 	struct mwl_txdesc *ds;
3445 	struct ieee80211_node *ni;
3446 	struct mwl_node *an;
3447 	int nreaped;
3448 	uint32_t status;
3449 
3450 	DPRINTF(sc, MWL_DEBUG_TX_PROC, "%s: tx queue %u\n", __func__, txq->qnum);
3451 	for (nreaped = 0;; nreaped++) {
3452 		MWL_TXQ_LOCK(txq);
3453 		bf = STAILQ_FIRST(&txq->active);
3454 		if (bf == NULL) {
3455 			MWL_TXQ_UNLOCK(txq);
3456 			break;
3457 		}
3458 		ds = bf->bf_desc;
3459 		MWL_TXDESC_SYNC(txq, ds,
3460 		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
3461 		if (ds->Status & htole32(EAGLE_TXD_STATUS_FW_OWNED)) {
3462 			MWL_TXQ_UNLOCK(txq);
3463 			break;
3464 		}
3465 		STAILQ_REMOVE_HEAD(&txq->active, bf_list);
3466 		MWL_TXQ_UNLOCK(txq);
3467 
3468 #ifdef MWL_DEBUG
3469 		if (sc->sc_debug & MWL_DEBUG_XMIT_DESC)
3470 			mwl_printtxbuf(bf, txq->qnum, nreaped);
3471 #endif
3472 		ni = bf->bf_node;
3473 		if (ni != NULL) {
3474 			an = MWL_NODE(ni);
3475 			status = le32toh(ds->Status);
3476 			if (status & EAGLE_TXD_STATUS_OK) {
3477 				uint16_t Format = le16toh(ds->Format);
3478 				uint8_t txant = MS(Format, EAGLE_TXD_ANTENNA);
3479 
3480 				sc->sc_stats.mst_ant_tx[txant]++;
3481 				if (status & EAGLE_TXD_STATUS_OK_RETRY)
3482 					sc->sc_stats.mst_tx_retries++;
3483 				if (status & EAGLE_TXD_STATUS_OK_MORE_RETRY)
3484 					sc->sc_stats.mst_tx_mretries++;
3485 				if (txq->qnum >= MWL_WME_AC_VO)
3486 					ic->ic_wme.wme_hipri_traffic++;
3487 				ni->ni_txrate = MS(Format, EAGLE_TXD_RATE);
3488 				if ((Format & EAGLE_TXD_FORMAT_HT) == 0) {
3489 					ni->ni_txrate = mwl_cvtlegacyrix(
3490 					    ni->ni_txrate);
3491 				} else
3492 					ni->ni_txrate |= IEEE80211_RATE_MCS;
3493 				sc->sc_stats.mst_tx_rate = ni->ni_txrate;
3494 			} else {
3495 				if (status & EAGLE_TXD_STATUS_FAILED_LINK_ERROR)
3496 					sc->sc_stats.mst_tx_linkerror++;
3497 				if (status & EAGLE_TXD_STATUS_FAILED_XRETRY)
3498 					sc->sc_stats.mst_tx_xretries++;
3499 				if (status & EAGLE_TXD_STATUS_FAILED_AGING)
3500 					sc->sc_stats.mst_tx_aging++;
3501 				if (bf->bf_m->m_flags & M_FF)
3502 					sc->sc_stats.mst_ff_txerr++;
3503 			}
3504 			/*
3505 			 * Do any tx complete callback.  Note this must
3506 			 * be done before releasing the node reference.
3507 			 * XXX no way to figure out if frame was ACK'd
3508 			 */
3509 			if (bf->bf_m->m_flags & M_TXCB) {
3510 				/* XXX strip fw len in case header inspected */
3511 				m_adj(bf->bf_m, sizeof(uint16_t));
3512 				ieee80211_process_callback(ni, bf->bf_m,
3513 					(status & EAGLE_TXD_STATUS_OK) == 0);
3514 			}
3515 			/*
3516 			 * Reclaim reference to node.
3517 			 *
3518 			 * NB: the node may be reclaimed here if, for example
3519 			 *     this is a DEAUTH message that was sent and the
3520 			 *     node was timed out due to inactivity.
3521 			 */
3522 			ieee80211_free_node(ni);
3523 		}
3524 		ds->Status = htole32(EAGLE_TXD_STATUS_IDLE);
3525 
3526 		bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap,
3527 		    BUS_DMASYNC_POSTWRITE);
3528 		bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap);
3529 		m_freem(bf->bf_m);
3530 
3531 		mwl_puttxbuf_tail(txq, bf);
3532 	}
3533 	return nreaped;
3534 #undef EAGLE_TXD_STATUS_MCAST
3535 }
3536 
3537 /*
3538  * Deferred processing of transmit interrupt; special-cased
3539  * for four hardware queues, 0-3.
3540  */
3541 static void
3542 mwl_tx_proc(void *arg, int npending)
3543 {
3544 	struct mwl_softc *sc = arg;
3545 	struct ifnet *ifp = sc->sc_ifp;
3546 	int nreaped;
3547 
3548 	/*
3549 	 * Process each active queue.
3550 	 */
3551 	nreaped = 0;
3552 	if (!STAILQ_EMPTY(&sc->sc_txq[0].active))
3553 		nreaped += mwl_tx_processq(sc, &sc->sc_txq[0]);
3554 	if (!STAILQ_EMPTY(&sc->sc_txq[1].active))
3555 		nreaped += mwl_tx_processq(sc, &sc->sc_txq[1]);
3556 	if (!STAILQ_EMPTY(&sc->sc_txq[2].active))
3557 		nreaped += mwl_tx_processq(sc, &sc->sc_txq[2]);
3558 	if (!STAILQ_EMPTY(&sc->sc_txq[3].active))
3559 		nreaped += mwl_tx_processq(sc, &sc->sc_txq[3]);
3560 
3561 	if (nreaped != 0) {
3562 		ifp->if_drv_flags &= ~IFF_DRV_OACTIVE;
3563 		sc->sc_tx_timer = 0;
3564 		if (!IFQ_IS_EMPTY(&ifp->if_snd)) {
3565 			/* NB: kick fw; the tx thread may have been preempted */
3566 			mwl_hal_txstart(sc->sc_mh, 0);
3567 			mwl_start(ifp);
3568 		}
3569 	}
3570 }
3571 
3572 static void
3573 mwl_tx_draintxq(struct mwl_softc *sc, struct mwl_txq *txq)
3574 {
3575 	struct ieee80211_node *ni;
3576 	struct mwl_txbuf *bf;
3577 	u_int ix;
3578 
3579 	/*
3580 	 * NB: this assumes output has been stopped and
3581 	 *     we do not need to block mwl_tx_tasklet
3582 	 */
3583 	for (ix = 0;; ix++) {
3584 		MWL_TXQ_LOCK(txq);
3585 		bf = STAILQ_FIRST(&txq->active);
3586 		if (bf == NULL) {
3587 			MWL_TXQ_UNLOCK(txq);
3588 			break;
3589 		}
3590 		STAILQ_REMOVE_HEAD(&txq->active, bf_list);
3591 		MWL_TXQ_UNLOCK(txq);
3592 #ifdef MWL_DEBUG
3593 		if (sc->sc_debug & MWL_DEBUG_RESET) {
3594 			struct ifnet *ifp = sc->sc_ifp;
3595 			struct ieee80211com *ic = ifp->if_l2com;
3596 			const struct mwltxrec *tr =
3597 			    mtod(bf->bf_m, const struct mwltxrec *);
3598 			mwl_printtxbuf(bf, txq->qnum, ix);
3599 			ieee80211_dump_pkt(ic, (const uint8_t *)&tr->wh,
3600 				bf->bf_m->m_len - sizeof(tr->fwlen), 0, -1);
3601 		}
3602 #endif /* MWL_DEBUG */
3603 		bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap);
3604 		ni = bf->bf_node;
3605 		if (ni != NULL) {
3606 			/*
3607 			 * Reclaim node reference.
3608 			 */
3609 			ieee80211_free_node(ni);
3610 		}
3611 		m_freem(bf->bf_m);
3612 
3613 		mwl_puttxbuf_tail(txq, bf);
3614 	}
3615 }
3616 
3617 /*
3618  * Drain the transmit queues and reclaim resources.
3619  */
3620 static void
3621 mwl_draintxq(struct mwl_softc *sc)
3622 {
3623 	struct ifnet *ifp = sc->sc_ifp;
3624 	int i;
3625 
3626 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
3627 		mwl_tx_draintxq(sc, &sc->sc_txq[i]);
3628 	ifp->if_drv_flags &= ~IFF_DRV_OACTIVE;
3629 	sc->sc_tx_timer = 0;
3630 }
3631 
3632 #ifdef MWL_DIAGAPI
3633 /*
3634  * Reset the transmit queues to a pristine state after a fw download.
3635  */
3636 static void
3637 mwl_resettxq(struct mwl_softc *sc)
3638 {
3639 	int i;
3640 
3641 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
3642 		mwl_txq_reset(sc, &sc->sc_txq[i]);
3643 }
3644 #endif /* MWL_DIAGAPI */
3645 
3646 /*
3647  * Clear the transmit queues of any frames submitted for the
3648  * specified vap.  This is done when the vap is deleted so we
3649  * don't potentially reference the vap after it is gone.
3650  * Note we cannot remove the frames; we only reclaim the node
3651  * reference.
3652  */
3653 static void
3654 mwl_cleartxq(struct mwl_softc *sc, struct ieee80211vap *vap)
3655 {
3656 	struct mwl_txq *txq;
3657 	struct mwl_txbuf *bf;
3658 	int i;
3659 
3660 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++) {
3661 		txq = &sc->sc_txq[i];
3662 		MWL_TXQ_LOCK(txq);
3663 		STAILQ_FOREACH(bf, &txq->active, bf_list) {
3664 			struct ieee80211_node *ni = bf->bf_node;
3665 			if (ni != NULL && ni->ni_vap == vap) {
3666 				bf->bf_node = NULL;
3667 				ieee80211_free_node(ni);
3668 			}
3669 		}
3670 		MWL_TXQ_UNLOCK(txq);
3671 	}
3672 }
3673 
3674 static int
3675 mwl_recv_action(struct ieee80211_node *ni, const struct ieee80211_frame *wh,
3676 	const uint8_t *frm, const uint8_t *efrm)
3677 {
3678 	struct mwl_softc *sc = ni->ni_ic->ic_ifp->if_softc;
3679 	const struct ieee80211_action *ia;
3680 
3681 	ia = (const struct ieee80211_action *) frm;
3682 	if (ia->ia_category == IEEE80211_ACTION_CAT_HT &&
3683 	    ia->ia_action == IEEE80211_ACTION_HT_MIMOPWRSAVE) {
3684 		const struct ieee80211_action_ht_mimopowersave *mps =
3685 		    (const struct ieee80211_action_ht_mimopowersave *) ia;
3686 
3687 		mwl_hal_setmimops(sc->sc_mh, ni->ni_macaddr,
3688 		    mps->am_control & IEEE80211_A_HT_MIMOPWRSAVE_ENA,
3689 		    MS(mps->am_control, IEEE80211_A_HT_MIMOPWRSAVE_MODE));
3690 		return 0;
3691 	} else
3692 		return sc->sc_recv_action(ni, wh, frm, efrm);
3693 }
3694 
3695 static int
3696 mwl_addba_request(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap,
3697 	int dialogtoken, int baparamset, int batimeout)
3698 {
3699 	struct mwl_softc *sc = ni->ni_ic->ic_ifp->if_softc;
3700 	struct ieee80211vap *vap = ni->ni_vap;
3701 	struct mwl_node *mn = MWL_NODE(ni);
3702 	struct mwl_bastate *bas;
3703 
3704 	bas = tap->txa_private;
3705 	if (bas == NULL) {
3706 		const MWL_HAL_BASTREAM *sp;
3707 		/*
3708 		 * Check for a free BA stream slot.
3709 		 */
3710 #if MWL_MAXBA > 3
3711 		if (mn->mn_ba[3].bastream == NULL)
3712 			bas = &mn->mn_ba[3];
3713 		else
3714 #endif
3715 #if MWL_MAXBA > 2
3716 		if (mn->mn_ba[2].bastream == NULL)
3717 			bas = &mn->mn_ba[2];
3718 		else
3719 #endif
3720 #if MWL_MAXBA > 1
3721 		if (mn->mn_ba[1].bastream == NULL)
3722 			bas = &mn->mn_ba[1];
3723 		else
3724 #endif
3725 #if MWL_MAXBA > 0
3726 		if (mn->mn_ba[0].bastream == NULL)
3727 			bas = &mn->mn_ba[0];
3728 		else
3729 #endif
3730 		{
3731 			/* sta already has max BA streams */
3732 			/* XXX assign BA stream to highest priority tid */
3733 			DPRINTF(sc, MWL_DEBUG_AMPDU,
3734 			    "%s: already has max bastreams\n", __func__);
3735 			sc->sc_stats.mst_ampdu_reject++;
3736 			return 0;
3737 		}
3738 		/* NB: no held reference to ni */
3739 		sp = mwl_hal_bastream_alloc(MWL_VAP(vap)->mv_hvap,
3740 		    (baparamset & IEEE80211_BAPS_POLICY_IMMEDIATE) != 0,
3741 		    ni->ni_macaddr, tap->txa_tid, ni->ni_htparam,
3742 		    ni, tap);
3743 		if (sp == NULL) {
3744 			/*
3745 			 * No available stream, return 0 so no
3746 			 * a-mpdu aggregation will be done.
3747 			 */
3748 			DPRINTF(sc, MWL_DEBUG_AMPDU,
3749 			    "%s: no bastream available\n", __func__);
3750 			sc->sc_stats.mst_ampdu_nostream++;
3751 			return 0;
3752 		}
3753 		DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: alloc bastream %p\n",
3754 		    __func__, sp);
3755 		/* NB: qos is left zero so we won't match in mwl_tx_start */
3756 		bas->bastream = sp;
3757 		tap->txa_private = bas;
3758 	}
3759 	/* fetch current seq# from the firmware; if available */
3760 	if (mwl_hal_bastream_get_seqno(sc->sc_mh, bas->bastream,
3761 	    vap->iv_opmode == IEEE80211_M_STA ? vap->iv_myaddr : ni->ni_macaddr,
3762 	    &tap->txa_start) != 0)
3763 		tap->txa_start = 0;
3764 	return sc->sc_addba_request(ni, tap, dialogtoken, baparamset, batimeout);
3765 }
3766 
3767 static int
3768 mwl_addba_response(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap,
3769 	int code, int baparamset, int batimeout)
3770 {
3771 	struct mwl_softc *sc = ni->ni_ic->ic_ifp->if_softc;
3772 	struct mwl_bastate *bas;
3773 
3774 	bas = tap->txa_private;
3775 	if (bas == NULL) {
3776 		/* XXX should not happen */
3777 		DPRINTF(sc, MWL_DEBUG_AMPDU,
3778 		    "%s: no BA stream allocated, TID %d\n",
3779 		    __func__, tap->txa_tid);
3780 		sc->sc_stats.mst_addba_nostream++;
3781 		return 0;
3782 	}
3783 	if (code == IEEE80211_STATUS_SUCCESS) {
3784 		struct ieee80211vap *vap = ni->ni_vap;
3785 		int bufsiz, error;
3786 
3787 		/*
3788 		 * Tell the firmware to setup the BA stream;
3789 		 * we know resources are available because we
3790 		 * pre-allocated one before forming the request.
3791 		 */
3792 		bufsiz = MS(baparamset, IEEE80211_BAPS_BUFSIZ);
3793 		if (bufsiz == 0)
3794 			bufsiz = IEEE80211_AGGR_BAWMAX;
3795 		error = mwl_hal_bastream_create(MWL_VAP(vap)->mv_hvap,
3796 		    bas->bastream, bufsiz, bufsiz, tap->txa_start);
3797 		if (error != 0) {
3798 			/*
3799 			 * Setup failed, return immediately so no a-mpdu
3800 			 * aggregation will be done.
3801 			 */
3802 			mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3803 			mwl_bastream_free(bas);
3804 			tap->txa_private = NULL;
3805 
3806 			DPRINTF(sc, MWL_DEBUG_AMPDU,
3807 			    "%s: create failed, error %d, bufsiz %d TID %d "
3808 			    "htparam 0x%x\n", __func__, error, bufsiz,
3809 			    tap->txa_tid, ni->ni_htparam);
3810 			sc->sc_stats.mst_bacreate_failed++;
3811 			return 0;
3812 		}
3813 		/* NB: cache txq to avoid ptr indirect */
3814 		mwl_bastream_setup(bas, tap->txa_tid, bas->bastream->txq);
3815 		DPRINTF(sc, MWL_DEBUG_AMPDU,
3816 		    "%s: bastream %p assigned to txq %d TID %d bufsiz %d "
3817 		    "htparam 0x%x\n", __func__, bas->bastream,
3818 		    bas->txq, tap->txa_tid, bufsiz, ni->ni_htparam);
3819 	} else {
3820 		/*
3821 		 * Other side NAK'd us; return the resources.
3822 		 */
3823 		DPRINTF(sc, MWL_DEBUG_AMPDU,
3824 		    "%s: request failed with code %d, destroy bastream %p\n",
3825 		    __func__, code, bas->bastream);
3826 		mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3827 		mwl_bastream_free(bas);
3828 		tap->txa_private = NULL;
3829 	}
3830 	/* NB: firmware sends BAR so we don't need to */
3831 	return sc->sc_addba_response(ni, tap, code, baparamset, batimeout);
3832 }
3833 
3834 static void
3835 mwl_addba_stop(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap)
3836 {
3837 	struct mwl_softc *sc = ni->ni_ic->ic_ifp->if_softc;
3838 	struct mwl_bastate *bas;
3839 
3840 	bas = tap->txa_private;
3841 	if (bas != NULL) {
3842 		DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: destroy bastream %p\n",
3843 		    __func__, bas->bastream);
3844 		mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3845 		mwl_bastream_free(bas);
3846 		tap->txa_private = NULL;
3847 	}
3848 	sc->sc_addba_stop(ni, tap);
3849 }
3850 
3851 /*
3852  * Setup the rx data structures.  This should only be
3853  * done once or we may get out of sync with the firmware.
3854  */
3855 static int
3856 mwl_startrecv(struct mwl_softc *sc)
3857 {
3858 	if (!sc->sc_recvsetup) {
3859 		struct mwl_rxbuf *bf, *prev;
3860 		struct mwl_rxdesc *ds;
3861 
3862 		prev = NULL;
3863 		STAILQ_FOREACH(bf, &sc->sc_rxbuf, bf_list) {
3864 			int error = mwl_rxbuf_init(sc, bf);
3865 			if (error != 0) {
3866 				DPRINTF(sc, MWL_DEBUG_RECV,
3867 					"%s: mwl_rxbuf_init failed %d\n",
3868 					__func__, error);
3869 				return error;
3870 			}
3871 			if (prev != NULL) {
3872 				ds = prev->bf_desc;
3873 				ds->pPhysNext = htole32(bf->bf_daddr);
3874 			}
3875 			prev = bf;
3876 		}
3877 		if (prev != NULL) {
3878 			ds = prev->bf_desc;
3879 			ds->pPhysNext =
3880 			    htole32(STAILQ_FIRST(&sc->sc_rxbuf)->bf_daddr);
3881 		}
3882 		sc->sc_recvsetup = 1;
3883 	}
3884 	mwl_mode_init(sc);		/* set filters, etc. */
3885 	return 0;
3886 }
3887 
3888 static MWL_HAL_APMODE
3889 mwl_getapmode(const struct ieee80211vap *vap, struct ieee80211_channel *chan)
3890 {
3891 	MWL_HAL_APMODE mode;
3892 
3893 	if (IEEE80211_IS_CHAN_HT(chan)) {
3894 		if (vap->iv_flags_ht & IEEE80211_FHT_PUREN)
3895 			mode = AP_MODE_N_ONLY;
3896 		else if (IEEE80211_IS_CHAN_5GHZ(chan))
3897 			mode = AP_MODE_AandN;
3898 		else if (vap->iv_flags & IEEE80211_F_PUREG)
3899 			mode = AP_MODE_GandN;
3900 		else
3901 			mode = AP_MODE_BandGandN;
3902 	} else if (IEEE80211_IS_CHAN_ANYG(chan)) {
3903 		if (vap->iv_flags & IEEE80211_F_PUREG)
3904 			mode = AP_MODE_G_ONLY;
3905 		else
3906 			mode = AP_MODE_MIXED;
3907 	} else if (IEEE80211_IS_CHAN_B(chan))
3908 		mode = AP_MODE_B_ONLY;
3909 	else if (IEEE80211_IS_CHAN_A(chan))
3910 		mode = AP_MODE_A_ONLY;
3911 	else
3912 		mode = AP_MODE_MIXED;		/* XXX should not happen? */
3913 	return mode;
3914 }
3915 
3916 static int
3917 mwl_setapmode(struct ieee80211vap *vap, struct ieee80211_channel *chan)
3918 {
3919 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
3920 	return mwl_hal_setapmode(hvap, mwl_getapmode(vap, chan));
3921 }
3922 
3923 /*
3924  * Set/change channels.
3925  */
3926 static int
3927 mwl_chan_set(struct mwl_softc *sc, struct ieee80211_channel *chan)
3928 {
3929 	struct mwl_hal *mh = sc->sc_mh;
3930 	struct ifnet *ifp = sc->sc_ifp;
3931 	struct ieee80211com *ic = ifp->if_l2com;
3932 	MWL_HAL_CHANNEL hchan;
3933 	int maxtxpow;
3934 
3935 	DPRINTF(sc, MWL_DEBUG_RESET, "%s: chan %u MHz/flags 0x%x\n",
3936 	    __func__, chan->ic_freq, chan->ic_flags);
3937 
3938 	/*
3939 	 * Convert to a HAL channel description with
3940 	 * the flags constrained to reflect the current
3941 	 * operating mode.
3942 	 */
3943 	mwl_mapchan(&hchan, chan);
3944 	mwl_hal_intrset(mh, 0);		/* disable interrupts */
3945 #if 0
3946 	mwl_draintxq(sc);		/* clear pending tx frames */
3947 #endif
3948 	mwl_hal_setchannel(mh, &hchan);
3949 	/*
3950 	 * Tx power is cap'd by the regulatory setting and
3951 	 * possibly a user-set limit.  We pass the min of
3952 	 * these to the hal to apply them to the cal data
3953 	 * for this channel.
3954 	 * XXX min bound?
3955 	 */
3956 	maxtxpow = 2*chan->ic_maxregpower;
3957 	if (maxtxpow > ic->ic_txpowlimit)
3958 		maxtxpow = ic->ic_txpowlimit;
3959 	mwl_hal_settxpower(mh, &hchan, maxtxpow / 2);
3960 	/* NB: potentially change mcast/mgt rates */
3961 	mwl_setcurchanrates(sc);
3962 
3963 	/*
3964 	 * Update internal state.
3965 	 */
3966 	sc->sc_tx_th.wt_chan_freq = htole16(chan->ic_freq);
3967 	sc->sc_rx_th.wr_chan_freq = htole16(chan->ic_freq);
3968 	if (IEEE80211_IS_CHAN_A(chan)) {
3969 		sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_A);
3970 		sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_A);
3971 	} else if (IEEE80211_IS_CHAN_ANYG(chan)) {
3972 		sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_G);
3973 		sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_G);
3974 	} else {
3975 		sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_B);
3976 		sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_B);
3977 	}
3978 	sc->sc_curchan = hchan;
3979 	mwl_hal_intrset(mh, sc->sc_imask);
3980 
3981 	return 0;
3982 }
3983 
3984 static void
3985 mwl_scan_start(struct ieee80211com *ic)
3986 {
3987 	struct ifnet *ifp = ic->ic_ifp;
3988 	struct mwl_softc *sc = ifp->if_softc;
3989 
3990 	DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__);
3991 }
3992 
3993 static void
3994 mwl_scan_end(struct ieee80211com *ic)
3995 {
3996 	struct ifnet *ifp = ic->ic_ifp;
3997 	struct mwl_softc *sc = ifp->if_softc;
3998 
3999 	DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__);
4000 }
4001 
4002 static void
4003 mwl_set_channel(struct ieee80211com *ic)
4004 {
4005 	struct ifnet *ifp = ic->ic_ifp;
4006 	struct mwl_softc *sc = ifp->if_softc;
4007 
4008 	(void) mwl_chan_set(sc, ic->ic_curchan);
4009 }
4010 
4011 /*
4012  * Handle a channel switch request.  We inform the firmware
4013  * and mark the global state to suppress various actions.
4014  * NB: we issue only one request to the fw; we may be called
4015  * multiple times if there are multiple vap's.
4016  */
4017 static void
4018 mwl_startcsa(struct ieee80211vap *vap)
4019 {
4020 	struct ieee80211com *ic = vap->iv_ic;
4021 	struct mwl_softc *sc = ic->ic_ifp->if_softc;
4022 	MWL_HAL_CHANNEL hchan;
4023 
4024 	if (sc->sc_csapending)
4025 		return;
4026 
4027 	mwl_mapchan(&hchan, ic->ic_csa_newchan);
4028 	/* 1 =>'s quiet channel */
4029 	mwl_hal_setchannelswitchie(sc->sc_mh, &hchan, 1, ic->ic_csa_count);
4030 	sc->sc_csapending = 1;
4031 }
4032 
4033 /*
4034  * Plumb any static WEP key for the station.  This is
4035  * necessary as we must propagate the key from the
4036  * global key table of the vap to each sta db entry.
4037  */
4038 static void
4039 mwl_setanywepkey(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN])
4040 {
4041 	if ((vap->iv_flags & (IEEE80211_F_PRIVACY|IEEE80211_F_WPA)) ==
4042 		IEEE80211_F_PRIVACY &&
4043 	    vap->iv_def_txkey != IEEE80211_KEYIX_NONE &&
4044 	    vap->iv_nw_keys[vap->iv_def_txkey].wk_keyix != IEEE80211_KEYIX_NONE)
4045 		(void) mwl_key_set(vap, &vap->iv_nw_keys[vap->iv_def_txkey], mac);
4046 }
4047 
4048 static int
4049 mwl_peerstadb(struct ieee80211_node *ni, int aid, int staid, MWL_HAL_PEERINFO *pi)
4050 {
4051 #define	WME(ie) ((const struct ieee80211_wme_info *) ie)
4052 	struct ieee80211vap *vap = ni->ni_vap;
4053 	struct mwl_hal_vap *hvap;
4054 	int error;
4055 
4056 	if (vap->iv_opmode == IEEE80211_M_WDS) {
4057 		/*
4058 		 * WDS vap's do not have a f/w vap; instead they piggyback
4059 		 * on an AP vap and we must install the sta db entry and
4060 		 * crypto state using that AP's handle (the WDS vap has none).
4061 		 */
4062 		hvap = MWL_VAP(vap)->mv_ap_hvap;
4063 	} else
4064 		hvap = MWL_VAP(vap)->mv_hvap;
4065 	error = mwl_hal_newstation(hvap, ni->ni_macaddr,
4066 	    aid, staid, pi,
4067 	    ni->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT),
4068 	    ni->ni_ies.wme_ie != NULL ? WME(ni->ni_ies.wme_ie)->wme_info : 0);
4069 	if (error == 0) {
4070 		/*
4071 		 * Setup security for this station.  For sta mode this is
4072 		 * needed even though do the same thing on transition to
4073 		 * AUTH state because the call to mwl_hal_newstation
4074 		 * clobbers the crypto state we setup.
4075 		 */
4076 		mwl_setanywepkey(vap, ni->ni_macaddr);
4077 	}
4078 	return error;
4079 #undef WME
4080 }
4081 
4082 static void
4083 mwl_setglobalkeys(struct ieee80211vap *vap)
4084 {
4085 	struct ieee80211_key *wk;
4086 
4087 	wk = &vap->iv_nw_keys[0];
4088 	for (; wk < &vap->iv_nw_keys[IEEE80211_WEP_NKID]; wk++)
4089 		if (wk->wk_keyix != IEEE80211_KEYIX_NONE)
4090 			(void) mwl_key_set(vap, wk, vap->iv_myaddr);
4091 }
4092 
4093 /*
4094  * Convert a legacy rate set to a firmware bitmask.
4095  */
4096 static uint32_t
4097 get_rate_bitmap(const struct ieee80211_rateset *rs)
4098 {
4099 	uint32_t rates;
4100 	int i;
4101 
4102 	rates = 0;
4103 	for (i = 0; i < rs->rs_nrates; i++)
4104 		switch (rs->rs_rates[i] & IEEE80211_RATE_VAL) {
4105 		case 2:	  rates |= 0x001; break;
4106 		case 4:	  rates |= 0x002; break;
4107 		case 11:  rates |= 0x004; break;
4108 		case 22:  rates |= 0x008; break;
4109 		case 44:  rates |= 0x010; break;
4110 		case 12:  rates |= 0x020; break;
4111 		case 18:  rates |= 0x040; break;
4112 		case 24:  rates |= 0x080; break;
4113 		case 36:  rates |= 0x100; break;
4114 		case 48:  rates |= 0x200; break;
4115 		case 72:  rates |= 0x400; break;
4116 		case 96:  rates |= 0x800; break;
4117 		case 108: rates |= 0x1000; break;
4118 		}
4119 	return rates;
4120 }
4121 
4122 /*
4123  * Construct an HT firmware bitmask from an HT rate set.
4124  */
4125 static uint32_t
4126 get_htrate_bitmap(const struct ieee80211_htrateset *rs)
4127 {
4128 	uint32_t rates;
4129 	int i;
4130 
4131 	rates = 0;
4132 	for (i = 0; i < rs->rs_nrates; i++) {
4133 		if (rs->rs_rates[i] < 16)
4134 			rates |= 1<<rs->rs_rates[i];
4135 	}
4136 	return rates;
4137 }
4138 
4139 /*
4140  * Craft station database entry for station.
4141  * NB: use host byte order here, the hal handles byte swapping.
4142  */
4143 static MWL_HAL_PEERINFO *
4144 mkpeerinfo(MWL_HAL_PEERINFO *pi, const struct ieee80211_node *ni)
4145 {
4146 	const struct ieee80211vap *vap = ni->ni_vap;
4147 
4148 	memset(pi, 0, sizeof(*pi));
4149 	pi->LegacyRateBitMap = get_rate_bitmap(&ni->ni_rates);
4150 	pi->CapInfo = ni->ni_capinfo;
4151 	if (ni->ni_flags & IEEE80211_NODE_HT) {
4152 		/* HT capabilities, etc */
4153 		pi->HTCapabilitiesInfo = ni->ni_htcap;
4154 		/* XXX pi.HTCapabilitiesInfo */
4155 	        pi->MacHTParamInfo = ni->ni_htparam;
4156 		pi->HTRateBitMap = get_htrate_bitmap(&ni->ni_htrates);
4157 		pi->AddHtInfo.ControlChan = ni->ni_htctlchan;
4158 		pi->AddHtInfo.AddChan = ni->ni_ht2ndchan;
4159 		pi->AddHtInfo.OpMode = ni->ni_htopmode;
4160 		pi->AddHtInfo.stbc = ni->ni_htstbc;
4161 
4162 		/* constrain according to local configuration */
4163 		if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI40) == 0)
4164 			pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI40;
4165 		if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI20) == 0)
4166 			pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI20;
4167 		if (ni->ni_chw != 40)
4168 			pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_CHWIDTH40;
4169 	}
4170 	return pi;
4171 }
4172 
4173 /*
4174  * Re-create the local sta db entry for a vap to ensure
4175  * up to date WME state is pushed to the firmware.  Because
4176  * this resets crypto state this must be followed by a
4177  * reload of any keys in the global key table.
4178  */
4179 static int
4180 mwl_localstadb(struct ieee80211vap *vap)
4181 {
4182 #define	WME(ie) ((const struct ieee80211_wme_info *) ie)
4183 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
4184 	struct ieee80211_node *bss;
4185 	MWL_HAL_PEERINFO pi;
4186 	int error;
4187 
4188 	switch (vap->iv_opmode) {
4189 	case IEEE80211_M_STA:
4190 		bss = vap->iv_bss;
4191 		error = mwl_hal_newstation(hvap, vap->iv_myaddr, 0, 0,
4192 		    vap->iv_state == IEEE80211_S_RUN ?
4193 			mkpeerinfo(&pi, bss) : NULL,
4194 		    (bss->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT)),
4195 		    bss->ni_ies.wme_ie != NULL ?
4196 			WME(bss->ni_ies.wme_ie)->wme_info : 0);
4197 		if (error == 0)
4198 			mwl_setglobalkeys(vap);
4199 		break;
4200 	case IEEE80211_M_HOSTAP:
4201 	case IEEE80211_M_MBSS:
4202 		error = mwl_hal_newstation(hvap, vap->iv_myaddr,
4203 		    0, 0, NULL, vap->iv_flags & IEEE80211_F_WME, 0);
4204 		if (error == 0)
4205 			mwl_setglobalkeys(vap);
4206 		break;
4207 	default:
4208 		error = 0;
4209 		break;
4210 	}
4211 	return error;
4212 #undef WME
4213 }
4214 
4215 static int
4216 mwl_newstate(struct ieee80211vap *vap, enum ieee80211_state nstate, int arg)
4217 {
4218 	struct mwl_vap *mvp = MWL_VAP(vap);
4219 	struct mwl_hal_vap *hvap = mvp->mv_hvap;
4220 	struct ieee80211com *ic = vap->iv_ic;
4221 	struct ieee80211_node *ni = NULL;
4222 	struct ifnet *ifp = ic->ic_ifp;
4223 	struct mwl_softc *sc = ifp->if_softc;
4224 	struct mwl_hal *mh = sc->sc_mh;
4225 	enum ieee80211_state ostate = vap->iv_state;
4226 	int error;
4227 
4228 	DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: %s -> %s\n",
4229 	    vap->iv_ifp->if_xname, __func__,
4230 	    ieee80211_state_name[ostate], ieee80211_state_name[nstate]);
4231 
4232 	callout_stop(&sc->sc_timer);
4233 	/*
4234 	 * Clear current radar detection state.
4235 	 */
4236 	if (ostate == IEEE80211_S_CAC) {
4237 		/* stop quiet mode radar detection */
4238 		mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_STOP);
4239 	} else if (sc->sc_radarena) {
4240 		/* stop in-service radar detection */
4241 		mwl_hal_setradardetection(mh, DR_DFS_DISABLE);
4242 		sc->sc_radarena = 0;
4243 	}
4244 	/*
4245 	 * Carry out per-state actions before doing net80211 work.
4246 	 */
4247 	if (nstate == IEEE80211_S_INIT) {
4248 		/* NB: only ap+sta vap's have a fw entity */
4249 		if (hvap != NULL)
4250 			mwl_hal_stop(hvap);
4251 	} else if (nstate == IEEE80211_S_SCAN) {
4252 		mwl_hal_start(hvap);
4253 		/* NB: this disables beacon frames */
4254 		mwl_hal_setinframode(hvap);
4255 	} else if (nstate == IEEE80211_S_AUTH) {
4256 		/*
4257 		 * Must create a sta db entry in case a WEP key needs to
4258 		 * be plumbed.  This entry will be overwritten if we
4259 		 * associate; otherwise it will be reclaimed on node free.
4260 		 */
4261 		ni = vap->iv_bss;
4262 		MWL_NODE(ni)->mn_hvap = hvap;
4263 		(void) mwl_peerstadb(ni, 0, 0, NULL);
4264 	} else if (nstate == IEEE80211_S_CSA) {
4265 		/* XXX move to below? */
4266 		if (vap->iv_opmode == IEEE80211_M_HOSTAP ||
4267 		    vap->iv_opmode == IEEE80211_M_MBSS)
4268 			mwl_startcsa(vap);
4269 	} else if (nstate == IEEE80211_S_CAC) {
4270 		/* XXX move to below? */
4271 		/* stop ap xmit and enable quiet mode radar detection */
4272 		mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_START);
4273 	}
4274 
4275 	/*
4276 	 * Invoke the parent method to do net80211 work.
4277 	 */
4278 	error = mvp->mv_newstate(vap, nstate, arg);
4279 
4280 	/*
4281 	 * Carry out work that must be done after net80211 runs;
4282 	 * this work requires up to date state (e.g. iv_bss).
4283 	 */
4284 	if (error == 0 && nstate == IEEE80211_S_RUN) {
4285 		/* NB: collect bss node again, it may have changed */
4286 		ni = vap->iv_bss;
4287 
4288 		DPRINTF(sc, MWL_DEBUG_STATE,
4289 		    "%s: %s(RUN): iv_flags 0x%08x bintvl %d bssid %s "
4290 		    "capinfo 0x%04x chan %d\n",
4291 		    vap->iv_ifp->if_xname, __func__, vap->iv_flags,
4292 		    ni->ni_intval, ether_sprintf(ni->ni_bssid), ni->ni_capinfo,
4293 		    ieee80211_chan2ieee(ic, ic->ic_curchan));
4294 
4295 		/*
4296 		 * Recreate local sta db entry to update WME/HT state.
4297 		 */
4298 		mwl_localstadb(vap);
4299 		switch (vap->iv_opmode) {
4300 		case IEEE80211_M_HOSTAP:
4301 		case IEEE80211_M_MBSS:
4302 			if (ostate == IEEE80211_S_CAC) {
4303 				/* enable in-service radar detection */
4304 				mwl_hal_setradardetection(mh,
4305 				    DR_IN_SERVICE_MONITOR_START);
4306 				sc->sc_radarena = 1;
4307 			}
4308 			/*
4309 			 * Allocate and setup the beacon frame
4310 			 * (and related state).
4311 			 */
4312 			error = mwl_reset_vap(vap, IEEE80211_S_RUN);
4313 			if (error != 0) {
4314 				DPRINTF(sc, MWL_DEBUG_STATE,
4315 				    "%s: beacon setup failed, error %d\n",
4316 				    __func__, error);
4317 				goto bad;
4318 			}
4319 			/* NB: must be after setting up beacon */
4320 			mwl_hal_start(hvap);
4321 			break;
4322 		case IEEE80211_M_STA:
4323 			DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: aid 0x%x\n",
4324 			    vap->iv_ifp->if_xname, __func__, ni->ni_associd);
4325 			/*
4326 			 * Set state now that we're associated.
4327 			 */
4328 			mwl_hal_setassocid(hvap, ni->ni_bssid, ni->ni_associd);
4329 			mwl_setrates(vap);
4330 			mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold);
4331 			if ((vap->iv_flags & IEEE80211_F_DWDS) &&
4332 			    sc->sc_ndwdsvaps++ == 0)
4333 				mwl_hal_setdwds(mh, 1);
4334 			break;
4335 		case IEEE80211_M_WDS:
4336 			DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: bssid %s\n",
4337 			    vap->iv_ifp->if_xname, __func__,
4338 			    ether_sprintf(ni->ni_bssid));
4339 			mwl_seteapolformat(vap);
4340 			break;
4341 		default:
4342 			break;
4343 		}
4344 		/*
4345 		 * Set CS mode according to operating channel;
4346 		 * this mostly an optimization for 5GHz.
4347 		 *
4348 		 * NB: must follow mwl_hal_start which resets csmode
4349 		 */
4350 		if (IEEE80211_IS_CHAN_5GHZ(ic->ic_bsschan))
4351 			mwl_hal_setcsmode(mh, CSMODE_AGGRESSIVE);
4352 		else
4353 			mwl_hal_setcsmode(mh, CSMODE_AUTO_ENA);
4354 		/*
4355 		 * Start timer to prod firmware.
4356 		 */
4357 		if (sc->sc_ageinterval != 0)
4358 			callout_reset(&sc->sc_timer, sc->sc_ageinterval*hz,
4359 			    mwl_agestations, sc);
4360 	} else if (nstate == IEEE80211_S_SLEEP) {
4361 		/* XXX set chip in power save */
4362 	} else if ((vap->iv_flags & IEEE80211_F_DWDS) &&
4363 	    --sc->sc_ndwdsvaps == 0)
4364 		mwl_hal_setdwds(mh, 0);
4365 bad:
4366 	return error;
4367 }
4368 
4369 /*
4370  * Manage station id's; these are separate from AID's
4371  * as AID's may have values out of the range of possible
4372  * station id's acceptable to the firmware.
4373  */
4374 static int
4375 allocstaid(struct mwl_softc *sc, int aid)
4376 {
4377 	int staid;
4378 
4379 	if (!(0 < aid && aid < MWL_MAXSTAID) || isset(sc->sc_staid, aid)) {
4380 		/* NB: don't use 0 */
4381 		for (staid = 1; staid < MWL_MAXSTAID; staid++)
4382 			if (isclr(sc->sc_staid, staid))
4383 				break;
4384 	} else
4385 		staid = aid;
4386 	setbit(sc->sc_staid, staid);
4387 	return staid;
4388 }
4389 
4390 static void
4391 delstaid(struct mwl_softc *sc, int staid)
4392 {
4393 	clrbit(sc->sc_staid, staid);
4394 }
4395 
4396 /*
4397  * Setup driver-specific state for a newly associated node.
4398  * Note that we're called also on a re-associate, the isnew
4399  * param tells us if this is the first time or not.
4400  */
4401 static void
4402 mwl_newassoc(struct ieee80211_node *ni, int isnew)
4403 {
4404 	struct ieee80211vap *vap = ni->ni_vap;
4405         struct mwl_softc *sc = vap->iv_ic->ic_ifp->if_softc;
4406 	struct mwl_node *mn = MWL_NODE(ni);
4407 	MWL_HAL_PEERINFO pi;
4408 	uint16_t aid;
4409 	int error;
4410 
4411 	aid = IEEE80211_AID(ni->ni_associd);
4412 	if (isnew) {
4413 		mn->mn_staid = allocstaid(sc, aid);
4414 		mn->mn_hvap = MWL_VAP(vap)->mv_hvap;
4415 	} else {
4416 		mn = MWL_NODE(ni);
4417 		/* XXX reset BA stream? */
4418 	}
4419 	DPRINTF(sc, MWL_DEBUG_NODE, "%s: mac %s isnew %d aid %d staid %d\n",
4420 	    __func__, ether_sprintf(ni->ni_macaddr), isnew, aid, mn->mn_staid);
4421 	error = mwl_peerstadb(ni, aid, mn->mn_staid, mkpeerinfo(&pi, ni));
4422 	if (error != 0) {
4423 		DPRINTF(sc, MWL_DEBUG_NODE,
4424 		    "%s: error %d creating sta db entry\n",
4425 		    __func__, error);
4426 		/* XXX how to deal with error? */
4427 	}
4428 }
4429 
4430 /*
4431  * Periodically poke the firmware to age out station state
4432  * (power save queues, pending tx aggregates).
4433  */
4434 static void
4435 mwl_agestations(void *arg)
4436 {
4437 	struct mwl_softc *sc = arg;
4438 
4439 	mwl_hal_setkeepalive(sc->sc_mh);
4440 	if (sc->sc_ageinterval != 0)		/* NB: catch dynamic changes */
4441 		callout_schedule(&sc->sc_timer, sc->sc_ageinterval*hz);
4442 }
4443 
4444 static const struct mwl_hal_channel *
4445 findhalchannel(const MWL_HAL_CHANNELINFO *ci, int ieee)
4446 {
4447 	int i;
4448 
4449 	for (i = 0; i < ci->nchannels; i++) {
4450 		const struct mwl_hal_channel *hc = &ci->channels[i];
4451 		if (hc->ieee == ieee)
4452 			return hc;
4453 	}
4454 	return NULL;
4455 }
4456 
4457 static int
4458 mwl_setregdomain(struct ieee80211com *ic, struct ieee80211_regdomain *rd,
4459 	int nchan, struct ieee80211_channel chans[])
4460 {
4461 	struct mwl_softc *sc = ic->ic_ifp->if_softc;
4462 	struct mwl_hal *mh = sc->sc_mh;
4463 	const MWL_HAL_CHANNELINFO *ci;
4464 	int i;
4465 
4466 	for (i = 0; i < nchan; i++) {
4467 		struct ieee80211_channel *c = &chans[i];
4468 		const struct mwl_hal_channel *hc;
4469 
4470 		if (IEEE80211_IS_CHAN_2GHZ(c)) {
4471 			mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_2DOT4GHZ,
4472 			    IEEE80211_IS_CHAN_HT40(c) ?
4473 				MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci);
4474 		} else if (IEEE80211_IS_CHAN_5GHZ(c)) {
4475 			mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_5GHZ,
4476 			    IEEE80211_IS_CHAN_HT40(c) ?
4477 				MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci);
4478 		} else {
4479 			if_printf(ic->ic_ifp,
4480 			    "%s: channel %u freq %u/0x%x not 2.4/5GHz\n",
4481 			    __func__, c->ic_ieee, c->ic_freq, c->ic_flags);
4482 			return EINVAL;
4483 		}
4484 		/*
4485 		 * Verify channel has cal data and cap tx power.
4486 		 */
4487 		hc = findhalchannel(ci, c->ic_ieee);
4488 		if (hc != NULL) {
4489 			if (c->ic_maxpower > 2*hc->maxTxPow)
4490 				c->ic_maxpower = 2*hc->maxTxPow;
4491 			goto next;
4492 		}
4493 		if (IEEE80211_IS_CHAN_HT40(c)) {
4494 			/*
4495 			 * Look for the extension channel since the
4496 			 * hal table only has the primary channel.
4497 			 */
4498 			hc = findhalchannel(ci, c->ic_extieee);
4499 			if (hc != NULL) {
4500 				if (c->ic_maxpower > 2*hc->maxTxPow)
4501 					c->ic_maxpower = 2*hc->maxTxPow;
4502 				goto next;
4503 			}
4504 		}
4505 		if_printf(ic->ic_ifp,
4506 		    "%s: no cal data for channel %u ext %u freq %u/0x%x\n",
4507 		    __func__, c->ic_ieee, c->ic_extieee,
4508 		    c->ic_freq, c->ic_flags);
4509 		return EINVAL;
4510 	next:
4511 		;
4512 	}
4513 	return 0;
4514 }
4515 
4516 #define	IEEE80211_CHAN_HTG	(IEEE80211_CHAN_HT|IEEE80211_CHAN_G)
4517 #define	IEEE80211_CHAN_HTA	(IEEE80211_CHAN_HT|IEEE80211_CHAN_A)
4518 
4519 static void
4520 addchan(struct ieee80211_channel *c, int freq, int flags, int ieee, int txpow)
4521 {
4522 	c->ic_freq = freq;
4523 	c->ic_flags = flags;
4524 	c->ic_ieee = ieee;
4525 	c->ic_minpower = 0;
4526 	c->ic_maxpower = 2*txpow;
4527 	c->ic_maxregpower = txpow;
4528 }
4529 
4530 static const struct ieee80211_channel *
4531 findchannel(const struct ieee80211_channel chans[], int nchans,
4532 	int freq, int flags)
4533 {
4534 	const struct ieee80211_channel *c;
4535 	int i;
4536 
4537 	for (i = 0; i < nchans; i++) {
4538 		c = &chans[i];
4539 		if (c->ic_freq == freq && c->ic_flags == flags)
4540 			return c;
4541 	}
4542 	return NULL;
4543 }
4544 
4545 static void
4546 addht40channels(struct ieee80211_channel chans[], int maxchans, int *nchans,
4547 	const MWL_HAL_CHANNELINFO *ci, int flags)
4548 {
4549 	struct ieee80211_channel *c;
4550 	const struct ieee80211_channel *extc;
4551 	const struct mwl_hal_channel *hc;
4552 	int i;
4553 
4554 	c = &chans[*nchans];
4555 
4556 	flags &= ~IEEE80211_CHAN_HT;
4557 	for (i = 0; i < ci->nchannels; i++) {
4558 		/*
4559 		 * Each entry defines an HT40 channel pair; find the
4560 		 * extension channel above and the insert the pair.
4561 		 */
4562 		hc = &ci->channels[i];
4563 		extc = findchannel(chans, *nchans, hc->freq+20,
4564 		    flags | IEEE80211_CHAN_HT20);
4565 		if (extc != NULL) {
4566 			if (*nchans >= maxchans)
4567 				break;
4568 			addchan(c, hc->freq, flags | IEEE80211_CHAN_HT40U,
4569 			    hc->ieee, hc->maxTxPow);
4570 			c->ic_extieee = extc->ic_ieee;
4571 			c++, (*nchans)++;
4572 			if (*nchans >= maxchans)
4573 				break;
4574 			addchan(c, extc->ic_freq, flags | IEEE80211_CHAN_HT40D,
4575 			    extc->ic_ieee, hc->maxTxPow);
4576 			c->ic_extieee = hc->ieee;
4577 			c++, (*nchans)++;
4578 		}
4579 	}
4580 }
4581 
4582 static void
4583 addchannels(struct ieee80211_channel chans[], int maxchans, int *nchans,
4584 	const MWL_HAL_CHANNELINFO *ci, int flags)
4585 {
4586 	struct ieee80211_channel *c;
4587 	int i;
4588 
4589 	c = &chans[*nchans];
4590 
4591 	for (i = 0; i < ci->nchannels; i++) {
4592 		const struct mwl_hal_channel *hc;
4593 
4594 		hc = &ci->channels[i];
4595 		if (*nchans >= maxchans)
4596 			break;
4597 		addchan(c, hc->freq, flags, hc->ieee, hc->maxTxPow);
4598 		c++, (*nchans)++;
4599 		if (flags == IEEE80211_CHAN_G || flags == IEEE80211_CHAN_HTG) {
4600 			/* g channel have a separate b-only entry */
4601 			if (*nchans >= maxchans)
4602 				break;
4603 			c[0] = c[-1];
4604 			c[-1].ic_flags = IEEE80211_CHAN_B;
4605 			c++, (*nchans)++;
4606 		}
4607 		if (flags == IEEE80211_CHAN_HTG) {
4608 			/* HT g channel have a separate g-only entry */
4609 			if (*nchans >= maxchans)
4610 				break;
4611 			c[-1].ic_flags = IEEE80211_CHAN_G;
4612 			c[0] = c[-1];
4613 			c[0].ic_flags &= ~IEEE80211_CHAN_HT;
4614 			c[0].ic_flags |= IEEE80211_CHAN_HT20;	/* HT20 */
4615 			c++, (*nchans)++;
4616 		}
4617 		if (flags == IEEE80211_CHAN_HTA) {
4618 			/* HT a channel have a separate a-only entry */
4619 			if (*nchans >= maxchans)
4620 				break;
4621 			c[-1].ic_flags = IEEE80211_CHAN_A;
4622 			c[0] = c[-1];
4623 			c[0].ic_flags &= ~IEEE80211_CHAN_HT;
4624 			c[0].ic_flags |= IEEE80211_CHAN_HT20;	/* HT20 */
4625 			c++, (*nchans)++;
4626 		}
4627 	}
4628 }
4629 
4630 static void
4631 getchannels(struct mwl_softc *sc, int maxchans, int *nchans,
4632 	struct ieee80211_channel chans[])
4633 {
4634 	const MWL_HAL_CHANNELINFO *ci;
4635 
4636 	/*
4637 	 * Use the channel info from the hal to craft the
4638 	 * channel list.  Note that we pass back an unsorted
4639 	 * list; the caller is required to sort it for us
4640 	 * (if desired).
4641 	 */
4642 	*nchans = 0;
4643 	if (mwl_hal_getchannelinfo(sc->sc_mh,
4644 	    MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0)
4645 		addchannels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTG);
4646 	if (mwl_hal_getchannelinfo(sc->sc_mh,
4647 	    MWL_FREQ_BAND_5GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0)
4648 		addchannels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTA);
4649 	if (mwl_hal_getchannelinfo(sc->sc_mh,
4650 	    MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0)
4651 		addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTG);
4652 	if (mwl_hal_getchannelinfo(sc->sc_mh,
4653 	    MWL_FREQ_BAND_5GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0)
4654 		addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTA);
4655 }
4656 
4657 static void
4658 mwl_getradiocaps(struct ieee80211com *ic,
4659 	int maxchans, int *nchans, struct ieee80211_channel chans[])
4660 {
4661 	struct mwl_softc *sc = ic->ic_ifp->if_softc;
4662 
4663 	getchannels(sc, maxchans, nchans, chans);
4664 }
4665 
4666 static int
4667 mwl_getchannels(struct mwl_softc *sc)
4668 {
4669 	struct ifnet *ifp = sc->sc_ifp;
4670 	struct ieee80211com *ic = ifp->if_l2com;
4671 
4672 	/*
4673 	 * Use the channel info from the hal to craft the
4674 	 * channel list for net80211.  Note that we pass up
4675 	 * an unsorted list; net80211 will sort it for us.
4676 	 */
4677 	memset(ic->ic_channels, 0, sizeof(ic->ic_channels));
4678 	ic->ic_nchans = 0;
4679 	getchannels(sc, IEEE80211_CHAN_MAX, &ic->ic_nchans, ic->ic_channels);
4680 
4681 	ic->ic_regdomain.regdomain = SKU_DEBUG;
4682 	ic->ic_regdomain.country = CTRY_DEFAULT;
4683 	ic->ic_regdomain.location = 'I';
4684 	ic->ic_regdomain.isocc[0] = ' ';	/* XXX? */
4685 	ic->ic_regdomain.isocc[1] = ' ';
4686 	return (ic->ic_nchans == 0 ? EIO : 0);
4687 }
4688 #undef IEEE80211_CHAN_HTA
4689 #undef IEEE80211_CHAN_HTG
4690 
4691 #ifdef MWL_DEBUG
4692 static void
4693 mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix)
4694 {
4695 	const struct mwl_rxdesc *ds = bf->bf_desc;
4696 	uint32_t status = le32toh(ds->Status);
4697 
4698 	printf("R[%2u] (DS.V:%p DS.P:%p) NEXT:%08x DATA:%08x RC:%02x%s\n"
4699 	       "      STAT:%02x LEN:%04x RSSI:%02x CHAN:%02x RATE:%02x QOS:%04x HT:%04x\n",
4700 	    ix, ds, (const struct mwl_desc *)bf->bf_daddr,
4701 	    le32toh(ds->pPhysNext), le32toh(ds->pPhysBuffData),
4702 	    ds->RxControl,
4703 	    ds->RxControl != EAGLE_RXD_CTRL_DRIVER_OWN ?
4704 	        "" : (status & EAGLE_RXD_STATUS_OK) ? " *" : " !",
4705 	    ds->Status, le16toh(ds->PktLen), ds->RSSI, ds->Channel,
4706 	    ds->Rate, le16toh(ds->QosCtrl), le16toh(ds->HtSig2));
4707 }
4708 
4709 static void
4710 mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix)
4711 {
4712 	const struct mwl_txdesc *ds = bf->bf_desc;
4713 	uint32_t status = le32toh(ds->Status);
4714 
4715 	printf("Q%u[%3u]", qnum, ix);
4716 	printf(" (DS.V:%p DS.P:%p)\n",
4717 	    ds, (const struct mwl_txdesc *)bf->bf_daddr);
4718 	printf("    NEXT:%08x DATA:%08x LEN:%04x STAT:%08x%s\n",
4719 	    le32toh(ds->pPhysNext),
4720 	    le32toh(ds->PktPtr), le16toh(ds->PktLen), status,
4721 	    status & EAGLE_TXD_STATUS_USED ?
4722 		"" : (status & 3) != 0 ? " *" : " !");
4723 	printf("    RATE:%02x PRI:%x QOS:%04x SAP:%08x FORMAT:%04x\n",
4724 	    ds->DataRate, ds->TxPriority, le16toh(ds->QosCtrl),
4725 	    le32toh(ds->SapPktInfo), le16toh(ds->Format));
4726 #if MWL_TXDESC > 1
4727 	printf("    MULTIFRAMES:%u LEN:%04x %04x %04x %04x %04x %04x\n"
4728 	    , le32toh(ds->multiframes)
4729 	    , le16toh(ds->PktLenArray[0]), le16toh(ds->PktLenArray[1])
4730 	    , le16toh(ds->PktLenArray[2]), le16toh(ds->PktLenArray[3])
4731 	    , le16toh(ds->PktLenArray[4]), le16toh(ds->PktLenArray[5])
4732 	);
4733 	printf("    DATA:%08x %08x %08x %08x %08x %08x\n"
4734 	    , le32toh(ds->PktPtrArray[0]), le32toh(ds->PktPtrArray[1])
4735 	    , le32toh(ds->PktPtrArray[2]), le32toh(ds->PktPtrArray[3])
4736 	    , le32toh(ds->PktPtrArray[4]), le32toh(ds->PktPtrArray[5])
4737 	);
4738 #endif
4739 #if 0
4740 { const uint8_t *cp = (const uint8_t *) ds;
4741   int i;
4742   for (i = 0; i < sizeof(struct mwl_txdesc); i++) {
4743 	printf("%02x ", cp[i]);
4744 	if (((i+1) % 16) == 0)
4745 		printf("\n");
4746   }
4747   printf("\n");
4748 }
4749 #endif
4750 }
4751 #endif /* MWL_DEBUG */
4752 
4753 #if 0
4754 static void
4755 mwl_txq_dump(struct mwl_txq *txq)
4756 {
4757 	struct mwl_txbuf *bf;
4758 	int i = 0;
4759 
4760 	MWL_TXQ_LOCK(txq);
4761 	STAILQ_FOREACH(bf, &txq->active, bf_list) {
4762 		struct mwl_txdesc *ds = bf->bf_desc;
4763 		MWL_TXDESC_SYNC(txq, ds,
4764 		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
4765 #ifdef MWL_DEBUG
4766 		mwl_printtxbuf(bf, txq->qnum, i);
4767 #endif
4768 		i++;
4769 	}
4770 	MWL_TXQ_UNLOCK(txq);
4771 }
4772 #endif
4773 
4774 static void
4775 mwl_watchdog(void *arg)
4776 {
4777 	struct mwl_softc *sc;
4778 	struct ifnet *ifp;
4779 
4780 	sc = arg;
4781 	callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc);
4782 	if (sc->sc_tx_timer == 0 || --sc->sc_tx_timer > 0)
4783 		return;
4784 
4785 	ifp = sc->sc_ifp;
4786 	if ((ifp->if_drv_flags & IFF_DRV_RUNNING) && !sc->sc_invalid) {
4787 		if (mwl_hal_setkeepalive(sc->sc_mh))
4788 			if_printf(ifp, "transmit timeout (firmware hung?)\n");
4789 		else
4790 			if_printf(ifp, "transmit timeout\n");
4791 #if 0
4792 		mwl_reset(ifp);
4793 mwl_txq_dump(&sc->sc_txq[0]);/*XXX*/
4794 #endif
4795 		ifp->if_oerrors++;
4796 		sc->sc_stats.mst_watchdog++;
4797 	}
4798 }
4799 
4800 #ifdef MWL_DIAGAPI
4801 /*
4802  * Diagnostic interface to the HAL.  This is used by various
4803  * tools to do things like retrieve register contents for
4804  * debugging.  The mechanism is intentionally opaque so that
4805  * it can change frequently w/o concern for compatiblity.
4806  */
4807 static int
4808 mwl_ioctl_diag(struct mwl_softc *sc, struct mwl_diag *md)
4809 {
4810 	struct mwl_hal *mh = sc->sc_mh;
4811 	u_int id = md->md_id & MWL_DIAG_ID;
4812 	void *indata = NULL;
4813 	void *outdata = NULL;
4814 	u_int32_t insize = md->md_in_size;
4815 	u_int32_t outsize = md->md_out_size;
4816 	int error = 0;
4817 
4818 	if (md->md_id & MWL_DIAG_IN) {
4819 		/*
4820 		 * Copy in data.
4821 		 */
4822 		indata = malloc(insize, M_TEMP, M_NOWAIT);
4823 		if (indata == NULL) {
4824 			error = ENOMEM;
4825 			goto bad;
4826 		}
4827 		error = copyin(md->md_in_data, indata, insize);
4828 		if (error)
4829 			goto bad;
4830 	}
4831 	if (md->md_id & MWL_DIAG_DYN) {
4832 		/*
4833 		 * Allocate a buffer for the results (otherwise the HAL
4834 		 * returns a pointer to a buffer where we can read the
4835 		 * results).  Note that we depend on the HAL leaving this
4836 		 * pointer for us to use below in reclaiming the buffer;
4837 		 * may want to be more defensive.
4838 		 */
4839 		outdata = malloc(outsize, M_TEMP, M_NOWAIT);
4840 		if (outdata == NULL) {
4841 			error = ENOMEM;
4842 			goto bad;
4843 		}
4844 	}
4845 	if (mwl_hal_getdiagstate(mh, id, indata, insize, &outdata, &outsize)) {
4846 		if (outsize < md->md_out_size)
4847 			md->md_out_size = outsize;
4848 		if (outdata != NULL)
4849 			error = copyout(outdata, md->md_out_data,
4850 					md->md_out_size);
4851 	} else {
4852 		error = EINVAL;
4853 	}
4854 bad:
4855 	if ((md->md_id & MWL_DIAG_IN) && indata != NULL)
4856 		free(indata, M_TEMP);
4857 	if ((md->md_id & MWL_DIAG_DYN) && outdata != NULL)
4858 		free(outdata, M_TEMP);
4859 	return error;
4860 }
4861 
4862 static int
4863 mwl_ioctl_reset(struct mwl_softc *sc, struct mwl_diag *md)
4864 {
4865 	struct mwl_hal *mh = sc->sc_mh;
4866 	int error;
4867 
4868 	MWL_LOCK_ASSERT(sc);
4869 
4870 	if (md->md_id == 0 && mwl_hal_fwload(mh, NULL) != 0) {
4871 		device_printf(sc->sc_dev, "unable to load firmware\n");
4872 		return EIO;
4873 	}
4874 	if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) {
4875 		device_printf(sc->sc_dev, "unable to fetch h/w specs\n");
4876 		return EIO;
4877 	}
4878 	error = mwl_setupdma(sc);
4879 	if (error != 0) {
4880 		/* NB: mwl_setupdma prints a msg */
4881 		return error;
4882 	}
4883 	/*
4884 	 * Reset tx/rx data structures; after reload we must
4885 	 * re-start the driver's notion of the next xmit/recv.
4886 	 */
4887 	mwl_draintxq(sc);		/* clear pending frames */
4888 	mwl_resettxq(sc);		/* rebuild tx q lists */
4889 	sc->sc_rxnext = NULL;		/* force rx to start at the list head */
4890 	return 0;
4891 }
4892 #endif /* MWL_DIAGAPI */
4893 
4894 static int
4895 mwl_ioctl(struct ifnet *ifp, u_long cmd, caddr_t data)
4896 {
4897 #define	IS_RUNNING(ifp) \
4898 	((ifp->if_flags & IFF_UP) && (ifp->if_drv_flags & IFF_DRV_RUNNING))
4899 	struct mwl_softc *sc = ifp->if_softc;
4900 	struct ieee80211com *ic = ifp->if_l2com;
4901 	struct ifreq *ifr = (struct ifreq *)data;
4902 	int error = 0, startall;
4903 
4904 	switch (cmd) {
4905 	case SIOCSIFFLAGS:
4906 		MWL_LOCK(sc);
4907 		startall = 0;
4908 		if (IS_RUNNING(ifp)) {
4909 			/*
4910 			 * To avoid rescanning another access point,
4911 			 * do not call mwl_init() here.  Instead,
4912 			 * only reflect promisc mode settings.
4913 			 */
4914 			mwl_mode_init(sc);
4915 		} else if (ifp->if_flags & IFF_UP) {
4916 			/*
4917 			 * Beware of being called during attach/detach
4918 			 * to reset promiscuous mode.  In that case we
4919 			 * will still be marked UP but not RUNNING.
4920 			 * However trying to re-init the interface
4921 			 * is the wrong thing to do as we've already
4922 			 * torn down much of our state.  There's
4923 			 * probably a better way to deal with this.
4924 			 */
4925 			if (!sc->sc_invalid) {
4926 				mwl_init_locked(sc);	/* XXX lose error */
4927 				startall = 1;
4928 			}
4929 		} else
4930 			mwl_stop_locked(ifp, 1);
4931 		MWL_UNLOCK(sc);
4932 		if (startall)
4933 			ieee80211_start_all(ic);
4934 		break;
4935 	case SIOCGMVSTATS:
4936 		mwl_hal_gethwstats(sc->sc_mh, &sc->sc_stats.hw_stats);
4937 		/* NB: embed these numbers to get a consistent view */
4938 		sc->sc_stats.mst_tx_packets = ifp->if_opackets;
4939 		sc->sc_stats.mst_rx_packets = ifp->if_ipackets;
4940 		/*
4941 		 * NB: Drop the softc lock in case of a page fault;
4942 		 * we'll accept any potential inconsisentcy in the
4943 		 * statistics.  The alternative is to copy the data
4944 		 * to a local structure.
4945 		 */
4946 		return copyout(&sc->sc_stats,
4947 				ifr->ifr_data, sizeof (sc->sc_stats));
4948 #ifdef MWL_DIAGAPI
4949 	case SIOCGMVDIAG:
4950 		/* XXX check privs */
4951 		return mwl_ioctl_diag(sc, (struct mwl_diag *) ifr);
4952 	case SIOCGMVRESET:
4953 		/* XXX check privs */
4954 		MWL_LOCK(sc);
4955 		error = mwl_ioctl_reset(sc,(struct mwl_diag *) ifr);
4956 		MWL_UNLOCK(sc);
4957 		break;
4958 #endif /* MWL_DIAGAPI */
4959 	case SIOCGIFMEDIA:
4960 		error = ifmedia_ioctl(ifp, ifr, &ic->ic_media, cmd);
4961 		break;
4962 	case SIOCGIFADDR:
4963 		error = ether_ioctl(ifp, cmd, data);
4964 		break;
4965 	default:
4966 		error = EINVAL;
4967 		break;
4968 	}
4969 	return error;
4970 #undef IS_RUNNING
4971 }
4972 
4973 #ifdef	MWL_DEBUG
4974 static int
4975 mwl_sysctl_debug(SYSCTL_HANDLER_ARGS)
4976 {
4977 	struct mwl_softc *sc = arg1;
4978 	int debug, error;
4979 
4980 	debug = sc->sc_debug | (mwl_hal_getdebug(sc->sc_mh) << 24);
4981 	error = sysctl_handle_int(oidp, &debug, 0, req);
4982 	if (error || !req->newptr)
4983 		return error;
4984 	mwl_hal_setdebug(sc->sc_mh, debug >> 24);
4985 	sc->sc_debug = debug & 0x00ffffff;
4986 	return 0;
4987 }
4988 #endif /* MWL_DEBUG */
4989 
4990 static void
4991 mwl_sysctlattach(struct mwl_softc *sc)
4992 {
4993 #ifdef	MWL_DEBUG
4994 	struct sysctl_ctx_list *ctx = device_get_sysctl_ctx(sc->sc_dev);
4995 	struct sysctl_oid *tree = device_get_sysctl_tree(sc->sc_dev);
4996 
4997 	sc->sc_debug = mwl_debug;
4998 	SYSCTL_ADD_PROC(ctx, SYSCTL_CHILDREN(tree), OID_AUTO,
4999 		"debug", CTLTYPE_INT | CTLFLAG_RW, sc, 0,
5000 		mwl_sysctl_debug, "I", "control debugging printfs");
5001 #endif
5002 }
5003 
5004 /*
5005  * Announce various information on device/driver attach.
5006  */
5007 static void
5008 mwl_announce(struct mwl_softc *sc)
5009 {
5010 	struct ifnet *ifp = sc->sc_ifp;
5011 
5012 	if_printf(ifp, "Rev A%d hardware, v%d.%d.%d.%d firmware (regioncode %d)\n",
5013 		sc->sc_hwspecs.hwVersion,
5014 		(sc->sc_hwspecs.fwReleaseNumber>>24) & 0xff,
5015 		(sc->sc_hwspecs.fwReleaseNumber>>16) & 0xff,
5016 		(sc->sc_hwspecs.fwReleaseNumber>>8) & 0xff,
5017 		(sc->sc_hwspecs.fwReleaseNumber>>0) & 0xff,
5018 		sc->sc_hwspecs.regionCode);
5019 	sc->sc_fwrelease = sc->sc_hwspecs.fwReleaseNumber;
5020 
5021 	if (bootverbose) {
5022 		int i;
5023 		for (i = 0; i <= WME_AC_VO; i++) {
5024 			struct mwl_txq *txq = sc->sc_ac2q[i];
5025 			if_printf(ifp, "Use hw queue %u for %s traffic\n",
5026 				txq->qnum, ieee80211_wme_acnames[i]);
5027 		}
5028 	}
5029 	if (bootverbose || mwl_rxdesc != MWL_RXDESC)
5030 		if_printf(ifp, "using %u rx descriptors\n", mwl_rxdesc);
5031 	if (bootverbose || mwl_rxbuf != MWL_RXBUF)
5032 		if_printf(ifp, "using %u rx buffers\n", mwl_rxbuf);
5033 	if (bootverbose || mwl_txbuf != MWL_TXBUF)
5034 		if_printf(ifp, "using %u tx buffers\n", mwl_txbuf);
5035 	if (bootverbose && mwl_hal_ismbsscapable(sc->sc_mh))
5036 		if_printf(ifp, "multi-bss support\n");
5037 #ifdef MWL_TX_NODROP
5038 	if (bootverbose)
5039 		if_printf(ifp, "no tx drop\n");
5040 #endif
5041 }
5042