xref: /freebsd/sys/dev/cxgb/cxgb_adapter.h (revision 6829dae12bb055451fa467da4589c43bd03b1e64)
1 /**************************************************************************
2 SPDX-License-Identifier: BSD-2-Clause-FreeBSD
3 
4 Copyright (c) 2007-2009, Chelsio Inc.
5 All rights reserved.
6 
7 Redistribution and use in source and binary forms, with or without
8 modification, are permitted provided that the following conditions are met:
9 
10  1. Redistributions of source code must retain the above copyright notice,
11     this list of conditions and the following disclaimer.
12 
13  2. Neither the name of the Chelsio Corporation nor the names of its
14     contributors may be used to endorse or promote products derived from
15     this software without specific prior written permission.
16 
17 THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
18 AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
19 IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
20 ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
21 LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
22 CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
23 SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
24 INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
25 CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
26 ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
27 POSSIBILITY OF SUCH DAMAGE.
28 
29 $FreeBSD$
30 
31 ***************************************************************************/
32 
33 
34 #ifndef _CXGB_ADAPTER_H_
35 #define _CXGB_ADAPTER_H_
36 
37 #include <sys/lock.h>
38 #include <sys/mutex.h>
39 #include <sys/rman.h>
40 #include <sys/mbuf.h>
41 #include <sys/socket.h>
42 #include <sys/sockio.h>
43 #include <sys/condvar.h>
44 #include <sys/buf_ring.h>
45 #include <sys/taskqueue.h>
46 
47 #include <net/ethernet.h>
48 #include <net/if.h>
49 #include <net/if_var.h>
50 #include <net/if_media.h>
51 #include <net/if_dl.h>
52 #include <netinet/in.h>
53 #include <netinet/tcp_lro.h>
54 
55 #include <machine/bus.h>
56 #include <machine/resource.h>
57 
58 #include <dev/pci/pcireg.h>
59 #include <dev/pci/pcivar.h>
60 
61 #include <cxgb_osdep.h>
62 
63 struct adapter;
64 struct sge_qset;
65 extern int cxgb_debug;
66 
67 #ifdef DEBUG_LOCKING
68 #define MTX_INIT(lock, lockname, class, flags) \
69 	do { \
70 		printf("initializing %s at %s:%d\n", lockname, __FILE__, __LINE__); \
71 		mtx_init((lock), lockname, class, flags);		\
72 	} while (0)
73 
74 #define MTX_DESTROY(lock) \
75 	do { \
76 		printf("destroying %s at %s:%d\n", (lock)->lock_object.lo_name, __FILE__, __LINE__); \
77 		mtx_destroy((lock));					\
78 	} while (0)
79 
80 #else
81 #define MTX_INIT mtx_init
82 #define MTX_DESTROY mtx_destroy
83 #endif
84 
85 enum {
86 	LF_NO = 0,
87 	LF_MAYBE,
88 	LF_YES
89 };
90 
91 struct port_info {
92 	struct adapter	*adapter;
93 	struct ifnet	*ifp;
94 	int		if_flags;
95 	int		flags;
96 	const struct port_type_info *port_type;
97 	struct cphy	phy;
98 	struct cmac	mac;
99 	struct timeval	last_refreshed;
100 	struct link_config link_config;
101 	struct ifmedia	media;
102 	struct mtx	lock;
103 	uint32_t	port_id;
104 	uint32_t	tx_chan;
105 	uint32_t	txpkt_intf;
106 	uint32_t        first_qset;
107 	uint32_t	nqsets;
108 	int		link_fault;
109 
110 	uint8_t		hw_addr[ETHER_ADDR_LEN];
111 	struct callout	link_check_ch;
112 	struct task	link_check_task;
113 	struct task	timer_reclaim_task;
114 	struct cdev     *port_cdev;
115 
116 #define PORT_LOCK_NAME_LEN 32
117 #define PORT_NAME_LEN 32
118 	char            lockbuf[PORT_LOCK_NAME_LEN];
119 	char            namebuf[PORT_NAME_LEN];
120 } __aligned(L1_CACHE_BYTES);
121 
122 enum {
123 	/* adapter flags */
124 	FULL_INIT_DONE	= (1 << 0),
125 	USING_MSI	= (1 << 1),
126 	USING_MSIX	= (1 << 2),
127 	QUEUES_BOUND	= (1 << 3),
128 	FW_UPTODATE	= (1 << 4),
129 	TPS_UPTODATE	= (1 << 5),
130 	CXGB_SHUTDOWN	= (1 << 6),
131 	CXGB_OFLD_INIT	= (1 << 7),
132 	TP_PARITY_INIT	= (1 << 8),
133 	CXGB_BUSY	= (1 << 9),
134 	TOM_INIT_DONE	= (1 << 10),
135 
136 	/* port flags */
137 	DOOMED		= (1 << 0),
138 };
139 #define IS_DOOMED(p)	(p->flags & DOOMED)
140 #define SET_DOOMED(p)	do {p->flags |= DOOMED;} while (0)
141 #define IS_BUSY(sc)	(sc->flags & CXGB_BUSY)
142 #define SET_BUSY(sc)	do {sc->flags |= CXGB_BUSY;} while (0)
143 #define CLR_BUSY(sc)	do {sc->flags &= ~CXGB_BUSY;} while (0)
144 
145 #define FL_Q_SIZE	4096
146 #define JUMBO_Q_SIZE	1024
147 #define RSPQ_Q_SIZE	2048
148 #define TX_ETH_Q_SIZE	1024
149 #define TX_OFLD_Q_SIZE	1024
150 #define TX_CTRL_Q_SIZE	256
151 
152 enum { TXQ_ETH = 0,
153        TXQ_OFLD = 1,
154        TXQ_CTRL = 2, };
155 
156 
157 /*
158  * work request size in bytes
159  */
160 #define WR_LEN (WR_FLITS * 8)
161 #define PIO_LEN (WR_LEN - sizeof(struct cpl_tx_pkt_lso))
162 
163 struct lro_state {
164 	unsigned short enabled;
165 	struct lro_ctrl ctrl;
166 };
167 
168 #define RX_BUNDLE_SIZE 8
169 
170 struct rsp_desc;
171 
172 struct sge_rspq {
173 	uint32_t	credits;
174 	uint32_t	size;
175 	uint32_t	cidx;
176 	uint32_t	gen;
177 	uint32_t	polling;
178 	uint32_t	holdoff_tmr;
179 	uint32_t	next_holdoff;
180 	uint32_t        imm_data;
181 	uint32_t        async_notif;
182 	uint32_t	cntxt_id;
183 	uint32_t        offload_pkts;
184 	uint32_t        pure_rsps;
185 	uint32_t        unhandled_irqs;
186 	uint32_t        starved;
187 
188 	bus_addr_t	phys_addr;
189 	bus_dma_tag_t	desc_tag;
190 	bus_dmamap_t	desc_map;
191 
192 	struct t3_mbuf_hdr rspq_mh;
193 	struct rsp_desc	*desc;
194 	struct mtx      lock;
195 #define RSPQ_NAME_LEN  32
196 	char            lockbuf[RSPQ_NAME_LEN];
197 	uint32_t	rspq_dump_start;
198 	uint32_t	rspq_dump_count;
199 };
200 
201 struct rx_desc;
202 struct rx_sw_desc;
203 
204 struct sge_fl {
205 	uint32_t	buf_size;
206 	uint32_t	credits;
207 	uint32_t	size;
208 	uint32_t	cidx;
209 	uint32_t	pidx;
210 	uint32_t	gen;
211 	uint32_t	db_pending;
212 	bus_addr_t	phys_addr;
213 	uint32_t	cntxt_id;
214 	uint32_t	empty;
215 	bus_dma_tag_t	desc_tag;
216 	bus_dmamap_t	desc_map;
217 	bus_dma_tag_t   entry_tag;
218 	uma_zone_t      zone;
219 	struct rx_desc	*desc;
220 	struct rx_sw_desc *sdesc;
221 	int             type;
222 };
223 
224 struct tx_desc;
225 struct tx_sw_desc;
226 
227 #define TXQ_TRANSMITTING    0x1
228 
229 struct sge_txq {
230 	uint64_t	flags;
231 	uint32_t	in_use;
232 	uint32_t	size;
233 	uint32_t	processed;
234 	uint32_t	cleaned;
235 	uint32_t	stop_thres;
236 	uint32_t	cidx;
237 	uint32_t	pidx;
238 	uint32_t	gen;
239 	uint32_t	unacked;
240 	uint32_t	db_pending;
241 	struct tx_desc	*desc;
242 	struct tx_sw_desc *sdesc;
243 	uint32_t	token;
244 	bus_addr_t	phys_addr;
245 	struct task     qresume_task;
246 	struct task     qreclaim_task;
247 	uint32_t	cntxt_id;
248 	uint64_t	stops;
249 	uint64_t	restarts;
250 	bus_dma_tag_t	desc_tag;
251 	bus_dmamap_t	desc_map;
252 	bus_dma_tag_t   entry_tag;
253 	struct mbufq	sendq;
254 
255 	struct buf_ring *txq_mr;
256 	struct ifaltq	*txq_ifq;
257 	struct callout	txq_timer;
258 	struct callout	txq_watchdog;
259 	uint64_t        txq_coalesced;
260 	uint32_t        txq_skipped;
261 	uint32_t        txq_enqueued;
262 	uint32_t	txq_dump_start;
263 	uint32_t	txq_dump_count;
264 	uint64_t	txq_direct_packets;
265 	uint64_t	txq_direct_bytes;
266 	uint64_t	txq_frees;
267 	struct sg_ent  txq_sgl[TX_MAX_SEGS / 2 + 1];
268 };
269 
270 #define SGE_PSTAT_MAX (SGE_PSTAT_VLANINS+1)
271 
272 #define QS_EXITING              0x1
273 #define QS_RUNNING              0x2
274 #define QS_BOUND                0x4
275 #define	QS_FLUSHING		0x8
276 #define	QS_TIMEOUT		0x10
277 
278 struct sge_qset {
279 	struct sge_rspq		rspq;
280 	struct sge_fl		fl[SGE_RXQ_PER_SET];
281 	struct lro_state        lro;
282 	struct sge_txq		txq[SGE_TXQ_PER_SET];
283 	uint32_t                txq_stopped;       /* which Tx queues are stopped */
284 	struct port_info        *port;
285 	struct adapter          *adap;
286 	int                     idx; /* qset # */
287 	int                     qs_flags;
288 	int			coalescing;
289 	struct cv		qs_cv;
290 	struct mtx		lock;
291 #define QS_NAME_LEN 32
292 	char                    namebuf[QS_NAME_LEN];
293 };
294 
295 struct sge {
296 	struct sge_qset	        qs[SGE_QSETS];
297 	struct mtx              reg_lock;
298 };
299 
300 struct filter_info;
301 
302 typedef int (*cpl_handler_t)(struct sge_qset *, struct rsp_desc *,
303     struct mbuf *);
304 
305 struct adapter {
306 	SLIST_ENTRY(adapter)	link;
307 	device_t		dev;
308 	int			flags;
309 
310 	/* PCI register resources */
311 	int			regs_rid;
312 	struct resource		*regs_res;
313 	int			udbs_rid;
314 	struct resource		*udbs_res;
315 	bus_space_handle_t	bh;
316 	bus_space_tag_t		bt;
317 	bus_size_t              mmio_len;
318 	uint32_t                link_width;
319 
320 	/* DMA resources */
321 	bus_dma_tag_t		parent_dmat;
322 	bus_dma_tag_t		rx_dmat;
323 	bus_dma_tag_t		rx_jumbo_dmat;
324 	bus_dma_tag_t		tx_dmat;
325 
326 	/* Interrupt resources */
327 	struct resource		*irq_res;
328 	int			irq_rid;
329 	void			*intr_tag;
330 
331 	uint32_t		msix_regs_rid;
332 	struct resource		*msix_regs_res;
333 
334 	struct resource		*msix_irq_res[SGE_QSETS];
335 	int			msix_irq_rid[SGE_QSETS];
336 	void			*msix_intr_tag[SGE_QSETS];
337 	uint8_t                 rxpkt_map[8]; /* maps RX_PKT interface values to port ids */
338 	uint8_t                 rrss_map[SGE_QSETS]; /* revers RSS map table */
339 	uint16_t                rspq_map[RSS_TABLE_SIZE];     /* maps 7-bit cookie to qidx */
340 	union {
341 		uint8_t                 fill[SGE_QSETS];
342 		uint64_t                coalesce;
343 	} u;
344 
345 #define tunq_fill u.fill
346 #define tunq_coalesce u.coalesce
347 
348 	struct filter_info      *filters;
349 
350 	/* Tasks */
351 	struct task		slow_intr_task;
352 	struct task		tick_task;
353 	struct taskqueue	*tq;
354 	struct callout		cxgb_tick_ch;
355 	struct callout		sge_timer_ch;
356 
357 	/* Register lock for use by the hardware layer */
358 	struct mtx		mdio_lock;
359 	struct mtx		elmer_lock;
360 
361 	/* Bookkeeping for the hardware layer */
362 	struct adapter_params  params;
363 	unsigned int slow_intr_mask;
364 	unsigned long irq_stats[IRQ_NUM_STATS];
365 
366 	struct sge              sge;
367 	struct mc7              pmrx;
368 	struct mc7              pmtx;
369 	struct mc7              cm;
370 	struct mc5              mc5;
371 
372 	struct port_info	port[MAX_NPORTS];
373 	device_t		portdev[MAX_NPORTS];
374 #ifdef TCP_OFFLOAD
375 	void 			*tom_softc;
376 	void 			*iwarp_softc;
377 #endif
378 	char                    fw_version[64];
379 	char                    port_types[MAX_NPORTS + 1];
380 	uint32_t                open_device_map;
381 #ifdef TCP_OFFLOAD
382 	int			offload_map;
383 #endif
384 	struct mtx              lock;
385 	driver_intr_t           *cxgb_intr;
386 	int                     msi_count;
387 
388 #define ADAPTER_LOCK_NAME_LEN	32
389 	char                    lockbuf[ADAPTER_LOCK_NAME_LEN];
390 	char                    reglockbuf[ADAPTER_LOCK_NAME_LEN];
391 	char                    mdiolockbuf[ADAPTER_LOCK_NAME_LEN];
392 	char                    elmerlockbuf[ADAPTER_LOCK_NAME_LEN];
393 
394 	int			timestamp;
395 
396 #ifdef TCP_OFFLOAD
397 #define NUM_CPL_HANDLERS	0xa7
398 	cpl_handler_t cpl_handler[NUM_CPL_HANDLERS] __aligned(CACHE_LINE_SIZE);
399 #endif
400 };
401 
402 struct t3_rx_mode {
403 
404 	uint32_t                idx;
405 	struct port_info        *port;
406 };
407 
408 #define MDIO_LOCK(adapter)	mtx_lock(&(adapter)->mdio_lock)
409 #define MDIO_UNLOCK(adapter)	mtx_unlock(&(adapter)->mdio_lock)
410 #define ELMR_LOCK(adapter)	mtx_lock(&(adapter)->elmer_lock)
411 #define ELMR_UNLOCK(adapter)	mtx_unlock(&(adapter)->elmer_lock)
412 
413 
414 #define PORT_LOCK(port)		     mtx_lock(&(port)->lock);
415 #define PORT_UNLOCK(port)	     mtx_unlock(&(port)->lock);
416 #define PORT_LOCK_INIT(port, name)   mtx_init(&(port)->lock, name, 0, MTX_DEF)
417 #define PORT_LOCK_DEINIT(port)       mtx_destroy(&(port)->lock)
418 #define PORT_LOCK_ASSERT_NOTOWNED(port) mtx_assert(&(port)->lock, MA_NOTOWNED)
419 #define PORT_LOCK_ASSERT_OWNED(port) mtx_assert(&(port)->lock, MA_OWNED)
420 
421 #define ADAPTER_LOCK(adap)	mtx_lock(&(adap)->lock);
422 #define ADAPTER_UNLOCK(adap)	mtx_unlock(&(adap)->lock);
423 #define ADAPTER_LOCK_INIT(adap, name) mtx_init(&(adap)->lock, name, 0, MTX_DEF)
424 #define ADAPTER_LOCK_DEINIT(adap) mtx_destroy(&(adap)->lock)
425 #define ADAPTER_LOCK_ASSERT_NOTOWNED(adap) mtx_assert(&(adap)->lock, MA_NOTOWNED)
426 #define ADAPTER_LOCK_ASSERT_OWNED(adap) mtx_assert(&(adap)->lock, MA_OWNED)
427 
428 
429 static __inline uint32_t
430 t3_read_reg(adapter_t *adapter, uint32_t reg_addr)
431 {
432 	return (bus_space_read_4(adapter->bt, adapter->bh, reg_addr));
433 }
434 
435 static __inline void
436 t3_write_reg(adapter_t *adapter, uint32_t reg_addr, uint32_t val)
437 {
438 	bus_space_write_4(adapter->bt, adapter->bh, reg_addr, val);
439 }
440 
441 static __inline void
442 t3_os_pci_read_config_4(adapter_t *adapter, int reg, uint32_t *val)
443 {
444 	*val = pci_read_config(adapter->dev, reg, 4);
445 }
446 
447 static __inline void
448 t3_os_pci_write_config_4(adapter_t *adapter, int reg, uint32_t val)
449 {
450 	pci_write_config(adapter->dev, reg, val, 4);
451 }
452 
453 static __inline void
454 t3_os_pci_read_config_2(adapter_t *adapter, int reg, uint16_t *val)
455 {
456 	*val = pci_read_config(adapter->dev, reg, 2);
457 }
458 
459 static __inline void
460 t3_os_pci_write_config_2(adapter_t *adapter, int reg, uint16_t val)
461 {
462 	pci_write_config(adapter->dev, reg, val, 2);
463 }
464 
465 static __inline uint8_t *
466 t3_get_next_mcaddr(struct t3_rx_mode *rm)
467 {
468 	uint8_t *macaddr = NULL;
469 	struct ifnet *ifp = rm->port->ifp;
470 	struct ifmultiaddr *ifma;
471 	int i = 0;
472 
473 	if_maddr_rlock(ifp);
474 	CK_STAILQ_FOREACH(ifma, &ifp->if_multiaddrs, ifma_link) {
475 		if (ifma->ifma_addr->sa_family != AF_LINK)
476 			continue;
477 		if (i == rm->idx) {
478 			macaddr = LLADDR((struct sockaddr_dl *)ifma->ifma_addr);
479 			break;
480 		}
481 		i++;
482 	}
483 	if_maddr_runlock(ifp);
484 
485 	rm->idx++;
486 	return (macaddr);
487 }
488 
489 static __inline void
490 t3_init_rx_mode(struct t3_rx_mode *rm, struct port_info *port)
491 {
492 	rm->idx = 0;
493 	rm->port = port;
494 }
495 
496 static __inline struct port_info *
497 adap2pinfo(struct adapter *adap, int idx)
498 {
499 	return &adap->port[idx];
500 }
501 
502 int t3_os_find_pci_capability(adapter_t *adapter, int cap);
503 int t3_os_pci_save_state(struct adapter *adapter);
504 int t3_os_pci_restore_state(struct adapter *adapter);
505 void t3_os_link_intr(struct port_info *);
506 void t3_os_link_changed(adapter_t *adapter, int port_id, int link_status,
507 			int speed, int duplex, int fc, int mac_was_reset);
508 void t3_os_phymod_changed(struct adapter *adap, int port_id);
509 void t3_sge_err_intr_handler(adapter_t *adapter);
510 #ifdef TCP_OFFLOAD
511 int t3_offload_tx(struct adapter *, struct mbuf *);
512 #endif
513 void t3_os_set_hw_addr(adapter_t *adapter, int port_idx, u8 hw_addr[]);
514 int t3_mgmt_tx(adapter_t *adap, struct mbuf *m);
515 int t3_register_cpl_handler(struct adapter *, int, cpl_handler_t);
516 
517 int t3_sge_alloc(struct adapter *);
518 int t3_sge_free(struct adapter *);
519 int t3_sge_alloc_qset(adapter_t *, uint32_t, int, int, const struct qset_params *,
520     int, struct port_info *);
521 void t3_free_sge_resources(adapter_t *, int);
522 void t3_sge_start(adapter_t *);
523 void t3_sge_stop(adapter_t *);
524 void t3b_intr(void *data);
525 void t3_intr_msi(void *data);
526 void t3_intr_msix(void *data);
527 
528 int t3_sge_init_adapter(adapter_t *);
529 int t3_sge_reset_adapter(adapter_t *);
530 int t3_sge_init_port(struct port_info *);
531 void t3_free_tx_desc(struct sge_qset *qs, int n, int qid);
532 
533 void t3_rx_eth(struct adapter *adap, struct mbuf *m, int ethpad);
534 
535 void t3_add_attach_sysctls(adapter_t *sc);
536 void t3_add_configured_sysctls(adapter_t *sc);
537 int t3_get_desc(const struct sge_qset *qs, unsigned int qnum, unsigned int idx,
538     unsigned char *data);
539 void t3_update_qset_coalesce(struct sge_qset *qs, const struct qset_params *p);
540 
541 /*
542  * XXX figure out how we can return this to being private to sge
543  */
544 #define desc_reclaimable(q) ((int)((q)->processed - (q)->cleaned - TX_MAX_DESC))
545 
546 #define container_of(p, stype, field) ((stype *)(((uint8_t *)(p)) - offsetof(stype, field)))
547 
548 static __inline struct sge_qset *
549 fl_to_qset(struct sge_fl *q, int qidx)
550 {
551 	return container_of(q, struct sge_qset, fl[qidx]);
552 }
553 
554 static __inline struct sge_qset *
555 rspq_to_qset(struct sge_rspq *q)
556 {
557 	return container_of(q, struct sge_qset, rspq);
558 }
559 
560 static __inline struct sge_qset *
561 txq_to_qset(struct sge_txq *q, int qidx)
562 {
563 	return container_of(q, struct sge_qset, txq[qidx]);
564 }
565 
566 #undef container_of
567 
568 #define OFFLOAD_DEVMAP_BIT (1 << MAX_NPORTS)
569 static inline int offload_running(adapter_t *adapter)
570 {
571         return isset(&adapter->open_device_map, OFFLOAD_DEVMAP_BIT);
572 }
573 
574 void cxgb_tx_watchdog(void *arg);
575 int cxgb_transmit(struct ifnet *ifp, struct mbuf *m);
576 void cxgb_qflush(struct ifnet *ifp);
577 void t3_iterate(void (*)(struct adapter *, void *), void *);
578 void cxgb_refresh_stats(struct port_info *);
579 
580 #ifdef NETDUMP
581 int cxgb_netdump_encap(struct sge_qset *qs, struct mbuf **m);
582 int cxgb_netdump_poll_rx(adapter_t *adap, struct sge_qset *qs);
583 int cxgb_netdump_poll_tx(struct sge_qset *qs);
584 #endif
585 
586 #endif
587