xref: /freebsd/sys/dev/ixl/i40e_adminq.c (revision ceebc2f348c028b21bf9bcc99f7a3c4b0cb7d926)
1 /******************************************************************************
2 
3   Copyright (c) 2013-2017, Intel Corporation
4   All rights reserved.
5 
6   Redistribution and use in source and binary forms, with or without
7   modification, are permitted provided that the following conditions are met:
8 
9    1. Redistributions of source code must retain the above copyright notice,
10       this list of conditions and the following disclaimer.
11 
12    2. Redistributions in binary form must reproduce the above copyright
13       notice, this list of conditions and the following disclaimer in the
14       documentation and/or other materials provided with the distribution.
15 
16    3. Neither the name of the Intel Corporation nor the names of its
17       contributors may be used to endorse or promote products derived from
18       this software without specific prior written permission.
19 
20   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
21   AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22   IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23   ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
24   LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25   CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26   SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27   INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28   CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29   ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30   POSSIBILITY OF SUCH DAMAGE.
31 
32 ******************************************************************************/
33 /*$FreeBSD$*/
34 
35 #include "i40e_status.h"
36 #include "i40e_type.h"
37 #include "i40e_register.h"
38 #include "i40e_adminq.h"
39 #include "i40e_prototype.h"
40 
41 /**
42  *  i40e_adminq_init_regs - Initialize AdminQ registers
43  *  @hw: pointer to the hardware structure
44  *
45  *  This assumes the alloc_asq and alloc_arq functions have already been called
46  **/
47 static void i40e_adminq_init_regs(struct i40e_hw *hw)
48 {
49 	/* set head and tail registers in our local struct */
50 	if (i40e_is_vf(hw)) {
51 		hw->aq.asq.tail = I40E_VF_ATQT1;
52 		hw->aq.asq.head = I40E_VF_ATQH1;
53 		hw->aq.asq.len  = I40E_VF_ATQLEN1;
54 		hw->aq.asq.bal  = I40E_VF_ATQBAL1;
55 		hw->aq.asq.bah  = I40E_VF_ATQBAH1;
56 		hw->aq.arq.tail = I40E_VF_ARQT1;
57 		hw->aq.arq.head = I40E_VF_ARQH1;
58 		hw->aq.arq.len  = I40E_VF_ARQLEN1;
59 		hw->aq.arq.bal  = I40E_VF_ARQBAL1;
60 		hw->aq.arq.bah  = I40E_VF_ARQBAH1;
61 	} else {
62 		hw->aq.asq.tail = I40E_PF_ATQT;
63 		hw->aq.asq.head = I40E_PF_ATQH;
64 		hw->aq.asq.len  = I40E_PF_ATQLEN;
65 		hw->aq.asq.bal  = I40E_PF_ATQBAL;
66 		hw->aq.asq.bah  = I40E_PF_ATQBAH;
67 		hw->aq.arq.tail = I40E_PF_ARQT;
68 		hw->aq.arq.head = I40E_PF_ARQH;
69 		hw->aq.arq.len  = I40E_PF_ARQLEN;
70 		hw->aq.arq.bal  = I40E_PF_ARQBAL;
71 		hw->aq.arq.bah  = I40E_PF_ARQBAH;
72 	}
73 }
74 
75 /**
76  *  i40e_alloc_adminq_asq_ring - Allocate Admin Queue send rings
77  *  @hw: pointer to the hardware structure
78  **/
79 enum i40e_status_code i40e_alloc_adminq_asq_ring(struct i40e_hw *hw)
80 {
81 	enum i40e_status_code ret_code;
82 
83 	ret_code = i40e_allocate_dma_mem(hw, &hw->aq.asq.desc_buf,
84 					 i40e_mem_atq_ring,
85 					 (hw->aq.num_asq_entries *
86 					 sizeof(struct i40e_aq_desc)),
87 					 I40E_ADMINQ_DESC_ALIGNMENT);
88 	if (ret_code)
89 		return ret_code;
90 
91 	ret_code = i40e_allocate_virt_mem(hw, &hw->aq.asq.cmd_buf,
92 					  (hw->aq.num_asq_entries *
93 					  sizeof(struct i40e_asq_cmd_details)));
94 	if (ret_code) {
95 		i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
96 		return ret_code;
97 	}
98 
99 	return ret_code;
100 }
101 
102 /**
103  *  i40e_alloc_adminq_arq_ring - Allocate Admin Queue receive rings
104  *  @hw: pointer to the hardware structure
105  **/
106 enum i40e_status_code i40e_alloc_adminq_arq_ring(struct i40e_hw *hw)
107 {
108 	enum i40e_status_code ret_code;
109 
110 	ret_code = i40e_allocate_dma_mem(hw, &hw->aq.arq.desc_buf,
111 					 i40e_mem_arq_ring,
112 					 (hw->aq.num_arq_entries *
113 					 sizeof(struct i40e_aq_desc)),
114 					 I40E_ADMINQ_DESC_ALIGNMENT);
115 
116 	return ret_code;
117 }
118 
119 /**
120  *  i40e_free_adminq_asq - Free Admin Queue send rings
121  *  @hw: pointer to the hardware structure
122  *
123  *  This assumes the posted send buffers have already been cleaned
124  *  and de-allocated
125  **/
126 void i40e_free_adminq_asq(struct i40e_hw *hw)
127 {
128 	i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
129 }
130 
131 /**
132  *  i40e_free_adminq_arq - Free Admin Queue receive rings
133  *  @hw: pointer to the hardware structure
134  *
135  *  This assumes the posted receive buffers have already been cleaned
136  *  and de-allocated
137  **/
138 void i40e_free_adminq_arq(struct i40e_hw *hw)
139 {
140 	i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf);
141 }
142 
143 /**
144  *  i40e_alloc_arq_bufs - Allocate pre-posted buffers for the receive queue
145  *  @hw: pointer to the hardware structure
146  **/
147 static enum i40e_status_code i40e_alloc_arq_bufs(struct i40e_hw *hw)
148 {
149 	enum i40e_status_code ret_code;
150 	struct i40e_aq_desc *desc;
151 	struct i40e_dma_mem *bi;
152 	int i;
153 
154 	/* We'll be allocating the buffer info memory first, then we can
155 	 * allocate the mapped buffers for the event processing
156 	 */
157 
158 	/* buffer_info structures do not need alignment */
159 	ret_code = i40e_allocate_virt_mem(hw, &hw->aq.arq.dma_head,
160 		(hw->aq.num_arq_entries * sizeof(struct i40e_dma_mem)));
161 	if (ret_code)
162 		goto alloc_arq_bufs;
163 	hw->aq.arq.r.arq_bi = (struct i40e_dma_mem *)hw->aq.arq.dma_head.va;
164 
165 	/* allocate the mapped buffers */
166 	for (i = 0; i < hw->aq.num_arq_entries; i++) {
167 		bi = &hw->aq.arq.r.arq_bi[i];
168 		ret_code = i40e_allocate_dma_mem(hw, bi,
169 						 i40e_mem_arq_buf,
170 						 hw->aq.arq_buf_size,
171 						 I40E_ADMINQ_DESC_ALIGNMENT);
172 		if (ret_code)
173 			goto unwind_alloc_arq_bufs;
174 
175 		/* now configure the descriptors for use */
176 		desc = I40E_ADMINQ_DESC(hw->aq.arq, i);
177 
178 		desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_BUF);
179 		if (hw->aq.arq_buf_size > I40E_AQ_LARGE_BUF)
180 			desc->flags |= CPU_TO_LE16(I40E_AQ_FLAG_LB);
181 		desc->opcode = 0;
182 		/* This is in accordance with Admin queue design, there is no
183 		 * register for buffer size configuration
184 		 */
185 		desc->datalen = CPU_TO_LE16((u16)bi->size);
186 		desc->retval = 0;
187 		desc->cookie_high = 0;
188 		desc->cookie_low = 0;
189 		desc->params.external.addr_high =
190 			CPU_TO_LE32(I40E_HI_DWORD(bi->pa));
191 		desc->params.external.addr_low =
192 			CPU_TO_LE32(I40E_LO_DWORD(bi->pa));
193 		desc->params.external.param0 = 0;
194 		desc->params.external.param1 = 0;
195 	}
196 
197 alloc_arq_bufs:
198 	return ret_code;
199 
200 unwind_alloc_arq_bufs:
201 	/* don't try to free the one that failed... */
202 	i--;
203 	for (; i >= 0; i--)
204 		i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]);
205 	i40e_free_virt_mem(hw, &hw->aq.arq.dma_head);
206 
207 	return ret_code;
208 }
209 
210 /**
211  *  i40e_alloc_asq_bufs - Allocate empty buffer structs for the send queue
212  *  @hw: pointer to the hardware structure
213  **/
214 static enum i40e_status_code i40e_alloc_asq_bufs(struct i40e_hw *hw)
215 {
216 	enum i40e_status_code ret_code;
217 	struct i40e_dma_mem *bi;
218 	int i;
219 
220 	/* No mapped memory needed yet, just the buffer info structures */
221 	ret_code = i40e_allocate_virt_mem(hw, &hw->aq.asq.dma_head,
222 		(hw->aq.num_asq_entries * sizeof(struct i40e_dma_mem)));
223 	if (ret_code)
224 		goto alloc_asq_bufs;
225 	hw->aq.asq.r.asq_bi = (struct i40e_dma_mem *)hw->aq.asq.dma_head.va;
226 
227 	/* allocate the mapped buffers */
228 	for (i = 0; i < hw->aq.num_asq_entries; i++) {
229 		bi = &hw->aq.asq.r.asq_bi[i];
230 		ret_code = i40e_allocate_dma_mem(hw, bi,
231 						 i40e_mem_asq_buf,
232 						 hw->aq.asq_buf_size,
233 						 I40E_ADMINQ_DESC_ALIGNMENT);
234 		if (ret_code)
235 			goto unwind_alloc_asq_bufs;
236 	}
237 alloc_asq_bufs:
238 	return ret_code;
239 
240 unwind_alloc_asq_bufs:
241 	/* don't try to free the one that failed... */
242 	i--;
243 	for (; i >= 0; i--)
244 		i40e_free_dma_mem(hw, &hw->aq.asq.r.asq_bi[i]);
245 	i40e_free_virt_mem(hw, &hw->aq.asq.dma_head);
246 
247 	return ret_code;
248 }
249 
250 /**
251  *  i40e_free_arq_bufs - Free receive queue buffer info elements
252  *  @hw: pointer to the hardware structure
253  **/
254 static void i40e_free_arq_bufs(struct i40e_hw *hw)
255 {
256 	int i;
257 
258 	/* free descriptors */
259 	for (i = 0; i < hw->aq.num_arq_entries; i++)
260 		i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]);
261 
262 	/* free the descriptor memory */
263 	i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf);
264 
265 	/* free the dma header */
266 	i40e_free_virt_mem(hw, &hw->aq.arq.dma_head);
267 }
268 
269 /**
270  *  i40e_free_asq_bufs - Free send queue buffer info elements
271  *  @hw: pointer to the hardware structure
272  **/
273 static void i40e_free_asq_bufs(struct i40e_hw *hw)
274 {
275 	int i;
276 
277 	/* only unmap if the address is non-NULL */
278 	for (i = 0; i < hw->aq.num_asq_entries; i++)
279 		if (hw->aq.asq.r.asq_bi[i].pa)
280 			i40e_free_dma_mem(hw, &hw->aq.asq.r.asq_bi[i]);
281 
282 	/* free the buffer info list */
283 	i40e_free_virt_mem(hw, &hw->aq.asq.cmd_buf);
284 
285 	/* free the descriptor memory */
286 	i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
287 
288 	/* free the dma header */
289 	i40e_free_virt_mem(hw, &hw->aq.asq.dma_head);
290 }
291 
292 /**
293  *  i40e_config_asq_regs - configure ASQ registers
294  *  @hw: pointer to the hardware structure
295  *
296  *  Configure base address and length registers for the transmit queue
297  **/
298 static enum i40e_status_code i40e_config_asq_regs(struct i40e_hw *hw)
299 {
300 	enum i40e_status_code ret_code = I40E_SUCCESS;
301 	u32 reg = 0;
302 
303 	/* Clear Head and Tail */
304 	wr32(hw, hw->aq.asq.head, 0);
305 	wr32(hw, hw->aq.asq.tail, 0);
306 
307 	/* set starting point */
308 	if (!i40e_is_vf(hw))
309 		wr32(hw, hw->aq.asq.len, (hw->aq.num_asq_entries |
310 					  I40E_PF_ATQLEN_ATQENABLE_MASK));
311 	if (i40e_is_vf(hw))
312 		wr32(hw, hw->aq.asq.len, (hw->aq.num_asq_entries |
313 					  I40E_VF_ATQLEN1_ATQENABLE_MASK));
314 	wr32(hw, hw->aq.asq.bal, I40E_LO_DWORD(hw->aq.asq.desc_buf.pa));
315 	wr32(hw, hw->aq.asq.bah, I40E_HI_DWORD(hw->aq.asq.desc_buf.pa));
316 
317 	/* Check one register to verify that config was applied */
318 	reg = rd32(hw, hw->aq.asq.bal);
319 	if (reg != I40E_LO_DWORD(hw->aq.asq.desc_buf.pa))
320 		ret_code = I40E_ERR_ADMIN_QUEUE_ERROR;
321 
322 	return ret_code;
323 }
324 
325 /**
326  *  i40e_config_arq_regs - ARQ register configuration
327  *  @hw: pointer to the hardware structure
328  *
329  * Configure base address and length registers for the receive (event queue)
330  **/
331 static enum i40e_status_code i40e_config_arq_regs(struct i40e_hw *hw)
332 {
333 	enum i40e_status_code ret_code = I40E_SUCCESS;
334 	u32 reg = 0;
335 
336 	/* Clear Head and Tail */
337 	wr32(hw, hw->aq.arq.head, 0);
338 	wr32(hw, hw->aq.arq.tail, 0);
339 
340 	/* set starting point */
341 	if (!i40e_is_vf(hw))
342 		wr32(hw, hw->aq.arq.len, (hw->aq.num_arq_entries |
343 					  I40E_PF_ARQLEN_ARQENABLE_MASK));
344 	if (i40e_is_vf(hw))
345 		wr32(hw, hw->aq.arq.len, (hw->aq.num_arq_entries |
346 					  I40E_VF_ARQLEN1_ARQENABLE_MASK));
347 	wr32(hw, hw->aq.arq.bal, I40E_LO_DWORD(hw->aq.arq.desc_buf.pa));
348 	wr32(hw, hw->aq.arq.bah, I40E_HI_DWORD(hw->aq.arq.desc_buf.pa));
349 
350 	/* Update tail in the HW to post pre-allocated buffers */
351 	wr32(hw, hw->aq.arq.tail, hw->aq.num_arq_entries - 1);
352 
353 	/* Check one register to verify that config was applied */
354 	reg = rd32(hw, hw->aq.arq.bal);
355 	if (reg != I40E_LO_DWORD(hw->aq.arq.desc_buf.pa))
356 		ret_code = I40E_ERR_ADMIN_QUEUE_ERROR;
357 
358 	return ret_code;
359 }
360 
361 /**
362  *  i40e_init_asq - main initialization routine for ASQ
363  *  @hw: pointer to the hardware structure
364  *
365  *  This is the main initialization routine for the Admin Send Queue
366  *  Prior to calling this function, drivers *MUST* set the following fields
367  *  in the hw->aq structure:
368  *     - hw->aq.num_asq_entries
369  *     - hw->aq.arq_buf_size
370  *
371  *  Do *NOT* hold the lock when calling this as the memory allocation routines
372  *  called are not going to be atomic context safe
373  **/
374 enum i40e_status_code i40e_init_asq(struct i40e_hw *hw)
375 {
376 	enum i40e_status_code ret_code = I40E_SUCCESS;
377 
378 	if (hw->aq.asq.count > 0) {
379 		/* queue already initialized */
380 		ret_code = I40E_ERR_NOT_READY;
381 		goto init_adminq_exit;
382 	}
383 
384 	/* verify input for valid configuration */
385 	if ((hw->aq.num_asq_entries == 0) ||
386 	    (hw->aq.asq_buf_size == 0)) {
387 		ret_code = I40E_ERR_CONFIG;
388 		goto init_adminq_exit;
389 	}
390 
391 	hw->aq.asq.next_to_use = 0;
392 	hw->aq.asq.next_to_clean = 0;
393 
394 	/* allocate the ring memory */
395 	ret_code = i40e_alloc_adminq_asq_ring(hw);
396 	if (ret_code != I40E_SUCCESS)
397 		goto init_adminq_exit;
398 
399 	/* allocate buffers in the rings */
400 	ret_code = i40e_alloc_asq_bufs(hw);
401 	if (ret_code != I40E_SUCCESS)
402 		goto init_adminq_free_rings;
403 
404 	/* initialize base registers */
405 	ret_code = i40e_config_asq_regs(hw);
406 	if (ret_code != I40E_SUCCESS)
407 		goto init_adminq_free_rings;
408 
409 	/* success! */
410 	hw->aq.asq.count = hw->aq.num_asq_entries;
411 	goto init_adminq_exit;
412 
413 init_adminq_free_rings:
414 	i40e_free_adminq_asq(hw);
415 
416 init_adminq_exit:
417 	return ret_code;
418 }
419 
420 /**
421  *  i40e_init_arq - initialize ARQ
422  *  @hw: pointer to the hardware structure
423  *
424  *  The main initialization routine for the Admin Receive (Event) Queue.
425  *  Prior to calling this function, drivers *MUST* set the following fields
426  *  in the hw->aq structure:
427  *     - hw->aq.num_asq_entries
428  *     - hw->aq.arq_buf_size
429  *
430  *  Do *NOT* hold the lock when calling this as the memory allocation routines
431  *  called are not going to be atomic context safe
432  **/
433 enum i40e_status_code i40e_init_arq(struct i40e_hw *hw)
434 {
435 	enum i40e_status_code ret_code = I40E_SUCCESS;
436 
437 	if (hw->aq.arq.count > 0) {
438 		/* queue already initialized */
439 		ret_code = I40E_ERR_NOT_READY;
440 		goto init_adminq_exit;
441 	}
442 
443 	/* verify input for valid configuration */
444 	if ((hw->aq.num_arq_entries == 0) ||
445 	    (hw->aq.arq_buf_size == 0)) {
446 		ret_code = I40E_ERR_CONFIG;
447 		goto init_adminq_exit;
448 	}
449 
450 	hw->aq.arq.next_to_use = 0;
451 	hw->aq.arq.next_to_clean = 0;
452 
453 	/* allocate the ring memory */
454 	ret_code = i40e_alloc_adminq_arq_ring(hw);
455 	if (ret_code != I40E_SUCCESS)
456 		goto init_adminq_exit;
457 
458 	/* allocate buffers in the rings */
459 	ret_code = i40e_alloc_arq_bufs(hw);
460 	if (ret_code != I40E_SUCCESS)
461 		goto init_adminq_free_rings;
462 
463 	/* initialize base registers */
464 	ret_code = i40e_config_arq_regs(hw);
465 	if (ret_code != I40E_SUCCESS)
466 		goto init_adminq_free_rings;
467 
468 	/* success! */
469 	hw->aq.arq.count = hw->aq.num_arq_entries;
470 	goto init_adminq_exit;
471 
472 init_adminq_free_rings:
473 	i40e_free_adminq_arq(hw);
474 
475 init_adminq_exit:
476 	return ret_code;
477 }
478 
479 /**
480  *  i40e_shutdown_asq - shutdown the ASQ
481  *  @hw: pointer to the hardware structure
482  *
483  *  The main shutdown routine for the Admin Send Queue
484  **/
485 enum i40e_status_code i40e_shutdown_asq(struct i40e_hw *hw)
486 {
487 	enum i40e_status_code ret_code = I40E_SUCCESS;
488 
489 	i40e_acquire_spinlock(&hw->aq.asq_spinlock);
490 
491 	if (hw->aq.asq.count == 0) {
492 		ret_code = I40E_ERR_NOT_READY;
493 		goto shutdown_asq_out;
494 	}
495 
496 	/* Stop firmware AdminQ processing */
497 	wr32(hw, hw->aq.asq.head, 0);
498 	wr32(hw, hw->aq.asq.tail, 0);
499 	wr32(hw, hw->aq.asq.len, 0);
500 	wr32(hw, hw->aq.asq.bal, 0);
501 	wr32(hw, hw->aq.asq.bah, 0);
502 
503 	hw->aq.asq.count = 0; /* to indicate uninitialized queue */
504 
505 	/* free ring buffers */
506 	i40e_free_asq_bufs(hw);
507 
508 shutdown_asq_out:
509 	i40e_release_spinlock(&hw->aq.asq_spinlock);
510 	return ret_code;
511 }
512 
513 /**
514  *  i40e_shutdown_arq - shutdown ARQ
515  *  @hw: pointer to the hardware structure
516  *
517  *  The main shutdown routine for the Admin Receive Queue
518  **/
519 enum i40e_status_code i40e_shutdown_arq(struct i40e_hw *hw)
520 {
521 	enum i40e_status_code ret_code = I40E_SUCCESS;
522 
523 	i40e_acquire_spinlock(&hw->aq.arq_spinlock);
524 
525 	if (hw->aq.arq.count == 0) {
526 		ret_code = I40E_ERR_NOT_READY;
527 		goto shutdown_arq_out;
528 	}
529 
530 	/* Stop firmware AdminQ processing */
531 	wr32(hw, hw->aq.arq.head, 0);
532 	wr32(hw, hw->aq.arq.tail, 0);
533 	wr32(hw, hw->aq.arq.len, 0);
534 	wr32(hw, hw->aq.arq.bal, 0);
535 	wr32(hw, hw->aq.arq.bah, 0);
536 
537 	hw->aq.arq.count = 0; /* to indicate uninitialized queue */
538 
539 	/* free ring buffers */
540 	i40e_free_arq_bufs(hw);
541 
542 shutdown_arq_out:
543 	i40e_release_spinlock(&hw->aq.arq_spinlock);
544 	return ret_code;
545 }
546 
547 /**
548  *  i40e_resume_aq - resume AQ processing from 0
549  *  @hw: pointer to the hardware structure
550  **/
551 static void i40e_resume_aq(struct i40e_hw *hw)
552 {
553 	/* Registers are reset after PF reset */
554 	hw->aq.asq.next_to_use = 0;
555 	hw->aq.asq.next_to_clean = 0;
556 
557 	i40e_config_asq_regs(hw);
558 
559 	hw->aq.arq.next_to_use = 0;
560 	hw->aq.arq.next_to_clean = 0;
561 
562 	i40e_config_arq_regs(hw);
563 }
564 
565 /**
566  *  i40e_init_adminq - main initialization routine for Admin Queue
567  *  @hw: pointer to the hardware structure
568  *
569  *  Prior to calling this function, drivers *MUST* set the following fields
570  *  in the hw->aq structure:
571  *     - hw->aq.num_asq_entries
572  *     - hw->aq.num_arq_entries
573  *     - hw->aq.arq_buf_size
574  *     - hw->aq.asq_buf_size
575  **/
576 enum i40e_status_code i40e_init_adminq(struct i40e_hw *hw)
577 {
578 	u16 cfg_ptr, oem_hi, oem_lo;
579 	u16 eetrack_lo, eetrack_hi;
580 	enum i40e_status_code ret_code;
581 	int retry = 0;
582 
583 	/* verify input for valid configuration */
584 	if ((hw->aq.num_arq_entries == 0) ||
585 	    (hw->aq.num_asq_entries == 0) ||
586 	    (hw->aq.arq_buf_size == 0) ||
587 	    (hw->aq.asq_buf_size == 0)) {
588 		ret_code = I40E_ERR_CONFIG;
589 		goto init_adminq_exit;
590 	}
591 	i40e_init_spinlock(&hw->aq.asq_spinlock);
592 	i40e_init_spinlock(&hw->aq.arq_spinlock);
593 
594 	/* Set up register offsets */
595 	i40e_adminq_init_regs(hw);
596 
597 	/* setup ASQ command write back timeout */
598 	hw->aq.asq_cmd_timeout = I40E_ASQ_CMD_TIMEOUT;
599 
600 	/* allocate the ASQ */
601 	ret_code = i40e_init_asq(hw);
602 	if (ret_code != I40E_SUCCESS)
603 		goto init_adminq_destroy_spinlocks;
604 
605 	/* allocate the ARQ */
606 	ret_code = i40e_init_arq(hw);
607 	if (ret_code != I40E_SUCCESS)
608 		goto init_adminq_free_asq;
609 
610 	/* VF has no need of firmware */
611 	if (i40e_is_vf(hw))
612 		goto init_adminq_exit;
613 	/* There are some cases where the firmware may not be quite ready
614 	 * for AdminQ operations, so we retry the AdminQ setup a few times
615 	 * if we see timeouts in this first AQ call.
616 	 */
617 	do {
618 		ret_code = i40e_aq_get_firmware_version(hw,
619 							&hw->aq.fw_maj_ver,
620 							&hw->aq.fw_min_ver,
621 							&hw->aq.fw_build,
622 							&hw->aq.api_maj_ver,
623 							&hw->aq.api_min_ver,
624 							NULL);
625 		if (ret_code != I40E_ERR_ADMIN_QUEUE_TIMEOUT)
626 			break;
627 		retry++;
628 		i40e_msec_delay(100);
629 		i40e_resume_aq(hw);
630 	} while (retry < 10);
631 	if (ret_code != I40E_SUCCESS)
632 		goto init_adminq_free_arq;
633 
634 	/* get the NVM version info */
635 	i40e_read_nvm_word(hw, I40E_SR_NVM_DEV_STARTER_VERSION,
636 			   &hw->nvm.version);
637 	i40e_read_nvm_word(hw, I40E_SR_NVM_EETRACK_LO, &eetrack_lo);
638 	i40e_read_nvm_word(hw, I40E_SR_NVM_EETRACK_HI, &eetrack_hi);
639 	hw->nvm.eetrack = (eetrack_hi << 16) | eetrack_lo;
640 	i40e_read_nvm_word(hw, I40E_SR_BOOT_CONFIG_PTR, &cfg_ptr);
641 	i40e_read_nvm_word(hw, (cfg_ptr + I40E_NVM_OEM_VER_OFF),
642 			   &oem_hi);
643 	i40e_read_nvm_word(hw, (cfg_ptr + (I40E_NVM_OEM_VER_OFF + 1)),
644 			   &oem_lo);
645 	hw->nvm.oem_ver = ((u32)oem_hi << 16) | oem_lo;
646 
647 	/* The ability to RX (not drop) 802.1ad frames was added in API 1.7 */
648 	if ((hw->aq.api_maj_ver > 1) ||
649 	    ((hw->aq.api_maj_ver == 1) &&
650 	     (hw->aq.api_min_ver >= 7)))
651 		hw->flags |= I40E_HW_FLAG_802_1AD_CAPABLE;
652 
653 	if (hw->mac.type == I40E_MAC_XL710 &&
654 	    hw->aq.api_maj_ver == I40E_FW_API_VERSION_MAJOR &&
655 	    hw->aq.api_min_ver >= I40E_MINOR_VER_GET_LINK_INFO_XL710) {
656 		hw->flags |= I40E_HW_FLAG_AQ_PHY_ACCESS_CAPABLE;
657 	}
658 
659 	/* Newer versions of firmware require lock when reading the NVM */
660 	if ((hw->aq.api_maj_ver > 1) ||
661 	    ((hw->aq.api_maj_ver == 1) &&
662 	     (hw->aq.api_min_ver >= 5)))
663 		hw->flags |= I40E_HW_FLAG_NVM_READ_REQUIRES_LOCK;
664 
665 	if (hw->aq.api_maj_ver > I40E_FW_API_VERSION_MAJOR) {
666 		ret_code = I40E_ERR_FIRMWARE_API_VERSION;
667 		goto init_adminq_free_arq;
668 	}
669 
670 	/* pre-emptive resource lock release */
671 	i40e_aq_release_resource(hw, I40E_NVM_RESOURCE_ID, 0, NULL);
672 	hw->nvm_release_on_done = FALSE;
673 	hw->nvmupd_state = I40E_NVMUPD_STATE_INIT;
674 
675 	ret_code = I40E_SUCCESS;
676 
677 	/* success! */
678 	goto init_adminq_exit;
679 
680 init_adminq_free_arq:
681 	i40e_shutdown_arq(hw);
682 init_adminq_free_asq:
683 	i40e_shutdown_asq(hw);
684 init_adminq_destroy_spinlocks:
685 	i40e_destroy_spinlock(&hw->aq.asq_spinlock);
686 	i40e_destroy_spinlock(&hw->aq.arq_spinlock);
687 
688 init_adminq_exit:
689 	return ret_code;
690 }
691 
692 /**
693  *  i40e_shutdown_adminq - shutdown routine for the Admin Queue
694  *  @hw: pointer to the hardware structure
695  **/
696 enum i40e_status_code i40e_shutdown_adminq(struct i40e_hw *hw)
697 {
698 	enum i40e_status_code ret_code = I40E_SUCCESS;
699 
700 	if (i40e_check_asq_alive(hw))
701 		i40e_aq_queue_shutdown(hw, TRUE);
702 
703 	i40e_shutdown_asq(hw);
704 	i40e_shutdown_arq(hw);
705 	i40e_destroy_spinlock(&hw->aq.asq_spinlock);
706 	i40e_destroy_spinlock(&hw->aq.arq_spinlock);
707 
708 	if (hw->nvm_buff.va)
709 		i40e_free_virt_mem(hw, &hw->nvm_buff);
710 
711 	return ret_code;
712 }
713 
714 /**
715  *  i40e_clean_asq - cleans Admin send queue
716  *  @hw: pointer to the hardware structure
717  *
718  *  returns the number of free desc
719  **/
720 u16 i40e_clean_asq(struct i40e_hw *hw)
721 {
722 	struct i40e_adminq_ring *asq = &(hw->aq.asq);
723 	struct i40e_asq_cmd_details *details;
724 	u16 ntc = asq->next_to_clean;
725 	struct i40e_aq_desc desc_cb;
726 	struct i40e_aq_desc *desc;
727 
728 	desc = I40E_ADMINQ_DESC(*asq, ntc);
729 	details = I40E_ADMINQ_DETAILS(*asq, ntc);
730 	while (rd32(hw, hw->aq.asq.head) != ntc) {
731 		i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
732 			   "ntc %d head %d.\n", ntc, rd32(hw, hw->aq.asq.head));
733 
734 		if (details->callback) {
735 			I40E_ADMINQ_CALLBACK cb_func =
736 					(I40E_ADMINQ_CALLBACK)details->callback;
737 			i40e_memcpy(&desc_cb, desc, sizeof(struct i40e_aq_desc),
738 				    I40E_DMA_TO_DMA);
739 			cb_func(hw, &desc_cb);
740 		}
741 		i40e_memset(desc, 0, sizeof(*desc), I40E_DMA_MEM);
742 		i40e_memset(details, 0, sizeof(*details), I40E_NONDMA_MEM);
743 		ntc++;
744 		if (ntc == asq->count)
745 			ntc = 0;
746 		desc = I40E_ADMINQ_DESC(*asq, ntc);
747 		details = I40E_ADMINQ_DETAILS(*asq, ntc);
748 	}
749 
750 	asq->next_to_clean = ntc;
751 
752 	return I40E_DESC_UNUSED(asq);
753 }
754 
755 /**
756  *  i40e_asq_done - check if FW has processed the Admin Send Queue
757  *  @hw: pointer to the hw struct
758  *
759  *  Returns TRUE if the firmware has processed all descriptors on the
760  *  admin send queue. Returns FALSE if there are still requests pending.
761  **/
762 bool i40e_asq_done(struct i40e_hw *hw)
763 {
764 	/* AQ designers suggest use of head for better
765 	 * timing reliability than DD bit
766 	 */
767 	return rd32(hw, hw->aq.asq.head) == hw->aq.asq.next_to_use;
768 
769 }
770 
771 /**
772  *  i40e_asq_send_command - send command to Admin Queue
773  *  @hw: pointer to the hw struct
774  *  @desc: prefilled descriptor describing the command (non DMA mem)
775  *  @buff: buffer to use for indirect commands
776  *  @buff_size: size of buffer for indirect commands
777  *  @cmd_details: pointer to command details structure
778  *
779  *  This is the main send command driver routine for the Admin Queue send
780  *  queue.  It runs the queue, cleans the queue, etc
781  **/
782 enum i40e_status_code i40e_asq_send_command(struct i40e_hw *hw,
783 				struct i40e_aq_desc *desc,
784 				void *buff, /* can be NULL */
785 				u16  buff_size,
786 				struct i40e_asq_cmd_details *cmd_details)
787 {
788 	enum i40e_status_code status = I40E_SUCCESS;
789 	struct i40e_dma_mem *dma_buff = NULL;
790 	struct i40e_asq_cmd_details *details;
791 	struct i40e_aq_desc *desc_on_ring;
792 	bool cmd_completed = FALSE;
793 	u16  retval = 0;
794 	u32  val = 0;
795 
796 	i40e_acquire_spinlock(&hw->aq.asq_spinlock);
797 
798 	hw->aq.asq_last_status = I40E_AQ_RC_OK;
799 
800 	if (hw->aq.asq.count == 0) {
801 		i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
802 			   "AQTX: Admin queue not initialized.\n");
803 		status = I40E_ERR_QUEUE_EMPTY;
804 		goto asq_send_command_error;
805 	}
806 
807 	val = rd32(hw, hw->aq.asq.head);
808 	if (val >= hw->aq.num_asq_entries) {
809 		i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
810 			   "AQTX: head overrun at %d\n", val);
811 		status = I40E_ERR_QUEUE_EMPTY;
812 		goto asq_send_command_error;
813 	}
814 
815 	details = I40E_ADMINQ_DETAILS(hw->aq.asq, hw->aq.asq.next_to_use);
816 	if (cmd_details) {
817 		i40e_memcpy(details,
818 			    cmd_details,
819 			    sizeof(struct i40e_asq_cmd_details),
820 			    I40E_NONDMA_TO_NONDMA);
821 
822 		/* If the cmd_details are defined copy the cookie.  The
823 		 * CPU_TO_LE32 is not needed here because the data is ignored
824 		 * by the FW, only used by the driver
825 		 */
826 		if (details->cookie) {
827 			desc->cookie_high =
828 				CPU_TO_LE32(I40E_HI_DWORD(details->cookie));
829 			desc->cookie_low =
830 				CPU_TO_LE32(I40E_LO_DWORD(details->cookie));
831 		}
832 	} else {
833 		i40e_memset(details, 0,
834 			    sizeof(struct i40e_asq_cmd_details),
835 			    I40E_NONDMA_MEM);
836 	}
837 
838 	/* clear requested flags and then set additional flags if defined */
839 	desc->flags &= ~CPU_TO_LE16(details->flags_dis);
840 	desc->flags |= CPU_TO_LE16(details->flags_ena);
841 
842 	if (buff_size > hw->aq.asq_buf_size) {
843 		i40e_debug(hw,
844 			   I40E_DEBUG_AQ_MESSAGE,
845 			   "AQTX: Invalid buffer size: %d.\n",
846 			   buff_size);
847 		status = I40E_ERR_INVALID_SIZE;
848 		goto asq_send_command_error;
849 	}
850 
851 	if (details->postpone && !details->async) {
852 		i40e_debug(hw,
853 			   I40E_DEBUG_AQ_MESSAGE,
854 			   "AQTX: Async flag not set along with postpone flag");
855 		status = I40E_ERR_PARAM;
856 		goto asq_send_command_error;
857 	}
858 
859 	/* call clean and check queue available function to reclaim the
860 	 * descriptors that were processed by FW, the function returns the
861 	 * number of desc available
862 	 */
863 	/* the clean function called here could be called in a separate thread
864 	 * in case of asynchronous completions
865 	 */
866 	if (i40e_clean_asq(hw) == 0) {
867 		i40e_debug(hw,
868 			   I40E_DEBUG_AQ_MESSAGE,
869 			   "AQTX: Error queue is full.\n");
870 		status = I40E_ERR_ADMIN_QUEUE_FULL;
871 		goto asq_send_command_error;
872 	}
873 
874 	/* initialize the temp desc pointer with the right desc */
875 	desc_on_ring = I40E_ADMINQ_DESC(hw->aq.asq, hw->aq.asq.next_to_use);
876 
877 	/* if the desc is available copy the temp desc to the right place */
878 	i40e_memcpy(desc_on_ring, desc, sizeof(struct i40e_aq_desc),
879 		    I40E_NONDMA_TO_DMA);
880 
881 	/* if buff is not NULL assume indirect command */
882 	if (buff != NULL) {
883 		dma_buff = &(hw->aq.asq.r.asq_bi[hw->aq.asq.next_to_use]);
884 		/* copy the user buff into the respective DMA buff */
885 		i40e_memcpy(dma_buff->va, buff, buff_size,
886 			    I40E_NONDMA_TO_DMA);
887 		desc_on_ring->datalen = CPU_TO_LE16(buff_size);
888 
889 		/* Update the address values in the desc with the pa value
890 		 * for respective buffer
891 		 */
892 		desc_on_ring->params.external.addr_high =
893 				CPU_TO_LE32(I40E_HI_DWORD(dma_buff->pa));
894 		desc_on_ring->params.external.addr_low =
895 				CPU_TO_LE32(I40E_LO_DWORD(dma_buff->pa));
896 	}
897 
898 	/* bump the tail */
899 	i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE, "AQTX: desc and buffer:\n");
900 	i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc_on_ring,
901 		      buff, buff_size);
902 	(hw->aq.asq.next_to_use)++;
903 	if (hw->aq.asq.next_to_use == hw->aq.asq.count)
904 		hw->aq.asq.next_to_use = 0;
905 	if (!details->postpone)
906 		wr32(hw, hw->aq.asq.tail, hw->aq.asq.next_to_use);
907 
908 	/* if cmd_details are not defined or async flag is not set,
909 	 * we need to wait for desc write back
910 	 */
911 	if (!details->async && !details->postpone) {
912 		u32 total_delay = 0;
913 
914 		do {
915 			/* AQ designers suggest use of head for better
916 			 * timing reliability than DD bit
917 			 */
918 			if (i40e_asq_done(hw))
919 				break;
920 			i40e_usec_delay(50);
921 			total_delay += 50;
922 		} while (total_delay < hw->aq.asq_cmd_timeout);
923 	}
924 
925 	/* if ready, copy the desc back to temp */
926 	if (i40e_asq_done(hw)) {
927 		i40e_memcpy(desc, desc_on_ring, sizeof(struct i40e_aq_desc),
928 			    I40E_DMA_TO_NONDMA);
929 		if (buff != NULL)
930 			i40e_memcpy(buff, dma_buff->va, buff_size,
931 				    I40E_DMA_TO_NONDMA);
932 		retval = LE16_TO_CPU(desc->retval);
933 		if (retval != 0) {
934 			i40e_debug(hw,
935 				   I40E_DEBUG_AQ_MESSAGE,
936 				   "AQTX: Command completed with error 0x%X.\n",
937 				   retval);
938 
939 			/* strip off FW internal code */
940 			retval &= 0xff;
941 		}
942 		cmd_completed = TRUE;
943 		if ((enum i40e_admin_queue_err)retval == I40E_AQ_RC_OK)
944 			status = I40E_SUCCESS;
945 		else
946 			status = I40E_ERR_ADMIN_QUEUE_ERROR;
947 		hw->aq.asq_last_status = (enum i40e_admin_queue_err)retval;
948 	}
949 
950 	i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
951 		   "AQTX: desc and buffer writeback:\n");
952 	i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc, buff, buff_size);
953 
954 	/* save writeback aq if requested */
955 	if (details->wb_desc)
956 		i40e_memcpy(details->wb_desc, desc_on_ring,
957 			    sizeof(struct i40e_aq_desc), I40E_DMA_TO_NONDMA);
958 
959 	/* update the error if time out occurred */
960 	if ((!cmd_completed) &&
961 	    (!details->async && !details->postpone)) {
962 		if (rd32(hw, hw->aq.asq.len) & I40E_GL_ATQLEN_ATQCRIT_MASK) {
963 			i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
964 				   "AQTX: AQ Critical error.\n");
965 			status = I40E_ERR_ADMIN_QUEUE_CRITICAL_ERROR;
966 		} else {
967 			i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
968 				   "AQTX: Writeback timeout.\n");
969 			status = I40E_ERR_ADMIN_QUEUE_TIMEOUT;
970 		}
971 	}
972 
973 asq_send_command_error:
974 	i40e_release_spinlock(&hw->aq.asq_spinlock);
975 	return status;
976 }
977 
978 /**
979  *  i40e_fill_default_direct_cmd_desc - AQ descriptor helper function
980  *  @desc:     pointer to the temp descriptor (non DMA mem)
981  *  @opcode:   the opcode can be used to decide which flags to turn off or on
982  *
983  *  Fill the desc with default values
984  **/
985 void i40e_fill_default_direct_cmd_desc(struct i40e_aq_desc *desc,
986 				       u16 opcode)
987 {
988 	/* zero out the desc */
989 	i40e_memset((void *)desc, 0, sizeof(struct i40e_aq_desc),
990 		    I40E_NONDMA_MEM);
991 	desc->opcode = CPU_TO_LE16(opcode);
992 	desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_SI);
993 }
994 
995 /**
996  *  i40e_clean_arq_element
997  *  @hw: pointer to the hw struct
998  *  @e: event info from the receive descriptor, includes any buffers
999  *  @pending: number of events that could be left to process
1000  *
1001  *  This function cleans one Admin Receive Queue element and returns
1002  *  the contents through e.  It can also return how many events are
1003  *  left to process through 'pending'
1004  **/
1005 enum i40e_status_code i40e_clean_arq_element(struct i40e_hw *hw,
1006 					     struct i40e_arq_event_info *e,
1007 					     u16 *pending)
1008 {
1009 	enum i40e_status_code ret_code = I40E_SUCCESS;
1010 	u16 ntc = hw->aq.arq.next_to_clean;
1011 	struct i40e_aq_desc *desc;
1012 	struct i40e_dma_mem *bi;
1013 	u16 desc_idx;
1014 	u16 datalen;
1015 	u16 flags;
1016 	u16 ntu;
1017 
1018 	/* pre-clean the event info */
1019 	i40e_memset(&e->desc, 0, sizeof(e->desc), I40E_NONDMA_MEM);
1020 
1021 	/* take the lock before we start messing with the ring */
1022 	i40e_acquire_spinlock(&hw->aq.arq_spinlock);
1023 
1024 	if (hw->aq.arq.count == 0) {
1025 		i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
1026 			   "AQRX: Admin queue not initialized.\n");
1027 		ret_code = I40E_ERR_QUEUE_EMPTY;
1028 		goto clean_arq_element_err;
1029 	}
1030 
1031 	/* set next_to_use to head */
1032 	if (!i40e_is_vf(hw))
1033 		ntu = rd32(hw, hw->aq.arq.head) & I40E_PF_ARQH_ARQH_MASK;
1034 	else
1035 		ntu = rd32(hw, hw->aq.arq.head) & I40E_VF_ARQH1_ARQH_MASK;
1036 	if (ntu == ntc) {
1037 		/* nothing to do - shouldn't need to update ring's values */
1038 		ret_code = I40E_ERR_ADMIN_QUEUE_NO_WORK;
1039 		goto clean_arq_element_out;
1040 	}
1041 
1042 	/* now clean the next descriptor */
1043 	desc = I40E_ADMINQ_DESC(hw->aq.arq, ntc);
1044 	desc_idx = ntc;
1045 
1046 	hw->aq.arq_last_status =
1047 		(enum i40e_admin_queue_err)LE16_TO_CPU(desc->retval);
1048 	flags = LE16_TO_CPU(desc->flags);
1049 	if (flags & I40E_AQ_FLAG_ERR) {
1050 		ret_code = I40E_ERR_ADMIN_QUEUE_ERROR;
1051 		i40e_debug(hw,
1052 			   I40E_DEBUG_AQ_MESSAGE,
1053 			   "AQRX: Event received with error 0x%X.\n",
1054 			   hw->aq.arq_last_status);
1055 	}
1056 
1057 	i40e_memcpy(&e->desc, desc, sizeof(struct i40e_aq_desc),
1058 		    I40E_DMA_TO_NONDMA);
1059 	datalen = LE16_TO_CPU(desc->datalen);
1060 	e->msg_len = min(datalen, e->buf_len);
1061 	if (e->msg_buf != NULL && (e->msg_len != 0))
1062 		i40e_memcpy(e->msg_buf,
1063 			    hw->aq.arq.r.arq_bi[desc_idx].va,
1064 			    e->msg_len, I40E_DMA_TO_NONDMA);
1065 
1066 	i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE, "AQRX: desc and buffer:\n");
1067 	i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc, e->msg_buf,
1068 		      hw->aq.arq_buf_size);
1069 
1070 	/* Restore the original datalen and buffer address in the desc,
1071 	 * FW updates datalen to indicate the event message
1072 	 * size
1073 	 */
1074 	bi = &hw->aq.arq.r.arq_bi[ntc];
1075 	i40e_memset((void *)desc, 0, sizeof(struct i40e_aq_desc), I40E_DMA_MEM);
1076 
1077 	desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_BUF);
1078 	if (hw->aq.arq_buf_size > I40E_AQ_LARGE_BUF)
1079 		desc->flags |= CPU_TO_LE16(I40E_AQ_FLAG_LB);
1080 	desc->datalen = CPU_TO_LE16((u16)bi->size);
1081 	desc->params.external.addr_high = CPU_TO_LE32(I40E_HI_DWORD(bi->pa));
1082 	desc->params.external.addr_low = CPU_TO_LE32(I40E_LO_DWORD(bi->pa));
1083 
1084 	/* set tail = the last cleaned desc index. */
1085 	wr32(hw, hw->aq.arq.tail, ntc);
1086 	/* ntc is updated to tail + 1 */
1087 	ntc++;
1088 	if (ntc == hw->aq.num_arq_entries)
1089 		ntc = 0;
1090 	hw->aq.arq.next_to_clean = ntc;
1091 	hw->aq.arq.next_to_use = ntu;
1092 
1093 	i40e_nvmupd_check_wait_event(hw, LE16_TO_CPU(e->desc.opcode), &e->desc);
1094 clean_arq_element_out:
1095 	/* Set pending if needed, unlock and return */
1096 	if (pending != NULL)
1097 		*pending = (ntc > ntu ? hw->aq.arq.count : 0) + (ntu - ntc);
1098 clean_arq_element_err:
1099 	i40e_release_spinlock(&hw->aq.arq_spinlock);
1100 
1101 	return ret_code;
1102 }
1103 
1104