xref: /freebsd/sys/dev/ixl/i40e_adminq.c (revision 65b71c48d226f3ef3eb1c4452dccc678a7d2ac96)
1 /******************************************************************************
2 
3   Copyright (c) 2013-2015, Intel Corporation
4   All rights reserved.
5 
6   Redistribution and use in source and binary forms, with or without
7   modification, are permitted provided that the following conditions are met:
8 
9    1. Redistributions of source code must retain the above copyright notice,
10       this list of conditions and the following disclaimer.
11 
12    2. Redistributions in binary form must reproduce the above copyright
13       notice, this list of conditions and the following disclaimer in the
14       documentation and/or other materials provided with the distribution.
15 
16    3. Neither the name of the Intel Corporation nor the names of its
17       contributors may be used to endorse or promote products derived from
18       this software without specific prior written permission.
19 
20   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
21   AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22   IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23   ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
24   LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25   CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26   SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27   INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28   CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29   ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30   POSSIBILITY OF SUCH DAMAGE.
31 
32 ******************************************************************************/
33 /*$FreeBSD$*/
34 
35 #include "i40e_status.h"
36 #include "i40e_type.h"
37 #include "i40e_register.h"
38 #include "i40e_adminq.h"
39 #include "i40e_prototype.h"
40 
41 /**
42  * i40e_is_nvm_update_op - return TRUE if this is an NVM update operation
43  * @desc: API request descriptor
44  **/
45 static INLINE bool i40e_is_nvm_update_op(struct i40e_aq_desc *desc)
46 {
47 	return (desc->opcode == CPU_TO_LE16(i40e_aqc_opc_nvm_erase) ||
48 		desc->opcode == CPU_TO_LE16(i40e_aqc_opc_nvm_update));
49 }
50 
51 /**
52  *  i40e_adminq_init_regs - Initialize AdminQ registers
53  *  @hw: pointer to the hardware structure
54  *
55  *  This assumes the alloc_asq and alloc_arq functions have already been called
56  **/
57 static void i40e_adminq_init_regs(struct i40e_hw *hw)
58 {
59 	/* set head and tail registers in our local struct */
60 	if (i40e_is_vf(hw)) {
61 		hw->aq.asq.tail = I40E_VF_ATQT1;
62 		hw->aq.asq.head = I40E_VF_ATQH1;
63 		hw->aq.asq.len  = I40E_VF_ATQLEN1;
64 		hw->aq.asq.bal  = I40E_VF_ATQBAL1;
65 		hw->aq.asq.bah  = I40E_VF_ATQBAH1;
66 		hw->aq.arq.tail = I40E_VF_ARQT1;
67 		hw->aq.arq.head = I40E_VF_ARQH1;
68 		hw->aq.arq.len  = I40E_VF_ARQLEN1;
69 		hw->aq.arq.bal  = I40E_VF_ARQBAL1;
70 		hw->aq.arq.bah  = I40E_VF_ARQBAH1;
71 	} else {
72 		hw->aq.asq.tail = I40E_PF_ATQT;
73 		hw->aq.asq.head = I40E_PF_ATQH;
74 		hw->aq.asq.len  = I40E_PF_ATQLEN;
75 		hw->aq.asq.bal  = I40E_PF_ATQBAL;
76 		hw->aq.asq.bah  = I40E_PF_ATQBAH;
77 		hw->aq.arq.tail = I40E_PF_ARQT;
78 		hw->aq.arq.head = I40E_PF_ARQH;
79 		hw->aq.arq.len  = I40E_PF_ARQLEN;
80 		hw->aq.arq.bal  = I40E_PF_ARQBAL;
81 		hw->aq.arq.bah  = I40E_PF_ARQBAH;
82 	}
83 }
84 
85 /**
86  *  i40e_alloc_adminq_asq_ring - Allocate Admin Queue send rings
87  *  @hw: pointer to the hardware structure
88  **/
89 enum i40e_status_code i40e_alloc_adminq_asq_ring(struct i40e_hw *hw)
90 {
91 	enum i40e_status_code ret_code;
92 
93 	ret_code = i40e_allocate_dma_mem(hw, &hw->aq.asq.desc_buf,
94 					 i40e_mem_atq_ring,
95 					 (hw->aq.num_asq_entries *
96 					 sizeof(struct i40e_aq_desc)),
97 					 I40E_ADMINQ_DESC_ALIGNMENT);
98 	if (ret_code)
99 		return ret_code;
100 
101 	ret_code = i40e_allocate_virt_mem(hw, &hw->aq.asq.cmd_buf,
102 					  (hw->aq.num_asq_entries *
103 					  sizeof(struct i40e_asq_cmd_details)));
104 	if (ret_code) {
105 		i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
106 		return ret_code;
107 	}
108 
109 	return ret_code;
110 }
111 
112 /**
113  *  i40e_alloc_adminq_arq_ring - Allocate Admin Queue receive rings
114  *  @hw: pointer to the hardware structure
115  **/
116 enum i40e_status_code i40e_alloc_adminq_arq_ring(struct i40e_hw *hw)
117 {
118 	enum i40e_status_code ret_code;
119 
120 	ret_code = i40e_allocate_dma_mem(hw, &hw->aq.arq.desc_buf,
121 					 i40e_mem_arq_ring,
122 					 (hw->aq.num_arq_entries *
123 					 sizeof(struct i40e_aq_desc)),
124 					 I40E_ADMINQ_DESC_ALIGNMENT);
125 
126 	return ret_code;
127 }
128 
129 /**
130  *  i40e_free_adminq_asq - Free Admin Queue send rings
131  *  @hw: pointer to the hardware structure
132  *
133  *  This assumes the posted send buffers have already been cleaned
134  *  and de-allocated
135  **/
136 void i40e_free_adminq_asq(struct i40e_hw *hw)
137 {
138 	i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
139 }
140 
141 /**
142  *  i40e_free_adminq_arq - Free Admin Queue receive rings
143  *  @hw: pointer to the hardware structure
144  *
145  *  This assumes the posted receive buffers have already been cleaned
146  *  and de-allocated
147  **/
148 void i40e_free_adminq_arq(struct i40e_hw *hw)
149 {
150 	i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf);
151 }
152 
153 /**
154  *  i40e_alloc_arq_bufs - Allocate pre-posted buffers for the receive queue
155  *  @hw: pointer to the hardware structure
156  **/
157 static enum i40e_status_code i40e_alloc_arq_bufs(struct i40e_hw *hw)
158 {
159 	enum i40e_status_code ret_code;
160 	struct i40e_aq_desc *desc;
161 	struct i40e_dma_mem *bi;
162 	int i;
163 
164 	/* We'll be allocating the buffer info memory first, then we can
165 	 * allocate the mapped buffers for the event processing
166 	 */
167 
168 	/* buffer_info structures do not need alignment */
169 	ret_code = i40e_allocate_virt_mem(hw, &hw->aq.arq.dma_head,
170 		(hw->aq.num_arq_entries * sizeof(struct i40e_dma_mem)));
171 	if (ret_code)
172 		goto alloc_arq_bufs;
173 	hw->aq.arq.r.arq_bi = (struct i40e_dma_mem *)hw->aq.arq.dma_head.va;
174 
175 	/* allocate the mapped buffers */
176 	for (i = 0; i < hw->aq.num_arq_entries; i++) {
177 		bi = &hw->aq.arq.r.arq_bi[i];
178 		ret_code = i40e_allocate_dma_mem(hw, bi,
179 						 i40e_mem_arq_buf,
180 						 hw->aq.arq_buf_size,
181 						 I40E_ADMINQ_DESC_ALIGNMENT);
182 		if (ret_code)
183 			goto unwind_alloc_arq_bufs;
184 
185 		/* now configure the descriptors for use */
186 		desc = I40E_ADMINQ_DESC(hw->aq.arq, i);
187 
188 		desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_BUF);
189 		if (hw->aq.arq_buf_size > I40E_AQ_LARGE_BUF)
190 			desc->flags |= CPU_TO_LE16(I40E_AQ_FLAG_LB);
191 		desc->opcode = 0;
192 		/* This is in accordance with Admin queue design, there is no
193 		 * register for buffer size configuration
194 		 */
195 		desc->datalen = CPU_TO_LE16((u16)bi->size);
196 		desc->retval = 0;
197 		desc->cookie_high = 0;
198 		desc->cookie_low = 0;
199 		desc->params.external.addr_high =
200 			CPU_TO_LE32(I40E_HI_DWORD(bi->pa));
201 		desc->params.external.addr_low =
202 			CPU_TO_LE32(I40E_LO_DWORD(bi->pa));
203 		desc->params.external.param0 = 0;
204 		desc->params.external.param1 = 0;
205 	}
206 
207 alloc_arq_bufs:
208 	return ret_code;
209 
210 unwind_alloc_arq_bufs:
211 	/* don't try to free the one that failed... */
212 	i--;
213 	for (; i >= 0; i--)
214 		i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]);
215 	i40e_free_virt_mem(hw, &hw->aq.arq.dma_head);
216 
217 	return ret_code;
218 }
219 
220 /**
221  *  i40e_alloc_asq_bufs - Allocate empty buffer structs for the send queue
222  *  @hw: pointer to the hardware structure
223  **/
224 static enum i40e_status_code i40e_alloc_asq_bufs(struct i40e_hw *hw)
225 {
226 	enum i40e_status_code ret_code;
227 	struct i40e_dma_mem *bi;
228 	int i;
229 
230 	/* No mapped memory needed yet, just the buffer info structures */
231 	ret_code = i40e_allocate_virt_mem(hw, &hw->aq.asq.dma_head,
232 		(hw->aq.num_asq_entries * sizeof(struct i40e_dma_mem)));
233 	if (ret_code)
234 		goto alloc_asq_bufs;
235 	hw->aq.asq.r.asq_bi = (struct i40e_dma_mem *)hw->aq.asq.dma_head.va;
236 
237 	/* allocate the mapped buffers */
238 	for (i = 0; i < hw->aq.num_asq_entries; i++) {
239 		bi = &hw->aq.asq.r.asq_bi[i];
240 		ret_code = i40e_allocate_dma_mem(hw, bi,
241 						 i40e_mem_asq_buf,
242 						 hw->aq.asq_buf_size,
243 						 I40E_ADMINQ_DESC_ALIGNMENT);
244 		if (ret_code)
245 			goto unwind_alloc_asq_bufs;
246 	}
247 alloc_asq_bufs:
248 	return ret_code;
249 
250 unwind_alloc_asq_bufs:
251 	/* don't try to free the one that failed... */
252 	i--;
253 	for (; i >= 0; i--)
254 		i40e_free_dma_mem(hw, &hw->aq.asq.r.asq_bi[i]);
255 	i40e_free_virt_mem(hw, &hw->aq.asq.dma_head);
256 
257 	return ret_code;
258 }
259 
260 /**
261  *  i40e_free_arq_bufs - Free receive queue buffer info elements
262  *  @hw: pointer to the hardware structure
263  **/
264 static void i40e_free_arq_bufs(struct i40e_hw *hw)
265 {
266 	int i;
267 
268 	/* free descriptors */
269 	for (i = 0; i < hw->aq.num_arq_entries; i++)
270 		i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]);
271 
272 	/* free the descriptor memory */
273 	i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf);
274 
275 	/* free the dma header */
276 	i40e_free_virt_mem(hw, &hw->aq.arq.dma_head);
277 }
278 
279 /**
280  *  i40e_free_asq_bufs - Free send queue buffer info elements
281  *  @hw: pointer to the hardware structure
282  **/
283 static void i40e_free_asq_bufs(struct i40e_hw *hw)
284 {
285 	int i;
286 
287 	/* only unmap if the address is non-NULL */
288 	for (i = 0; i < hw->aq.num_asq_entries; i++)
289 		if (hw->aq.asq.r.asq_bi[i].pa)
290 			i40e_free_dma_mem(hw, &hw->aq.asq.r.asq_bi[i]);
291 
292 	/* free the buffer info list */
293 	i40e_free_virt_mem(hw, &hw->aq.asq.cmd_buf);
294 
295 	/* free the descriptor memory */
296 	i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
297 
298 	/* free the dma header */
299 	i40e_free_virt_mem(hw, &hw->aq.asq.dma_head);
300 }
301 
302 /**
303  *  i40e_config_asq_regs - configure ASQ registers
304  *  @hw: pointer to the hardware structure
305  *
306  *  Configure base address and length registers for the transmit queue
307  **/
308 static enum i40e_status_code i40e_config_asq_regs(struct i40e_hw *hw)
309 {
310 	enum i40e_status_code ret_code = I40E_SUCCESS;
311 	u32 reg = 0;
312 
313 	/* Clear Head and Tail */
314 	wr32(hw, hw->aq.asq.head, 0);
315 	wr32(hw, hw->aq.asq.tail, 0);
316 
317 	/* set starting point */
318 	if (!i40e_is_vf(hw))
319 		wr32(hw, hw->aq.asq.len, (hw->aq.num_asq_entries |
320 					  I40E_PF_ATQLEN_ATQENABLE_MASK));
321 	if (i40e_is_vf(hw))
322 		wr32(hw, hw->aq.asq.len, (hw->aq.num_asq_entries |
323 					  I40E_VF_ATQLEN1_ATQENABLE_MASK));
324 	wr32(hw, hw->aq.asq.bal, I40E_LO_DWORD(hw->aq.asq.desc_buf.pa));
325 	wr32(hw, hw->aq.asq.bah, I40E_HI_DWORD(hw->aq.asq.desc_buf.pa));
326 
327 	/* Check one register to verify that config was applied */
328 	reg = rd32(hw, hw->aq.asq.bal);
329 	if (reg != I40E_LO_DWORD(hw->aq.asq.desc_buf.pa))
330 		ret_code = I40E_ERR_ADMIN_QUEUE_ERROR;
331 
332 	return ret_code;
333 }
334 
335 /**
336  *  i40e_config_arq_regs - ARQ register configuration
337  *  @hw: pointer to the hardware structure
338  *
339  * Configure base address and length registers for the receive (event queue)
340  **/
341 static enum i40e_status_code i40e_config_arq_regs(struct i40e_hw *hw)
342 {
343 	enum i40e_status_code ret_code = I40E_SUCCESS;
344 	u32 reg = 0;
345 
346 	/* Clear Head and Tail */
347 	wr32(hw, hw->aq.arq.head, 0);
348 	wr32(hw, hw->aq.arq.tail, 0);
349 
350 	/* set starting point */
351 	if (!i40e_is_vf(hw))
352 		wr32(hw, hw->aq.arq.len, (hw->aq.num_arq_entries |
353 					  I40E_PF_ARQLEN_ARQENABLE_MASK));
354 	if (i40e_is_vf(hw))
355 		wr32(hw, hw->aq.arq.len, (hw->aq.num_arq_entries |
356 					  I40E_VF_ARQLEN1_ARQENABLE_MASK));
357 	wr32(hw, hw->aq.arq.bal, I40E_LO_DWORD(hw->aq.arq.desc_buf.pa));
358 	wr32(hw, hw->aq.arq.bah, I40E_HI_DWORD(hw->aq.arq.desc_buf.pa));
359 
360 	/* Update tail in the HW to post pre-allocated buffers */
361 	wr32(hw, hw->aq.arq.tail, hw->aq.num_arq_entries - 1);
362 
363 	/* Check one register to verify that config was applied */
364 	reg = rd32(hw, hw->aq.arq.bal);
365 	if (reg != I40E_LO_DWORD(hw->aq.arq.desc_buf.pa))
366 		ret_code = I40E_ERR_ADMIN_QUEUE_ERROR;
367 
368 	return ret_code;
369 }
370 
371 /**
372  *  i40e_init_asq - main initialization routine for ASQ
373  *  @hw: pointer to the hardware structure
374  *
375  *  This is the main initialization routine for the Admin Send Queue
376  *  Prior to calling this function, drivers *MUST* set the following fields
377  *  in the hw->aq structure:
378  *     - hw->aq.num_asq_entries
379  *     - hw->aq.arq_buf_size
380  *
381  *  Do *NOT* hold the lock when calling this as the memory allocation routines
382  *  called are not going to be atomic context safe
383  **/
384 enum i40e_status_code i40e_init_asq(struct i40e_hw *hw)
385 {
386 	enum i40e_status_code ret_code = I40E_SUCCESS;
387 
388 	if (hw->aq.asq.count > 0) {
389 		/* queue already initialized */
390 		ret_code = I40E_ERR_NOT_READY;
391 		goto init_adminq_exit;
392 	}
393 
394 	/* verify input for valid configuration */
395 	if ((hw->aq.num_asq_entries == 0) ||
396 	    (hw->aq.asq_buf_size == 0)) {
397 		ret_code = I40E_ERR_CONFIG;
398 		goto init_adminq_exit;
399 	}
400 
401 	hw->aq.asq.next_to_use = 0;
402 	hw->aq.asq.next_to_clean = 0;
403 	hw->aq.asq.count = hw->aq.num_asq_entries;
404 
405 	/* allocate the ring memory */
406 	ret_code = i40e_alloc_adminq_asq_ring(hw);
407 	if (ret_code != I40E_SUCCESS)
408 		goto init_adminq_exit;
409 
410 	/* allocate buffers in the rings */
411 	ret_code = i40e_alloc_asq_bufs(hw);
412 	if (ret_code != I40E_SUCCESS)
413 		goto init_adminq_free_rings;
414 
415 	/* initialize base registers */
416 	ret_code = i40e_config_asq_regs(hw);
417 	if (ret_code != I40E_SUCCESS)
418 		goto init_adminq_free_rings;
419 
420 	/* success! */
421 	goto init_adminq_exit;
422 
423 init_adminq_free_rings:
424 	i40e_free_adminq_asq(hw);
425 
426 init_adminq_exit:
427 	return ret_code;
428 }
429 
430 /**
431  *  i40e_init_arq - initialize ARQ
432  *  @hw: pointer to the hardware structure
433  *
434  *  The main initialization routine for the Admin Receive (Event) Queue.
435  *  Prior to calling this function, drivers *MUST* set the following fields
436  *  in the hw->aq structure:
437  *     - hw->aq.num_asq_entries
438  *     - hw->aq.arq_buf_size
439  *
440  *  Do *NOT* hold the lock when calling this as the memory allocation routines
441  *  called are not going to be atomic context safe
442  **/
443 enum i40e_status_code i40e_init_arq(struct i40e_hw *hw)
444 {
445 	enum i40e_status_code ret_code = I40E_SUCCESS;
446 
447 	if (hw->aq.arq.count > 0) {
448 		/* queue already initialized */
449 		ret_code = I40E_ERR_NOT_READY;
450 		goto init_adminq_exit;
451 	}
452 
453 	/* verify input for valid configuration */
454 	if ((hw->aq.num_arq_entries == 0) ||
455 	    (hw->aq.arq_buf_size == 0)) {
456 		ret_code = I40E_ERR_CONFIG;
457 		goto init_adminq_exit;
458 	}
459 
460 	hw->aq.arq.next_to_use = 0;
461 	hw->aq.arq.next_to_clean = 0;
462 	hw->aq.arq.count = hw->aq.num_arq_entries;
463 
464 	/* allocate the ring memory */
465 	ret_code = i40e_alloc_adminq_arq_ring(hw);
466 	if (ret_code != I40E_SUCCESS)
467 		goto init_adminq_exit;
468 
469 	/* allocate buffers in the rings */
470 	ret_code = i40e_alloc_arq_bufs(hw);
471 	if (ret_code != I40E_SUCCESS)
472 		goto init_adminq_free_rings;
473 
474 	/* initialize base registers */
475 	ret_code = i40e_config_arq_regs(hw);
476 	if (ret_code != I40E_SUCCESS)
477 		goto init_adminq_free_rings;
478 
479 	/* success! */
480 	goto init_adminq_exit;
481 
482 init_adminq_free_rings:
483 	i40e_free_adminq_arq(hw);
484 
485 init_adminq_exit:
486 	return ret_code;
487 }
488 
489 /**
490  *  i40e_shutdown_asq - shutdown the ASQ
491  *  @hw: pointer to the hardware structure
492  *
493  *  The main shutdown routine for the Admin Send Queue
494  **/
495 enum i40e_status_code i40e_shutdown_asq(struct i40e_hw *hw)
496 {
497 	enum i40e_status_code ret_code = I40E_SUCCESS;
498 
499 	if (hw->aq.asq.count == 0)
500 		return I40E_ERR_NOT_READY;
501 
502 	/* Stop firmware AdminQ processing */
503 	wr32(hw, hw->aq.asq.head, 0);
504 	wr32(hw, hw->aq.asq.tail, 0);
505 	wr32(hw, hw->aq.asq.len, 0);
506 	wr32(hw, hw->aq.asq.bal, 0);
507 	wr32(hw, hw->aq.asq.bah, 0);
508 
509 	/* make sure spinlock is available */
510 	i40e_acquire_spinlock(&hw->aq.asq_spinlock);
511 
512 	hw->aq.asq.count = 0; /* to indicate uninitialized queue */
513 
514 	/* free ring buffers */
515 	i40e_free_asq_bufs(hw);
516 
517 	i40e_release_spinlock(&hw->aq.asq_spinlock);
518 
519 	return ret_code;
520 }
521 
522 /**
523  *  i40e_shutdown_arq - shutdown ARQ
524  *  @hw: pointer to the hardware structure
525  *
526  *  The main shutdown routine for the Admin Receive Queue
527  **/
528 enum i40e_status_code i40e_shutdown_arq(struct i40e_hw *hw)
529 {
530 	enum i40e_status_code ret_code = I40E_SUCCESS;
531 
532 	if (hw->aq.arq.count == 0)
533 		return I40E_ERR_NOT_READY;
534 
535 	/* Stop firmware AdminQ processing */
536 	wr32(hw, hw->aq.arq.head, 0);
537 	wr32(hw, hw->aq.arq.tail, 0);
538 	wr32(hw, hw->aq.arq.len, 0);
539 	wr32(hw, hw->aq.arq.bal, 0);
540 	wr32(hw, hw->aq.arq.bah, 0);
541 
542 	/* make sure spinlock is available */
543 	i40e_acquire_spinlock(&hw->aq.arq_spinlock);
544 
545 	hw->aq.arq.count = 0; /* to indicate uninitialized queue */
546 
547 	/* free ring buffers */
548 	i40e_free_arq_bufs(hw);
549 
550 	i40e_release_spinlock(&hw->aq.arq_spinlock);
551 
552 	return ret_code;
553 }
554 
555 /**
556  *  i40e_init_adminq - main initialization routine for Admin Queue
557  *  @hw: pointer to the hardware structure
558  *
559  *  Prior to calling this function, drivers *MUST* set the following fields
560  *  in the hw->aq structure:
561  *     - hw->aq.num_asq_entries
562  *     - hw->aq.num_arq_entries
563  *     - hw->aq.arq_buf_size
564  *     - hw->aq.asq_buf_size
565  **/
566 enum i40e_status_code i40e_init_adminq(struct i40e_hw *hw)
567 {
568 	enum i40e_status_code ret_code;
569 	u16 eetrack_lo, eetrack_hi;
570 	u16 cfg_ptr, oem_hi, oem_lo;
571 	int retry = 0;
572 	/* verify input for valid configuration */
573 	if ((hw->aq.num_arq_entries == 0) ||
574 	    (hw->aq.num_asq_entries == 0) ||
575 	    (hw->aq.arq_buf_size == 0) ||
576 	    (hw->aq.asq_buf_size == 0)) {
577 		ret_code = I40E_ERR_CONFIG;
578 		goto init_adminq_exit;
579 	}
580 
581 	/* initialize spin locks */
582 	i40e_init_spinlock(&hw->aq.asq_spinlock);
583 	i40e_init_spinlock(&hw->aq.arq_spinlock);
584 
585 	/* Set up register offsets */
586 	i40e_adminq_init_regs(hw);
587 
588 	/* setup ASQ command write back timeout */
589 	hw->aq.asq_cmd_timeout = I40E_ASQ_CMD_TIMEOUT;
590 
591 	/* allocate the ASQ */
592 	ret_code = i40e_init_asq(hw);
593 	if (ret_code != I40E_SUCCESS)
594 		goto init_adminq_destroy_spinlocks;
595 
596 	/* allocate the ARQ */
597 	ret_code = i40e_init_arq(hw);
598 	if (ret_code != I40E_SUCCESS)
599 		goto init_adminq_free_asq;
600 
601 	/* VF has no need of firmware */
602 	if (i40e_is_vf(hw))
603 		goto init_adminq_exit;
604 	/* There are some cases where the firmware may not be quite ready
605 	 * for AdminQ operations, so we retry the AdminQ setup a few times
606 	 * if we see timeouts in this first AQ call.
607 	 */
608 	do {
609 		ret_code = i40e_aq_get_firmware_version(hw,
610 							&hw->aq.fw_maj_ver,
611 							&hw->aq.fw_min_ver,
612 							&hw->aq.fw_build,
613 							&hw->aq.api_maj_ver,
614 							&hw->aq.api_min_ver,
615 							NULL);
616 		if (ret_code != I40E_ERR_ADMIN_QUEUE_TIMEOUT)
617 			break;
618 		retry++;
619 		i40e_msec_delay(100);
620 		i40e_resume_aq(hw);
621 	} while (retry < 10);
622 	if (ret_code != I40E_SUCCESS)
623 		goto init_adminq_free_arq;
624 
625 	/* get the NVM version info */
626 	i40e_read_nvm_word(hw, I40E_SR_NVM_DEV_STARTER_VERSION,
627 			   &hw->nvm.version);
628 	i40e_read_nvm_word(hw, I40E_SR_NVM_EETRACK_LO, &eetrack_lo);
629 	i40e_read_nvm_word(hw, I40E_SR_NVM_EETRACK_HI, &eetrack_hi);
630 	hw->nvm.eetrack = (eetrack_hi << 16) | eetrack_lo;
631 	i40e_read_nvm_word(hw, I40E_SR_BOOT_CONFIG_PTR, &cfg_ptr);
632 	i40e_read_nvm_word(hw, (cfg_ptr + I40E_NVM_OEM_VER_OFF),
633 			   &oem_hi);
634 	i40e_read_nvm_word(hw, (cfg_ptr + (I40E_NVM_OEM_VER_OFF + 1)),
635 			   &oem_lo);
636 	hw->nvm.oem_ver = ((u32)oem_hi << 16) | oem_lo;
637 
638 	if (hw->aq.api_maj_ver > I40E_FW_API_VERSION_MAJOR) {
639 		ret_code = I40E_ERR_FIRMWARE_API_VERSION;
640 		goto init_adminq_free_arq;
641 	}
642 
643 	/* pre-emptive resource lock release */
644 	i40e_aq_release_resource(hw, I40E_NVM_RESOURCE_ID, 0, NULL);
645 	hw->aq.nvm_release_on_done = FALSE;
646 	hw->nvmupd_state = I40E_NVMUPD_STATE_INIT;
647 
648 	ret_code = i40e_aq_set_hmc_resource_profile(hw,
649 						    I40E_HMC_PROFILE_DEFAULT,
650 						    0,
651 						    NULL);
652 	ret_code = I40E_SUCCESS;
653 
654 	/* success! */
655 	goto init_adminq_exit;
656 
657 init_adminq_free_arq:
658 	i40e_shutdown_arq(hw);
659 init_adminq_free_asq:
660 	i40e_shutdown_asq(hw);
661 init_adminq_destroy_spinlocks:
662 	i40e_destroy_spinlock(&hw->aq.asq_spinlock);
663 	i40e_destroy_spinlock(&hw->aq.arq_spinlock);
664 
665 init_adminq_exit:
666 	return ret_code;
667 }
668 
669 /**
670  *  i40e_shutdown_adminq - shutdown routine for the Admin Queue
671  *  @hw: pointer to the hardware structure
672  **/
673 enum i40e_status_code i40e_shutdown_adminq(struct i40e_hw *hw)
674 {
675 	enum i40e_status_code ret_code = I40E_SUCCESS;
676 
677 	if (i40e_check_asq_alive(hw))
678 		i40e_aq_queue_shutdown(hw, TRUE);
679 
680 	i40e_shutdown_asq(hw);
681 	i40e_shutdown_arq(hw);
682 
683 	/* destroy the spinlocks */
684 	i40e_destroy_spinlock(&hw->aq.asq_spinlock);
685 	i40e_destroy_spinlock(&hw->aq.arq_spinlock);
686 
687 	if (hw->nvm_buff.va)
688 		i40e_free_virt_mem(hw, &hw->nvm_buff);
689 
690 	return ret_code;
691 }
692 
693 /**
694  *  i40e_clean_asq - cleans Admin send queue
695  *  @hw: pointer to the hardware structure
696  *
697  *  returns the number of free desc
698  **/
699 u16 i40e_clean_asq(struct i40e_hw *hw)
700 {
701 	struct i40e_adminq_ring *asq = &(hw->aq.asq);
702 	struct i40e_asq_cmd_details *details;
703 	u16 ntc = asq->next_to_clean;
704 	struct i40e_aq_desc desc_cb;
705 	struct i40e_aq_desc *desc;
706 
707 	desc = I40E_ADMINQ_DESC(*asq, ntc);
708 	details = I40E_ADMINQ_DETAILS(*asq, ntc);
709 
710 	while (rd32(hw, hw->aq.asq.head) != ntc) {
711 		i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
712 			   "ntc %d head %d.\n", ntc, rd32(hw, hw->aq.asq.head));
713 
714 		if (details->callback) {
715 			I40E_ADMINQ_CALLBACK cb_func =
716 					(I40E_ADMINQ_CALLBACK)details->callback;
717 			i40e_memcpy(&desc_cb, desc, sizeof(struct i40e_aq_desc),
718 				    I40E_DMA_TO_DMA);
719 			cb_func(hw, &desc_cb);
720 		}
721 		i40e_memset(desc, 0, sizeof(*desc), I40E_DMA_MEM);
722 		i40e_memset(details, 0, sizeof(*details), I40E_NONDMA_MEM);
723 		ntc++;
724 		if (ntc == asq->count)
725 			ntc = 0;
726 		desc = I40E_ADMINQ_DESC(*asq, ntc);
727 		details = I40E_ADMINQ_DETAILS(*asq, ntc);
728 	}
729 
730 	asq->next_to_clean = ntc;
731 
732 	return I40E_DESC_UNUSED(asq);
733 }
734 
735 /**
736  *  i40e_asq_done - check if FW has processed the Admin Send Queue
737  *  @hw: pointer to the hw struct
738  *
739  *  Returns TRUE if the firmware has processed all descriptors on the
740  *  admin send queue. Returns FALSE if there are still requests pending.
741  **/
742 bool i40e_asq_done(struct i40e_hw *hw)
743 {
744 	/* AQ designers suggest use of head for better
745 	 * timing reliability than DD bit
746 	 */
747 	return rd32(hw, hw->aq.asq.head) == hw->aq.asq.next_to_use;
748 
749 }
750 
751 /**
752  *  i40e_asq_send_command - send command to Admin Queue
753  *  @hw: pointer to the hw struct
754  *  @desc: prefilled descriptor describing the command (non DMA mem)
755  *  @buff: buffer to use for indirect commands
756  *  @buff_size: size of buffer for indirect commands
757  *  @cmd_details: pointer to command details structure
758  *
759  *  This is the main send command driver routine for the Admin Queue send
760  *  queue.  It runs the queue, cleans the queue, etc
761  **/
762 enum i40e_status_code i40e_asq_send_command(struct i40e_hw *hw,
763 				struct i40e_aq_desc *desc,
764 				void *buff, /* can be NULL */
765 				u16  buff_size,
766 				struct i40e_asq_cmd_details *cmd_details)
767 {
768 	enum i40e_status_code status = I40E_SUCCESS;
769 	struct i40e_dma_mem *dma_buff = NULL;
770 	struct i40e_asq_cmd_details *details;
771 	struct i40e_aq_desc *desc_on_ring;
772 	bool cmd_completed = FALSE;
773 	u16  retval = 0;
774 	u32  val = 0;
775 
776 	hw->aq.asq_last_status = I40E_AQ_RC_OK;
777 
778 	val = rd32(hw, hw->aq.asq.head);
779 	if (val >= hw->aq.num_asq_entries) {
780 		i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
781 			   "AQTX: head overrun at %d\n", val);
782 		status = I40E_ERR_QUEUE_EMPTY;
783 		goto asq_send_command_exit;
784 	}
785 
786 	if (hw->aq.asq.count == 0) {
787 		i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
788 			   "AQTX: Admin queue not initialized.\n");
789 		status = I40E_ERR_QUEUE_EMPTY;
790 		goto asq_send_command_exit;
791 	}
792 
793 	details = I40E_ADMINQ_DETAILS(hw->aq.asq, hw->aq.asq.next_to_use);
794 	if (cmd_details) {
795 		i40e_memcpy(details,
796 			    cmd_details,
797 			    sizeof(struct i40e_asq_cmd_details),
798 			    I40E_NONDMA_TO_NONDMA);
799 
800 		/* If the cmd_details are defined copy the cookie.  The
801 		 * CPU_TO_LE32 is not needed here because the data is ignored
802 		 * by the FW, only used by the driver
803 		 */
804 		if (details->cookie) {
805 			desc->cookie_high =
806 				CPU_TO_LE32(I40E_HI_DWORD(details->cookie));
807 			desc->cookie_low =
808 				CPU_TO_LE32(I40E_LO_DWORD(details->cookie));
809 		}
810 	} else {
811 		i40e_memset(details, 0,
812 			    sizeof(struct i40e_asq_cmd_details),
813 			    I40E_NONDMA_MEM);
814 	}
815 
816 	/* clear requested flags and then set additional flags if defined */
817 	desc->flags &= ~CPU_TO_LE16(details->flags_dis);
818 	desc->flags |= CPU_TO_LE16(details->flags_ena);
819 
820 	i40e_acquire_spinlock(&hw->aq.asq_spinlock);
821 
822 	if (buff_size > hw->aq.asq_buf_size) {
823 		i40e_debug(hw,
824 			   I40E_DEBUG_AQ_MESSAGE,
825 			   "AQTX: Invalid buffer size: %d.\n",
826 			   buff_size);
827 		status = I40E_ERR_INVALID_SIZE;
828 		goto asq_send_command_error;
829 	}
830 
831 	if (details->postpone && !details->async) {
832 		i40e_debug(hw,
833 			   I40E_DEBUG_AQ_MESSAGE,
834 			   "AQTX: Async flag not set along with postpone flag");
835 		status = I40E_ERR_PARAM;
836 		goto asq_send_command_error;
837 	}
838 
839 	/* call clean and check queue available function to reclaim the
840 	 * descriptors that were processed by FW, the function returns the
841 	 * number of desc available
842 	 */
843 	/* the clean function called here could be called in a separate thread
844 	 * in case of asynchronous completions
845 	 */
846 	if (i40e_clean_asq(hw) == 0) {
847 		i40e_debug(hw,
848 			   I40E_DEBUG_AQ_MESSAGE,
849 			   "AQTX: Error queue is full.\n");
850 		status = I40E_ERR_ADMIN_QUEUE_FULL;
851 		goto asq_send_command_error;
852 	}
853 
854 	/* initialize the temp desc pointer with the right desc */
855 	desc_on_ring = I40E_ADMINQ_DESC(hw->aq.asq, hw->aq.asq.next_to_use);
856 
857 	/* if the desc is available copy the temp desc to the right place */
858 	i40e_memcpy(desc_on_ring, desc, sizeof(struct i40e_aq_desc),
859 		    I40E_NONDMA_TO_DMA);
860 
861 	/* if buff is not NULL assume indirect command */
862 	if (buff != NULL) {
863 		dma_buff = &(hw->aq.asq.r.asq_bi[hw->aq.asq.next_to_use]);
864 		/* copy the user buff into the respective DMA buff */
865 		i40e_memcpy(dma_buff->va, buff, buff_size,
866 			    I40E_NONDMA_TO_DMA);
867 		desc_on_ring->datalen = CPU_TO_LE16(buff_size);
868 
869 		/* Update the address values in the desc with the pa value
870 		 * for respective buffer
871 		 */
872 		desc_on_ring->params.external.addr_high =
873 				CPU_TO_LE32(I40E_HI_DWORD(dma_buff->pa));
874 		desc_on_ring->params.external.addr_low =
875 				CPU_TO_LE32(I40E_LO_DWORD(dma_buff->pa));
876 	}
877 
878 	/* bump the tail */
879 	i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE, "AQTX: desc and buffer:\n");
880 	i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc_on_ring,
881 		      buff, buff_size);
882 	(hw->aq.asq.next_to_use)++;
883 	if (hw->aq.asq.next_to_use == hw->aq.asq.count)
884 		hw->aq.asq.next_to_use = 0;
885 	if (!details->postpone)
886 		wr32(hw, hw->aq.asq.tail, hw->aq.asq.next_to_use);
887 
888 	/* if cmd_details are not defined or async flag is not set,
889 	 * we need to wait for desc write back
890 	 */
891 	if (!details->async && !details->postpone) {
892 		u32 total_delay = 0;
893 
894 		do {
895 			/* AQ designers suggest use of head for better
896 			 * timing reliability than DD bit
897 			 */
898 			if (i40e_asq_done(hw))
899 				break;
900 			/* ugh! delay while spin_lock */
901 			i40e_msec_delay(1);
902 			total_delay++;
903 		} while (total_delay < hw->aq.asq_cmd_timeout);
904 	}
905 
906 	/* if ready, copy the desc back to temp */
907 	if (i40e_asq_done(hw)) {
908 		i40e_memcpy(desc, desc_on_ring, sizeof(struct i40e_aq_desc),
909 			    I40E_DMA_TO_NONDMA);
910 		if (buff != NULL)
911 			i40e_memcpy(buff, dma_buff->va, buff_size,
912 				    I40E_DMA_TO_NONDMA);
913 		retval = LE16_TO_CPU(desc->retval);
914 		if (retval != 0) {
915 			i40e_debug(hw,
916 				   I40E_DEBUG_AQ_MESSAGE,
917 				   "AQTX: Command completed with error 0x%X.\n",
918 				   retval);
919 
920 			/* strip off FW internal code */
921 			retval &= 0xff;
922 		}
923 		cmd_completed = TRUE;
924 		if ((enum i40e_admin_queue_err)retval == I40E_AQ_RC_OK)
925 			status = I40E_SUCCESS;
926 		else
927 			status = I40E_ERR_ADMIN_QUEUE_ERROR;
928 		hw->aq.asq_last_status = (enum i40e_admin_queue_err)retval;
929 	}
930 
931 	i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
932 		   "AQTX: desc and buffer writeback:\n");
933 	i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc, buff, buff_size);
934 
935 	/* save writeback aq if requested */
936 	if (details->wb_desc)
937 		i40e_memcpy(details->wb_desc, desc_on_ring,
938 			    sizeof(struct i40e_aq_desc), I40E_DMA_TO_NONDMA);
939 
940 	/* update the error if time out occurred */
941 	if ((!cmd_completed) &&
942 	    (!details->async && !details->postpone)) {
943 		i40e_debug(hw,
944 			   I40E_DEBUG_AQ_MESSAGE,
945 			   "AQTX: Writeback timeout.\n");
946 		status = I40E_ERR_ADMIN_QUEUE_TIMEOUT;
947 	}
948 
949 asq_send_command_error:
950 	i40e_release_spinlock(&hw->aq.asq_spinlock);
951 asq_send_command_exit:
952 	return status;
953 }
954 
955 /**
956  *  i40e_fill_default_direct_cmd_desc - AQ descriptor helper function
957  *  @desc:     pointer to the temp descriptor (non DMA mem)
958  *  @opcode:   the opcode can be used to decide which flags to turn off or on
959  *
960  *  Fill the desc with default values
961  **/
962 void i40e_fill_default_direct_cmd_desc(struct i40e_aq_desc *desc,
963 				       u16 opcode)
964 {
965 	/* zero out the desc */
966 	i40e_memset((void *)desc, 0, sizeof(struct i40e_aq_desc),
967 		    I40E_NONDMA_MEM);
968 	desc->opcode = CPU_TO_LE16(opcode);
969 	desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_SI);
970 }
971 
972 /**
973  *  i40e_clean_arq_element
974  *  @hw: pointer to the hw struct
975  *  @e: event info from the receive descriptor, includes any buffers
976  *  @pending: number of events that could be left to process
977  *
978  *  This function cleans one Admin Receive Queue element and returns
979  *  the contents through e.  It can also return how many events are
980  *  left to process through 'pending'
981  **/
982 enum i40e_status_code i40e_clean_arq_element(struct i40e_hw *hw,
983 					     struct i40e_arq_event_info *e,
984 					     u16 *pending)
985 {
986 	enum i40e_status_code ret_code = I40E_SUCCESS;
987 	u16 ntc = hw->aq.arq.next_to_clean;
988 	struct i40e_aq_desc *desc;
989 	struct i40e_dma_mem *bi;
990 	u16 desc_idx;
991 	u16 datalen;
992 	u16 flags;
993 	u16 ntu;
994 
995 	/* take the lock before we start messing with the ring */
996 	i40e_acquire_spinlock(&hw->aq.arq_spinlock);
997 
998 	/* set next_to_use to head */
999 	if (!i40e_is_vf(hw))
1000 		ntu = (rd32(hw, hw->aq.arq.head) & I40E_PF_ARQH_ARQH_MASK);
1001 	if (i40e_is_vf(hw))
1002 		ntu = (rd32(hw, hw->aq.arq.head) & I40E_VF_ARQH1_ARQH_MASK);
1003 	if (ntu == ntc) {
1004 		/* nothing to do - shouldn't need to update ring's values */
1005 		ret_code = I40E_ERR_ADMIN_QUEUE_NO_WORK;
1006 		goto clean_arq_element_out;
1007 	}
1008 
1009 	/* now clean the next descriptor */
1010 	desc = I40E_ADMINQ_DESC(hw->aq.arq, ntc);
1011 	desc_idx = ntc;
1012 
1013 	flags = LE16_TO_CPU(desc->flags);
1014 	if (flags & I40E_AQ_FLAG_ERR) {
1015 		ret_code = I40E_ERR_ADMIN_QUEUE_ERROR;
1016 		hw->aq.arq_last_status =
1017 			(enum i40e_admin_queue_err)LE16_TO_CPU(desc->retval);
1018 		i40e_debug(hw,
1019 			   I40E_DEBUG_AQ_MESSAGE,
1020 			   "AQRX: Event received with error 0x%X.\n",
1021 			   hw->aq.arq_last_status);
1022 	}
1023 
1024 	i40e_memcpy(&e->desc, desc, sizeof(struct i40e_aq_desc),
1025 		    I40E_DMA_TO_NONDMA);
1026 	datalen = LE16_TO_CPU(desc->datalen);
1027 	e->msg_len = min(datalen, e->buf_len);
1028 	if (e->msg_buf != NULL && (e->msg_len != 0))
1029 		i40e_memcpy(e->msg_buf,
1030 			    hw->aq.arq.r.arq_bi[desc_idx].va,
1031 			    e->msg_len, I40E_DMA_TO_NONDMA);
1032 
1033 	i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE, "AQRX: desc and buffer:\n");
1034 	i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc, e->msg_buf,
1035 		      hw->aq.arq_buf_size);
1036 
1037 	/* Restore the original datalen and buffer address in the desc,
1038 	 * FW updates datalen to indicate the event message
1039 	 * size
1040 	 */
1041 	bi = &hw->aq.arq.r.arq_bi[ntc];
1042 	i40e_memset((void *)desc, 0, sizeof(struct i40e_aq_desc), I40E_DMA_MEM);
1043 
1044 	desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_BUF);
1045 	if (hw->aq.arq_buf_size > I40E_AQ_LARGE_BUF)
1046 		desc->flags |= CPU_TO_LE16(I40E_AQ_FLAG_LB);
1047 	desc->datalen = CPU_TO_LE16((u16)bi->size);
1048 	desc->params.external.addr_high = CPU_TO_LE32(I40E_HI_DWORD(bi->pa));
1049 	desc->params.external.addr_low = CPU_TO_LE32(I40E_LO_DWORD(bi->pa));
1050 
1051 	/* set tail = the last cleaned desc index. */
1052 	wr32(hw, hw->aq.arq.tail, ntc);
1053 	/* ntc is updated to tail + 1 */
1054 	ntc++;
1055 	if (ntc == hw->aq.num_arq_entries)
1056 		ntc = 0;
1057 	hw->aq.arq.next_to_clean = ntc;
1058 	hw->aq.arq.next_to_use = ntu;
1059 
1060 clean_arq_element_out:
1061 	/* Set pending if needed, unlock and return */
1062 	if (pending != NULL)
1063 		*pending = (ntc > ntu ? hw->aq.arq.count : 0) + (ntu - ntc);
1064 	i40e_release_spinlock(&hw->aq.arq_spinlock);
1065 
1066 	if (i40e_is_nvm_update_op(&e->desc)) {
1067 		if (hw->aq.nvm_release_on_done) {
1068 			i40e_release_nvm(hw);
1069 			hw->aq.nvm_release_on_done = FALSE;
1070 		}
1071 
1072 		switch (hw->nvmupd_state) {
1073 		case I40E_NVMUPD_STATE_INIT_WAIT:
1074 			hw->nvmupd_state = I40E_NVMUPD_STATE_INIT;
1075 			break;
1076 
1077 		case I40E_NVMUPD_STATE_WRITE_WAIT:
1078 			hw->nvmupd_state = I40E_NVMUPD_STATE_WRITING;
1079 			break;
1080 
1081 		default:
1082 			break;
1083 		}
1084 	}
1085 
1086 	return ret_code;
1087 }
1088 
1089 void i40e_resume_aq(struct i40e_hw *hw)
1090 {
1091 	/* Registers are reset after PF reset */
1092 	hw->aq.asq.next_to_use = 0;
1093 	hw->aq.asq.next_to_clean = 0;
1094 
1095 	i40e_config_asq_regs(hw);
1096 
1097 	hw->aq.arq.next_to_use = 0;
1098 	hw->aq.arq.next_to_clean = 0;
1099 
1100 	i40e_config_arq_regs(hw);
1101 }
1102