xref: /freebsd/sys/dev/ixl/i40e_adminq.c (revision c6db8143eda5c775467145ac73e8ebec47afdd8f)
1 /******************************************************************************
2 
3   Copyright (c) 2013-2014, Intel Corporation
4   All rights reserved.
5 
6   Redistribution and use in source and binary forms, with or without
7   modification, are permitted provided that the following conditions are met:
8 
9    1. Redistributions of source code must retain the above copyright notice,
10       this list of conditions and the following disclaimer.
11 
12    2. Redistributions in binary form must reproduce the above copyright
13       notice, this list of conditions and the following disclaimer in the
14       documentation and/or other materials provided with the distribution.
15 
16    3. Neither the name of the Intel Corporation nor the names of its
17       contributors may be used to endorse or promote products derived from
18       this software without specific prior written permission.
19 
20   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
21   AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22   IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23   ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
24   LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25   CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26   SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27   INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28   CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29   ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30   POSSIBILITY OF SUCH DAMAGE.
31 
32 ******************************************************************************/
33 /*$FreeBSD$*/
34 
35 #include "i40e_status.h"
36 #include "i40e_type.h"
37 #include "i40e_register.h"
38 #include "i40e_adminq.h"
39 #include "i40e_prototype.h"
40 
41 /**
42  * i40e_is_nvm_update_op - return TRUE if this is an NVM update operation
43  * @desc: API request descriptor
44  **/
45 static INLINE bool i40e_is_nvm_update_op(struct i40e_aq_desc *desc)
46 {
47 	return (desc->opcode == CPU_TO_LE16(i40e_aqc_opc_nvm_erase) ||
48 		desc->opcode == CPU_TO_LE16(i40e_aqc_opc_nvm_update));
49 }
50 
51 /**
52  *  i40e_adminq_init_regs - Initialize AdminQ registers
53  *  @hw: pointer to the hardware structure
54  *
55  *  This assumes the alloc_asq and alloc_arq functions have already been called
56  **/
57 static void i40e_adminq_init_regs(struct i40e_hw *hw)
58 {
59 	/* set head and tail registers in our local struct */
60 	if (i40e_is_vf(hw)) {
61 		hw->aq.asq.tail = I40E_VF_ATQT1;
62 		hw->aq.asq.head = I40E_VF_ATQH1;
63 		hw->aq.asq.len  = I40E_VF_ATQLEN1;
64 		hw->aq.asq.bal  = I40E_VF_ATQBAL1;
65 		hw->aq.asq.bah  = I40E_VF_ATQBAH1;
66 		hw->aq.arq.tail = I40E_VF_ARQT1;
67 		hw->aq.arq.head = I40E_VF_ARQH1;
68 		hw->aq.arq.len  = I40E_VF_ARQLEN1;
69 		hw->aq.arq.bal  = I40E_VF_ARQBAL1;
70 		hw->aq.arq.bah  = I40E_VF_ARQBAH1;
71 	} else {
72 		hw->aq.asq.tail = I40E_PF_ATQT;
73 		hw->aq.asq.head = I40E_PF_ATQH;
74 		hw->aq.asq.len  = I40E_PF_ATQLEN;
75 		hw->aq.asq.bal  = I40E_PF_ATQBAL;
76 		hw->aq.asq.bah  = I40E_PF_ATQBAH;
77 		hw->aq.arq.tail = I40E_PF_ARQT;
78 		hw->aq.arq.head = I40E_PF_ARQH;
79 		hw->aq.arq.len  = I40E_PF_ARQLEN;
80 		hw->aq.arq.bal  = I40E_PF_ARQBAL;
81 		hw->aq.arq.bah  = I40E_PF_ARQBAH;
82 	}
83 }
84 
85 /**
86  *  i40e_alloc_adminq_asq_ring - Allocate Admin Queue send rings
87  *  @hw: pointer to the hardware structure
88  **/
89 enum i40e_status_code i40e_alloc_adminq_asq_ring(struct i40e_hw *hw)
90 {
91 	enum i40e_status_code ret_code;
92 
93 	ret_code = i40e_allocate_dma_mem(hw, &hw->aq.asq.desc_buf,
94 					 i40e_mem_atq_ring,
95 					 (hw->aq.num_asq_entries *
96 					 sizeof(struct i40e_aq_desc)),
97 					 I40E_ADMINQ_DESC_ALIGNMENT);
98 	if (ret_code)
99 		return ret_code;
100 
101 	ret_code = i40e_allocate_virt_mem(hw, &hw->aq.asq.cmd_buf,
102 					  (hw->aq.num_asq_entries *
103 					  sizeof(struct i40e_asq_cmd_details)));
104 	if (ret_code) {
105 		i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
106 		return ret_code;
107 	}
108 
109 	return ret_code;
110 }
111 
112 /**
113  *  i40e_alloc_adminq_arq_ring - Allocate Admin Queue receive rings
114  *  @hw: pointer to the hardware structure
115  **/
116 enum i40e_status_code i40e_alloc_adminq_arq_ring(struct i40e_hw *hw)
117 {
118 	enum i40e_status_code ret_code;
119 
120 	ret_code = i40e_allocate_dma_mem(hw, &hw->aq.arq.desc_buf,
121 					 i40e_mem_arq_ring,
122 					 (hw->aq.num_arq_entries *
123 					 sizeof(struct i40e_aq_desc)),
124 					 I40E_ADMINQ_DESC_ALIGNMENT);
125 
126 	return ret_code;
127 }
128 
129 /**
130  *  i40e_free_adminq_asq - Free Admin Queue send rings
131  *  @hw: pointer to the hardware structure
132  *
133  *  This assumes the posted send buffers have already been cleaned
134  *  and de-allocated
135  **/
136 void i40e_free_adminq_asq(struct i40e_hw *hw)
137 {
138 	i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
139 }
140 
141 /**
142  *  i40e_free_adminq_arq - Free Admin Queue receive rings
143  *  @hw: pointer to the hardware structure
144  *
145  *  This assumes the posted receive buffers have already been cleaned
146  *  and de-allocated
147  **/
148 void i40e_free_adminq_arq(struct i40e_hw *hw)
149 {
150 	i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf);
151 }
152 
153 /**
154  *  i40e_alloc_arq_bufs - Allocate pre-posted buffers for the receive queue
155  *  @hw: pointer to the hardware structure
156  **/
157 static enum i40e_status_code i40e_alloc_arq_bufs(struct i40e_hw *hw)
158 {
159 	enum i40e_status_code ret_code;
160 	struct i40e_aq_desc *desc;
161 	struct i40e_dma_mem *bi;
162 	int i;
163 
164 	/* We'll be allocating the buffer info memory first, then we can
165 	 * allocate the mapped buffers for the event processing
166 	 */
167 
168 	/* buffer_info structures do not need alignment */
169 	ret_code = i40e_allocate_virt_mem(hw, &hw->aq.arq.dma_head,
170 		(hw->aq.num_arq_entries * sizeof(struct i40e_dma_mem)));
171 	if (ret_code)
172 		goto alloc_arq_bufs;
173 	hw->aq.arq.r.arq_bi = (struct i40e_dma_mem *)hw->aq.arq.dma_head.va;
174 
175 	/* allocate the mapped buffers */
176 	for (i = 0; i < hw->aq.num_arq_entries; i++) {
177 		bi = &hw->aq.arq.r.arq_bi[i];
178 		ret_code = i40e_allocate_dma_mem(hw, bi,
179 						 i40e_mem_arq_buf,
180 						 hw->aq.arq_buf_size,
181 						 I40E_ADMINQ_DESC_ALIGNMENT);
182 		if (ret_code)
183 			goto unwind_alloc_arq_bufs;
184 
185 		/* now configure the descriptors for use */
186 		desc = I40E_ADMINQ_DESC(hw->aq.arq, i);
187 
188 		desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_BUF);
189 		if (hw->aq.arq_buf_size > I40E_AQ_LARGE_BUF)
190 			desc->flags |= CPU_TO_LE16(I40E_AQ_FLAG_LB);
191 		desc->opcode = 0;
192 		/* This is in accordance with Admin queue design, there is no
193 		 * register for buffer size configuration
194 		 */
195 		desc->datalen = CPU_TO_LE16((u16)bi->size);
196 		desc->retval = 0;
197 		desc->cookie_high = 0;
198 		desc->cookie_low = 0;
199 		desc->params.external.addr_high =
200 			CPU_TO_LE32(I40E_HI_DWORD(bi->pa));
201 		desc->params.external.addr_low =
202 			CPU_TO_LE32(I40E_LO_DWORD(bi->pa));
203 		desc->params.external.param0 = 0;
204 		desc->params.external.param1 = 0;
205 	}
206 
207 alloc_arq_bufs:
208 	return ret_code;
209 
210 unwind_alloc_arq_bufs:
211 	/* don't try to free the one that failed... */
212 	i--;
213 	for (; i >= 0; i--)
214 		i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]);
215 	i40e_free_virt_mem(hw, &hw->aq.arq.dma_head);
216 
217 	return ret_code;
218 }
219 
220 /**
221  *  i40e_alloc_asq_bufs - Allocate empty buffer structs for the send queue
222  *  @hw: pointer to the hardware structure
223  **/
224 static enum i40e_status_code i40e_alloc_asq_bufs(struct i40e_hw *hw)
225 {
226 	enum i40e_status_code ret_code;
227 	struct i40e_dma_mem *bi;
228 	int i;
229 
230 	/* No mapped memory needed yet, just the buffer info structures */
231 	ret_code = i40e_allocate_virt_mem(hw, &hw->aq.asq.dma_head,
232 		(hw->aq.num_asq_entries * sizeof(struct i40e_dma_mem)));
233 	if (ret_code)
234 		goto alloc_asq_bufs;
235 	hw->aq.asq.r.asq_bi = (struct i40e_dma_mem *)hw->aq.asq.dma_head.va;
236 
237 	/* allocate the mapped buffers */
238 	for (i = 0; i < hw->aq.num_asq_entries; i++) {
239 		bi = &hw->aq.asq.r.asq_bi[i];
240 		ret_code = i40e_allocate_dma_mem(hw, bi,
241 						 i40e_mem_asq_buf,
242 						 hw->aq.asq_buf_size,
243 						 I40E_ADMINQ_DESC_ALIGNMENT);
244 		if (ret_code)
245 			goto unwind_alloc_asq_bufs;
246 	}
247 alloc_asq_bufs:
248 	return ret_code;
249 
250 unwind_alloc_asq_bufs:
251 	/* don't try to free the one that failed... */
252 	i--;
253 	for (; i >= 0; i--)
254 		i40e_free_dma_mem(hw, &hw->aq.asq.r.asq_bi[i]);
255 	i40e_free_virt_mem(hw, &hw->aq.asq.dma_head);
256 
257 	return ret_code;
258 }
259 
260 /**
261  *  i40e_free_arq_bufs - Free receive queue buffer info elements
262  *  @hw: pointer to the hardware structure
263  **/
264 static void i40e_free_arq_bufs(struct i40e_hw *hw)
265 {
266 	int i;
267 
268 	/* free descriptors */
269 	for (i = 0; i < hw->aq.num_arq_entries; i++)
270 		i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]);
271 
272 	/* free the descriptor memory */
273 	i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf);
274 
275 	/* free the dma header */
276 	i40e_free_virt_mem(hw, &hw->aq.arq.dma_head);
277 }
278 
279 /**
280  *  i40e_free_asq_bufs - Free send queue buffer info elements
281  *  @hw: pointer to the hardware structure
282  **/
283 static void i40e_free_asq_bufs(struct i40e_hw *hw)
284 {
285 	int i;
286 
287 	/* only unmap if the address is non-NULL */
288 	for (i = 0; i < hw->aq.num_asq_entries; i++)
289 		if (hw->aq.asq.r.asq_bi[i].pa)
290 			i40e_free_dma_mem(hw, &hw->aq.asq.r.asq_bi[i]);
291 
292 	/* free the buffer info list */
293 	i40e_free_virt_mem(hw, &hw->aq.asq.cmd_buf);
294 
295 	/* free the descriptor memory */
296 	i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
297 
298 	/* free the dma header */
299 	i40e_free_virt_mem(hw, &hw->aq.asq.dma_head);
300 }
301 
302 /**
303  *  i40e_config_asq_regs - configure ASQ registers
304  *  @hw: pointer to the hardware structure
305  *
306  *  Configure base address and length registers for the transmit queue
307  **/
308 static enum i40e_status_code i40e_config_asq_regs(struct i40e_hw *hw)
309 {
310 	enum i40e_status_code ret_code = I40E_SUCCESS;
311 	u32 reg = 0;
312 
313 	/* Clear Head and Tail */
314 	wr32(hw, hw->aq.asq.head, 0);
315 	wr32(hw, hw->aq.asq.tail, 0);
316 
317 	/* set starting point */
318 	wr32(hw, hw->aq.asq.len, (hw->aq.num_asq_entries |
319 				  I40E_PF_ATQLEN_ATQENABLE_MASK));
320 	wr32(hw, hw->aq.asq.bal, I40E_LO_DWORD(hw->aq.asq.desc_buf.pa));
321 	wr32(hw, hw->aq.asq.bah, I40E_HI_DWORD(hw->aq.asq.desc_buf.pa));
322 
323 	/* Check one register to verify that config was applied */
324 	reg = rd32(hw, hw->aq.asq.bal);
325 	if (reg != I40E_LO_DWORD(hw->aq.asq.desc_buf.pa))
326 		ret_code = I40E_ERR_ADMIN_QUEUE_ERROR;
327 
328 	return ret_code;
329 }
330 
331 /**
332  *  i40e_config_arq_regs - ARQ register configuration
333  *  @hw: pointer to the hardware structure
334  *
335  * Configure base address and length registers for the receive (event queue)
336  **/
337 static enum i40e_status_code i40e_config_arq_regs(struct i40e_hw *hw)
338 {
339 	enum i40e_status_code ret_code = I40E_SUCCESS;
340 	u32 reg = 0;
341 
342 	/* Clear Head and Tail */
343 	wr32(hw, hw->aq.arq.head, 0);
344 	wr32(hw, hw->aq.arq.tail, 0);
345 
346 	/* set starting point */
347 	wr32(hw, hw->aq.arq.len, (hw->aq.num_arq_entries |
348 				  I40E_PF_ARQLEN_ARQENABLE_MASK));
349 	wr32(hw, hw->aq.arq.bal, I40E_LO_DWORD(hw->aq.arq.desc_buf.pa));
350 	wr32(hw, hw->aq.arq.bah, I40E_HI_DWORD(hw->aq.arq.desc_buf.pa));
351 
352 	/* Update tail in the HW to post pre-allocated buffers */
353 	wr32(hw, hw->aq.arq.tail, hw->aq.num_arq_entries - 1);
354 
355 	/* Check one register to verify that config was applied */
356 	reg = rd32(hw, hw->aq.arq.bal);
357 	if (reg != I40E_LO_DWORD(hw->aq.arq.desc_buf.pa))
358 		ret_code = I40E_ERR_ADMIN_QUEUE_ERROR;
359 
360 	return ret_code;
361 }
362 
363 /**
364  *  i40e_init_asq - main initialization routine for ASQ
365  *  @hw: pointer to the hardware structure
366  *
367  *  This is the main initialization routine for the Admin Send Queue
368  *  Prior to calling this function, drivers *MUST* set the following fields
369  *  in the hw->aq structure:
370  *     - hw->aq.num_asq_entries
371  *     - hw->aq.arq_buf_size
372  *
373  *  Do *NOT* hold the lock when calling this as the memory allocation routines
374  *  called are not going to be atomic context safe
375  **/
376 enum i40e_status_code i40e_init_asq(struct i40e_hw *hw)
377 {
378 	enum i40e_status_code ret_code = I40E_SUCCESS;
379 
380 	if (hw->aq.asq.count > 0) {
381 		/* queue already initialized */
382 		ret_code = I40E_ERR_NOT_READY;
383 		goto init_adminq_exit;
384 	}
385 
386 	/* verify input for valid configuration */
387 	if ((hw->aq.num_asq_entries == 0) ||
388 	    (hw->aq.asq_buf_size == 0)) {
389 		ret_code = I40E_ERR_CONFIG;
390 		goto init_adminq_exit;
391 	}
392 
393 	hw->aq.asq.next_to_use = 0;
394 	hw->aq.asq.next_to_clean = 0;
395 	hw->aq.asq.count = hw->aq.num_asq_entries;
396 
397 	/* allocate the ring memory */
398 	ret_code = i40e_alloc_adminq_asq_ring(hw);
399 	if (ret_code != I40E_SUCCESS)
400 		goto init_adminq_exit;
401 
402 	/* allocate buffers in the rings */
403 	ret_code = i40e_alloc_asq_bufs(hw);
404 	if (ret_code != I40E_SUCCESS)
405 		goto init_adminq_free_rings;
406 
407 	/* initialize base registers */
408 	ret_code = i40e_config_asq_regs(hw);
409 	if (ret_code != I40E_SUCCESS)
410 		goto init_adminq_free_rings;
411 
412 	/* success! */
413 	goto init_adminq_exit;
414 
415 init_adminq_free_rings:
416 	i40e_free_adminq_asq(hw);
417 
418 init_adminq_exit:
419 	return ret_code;
420 }
421 
422 /**
423  *  i40e_init_arq - initialize ARQ
424  *  @hw: pointer to the hardware structure
425  *
426  *  The main initialization routine for the Admin Receive (Event) Queue.
427  *  Prior to calling this function, drivers *MUST* set the following fields
428  *  in the hw->aq structure:
429  *     - hw->aq.num_asq_entries
430  *     - hw->aq.arq_buf_size
431  *
432  *  Do *NOT* hold the lock when calling this as the memory allocation routines
433  *  called are not going to be atomic context safe
434  **/
435 enum i40e_status_code i40e_init_arq(struct i40e_hw *hw)
436 {
437 	enum i40e_status_code ret_code = I40E_SUCCESS;
438 
439 	if (hw->aq.arq.count > 0) {
440 		/* queue already initialized */
441 		ret_code = I40E_ERR_NOT_READY;
442 		goto init_adminq_exit;
443 	}
444 
445 	/* verify input for valid configuration */
446 	if ((hw->aq.num_arq_entries == 0) ||
447 	    (hw->aq.arq_buf_size == 0)) {
448 		ret_code = I40E_ERR_CONFIG;
449 		goto init_adminq_exit;
450 	}
451 
452 	hw->aq.arq.next_to_use = 0;
453 	hw->aq.arq.next_to_clean = 0;
454 	hw->aq.arq.count = hw->aq.num_arq_entries;
455 
456 	/* allocate the ring memory */
457 	ret_code = i40e_alloc_adminq_arq_ring(hw);
458 	if (ret_code != I40E_SUCCESS)
459 		goto init_adminq_exit;
460 
461 	/* allocate buffers in the rings */
462 	ret_code = i40e_alloc_arq_bufs(hw);
463 	if (ret_code != I40E_SUCCESS)
464 		goto init_adminq_free_rings;
465 
466 	/* initialize base registers */
467 	ret_code = i40e_config_arq_regs(hw);
468 	if (ret_code != I40E_SUCCESS)
469 		goto init_adminq_free_rings;
470 
471 	/* success! */
472 	goto init_adminq_exit;
473 
474 init_adminq_free_rings:
475 	i40e_free_adminq_arq(hw);
476 
477 init_adminq_exit:
478 	return ret_code;
479 }
480 
481 /**
482  *  i40e_shutdown_asq - shutdown the ASQ
483  *  @hw: pointer to the hardware structure
484  *
485  *  The main shutdown routine for the Admin Send Queue
486  **/
487 enum i40e_status_code i40e_shutdown_asq(struct i40e_hw *hw)
488 {
489 	enum i40e_status_code ret_code = I40E_SUCCESS;
490 
491 	if (hw->aq.asq.count == 0)
492 		return I40E_ERR_NOT_READY;
493 
494 	/* Stop firmware AdminQ processing */
495 	wr32(hw, hw->aq.asq.head, 0);
496 	wr32(hw, hw->aq.asq.tail, 0);
497 	wr32(hw, hw->aq.asq.len, 0);
498 	wr32(hw, hw->aq.asq.bal, 0);
499 	wr32(hw, hw->aq.asq.bah, 0);
500 
501 	/* make sure spinlock is available */
502 	i40e_acquire_spinlock(&hw->aq.asq_spinlock);
503 
504 	hw->aq.asq.count = 0; /* to indicate uninitialized queue */
505 
506 	/* free ring buffers */
507 	i40e_free_asq_bufs(hw);
508 
509 	i40e_release_spinlock(&hw->aq.asq_spinlock);
510 
511 	return ret_code;
512 }
513 
514 /**
515  *  i40e_shutdown_arq - shutdown ARQ
516  *  @hw: pointer to the hardware structure
517  *
518  *  The main shutdown routine for the Admin Receive Queue
519  **/
520 enum i40e_status_code i40e_shutdown_arq(struct i40e_hw *hw)
521 {
522 	enum i40e_status_code ret_code = I40E_SUCCESS;
523 
524 	if (hw->aq.arq.count == 0)
525 		return I40E_ERR_NOT_READY;
526 
527 	/* Stop firmware AdminQ processing */
528 	wr32(hw, hw->aq.arq.head, 0);
529 	wr32(hw, hw->aq.arq.tail, 0);
530 	wr32(hw, hw->aq.arq.len, 0);
531 	wr32(hw, hw->aq.arq.bal, 0);
532 	wr32(hw, hw->aq.arq.bah, 0);
533 
534 	/* make sure spinlock is available */
535 	i40e_acquire_spinlock(&hw->aq.arq_spinlock);
536 
537 	hw->aq.arq.count = 0; /* to indicate uninitialized queue */
538 
539 	/* free ring buffers */
540 	i40e_free_arq_bufs(hw);
541 
542 	i40e_release_spinlock(&hw->aq.arq_spinlock);
543 
544 	return ret_code;
545 }
546 
547 /**
548  *  i40e_init_adminq - main initialization routine for Admin Queue
549  *  @hw: pointer to the hardware structure
550  *
551  *  Prior to calling this function, drivers *MUST* set the following fields
552  *  in the hw->aq structure:
553  *     - hw->aq.num_asq_entries
554  *     - hw->aq.num_arq_entries
555  *     - hw->aq.arq_buf_size
556  *     - hw->aq.asq_buf_size
557  **/
558 enum i40e_status_code i40e_init_adminq(struct i40e_hw *hw)
559 {
560 	enum i40e_status_code ret_code;
561 	u16 eetrack_lo, eetrack_hi;
562 	int retry = 0;
563 	/* verify input for valid configuration */
564 	if ((hw->aq.num_arq_entries == 0) ||
565 	    (hw->aq.num_asq_entries == 0) ||
566 	    (hw->aq.arq_buf_size == 0) ||
567 	    (hw->aq.asq_buf_size == 0)) {
568 		ret_code = I40E_ERR_CONFIG;
569 		goto init_adminq_exit;
570 	}
571 
572 	/* initialize spin locks */
573 	i40e_init_spinlock(&hw->aq.asq_spinlock);
574 	i40e_init_spinlock(&hw->aq.arq_spinlock);
575 
576 	/* Set up register offsets */
577 	i40e_adminq_init_regs(hw);
578 
579 	/* setup ASQ command write back timeout */
580 	hw->aq.asq_cmd_timeout = I40E_ASQ_CMD_TIMEOUT;
581 
582 	/* allocate the ASQ */
583 	ret_code = i40e_init_asq(hw);
584 	if (ret_code != I40E_SUCCESS)
585 		goto init_adminq_destroy_spinlocks;
586 
587 	/* allocate the ARQ */
588 	ret_code = i40e_init_arq(hw);
589 	if (ret_code != I40E_SUCCESS)
590 		goto init_adminq_free_asq;
591 
592         if (i40e_is_vf(hw))  /* VF has no need of firmware */
593                 goto init_adminq_exit;
594 
595 /* There are some cases where the firmware may not be quite ready
596 	 * for AdminQ operations, so we retry the AdminQ setup a few times
597 	 * if we see timeouts in this first AQ call.
598 	 */
599 	do {
600 		ret_code = i40e_aq_get_firmware_version(hw,
601 							&hw->aq.fw_maj_ver,
602 							&hw->aq.fw_min_ver,
603 							&hw->aq.api_maj_ver,
604 							&hw->aq.api_min_ver,
605 							NULL);
606 		if (ret_code != I40E_ERR_ADMIN_QUEUE_TIMEOUT)
607 			break;
608 		retry++;
609 		i40e_msec_delay(100);
610 		i40e_resume_aq(hw);
611 	} while (retry < 10);
612 	if (ret_code != I40E_SUCCESS)
613 		goto init_adminq_free_arq;
614 
615 	/* get the NVM version info */
616 	i40e_read_nvm_word(hw, I40E_SR_NVM_IMAGE_VERSION, &hw->nvm.version);
617 	i40e_read_nvm_word(hw, I40E_SR_NVM_EETRACK_LO, &eetrack_lo);
618 	i40e_read_nvm_word(hw, I40E_SR_NVM_EETRACK_HI, &eetrack_hi);
619 	hw->nvm.eetrack = (eetrack_hi << 16) | eetrack_lo;
620 
621 	if (hw->aq.api_maj_ver > I40E_FW_API_VERSION_MAJOR) {
622 		ret_code = I40E_ERR_FIRMWARE_API_VERSION;
623 		goto init_adminq_free_arq;
624 	}
625 
626 	/* pre-emptive resource lock release */
627 	i40e_aq_release_resource(hw, I40E_NVM_RESOURCE_ID, 0, NULL);
628 	hw->aq.nvm_busy = FALSE;
629 
630 	ret_code = i40e_aq_set_hmc_resource_profile(hw,
631 						    I40E_HMC_PROFILE_DEFAULT,
632 						    0,
633 						    NULL);
634 	ret_code = I40E_SUCCESS;
635 
636 	/* success! */
637 	goto init_adminq_exit;
638 
639 init_adminq_free_arq:
640 	i40e_shutdown_arq(hw);
641 init_adminq_free_asq:
642 	i40e_shutdown_asq(hw);
643 init_adminq_destroy_spinlocks:
644 	i40e_destroy_spinlock(&hw->aq.asq_spinlock);
645 	i40e_destroy_spinlock(&hw->aq.arq_spinlock);
646 
647 init_adminq_exit:
648 	return ret_code;
649 }
650 
651 /**
652  *  i40e_shutdown_adminq - shutdown routine for the Admin Queue
653  *  @hw: pointer to the hardware structure
654  **/
655 enum i40e_status_code i40e_shutdown_adminq(struct i40e_hw *hw)
656 {
657 	enum i40e_status_code ret_code = I40E_SUCCESS;
658 
659 	if (i40e_check_asq_alive(hw))
660 		i40e_aq_queue_shutdown(hw, TRUE);
661 
662 	i40e_shutdown_asq(hw);
663 	i40e_shutdown_arq(hw);
664 
665 	/* destroy the spinlocks */
666 	i40e_destroy_spinlock(&hw->aq.asq_spinlock);
667 	i40e_destroy_spinlock(&hw->aq.arq_spinlock);
668 
669 	return ret_code;
670 }
671 
672 /**
673  *  i40e_clean_asq - cleans Admin send queue
674  *  @hw: pointer to the hardware structure
675  *
676  *  returns the number of free desc
677  **/
678 u16 i40e_clean_asq(struct i40e_hw *hw)
679 {
680 	struct i40e_adminq_ring *asq = &(hw->aq.asq);
681 	struct i40e_asq_cmd_details *details;
682 	u16 ntc = asq->next_to_clean;
683 	struct i40e_aq_desc desc_cb;
684 	struct i40e_aq_desc *desc;
685 
686 	desc = I40E_ADMINQ_DESC(*asq, ntc);
687 	details = I40E_ADMINQ_DETAILS(*asq, ntc);
688 	while (rd32(hw, hw->aq.asq.head) != ntc) {
689 		i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
690 			   "%s: ntc %d head %d.\n", __FUNCTION__, ntc,
691 			   rd32(hw, hw->aq.asq.head));
692 
693 		if (details->callback) {
694 			I40E_ADMINQ_CALLBACK cb_func =
695 					(I40E_ADMINQ_CALLBACK)details->callback;
696 			i40e_memcpy(&desc_cb, desc,
697 			            sizeof(struct i40e_aq_desc), I40E_DMA_TO_DMA);
698 			cb_func(hw, &desc_cb);
699 		}
700 		i40e_memset(desc, 0, sizeof(*desc), I40E_DMA_MEM);
701 		i40e_memset(details, 0, sizeof(*details), I40E_NONDMA_MEM);
702 		ntc++;
703 		if (ntc == asq->count)
704 			ntc = 0;
705 		desc = I40E_ADMINQ_DESC(*asq, ntc);
706 		details = I40E_ADMINQ_DETAILS(*asq, ntc);
707 	}
708 
709 	asq->next_to_clean = ntc;
710 
711 	return I40E_DESC_UNUSED(asq);
712 }
713 
714 /**
715  *  i40e_asq_done - check if FW has processed the Admin Send Queue
716  *  @hw: pointer to the hw struct
717  *
718  *  Returns TRUE if the firmware has processed all descriptors on the
719  *  admin send queue. Returns FALSE if there are still requests pending.
720  **/
721 bool i40e_asq_done(struct i40e_hw *hw)
722 {
723 	/* AQ designers suggest use of head for better
724 	 * timing reliability than DD bit
725 	 */
726 	return rd32(hw, hw->aq.asq.head) == hw->aq.asq.next_to_use;
727 
728 }
729 
730 /**
731  *  i40e_asq_send_command - send command to Admin Queue
732  *  @hw: pointer to the hw struct
733  *  @desc: prefilled descriptor describing the command (non DMA mem)
734  *  @buff: buffer to use for indirect commands
735  *  @buff_size: size of buffer for indirect commands
736  *  @cmd_details: pointer to command details structure
737  *
738  *  This is the main send command driver routine for the Admin Queue send
739  *  queue.  It runs the queue, cleans the queue, etc
740  **/
741 enum i40e_status_code i40e_asq_send_command(struct i40e_hw *hw,
742 				struct i40e_aq_desc *desc,
743 				void *buff, /* can be NULL */
744 				u16  buff_size,
745 				struct i40e_asq_cmd_details *cmd_details)
746 {
747 	enum i40e_status_code status = I40E_SUCCESS;
748 	struct i40e_dma_mem *dma_buff = NULL;
749 	struct i40e_asq_cmd_details *details;
750 	struct i40e_aq_desc *desc_on_ring;
751 	bool cmd_completed = FALSE;
752 	u16  retval = 0;
753 	u32  val = 0;
754 
755 	val = rd32(hw, hw->aq.asq.head);
756 	if (val >= hw->aq.num_asq_entries) {
757 		i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
758 			   "AQTX: head overrun at %d\n", val);
759 		status = I40E_ERR_QUEUE_EMPTY;
760 		goto asq_send_command_exit;
761 	}
762 
763 	if (hw->aq.asq.count == 0) {
764 		i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
765 			   "AQTX: Admin queue not initialized.\n");
766 		status = I40E_ERR_QUEUE_EMPTY;
767 		goto asq_send_command_exit;
768 	}
769 
770 	if (i40e_is_nvm_update_op(desc) && hw->aq.nvm_busy) {
771 		i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE, "AQTX: NVM busy.\n");
772 		status = I40E_ERR_NVM;
773 		goto asq_send_command_exit;
774 	}
775 
776 	details = I40E_ADMINQ_DETAILS(hw->aq.asq, hw->aq.asq.next_to_use);
777 	if (cmd_details) {
778 		i40e_memcpy(details,
779 			    cmd_details,
780 			    sizeof(struct i40e_asq_cmd_details),
781 			    I40E_NONDMA_TO_NONDMA);
782 
783 		/* If the cmd_details are defined copy the cookie.  The
784 		 * CPU_TO_LE32 is not needed here because the data is ignored
785 		 * by the FW, only used by the driver
786 		 */
787 		if (details->cookie) {
788 			desc->cookie_high =
789 				CPU_TO_LE32(I40E_HI_DWORD(details->cookie));
790 			desc->cookie_low =
791 				CPU_TO_LE32(I40E_LO_DWORD(details->cookie));
792 		}
793 	} else {
794 		i40e_memset(details, 0,
795 			    sizeof(struct i40e_asq_cmd_details),
796 			    I40E_NONDMA_MEM);
797 	}
798 
799 	/* clear requested flags and then set additional flags if defined */
800 	desc->flags &= ~CPU_TO_LE16(details->flags_dis);
801 	desc->flags |= CPU_TO_LE16(details->flags_ena);
802 
803 	i40e_acquire_spinlock(&hw->aq.asq_spinlock);
804 
805 	if (buff_size > hw->aq.asq_buf_size) {
806 		i40e_debug(hw,
807 			   I40E_DEBUG_AQ_MESSAGE,
808 			   "AQTX: Invalid buffer size: %d.\n",
809 			   buff_size);
810 		status = I40E_ERR_INVALID_SIZE;
811 		goto asq_send_command_error;
812 	}
813 
814 	if (details->postpone && !details->async) {
815 		i40e_debug(hw,
816 			   I40E_DEBUG_AQ_MESSAGE,
817 			   "AQTX: Async flag not set along with postpone flag");
818 		status = I40E_ERR_PARAM;
819 		goto asq_send_command_error;
820 	}
821 
822 	/* call clean and check queue available function to reclaim the
823 	 * descriptors that were processed by FW, the function returns the
824 	 * number of desc available
825 	 */
826 	/* the clean function called here could be called in a separate thread
827 	 * in case of asynchronous completions
828 	 */
829 	if (i40e_clean_asq(hw) == 0) {
830 		i40e_debug(hw,
831 			   I40E_DEBUG_AQ_MESSAGE,
832 			   "AQTX: Error queue is full.\n");
833 		status = I40E_ERR_ADMIN_QUEUE_FULL;
834 		goto asq_send_command_error;
835 	}
836 
837 	/* initialize the temp desc pointer with the right desc */
838 	desc_on_ring = I40E_ADMINQ_DESC(hw->aq.asq, hw->aq.asq.next_to_use);
839 
840 	/* if the desc is available copy the temp desc to the right place */
841 	i40e_memcpy(desc_on_ring, desc, sizeof(struct i40e_aq_desc),
842 		    I40E_NONDMA_TO_DMA);
843 
844 	/* if buff is not NULL assume indirect command */
845 	if (buff != NULL) {
846 		dma_buff = &(hw->aq.asq.r.asq_bi[hw->aq.asq.next_to_use]);
847 		/* copy the user buff into the respective DMA buff */
848 		i40e_memcpy(dma_buff->va, buff, buff_size,
849 			    I40E_NONDMA_TO_DMA);
850 		desc_on_ring->datalen = CPU_TO_LE16(buff_size);
851 
852 		/* Update the address values in the desc with the pa value
853 		 * for respective buffer
854 		 */
855 		desc_on_ring->params.external.addr_high =
856 				CPU_TO_LE32(I40E_HI_DWORD(dma_buff->pa));
857 		desc_on_ring->params.external.addr_low =
858 				CPU_TO_LE32(I40E_LO_DWORD(dma_buff->pa));
859 	}
860 
861 	/* bump the tail */
862 	i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE, "AQTX: desc and buffer:\n");
863 	i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc_on_ring,
864 		      buff, buff_size);
865 	(hw->aq.asq.next_to_use)++;
866 	if (hw->aq.asq.next_to_use == hw->aq.asq.count)
867 		hw->aq.asq.next_to_use = 0;
868 	if (!details->postpone)
869 		wr32(hw, hw->aq.asq.tail, hw->aq.asq.next_to_use);
870 
871 	/* if cmd_details are not defined or async flag is not set,
872 	 * we need to wait for desc write back
873 	 */
874 	if (!details->async && !details->postpone) {
875 		u32 total_delay = 0;
876 
877 		do {
878 			/* AQ designers suggest use of head for better
879 			 * timing reliability than DD bit
880 			 */
881 			if (i40e_asq_done(hw))
882 				break;
883 			/* ugh! delay while spin_lock */
884 			i40e_msec_delay(1);
885 			total_delay++;
886 		} while (total_delay < hw->aq.asq_cmd_timeout);
887 	}
888 
889 	/* if ready, copy the desc back to temp */
890 	if (i40e_asq_done(hw)) {
891 		i40e_memcpy(desc, desc_on_ring, sizeof(struct i40e_aq_desc),
892 			    I40E_DMA_TO_NONDMA);
893 		if (buff != NULL)
894 			i40e_memcpy(buff, dma_buff->va, buff_size,
895 				    I40E_DMA_TO_NONDMA);
896 		retval = LE16_TO_CPU(desc->retval);
897 		if (retval != 0) {
898 			i40e_debug(hw,
899 				   I40E_DEBUG_AQ_MESSAGE,
900 				   "AQTX: Command completed with error 0x%X.\n",
901 				   retval);
902 
903 			/* strip off FW internal code */
904 			retval &= 0xff;
905 		}
906 		cmd_completed = TRUE;
907 		if ((enum i40e_admin_queue_err)retval == I40E_AQ_RC_OK)
908 			status = I40E_SUCCESS;
909 		else
910 			status = I40E_ERR_ADMIN_QUEUE_ERROR;
911 		hw->aq.asq_last_status = (enum i40e_admin_queue_err)retval;
912 	}
913 
914 	i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
915 		   "AQTX: desc and buffer writeback:\n");
916 	i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc, buff, buff_size);
917 
918 	/* update the error if time out occurred */
919 	if ((!cmd_completed) &&
920 	    (!details->async && !details->postpone)) {
921 		i40e_debug(hw,
922 			   I40E_DEBUG_AQ_MESSAGE,
923 			   "AQTX: Writeback timeout.\n");
924 		status = I40E_ERR_ADMIN_QUEUE_TIMEOUT;
925 	}
926 
927 	if (!status && i40e_is_nvm_update_op(desc))
928 		hw->aq.nvm_busy = TRUE;
929 
930 asq_send_command_error:
931 	i40e_release_spinlock(&hw->aq.asq_spinlock);
932 asq_send_command_exit:
933 	return status;
934 }
935 
936 /**
937  *  i40e_fill_default_direct_cmd_desc - AQ descriptor helper function
938  *  @desc:     pointer to the temp descriptor (non DMA mem)
939  *  @opcode:   the opcode can be used to decide which flags to turn off or on
940  *
941  *  Fill the desc with default values
942  **/
943 void i40e_fill_default_direct_cmd_desc(struct i40e_aq_desc *desc,
944 				       u16 opcode)
945 {
946 	/* zero out the desc */
947 	i40e_memset((void *)desc, 0, sizeof(struct i40e_aq_desc),
948 		    I40E_NONDMA_MEM);
949 	desc->opcode = CPU_TO_LE16(opcode);
950 	desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_SI);
951 }
952 
953 /**
954  *  i40e_clean_arq_element
955  *  @hw: pointer to the hw struct
956  *  @e: event info from the receive descriptor, includes any buffers
957  *  @pending: number of events that could be left to process
958  *
959  *  This function cleans one Admin Receive Queue element and returns
960  *  the contents through e.  It can also return how many events are
961  *  left to process through 'pending'
962  **/
963 enum i40e_status_code i40e_clean_arq_element(struct i40e_hw *hw,
964 					     struct i40e_arq_event_info *e,
965 					     u16 *pending)
966 {
967 	enum i40e_status_code ret_code = I40E_SUCCESS;
968 	u16 ntc = hw->aq.arq.next_to_clean;
969 	struct i40e_aq_desc *desc;
970 	struct i40e_dma_mem *bi;
971 	u16 desc_idx;
972 	u16 datalen;
973 	u16 flags;
974 	u16 ntu;
975 
976 	/* take the lock before we start messing with the ring */
977 	i40e_acquire_spinlock(&hw->aq.arq_spinlock);
978 
979 	/* set next_to_use to head */
980 	ntu = (rd32(hw, hw->aq.arq.head) & I40E_PF_ARQH_ARQH_MASK);
981 	if (ntu == ntc) {
982 		/* nothing to do - shouldn't need to update ring's values */
983 		ret_code = I40E_ERR_ADMIN_QUEUE_NO_WORK;
984 		goto clean_arq_element_out;
985 	}
986 
987 	/* now clean the next descriptor */
988 	desc = I40E_ADMINQ_DESC(hw->aq.arq, ntc);
989 	desc_idx = ntc;
990 
991 	flags = LE16_TO_CPU(desc->flags);
992 	if (flags & I40E_AQ_FLAG_ERR) {
993 		ret_code = I40E_ERR_ADMIN_QUEUE_ERROR;
994 		hw->aq.arq_last_status =
995 			(enum i40e_admin_queue_err)LE16_TO_CPU(desc->retval);
996 		i40e_debug(hw,
997 			   I40E_DEBUG_AQ_MESSAGE,
998 			   "AQRX: Event received with error 0x%X.\n",
999 			   hw->aq.arq_last_status);
1000 	}
1001 
1002 	i40e_memcpy(&e->desc, desc, sizeof(struct i40e_aq_desc),
1003 		    I40E_DMA_TO_NONDMA);
1004 	datalen = LE16_TO_CPU(desc->datalen);
1005 	e->msg_len = min(datalen, e->buf_len);
1006 	if (e->msg_buf != NULL && (e->msg_len != 0))
1007 		i40e_memcpy(e->msg_buf,
1008 			    hw->aq.arq.r.arq_bi[desc_idx].va,
1009 			    e->msg_len, I40E_DMA_TO_NONDMA);
1010 
1011 	i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE, "AQRX: desc and buffer:\n");
1012 	i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc, e->msg_buf,
1013 		      hw->aq.arq_buf_size);
1014 
1015 	/* Restore the original datalen and buffer address in the desc,
1016 	 * FW updates datalen to indicate the event message
1017 	 * size
1018 	 */
1019 	bi = &hw->aq.arq.r.arq_bi[ntc];
1020 	i40e_memset((void *)desc, 0, sizeof(struct i40e_aq_desc), I40E_DMA_MEM);
1021 
1022 	desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_BUF);
1023 	if (hw->aq.arq_buf_size > I40E_AQ_LARGE_BUF)
1024 		desc->flags |= CPU_TO_LE16(I40E_AQ_FLAG_LB);
1025 	desc->datalen = CPU_TO_LE16((u16)bi->size);
1026 	desc->params.external.addr_high = CPU_TO_LE32(I40E_HI_DWORD(bi->pa));
1027 	desc->params.external.addr_low = CPU_TO_LE32(I40E_LO_DWORD(bi->pa));
1028 
1029 	/* set tail = the last cleaned desc index. */
1030 	wr32(hw, hw->aq.arq.tail, ntc);
1031 	/* ntc is updated to tail + 1 */
1032 	ntc++;
1033 	if (ntc == hw->aq.num_arq_entries)
1034 		ntc = 0;
1035 	hw->aq.arq.next_to_clean = ntc;
1036 	hw->aq.arq.next_to_use = ntu;
1037 
1038 clean_arq_element_out:
1039 	/* Set pending if needed, unlock and return */
1040 	if (pending != NULL)
1041 		*pending = (ntc > ntu ? hw->aq.arq.count : 0) + (ntu - ntc);
1042 	i40e_release_spinlock(&hw->aq.arq_spinlock);
1043 
1044 	if (i40e_is_nvm_update_op(&e->desc)) {
1045 		hw->aq.nvm_busy = FALSE;
1046 		if (hw->aq.nvm_release_on_done) {
1047 			i40e_release_nvm(hw);
1048 			hw->aq.nvm_release_on_done = FALSE;
1049 		}
1050 	}
1051 
1052 	return ret_code;
1053 }
1054 
1055 void i40e_resume_aq(struct i40e_hw *hw)
1056 {
1057 	/* Registers are reset after PF reset */
1058 	hw->aq.asq.next_to_use = 0;
1059 	hw->aq.asq.next_to_clean = 0;
1060 
1061 #if (I40E_VF_ATQLEN_ATQENABLE_MASK != I40E_PF_ATQLEN_ATQENABLE_MASK)
1062 #error I40E_VF_ATQLEN_ATQENABLE_MASK != I40E_PF_ATQLEN_ATQENABLE_MASK
1063 #endif
1064 	i40e_config_asq_regs(hw);
1065 
1066 	hw->aq.arq.next_to_use = 0;
1067 	hw->aq.arq.next_to_clean = 0;
1068 
1069 	i40e_config_arq_regs(hw);
1070 }
1071