xref: /freebsd/sys/dev/ixl/i40e_adminq.c (revision 8ef24a0d4b28fe230e20637f56869cc4148cd2ca)
1 /******************************************************************************
2 
3   Copyright (c) 2013-2015, Intel Corporation
4   All rights reserved.
5 
6   Redistribution and use in source and binary forms, with or without
7   modification, are permitted provided that the following conditions are met:
8 
9    1. Redistributions of source code must retain the above copyright notice,
10       this list of conditions and the following disclaimer.
11 
12    2. Redistributions in binary form must reproduce the above copyright
13       notice, this list of conditions and the following disclaimer in the
14       documentation and/or other materials provided with the distribution.
15 
16    3. Neither the name of the Intel Corporation nor the names of its
17       contributors may be used to endorse or promote products derived from
18       this software without specific prior written permission.
19 
20   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
21   AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22   IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23   ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
24   LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25   CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26   SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27   INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28   CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29   ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30   POSSIBILITY OF SUCH DAMAGE.
31 
32 ******************************************************************************/
33 /*$FreeBSD$*/
34 
35 #include "i40e_status.h"
36 #include "i40e_type.h"
37 #include "i40e_register.h"
38 #include "i40e_adminq.h"
39 #include "i40e_prototype.h"
40 
41 /**
42  * i40e_is_nvm_update_op - return TRUE if this is an NVM update operation
43  * @desc: API request descriptor
44  **/
45 static INLINE bool i40e_is_nvm_update_op(struct i40e_aq_desc *desc)
46 {
47 	return (desc->opcode == CPU_TO_LE16(i40e_aqc_opc_nvm_erase)) ||
48 		(desc->opcode == CPU_TO_LE16(i40e_aqc_opc_nvm_update));
49 }
50 
51 /**
52  *  i40e_adminq_init_regs - Initialize AdminQ registers
53  *  @hw: pointer to the hardware structure
54  *
55  *  This assumes the alloc_asq and alloc_arq functions have already been called
56  **/
57 static void i40e_adminq_init_regs(struct i40e_hw *hw)
58 {
59 	/* set head and tail registers in our local struct */
60 	if (i40e_is_vf(hw)) {
61 		hw->aq.asq.tail = I40E_VF_ATQT1;
62 		hw->aq.asq.head = I40E_VF_ATQH1;
63 		hw->aq.asq.len  = I40E_VF_ATQLEN1;
64 		hw->aq.asq.bal  = I40E_VF_ATQBAL1;
65 		hw->aq.asq.bah  = I40E_VF_ATQBAH1;
66 		hw->aq.arq.tail = I40E_VF_ARQT1;
67 		hw->aq.arq.head = I40E_VF_ARQH1;
68 		hw->aq.arq.len  = I40E_VF_ARQLEN1;
69 		hw->aq.arq.bal  = I40E_VF_ARQBAL1;
70 		hw->aq.arq.bah  = I40E_VF_ARQBAH1;
71 	} else {
72 		hw->aq.asq.tail = I40E_PF_ATQT;
73 		hw->aq.asq.head = I40E_PF_ATQH;
74 		hw->aq.asq.len  = I40E_PF_ATQLEN;
75 		hw->aq.asq.bal  = I40E_PF_ATQBAL;
76 		hw->aq.asq.bah  = I40E_PF_ATQBAH;
77 		hw->aq.arq.tail = I40E_PF_ARQT;
78 		hw->aq.arq.head = I40E_PF_ARQH;
79 		hw->aq.arq.len  = I40E_PF_ARQLEN;
80 		hw->aq.arq.bal  = I40E_PF_ARQBAL;
81 		hw->aq.arq.bah  = I40E_PF_ARQBAH;
82 	}
83 }
84 
85 /**
86  *  i40e_alloc_adminq_asq_ring - Allocate Admin Queue send rings
87  *  @hw: pointer to the hardware structure
88  **/
89 enum i40e_status_code i40e_alloc_adminq_asq_ring(struct i40e_hw *hw)
90 {
91 	enum i40e_status_code ret_code;
92 
93 	ret_code = i40e_allocate_dma_mem(hw, &hw->aq.asq.desc_buf,
94 					 i40e_mem_atq_ring,
95 					 (hw->aq.num_asq_entries *
96 					 sizeof(struct i40e_aq_desc)),
97 					 I40E_ADMINQ_DESC_ALIGNMENT);
98 	if (ret_code)
99 		return ret_code;
100 
101 	ret_code = i40e_allocate_virt_mem(hw, &hw->aq.asq.cmd_buf,
102 					  (hw->aq.num_asq_entries *
103 					  sizeof(struct i40e_asq_cmd_details)));
104 	if (ret_code) {
105 		i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
106 		return ret_code;
107 	}
108 
109 	return ret_code;
110 }
111 
112 /**
113  *  i40e_alloc_adminq_arq_ring - Allocate Admin Queue receive rings
114  *  @hw: pointer to the hardware structure
115  **/
116 enum i40e_status_code i40e_alloc_adminq_arq_ring(struct i40e_hw *hw)
117 {
118 	enum i40e_status_code ret_code;
119 
120 	ret_code = i40e_allocate_dma_mem(hw, &hw->aq.arq.desc_buf,
121 					 i40e_mem_arq_ring,
122 					 (hw->aq.num_arq_entries *
123 					 sizeof(struct i40e_aq_desc)),
124 					 I40E_ADMINQ_DESC_ALIGNMENT);
125 
126 	return ret_code;
127 }
128 
129 /**
130  *  i40e_free_adminq_asq - Free Admin Queue send rings
131  *  @hw: pointer to the hardware structure
132  *
133  *  This assumes the posted send buffers have already been cleaned
134  *  and de-allocated
135  **/
136 void i40e_free_adminq_asq(struct i40e_hw *hw)
137 {
138 	i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
139 }
140 
141 /**
142  *  i40e_free_adminq_arq - Free Admin Queue receive rings
143  *  @hw: pointer to the hardware structure
144  *
145  *  This assumes the posted receive buffers have already been cleaned
146  *  and de-allocated
147  **/
148 void i40e_free_adminq_arq(struct i40e_hw *hw)
149 {
150 	i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf);
151 }
152 
153 /**
154  *  i40e_alloc_arq_bufs - Allocate pre-posted buffers for the receive queue
155  *  @hw: pointer to the hardware structure
156  **/
157 static enum i40e_status_code i40e_alloc_arq_bufs(struct i40e_hw *hw)
158 {
159 	enum i40e_status_code ret_code;
160 	struct i40e_aq_desc *desc;
161 	struct i40e_dma_mem *bi;
162 	int i;
163 
164 	/* We'll be allocating the buffer info memory first, then we can
165 	 * allocate the mapped buffers for the event processing
166 	 */
167 
168 	/* buffer_info structures do not need alignment */
169 	ret_code = i40e_allocate_virt_mem(hw, &hw->aq.arq.dma_head,
170 		(hw->aq.num_arq_entries * sizeof(struct i40e_dma_mem)));
171 	if (ret_code)
172 		goto alloc_arq_bufs;
173 	hw->aq.arq.r.arq_bi = (struct i40e_dma_mem *)hw->aq.arq.dma_head.va;
174 
175 	/* allocate the mapped buffers */
176 	for (i = 0; i < hw->aq.num_arq_entries; i++) {
177 		bi = &hw->aq.arq.r.arq_bi[i];
178 		ret_code = i40e_allocate_dma_mem(hw, bi,
179 						 i40e_mem_arq_buf,
180 						 hw->aq.arq_buf_size,
181 						 I40E_ADMINQ_DESC_ALIGNMENT);
182 		if (ret_code)
183 			goto unwind_alloc_arq_bufs;
184 
185 		/* now configure the descriptors for use */
186 		desc = I40E_ADMINQ_DESC(hw->aq.arq, i);
187 
188 		desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_BUF);
189 		if (hw->aq.arq_buf_size > I40E_AQ_LARGE_BUF)
190 			desc->flags |= CPU_TO_LE16(I40E_AQ_FLAG_LB);
191 		desc->opcode = 0;
192 		/* This is in accordance with Admin queue design, there is no
193 		 * register for buffer size configuration
194 		 */
195 		desc->datalen = CPU_TO_LE16((u16)bi->size);
196 		desc->retval = 0;
197 		desc->cookie_high = 0;
198 		desc->cookie_low = 0;
199 		desc->params.external.addr_high =
200 			CPU_TO_LE32(I40E_HI_DWORD(bi->pa));
201 		desc->params.external.addr_low =
202 			CPU_TO_LE32(I40E_LO_DWORD(bi->pa));
203 		desc->params.external.param0 = 0;
204 		desc->params.external.param1 = 0;
205 	}
206 
207 alloc_arq_bufs:
208 	return ret_code;
209 
210 unwind_alloc_arq_bufs:
211 	/* don't try to free the one that failed... */
212 	i--;
213 	for (; i >= 0; i--)
214 		i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]);
215 	i40e_free_virt_mem(hw, &hw->aq.arq.dma_head);
216 
217 	return ret_code;
218 }
219 
220 /**
221  *  i40e_alloc_asq_bufs - Allocate empty buffer structs for the send queue
222  *  @hw: pointer to the hardware structure
223  **/
224 static enum i40e_status_code i40e_alloc_asq_bufs(struct i40e_hw *hw)
225 {
226 	enum i40e_status_code ret_code;
227 	struct i40e_dma_mem *bi;
228 	int i;
229 
230 	/* No mapped memory needed yet, just the buffer info structures */
231 	ret_code = i40e_allocate_virt_mem(hw, &hw->aq.asq.dma_head,
232 		(hw->aq.num_asq_entries * sizeof(struct i40e_dma_mem)));
233 	if (ret_code)
234 		goto alloc_asq_bufs;
235 	hw->aq.asq.r.asq_bi = (struct i40e_dma_mem *)hw->aq.asq.dma_head.va;
236 
237 	/* allocate the mapped buffers */
238 	for (i = 0; i < hw->aq.num_asq_entries; i++) {
239 		bi = &hw->aq.asq.r.asq_bi[i];
240 		ret_code = i40e_allocate_dma_mem(hw, bi,
241 						 i40e_mem_asq_buf,
242 						 hw->aq.asq_buf_size,
243 						 I40E_ADMINQ_DESC_ALIGNMENT);
244 		if (ret_code)
245 			goto unwind_alloc_asq_bufs;
246 	}
247 alloc_asq_bufs:
248 	return ret_code;
249 
250 unwind_alloc_asq_bufs:
251 	/* don't try to free the one that failed... */
252 	i--;
253 	for (; i >= 0; i--)
254 		i40e_free_dma_mem(hw, &hw->aq.asq.r.asq_bi[i]);
255 	i40e_free_virt_mem(hw, &hw->aq.asq.dma_head);
256 
257 	return ret_code;
258 }
259 
260 /**
261  *  i40e_free_arq_bufs - Free receive queue buffer info elements
262  *  @hw: pointer to the hardware structure
263  **/
264 static void i40e_free_arq_bufs(struct i40e_hw *hw)
265 {
266 	int i;
267 
268 	/* free descriptors */
269 	for (i = 0; i < hw->aq.num_arq_entries; i++)
270 		i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]);
271 
272 	/* free the descriptor memory */
273 	i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf);
274 
275 	/* free the dma header */
276 	i40e_free_virt_mem(hw, &hw->aq.arq.dma_head);
277 }
278 
279 /**
280  *  i40e_free_asq_bufs - Free send queue buffer info elements
281  *  @hw: pointer to the hardware structure
282  **/
283 static void i40e_free_asq_bufs(struct i40e_hw *hw)
284 {
285 	int i;
286 
287 	/* only unmap if the address is non-NULL */
288 	for (i = 0; i < hw->aq.num_asq_entries; i++)
289 		if (hw->aq.asq.r.asq_bi[i].pa)
290 			i40e_free_dma_mem(hw, &hw->aq.asq.r.asq_bi[i]);
291 
292 	/* free the buffer info list */
293 	i40e_free_virt_mem(hw, &hw->aq.asq.cmd_buf);
294 
295 	/* free the descriptor memory */
296 	i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
297 
298 	/* free the dma header */
299 	i40e_free_virt_mem(hw, &hw->aq.asq.dma_head);
300 }
301 
302 /**
303  *  i40e_config_asq_regs - configure ASQ registers
304  *  @hw: pointer to the hardware structure
305  *
306  *  Configure base address and length registers for the transmit queue
307  **/
308 static enum i40e_status_code i40e_config_asq_regs(struct i40e_hw *hw)
309 {
310 	enum i40e_status_code ret_code = I40E_SUCCESS;
311 	u32 reg = 0;
312 
313 	/* Clear Head and Tail */
314 	wr32(hw, hw->aq.asq.head, 0);
315 	wr32(hw, hw->aq.asq.tail, 0);
316 
317 	/* set starting point */
318 	if (!i40e_is_vf(hw))
319 		wr32(hw, hw->aq.asq.len, (hw->aq.num_asq_entries |
320 					  I40E_PF_ATQLEN_ATQENABLE_MASK));
321 	if (i40e_is_vf(hw))
322 		wr32(hw, hw->aq.asq.len, (hw->aq.num_asq_entries |
323 					  I40E_VF_ATQLEN1_ATQENABLE_MASK));
324 	wr32(hw, hw->aq.asq.bal, I40E_LO_DWORD(hw->aq.asq.desc_buf.pa));
325 	wr32(hw, hw->aq.asq.bah, I40E_HI_DWORD(hw->aq.asq.desc_buf.pa));
326 
327 	/* Check one register to verify that config was applied */
328 	reg = rd32(hw, hw->aq.asq.bal);
329 	if (reg != I40E_LO_DWORD(hw->aq.asq.desc_buf.pa))
330 		ret_code = I40E_ERR_ADMIN_QUEUE_ERROR;
331 
332 	return ret_code;
333 }
334 
335 /**
336  *  i40e_config_arq_regs - ARQ register configuration
337  *  @hw: pointer to the hardware structure
338  *
339  * Configure base address and length registers for the receive (event queue)
340  **/
341 static enum i40e_status_code i40e_config_arq_regs(struct i40e_hw *hw)
342 {
343 	enum i40e_status_code ret_code = I40E_SUCCESS;
344 	u32 reg = 0;
345 
346 	/* Clear Head and Tail */
347 	wr32(hw, hw->aq.arq.head, 0);
348 	wr32(hw, hw->aq.arq.tail, 0);
349 
350 	/* set starting point */
351 	if (!i40e_is_vf(hw))
352 		wr32(hw, hw->aq.arq.len, (hw->aq.num_arq_entries |
353 					  I40E_PF_ARQLEN_ARQENABLE_MASK));
354 	if (i40e_is_vf(hw))
355 		wr32(hw, hw->aq.arq.len, (hw->aq.num_arq_entries |
356 					  I40E_VF_ARQLEN1_ARQENABLE_MASK));
357 	wr32(hw, hw->aq.arq.bal, I40E_LO_DWORD(hw->aq.arq.desc_buf.pa));
358 	wr32(hw, hw->aq.arq.bah, I40E_HI_DWORD(hw->aq.arq.desc_buf.pa));
359 
360 	/* Update tail in the HW to post pre-allocated buffers */
361 	wr32(hw, hw->aq.arq.tail, hw->aq.num_arq_entries - 1);
362 
363 	/* Check one register to verify that config was applied */
364 	reg = rd32(hw, hw->aq.arq.bal);
365 	if (reg != I40E_LO_DWORD(hw->aq.arq.desc_buf.pa))
366 		ret_code = I40E_ERR_ADMIN_QUEUE_ERROR;
367 
368 	return ret_code;
369 }
370 
371 /**
372  *  i40e_init_asq - main initialization routine for ASQ
373  *  @hw: pointer to the hardware structure
374  *
375  *  This is the main initialization routine for the Admin Send Queue
376  *  Prior to calling this function, drivers *MUST* set the following fields
377  *  in the hw->aq structure:
378  *     - hw->aq.num_asq_entries
379  *     - hw->aq.arq_buf_size
380  *
381  *  Do *NOT* hold the lock when calling this as the memory allocation routines
382  *  called are not going to be atomic context safe
383  **/
384 enum i40e_status_code i40e_init_asq(struct i40e_hw *hw)
385 {
386 	enum i40e_status_code ret_code = I40E_SUCCESS;
387 
388 	if (hw->aq.asq.count > 0) {
389 		/* queue already initialized */
390 		ret_code = I40E_ERR_NOT_READY;
391 		goto init_adminq_exit;
392 	}
393 
394 	/* verify input for valid configuration */
395 	if ((hw->aq.num_asq_entries == 0) ||
396 	    (hw->aq.asq_buf_size == 0)) {
397 		ret_code = I40E_ERR_CONFIG;
398 		goto init_adminq_exit;
399 	}
400 
401 	hw->aq.asq.next_to_use = 0;
402 	hw->aq.asq.next_to_clean = 0;
403 
404 	/* allocate the ring memory */
405 	ret_code = i40e_alloc_adminq_asq_ring(hw);
406 	if (ret_code != I40E_SUCCESS)
407 		goto init_adminq_exit;
408 
409 	/* allocate buffers in the rings */
410 	ret_code = i40e_alloc_asq_bufs(hw);
411 	if (ret_code != I40E_SUCCESS)
412 		goto init_adminq_free_rings;
413 
414 	/* initialize base registers */
415 	ret_code = i40e_config_asq_regs(hw);
416 	if (ret_code != I40E_SUCCESS)
417 		goto init_adminq_free_rings;
418 
419 	/* success! */
420 	hw->aq.asq.count = hw->aq.num_asq_entries;
421 	goto init_adminq_exit;
422 
423 init_adminq_free_rings:
424 	i40e_free_adminq_asq(hw);
425 
426 init_adminq_exit:
427 	return ret_code;
428 }
429 
430 /**
431  *  i40e_init_arq - initialize ARQ
432  *  @hw: pointer to the hardware structure
433  *
434  *  The main initialization routine for the Admin Receive (Event) Queue.
435  *  Prior to calling this function, drivers *MUST* set the following fields
436  *  in the hw->aq structure:
437  *     - hw->aq.num_asq_entries
438  *     - hw->aq.arq_buf_size
439  *
440  *  Do *NOT* hold the lock when calling this as the memory allocation routines
441  *  called are not going to be atomic context safe
442  **/
443 enum i40e_status_code i40e_init_arq(struct i40e_hw *hw)
444 {
445 	enum i40e_status_code ret_code = I40E_SUCCESS;
446 
447 	if (hw->aq.arq.count > 0) {
448 		/* queue already initialized */
449 		ret_code = I40E_ERR_NOT_READY;
450 		goto init_adminq_exit;
451 	}
452 
453 	/* verify input for valid configuration */
454 	if ((hw->aq.num_arq_entries == 0) ||
455 	    (hw->aq.arq_buf_size == 0)) {
456 		ret_code = I40E_ERR_CONFIG;
457 		goto init_adminq_exit;
458 	}
459 
460 	hw->aq.arq.next_to_use = 0;
461 	hw->aq.arq.next_to_clean = 0;
462 
463 	/* allocate the ring memory */
464 	ret_code = i40e_alloc_adminq_arq_ring(hw);
465 	if (ret_code != I40E_SUCCESS)
466 		goto init_adminq_exit;
467 
468 	/* allocate buffers in the rings */
469 	ret_code = i40e_alloc_arq_bufs(hw);
470 	if (ret_code != I40E_SUCCESS)
471 		goto init_adminq_free_rings;
472 
473 	/* initialize base registers */
474 	ret_code = i40e_config_arq_regs(hw);
475 	if (ret_code != I40E_SUCCESS)
476 		goto init_adminq_free_rings;
477 
478 	/* success! */
479 	hw->aq.arq.count = hw->aq.num_arq_entries;
480 	goto init_adminq_exit;
481 
482 init_adminq_free_rings:
483 	i40e_free_adminq_arq(hw);
484 
485 init_adminq_exit:
486 	return ret_code;
487 }
488 
489 /**
490  *  i40e_shutdown_asq - shutdown the ASQ
491  *  @hw: pointer to the hardware structure
492  *
493  *  The main shutdown routine for the Admin Send Queue
494  **/
495 enum i40e_status_code i40e_shutdown_asq(struct i40e_hw *hw)
496 {
497 	enum i40e_status_code ret_code = I40E_SUCCESS;
498 
499 	i40e_acquire_spinlock(&hw->aq.asq_spinlock);
500 
501 	if (hw->aq.asq.count == 0) {
502 		ret_code = I40E_ERR_NOT_READY;
503 		goto shutdown_asq_out;
504 	}
505 
506 	/* Stop firmware AdminQ processing */
507 	wr32(hw, hw->aq.asq.head, 0);
508 	wr32(hw, hw->aq.asq.tail, 0);
509 	wr32(hw, hw->aq.asq.len, 0);
510 	wr32(hw, hw->aq.asq.bal, 0);
511 	wr32(hw, hw->aq.asq.bah, 0);
512 
513 	hw->aq.asq.count = 0; /* to indicate uninitialized queue */
514 
515 	/* free ring buffers */
516 	i40e_free_asq_bufs(hw);
517 
518 shutdown_asq_out:
519 	i40e_release_spinlock(&hw->aq.asq_spinlock);
520 	return ret_code;
521 }
522 
523 /**
524  *  i40e_shutdown_arq - shutdown ARQ
525  *  @hw: pointer to the hardware structure
526  *
527  *  The main shutdown routine for the Admin Receive Queue
528  **/
529 enum i40e_status_code i40e_shutdown_arq(struct i40e_hw *hw)
530 {
531 	enum i40e_status_code ret_code = I40E_SUCCESS;
532 
533 	i40e_acquire_spinlock(&hw->aq.arq_spinlock);
534 
535 	if (hw->aq.arq.count == 0) {
536 		ret_code = I40E_ERR_NOT_READY;
537 		goto shutdown_arq_out;
538 	}
539 
540 	/* Stop firmware AdminQ processing */
541 	wr32(hw, hw->aq.arq.head, 0);
542 	wr32(hw, hw->aq.arq.tail, 0);
543 	wr32(hw, hw->aq.arq.len, 0);
544 	wr32(hw, hw->aq.arq.bal, 0);
545 	wr32(hw, hw->aq.arq.bah, 0);
546 
547 	hw->aq.arq.count = 0; /* to indicate uninitialized queue */
548 
549 	/* free ring buffers */
550 	i40e_free_arq_bufs(hw);
551 
552 shutdown_arq_out:
553 	i40e_release_spinlock(&hw->aq.arq_spinlock);
554 	return ret_code;
555 }
556 
557 /**
558  *  i40e_resume_aq - resume AQ processing from 0
559  *  @hw: pointer to the hardware structure
560  **/
561 static void i40e_resume_aq(struct i40e_hw *hw)
562 {
563 	/* Registers are reset after PF reset */
564 	hw->aq.asq.next_to_use = 0;
565 	hw->aq.asq.next_to_clean = 0;
566 
567 	i40e_config_asq_regs(hw);
568 
569 	hw->aq.arq.next_to_use = 0;
570 	hw->aq.arq.next_to_clean = 0;
571 
572 	i40e_config_arq_regs(hw);
573 }
574 
575 /**
576  *  i40e_init_adminq - main initialization routine for Admin Queue
577  *  @hw: pointer to the hardware structure
578  *
579  *  Prior to calling this function, drivers *MUST* set the following fields
580  *  in the hw->aq structure:
581  *     - hw->aq.num_asq_entries
582  *     - hw->aq.num_arq_entries
583  *     - hw->aq.arq_buf_size
584  *     - hw->aq.asq_buf_size
585  **/
586 enum i40e_status_code i40e_init_adminq(struct i40e_hw *hw)
587 {
588 	u16 cfg_ptr, oem_hi, oem_lo;
589 	u16 eetrack_lo, eetrack_hi;
590 	enum i40e_status_code ret_code;
591 	int retry = 0;
592 
593 	/* verify input for valid configuration */
594 	if ((hw->aq.num_arq_entries == 0) ||
595 	    (hw->aq.num_asq_entries == 0) ||
596 	    (hw->aq.arq_buf_size == 0) ||
597 	    (hw->aq.asq_buf_size == 0)) {
598 		ret_code = I40E_ERR_CONFIG;
599 		goto init_adminq_exit;
600 	}
601 	i40e_init_spinlock(&hw->aq.asq_spinlock);
602 	i40e_init_spinlock(&hw->aq.arq_spinlock);
603 
604 	/* Set up register offsets */
605 	i40e_adminq_init_regs(hw);
606 
607 	/* setup ASQ command write back timeout */
608 	hw->aq.asq_cmd_timeout = I40E_ASQ_CMD_TIMEOUT;
609 
610 	/* allocate the ASQ */
611 	ret_code = i40e_init_asq(hw);
612 	if (ret_code != I40E_SUCCESS)
613 		goto init_adminq_destroy_spinlocks;
614 
615 	/* allocate the ARQ */
616 	ret_code = i40e_init_arq(hw);
617 	if (ret_code != I40E_SUCCESS)
618 		goto init_adminq_free_asq;
619 
620 	/* VF has no need of firmware */
621 	if (i40e_is_vf(hw))
622 		goto init_adminq_exit;
623 	/* There are some cases where the firmware may not be quite ready
624 	 * for AdminQ operations, so we retry the AdminQ setup a few times
625 	 * if we see timeouts in this first AQ call.
626 	 */
627 	do {
628 		ret_code = i40e_aq_get_firmware_version(hw,
629 							&hw->aq.fw_maj_ver,
630 							&hw->aq.fw_min_ver,
631 							&hw->aq.fw_build,
632 							&hw->aq.api_maj_ver,
633 							&hw->aq.api_min_ver,
634 							NULL);
635 		if (ret_code != I40E_ERR_ADMIN_QUEUE_TIMEOUT)
636 			break;
637 		retry++;
638 		i40e_msec_delay(100);
639 		i40e_resume_aq(hw);
640 	} while (retry < 10);
641 	if (ret_code != I40E_SUCCESS)
642 		goto init_adminq_free_arq;
643 
644 	/* get the NVM version info */
645 	i40e_read_nvm_word(hw, I40E_SR_NVM_DEV_STARTER_VERSION,
646 			   &hw->nvm.version);
647 	i40e_read_nvm_word(hw, I40E_SR_NVM_EETRACK_LO, &eetrack_lo);
648 	i40e_read_nvm_word(hw, I40E_SR_NVM_EETRACK_HI, &eetrack_hi);
649 	hw->nvm.eetrack = (eetrack_hi << 16) | eetrack_lo;
650 	i40e_read_nvm_word(hw, I40E_SR_BOOT_CONFIG_PTR, &cfg_ptr);
651 	i40e_read_nvm_word(hw, (cfg_ptr + I40E_NVM_OEM_VER_OFF),
652 			   &oem_hi);
653 	i40e_read_nvm_word(hw, (cfg_ptr + (I40E_NVM_OEM_VER_OFF + 1)),
654 			   &oem_lo);
655 	hw->nvm.oem_ver = ((u32)oem_hi << 16) | oem_lo;
656 
657 	if (hw->aq.api_maj_ver > I40E_FW_API_VERSION_MAJOR) {
658 		ret_code = I40E_ERR_FIRMWARE_API_VERSION;
659 		goto init_adminq_free_arq;
660 	}
661 
662 	/* pre-emptive resource lock release */
663 	i40e_aq_release_resource(hw, I40E_NVM_RESOURCE_ID, 0, NULL);
664 	hw->aq.nvm_release_on_done = FALSE;
665 	hw->nvmupd_state = I40E_NVMUPD_STATE_INIT;
666 
667 	ret_code = i40e_aq_set_hmc_resource_profile(hw,
668 						    I40E_HMC_PROFILE_DEFAULT,
669 						    0,
670 						    NULL);
671 	ret_code = I40E_SUCCESS;
672 
673 	/* success! */
674 	goto init_adminq_exit;
675 
676 init_adminq_free_arq:
677 	i40e_shutdown_arq(hw);
678 init_adminq_free_asq:
679 	i40e_shutdown_asq(hw);
680 init_adminq_destroy_spinlocks:
681 	i40e_destroy_spinlock(&hw->aq.asq_spinlock);
682 	i40e_destroy_spinlock(&hw->aq.arq_spinlock);
683 
684 init_adminq_exit:
685 	return ret_code;
686 }
687 
688 /**
689  *  i40e_shutdown_adminq - shutdown routine for the Admin Queue
690  *  @hw: pointer to the hardware structure
691  **/
692 enum i40e_status_code i40e_shutdown_adminq(struct i40e_hw *hw)
693 {
694 	enum i40e_status_code ret_code = I40E_SUCCESS;
695 
696 	if (i40e_check_asq_alive(hw))
697 		i40e_aq_queue_shutdown(hw, TRUE);
698 
699 	i40e_shutdown_asq(hw);
700 	i40e_shutdown_arq(hw);
701 	i40e_destroy_spinlock(&hw->aq.asq_spinlock);
702 	i40e_destroy_spinlock(&hw->aq.arq_spinlock);
703 
704 	if (hw->nvm_buff.va)
705 		i40e_free_virt_mem(hw, &hw->nvm_buff);
706 
707 	return ret_code;
708 }
709 
710 /**
711  *  i40e_clean_asq - cleans Admin send queue
712  *  @hw: pointer to the hardware structure
713  *
714  *  returns the number of free desc
715  **/
716 u16 i40e_clean_asq(struct i40e_hw *hw)
717 {
718 	struct i40e_adminq_ring *asq = &(hw->aq.asq);
719 	struct i40e_asq_cmd_details *details;
720 	u16 ntc = asq->next_to_clean;
721 	struct i40e_aq_desc desc_cb;
722 	struct i40e_aq_desc *desc;
723 
724 	desc = I40E_ADMINQ_DESC(*asq, ntc);
725 	details = I40E_ADMINQ_DETAILS(*asq, ntc);
726 	while (rd32(hw, hw->aq.asq.head) != ntc) {
727 		i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
728 			   "ntc %d head %d.\n", ntc, rd32(hw, hw->aq.asq.head));
729 
730 		if (details->callback) {
731 			I40E_ADMINQ_CALLBACK cb_func =
732 					(I40E_ADMINQ_CALLBACK)details->callback;
733 			i40e_memcpy(&desc_cb, desc, sizeof(struct i40e_aq_desc),
734 				    I40E_DMA_TO_DMA);
735 			cb_func(hw, &desc_cb);
736 		}
737 		i40e_memset(desc, 0, sizeof(*desc), I40E_DMA_MEM);
738 		i40e_memset(details, 0, sizeof(*details), I40E_NONDMA_MEM);
739 		ntc++;
740 		if (ntc == asq->count)
741 			ntc = 0;
742 		desc = I40E_ADMINQ_DESC(*asq, ntc);
743 		details = I40E_ADMINQ_DETAILS(*asq, ntc);
744 	}
745 
746 	asq->next_to_clean = ntc;
747 
748 	return I40E_DESC_UNUSED(asq);
749 }
750 
751 /**
752  *  i40e_asq_done - check if FW has processed the Admin Send Queue
753  *  @hw: pointer to the hw struct
754  *
755  *  Returns TRUE if the firmware has processed all descriptors on the
756  *  admin send queue. Returns FALSE if there are still requests pending.
757  **/
758 bool i40e_asq_done(struct i40e_hw *hw)
759 {
760 	/* AQ designers suggest use of head for better
761 	 * timing reliability than DD bit
762 	 */
763 	return rd32(hw, hw->aq.asq.head) == hw->aq.asq.next_to_use;
764 
765 }
766 
767 /**
768  *  i40e_asq_send_command - send command to Admin Queue
769  *  @hw: pointer to the hw struct
770  *  @desc: prefilled descriptor describing the command (non DMA mem)
771  *  @buff: buffer to use for indirect commands
772  *  @buff_size: size of buffer for indirect commands
773  *  @cmd_details: pointer to command details structure
774  *
775  *  This is the main send command driver routine for the Admin Queue send
776  *  queue.  It runs the queue, cleans the queue, etc
777  **/
778 enum i40e_status_code i40e_asq_send_command(struct i40e_hw *hw,
779 				struct i40e_aq_desc *desc,
780 				void *buff, /* can be NULL */
781 				u16  buff_size,
782 				struct i40e_asq_cmd_details *cmd_details)
783 {
784 	enum i40e_status_code status = I40E_SUCCESS;
785 	struct i40e_dma_mem *dma_buff = NULL;
786 	struct i40e_asq_cmd_details *details;
787 	struct i40e_aq_desc *desc_on_ring;
788 	bool cmd_completed = FALSE;
789 	u16  retval = 0;
790 	u32  val = 0;
791 
792 	i40e_acquire_spinlock(&hw->aq.asq_spinlock);
793 
794 	hw->aq.asq_last_status = I40E_AQ_RC_OK;
795 
796 	if (hw->aq.asq.count == 0) {
797 		i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
798 			   "AQTX: Admin queue not initialized.\n");
799 		status = I40E_ERR_QUEUE_EMPTY;
800 		goto asq_send_command_error;
801 	}
802 
803 	val = rd32(hw, hw->aq.asq.head);
804 	if (val >= hw->aq.num_asq_entries) {
805 		i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
806 			   "AQTX: head overrun at %d\n", val);
807 		status = I40E_ERR_QUEUE_EMPTY;
808 		goto asq_send_command_error;
809 	}
810 
811 	details = I40E_ADMINQ_DETAILS(hw->aq.asq, hw->aq.asq.next_to_use);
812 	if (cmd_details) {
813 		i40e_memcpy(details,
814 			    cmd_details,
815 			    sizeof(struct i40e_asq_cmd_details),
816 			    I40E_NONDMA_TO_NONDMA);
817 
818 		/* If the cmd_details are defined copy the cookie.  The
819 		 * CPU_TO_LE32 is not needed here because the data is ignored
820 		 * by the FW, only used by the driver
821 		 */
822 		if (details->cookie) {
823 			desc->cookie_high =
824 				CPU_TO_LE32(I40E_HI_DWORD(details->cookie));
825 			desc->cookie_low =
826 				CPU_TO_LE32(I40E_LO_DWORD(details->cookie));
827 		}
828 	} else {
829 		i40e_memset(details, 0,
830 			    sizeof(struct i40e_asq_cmd_details),
831 			    I40E_NONDMA_MEM);
832 	}
833 
834 	/* clear requested flags and then set additional flags if defined */
835 	desc->flags &= ~CPU_TO_LE16(details->flags_dis);
836 	desc->flags |= CPU_TO_LE16(details->flags_ena);
837 
838 	if (buff_size > hw->aq.asq_buf_size) {
839 		i40e_debug(hw,
840 			   I40E_DEBUG_AQ_MESSAGE,
841 			   "AQTX: Invalid buffer size: %d.\n",
842 			   buff_size);
843 		status = I40E_ERR_INVALID_SIZE;
844 		goto asq_send_command_error;
845 	}
846 
847 	if (details->postpone && !details->async) {
848 		i40e_debug(hw,
849 			   I40E_DEBUG_AQ_MESSAGE,
850 			   "AQTX: Async flag not set along with postpone flag");
851 		status = I40E_ERR_PARAM;
852 		goto asq_send_command_error;
853 	}
854 
855 	/* call clean and check queue available function to reclaim the
856 	 * descriptors that were processed by FW, the function returns the
857 	 * number of desc available
858 	 */
859 	/* the clean function called here could be called in a separate thread
860 	 * in case of asynchronous completions
861 	 */
862 	if (i40e_clean_asq(hw) == 0) {
863 		i40e_debug(hw,
864 			   I40E_DEBUG_AQ_MESSAGE,
865 			   "AQTX: Error queue is full.\n");
866 		status = I40E_ERR_ADMIN_QUEUE_FULL;
867 		goto asq_send_command_error;
868 	}
869 
870 	/* initialize the temp desc pointer with the right desc */
871 	desc_on_ring = I40E_ADMINQ_DESC(hw->aq.asq, hw->aq.asq.next_to_use);
872 
873 	/* if the desc is available copy the temp desc to the right place */
874 	i40e_memcpy(desc_on_ring, desc, sizeof(struct i40e_aq_desc),
875 		    I40E_NONDMA_TO_DMA);
876 
877 	/* if buff is not NULL assume indirect command */
878 	if (buff != NULL) {
879 		dma_buff = &(hw->aq.asq.r.asq_bi[hw->aq.asq.next_to_use]);
880 		/* copy the user buff into the respective DMA buff */
881 		i40e_memcpy(dma_buff->va, buff, buff_size,
882 			    I40E_NONDMA_TO_DMA);
883 		desc_on_ring->datalen = CPU_TO_LE16(buff_size);
884 
885 		/* Update the address values in the desc with the pa value
886 		 * for respective buffer
887 		 */
888 		desc_on_ring->params.external.addr_high =
889 				CPU_TO_LE32(I40E_HI_DWORD(dma_buff->pa));
890 		desc_on_ring->params.external.addr_low =
891 				CPU_TO_LE32(I40E_LO_DWORD(dma_buff->pa));
892 	}
893 
894 	/* bump the tail */
895 	i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE, "AQTX: desc and buffer:\n");
896 	i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc_on_ring,
897 		      buff, buff_size);
898 	(hw->aq.asq.next_to_use)++;
899 	if (hw->aq.asq.next_to_use == hw->aq.asq.count)
900 		hw->aq.asq.next_to_use = 0;
901 	if (!details->postpone)
902 		wr32(hw, hw->aq.asq.tail, hw->aq.asq.next_to_use);
903 
904 	/* if cmd_details are not defined or async flag is not set,
905 	 * we need to wait for desc write back
906 	 */
907 	if (!details->async && !details->postpone) {
908 		u32 total_delay = 0;
909 
910 		do {
911 			/* AQ designers suggest use of head for better
912 			 * timing reliability than DD bit
913 			 */
914 			if (i40e_asq_done(hw))
915 				break;
916 			i40e_msec_delay(1);
917 			total_delay++;
918 		} while (total_delay < hw->aq.asq_cmd_timeout);
919 	}
920 
921 	/* if ready, copy the desc back to temp */
922 	if (i40e_asq_done(hw)) {
923 		i40e_memcpy(desc, desc_on_ring, sizeof(struct i40e_aq_desc),
924 			    I40E_DMA_TO_NONDMA);
925 		if (buff != NULL)
926 			i40e_memcpy(buff, dma_buff->va, buff_size,
927 				    I40E_DMA_TO_NONDMA);
928 		retval = LE16_TO_CPU(desc->retval);
929 		if (retval != 0) {
930 			i40e_debug(hw,
931 				   I40E_DEBUG_AQ_MESSAGE,
932 				   "AQTX: Command completed with error 0x%X.\n",
933 				   retval);
934 
935 			/* strip off FW internal code */
936 			retval &= 0xff;
937 		}
938 		cmd_completed = TRUE;
939 		if ((enum i40e_admin_queue_err)retval == I40E_AQ_RC_OK)
940 			status = I40E_SUCCESS;
941 		else
942 			status = I40E_ERR_ADMIN_QUEUE_ERROR;
943 		hw->aq.asq_last_status = (enum i40e_admin_queue_err)retval;
944 	}
945 
946 	i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
947 		   "AQTX: desc and buffer writeback:\n");
948 	i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc, buff, buff_size);
949 
950 	/* save writeback aq if requested */
951 	if (details->wb_desc)
952 		i40e_memcpy(details->wb_desc, desc_on_ring,
953 			    sizeof(struct i40e_aq_desc), I40E_DMA_TO_NONDMA);
954 
955 	/* update the error if time out occurred */
956 	if ((!cmd_completed) &&
957 	    (!details->async && !details->postpone)) {
958 		i40e_debug(hw,
959 			   I40E_DEBUG_AQ_MESSAGE,
960 			   "AQTX: Writeback timeout.\n");
961 		status = I40E_ERR_ADMIN_QUEUE_TIMEOUT;
962 	}
963 
964 asq_send_command_error:
965 	i40e_release_spinlock(&hw->aq.asq_spinlock);
966 	return status;
967 }
968 
969 /**
970  *  i40e_fill_default_direct_cmd_desc - AQ descriptor helper function
971  *  @desc:     pointer to the temp descriptor (non DMA mem)
972  *  @opcode:   the opcode can be used to decide which flags to turn off or on
973  *
974  *  Fill the desc with default values
975  **/
976 void i40e_fill_default_direct_cmd_desc(struct i40e_aq_desc *desc,
977 				       u16 opcode)
978 {
979 	/* zero out the desc */
980 	i40e_memset((void *)desc, 0, sizeof(struct i40e_aq_desc),
981 		    I40E_NONDMA_MEM);
982 	desc->opcode = CPU_TO_LE16(opcode);
983 	desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_SI);
984 }
985 
986 /**
987  *  i40e_clean_arq_element
988  *  @hw: pointer to the hw struct
989  *  @e: event info from the receive descriptor, includes any buffers
990  *  @pending: number of events that could be left to process
991  *
992  *  This function cleans one Admin Receive Queue element and returns
993  *  the contents through e.  It can also return how many events are
994  *  left to process through 'pending'
995  **/
996 enum i40e_status_code i40e_clean_arq_element(struct i40e_hw *hw,
997 					     struct i40e_arq_event_info *e,
998 					     u16 *pending)
999 {
1000 	enum i40e_status_code ret_code = I40E_SUCCESS;
1001 	u16 ntc = hw->aq.arq.next_to_clean;
1002 	struct i40e_aq_desc *desc;
1003 	struct i40e_dma_mem *bi;
1004 	u16 desc_idx;
1005 	u16 datalen;
1006 	u16 flags;
1007 	u16 ntu;
1008 
1009 	/* pre-clean the event info */
1010 	i40e_memset(&e->desc, 0, sizeof(e->desc), I40E_NONDMA_MEM);
1011 
1012 	/* take the lock before we start messing with the ring */
1013 	i40e_acquire_spinlock(&hw->aq.arq_spinlock);
1014 
1015 	if (hw->aq.arq.count == 0) {
1016 		i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
1017 			   "AQRX: Admin queue not initialized.\n");
1018 		ret_code = I40E_ERR_QUEUE_EMPTY;
1019 		goto clean_arq_element_err;
1020 	}
1021 
1022 	/* set next_to_use to head */
1023 	if (!i40e_is_vf(hw))
1024 		ntu = (rd32(hw, hw->aq.arq.head) & I40E_PF_ARQH_ARQH_MASK);
1025 	if (i40e_is_vf(hw))
1026 		ntu = (rd32(hw, hw->aq.arq.head) & I40E_VF_ARQH1_ARQH_MASK);
1027 	if (ntu == ntc) {
1028 		/* nothing to do - shouldn't need to update ring's values */
1029 		ret_code = I40E_ERR_ADMIN_QUEUE_NO_WORK;
1030 		goto clean_arq_element_out;
1031 	}
1032 
1033 	/* now clean the next descriptor */
1034 	desc = I40E_ADMINQ_DESC(hw->aq.arq, ntc);
1035 	desc_idx = ntc;
1036 
1037 	flags = LE16_TO_CPU(desc->flags);
1038 	if (flags & I40E_AQ_FLAG_ERR) {
1039 		ret_code = I40E_ERR_ADMIN_QUEUE_ERROR;
1040 		hw->aq.arq_last_status =
1041 			(enum i40e_admin_queue_err)LE16_TO_CPU(desc->retval);
1042 		i40e_debug(hw,
1043 			   I40E_DEBUG_AQ_MESSAGE,
1044 			   "AQRX: Event received with error 0x%X.\n",
1045 			   hw->aq.arq_last_status);
1046 	}
1047 
1048 	i40e_memcpy(&e->desc, desc, sizeof(struct i40e_aq_desc),
1049 		    I40E_DMA_TO_NONDMA);
1050 	datalen = LE16_TO_CPU(desc->datalen);
1051 	e->msg_len = min(datalen, e->buf_len);
1052 	if (e->msg_buf != NULL && (e->msg_len != 0))
1053 		i40e_memcpy(e->msg_buf,
1054 			    hw->aq.arq.r.arq_bi[desc_idx].va,
1055 			    e->msg_len, I40E_DMA_TO_NONDMA);
1056 
1057 	i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE, "AQRX: desc and buffer:\n");
1058 	i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc, e->msg_buf,
1059 		      hw->aq.arq_buf_size);
1060 
1061 	/* Restore the original datalen and buffer address in the desc,
1062 	 * FW updates datalen to indicate the event message
1063 	 * size
1064 	 */
1065 	bi = &hw->aq.arq.r.arq_bi[ntc];
1066 	i40e_memset((void *)desc, 0, sizeof(struct i40e_aq_desc), I40E_DMA_MEM);
1067 
1068 	desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_BUF);
1069 	if (hw->aq.arq_buf_size > I40E_AQ_LARGE_BUF)
1070 		desc->flags |= CPU_TO_LE16(I40E_AQ_FLAG_LB);
1071 	desc->datalen = CPU_TO_LE16((u16)bi->size);
1072 	desc->params.external.addr_high = CPU_TO_LE32(I40E_HI_DWORD(bi->pa));
1073 	desc->params.external.addr_low = CPU_TO_LE32(I40E_LO_DWORD(bi->pa));
1074 
1075 	/* set tail = the last cleaned desc index. */
1076 	wr32(hw, hw->aq.arq.tail, ntc);
1077 	/* ntc is updated to tail + 1 */
1078 	ntc++;
1079 	if (ntc == hw->aq.num_arq_entries)
1080 		ntc = 0;
1081 	hw->aq.arq.next_to_clean = ntc;
1082 	hw->aq.arq.next_to_use = ntu;
1083 
1084 	if (i40e_is_nvm_update_op(&e->desc)) {
1085 		if (hw->aq.nvm_release_on_done) {
1086 			i40e_release_nvm(hw);
1087 			hw->aq.nvm_release_on_done = FALSE;
1088 		}
1089 
1090 		switch (hw->nvmupd_state) {
1091 		case I40E_NVMUPD_STATE_INIT_WAIT:
1092 			hw->nvmupd_state = I40E_NVMUPD_STATE_INIT;
1093 			break;
1094 
1095 		case I40E_NVMUPD_STATE_WRITE_WAIT:
1096 			hw->nvmupd_state = I40E_NVMUPD_STATE_WRITING;
1097 			break;
1098 
1099 		default:
1100 			break;
1101 		}
1102 	}
1103 
1104 clean_arq_element_out:
1105 	/* Set pending if needed, unlock and return */
1106 	if (pending != NULL)
1107 		*pending = (ntc > ntu ? hw->aq.arq.count : 0) + (ntu - ntc);
1108 clean_arq_element_err:
1109 	i40e_release_spinlock(&hw->aq.arq_spinlock);
1110 
1111 	return ret_code;
1112 }
1113 
1114