1 /****************************************************************************** 2 3 Copyright (c) 2013-2018, Intel Corporation 4 All rights reserved. 5 6 Redistribution and use in source and binary forms, with or without 7 modification, are permitted provided that the following conditions are met: 8 9 1. Redistributions of source code must retain the above copyright notice, 10 this list of conditions and the following disclaimer. 11 12 2. Redistributions in binary form must reproduce the above copyright 13 notice, this list of conditions and the following disclaimer in the 14 documentation and/or other materials provided with the distribution. 15 16 3. Neither the name of the Intel Corporation nor the names of its 17 contributors may be used to endorse or promote products derived from 18 this software without specific prior written permission. 19 20 THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" 21 AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 22 IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 23 ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE 24 LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 25 CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 26 SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 27 INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 28 CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 29 ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 30 POSSIBILITY OF SUCH DAMAGE. 31 32 ******************************************************************************/ 33 /*$FreeBSD$*/ 34 35 #include "i40e_status.h" 36 #include "i40e_type.h" 37 #include "i40e_register.h" 38 #include "i40e_adminq.h" 39 #include "i40e_prototype.h" 40 41 /** 42 * i40e_adminq_init_regs - Initialize AdminQ registers 43 * @hw: pointer to the hardware structure 44 * 45 * This assumes the alloc_asq and alloc_arq functions have already been called 46 **/ 47 static void i40e_adminq_init_regs(struct i40e_hw *hw) 48 { 49 /* set head and tail registers in our local struct */ 50 if (i40e_is_vf(hw)) { 51 hw->aq.asq.tail = I40E_VF_ATQT1; 52 hw->aq.asq.head = I40E_VF_ATQH1; 53 hw->aq.asq.len = I40E_VF_ATQLEN1; 54 hw->aq.asq.bal = I40E_VF_ATQBAL1; 55 hw->aq.asq.bah = I40E_VF_ATQBAH1; 56 hw->aq.arq.tail = I40E_VF_ARQT1; 57 hw->aq.arq.head = I40E_VF_ARQH1; 58 hw->aq.arq.len = I40E_VF_ARQLEN1; 59 hw->aq.arq.bal = I40E_VF_ARQBAL1; 60 hw->aq.arq.bah = I40E_VF_ARQBAH1; 61 } else { 62 hw->aq.asq.tail = I40E_PF_ATQT; 63 hw->aq.asq.head = I40E_PF_ATQH; 64 hw->aq.asq.len = I40E_PF_ATQLEN; 65 hw->aq.asq.bal = I40E_PF_ATQBAL; 66 hw->aq.asq.bah = I40E_PF_ATQBAH; 67 hw->aq.arq.tail = I40E_PF_ARQT; 68 hw->aq.arq.head = I40E_PF_ARQH; 69 hw->aq.arq.len = I40E_PF_ARQLEN; 70 hw->aq.arq.bal = I40E_PF_ARQBAL; 71 hw->aq.arq.bah = I40E_PF_ARQBAH; 72 } 73 } 74 75 /** 76 * i40e_alloc_adminq_asq_ring - Allocate Admin Queue send rings 77 * @hw: pointer to the hardware structure 78 **/ 79 enum i40e_status_code i40e_alloc_adminq_asq_ring(struct i40e_hw *hw) 80 { 81 enum i40e_status_code ret_code; 82 83 ret_code = i40e_allocate_dma_mem(hw, &hw->aq.asq.desc_buf, 84 i40e_mem_atq_ring, 85 (hw->aq.num_asq_entries * 86 sizeof(struct i40e_aq_desc)), 87 I40E_ADMINQ_DESC_ALIGNMENT); 88 if (ret_code) 89 return ret_code; 90 91 ret_code = i40e_allocate_virt_mem(hw, &hw->aq.asq.cmd_buf, 92 (hw->aq.num_asq_entries * 93 sizeof(struct i40e_asq_cmd_details))); 94 if (ret_code) { 95 i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf); 96 return ret_code; 97 } 98 99 return ret_code; 100 } 101 102 /** 103 * i40e_alloc_adminq_arq_ring - Allocate Admin Queue receive rings 104 * @hw: pointer to the hardware structure 105 **/ 106 enum i40e_status_code i40e_alloc_adminq_arq_ring(struct i40e_hw *hw) 107 { 108 enum i40e_status_code ret_code; 109 110 ret_code = i40e_allocate_dma_mem(hw, &hw->aq.arq.desc_buf, 111 i40e_mem_arq_ring, 112 (hw->aq.num_arq_entries * 113 sizeof(struct i40e_aq_desc)), 114 I40E_ADMINQ_DESC_ALIGNMENT); 115 116 return ret_code; 117 } 118 119 /** 120 * i40e_free_adminq_asq - Free Admin Queue send rings 121 * @hw: pointer to the hardware structure 122 * 123 * This assumes the posted send buffers have already been cleaned 124 * and de-allocated 125 **/ 126 void i40e_free_adminq_asq(struct i40e_hw *hw) 127 { 128 i40e_free_virt_mem(hw, &hw->aq.asq.cmd_buf); 129 i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf); 130 } 131 132 /** 133 * i40e_free_adminq_arq - Free Admin Queue receive rings 134 * @hw: pointer to the hardware structure 135 * 136 * This assumes the posted receive buffers have already been cleaned 137 * and de-allocated 138 **/ 139 void i40e_free_adminq_arq(struct i40e_hw *hw) 140 { 141 i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf); 142 } 143 144 /** 145 * i40e_alloc_arq_bufs - Allocate pre-posted buffers for the receive queue 146 * @hw: pointer to the hardware structure 147 **/ 148 static enum i40e_status_code i40e_alloc_arq_bufs(struct i40e_hw *hw) 149 { 150 enum i40e_status_code ret_code; 151 struct i40e_aq_desc *desc; 152 struct i40e_dma_mem *bi; 153 int i; 154 155 /* We'll be allocating the buffer info memory first, then we can 156 * allocate the mapped buffers for the event processing 157 */ 158 159 /* buffer_info structures do not need alignment */ 160 ret_code = i40e_allocate_virt_mem(hw, &hw->aq.arq.dma_head, 161 (hw->aq.num_arq_entries * sizeof(struct i40e_dma_mem))); 162 if (ret_code) 163 goto alloc_arq_bufs; 164 hw->aq.arq.r.arq_bi = (struct i40e_dma_mem *)hw->aq.arq.dma_head.va; 165 166 /* allocate the mapped buffers */ 167 for (i = 0; i < hw->aq.num_arq_entries; i++) { 168 bi = &hw->aq.arq.r.arq_bi[i]; 169 ret_code = i40e_allocate_dma_mem(hw, bi, 170 i40e_mem_arq_buf, 171 hw->aq.arq_buf_size, 172 I40E_ADMINQ_DESC_ALIGNMENT); 173 if (ret_code) 174 goto unwind_alloc_arq_bufs; 175 176 /* now configure the descriptors for use */ 177 desc = I40E_ADMINQ_DESC(hw->aq.arq, i); 178 179 desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_BUF); 180 if (hw->aq.arq_buf_size > I40E_AQ_LARGE_BUF) 181 desc->flags |= CPU_TO_LE16(I40E_AQ_FLAG_LB); 182 desc->opcode = 0; 183 /* This is in accordance with Admin queue design, there is no 184 * register for buffer size configuration 185 */ 186 desc->datalen = CPU_TO_LE16((u16)bi->size); 187 desc->retval = 0; 188 desc->cookie_high = 0; 189 desc->cookie_low = 0; 190 desc->params.external.addr_high = 191 CPU_TO_LE32(I40E_HI_DWORD(bi->pa)); 192 desc->params.external.addr_low = 193 CPU_TO_LE32(I40E_LO_DWORD(bi->pa)); 194 desc->params.external.param0 = 0; 195 desc->params.external.param1 = 0; 196 } 197 198 alloc_arq_bufs: 199 return ret_code; 200 201 unwind_alloc_arq_bufs: 202 /* don't try to free the one that failed... */ 203 i--; 204 for (; i >= 0; i--) 205 i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]); 206 i40e_free_virt_mem(hw, &hw->aq.arq.dma_head); 207 208 return ret_code; 209 } 210 211 /** 212 * i40e_alloc_asq_bufs - Allocate empty buffer structs for the send queue 213 * @hw: pointer to the hardware structure 214 **/ 215 static enum i40e_status_code i40e_alloc_asq_bufs(struct i40e_hw *hw) 216 { 217 enum i40e_status_code ret_code; 218 struct i40e_dma_mem *bi; 219 int i; 220 221 /* No mapped memory needed yet, just the buffer info structures */ 222 ret_code = i40e_allocate_virt_mem(hw, &hw->aq.asq.dma_head, 223 (hw->aq.num_asq_entries * sizeof(struct i40e_dma_mem))); 224 if (ret_code) 225 goto alloc_asq_bufs; 226 hw->aq.asq.r.asq_bi = (struct i40e_dma_mem *)hw->aq.asq.dma_head.va; 227 228 /* allocate the mapped buffers */ 229 for (i = 0; i < hw->aq.num_asq_entries; i++) { 230 bi = &hw->aq.asq.r.asq_bi[i]; 231 ret_code = i40e_allocate_dma_mem(hw, bi, 232 i40e_mem_asq_buf, 233 hw->aq.asq_buf_size, 234 I40E_ADMINQ_DESC_ALIGNMENT); 235 if (ret_code) 236 goto unwind_alloc_asq_bufs; 237 } 238 alloc_asq_bufs: 239 return ret_code; 240 241 unwind_alloc_asq_bufs: 242 /* don't try to free the one that failed... */ 243 i--; 244 for (; i >= 0; i--) 245 i40e_free_dma_mem(hw, &hw->aq.asq.r.asq_bi[i]); 246 i40e_free_virt_mem(hw, &hw->aq.asq.dma_head); 247 248 return ret_code; 249 } 250 251 /** 252 * i40e_free_arq_bufs - Free receive queue buffer info elements 253 * @hw: pointer to the hardware structure 254 **/ 255 static void i40e_free_arq_bufs(struct i40e_hw *hw) 256 { 257 int i; 258 259 /* free descriptors */ 260 for (i = 0; i < hw->aq.num_arq_entries; i++) 261 i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]); 262 263 /* free the descriptor memory */ 264 i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf); 265 266 /* free the dma header */ 267 i40e_free_virt_mem(hw, &hw->aq.arq.dma_head); 268 } 269 270 /** 271 * i40e_free_asq_bufs - Free send queue buffer info elements 272 * @hw: pointer to the hardware structure 273 **/ 274 static void i40e_free_asq_bufs(struct i40e_hw *hw) 275 { 276 int i; 277 278 /* only unmap if the address is non-NULL */ 279 for (i = 0; i < hw->aq.num_asq_entries; i++) 280 if (hw->aq.asq.r.asq_bi[i].pa) 281 i40e_free_dma_mem(hw, &hw->aq.asq.r.asq_bi[i]); 282 283 /* free the buffer info list */ 284 i40e_free_virt_mem(hw, &hw->aq.asq.cmd_buf); 285 286 /* free the descriptor memory */ 287 i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf); 288 289 /* free the dma header */ 290 i40e_free_virt_mem(hw, &hw->aq.asq.dma_head); 291 } 292 293 /** 294 * i40e_config_asq_regs - configure ASQ registers 295 * @hw: pointer to the hardware structure 296 * 297 * Configure base address and length registers for the transmit queue 298 **/ 299 static enum i40e_status_code i40e_config_asq_regs(struct i40e_hw *hw) 300 { 301 enum i40e_status_code ret_code = I40E_SUCCESS; 302 u32 reg = 0; 303 304 /* Clear Head and Tail */ 305 wr32(hw, hw->aq.asq.head, 0); 306 wr32(hw, hw->aq.asq.tail, 0); 307 308 /* set starting point */ 309 if (!i40e_is_vf(hw)) 310 wr32(hw, hw->aq.asq.len, (hw->aq.num_asq_entries | 311 I40E_PF_ATQLEN_ATQENABLE_MASK)); 312 if (i40e_is_vf(hw)) 313 wr32(hw, hw->aq.asq.len, (hw->aq.num_asq_entries | 314 I40E_VF_ATQLEN1_ATQENABLE_MASK)); 315 wr32(hw, hw->aq.asq.bal, I40E_LO_DWORD(hw->aq.asq.desc_buf.pa)); 316 wr32(hw, hw->aq.asq.bah, I40E_HI_DWORD(hw->aq.asq.desc_buf.pa)); 317 318 /* Check one register to verify that config was applied */ 319 reg = rd32(hw, hw->aq.asq.bal); 320 if (reg != I40E_LO_DWORD(hw->aq.asq.desc_buf.pa)) 321 ret_code = I40E_ERR_ADMIN_QUEUE_ERROR; 322 323 return ret_code; 324 } 325 326 /** 327 * i40e_config_arq_regs - ARQ register configuration 328 * @hw: pointer to the hardware structure 329 * 330 * Configure base address and length registers for the receive (event queue) 331 **/ 332 static enum i40e_status_code i40e_config_arq_regs(struct i40e_hw *hw) 333 { 334 enum i40e_status_code ret_code = I40E_SUCCESS; 335 u32 reg = 0; 336 337 /* Clear Head and Tail */ 338 wr32(hw, hw->aq.arq.head, 0); 339 wr32(hw, hw->aq.arq.tail, 0); 340 341 /* set starting point */ 342 if (!i40e_is_vf(hw)) 343 wr32(hw, hw->aq.arq.len, (hw->aq.num_arq_entries | 344 I40E_PF_ARQLEN_ARQENABLE_MASK)); 345 if (i40e_is_vf(hw)) 346 wr32(hw, hw->aq.arq.len, (hw->aq.num_arq_entries | 347 I40E_VF_ARQLEN1_ARQENABLE_MASK)); 348 wr32(hw, hw->aq.arq.bal, I40E_LO_DWORD(hw->aq.arq.desc_buf.pa)); 349 wr32(hw, hw->aq.arq.bah, I40E_HI_DWORD(hw->aq.arq.desc_buf.pa)); 350 351 /* Update tail in the HW to post pre-allocated buffers */ 352 wr32(hw, hw->aq.arq.tail, hw->aq.num_arq_entries - 1); 353 354 /* Check one register to verify that config was applied */ 355 reg = rd32(hw, hw->aq.arq.bal); 356 if (reg != I40E_LO_DWORD(hw->aq.arq.desc_buf.pa)) 357 ret_code = I40E_ERR_ADMIN_QUEUE_ERROR; 358 359 return ret_code; 360 } 361 362 /** 363 * i40e_init_asq - main initialization routine for ASQ 364 * @hw: pointer to the hardware structure 365 * 366 * This is the main initialization routine for the Admin Send Queue 367 * Prior to calling this function, drivers *MUST* set the following fields 368 * in the hw->aq structure: 369 * - hw->aq.num_asq_entries 370 * - hw->aq.arq_buf_size 371 * 372 * Do *NOT* hold the lock when calling this as the memory allocation routines 373 * called are not going to be atomic context safe 374 **/ 375 enum i40e_status_code i40e_init_asq(struct i40e_hw *hw) 376 { 377 enum i40e_status_code ret_code = I40E_SUCCESS; 378 379 if (hw->aq.asq.count > 0) { 380 /* queue already initialized */ 381 ret_code = I40E_ERR_NOT_READY; 382 goto init_adminq_exit; 383 } 384 385 /* verify input for valid configuration */ 386 if ((hw->aq.num_asq_entries == 0) || 387 (hw->aq.asq_buf_size == 0)) { 388 ret_code = I40E_ERR_CONFIG; 389 goto init_adminq_exit; 390 } 391 392 hw->aq.asq.next_to_use = 0; 393 hw->aq.asq.next_to_clean = 0; 394 395 /* allocate the ring memory */ 396 ret_code = i40e_alloc_adminq_asq_ring(hw); 397 if (ret_code != I40E_SUCCESS) 398 goto init_adminq_exit; 399 400 /* allocate buffers in the rings */ 401 ret_code = i40e_alloc_asq_bufs(hw); 402 if (ret_code != I40E_SUCCESS) 403 goto init_adminq_free_rings; 404 405 /* initialize base registers */ 406 ret_code = i40e_config_asq_regs(hw); 407 if (ret_code != I40E_SUCCESS) 408 goto init_config_regs; 409 410 /* success! */ 411 hw->aq.asq.count = hw->aq.num_asq_entries; 412 goto init_adminq_exit; 413 414 init_adminq_free_rings: 415 i40e_free_adminq_asq(hw); 416 return ret_code; 417 418 init_config_regs: 419 i40e_free_asq_bufs(hw); 420 421 init_adminq_exit: 422 return ret_code; 423 } 424 425 /** 426 * i40e_init_arq - initialize ARQ 427 * @hw: pointer to the hardware structure 428 * 429 * The main initialization routine for the Admin Receive (Event) Queue. 430 * Prior to calling this function, drivers *MUST* set the following fields 431 * in the hw->aq structure: 432 * - hw->aq.num_asq_entries 433 * - hw->aq.arq_buf_size 434 * 435 * Do *NOT* hold the lock when calling this as the memory allocation routines 436 * called are not going to be atomic context safe 437 **/ 438 enum i40e_status_code i40e_init_arq(struct i40e_hw *hw) 439 { 440 enum i40e_status_code ret_code = I40E_SUCCESS; 441 442 if (hw->aq.arq.count > 0) { 443 /* queue already initialized */ 444 ret_code = I40E_ERR_NOT_READY; 445 goto init_adminq_exit; 446 } 447 448 /* verify input for valid configuration */ 449 if ((hw->aq.num_arq_entries == 0) || 450 (hw->aq.arq_buf_size == 0)) { 451 ret_code = I40E_ERR_CONFIG; 452 goto init_adminq_exit; 453 } 454 455 hw->aq.arq.next_to_use = 0; 456 hw->aq.arq.next_to_clean = 0; 457 458 /* allocate the ring memory */ 459 ret_code = i40e_alloc_adminq_arq_ring(hw); 460 if (ret_code != I40E_SUCCESS) 461 goto init_adminq_exit; 462 463 /* allocate buffers in the rings */ 464 ret_code = i40e_alloc_arq_bufs(hw); 465 if (ret_code != I40E_SUCCESS) 466 goto init_adminq_free_rings; 467 468 /* initialize base registers */ 469 ret_code = i40e_config_arq_regs(hw); 470 if (ret_code != I40E_SUCCESS) 471 goto init_adminq_free_rings; 472 473 /* success! */ 474 hw->aq.arq.count = hw->aq.num_arq_entries; 475 goto init_adminq_exit; 476 477 init_adminq_free_rings: 478 i40e_free_adminq_arq(hw); 479 480 init_adminq_exit: 481 return ret_code; 482 } 483 484 /** 485 * i40e_shutdown_asq - shutdown the ASQ 486 * @hw: pointer to the hardware structure 487 * 488 * The main shutdown routine for the Admin Send Queue 489 **/ 490 enum i40e_status_code i40e_shutdown_asq(struct i40e_hw *hw) 491 { 492 enum i40e_status_code ret_code = I40E_SUCCESS; 493 494 i40e_acquire_spinlock(&hw->aq.asq_spinlock); 495 496 if (hw->aq.asq.count == 0) { 497 ret_code = I40E_ERR_NOT_READY; 498 goto shutdown_asq_out; 499 } 500 501 /* Stop firmware AdminQ processing */ 502 wr32(hw, hw->aq.asq.head, 0); 503 wr32(hw, hw->aq.asq.tail, 0); 504 wr32(hw, hw->aq.asq.len, 0); 505 wr32(hw, hw->aq.asq.bal, 0); 506 wr32(hw, hw->aq.asq.bah, 0); 507 508 hw->aq.asq.count = 0; /* to indicate uninitialized queue */ 509 510 /* free ring buffers */ 511 i40e_free_asq_bufs(hw); 512 513 shutdown_asq_out: 514 i40e_release_spinlock(&hw->aq.asq_spinlock); 515 return ret_code; 516 } 517 518 /** 519 * i40e_shutdown_arq - shutdown ARQ 520 * @hw: pointer to the hardware structure 521 * 522 * The main shutdown routine for the Admin Receive Queue 523 **/ 524 enum i40e_status_code i40e_shutdown_arq(struct i40e_hw *hw) 525 { 526 enum i40e_status_code ret_code = I40E_SUCCESS; 527 528 i40e_acquire_spinlock(&hw->aq.arq_spinlock); 529 530 if (hw->aq.arq.count == 0) { 531 ret_code = I40E_ERR_NOT_READY; 532 goto shutdown_arq_out; 533 } 534 535 /* Stop firmware AdminQ processing */ 536 wr32(hw, hw->aq.arq.head, 0); 537 wr32(hw, hw->aq.arq.tail, 0); 538 wr32(hw, hw->aq.arq.len, 0); 539 wr32(hw, hw->aq.arq.bal, 0); 540 wr32(hw, hw->aq.arq.bah, 0); 541 542 hw->aq.arq.count = 0; /* to indicate uninitialized queue */ 543 544 /* free ring buffers */ 545 i40e_free_arq_bufs(hw); 546 547 shutdown_arq_out: 548 i40e_release_spinlock(&hw->aq.arq_spinlock); 549 return ret_code; 550 } 551 552 /** 553 * i40e_resume_aq - resume AQ processing from 0 554 * @hw: pointer to the hardware structure 555 **/ 556 static void i40e_resume_aq(struct i40e_hw *hw) 557 { 558 /* Registers are reset after PF reset */ 559 hw->aq.asq.next_to_use = 0; 560 hw->aq.asq.next_to_clean = 0; 561 562 i40e_config_asq_regs(hw); 563 564 hw->aq.arq.next_to_use = 0; 565 hw->aq.arq.next_to_clean = 0; 566 567 i40e_config_arq_regs(hw); 568 } 569 570 /** 571 * i40e_set_hw_flags - set HW flags 572 * @hw: pointer to the hardware structure 573 **/ 574 static void i40e_set_hw_flags(struct i40e_hw *hw) 575 { 576 struct i40e_adminq_info *aq = &hw->aq; 577 578 hw->flags = 0; 579 580 switch (hw->mac.type) { 581 case I40E_MAC_XL710: 582 if (aq->api_maj_ver > 1 || 583 (aq->api_maj_ver == 1 && 584 aq->api_min_ver >= I40E_MINOR_VER_GET_LINK_INFO_XL710)) { 585 hw->flags |= I40E_HW_FLAG_AQ_PHY_ACCESS_CAPABLE; 586 hw->flags |= I40E_HW_FLAG_FW_LLDP_STOPPABLE; 587 /* The ability to RX (not drop) 802.1ad frames */ 588 hw->flags |= I40E_HW_FLAG_802_1AD_CAPABLE; 589 } 590 break; 591 case I40E_MAC_X722: 592 hw->flags |= I40E_HW_FLAG_AQ_SRCTL_ACCESS_ENABLE | 593 I40E_HW_FLAG_NVM_READ_REQUIRES_LOCK; 594 595 if (aq->api_maj_ver > 1 || 596 (aq->api_maj_ver == 1 && 597 aq->api_min_ver >= I40E_MINOR_VER_FW_LLDP_STOPPABLE_X722)) 598 hw->flags |= I40E_HW_FLAG_FW_LLDP_STOPPABLE; 599 600 if (aq->api_maj_ver > 1 || 601 (aq->api_maj_ver == 1 && 602 aq->api_min_ver >= I40E_MINOR_VER_GET_LINK_INFO_X722)) 603 hw->flags |= I40E_HW_FLAG_AQ_PHY_ACCESS_CAPABLE; 604 605 if (aq->api_maj_ver > 1 || 606 (aq->api_maj_ver == 1 && 607 aq->api_min_ver >= I40E_MINOR_VER_FW_REQUEST_FEC_X722)) 608 hw->flags |= I40E_HW_FLAG_X722_FEC_REQUEST_CAPABLE; 609 610 /* fall through */ 611 default: 612 break; 613 } 614 615 /* Newer versions of firmware require lock when reading the NVM */ 616 if (aq->api_maj_ver > 1 || 617 (aq->api_maj_ver == 1 && 618 aq->api_min_ver >= 5)) 619 hw->flags |= I40E_HW_FLAG_NVM_READ_REQUIRES_LOCK; 620 621 if (aq->api_maj_ver > 1 || 622 (aq->api_maj_ver == 1 && 623 aq->api_min_ver >= 8)) { 624 hw->flags |= I40E_HW_FLAG_FW_LLDP_PERSISTENT; 625 hw->flags |= I40E_HW_FLAG_DROP_MODE; 626 } 627 628 if (aq->api_maj_ver > 1 || 629 (aq->api_maj_ver == 1 && 630 aq->api_min_ver >= 9)) 631 hw->flags |= I40E_HW_FLAG_AQ_PHY_ACCESS_EXTENDED; 632 } 633 634 /** 635 * i40e_init_adminq - main initialization routine for Admin Queue 636 * @hw: pointer to the hardware structure 637 * 638 * Prior to calling this function, drivers *MUST* set the following fields 639 * in the hw->aq structure: 640 * - hw->aq.num_asq_entries 641 * - hw->aq.num_arq_entries 642 * - hw->aq.arq_buf_size 643 * - hw->aq.asq_buf_size 644 **/ 645 enum i40e_status_code i40e_init_adminq(struct i40e_hw *hw) 646 { 647 struct i40e_adminq_info *aq = &hw->aq; 648 enum i40e_status_code ret_code; 649 u16 cfg_ptr, oem_hi, oem_lo; 650 u16 eetrack_lo, eetrack_hi; 651 int retry = 0; 652 653 /* verify input for valid configuration */ 654 if (aq->num_arq_entries == 0 || 655 aq->num_asq_entries == 0 || 656 aq->arq_buf_size == 0 || 657 aq->asq_buf_size == 0) { 658 ret_code = I40E_ERR_CONFIG; 659 goto init_adminq_exit; 660 } 661 i40e_init_spinlock(&aq->asq_spinlock); 662 i40e_init_spinlock(&aq->arq_spinlock); 663 664 /* Set up register offsets */ 665 i40e_adminq_init_regs(hw); 666 667 /* setup ASQ command write back timeout */ 668 hw->aq.asq_cmd_timeout = I40E_ASQ_CMD_TIMEOUT; 669 670 /* allocate the ASQ */ 671 ret_code = i40e_init_asq(hw); 672 if (ret_code != I40E_SUCCESS) 673 goto init_adminq_destroy_spinlocks; 674 675 /* allocate the ARQ */ 676 ret_code = i40e_init_arq(hw); 677 if (ret_code != I40E_SUCCESS) 678 goto init_adminq_free_asq; 679 680 /* VF has no need of firmware */ 681 if (i40e_is_vf(hw)) 682 goto init_adminq_exit; 683 /* There are some cases where the firmware may not be quite ready 684 * for AdminQ operations, so we retry the AdminQ setup a few times 685 * if we see timeouts in this first AQ call. 686 */ 687 do { 688 ret_code = i40e_aq_get_firmware_version(hw, 689 &aq->fw_maj_ver, 690 &aq->fw_min_ver, 691 &aq->fw_build, 692 &aq->api_maj_ver, 693 &aq->api_min_ver, 694 NULL); 695 if (ret_code != I40E_ERR_ADMIN_QUEUE_TIMEOUT) 696 break; 697 retry++; 698 i40e_msec_delay(100); 699 i40e_resume_aq(hw); 700 } while (retry < 10); 701 if (ret_code != I40E_SUCCESS) 702 goto init_adminq_free_arq; 703 704 /* 705 * Some features were introduced in different FW API version 706 * for different MAC type. 707 */ 708 i40e_set_hw_flags(hw); 709 710 /* get the NVM version info */ 711 i40e_read_nvm_word(hw, I40E_SR_NVM_DEV_STARTER_VERSION, 712 &hw->nvm.version); 713 i40e_read_nvm_word(hw, I40E_SR_NVM_EETRACK_LO, &eetrack_lo); 714 i40e_read_nvm_word(hw, I40E_SR_NVM_EETRACK_HI, &eetrack_hi); 715 hw->nvm.eetrack = (eetrack_hi << 16) | eetrack_lo; 716 i40e_read_nvm_word(hw, I40E_SR_BOOT_CONFIG_PTR, &cfg_ptr); 717 i40e_read_nvm_word(hw, (cfg_ptr + I40E_NVM_OEM_VER_OFF), 718 &oem_hi); 719 i40e_read_nvm_word(hw, (cfg_ptr + (I40E_NVM_OEM_VER_OFF + 1)), 720 &oem_lo); 721 hw->nvm.oem_ver = ((u32)oem_hi << 16) | oem_lo; 722 723 if (aq->api_maj_ver > I40E_FW_API_VERSION_MAJOR) { 724 ret_code = I40E_ERR_FIRMWARE_API_VERSION; 725 goto init_adminq_free_arq; 726 } 727 728 /* pre-emptive resource lock release */ 729 i40e_aq_release_resource(hw, I40E_NVM_RESOURCE_ID, 0, NULL); 730 hw->nvm_release_on_done = FALSE; 731 hw->nvmupd_state = I40E_NVMUPD_STATE_INIT; 732 733 ret_code = I40E_SUCCESS; 734 735 /* success! */ 736 goto init_adminq_exit; 737 738 init_adminq_free_arq: 739 i40e_shutdown_arq(hw); 740 init_adminq_free_asq: 741 i40e_shutdown_asq(hw); 742 init_adminq_destroy_spinlocks: 743 i40e_destroy_spinlock(&aq->asq_spinlock); 744 i40e_destroy_spinlock(&aq->arq_spinlock); 745 746 init_adminq_exit: 747 return ret_code; 748 } 749 750 /** 751 * i40e_shutdown_adminq - shutdown routine for the Admin Queue 752 * @hw: pointer to the hardware structure 753 **/ 754 enum i40e_status_code i40e_shutdown_adminq(struct i40e_hw *hw) 755 { 756 enum i40e_status_code ret_code = I40E_SUCCESS; 757 758 if (i40e_check_asq_alive(hw)) 759 i40e_aq_queue_shutdown(hw, TRUE); 760 761 i40e_shutdown_asq(hw); 762 i40e_shutdown_arq(hw); 763 i40e_destroy_spinlock(&hw->aq.asq_spinlock); 764 i40e_destroy_spinlock(&hw->aq.arq_spinlock); 765 766 if (hw->nvm_buff.va) 767 i40e_free_virt_mem(hw, &hw->nvm_buff); 768 769 return ret_code; 770 } 771 772 /** 773 * i40e_clean_asq - cleans Admin send queue 774 * @hw: pointer to the hardware structure 775 * 776 * returns the number of free desc 777 **/ 778 u16 i40e_clean_asq(struct i40e_hw *hw) 779 { 780 struct i40e_adminq_ring *asq = &(hw->aq.asq); 781 struct i40e_asq_cmd_details *details; 782 u16 ntc = asq->next_to_clean; 783 struct i40e_aq_desc desc_cb; 784 struct i40e_aq_desc *desc; 785 786 desc = I40E_ADMINQ_DESC(*asq, ntc); 787 details = I40E_ADMINQ_DETAILS(*asq, ntc); 788 while (rd32(hw, hw->aq.asq.head) != ntc) { 789 i40e_debug(hw, I40E_DEBUG_AQ_COMMAND, 790 "ntc %d head %d.\n", ntc, rd32(hw, hw->aq.asq.head)); 791 792 if (details->callback) { 793 I40E_ADMINQ_CALLBACK cb_func = 794 (I40E_ADMINQ_CALLBACK)details->callback; 795 i40e_memcpy(&desc_cb, desc, sizeof(struct i40e_aq_desc), 796 I40E_DMA_TO_DMA); 797 cb_func(hw, &desc_cb); 798 } 799 i40e_memset(desc, 0, sizeof(*desc), I40E_DMA_MEM); 800 i40e_memset(details, 0, sizeof(*details), I40E_NONDMA_MEM); 801 ntc++; 802 if (ntc == asq->count) 803 ntc = 0; 804 desc = I40E_ADMINQ_DESC(*asq, ntc); 805 details = I40E_ADMINQ_DETAILS(*asq, ntc); 806 } 807 808 asq->next_to_clean = ntc; 809 810 return I40E_DESC_UNUSED(asq); 811 } 812 813 /** 814 * i40e_asq_done - check if FW has processed the Admin Send Queue 815 * @hw: pointer to the hw struct 816 * 817 * Returns TRUE if the firmware has processed all descriptors on the 818 * admin send queue. Returns FALSE if there are still requests pending. 819 **/ 820 bool i40e_asq_done(struct i40e_hw *hw) 821 { 822 /* AQ designers suggest use of head for better 823 * timing reliability than DD bit 824 */ 825 return rd32(hw, hw->aq.asq.head) == hw->aq.asq.next_to_use; 826 827 } 828 829 /** 830 * i40e_asq_send_command - send command to Admin Queue 831 * @hw: pointer to the hw struct 832 * @desc: prefilled descriptor describing the command (non DMA mem) 833 * @buff: buffer to use for indirect commands 834 * @buff_size: size of buffer for indirect commands 835 * @cmd_details: pointer to command details structure 836 * 837 * This is the main send command driver routine for the Admin Queue send 838 * queue. It runs the queue, cleans the queue, etc 839 **/ 840 enum i40e_status_code i40e_asq_send_command(struct i40e_hw *hw, 841 struct i40e_aq_desc *desc, 842 void *buff, /* can be NULL */ 843 u16 buff_size, 844 struct i40e_asq_cmd_details *cmd_details) 845 { 846 enum i40e_status_code status = I40E_SUCCESS; 847 struct i40e_dma_mem *dma_buff = NULL; 848 struct i40e_asq_cmd_details *details; 849 struct i40e_aq_desc *desc_on_ring; 850 bool cmd_completed = FALSE; 851 u16 retval = 0; 852 u32 val = 0; 853 854 i40e_acquire_spinlock(&hw->aq.asq_spinlock); 855 856 hw->aq.asq_last_status = I40E_AQ_RC_OK; 857 858 if (hw->aq.asq.count == 0) { 859 i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE, 860 "AQTX: Admin queue not initialized.\n"); 861 status = I40E_ERR_QUEUE_EMPTY; 862 goto asq_send_command_error; 863 } 864 865 val = rd32(hw, hw->aq.asq.head); 866 if (val >= hw->aq.num_asq_entries) { 867 i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE, 868 "AQTX: head overrun at %d\n", val); 869 status = I40E_ERR_ADMIN_QUEUE_FULL; 870 goto asq_send_command_error; 871 } 872 873 details = I40E_ADMINQ_DETAILS(hw->aq.asq, hw->aq.asq.next_to_use); 874 if (cmd_details) { 875 i40e_memcpy(details, 876 cmd_details, 877 sizeof(struct i40e_asq_cmd_details), 878 I40E_NONDMA_TO_NONDMA); 879 880 /* If the cmd_details are defined copy the cookie. The 881 * CPU_TO_LE32 is not needed here because the data is ignored 882 * by the FW, only used by the driver 883 */ 884 if (details->cookie) { 885 desc->cookie_high = 886 CPU_TO_LE32(I40E_HI_DWORD(details->cookie)); 887 desc->cookie_low = 888 CPU_TO_LE32(I40E_LO_DWORD(details->cookie)); 889 } 890 } else { 891 i40e_memset(details, 0, 892 sizeof(struct i40e_asq_cmd_details), 893 I40E_NONDMA_MEM); 894 } 895 896 /* clear requested flags and then set additional flags if defined */ 897 desc->flags &= ~CPU_TO_LE16(details->flags_dis); 898 desc->flags |= CPU_TO_LE16(details->flags_ena); 899 900 if (buff_size > hw->aq.asq_buf_size) { 901 i40e_debug(hw, 902 I40E_DEBUG_AQ_MESSAGE, 903 "AQTX: Invalid buffer size: %d.\n", 904 buff_size); 905 status = I40E_ERR_INVALID_SIZE; 906 goto asq_send_command_error; 907 } 908 909 if (details->postpone && !details->async) { 910 i40e_debug(hw, 911 I40E_DEBUG_AQ_MESSAGE, 912 "AQTX: Async flag not set along with postpone flag"); 913 status = I40E_ERR_PARAM; 914 goto asq_send_command_error; 915 } 916 917 /* call clean and check queue available function to reclaim the 918 * descriptors that were processed by FW, the function returns the 919 * number of desc available 920 */ 921 /* the clean function called here could be called in a separate thread 922 * in case of asynchronous completions 923 */ 924 if (i40e_clean_asq(hw) == 0) { 925 i40e_debug(hw, 926 I40E_DEBUG_AQ_MESSAGE, 927 "AQTX: Error queue is full.\n"); 928 status = I40E_ERR_ADMIN_QUEUE_FULL; 929 goto asq_send_command_error; 930 } 931 932 /* initialize the temp desc pointer with the right desc */ 933 desc_on_ring = I40E_ADMINQ_DESC(hw->aq.asq, hw->aq.asq.next_to_use); 934 935 /* if the desc is available copy the temp desc to the right place */ 936 i40e_memcpy(desc_on_ring, desc, sizeof(struct i40e_aq_desc), 937 I40E_NONDMA_TO_DMA); 938 939 /* if buff is not NULL assume indirect command */ 940 if (buff != NULL) { 941 dma_buff = &(hw->aq.asq.r.asq_bi[hw->aq.asq.next_to_use]); 942 /* copy the user buff into the respective DMA buff */ 943 i40e_memcpy(dma_buff->va, buff, buff_size, 944 I40E_NONDMA_TO_DMA); 945 desc_on_ring->datalen = CPU_TO_LE16(buff_size); 946 947 /* Update the address values in the desc with the pa value 948 * for respective buffer 949 */ 950 desc_on_ring->params.external.addr_high = 951 CPU_TO_LE32(I40E_HI_DWORD(dma_buff->pa)); 952 desc_on_ring->params.external.addr_low = 953 CPU_TO_LE32(I40E_LO_DWORD(dma_buff->pa)); 954 } 955 956 /* bump the tail */ 957 i40e_debug(hw, I40E_DEBUG_AQ_COMMAND, "AQTX: desc and buffer:\n"); 958 i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc_on_ring, 959 buff, buff_size); 960 (hw->aq.asq.next_to_use)++; 961 if (hw->aq.asq.next_to_use == hw->aq.asq.count) 962 hw->aq.asq.next_to_use = 0; 963 if (!details->postpone) 964 wr32(hw, hw->aq.asq.tail, hw->aq.asq.next_to_use); 965 966 /* if cmd_details are not defined or async flag is not set, 967 * we need to wait for desc write back 968 */ 969 if (!details->async && !details->postpone) { 970 u32 total_delay = 0; 971 972 do { 973 /* AQ designers suggest use of head for better 974 * timing reliability than DD bit 975 */ 976 if (i40e_asq_done(hw)) 977 break; 978 i40e_usec_delay(50); 979 total_delay += 50; 980 } while (total_delay < hw->aq.asq_cmd_timeout); 981 } 982 983 /* if ready, copy the desc back to temp */ 984 if (i40e_asq_done(hw)) { 985 i40e_memcpy(desc, desc_on_ring, sizeof(struct i40e_aq_desc), 986 I40E_DMA_TO_NONDMA); 987 if (buff != NULL) 988 i40e_memcpy(buff, dma_buff->va, buff_size, 989 I40E_DMA_TO_NONDMA); 990 retval = LE16_TO_CPU(desc->retval); 991 if (retval != 0) { 992 i40e_debug(hw, 993 I40E_DEBUG_AQ_MESSAGE, 994 "AQTX: Command completed with error 0x%X.\n", 995 retval); 996 997 /* strip off FW internal code */ 998 retval &= 0xff; 999 } 1000 cmd_completed = TRUE; 1001 if ((enum i40e_admin_queue_err)retval == I40E_AQ_RC_OK) 1002 status = I40E_SUCCESS; 1003 else if ((enum i40e_admin_queue_err)retval == I40E_AQ_RC_EBUSY) 1004 status = I40E_ERR_NOT_READY; 1005 else 1006 status = I40E_ERR_ADMIN_QUEUE_ERROR; 1007 hw->aq.asq_last_status = (enum i40e_admin_queue_err)retval; 1008 } 1009 1010 i40e_debug(hw, I40E_DEBUG_AQ_COMMAND, 1011 "AQTX: desc and buffer writeback:\n"); 1012 i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc, buff, buff_size); 1013 1014 /* save writeback aq if requested */ 1015 if (details->wb_desc) 1016 i40e_memcpy(details->wb_desc, desc_on_ring, 1017 sizeof(struct i40e_aq_desc), I40E_DMA_TO_NONDMA); 1018 1019 /* update the error if time out occurred */ 1020 if ((!cmd_completed) && 1021 (!details->async && !details->postpone)) { 1022 if (rd32(hw, hw->aq.asq.len) & I40E_GL_ATQLEN_ATQCRIT_MASK) { 1023 i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE, 1024 "AQTX: AQ Critical error.\n"); 1025 status = I40E_ERR_ADMIN_QUEUE_CRITICAL_ERROR; 1026 } else { 1027 i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE, 1028 "AQTX: Writeback timeout.\n"); 1029 status = I40E_ERR_ADMIN_QUEUE_TIMEOUT; 1030 } 1031 } 1032 1033 asq_send_command_error: 1034 i40e_release_spinlock(&hw->aq.asq_spinlock); 1035 return status; 1036 } 1037 1038 /** 1039 * i40e_fill_default_direct_cmd_desc - AQ descriptor helper function 1040 * @desc: pointer to the temp descriptor (non DMA mem) 1041 * @opcode: the opcode can be used to decide which flags to turn off or on 1042 * 1043 * Fill the desc with default values 1044 **/ 1045 void i40e_fill_default_direct_cmd_desc(struct i40e_aq_desc *desc, 1046 u16 opcode) 1047 { 1048 /* zero out the desc */ 1049 i40e_memset((void *)desc, 0, sizeof(struct i40e_aq_desc), 1050 I40E_NONDMA_MEM); 1051 desc->opcode = CPU_TO_LE16(opcode); 1052 desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_SI); 1053 } 1054 1055 /** 1056 * i40e_clean_arq_element 1057 * @hw: pointer to the hw struct 1058 * @e: event info from the receive descriptor, includes any buffers 1059 * @pending: number of events that could be left to process 1060 * 1061 * This function cleans one Admin Receive Queue element and returns 1062 * the contents through e. It can also return how many events are 1063 * left to process through 'pending' 1064 **/ 1065 enum i40e_status_code i40e_clean_arq_element(struct i40e_hw *hw, 1066 struct i40e_arq_event_info *e, 1067 u16 *pending) 1068 { 1069 enum i40e_status_code ret_code = I40E_SUCCESS; 1070 u16 ntc = hw->aq.arq.next_to_clean; 1071 struct i40e_aq_desc *desc; 1072 struct i40e_dma_mem *bi; 1073 u16 desc_idx; 1074 u16 datalen; 1075 u16 flags; 1076 u16 ntu; 1077 1078 /* pre-clean the event info */ 1079 i40e_memset(&e->desc, 0, sizeof(e->desc), I40E_NONDMA_MEM); 1080 1081 /* take the lock before we start messing with the ring */ 1082 i40e_acquire_spinlock(&hw->aq.arq_spinlock); 1083 1084 if (hw->aq.arq.count == 0) { 1085 i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE, 1086 "AQRX: Admin queue not initialized.\n"); 1087 ret_code = I40E_ERR_QUEUE_EMPTY; 1088 goto clean_arq_element_err; 1089 } 1090 1091 /* set next_to_use to head */ 1092 if (!i40e_is_vf(hw)) 1093 ntu = rd32(hw, hw->aq.arq.head) & I40E_PF_ARQH_ARQH_MASK; 1094 else 1095 ntu = rd32(hw, hw->aq.arq.head) & I40E_VF_ARQH1_ARQH_MASK; 1096 if (ntu == ntc) { 1097 /* nothing to do - shouldn't need to update ring's values */ 1098 ret_code = I40E_ERR_ADMIN_QUEUE_NO_WORK; 1099 goto clean_arq_element_out; 1100 } 1101 1102 /* now clean the next descriptor */ 1103 desc = I40E_ADMINQ_DESC(hw->aq.arq, ntc); 1104 desc_idx = ntc; 1105 1106 hw->aq.arq_last_status = 1107 (enum i40e_admin_queue_err)LE16_TO_CPU(desc->retval); 1108 flags = LE16_TO_CPU(desc->flags); 1109 if (flags & I40E_AQ_FLAG_ERR) { 1110 ret_code = I40E_ERR_ADMIN_QUEUE_ERROR; 1111 i40e_debug(hw, 1112 I40E_DEBUG_AQ_MESSAGE, 1113 "AQRX: Event received with error 0x%X.\n", 1114 hw->aq.arq_last_status); 1115 } 1116 1117 i40e_memcpy(&e->desc, desc, sizeof(struct i40e_aq_desc), 1118 I40E_DMA_TO_NONDMA); 1119 datalen = LE16_TO_CPU(desc->datalen); 1120 e->msg_len = min(datalen, e->buf_len); 1121 if (e->msg_buf != NULL && (e->msg_len != 0)) 1122 i40e_memcpy(e->msg_buf, 1123 hw->aq.arq.r.arq_bi[desc_idx].va, 1124 e->msg_len, I40E_DMA_TO_NONDMA); 1125 1126 i40e_debug(hw, I40E_DEBUG_AQ_COMMAND, "AQRX: desc and buffer:\n"); 1127 i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc, e->msg_buf, 1128 hw->aq.arq_buf_size); 1129 1130 /* Restore the original datalen and buffer address in the desc, 1131 * FW updates datalen to indicate the event message 1132 * size 1133 */ 1134 bi = &hw->aq.arq.r.arq_bi[ntc]; 1135 i40e_memset((void *)desc, 0, sizeof(struct i40e_aq_desc), I40E_DMA_MEM); 1136 1137 desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_BUF); 1138 if (hw->aq.arq_buf_size > I40E_AQ_LARGE_BUF) 1139 desc->flags |= CPU_TO_LE16(I40E_AQ_FLAG_LB); 1140 desc->datalen = CPU_TO_LE16((u16)bi->size); 1141 desc->params.external.addr_high = CPU_TO_LE32(I40E_HI_DWORD(bi->pa)); 1142 desc->params.external.addr_low = CPU_TO_LE32(I40E_LO_DWORD(bi->pa)); 1143 1144 /* set tail = the last cleaned desc index. */ 1145 wr32(hw, hw->aq.arq.tail, ntc); 1146 /* ntc is updated to tail + 1 */ 1147 ntc++; 1148 if (ntc == hw->aq.num_arq_entries) 1149 ntc = 0; 1150 hw->aq.arq.next_to_clean = ntc; 1151 hw->aq.arq.next_to_use = ntu; 1152 1153 i40e_nvmupd_check_wait_event(hw, LE16_TO_CPU(e->desc.opcode), &e->desc); 1154 clean_arq_element_out: 1155 /* Set pending if needed, unlock and return */ 1156 if (pending != NULL) 1157 *pending = (ntc > ntu ? hw->aq.arq.count : 0) + (ntu - ntc); 1158 clean_arq_element_err: 1159 i40e_release_spinlock(&hw->aq.arq_spinlock); 1160 1161 return ret_code; 1162 } 1163 1164