1 /* 2 * Copyright (c) 2018-2019 Cavium, Inc. 3 * All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 9 * 1. Redistributions of source code must retain the above copyright 10 * notice, this list of conditions and the following disclaimer. 11 * 2. Redistributions in binary form must reproduce the above copyright 12 * notice, this list of conditions and the following disclaimer in the 13 * documentation and/or other materials provided with the distribution. 14 * 15 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" 16 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 17 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 18 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE 19 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 20 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 21 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 22 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 23 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 24 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 25 * POSSIBILITY OF SUCH DAMAGE. 26 */ 27 28 /* 29 * File: qlnxr_def.h 30 * Author: David C Somayajulu 31 */ 32 33 #ifndef __QLNX_DEF_H_ 34 #define __QLNX_DEF_H_ 35 36 #include <sys/ktr.h> 37 38 #include <linux/list.h> 39 #include <linux/spinlock.h> 40 #include <linux/idr.h> 41 #include <linux/completion.h> 42 #include <linux/sched.h> 43 #include <linux/pci.h> 44 #include <linux/dma-mapping.h> 45 #include <linux/wait.h> 46 #include <linux/kref.h> 47 #include <linux/timer.h> 48 #include <linux/io.h> 49 #include <linux/fs.h> 50 #include <sys/vmem.h> 51 52 #include <asm/byteorder.h> 53 54 #include <netinet/in.h> 55 #include <net/ipv6.h> 56 #include <netinet/toecore.h> 57 58 #include <rdma/ib_smi.h> 59 #include <rdma/ib_user_verbs.h> 60 #include <rdma/ib_addr.h> 61 #include <rdma/ib_verbs.h> 62 #include <rdma/iw_cm.h> 63 #include <rdma/ib_umem.h> 64 #include <rdma/ib_mad.h> 65 #include <rdma/ib_sa.h> 66 #include <rdma/uverbs_ioctl.h> 67 68 #include "qlnx_os.h" 69 #include "bcm_osal.h" 70 71 #include "reg_addr.h" 72 #include "ecore_gtt_reg_addr.h" 73 #include "ecore.h" 74 #include "ecore_chain.h" 75 #include "ecore_status.h" 76 #include "ecore_hw.h" 77 #include "ecore_rt_defs.h" 78 #include "ecore_init_ops.h" 79 #include "ecore_int.h" 80 #include "ecore_cxt.h" 81 #include "ecore_spq.h" 82 #include "ecore_init_fw_funcs.h" 83 #include "ecore_sp_commands.h" 84 #include "ecore_dev_api.h" 85 #include "ecore_l2_api.h" 86 #ifdef CONFIG_ECORE_SRIOV 87 #include "ecore_sriov.h" 88 #include "ecore_vf.h" 89 #endif 90 #ifdef CONFIG_ECORE_LL2 91 #include "ecore_ll2.h" 92 #endif 93 #ifdef CONFIG_ECORE_FCOE 94 #include "ecore_fcoe.h" 95 #endif 96 #ifdef CONFIG_ECORE_ISCSI 97 #include "ecore_iscsi.h" 98 #endif 99 #include "ecore_mcp.h" 100 #include "ecore_hw_defs.h" 101 #include "mcp_public.h" 102 103 #ifdef CONFIG_ECORE_RDMA 104 #include "ecore_rdma.h" 105 #include "ecore_rdma_api.h" 106 #endif 107 108 #ifdef CONFIG_ECORE_ROCE 109 #include "ecore_roce.h" 110 #endif 111 112 #ifdef CONFIG_ECORE_IWARP 113 #include "ecore_iwarp.h" 114 #endif 115 116 #include "ecore_iro.h" 117 #include "nvm_cfg.h" 118 119 #include "ecore_dbg_fw_funcs.h" 120 #include "rdma_common.h" 121 122 #include "qlnx_ioctl.h" 123 #include "qlnx_def.h" 124 #include "qlnx_rdma.h" 125 #include "qlnxr_verbs.h" 126 #include "qlnxr_user.h" 127 #include "qlnx_ver.h" 128 #include <sys/smp.h> 129 130 #define QLNXR_ROCE_INTERFACE_VERSION 1801 131 132 #define QLNXR_MODULE_VERSION "8.18.1.0" 133 #define QLNXR_NODE_DESC "QLogic 579xx RoCE HCA" 134 135 #define OC_SKH_DEVICE_PF 0x720 136 #define OC_SKH_DEVICE_VF 0x728 137 #define QLNXR_MAX_AH 512 138 139 /* QLNXR Limitations */ 140 141 /* SQ/RQ Limitations 142 * An S/RQ PBL contains a list a pointers to pages. Each page contains S/RQE 143 * elements. Several S/RQE elements make an S/RQE, up to a certain maximum that 144 * is different between SQ and RQ. The size of the PBL was chosen such as not to 145 * limit the MAX_WR supported by ECORE, and rounded up to a power of two. 146 */ 147 /* SQ */ 148 #define QLNXR_MAX_SQ_PBL (0x8000) /* 2^15 bytes */ 149 #define QLNXR_MAX_SQ_PBL_ENTRIES (0x10000 / sizeof(void *)) /* number */ 150 #define QLNXR_SQE_ELEMENT_SIZE (sizeof(struct rdma_sq_sge)) /* bytes */ 151 #define QLNXR_MAX_SQE_ELEMENTS_PER_SQE (ROCE_REQ_MAX_SINGLE_SQ_WQE_SIZE / \ 152 QLNXR_SQE_ELEMENT_SIZE) /* number */ 153 #define QLNXR_MAX_SQE_ELEMENTS_PER_PAGE ((RDMA_RING_PAGE_SIZE) / \ 154 QLNXR_SQE_ELEMENT_SIZE) /* number */ 155 #define QLNXR_MAX_SQE ((QLNXR_MAX_SQ_PBL_ENTRIES) * (RDMA_RING_PAGE_SIZE) / \ 156 (QLNXR_SQE_ELEMENT_SIZE) / (QLNXR_MAX_SQE_ELEMENTS_PER_SQE)) 157 /* RQ */ 158 #define QLNXR_MAX_RQ_PBL (0x2000) /* 2^13 bytes */ 159 #define QLNXR_MAX_RQ_PBL_ENTRIES (0x10000 / sizeof(void *)) /* number */ 160 #define QLNXR_RQE_ELEMENT_SIZE (sizeof(struct rdma_rq_sge)) /* bytes */ 161 #define QLNXR_MAX_RQE_ELEMENTS_PER_RQE (RDMA_MAX_SGE_PER_RQ_WQE) /* number */ 162 #define QLNXR_MAX_RQE_ELEMENTS_PER_PAGE ((RDMA_RING_PAGE_SIZE) / \ 163 QLNXR_RQE_ELEMENT_SIZE) /* number */ 164 #define QLNXR_MAX_RQE ((QLNXR_MAX_RQ_PBL_ENTRIES) * (RDMA_RING_PAGE_SIZE) / \ 165 (QLNXR_RQE_ELEMENT_SIZE) / (QLNXR_MAX_RQE_ELEMENTS_PER_RQE)) 166 167 /* CQE Limitation 168 * Although FW supports two layer PBL we use single layer since it is more 169 * than enough. For that layer we use a maximum size of 512 kB, again, because 170 * it reaches the maximum number of page pointers. Notice is the '-1' in the 171 * calculation that comes from having a u16 for the number of pages i.e. 0xffff 172 * is the maximum number of pages (in single layer). 173 */ 174 #define QLNXR_CQE_SIZE (sizeof(union rdma_cqe)) 175 #define QLNXR_MAX_CQE_PBL_SIZE (512*1024) /* 512kB */ 176 #define QLNXR_MAX_CQE_PBL_ENTRIES (((QLNXR_MAX_CQE_PBL_SIZE) / \ 177 sizeof(u64)) - 1) /* 64k -1 */ 178 #define QLNXR_MAX_CQES ((u32)((QLNXR_MAX_CQE_PBL_ENTRIES) * (ECORE_CHAIN_PAGE_SIZE)\ 179 / QLNXR_CQE_SIZE)) /* 8M -4096/32 = 8,388,480 */ 180 181 /* CNQ size Limitation 182 * The maximum CNQ size is not reachable because the FW supports a chain of u16 183 * (specifically 64k-1). The FW can buffer CNQ elements avoiding an overflow, on 184 * the expense of performance. Hence we set it to an arbitrarily smaller value 185 * than the maximum. 186 */ 187 #define QLNXR_ROCE_MAX_CNQ_SIZE (0x4000) /* 2^16 */ 188 189 #define QLNXR_MAX_PORT (1) 190 #define QLNXR_PORT (1) 191 192 #define QLNXR_UVERBS(CMD_NAME) (1ull << IB_USER_VERBS_CMD_##CMD_NAME) 193 194 #define convert_to_64bit(lo, hi) ((u64)hi << 32 | (u64)lo) 195 196 /* The following number is used to determine if a handle recevied from the FW 197 * actually point to a CQ/QP. 198 */ 199 #define QLNXR_CQ_MAGIC_NUMBER (0x11223344) 200 #define QLNXR_QP_MAGIC_NUMBER (0x77889900) 201 202 /* Fast path debug prints */ 203 #define FP_DP_VERBOSE(...) 204 /* #define FP_DP_VERBOSE(...) DP_VERBOSE(__VA_ARGS__) */ 205 206 #define FW_PAGE_SIZE (RDMA_RING_PAGE_SIZE) 207 208 #define QLNXR_MSG_INIT 0x10000, 209 #define QLNXR_MSG_FAIL 0x10000, 210 #define QLNXR_MSG_CQ 0x20000, 211 #define QLNXR_MSG_RQ 0x40000, 212 #define QLNXR_MSG_SQ 0x80000, 213 #define QLNXR_MSG_QP (QLNXR_MSG_SQ | QLNXR_MSG_RQ), 214 #define QLNXR_MSG_MR 0x100000, 215 #define QLNXR_MSG_GSI 0x200000, 216 #define QLNXR_MSG_MISC 0x400000, 217 #define QLNXR_MSG_SRQ 0x800000, 218 #define QLNXR_MSG_IWARP 0x1000000, 219 220 #define QLNXR_ROCE_PKEY_MAX 1 221 #define QLNXR_ROCE_PKEY_TABLE_LEN 1 222 #define QLNXR_ROCE_PKEY_DEFAULT 0xffff 223 224 #define QLNXR_MAX_SGID 128 /* TBD - add more source gids... */ 225 226 #define QLNXR_ENET_STATE_BIT (0) 227 228 #define QLNXR_MAX_MSIX (16) 229 230 struct qlnxr_cnq { 231 struct qlnxr_dev *dev; 232 struct ecore_chain pbl; 233 struct ecore_sb_info *sb; 234 char name[32]; 235 u64 n_comp; 236 __le16 *hw_cons_ptr; 237 u8 index; 238 int irq_rid; 239 struct resource *irq; 240 void *irq_handle; 241 }; 242 243 struct qlnxr_device_attr { 244 /* Vendor specific information */ 245 u32 vendor_id; 246 u32 vendor_part_id; 247 u32 hw_ver; 248 u64 fw_ver; 249 250 u64 node_guid; /* node GUID */ 251 u64 sys_image_guid; /* System image GUID */ 252 253 u8 max_cnq; 254 u8 max_sge; /* Maximum # of scatter/gather entries 255 * per Work Request supported 256 */ 257 u16 max_inline; 258 u32 max_sqe; /* Maximum number of send outstanding send work 259 * requests on any Work Queue supported 260 */ 261 u32 max_rqe; /* Maximum number of receive outstanding receive 262 * work requests on any Work Queue supported 263 */ 264 u8 max_qp_resp_rd_atomic_resc; /* Maximum number of RDMA Reads 265 * & atomic operation that can 266 * be outstanding per QP 267 */ 268 269 u8 max_qp_req_rd_atomic_resc; /* The maximum depth per QP for 270 * initiation of RDMA Read 271 * & atomic operations 272 */ 273 u64 max_dev_resp_rd_atomic_resc; 274 u32 max_cq; 275 u32 max_qp; 276 u32 max_mr; /* Maximum # of MRs supported */ 277 u64 max_mr_size; /* Size (in bytes) of largest contiguous memory 278 * block that can be registered by this device 279 */ 280 u32 max_cqe; 281 u32 max_mw; /* Maximum # of memory windows supported */ 282 u32 max_fmr; 283 u32 max_mr_mw_fmr_pbl; 284 u64 max_mr_mw_fmr_size; 285 u32 max_pd; /* Maximum # of protection domains supported */ 286 u32 max_ah; 287 u8 max_pkey; 288 u32 max_srq; /* Maximum number of SRQs */ 289 u32 max_srq_wr; /* Maximum number of WRs per SRQ */ 290 u8 max_srq_sge; /* Maximum number of SGE per WQE */ 291 u8 max_stats_queues; /* Maximum number of statistics queues */ 292 u32 dev_caps; 293 294 /* Abilty to support RNR-NAK generation */ 295 296 #define QLNXR_ROCE_DEV_CAP_RNR_NAK_MASK 0x1 297 #define QLNXR_ROCE_DEV_CAP_RNR_NAK_SHIFT 0 298 /* Abilty to support shutdown port */ 299 #define QLNXR_ROCE_DEV_CAP_SHUTDOWN_PORT_MASK 0x1 300 #define QLNXR_ROCE_DEV_CAP_SHUTDOWN_PORT_SHIFT 1 301 /* Abilty to support port active event */ 302 #define QLNXR_ROCE_DEV_CAP_PORT_ACTIVE_EVENT_MASK 0x1 303 #define QLNXR_ROCE_DEV_CAP_PORT_ACTIVE_EVENT_SHIFT 2 304 /* Abilty to support port change event */ 305 #define QLNXR_ROCE_DEV_CAP_PORT_CHANGE_EVENT_MASK 0x1 306 #define QLNXR_ROCE_DEV_CAP_PORT_CHANGE_EVENT_SHIFT 3 307 /* Abilty to support system image GUID */ 308 #define QLNXR_ROCE_DEV_CAP_SYS_IMAGE_MASK 0x1 309 #define QLNXR_ROCE_DEV_CAP_SYS_IMAGE_SHIFT 4 310 /* Abilty to support bad P_Key counter support */ 311 #define QLNXR_ROCE_DEV_CAP_BAD_PKEY_CNT_MASK 0x1 312 #define QLNXR_ROCE_DEV_CAP_BAD_PKEY_CNT_SHIFT 5 313 /* Abilty to support atomic operations */ 314 #define QLNXR_ROCE_DEV_CAP_ATOMIC_OP_MASK 0x1 315 #define QLNXR_ROCE_DEV_CAP_ATOMIC_OP_SHIFT 6 316 #define QLNXR_ROCE_DEV_CAP_RESIZE_CQ_MASK 0x1 317 #define QLNXR_ROCE_DEV_CAP_RESIZE_CQ_SHIFT 7 318 /* Abilty to support modifying the maximum number of 319 * outstanding work requests per QP 320 */ 321 #define QLNXR_ROCE_DEV_CAP_RESIZE_MAX_WR_MASK 0x1 322 #define QLNXR_ROCE_DEV_CAP_RESIZE_MAX_WR_SHIFT 8 323 324 /* Abilty to support automatic path migration */ 325 #define QLNXR_ROCE_DEV_CAP_AUTO_PATH_MIG_MASK 0x1 326 #define QLNXR_ROCE_DEV_CAP_AUTO_PATH_MIG_SHIFT 9 327 /* Abilty to support the base memory management extensions */ 328 #define QLNXR_ROCE_DEV_CAP_BASE_MEMORY_EXT_MASK 0x1 329 #define QLNXR_ROCE_DEV_CAP_BASE_MEMORY_EXT_SHIFT 10 330 #define QLNXR_ROCE_DEV_CAP_BASE_QUEUE_EXT_MASK 0x1 331 #define QLNXR_ROCE_DEV_CAP_BASE_QUEUE_EXT_SHIFT 11 332 /* Abilty to support multipile page sizes per memory region */ 333 #define QLNXR_ROCE_DEV_CAP_MULTI_PAGE_PER_MR_EXT_MASK 0x1 334 #define QLNXR_ROCE_DEV_CAP_MULTI_PAGE_PER_MR_EXT_SHIFT 12 335 /* Abilty to support block list physical buffer list */ 336 #define QLNXR_ROCE_DEV_CAP_BLOCK_MODE_MASK 0x1 337 #define QLNXR_ROCE_DEV_CAP_BLOCK_MODE_SHIFT 13 338 /* Abilty to support zero based virtual addresses */ 339 #define QLNXR_ROCE_DEV_CAP_ZBVA_MASK 0x1 340 #define QLNXR_ROCE_DEV_CAP_ZBVA_SHIFT 14 341 /* Abilty to support local invalidate fencing */ 342 #define QLNXR_ROCE_DEV_CAP_LOCAL_INV_FENCE_MASK 0x1 343 #define QLNXR_ROCE_DEV_CAP_LOCAL_INV_FENCE_SHIFT 15 344 /* Abilty to support Loopback on QP */ 345 #define QLNXR_ROCE_DEV_CAP_LB_INDICATOR_MASK 0x1 346 #define QLNXR_ROCE_DEV_CAP_LB_INDICATOR_SHIFT 16 347 u64 page_size_caps; 348 u8 dev_ack_delay; 349 u32 reserved_lkey; /* Value of reserved L_key */ 350 u32 bad_pkey_counter;/* Bad P_key counter support 351 * indicator 352 */ 353 struct ecore_rdma_events events; 354 }; 355 356 struct qlnxr_dev { 357 struct ib_device ibdev; 358 qlnx_host_t *ha; 359 struct ecore_dev *cdev; 360 361 /* Added to extend Applications Support */ 362 struct pci_dev pdev; 363 uint32_t dp_module; 364 uint8_t dp_level; 365 366 void *rdma_ctx; 367 368 struct mtx idr_lock; 369 struct idr qpidr; 370 371 uint32_t wq_multiplier; 372 int num_cnq; 373 374 struct ecore_sb_info sb_array[QLNXR_MAX_MSIX]; 375 struct qlnxr_cnq cnq_array[QLNXR_MAX_MSIX]; 376 377 int sb_start; 378 379 int gsi_qp_created; 380 struct qlnxr_cq *gsi_sqcq; 381 struct qlnxr_cq *gsi_rqcq; 382 struct qlnxr_qp *gsi_qp; 383 384 /* TBD: we'll need an array of these probablly per DPI... */ 385 void __iomem *db_addr; 386 uint64_t db_phys_addr; 387 uint32_t db_size; 388 uint16_t dpi; 389 390 uint64_t guid; 391 enum ib_atomic_cap atomic_cap; 392 393 union ib_gid sgid_tbl[QLNXR_MAX_SGID]; 394 struct mtx sgid_lock; 395 struct notifier_block nb_inet; 396 struct notifier_block nb_inet6; 397 398 uint8_t mr_key; 399 struct list_head entry; 400 401 struct dentry *dbgfs; 402 403 uint8_t gsi_ll2_mac_address[ETH_ALEN]; 404 uint8_t gsi_ll2_handle; 405 406 unsigned long enet_state; 407 408 struct workqueue_struct *iwarp_wq; 409 410 volatile uint32_t pd_count; 411 struct qlnxr_device_attr attr; 412 uint8_t user_dpm_enabled; 413 }; 414 415 typedef struct qlnxr_dev qlnxr_dev_t; 416 417 struct qlnxr_pd { 418 struct ib_pd ibpd; 419 u32 pd_id; 420 struct qlnxr_ucontext *uctx; 421 }; 422 423 struct qlnxr_ucontext { 424 struct ib_ucontext ibucontext; 425 struct qlnxr_dev *dev; 426 struct qlnxr_pd *pd; 427 u64 dpi_addr; 428 u64 dpi_phys_addr; 429 u32 dpi_size; 430 u16 dpi; 431 432 struct list_head mm_head; 433 struct mutex mm_list_lock; 434 }; 435 436 struct qlnxr_dev_attr { 437 struct ib_device_attr ib_attr; 438 }; 439 440 struct qlnxr_dma_mem { 441 void *va; 442 dma_addr_t pa; 443 u32 size; 444 }; 445 446 struct qlnxr_pbl { 447 struct list_head list_entry; 448 void *va; 449 dma_addr_t pa; 450 }; 451 452 struct qlnxr_queue_info { 453 void *va; 454 dma_addr_t dma; 455 u32 size; 456 u16 len; 457 u16 entry_size; /* Size of an element in the queue */ 458 u16 id; /* qid, where to ring the doorbell. */ 459 u16 head, tail; 460 bool created; 461 }; 462 463 struct qlnxr_eq { 464 struct qlnxr_queue_info q; 465 u32 vector; 466 int cq_cnt; 467 struct qlnxr_dev *dev; 468 char irq_name[32]; 469 }; 470 471 struct qlnxr_mq { 472 struct qlnxr_queue_info sq; 473 struct qlnxr_queue_info cq; 474 bool rearm_cq; 475 }; 476 477 struct phy_info { 478 u16 auto_speeds_supported; 479 u16 fixed_speeds_supported; 480 u16 phy_type; 481 u16 interface_type; 482 }; 483 484 union db_prod64 { 485 struct rdma_pwm_val32_data data; 486 u64 raw; 487 }; 488 489 enum qlnxr_cq_type { 490 QLNXR_CQ_TYPE_GSI, 491 QLNXR_CQ_TYPE_KERNEL, 492 QLNXR_CQ_TYPE_USER 493 }; 494 495 struct qlnxr_pbl_info { 496 u32 num_pbls; 497 u32 num_pbes; 498 u32 pbl_size; 499 u32 pbe_size; 500 bool two_layered; 501 }; 502 503 struct qlnxr_userq { 504 struct ib_umem *umem; 505 struct qlnxr_pbl_info pbl_info; 506 struct qlnxr_pbl *pbl_tbl; 507 u64 buf_addr; 508 size_t buf_len; 509 }; 510 511 struct qlnxr_cq { 512 struct ib_cq ibcq; /* must be first */ 513 514 enum qlnxr_cq_type cq_type; 515 uint32_t sig; 516 uint16_t icid; 517 518 /* relevant to cqs created from kernel space only (ULPs) */ 519 spinlock_t cq_lock; 520 uint8_t arm_flags; 521 struct ecore_chain pbl; 522 523 void __iomem *db_addr; /* db address for cons update*/ 524 union db_prod64 db; 525 526 uint8_t pbl_toggle; 527 union rdma_cqe *latest_cqe; 528 union rdma_cqe *toggle_cqe; 529 530 /* TODO: remove since it is redundant with 32 bit chains */ 531 uint32_t cq_cons; 532 533 /* relevant to cqs created from user space only (applications) */ 534 struct qlnxr_userq q; 535 536 /* destroy-IRQ handler race prevention */ 537 uint8_t destroyed; 538 uint16_t cnq_notif; 539 }; 540 541 struct qlnxr_ah { 542 struct ib_ah ibah; 543 struct ib_ah_attr attr; 544 }; 545 546 union db_prod32 { 547 struct rdma_pwm_val16_data data; 548 u32 raw; 549 }; 550 551 struct qlnxr_qp_hwq_info { 552 /* WQE Elements*/ 553 struct ecore_chain pbl; 554 u64 p_phys_addr_tbl; 555 u32 max_sges; 556 557 /* WQE */ 558 u16 prod; /* WQE prod index for SW ring */ 559 u16 cons; /* WQE cons index for SW ring */ 560 u16 wqe_cons; 561 u16 gsi_cons; /* filled in by GSI implementation */ 562 u16 max_wr; 563 564 /* DB */ 565 void __iomem *db; /* Doorbell address */ 566 union db_prod32 db_data; /* Doorbell data */ 567 568 /* Required for iwarp_only */ 569 void __iomem *iwarp_db2; /* Doorbell address */ 570 union db_prod32 iwarp_db2_data; /* Doorbell data */ 571 }; 572 573 #define QLNXR_INC_SW_IDX(p_info, index) \ 574 do { \ 575 p_info->index = (p_info->index + 1) & \ 576 ecore_chain_get_capacity(p_info->pbl) \ 577 } while (0) 578 579 struct qlnxr_srq_hwq_info { 580 u32 max_sges; 581 u32 max_wr; 582 struct ecore_chain pbl; 583 u64 p_phys_addr_tbl; 584 u32 wqe_prod; /* WQE prod index in HW ring */ 585 u32 sge_prod; /* SGE prod index in HW ring */ 586 u32 wr_prod_cnt; /* wr producer count */ 587 u32 wr_cons_cnt; /* wr consumer count */ 588 u32 num_elems; 589 590 u32 *virt_prod_pair_addr; /* producer pair virtual address */ 591 dma_addr_t phy_prod_pair_addr; /* producer pair physical address */ 592 }; 593 594 struct qlnxr_srq { 595 struct ib_srq ibsrq; 596 struct qlnxr_dev *dev; 597 /* relevant to cqs created from user space only (applications) */ 598 struct qlnxr_userq usrq; 599 struct qlnxr_srq_hwq_info hw_srq; 600 struct ib_umem *prod_umem; 601 u16 srq_id; 602 /* lock to protect srq recv post */ 603 spinlock_t lock; 604 }; 605 606 enum qlnxr_qp_err_bitmap { 607 QLNXR_QP_ERR_SQ_FULL = 1 << 0, 608 QLNXR_QP_ERR_RQ_FULL = 1 << 1, 609 QLNXR_QP_ERR_BAD_SR = 1 << 2, 610 QLNXR_QP_ERR_BAD_RR = 1 << 3, 611 QLNXR_QP_ERR_SQ_PBL_FULL = 1 << 4, 612 QLNXR_QP_ERR_RQ_PBL_FULL = 1 << 5, 613 }; 614 615 struct mr_info { 616 struct qlnxr_pbl *pbl_table; 617 struct qlnxr_pbl_info pbl_info; 618 struct list_head free_pbl_list; 619 struct list_head inuse_pbl_list; 620 u32 completed; 621 u32 completed_handled; 622 }; 623 624 struct qlnxr_qp { 625 struct ib_qp ibqp; /* must be first */ 626 struct qlnxr_dev *dev; 627 struct qlnxr_iw_ep *ep; 628 struct qlnxr_qp_hwq_info sq; 629 struct qlnxr_qp_hwq_info rq; 630 631 u32 max_inline_data; 632 633 spinlock_t q_lock ____cacheline_aligned; 634 635 struct qlnxr_cq *sq_cq; 636 struct qlnxr_cq *rq_cq; 637 struct qlnxr_srq *srq; 638 enum ecore_roce_qp_state state; /* QP state */ 639 u32 id; 640 struct qlnxr_pd *pd; 641 enum ib_qp_type qp_type; 642 struct ecore_rdma_qp *ecore_qp; 643 u32 qp_id; 644 u16 icid; 645 u16 mtu; 646 int sgid_idx; 647 u32 rq_psn; 648 u32 sq_psn; 649 u32 qkey; 650 u32 dest_qp_num; 651 u32 sig; /* unique siganture to identify valid QP */ 652 653 /* relevant to qps created from kernel space only (ULPs) */ 654 u8 prev_wqe_size; 655 u16 wqe_cons; 656 u32 err_bitmap; 657 bool signaled; 658 /* SQ shadow */ 659 struct { 660 u64 wr_id; 661 enum ib_wc_opcode opcode; 662 u32 bytes_len; 663 u8 wqe_size; 664 bool signaled; 665 dma_addr_t icrc_mapping; 666 u32 *icrc; 667 struct qlnxr_mr *mr; 668 } *wqe_wr_id; 669 670 /* RQ shadow */ 671 struct { 672 u64 wr_id; 673 struct ib_sge sg_list[RDMA_MAX_SGE_PER_RQ_WQE]; 674 uint8_t wqe_size; 675 676 /* for GSI only */ 677 u8 smac[ETH_ALEN]; 678 u16 vlan_id; 679 int rc; 680 } *rqe_wr_id; 681 682 /* relevant to qps created from user space only (applications) */ 683 struct qlnxr_userq usq; 684 struct qlnxr_userq urq; 685 atomic_t refcnt; 686 bool destroyed; 687 }; 688 689 enum qlnxr_mr_type { 690 QLNXR_MR_USER, 691 QLNXR_MR_KERNEL, 692 QLNXR_MR_DMA, 693 QLNXR_MR_FRMR 694 }; 695 696 struct qlnxr_mr { 697 struct ib_mr ibmr; 698 struct ib_umem *umem; 699 700 struct ecore_rdma_register_tid_in_params hw_mr; 701 enum qlnxr_mr_type type; 702 703 struct qlnxr_dev *dev; 704 struct mr_info info; 705 706 u64 *pages; 707 u32 npages; 708 709 u64 *iova_start; /* valid only for kernel_mr */ 710 }; 711 712 struct qlnxr_mm { 713 struct { 714 u64 phy_addr; 715 unsigned long len; 716 } key; 717 struct list_head entry; 718 }; 719 720 struct qlnxr_iw_listener { 721 struct qlnxr_dev *dev; 722 struct iw_cm_id *cm_id; 723 int backlog; 724 void *ecore_handle; 725 }; 726 727 struct qlnxr_iw_ep { 728 struct qlnxr_dev *dev; 729 struct iw_cm_id *cm_id; 730 struct qlnxr_qp *qp; 731 void *ecore_context; 732 u8 during_connect; 733 }; 734 735 static inline void 736 qlnxr_inc_sw_cons(struct qlnxr_qp_hwq_info *info) 737 { 738 info->cons = (info->cons + 1) % info->max_wr; 739 info->wqe_cons++; 740 } 741 742 static inline void 743 qlnxr_inc_sw_prod(struct qlnxr_qp_hwq_info *info) 744 { 745 info->prod = (info->prod + 1) % info->max_wr; 746 } 747 748 static inline struct qlnxr_dev * 749 get_qlnxr_dev(struct ib_device *ibdev) 750 { 751 return container_of(ibdev, struct qlnxr_dev, ibdev); 752 } 753 754 static inline struct qlnxr_ucontext * 755 get_qlnxr_ucontext(struct ib_ucontext *ibucontext) 756 { 757 return container_of(ibucontext, struct qlnxr_ucontext, ibucontext); 758 } 759 760 static inline struct qlnxr_pd * 761 get_qlnxr_pd(struct ib_pd *ibpd) 762 { 763 return container_of(ibpd, struct qlnxr_pd, ibpd); 764 } 765 766 static inline struct qlnxr_cq * 767 get_qlnxr_cq(struct ib_cq *ibcq) 768 { 769 return container_of(ibcq, struct qlnxr_cq, ibcq); 770 } 771 772 static inline struct qlnxr_qp * 773 get_qlnxr_qp(struct ib_qp *ibqp) 774 { 775 return container_of(ibqp, struct qlnxr_qp, ibqp); 776 } 777 778 static inline struct qlnxr_mr * 779 get_qlnxr_mr(struct ib_mr *ibmr) 780 { 781 return container_of(ibmr, struct qlnxr_mr, ibmr); 782 } 783 784 static inline struct qlnxr_ah * 785 get_qlnxr_ah(struct ib_ah *ibah) 786 { 787 return container_of(ibah, struct qlnxr_ah, ibah); 788 } 789 790 static inline struct qlnxr_srq * 791 get_qlnxr_srq(struct ib_srq *ibsrq) 792 { 793 return container_of(ibsrq, struct qlnxr_srq, ibsrq); 794 } 795 796 static inline bool qlnxr_qp_has_srq(struct qlnxr_qp *qp) 797 { 798 return !!qp->srq; 799 } 800 801 static inline bool qlnxr_qp_has_sq(struct qlnxr_qp *qp) 802 { 803 if (qp->qp_type == IB_QPT_GSI) 804 return 0; 805 806 return 1; 807 } 808 809 static inline bool qlnxr_qp_has_rq(struct qlnxr_qp *qp) 810 { 811 if (qp->qp_type == IB_QPT_GSI || qlnxr_qp_has_srq(qp)) 812 return 0; 813 814 return 1; 815 } 816 817 #define SET_FIELD2(value, name, flag) \ 818 do { \ 819 (value) |= ((flag) << (name ## _SHIFT)); \ 820 } while (0) 821 822 #define QLNXR_RESP_IMM (RDMA_CQE_RESPONDER_IMM_FLG_MASK << \ 823 RDMA_CQE_RESPONDER_IMM_FLG_SHIFT) 824 #define QLNXR_RESP_RDMA (RDMA_CQE_RESPONDER_RDMA_FLG_MASK << \ 825 RDMA_CQE_RESPONDER_RDMA_FLG_SHIFT) 826 #define QLNXR_RESP_INV (RDMA_CQE_RESPONDER_INV_FLG_MASK << \ 827 RDMA_CQE_RESPONDER_INV_FLG_SHIFT) 828 829 #define QLNXR_RESP_RDMA_IMM (QLNXR_RESP_IMM | QLNXR_RESP_RDMA) 830 831 static inline int 832 qlnxr_get_dmac(struct qlnxr_dev *dev, struct ib_ah_attr *ah_attr, u8 *mac_addr) 833 { 834 #ifdef DEFINE_NO_IP_BASED_GIDS 835 u8 *guid = &ah_attr->grh.dgid.raw[8]; /* GID's 64 MSBs are the GUID */ 836 #endif 837 union ib_gid zero_sgid = { { 0 } }; 838 struct in6_addr in6; 839 840 if (!memcmp(&ah_attr->grh.dgid, &zero_sgid, sizeof(union ib_gid))) { 841 memset(mac_addr, 0x00, ETH_ALEN); 842 return -EINVAL; 843 } 844 845 memcpy(&in6, ah_attr->grh.dgid.raw, sizeof(in6)); 846 847 #ifdef DEFINE_NO_IP_BASED_GIDS 848 /* get the MAC address from the GUID i.e. EUI-64 to MAC address */ 849 mac_addr[0] = guid[0] ^ 2; /* toggle the local/universal bit to local */ 850 mac_addr[1] = guid[1]; 851 mac_addr[2] = guid[2]; 852 mac_addr[3] = guid[5]; 853 mac_addr[4] = guid[6]; 854 mac_addr[5] = guid[7]; 855 #else 856 memcpy(mac_addr, ah_attr->dmac, ETH_ALEN); 857 #endif 858 return 0; 859 } 860 861 extern int qlnx_rdma_ll2_set_mac_filter(void *rdma_ctx, uint8_t *old_mac_address, 862 uint8_t *new_mac_address); 863 864 #define QLNXR_ROCE_PKEY_MAX 1 865 #define QLNXR_ROCE_PKEY_TABLE_LEN 1 866 #define QLNXR_ROCE_PKEY_DEFAULT 0xffff 867 868 #define QLNX_IS_IWARP(rdev) IS_IWARP(ECORE_LEADING_HWFN(rdev->cdev)) 869 #define QLNX_IS_ROCE(rdev) IS_ROCE(ECORE_LEADING_HWFN(rdev->cdev)) 870 871 #define MAX_RXMIT_CONNS 16 872 873 #endif /* #ifndef __QLNX_DEF_H_ */ 874