1 // SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause) 2 /* Copyright 2021 Marvell. All rights reserved. */ 3 4 #include <linux/types.h> 5 #include <asm/byteorder.h> 6 #include <asm/param.h> 7 #include <linux/delay.h> 8 #include <linux/dma-mapping.h> 9 #include <linux/etherdevice.h> 10 #include <linux/kernel.h> 11 #include <linux/log2.h> 12 #include <linux/module.h> 13 #include <linux/pci.h> 14 #include <linux/stddef.h> 15 #include <linux/string.h> 16 #include <linux/errno.h> 17 #include <linux/list.h> 18 #include <linux/qed/qed_nvmetcp_if.h> 19 #include "qed.h" 20 #include "qed_cxt.h" 21 #include "qed_dev_api.h" 22 #include "qed_hsi.h" 23 #include "qed_hw.h" 24 #include "qed_int.h" 25 #include "qed_nvmetcp.h" 26 #include "qed_ll2.h" 27 #include "qed_mcp.h" 28 #include "qed_sp.h" 29 #include "qed_reg_addr.h" 30 #include "qed_nvmetcp_fw_funcs.h" 31 32 static int qed_nvmetcp_async_event(struct qed_hwfn *p_hwfn, u8 fw_event_code, 33 u16 echo, union event_ring_data *data, 34 u8 fw_return_code) 35 { 36 if (p_hwfn->p_nvmetcp_info->event_cb) { 37 struct qed_nvmetcp_info *p_nvmetcp = p_hwfn->p_nvmetcp_info; 38 39 return p_nvmetcp->event_cb(p_nvmetcp->event_context, 40 fw_event_code, data); 41 } else { 42 DP_NOTICE(p_hwfn, "nvmetcp async completion is not set\n"); 43 44 return -EINVAL; 45 } 46 } 47 48 static int qed_sp_nvmetcp_func_start(struct qed_hwfn *p_hwfn, 49 enum spq_mode comp_mode, 50 struct qed_spq_comp_cb *p_comp_addr, 51 void *event_context, 52 nvmetcp_event_cb_t async_event_cb) 53 { 54 struct nvmetcp_init_ramrod_params *p_ramrod = NULL; 55 struct qed_nvmetcp_pf_params *p_params = NULL; 56 struct scsi_init_func_queues *p_queue = NULL; 57 struct nvmetcp_spe_func_init *p_init = NULL; 58 struct qed_sp_init_data init_data = {}; 59 struct qed_spq_entry *p_ent = NULL; 60 int rc = 0; 61 u16 val; 62 u8 i; 63 64 /* Get SPQ entry */ 65 init_data.cid = qed_spq_get_cid(p_hwfn); 66 init_data.opaque_fid = p_hwfn->hw_info.opaque_fid; 67 init_data.comp_mode = comp_mode; 68 init_data.p_comp_data = p_comp_addr; 69 rc = qed_sp_init_request(p_hwfn, &p_ent, 70 NVMETCP_RAMROD_CMD_ID_INIT_FUNC, 71 PROTOCOLID_TCP_ULP, &init_data); 72 if (rc) 73 return rc; 74 75 p_ramrod = &p_ent->ramrod.nvmetcp_init; 76 p_init = &p_ramrod->nvmetcp_init_spe; 77 p_params = &p_hwfn->pf_params.nvmetcp_pf_params; 78 p_queue = &p_init->q_params; 79 p_init->num_sq_pages_in_ring = p_params->num_sq_pages_in_ring; 80 p_init->num_r2tq_pages_in_ring = p_params->num_r2tq_pages_in_ring; 81 p_init->num_uhq_pages_in_ring = p_params->num_uhq_pages_in_ring; 82 p_init->ll2_rx_queue_id = RESC_START(p_hwfn, QED_LL2_RAM_QUEUE) + 83 p_params->ll2_ooo_queue_id; 84 SET_FIELD(p_init->flags, NVMETCP_SPE_FUNC_INIT_NVMETCP_MODE, 1); 85 p_init->func_params.log_page_size = ilog2(PAGE_SIZE); 86 p_init->func_params.num_tasks = cpu_to_le16(p_params->num_tasks); 87 p_init->debug_flags = p_params->debug_mode; 88 DMA_REGPAIR_LE(p_queue->glbl_q_params_addr, 89 p_params->glbl_q_params_addr); 90 p_queue->cq_num_entries = cpu_to_le16(QED_NVMETCP_FW_CQ_SIZE); 91 p_queue->num_queues = p_params->num_queues; 92 val = RESC_START(p_hwfn, QED_CMDQS_CQS); 93 p_queue->queue_relative_offset = cpu_to_le16((u16)val); 94 p_queue->cq_sb_pi = p_params->gl_rq_pi; 95 96 for (i = 0; i < p_params->num_queues; i++) { 97 val = qed_get_igu_sb_id(p_hwfn, i); 98 p_queue->cq_cmdq_sb_num_arr[i] = cpu_to_le16(val); 99 } 100 101 SET_FIELD(p_queue->q_validity, 102 SCSI_INIT_FUNC_QUEUES_CMD_VALID, 0); 103 p_queue->cmdq_num_entries = 0; 104 p_queue->bdq_resource_id = (u8)RESC_START(p_hwfn, QED_BDQ); 105 p_ramrod->tcp_init.two_msl_timer = cpu_to_le32(QED_TCP_TWO_MSL_TIMER); 106 p_ramrod->tcp_init.tx_sws_timer = cpu_to_le16(QED_TCP_SWS_TIMER); 107 p_init->half_way_close_timeout = cpu_to_le16(QED_TCP_HALF_WAY_CLOSE_TIMEOUT); 108 p_ramrod->tcp_init.max_fin_rt = QED_TCP_MAX_FIN_RT; 109 SET_FIELD(p_ramrod->nvmetcp_init_spe.params, 110 NVMETCP_SPE_FUNC_INIT_MAX_SYN_RT, QED_TCP_MAX_FIN_RT); 111 p_hwfn->p_nvmetcp_info->event_context = event_context; 112 p_hwfn->p_nvmetcp_info->event_cb = async_event_cb; 113 qed_spq_register_async_cb(p_hwfn, PROTOCOLID_TCP_ULP, 114 qed_nvmetcp_async_event); 115 116 return qed_spq_post(p_hwfn, p_ent, NULL); 117 } 118 119 static int qed_sp_nvmetcp_func_stop(struct qed_hwfn *p_hwfn, 120 enum spq_mode comp_mode, 121 struct qed_spq_comp_cb *p_comp_addr) 122 { 123 struct qed_spq_entry *p_ent = NULL; 124 struct qed_sp_init_data init_data; 125 int rc; 126 127 /* Get SPQ entry */ 128 memset(&init_data, 0, sizeof(init_data)); 129 init_data.cid = qed_spq_get_cid(p_hwfn); 130 init_data.opaque_fid = p_hwfn->hw_info.opaque_fid; 131 init_data.comp_mode = comp_mode; 132 init_data.p_comp_data = p_comp_addr; 133 rc = qed_sp_init_request(p_hwfn, &p_ent, 134 NVMETCP_RAMROD_CMD_ID_DESTROY_FUNC, 135 PROTOCOLID_TCP_ULP, &init_data); 136 if (rc) 137 return rc; 138 139 rc = qed_spq_post(p_hwfn, p_ent, NULL); 140 qed_spq_unregister_async_cb(p_hwfn, PROTOCOLID_TCP_ULP); 141 142 return rc; 143 } 144 145 static int qed_fill_nvmetcp_dev_info(struct qed_dev *cdev, 146 struct qed_dev_nvmetcp_info *info) 147 { 148 struct qed_hwfn *hwfn = QED_AFFIN_HWFN(cdev); 149 int rc; 150 151 memset(info, 0, sizeof(*info)); 152 rc = qed_fill_dev_info(cdev, &info->common); 153 info->port_id = MFW_PORT(hwfn); 154 info->num_cqs = FEAT_NUM(hwfn, QED_NVMETCP_CQ); 155 156 return rc; 157 } 158 159 static void qed_register_nvmetcp_ops(struct qed_dev *cdev, 160 struct qed_nvmetcp_cb_ops *ops, 161 void *cookie) 162 { 163 cdev->protocol_ops.nvmetcp = ops; 164 cdev->ops_cookie = cookie; 165 } 166 167 static int qed_nvmetcp_stop(struct qed_dev *cdev) 168 { 169 int rc; 170 171 if (!(cdev->flags & QED_FLAG_STORAGE_STARTED)) { 172 DP_NOTICE(cdev, "nvmetcp already stopped\n"); 173 174 return 0; 175 } 176 177 if (!hash_empty(cdev->connections)) { 178 DP_NOTICE(cdev, 179 "Can't stop nvmetcp - not all connections were returned\n"); 180 181 return -EINVAL; 182 } 183 184 /* Stop the nvmetcp */ 185 rc = qed_sp_nvmetcp_func_stop(QED_AFFIN_HWFN(cdev), QED_SPQ_MODE_EBLOCK, 186 NULL); 187 cdev->flags &= ~QED_FLAG_STORAGE_STARTED; 188 189 return rc; 190 } 191 192 static int qed_nvmetcp_start(struct qed_dev *cdev, 193 struct qed_nvmetcp_tid *tasks, 194 void *event_context, 195 nvmetcp_event_cb_t async_event_cb) 196 { 197 struct qed_tid_mem *tid_info; 198 int rc; 199 200 if (cdev->flags & QED_FLAG_STORAGE_STARTED) { 201 DP_NOTICE(cdev, "nvmetcp already started;\n"); 202 203 return 0; 204 } 205 206 rc = qed_sp_nvmetcp_func_start(QED_AFFIN_HWFN(cdev), 207 QED_SPQ_MODE_EBLOCK, NULL, 208 event_context, async_event_cb); 209 if (rc) { 210 DP_NOTICE(cdev, "Failed to start nvmetcp\n"); 211 212 return rc; 213 } 214 215 cdev->flags |= QED_FLAG_STORAGE_STARTED; 216 hash_init(cdev->connections); 217 218 if (!tasks) 219 return 0; 220 221 tid_info = kzalloc(sizeof(*tid_info), GFP_KERNEL); 222 if (!tid_info) { 223 qed_nvmetcp_stop(cdev); 224 225 return -ENOMEM; 226 } 227 228 rc = qed_cxt_get_tid_mem_info(QED_AFFIN_HWFN(cdev), tid_info); 229 if (rc) { 230 DP_NOTICE(cdev, "Failed to gather task information\n"); 231 qed_nvmetcp_stop(cdev); 232 kfree(tid_info); 233 234 return rc; 235 } 236 237 /* Fill task information */ 238 tasks->size = tid_info->tid_size; 239 tasks->num_tids_per_block = tid_info->num_tids_per_block; 240 memcpy(tasks->blocks, tid_info->blocks, 241 MAX_TID_BLOCKS_NVMETCP * sizeof(u8 *)); 242 kfree(tid_info); 243 244 return 0; 245 } 246 247 static struct qed_hash_nvmetcp_con *qed_nvmetcp_get_hash(struct qed_dev *cdev, 248 u32 handle) 249 { 250 struct qed_hash_nvmetcp_con *hash_con = NULL; 251 252 if (!(cdev->flags & QED_FLAG_STORAGE_STARTED)) 253 return NULL; 254 255 hash_for_each_possible(cdev->connections, hash_con, node, handle) { 256 if (hash_con->con->icid == handle) 257 break; 258 } 259 260 if (!hash_con || hash_con->con->icid != handle) 261 return NULL; 262 263 return hash_con; 264 } 265 266 static int qed_sp_nvmetcp_conn_offload(struct qed_hwfn *p_hwfn, 267 struct qed_nvmetcp_conn *p_conn, 268 enum spq_mode comp_mode, 269 struct qed_spq_comp_cb *p_comp_addr) 270 { 271 struct nvmetcp_spe_conn_offload *p_ramrod = NULL; 272 struct tcp_offload_params_opt2 *p_tcp = NULL; 273 struct qed_sp_init_data init_data = { 0 }; 274 struct qed_spq_entry *p_ent = NULL; 275 dma_addr_t r2tq_pbl_addr; 276 dma_addr_t xhq_pbl_addr; 277 dma_addr_t uhq_pbl_addr; 278 u16 physical_q; 279 int rc = 0; 280 u8 i; 281 282 /* Get SPQ entry */ 283 init_data.cid = p_conn->icid; 284 init_data.opaque_fid = p_hwfn->hw_info.opaque_fid; 285 init_data.comp_mode = comp_mode; 286 init_data.p_comp_data = p_comp_addr; 287 rc = qed_sp_init_request(p_hwfn, &p_ent, 288 NVMETCP_RAMROD_CMD_ID_OFFLOAD_CONN, 289 PROTOCOLID_TCP_ULP, &init_data); 290 if (rc) 291 return rc; 292 293 p_ramrod = &p_ent->ramrod.nvmetcp_conn_offload; 294 295 /* Transmission PQ is the first of the PF */ 296 physical_q = qed_get_cm_pq_idx(p_hwfn, PQ_FLAGS_OFLD); 297 p_conn->physical_q0 = cpu_to_le16(physical_q); 298 p_ramrod->nvmetcp.physical_q0 = cpu_to_le16(physical_q); 299 300 /* nvmetcp Pure-ACK PQ */ 301 physical_q = qed_get_cm_pq_idx(p_hwfn, PQ_FLAGS_ACK); 302 p_conn->physical_q1 = cpu_to_le16(physical_q); 303 p_ramrod->nvmetcp.physical_q1 = cpu_to_le16(physical_q); 304 p_ramrod->conn_id = cpu_to_le16(p_conn->conn_id); 305 DMA_REGPAIR_LE(p_ramrod->nvmetcp.sq_pbl_addr, p_conn->sq_pbl_addr); 306 r2tq_pbl_addr = qed_chain_get_pbl_phys(&p_conn->r2tq); 307 DMA_REGPAIR_LE(p_ramrod->nvmetcp.r2tq_pbl_addr, r2tq_pbl_addr); 308 xhq_pbl_addr = qed_chain_get_pbl_phys(&p_conn->xhq); 309 DMA_REGPAIR_LE(p_ramrod->nvmetcp.xhq_pbl_addr, xhq_pbl_addr); 310 uhq_pbl_addr = qed_chain_get_pbl_phys(&p_conn->uhq); 311 DMA_REGPAIR_LE(p_ramrod->nvmetcp.uhq_pbl_addr, uhq_pbl_addr); 312 p_ramrod->nvmetcp.flags = p_conn->offl_flags; 313 p_ramrod->nvmetcp.default_cq = p_conn->default_cq; 314 p_ramrod->nvmetcp.initial_ack = 0; 315 DMA_REGPAIR_LE(p_ramrod->nvmetcp.nvmetcp.cccid_itid_table_addr, 316 p_conn->nvmetcp_cccid_itid_table_addr); 317 p_ramrod->nvmetcp.nvmetcp.cccid_max_range = 318 cpu_to_le16(p_conn->nvmetcp_cccid_max_range); 319 p_tcp = &p_ramrod->tcp; 320 qed_set_fw_mac_addr(&p_tcp->remote_mac_addr_hi, 321 &p_tcp->remote_mac_addr_mid, 322 &p_tcp->remote_mac_addr_lo, p_conn->remote_mac); 323 qed_set_fw_mac_addr(&p_tcp->local_mac_addr_hi, 324 &p_tcp->local_mac_addr_mid, 325 &p_tcp->local_mac_addr_lo, p_conn->local_mac); 326 p_tcp->vlan_id = cpu_to_le16(p_conn->vlan_id); 327 p_tcp->flags = cpu_to_le16(p_conn->tcp_flags); 328 p_tcp->ip_version = p_conn->ip_version; 329 if (p_tcp->ip_version == TCP_IPV6) { 330 for (i = 0; i < 4; i++) { 331 p_tcp->remote_ip[i] = cpu_to_le32(p_conn->remote_ip[i]); 332 p_tcp->local_ip[i] = cpu_to_le32(p_conn->local_ip[i]); 333 } 334 } else { 335 p_tcp->remote_ip[0] = cpu_to_le32(p_conn->remote_ip[0]); 336 p_tcp->local_ip[0] = cpu_to_le32(p_conn->local_ip[0]); 337 } 338 339 p_tcp->flow_label = cpu_to_le32(p_conn->flow_label); 340 p_tcp->ttl = p_conn->ttl; 341 p_tcp->tos_or_tc = p_conn->tos_or_tc; 342 p_tcp->remote_port = cpu_to_le16(p_conn->remote_port); 343 p_tcp->local_port = cpu_to_le16(p_conn->local_port); 344 p_tcp->mss = cpu_to_le16(p_conn->mss); 345 p_tcp->rcv_wnd_scale = p_conn->rcv_wnd_scale; 346 p_tcp->connect_mode = p_conn->connect_mode; 347 p_tcp->cwnd = cpu_to_le32(p_conn->cwnd); 348 p_tcp->ka_max_probe_cnt = p_conn->ka_max_probe_cnt; 349 p_tcp->ka_timeout = cpu_to_le32(p_conn->ka_timeout); 350 p_tcp->max_rt_time = cpu_to_le32(p_conn->max_rt_time); 351 p_tcp->ka_interval = cpu_to_le32(p_conn->ka_interval); 352 353 return qed_spq_post(p_hwfn, p_ent, NULL); 354 } 355 356 static int qed_sp_nvmetcp_conn_update(struct qed_hwfn *p_hwfn, 357 struct qed_nvmetcp_conn *p_conn, 358 enum spq_mode comp_mode, 359 struct qed_spq_comp_cb *p_comp_addr) 360 { 361 struct nvmetcp_conn_update_ramrod_params *p_ramrod = NULL; 362 struct qed_spq_entry *p_ent = NULL; 363 struct qed_sp_init_data init_data; 364 int rc = -EINVAL; 365 u32 dval; 366 367 /* Get SPQ entry */ 368 memset(&init_data, 0, sizeof(init_data)); 369 init_data.cid = p_conn->icid; 370 init_data.opaque_fid = p_hwfn->hw_info.opaque_fid; 371 init_data.comp_mode = comp_mode; 372 init_data.p_comp_data = p_comp_addr; 373 374 rc = qed_sp_init_request(p_hwfn, &p_ent, 375 NVMETCP_RAMROD_CMD_ID_UPDATE_CONN, 376 PROTOCOLID_TCP_ULP, &init_data); 377 if (rc) 378 return rc; 379 380 p_ramrod = &p_ent->ramrod.nvmetcp_conn_update; 381 p_ramrod->conn_id = cpu_to_le16(p_conn->conn_id); 382 p_ramrod->flags = p_conn->update_flag; 383 p_ramrod->max_seq_size = cpu_to_le32(p_conn->max_seq_size); 384 dval = p_conn->max_recv_pdu_length; 385 p_ramrod->max_recv_pdu_length = cpu_to_le32(dval); 386 dval = p_conn->max_send_pdu_length; 387 p_ramrod->max_send_pdu_length = cpu_to_le32(dval); 388 p_ramrod->first_seq_length = cpu_to_le32(p_conn->first_seq_length); 389 390 return qed_spq_post(p_hwfn, p_ent, NULL); 391 } 392 393 static int qed_sp_nvmetcp_conn_terminate(struct qed_hwfn *p_hwfn, 394 struct qed_nvmetcp_conn *p_conn, 395 enum spq_mode comp_mode, 396 struct qed_spq_comp_cb *p_comp_addr) 397 { 398 struct nvmetcp_spe_conn_termination *p_ramrod = NULL; 399 struct qed_spq_entry *p_ent = NULL; 400 struct qed_sp_init_data init_data; 401 int rc = -EINVAL; 402 403 /* Get SPQ entry */ 404 memset(&init_data, 0, sizeof(init_data)); 405 init_data.cid = p_conn->icid; 406 init_data.opaque_fid = p_hwfn->hw_info.opaque_fid; 407 init_data.comp_mode = comp_mode; 408 init_data.p_comp_data = p_comp_addr; 409 rc = qed_sp_init_request(p_hwfn, &p_ent, 410 NVMETCP_RAMROD_CMD_ID_TERMINATION_CONN, 411 PROTOCOLID_TCP_ULP, &init_data); 412 if (rc) 413 return rc; 414 415 p_ramrod = &p_ent->ramrod.nvmetcp_conn_terminate; 416 p_ramrod->conn_id = cpu_to_le16(p_conn->conn_id); 417 p_ramrod->abortive = p_conn->abortive_dsconnect; 418 419 return qed_spq_post(p_hwfn, p_ent, NULL); 420 } 421 422 static int qed_sp_nvmetcp_conn_clear_sq(struct qed_hwfn *p_hwfn, 423 struct qed_nvmetcp_conn *p_conn, 424 enum spq_mode comp_mode, 425 struct qed_spq_comp_cb *p_comp_addr) 426 { 427 struct qed_spq_entry *p_ent = NULL; 428 struct qed_sp_init_data init_data; 429 int rc = -EINVAL; 430 431 /* Get SPQ entry */ 432 memset(&init_data, 0, sizeof(init_data)); 433 init_data.cid = p_conn->icid; 434 init_data.opaque_fid = p_hwfn->hw_info.opaque_fid; 435 init_data.comp_mode = comp_mode; 436 init_data.p_comp_data = p_comp_addr; 437 rc = qed_sp_init_request(p_hwfn, &p_ent, 438 NVMETCP_RAMROD_CMD_ID_CLEAR_SQ, 439 PROTOCOLID_TCP_ULP, &init_data); 440 if (rc) 441 return rc; 442 443 return qed_spq_post(p_hwfn, p_ent, NULL); 444 } 445 446 static void __iomem *qed_nvmetcp_get_db_addr(struct qed_hwfn *p_hwfn, u32 cid) 447 { 448 return (u8 __iomem *)p_hwfn->doorbells + 449 qed_db_addr(cid, DQ_DEMS_LEGACY); 450 } 451 452 static int qed_nvmetcp_allocate_connection(struct qed_hwfn *p_hwfn, 453 struct qed_nvmetcp_conn **p_out_conn) 454 { 455 struct qed_chain_init_params params = { 456 .mode = QED_CHAIN_MODE_PBL, 457 .intended_use = QED_CHAIN_USE_TO_CONSUME_PRODUCE, 458 .cnt_type = QED_CHAIN_CNT_TYPE_U16, 459 }; 460 struct qed_nvmetcp_pf_params *p_params = NULL; 461 struct qed_nvmetcp_conn *p_conn = NULL; 462 int rc = 0; 463 464 /* Try finding a free connection that can be used */ 465 spin_lock_bh(&p_hwfn->p_nvmetcp_info->lock); 466 if (!list_empty(&p_hwfn->p_nvmetcp_info->free_list)) 467 p_conn = list_first_entry(&p_hwfn->p_nvmetcp_info->free_list, 468 struct qed_nvmetcp_conn, list_entry); 469 if (p_conn) { 470 list_del(&p_conn->list_entry); 471 spin_unlock_bh(&p_hwfn->p_nvmetcp_info->lock); 472 *p_out_conn = p_conn; 473 474 return 0; 475 } 476 spin_unlock_bh(&p_hwfn->p_nvmetcp_info->lock); 477 478 /* Need to allocate a new connection */ 479 p_params = &p_hwfn->pf_params.nvmetcp_pf_params; 480 p_conn = kzalloc(sizeof(*p_conn), GFP_KERNEL); 481 if (!p_conn) 482 return -ENOMEM; 483 484 params.num_elems = p_params->num_r2tq_pages_in_ring * 485 QED_CHAIN_PAGE_SIZE / sizeof(struct nvmetcp_wqe); 486 params.elem_size = sizeof(struct nvmetcp_wqe); 487 rc = qed_chain_alloc(p_hwfn->cdev, &p_conn->r2tq, ¶ms); 488 if (rc) 489 goto nomem_r2tq; 490 491 params.num_elems = p_params->num_uhq_pages_in_ring * 492 QED_CHAIN_PAGE_SIZE / sizeof(struct iscsi_uhqe); 493 params.elem_size = sizeof(struct iscsi_uhqe); 494 rc = qed_chain_alloc(p_hwfn->cdev, &p_conn->uhq, ¶ms); 495 if (rc) 496 goto nomem_uhq; 497 498 params.elem_size = sizeof(struct iscsi_xhqe); 499 rc = qed_chain_alloc(p_hwfn->cdev, &p_conn->xhq, ¶ms); 500 if (rc) 501 goto nomem; 502 503 p_conn->free_on_delete = true; 504 *p_out_conn = p_conn; 505 506 return 0; 507 508 nomem: 509 qed_chain_free(p_hwfn->cdev, &p_conn->uhq); 510 nomem_uhq: 511 qed_chain_free(p_hwfn->cdev, &p_conn->r2tq); 512 nomem_r2tq: 513 kfree(p_conn); 514 515 return -ENOMEM; 516 } 517 518 static int qed_nvmetcp_acquire_connection(struct qed_hwfn *p_hwfn, 519 struct qed_nvmetcp_conn **p_out_conn) 520 { 521 struct qed_nvmetcp_conn *p_conn = NULL; 522 int rc = 0; 523 u32 icid; 524 525 spin_lock_bh(&p_hwfn->p_nvmetcp_info->lock); 526 rc = qed_cxt_acquire_cid(p_hwfn, PROTOCOLID_TCP_ULP, &icid); 527 spin_unlock_bh(&p_hwfn->p_nvmetcp_info->lock); 528 529 if (rc) 530 return rc; 531 532 rc = qed_nvmetcp_allocate_connection(p_hwfn, &p_conn); 533 if (rc) { 534 spin_lock_bh(&p_hwfn->p_nvmetcp_info->lock); 535 qed_cxt_release_cid(p_hwfn, icid); 536 spin_unlock_bh(&p_hwfn->p_nvmetcp_info->lock); 537 538 return rc; 539 } 540 541 p_conn->icid = icid; 542 p_conn->conn_id = (u16)icid; 543 p_conn->fw_cid = (p_hwfn->hw_info.opaque_fid << 16) | icid; 544 *p_out_conn = p_conn; 545 546 return rc; 547 } 548 549 static void qed_nvmetcp_release_connection(struct qed_hwfn *p_hwfn, 550 struct qed_nvmetcp_conn *p_conn) 551 { 552 spin_lock_bh(&p_hwfn->p_nvmetcp_info->lock); 553 list_add_tail(&p_conn->list_entry, &p_hwfn->p_nvmetcp_info->free_list); 554 qed_cxt_release_cid(p_hwfn, p_conn->icid); 555 spin_unlock_bh(&p_hwfn->p_nvmetcp_info->lock); 556 } 557 558 static void qed_nvmetcp_free_connection(struct qed_hwfn *p_hwfn, 559 struct qed_nvmetcp_conn *p_conn) 560 { 561 qed_chain_free(p_hwfn->cdev, &p_conn->xhq); 562 qed_chain_free(p_hwfn->cdev, &p_conn->uhq); 563 qed_chain_free(p_hwfn->cdev, &p_conn->r2tq); 564 kfree(p_conn); 565 } 566 567 int qed_nvmetcp_alloc(struct qed_hwfn *p_hwfn) 568 { 569 struct qed_nvmetcp_info *p_nvmetcp_info; 570 571 p_nvmetcp_info = kzalloc(sizeof(*p_nvmetcp_info), GFP_KERNEL); 572 if (!p_nvmetcp_info) 573 return -ENOMEM; 574 575 INIT_LIST_HEAD(&p_nvmetcp_info->free_list); 576 p_hwfn->p_nvmetcp_info = p_nvmetcp_info; 577 578 return 0; 579 } 580 581 void qed_nvmetcp_setup(struct qed_hwfn *p_hwfn) 582 { 583 spin_lock_init(&p_hwfn->p_nvmetcp_info->lock); 584 } 585 586 void qed_nvmetcp_free(struct qed_hwfn *p_hwfn) 587 { 588 struct qed_nvmetcp_conn *p_conn = NULL; 589 590 if (!p_hwfn->p_nvmetcp_info) 591 return; 592 593 while (!list_empty(&p_hwfn->p_nvmetcp_info->free_list)) { 594 p_conn = list_first_entry(&p_hwfn->p_nvmetcp_info->free_list, 595 struct qed_nvmetcp_conn, list_entry); 596 if (p_conn) { 597 list_del(&p_conn->list_entry); 598 qed_nvmetcp_free_connection(p_hwfn, p_conn); 599 } 600 } 601 602 kfree(p_hwfn->p_nvmetcp_info); 603 p_hwfn->p_nvmetcp_info = NULL; 604 } 605 606 static int qed_nvmetcp_acquire_conn(struct qed_dev *cdev, 607 u32 *handle, 608 u32 *fw_cid, void __iomem **p_doorbell) 609 { 610 struct qed_hash_nvmetcp_con *hash_con; 611 int rc; 612 613 /* Allocate a hashed connection */ 614 hash_con = kzalloc(sizeof(*hash_con), GFP_ATOMIC); 615 if (!hash_con) 616 return -ENOMEM; 617 618 /* Acquire the connection */ 619 rc = qed_nvmetcp_acquire_connection(QED_AFFIN_HWFN(cdev), 620 &hash_con->con); 621 if (rc) { 622 DP_NOTICE(cdev, "Failed to acquire Connection\n"); 623 kfree(hash_con); 624 625 return rc; 626 } 627 628 /* Added the connection to hash table */ 629 *handle = hash_con->con->icid; 630 *fw_cid = hash_con->con->fw_cid; 631 hash_add(cdev->connections, &hash_con->node, *handle); 632 if (p_doorbell) 633 *p_doorbell = qed_nvmetcp_get_db_addr(QED_AFFIN_HWFN(cdev), 634 *handle); 635 636 return 0; 637 } 638 639 static int qed_nvmetcp_release_conn(struct qed_dev *cdev, u32 handle) 640 { 641 struct qed_hash_nvmetcp_con *hash_con; 642 643 hash_con = qed_nvmetcp_get_hash(cdev, handle); 644 if (!hash_con) { 645 DP_NOTICE(cdev, "Failed to find connection for handle %d\n", 646 handle); 647 648 return -EINVAL; 649 } 650 651 hlist_del(&hash_con->node); 652 qed_nvmetcp_release_connection(QED_AFFIN_HWFN(cdev), hash_con->con); 653 kfree(hash_con); 654 655 return 0; 656 } 657 658 static int qed_nvmetcp_offload_conn(struct qed_dev *cdev, u32 handle, 659 struct qed_nvmetcp_params_offload *conn_info) 660 { 661 struct qed_hash_nvmetcp_con *hash_con; 662 struct qed_nvmetcp_conn *con; 663 664 hash_con = qed_nvmetcp_get_hash(cdev, handle); 665 if (!hash_con) { 666 DP_NOTICE(cdev, "Failed to find connection for handle %d\n", 667 handle); 668 669 return -EINVAL; 670 } 671 672 /* Update the connection with information from the params */ 673 con = hash_con->con; 674 675 /* FW initializations */ 676 con->layer_code = NVMETCP_SLOW_PATH_LAYER_CODE; 677 con->sq_pbl_addr = conn_info->sq_pbl_addr; 678 con->nvmetcp_cccid_max_range = conn_info->nvmetcp_cccid_max_range; 679 con->nvmetcp_cccid_itid_table_addr = conn_info->nvmetcp_cccid_itid_table_addr; 680 con->default_cq = conn_info->default_cq; 681 SET_FIELD(con->offl_flags, NVMETCP_CONN_OFFLOAD_PARAMS_TARGET_MODE, 0); 682 SET_FIELD(con->offl_flags, NVMETCP_CONN_OFFLOAD_PARAMS_NVMETCP_MODE, 1); 683 SET_FIELD(con->offl_flags, NVMETCP_CONN_OFFLOAD_PARAMS_TCP_ON_CHIP_1B, 1); 684 685 /* Networking and TCP stack initializations */ 686 ether_addr_copy(con->local_mac, conn_info->src.mac); 687 ether_addr_copy(con->remote_mac, conn_info->dst.mac); 688 memcpy(con->local_ip, conn_info->src.ip, sizeof(con->local_ip)); 689 memcpy(con->remote_ip, conn_info->dst.ip, sizeof(con->remote_ip)); 690 con->local_port = conn_info->src.port; 691 con->remote_port = conn_info->dst.port; 692 con->vlan_id = conn_info->vlan_id; 693 694 if (conn_info->timestamp_en) 695 SET_FIELD(con->tcp_flags, TCP_OFFLOAD_PARAMS_OPT2_TS_EN, 1); 696 697 if (conn_info->delayed_ack_en) 698 SET_FIELD(con->tcp_flags, TCP_OFFLOAD_PARAMS_OPT2_DA_EN, 1); 699 700 if (conn_info->tcp_keep_alive_en) 701 SET_FIELD(con->tcp_flags, TCP_OFFLOAD_PARAMS_OPT2_KA_EN, 1); 702 703 if (conn_info->ecn_en) 704 SET_FIELD(con->tcp_flags, TCP_OFFLOAD_PARAMS_OPT2_ECN_EN, 1); 705 706 con->ip_version = conn_info->ip_version; 707 con->flow_label = QED_TCP_FLOW_LABEL; 708 con->ka_max_probe_cnt = conn_info->ka_max_probe_cnt; 709 con->ka_timeout = conn_info->ka_timeout; 710 con->ka_interval = conn_info->ka_interval; 711 con->max_rt_time = conn_info->max_rt_time; 712 con->ttl = conn_info->ttl; 713 con->tos_or_tc = conn_info->tos_or_tc; 714 con->mss = conn_info->mss; 715 con->cwnd = conn_info->cwnd; 716 con->rcv_wnd_scale = conn_info->rcv_wnd_scale; 717 con->connect_mode = 0; 718 719 return qed_sp_nvmetcp_conn_offload(QED_AFFIN_HWFN(cdev), con, 720 QED_SPQ_MODE_EBLOCK, NULL); 721 } 722 723 static int qed_nvmetcp_update_conn(struct qed_dev *cdev, 724 u32 handle, 725 struct qed_nvmetcp_params_update *conn_info) 726 { 727 struct qed_hash_nvmetcp_con *hash_con; 728 struct qed_nvmetcp_conn *con; 729 730 hash_con = qed_nvmetcp_get_hash(cdev, handle); 731 if (!hash_con) { 732 DP_NOTICE(cdev, "Failed to find connection for handle %d\n", 733 handle); 734 735 return -EINVAL; 736 } 737 738 /* Update the connection with information from the params */ 739 con = hash_con->con; 740 SET_FIELD(con->update_flag, 741 ISCSI_CONN_UPDATE_RAMROD_PARAMS_INITIAL_R2T, 0); 742 SET_FIELD(con->update_flag, 743 ISCSI_CONN_UPDATE_RAMROD_PARAMS_IMMEDIATE_DATA, 1); 744 if (conn_info->hdr_digest_en) 745 SET_FIELD(con->update_flag, ISCSI_CONN_UPDATE_RAMROD_PARAMS_HD_EN, 1); 746 747 if (conn_info->data_digest_en) 748 SET_FIELD(con->update_flag, ISCSI_CONN_UPDATE_RAMROD_PARAMS_DD_EN, 1); 749 750 /* Placeholder - initialize pfv, cpda, hpda */ 751 752 con->max_seq_size = conn_info->max_io_size; 753 con->max_recv_pdu_length = conn_info->max_recv_pdu_length; 754 con->max_send_pdu_length = conn_info->max_send_pdu_length; 755 con->first_seq_length = conn_info->max_io_size; 756 757 return qed_sp_nvmetcp_conn_update(QED_AFFIN_HWFN(cdev), con, 758 QED_SPQ_MODE_EBLOCK, NULL); 759 } 760 761 static int qed_nvmetcp_clear_conn_sq(struct qed_dev *cdev, u32 handle) 762 { 763 struct qed_hash_nvmetcp_con *hash_con; 764 765 hash_con = qed_nvmetcp_get_hash(cdev, handle); 766 if (!hash_con) { 767 DP_NOTICE(cdev, "Failed to find connection for handle %d\n", 768 handle); 769 770 return -EINVAL; 771 } 772 773 return qed_sp_nvmetcp_conn_clear_sq(QED_AFFIN_HWFN(cdev), hash_con->con, 774 QED_SPQ_MODE_EBLOCK, NULL); 775 } 776 777 static int qed_nvmetcp_destroy_conn(struct qed_dev *cdev, 778 u32 handle, u8 abrt_conn) 779 { 780 struct qed_hash_nvmetcp_con *hash_con; 781 782 hash_con = qed_nvmetcp_get_hash(cdev, handle); 783 if (!hash_con) { 784 DP_NOTICE(cdev, "Failed to find connection for handle %d\n", 785 handle); 786 787 return -EINVAL; 788 } 789 790 hash_con->con->abortive_dsconnect = abrt_conn; 791 792 return qed_sp_nvmetcp_conn_terminate(QED_AFFIN_HWFN(cdev), hash_con->con, 793 QED_SPQ_MODE_EBLOCK, NULL); 794 } 795 796 static const struct qed_nvmetcp_ops qed_nvmetcp_ops_pass = { 797 .common = &qed_common_ops_pass, 798 .ll2 = &qed_ll2_ops_pass, 799 .fill_dev_info = &qed_fill_nvmetcp_dev_info, 800 .register_ops = &qed_register_nvmetcp_ops, 801 .start = &qed_nvmetcp_start, 802 .stop = &qed_nvmetcp_stop, 803 .acquire_conn = &qed_nvmetcp_acquire_conn, 804 .release_conn = &qed_nvmetcp_release_conn, 805 .offload_conn = &qed_nvmetcp_offload_conn, 806 .update_conn = &qed_nvmetcp_update_conn, 807 .destroy_conn = &qed_nvmetcp_destroy_conn, 808 .clear_sq = &qed_nvmetcp_clear_conn_sq, 809 .add_src_tcp_port_filter = &qed_llh_add_src_tcp_port_filter, 810 .remove_src_tcp_port_filter = &qed_llh_remove_src_tcp_port_filter, 811 .add_dst_tcp_port_filter = &qed_llh_add_dst_tcp_port_filter, 812 .remove_dst_tcp_port_filter = &qed_llh_remove_dst_tcp_port_filter, 813 .clear_all_filters = &qed_llh_clear_all_filters, 814 .init_read_io = &init_nvmetcp_host_read_task, 815 .init_write_io = &init_nvmetcp_host_write_task, 816 .init_icreq_exchange = &init_nvmetcp_init_conn_req_task, 817 .init_task_cleanup = &init_cleanup_task_nvmetcp 818 }; 819 820 const struct qed_nvmetcp_ops *qed_get_nvmetcp_ops(void) 821 { 822 return &qed_nvmetcp_ops_pass; 823 } 824 EXPORT_SYMBOL(qed_get_nvmetcp_ops); 825 826 void qed_put_nvmetcp_ops(void) 827 { 828 } 829 EXPORT_SYMBOL(qed_put_nvmetcp_ops); 830