xref: /linux/drivers/scsi/elx/efct/efct_lio.c (revision 3a39d672e7f48b8d6b91a09afa4b55352773b4b5)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (C) 2021 Broadcom. All Rights Reserved. The term
4  * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries.
5  */
6 
7 #include <target/target_core_base.h>
8 #include <target/target_core_fabric.h>
9 #include "efct_driver.h"
10 #include "efct_lio.h"
11 
12 /*
13  * lio_wq is used to call the LIO backed during creation or deletion of
14  * sessions. This brings serialization to the session management as we create
15  * single threaded work queue.
16  */
17 static struct workqueue_struct *lio_wq;
18 
19 static int
efct_format_wwn(char * str,size_t len,const char * pre,u64 wwn)20 efct_format_wwn(char *str, size_t len, const char *pre, u64 wwn)
21 {
22 	u8 a[8];
23 
24 	put_unaligned_be64(wwn, a);
25 	return snprintf(str, len, "%s%8phC", pre, a);
26 }
27 
28 static int
efct_lio_parse_wwn(const char * name,u64 * wwp,u8 npiv)29 efct_lio_parse_wwn(const char *name, u64 *wwp, u8 npiv)
30 {
31 	int num;
32 	u8 b[8];
33 
34 	if (npiv) {
35 		num = sscanf(name,
36 			     "%02hhx%02hhx%02hhx%02hhx%02hhx%02hhx%02hhx%02hhx",
37 			     &b[0], &b[1], &b[2], &b[3], &b[4], &b[5], &b[6],
38 			     &b[7]);
39 	} else {
40 		num = sscanf(name,
41 		      "%02hhx:%02hhx:%02hhx:%02hhx:%02hhx:%02hhx:%02hhx:%02hhx",
42 			     &b[0], &b[1], &b[2], &b[3], &b[4], &b[5], &b[6],
43 			     &b[7]);
44 	}
45 
46 	if (num != 8)
47 		return -EINVAL;
48 
49 	*wwp = get_unaligned_be64(b);
50 	return 0;
51 }
52 
53 static int
efct_lio_parse_npiv_wwn(const char * name,size_t size,u64 * wwpn,u64 * wwnn)54 efct_lio_parse_npiv_wwn(const char *name, size_t size, u64 *wwpn, u64 *wwnn)
55 {
56 	unsigned int cnt = size;
57 	int rc;
58 
59 	*wwpn = *wwnn = 0;
60 	if (name[cnt - 1] == '\n' || name[cnt - 1] == 0)
61 		cnt--;
62 
63 	/* validate we have enough characters for WWPN */
64 	if ((cnt != (16 + 1 + 16)) || (name[16] != ':'))
65 		return -EINVAL;
66 
67 	rc = efct_lio_parse_wwn(&name[0], wwpn, 1);
68 	if (rc)
69 		return rc;
70 
71 	rc = efct_lio_parse_wwn(&name[17], wwnn, 1);
72 	if (rc)
73 		return rc;
74 
75 	return 0;
76 }
77 
78 static ssize_t
efct_lio_tpg_enable_show(struct config_item * item,char * page)79 efct_lio_tpg_enable_show(struct config_item *item, char *page)
80 {
81 	struct se_portal_group *se_tpg = to_tpg(item);
82 	struct efct_lio_tpg *tpg =
83 		container_of(se_tpg, struct efct_lio_tpg, tpg);
84 
85 	return snprintf(page, PAGE_SIZE, "%d\n", tpg->enabled);
86 }
87 
88 static ssize_t
efct_lio_tpg_enable_store(struct config_item * item,const char * page,size_t count)89 efct_lio_tpg_enable_store(struct config_item *item, const char *page,
90 			  size_t count)
91 {
92 	struct se_portal_group *se_tpg = to_tpg(item);
93 	struct efct_lio_tpg *tpg =
94 		container_of(se_tpg, struct efct_lio_tpg, tpg);
95 	struct efct *efct;
96 	struct efc *efc;
97 	unsigned long op;
98 
99 	if (!tpg->nport || !tpg->nport->efct) {
100 		pr_err("%s: Unable to find EFCT device\n", __func__);
101 		return -EINVAL;
102 	}
103 
104 	efct = tpg->nport->efct;
105 	efc = efct->efcport;
106 
107 	if (kstrtoul(page, 0, &op) < 0)
108 		return -EINVAL;
109 
110 	if (op == 1) {
111 		int ret;
112 
113 		tpg->enabled = true;
114 		efc_log_debug(efct, "enable portal group %d\n", tpg->tpgt);
115 
116 		ret = efct_xport_control(efct->xport, EFCT_XPORT_PORT_ONLINE);
117 		if (ret) {
118 			efct->tgt_efct.lio_nport = NULL;
119 			efc_log_debug(efct, "cannot bring port online\n");
120 			return ret;
121 		}
122 	} else if (op == 0) {
123 		efc_log_debug(efct, "disable portal group %d\n", tpg->tpgt);
124 
125 		if (efc->domain && efc->domain->nport)
126 			efct_scsi_tgt_del_nport(efc, efc->domain->nport);
127 
128 		tpg->enabled = false;
129 	} else {
130 		return -EINVAL;
131 	}
132 
133 	return count;
134 }
135 
136 static ssize_t
efct_lio_npiv_tpg_enable_show(struct config_item * item,char * page)137 efct_lio_npiv_tpg_enable_show(struct config_item *item, char *page)
138 {
139 	struct se_portal_group *se_tpg = to_tpg(item);
140 	struct efct_lio_tpg *tpg =
141 		container_of(se_tpg, struct efct_lio_tpg, tpg);
142 
143 	return snprintf(page, PAGE_SIZE, "%d\n", tpg->enabled);
144 }
145 
146 static ssize_t
efct_lio_npiv_tpg_enable_store(struct config_item * item,const char * page,size_t count)147 efct_lio_npiv_tpg_enable_store(struct config_item *item, const char *page,
148 			       size_t count)
149 {
150 	struct se_portal_group *se_tpg = to_tpg(item);
151 	struct efct_lio_tpg *tpg =
152 		container_of(se_tpg, struct efct_lio_tpg, tpg);
153 	struct efct_lio_vport *lio_vport = tpg->vport;
154 	struct efct *efct;
155 	struct efc *efc;
156 	unsigned long op;
157 
158 	if (kstrtoul(page, 0, &op) < 0)
159 		return -EINVAL;
160 
161 	if (!lio_vport) {
162 		pr_err("Unable to find vport\n");
163 		return -EINVAL;
164 	}
165 
166 	efct = lio_vport->efct;
167 	efc = efct->efcport;
168 
169 	if (op == 1) {
170 		tpg->enabled = true;
171 		efc_log_debug(efct, "enable portal group %d\n", tpg->tpgt);
172 
173 		if (efc->domain) {
174 			int ret;
175 
176 			ret = efc_nport_vport_new(efc->domain,
177 						  lio_vport->npiv_wwpn,
178 						  lio_vport->npiv_wwnn,
179 						  U32_MAX, false, true,
180 						  NULL, NULL);
181 			if (ret != 0) {
182 				efc_log_err(efct, "Failed to create Vport\n");
183 				return ret;
184 			}
185 			return count;
186 		}
187 
188 		if (!(efc_vport_create_spec(efc, lio_vport->npiv_wwnn,
189 					    lio_vport->npiv_wwpn, U32_MAX,
190 					    false, true, NULL, NULL)))
191 			return -ENOMEM;
192 
193 	} else if (op == 0) {
194 		efc_log_debug(efct, "disable portal group %d\n", tpg->tpgt);
195 
196 		tpg->enabled = false;
197 		/* only physical nport should exist, free lio_nport
198 		 * allocated in efct_lio_make_nport
199 		 */
200 		if (efc->domain) {
201 			efc_nport_vport_del(efct->efcport, efc->domain,
202 					    lio_vport->npiv_wwpn,
203 					    lio_vport->npiv_wwnn);
204 			return count;
205 		}
206 	} else {
207 		return -EINVAL;
208 	}
209 	return count;
210 }
211 
efct_lio_get_fabric_wwn(struct se_portal_group * se_tpg)212 static char *efct_lio_get_fabric_wwn(struct se_portal_group *se_tpg)
213 {
214 	struct efct_lio_tpg *tpg =
215 		container_of(se_tpg, struct efct_lio_tpg, tpg);
216 
217 	return tpg->nport->wwpn_str;
218 }
219 
efct_lio_get_npiv_fabric_wwn(struct se_portal_group * se_tpg)220 static char *efct_lio_get_npiv_fabric_wwn(struct se_portal_group *se_tpg)
221 {
222 	struct efct_lio_tpg *tpg =
223 		container_of(se_tpg, struct efct_lio_tpg, tpg);
224 
225 	return tpg->vport->wwpn_str;
226 }
227 
efct_lio_get_tag(struct se_portal_group * se_tpg)228 static u16 efct_lio_get_tag(struct se_portal_group *se_tpg)
229 {
230 	struct efct_lio_tpg *tpg =
231 		container_of(se_tpg, struct efct_lio_tpg, tpg);
232 
233 	return tpg->tpgt;
234 }
235 
efct_lio_get_npiv_tag(struct se_portal_group * se_tpg)236 static u16 efct_lio_get_npiv_tag(struct se_portal_group *se_tpg)
237 {
238 	struct efct_lio_tpg *tpg =
239 		container_of(se_tpg, struct efct_lio_tpg, tpg);
240 
241 	return tpg->tpgt;
242 }
243 
efct_lio_check_demo_mode(struct se_portal_group * se_tpg)244 static int efct_lio_check_demo_mode(struct se_portal_group *se_tpg)
245 {
246 	return 1;
247 }
248 
efct_lio_check_demo_mode_cache(struct se_portal_group * se_tpg)249 static int efct_lio_check_demo_mode_cache(struct se_portal_group *se_tpg)
250 {
251 	return 1;
252 }
253 
efct_lio_check_demo_write_protect(struct se_portal_group * se_tpg)254 static int efct_lio_check_demo_write_protect(struct se_portal_group *se_tpg)
255 {
256 	struct efct_lio_tpg *tpg =
257 		container_of(se_tpg, struct efct_lio_tpg, tpg);
258 
259 	return tpg->tpg_attrib.demo_mode_write_protect;
260 }
261 
262 static int
efct_lio_npiv_check_demo_write_protect(struct se_portal_group * se_tpg)263 efct_lio_npiv_check_demo_write_protect(struct se_portal_group *se_tpg)
264 {
265 	struct efct_lio_tpg *tpg =
266 		container_of(se_tpg, struct efct_lio_tpg, tpg);
267 
268 	return tpg->tpg_attrib.demo_mode_write_protect;
269 }
270 
efct_lio_check_prod_write_protect(struct se_portal_group * se_tpg)271 static int efct_lio_check_prod_write_protect(struct se_portal_group *se_tpg)
272 {
273 	struct efct_lio_tpg *tpg =
274 		container_of(se_tpg, struct efct_lio_tpg, tpg);
275 
276 	return tpg->tpg_attrib.prod_mode_write_protect;
277 }
278 
279 static int
efct_lio_npiv_check_prod_write_protect(struct se_portal_group * se_tpg)280 efct_lio_npiv_check_prod_write_protect(struct se_portal_group *se_tpg)
281 {
282 	struct efct_lio_tpg *tpg =
283 		container_of(se_tpg, struct efct_lio_tpg, tpg);
284 
285 	return tpg->tpg_attrib.prod_mode_write_protect;
286 }
287 
efct_lio_check_stop_free(struct se_cmd * se_cmd)288 static int efct_lio_check_stop_free(struct se_cmd *se_cmd)
289 {
290 	struct efct_scsi_tgt_io *ocp =
291 		container_of(se_cmd, struct efct_scsi_tgt_io, cmd);
292 	struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
293 
294 	efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_CHK_STOP_FREE);
295 	return target_put_sess_cmd(se_cmd);
296 }
297 
298 static int
efct_lio_abort_tgt_cb(struct efct_io * io,enum efct_scsi_io_status scsi_status,u32 flags,void * arg)299 efct_lio_abort_tgt_cb(struct efct_io *io,
300 		      enum efct_scsi_io_status scsi_status,
301 		      u32 flags, void *arg)
302 {
303 	efct_lio_io_printf(io, "Abort done, status:%d\n", scsi_status);
304 	return 0;
305 }
306 
307 static void
efct_lio_aborted_task(struct se_cmd * se_cmd)308 efct_lio_aborted_task(struct se_cmd *se_cmd)
309 {
310 	struct efct_scsi_tgt_io *ocp =
311 		container_of(se_cmd, struct efct_scsi_tgt_io, cmd);
312 	struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
313 
314 	efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_ABORTED_TASK);
315 
316 	if (ocp->rsp_sent)
317 		return;
318 
319 	/* command has been aborted, cleanup here */
320 	ocp->aborting = true;
321 	ocp->err = EFCT_SCSI_STATUS_ABORTED;
322 	/* terminate the exchange */
323 	efct_scsi_tgt_abort_io(io, efct_lio_abort_tgt_cb, NULL);
324 }
325 
efct_lio_release_cmd(struct se_cmd * se_cmd)326 static void efct_lio_release_cmd(struct se_cmd *se_cmd)
327 {
328 	struct efct_scsi_tgt_io *ocp =
329 		container_of(se_cmd, struct efct_scsi_tgt_io, cmd);
330 	struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
331 	struct efct *efct = io->efct;
332 
333 	efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_RELEASE_CMD);
334 	efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_CMPL_CMD);
335 	efct_scsi_io_complete(io);
336 	atomic_sub_return(1, &efct->tgt_efct.ios_in_use);
337 }
338 
efct_lio_close_session(struct se_session * se_sess)339 static void efct_lio_close_session(struct se_session *se_sess)
340 {
341 	struct efc_node *node = se_sess->fabric_sess_ptr;
342 
343 	pr_debug("se_sess=%p node=%p", se_sess, node);
344 
345 	if (!node) {
346 		pr_debug("node is NULL");
347 		return;
348 	}
349 
350 	efc_node_post_shutdown(node, NULL);
351 }
352 
efct_lio_get_cmd_state(struct se_cmd * cmd)353 static int efct_lio_get_cmd_state(struct se_cmd *cmd)
354 {
355 	struct efct_scsi_tgt_io *ocp =
356 		container_of(cmd, struct efct_scsi_tgt_io, cmd);
357 	struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
358 
359 	return io->tgt_io.state;
360 }
361 
362 static int
efct_lio_sg_map(struct efct_io * io)363 efct_lio_sg_map(struct efct_io *io)
364 {
365 	struct efct_scsi_tgt_io *ocp = &io->tgt_io;
366 	struct se_cmd *cmd = &ocp->cmd;
367 
368 	ocp->seg_map_cnt = dma_map_sg(&io->efct->pci->dev, cmd->t_data_sg,
369 				      cmd->t_data_nents, cmd->data_direction);
370 	if (ocp->seg_map_cnt == 0)
371 		return -EFAULT;
372 	return 0;
373 }
374 
375 static void
efct_lio_sg_unmap(struct efct_io * io)376 efct_lio_sg_unmap(struct efct_io *io)
377 {
378 	struct efct_scsi_tgt_io *ocp = &io->tgt_io;
379 	struct se_cmd *cmd = &ocp->cmd;
380 
381 	if (WARN_ON(!ocp->seg_map_cnt || !cmd->t_data_sg))
382 		return;
383 
384 	dma_unmap_sg(&io->efct->pci->dev, cmd->t_data_sg,
385 		     ocp->seg_map_cnt, cmd->data_direction);
386 	ocp->seg_map_cnt = 0;
387 }
388 
389 static int
efct_lio_status_done(struct efct_io * io,enum efct_scsi_io_status scsi_status,u32 flags,void * arg)390 efct_lio_status_done(struct efct_io *io,
391 		     enum efct_scsi_io_status scsi_status,
392 		     u32 flags, void *arg)
393 {
394 	struct efct_scsi_tgt_io *ocp = &io->tgt_io;
395 
396 	efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_RSP_DONE);
397 	if (scsi_status != EFCT_SCSI_STATUS_GOOD) {
398 		efct_lio_io_printf(io, "callback completed with error=%d\n",
399 				   scsi_status);
400 		ocp->err = scsi_status;
401 	}
402 	if (ocp->seg_map_cnt)
403 		efct_lio_sg_unmap(io);
404 
405 	efct_lio_io_printf(io, "status=%d, err=%d flags=0x%x, dir=%d\n",
406 			   scsi_status, ocp->err, flags, ocp->ddir);
407 
408 	efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_GENERIC_FREE);
409 	transport_generic_free_cmd(&io->tgt_io.cmd, 0);
410 	return 0;
411 }
412 
413 static int
414 efct_lio_datamove_done(struct efct_io *io, enum efct_scsi_io_status scsi_status,
415 		       u32 flags, void *arg);
416 
417 static int
efct_lio_write_pending(struct se_cmd * cmd)418 efct_lio_write_pending(struct se_cmd *cmd)
419 {
420 	struct efct_scsi_tgt_io *ocp =
421 		container_of(cmd, struct efct_scsi_tgt_io, cmd);
422 	struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
423 	struct efct_scsi_sgl *sgl = io->sgl;
424 	struct scatterlist *sg;
425 	u32 flags = 0, cnt, curcnt;
426 	u64 length = 0;
427 
428 	efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_WRITE_PENDING);
429 	efct_lio_io_printf(io, "trans_state=0x%x se_cmd_flags=0x%x\n",
430 			   cmd->transport_state, cmd->se_cmd_flags);
431 
432 	if (ocp->seg_cnt == 0) {
433 		ocp->seg_cnt = cmd->t_data_nents;
434 		ocp->cur_seg = 0;
435 		if (efct_lio_sg_map(io)) {
436 			efct_lio_io_printf(io, "efct_lio_sg_map failed\n");
437 			return -EFAULT;
438 		}
439 	}
440 	curcnt = (ocp->seg_map_cnt - ocp->cur_seg);
441 	curcnt = (curcnt < io->sgl_allocated) ? curcnt : io->sgl_allocated;
442 	/* find current sg */
443 	for (cnt = 0, sg = cmd->t_data_sg; cnt < ocp->cur_seg; cnt++,
444 	     sg = sg_next(sg))
445 		;/* do nothing */
446 
447 	for (cnt = 0; cnt < curcnt; cnt++, sg = sg_next(sg)) {
448 		sgl[cnt].addr = sg_dma_address(sg);
449 		sgl[cnt].dif_addr = 0;
450 		sgl[cnt].len = sg_dma_len(sg);
451 		length += sgl[cnt].len;
452 		ocp->cur_seg++;
453 	}
454 
455 	if (ocp->cur_seg == ocp->seg_cnt)
456 		flags = EFCT_SCSI_LAST_DATAPHASE;
457 
458 	return efct_scsi_recv_wr_data(io, flags, sgl, curcnt, length,
459 				    efct_lio_datamove_done, NULL);
460 }
461 
462 static int
efct_lio_queue_data_in(struct se_cmd * cmd)463 efct_lio_queue_data_in(struct se_cmd *cmd)
464 {
465 	struct efct_scsi_tgt_io *ocp =
466 		container_of(cmd, struct efct_scsi_tgt_io, cmd);
467 	struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
468 	struct efct_scsi_sgl *sgl = io->sgl;
469 	struct scatterlist *sg = NULL;
470 	uint flags = 0, cnt = 0, curcnt = 0;
471 	u64 length = 0;
472 
473 	efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_QUEUE_DATA_IN);
474 
475 	if (ocp->seg_cnt == 0) {
476 		if (cmd->data_length) {
477 			ocp->seg_cnt = cmd->t_data_nents;
478 			ocp->cur_seg = 0;
479 			if (efct_lio_sg_map(io)) {
480 				efct_lio_io_printf(io,
481 						   "efct_lio_sg_map failed\n");
482 				return -EAGAIN;
483 			}
484 		} else {
485 			/* If command length is 0, send the response status */
486 			struct efct_scsi_cmd_resp rsp;
487 
488 			memset(&rsp, 0, sizeof(rsp));
489 			efct_lio_io_printf(io,
490 					   "cmd : %p length 0, send status\n",
491 					   cmd);
492 			return efct_scsi_send_resp(io, 0, &rsp,
493 						   efct_lio_status_done, NULL);
494 		}
495 	}
496 	curcnt = min(ocp->seg_map_cnt - ocp->cur_seg, io->sgl_allocated);
497 
498 	while (cnt < curcnt) {
499 		sg = &cmd->t_data_sg[ocp->cur_seg];
500 		sgl[cnt].addr = sg_dma_address(sg);
501 		sgl[cnt].dif_addr = 0;
502 		if (ocp->transferred_len + sg_dma_len(sg) >= cmd->data_length)
503 			sgl[cnt].len = cmd->data_length - ocp->transferred_len;
504 		else
505 			sgl[cnt].len = sg_dma_len(sg);
506 
507 		ocp->transferred_len += sgl[cnt].len;
508 		length += sgl[cnt].len;
509 		ocp->cur_seg++;
510 		cnt++;
511 		if (ocp->transferred_len == cmd->data_length)
512 			break;
513 	}
514 
515 	if (ocp->transferred_len == cmd->data_length) {
516 		flags = EFCT_SCSI_LAST_DATAPHASE;
517 		ocp->seg_cnt = ocp->cur_seg;
518 	}
519 
520 	/* If there is residual, disable Auto Good Response */
521 	if (cmd->residual_count)
522 		flags |= EFCT_SCSI_NO_AUTO_RESPONSE;
523 
524 	efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_SEND_RD_DATA);
525 
526 	return efct_scsi_send_rd_data(io, flags, sgl, curcnt, length,
527 				    efct_lio_datamove_done, NULL);
528 }
529 
530 static void
efct_lio_send_resp(struct efct_io * io,enum efct_scsi_io_status scsi_status,u32 flags)531 efct_lio_send_resp(struct efct_io *io, enum efct_scsi_io_status scsi_status,
532 		   u32 flags)
533 {
534 	struct efct_scsi_cmd_resp rsp;
535 	struct efct_scsi_tgt_io *ocp = &io->tgt_io;
536 	struct se_cmd *cmd = &io->tgt_io.cmd;
537 	int rc;
538 
539 	if (flags & EFCT_SCSI_IO_CMPL_RSP_SENT) {
540 		ocp->rsp_sent = true;
541 		efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_GENERIC_FREE);
542 		transport_generic_free_cmd(&io->tgt_io.cmd, 0);
543 		return;
544 	}
545 
546 	/* send check condition if an error occurred */
547 	memset(&rsp, 0, sizeof(rsp));
548 	rsp.scsi_status = cmd->scsi_status;
549 	rsp.sense_data = (uint8_t *)io->tgt_io.sense_buffer;
550 	rsp.sense_data_length = cmd->scsi_sense_length;
551 
552 	/* Check for residual underrun or overrun */
553 	if (cmd->se_cmd_flags & SCF_OVERFLOW_BIT)
554 		rsp.residual = -cmd->residual_count;
555 	else if (cmd->se_cmd_flags & SCF_UNDERFLOW_BIT)
556 		rsp.residual = cmd->residual_count;
557 
558 	rc = efct_scsi_send_resp(io, 0, &rsp, efct_lio_status_done, NULL);
559 	efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_SEND_RSP);
560 	if (rc != 0) {
561 		efct_lio_io_printf(io, "Read done, send rsp failed %d\n", rc);
562 		efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_GENERIC_FREE);
563 		transport_generic_free_cmd(&io->tgt_io.cmd, 0);
564 	} else {
565 		ocp->rsp_sent = true;
566 	}
567 }
568 
569 static int
efct_lio_datamove_done(struct efct_io * io,enum efct_scsi_io_status scsi_status,u32 flags,void * arg)570 efct_lio_datamove_done(struct efct_io *io, enum efct_scsi_io_status scsi_status,
571 		       u32 flags, void *arg)
572 {
573 	struct efct_scsi_tgt_io *ocp = &io->tgt_io;
574 
575 	efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_DATA_DONE);
576 	if (scsi_status != EFCT_SCSI_STATUS_GOOD) {
577 		efct_lio_io_printf(io, "callback completed with error=%d\n",
578 				   scsi_status);
579 		ocp->err = scsi_status;
580 	}
581 	efct_lio_io_printf(io, "seg_map_cnt=%d\n", ocp->seg_map_cnt);
582 	if (ocp->seg_map_cnt) {
583 		if (ocp->err == EFCT_SCSI_STATUS_GOOD &&
584 		    ocp->cur_seg < ocp->seg_cnt) {
585 			int rc;
586 
587 			efct_lio_io_printf(io, "continuing cmd at segm=%d\n",
588 					   ocp->cur_seg);
589 			if (ocp->ddir == DMA_TO_DEVICE)
590 				rc = efct_lio_write_pending(&ocp->cmd);
591 			else
592 				rc = efct_lio_queue_data_in(&ocp->cmd);
593 			if (!rc)
594 				return 0;
595 
596 			ocp->err = EFCT_SCSI_STATUS_ERROR;
597 			efct_lio_io_printf(io, "could not continue command\n");
598 		}
599 		efct_lio_sg_unmap(io);
600 	}
601 
602 	if (io->tgt_io.aborting) {
603 		efct_lio_io_printf(io, "IO done aborted\n");
604 		return 0;
605 	}
606 
607 	if (ocp->ddir == DMA_TO_DEVICE) {
608 		efct_lio_io_printf(io, "Write done, trans_state=0x%x\n",
609 				   io->tgt_io.cmd.transport_state);
610 		if (scsi_status != EFCT_SCSI_STATUS_GOOD) {
611 			transport_generic_request_failure(&io->tgt_io.cmd,
612 					TCM_CHECK_CONDITION_ABORT_CMD);
613 			efct_set_lio_io_state(io,
614 				EFCT_LIO_STATE_TGT_GENERIC_REQ_FAILURE);
615 		} else {
616 			efct_set_lio_io_state(io,
617 						EFCT_LIO_STATE_TGT_EXECUTE_CMD);
618 			target_execute_cmd(&io->tgt_io.cmd);
619 		}
620 	} else {
621 		efct_lio_send_resp(io, scsi_status, flags);
622 	}
623 	return 0;
624 }
625 
626 static int
efct_lio_tmf_done(struct efct_io * io,enum efct_scsi_io_status scsi_status,u32 flags,void * arg)627 efct_lio_tmf_done(struct efct_io *io, enum efct_scsi_io_status scsi_status,
628 		  u32 flags, void *arg)
629 {
630 	efct_lio_tmfio_printf(io, "cmd=%p status=%d, flags=0x%x\n",
631 			      &io->tgt_io.cmd, scsi_status, flags);
632 
633 	efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_GENERIC_FREE);
634 	transport_generic_free_cmd(&io->tgt_io.cmd, 0);
635 	return 0;
636 }
637 
638 static int
efct_lio_null_tmf_done(struct efct_io * tmfio,enum efct_scsi_io_status scsi_status,u32 flags,void * arg)639 efct_lio_null_tmf_done(struct efct_io *tmfio,
640 		       enum efct_scsi_io_status scsi_status,
641 		      u32 flags, void *arg)
642 {
643 	efct_lio_tmfio_printf(tmfio, "cmd=%p status=%d, flags=0x%x\n",
644 			      &tmfio->tgt_io.cmd, scsi_status, flags);
645 
646 	/* free struct efct_io only, no active se_cmd */
647 	efct_scsi_io_complete(tmfio);
648 	return 0;
649 }
650 
651 static int
efct_lio_queue_status(struct se_cmd * cmd)652 efct_lio_queue_status(struct se_cmd *cmd)
653 {
654 	struct efct_scsi_cmd_resp rsp;
655 	struct efct_scsi_tgt_io *ocp =
656 		container_of(cmd, struct efct_scsi_tgt_io, cmd);
657 	struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
658 	int rc = 0;
659 
660 	efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_QUEUE_STATUS);
661 	efct_lio_io_printf(io,
662 		"status=0x%x trans_state=0x%x se_cmd_flags=0x%x sns_len=%d\n",
663 		cmd->scsi_status, cmd->transport_state, cmd->se_cmd_flags,
664 		cmd->scsi_sense_length);
665 
666 	memset(&rsp, 0, sizeof(rsp));
667 	rsp.scsi_status = cmd->scsi_status;
668 	rsp.sense_data = (u8 *)io->tgt_io.sense_buffer;
669 	rsp.sense_data_length = cmd->scsi_sense_length;
670 
671 	/* Check for residual underrun or overrun, mark negitive value for
672 	 * underrun to recognize in HW
673 	 */
674 	if (cmd->se_cmd_flags & SCF_OVERFLOW_BIT)
675 		rsp.residual = -cmd->residual_count;
676 	else if (cmd->se_cmd_flags & SCF_UNDERFLOW_BIT)
677 		rsp.residual = cmd->residual_count;
678 
679 	rc = efct_scsi_send_resp(io, 0, &rsp, efct_lio_status_done, NULL);
680 	efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_SEND_RSP);
681 	if (rc == 0)
682 		ocp->rsp_sent = true;
683 	return rc;
684 }
685 
efct_lio_queue_tm_rsp(struct se_cmd * cmd)686 static void efct_lio_queue_tm_rsp(struct se_cmd *cmd)
687 {
688 	struct efct_scsi_tgt_io *ocp =
689 		container_of(cmd, struct efct_scsi_tgt_io, cmd);
690 	struct efct_io *tmfio = container_of(ocp, struct efct_io, tgt_io);
691 	struct se_tmr_req *se_tmr = cmd->se_tmr_req;
692 	u8 rspcode;
693 
694 	efct_lio_tmfio_printf(tmfio, "cmd=%p function=0x%x tmr->response=%d\n",
695 			      cmd, se_tmr->function, se_tmr->response);
696 	switch (se_tmr->response) {
697 	case TMR_FUNCTION_COMPLETE:
698 		rspcode = EFCT_SCSI_TMF_FUNCTION_COMPLETE;
699 		break;
700 	case TMR_TASK_DOES_NOT_EXIST:
701 		rspcode = EFCT_SCSI_TMF_FUNCTION_IO_NOT_FOUND;
702 		break;
703 	case TMR_LUN_DOES_NOT_EXIST:
704 		rspcode = EFCT_SCSI_TMF_INCORRECT_LOGICAL_UNIT_NUMBER;
705 		break;
706 	case TMR_FUNCTION_REJECTED:
707 	default:
708 		rspcode = EFCT_SCSI_TMF_FUNCTION_REJECTED;
709 		break;
710 	}
711 	efct_scsi_send_tmf_resp(tmfio, rspcode, NULL, efct_lio_tmf_done, NULL);
712 }
713 
efct_find_wwpn(u64 wwpn)714 static struct efct *efct_find_wwpn(u64 wwpn)
715 {
716 	struct efct *efct;
717 
718 	 /* Search for the HBA that has this WWPN */
719 	list_for_each_entry(efct, &efct_devices, list_entry) {
720 
721 		if (wwpn == efct_get_wwpn(&efct->hw))
722 			return efct;
723 	}
724 
725 	return NULL;
726 }
727 
728 static struct se_wwn *
efct_lio_make_nport(struct target_fabric_configfs * tf,struct config_group * group,const char * name)729 efct_lio_make_nport(struct target_fabric_configfs *tf,
730 		    struct config_group *group, const char *name)
731 {
732 	struct efct_lio_nport *lio_nport;
733 	struct efct *efct;
734 	int ret;
735 	u64 wwpn;
736 
737 	ret = efct_lio_parse_wwn(name, &wwpn, 0);
738 	if (ret)
739 		return ERR_PTR(ret);
740 
741 	efct = efct_find_wwpn(wwpn);
742 	if (!efct) {
743 		pr_err("cannot find EFCT for base wwpn %s\n", name);
744 		return ERR_PTR(-ENXIO);
745 	}
746 
747 	lio_nport = kzalloc(sizeof(*lio_nport), GFP_KERNEL);
748 	if (!lio_nport)
749 		return ERR_PTR(-ENOMEM);
750 
751 	lio_nport->efct = efct;
752 	lio_nport->wwpn = wwpn;
753 	efct_format_wwn(lio_nport->wwpn_str, sizeof(lio_nport->wwpn_str),
754 			"naa.", wwpn);
755 	efct->tgt_efct.lio_nport = lio_nport;
756 
757 	return &lio_nport->nport_wwn;
758 }
759 
760 static struct se_wwn *
efct_lio_npiv_make_nport(struct target_fabric_configfs * tf,struct config_group * group,const char * name)761 efct_lio_npiv_make_nport(struct target_fabric_configfs *tf,
762 			 struct config_group *group, const char *name)
763 {
764 	struct efct_lio_vport *lio_vport;
765 	struct efct *efct;
766 	int ret;
767 	u64 p_wwpn, npiv_wwpn, npiv_wwnn;
768 	char *p, *pbuf, tmp[128];
769 	struct efct_lio_vport_list_t *vport_list;
770 	struct fc_vport *new_fc_vport;
771 	struct fc_vport_identifiers vport_id;
772 	unsigned long flags = 0;
773 
774 	snprintf(tmp, sizeof(tmp), "%s", name);
775 	pbuf = &tmp[0];
776 
777 	p = strsep(&pbuf, "@");
778 
779 	if (!p || !pbuf) {
780 		pr_err("Unable to find separator operator(@)\n");
781 		return ERR_PTR(-EINVAL);
782 	}
783 
784 	ret = efct_lio_parse_wwn(p, &p_wwpn, 0);
785 	if (ret)
786 		return ERR_PTR(ret);
787 
788 	ret = efct_lio_parse_npiv_wwn(pbuf, strlen(pbuf), &npiv_wwpn,
789 				      &npiv_wwnn);
790 	if (ret)
791 		return ERR_PTR(ret);
792 
793 	efct = efct_find_wwpn(p_wwpn);
794 	if (!efct) {
795 		pr_err("cannot find EFCT for base wwpn %s\n", name);
796 		return ERR_PTR(-ENXIO);
797 	}
798 
799 	lio_vport = kzalloc(sizeof(*lio_vport), GFP_KERNEL);
800 	if (!lio_vport)
801 		return ERR_PTR(-ENOMEM);
802 
803 	lio_vport->efct = efct;
804 	lio_vport->wwpn = p_wwpn;
805 	lio_vport->npiv_wwpn = npiv_wwpn;
806 	lio_vport->npiv_wwnn = npiv_wwnn;
807 
808 	efct_format_wwn(lio_vport->wwpn_str, sizeof(lio_vport->wwpn_str),
809 			"naa.", npiv_wwpn);
810 
811 	vport_list = kzalloc(sizeof(*vport_list), GFP_KERNEL);
812 	if (!vport_list) {
813 		kfree(lio_vport);
814 		return ERR_PTR(-ENOMEM);
815 	}
816 
817 	vport_list->lio_vport = lio_vport;
818 
819 	memset(&vport_id, 0, sizeof(vport_id));
820 	vport_id.port_name = npiv_wwpn;
821 	vport_id.node_name = npiv_wwnn;
822 	vport_id.roles = FC_PORT_ROLE_FCP_INITIATOR;
823 	vport_id.vport_type = FC_PORTTYPE_NPIV;
824 	vport_id.disable = false;
825 
826 	new_fc_vport = fc_vport_create(efct->shost, 0, &vport_id);
827 	if (!new_fc_vport) {
828 		efc_log_err(efct, "fc_vport_create failed\n");
829 		kfree(lio_vport);
830 		kfree(vport_list);
831 		return ERR_PTR(-ENOMEM);
832 	}
833 
834 	lio_vport->fc_vport = new_fc_vport;
835 	spin_lock_irqsave(&efct->tgt_efct.efct_lio_lock, flags);
836 	INIT_LIST_HEAD(&vport_list->list_entry);
837 	list_add_tail(&vport_list->list_entry, &efct->tgt_efct.vport_list);
838 	spin_unlock_irqrestore(&efct->tgt_efct.efct_lio_lock, flags);
839 
840 	return &lio_vport->vport_wwn;
841 }
842 
843 static void
efct_lio_drop_nport(struct se_wwn * wwn)844 efct_lio_drop_nport(struct se_wwn *wwn)
845 {
846 	struct efct_lio_nport *lio_nport =
847 		container_of(wwn, struct efct_lio_nport, nport_wwn);
848 	struct efct *efct = lio_nport->efct;
849 
850 	/* only physical nport should exist, free lio_nport allocated
851 	 * in efct_lio_make_nport.
852 	 */
853 	kfree(efct->tgt_efct.lio_nport);
854 	efct->tgt_efct.lio_nport = NULL;
855 }
856 
857 static void
efct_lio_npiv_drop_nport(struct se_wwn * wwn)858 efct_lio_npiv_drop_nport(struct se_wwn *wwn)
859 {
860 	struct efct_lio_vport *lio_vport =
861 		container_of(wwn, struct efct_lio_vport, vport_wwn);
862 	struct efct_lio_vport_list_t *vport, *next_vport;
863 	struct efct *efct = lio_vport->efct;
864 	unsigned long flags = 0;
865 
866 	if (lio_vport->fc_vport)
867 		fc_vport_terminate(lio_vport->fc_vport);
868 
869 	spin_lock_irqsave(&efct->tgt_efct.efct_lio_lock, flags);
870 
871 	list_for_each_entry_safe(vport, next_vport, &efct->tgt_efct.vport_list,
872 				 list_entry) {
873 		if (vport->lio_vport == lio_vport) {
874 			list_del(&vport->list_entry);
875 			kfree(vport->lio_vport);
876 			kfree(vport);
877 			break;
878 		}
879 	}
880 	spin_unlock_irqrestore(&efct->tgt_efct.efct_lio_lock, flags);
881 }
882 
883 static struct se_portal_group *
efct_lio_make_tpg(struct se_wwn * wwn,const char * name)884 efct_lio_make_tpg(struct se_wwn *wwn, const char *name)
885 {
886 	struct efct_lio_nport *lio_nport =
887 		container_of(wwn, struct efct_lio_nport, nport_wwn);
888 	struct efct_lio_tpg *tpg;
889 	struct efct *efct;
890 	unsigned long n;
891 	int ret;
892 
893 	if (strstr(name, "tpgt_") != name)
894 		return ERR_PTR(-EINVAL);
895 	if (kstrtoul(name + 5, 10, &n) || n > USHRT_MAX)
896 		return ERR_PTR(-EINVAL);
897 
898 	tpg = kzalloc(sizeof(*tpg), GFP_KERNEL);
899 	if (!tpg)
900 		return ERR_PTR(-ENOMEM);
901 
902 	tpg->nport = lio_nport;
903 	tpg->tpgt = n;
904 	tpg->enabled = false;
905 
906 	tpg->tpg_attrib.generate_node_acls = 1;
907 	tpg->tpg_attrib.demo_mode_write_protect = 1;
908 	tpg->tpg_attrib.cache_dynamic_acls = 1;
909 	tpg->tpg_attrib.demo_mode_login_only = 1;
910 	tpg->tpg_attrib.session_deletion_wait = 1;
911 
912 	ret = core_tpg_register(wwn, &tpg->tpg, SCSI_PROTOCOL_FCP);
913 	if (ret < 0) {
914 		kfree(tpg);
915 		return NULL;
916 	}
917 	efct = lio_nport->efct;
918 	efct->tgt_efct.tpg = tpg;
919 	efc_log_debug(efct, "create portal group %d\n", tpg->tpgt);
920 
921 	xa_init(&efct->lookup);
922 	return &tpg->tpg;
923 }
924 
925 static void
efct_lio_drop_tpg(struct se_portal_group * se_tpg)926 efct_lio_drop_tpg(struct se_portal_group *se_tpg)
927 {
928 	struct efct_lio_tpg *tpg =
929 		container_of(se_tpg, struct efct_lio_tpg, tpg);
930 
931 	struct efct *efct = tpg->nport->efct;
932 
933 	efc_log_debug(efct, "drop portal group %d\n", tpg->tpgt);
934 	tpg->nport->efct->tgt_efct.tpg = NULL;
935 	core_tpg_deregister(se_tpg);
936 	xa_destroy(&efct->lookup);
937 	kfree(tpg);
938 }
939 
940 static struct se_portal_group *
efct_lio_npiv_make_tpg(struct se_wwn * wwn,const char * name)941 efct_lio_npiv_make_tpg(struct se_wwn *wwn, const char *name)
942 {
943 	struct efct_lio_vport *lio_vport =
944 		container_of(wwn, struct efct_lio_vport, vport_wwn);
945 	struct efct_lio_tpg *tpg;
946 	struct efct *efct;
947 	unsigned long n;
948 	int ret;
949 
950 	efct = lio_vport->efct;
951 	if (strstr(name, "tpgt_") != name)
952 		return ERR_PTR(-EINVAL);
953 	if (kstrtoul(name + 5, 10, &n) || n > USHRT_MAX)
954 		return ERR_PTR(-EINVAL);
955 
956 	if (n != 1) {
957 		efc_log_err(efct, "Invalid tpgt index: %ld provided\n", n);
958 		return ERR_PTR(-EINVAL);
959 	}
960 
961 	tpg = kzalloc(sizeof(*tpg), GFP_KERNEL);
962 	if (!tpg)
963 		return ERR_PTR(-ENOMEM);
964 
965 	tpg->vport = lio_vport;
966 	tpg->tpgt = n;
967 	tpg->enabled = false;
968 
969 	tpg->tpg_attrib.generate_node_acls = 1;
970 	tpg->tpg_attrib.demo_mode_write_protect = 1;
971 	tpg->tpg_attrib.cache_dynamic_acls = 1;
972 	tpg->tpg_attrib.demo_mode_login_only = 1;
973 	tpg->tpg_attrib.session_deletion_wait = 1;
974 
975 	ret = core_tpg_register(wwn, &tpg->tpg, SCSI_PROTOCOL_FCP);
976 
977 	if (ret < 0) {
978 		kfree(tpg);
979 		return NULL;
980 	}
981 	lio_vport->tpg = tpg;
982 	efc_log_debug(efct, "create vport portal group %d\n", tpg->tpgt);
983 
984 	return &tpg->tpg;
985 }
986 
987 static void
efct_lio_npiv_drop_tpg(struct se_portal_group * se_tpg)988 efct_lio_npiv_drop_tpg(struct se_portal_group *se_tpg)
989 {
990 	struct efct_lio_tpg *tpg =
991 		container_of(se_tpg, struct efct_lio_tpg, tpg);
992 
993 	efc_log_debug(tpg->vport->efct, "drop npiv portal group %d\n",
994 		       tpg->tpgt);
995 	core_tpg_deregister(se_tpg);
996 	kfree(tpg);
997 }
998 
999 static int
efct_lio_init_nodeacl(struct se_node_acl * se_nacl,const char * name)1000 efct_lio_init_nodeacl(struct se_node_acl *se_nacl, const char *name)
1001 {
1002 	struct efct_lio_nacl *nacl;
1003 	u64 wwnn;
1004 
1005 	if (efct_lio_parse_wwn(name, &wwnn, 0) < 0)
1006 		return -EINVAL;
1007 
1008 	nacl = container_of(se_nacl, struct efct_lio_nacl, se_node_acl);
1009 	nacl->nport_wwnn = wwnn;
1010 
1011 	efct_format_wwn(nacl->nport_name, sizeof(nacl->nport_name), "", wwnn);
1012 	return 0;
1013 }
1014 
efct_lio_check_demo_mode_login_only(struct se_portal_group * stpg)1015 static int efct_lio_check_demo_mode_login_only(struct se_portal_group *stpg)
1016 {
1017 	struct efct_lio_tpg *tpg = container_of(stpg, struct efct_lio_tpg, tpg);
1018 
1019 	return tpg->tpg_attrib.demo_mode_login_only;
1020 }
1021 
1022 static int
efct_lio_npiv_check_demo_mode_login_only(struct se_portal_group * stpg)1023 efct_lio_npiv_check_demo_mode_login_only(struct se_portal_group *stpg)
1024 {
1025 	struct efct_lio_tpg *tpg = container_of(stpg, struct efct_lio_tpg, tpg);
1026 
1027 	return tpg->tpg_attrib.demo_mode_login_only;
1028 }
1029 
1030 static struct efct_lio_tpg *
efct_get_vport_tpg(struct efc_node * node)1031 efct_get_vport_tpg(struct efc_node *node)
1032 {
1033 	struct efct *efct;
1034 	u64 wwpn = node->nport->wwpn;
1035 	struct efct_lio_vport_list_t *vport, *next;
1036 	struct efct_lio_vport *lio_vport = NULL;
1037 	struct efct_lio_tpg *tpg = NULL;
1038 	unsigned long flags = 0;
1039 
1040 	efct = node->efc->base;
1041 	spin_lock_irqsave(&efct->tgt_efct.efct_lio_lock, flags);
1042 	list_for_each_entry_safe(vport, next, &efct->tgt_efct.vport_list,
1043 				 list_entry) {
1044 		lio_vport = vport->lio_vport;
1045 		if (wwpn && lio_vport && lio_vport->npiv_wwpn == wwpn) {
1046 			efc_log_debug(efct, "found tpg on vport\n");
1047 			tpg = lio_vport->tpg;
1048 			break;
1049 		}
1050 	}
1051 	spin_unlock_irqrestore(&efct->tgt_efct.efct_lio_lock, flags);
1052 	return tpg;
1053 }
1054 
1055 static void
_efct_tgt_node_free(struct kref * arg)1056 _efct_tgt_node_free(struct kref *arg)
1057 {
1058 	struct efct_node *tgt_node = container_of(arg, struct efct_node, ref);
1059 	struct efc_node *node = tgt_node->node;
1060 
1061 	efc_scsi_del_initiator_complete(node->efc, node);
1062 	kfree(tgt_node);
1063 }
1064 
efct_session_cb(struct se_portal_group * se_tpg,struct se_session * se_sess,void * private)1065 static int efct_session_cb(struct se_portal_group *se_tpg,
1066 			   struct se_session *se_sess, void *private)
1067 {
1068 	struct efc_node *node = private;
1069 	struct efct_node *tgt_node;
1070 	struct efct *efct = node->efc->base;
1071 
1072 	tgt_node = kzalloc(sizeof(*tgt_node), GFP_KERNEL);
1073 	if (!tgt_node)
1074 		return -ENOMEM;
1075 
1076 	kref_init(&tgt_node->ref);
1077 	tgt_node->release = _efct_tgt_node_free;
1078 
1079 	tgt_node->session = se_sess;
1080 	node->tgt_node = tgt_node;
1081 	tgt_node->efct = efct;
1082 
1083 	tgt_node->node = node;
1084 
1085 	tgt_node->node_fc_id = node->rnode.fc_id;
1086 	tgt_node->port_fc_id = node->nport->fc_id;
1087 	tgt_node->vpi = node->nport->indicator;
1088 	tgt_node->rpi = node->rnode.indicator;
1089 
1090 	spin_lock_init(&tgt_node->active_ios_lock);
1091 	INIT_LIST_HEAD(&tgt_node->active_ios);
1092 
1093 	return 0;
1094 }
1095 
efct_scsi_tgt_new_device(struct efct * efct)1096 int efct_scsi_tgt_new_device(struct efct *efct)
1097 {
1098 	u32 total_ios;
1099 
1100 	/* Get the max settings */
1101 	efct->tgt_efct.max_sge = sli_get_max_sge(&efct->hw.sli);
1102 	efct->tgt_efct.max_sgl = sli_get_max_sgl(&efct->hw.sli);
1103 
1104 	/* initialize IO watermark fields */
1105 	atomic_set(&efct->tgt_efct.ios_in_use, 0);
1106 	total_ios = efct->hw.config.n_io;
1107 	efc_log_debug(efct, "total_ios=%d\n", total_ios);
1108 	efct->tgt_efct.watermark_min =
1109 			(total_ios * EFCT_WATERMARK_LOW_PCT) / 100;
1110 	efct->tgt_efct.watermark_max =
1111 			(total_ios * EFCT_WATERMARK_HIGH_PCT) / 100;
1112 	atomic_set(&efct->tgt_efct.io_high_watermark,
1113 		   efct->tgt_efct.watermark_max);
1114 	atomic_set(&efct->tgt_efct.watermark_hit, 0);
1115 	atomic_set(&efct->tgt_efct.initiator_count, 0);
1116 
1117 	lio_wq = alloc_ordered_workqueue("%s", WQ_MEM_RECLAIM,
1118 					 "efct_lio_worker");
1119 	if (!lio_wq) {
1120 		efc_log_err(efct, "workqueue create failed\n");
1121 		return -EIO;
1122 	}
1123 
1124 	spin_lock_init(&efct->tgt_efct.efct_lio_lock);
1125 	INIT_LIST_HEAD(&efct->tgt_efct.vport_list);
1126 
1127 	return 0;
1128 }
1129 
efct_scsi_tgt_del_device(struct efct * efct)1130 int efct_scsi_tgt_del_device(struct efct *efct)
1131 {
1132 	flush_workqueue(lio_wq);
1133 
1134 	return 0;
1135 }
1136 
1137 int
efct_scsi_tgt_new_nport(struct efc * efc,struct efc_nport * nport)1138 efct_scsi_tgt_new_nport(struct efc *efc, struct efc_nport *nport)
1139 {
1140 	struct efct *efct = nport->efc->base;
1141 
1142 	efc_log_debug(efct, "New SPORT: %s bound to %s\n", nport->display_name,
1143 		       efct->tgt_efct.lio_nport->wwpn_str);
1144 
1145 	return 0;
1146 }
1147 
1148 void
efct_scsi_tgt_del_nport(struct efc * efc,struct efc_nport * nport)1149 efct_scsi_tgt_del_nport(struct efc *efc, struct efc_nport *nport)
1150 {
1151 	efc_log_debug(efc, "Del SPORT: %s\n", nport->display_name);
1152 }
1153 
efct_lio_setup_session(struct work_struct * work)1154 static void efct_lio_setup_session(struct work_struct *work)
1155 {
1156 	struct efct_lio_wq_data *wq_data =
1157 		container_of(work, struct efct_lio_wq_data, work);
1158 	struct efct *efct = wq_data->efct;
1159 	struct efc_node *node = wq_data->ptr;
1160 	char wwpn[WWN_NAME_LEN];
1161 	struct efct_lio_tpg *tpg;
1162 	struct efct_node *tgt_node;
1163 	struct se_portal_group *se_tpg;
1164 	struct se_session *se_sess;
1165 	int watermark;
1166 	int ini_count;
1167 	u64 id;
1168 
1169 	/* Check to see if it's belongs to vport,
1170 	 * if not get physical port
1171 	 */
1172 	tpg = efct_get_vport_tpg(node);
1173 	if (tpg) {
1174 		se_tpg = &tpg->tpg;
1175 	} else if (efct->tgt_efct.tpg) {
1176 		tpg = efct->tgt_efct.tpg;
1177 		se_tpg = &tpg->tpg;
1178 	} else {
1179 		efc_log_err(efct, "failed to init session\n");
1180 		return;
1181 	}
1182 
1183 	/*
1184 	 * Format the FCP Initiator port_name into colon
1185 	 * separated values to match the format by our explicit
1186 	 * ConfigFS NodeACLs.
1187 	 */
1188 	efct_format_wwn(wwpn, sizeof(wwpn), "",	efc_node_get_wwpn(node));
1189 
1190 	se_sess = target_setup_session(se_tpg, 0, 0, TARGET_PROT_NORMAL, wwpn,
1191 				       node, efct_session_cb);
1192 	if (IS_ERR(se_sess)) {
1193 		efc_log_err(efct, "failed to setup session\n");
1194 		kfree(wq_data);
1195 		efc_scsi_sess_reg_complete(node, -EIO);
1196 		return;
1197 	}
1198 
1199 	tgt_node = node->tgt_node;
1200 	id = (u64) tgt_node->port_fc_id << 32 | tgt_node->node_fc_id;
1201 
1202 	efc_log_debug(efct, "new initiator sess=%p node=%p id: %llx\n",
1203 		      se_sess, node, id);
1204 
1205 	if (xa_err(xa_store(&efct->lookup, id, tgt_node, GFP_KERNEL)))
1206 		efc_log_err(efct, "Node lookup store failed\n");
1207 
1208 	efc_scsi_sess_reg_complete(node, 0);
1209 
1210 	/* update IO watermark: increment initiator count */
1211 	ini_count = atomic_add_return(1, &efct->tgt_efct.initiator_count);
1212 	watermark = efct->tgt_efct.watermark_max -
1213 		    ini_count * EFCT_IO_WATERMARK_PER_INITIATOR;
1214 	watermark = (efct->tgt_efct.watermark_min > watermark) ?
1215 			efct->tgt_efct.watermark_min : watermark;
1216 	atomic_set(&efct->tgt_efct.io_high_watermark, watermark);
1217 
1218 	kfree(wq_data);
1219 }
1220 
efct_scsi_new_initiator(struct efc * efc,struct efc_node * node)1221 int efct_scsi_new_initiator(struct efc *efc, struct efc_node *node)
1222 {
1223 	struct efct *efct = node->efc->base;
1224 	struct efct_lio_wq_data *wq_data;
1225 
1226 	/*
1227 	 * Since LIO only supports initiator validation at thread level,
1228 	 * we are open minded and accept all callers.
1229 	 */
1230 	wq_data = kzalloc(sizeof(*wq_data), GFP_ATOMIC);
1231 	if (!wq_data)
1232 		return -ENOMEM;
1233 
1234 	wq_data->ptr = node;
1235 	wq_data->efct = efct;
1236 	INIT_WORK(&wq_data->work, efct_lio_setup_session);
1237 	queue_work(lio_wq, &wq_data->work);
1238 	return EFC_SCSI_CALL_ASYNC;
1239 }
1240 
efct_lio_remove_session(struct work_struct * work)1241 static void efct_lio_remove_session(struct work_struct *work)
1242 {
1243 	struct efct_lio_wq_data *wq_data =
1244 		container_of(work, struct efct_lio_wq_data, work);
1245 	struct efct *efct = wq_data->efct;
1246 	struct efc_node *node = wq_data->ptr;
1247 	struct efct_node *tgt_node;
1248 	struct se_session *se_sess;
1249 
1250 	tgt_node = node->tgt_node;
1251 	if (!tgt_node) {
1252 		/* base driver has sent back-to-back requests
1253 		 * to unreg session with no intervening
1254 		 * register
1255 		 */
1256 		efc_log_err(efct, "unreg session for NULL session\n");
1257 		efc_scsi_del_initiator_complete(node->efc, node);
1258 		return;
1259 	}
1260 
1261 	se_sess = tgt_node->session;
1262 	efc_log_debug(efct, "unreg session se_sess=%p node=%p\n",
1263 		       se_sess, node);
1264 
1265 	/* first flag all session commands to complete */
1266 	target_stop_session(se_sess);
1267 
1268 	/* now wait for session commands to complete */
1269 	target_wait_for_sess_cmds(se_sess);
1270 	target_remove_session(se_sess);
1271 	tgt_node->session = NULL;
1272 	node->tgt_node = NULL;
1273 	kref_put(&tgt_node->ref, tgt_node->release);
1274 
1275 	kfree(wq_data);
1276 }
1277 
efct_scsi_del_initiator(struct efc * efc,struct efc_node * node,int reason)1278 int efct_scsi_del_initiator(struct efc *efc, struct efc_node *node, int reason)
1279 {
1280 	struct efct *efct = node->efc->base;
1281 	struct efct_node *tgt_node = node->tgt_node;
1282 	struct efct_lio_wq_data *wq_data;
1283 	int watermark;
1284 	int ini_count;
1285 	u64 id;
1286 
1287 	if (reason == EFCT_SCSI_INITIATOR_MISSING)
1288 		return EFC_SCSI_CALL_COMPLETE;
1289 
1290 	if (!tgt_node) {
1291 		efc_log_err(efct, "tgt_node is NULL\n");
1292 		return -EIO;
1293 	}
1294 
1295 	wq_data = kzalloc(sizeof(*wq_data), GFP_ATOMIC);
1296 	if (!wq_data)
1297 		return -ENOMEM;
1298 
1299 	id = (u64) tgt_node->port_fc_id << 32 | tgt_node->node_fc_id;
1300 	xa_erase(&efct->lookup, id);
1301 
1302 	wq_data->ptr = node;
1303 	wq_data->efct = efct;
1304 	INIT_WORK(&wq_data->work, efct_lio_remove_session);
1305 	queue_work(lio_wq, &wq_data->work);
1306 
1307 	/*
1308 	 * update IO watermark: decrement initiator count
1309 	 */
1310 	ini_count = atomic_sub_return(1, &efct->tgt_efct.initiator_count);
1311 
1312 	watermark = efct->tgt_efct.watermark_max -
1313 		    ini_count * EFCT_IO_WATERMARK_PER_INITIATOR;
1314 	watermark = (efct->tgt_efct.watermark_min > watermark) ?
1315 			efct->tgt_efct.watermark_min : watermark;
1316 	atomic_set(&efct->tgt_efct.io_high_watermark, watermark);
1317 
1318 	return EFC_SCSI_CALL_ASYNC;
1319 }
1320 
efct_scsi_recv_cmd(struct efct_io * io,uint64_t lun,u8 * cdb,u32 cdb_len,u32 flags)1321 void efct_scsi_recv_cmd(struct efct_io *io, uint64_t lun, u8 *cdb,
1322 		       u32 cdb_len, u32 flags)
1323 {
1324 	struct efct_scsi_tgt_io *ocp = &io->tgt_io;
1325 	struct se_cmd *se_cmd = &io->tgt_io.cmd;
1326 	struct efct *efct = io->efct;
1327 	char *ddir;
1328 	struct efct_node *tgt_node;
1329 	struct se_session *se_sess;
1330 	int rc = 0;
1331 
1332 	memset(ocp, 0, sizeof(struct efct_scsi_tgt_io));
1333 	efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_RECV_CMD);
1334 	atomic_add_return(1, &efct->tgt_efct.ios_in_use);
1335 
1336 	/* set target timeout */
1337 	io->timeout = efct->target_io_timer_sec;
1338 
1339 	if (flags & EFCT_SCSI_CMD_SIMPLE)
1340 		ocp->task_attr = TCM_SIMPLE_TAG;
1341 	else if (flags & EFCT_SCSI_CMD_HEAD_OF_QUEUE)
1342 		ocp->task_attr = TCM_HEAD_TAG;
1343 	else if (flags & EFCT_SCSI_CMD_ORDERED)
1344 		ocp->task_attr = TCM_ORDERED_TAG;
1345 	else if (flags & EFCT_SCSI_CMD_ACA)
1346 		ocp->task_attr = TCM_ACA_TAG;
1347 
1348 	switch (flags & (EFCT_SCSI_CMD_DIR_IN | EFCT_SCSI_CMD_DIR_OUT)) {
1349 	case EFCT_SCSI_CMD_DIR_IN:
1350 		ddir = "FROM_INITIATOR";
1351 		ocp->ddir = DMA_TO_DEVICE;
1352 		break;
1353 	case EFCT_SCSI_CMD_DIR_OUT:
1354 		ddir = "TO_INITIATOR";
1355 		ocp->ddir = DMA_FROM_DEVICE;
1356 		break;
1357 	case EFCT_SCSI_CMD_DIR_IN | EFCT_SCSI_CMD_DIR_OUT:
1358 		ddir = "BIDIR";
1359 		ocp->ddir = DMA_BIDIRECTIONAL;
1360 		break;
1361 	default:
1362 		ddir = "NONE";
1363 		ocp->ddir = DMA_NONE;
1364 		break;
1365 	}
1366 
1367 	ocp->lun = lun;
1368 	efct_lio_io_printf(io, "new cmd=0x%x ddir=%s dl=%u\n",
1369 			   cdb[0], ddir, io->exp_xfer_len);
1370 
1371 	tgt_node = io->node;
1372 	se_sess = tgt_node->session;
1373 	if (!se_sess) {
1374 		efc_log_err(efct, "No session found to submit IO se_cmd: %p\n",
1375 			    &ocp->cmd);
1376 		efct_scsi_io_free(io);
1377 		return;
1378 	}
1379 
1380 	efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_SUBMIT_CMD);
1381 	rc = target_init_cmd(se_cmd, se_sess, &io->tgt_io.sense_buffer[0],
1382 			     ocp->lun, io->exp_xfer_len, ocp->task_attr,
1383 			     ocp->ddir, TARGET_SCF_ACK_KREF);
1384 	if (rc) {
1385 		efc_log_err(efct, "failed to init cmd se_cmd: %p\n", se_cmd);
1386 		efct_scsi_io_free(io);
1387 		return;
1388 	}
1389 
1390 	if (target_submit_prep(se_cmd, cdb, NULL, 0, NULL, 0,
1391 				NULL, 0, GFP_ATOMIC))
1392 		return;
1393 
1394 	target_submit(se_cmd);
1395 }
1396 
1397 int
efct_scsi_recv_tmf(struct efct_io * tmfio,u32 lun,enum efct_scsi_tmf_cmd cmd,struct efct_io * io_to_abort,u32 flags)1398 efct_scsi_recv_tmf(struct efct_io *tmfio, u32 lun, enum efct_scsi_tmf_cmd cmd,
1399 		   struct efct_io *io_to_abort, u32 flags)
1400 {
1401 	unsigned char tmr_func;
1402 	struct efct *efct = tmfio->efct;
1403 	struct efct_scsi_tgt_io *ocp = &tmfio->tgt_io;
1404 	struct efct_node *tgt_node;
1405 	struct se_session *se_sess;
1406 	int rc;
1407 
1408 	memset(ocp, 0, sizeof(struct efct_scsi_tgt_io));
1409 	efct_set_lio_io_state(tmfio, EFCT_LIO_STATE_SCSI_RECV_TMF);
1410 	atomic_add_return(1, &efct->tgt_efct.ios_in_use);
1411 	efct_lio_tmfio_printf(tmfio, "%s: new tmf %x lun=%u\n",
1412 			      tmfio->display_name, cmd, lun);
1413 
1414 	switch (cmd) {
1415 	case EFCT_SCSI_TMF_ABORT_TASK:
1416 		tmr_func = TMR_ABORT_TASK;
1417 		break;
1418 	case EFCT_SCSI_TMF_ABORT_TASK_SET:
1419 		tmr_func = TMR_ABORT_TASK_SET;
1420 		break;
1421 	case EFCT_SCSI_TMF_CLEAR_TASK_SET:
1422 		tmr_func = TMR_CLEAR_TASK_SET;
1423 		break;
1424 	case EFCT_SCSI_TMF_LOGICAL_UNIT_RESET:
1425 		tmr_func = TMR_LUN_RESET;
1426 		break;
1427 	case EFCT_SCSI_TMF_CLEAR_ACA:
1428 		tmr_func = TMR_CLEAR_ACA;
1429 		break;
1430 	case EFCT_SCSI_TMF_TARGET_RESET:
1431 		tmr_func = TMR_TARGET_WARM_RESET;
1432 		break;
1433 	case EFCT_SCSI_TMF_QUERY_ASYNCHRONOUS_EVENT:
1434 	case EFCT_SCSI_TMF_QUERY_TASK_SET:
1435 	default:
1436 		goto tmf_fail;
1437 	}
1438 
1439 	tmfio->tgt_io.tmf = tmr_func;
1440 	tmfio->tgt_io.lun = lun;
1441 	tmfio->tgt_io.io_to_abort = io_to_abort;
1442 
1443 	tgt_node = tmfio->node;
1444 
1445 	se_sess = tgt_node->session;
1446 	if (!se_sess)
1447 		return 0;
1448 
1449 	rc = target_submit_tmr(&ocp->cmd, se_sess, NULL, lun, ocp, tmr_func,
1450 			GFP_ATOMIC, tmfio->init_task_tag, TARGET_SCF_ACK_KREF);
1451 
1452 	efct_set_lio_io_state(tmfio, EFCT_LIO_STATE_TGT_SUBMIT_TMR);
1453 	if (rc)
1454 		goto tmf_fail;
1455 
1456 	return 0;
1457 
1458 tmf_fail:
1459 	efct_scsi_send_tmf_resp(tmfio, EFCT_SCSI_TMF_FUNCTION_REJECTED,
1460 				NULL, efct_lio_null_tmf_done, NULL);
1461 	return 0;
1462 }
1463 
1464 /* Start items for efct_lio_tpg_attrib_cit */
1465 
1466 #define DEF_EFCT_TPG_ATTRIB(name)					  \
1467 									  \
1468 static ssize_t efct_lio_tpg_attrib_##name##_show(			  \
1469 		struct config_item *item, char *page)			  \
1470 {									  \
1471 	struct se_portal_group *se_tpg = to_tpg(item);			  \
1472 	struct efct_lio_tpg *tpg = container_of(se_tpg,			  \
1473 			struct efct_lio_tpg, tpg);			  \
1474 									  \
1475 	return sprintf(page, "%u\n", tpg->tpg_attrib.name);		  \
1476 }									  \
1477 									  \
1478 static ssize_t efct_lio_tpg_attrib_##name##_store(			  \
1479 		struct config_item *item, const char *page, size_t count) \
1480 {									  \
1481 	struct se_portal_group *se_tpg = to_tpg(item);			  \
1482 	struct efct_lio_tpg *tpg = container_of(se_tpg,			  \
1483 					struct efct_lio_tpg, tpg);	  \
1484 	struct efct_lio_tpg_attrib *a = &tpg->tpg_attrib;		  \
1485 	unsigned long val;						  \
1486 	int ret;							  \
1487 									  \
1488 	ret = kstrtoul(page, 0, &val);					  \
1489 	if (ret < 0) {							  \
1490 		pr_err("kstrtoul() failed with ret: %d\n", ret);	  \
1491 		return ret;						  \
1492 	}								  \
1493 									  \
1494 	if (val != 0 && val != 1) {					  \
1495 		pr_err("Illegal boolean value %lu\n", val);		  \
1496 		return -EINVAL;						  \
1497 	}								  \
1498 									  \
1499 	a->name = val;							  \
1500 									  \
1501 	return count;							  \
1502 }									  \
1503 CONFIGFS_ATTR(efct_lio_tpg_attrib_, name)
1504 
1505 DEF_EFCT_TPG_ATTRIB(generate_node_acls);
1506 DEF_EFCT_TPG_ATTRIB(cache_dynamic_acls);
1507 DEF_EFCT_TPG_ATTRIB(demo_mode_write_protect);
1508 DEF_EFCT_TPG_ATTRIB(prod_mode_write_protect);
1509 DEF_EFCT_TPG_ATTRIB(demo_mode_login_only);
1510 DEF_EFCT_TPG_ATTRIB(session_deletion_wait);
1511 
1512 static struct configfs_attribute *efct_lio_tpg_attrib_attrs[] = {
1513 	&efct_lio_tpg_attrib_attr_generate_node_acls,
1514 	&efct_lio_tpg_attrib_attr_cache_dynamic_acls,
1515 	&efct_lio_tpg_attrib_attr_demo_mode_write_protect,
1516 	&efct_lio_tpg_attrib_attr_prod_mode_write_protect,
1517 	&efct_lio_tpg_attrib_attr_demo_mode_login_only,
1518 	&efct_lio_tpg_attrib_attr_session_deletion_wait,
1519 	NULL,
1520 };
1521 
1522 #define DEF_EFCT_NPIV_TPG_ATTRIB(name)					   \
1523 									   \
1524 static ssize_t efct_lio_npiv_tpg_attrib_##name##_show(			   \
1525 		struct config_item *item, char *page)			   \
1526 {									   \
1527 	struct se_portal_group *se_tpg = to_tpg(item);			   \
1528 	struct efct_lio_tpg *tpg = container_of(se_tpg,			   \
1529 			struct efct_lio_tpg, tpg);			   \
1530 									   \
1531 	return sprintf(page, "%u\n", tpg->tpg_attrib.name);		   \
1532 }									   \
1533 									   \
1534 static ssize_t efct_lio_npiv_tpg_attrib_##name##_store(			   \
1535 		struct config_item *item, const char *page, size_t count)  \
1536 {									   \
1537 	struct se_portal_group *se_tpg = to_tpg(item);			   \
1538 	struct efct_lio_tpg *tpg = container_of(se_tpg,			   \
1539 			struct efct_lio_tpg, tpg);			   \
1540 	struct efct_lio_tpg_attrib *a = &tpg->tpg_attrib;		   \
1541 	unsigned long val;						   \
1542 	int ret;							   \
1543 									   \
1544 	ret = kstrtoul(page, 0, &val);					   \
1545 	if (ret < 0) {							   \
1546 		pr_err("kstrtoul() failed with ret: %d\n", ret);	   \
1547 		return ret;						   \
1548 	}								   \
1549 									   \
1550 	if (val != 0 && val != 1) {					   \
1551 		pr_err("Illegal boolean value %lu\n", val);		   \
1552 		return -EINVAL;						   \
1553 	}								   \
1554 									   \
1555 	a->name = val;							   \
1556 									   \
1557 	return count;							   \
1558 }									   \
1559 CONFIGFS_ATTR(efct_lio_npiv_tpg_attrib_, name)
1560 
1561 DEF_EFCT_NPIV_TPG_ATTRIB(generate_node_acls);
1562 DEF_EFCT_NPIV_TPG_ATTRIB(cache_dynamic_acls);
1563 DEF_EFCT_NPIV_TPG_ATTRIB(demo_mode_write_protect);
1564 DEF_EFCT_NPIV_TPG_ATTRIB(prod_mode_write_protect);
1565 DEF_EFCT_NPIV_TPG_ATTRIB(demo_mode_login_only);
1566 DEF_EFCT_NPIV_TPG_ATTRIB(session_deletion_wait);
1567 
1568 static struct configfs_attribute *efct_lio_npiv_tpg_attrib_attrs[] = {
1569 	&efct_lio_npiv_tpg_attrib_attr_generate_node_acls,
1570 	&efct_lio_npiv_tpg_attrib_attr_cache_dynamic_acls,
1571 	&efct_lio_npiv_tpg_attrib_attr_demo_mode_write_protect,
1572 	&efct_lio_npiv_tpg_attrib_attr_prod_mode_write_protect,
1573 	&efct_lio_npiv_tpg_attrib_attr_demo_mode_login_only,
1574 	&efct_lio_npiv_tpg_attrib_attr_session_deletion_wait,
1575 	NULL,
1576 };
1577 
1578 CONFIGFS_ATTR(efct_lio_tpg_, enable);
1579 static struct configfs_attribute *efct_lio_tpg_attrs[] = {
1580 				&efct_lio_tpg_attr_enable, NULL };
1581 CONFIGFS_ATTR(efct_lio_npiv_tpg_, enable);
1582 static struct configfs_attribute *efct_lio_npiv_tpg_attrs[] = {
1583 				&efct_lio_npiv_tpg_attr_enable, NULL };
1584 
1585 static const struct target_core_fabric_ops efct_lio_ops = {
1586 	.module				= THIS_MODULE,
1587 	.fabric_name			= "efct",
1588 	.node_acl_size			= sizeof(struct efct_lio_nacl),
1589 	.max_data_sg_nents		= 65535,
1590 	.tpg_get_wwn			= efct_lio_get_fabric_wwn,
1591 	.tpg_get_tag			= efct_lio_get_tag,
1592 	.fabric_init_nodeacl		= efct_lio_init_nodeacl,
1593 	.tpg_check_demo_mode		= efct_lio_check_demo_mode,
1594 	.tpg_check_demo_mode_cache      = efct_lio_check_demo_mode_cache,
1595 	.tpg_check_demo_mode_write_protect = efct_lio_check_demo_write_protect,
1596 	.tpg_check_prod_mode_write_protect = efct_lio_check_prod_write_protect,
1597 	.check_stop_free		= efct_lio_check_stop_free,
1598 	.aborted_task			= efct_lio_aborted_task,
1599 	.release_cmd			= efct_lio_release_cmd,
1600 	.close_session			= efct_lio_close_session,
1601 	.write_pending			= efct_lio_write_pending,
1602 	.get_cmd_state			= efct_lio_get_cmd_state,
1603 	.queue_data_in			= efct_lio_queue_data_in,
1604 	.queue_status			= efct_lio_queue_status,
1605 	.queue_tm_rsp			= efct_lio_queue_tm_rsp,
1606 	.fabric_make_wwn		= efct_lio_make_nport,
1607 	.fabric_drop_wwn		= efct_lio_drop_nport,
1608 	.fabric_make_tpg		= efct_lio_make_tpg,
1609 	.fabric_drop_tpg		= efct_lio_drop_tpg,
1610 	.tpg_check_demo_mode_login_only = efct_lio_check_demo_mode_login_only,
1611 	.tpg_check_prot_fabric_only	= NULL,
1612 	.sess_get_initiator_sid		= NULL,
1613 	.tfc_tpg_base_attrs		= efct_lio_tpg_attrs,
1614 	.tfc_tpg_attrib_attrs           = efct_lio_tpg_attrib_attrs,
1615 	.default_submit_type		= TARGET_DIRECT_SUBMIT,
1616 	.direct_submit_supp		= 1,
1617 };
1618 
1619 static const struct target_core_fabric_ops efct_lio_npiv_ops = {
1620 	.module				= THIS_MODULE,
1621 	.fabric_name			= "efct_npiv",
1622 	.node_acl_size			= sizeof(struct efct_lio_nacl),
1623 	.max_data_sg_nents		= 65535,
1624 	.tpg_get_wwn			= efct_lio_get_npiv_fabric_wwn,
1625 	.tpg_get_tag			= efct_lio_get_npiv_tag,
1626 	.fabric_init_nodeacl		= efct_lio_init_nodeacl,
1627 	.tpg_check_demo_mode		= efct_lio_check_demo_mode,
1628 	.tpg_check_demo_mode_cache      = efct_lio_check_demo_mode_cache,
1629 	.tpg_check_demo_mode_write_protect =
1630 					efct_lio_npiv_check_demo_write_protect,
1631 	.tpg_check_prod_mode_write_protect =
1632 					efct_lio_npiv_check_prod_write_protect,
1633 	.check_stop_free		= efct_lio_check_stop_free,
1634 	.aborted_task			= efct_lio_aborted_task,
1635 	.release_cmd			= efct_lio_release_cmd,
1636 	.close_session			= efct_lio_close_session,
1637 	.write_pending			= efct_lio_write_pending,
1638 	.get_cmd_state			= efct_lio_get_cmd_state,
1639 	.queue_data_in			= efct_lio_queue_data_in,
1640 	.queue_status			= efct_lio_queue_status,
1641 	.queue_tm_rsp			= efct_lio_queue_tm_rsp,
1642 	.fabric_make_wwn		= efct_lio_npiv_make_nport,
1643 	.fabric_drop_wwn		= efct_lio_npiv_drop_nport,
1644 	.fabric_make_tpg		= efct_lio_npiv_make_tpg,
1645 	.fabric_drop_tpg		= efct_lio_npiv_drop_tpg,
1646 	.tpg_check_demo_mode_login_only =
1647 				efct_lio_npiv_check_demo_mode_login_only,
1648 	.tpg_check_prot_fabric_only	= NULL,
1649 	.sess_get_initiator_sid		= NULL,
1650 	.tfc_tpg_base_attrs		= efct_lio_npiv_tpg_attrs,
1651 	.tfc_tpg_attrib_attrs		= efct_lio_npiv_tpg_attrib_attrs,
1652 
1653 	.default_submit_type		= TARGET_DIRECT_SUBMIT,
1654 	.direct_submit_supp		= 1,
1655 };
1656 
efct_scsi_tgt_driver_init(void)1657 int efct_scsi_tgt_driver_init(void)
1658 {
1659 	int rc;
1660 
1661 	/* Register the top level struct config_item_type with TCM core */
1662 	rc = target_register_template(&efct_lio_ops);
1663 	if (rc < 0) {
1664 		pr_err("target_fabric_configfs_register failed with %d\n", rc);
1665 		return rc;
1666 	}
1667 	rc = target_register_template(&efct_lio_npiv_ops);
1668 	if (rc < 0) {
1669 		pr_err("target_fabric_configfs_register failed with %d\n", rc);
1670 		target_unregister_template(&efct_lio_ops);
1671 		return rc;
1672 	}
1673 	return 0;
1674 }
1675 
efct_scsi_tgt_driver_exit(void)1676 int efct_scsi_tgt_driver_exit(void)
1677 {
1678 	target_unregister_template(&efct_lio_ops);
1679 	target_unregister_template(&efct_lio_npiv_ops);
1680 	return 0;
1681 }
1682