1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * Copyright (C) 2021 Broadcom. All Rights Reserved. The term
4 * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries.
5 */
6
7 #include <target/target_core_base.h>
8 #include <target/target_core_fabric.h>
9 #include "efct_driver.h"
10 #include "efct_lio.h"
11
12 /*
13 * lio_wq is used to call the LIO backed during creation or deletion of
14 * sessions. This brings serialization to the session management as we create
15 * single threaded work queue.
16 */
17 static struct workqueue_struct *lio_wq;
18
19 static int
efct_format_wwn(char * str,size_t len,const char * pre,u64 wwn)20 efct_format_wwn(char *str, size_t len, const char *pre, u64 wwn)
21 {
22 u8 a[8];
23
24 put_unaligned_be64(wwn, a);
25 return snprintf(str, len, "%s%8phC", pre, a);
26 }
27
28 static int
efct_lio_parse_wwn(const char * name,u64 * wwp,u8 npiv)29 efct_lio_parse_wwn(const char *name, u64 *wwp, u8 npiv)
30 {
31 int num;
32 u8 b[8];
33
34 if (npiv) {
35 num = sscanf(name,
36 "%02hhx%02hhx%02hhx%02hhx%02hhx%02hhx%02hhx%02hhx",
37 &b[0], &b[1], &b[2], &b[3], &b[4], &b[5], &b[6],
38 &b[7]);
39 } else {
40 num = sscanf(name,
41 "%02hhx:%02hhx:%02hhx:%02hhx:%02hhx:%02hhx:%02hhx:%02hhx",
42 &b[0], &b[1], &b[2], &b[3], &b[4], &b[5], &b[6],
43 &b[7]);
44 }
45
46 if (num != 8)
47 return -EINVAL;
48
49 *wwp = get_unaligned_be64(b);
50 return 0;
51 }
52
53 static int
efct_lio_parse_npiv_wwn(const char * name,size_t size,u64 * wwpn,u64 * wwnn)54 efct_lio_parse_npiv_wwn(const char *name, size_t size, u64 *wwpn, u64 *wwnn)
55 {
56 unsigned int cnt = size;
57 int rc;
58
59 *wwpn = *wwnn = 0;
60 if (name[cnt - 1] == '\n' || name[cnt - 1] == 0)
61 cnt--;
62
63 /* validate we have enough characters for WWPN */
64 if ((cnt != (16 + 1 + 16)) || (name[16] != ':'))
65 return -EINVAL;
66
67 rc = efct_lio_parse_wwn(&name[0], wwpn, 1);
68 if (rc)
69 return rc;
70
71 rc = efct_lio_parse_wwn(&name[17], wwnn, 1);
72 if (rc)
73 return rc;
74
75 return 0;
76 }
77
78 static ssize_t
efct_lio_tpg_enable_show(struct config_item * item,char * page)79 efct_lio_tpg_enable_show(struct config_item *item, char *page)
80 {
81 struct se_portal_group *se_tpg = to_tpg(item);
82 struct efct_lio_tpg *tpg =
83 container_of(se_tpg, struct efct_lio_tpg, tpg);
84
85 return snprintf(page, PAGE_SIZE, "%d\n", tpg->enabled);
86 }
87
88 static ssize_t
efct_lio_tpg_enable_store(struct config_item * item,const char * page,size_t count)89 efct_lio_tpg_enable_store(struct config_item *item, const char *page,
90 size_t count)
91 {
92 struct se_portal_group *se_tpg = to_tpg(item);
93 struct efct_lio_tpg *tpg =
94 container_of(se_tpg, struct efct_lio_tpg, tpg);
95 struct efct *efct;
96 struct efc *efc;
97 unsigned long op;
98
99 if (!tpg->nport || !tpg->nport->efct) {
100 pr_err("%s: Unable to find EFCT device\n", __func__);
101 return -EINVAL;
102 }
103
104 efct = tpg->nport->efct;
105 efc = efct->efcport;
106
107 if (kstrtoul(page, 0, &op) < 0)
108 return -EINVAL;
109
110 if (op == 1) {
111 int ret;
112
113 tpg->enabled = true;
114 efc_log_debug(efct, "enable portal group %d\n", tpg->tpgt);
115
116 ret = efct_xport_control(efct->xport, EFCT_XPORT_PORT_ONLINE);
117 if (ret) {
118 efct->tgt_efct.lio_nport = NULL;
119 efc_log_debug(efct, "cannot bring port online\n");
120 return ret;
121 }
122 } else if (op == 0) {
123 efc_log_debug(efct, "disable portal group %d\n", tpg->tpgt);
124
125 if (efc->domain && efc->domain->nport)
126 efct_scsi_tgt_del_nport(efc, efc->domain->nport);
127
128 tpg->enabled = false;
129 } else {
130 return -EINVAL;
131 }
132
133 return count;
134 }
135
136 static ssize_t
efct_lio_npiv_tpg_enable_show(struct config_item * item,char * page)137 efct_lio_npiv_tpg_enable_show(struct config_item *item, char *page)
138 {
139 struct se_portal_group *se_tpg = to_tpg(item);
140 struct efct_lio_tpg *tpg =
141 container_of(se_tpg, struct efct_lio_tpg, tpg);
142
143 return snprintf(page, PAGE_SIZE, "%d\n", tpg->enabled);
144 }
145
146 static ssize_t
efct_lio_npiv_tpg_enable_store(struct config_item * item,const char * page,size_t count)147 efct_lio_npiv_tpg_enable_store(struct config_item *item, const char *page,
148 size_t count)
149 {
150 struct se_portal_group *se_tpg = to_tpg(item);
151 struct efct_lio_tpg *tpg =
152 container_of(se_tpg, struct efct_lio_tpg, tpg);
153 struct efct_lio_vport *lio_vport = tpg->vport;
154 struct efct *efct;
155 struct efc *efc;
156 unsigned long op;
157
158 if (kstrtoul(page, 0, &op) < 0)
159 return -EINVAL;
160
161 if (!lio_vport) {
162 pr_err("Unable to find vport\n");
163 return -EINVAL;
164 }
165
166 efct = lio_vport->efct;
167 efc = efct->efcport;
168
169 if (op == 1) {
170 tpg->enabled = true;
171 efc_log_debug(efct, "enable portal group %d\n", tpg->tpgt);
172
173 if (efc->domain) {
174 int ret;
175
176 ret = efc_nport_vport_new(efc->domain,
177 lio_vport->npiv_wwpn,
178 lio_vport->npiv_wwnn,
179 U32_MAX, false, true,
180 NULL, NULL);
181 if (ret != 0) {
182 efc_log_err(efct, "Failed to create Vport\n");
183 return ret;
184 }
185 return count;
186 }
187
188 if (!(efc_vport_create_spec(efc, lio_vport->npiv_wwnn,
189 lio_vport->npiv_wwpn, U32_MAX,
190 false, true, NULL, NULL)))
191 return -ENOMEM;
192
193 } else if (op == 0) {
194 efc_log_debug(efct, "disable portal group %d\n", tpg->tpgt);
195
196 tpg->enabled = false;
197 /* only physical nport should exist, free lio_nport
198 * allocated in efct_lio_make_nport
199 */
200 if (efc->domain) {
201 efc_nport_vport_del(efct->efcport, efc->domain,
202 lio_vport->npiv_wwpn,
203 lio_vport->npiv_wwnn);
204 return count;
205 }
206 } else {
207 return -EINVAL;
208 }
209 return count;
210 }
211
efct_lio_get_fabric_wwn(struct se_portal_group * se_tpg)212 static char *efct_lio_get_fabric_wwn(struct se_portal_group *se_tpg)
213 {
214 struct efct_lio_tpg *tpg =
215 container_of(se_tpg, struct efct_lio_tpg, tpg);
216
217 return tpg->nport->wwpn_str;
218 }
219
efct_lio_get_npiv_fabric_wwn(struct se_portal_group * se_tpg)220 static char *efct_lio_get_npiv_fabric_wwn(struct se_portal_group *se_tpg)
221 {
222 struct efct_lio_tpg *tpg =
223 container_of(se_tpg, struct efct_lio_tpg, tpg);
224
225 return tpg->vport->wwpn_str;
226 }
227
efct_lio_get_tag(struct se_portal_group * se_tpg)228 static u16 efct_lio_get_tag(struct se_portal_group *se_tpg)
229 {
230 struct efct_lio_tpg *tpg =
231 container_of(se_tpg, struct efct_lio_tpg, tpg);
232
233 return tpg->tpgt;
234 }
235
efct_lio_get_npiv_tag(struct se_portal_group * se_tpg)236 static u16 efct_lio_get_npiv_tag(struct se_portal_group *se_tpg)
237 {
238 struct efct_lio_tpg *tpg =
239 container_of(se_tpg, struct efct_lio_tpg, tpg);
240
241 return tpg->tpgt;
242 }
243
efct_lio_check_demo_mode(struct se_portal_group * se_tpg)244 static int efct_lio_check_demo_mode(struct se_portal_group *se_tpg)
245 {
246 return 1;
247 }
248
efct_lio_check_demo_mode_cache(struct se_portal_group * se_tpg)249 static int efct_lio_check_demo_mode_cache(struct se_portal_group *se_tpg)
250 {
251 return 1;
252 }
253
efct_lio_check_demo_write_protect(struct se_portal_group * se_tpg)254 static int efct_lio_check_demo_write_protect(struct se_portal_group *se_tpg)
255 {
256 struct efct_lio_tpg *tpg =
257 container_of(se_tpg, struct efct_lio_tpg, tpg);
258
259 return tpg->tpg_attrib.demo_mode_write_protect;
260 }
261
262 static int
efct_lio_npiv_check_demo_write_protect(struct se_portal_group * se_tpg)263 efct_lio_npiv_check_demo_write_protect(struct se_portal_group *se_tpg)
264 {
265 struct efct_lio_tpg *tpg =
266 container_of(se_tpg, struct efct_lio_tpg, tpg);
267
268 return tpg->tpg_attrib.demo_mode_write_protect;
269 }
270
efct_lio_check_prod_write_protect(struct se_portal_group * se_tpg)271 static int efct_lio_check_prod_write_protect(struct se_portal_group *se_tpg)
272 {
273 struct efct_lio_tpg *tpg =
274 container_of(se_tpg, struct efct_lio_tpg, tpg);
275
276 return tpg->tpg_attrib.prod_mode_write_protect;
277 }
278
279 static int
efct_lio_npiv_check_prod_write_protect(struct se_portal_group * se_tpg)280 efct_lio_npiv_check_prod_write_protect(struct se_portal_group *se_tpg)
281 {
282 struct efct_lio_tpg *tpg =
283 container_of(se_tpg, struct efct_lio_tpg, tpg);
284
285 return tpg->tpg_attrib.prod_mode_write_protect;
286 }
287
efct_lio_tpg_get_inst_index(struct se_portal_group * se_tpg)288 static u32 efct_lio_tpg_get_inst_index(struct se_portal_group *se_tpg)
289 {
290 return 1;
291 }
292
efct_lio_check_stop_free(struct se_cmd * se_cmd)293 static int efct_lio_check_stop_free(struct se_cmd *se_cmd)
294 {
295 struct efct_scsi_tgt_io *ocp =
296 container_of(se_cmd, struct efct_scsi_tgt_io, cmd);
297 struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
298
299 efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_CHK_STOP_FREE);
300 return target_put_sess_cmd(se_cmd);
301 }
302
303 static int
efct_lio_abort_tgt_cb(struct efct_io * io,enum efct_scsi_io_status scsi_status,u32 flags,void * arg)304 efct_lio_abort_tgt_cb(struct efct_io *io,
305 enum efct_scsi_io_status scsi_status,
306 u32 flags, void *arg)
307 {
308 efct_lio_io_printf(io, "Abort done, status:%d\n", scsi_status);
309 return 0;
310 }
311
312 static void
efct_lio_aborted_task(struct se_cmd * se_cmd)313 efct_lio_aborted_task(struct se_cmd *se_cmd)
314 {
315 struct efct_scsi_tgt_io *ocp =
316 container_of(se_cmd, struct efct_scsi_tgt_io, cmd);
317 struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
318
319 efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_ABORTED_TASK);
320
321 if (ocp->rsp_sent)
322 return;
323
324 /* command has been aborted, cleanup here */
325 ocp->aborting = true;
326 ocp->err = EFCT_SCSI_STATUS_ABORTED;
327 /* terminate the exchange */
328 efct_scsi_tgt_abort_io(io, efct_lio_abort_tgt_cb, NULL);
329 }
330
efct_lio_release_cmd(struct se_cmd * se_cmd)331 static void efct_lio_release_cmd(struct se_cmd *se_cmd)
332 {
333 struct efct_scsi_tgt_io *ocp =
334 container_of(se_cmd, struct efct_scsi_tgt_io, cmd);
335 struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
336 struct efct *efct = io->efct;
337
338 efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_RELEASE_CMD);
339 efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_CMPL_CMD);
340 efct_scsi_io_complete(io);
341 atomic_sub_return(1, &efct->tgt_efct.ios_in_use);
342 }
343
efct_lio_close_session(struct se_session * se_sess)344 static void efct_lio_close_session(struct se_session *se_sess)
345 {
346 struct efc_node *node = se_sess->fabric_sess_ptr;
347
348 pr_debug("se_sess=%p node=%p", se_sess, node);
349
350 if (!node) {
351 pr_debug("node is NULL");
352 return;
353 }
354
355 efc_node_post_shutdown(node, NULL);
356 }
357
efct_lio_sess_get_index(struct se_session * se_sess)358 static u32 efct_lio_sess_get_index(struct se_session *se_sess)
359 {
360 return 0;
361 }
362
efct_lio_set_default_node_attrs(struct se_node_acl * nacl)363 static void efct_lio_set_default_node_attrs(struct se_node_acl *nacl)
364 {
365 }
366
efct_lio_get_cmd_state(struct se_cmd * cmd)367 static int efct_lio_get_cmd_state(struct se_cmd *cmd)
368 {
369 struct efct_scsi_tgt_io *ocp =
370 container_of(cmd, struct efct_scsi_tgt_io, cmd);
371 struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
372
373 if (!io)
374 return 0;
375
376 return io->tgt_io.state;
377 }
378
379 static int
efct_lio_sg_map(struct efct_io * io)380 efct_lio_sg_map(struct efct_io *io)
381 {
382 struct efct_scsi_tgt_io *ocp = &io->tgt_io;
383 struct se_cmd *cmd = &ocp->cmd;
384
385 ocp->seg_map_cnt = pci_map_sg(io->efct->pci, cmd->t_data_sg,
386 cmd->t_data_nents, cmd->data_direction);
387 if (ocp->seg_map_cnt == 0)
388 return -EFAULT;
389 return 0;
390 }
391
392 static void
efct_lio_sg_unmap(struct efct_io * io)393 efct_lio_sg_unmap(struct efct_io *io)
394 {
395 struct efct_scsi_tgt_io *ocp = &io->tgt_io;
396 struct se_cmd *cmd = &ocp->cmd;
397
398 if (WARN_ON(!ocp->seg_map_cnt || !cmd->t_data_sg))
399 return;
400
401 pci_unmap_sg(io->efct->pci, cmd->t_data_sg,
402 ocp->seg_map_cnt, cmd->data_direction);
403 ocp->seg_map_cnt = 0;
404 }
405
406 static int
efct_lio_status_done(struct efct_io * io,enum efct_scsi_io_status scsi_status,u32 flags,void * arg)407 efct_lio_status_done(struct efct_io *io,
408 enum efct_scsi_io_status scsi_status,
409 u32 flags, void *arg)
410 {
411 struct efct_scsi_tgt_io *ocp = &io->tgt_io;
412
413 efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_RSP_DONE);
414 if (scsi_status != EFCT_SCSI_STATUS_GOOD) {
415 efct_lio_io_printf(io, "callback completed with error=%d\n",
416 scsi_status);
417 ocp->err = scsi_status;
418 }
419 if (ocp->seg_map_cnt)
420 efct_lio_sg_unmap(io);
421
422 efct_lio_io_printf(io, "status=%d, err=%d flags=0x%x, dir=%d\n",
423 scsi_status, ocp->err, flags, ocp->ddir);
424
425 efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_GENERIC_FREE);
426 transport_generic_free_cmd(&io->tgt_io.cmd, 0);
427 return 0;
428 }
429
430 static int
431 efct_lio_datamove_done(struct efct_io *io, enum efct_scsi_io_status scsi_status,
432 u32 flags, void *arg);
433
434 static int
efct_lio_write_pending(struct se_cmd * cmd)435 efct_lio_write_pending(struct se_cmd *cmd)
436 {
437 struct efct_scsi_tgt_io *ocp =
438 container_of(cmd, struct efct_scsi_tgt_io, cmd);
439 struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
440 struct efct_scsi_sgl *sgl = io->sgl;
441 struct scatterlist *sg;
442 u32 flags = 0, cnt, curcnt;
443 u64 length = 0;
444
445 efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_WRITE_PENDING);
446 efct_lio_io_printf(io, "trans_state=0x%x se_cmd_flags=0x%x\n",
447 cmd->transport_state, cmd->se_cmd_flags);
448
449 if (ocp->seg_cnt == 0) {
450 ocp->seg_cnt = cmd->t_data_nents;
451 ocp->cur_seg = 0;
452 if (efct_lio_sg_map(io)) {
453 efct_lio_io_printf(io, "efct_lio_sg_map failed\n");
454 return -EFAULT;
455 }
456 }
457 curcnt = (ocp->seg_map_cnt - ocp->cur_seg);
458 curcnt = (curcnt < io->sgl_allocated) ? curcnt : io->sgl_allocated;
459 /* find current sg */
460 for (cnt = 0, sg = cmd->t_data_sg; cnt < ocp->cur_seg; cnt++,
461 sg = sg_next(sg))
462 ;/* do nothing */
463
464 for (cnt = 0; cnt < curcnt; cnt++, sg = sg_next(sg)) {
465 sgl[cnt].addr = sg_dma_address(sg);
466 sgl[cnt].dif_addr = 0;
467 sgl[cnt].len = sg_dma_len(sg);
468 length += sgl[cnt].len;
469 ocp->cur_seg++;
470 }
471
472 if (ocp->cur_seg == ocp->seg_cnt)
473 flags = EFCT_SCSI_LAST_DATAPHASE;
474
475 return efct_scsi_recv_wr_data(io, flags, sgl, curcnt, length,
476 efct_lio_datamove_done, NULL);
477 }
478
479 static int
efct_lio_queue_data_in(struct se_cmd * cmd)480 efct_lio_queue_data_in(struct se_cmd *cmd)
481 {
482 struct efct_scsi_tgt_io *ocp =
483 container_of(cmd, struct efct_scsi_tgt_io, cmd);
484 struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
485 struct efct_scsi_sgl *sgl = io->sgl;
486 struct scatterlist *sg = NULL;
487 uint flags = 0, cnt = 0, curcnt = 0;
488 u64 length = 0;
489
490 efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_QUEUE_DATA_IN);
491
492 if (ocp->seg_cnt == 0) {
493 if (cmd->data_length) {
494 ocp->seg_cnt = cmd->t_data_nents;
495 ocp->cur_seg = 0;
496 if (efct_lio_sg_map(io)) {
497 efct_lio_io_printf(io,
498 "efct_lio_sg_map failed\n");
499 return -EAGAIN;
500 }
501 } else {
502 /* If command length is 0, send the response status */
503 struct efct_scsi_cmd_resp rsp;
504
505 memset(&rsp, 0, sizeof(rsp));
506 efct_lio_io_printf(io,
507 "cmd : %p length 0, send status\n",
508 cmd);
509 return efct_scsi_send_resp(io, 0, &rsp,
510 efct_lio_status_done, NULL);
511 }
512 }
513 curcnt = min(ocp->seg_map_cnt - ocp->cur_seg, io->sgl_allocated);
514
515 while (cnt < curcnt) {
516 sg = &cmd->t_data_sg[ocp->cur_seg];
517 sgl[cnt].addr = sg_dma_address(sg);
518 sgl[cnt].dif_addr = 0;
519 if (ocp->transferred_len + sg_dma_len(sg) >= cmd->data_length)
520 sgl[cnt].len = cmd->data_length - ocp->transferred_len;
521 else
522 sgl[cnt].len = sg_dma_len(sg);
523
524 ocp->transferred_len += sgl[cnt].len;
525 length += sgl[cnt].len;
526 ocp->cur_seg++;
527 cnt++;
528 if (ocp->transferred_len == cmd->data_length)
529 break;
530 }
531
532 if (ocp->transferred_len == cmd->data_length) {
533 flags = EFCT_SCSI_LAST_DATAPHASE;
534 ocp->seg_cnt = ocp->cur_seg;
535 }
536
537 /* If there is residual, disable Auto Good Response */
538 if (cmd->residual_count)
539 flags |= EFCT_SCSI_NO_AUTO_RESPONSE;
540
541 efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_SEND_RD_DATA);
542
543 return efct_scsi_send_rd_data(io, flags, sgl, curcnt, length,
544 efct_lio_datamove_done, NULL);
545 }
546
547 static void
efct_lio_send_resp(struct efct_io * io,enum efct_scsi_io_status scsi_status,u32 flags)548 efct_lio_send_resp(struct efct_io *io, enum efct_scsi_io_status scsi_status,
549 u32 flags)
550 {
551 struct efct_scsi_cmd_resp rsp;
552 struct efct_scsi_tgt_io *ocp = &io->tgt_io;
553 struct se_cmd *cmd = &io->tgt_io.cmd;
554 int rc;
555
556 if (flags & EFCT_SCSI_IO_CMPL_RSP_SENT) {
557 ocp->rsp_sent = true;
558 efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_GENERIC_FREE);
559 transport_generic_free_cmd(&io->tgt_io.cmd, 0);
560 return;
561 }
562
563 /* send check condition if an error occurred */
564 memset(&rsp, 0, sizeof(rsp));
565 rsp.scsi_status = cmd->scsi_status;
566 rsp.sense_data = (uint8_t *)io->tgt_io.sense_buffer;
567 rsp.sense_data_length = cmd->scsi_sense_length;
568
569 /* Check for residual underrun or overrun */
570 if (cmd->se_cmd_flags & SCF_OVERFLOW_BIT)
571 rsp.residual = -cmd->residual_count;
572 else if (cmd->se_cmd_flags & SCF_UNDERFLOW_BIT)
573 rsp.residual = cmd->residual_count;
574
575 rc = efct_scsi_send_resp(io, 0, &rsp, efct_lio_status_done, NULL);
576 efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_SEND_RSP);
577 if (rc != 0) {
578 efct_lio_io_printf(io, "Read done, send rsp failed %d\n", rc);
579 efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_GENERIC_FREE);
580 transport_generic_free_cmd(&io->tgt_io.cmd, 0);
581 } else {
582 ocp->rsp_sent = true;
583 }
584 }
585
586 static int
efct_lio_datamove_done(struct efct_io * io,enum efct_scsi_io_status scsi_status,u32 flags,void * arg)587 efct_lio_datamove_done(struct efct_io *io, enum efct_scsi_io_status scsi_status,
588 u32 flags, void *arg)
589 {
590 struct efct_scsi_tgt_io *ocp = &io->tgt_io;
591
592 efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_DATA_DONE);
593 if (scsi_status != EFCT_SCSI_STATUS_GOOD) {
594 efct_lio_io_printf(io, "callback completed with error=%d\n",
595 scsi_status);
596 ocp->err = scsi_status;
597 }
598 efct_lio_io_printf(io, "seg_map_cnt=%d\n", ocp->seg_map_cnt);
599 if (ocp->seg_map_cnt) {
600 if (ocp->err == EFCT_SCSI_STATUS_GOOD &&
601 ocp->cur_seg < ocp->seg_cnt) {
602 int rc;
603
604 efct_lio_io_printf(io, "continuing cmd at segm=%d\n",
605 ocp->cur_seg);
606 if (ocp->ddir == DMA_TO_DEVICE)
607 rc = efct_lio_write_pending(&ocp->cmd);
608 else
609 rc = efct_lio_queue_data_in(&ocp->cmd);
610 if (!rc)
611 return 0;
612
613 ocp->err = EFCT_SCSI_STATUS_ERROR;
614 efct_lio_io_printf(io, "could not continue command\n");
615 }
616 efct_lio_sg_unmap(io);
617 }
618
619 if (io->tgt_io.aborting) {
620 efct_lio_io_printf(io, "IO done aborted\n");
621 return 0;
622 }
623
624 if (ocp->ddir == DMA_TO_DEVICE) {
625 efct_lio_io_printf(io, "Write done, trans_state=0x%x\n",
626 io->tgt_io.cmd.transport_state);
627 if (scsi_status != EFCT_SCSI_STATUS_GOOD) {
628 transport_generic_request_failure(&io->tgt_io.cmd,
629 TCM_CHECK_CONDITION_ABORT_CMD);
630 efct_set_lio_io_state(io,
631 EFCT_LIO_STATE_TGT_GENERIC_REQ_FAILURE);
632 } else {
633 efct_set_lio_io_state(io,
634 EFCT_LIO_STATE_TGT_EXECUTE_CMD);
635 target_execute_cmd(&io->tgt_io.cmd);
636 }
637 } else {
638 efct_lio_send_resp(io, scsi_status, flags);
639 }
640 return 0;
641 }
642
643 static int
efct_lio_tmf_done(struct efct_io * io,enum efct_scsi_io_status scsi_status,u32 flags,void * arg)644 efct_lio_tmf_done(struct efct_io *io, enum efct_scsi_io_status scsi_status,
645 u32 flags, void *arg)
646 {
647 efct_lio_tmfio_printf(io, "cmd=%p status=%d, flags=0x%x\n",
648 &io->tgt_io.cmd, scsi_status, flags);
649
650 efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_GENERIC_FREE);
651 transport_generic_free_cmd(&io->tgt_io.cmd, 0);
652 return 0;
653 }
654
655 static int
efct_lio_null_tmf_done(struct efct_io * tmfio,enum efct_scsi_io_status scsi_status,u32 flags,void * arg)656 efct_lio_null_tmf_done(struct efct_io *tmfio,
657 enum efct_scsi_io_status scsi_status,
658 u32 flags, void *arg)
659 {
660 efct_lio_tmfio_printf(tmfio, "cmd=%p status=%d, flags=0x%x\n",
661 &tmfio->tgt_io.cmd, scsi_status, flags);
662
663 /* free struct efct_io only, no active se_cmd */
664 efct_scsi_io_complete(tmfio);
665 return 0;
666 }
667
668 static int
efct_lio_queue_status(struct se_cmd * cmd)669 efct_lio_queue_status(struct se_cmd *cmd)
670 {
671 struct efct_scsi_cmd_resp rsp;
672 struct efct_scsi_tgt_io *ocp =
673 container_of(cmd, struct efct_scsi_tgt_io, cmd);
674 struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
675 int rc = 0;
676
677 efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_QUEUE_STATUS);
678 efct_lio_io_printf(io,
679 "status=0x%x trans_state=0x%x se_cmd_flags=0x%x sns_len=%d\n",
680 cmd->scsi_status, cmd->transport_state, cmd->se_cmd_flags,
681 cmd->scsi_sense_length);
682
683 memset(&rsp, 0, sizeof(rsp));
684 rsp.scsi_status = cmd->scsi_status;
685 rsp.sense_data = (u8 *)io->tgt_io.sense_buffer;
686 rsp.sense_data_length = cmd->scsi_sense_length;
687
688 /* Check for residual underrun or overrun, mark negitive value for
689 * underrun to recognize in HW
690 */
691 if (cmd->se_cmd_flags & SCF_OVERFLOW_BIT)
692 rsp.residual = -cmd->residual_count;
693 else if (cmd->se_cmd_flags & SCF_UNDERFLOW_BIT)
694 rsp.residual = cmd->residual_count;
695
696 rc = efct_scsi_send_resp(io, 0, &rsp, efct_lio_status_done, NULL);
697 efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_SEND_RSP);
698 if (rc == 0)
699 ocp->rsp_sent = true;
700 return rc;
701 }
702
efct_lio_queue_tm_rsp(struct se_cmd * cmd)703 static void efct_lio_queue_tm_rsp(struct se_cmd *cmd)
704 {
705 struct efct_scsi_tgt_io *ocp =
706 container_of(cmd, struct efct_scsi_tgt_io, cmd);
707 struct efct_io *tmfio = container_of(ocp, struct efct_io, tgt_io);
708 struct se_tmr_req *se_tmr = cmd->se_tmr_req;
709 u8 rspcode;
710
711 efct_lio_tmfio_printf(tmfio, "cmd=%p function=0x%x tmr->response=%d\n",
712 cmd, se_tmr->function, se_tmr->response);
713 switch (se_tmr->response) {
714 case TMR_FUNCTION_COMPLETE:
715 rspcode = EFCT_SCSI_TMF_FUNCTION_COMPLETE;
716 break;
717 case TMR_TASK_DOES_NOT_EXIST:
718 rspcode = EFCT_SCSI_TMF_FUNCTION_IO_NOT_FOUND;
719 break;
720 case TMR_LUN_DOES_NOT_EXIST:
721 rspcode = EFCT_SCSI_TMF_INCORRECT_LOGICAL_UNIT_NUMBER;
722 break;
723 case TMR_FUNCTION_REJECTED:
724 default:
725 rspcode = EFCT_SCSI_TMF_FUNCTION_REJECTED;
726 break;
727 }
728 efct_scsi_send_tmf_resp(tmfio, rspcode, NULL, efct_lio_tmf_done, NULL);
729 }
730
efct_find_wwpn(u64 wwpn)731 static struct efct *efct_find_wwpn(u64 wwpn)
732 {
733 struct efct *efct;
734
735 /* Search for the HBA that has this WWPN */
736 list_for_each_entry(efct, &efct_devices, list_entry) {
737
738 if (wwpn == efct_get_wwpn(&efct->hw))
739 return efct;
740 }
741
742 return NULL;
743 }
744
745 static struct se_wwn *
efct_lio_make_nport(struct target_fabric_configfs * tf,struct config_group * group,const char * name)746 efct_lio_make_nport(struct target_fabric_configfs *tf,
747 struct config_group *group, const char *name)
748 {
749 struct efct_lio_nport *lio_nport;
750 struct efct *efct;
751 int ret;
752 u64 wwpn;
753
754 ret = efct_lio_parse_wwn(name, &wwpn, 0);
755 if (ret)
756 return ERR_PTR(ret);
757
758 efct = efct_find_wwpn(wwpn);
759 if (!efct) {
760 pr_err("cannot find EFCT for base wwpn %s\n", name);
761 return ERR_PTR(-ENXIO);
762 }
763
764 lio_nport = kzalloc(sizeof(*lio_nport), GFP_KERNEL);
765 if (!lio_nport)
766 return ERR_PTR(-ENOMEM);
767
768 lio_nport->efct = efct;
769 lio_nport->wwpn = wwpn;
770 efct_format_wwn(lio_nport->wwpn_str, sizeof(lio_nport->wwpn_str),
771 "naa.", wwpn);
772 efct->tgt_efct.lio_nport = lio_nport;
773
774 return &lio_nport->nport_wwn;
775 }
776
777 static struct se_wwn *
efct_lio_npiv_make_nport(struct target_fabric_configfs * tf,struct config_group * group,const char * name)778 efct_lio_npiv_make_nport(struct target_fabric_configfs *tf,
779 struct config_group *group, const char *name)
780 {
781 struct efct_lio_vport *lio_vport;
782 struct efct *efct;
783 int ret;
784 u64 p_wwpn, npiv_wwpn, npiv_wwnn;
785 char *p, *pbuf, tmp[128];
786 struct efct_lio_vport_list_t *vport_list;
787 struct fc_vport *new_fc_vport;
788 struct fc_vport_identifiers vport_id;
789 unsigned long flags = 0;
790
791 snprintf(tmp, sizeof(tmp), "%s", name);
792 pbuf = &tmp[0];
793
794 p = strsep(&pbuf, "@");
795
796 if (!p || !pbuf) {
797 pr_err("Unable to find separator operator(@)\n");
798 return ERR_PTR(-EINVAL);
799 }
800
801 ret = efct_lio_parse_wwn(p, &p_wwpn, 0);
802 if (ret)
803 return ERR_PTR(ret);
804
805 ret = efct_lio_parse_npiv_wwn(pbuf, strlen(pbuf), &npiv_wwpn,
806 &npiv_wwnn);
807 if (ret)
808 return ERR_PTR(ret);
809
810 efct = efct_find_wwpn(p_wwpn);
811 if (!efct) {
812 pr_err("cannot find EFCT for base wwpn %s\n", name);
813 return ERR_PTR(-ENXIO);
814 }
815
816 lio_vport = kzalloc(sizeof(*lio_vport), GFP_KERNEL);
817 if (!lio_vport)
818 return ERR_PTR(-ENOMEM);
819
820 lio_vport->efct = efct;
821 lio_vport->wwpn = p_wwpn;
822 lio_vport->npiv_wwpn = npiv_wwpn;
823 lio_vport->npiv_wwnn = npiv_wwnn;
824
825 efct_format_wwn(lio_vport->wwpn_str, sizeof(lio_vport->wwpn_str),
826 "naa.", npiv_wwpn);
827
828 vport_list = kzalloc(sizeof(*vport_list), GFP_KERNEL);
829 if (!vport_list) {
830 kfree(lio_vport);
831 return ERR_PTR(-ENOMEM);
832 }
833
834 vport_list->lio_vport = lio_vport;
835
836 memset(&vport_id, 0, sizeof(vport_id));
837 vport_id.port_name = npiv_wwpn;
838 vport_id.node_name = npiv_wwnn;
839 vport_id.roles = FC_PORT_ROLE_FCP_INITIATOR;
840 vport_id.vport_type = FC_PORTTYPE_NPIV;
841 vport_id.disable = false;
842
843 new_fc_vport = fc_vport_create(efct->shost, 0, &vport_id);
844 if (!new_fc_vport) {
845 efc_log_err(efct, "fc_vport_create failed\n");
846 kfree(lio_vport);
847 kfree(vport_list);
848 return ERR_PTR(-ENOMEM);
849 }
850
851 lio_vport->fc_vport = new_fc_vport;
852 spin_lock_irqsave(&efct->tgt_efct.efct_lio_lock, flags);
853 INIT_LIST_HEAD(&vport_list->list_entry);
854 list_add_tail(&vport_list->list_entry, &efct->tgt_efct.vport_list);
855 spin_unlock_irqrestore(&efct->tgt_efct.efct_lio_lock, flags);
856
857 return &lio_vport->vport_wwn;
858 }
859
860 static void
efct_lio_drop_nport(struct se_wwn * wwn)861 efct_lio_drop_nport(struct se_wwn *wwn)
862 {
863 struct efct_lio_nport *lio_nport =
864 container_of(wwn, struct efct_lio_nport, nport_wwn);
865 struct efct *efct = lio_nport->efct;
866
867 /* only physical nport should exist, free lio_nport allocated
868 * in efct_lio_make_nport.
869 */
870 kfree(efct->tgt_efct.lio_nport);
871 efct->tgt_efct.lio_nport = NULL;
872 }
873
874 static void
efct_lio_npiv_drop_nport(struct se_wwn * wwn)875 efct_lio_npiv_drop_nport(struct se_wwn *wwn)
876 {
877 struct efct_lio_vport *lio_vport =
878 container_of(wwn, struct efct_lio_vport, vport_wwn);
879 struct efct_lio_vport_list_t *vport, *next_vport;
880 struct efct *efct = lio_vport->efct;
881 unsigned long flags = 0;
882
883 if (lio_vport->fc_vport)
884 fc_vport_terminate(lio_vport->fc_vport);
885
886 spin_lock_irqsave(&efct->tgt_efct.efct_lio_lock, flags);
887
888 list_for_each_entry_safe(vport, next_vport, &efct->tgt_efct.vport_list,
889 list_entry) {
890 if (vport->lio_vport == lio_vport) {
891 list_del(&vport->list_entry);
892 kfree(vport->lio_vport);
893 kfree(vport);
894 break;
895 }
896 }
897 spin_unlock_irqrestore(&efct->tgt_efct.efct_lio_lock, flags);
898 }
899
900 static struct se_portal_group *
efct_lio_make_tpg(struct se_wwn * wwn,const char * name)901 efct_lio_make_tpg(struct se_wwn *wwn, const char *name)
902 {
903 struct efct_lio_nport *lio_nport =
904 container_of(wwn, struct efct_lio_nport, nport_wwn);
905 struct efct_lio_tpg *tpg;
906 struct efct *efct;
907 unsigned long n;
908 int ret;
909
910 if (strstr(name, "tpgt_") != name)
911 return ERR_PTR(-EINVAL);
912 if (kstrtoul(name + 5, 10, &n) || n > USHRT_MAX)
913 return ERR_PTR(-EINVAL);
914
915 tpg = kzalloc(sizeof(*tpg), GFP_KERNEL);
916 if (!tpg)
917 return ERR_PTR(-ENOMEM);
918
919 tpg->nport = lio_nport;
920 tpg->tpgt = n;
921 tpg->enabled = false;
922
923 tpg->tpg_attrib.generate_node_acls = 1;
924 tpg->tpg_attrib.demo_mode_write_protect = 1;
925 tpg->tpg_attrib.cache_dynamic_acls = 1;
926 tpg->tpg_attrib.demo_mode_login_only = 1;
927 tpg->tpg_attrib.session_deletion_wait = 1;
928
929 ret = core_tpg_register(wwn, &tpg->tpg, SCSI_PROTOCOL_FCP);
930 if (ret < 0) {
931 kfree(tpg);
932 return NULL;
933 }
934 efct = lio_nport->efct;
935 efct->tgt_efct.tpg = tpg;
936 efc_log_debug(efct, "create portal group %d\n", tpg->tpgt);
937
938 xa_init(&efct->lookup);
939 return &tpg->tpg;
940 }
941
942 static void
efct_lio_drop_tpg(struct se_portal_group * se_tpg)943 efct_lio_drop_tpg(struct se_portal_group *se_tpg)
944 {
945 struct efct_lio_tpg *tpg =
946 container_of(se_tpg, struct efct_lio_tpg, tpg);
947
948 struct efct *efct = tpg->nport->efct;
949
950 efc_log_debug(efct, "drop portal group %d\n", tpg->tpgt);
951 tpg->nport->efct->tgt_efct.tpg = NULL;
952 core_tpg_deregister(se_tpg);
953 xa_destroy(&efct->lookup);
954 kfree(tpg);
955 }
956
957 static struct se_portal_group *
efct_lio_npiv_make_tpg(struct se_wwn * wwn,const char * name)958 efct_lio_npiv_make_tpg(struct se_wwn *wwn, const char *name)
959 {
960 struct efct_lio_vport *lio_vport =
961 container_of(wwn, struct efct_lio_vport, vport_wwn);
962 struct efct_lio_tpg *tpg;
963 struct efct *efct;
964 unsigned long n;
965 int ret;
966
967 efct = lio_vport->efct;
968 if (strstr(name, "tpgt_") != name)
969 return ERR_PTR(-EINVAL);
970 if (kstrtoul(name + 5, 10, &n) || n > USHRT_MAX)
971 return ERR_PTR(-EINVAL);
972
973 if (n != 1) {
974 efc_log_err(efct, "Invalid tpgt index: %ld provided\n", n);
975 return ERR_PTR(-EINVAL);
976 }
977
978 tpg = kzalloc(sizeof(*tpg), GFP_KERNEL);
979 if (!tpg)
980 return ERR_PTR(-ENOMEM);
981
982 tpg->vport = lio_vport;
983 tpg->tpgt = n;
984 tpg->enabled = false;
985
986 tpg->tpg_attrib.generate_node_acls = 1;
987 tpg->tpg_attrib.demo_mode_write_protect = 1;
988 tpg->tpg_attrib.cache_dynamic_acls = 1;
989 tpg->tpg_attrib.demo_mode_login_only = 1;
990 tpg->tpg_attrib.session_deletion_wait = 1;
991
992 ret = core_tpg_register(wwn, &tpg->tpg, SCSI_PROTOCOL_FCP);
993
994 if (ret < 0) {
995 kfree(tpg);
996 return NULL;
997 }
998 lio_vport->tpg = tpg;
999 efc_log_debug(efct, "create vport portal group %d\n", tpg->tpgt);
1000
1001 return &tpg->tpg;
1002 }
1003
1004 static void
efct_lio_npiv_drop_tpg(struct se_portal_group * se_tpg)1005 efct_lio_npiv_drop_tpg(struct se_portal_group *se_tpg)
1006 {
1007 struct efct_lio_tpg *tpg =
1008 container_of(se_tpg, struct efct_lio_tpg, tpg);
1009
1010 efc_log_debug(tpg->vport->efct, "drop npiv portal group %d\n",
1011 tpg->tpgt);
1012 core_tpg_deregister(se_tpg);
1013 kfree(tpg);
1014 }
1015
1016 static int
efct_lio_init_nodeacl(struct se_node_acl * se_nacl,const char * name)1017 efct_lio_init_nodeacl(struct se_node_acl *se_nacl, const char *name)
1018 {
1019 struct efct_lio_nacl *nacl;
1020 u64 wwnn;
1021
1022 if (efct_lio_parse_wwn(name, &wwnn, 0) < 0)
1023 return -EINVAL;
1024
1025 nacl = container_of(se_nacl, struct efct_lio_nacl, se_node_acl);
1026 nacl->nport_wwnn = wwnn;
1027
1028 efct_format_wwn(nacl->nport_name, sizeof(nacl->nport_name), "", wwnn);
1029 return 0;
1030 }
1031
efct_lio_check_demo_mode_login_only(struct se_portal_group * stpg)1032 static int efct_lio_check_demo_mode_login_only(struct se_portal_group *stpg)
1033 {
1034 struct efct_lio_tpg *tpg = container_of(stpg, struct efct_lio_tpg, tpg);
1035
1036 return tpg->tpg_attrib.demo_mode_login_only;
1037 }
1038
1039 static int
efct_lio_npiv_check_demo_mode_login_only(struct se_portal_group * stpg)1040 efct_lio_npiv_check_demo_mode_login_only(struct se_portal_group *stpg)
1041 {
1042 struct efct_lio_tpg *tpg = container_of(stpg, struct efct_lio_tpg, tpg);
1043
1044 return tpg->tpg_attrib.demo_mode_login_only;
1045 }
1046
1047 static struct efct_lio_tpg *
efct_get_vport_tpg(struct efc_node * node)1048 efct_get_vport_tpg(struct efc_node *node)
1049 {
1050 struct efct *efct;
1051 u64 wwpn = node->nport->wwpn;
1052 struct efct_lio_vport_list_t *vport, *next;
1053 struct efct_lio_vport *lio_vport = NULL;
1054 struct efct_lio_tpg *tpg = NULL;
1055 unsigned long flags = 0;
1056
1057 efct = node->efc->base;
1058 spin_lock_irqsave(&efct->tgt_efct.efct_lio_lock, flags);
1059 list_for_each_entry_safe(vport, next, &efct->tgt_efct.vport_list,
1060 list_entry) {
1061 lio_vport = vport->lio_vport;
1062 if (wwpn && lio_vport && lio_vport->npiv_wwpn == wwpn) {
1063 efc_log_debug(efct, "found tpg on vport\n");
1064 tpg = lio_vport->tpg;
1065 break;
1066 }
1067 }
1068 spin_unlock_irqrestore(&efct->tgt_efct.efct_lio_lock, flags);
1069 return tpg;
1070 }
1071
1072 static void
_efct_tgt_node_free(struct kref * arg)1073 _efct_tgt_node_free(struct kref *arg)
1074 {
1075 struct efct_node *tgt_node = container_of(arg, struct efct_node, ref);
1076 struct efc_node *node = tgt_node->node;
1077
1078 efc_scsi_del_initiator_complete(node->efc, node);
1079 kfree(tgt_node);
1080 }
1081
efct_session_cb(struct se_portal_group * se_tpg,struct se_session * se_sess,void * private)1082 static int efct_session_cb(struct se_portal_group *se_tpg,
1083 struct se_session *se_sess, void *private)
1084 {
1085 struct efc_node *node = private;
1086 struct efct_node *tgt_node;
1087 struct efct *efct = node->efc->base;
1088
1089 tgt_node = kzalloc(sizeof(*tgt_node), GFP_KERNEL);
1090 if (!tgt_node)
1091 return -ENOMEM;
1092
1093 kref_init(&tgt_node->ref);
1094 tgt_node->release = _efct_tgt_node_free;
1095
1096 tgt_node->session = se_sess;
1097 node->tgt_node = tgt_node;
1098 tgt_node->efct = efct;
1099
1100 tgt_node->node = node;
1101
1102 tgt_node->node_fc_id = node->rnode.fc_id;
1103 tgt_node->port_fc_id = node->nport->fc_id;
1104 tgt_node->vpi = node->nport->indicator;
1105 tgt_node->rpi = node->rnode.indicator;
1106
1107 spin_lock_init(&tgt_node->active_ios_lock);
1108 INIT_LIST_HEAD(&tgt_node->active_ios);
1109
1110 return 0;
1111 }
1112
efct_scsi_tgt_new_device(struct efct * efct)1113 int efct_scsi_tgt_new_device(struct efct *efct)
1114 {
1115 u32 total_ios;
1116
1117 /* Get the max settings */
1118 efct->tgt_efct.max_sge = sli_get_max_sge(&efct->hw.sli);
1119 efct->tgt_efct.max_sgl = sli_get_max_sgl(&efct->hw.sli);
1120
1121 /* initialize IO watermark fields */
1122 atomic_set(&efct->tgt_efct.ios_in_use, 0);
1123 total_ios = efct->hw.config.n_io;
1124 efc_log_debug(efct, "total_ios=%d\n", total_ios);
1125 efct->tgt_efct.watermark_min =
1126 (total_ios * EFCT_WATERMARK_LOW_PCT) / 100;
1127 efct->tgt_efct.watermark_max =
1128 (total_ios * EFCT_WATERMARK_HIGH_PCT) / 100;
1129 atomic_set(&efct->tgt_efct.io_high_watermark,
1130 efct->tgt_efct.watermark_max);
1131 atomic_set(&efct->tgt_efct.watermark_hit, 0);
1132 atomic_set(&efct->tgt_efct.initiator_count, 0);
1133
1134 lio_wq = create_singlethread_workqueue("efct_lio_worker");
1135 if (!lio_wq) {
1136 efc_log_err(efct, "workqueue create failed\n");
1137 return -EIO;
1138 }
1139
1140 spin_lock_init(&efct->tgt_efct.efct_lio_lock);
1141 INIT_LIST_HEAD(&efct->tgt_efct.vport_list);
1142
1143 return 0;
1144 }
1145
efct_scsi_tgt_del_device(struct efct * efct)1146 int efct_scsi_tgt_del_device(struct efct *efct)
1147 {
1148 flush_workqueue(lio_wq);
1149
1150 return 0;
1151 }
1152
1153 int
efct_scsi_tgt_new_nport(struct efc * efc,struct efc_nport * nport)1154 efct_scsi_tgt_new_nport(struct efc *efc, struct efc_nport *nport)
1155 {
1156 struct efct *efct = nport->efc->base;
1157
1158 efc_log_debug(efct, "New SPORT: %s bound to %s\n", nport->display_name,
1159 efct->tgt_efct.lio_nport->wwpn_str);
1160
1161 return 0;
1162 }
1163
1164 void
efct_scsi_tgt_del_nport(struct efc * efc,struct efc_nport * nport)1165 efct_scsi_tgt_del_nport(struct efc *efc, struct efc_nport *nport)
1166 {
1167 efc_log_debug(efc, "Del SPORT: %s\n", nport->display_name);
1168 }
1169
efct_lio_setup_session(struct work_struct * work)1170 static void efct_lio_setup_session(struct work_struct *work)
1171 {
1172 struct efct_lio_wq_data *wq_data =
1173 container_of(work, struct efct_lio_wq_data, work);
1174 struct efct *efct = wq_data->efct;
1175 struct efc_node *node = wq_data->ptr;
1176 char wwpn[WWN_NAME_LEN];
1177 struct efct_lio_tpg *tpg;
1178 struct efct_node *tgt_node;
1179 struct se_portal_group *se_tpg;
1180 struct se_session *se_sess;
1181 int watermark;
1182 int ini_count;
1183 u64 id;
1184
1185 /* Check to see if it's belongs to vport,
1186 * if not get physical port
1187 */
1188 tpg = efct_get_vport_tpg(node);
1189 if (tpg) {
1190 se_tpg = &tpg->tpg;
1191 } else if (efct->tgt_efct.tpg) {
1192 tpg = efct->tgt_efct.tpg;
1193 se_tpg = &tpg->tpg;
1194 } else {
1195 efc_log_err(efct, "failed to init session\n");
1196 return;
1197 }
1198
1199 /*
1200 * Format the FCP Initiator port_name into colon
1201 * separated values to match the format by our explicit
1202 * ConfigFS NodeACLs.
1203 */
1204 efct_format_wwn(wwpn, sizeof(wwpn), "", efc_node_get_wwpn(node));
1205
1206 se_sess = target_setup_session(se_tpg, 0, 0, TARGET_PROT_NORMAL, wwpn,
1207 node, efct_session_cb);
1208 if (IS_ERR(se_sess)) {
1209 efc_log_err(efct, "failed to setup session\n");
1210 kfree(wq_data);
1211 efc_scsi_sess_reg_complete(node, -EIO);
1212 return;
1213 }
1214
1215 tgt_node = node->tgt_node;
1216 id = (u64) tgt_node->port_fc_id << 32 | tgt_node->node_fc_id;
1217
1218 efc_log_debug(efct, "new initiator sess=%p node=%p id: %llx\n",
1219 se_sess, node, id);
1220
1221 if (xa_err(xa_store(&efct->lookup, id, tgt_node, GFP_KERNEL)))
1222 efc_log_err(efct, "Node lookup store failed\n");
1223
1224 efc_scsi_sess_reg_complete(node, 0);
1225
1226 /* update IO watermark: increment initiator count */
1227 ini_count = atomic_add_return(1, &efct->tgt_efct.initiator_count);
1228 watermark = efct->tgt_efct.watermark_max -
1229 ini_count * EFCT_IO_WATERMARK_PER_INITIATOR;
1230 watermark = (efct->tgt_efct.watermark_min > watermark) ?
1231 efct->tgt_efct.watermark_min : watermark;
1232 atomic_set(&efct->tgt_efct.io_high_watermark, watermark);
1233
1234 kfree(wq_data);
1235 }
1236
efct_scsi_new_initiator(struct efc * efc,struct efc_node * node)1237 int efct_scsi_new_initiator(struct efc *efc, struct efc_node *node)
1238 {
1239 struct efct *efct = node->efc->base;
1240 struct efct_lio_wq_data *wq_data;
1241
1242 /*
1243 * Since LIO only supports initiator validation at thread level,
1244 * we are open minded and accept all callers.
1245 */
1246 wq_data = kzalloc(sizeof(*wq_data), GFP_ATOMIC);
1247 if (!wq_data)
1248 return -ENOMEM;
1249
1250 wq_data->ptr = node;
1251 wq_data->efct = efct;
1252 INIT_WORK(&wq_data->work, efct_lio_setup_session);
1253 queue_work(lio_wq, &wq_data->work);
1254 return EFC_SCSI_CALL_ASYNC;
1255 }
1256
efct_lio_remove_session(struct work_struct * work)1257 static void efct_lio_remove_session(struct work_struct *work)
1258 {
1259 struct efct_lio_wq_data *wq_data =
1260 container_of(work, struct efct_lio_wq_data, work);
1261 struct efct *efct = wq_data->efct;
1262 struct efc_node *node = wq_data->ptr;
1263 struct efct_node *tgt_node;
1264 struct se_session *se_sess;
1265
1266 tgt_node = node->tgt_node;
1267 if (!tgt_node) {
1268 /* base driver has sent back-to-back requests
1269 * to unreg session with no intervening
1270 * register
1271 */
1272 efc_log_err(efct, "unreg session for NULL session\n");
1273 efc_scsi_del_initiator_complete(node->efc, node);
1274 return;
1275 }
1276
1277 se_sess = tgt_node->session;
1278 efc_log_debug(efct, "unreg session se_sess=%p node=%p\n",
1279 se_sess, node);
1280
1281 /* first flag all session commands to complete */
1282 target_stop_session(se_sess);
1283
1284 /* now wait for session commands to complete */
1285 target_wait_for_sess_cmds(se_sess);
1286 target_remove_session(se_sess);
1287 tgt_node->session = NULL;
1288 node->tgt_node = NULL;
1289 kref_put(&tgt_node->ref, tgt_node->release);
1290
1291 kfree(wq_data);
1292 }
1293
efct_scsi_del_initiator(struct efc * efc,struct efc_node * node,int reason)1294 int efct_scsi_del_initiator(struct efc *efc, struct efc_node *node, int reason)
1295 {
1296 struct efct *efct = node->efc->base;
1297 struct efct_node *tgt_node = node->tgt_node;
1298 struct efct_lio_wq_data *wq_data;
1299 int watermark;
1300 int ini_count;
1301 u64 id;
1302
1303 if (reason == EFCT_SCSI_INITIATOR_MISSING)
1304 return EFC_SCSI_CALL_COMPLETE;
1305
1306 if (!tgt_node) {
1307 efc_log_err(efct, "tgt_node is NULL\n");
1308 return -EIO;
1309 }
1310
1311 wq_data = kzalloc(sizeof(*wq_data), GFP_ATOMIC);
1312 if (!wq_data)
1313 return -ENOMEM;
1314
1315 id = (u64) tgt_node->port_fc_id << 32 | tgt_node->node_fc_id;
1316 xa_erase(&efct->lookup, id);
1317
1318 wq_data->ptr = node;
1319 wq_data->efct = efct;
1320 INIT_WORK(&wq_data->work, efct_lio_remove_session);
1321 queue_work(lio_wq, &wq_data->work);
1322
1323 /*
1324 * update IO watermark: decrement initiator count
1325 */
1326 ini_count = atomic_sub_return(1, &efct->tgt_efct.initiator_count);
1327
1328 watermark = efct->tgt_efct.watermark_max -
1329 ini_count * EFCT_IO_WATERMARK_PER_INITIATOR;
1330 watermark = (efct->tgt_efct.watermark_min > watermark) ?
1331 efct->tgt_efct.watermark_min : watermark;
1332 atomic_set(&efct->tgt_efct.io_high_watermark, watermark);
1333
1334 return EFC_SCSI_CALL_ASYNC;
1335 }
1336
efct_scsi_recv_cmd(struct efct_io * io,uint64_t lun,u8 * cdb,u32 cdb_len,u32 flags)1337 void efct_scsi_recv_cmd(struct efct_io *io, uint64_t lun, u8 *cdb,
1338 u32 cdb_len, u32 flags)
1339 {
1340 struct efct_scsi_tgt_io *ocp = &io->tgt_io;
1341 struct se_cmd *se_cmd = &io->tgt_io.cmd;
1342 struct efct *efct = io->efct;
1343 char *ddir;
1344 struct efct_node *tgt_node;
1345 struct se_session *se_sess;
1346 int rc = 0;
1347
1348 memset(ocp, 0, sizeof(struct efct_scsi_tgt_io));
1349 efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_RECV_CMD);
1350 atomic_add_return(1, &efct->tgt_efct.ios_in_use);
1351
1352 /* set target timeout */
1353 io->timeout = efct->target_io_timer_sec;
1354
1355 if (flags & EFCT_SCSI_CMD_SIMPLE)
1356 ocp->task_attr = TCM_SIMPLE_TAG;
1357 else if (flags & EFCT_SCSI_CMD_HEAD_OF_QUEUE)
1358 ocp->task_attr = TCM_HEAD_TAG;
1359 else if (flags & EFCT_SCSI_CMD_ORDERED)
1360 ocp->task_attr = TCM_ORDERED_TAG;
1361 else if (flags & EFCT_SCSI_CMD_ACA)
1362 ocp->task_attr = TCM_ACA_TAG;
1363
1364 switch (flags & (EFCT_SCSI_CMD_DIR_IN | EFCT_SCSI_CMD_DIR_OUT)) {
1365 case EFCT_SCSI_CMD_DIR_IN:
1366 ddir = "FROM_INITIATOR";
1367 ocp->ddir = DMA_TO_DEVICE;
1368 break;
1369 case EFCT_SCSI_CMD_DIR_OUT:
1370 ddir = "TO_INITIATOR";
1371 ocp->ddir = DMA_FROM_DEVICE;
1372 break;
1373 case EFCT_SCSI_CMD_DIR_IN | EFCT_SCSI_CMD_DIR_OUT:
1374 ddir = "BIDIR";
1375 ocp->ddir = DMA_BIDIRECTIONAL;
1376 break;
1377 default:
1378 ddir = "NONE";
1379 ocp->ddir = DMA_NONE;
1380 break;
1381 }
1382
1383 ocp->lun = lun;
1384 efct_lio_io_printf(io, "new cmd=0x%x ddir=%s dl=%u\n",
1385 cdb[0], ddir, io->exp_xfer_len);
1386
1387 tgt_node = io->node;
1388 se_sess = tgt_node->session;
1389 if (!se_sess) {
1390 efc_log_err(efct, "No session found to submit IO se_cmd: %p\n",
1391 &ocp->cmd);
1392 efct_scsi_io_free(io);
1393 return;
1394 }
1395
1396 efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_SUBMIT_CMD);
1397 rc = target_init_cmd(se_cmd, se_sess, &io->tgt_io.sense_buffer[0],
1398 ocp->lun, io->exp_xfer_len, ocp->task_attr,
1399 ocp->ddir, TARGET_SCF_ACK_KREF);
1400 if (rc) {
1401 efc_log_err(efct, "failed to init cmd se_cmd: %p\n", se_cmd);
1402 efct_scsi_io_free(io);
1403 return;
1404 }
1405
1406 if (target_submit_prep(se_cmd, cdb, NULL, 0, NULL, 0,
1407 NULL, 0, GFP_ATOMIC))
1408 return;
1409
1410 target_submit(se_cmd);
1411 }
1412
1413 int
efct_scsi_recv_tmf(struct efct_io * tmfio,u32 lun,enum efct_scsi_tmf_cmd cmd,struct efct_io * io_to_abort,u32 flags)1414 efct_scsi_recv_tmf(struct efct_io *tmfio, u32 lun, enum efct_scsi_tmf_cmd cmd,
1415 struct efct_io *io_to_abort, u32 flags)
1416 {
1417 unsigned char tmr_func;
1418 struct efct *efct = tmfio->efct;
1419 struct efct_scsi_tgt_io *ocp = &tmfio->tgt_io;
1420 struct efct_node *tgt_node;
1421 struct se_session *se_sess;
1422 int rc;
1423
1424 memset(ocp, 0, sizeof(struct efct_scsi_tgt_io));
1425 efct_set_lio_io_state(tmfio, EFCT_LIO_STATE_SCSI_RECV_TMF);
1426 atomic_add_return(1, &efct->tgt_efct.ios_in_use);
1427 efct_lio_tmfio_printf(tmfio, "%s: new tmf %x lun=%u\n",
1428 tmfio->display_name, cmd, lun);
1429
1430 switch (cmd) {
1431 case EFCT_SCSI_TMF_ABORT_TASK:
1432 tmr_func = TMR_ABORT_TASK;
1433 break;
1434 case EFCT_SCSI_TMF_ABORT_TASK_SET:
1435 tmr_func = TMR_ABORT_TASK_SET;
1436 break;
1437 case EFCT_SCSI_TMF_CLEAR_TASK_SET:
1438 tmr_func = TMR_CLEAR_TASK_SET;
1439 break;
1440 case EFCT_SCSI_TMF_LOGICAL_UNIT_RESET:
1441 tmr_func = TMR_LUN_RESET;
1442 break;
1443 case EFCT_SCSI_TMF_CLEAR_ACA:
1444 tmr_func = TMR_CLEAR_ACA;
1445 break;
1446 case EFCT_SCSI_TMF_TARGET_RESET:
1447 tmr_func = TMR_TARGET_WARM_RESET;
1448 break;
1449 case EFCT_SCSI_TMF_QUERY_ASYNCHRONOUS_EVENT:
1450 case EFCT_SCSI_TMF_QUERY_TASK_SET:
1451 default:
1452 goto tmf_fail;
1453 }
1454
1455 tmfio->tgt_io.tmf = tmr_func;
1456 tmfio->tgt_io.lun = lun;
1457 tmfio->tgt_io.io_to_abort = io_to_abort;
1458
1459 tgt_node = tmfio->node;
1460
1461 se_sess = tgt_node->session;
1462 if (!se_sess)
1463 return 0;
1464
1465 rc = target_submit_tmr(&ocp->cmd, se_sess, NULL, lun, ocp, tmr_func,
1466 GFP_ATOMIC, tmfio->init_task_tag, TARGET_SCF_ACK_KREF);
1467
1468 efct_set_lio_io_state(tmfio, EFCT_LIO_STATE_TGT_SUBMIT_TMR);
1469 if (rc)
1470 goto tmf_fail;
1471
1472 return 0;
1473
1474 tmf_fail:
1475 efct_scsi_send_tmf_resp(tmfio, EFCT_SCSI_TMF_FUNCTION_REJECTED,
1476 NULL, efct_lio_null_tmf_done, NULL);
1477 return 0;
1478 }
1479
1480 /* Start items for efct_lio_tpg_attrib_cit */
1481
1482 #define DEF_EFCT_TPG_ATTRIB(name) \
1483 \
1484 static ssize_t efct_lio_tpg_attrib_##name##_show( \
1485 struct config_item *item, char *page) \
1486 { \
1487 struct se_portal_group *se_tpg = to_tpg(item); \
1488 struct efct_lio_tpg *tpg = container_of(se_tpg, \
1489 struct efct_lio_tpg, tpg); \
1490 \
1491 return sprintf(page, "%u\n", tpg->tpg_attrib.name); \
1492 } \
1493 \
1494 static ssize_t efct_lio_tpg_attrib_##name##_store( \
1495 struct config_item *item, const char *page, size_t count) \
1496 { \
1497 struct se_portal_group *se_tpg = to_tpg(item); \
1498 struct efct_lio_tpg *tpg = container_of(se_tpg, \
1499 struct efct_lio_tpg, tpg); \
1500 struct efct_lio_tpg_attrib *a = &tpg->tpg_attrib; \
1501 unsigned long val; \
1502 int ret; \
1503 \
1504 ret = kstrtoul(page, 0, &val); \
1505 if (ret < 0) { \
1506 pr_err("kstrtoul() failed with ret: %d\n", ret); \
1507 return ret; \
1508 } \
1509 \
1510 if (val != 0 && val != 1) { \
1511 pr_err("Illegal boolean value %lu\n", val); \
1512 return -EINVAL; \
1513 } \
1514 \
1515 a->name = val; \
1516 \
1517 return count; \
1518 } \
1519 CONFIGFS_ATTR(efct_lio_tpg_attrib_, name)
1520
1521 DEF_EFCT_TPG_ATTRIB(generate_node_acls);
1522 DEF_EFCT_TPG_ATTRIB(cache_dynamic_acls);
1523 DEF_EFCT_TPG_ATTRIB(demo_mode_write_protect);
1524 DEF_EFCT_TPG_ATTRIB(prod_mode_write_protect);
1525 DEF_EFCT_TPG_ATTRIB(demo_mode_login_only);
1526 DEF_EFCT_TPG_ATTRIB(session_deletion_wait);
1527
1528 static struct configfs_attribute *efct_lio_tpg_attrib_attrs[] = {
1529 &efct_lio_tpg_attrib_attr_generate_node_acls,
1530 &efct_lio_tpg_attrib_attr_cache_dynamic_acls,
1531 &efct_lio_tpg_attrib_attr_demo_mode_write_protect,
1532 &efct_lio_tpg_attrib_attr_prod_mode_write_protect,
1533 &efct_lio_tpg_attrib_attr_demo_mode_login_only,
1534 &efct_lio_tpg_attrib_attr_session_deletion_wait,
1535 NULL,
1536 };
1537
1538 #define DEF_EFCT_NPIV_TPG_ATTRIB(name) \
1539 \
1540 static ssize_t efct_lio_npiv_tpg_attrib_##name##_show( \
1541 struct config_item *item, char *page) \
1542 { \
1543 struct se_portal_group *se_tpg = to_tpg(item); \
1544 struct efct_lio_tpg *tpg = container_of(se_tpg, \
1545 struct efct_lio_tpg, tpg); \
1546 \
1547 return sprintf(page, "%u\n", tpg->tpg_attrib.name); \
1548 } \
1549 \
1550 static ssize_t efct_lio_npiv_tpg_attrib_##name##_store( \
1551 struct config_item *item, const char *page, size_t count) \
1552 { \
1553 struct se_portal_group *se_tpg = to_tpg(item); \
1554 struct efct_lio_tpg *tpg = container_of(se_tpg, \
1555 struct efct_lio_tpg, tpg); \
1556 struct efct_lio_tpg_attrib *a = &tpg->tpg_attrib; \
1557 unsigned long val; \
1558 int ret; \
1559 \
1560 ret = kstrtoul(page, 0, &val); \
1561 if (ret < 0) { \
1562 pr_err("kstrtoul() failed with ret: %d\n", ret); \
1563 return ret; \
1564 } \
1565 \
1566 if (val != 0 && val != 1) { \
1567 pr_err("Illegal boolean value %lu\n", val); \
1568 return -EINVAL; \
1569 } \
1570 \
1571 a->name = val; \
1572 \
1573 return count; \
1574 } \
1575 CONFIGFS_ATTR(efct_lio_npiv_tpg_attrib_, name)
1576
1577 DEF_EFCT_NPIV_TPG_ATTRIB(generate_node_acls);
1578 DEF_EFCT_NPIV_TPG_ATTRIB(cache_dynamic_acls);
1579 DEF_EFCT_NPIV_TPG_ATTRIB(demo_mode_write_protect);
1580 DEF_EFCT_NPIV_TPG_ATTRIB(prod_mode_write_protect);
1581 DEF_EFCT_NPIV_TPG_ATTRIB(demo_mode_login_only);
1582 DEF_EFCT_NPIV_TPG_ATTRIB(session_deletion_wait);
1583
1584 static struct configfs_attribute *efct_lio_npiv_tpg_attrib_attrs[] = {
1585 &efct_lio_npiv_tpg_attrib_attr_generate_node_acls,
1586 &efct_lio_npiv_tpg_attrib_attr_cache_dynamic_acls,
1587 &efct_lio_npiv_tpg_attrib_attr_demo_mode_write_protect,
1588 &efct_lio_npiv_tpg_attrib_attr_prod_mode_write_protect,
1589 &efct_lio_npiv_tpg_attrib_attr_demo_mode_login_only,
1590 &efct_lio_npiv_tpg_attrib_attr_session_deletion_wait,
1591 NULL,
1592 };
1593
1594 CONFIGFS_ATTR(efct_lio_tpg_, enable);
1595 static struct configfs_attribute *efct_lio_tpg_attrs[] = {
1596 &efct_lio_tpg_attr_enable, NULL };
1597 CONFIGFS_ATTR(efct_lio_npiv_tpg_, enable);
1598 static struct configfs_attribute *efct_lio_npiv_tpg_attrs[] = {
1599 &efct_lio_npiv_tpg_attr_enable, NULL };
1600
1601 static const struct target_core_fabric_ops efct_lio_ops = {
1602 .module = THIS_MODULE,
1603 .fabric_name = "efct",
1604 .node_acl_size = sizeof(struct efct_lio_nacl),
1605 .max_data_sg_nents = 65535,
1606 .tpg_get_wwn = efct_lio_get_fabric_wwn,
1607 .tpg_get_tag = efct_lio_get_tag,
1608 .fabric_init_nodeacl = efct_lio_init_nodeacl,
1609 .tpg_check_demo_mode = efct_lio_check_demo_mode,
1610 .tpg_check_demo_mode_cache = efct_lio_check_demo_mode_cache,
1611 .tpg_check_demo_mode_write_protect = efct_lio_check_demo_write_protect,
1612 .tpg_check_prod_mode_write_protect = efct_lio_check_prod_write_protect,
1613 .tpg_get_inst_index = efct_lio_tpg_get_inst_index,
1614 .check_stop_free = efct_lio_check_stop_free,
1615 .aborted_task = efct_lio_aborted_task,
1616 .release_cmd = efct_lio_release_cmd,
1617 .close_session = efct_lio_close_session,
1618 .sess_get_index = efct_lio_sess_get_index,
1619 .write_pending = efct_lio_write_pending,
1620 .set_default_node_attributes = efct_lio_set_default_node_attrs,
1621 .get_cmd_state = efct_lio_get_cmd_state,
1622 .queue_data_in = efct_lio_queue_data_in,
1623 .queue_status = efct_lio_queue_status,
1624 .queue_tm_rsp = efct_lio_queue_tm_rsp,
1625 .fabric_make_wwn = efct_lio_make_nport,
1626 .fabric_drop_wwn = efct_lio_drop_nport,
1627 .fabric_make_tpg = efct_lio_make_tpg,
1628 .fabric_drop_tpg = efct_lio_drop_tpg,
1629 .tpg_check_demo_mode_login_only = efct_lio_check_demo_mode_login_only,
1630 .tpg_check_prot_fabric_only = NULL,
1631 .sess_get_initiator_sid = NULL,
1632 .tfc_tpg_base_attrs = efct_lio_tpg_attrs,
1633 .tfc_tpg_attrib_attrs = efct_lio_tpg_attrib_attrs,
1634 };
1635
1636 static const struct target_core_fabric_ops efct_lio_npiv_ops = {
1637 .module = THIS_MODULE,
1638 .fabric_name = "efct_npiv",
1639 .node_acl_size = sizeof(struct efct_lio_nacl),
1640 .max_data_sg_nents = 65535,
1641 .tpg_get_wwn = efct_lio_get_npiv_fabric_wwn,
1642 .tpg_get_tag = efct_lio_get_npiv_tag,
1643 .fabric_init_nodeacl = efct_lio_init_nodeacl,
1644 .tpg_check_demo_mode = efct_lio_check_demo_mode,
1645 .tpg_check_demo_mode_cache = efct_lio_check_demo_mode_cache,
1646 .tpg_check_demo_mode_write_protect =
1647 efct_lio_npiv_check_demo_write_protect,
1648 .tpg_check_prod_mode_write_protect =
1649 efct_lio_npiv_check_prod_write_protect,
1650 .tpg_get_inst_index = efct_lio_tpg_get_inst_index,
1651 .check_stop_free = efct_lio_check_stop_free,
1652 .aborted_task = efct_lio_aborted_task,
1653 .release_cmd = efct_lio_release_cmd,
1654 .close_session = efct_lio_close_session,
1655 .sess_get_index = efct_lio_sess_get_index,
1656 .write_pending = efct_lio_write_pending,
1657 .set_default_node_attributes = efct_lio_set_default_node_attrs,
1658 .get_cmd_state = efct_lio_get_cmd_state,
1659 .queue_data_in = efct_lio_queue_data_in,
1660 .queue_status = efct_lio_queue_status,
1661 .queue_tm_rsp = efct_lio_queue_tm_rsp,
1662 .fabric_make_wwn = efct_lio_npiv_make_nport,
1663 .fabric_drop_wwn = efct_lio_npiv_drop_nport,
1664 .fabric_make_tpg = efct_lio_npiv_make_tpg,
1665 .fabric_drop_tpg = efct_lio_npiv_drop_tpg,
1666 .tpg_check_demo_mode_login_only =
1667 efct_lio_npiv_check_demo_mode_login_only,
1668 .tpg_check_prot_fabric_only = NULL,
1669 .sess_get_initiator_sid = NULL,
1670 .tfc_tpg_base_attrs = efct_lio_npiv_tpg_attrs,
1671 .tfc_tpg_attrib_attrs = efct_lio_npiv_tpg_attrib_attrs,
1672 };
1673
efct_scsi_tgt_driver_init(void)1674 int efct_scsi_tgt_driver_init(void)
1675 {
1676 int rc;
1677
1678 /* Register the top level struct config_item_type with TCM core */
1679 rc = target_register_template(&efct_lio_ops);
1680 if (rc < 0) {
1681 pr_err("target_fabric_configfs_register failed with %d\n", rc);
1682 return rc;
1683 }
1684 rc = target_register_template(&efct_lio_npiv_ops);
1685 if (rc < 0) {
1686 pr_err("target_fabric_configfs_register failed with %d\n", rc);
1687 target_unregister_template(&efct_lio_ops);
1688 return rc;
1689 }
1690 return 0;
1691 }
1692
efct_scsi_tgt_driver_exit(void)1693 int efct_scsi_tgt_driver_exit(void)
1694 {
1695 target_unregister_template(&efct_lio_ops);
1696 target_unregister_template(&efct_lio_npiv_ops);
1697 return 0;
1698 }
1699