1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3 * QLogic Fibre Channel HBA Driver
4 * Copyright (c) 2003-2014 QLogic Corporation
5 */
6 #include "qla_def.h"
7 #include "qla_target.h"
8
9 #include <linux/kthread.h>
10 #include <linux/vmalloc.h>
11 #include <linux/slab.h>
12 #include <linux/delay.h>
13
14 static int qla24xx_vport_disable(struct fc_vport *, bool);
15
16 /* SYSFS attributes --------------------------------------------------------- */
17
18 static ssize_t
qla2x00_sysfs_read_fw_dump(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)19 qla2x00_sysfs_read_fw_dump(struct file *filp, struct kobject *kobj,
20 struct bin_attribute *bin_attr,
21 char *buf, loff_t off, size_t count)
22 {
23 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
24 struct device, kobj)));
25 struct qla_hw_data *ha = vha->hw;
26 int rval = 0;
27
28 if (!(ha->fw_dump_reading || ha->mctp_dump_reading ||
29 ha->mpi_fw_dump_reading))
30 return 0;
31
32 mutex_lock(&ha->optrom_mutex);
33 if (IS_P3P_TYPE(ha)) {
34 if (off < ha->md_template_size) {
35 rval = memory_read_from_buffer(buf, count,
36 &off, ha->md_tmplt_hdr, ha->md_template_size);
37 } else {
38 off -= ha->md_template_size;
39 rval = memory_read_from_buffer(buf, count,
40 &off, ha->md_dump, ha->md_dump_size);
41 }
42 } else if (ha->mctp_dumped && ha->mctp_dump_reading) {
43 rval = memory_read_from_buffer(buf, count, &off, ha->mctp_dump,
44 MCTP_DUMP_SIZE);
45 } else if (ha->mpi_fw_dumped && ha->mpi_fw_dump_reading) {
46 rval = memory_read_from_buffer(buf, count, &off,
47 ha->mpi_fw_dump,
48 ha->mpi_fw_dump_len);
49 } else if (ha->fw_dump_reading) {
50 rval = memory_read_from_buffer(buf, count, &off, ha->fw_dump,
51 ha->fw_dump_len);
52 } else {
53 rval = 0;
54 }
55 mutex_unlock(&ha->optrom_mutex);
56 return rval;
57 }
58
59 static ssize_t
qla2x00_sysfs_write_fw_dump(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)60 qla2x00_sysfs_write_fw_dump(struct file *filp, struct kobject *kobj,
61 struct bin_attribute *bin_attr,
62 char *buf, loff_t off, size_t count)
63 {
64 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
65 struct device, kobj)));
66 struct qla_hw_data *ha = vha->hw;
67 int reading;
68
69 if (off != 0)
70 return (0);
71
72 reading = simple_strtol(buf, NULL, 10);
73 switch (reading) {
74 case 0:
75 if (!ha->fw_dump_reading)
76 break;
77
78 ql_log(ql_log_info, vha, 0x705d,
79 "Firmware dump cleared on (%ld).\n", vha->host_no);
80
81 if (IS_P3P_TYPE(ha)) {
82 qla82xx_md_free(vha);
83 qla82xx_md_prep(vha);
84 }
85 ha->fw_dump_reading = 0;
86 ha->fw_dumped = false;
87 break;
88 case 1:
89 if (ha->fw_dumped && !ha->fw_dump_reading) {
90 ha->fw_dump_reading = 1;
91
92 ql_log(ql_log_info, vha, 0x705e,
93 "Raw firmware dump ready for read on (%ld).\n",
94 vha->host_no);
95 }
96 break;
97 case 2:
98 qla2x00_alloc_fw_dump(vha);
99 break;
100 case 3:
101 if (IS_QLA82XX(ha)) {
102 qla82xx_idc_lock(ha);
103 qla82xx_set_reset_owner(vha);
104 qla82xx_idc_unlock(ha);
105 } else if (IS_QLA8044(ha)) {
106 qla8044_idc_lock(ha);
107 qla82xx_set_reset_owner(vha);
108 qla8044_idc_unlock(ha);
109 } else {
110 qla2x00_system_error(vha);
111 }
112 break;
113 case 4:
114 if (IS_P3P_TYPE(ha)) {
115 if (ha->md_tmplt_hdr)
116 ql_dbg(ql_dbg_user, vha, 0x705b,
117 "MiniDump supported with this firmware.\n");
118 else
119 ql_dbg(ql_dbg_user, vha, 0x709d,
120 "MiniDump not supported with this firmware.\n");
121 }
122 break;
123 case 5:
124 if (IS_P3P_TYPE(ha))
125 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
126 break;
127 case 6:
128 if (!ha->mctp_dump_reading)
129 break;
130 ql_log(ql_log_info, vha, 0x70c1,
131 "MCTP dump cleared on (%ld).\n", vha->host_no);
132 ha->mctp_dump_reading = 0;
133 ha->mctp_dumped = 0;
134 break;
135 case 7:
136 if (ha->mctp_dumped && !ha->mctp_dump_reading) {
137 ha->mctp_dump_reading = 1;
138 ql_log(ql_log_info, vha, 0x70c2,
139 "Raw mctp dump ready for read on (%ld).\n",
140 vha->host_no);
141 }
142 break;
143 case 8:
144 if (!ha->mpi_fw_dump_reading)
145 break;
146 ql_log(ql_log_info, vha, 0x70e7,
147 "MPI firmware dump cleared on (%ld).\n", vha->host_no);
148 ha->mpi_fw_dump_reading = 0;
149 ha->mpi_fw_dumped = 0;
150 break;
151 case 9:
152 if (ha->mpi_fw_dumped && !ha->mpi_fw_dump_reading) {
153 ha->mpi_fw_dump_reading = 1;
154 ql_log(ql_log_info, vha, 0x70e8,
155 "Raw MPI firmware dump ready for read on (%ld).\n",
156 vha->host_no);
157 }
158 break;
159 case 10:
160 if (IS_QLA27XX(ha) || IS_QLA28XX(ha)) {
161 ql_log(ql_log_info, vha, 0x70e9,
162 "Issuing MPI firmware dump on host#%ld.\n",
163 vha->host_no);
164 ha->isp_ops->mpi_fw_dump(vha, 0);
165 }
166 break;
167 }
168 return count;
169 }
170
171 static struct bin_attribute sysfs_fw_dump_attr = {
172 .attr = {
173 .name = "fw_dump",
174 .mode = S_IRUSR | S_IWUSR,
175 },
176 .size = 0,
177 .read = qla2x00_sysfs_read_fw_dump,
178 .write = qla2x00_sysfs_write_fw_dump,
179 };
180
181 static ssize_t
qla2x00_sysfs_read_nvram(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)182 qla2x00_sysfs_read_nvram(struct file *filp, struct kobject *kobj,
183 struct bin_attribute *bin_attr,
184 char *buf, loff_t off, size_t count)
185 {
186 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
187 struct device, kobj)));
188 struct qla_hw_data *ha = vha->hw;
189 uint32_t faddr;
190 struct active_regions active_regions = { };
191
192 if (!capable(CAP_SYS_ADMIN))
193 return 0;
194
195 mutex_lock(&ha->optrom_mutex);
196 if (qla2x00_chip_is_down(vha)) {
197 mutex_unlock(&ha->optrom_mutex);
198 return -EAGAIN;
199 }
200
201 if (!IS_NOCACHE_VPD_TYPE(ha)) {
202 mutex_unlock(&ha->optrom_mutex);
203 goto skip;
204 }
205
206 faddr = ha->flt_region_nvram;
207 if (IS_QLA28XX(ha)) {
208 qla28xx_get_aux_images(vha, &active_regions);
209 if (active_regions.aux.vpd_nvram == QLA27XX_SECONDARY_IMAGE)
210 faddr = ha->flt_region_nvram_sec;
211 }
212 ha->isp_ops->read_optrom(vha, ha->nvram, faddr << 2, ha->nvram_size);
213
214 mutex_unlock(&ha->optrom_mutex);
215
216 skip:
217 return memory_read_from_buffer(buf, count, &off, ha->nvram,
218 ha->nvram_size);
219 }
220
221 static ssize_t
qla2x00_sysfs_write_nvram(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)222 qla2x00_sysfs_write_nvram(struct file *filp, struct kobject *kobj,
223 struct bin_attribute *bin_attr,
224 char *buf, loff_t off, size_t count)
225 {
226 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
227 struct device, kobj)));
228 struct qla_hw_data *ha = vha->hw;
229 uint16_t cnt;
230
231 if (!capable(CAP_SYS_ADMIN) || off != 0 || count != ha->nvram_size ||
232 !ha->isp_ops->write_nvram)
233 return -EINVAL;
234
235 /* Checksum NVRAM. */
236 if (IS_FWI2_CAPABLE(ha)) {
237 __le32 *iter = (__force __le32 *)buf;
238 uint32_t chksum;
239
240 chksum = 0;
241 for (cnt = 0; cnt < ((count >> 2) - 1); cnt++, iter++)
242 chksum += le32_to_cpu(*iter);
243 chksum = ~chksum + 1;
244 *iter = cpu_to_le32(chksum);
245 } else {
246 uint8_t *iter;
247 uint8_t chksum;
248
249 iter = (uint8_t *)buf;
250 chksum = 0;
251 for (cnt = 0; cnt < count - 1; cnt++)
252 chksum += *iter++;
253 chksum = ~chksum + 1;
254 *iter = chksum;
255 }
256
257 if (qla2x00_wait_for_hba_online(vha) != QLA_SUCCESS) {
258 ql_log(ql_log_warn, vha, 0x705f,
259 "HBA not online, failing NVRAM update.\n");
260 return -EAGAIN;
261 }
262
263 mutex_lock(&ha->optrom_mutex);
264 if (qla2x00_chip_is_down(vha)) {
265 mutex_unlock(&ha->optrom_mutex);
266 return -EAGAIN;
267 }
268
269 /* Write NVRAM. */
270 ha->isp_ops->write_nvram(vha, buf, ha->nvram_base, count);
271 ha->isp_ops->read_nvram(vha, ha->nvram, ha->nvram_base,
272 count);
273 mutex_unlock(&ha->optrom_mutex);
274
275 ql_dbg(ql_dbg_user, vha, 0x7060,
276 "Setting ISP_ABORT_NEEDED\n");
277 /* NVRAM settings take effect immediately. */
278 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
279 qla2xxx_wake_dpc(vha);
280 qla2x00_wait_for_chip_reset(vha);
281
282 return count;
283 }
284
285 static struct bin_attribute sysfs_nvram_attr = {
286 .attr = {
287 .name = "nvram",
288 .mode = S_IRUSR | S_IWUSR,
289 },
290 .size = 512,
291 .read = qla2x00_sysfs_read_nvram,
292 .write = qla2x00_sysfs_write_nvram,
293 };
294
295 static ssize_t
qla2x00_sysfs_read_optrom(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)296 qla2x00_sysfs_read_optrom(struct file *filp, struct kobject *kobj,
297 struct bin_attribute *bin_attr,
298 char *buf, loff_t off, size_t count)
299 {
300 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
301 struct device, kobj)));
302 struct qla_hw_data *ha = vha->hw;
303 ssize_t rval = 0;
304
305 mutex_lock(&ha->optrom_mutex);
306
307 if (ha->optrom_state != QLA_SREADING)
308 goto out;
309
310 rval = memory_read_from_buffer(buf, count, &off, ha->optrom_buffer,
311 ha->optrom_region_size);
312
313 out:
314 mutex_unlock(&ha->optrom_mutex);
315
316 return rval;
317 }
318
319 static ssize_t
qla2x00_sysfs_write_optrom(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)320 qla2x00_sysfs_write_optrom(struct file *filp, struct kobject *kobj,
321 struct bin_attribute *bin_attr,
322 char *buf, loff_t off, size_t count)
323 {
324 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
325 struct device, kobj)));
326 struct qla_hw_data *ha = vha->hw;
327
328 mutex_lock(&ha->optrom_mutex);
329
330 if (ha->optrom_state != QLA_SWRITING) {
331 mutex_unlock(&ha->optrom_mutex);
332 return -EINVAL;
333 }
334 if (off > ha->optrom_region_size) {
335 mutex_unlock(&ha->optrom_mutex);
336 return -ERANGE;
337 }
338 if (off + count > ha->optrom_region_size)
339 count = ha->optrom_region_size - off;
340
341 memcpy(&ha->optrom_buffer[off], buf, count);
342 mutex_unlock(&ha->optrom_mutex);
343
344 return count;
345 }
346
347 static struct bin_attribute sysfs_optrom_attr = {
348 .attr = {
349 .name = "optrom",
350 .mode = S_IRUSR | S_IWUSR,
351 },
352 .size = 0,
353 .read = qla2x00_sysfs_read_optrom,
354 .write = qla2x00_sysfs_write_optrom,
355 };
356
357 static ssize_t
qla2x00_sysfs_write_optrom_ctl(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)358 qla2x00_sysfs_write_optrom_ctl(struct file *filp, struct kobject *kobj,
359 struct bin_attribute *bin_attr,
360 char *buf, loff_t off, size_t count)
361 {
362 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
363 struct device, kobj)));
364 struct qla_hw_data *ha = vha->hw;
365 uint32_t start = 0;
366 uint32_t size = ha->optrom_size;
367 int val, valid;
368 ssize_t rval = count;
369
370 if (off)
371 return -EINVAL;
372
373 if (unlikely(pci_channel_offline(ha->pdev)))
374 return -EAGAIN;
375
376 if (sscanf(buf, "%d:%x:%x", &val, &start, &size) < 1)
377 return -EINVAL;
378 if (start > ha->optrom_size)
379 return -EINVAL;
380 if (size > ha->optrom_size - start)
381 size = ha->optrom_size - start;
382
383 mutex_lock(&ha->optrom_mutex);
384 if (qla2x00_chip_is_down(vha)) {
385 mutex_unlock(&ha->optrom_mutex);
386 return -EAGAIN;
387 }
388 switch (val) {
389 case 0:
390 if (ha->optrom_state != QLA_SREADING &&
391 ha->optrom_state != QLA_SWRITING) {
392 rval = -EINVAL;
393 goto out;
394 }
395 ha->optrom_state = QLA_SWAITING;
396
397 ql_dbg(ql_dbg_user, vha, 0x7061,
398 "Freeing flash region allocation -- 0x%x bytes.\n",
399 ha->optrom_region_size);
400
401 vfree(ha->optrom_buffer);
402 ha->optrom_buffer = NULL;
403 break;
404 case 1:
405 if (ha->optrom_state != QLA_SWAITING) {
406 rval = -EINVAL;
407 goto out;
408 }
409
410 ha->optrom_region_start = start;
411 ha->optrom_region_size = size;
412
413 ha->optrom_state = QLA_SREADING;
414 ha->optrom_buffer = vzalloc(ha->optrom_region_size);
415 if (ha->optrom_buffer == NULL) {
416 ql_log(ql_log_warn, vha, 0x7062,
417 "Unable to allocate memory for optrom retrieval "
418 "(%x).\n", ha->optrom_region_size);
419
420 ha->optrom_state = QLA_SWAITING;
421 rval = -ENOMEM;
422 goto out;
423 }
424
425 if (qla2x00_wait_for_hba_online(vha) != QLA_SUCCESS) {
426 ql_log(ql_log_warn, vha, 0x7063,
427 "HBA not online, failing NVRAM update.\n");
428 rval = -EAGAIN;
429 goto out;
430 }
431
432 ql_dbg(ql_dbg_user, vha, 0x7064,
433 "Reading flash region -- 0x%x/0x%x.\n",
434 ha->optrom_region_start, ha->optrom_region_size);
435
436 ha->isp_ops->read_optrom(vha, ha->optrom_buffer,
437 ha->optrom_region_start, ha->optrom_region_size);
438 break;
439 case 2:
440 if (ha->optrom_state != QLA_SWAITING) {
441 rval = -EINVAL;
442 goto out;
443 }
444
445 /*
446 * We need to be more restrictive on which FLASH regions are
447 * allowed to be updated via user-space. Regions accessible
448 * via this method include:
449 *
450 * ISP21xx/ISP22xx/ISP23xx type boards:
451 *
452 * 0x000000 -> 0x020000 -- Boot code.
453 *
454 * ISP2322/ISP24xx type boards:
455 *
456 * 0x000000 -> 0x07ffff -- Boot code.
457 * 0x080000 -> 0x0fffff -- Firmware.
458 *
459 * ISP25xx type boards:
460 *
461 * 0x000000 -> 0x07ffff -- Boot code.
462 * 0x080000 -> 0x0fffff -- Firmware.
463 * 0x120000 -> 0x12ffff -- VPD and HBA parameters.
464 *
465 * > ISP25xx type boards:
466 *
467 * None -- should go through BSG.
468 */
469 valid = 0;
470 if (ha->optrom_size == OPTROM_SIZE_2300 && start == 0)
471 valid = 1;
472 else if (IS_QLA24XX_TYPE(ha) || IS_QLA25XX(ha))
473 valid = 1;
474 if (!valid) {
475 ql_log(ql_log_warn, vha, 0x7065,
476 "Invalid start region 0x%x/0x%x.\n", start, size);
477 rval = -EINVAL;
478 goto out;
479 }
480
481 ha->optrom_region_start = start;
482 ha->optrom_region_size = size;
483
484 ha->optrom_state = QLA_SWRITING;
485 ha->optrom_buffer = vzalloc(ha->optrom_region_size);
486 if (ha->optrom_buffer == NULL) {
487 ql_log(ql_log_warn, vha, 0x7066,
488 "Unable to allocate memory for optrom update "
489 "(%x)\n", ha->optrom_region_size);
490
491 ha->optrom_state = QLA_SWAITING;
492 rval = -ENOMEM;
493 goto out;
494 }
495
496 ql_dbg(ql_dbg_user, vha, 0x7067,
497 "Staging flash region write -- 0x%x/0x%x.\n",
498 ha->optrom_region_start, ha->optrom_region_size);
499
500 break;
501 case 3:
502 if (ha->optrom_state != QLA_SWRITING) {
503 rval = -EINVAL;
504 goto out;
505 }
506
507 if (qla2x00_wait_for_hba_online(vha) != QLA_SUCCESS) {
508 ql_log(ql_log_warn, vha, 0x7068,
509 "HBA not online, failing flash update.\n");
510 rval = -EAGAIN;
511 goto out;
512 }
513
514 ql_dbg(ql_dbg_user, vha, 0x7069,
515 "Writing flash region -- 0x%x/0x%x.\n",
516 ha->optrom_region_start, ha->optrom_region_size);
517
518 rval = ha->isp_ops->write_optrom(vha, ha->optrom_buffer,
519 ha->optrom_region_start, ha->optrom_region_size);
520 if (rval)
521 rval = -EIO;
522 break;
523 default:
524 rval = -EINVAL;
525 }
526
527 out:
528 mutex_unlock(&ha->optrom_mutex);
529 return rval;
530 }
531
532 static struct bin_attribute sysfs_optrom_ctl_attr = {
533 .attr = {
534 .name = "optrom_ctl",
535 .mode = S_IWUSR,
536 },
537 .size = 0,
538 .write = qla2x00_sysfs_write_optrom_ctl,
539 };
540
541 static ssize_t
qla2x00_sysfs_read_vpd(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)542 qla2x00_sysfs_read_vpd(struct file *filp, struct kobject *kobj,
543 struct bin_attribute *bin_attr,
544 char *buf, loff_t off, size_t count)
545 {
546 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
547 struct device, kobj)));
548 struct qla_hw_data *ha = vha->hw;
549 uint32_t faddr;
550 struct active_regions active_regions = { };
551
552 if (unlikely(pci_channel_offline(ha->pdev)))
553 return -EAGAIN;
554
555 if (!capable(CAP_SYS_ADMIN))
556 return -EINVAL;
557
558 if (!IS_NOCACHE_VPD_TYPE(ha))
559 goto skip;
560
561 faddr = ha->flt_region_vpd << 2;
562
563 if (IS_QLA28XX(ha)) {
564 qla28xx_get_aux_images(vha, &active_regions);
565 if (active_regions.aux.vpd_nvram == QLA27XX_SECONDARY_IMAGE)
566 faddr = ha->flt_region_vpd_sec << 2;
567
568 ql_dbg(ql_dbg_init, vha, 0x7070,
569 "Loading %s nvram image.\n",
570 active_regions.aux.vpd_nvram == QLA27XX_PRIMARY_IMAGE ?
571 "primary" : "secondary");
572 }
573
574 mutex_lock(&ha->optrom_mutex);
575 if (qla2x00_chip_is_down(vha)) {
576 mutex_unlock(&ha->optrom_mutex);
577 return -EAGAIN;
578 }
579
580 ha->isp_ops->read_optrom(vha, ha->vpd, faddr, ha->vpd_size);
581 mutex_unlock(&ha->optrom_mutex);
582
583 ha->isp_ops->read_optrom(vha, ha->vpd, faddr, ha->vpd_size);
584 skip:
585 return memory_read_from_buffer(buf, count, &off, ha->vpd, ha->vpd_size);
586 }
587
588 static ssize_t
qla2x00_sysfs_write_vpd(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)589 qla2x00_sysfs_write_vpd(struct file *filp, struct kobject *kobj,
590 struct bin_attribute *bin_attr,
591 char *buf, loff_t off, size_t count)
592 {
593 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
594 struct device, kobj)));
595 struct qla_hw_data *ha = vha->hw;
596 uint8_t *tmp_data;
597
598 if (unlikely(pci_channel_offline(ha->pdev)))
599 return 0;
600
601 if (qla2x00_chip_is_down(vha))
602 return 0;
603
604 if (!capable(CAP_SYS_ADMIN) || off != 0 || count != ha->vpd_size ||
605 !ha->isp_ops->write_nvram)
606 return 0;
607
608 if (qla2x00_wait_for_hba_online(vha) != QLA_SUCCESS) {
609 ql_log(ql_log_warn, vha, 0x706a,
610 "HBA not online, failing VPD update.\n");
611 return -EAGAIN;
612 }
613
614 mutex_lock(&ha->optrom_mutex);
615 if (qla2x00_chip_is_down(vha)) {
616 mutex_unlock(&ha->optrom_mutex);
617 return -EAGAIN;
618 }
619
620 /* Write NVRAM. */
621 ha->isp_ops->write_nvram(vha, buf, ha->vpd_base, count);
622 ha->isp_ops->read_nvram(vha, ha->vpd, ha->vpd_base, count);
623
624 /* Update flash version information for 4Gb & above. */
625 if (!IS_FWI2_CAPABLE(ha)) {
626 mutex_unlock(&ha->optrom_mutex);
627 return -EINVAL;
628 }
629
630 tmp_data = vmalloc(256);
631 if (!tmp_data) {
632 mutex_unlock(&ha->optrom_mutex);
633 ql_log(ql_log_warn, vha, 0x706b,
634 "Unable to allocate memory for VPD information update.\n");
635 return -ENOMEM;
636 }
637 ha->isp_ops->get_flash_version(vha, tmp_data);
638 vfree(tmp_data);
639
640 mutex_unlock(&ha->optrom_mutex);
641
642 return count;
643 }
644
645 static struct bin_attribute sysfs_vpd_attr = {
646 .attr = {
647 .name = "vpd",
648 .mode = S_IRUSR | S_IWUSR,
649 },
650 .size = 0,
651 .read = qla2x00_sysfs_read_vpd,
652 .write = qla2x00_sysfs_write_vpd,
653 };
654
655 static ssize_t
qla2x00_sysfs_read_sfp(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)656 qla2x00_sysfs_read_sfp(struct file *filp, struct kobject *kobj,
657 struct bin_attribute *bin_attr,
658 char *buf, loff_t off, size_t count)
659 {
660 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
661 struct device, kobj)));
662 int rval;
663
664 if (!capable(CAP_SYS_ADMIN) || off != 0 || count < SFP_DEV_SIZE)
665 return 0;
666
667 mutex_lock(&vha->hw->optrom_mutex);
668 if (qla2x00_chip_is_down(vha)) {
669 mutex_unlock(&vha->hw->optrom_mutex);
670 return 0;
671 }
672
673 rval = qla2x00_read_sfp_dev(vha, buf, count);
674 mutex_unlock(&vha->hw->optrom_mutex);
675
676 if (rval)
677 return -EIO;
678
679 return count;
680 }
681
682 static struct bin_attribute sysfs_sfp_attr = {
683 .attr = {
684 .name = "sfp",
685 .mode = S_IRUSR | S_IWUSR,
686 },
687 .size = SFP_DEV_SIZE,
688 .read = qla2x00_sysfs_read_sfp,
689 };
690
691 static ssize_t
qla2x00_sysfs_write_reset(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)692 qla2x00_sysfs_write_reset(struct file *filp, struct kobject *kobj,
693 struct bin_attribute *bin_attr,
694 char *buf, loff_t off, size_t count)
695 {
696 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
697 struct device, kobj)));
698 struct qla_hw_data *ha = vha->hw;
699 struct scsi_qla_host *base_vha = pci_get_drvdata(ha->pdev);
700 int type;
701 uint32_t idc_control;
702 uint8_t *tmp_data = NULL;
703
704 if (off != 0)
705 return -EINVAL;
706
707 type = simple_strtol(buf, NULL, 10);
708 switch (type) {
709 case 0x2025c:
710 ql_log(ql_log_info, vha, 0x706e,
711 "Issuing ISP reset.\n");
712
713 if (vha->hw->flags.port_isolated) {
714 ql_log(ql_log_info, vha, 0x706e,
715 "Port is isolated, returning.\n");
716 return -EINVAL;
717 }
718
719 scsi_block_requests(vha->host);
720 if (IS_QLA82XX(ha)) {
721 ha->flags.isp82xx_no_md_cap = 1;
722 qla82xx_idc_lock(ha);
723 qla82xx_set_reset_owner(vha);
724 qla82xx_idc_unlock(ha);
725 } else if (IS_QLA8044(ha)) {
726 qla8044_idc_lock(ha);
727 idc_control = qla8044_rd_reg(ha,
728 QLA8044_IDC_DRV_CTRL);
729 qla8044_wr_reg(ha, QLA8044_IDC_DRV_CTRL,
730 (idc_control | GRACEFUL_RESET_BIT1));
731 qla82xx_set_reset_owner(vha);
732 qla8044_idc_unlock(ha);
733 } else {
734 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
735 qla2xxx_wake_dpc(vha);
736 }
737 qla2x00_wait_for_chip_reset(vha);
738 scsi_unblock_requests(vha->host);
739 break;
740 case 0x2025d:
741 if (!IS_QLA81XX(ha) && !IS_QLA83XX(ha) &&
742 !IS_QLA27XX(ha) && !IS_QLA28XX(ha))
743 return -EPERM;
744
745 ql_log(ql_log_info, vha, 0x706f,
746 "Issuing MPI reset.\n");
747
748 if (IS_QLA83XX(ha)) {
749 uint32_t idc_control;
750
751 qla83xx_idc_lock(vha, 0);
752 __qla83xx_get_idc_control(vha, &idc_control);
753 idc_control |= QLA83XX_IDC_GRACEFUL_RESET;
754 __qla83xx_set_idc_control(vha, idc_control);
755 qla83xx_wr_reg(vha, QLA83XX_IDC_DEV_STATE,
756 QLA8XXX_DEV_NEED_RESET);
757 qla83xx_idc_audit(vha, IDC_AUDIT_TIMESTAMP);
758 qla83xx_idc_unlock(vha, 0);
759 break;
760 } else {
761 /* Make sure FC side is not in reset */
762 WARN_ON_ONCE(qla2x00_wait_for_hba_online(vha) !=
763 QLA_SUCCESS);
764
765 /* Issue MPI reset */
766 scsi_block_requests(vha->host);
767 if (qla81xx_restart_mpi_firmware(vha) != QLA_SUCCESS)
768 ql_log(ql_log_warn, vha, 0x7070,
769 "MPI reset failed.\n");
770 scsi_unblock_requests(vha->host);
771 break;
772 }
773 break;
774 case 0x2025e:
775 if (!IS_P3P_TYPE(ha) || vha != base_vha) {
776 ql_log(ql_log_info, vha, 0x7071,
777 "FCoE ctx reset not supported.\n");
778 return -EPERM;
779 }
780
781 ql_log(ql_log_info, vha, 0x7072,
782 "Issuing FCoE ctx reset.\n");
783 set_bit(FCOE_CTX_RESET_NEEDED, &vha->dpc_flags);
784 qla2xxx_wake_dpc(vha);
785 qla2x00_wait_for_fcoe_ctx_reset(vha);
786 break;
787 case 0x2025f:
788 if (!IS_QLA8031(ha))
789 return -EPERM;
790 ql_log(ql_log_info, vha, 0x70bc,
791 "Disabling Reset by IDC control\n");
792 qla83xx_idc_lock(vha, 0);
793 __qla83xx_get_idc_control(vha, &idc_control);
794 idc_control |= QLA83XX_IDC_RESET_DISABLED;
795 __qla83xx_set_idc_control(vha, idc_control);
796 qla83xx_idc_unlock(vha, 0);
797 break;
798 case 0x20260:
799 if (!IS_QLA8031(ha))
800 return -EPERM;
801 ql_log(ql_log_info, vha, 0x70bd,
802 "Enabling Reset by IDC control\n");
803 qla83xx_idc_lock(vha, 0);
804 __qla83xx_get_idc_control(vha, &idc_control);
805 idc_control &= ~QLA83XX_IDC_RESET_DISABLED;
806 __qla83xx_set_idc_control(vha, idc_control);
807 qla83xx_idc_unlock(vha, 0);
808 break;
809 case 0x20261:
810 ql_dbg(ql_dbg_user, vha, 0x70e0,
811 "Updating cache versions without reset ");
812
813 tmp_data = vmalloc(256);
814 if (!tmp_data) {
815 ql_log(ql_log_warn, vha, 0x70e1,
816 "Unable to allocate memory for VPD information update.\n");
817 return -ENOMEM;
818 }
819 ha->isp_ops->get_flash_version(vha, tmp_data);
820 vfree(tmp_data);
821 break;
822 }
823 return count;
824 }
825
826 static struct bin_attribute sysfs_reset_attr = {
827 .attr = {
828 .name = "reset",
829 .mode = S_IWUSR,
830 },
831 .size = 0,
832 .write = qla2x00_sysfs_write_reset,
833 };
834
835 static ssize_t
qla2x00_issue_logo(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)836 qla2x00_issue_logo(struct file *filp, struct kobject *kobj,
837 struct bin_attribute *bin_attr,
838 char *buf, loff_t off, size_t count)
839 {
840 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
841 struct device, kobj)));
842 int type;
843 port_id_t did;
844
845 if (!capable(CAP_SYS_ADMIN))
846 return 0;
847
848 if (unlikely(pci_channel_offline(vha->hw->pdev)))
849 return 0;
850
851 if (qla2x00_chip_is_down(vha))
852 return 0;
853
854 type = simple_strtol(buf, NULL, 10);
855
856 did.b.domain = (type & 0x00ff0000) >> 16;
857 did.b.area = (type & 0x0000ff00) >> 8;
858 did.b.al_pa = (type & 0x000000ff);
859
860 ql_log(ql_log_info, vha, 0xd04d, "portid=%02x%02x%02x done\n",
861 did.b.domain, did.b.area, did.b.al_pa);
862
863 ql_log(ql_log_info, vha, 0x70e4, "%s: %d\n", __func__, type);
864
865 qla24xx_els_dcmd_iocb(vha, ELS_DCMD_LOGO, did);
866 return count;
867 }
868
869 static struct bin_attribute sysfs_issue_logo_attr = {
870 .attr = {
871 .name = "issue_logo",
872 .mode = S_IWUSR,
873 },
874 .size = 0,
875 .write = qla2x00_issue_logo,
876 };
877
878 static ssize_t
qla2x00_sysfs_read_xgmac_stats(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)879 qla2x00_sysfs_read_xgmac_stats(struct file *filp, struct kobject *kobj,
880 struct bin_attribute *bin_attr,
881 char *buf, loff_t off, size_t count)
882 {
883 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
884 struct device, kobj)));
885 struct qla_hw_data *ha = vha->hw;
886 int rval;
887 uint16_t actual_size;
888
889 if (!capable(CAP_SYS_ADMIN) || off != 0 || count > XGMAC_DATA_SIZE)
890 return 0;
891
892 if (unlikely(pci_channel_offline(ha->pdev)))
893 return 0;
894 mutex_lock(&vha->hw->optrom_mutex);
895 if (qla2x00_chip_is_down(vha)) {
896 mutex_unlock(&vha->hw->optrom_mutex);
897 return 0;
898 }
899
900 if (ha->xgmac_data)
901 goto do_read;
902
903 ha->xgmac_data = dma_alloc_coherent(&ha->pdev->dev, XGMAC_DATA_SIZE,
904 &ha->xgmac_data_dma, GFP_KERNEL);
905 if (!ha->xgmac_data) {
906 mutex_unlock(&vha->hw->optrom_mutex);
907 ql_log(ql_log_warn, vha, 0x7076,
908 "Unable to allocate memory for XGMAC read-data.\n");
909 return 0;
910 }
911
912 do_read:
913 actual_size = 0;
914 memset(ha->xgmac_data, 0, XGMAC_DATA_SIZE);
915
916 rval = qla2x00_get_xgmac_stats(vha, ha->xgmac_data_dma,
917 XGMAC_DATA_SIZE, &actual_size);
918
919 mutex_unlock(&vha->hw->optrom_mutex);
920 if (rval != QLA_SUCCESS) {
921 ql_log(ql_log_warn, vha, 0x7077,
922 "Unable to read XGMAC data (%x).\n", rval);
923 count = 0;
924 }
925
926 count = actual_size > count ? count : actual_size;
927 memcpy(buf, ha->xgmac_data, count);
928
929 return count;
930 }
931
932 static struct bin_attribute sysfs_xgmac_stats_attr = {
933 .attr = {
934 .name = "xgmac_stats",
935 .mode = S_IRUSR,
936 },
937 .size = 0,
938 .read = qla2x00_sysfs_read_xgmac_stats,
939 };
940
941 static ssize_t
qla2x00_sysfs_read_dcbx_tlv(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)942 qla2x00_sysfs_read_dcbx_tlv(struct file *filp, struct kobject *kobj,
943 struct bin_attribute *bin_attr,
944 char *buf, loff_t off, size_t count)
945 {
946 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
947 struct device, kobj)));
948 struct qla_hw_data *ha = vha->hw;
949 int rval;
950
951 if (!capable(CAP_SYS_ADMIN) || off != 0 || count > DCBX_TLV_DATA_SIZE)
952 return 0;
953
954 mutex_lock(&vha->hw->optrom_mutex);
955 if (ha->dcbx_tlv)
956 goto do_read;
957 if (qla2x00_chip_is_down(vha)) {
958 mutex_unlock(&vha->hw->optrom_mutex);
959 return 0;
960 }
961
962 ha->dcbx_tlv = dma_alloc_coherent(&ha->pdev->dev, DCBX_TLV_DATA_SIZE,
963 &ha->dcbx_tlv_dma, GFP_KERNEL);
964 if (!ha->dcbx_tlv) {
965 mutex_unlock(&vha->hw->optrom_mutex);
966 ql_log(ql_log_warn, vha, 0x7078,
967 "Unable to allocate memory for DCBX TLV read-data.\n");
968 return -ENOMEM;
969 }
970
971 do_read:
972 memset(ha->dcbx_tlv, 0, DCBX_TLV_DATA_SIZE);
973
974 rval = qla2x00_get_dcbx_params(vha, ha->dcbx_tlv_dma,
975 DCBX_TLV_DATA_SIZE);
976
977 mutex_unlock(&vha->hw->optrom_mutex);
978
979 if (rval != QLA_SUCCESS) {
980 ql_log(ql_log_warn, vha, 0x7079,
981 "Unable to read DCBX TLV (%x).\n", rval);
982 return -EIO;
983 }
984
985 memcpy(buf, ha->dcbx_tlv, count);
986
987 return count;
988 }
989
990 static struct bin_attribute sysfs_dcbx_tlv_attr = {
991 .attr = {
992 .name = "dcbx_tlv",
993 .mode = S_IRUSR,
994 },
995 .size = 0,
996 .read = qla2x00_sysfs_read_dcbx_tlv,
997 };
998
999 static struct sysfs_entry {
1000 char *name;
1001 struct bin_attribute *attr;
1002 int type;
1003 } bin_file_entries[] = {
1004 { "fw_dump", &sysfs_fw_dump_attr, },
1005 { "nvram", &sysfs_nvram_attr, },
1006 { "optrom", &sysfs_optrom_attr, },
1007 { "optrom_ctl", &sysfs_optrom_ctl_attr, },
1008 { "vpd", &sysfs_vpd_attr, 1 },
1009 { "sfp", &sysfs_sfp_attr, 1 },
1010 { "reset", &sysfs_reset_attr, },
1011 { "issue_logo", &sysfs_issue_logo_attr, },
1012 { "xgmac_stats", &sysfs_xgmac_stats_attr, 3 },
1013 { "dcbx_tlv", &sysfs_dcbx_tlv_attr, 3 },
1014 { NULL },
1015 };
1016
1017 void
qla2x00_alloc_sysfs_attr(scsi_qla_host_t * vha)1018 qla2x00_alloc_sysfs_attr(scsi_qla_host_t *vha)
1019 {
1020 struct Scsi_Host *host = vha->host;
1021 struct sysfs_entry *iter;
1022 int ret;
1023
1024 for (iter = bin_file_entries; iter->name; iter++) {
1025 if (iter->type && !IS_FWI2_CAPABLE(vha->hw))
1026 continue;
1027 if (iter->type == 2 && !IS_QLA25XX(vha->hw))
1028 continue;
1029 if (iter->type == 3 && !(IS_CNA_CAPABLE(vha->hw)))
1030 continue;
1031
1032 ret = sysfs_create_bin_file(&host->shost_gendev.kobj,
1033 iter->attr);
1034 if (ret)
1035 ql_log(ql_log_warn, vha, 0x00f3,
1036 "Unable to create sysfs %s binary attribute (%d).\n",
1037 iter->name, ret);
1038 else
1039 ql_dbg(ql_dbg_init, vha, 0x00f4,
1040 "Successfully created sysfs %s binary attribute.\n",
1041 iter->name);
1042 }
1043 }
1044
1045 void
qla2x00_free_sysfs_attr(scsi_qla_host_t * vha,bool stop_beacon)1046 qla2x00_free_sysfs_attr(scsi_qla_host_t *vha, bool stop_beacon)
1047 {
1048 struct Scsi_Host *host = vha->host;
1049 struct sysfs_entry *iter;
1050 struct qla_hw_data *ha = vha->hw;
1051
1052 for (iter = bin_file_entries; iter->name; iter++) {
1053 if (iter->type && !IS_FWI2_CAPABLE(ha))
1054 continue;
1055 if (iter->type == 2 && !IS_QLA25XX(ha))
1056 continue;
1057 if (iter->type == 3 && !(IS_CNA_CAPABLE(ha)))
1058 continue;
1059
1060 sysfs_remove_bin_file(&host->shost_gendev.kobj,
1061 iter->attr);
1062 }
1063
1064 if (stop_beacon && ha->beacon_blink_led == 1)
1065 ha->isp_ops->beacon_off(vha);
1066 }
1067
1068 /* Scsi_Host attributes. */
1069
1070 static ssize_t
qla2x00_driver_version_show(struct device * dev,struct device_attribute * attr,char * buf)1071 qla2x00_driver_version_show(struct device *dev,
1072 struct device_attribute *attr, char *buf)
1073 {
1074 return scnprintf(buf, PAGE_SIZE, "%s\n", qla2x00_version_str);
1075 }
1076
1077 static ssize_t
qla2x00_fw_version_show(struct device * dev,struct device_attribute * attr,char * buf)1078 qla2x00_fw_version_show(struct device *dev,
1079 struct device_attribute *attr, char *buf)
1080 {
1081 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1082 struct qla_hw_data *ha = vha->hw;
1083 char fw_str[128];
1084
1085 return scnprintf(buf, PAGE_SIZE, "%s\n",
1086 ha->isp_ops->fw_version_str(vha, fw_str, sizeof(fw_str)));
1087 }
1088
1089 static ssize_t
qla2x00_serial_num_show(struct device * dev,struct device_attribute * attr,char * buf)1090 qla2x00_serial_num_show(struct device *dev, struct device_attribute *attr,
1091 char *buf)
1092 {
1093 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1094 struct qla_hw_data *ha = vha->hw;
1095 uint32_t sn;
1096
1097 if (IS_QLAFX00(vha->hw)) {
1098 return scnprintf(buf, PAGE_SIZE, "%s\n",
1099 vha->hw->mr.serial_num);
1100 } else if (IS_FWI2_CAPABLE(ha)) {
1101 qla2xxx_get_vpd_field(vha, "SN", buf, PAGE_SIZE - 1);
1102 return strlen(strcat(buf, "\n"));
1103 }
1104
1105 sn = ((ha->serial0 & 0x1f) << 16) | (ha->serial2 << 8) | ha->serial1;
1106 return scnprintf(buf, PAGE_SIZE, "%c%05d\n", 'A' + sn / 100000,
1107 sn % 100000);
1108 }
1109
1110 static ssize_t
qla2x00_isp_name_show(struct device * dev,struct device_attribute * attr,char * buf)1111 qla2x00_isp_name_show(struct device *dev, struct device_attribute *attr,
1112 char *buf)
1113 {
1114 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1115
1116 return scnprintf(buf, PAGE_SIZE, "ISP%04X\n", vha->hw->pdev->device);
1117 }
1118
1119 static ssize_t
qla2x00_isp_id_show(struct device * dev,struct device_attribute * attr,char * buf)1120 qla2x00_isp_id_show(struct device *dev, struct device_attribute *attr,
1121 char *buf)
1122 {
1123 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1124 struct qla_hw_data *ha = vha->hw;
1125
1126 if (IS_QLAFX00(vha->hw))
1127 return scnprintf(buf, PAGE_SIZE, "%s\n",
1128 vha->hw->mr.hw_version);
1129
1130 return scnprintf(buf, PAGE_SIZE, "%04x %04x %04x %04x\n",
1131 ha->product_id[0], ha->product_id[1], ha->product_id[2],
1132 ha->product_id[3]);
1133 }
1134
1135 static ssize_t
qla2x00_model_name_show(struct device * dev,struct device_attribute * attr,char * buf)1136 qla2x00_model_name_show(struct device *dev, struct device_attribute *attr,
1137 char *buf)
1138 {
1139 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1140
1141 return scnprintf(buf, PAGE_SIZE, "%s\n", vha->hw->model_number);
1142 }
1143
1144 static ssize_t
qla2x00_model_desc_show(struct device * dev,struct device_attribute * attr,char * buf)1145 qla2x00_model_desc_show(struct device *dev, struct device_attribute *attr,
1146 char *buf)
1147 {
1148 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1149
1150 return scnprintf(buf, PAGE_SIZE, "%s\n", vha->hw->model_desc);
1151 }
1152
1153 static ssize_t
qla2x00_pci_info_show(struct device * dev,struct device_attribute * attr,char * buf)1154 qla2x00_pci_info_show(struct device *dev, struct device_attribute *attr,
1155 char *buf)
1156 {
1157 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1158 char pci_info[30];
1159
1160 return scnprintf(buf, PAGE_SIZE, "%s\n",
1161 vha->hw->isp_ops->pci_info_str(vha, pci_info,
1162 sizeof(pci_info)));
1163 }
1164
1165 static ssize_t
qla2x00_link_state_show(struct device * dev,struct device_attribute * attr,char * buf)1166 qla2x00_link_state_show(struct device *dev, struct device_attribute *attr,
1167 char *buf)
1168 {
1169 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1170 struct qla_hw_data *ha = vha->hw;
1171 int len = 0;
1172
1173 if (atomic_read(&vha->loop_state) == LOOP_DOWN ||
1174 atomic_read(&vha->loop_state) == LOOP_DEAD ||
1175 vha->device_flags & DFLG_NO_CABLE)
1176 len = scnprintf(buf, PAGE_SIZE, "Link Down\n");
1177 else if (atomic_read(&vha->loop_state) != LOOP_READY ||
1178 qla2x00_chip_is_down(vha))
1179 len = scnprintf(buf, PAGE_SIZE, "Unknown Link State\n");
1180 else {
1181 len = scnprintf(buf, PAGE_SIZE, "Link Up - ");
1182
1183 switch (ha->current_topology) {
1184 case ISP_CFG_NL:
1185 len += scnprintf(buf + len, PAGE_SIZE-len, "Loop\n");
1186 break;
1187 case ISP_CFG_FL:
1188 len += scnprintf(buf + len, PAGE_SIZE-len, "FL_Port\n");
1189 break;
1190 case ISP_CFG_N:
1191 len += scnprintf(buf + len, PAGE_SIZE-len,
1192 "N_Port to N_Port\n");
1193 break;
1194 case ISP_CFG_F:
1195 len += scnprintf(buf + len, PAGE_SIZE-len, "F_Port\n");
1196 break;
1197 default:
1198 len += scnprintf(buf + len, PAGE_SIZE-len, "Loop\n");
1199 break;
1200 }
1201 }
1202 return len;
1203 }
1204
1205 static ssize_t
qla2x00_zio_show(struct device * dev,struct device_attribute * attr,char * buf)1206 qla2x00_zio_show(struct device *dev, struct device_attribute *attr,
1207 char *buf)
1208 {
1209 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1210 int len = 0;
1211
1212 switch (vha->hw->zio_mode) {
1213 case QLA_ZIO_MODE_6:
1214 len += scnprintf(buf + len, PAGE_SIZE-len, "Mode 6\n");
1215 break;
1216 case QLA_ZIO_DISABLED:
1217 len += scnprintf(buf + len, PAGE_SIZE-len, "Disabled\n");
1218 break;
1219 }
1220 return len;
1221 }
1222
1223 static ssize_t
qla2x00_zio_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1224 qla2x00_zio_store(struct device *dev, struct device_attribute *attr,
1225 const char *buf, size_t count)
1226 {
1227 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1228 struct qla_hw_data *ha = vha->hw;
1229 int val = 0;
1230 uint16_t zio_mode;
1231
1232 if (!IS_ZIO_SUPPORTED(ha))
1233 return -ENOTSUPP;
1234
1235 if (sscanf(buf, "%d", &val) != 1)
1236 return -EINVAL;
1237
1238 if (val)
1239 zio_mode = QLA_ZIO_MODE_6;
1240 else
1241 zio_mode = QLA_ZIO_DISABLED;
1242
1243 /* Update per-hba values and queue a reset. */
1244 if (zio_mode != QLA_ZIO_DISABLED || ha->zio_mode != QLA_ZIO_DISABLED) {
1245 ha->zio_mode = zio_mode;
1246 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
1247 }
1248 return strlen(buf);
1249 }
1250
1251 static ssize_t
qla2x00_zio_timer_show(struct device * dev,struct device_attribute * attr,char * buf)1252 qla2x00_zio_timer_show(struct device *dev, struct device_attribute *attr,
1253 char *buf)
1254 {
1255 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1256
1257 return scnprintf(buf, PAGE_SIZE, "%d us\n", vha->hw->zio_timer * 100);
1258 }
1259
1260 static ssize_t
qla2x00_zio_timer_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1261 qla2x00_zio_timer_store(struct device *dev, struct device_attribute *attr,
1262 const char *buf, size_t count)
1263 {
1264 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1265 int val = 0;
1266 uint16_t zio_timer;
1267
1268 if (sscanf(buf, "%d", &val) != 1)
1269 return -EINVAL;
1270 if (val > 25500 || val < 100)
1271 return -ERANGE;
1272
1273 zio_timer = (uint16_t)(val / 100);
1274 vha->hw->zio_timer = zio_timer;
1275
1276 return strlen(buf);
1277 }
1278
1279 static ssize_t
qla_zio_threshold_show(struct device * dev,struct device_attribute * attr,char * buf)1280 qla_zio_threshold_show(struct device *dev, struct device_attribute *attr,
1281 char *buf)
1282 {
1283 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1284
1285 return scnprintf(buf, PAGE_SIZE, "%d exchanges\n",
1286 vha->hw->last_zio_threshold);
1287 }
1288
1289 static ssize_t
qla_zio_threshold_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1290 qla_zio_threshold_store(struct device *dev, struct device_attribute *attr,
1291 const char *buf, size_t count)
1292 {
1293 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1294 int val = 0;
1295
1296 if (vha->hw->zio_mode != QLA_ZIO_MODE_6)
1297 return -EINVAL;
1298 if (sscanf(buf, "%d", &val) != 1)
1299 return -EINVAL;
1300 if (val < 0 || val > 256)
1301 return -ERANGE;
1302
1303 atomic_set(&vha->hw->zio_threshold, val);
1304 return strlen(buf);
1305 }
1306
1307 static ssize_t
qla2x00_beacon_show(struct device * dev,struct device_attribute * attr,char * buf)1308 qla2x00_beacon_show(struct device *dev, struct device_attribute *attr,
1309 char *buf)
1310 {
1311 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1312 int len = 0;
1313
1314 if (vha->hw->beacon_blink_led)
1315 len += scnprintf(buf + len, PAGE_SIZE-len, "Enabled\n");
1316 else
1317 len += scnprintf(buf + len, PAGE_SIZE-len, "Disabled\n");
1318 return len;
1319 }
1320
1321 static ssize_t
qla2x00_beacon_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1322 qla2x00_beacon_store(struct device *dev, struct device_attribute *attr,
1323 const char *buf, size_t count)
1324 {
1325 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1326 struct qla_hw_data *ha = vha->hw;
1327 int val = 0;
1328 int rval;
1329
1330 if (IS_QLA2100(ha) || IS_QLA2200(ha))
1331 return -EPERM;
1332
1333 if (sscanf(buf, "%d", &val) != 1)
1334 return -EINVAL;
1335
1336 mutex_lock(&vha->hw->optrom_mutex);
1337 if (qla2x00_chip_is_down(vha)) {
1338 mutex_unlock(&vha->hw->optrom_mutex);
1339 ql_log(ql_log_warn, vha, 0x707a,
1340 "Abort ISP active -- ignoring beacon request.\n");
1341 return -EBUSY;
1342 }
1343
1344 if (val)
1345 rval = ha->isp_ops->beacon_on(vha);
1346 else
1347 rval = ha->isp_ops->beacon_off(vha);
1348
1349 if (rval != QLA_SUCCESS)
1350 count = 0;
1351
1352 mutex_unlock(&vha->hw->optrom_mutex);
1353
1354 return count;
1355 }
1356
1357 static ssize_t
qla2x00_beacon_config_show(struct device * dev,struct device_attribute * attr,char * buf)1358 qla2x00_beacon_config_show(struct device *dev, struct device_attribute *attr,
1359 char *buf)
1360 {
1361 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1362 struct qla_hw_data *ha = vha->hw;
1363 uint16_t led[3] = { 0 };
1364
1365 if (!IS_QLA2031(ha) && !IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1366 return -EPERM;
1367
1368 if (ql26xx_led_config(vha, 0, led))
1369 return scnprintf(buf, PAGE_SIZE, "\n");
1370
1371 return scnprintf(buf, PAGE_SIZE, "%#04hx %#04hx %#04hx\n",
1372 led[0], led[1], led[2]);
1373 }
1374
1375 static ssize_t
qla2x00_beacon_config_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1376 qla2x00_beacon_config_store(struct device *dev, struct device_attribute *attr,
1377 const char *buf, size_t count)
1378 {
1379 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1380 struct qla_hw_data *ha = vha->hw;
1381 uint16_t options = BIT_0;
1382 uint16_t led[3] = { 0 };
1383 uint16_t word[4];
1384 int n;
1385
1386 if (!IS_QLA2031(ha) && !IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1387 return -EPERM;
1388
1389 n = sscanf(buf, "%hx %hx %hx %hx", word+0, word+1, word+2, word+3);
1390 if (n == 4) {
1391 if (word[0] == 3) {
1392 options |= BIT_3|BIT_2|BIT_1;
1393 led[0] = word[1];
1394 led[1] = word[2];
1395 led[2] = word[3];
1396 goto write;
1397 }
1398 return -EINVAL;
1399 }
1400
1401 if (n == 2) {
1402 /* check led index */
1403 if (word[0] == 0) {
1404 options |= BIT_2;
1405 led[0] = word[1];
1406 goto write;
1407 }
1408 if (word[0] == 1) {
1409 options |= BIT_3;
1410 led[1] = word[1];
1411 goto write;
1412 }
1413 if (word[0] == 2) {
1414 options |= BIT_1;
1415 led[2] = word[1];
1416 goto write;
1417 }
1418 return -EINVAL;
1419 }
1420
1421 return -EINVAL;
1422
1423 write:
1424 if (ql26xx_led_config(vha, options, led))
1425 return -EFAULT;
1426
1427 return count;
1428 }
1429
1430 static ssize_t
qla2x00_optrom_bios_version_show(struct device * dev,struct device_attribute * attr,char * buf)1431 qla2x00_optrom_bios_version_show(struct device *dev,
1432 struct device_attribute *attr, char *buf)
1433 {
1434 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1435 struct qla_hw_data *ha = vha->hw;
1436
1437 return scnprintf(buf, PAGE_SIZE, "%d.%02d\n", ha->bios_revision[1],
1438 ha->bios_revision[0]);
1439 }
1440
1441 static ssize_t
qla2x00_optrom_efi_version_show(struct device * dev,struct device_attribute * attr,char * buf)1442 qla2x00_optrom_efi_version_show(struct device *dev,
1443 struct device_attribute *attr, char *buf)
1444 {
1445 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1446 struct qla_hw_data *ha = vha->hw;
1447
1448 return scnprintf(buf, PAGE_SIZE, "%d.%02d\n", ha->efi_revision[1],
1449 ha->efi_revision[0]);
1450 }
1451
1452 static ssize_t
qla2x00_optrom_fcode_version_show(struct device * dev,struct device_attribute * attr,char * buf)1453 qla2x00_optrom_fcode_version_show(struct device *dev,
1454 struct device_attribute *attr, char *buf)
1455 {
1456 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1457 struct qla_hw_data *ha = vha->hw;
1458
1459 return scnprintf(buf, PAGE_SIZE, "%d.%02d\n", ha->fcode_revision[1],
1460 ha->fcode_revision[0]);
1461 }
1462
1463 static ssize_t
qla2x00_optrom_fw_version_show(struct device * dev,struct device_attribute * attr,char * buf)1464 qla2x00_optrom_fw_version_show(struct device *dev,
1465 struct device_attribute *attr, char *buf)
1466 {
1467 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1468 struct qla_hw_data *ha = vha->hw;
1469
1470 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d %d\n",
1471 ha->fw_revision[0], ha->fw_revision[1], ha->fw_revision[2],
1472 ha->fw_revision[3]);
1473 }
1474
1475 static ssize_t
qla2x00_optrom_gold_fw_version_show(struct device * dev,struct device_attribute * attr,char * buf)1476 qla2x00_optrom_gold_fw_version_show(struct device *dev,
1477 struct device_attribute *attr, char *buf)
1478 {
1479 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1480 struct qla_hw_data *ha = vha->hw;
1481
1482 if (!IS_QLA81XX(ha) && !IS_QLA83XX(ha) &&
1483 !IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1484 return scnprintf(buf, PAGE_SIZE, "\n");
1485
1486 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d (%d)\n",
1487 ha->gold_fw_version[0], ha->gold_fw_version[1],
1488 ha->gold_fw_version[2], ha->gold_fw_version[3]);
1489 }
1490
1491 static ssize_t
qla2x00_total_isp_aborts_show(struct device * dev,struct device_attribute * attr,char * buf)1492 qla2x00_total_isp_aborts_show(struct device *dev,
1493 struct device_attribute *attr, char *buf)
1494 {
1495 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1496
1497 return scnprintf(buf, PAGE_SIZE, "%d\n",
1498 vha->qla_stats.total_isp_aborts);
1499 }
1500
1501 static ssize_t
qla24xx_84xx_fw_version_show(struct device * dev,struct device_attribute * attr,char * buf)1502 qla24xx_84xx_fw_version_show(struct device *dev,
1503 struct device_attribute *attr, char *buf)
1504 {
1505 int rval = QLA_SUCCESS;
1506 uint16_t status[2] = { 0 };
1507 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1508 struct qla_hw_data *ha = vha->hw;
1509
1510 if (!IS_QLA84XX(ha))
1511 return scnprintf(buf, PAGE_SIZE, "\n");
1512
1513 if (!ha->cs84xx->op_fw_version) {
1514 rval = qla84xx_verify_chip(vha, status);
1515
1516 if (!rval && !status[0])
1517 return scnprintf(buf, PAGE_SIZE, "%u\n",
1518 (uint32_t)ha->cs84xx->op_fw_version);
1519 }
1520
1521 return scnprintf(buf, PAGE_SIZE, "\n");
1522 }
1523
1524 static ssize_t
qla2x00_serdes_version_show(struct device * dev,struct device_attribute * attr,char * buf)1525 qla2x00_serdes_version_show(struct device *dev, struct device_attribute *attr,
1526 char *buf)
1527 {
1528 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1529 struct qla_hw_data *ha = vha->hw;
1530
1531 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1532 return scnprintf(buf, PAGE_SIZE, "\n");
1533
1534 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d\n",
1535 ha->serdes_version[0], ha->serdes_version[1],
1536 ha->serdes_version[2]);
1537 }
1538
1539 static ssize_t
qla2x00_mpi_version_show(struct device * dev,struct device_attribute * attr,char * buf)1540 qla2x00_mpi_version_show(struct device *dev, struct device_attribute *attr,
1541 char *buf)
1542 {
1543 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1544 struct qla_hw_data *ha = vha->hw;
1545
1546 if (!IS_QLA81XX(ha) && !IS_QLA8031(ha) && !IS_QLA8044(ha) &&
1547 !IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1548 return scnprintf(buf, PAGE_SIZE, "\n");
1549
1550 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d (%x)\n",
1551 ha->mpi_version[0], ha->mpi_version[1], ha->mpi_version[2],
1552 ha->mpi_capabilities);
1553 }
1554
1555 static ssize_t
qla2x00_phy_version_show(struct device * dev,struct device_attribute * attr,char * buf)1556 qla2x00_phy_version_show(struct device *dev, struct device_attribute *attr,
1557 char *buf)
1558 {
1559 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1560 struct qla_hw_data *ha = vha->hw;
1561
1562 if (!IS_QLA81XX(ha) && !IS_QLA8031(ha))
1563 return scnprintf(buf, PAGE_SIZE, "\n");
1564
1565 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d\n",
1566 ha->phy_version[0], ha->phy_version[1], ha->phy_version[2]);
1567 }
1568
1569 static ssize_t
qla2x00_flash_block_size_show(struct device * dev,struct device_attribute * attr,char * buf)1570 qla2x00_flash_block_size_show(struct device *dev,
1571 struct device_attribute *attr, char *buf)
1572 {
1573 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1574 struct qla_hw_data *ha = vha->hw;
1575
1576 return scnprintf(buf, PAGE_SIZE, "0x%x\n", ha->fdt_block_size);
1577 }
1578
1579 static ssize_t
qla2x00_vlan_id_show(struct device * dev,struct device_attribute * attr,char * buf)1580 qla2x00_vlan_id_show(struct device *dev, struct device_attribute *attr,
1581 char *buf)
1582 {
1583 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1584
1585 if (!IS_CNA_CAPABLE(vha->hw))
1586 return scnprintf(buf, PAGE_SIZE, "\n");
1587
1588 return scnprintf(buf, PAGE_SIZE, "%d\n", vha->fcoe_vlan_id);
1589 }
1590
1591 static ssize_t
qla2x00_vn_port_mac_address_show(struct device * dev,struct device_attribute * attr,char * buf)1592 qla2x00_vn_port_mac_address_show(struct device *dev,
1593 struct device_attribute *attr, char *buf)
1594 {
1595 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1596
1597 if (!IS_CNA_CAPABLE(vha->hw))
1598 return scnprintf(buf, PAGE_SIZE, "\n");
1599
1600 return scnprintf(buf, PAGE_SIZE, "%pMR\n", vha->fcoe_vn_port_mac);
1601 }
1602
1603 static ssize_t
qla2x00_fabric_param_show(struct device * dev,struct device_attribute * attr,char * buf)1604 qla2x00_fabric_param_show(struct device *dev, struct device_attribute *attr,
1605 char *buf)
1606 {
1607 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1608
1609 return scnprintf(buf, PAGE_SIZE, "%d\n", vha->hw->switch_cap);
1610 }
1611
1612 static ssize_t
qla2x00_thermal_temp_show(struct device * dev,struct device_attribute * attr,char * buf)1613 qla2x00_thermal_temp_show(struct device *dev,
1614 struct device_attribute *attr, char *buf)
1615 {
1616 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1617 uint16_t temp = 0;
1618 int rc;
1619
1620 mutex_lock(&vha->hw->optrom_mutex);
1621 if (qla2x00_chip_is_down(vha)) {
1622 mutex_unlock(&vha->hw->optrom_mutex);
1623 ql_log(ql_log_warn, vha, 0x70dc, "ISP reset active.\n");
1624 goto done;
1625 }
1626
1627 if (vha->hw->flags.eeh_busy) {
1628 mutex_unlock(&vha->hw->optrom_mutex);
1629 ql_log(ql_log_warn, vha, 0x70dd, "PCI EEH busy.\n");
1630 goto done;
1631 }
1632
1633 rc = qla2x00_get_thermal_temp(vha, &temp);
1634 mutex_unlock(&vha->hw->optrom_mutex);
1635 if (rc == QLA_SUCCESS)
1636 return scnprintf(buf, PAGE_SIZE, "%d\n", temp);
1637
1638 done:
1639 return scnprintf(buf, PAGE_SIZE, "\n");
1640 }
1641
1642 static ssize_t
qla2x00_fw_state_show(struct device * dev,struct device_attribute * attr,char * buf)1643 qla2x00_fw_state_show(struct device *dev, struct device_attribute *attr,
1644 char *buf)
1645 {
1646 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1647 int rval = QLA_FUNCTION_FAILED;
1648 uint16_t state[6];
1649 uint32_t pstate;
1650
1651 if (IS_QLAFX00(vha->hw)) {
1652 pstate = qlafx00_fw_state_show(dev, attr, buf);
1653 return scnprintf(buf, PAGE_SIZE, "0x%x\n", pstate);
1654 }
1655
1656 mutex_lock(&vha->hw->optrom_mutex);
1657 if (qla2x00_chip_is_down(vha)) {
1658 mutex_unlock(&vha->hw->optrom_mutex);
1659 ql_log(ql_log_warn, vha, 0x707c,
1660 "ISP reset active.\n");
1661 goto out;
1662 } else if (vha->hw->flags.eeh_busy) {
1663 mutex_unlock(&vha->hw->optrom_mutex);
1664 goto out;
1665 }
1666
1667 rval = qla2x00_get_firmware_state(vha, state);
1668 mutex_unlock(&vha->hw->optrom_mutex);
1669 out:
1670 if (rval != QLA_SUCCESS) {
1671 memset(state, -1, sizeof(state));
1672 rval = qla2x00_get_firmware_state(vha, state);
1673 }
1674
1675 return scnprintf(buf, PAGE_SIZE, "0x%x 0x%x 0x%x 0x%x 0x%x 0x%x\n",
1676 state[0], state[1], state[2], state[3], state[4], state[5]);
1677 }
1678
1679 static ssize_t
qla2x00_diag_requests_show(struct device * dev,struct device_attribute * attr,char * buf)1680 qla2x00_diag_requests_show(struct device *dev,
1681 struct device_attribute *attr, char *buf)
1682 {
1683 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1684
1685 if (!IS_BIDI_CAPABLE(vha->hw))
1686 return scnprintf(buf, PAGE_SIZE, "\n");
1687
1688 return scnprintf(buf, PAGE_SIZE, "%llu\n", vha->bidi_stats.io_count);
1689 }
1690
1691 static ssize_t
qla2x00_diag_megabytes_show(struct device * dev,struct device_attribute * attr,char * buf)1692 qla2x00_diag_megabytes_show(struct device *dev,
1693 struct device_attribute *attr, char *buf)
1694 {
1695 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1696
1697 if (!IS_BIDI_CAPABLE(vha->hw))
1698 return scnprintf(buf, PAGE_SIZE, "\n");
1699
1700 return scnprintf(buf, PAGE_SIZE, "%llu\n",
1701 vha->bidi_stats.transfer_bytes >> 20);
1702 }
1703
1704 static ssize_t
qla2x00_fw_dump_size_show(struct device * dev,struct device_attribute * attr,char * buf)1705 qla2x00_fw_dump_size_show(struct device *dev, struct device_attribute *attr,
1706 char *buf)
1707 {
1708 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1709 struct qla_hw_data *ha = vha->hw;
1710 uint32_t size;
1711
1712 if (!ha->fw_dumped)
1713 size = 0;
1714 else if (IS_P3P_TYPE(ha))
1715 size = ha->md_template_size + ha->md_dump_size;
1716 else
1717 size = ha->fw_dump_len;
1718
1719 return scnprintf(buf, PAGE_SIZE, "%d\n", size);
1720 }
1721
1722 static ssize_t
qla2x00_allow_cna_fw_dump_show(struct device * dev,struct device_attribute * attr,char * buf)1723 qla2x00_allow_cna_fw_dump_show(struct device *dev,
1724 struct device_attribute *attr, char *buf)
1725 {
1726 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1727
1728 if (!IS_P3P_TYPE(vha->hw))
1729 return scnprintf(buf, PAGE_SIZE, "\n");
1730 else
1731 return scnprintf(buf, PAGE_SIZE, "%s\n",
1732 vha->hw->allow_cna_fw_dump ? "true" : "false");
1733 }
1734
1735 static ssize_t
qla2x00_allow_cna_fw_dump_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1736 qla2x00_allow_cna_fw_dump_store(struct device *dev,
1737 struct device_attribute *attr, const char *buf, size_t count)
1738 {
1739 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1740 int val = 0;
1741
1742 if (!IS_P3P_TYPE(vha->hw))
1743 return -EINVAL;
1744
1745 if (sscanf(buf, "%d", &val) != 1)
1746 return -EINVAL;
1747
1748 vha->hw->allow_cna_fw_dump = val != 0;
1749
1750 return strlen(buf);
1751 }
1752
1753 static ssize_t
qla2x00_pep_version_show(struct device * dev,struct device_attribute * attr,char * buf)1754 qla2x00_pep_version_show(struct device *dev, struct device_attribute *attr,
1755 char *buf)
1756 {
1757 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1758 struct qla_hw_data *ha = vha->hw;
1759
1760 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1761 return scnprintf(buf, PAGE_SIZE, "\n");
1762
1763 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d\n",
1764 ha->pep_version[0], ha->pep_version[1], ha->pep_version[2]);
1765 }
1766
1767 static ssize_t
qla2x00_min_supported_speed_show(struct device * dev,struct device_attribute * attr,char * buf)1768 qla2x00_min_supported_speed_show(struct device *dev,
1769 struct device_attribute *attr, char *buf)
1770 {
1771 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1772 struct qla_hw_data *ha = vha->hw;
1773
1774 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1775 return scnprintf(buf, PAGE_SIZE, "\n");
1776
1777 return scnprintf(buf, PAGE_SIZE, "%s\n",
1778 ha->min_supported_speed == 6 ? "64Gps" :
1779 ha->min_supported_speed == 5 ? "32Gps" :
1780 ha->min_supported_speed == 4 ? "16Gps" :
1781 ha->min_supported_speed == 3 ? "8Gps" :
1782 ha->min_supported_speed == 2 ? "4Gps" :
1783 ha->min_supported_speed != 0 ? "unknown" : "");
1784 }
1785
1786 static ssize_t
qla2x00_max_supported_speed_show(struct device * dev,struct device_attribute * attr,char * buf)1787 qla2x00_max_supported_speed_show(struct device *dev,
1788 struct device_attribute *attr, char *buf)
1789 {
1790 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1791 struct qla_hw_data *ha = vha->hw;
1792
1793 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1794 return scnprintf(buf, PAGE_SIZE, "\n");
1795
1796 return scnprintf(buf, PAGE_SIZE, "%s\n",
1797 ha->max_supported_speed == 2 ? "64Gps" :
1798 ha->max_supported_speed == 1 ? "32Gps" :
1799 ha->max_supported_speed == 0 ? "16Gps" : "unknown");
1800 }
1801
1802 static ssize_t
qla2x00_port_speed_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1803 qla2x00_port_speed_store(struct device *dev, struct device_attribute *attr,
1804 const char *buf, size_t count)
1805 {
1806 struct scsi_qla_host *vha = shost_priv(dev_to_shost(dev));
1807 ulong type, speed;
1808 int oldspeed, rval;
1809 int mode = QLA_SET_DATA_RATE_LR;
1810 struct qla_hw_data *ha = vha->hw;
1811
1812 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha)) {
1813 ql_log(ql_log_warn, vha, 0x70d8,
1814 "Speed setting not supported \n");
1815 return -EINVAL;
1816 }
1817
1818 rval = kstrtol(buf, 10, &type);
1819 if (rval)
1820 return rval;
1821 speed = type;
1822 if (type == 40 || type == 80 || type == 160 ||
1823 type == 320) {
1824 ql_dbg(ql_dbg_user, vha, 0x70d9,
1825 "Setting will be affected after a loss of sync\n");
1826 type = type/10;
1827 mode = QLA_SET_DATA_RATE_NOLR;
1828 }
1829
1830 oldspeed = ha->set_data_rate;
1831
1832 switch (type) {
1833 case 0:
1834 ha->set_data_rate = PORT_SPEED_AUTO;
1835 break;
1836 case 4:
1837 ha->set_data_rate = PORT_SPEED_4GB;
1838 break;
1839 case 8:
1840 ha->set_data_rate = PORT_SPEED_8GB;
1841 break;
1842 case 16:
1843 ha->set_data_rate = PORT_SPEED_16GB;
1844 break;
1845 case 32:
1846 ha->set_data_rate = PORT_SPEED_32GB;
1847 break;
1848 default:
1849 ql_log(ql_log_warn, vha, 0x1199,
1850 "Unrecognized speed setting:%lx. Setting Autoneg\n",
1851 speed);
1852 ha->set_data_rate = PORT_SPEED_AUTO;
1853 }
1854
1855 if (qla2x00_chip_is_down(vha) || (oldspeed == ha->set_data_rate))
1856 return -EINVAL;
1857
1858 ql_log(ql_log_info, vha, 0x70da,
1859 "Setting speed to %lx Gbps \n", type);
1860
1861 rval = qla2x00_set_data_rate(vha, mode);
1862 if (rval != QLA_SUCCESS)
1863 return -EIO;
1864
1865 return strlen(buf);
1866 }
1867
1868 static const struct {
1869 u16 rate;
1870 char *str;
1871 } port_speed_str[] = {
1872 { PORT_SPEED_4GB, "4" },
1873 { PORT_SPEED_8GB, "8" },
1874 { PORT_SPEED_16GB, "16" },
1875 { PORT_SPEED_32GB, "32" },
1876 { PORT_SPEED_64GB, "64" },
1877 { PORT_SPEED_10GB, "10" },
1878 };
1879
1880 static ssize_t
qla2x00_port_speed_show(struct device * dev,struct device_attribute * attr,char * buf)1881 qla2x00_port_speed_show(struct device *dev, struct device_attribute *attr,
1882 char *buf)
1883 {
1884 struct scsi_qla_host *vha = shost_priv(dev_to_shost(dev));
1885 struct qla_hw_data *ha = vha->hw;
1886 ssize_t rval;
1887 u16 i;
1888 char *speed = "Unknown";
1889
1890 rval = qla2x00_get_data_rate(vha);
1891 if (rval != QLA_SUCCESS) {
1892 ql_log(ql_log_warn, vha, 0x70db,
1893 "Unable to get port speed rval:%zd\n", rval);
1894 return -EINVAL;
1895 }
1896
1897 for (i = 0; i < ARRAY_SIZE(port_speed_str); i++) {
1898 if (port_speed_str[i].rate != ha->link_data_rate)
1899 continue;
1900 speed = port_speed_str[i].str;
1901 break;
1902 }
1903
1904 return scnprintf(buf, PAGE_SIZE, "%s\n", speed);
1905 }
1906
1907 static ssize_t
qla2x00_mpi_pause_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1908 qla2x00_mpi_pause_store(struct device *dev,
1909 struct device_attribute *attr, const char *buf, size_t count)
1910 {
1911 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1912 int rval = 0;
1913
1914 if (sscanf(buf, "%d", &rval) != 1)
1915 return -EINVAL;
1916
1917 ql_log(ql_log_warn, vha, 0x7089, "Pausing MPI...\n");
1918
1919 rval = qla83xx_wr_reg(vha, 0x002012d4, 0x30000001);
1920
1921 if (rval != QLA_SUCCESS) {
1922 ql_log(ql_log_warn, vha, 0x708a, "Unable to pause MPI.\n");
1923 count = 0;
1924 }
1925
1926 return count;
1927 }
1928
1929 static DEVICE_ATTR(mpi_pause, S_IWUSR, NULL, qla2x00_mpi_pause_store);
1930
1931 /* ----- */
1932
1933 static ssize_t
qlini_mode_show(struct device * dev,struct device_attribute * attr,char * buf)1934 qlini_mode_show(struct device *dev, struct device_attribute *attr, char *buf)
1935 {
1936 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1937 int len = 0;
1938
1939 len += scnprintf(buf + len, PAGE_SIZE-len,
1940 "Supported options: enabled | disabled | dual | exclusive\n");
1941
1942 /* --- */
1943 len += scnprintf(buf + len, PAGE_SIZE-len, "Current selection: ");
1944
1945 switch (vha->qlini_mode) {
1946 case QLA2XXX_INI_MODE_EXCLUSIVE:
1947 len += scnprintf(buf + len, PAGE_SIZE-len,
1948 QLA2XXX_INI_MODE_STR_EXCLUSIVE);
1949 break;
1950 case QLA2XXX_INI_MODE_DISABLED:
1951 len += scnprintf(buf + len, PAGE_SIZE-len,
1952 QLA2XXX_INI_MODE_STR_DISABLED);
1953 break;
1954 case QLA2XXX_INI_MODE_ENABLED:
1955 len += scnprintf(buf + len, PAGE_SIZE-len,
1956 QLA2XXX_INI_MODE_STR_ENABLED);
1957 break;
1958 case QLA2XXX_INI_MODE_DUAL:
1959 len += scnprintf(buf + len, PAGE_SIZE-len,
1960 QLA2XXX_INI_MODE_STR_DUAL);
1961 break;
1962 }
1963 len += scnprintf(buf + len, PAGE_SIZE-len, "\n");
1964
1965 return len;
1966 }
1967
1968 static char *mode_to_str[] = {
1969 "exclusive",
1970 "disabled",
1971 "enabled",
1972 "dual",
1973 };
1974
1975 #define NEED_EXCH_OFFLOAD(_exchg) ((_exchg) > FW_DEF_EXCHANGES_CNT)
qla_set_ini_mode(scsi_qla_host_t * vha,int op)1976 static void qla_set_ini_mode(scsi_qla_host_t *vha, int op)
1977 {
1978 enum {
1979 NO_ACTION,
1980 MODE_CHANGE_ACCEPT,
1981 MODE_CHANGE_NO_ACTION,
1982 TARGET_STILL_ACTIVE,
1983 };
1984 int action = NO_ACTION;
1985 int set_mode = 0;
1986 u8 eo_toggle = 0; /* exchange offload flipped */
1987
1988 switch (vha->qlini_mode) {
1989 case QLA2XXX_INI_MODE_DISABLED:
1990 switch (op) {
1991 case QLA2XXX_INI_MODE_DISABLED:
1992 if (qla_tgt_mode_enabled(vha)) {
1993 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld) !=
1994 vha->hw->flags.exchoffld_enabled)
1995 eo_toggle = 1;
1996 if (((vha->ql2xexchoffld !=
1997 vha->u_ql2xexchoffld) &&
1998 NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld)) ||
1999 eo_toggle) {
2000 /*
2001 * The number of exchange to be offload
2002 * was tweaked or offload option was
2003 * flipped
2004 */
2005 action = MODE_CHANGE_ACCEPT;
2006 } else {
2007 action = MODE_CHANGE_NO_ACTION;
2008 }
2009 } else {
2010 action = MODE_CHANGE_NO_ACTION;
2011 }
2012 break;
2013 case QLA2XXX_INI_MODE_EXCLUSIVE:
2014 if (qla_tgt_mode_enabled(vha)) {
2015 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld) !=
2016 vha->hw->flags.exchoffld_enabled)
2017 eo_toggle = 1;
2018 if (((vha->ql2xexchoffld !=
2019 vha->u_ql2xexchoffld) &&
2020 NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld)) ||
2021 eo_toggle) {
2022 /*
2023 * The number of exchange to be offload
2024 * was tweaked or offload option was
2025 * flipped
2026 */
2027 action = MODE_CHANGE_ACCEPT;
2028 } else {
2029 action = MODE_CHANGE_NO_ACTION;
2030 }
2031 } else {
2032 action = MODE_CHANGE_ACCEPT;
2033 }
2034 break;
2035 case QLA2XXX_INI_MODE_DUAL:
2036 action = MODE_CHANGE_ACCEPT;
2037 /* active_mode is target only, reset it to dual */
2038 if (qla_tgt_mode_enabled(vha)) {
2039 set_mode = 1;
2040 action = MODE_CHANGE_ACCEPT;
2041 } else {
2042 action = MODE_CHANGE_NO_ACTION;
2043 }
2044 break;
2045
2046 case QLA2XXX_INI_MODE_ENABLED:
2047 if (qla_tgt_mode_enabled(vha))
2048 action = TARGET_STILL_ACTIVE;
2049 else {
2050 action = MODE_CHANGE_ACCEPT;
2051 set_mode = 1;
2052 }
2053 break;
2054 }
2055 break;
2056
2057 case QLA2XXX_INI_MODE_EXCLUSIVE:
2058 switch (op) {
2059 case QLA2XXX_INI_MODE_EXCLUSIVE:
2060 if (qla_tgt_mode_enabled(vha)) {
2061 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld) !=
2062 vha->hw->flags.exchoffld_enabled)
2063 eo_toggle = 1;
2064 if (((vha->ql2xexchoffld !=
2065 vha->u_ql2xexchoffld) &&
2066 NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld)) ||
2067 eo_toggle)
2068 /*
2069 * The number of exchange to be offload
2070 * was tweaked or offload option was
2071 * flipped
2072 */
2073 action = MODE_CHANGE_ACCEPT;
2074 else
2075 action = NO_ACTION;
2076 } else
2077 action = NO_ACTION;
2078
2079 break;
2080
2081 case QLA2XXX_INI_MODE_DISABLED:
2082 if (qla_tgt_mode_enabled(vha)) {
2083 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld) !=
2084 vha->hw->flags.exchoffld_enabled)
2085 eo_toggle = 1;
2086 if (((vha->ql2xexchoffld !=
2087 vha->u_ql2xexchoffld) &&
2088 NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld)) ||
2089 eo_toggle)
2090 action = MODE_CHANGE_ACCEPT;
2091 else
2092 action = MODE_CHANGE_NO_ACTION;
2093 } else
2094 action = MODE_CHANGE_NO_ACTION;
2095 break;
2096
2097 case QLA2XXX_INI_MODE_DUAL: /* exclusive -> dual */
2098 if (qla_tgt_mode_enabled(vha)) {
2099 action = MODE_CHANGE_ACCEPT;
2100 set_mode = 1;
2101 } else
2102 action = MODE_CHANGE_ACCEPT;
2103 break;
2104
2105 case QLA2XXX_INI_MODE_ENABLED:
2106 if (qla_tgt_mode_enabled(vha))
2107 action = TARGET_STILL_ACTIVE;
2108 else {
2109 if (vha->hw->flags.fw_started)
2110 action = MODE_CHANGE_NO_ACTION;
2111 else
2112 action = MODE_CHANGE_ACCEPT;
2113 }
2114 break;
2115 }
2116 break;
2117
2118 case QLA2XXX_INI_MODE_ENABLED:
2119 switch (op) {
2120 case QLA2XXX_INI_MODE_ENABLED:
2121 if (NEED_EXCH_OFFLOAD(vha->u_ql2xiniexchg) !=
2122 vha->hw->flags.exchoffld_enabled)
2123 eo_toggle = 1;
2124 if (((vha->ql2xiniexchg != vha->u_ql2xiniexchg) &&
2125 NEED_EXCH_OFFLOAD(vha->u_ql2xiniexchg)) ||
2126 eo_toggle)
2127 action = MODE_CHANGE_ACCEPT;
2128 else
2129 action = NO_ACTION;
2130 break;
2131 case QLA2XXX_INI_MODE_DUAL:
2132 case QLA2XXX_INI_MODE_DISABLED:
2133 action = MODE_CHANGE_ACCEPT;
2134 break;
2135 default:
2136 action = MODE_CHANGE_NO_ACTION;
2137 break;
2138 }
2139 break;
2140
2141 case QLA2XXX_INI_MODE_DUAL:
2142 switch (op) {
2143 case QLA2XXX_INI_MODE_DUAL:
2144 if (qla_tgt_mode_enabled(vha) ||
2145 qla_dual_mode_enabled(vha)) {
2146 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld +
2147 vha->u_ql2xiniexchg) !=
2148 vha->hw->flags.exchoffld_enabled)
2149 eo_toggle = 1;
2150
2151 if ((((vha->ql2xexchoffld +
2152 vha->ql2xiniexchg) !=
2153 (vha->u_ql2xiniexchg +
2154 vha->u_ql2xexchoffld)) &&
2155 NEED_EXCH_OFFLOAD(vha->u_ql2xiniexchg +
2156 vha->u_ql2xexchoffld)) || eo_toggle)
2157 action = MODE_CHANGE_ACCEPT;
2158 else
2159 action = NO_ACTION;
2160 } else {
2161 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld +
2162 vha->u_ql2xiniexchg) !=
2163 vha->hw->flags.exchoffld_enabled)
2164 eo_toggle = 1;
2165
2166 if ((((vha->ql2xexchoffld + vha->ql2xiniexchg)
2167 != (vha->u_ql2xiniexchg +
2168 vha->u_ql2xexchoffld)) &&
2169 NEED_EXCH_OFFLOAD(vha->u_ql2xiniexchg +
2170 vha->u_ql2xexchoffld)) || eo_toggle)
2171 action = MODE_CHANGE_NO_ACTION;
2172 else
2173 action = NO_ACTION;
2174 }
2175 break;
2176
2177 case QLA2XXX_INI_MODE_DISABLED:
2178 if (qla_tgt_mode_enabled(vha) ||
2179 qla_dual_mode_enabled(vha)) {
2180 /* turning off initiator mode */
2181 set_mode = 1;
2182 action = MODE_CHANGE_ACCEPT;
2183 } else {
2184 action = MODE_CHANGE_NO_ACTION;
2185 }
2186 break;
2187
2188 case QLA2XXX_INI_MODE_EXCLUSIVE:
2189 if (qla_tgt_mode_enabled(vha) ||
2190 qla_dual_mode_enabled(vha)) {
2191 set_mode = 1;
2192 action = MODE_CHANGE_ACCEPT;
2193 } else {
2194 action = MODE_CHANGE_ACCEPT;
2195 }
2196 break;
2197
2198 case QLA2XXX_INI_MODE_ENABLED:
2199 if (qla_tgt_mode_enabled(vha) ||
2200 qla_dual_mode_enabled(vha)) {
2201 action = TARGET_STILL_ACTIVE;
2202 } else {
2203 action = MODE_CHANGE_ACCEPT;
2204 }
2205 }
2206 break;
2207 }
2208
2209 switch (action) {
2210 case MODE_CHANGE_ACCEPT:
2211 ql_log(ql_log_warn, vha, 0xffff,
2212 "Mode change accepted. From %s to %s, Tgt exchg %d|%d. ini exchg %d|%d\n",
2213 mode_to_str[vha->qlini_mode], mode_to_str[op],
2214 vha->ql2xexchoffld, vha->u_ql2xexchoffld,
2215 vha->ql2xiniexchg, vha->u_ql2xiniexchg);
2216
2217 vha->qlini_mode = op;
2218 vha->ql2xexchoffld = vha->u_ql2xexchoffld;
2219 vha->ql2xiniexchg = vha->u_ql2xiniexchg;
2220 if (set_mode)
2221 qlt_set_mode(vha);
2222 vha->flags.online = 1;
2223 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
2224 break;
2225
2226 case MODE_CHANGE_NO_ACTION:
2227 ql_log(ql_log_warn, vha, 0xffff,
2228 "Mode is set. No action taken. From %s to %s, Tgt exchg %d|%d. ini exchg %d|%d\n",
2229 mode_to_str[vha->qlini_mode], mode_to_str[op],
2230 vha->ql2xexchoffld, vha->u_ql2xexchoffld,
2231 vha->ql2xiniexchg, vha->u_ql2xiniexchg);
2232 vha->qlini_mode = op;
2233 vha->ql2xexchoffld = vha->u_ql2xexchoffld;
2234 vha->ql2xiniexchg = vha->u_ql2xiniexchg;
2235 break;
2236
2237 case TARGET_STILL_ACTIVE:
2238 ql_log(ql_log_warn, vha, 0xffff,
2239 "Target Mode is active. Unable to change Mode.\n");
2240 break;
2241
2242 case NO_ACTION:
2243 default:
2244 ql_log(ql_log_warn, vha, 0xffff,
2245 "Mode unchange. No action taken. %d|%d pct %d|%d.\n",
2246 vha->qlini_mode, op,
2247 vha->ql2xexchoffld, vha->u_ql2xexchoffld);
2248 break;
2249 }
2250 }
2251
2252 static ssize_t
qlini_mode_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)2253 qlini_mode_store(struct device *dev, struct device_attribute *attr,
2254 const char *buf, size_t count)
2255 {
2256 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2257 int ini;
2258
2259 if (!buf)
2260 return -EINVAL;
2261
2262 if (strncasecmp(QLA2XXX_INI_MODE_STR_EXCLUSIVE, buf,
2263 strlen(QLA2XXX_INI_MODE_STR_EXCLUSIVE)) == 0)
2264 ini = QLA2XXX_INI_MODE_EXCLUSIVE;
2265 else if (strncasecmp(QLA2XXX_INI_MODE_STR_DISABLED, buf,
2266 strlen(QLA2XXX_INI_MODE_STR_DISABLED)) == 0)
2267 ini = QLA2XXX_INI_MODE_DISABLED;
2268 else if (strncasecmp(QLA2XXX_INI_MODE_STR_ENABLED, buf,
2269 strlen(QLA2XXX_INI_MODE_STR_ENABLED)) == 0)
2270 ini = QLA2XXX_INI_MODE_ENABLED;
2271 else if (strncasecmp(QLA2XXX_INI_MODE_STR_DUAL, buf,
2272 strlen(QLA2XXX_INI_MODE_STR_DUAL)) == 0)
2273 ini = QLA2XXX_INI_MODE_DUAL;
2274 else
2275 return -EINVAL;
2276
2277 qla_set_ini_mode(vha, ini);
2278 return strlen(buf);
2279 }
2280
2281 static ssize_t
ql2xexchoffld_show(struct device * dev,struct device_attribute * attr,char * buf)2282 ql2xexchoffld_show(struct device *dev, struct device_attribute *attr,
2283 char *buf)
2284 {
2285 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2286 int len = 0;
2287
2288 len += scnprintf(buf + len, PAGE_SIZE-len,
2289 "target exchange: new %d : current: %d\n\n",
2290 vha->u_ql2xexchoffld, vha->ql2xexchoffld);
2291
2292 len += scnprintf(buf + len, PAGE_SIZE-len,
2293 "Please (re)set operating mode via \"/sys/class/scsi_host/host%ld/qlini_mode\" to load new setting.\n",
2294 vha->host_no);
2295
2296 return len;
2297 }
2298
2299 static ssize_t
ql2xexchoffld_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)2300 ql2xexchoffld_store(struct device *dev, struct device_attribute *attr,
2301 const char *buf, size_t count)
2302 {
2303 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2304 int val = 0;
2305
2306 if (sscanf(buf, "%d", &val) != 1)
2307 return -EINVAL;
2308
2309 if (val > FW_MAX_EXCHANGES_CNT)
2310 val = FW_MAX_EXCHANGES_CNT;
2311 else if (val < 0)
2312 val = 0;
2313
2314 vha->u_ql2xexchoffld = val;
2315 return strlen(buf);
2316 }
2317
2318 static ssize_t
ql2xiniexchg_show(struct device * dev,struct device_attribute * attr,char * buf)2319 ql2xiniexchg_show(struct device *dev, struct device_attribute *attr,
2320 char *buf)
2321 {
2322 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2323 int len = 0;
2324
2325 len += scnprintf(buf + len, PAGE_SIZE-len,
2326 "target exchange: new %d : current: %d\n\n",
2327 vha->u_ql2xiniexchg, vha->ql2xiniexchg);
2328
2329 len += scnprintf(buf + len, PAGE_SIZE-len,
2330 "Please (re)set operating mode via \"/sys/class/scsi_host/host%ld/qlini_mode\" to load new setting.\n",
2331 vha->host_no);
2332
2333 return len;
2334 }
2335
2336 static ssize_t
ql2xiniexchg_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)2337 ql2xiniexchg_store(struct device *dev, struct device_attribute *attr,
2338 const char *buf, size_t count)
2339 {
2340 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2341 int val = 0;
2342
2343 if (sscanf(buf, "%d", &val) != 1)
2344 return -EINVAL;
2345
2346 if (val > FW_MAX_EXCHANGES_CNT)
2347 val = FW_MAX_EXCHANGES_CNT;
2348 else if (val < 0)
2349 val = 0;
2350
2351 vha->u_ql2xiniexchg = val;
2352 return strlen(buf);
2353 }
2354
2355 static ssize_t
qla2x00_dif_bundle_statistics_show(struct device * dev,struct device_attribute * attr,char * buf)2356 qla2x00_dif_bundle_statistics_show(struct device *dev,
2357 struct device_attribute *attr, char *buf)
2358 {
2359 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2360 struct qla_hw_data *ha = vha->hw;
2361
2362 return scnprintf(buf, PAGE_SIZE,
2363 "cross=%llu read=%llu write=%llu kalloc=%llu dma_alloc=%llu unusable=%u\n",
2364 ha->dif_bundle_crossed_pages, ha->dif_bundle_reads,
2365 ha->dif_bundle_writes, ha->dif_bundle_kallocs,
2366 ha->dif_bundle_dma_allocs, ha->pool.unusable.count);
2367 }
2368
2369 static ssize_t
qla2x00_fw_attr_show(struct device * dev,struct device_attribute * attr,char * buf)2370 qla2x00_fw_attr_show(struct device *dev,
2371 struct device_attribute *attr, char *buf)
2372 {
2373 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2374 struct qla_hw_data *ha = vha->hw;
2375
2376 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha))
2377 return scnprintf(buf, PAGE_SIZE, "\n");
2378
2379 return scnprintf(buf, PAGE_SIZE, "%llx\n",
2380 (uint64_t)ha->fw_attributes_ext[1] << 48 |
2381 (uint64_t)ha->fw_attributes_ext[0] << 32 |
2382 (uint64_t)ha->fw_attributes_h << 16 |
2383 (uint64_t)ha->fw_attributes);
2384 }
2385
2386 static ssize_t
qla2x00_port_no_show(struct device * dev,struct device_attribute * attr,char * buf)2387 qla2x00_port_no_show(struct device *dev, struct device_attribute *attr,
2388 char *buf)
2389 {
2390 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2391
2392 return scnprintf(buf, PAGE_SIZE, "%u\n", vha->hw->port_no);
2393 }
2394
2395 static ssize_t
qla2x00_dport_diagnostics_show(struct device * dev,struct device_attribute * attr,char * buf)2396 qla2x00_dport_diagnostics_show(struct device *dev,
2397 struct device_attribute *attr, char *buf)
2398 {
2399 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2400
2401 if (!IS_QLA83XX(vha->hw) && !IS_QLA27XX(vha->hw) &&
2402 !IS_QLA28XX(vha->hw))
2403 return scnprintf(buf, PAGE_SIZE, "\n");
2404
2405 if (!*vha->dport_data)
2406 return scnprintf(buf, PAGE_SIZE, "\n");
2407
2408 return scnprintf(buf, PAGE_SIZE, "%04x %04x %04x %04x\n",
2409 vha->dport_data[0], vha->dport_data[1],
2410 vha->dport_data[2], vha->dport_data[3]);
2411 }
2412 static DEVICE_ATTR(dport_diagnostics, 0444,
2413 qla2x00_dport_diagnostics_show, NULL);
2414
2415 static DEVICE_ATTR(driver_version, S_IRUGO, qla2x00_driver_version_show, NULL);
2416 static DEVICE_ATTR(fw_version, S_IRUGO, qla2x00_fw_version_show, NULL);
2417 static DEVICE_ATTR(serial_num, S_IRUGO, qla2x00_serial_num_show, NULL);
2418 static DEVICE_ATTR(isp_name, S_IRUGO, qla2x00_isp_name_show, NULL);
2419 static DEVICE_ATTR(isp_id, S_IRUGO, qla2x00_isp_id_show, NULL);
2420 static DEVICE_ATTR(model_name, S_IRUGO, qla2x00_model_name_show, NULL);
2421 static DEVICE_ATTR(model_desc, S_IRUGO, qla2x00_model_desc_show, NULL);
2422 static DEVICE_ATTR(pci_info, S_IRUGO, qla2x00_pci_info_show, NULL);
2423 static DEVICE_ATTR(link_state, S_IRUGO, qla2x00_link_state_show, NULL);
2424 static DEVICE_ATTR(zio, S_IRUGO | S_IWUSR, qla2x00_zio_show, qla2x00_zio_store);
2425 static DEVICE_ATTR(zio_timer, S_IRUGO | S_IWUSR, qla2x00_zio_timer_show,
2426 qla2x00_zio_timer_store);
2427 static DEVICE_ATTR(beacon, S_IRUGO | S_IWUSR, qla2x00_beacon_show,
2428 qla2x00_beacon_store);
2429 static DEVICE_ATTR(beacon_config, 0644, qla2x00_beacon_config_show,
2430 qla2x00_beacon_config_store);
2431 static DEVICE_ATTR(optrom_bios_version, S_IRUGO,
2432 qla2x00_optrom_bios_version_show, NULL);
2433 static DEVICE_ATTR(optrom_efi_version, S_IRUGO,
2434 qla2x00_optrom_efi_version_show, NULL);
2435 static DEVICE_ATTR(optrom_fcode_version, S_IRUGO,
2436 qla2x00_optrom_fcode_version_show, NULL);
2437 static DEVICE_ATTR(optrom_fw_version, S_IRUGO, qla2x00_optrom_fw_version_show,
2438 NULL);
2439 static DEVICE_ATTR(optrom_gold_fw_version, S_IRUGO,
2440 qla2x00_optrom_gold_fw_version_show, NULL);
2441 static DEVICE_ATTR(84xx_fw_version, S_IRUGO, qla24xx_84xx_fw_version_show,
2442 NULL);
2443 static DEVICE_ATTR(total_isp_aborts, S_IRUGO, qla2x00_total_isp_aborts_show,
2444 NULL);
2445 static DEVICE_ATTR(serdes_version, 0444, qla2x00_serdes_version_show, NULL);
2446 static DEVICE_ATTR(mpi_version, S_IRUGO, qla2x00_mpi_version_show, NULL);
2447 static DEVICE_ATTR(phy_version, S_IRUGO, qla2x00_phy_version_show, NULL);
2448 static DEVICE_ATTR(flash_block_size, S_IRUGO, qla2x00_flash_block_size_show,
2449 NULL);
2450 static DEVICE_ATTR(vlan_id, S_IRUGO, qla2x00_vlan_id_show, NULL);
2451 static DEVICE_ATTR(vn_port_mac_address, S_IRUGO,
2452 qla2x00_vn_port_mac_address_show, NULL);
2453 static DEVICE_ATTR(fabric_param, S_IRUGO, qla2x00_fabric_param_show, NULL);
2454 static DEVICE_ATTR(fw_state, S_IRUGO, qla2x00_fw_state_show, NULL);
2455 static DEVICE_ATTR(thermal_temp, S_IRUGO, qla2x00_thermal_temp_show, NULL);
2456 static DEVICE_ATTR(diag_requests, S_IRUGO, qla2x00_diag_requests_show, NULL);
2457 static DEVICE_ATTR(diag_megabytes, S_IRUGO, qla2x00_diag_megabytes_show, NULL);
2458 static DEVICE_ATTR(fw_dump_size, S_IRUGO, qla2x00_fw_dump_size_show, NULL);
2459 static DEVICE_ATTR(allow_cna_fw_dump, S_IRUGO | S_IWUSR,
2460 qla2x00_allow_cna_fw_dump_show,
2461 qla2x00_allow_cna_fw_dump_store);
2462 static DEVICE_ATTR(pep_version, S_IRUGO, qla2x00_pep_version_show, NULL);
2463 static DEVICE_ATTR(min_supported_speed, 0444,
2464 qla2x00_min_supported_speed_show, NULL);
2465 static DEVICE_ATTR(max_supported_speed, 0444,
2466 qla2x00_max_supported_speed_show, NULL);
2467 static DEVICE_ATTR(zio_threshold, 0644,
2468 qla_zio_threshold_show,
2469 qla_zio_threshold_store);
2470 static DEVICE_ATTR_RW(qlini_mode);
2471 static DEVICE_ATTR_RW(ql2xexchoffld);
2472 static DEVICE_ATTR_RW(ql2xiniexchg);
2473 static DEVICE_ATTR(dif_bundle_statistics, 0444,
2474 qla2x00_dif_bundle_statistics_show, NULL);
2475 static DEVICE_ATTR(port_speed, 0644, qla2x00_port_speed_show,
2476 qla2x00_port_speed_store);
2477 static DEVICE_ATTR(port_no, 0444, qla2x00_port_no_show, NULL);
2478 static DEVICE_ATTR(fw_attr, 0444, qla2x00_fw_attr_show, NULL);
2479
2480 static struct attribute *qla2x00_host_attrs[] = {
2481 &dev_attr_driver_version.attr,
2482 &dev_attr_fw_version.attr,
2483 &dev_attr_serial_num.attr,
2484 &dev_attr_isp_name.attr,
2485 &dev_attr_isp_id.attr,
2486 &dev_attr_model_name.attr,
2487 &dev_attr_model_desc.attr,
2488 &dev_attr_pci_info.attr,
2489 &dev_attr_link_state.attr,
2490 &dev_attr_zio.attr,
2491 &dev_attr_zio_timer.attr,
2492 &dev_attr_beacon.attr,
2493 &dev_attr_beacon_config.attr,
2494 &dev_attr_optrom_bios_version.attr,
2495 &dev_attr_optrom_efi_version.attr,
2496 &dev_attr_optrom_fcode_version.attr,
2497 &dev_attr_optrom_fw_version.attr,
2498 &dev_attr_84xx_fw_version.attr,
2499 &dev_attr_total_isp_aborts.attr,
2500 &dev_attr_serdes_version.attr,
2501 &dev_attr_mpi_version.attr,
2502 &dev_attr_phy_version.attr,
2503 &dev_attr_flash_block_size.attr,
2504 &dev_attr_vlan_id.attr,
2505 &dev_attr_vn_port_mac_address.attr,
2506 &dev_attr_fabric_param.attr,
2507 &dev_attr_fw_state.attr,
2508 &dev_attr_optrom_gold_fw_version.attr,
2509 &dev_attr_thermal_temp.attr,
2510 &dev_attr_diag_requests.attr,
2511 &dev_attr_diag_megabytes.attr,
2512 &dev_attr_fw_dump_size.attr,
2513 &dev_attr_allow_cna_fw_dump.attr,
2514 &dev_attr_pep_version.attr,
2515 &dev_attr_min_supported_speed.attr,
2516 &dev_attr_max_supported_speed.attr,
2517 &dev_attr_zio_threshold.attr,
2518 &dev_attr_dif_bundle_statistics.attr,
2519 &dev_attr_port_speed.attr,
2520 &dev_attr_port_no.attr,
2521 &dev_attr_fw_attr.attr,
2522 &dev_attr_dport_diagnostics.attr,
2523 &dev_attr_mpi_pause.attr,
2524 &dev_attr_qlini_mode.attr,
2525 &dev_attr_ql2xiniexchg.attr,
2526 &dev_attr_ql2xexchoffld.attr,
2527 NULL,
2528 };
2529
qla_host_attr_is_visible(struct kobject * kobj,struct attribute * attr,int i)2530 static umode_t qla_host_attr_is_visible(struct kobject *kobj,
2531 struct attribute *attr, int i)
2532 {
2533 if (ql2x_ini_mode != QLA2XXX_INI_MODE_DUAL &&
2534 (attr == &dev_attr_qlini_mode.attr ||
2535 attr == &dev_attr_ql2xiniexchg.attr ||
2536 attr == &dev_attr_ql2xexchoffld.attr))
2537 return 0;
2538 return attr->mode;
2539 }
2540
2541 static const struct attribute_group qla2x00_host_attr_group = {
2542 .is_visible = qla_host_attr_is_visible,
2543 .attrs = qla2x00_host_attrs
2544 };
2545
2546 const struct attribute_group *qla2x00_host_groups[] = {
2547 &qla2x00_host_attr_group,
2548 NULL
2549 };
2550
2551 /* Host attributes. */
2552
2553 static void
qla2x00_get_host_port_id(struct Scsi_Host * shost)2554 qla2x00_get_host_port_id(struct Scsi_Host *shost)
2555 {
2556 scsi_qla_host_t *vha = shost_priv(shost);
2557
2558 fc_host_port_id(shost) = vha->d_id.b.domain << 16 |
2559 vha->d_id.b.area << 8 | vha->d_id.b.al_pa;
2560 }
2561
2562 static void
qla2x00_get_host_speed(struct Scsi_Host * shost)2563 qla2x00_get_host_speed(struct Scsi_Host *shost)
2564 {
2565 scsi_qla_host_t *vha = shost_priv(shost);
2566 u32 speed;
2567
2568 if (IS_QLAFX00(vha->hw)) {
2569 qlafx00_get_host_speed(shost);
2570 return;
2571 }
2572
2573 switch (vha->hw->link_data_rate) {
2574 case PORT_SPEED_1GB:
2575 speed = FC_PORTSPEED_1GBIT;
2576 break;
2577 case PORT_SPEED_2GB:
2578 speed = FC_PORTSPEED_2GBIT;
2579 break;
2580 case PORT_SPEED_4GB:
2581 speed = FC_PORTSPEED_4GBIT;
2582 break;
2583 case PORT_SPEED_8GB:
2584 speed = FC_PORTSPEED_8GBIT;
2585 break;
2586 case PORT_SPEED_10GB:
2587 speed = FC_PORTSPEED_10GBIT;
2588 break;
2589 case PORT_SPEED_16GB:
2590 speed = FC_PORTSPEED_16GBIT;
2591 break;
2592 case PORT_SPEED_32GB:
2593 speed = FC_PORTSPEED_32GBIT;
2594 break;
2595 case PORT_SPEED_64GB:
2596 speed = FC_PORTSPEED_64GBIT;
2597 break;
2598 default:
2599 speed = FC_PORTSPEED_UNKNOWN;
2600 break;
2601 }
2602
2603 fc_host_speed(shost) = speed;
2604 }
2605
2606 static void
qla2x00_get_host_port_type(struct Scsi_Host * shost)2607 qla2x00_get_host_port_type(struct Scsi_Host *shost)
2608 {
2609 scsi_qla_host_t *vha = shost_priv(shost);
2610 uint32_t port_type;
2611
2612 if (vha->vp_idx) {
2613 fc_host_port_type(shost) = FC_PORTTYPE_NPIV;
2614 return;
2615 }
2616 switch (vha->hw->current_topology) {
2617 case ISP_CFG_NL:
2618 port_type = FC_PORTTYPE_LPORT;
2619 break;
2620 case ISP_CFG_FL:
2621 port_type = FC_PORTTYPE_NLPORT;
2622 break;
2623 case ISP_CFG_N:
2624 port_type = FC_PORTTYPE_PTP;
2625 break;
2626 case ISP_CFG_F:
2627 port_type = FC_PORTTYPE_NPORT;
2628 break;
2629 default:
2630 port_type = FC_PORTTYPE_UNKNOWN;
2631 break;
2632 }
2633
2634 fc_host_port_type(shost) = port_type;
2635 }
2636
2637 static void
qla2x00_get_starget_node_name(struct scsi_target * starget)2638 qla2x00_get_starget_node_name(struct scsi_target *starget)
2639 {
2640 struct Scsi_Host *host = dev_to_shost(starget->dev.parent);
2641 scsi_qla_host_t *vha = shost_priv(host);
2642 fc_port_t *fcport;
2643 u64 node_name = 0;
2644
2645 list_for_each_entry(fcport, &vha->vp_fcports, list) {
2646 if (fcport->rport &&
2647 starget->id == fcport->rport->scsi_target_id) {
2648 node_name = wwn_to_u64(fcport->node_name);
2649 break;
2650 }
2651 }
2652
2653 fc_starget_node_name(starget) = node_name;
2654 }
2655
2656 static void
qla2x00_get_starget_port_name(struct scsi_target * starget)2657 qla2x00_get_starget_port_name(struct scsi_target *starget)
2658 {
2659 struct Scsi_Host *host = dev_to_shost(starget->dev.parent);
2660 scsi_qla_host_t *vha = shost_priv(host);
2661 fc_port_t *fcport;
2662 u64 port_name = 0;
2663
2664 list_for_each_entry(fcport, &vha->vp_fcports, list) {
2665 if (fcport->rport &&
2666 starget->id == fcport->rport->scsi_target_id) {
2667 port_name = wwn_to_u64(fcport->port_name);
2668 break;
2669 }
2670 }
2671
2672 fc_starget_port_name(starget) = port_name;
2673 }
2674
2675 static void
qla2x00_get_starget_port_id(struct scsi_target * starget)2676 qla2x00_get_starget_port_id(struct scsi_target *starget)
2677 {
2678 struct Scsi_Host *host = dev_to_shost(starget->dev.parent);
2679 scsi_qla_host_t *vha = shost_priv(host);
2680 fc_port_t *fcport;
2681 uint32_t port_id = ~0U;
2682
2683 list_for_each_entry(fcport, &vha->vp_fcports, list) {
2684 if (fcport->rport &&
2685 starget->id == fcport->rport->scsi_target_id) {
2686 port_id = fcport->d_id.b.domain << 16 |
2687 fcport->d_id.b.area << 8 | fcport->d_id.b.al_pa;
2688 break;
2689 }
2690 }
2691
2692 fc_starget_port_id(starget) = port_id;
2693 }
2694
2695 static inline void
qla2x00_set_rport_loss_tmo(struct fc_rport * rport,uint32_t timeout)2696 qla2x00_set_rport_loss_tmo(struct fc_rport *rport, uint32_t timeout)
2697 {
2698 fc_port_t *fcport = *(fc_port_t **)rport->dd_data;
2699
2700 rport->dev_loss_tmo = timeout ? timeout : 1;
2701
2702 if (IS_ENABLED(CONFIG_NVME_FC) && fcport && fcport->nvme_remote_port)
2703 nvme_fc_set_remoteport_devloss(fcport->nvme_remote_port,
2704 rport->dev_loss_tmo);
2705 }
2706
2707 static void
qla2x00_dev_loss_tmo_callbk(struct fc_rport * rport)2708 qla2x00_dev_loss_tmo_callbk(struct fc_rport *rport)
2709 {
2710 struct Scsi_Host *host = rport_to_shost(rport);
2711 fc_port_t *fcport = *(fc_port_t **)rport->dd_data;
2712 unsigned long flags;
2713
2714 if (!fcport)
2715 return;
2716
2717 ql_dbg(ql_dbg_async, fcport->vha, 0x5101,
2718 DBG_FCPORT_PRFMT(fcport, "dev_loss_tmo expiry, rport_state=%d",
2719 rport->port_state));
2720
2721 /*
2722 * Now that the rport has been deleted, set the fcport state to
2723 * FCS_DEVICE_DEAD, if the fcport is still lost.
2724 */
2725 if (fcport->scan_state != QLA_FCPORT_FOUND)
2726 qla2x00_set_fcport_state(fcport, FCS_DEVICE_DEAD);
2727
2728 /*
2729 * Transport has effectively 'deleted' the rport, clear
2730 * all local references.
2731 */
2732 spin_lock_irqsave(host->host_lock, flags);
2733 /* Confirm port has not reappeared before clearing pointers. */
2734 if (rport->port_state != FC_PORTSTATE_ONLINE) {
2735 fcport->rport = NULL;
2736 *((fc_port_t **)rport->dd_data) = NULL;
2737 }
2738 spin_unlock_irqrestore(host->host_lock, flags);
2739
2740 if (test_bit(ABORT_ISP_ACTIVE, &fcport->vha->dpc_flags))
2741 return;
2742
2743 if (unlikely(pci_channel_offline(fcport->vha->hw->pdev))) {
2744 /* Will wait for wind down of adapter */
2745 ql_dbg(ql_dbg_aer, fcport->vha, 0x900c,
2746 "%s pci offline detected (id %06x)\n", __func__,
2747 fcport->d_id.b24);
2748 qla_pci_set_eeh_busy(fcport->vha);
2749 qla2x00_eh_wait_for_pending_commands(fcport->vha, fcport->d_id.b24,
2750 0, WAIT_TARGET);
2751 return;
2752 }
2753 }
2754
2755 static void
qla2x00_terminate_rport_io(struct fc_rport * rport)2756 qla2x00_terminate_rport_io(struct fc_rport *rport)
2757 {
2758 fc_port_t *fcport = *(fc_port_t **)rport->dd_data;
2759 scsi_qla_host_t *vha;
2760
2761 if (!fcport)
2762 return;
2763
2764 if (test_bit(UNLOADING, &fcport->vha->dpc_flags))
2765 return;
2766
2767 if (test_bit(ABORT_ISP_ACTIVE, &fcport->vha->dpc_flags))
2768 return;
2769 vha = fcport->vha;
2770
2771 if (unlikely(pci_channel_offline(fcport->vha->hw->pdev))) {
2772 /* Will wait for wind down of adapter */
2773 ql_dbg(ql_dbg_aer, fcport->vha, 0x900b,
2774 "%s pci offline detected (id %06x)\n", __func__,
2775 fcport->d_id.b24);
2776 qla_pci_set_eeh_busy(vha);
2777 qla2x00_eh_wait_for_pending_commands(fcport->vha, fcport->d_id.b24,
2778 0, WAIT_TARGET);
2779 return;
2780 }
2781 /*
2782 * At this point all fcport's software-states are cleared. Perform any
2783 * final cleanup of firmware resources (PCBs and XCBs).
2784 *
2785 * Attempt to cleanup only lost devices.
2786 */
2787 if (fcport->loop_id != FC_NO_LOOP_ID) {
2788 if (IS_FWI2_CAPABLE(fcport->vha->hw) &&
2789 fcport->scan_state != QLA_FCPORT_FOUND) {
2790 if (fcport->loop_id != FC_NO_LOOP_ID)
2791 fcport->logout_on_delete = 1;
2792
2793 if (!EDIF_NEGOTIATION_PENDING(fcport)) {
2794 ql_dbg(ql_dbg_disc, fcport->vha, 0x911e,
2795 "%s %d schedule session deletion\n", __func__,
2796 __LINE__);
2797 qlt_schedule_sess_for_deletion(fcport);
2798 }
2799 } else if (!IS_FWI2_CAPABLE(fcport->vha->hw)) {
2800 qla2x00_port_logout(fcport->vha, fcport);
2801 }
2802 }
2803
2804 /* check for any straggling io left behind */
2805 if (qla2x00_eh_wait_for_pending_commands(fcport->vha, fcport->d_id.b24, 0, WAIT_TARGET)) {
2806 ql_log(ql_log_warn, vha, 0x300b,
2807 "IO not return. Resetting. \n");
2808 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
2809 qla2xxx_wake_dpc(vha);
2810 qla2x00_wait_for_chip_reset(vha);
2811 }
2812 }
2813
2814 static int
qla2x00_issue_lip(struct Scsi_Host * shost)2815 qla2x00_issue_lip(struct Scsi_Host *shost)
2816 {
2817 scsi_qla_host_t *vha = shost_priv(shost);
2818
2819 if (IS_QLAFX00(vha->hw))
2820 return 0;
2821
2822 if (vha->hw->flags.port_isolated)
2823 return 0;
2824
2825 qla2x00_loop_reset(vha);
2826 return 0;
2827 }
2828
2829 static struct fc_host_statistics *
qla2x00_get_fc_host_stats(struct Scsi_Host * shost)2830 qla2x00_get_fc_host_stats(struct Scsi_Host *shost)
2831 {
2832 scsi_qla_host_t *vha = shost_priv(shost);
2833 struct qla_hw_data *ha = vha->hw;
2834 struct scsi_qla_host *base_vha = pci_get_drvdata(ha->pdev);
2835 int rval;
2836 struct link_statistics *stats;
2837 dma_addr_t stats_dma;
2838 struct fc_host_statistics *p = &vha->fc_host_stat;
2839 struct qla_qpair *qpair;
2840 int i;
2841 u64 ib = 0, ob = 0, ir = 0, or = 0;
2842
2843 memset(p, -1, sizeof(*p));
2844
2845 if (IS_QLAFX00(vha->hw))
2846 goto done;
2847
2848 if (test_bit(UNLOADING, &vha->dpc_flags))
2849 goto done;
2850
2851 if (unlikely(pci_channel_offline(ha->pdev)))
2852 goto done;
2853
2854 if (qla2x00_chip_is_down(vha))
2855 goto done;
2856
2857 stats = dma_alloc_coherent(&ha->pdev->dev, sizeof(*stats), &stats_dma,
2858 GFP_KERNEL);
2859 if (!stats) {
2860 ql_log(ql_log_warn, vha, 0x707d,
2861 "Failed to allocate memory for stats.\n");
2862 goto done;
2863 }
2864
2865 rval = QLA_FUNCTION_FAILED;
2866 if (IS_FWI2_CAPABLE(ha)) {
2867 rval = qla24xx_get_isp_stats(base_vha, stats, stats_dma, 0);
2868 } else if (atomic_read(&base_vha->loop_state) == LOOP_READY &&
2869 !ha->dpc_active) {
2870 /* Must be in a 'READY' state for statistics retrieval. */
2871 rval = qla2x00_get_link_status(base_vha, base_vha->loop_id,
2872 stats, stats_dma);
2873 }
2874
2875 if (rval != QLA_SUCCESS)
2876 goto done_free;
2877
2878 /* --- */
2879 for (i = 0; i < vha->hw->max_qpairs; i++) {
2880 qpair = vha->hw->queue_pair_map[i];
2881 if (!qpair)
2882 continue;
2883 ir += qpair->counters.input_requests;
2884 or += qpair->counters.output_requests;
2885 ib += qpair->counters.input_bytes;
2886 ob += qpair->counters.output_bytes;
2887 }
2888 ir += ha->base_qpair->counters.input_requests;
2889 or += ha->base_qpair->counters.output_requests;
2890 ib += ha->base_qpair->counters.input_bytes;
2891 ob += ha->base_qpair->counters.output_bytes;
2892
2893 ir += vha->qla_stats.input_requests;
2894 or += vha->qla_stats.output_requests;
2895 ib += vha->qla_stats.input_bytes;
2896 ob += vha->qla_stats.output_bytes;
2897 /* --- */
2898
2899 p->link_failure_count = le32_to_cpu(stats->link_fail_cnt);
2900 p->loss_of_sync_count = le32_to_cpu(stats->loss_sync_cnt);
2901 p->loss_of_signal_count = le32_to_cpu(stats->loss_sig_cnt);
2902 p->prim_seq_protocol_err_count = le32_to_cpu(stats->prim_seq_err_cnt);
2903 p->invalid_tx_word_count = le32_to_cpu(stats->inval_xmit_word_cnt);
2904 p->invalid_crc_count = le32_to_cpu(stats->inval_crc_cnt);
2905 if (IS_FWI2_CAPABLE(ha)) {
2906 p->lip_count = le32_to_cpu(stats->lip_cnt);
2907 p->tx_frames = le32_to_cpu(stats->tx_frames);
2908 p->rx_frames = le32_to_cpu(stats->rx_frames);
2909 p->dumped_frames = le32_to_cpu(stats->discarded_frames);
2910 p->nos_count = le32_to_cpu(stats->nos_rcvd);
2911 p->error_frames =
2912 le32_to_cpu(stats->dropped_frames) +
2913 le32_to_cpu(stats->discarded_frames);
2914 if (IS_QLA83XX(ha) || IS_QLA27XX(ha) || IS_QLA28XX(ha)) {
2915 p->rx_words = le64_to_cpu(stats->fpm_recv_word_cnt);
2916 p->tx_words = le64_to_cpu(stats->fpm_xmit_word_cnt);
2917 } else {
2918 p->rx_words = ib >> 2;
2919 p->tx_words = ob >> 2;
2920 }
2921 }
2922
2923 p->fcp_control_requests = vha->qla_stats.control_requests;
2924 p->fcp_input_requests = ir;
2925 p->fcp_output_requests = or;
2926 p->fcp_input_megabytes = ib >> 20;
2927 p->fcp_output_megabytes = ob >> 20;
2928 p->seconds_since_last_reset =
2929 get_jiffies_64() - vha->qla_stats.jiffies_at_last_reset;
2930 do_div(p->seconds_since_last_reset, HZ);
2931
2932 done_free:
2933 dma_free_coherent(&ha->pdev->dev, sizeof(struct link_statistics),
2934 stats, stats_dma);
2935 done:
2936 return p;
2937 }
2938
2939 static void
qla2x00_reset_host_stats(struct Scsi_Host * shost)2940 qla2x00_reset_host_stats(struct Scsi_Host *shost)
2941 {
2942 scsi_qla_host_t *vha = shost_priv(shost);
2943 struct qla_hw_data *ha = vha->hw;
2944 struct scsi_qla_host *base_vha = pci_get_drvdata(ha->pdev);
2945 struct link_statistics *stats;
2946 dma_addr_t stats_dma;
2947 int i;
2948 struct qla_qpair *qpair;
2949
2950 memset(&vha->qla_stats, 0, sizeof(vha->qla_stats));
2951 memset(&vha->fc_host_stat, 0, sizeof(vha->fc_host_stat));
2952 for (i = 0; i < vha->hw->max_qpairs; i++) {
2953 qpair = vha->hw->queue_pair_map[i];
2954 if (!qpair)
2955 continue;
2956 memset(&qpair->counters, 0, sizeof(qpair->counters));
2957 }
2958 memset(&ha->base_qpair->counters, 0, sizeof(qpair->counters));
2959
2960 vha->qla_stats.jiffies_at_last_reset = get_jiffies_64();
2961
2962 if (IS_FWI2_CAPABLE(ha)) {
2963 int rval;
2964
2965 stats = dma_alloc_coherent(&ha->pdev->dev,
2966 sizeof(*stats), &stats_dma, GFP_KERNEL);
2967 if (!stats) {
2968 ql_log(ql_log_warn, vha, 0x70d7,
2969 "Failed to allocate memory for stats.\n");
2970 return;
2971 }
2972
2973 /* reset firmware statistics */
2974 rval = qla24xx_get_isp_stats(base_vha, stats, stats_dma, BIT_0);
2975 if (rval != QLA_SUCCESS)
2976 ql_log(ql_log_warn, vha, 0x70de,
2977 "Resetting ISP statistics failed: rval = %d\n",
2978 rval);
2979
2980 dma_free_coherent(&ha->pdev->dev, sizeof(*stats),
2981 stats, stats_dma);
2982 }
2983 }
2984
2985 static void
qla2x00_get_host_symbolic_name(struct Scsi_Host * shost)2986 qla2x00_get_host_symbolic_name(struct Scsi_Host *shost)
2987 {
2988 scsi_qla_host_t *vha = shost_priv(shost);
2989
2990 qla2x00_get_sym_node_name(vha, fc_host_symbolic_name(shost),
2991 sizeof(fc_host_symbolic_name(shost)));
2992 }
2993
2994 static void
qla2x00_set_host_system_hostname(struct Scsi_Host * shost)2995 qla2x00_set_host_system_hostname(struct Scsi_Host *shost)
2996 {
2997 scsi_qla_host_t *vha = shost_priv(shost);
2998
2999 set_bit(REGISTER_FDMI_NEEDED, &vha->dpc_flags);
3000 }
3001
3002 static void
qla2x00_get_host_fabric_name(struct Scsi_Host * shost)3003 qla2x00_get_host_fabric_name(struct Scsi_Host *shost)
3004 {
3005 scsi_qla_host_t *vha = shost_priv(shost);
3006 static const uint8_t node_name[WWN_SIZE] = {
3007 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF
3008 };
3009 u64 fabric_name = wwn_to_u64(node_name);
3010
3011 if (vha->device_flags & SWITCH_FOUND)
3012 fabric_name = wwn_to_u64(vha->fabric_node_name);
3013
3014 fc_host_fabric_name(shost) = fabric_name;
3015 }
3016
3017 static void
qla2x00_get_host_port_state(struct Scsi_Host * shost)3018 qla2x00_get_host_port_state(struct Scsi_Host *shost)
3019 {
3020 scsi_qla_host_t *vha = shost_priv(shost);
3021 struct scsi_qla_host *base_vha = pci_get_drvdata(vha->hw->pdev);
3022
3023 if (!base_vha->flags.online) {
3024 fc_host_port_state(shost) = FC_PORTSTATE_OFFLINE;
3025 return;
3026 }
3027
3028 switch (atomic_read(&base_vha->loop_state)) {
3029 case LOOP_UPDATE:
3030 fc_host_port_state(shost) = FC_PORTSTATE_DIAGNOSTICS;
3031 break;
3032 case LOOP_DOWN:
3033 if (test_bit(LOOP_RESYNC_NEEDED, &base_vha->dpc_flags))
3034 fc_host_port_state(shost) = FC_PORTSTATE_DIAGNOSTICS;
3035 else
3036 fc_host_port_state(shost) = FC_PORTSTATE_LINKDOWN;
3037 break;
3038 case LOOP_DEAD:
3039 fc_host_port_state(shost) = FC_PORTSTATE_LINKDOWN;
3040 break;
3041 case LOOP_READY:
3042 fc_host_port_state(shost) = FC_PORTSTATE_ONLINE;
3043 break;
3044 default:
3045 fc_host_port_state(shost) = FC_PORTSTATE_UNKNOWN;
3046 break;
3047 }
3048 }
3049
3050 static int
qla24xx_vport_create(struct fc_vport * fc_vport,bool disable)3051 qla24xx_vport_create(struct fc_vport *fc_vport, bool disable)
3052 {
3053 int ret = 0;
3054 uint8_t qos = 0;
3055 scsi_qla_host_t *base_vha = shost_priv(fc_vport->shost);
3056 scsi_qla_host_t *vha = NULL;
3057 struct qla_hw_data *ha = base_vha->hw;
3058 int cnt;
3059 struct req_que *req = ha->req_q_map[0];
3060 struct qla_qpair *qpair;
3061
3062 ret = qla24xx_vport_create_req_sanity_check(fc_vport);
3063 if (ret) {
3064 ql_log(ql_log_warn, vha, 0x707e,
3065 "Vport sanity check failed, status %x\n", ret);
3066 return (ret);
3067 }
3068
3069 vha = qla24xx_create_vhost(fc_vport);
3070 if (vha == NULL) {
3071 ql_log(ql_log_warn, vha, 0x707f, "Vport create host failed.\n");
3072 return FC_VPORT_FAILED;
3073 }
3074 if (disable) {
3075 atomic_set(&vha->vp_state, VP_OFFLINE);
3076 fc_vport_set_state(fc_vport, FC_VPORT_DISABLED);
3077 } else
3078 atomic_set(&vha->vp_state, VP_FAILED);
3079
3080 /* ready to create vport */
3081 ql_log(ql_log_info, vha, 0x7080,
3082 "VP entry id %d assigned.\n", vha->vp_idx);
3083
3084 /* initialized vport states */
3085 atomic_set(&vha->loop_state, LOOP_DOWN);
3086 vha->vp_err_state = VP_ERR_PORTDWN;
3087 vha->vp_prev_err_state = VP_ERR_UNKWN;
3088 /* Check if physical ha port is Up */
3089 if (atomic_read(&base_vha->loop_state) == LOOP_DOWN ||
3090 atomic_read(&base_vha->loop_state) == LOOP_DEAD) {
3091 /* Don't retry or attempt login of this virtual port */
3092 ql_dbg(ql_dbg_user, vha, 0x7081,
3093 "Vport loop state is not UP.\n");
3094 atomic_set(&vha->loop_state, LOOP_DEAD);
3095 if (!disable)
3096 fc_vport_set_state(fc_vport, FC_VPORT_LINKDOWN);
3097 }
3098
3099 if (IS_T10_PI_CAPABLE(ha) && ql2xenabledif) {
3100 if (ha->fw_attributes & BIT_4) {
3101 int prot = 0, guard;
3102
3103 vha->flags.difdix_supported = 1;
3104 ql_dbg(ql_dbg_user, vha, 0x7082,
3105 "Registered for DIF/DIX type 1 and 3 protection.\n");
3106 scsi_host_set_prot(vha->host,
3107 prot | SHOST_DIF_TYPE1_PROTECTION
3108 | SHOST_DIF_TYPE2_PROTECTION
3109 | SHOST_DIF_TYPE3_PROTECTION
3110 | SHOST_DIX_TYPE1_PROTECTION
3111 | SHOST_DIX_TYPE2_PROTECTION
3112 | SHOST_DIX_TYPE3_PROTECTION);
3113
3114 guard = SHOST_DIX_GUARD_CRC;
3115
3116 if (IS_PI_IPGUARD_CAPABLE(ha) &&
3117 (ql2xenabledif > 1 || IS_PI_DIFB_DIX0_CAPABLE(ha)))
3118 guard |= SHOST_DIX_GUARD_IP;
3119
3120 scsi_host_set_guard(vha->host, guard);
3121 } else
3122 vha->flags.difdix_supported = 0;
3123 }
3124
3125 if (scsi_add_host_with_dma(vha->host, &fc_vport->dev,
3126 &ha->pdev->dev)) {
3127 ql_dbg(ql_dbg_user, vha, 0x7083,
3128 "scsi_add_host failure for VP[%d].\n", vha->vp_idx);
3129 goto vport_create_failed_2;
3130 }
3131
3132 /* initialize attributes */
3133 fc_host_dev_loss_tmo(vha->host) = ha->port_down_retry_count;
3134 fc_host_node_name(vha->host) = wwn_to_u64(vha->node_name);
3135 fc_host_port_name(vha->host) = wwn_to_u64(vha->port_name);
3136 fc_host_supported_classes(vha->host) =
3137 fc_host_supported_classes(base_vha->host);
3138 fc_host_supported_speeds(vha->host) =
3139 fc_host_supported_speeds(base_vha->host);
3140
3141 qlt_vport_create(vha, ha);
3142 qla24xx_vport_disable(fc_vport, disable);
3143
3144 if (!ql2xmqsupport || !ha->npiv_info)
3145 goto vport_queue;
3146
3147 /* Create a request queue in QoS mode for the vport */
3148 for (cnt = 0; cnt < ha->nvram_npiv_size; cnt++) {
3149 if (memcmp(ha->npiv_info[cnt].port_name, vha->port_name, 8) == 0
3150 && memcmp(ha->npiv_info[cnt].node_name, vha->node_name,
3151 8) == 0) {
3152 qos = ha->npiv_info[cnt].q_qos;
3153 break;
3154 }
3155 }
3156
3157 if (qos) {
3158 qpair = qla2xxx_create_qpair(vha, qos, vha->vp_idx, true);
3159 if (!qpair)
3160 ql_log(ql_log_warn, vha, 0x7084,
3161 "Can't create qpair for VP[%d]\n",
3162 vha->vp_idx);
3163 else {
3164 ql_dbg(ql_dbg_multiq, vha, 0xc001,
3165 "Queue pair: %d Qos: %d) created for VP[%d]\n",
3166 qpair->id, qos, vha->vp_idx);
3167 ql_dbg(ql_dbg_user, vha, 0x7085,
3168 "Queue Pair: %d Qos: %d) created for VP[%d]\n",
3169 qpair->id, qos, vha->vp_idx);
3170 req = qpair->req;
3171 vha->qpair = qpair;
3172 }
3173 }
3174
3175 vport_queue:
3176 vha->req = req;
3177 return 0;
3178
3179 vport_create_failed_2:
3180 qla24xx_disable_vp(vha);
3181 qla24xx_deallocate_vp_id(vha);
3182 scsi_host_put(vha->host);
3183 return FC_VPORT_FAILED;
3184 }
3185
3186 static int
qla24xx_vport_delete(struct fc_vport * fc_vport)3187 qla24xx_vport_delete(struct fc_vport *fc_vport)
3188 {
3189 scsi_qla_host_t *vha = fc_vport->dd_data;
3190 struct qla_hw_data *ha = vha->hw;
3191 uint16_t id = vha->vp_idx;
3192
3193 set_bit(VPORT_DELETE, &vha->dpc_flags);
3194
3195 while (test_bit(LOOP_RESYNC_ACTIVE, &vha->dpc_flags))
3196 msleep(1000);
3197
3198
3199 qla24xx_disable_vp(vha);
3200 qla2x00_wait_for_sess_deletion(vha);
3201
3202 qla_nvme_delete(vha);
3203 qla_enode_stop(vha);
3204 qla_edb_stop(vha);
3205
3206 vha->flags.delete_progress = 1;
3207
3208 qlt_remove_target(ha, vha);
3209
3210 fc_remove_host(vha->host);
3211
3212 scsi_remove_host(vha->host);
3213
3214 /* Allow timer to run to drain queued items, when removing vp */
3215 qla24xx_deallocate_vp_id(vha);
3216
3217 if (vha->timer_active) {
3218 qla2x00_vp_stop_timer(vha);
3219 ql_dbg(ql_dbg_user, vha, 0x7086,
3220 "Timer for the VP[%d] has stopped\n", vha->vp_idx);
3221 }
3222
3223 qla2x00_free_fcports(vha);
3224
3225 mutex_lock(&ha->vport_lock);
3226 ha->cur_vport_count--;
3227 clear_bit(vha->vp_idx, ha->vp_idx_map);
3228 mutex_unlock(&ha->vport_lock);
3229
3230 dma_free_coherent(&ha->pdev->dev, vha->gnl.size, vha->gnl.l,
3231 vha->gnl.ldma);
3232
3233 vha->gnl.l = NULL;
3234
3235 vfree(vha->scan.l);
3236
3237 if (vha->qpair && vha->qpair->vp_idx == vha->vp_idx) {
3238 if (qla2xxx_delete_qpair(vha, vha->qpair) != QLA_SUCCESS)
3239 ql_log(ql_log_warn, vha, 0x7087,
3240 "Queue Pair delete failed.\n");
3241 }
3242
3243 ql_log(ql_log_info, vha, 0x7088, "VP[%d] deleted.\n", id);
3244 scsi_host_put(vha->host);
3245 return 0;
3246 }
3247
3248 static int
qla24xx_vport_disable(struct fc_vport * fc_vport,bool disable)3249 qla24xx_vport_disable(struct fc_vport *fc_vport, bool disable)
3250 {
3251 scsi_qla_host_t *vha = fc_vport->dd_data;
3252
3253 if (disable)
3254 qla24xx_disable_vp(vha);
3255 else
3256 qla24xx_enable_vp(vha);
3257
3258 return 0;
3259 }
3260
3261 struct fc_function_template qla2xxx_transport_functions = {
3262
3263 .show_host_node_name = 1,
3264 .show_host_port_name = 1,
3265 .show_host_supported_classes = 1,
3266 .show_host_supported_speeds = 1,
3267
3268 .get_host_port_id = qla2x00_get_host_port_id,
3269 .show_host_port_id = 1,
3270 .get_host_speed = qla2x00_get_host_speed,
3271 .show_host_speed = 1,
3272 .get_host_port_type = qla2x00_get_host_port_type,
3273 .show_host_port_type = 1,
3274 .get_host_symbolic_name = qla2x00_get_host_symbolic_name,
3275 .show_host_symbolic_name = 1,
3276 .set_host_system_hostname = qla2x00_set_host_system_hostname,
3277 .show_host_system_hostname = 1,
3278 .get_host_fabric_name = qla2x00_get_host_fabric_name,
3279 .show_host_fabric_name = 1,
3280 .get_host_port_state = qla2x00_get_host_port_state,
3281 .show_host_port_state = 1,
3282
3283 .dd_fcrport_size = sizeof(struct fc_port *),
3284 .show_rport_supported_classes = 1,
3285
3286 .get_starget_node_name = qla2x00_get_starget_node_name,
3287 .show_starget_node_name = 1,
3288 .get_starget_port_name = qla2x00_get_starget_port_name,
3289 .show_starget_port_name = 1,
3290 .get_starget_port_id = qla2x00_get_starget_port_id,
3291 .show_starget_port_id = 1,
3292
3293 .set_rport_dev_loss_tmo = qla2x00_set_rport_loss_tmo,
3294 .show_rport_dev_loss_tmo = 1,
3295
3296 .issue_fc_host_lip = qla2x00_issue_lip,
3297 .dev_loss_tmo_callbk = qla2x00_dev_loss_tmo_callbk,
3298 .terminate_rport_io = qla2x00_terminate_rport_io,
3299 .get_fc_host_stats = qla2x00_get_fc_host_stats,
3300 .reset_fc_host_stats = qla2x00_reset_host_stats,
3301
3302 .vport_create = qla24xx_vport_create,
3303 .vport_disable = qla24xx_vport_disable,
3304 .vport_delete = qla24xx_vport_delete,
3305 .bsg_request = qla24xx_bsg_request,
3306 .bsg_timeout = qla24xx_bsg_timeout,
3307 };
3308
3309 struct fc_function_template qla2xxx_transport_vport_functions = {
3310
3311 .show_host_node_name = 1,
3312 .show_host_port_name = 1,
3313 .show_host_supported_classes = 1,
3314
3315 .get_host_port_id = qla2x00_get_host_port_id,
3316 .show_host_port_id = 1,
3317 .get_host_speed = qla2x00_get_host_speed,
3318 .show_host_speed = 1,
3319 .get_host_port_type = qla2x00_get_host_port_type,
3320 .show_host_port_type = 1,
3321 .get_host_symbolic_name = qla2x00_get_host_symbolic_name,
3322 .show_host_symbolic_name = 1,
3323 .set_host_system_hostname = qla2x00_set_host_system_hostname,
3324 .show_host_system_hostname = 1,
3325 .get_host_fabric_name = qla2x00_get_host_fabric_name,
3326 .show_host_fabric_name = 1,
3327 .get_host_port_state = qla2x00_get_host_port_state,
3328 .show_host_port_state = 1,
3329
3330 .dd_fcrport_size = sizeof(struct fc_port *),
3331 .show_rport_supported_classes = 1,
3332
3333 .get_starget_node_name = qla2x00_get_starget_node_name,
3334 .show_starget_node_name = 1,
3335 .get_starget_port_name = qla2x00_get_starget_port_name,
3336 .show_starget_port_name = 1,
3337 .get_starget_port_id = qla2x00_get_starget_port_id,
3338 .show_starget_port_id = 1,
3339
3340 .set_rport_dev_loss_tmo = qla2x00_set_rport_loss_tmo,
3341 .show_rport_dev_loss_tmo = 1,
3342
3343 .issue_fc_host_lip = qla2x00_issue_lip,
3344 .dev_loss_tmo_callbk = qla2x00_dev_loss_tmo_callbk,
3345 .terminate_rport_io = qla2x00_terminate_rport_io,
3346 .get_fc_host_stats = qla2x00_get_fc_host_stats,
3347 .reset_fc_host_stats = qla2x00_reset_host_stats,
3348
3349 .bsg_request = qla24xx_bsg_request,
3350 .bsg_timeout = qla24xx_bsg_timeout,
3351 };
3352
3353 static uint
qla2x00_get_host_supported_speeds(scsi_qla_host_t * vha,uint speeds)3354 qla2x00_get_host_supported_speeds(scsi_qla_host_t *vha, uint speeds)
3355 {
3356 uint supported_speeds = FC_PORTSPEED_UNKNOWN;
3357
3358 if (speeds & FDMI_PORT_SPEED_64GB)
3359 supported_speeds |= FC_PORTSPEED_64GBIT;
3360 if (speeds & FDMI_PORT_SPEED_32GB)
3361 supported_speeds |= FC_PORTSPEED_32GBIT;
3362 if (speeds & FDMI_PORT_SPEED_16GB)
3363 supported_speeds |= FC_PORTSPEED_16GBIT;
3364 if (speeds & FDMI_PORT_SPEED_8GB)
3365 supported_speeds |= FC_PORTSPEED_8GBIT;
3366 if (speeds & FDMI_PORT_SPEED_4GB)
3367 supported_speeds |= FC_PORTSPEED_4GBIT;
3368 if (speeds & FDMI_PORT_SPEED_2GB)
3369 supported_speeds |= FC_PORTSPEED_2GBIT;
3370 if (speeds & FDMI_PORT_SPEED_1GB)
3371 supported_speeds |= FC_PORTSPEED_1GBIT;
3372
3373 return supported_speeds;
3374 }
3375
3376 void
qla2x00_init_host_attr(scsi_qla_host_t * vha)3377 qla2x00_init_host_attr(scsi_qla_host_t *vha)
3378 {
3379 struct qla_hw_data *ha = vha->hw;
3380 u32 speeds = 0, fdmi_speed = 0;
3381
3382 fc_host_dev_loss_tmo(vha->host) = ha->port_down_retry_count;
3383 fc_host_node_name(vha->host) = wwn_to_u64(vha->node_name);
3384 fc_host_port_name(vha->host) = wwn_to_u64(vha->port_name);
3385 fc_host_supported_classes(vha->host) = ha->base_qpair->enable_class_2 ?
3386 (FC_COS_CLASS2|FC_COS_CLASS3) : FC_COS_CLASS3;
3387 fc_host_max_npiv_vports(vha->host) = ha->max_npiv_vports;
3388 fc_host_npiv_vports_inuse(vha->host) = ha->cur_vport_count;
3389
3390 fdmi_speed = qla25xx_fdmi_port_speed_capability(ha);
3391 speeds = qla2x00_get_host_supported_speeds(vha, fdmi_speed);
3392
3393 fc_host_supported_speeds(vha->host) = speeds;
3394 }
3395