Lines Matching +full:hi +full:- +full:fi
5 * Copyright (c) 2001-2013 ATTO Technology, Inc.
21 * LIMITATION, ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT,
40 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301,
48 * allocate a DMA-able memory area to communicate with the firmware. In
50 * we will allocate a global buffer the first time we need it and re-use
85 a->fm_api_command_done = 1; in complete_fm_api_req()
86 wake_up_interruptible(&a->fm_api_waiter); in complete_fm_api_req()
92 struct esas2r_adapter *a = (struct esas2r_adapter *)sgc->adapter; in get_physaddr_fm_api()
93 int offset = sgc->cur_offset - a->save_offset; in get_physaddr_fm_api()
95 (*addr) = a->firmware.phys + offset; in get_physaddr_fm_api()
96 return a->firmware.orig_len - offset; in get_physaddr_fm_api()
101 struct esas2r_adapter *a = (struct esas2r_adapter *)sgc->adapter; in get_physaddr_fm_api_header()
102 int offset = sgc->cur_offset - a->save_offset; in get_physaddr_fm_api_header()
104 (*addr) = a->firmware.header_buff_phys + offset; in get_physaddr_fm_api_header()
105 return sizeof(struct esas2r_flash_img) - offset; in get_physaddr_fm_api_header()
109 static void do_fm_api(struct esas2r_adapter *a, struct esas2r_flash_img *fi) in do_fm_api() argument
113 if (mutex_lock_interruptible(&a->fm_api_mutex)) { in do_fm_api()
114 fi->status = FI_STAT_BUSY; in do_fm_api()
120 fi->status = FI_STAT_BUSY; in do_fm_api()
124 if (fi == &a->firmware.header) { in do_fm_api()
125 a->firmware.header_buff = dma_alloc_coherent(&a->pcid->dev, in do_fm_api()
129 (dma_addr_t *)&a-> in do_fm_api()
134 if (a->firmware.header_buff == NULL) { in do_fm_api()
136 fi->status = FI_STAT_BUSY; in do_fm_api()
140 memcpy(a->firmware.header_buff, fi, in do_fm_api()
142 a->save_offset = a->firmware.header_buff; in do_fm_api()
143 a->fm_api_sgc.get_phys_addr = in do_fm_api()
146 a->save_offset = (u8 *)fi; in do_fm_api()
147 a->fm_api_sgc.get_phys_addr = in do_fm_api()
151 rq->comp_cb = complete_fm_api_req; in do_fm_api()
152 a->fm_api_command_done = 0; in do_fm_api()
153 a->fm_api_sgc.cur_offset = a->save_offset; in do_fm_api()
155 if (!esas2r_fm_api(a, (struct esas2r_flash_img *)a->save_offset, rq, in do_fm_api()
156 &a->fm_api_sgc)) in do_fm_api()
160 while (!a->fm_api_command_done) in do_fm_api()
161 wait_event_interruptible(a->fm_api_waiter, in do_fm_api()
162 a->fm_api_command_done); in do_fm_api()
164 if (fi == &a->firmware.header) { in do_fm_api()
165 memcpy(fi, a->firmware.header_buff, in do_fm_api()
168 dma_free_coherent(&a->pcid->dev, in do_fm_api()
170 a->firmware.header_buff, in do_fm_api()
171 (dma_addr_t)a->firmware.header_buff_phys); in do_fm_api()
176 mutex_unlock(&a->fm_api_mutex); in do_fm_api()
184 a->nvram_command_done = 1; in complete_nvr_req()
185 wake_up_interruptible(&a->nvram_waiter); in complete_nvr_req()
192 int offset = (u8 *)sgc->cur_offset - esas2r_buffered_ioctl; in get_physaddr_buffered_ioctl()
195 return esas2r_buffered_ioctl_size - offset; in get_physaddr_buffered_ioctl()
201 a->buffered_ioctl_done = 1; in complete_buffered_ioctl_req()
202 wake_up_interruptible(&a->buffered_ioctl_waiter); in complete_buffered_ioctl_req()
207 struct esas2r_adapter *a = bi->a; in handle_buffered_ioctl()
217 if (esas2r_buffered_ioctl_size < bi->length) { in handle_buffered_ioctl()
218 /* free the too-small buffer and get a new one */ in handle_buffered_ioctl()
219 dma_free_coherent(&a->pcid->dev, in handle_buffered_ioctl()
228 esas2r_buffered_ioctl_size = bi->length; in handle_buffered_ioctl()
229 esas2r_buffered_ioctl_pcid = a->pcid; in handle_buffered_ioctl()
230 esas2r_buffered_ioctl = dma_alloc_coherent(&a->pcid->dev, in handle_buffered_ioctl()
242 bi->length); in handle_buffered_ioctl()
249 memcpy(esas2r_buffered_ioctl, bi->ioctl, bi->length); in handle_buffered_ioctl()
257 esas2r_debug("buffered ioctl - no requests"); in handle_buffered_ioctl()
261 a->buffered_ioctl_done = 0; in handle_buffered_ioctl()
262 rq->comp_cb = complete_buffered_ioctl_req; in handle_buffered_ioctl()
263 sgc.cur_offset = esas2r_buffered_ioctl + bi->offset; in handle_buffered_ioctl()
267 if (!(*bi->callback)(a, rq, &sgc, bi->context)) { in handle_buffered_ioctl()
269 a->buffered_ioctl_done = 0; in handle_buffered_ioctl()
274 while (!a->buffered_ioctl_done) in handle_buffered_ioctl()
275 wait_event_interruptible(a->buffered_ioctl_waiter, in handle_buffered_ioctl()
276 a->buffered_ioctl_done); in handle_buffered_ioctl()
279 if (result == IOCTL_SUCCESS && bi->done_callback) in handle_buffered_ioctl()
280 (*bi->done_callback)(a, rq, bi->done_context); in handle_buffered_ioctl()
286 memcpy(bi->ioctl, esas2r_buffered_ioctl, bi->length); in handle_buffered_ioctl()
300 esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); in smp_ioctl_callback()
301 esas2r_build_ioctl_req(a, rq, sgc->length, VDA_IOCTL_SMP); in smp_ioctl_callback()
304 si->status = ATTO_STS_OUT_OF_RSRC; in smp_ioctl_callback()
321 + le32_to_cpu(si->req_length) in handle_smp_ioctl()
322 + le32_to_cpu(si->rsp_length); in handle_smp_ioctl()
333 rq->target_id = le16_to_cpu(rq->func_rsp.ioctl_rsp.csmi.target_id); in esas2r_csmi_ioctl_tunnel_comp_cb()
334 rq->vrq->scsi.flags |= cpu_to_le32(rq->func_rsp.ioctl_rsp.csmi.lun); in esas2r_csmi_ioctl_tunnel_comp_cb()
337 (*rq->aux_req_cb)(a, rq); in esas2r_csmi_ioctl_tunnel_comp_cb()
348 struct atto_vda_ioctl_req *ioctl = &rq->vrq->ioctl; in csmi_ioctl_tunnel()
350 if (test_bit(AF_DEGRADED_MODE, &a->flags)) in csmi_ioctl_tunnel()
353 esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); in csmi_ioctl_tunnel()
354 esas2r_build_ioctl_req(a, rq, sgc->length, VDA_IOCTL_CSMI); in csmi_ioctl_tunnel()
355 ioctl->csmi.ctrl_code = cpu_to_le32(ctrl_code); in csmi_ioctl_tunnel()
356 ioctl->csmi.target_id = cpu_to_le16(target_id); in csmi_ioctl_tunnel()
357 ioctl->csmi.lun = (u8)le32_to_cpu(rq->vrq->scsi.flags); in csmi_ioctl_tunnel()
363 rq->aux_req_cx = ci; in csmi_ioctl_tunnel()
364 rq->aux_req_cb = rq->comp_cb; in csmi_ioctl_tunnel()
365 rq->comp_cb = esas2r_csmi_ioctl_tunnel_comp_cb; in csmi_ioctl_tunnel()
404 if (ci->control_code == CSMI_CC_GET_DEV_ADDR) { in csmi_ioctl_callback()
405 struct atto_csmi_get_dev_addr *gda = &ci->data.dev_addr; in csmi_ioctl_callback()
407 path = gda->path_id; in csmi_ioctl_callback()
408 tid = gda->target_id; in csmi_ioctl_callback()
409 lun = gda->lun; in csmi_ioctl_callback()
410 } else if (ci->control_code == CSMI_CC_TASK_MGT) { in csmi_ioctl_callback()
411 struct atto_csmi_task_mgmt *tm = &ci->data.tsk_mgt; in csmi_ioctl_callback()
413 path = tm->path_id; in csmi_ioctl_callback()
414 tid = tm->target_id; in csmi_ioctl_callback()
415 lun = tm->lun; in csmi_ioctl_callback()
419 rq->func_rsp.ioctl_rsp.csmi.csmi_status = cpu_to_le32( in csmi_ioctl_callback()
424 rq->target_id = tid; in csmi_ioctl_callback()
425 rq->vrq->scsi.flags |= cpu_to_le32(lun); in csmi_ioctl_callback()
427 switch (ci->control_code) { in csmi_ioctl_callback()
430 struct atto_csmi_get_driver_info *gdi = &ioctl_csmi->drvr_info; in csmi_ioctl_callback()
432 strcpy(gdi->description, esas2r_get_model_name(a)); in csmi_ioctl_callback()
433 gdi->csmi_major_rev = CSMI_MAJOR_REV; in csmi_ioctl_callback()
434 gdi->csmi_minor_rev = CSMI_MINOR_REV; in csmi_ioctl_callback()
440 struct atto_csmi_get_cntlr_cfg *gcc = &ioctl_csmi->cntlr_cfg; in csmi_ioctl_callback()
442 gcc->base_io_addr = 0; in csmi_ioctl_callback()
443 pci_read_config_dword(a->pcid, PCI_BASE_ADDRESS_2, in csmi_ioctl_callback()
444 &gcc->base_memaddr_lo); in csmi_ioctl_callback()
445 pci_read_config_dword(a->pcid, PCI_BASE_ADDRESS_3, in csmi_ioctl_callback()
446 &gcc->base_memaddr_hi); in csmi_ioctl_callback()
447 gcc->board_id = MAKEDWORD(a->pcid->subsystem_device, in csmi_ioctl_callback()
448 a->pcid->subsystem_vendor); in csmi_ioctl_callback()
449 gcc->slot_num = CSMI_SLOT_NUM_UNKNOWN; in csmi_ioctl_callback()
450 gcc->cntlr_class = CSMI_CNTLR_CLASS_HBA; in csmi_ioctl_callback()
451 gcc->io_bus_type = CSMI_BUS_TYPE_PCI; in csmi_ioctl_callback()
452 gcc->pci_addr.bus_num = a->pcid->bus->number; in csmi_ioctl_callback()
453 gcc->pci_addr.device_num = PCI_SLOT(a->pcid->devfn); in csmi_ioctl_callback()
454 gcc->pci_addr.function_num = PCI_FUNC(a->pcid->devfn); in csmi_ioctl_callback()
456 memset(gcc->serial_num, 0, sizeof(gcc->serial_num)); in csmi_ioctl_callback()
458 gcc->major_rev = LOBYTE(LOWORD(a->fw_version)); in csmi_ioctl_callback()
459 gcc->minor_rev = HIBYTE(LOWORD(a->fw_version)); in csmi_ioctl_callback()
460 gcc->build_rev = LOBYTE(HIWORD(a->fw_version)); in csmi_ioctl_callback()
461 gcc->release_rev = HIBYTE(HIWORD(a->fw_version)); in csmi_ioctl_callback()
462 gcc->bios_major_rev = HIBYTE(HIWORD(a->flash_ver)); in csmi_ioctl_callback()
463 gcc->bios_minor_rev = LOBYTE(HIWORD(a->flash_ver)); in csmi_ioctl_callback()
464 gcc->bios_build_rev = LOWORD(a->flash_ver); in csmi_ioctl_callback()
466 if (test_bit(AF2_THUNDERLINK, &a->flags2)) in csmi_ioctl_callback()
467 gcc->cntlr_flags = CSMI_CNTLRF_SAS_HBA in csmi_ioctl_callback()
470 gcc->cntlr_flags = CSMI_CNTLRF_SAS_RAID in csmi_ioctl_callback()
473 gcc->rrom_major_rev = 0; in csmi_ioctl_callback()
474 gcc->rrom_minor_rev = 0; in csmi_ioctl_callback()
475 gcc->rrom_build_rev = 0; in csmi_ioctl_callback()
476 gcc->rrom_release_rev = 0; in csmi_ioctl_callback()
477 gcc->rrom_biosmajor_rev = 0; in csmi_ioctl_callback()
478 gcc->rrom_biosminor_rev = 0; in csmi_ioctl_callback()
479 gcc->rrom_biosbuild_rev = 0; in csmi_ioctl_callback()
480 gcc->rrom_biosrelease_rev = 0; in csmi_ioctl_callback()
486 struct atto_csmi_get_cntlr_sts *gcs = &ioctl_csmi->cntlr_sts; in csmi_ioctl_callback()
488 if (test_bit(AF_DEGRADED_MODE, &a->flags)) in csmi_ioctl_callback()
489 gcs->status = CSMI_CNTLR_STS_FAILED; in csmi_ioctl_callback()
491 gcs->status = CSMI_CNTLR_STS_GOOD; in csmi_ioctl_callback()
493 gcs->offline_reason = CSMI_OFFLINE_NO_REASON; in csmi_ioctl_callback()
515 ci->control_code, in csmi_ioctl_callback()
525 struct atto_csmi_get_scsi_addr *gsa = &ioctl_csmi->scsi_addr; in csmi_ioctl_callback()
529 memcpy(&lun, gsa->sas_lun, sizeof(struct scsi_lun)); in csmi_ioctl_callback()
537 spin_lock_irqsave(&a->mem_lock, flags); in csmi_ioctl_callback()
538 t = esas2r_targ_db_find_by_sas_addr(a, (u64 *)gsa->sas_addr); in csmi_ioctl_callback()
539 spin_unlock_irqrestore(&a->mem_lock, flags); in csmi_ioctl_callback()
546 gsa->host_index = 0xFF; in csmi_ioctl_callback()
547 gsa->lun = gsa->sas_lun[1]; in csmi_ioctl_callback()
548 rq->target_id = esas2r_targ_get_id(t, a); in csmi_ioctl_callback()
554 struct atto_csmi_get_dev_addr *gda = &ioctl_csmi->dev_addr; in csmi_ioctl_callback()
557 t = a->targetdb + rq->target_id; in csmi_ioctl_callback()
559 if (t >= a->targetdb_end in csmi_ioctl_callback()
560 || t->target_state != TS_PRESENT in csmi_ioctl_callback()
561 || t->sas_addr == 0) { in csmi_ioctl_callback()
567 *(u64 *)gda->sas_addr = t->sas_addr; in csmi_ioctl_callback()
568 memset(gda->sas_lun, 0, sizeof(gda->sas_lun)); in csmi_ioctl_callback()
569 gda->sas_lun[1] = (u8)le32_to_cpu(rq->vrq->scsi.flags); in csmi_ioctl_callback()
576 t = a->targetdb + rq->target_id; in csmi_ioctl_callback()
578 if (t >= a->targetdb_end in csmi_ioctl_callback()
579 || t->target_state != TS_PRESENT in csmi_ioctl_callback()
580 || !(t->flags & TF_PASS_THRU)) { in csmi_ioctl_callback()
586 ci->control_code, in csmi_ioctl_callback()
587 t->phys_targ_id)) { in csmi_ioctl_callback()
600 rq->func_rsp.ioctl_rsp.csmi.csmi_status = cpu_to_le32(sts); in csmi_ioctl_callback()
613 switch (ci->control_code) { in csmi_ioctl_done_callback()
617 &ioctl_csmi->drvr_info; in csmi_ioctl_done_callback()
619 strcpy(gdi->name, ESAS2R_VERSION_STR); in csmi_ioctl_done_callback()
621 gdi->major_rev = ESAS2R_MAJOR_REV; in csmi_ioctl_done_callback()
622 gdi->minor_rev = ESAS2R_MINOR_REV; in csmi_ioctl_done_callback()
623 gdi->build_rev = 0; in csmi_ioctl_done_callback()
624 gdi->release_rev = 0; in csmi_ioctl_done_callback()
630 struct atto_csmi_get_scsi_addr *gsa = &ioctl_csmi->scsi_addr; in csmi_ioctl_done_callback()
632 if (le32_to_cpu(rq->func_rsp.ioctl_rsp.csmi.csmi_status) == in csmi_ioctl_done_callback()
634 gsa->target_id = rq->target_id; in csmi_ioctl_done_callback()
635 gsa->path_id = 0; in csmi_ioctl_done_callback()
642 ci->status = le32_to_cpu(rq->func_rsp.ioctl_rsp.csmi.csmi_status); in csmi_ioctl_done_callback()
653 bi.ioctl = &ci->data; in handle_csmi_ioctl()
668 struct atto_ioctl *hi, in hba_ioctl_tunnel() argument
672 esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); in hba_ioctl_tunnel()
674 esas2r_build_ioctl_req(a, rq, sgc->length, VDA_IOCTL_HBA); in hba_ioctl_tunnel()
677 hi->status = ATTO_STS_OUT_OF_RSRC; in hba_ioctl_tunnel()
690 struct atto_ioctl *hi = (struct atto_ioctl *)rq->aux_req_cx; in scsi_passthru_comp_cb() local
691 struct atto_hba_scsi_pass_thru *spt = &hi->data.scsi_pass_thru; in scsi_passthru_comp_cb()
694 spt->scsi_status = rq->func_rsp.scsi_rsp.scsi_stat; in scsi_passthru_comp_cb()
695 spt->sense_length = rq->sense_len; in scsi_passthru_comp_cb()
696 spt->residual_length = in scsi_passthru_comp_cb()
697 le32_to_cpu(rq->func_rsp.scsi_rsp.residual_length); in scsi_passthru_comp_cb()
699 switch (rq->req_stat) { in scsi_passthru_comp_cb()
734 spt->req_status = sts; in scsi_passthru_comp_cb()
737 spt->target_id = in scsi_passthru_comp_cb()
738 esas2r_targ_db_find_next_present(a, (u16)spt->target_id); in scsi_passthru_comp_cb()
741 (*rq->aux_req_cb)(a, rq); in scsi_passthru_comp_cb()
749 struct atto_ioctl *hi = (struct atto_ioctl *)esas2r_buffered_ioctl; in hba_ioctl_callback() local
751 hi->status = ATTO_STS_SUCCESS; in hba_ioctl_callback()
753 switch (hi->function) { in hba_ioctl_callback()
756 u8 *class_code = (u8 *)&a->pcid->class; in hba_ioctl_callback()
759 &hi->data.get_adap_info; in hba_ioctl_callback()
761 if (hi->flags & HBAF_TUNNEL) { in hba_ioctl_callback()
762 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
766 if (hi->version > ATTO_VER_GET_ADAP_INFO0) { in hba_ioctl_callback()
767 hi->status = ATTO_STS_INV_VERSION; in hba_ioctl_callback()
768 hi->version = ATTO_VER_GET_ADAP_INFO0; in hba_ioctl_callback()
774 gai->pci.vendor_id = a->pcid->vendor; in hba_ioctl_callback()
775 gai->pci.device_id = a->pcid->device; in hba_ioctl_callback()
776 gai->pci.ss_vendor_id = a->pcid->subsystem_vendor; in hba_ioctl_callback()
777 gai->pci.ss_device_id = a->pcid->subsystem_device; in hba_ioctl_callback()
778 gai->pci.class_code[0] = class_code[0]; in hba_ioctl_callback()
779 gai->pci.class_code[1] = class_code[1]; in hba_ioctl_callback()
780 gai->pci.class_code[2] = class_code[2]; in hba_ioctl_callback()
781 gai->pci.rev_id = a->pcid->revision; in hba_ioctl_callback()
782 gai->pci.bus_num = a->pcid->bus->number; in hba_ioctl_callback()
783 gai->pci.dev_num = PCI_SLOT(a->pcid->devfn); in hba_ioctl_callback()
784 gai->pci.func_num = PCI_FUNC(a->pcid->devfn); in hba_ioctl_callback()
786 if (pci_is_pcie(a->pcid)) { in hba_ioctl_callback()
790 pcie_capability_read_word(a->pcid, PCI_EXP_LNKSTA, in hba_ioctl_callback()
792 pcie_capability_read_dword(a->pcid, PCI_EXP_LNKCAP, in hba_ioctl_callback()
795 gai->pci.link_speed_curr = in hba_ioctl_callback()
797 gai->pci.link_speed_max = in hba_ioctl_callback()
799 gai->pci.link_width_curr = in hba_ioctl_callback()
802 gai->pci.link_width_max = in hba_ioctl_callback()
807 gai->pci.msi_vector_cnt = 1; in hba_ioctl_callback()
809 if (a->pcid->msix_enabled) in hba_ioctl_callback()
810 gai->pci.interrupt_mode = ATTO_GAI_PCIIM_MSIX; in hba_ioctl_callback()
811 else if (a->pcid->msi_enabled) in hba_ioctl_callback()
812 gai->pci.interrupt_mode = ATTO_GAI_PCIIM_MSI; in hba_ioctl_callback()
814 gai->pci.interrupt_mode = ATTO_GAI_PCIIM_LEGACY; in hba_ioctl_callback()
816 gai->adap_type = ATTO_GAI_AT_ESASRAID2; in hba_ioctl_callback()
818 if (test_bit(AF2_THUNDERLINK, &a->flags2)) in hba_ioctl_callback()
819 gai->adap_type = ATTO_GAI_AT_TLSASHBA; in hba_ioctl_callback()
821 if (test_bit(AF_DEGRADED_MODE, &a->flags)) in hba_ioctl_callback()
822 gai->adap_flags |= ATTO_GAI_AF_DEGRADED; in hba_ioctl_callback()
824 gai->adap_flags |= ATTO_GAI_AF_SPT_SUPP | in hba_ioctl_callback()
827 if (a->pcid->subsystem_device == ATTO_ESAS_R60F in hba_ioctl_callback()
828 || a->pcid->subsystem_device == ATTO_ESAS_R608 in hba_ioctl_callback()
829 || a->pcid->subsystem_device == ATTO_ESAS_R644 in hba_ioctl_callback()
830 || a->pcid->subsystem_device == ATTO_TSSC_3808E) in hba_ioctl_callback()
831 gai->adap_flags |= ATTO_GAI_AF_VIRT_SES; in hba_ioctl_callback()
833 gai->num_ports = ESAS2R_NUM_PHYS; in hba_ioctl_callback()
834 gai->num_phys = ESAS2R_NUM_PHYS; in hba_ioctl_callback()
836 strcpy(gai->firmware_rev, a->fw_rev); in hba_ioctl_callback()
837 strcpy(gai->flash_rev, a->flash_rev); in hba_ioctl_callback()
838 strcpy(gai->model_name_short, esas2r_get_model_name_short(a)); in hba_ioctl_callback()
839 strcpy(gai->model_name, esas2r_get_model_name(a)); in hba_ioctl_callback()
841 gai->num_targets = ESAS2R_MAX_TARGETS; in hba_ioctl_callback()
843 gai->num_busses = 1; in hba_ioctl_callback()
844 gai->num_targsper_bus = gai->num_targets; in hba_ioctl_callback()
845 gai->num_lunsper_targ = 256; in hba_ioctl_callback()
847 if (a->pcid->subsystem_device == ATTO_ESAS_R6F0 in hba_ioctl_callback()
848 || a->pcid->subsystem_device == ATTO_ESAS_R60F) in hba_ioctl_callback()
849 gai->num_connectors = 4; in hba_ioctl_callback()
851 gai->num_connectors = 2; in hba_ioctl_callback()
853 gai->adap_flags2 |= ATTO_GAI_AF2_ADAP_CTRL_SUPP; in hba_ioctl_callback()
855 gai->num_targets_backend = a->num_targets_backend; in hba_ioctl_callback()
857 gai->tunnel_flags = a->ioctl_tunnel in hba_ioctl_callback()
871 &hi->data.get_adap_addr; in hba_ioctl_callback()
873 if (hi->flags & HBAF_TUNNEL) { in hba_ioctl_callback()
874 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
878 if (hi->version > ATTO_VER_GET_ADAP_ADDR0) { in hba_ioctl_callback()
879 hi->status = ATTO_STS_INV_VERSION; in hba_ioctl_callback()
880 hi->version = ATTO_VER_GET_ADAP_ADDR0; in hba_ioctl_callback()
881 } else if (gaa->addr_type == ATTO_GAA_AT_PORT in hba_ioctl_callback()
882 || gaa->addr_type == ATTO_GAA_AT_NODE) { in hba_ioctl_callback()
883 if (gaa->addr_type == ATTO_GAA_AT_PORT in hba_ioctl_callback()
884 && gaa->port_id >= ESAS2R_NUM_PHYS) { in hba_ioctl_callback()
885 hi->status = ATTO_STS_NOT_APPL; in hba_ioctl_callback()
887 memcpy((u64 *)gaa->address, in hba_ioctl_callback()
888 &a->nvram->sas_addr[0], sizeof(u64)); in hba_ioctl_callback()
889 gaa->addr_len = sizeof(u64); in hba_ioctl_callback()
892 hi->status = ATTO_STS_INV_PARAM; in hba_ioctl_callback()
900 if (hi->flags & HBAF_TUNNEL) { in hba_ioctl_callback()
901 if (hba_ioctl_tunnel(a, hi, rq, sgc)) in hba_ioctl_callback()
907 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
914 struct atto_hba_trace *trc = &hi->data.trace; in hba_ioctl_callback()
916 if (hi->flags & HBAF_TUNNEL) { in hba_ioctl_callback()
917 if (hba_ioctl_tunnel(a, hi, rq, sgc)) in hba_ioctl_callback()
923 if (hi->version > ATTO_VER_TRACE1) { in hba_ioctl_callback()
924 hi->status = ATTO_STS_INV_VERSION; in hba_ioctl_callback()
925 hi->version = ATTO_VER_TRACE1; in hba_ioctl_callback()
929 if (trc->trace_type == ATTO_TRC_TT_FWCOREDUMP in hba_ioctl_callback()
930 && hi->version >= ATTO_VER_TRACE1) { in hba_ioctl_callback()
931 if (trc->trace_func == ATTO_TRC_TF_UPLOAD) { in hba_ioctl_callback()
932 u32 len = hi->data_length; in hba_ioctl_callback()
933 u32 offset = trc->current_offset; in hba_ioctl_callback()
937 if (!test_bit(AF2_COREDUMP_SAVED, &a->flags2)) in hba_ioctl_callback()
946 hi->status = ATTO_STS_INV_PARAM; in hba_ioctl_callback()
950 memcpy(trc->contents, in hba_ioctl_callback()
951 a->fw_coredump_buff + offset, in hba_ioctl_callback()
953 hi->data_length = len; in hba_ioctl_callback()
954 } else if (trc->trace_func == ATTO_TRC_TF_RESET) { in hba_ioctl_callback()
955 memset(a->fw_coredump_buff, 0, in hba_ioctl_callback()
958 clear_bit(AF2_COREDUMP_SAVED, &a->flags2); in hba_ioctl_callback()
959 } else if (trc->trace_func != ATTO_TRC_TF_GET_INFO) { in hba_ioctl_callback()
960 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
965 trc->trace_mask = 0; in hba_ioctl_callback()
966 trc->current_offset = 0; in hba_ioctl_callback()
967 trc->total_length = ESAS2R_FWCOREDUMP_SZ; in hba_ioctl_callback()
970 if (!test_bit(AF2_COREDUMP_SAVED, &a->flags2)) in hba_ioctl_callback()
971 trc->total_length = 0; in hba_ioctl_callback()
973 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
981 struct atto_hba_scsi_pass_thru *spt = &hi->data.scsi_pass_thru; in hba_ioctl_callback()
984 memcpy(&lun, spt->lun, sizeof(struct scsi_lun)); in hba_ioctl_callback()
986 if (hi->flags & HBAF_TUNNEL) { in hba_ioctl_callback()
987 if (hba_ioctl_tunnel(a, hi, rq, sgc)) in hba_ioctl_callback()
993 if (hi->version > ATTO_VER_SCSI_PASS_THRU0) { in hba_ioctl_callback()
994 hi->status = ATTO_STS_INV_VERSION; in hba_ioctl_callback()
995 hi->version = ATTO_VER_SCSI_PASS_THRU0; in hba_ioctl_callback()
999 if (spt->target_id >= ESAS2R_MAX_TARGETS || !check_lun(lun)) { in hba_ioctl_callback()
1000 hi->status = ATTO_STS_INV_PARAM; in hba_ioctl_callback()
1006 sgc->length = hi->data_length; in hba_ioctl_callback()
1007 sgc->cur_offset += offsetof(struct atto_ioctl, data.byte) in hba_ioctl_callback()
1011 rq->target_id = (u16)spt->target_id; in hba_ioctl_callback()
1012 rq->vrq->scsi.flags |= cpu_to_le32(spt->lun[1]); in hba_ioctl_callback()
1013 memcpy(rq->vrq->scsi.cdb, spt->cdb, 16); in hba_ioctl_callback()
1014 rq->vrq->scsi.length = cpu_to_le32(hi->data_length); in hba_ioctl_callback()
1015 rq->sense_len = spt->sense_length; in hba_ioctl_callback()
1016 rq->sense_buf = (u8 *)spt->sense_data; in hba_ioctl_callback()
1017 /* NOTE: we ignore spt->timeout */ in hba_ioctl_callback()
1024 rq->aux_req_cx = hi; in hba_ioctl_callback()
1025 rq->aux_req_cb = rq->comp_cb; in hba_ioctl_callback()
1026 rq->comp_cb = scsi_passthru_comp_cb; in hba_ioctl_callback()
1028 if (spt->flags & ATTO_SPTF_DATA_IN) { in hba_ioctl_callback()
1029 rq->vrq->scsi.flags |= cpu_to_le32(FCP_CMND_RDD); in hba_ioctl_callback()
1030 } else if (spt->flags & ATTO_SPTF_DATA_OUT) { in hba_ioctl_callback()
1031 rq->vrq->scsi.flags |= cpu_to_le32(FCP_CMND_WRD); in hba_ioctl_callback()
1033 if (sgc->length) { in hba_ioctl_callback()
1034 hi->status = ATTO_STS_INV_PARAM; in hba_ioctl_callback()
1039 if (spt->flags & ATTO_SPTF_ORDERED_Q) in hba_ioctl_callback()
1040 rq->vrq->scsi.flags |= in hba_ioctl_callback()
1042 else if (spt->flags & ATTO_SPTF_HEAD_OF_Q) in hba_ioctl_callback()
1043 rq->vrq->scsi.flags |= cpu_to_le32(FCP_CMND_TA_HEAD_Q); in hba_ioctl_callback()
1047 hi->status = ATTO_STS_OUT_OF_RSRC; in hba_ioctl_callback()
1059 &hi->data.get_dev_addr; in hba_ioctl_callback()
1062 if (hi->flags & HBAF_TUNNEL) { in hba_ioctl_callback()
1063 if (hba_ioctl_tunnel(a, hi, rq, sgc)) in hba_ioctl_callback()
1069 if (hi->version > ATTO_VER_GET_DEV_ADDR0) { in hba_ioctl_callback()
1070 hi->status = ATTO_STS_INV_VERSION; in hba_ioctl_callback()
1071 hi->version = ATTO_VER_GET_DEV_ADDR0; in hba_ioctl_callback()
1075 if (gda->target_id >= ESAS2R_MAX_TARGETS) { in hba_ioctl_callback()
1076 hi->status = ATTO_STS_INV_PARAM; in hba_ioctl_callback()
1080 t = a->targetdb + (u16)gda->target_id; in hba_ioctl_callback()
1082 if (t->target_state != TS_PRESENT) { in hba_ioctl_callback()
1083 hi->status = ATTO_STS_FAILED; in hba_ioctl_callback()
1084 } else if (gda->addr_type == ATTO_GDA_AT_PORT) { in hba_ioctl_callback()
1085 if (t->sas_addr == 0) { in hba_ioctl_callback()
1086 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
1088 *(u64 *)gda->address = t->sas_addr; in hba_ioctl_callback()
1090 gda->addr_len = sizeof(u64); in hba_ioctl_callback()
1092 } else if (gda->addr_type == ATTO_GDA_AT_NODE) { in hba_ioctl_callback()
1093 hi->status = ATTO_STS_NOT_APPL; in hba_ioctl_callback()
1095 hi->status = ATTO_STS_INV_PARAM; in hba_ioctl_callback()
1100 gda->target_id = in hba_ioctl_callback()
1102 (u16)gda->target_id); in hba_ioctl_callback()
1109 if (hba_ioctl_tunnel(a, hi, rq, sgc)) in hba_ioctl_callback()
1117 struct atto_hba_adap_ctrl *ac = &hi->data.adap_ctrl; in hba_ioctl_callback()
1119 if (hi->flags & HBAF_TUNNEL) { in hba_ioctl_callback()
1120 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
1124 if (hi->version > ATTO_VER_ADAP_CTRL0) { in hba_ioctl_callback()
1125 hi->status = ATTO_STS_INV_VERSION; in hba_ioctl_callback()
1126 hi->version = ATTO_VER_ADAP_CTRL0; in hba_ioctl_callback()
1130 if (ac->adap_func == ATTO_AC_AF_HARD_RST) { in hba_ioctl_callback()
1132 } else if (ac->adap_func != ATTO_AC_AF_GET_STATE) { in hba_ioctl_callback()
1133 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
1137 if (test_bit(AF_CHPRST_NEEDED, &a->flags)) in hba_ioctl_callback()
1138 ac->adap_state = ATTO_AC_AS_RST_SCHED; in hba_ioctl_callback()
1139 else if (test_bit(AF_CHPRST_PENDING, &a->flags)) in hba_ioctl_callback()
1140 ac->adap_state = ATTO_AC_AS_RST_IN_PROG; in hba_ioctl_callback()
1141 else if (test_bit(AF_DISC_PENDING, &a->flags)) in hba_ioctl_callback()
1142 ac->adap_state = ATTO_AC_AS_RST_DISC; in hba_ioctl_callback()
1143 else if (test_bit(AF_DISABLED, &a->flags)) in hba_ioctl_callback()
1144 ac->adap_state = ATTO_AC_AS_DISABLED; in hba_ioctl_callback()
1145 else if (test_bit(AF_DEGRADED_MODE, &a->flags)) in hba_ioctl_callback()
1146 ac->adap_state = ATTO_AC_AS_DEGRADED; in hba_ioctl_callback()
1148 ac->adap_state = ATTO_AC_AS_OK; in hba_ioctl_callback()
1155 struct atto_hba_get_device_info *gdi = &hi->data.get_dev_info; in hba_ioctl_callback()
1158 if (hi->flags & HBAF_TUNNEL) { in hba_ioctl_callback()
1159 if (hba_ioctl_tunnel(a, hi, rq, sgc)) in hba_ioctl_callback()
1165 if (hi->version > ATTO_VER_GET_DEV_INFO0) { in hba_ioctl_callback()
1166 hi->status = ATTO_STS_INV_VERSION; in hba_ioctl_callback()
1167 hi->version = ATTO_VER_GET_DEV_INFO0; in hba_ioctl_callback()
1171 if (gdi->target_id >= ESAS2R_MAX_TARGETS) { in hba_ioctl_callback()
1172 hi->status = ATTO_STS_INV_PARAM; in hba_ioctl_callback()
1176 t = a->targetdb + (u16)gdi->target_id; in hba_ioctl_callback()
1180 gdi->target_id = in hba_ioctl_callback()
1182 (u16)gdi->target_id); in hba_ioctl_callback()
1184 if (t->target_state != TS_PRESENT) { in hba_ioctl_callback()
1185 hi->status = ATTO_STS_FAILED; in hba_ioctl_callback()
1189 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
1195 hi->status = ATTO_STS_INV_FUNC; in hba_ioctl_callback()
1208 esas2r_debug("hba_ioctl_done_callback %d", a->index); in hba_ioctl_done_callback()
1210 if (ioctl_hba->function == ATTO_FUNC_GET_ADAP_INFO) { in hba_ioctl_done_callback()
1212 &ioctl_hba->data.get_adap_info; in hba_ioctl_done_callback()
1216 gai->drvr_rev_major = ESAS2R_MAJOR_REV; in hba_ioctl_done_callback()
1217 gai->drvr_rev_minor = ESAS2R_MINOR_REV; in hba_ioctl_done_callback()
1219 strcpy(gai->drvr_rev_ascii, ESAS2R_VERSION_STR); in hba_ioctl_done_callback()
1220 strcpy(gai->drvr_name, ESAS2R_DRVR_NAME); in hba_ioctl_done_callback()
1222 gai->num_busses = 1; in hba_ioctl_done_callback()
1223 gai->num_targsper_bus = ESAS2R_MAX_ID + 1; in hba_ioctl_done_callback()
1224 gai->num_lunsper_targ = 1; in hba_ioctl_done_callback()
1237 bi.length = sizeof(struct atto_ioctl) + ioctl_hba->data_length; in handle_hba_ioctl()
1253 a->nvram_command_done = 0; in esas2r_write_params()
1254 rq->comp_cb = complete_nvr_req; in esas2r_write_params()
1258 while (!a->nvram_command_done) in esas2r_write_params()
1259 wait_event_interruptible(a->nvram_waiter, in esas2r_write_params()
1260 a->nvram_command_done); in esas2r_write_params()
1264 if (rq->req_stat == RS_SUCCESS) in esas2r_write_params()
1271 /* This function only cares about ATTO-specific ioctls (atto_express_ioctl) */
1285 return -ENOTSUPP; in esas2r_ioctl_handler()
1297 if (memcmp(ioctl->header.signature, in esas2r_ioctl_handler()
1303 return -ENOTSUPP; in esas2r_ioctl_handler()
1308 ioctl->header.return_code = IOCTL_SUCCESS; in esas2r_ioctl_handler()
1319 ioctl->data.chanlist.num_channels = 0; in esas2r_ioctl_handler()
1323 ioctl->data.chanlist.num_channels++; in esas2r_ioctl_handler()
1324 ioctl->data.chanlist.channel[k] = i; in esas2r_ioctl_handler()
1335 if (ioctl->header.channel == 0xFF) { in esas2r_ioctl_handler()
1338 if (ioctl->header.channel >= MAX_ADAPTERS || in esas2r_ioctl_handler()
1339 esas2r_adapters[ioctl->header.channel] == NULL) { in esas2r_ioctl_handler()
1340 ioctl->header.return_code = IOCTL_BAD_CHANNEL; in esas2r_ioctl_handler()
1344 return -ENOTSUPP; in esas2r_ioctl_handler()
1346 a = esas2r_adapters[ioctl->header.channel]; in esas2r_ioctl_handler()
1352 if (ioctl->data.fwrw.img_type == FW_IMG_FM_API) { in esas2r_ioctl_handler()
1354 (char *)ioctl->data.fwrw.image, in esas2r_ioctl_handler()
1361 (char *)ioctl->data.fwrw. in esas2r_ioctl_handler()
1367 } else if (ioctl->data.fwrw.img_type == FW_IMG_FS_API) { in esas2r_ioctl_handler()
1369 (char *)ioctl->data.fwrw.image, in esas2r_ioctl_handler()
1376 (char *)ioctl->data.fwrw. in esas2r_ioctl_handler()
1383 ioctl->header.return_code = IOCTL_BAD_FLASH_IMGTYPE; in esas2r_ioctl_handler()
1390 memcpy(ioctl->data.prw.data_buffer, a->nvram, in esas2r_ioctl_handler()
1392 ioctl->data.prw.code = 1; in esas2r_ioctl_handler()
1402 return -ENOMEM; in esas2r_ioctl_handler()
1406 (struct esas2r_sas_nvram *)ioctl->data.prw.data_buffer); in esas2r_ioctl_handler()
1407 ioctl->data.prw.code = code; in esas2r_ioctl_handler()
1416 (struct esas2r_sas_nvram *)ioctl->data.prw.data_buffer); in esas2r_ioctl_handler()
1417 ioctl->data.prw.code = 1; in esas2r_ioctl_handler()
1422 ioctl->data.chaninfo.major_rev = ESAS2R_MAJOR_REV; in esas2r_ioctl_handler()
1423 ioctl->data.chaninfo.minor_rev = ESAS2R_MINOR_REV; in esas2r_ioctl_handler()
1424 ioctl->data.chaninfo.IRQ = a->pcid->irq; in esas2r_ioctl_handler()
1425 ioctl->data.chaninfo.device_id = a->pcid->device; in esas2r_ioctl_handler()
1426 ioctl->data.chaninfo.vendor_id = a->pcid->vendor; in esas2r_ioctl_handler()
1427 ioctl->data.chaninfo.ven_dev_id = a->pcid->subsystem_device; in esas2r_ioctl_handler()
1428 ioctl->data.chaninfo.revision_id = a->pcid->revision; in esas2r_ioctl_handler()
1429 ioctl->data.chaninfo.pci_bus = a->pcid->bus->number; in esas2r_ioctl_handler()
1430 ioctl->data.chaninfo.pci_dev_func = a->pcid->devfn; in esas2r_ioctl_handler()
1431 ioctl->data.chaninfo.core_rev = 0; in esas2r_ioctl_handler()
1432 ioctl->data.chaninfo.host_no = a->host->host_no; in esas2r_ioctl_handler()
1433 ioctl->data.chaninfo.hbaapi_rev = 0; in esas2r_ioctl_handler()
1437 ioctl->header.return_code = handle_smp_ioctl(a, in esas2r_ioctl_handler()
1438 &ioctl->data. in esas2r_ioctl_handler()
1443 ioctl->header.return_code = in esas2r_ioctl_handler()
1444 handle_csmi_ioctl(a, &ioctl->data.csmi); in esas2r_ioctl_handler()
1448 ioctl->header.return_code = handle_hba_ioctl(a, in esas2r_ioctl_handler()
1449 &ioctl->data. in esas2r_ioctl_handler()
1455 (char *)&ioctl->data.ioctl_vda, in esas2r_ioctl_handler()
1458 ioctl->data.ioctl_vda.data_length); in esas2r_ioctl_handler()
1462 (char *)&ioctl->data.ioctl_vda, in esas2r_ioctl_handler()
1465 ioctl->data.ioctl_vda.data_length); in esas2r_ioctl_handler()
1475 ioctl->data.modinfo.adapter = a; in esas2r_ioctl_handler()
1476 ioctl->data.modinfo.pci_dev = a->pcid; in esas2r_ioctl_handler()
1477 ioctl->data.modinfo.scsi_host = a->host; in esas2r_ioctl_handler()
1478 ioctl->data.modinfo.host_no = a->host->host_no; in esas2r_ioctl_handler()
1484 ioctl->header.return_code = IOCTL_ERR_INVCMD; in esas2r_ioctl_handler()
1494 case -ENOMEM: in esas2r_ioctl_handler()
1495 case -EBUSY: in esas2r_ioctl_handler()
1496 ioctl->header.return_code = IOCTL_OUT_OF_RESOURCES; in esas2r_ioctl_handler()
1499 case -ENOSYS: in esas2r_ioctl_handler()
1500 case -EINVAL: in esas2r_ioctl_handler()
1501 ioctl->header.return_code = IOCTL_INVALID_PARAM; in esas2r_ioctl_handler()
1505 ioctl->header.return_code = IOCTL_GENERAL_ERROR; in esas2r_ioctl_handler()
1519 return -EFAULT; in esas2r_ioctl_handler()
1529 return esas2r_ioctl_handler(sd->host->hostdata, cmd, arg); in esas2r_ioctl()
1534 if (a->firmware.data) { in free_fw_buffers()
1535 dma_free_coherent(&a->pcid->dev, in free_fw_buffers()
1536 (size_t)a->firmware.orig_len, in free_fw_buffers()
1537 a->firmware.data, in free_fw_buffers()
1538 (dma_addr_t)a->firmware.phys); in free_fw_buffers()
1540 a->firmware.data = NULL; in free_fw_buffers()
1548 a->firmware.orig_len = length; in allocate_fw_buffers()
1550 a->firmware.data = dma_alloc_coherent(&a->pcid->dev, in allocate_fw_buffers()
1552 (dma_addr_t *)&a->firmware.phys, in allocate_fw_buffers()
1555 if (!a->firmware.data) { in allocate_fw_buffers()
1568 if (a->firmware.state == FW_STATUS_ST) { in esas2r_read_fw()
1569 int size = min_t(int, count, sizeof(a->firmware.header)); in esas2r_read_fw()
1571 memcpy(buf, &a->firmware.header, size); in esas2r_read_fw()
1581 if (a->firmware.state == FW_COMMAND_ST) { in esas2r_read_fw()
1582 u32 length = a->firmware.header.length; in esas2r_read_fw()
1590 if (a->firmware.header.action == FI_ACT_UP) { in esas2r_read_fw()
1592 return -ENOMEM; in esas2r_read_fw()
1597 memcpy(a->firmware.data, in esas2r_read_fw()
1598 &a->firmware.header, in esas2r_read_fw()
1599 sizeof(a->firmware.header)); in esas2r_read_fw()
1602 (struct esas2r_flash_img *)a->firmware.data); in esas2r_read_fw()
1603 } else if (a->firmware.header.action == FI_ACT_UPSZ) { in esas2r_read_fw()
1606 (int)sizeof(a->firmware.header)); in esas2r_read_fw()
1607 do_fm_api(a, &a->firmware.header); in esas2r_read_fw()
1608 memcpy(buf, &a->firmware.header, size); in esas2r_read_fw()
1613 a->firmware.header.action); in esas2r_read_fw()
1614 return -ENOSYS; in esas2r_read_fw()
1619 count = length - off; in esas2r_read_fw()
1624 if (!a->firmware.data) { in esas2r_read_fw()
1627 return -ENOMEM; in esas2r_read_fw()
1634 memcpy(buf, &a->firmware.data[off], count); in esas2r_read_fw()
1649 a->firmware.state); in esas2r_read_fw()
1651 return -EINVAL; in esas2r_read_fw()
1668 a->firmware.state = FW_INVALID_ST; in esas2r_write_fw()
1673 || header->fi_version > FI_VERSION_1) { in esas2r_write_fw()
1676 return -EINVAL; in esas2r_write_fw()
1681 if (header->fi_version == FI_VERSION_1) in esas2r_write_fw()
1687 return -EINVAL; in esas2r_write_fw()
1691 length = header->length; in esas2r_write_fw()
1696 length, header->fi_version); in esas2r_write_fw()
1697 return -EINVAL; in esas2r_write_fw()
1706 if (header->action == FI_ACT_DOWN) { in esas2r_write_fw()
1708 return -ENOMEM; in esas2r_write_fw()
1714 memcpy(&a->firmware.header, in esas2r_write_fw()
1717 } else if (header->action == FI_ACT_UP in esas2r_write_fw()
1718 || header->action == FI_ACT_UPSZ) { in esas2r_write_fw()
1720 memcpy(&a->firmware.header, in esas2r_write_fw()
1724 a->firmware.state = FW_COMMAND_ST; in esas2r_write_fw()
1728 count, header->action); in esas2r_write_fw()
1738 a->firmware.header.action); in esas2r_write_fw()
1739 return -ENOSYS; in esas2r_write_fw()
1742 length = a->firmware.header.length; in esas2r_write_fw()
1752 count = length - off; in esas2r_write_fw()
1764 if (a->firmware.header.action == FI_ACT_UP) in esas2r_write_fw()
1767 if (!a->firmware.data) { in esas2r_write_fw()
1770 return -ENOMEM; in esas2r_write_fw()
1773 memcpy(&a->firmware.data[off], buf, count); in esas2r_write_fw()
1777 (struct esas2r_flash_img *)a->firmware.data); in esas2r_write_fw()
1783 memcpy(&a->firmware.header, in esas2r_write_fw()
1784 a->firmware.data, in esas2r_write_fw()
1785 sizeof(a->firmware.header)); in esas2r_write_fw()
1787 a->firmware.state = FW_STATUS_ST; in esas2r_write_fw()
1808 a->vda_command_done = 1; in vda_complete_req()
1809 wake_up_interruptible(&a->vda_waiter); in vda_complete_req()
1815 struct esas2r_adapter *a = (struct esas2r_adapter *)sgc->adapter; in get_physaddr_vda()
1816 int offset = (u8 *)sgc->cur_offset - (u8 *)a->vda_buffer; in get_physaddr_vda()
1818 (*addr) = a->ppvda_buffer + offset; in get_physaddr_vda()
1819 return VDA_MAX_BUFFER_SIZE - offset; in get_physaddr_vda()
1825 if (!a->vda_buffer) in esas2r_read_vda()
1826 return -ENOMEM; in esas2r_read_vda()
1831 (struct atto_ioctl_vda *)a->vda_buffer; in esas2r_read_vda()
1845 return -EBUSY; in esas2r_read_vda()
1848 rq->comp_cb = vda_complete_req; in esas2r_read_vda()
1852 sgc.cur_offset = a->vda_buffer + VDA_BUFFER_HEADER_SZ; in esas2r_read_vda()
1855 a->vda_command_done = 0; in esas2r_read_vda()
1863 while (!a->vda_command_done) in esas2r_read_vda()
1864 wait_event_interruptible(a->vda_waiter, in esas2r_read_vda()
1865 a->vda_command_done); in esas2r_read_vda()
1875 count = VDA_MAX_BUFFER_SIZE - off; in esas2r_read_vda()
1880 memcpy(buf, a->vda_buffer + off, count); in esas2r_read_vda()
1894 if (!a->vda_buffer) { in esas2r_write_vda()
1896 a->vda_buffer = dma_alloc_coherent(&a->pcid->dev, in esas2r_write_vda()
1902 a->ppvda_buffer = dma_addr; in esas2r_write_vda()
1905 if (!a->vda_buffer) in esas2r_write_vda()
1906 return -ENOMEM; in esas2r_write_vda()
1912 count = VDA_MAX_BUFFER_SIZE - off; in esas2r_write_vda()
1917 memcpy(a->vda_buffer + off, buf, count); in esas2r_write_vda()
1926 a->fs_api_command_done = 1; in fs_api_complete_req()
1928 wake_up_interruptible(&a->fs_api_waiter); in fs_api_complete_req()
1934 struct esas2r_adapter *a = (struct esas2r_adapter *)sgc->adapter; in get_physaddr_fs_api()
1936 (struct esas2r_ioctl_fs *)a->fs_api_buffer; in get_physaddr_fs_api()
1937 u32 offset = (u8 *)sgc->cur_offset - (u8 *)fs; in get_physaddr_fs_api()
1939 (*addr) = a->ppfs_api_buffer + offset; in get_physaddr_fs_api()
1941 return a->fs_api_buffer_size - offset; in get_physaddr_fs_api()
1947 if (!a->fs_api_buffer) in esas2r_read_fs()
1948 return -ENOMEM; in esas2r_read_fs()
1954 (struct esas2r_ioctl_fs *)a->fs_api_buffer; in esas2r_read_fs()
1957 if (mutex_lock_interruptible(&a->fs_api_mutex)) { in esas2r_read_fs()
1959 fs->status = ATTO_STS_OUT_OF_RSRC; in esas2r_read_fs()
1960 return -EBUSY; in esas2r_read_fs()
1973 mutex_unlock(&a->fs_api_mutex); in esas2r_read_fs()
1977 rq->comp_cb = fs_api_complete_req; in esas2r_read_fs()
1981 sgc.cur_offset = fs->data; in esas2r_read_fs()
1984 a->fs_api_command_done = 0; in esas2r_read_fs()
1987 if (fs->status == ATTO_STS_OUT_OF_RSRC) in esas2r_read_fs()
1988 count = -EBUSY; in esas2r_read_fs()
1995 while (!a->fs_api_command_done) in esas2r_read_fs()
1996 wait_event_interruptible(a->fs_api_waiter, in esas2r_read_fs()
1997 a->fs_api_command_done); in esas2r_read_fs()
2001 mutex_unlock(&a->fs_api_mutex); in esas2r_read_fs()
2009 if (off > a->fs_api_buffer_size) in esas2r_read_fs()
2012 if (count + off > a->fs_api_buffer_size) in esas2r_read_fs()
2013 count = a->fs_api_buffer_size - off; in esas2r_read_fs()
2018 memcpy(buf, a->fs_api_buffer + off, count); in esas2r_read_fs()
2029 u32 length = fs->command.length + offsetof( in esas2r_write_fs()
2038 if (fs->command.command == ESAS2R_FS_CMD_BEGINW) in esas2r_write_fs()
2048 return -EINVAL; in esas2r_write_fs()
2051 if (a->fs_api_buffer) { in esas2r_write_fs()
2052 if (a->fs_api_buffer_size < length) { in esas2r_write_fs()
2053 /* Free too-small buffer and get a new one */ in esas2r_write_fs()
2054 dma_free_coherent(&a->pcid->dev, in esas2r_write_fs()
2055 (size_t)a->fs_api_buffer_size, in esas2r_write_fs()
2056 a->fs_api_buffer, in esas2r_write_fs()
2057 (dma_addr_t)a->ppfs_api_buffer); in esas2r_write_fs()
2063 a->fs_api_buffer_size = length; in esas2r_write_fs()
2065 a->fs_api_buffer = dma_alloc_coherent(&a->pcid->dev, in esas2r_write_fs()
2066 (size_t)a->fs_api_buffer_size, in esas2r_write_fs()
2067 (dma_addr_t *)&a->ppfs_api_buffer, in esas2r_write_fs()
2072 if (!a->fs_api_buffer) in esas2r_write_fs()
2073 return -ENOMEM; in esas2r_write_fs()
2075 if (off > a->fs_api_buffer_size) in esas2r_write_fs()
2078 if (count + off > a->fs_api_buffer_size) in esas2r_write_fs()
2079 count = a->fs_api_buffer_size - off; in esas2r_write_fs()
2084 memcpy(a->fs_api_buffer + off, buf, count); in esas2r_write_fs()