Lines Matching +full:hi +full:- +full:fi
5 * Copyright (c) 2001-2013 ATTO Technology, Inc.
21 * LIMITATION, ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT,
40 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301,
48 * allocate a DMA-able memory area to communicate with the firmware. In
50 * we will allocate a global buffer the first time we need it and re-use
85 a->fm_api_command_done = 1; in complete_fm_api_req()
86 wake_up_interruptible(&a->fm_api_waiter); in complete_fm_api_req()
92 struct esas2r_adapter *a = (struct esas2r_adapter *)sgc->adapter; in get_physaddr_fm_api()
93 int offset = sgc->cur_offset - a->save_offset; in get_physaddr_fm_api()
95 (*addr) = a->firmware.phys + offset; in get_physaddr_fm_api()
96 return a->firmware.orig_len - offset; in get_physaddr_fm_api()
101 struct esas2r_adapter *a = (struct esas2r_adapter *)sgc->adapter; in get_physaddr_fm_api_header()
102 int offset = sgc->cur_offset - a->save_offset; in get_physaddr_fm_api_header()
104 (*addr) = a->firmware.header_buff_phys + offset; in get_physaddr_fm_api_header()
105 return sizeof(struct esas2r_flash_img) - offset; in get_physaddr_fm_api_header()
109 static void do_fm_api(struct esas2r_adapter *a, struct esas2r_flash_img *fi) in do_fm_api() argument
113 if (mutex_lock_interruptible(&a->fm_api_mutex)) { in do_fm_api()
114 fi->status = FI_STAT_BUSY; in do_fm_api()
120 fi->status = FI_STAT_BUSY; in do_fm_api()
124 if (fi == &a->firmware.header) { in do_fm_api()
125 a->firmware.header_buff = dma_alloc_coherent(&a->pcid->dev, in do_fm_api()
129 (dma_addr_t *)&a-> in do_fm_api()
134 if (a->firmware.header_buff == NULL) { in do_fm_api()
136 fi->status = FI_STAT_BUSY; in do_fm_api()
140 memcpy(a->firmware.header_buff, fi, in do_fm_api()
142 a->save_offset = a->firmware.header_buff; in do_fm_api()
143 a->fm_api_sgc.get_phys_addr = in do_fm_api()
146 a->save_offset = (u8 *)fi; in do_fm_api()
147 a->fm_api_sgc.get_phys_addr = in do_fm_api()
151 rq->comp_cb = complete_fm_api_req; in do_fm_api()
152 a->fm_api_command_done = 0; in do_fm_api()
153 a->fm_api_sgc.cur_offset = a->save_offset; in do_fm_api()
155 if (!esas2r_fm_api(a, (struct esas2r_flash_img *)a->save_offset, rq, in do_fm_api()
156 &a->fm_api_sgc)) in do_fm_api()
160 while (!a->fm_api_command_done) in do_fm_api()
161 wait_event_interruptible(a->fm_api_waiter, in do_fm_api()
162 a->fm_api_command_done); in do_fm_api()
164 if (fi == &a->firmware.header) { in do_fm_api()
165 memcpy(fi, a->firmware.header_buff, in do_fm_api()
168 dma_free_coherent(&a->pcid->dev, in do_fm_api()
170 a->firmware.header_buff, in do_fm_api()
171 (dma_addr_t)a->firmware.header_buff_phys); in do_fm_api()
176 mutex_unlock(&a->fm_api_mutex); in do_fm_api()
184 a->nvram_command_done = 1; in complete_nvr_req()
185 wake_up_interruptible(&a->nvram_waiter); in complete_nvr_req()
192 int offset = (u8 *)sgc->cur_offset - esas2r_buffered_ioctl; in get_physaddr_buffered_ioctl()
195 return esas2r_buffered_ioctl_size - offset; in get_physaddr_buffered_ioctl()
201 a->buffered_ioctl_done = 1; in complete_buffered_ioctl_req()
202 wake_up_interruptible(&a->buffered_ioctl_waiter); in complete_buffered_ioctl_req()
207 struct esas2r_adapter *a = bi->a; in handle_buffered_ioctl()
217 if (esas2r_buffered_ioctl_size < bi->length) { in handle_buffered_ioctl()
218 /* free the too-small buffer and get a new one */ in handle_buffered_ioctl()
219 dma_free_coherent(&a->pcid->dev, in handle_buffered_ioctl()
228 esas2r_buffered_ioctl_size = bi->length; in handle_buffered_ioctl()
229 esas2r_buffered_ioctl_pcid = a->pcid; in handle_buffered_ioctl()
230 esas2r_buffered_ioctl = dma_alloc_coherent(&a->pcid->dev, in handle_buffered_ioctl()
242 bi->length); in handle_buffered_ioctl()
249 memcpy(esas2r_buffered_ioctl, bi->ioctl, bi->length); in handle_buffered_ioctl()
257 esas2r_debug("buffered ioctl - no requests"); in handle_buffered_ioctl()
261 a->buffered_ioctl_done = 0; in handle_buffered_ioctl()
262 rq->comp_cb = complete_buffered_ioctl_req; in handle_buffered_ioctl()
263 sgc.cur_offset = esas2r_buffered_ioctl + bi->offset; in handle_buffered_ioctl()
267 if (!(*bi->callback)(a, rq, &sgc, bi->context)) { in handle_buffered_ioctl()
269 a->buffered_ioctl_done = 0; in handle_buffered_ioctl()
274 while (!a->buffered_ioctl_done) in handle_buffered_ioctl()
275 wait_event_interruptible(a->buffered_ioctl_waiter, in handle_buffered_ioctl()
276 a->buffered_ioctl_done); in handle_buffered_ioctl()
279 if (result == IOCTL_SUCCESS && bi->done_callback) in handle_buffered_ioctl()
280 (*bi->done_callback)(a, rq, bi->done_context); in handle_buffered_ioctl()
286 memcpy(bi->ioctl, esas2r_buffered_ioctl, bi->length); in handle_buffered_ioctl()
300 esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); in smp_ioctl_callback()
301 esas2r_build_ioctl_req(a, rq, sgc->length, VDA_IOCTL_SMP); in smp_ioctl_callback()
304 si->status = ATTO_STS_OUT_OF_RSRC; in smp_ioctl_callback()
321 + le32_to_cpu(si->req_length) in handle_smp_ioctl()
322 + le32_to_cpu(si->rsp_length); in handle_smp_ioctl()
333 rq->target_id = le16_to_cpu(rq->func_rsp.ioctl_rsp.csmi.target_id); in esas2r_csmi_ioctl_tunnel_comp_cb()
334 rq->vrq->scsi.flags |= cpu_to_le32(rq->func_rsp.ioctl_rsp.csmi.lun); in esas2r_csmi_ioctl_tunnel_comp_cb()
337 (*rq->aux_req_cb)(a, rq); in esas2r_csmi_ioctl_tunnel_comp_cb()
348 struct atto_vda_ioctl_req *ioctl = &rq->vrq->ioctl; in csmi_ioctl_tunnel()
350 if (test_bit(AF_DEGRADED_MODE, &a->flags)) in csmi_ioctl_tunnel()
353 esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); in csmi_ioctl_tunnel()
354 esas2r_build_ioctl_req(a, rq, sgc->length, VDA_IOCTL_CSMI); in csmi_ioctl_tunnel()
355 ioctl->csmi.ctrl_code = cpu_to_le32(ctrl_code); in csmi_ioctl_tunnel()
356 ioctl->csmi.target_id = cpu_to_le16(target_id); in csmi_ioctl_tunnel()
357 ioctl->csmi.lun = (u8)le32_to_cpu(rq->vrq->scsi.flags); in csmi_ioctl_tunnel()
363 rq->aux_req_cx = ci; in csmi_ioctl_tunnel()
364 rq->aux_req_cb = rq->comp_cb; in csmi_ioctl_tunnel()
365 rq->comp_cb = esas2r_csmi_ioctl_tunnel_comp_cb; in csmi_ioctl_tunnel()
404 if (ci->control_code == CSMI_CC_GET_DEV_ADDR) { in csmi_ioctl_callback()
405 struct atto_csmi_get_dev_addr *gda = &ci->data.dev_addr; in csmi_ioctl_callback()
407 path = gda->path_id; in csmi_ioctl_callback()
408 tid = gda->target_id; in csmi_ioctl_callback()
409 lun = gda->lun; in csmi_ioctl_callback()
410 } else if (ci->control_code == CSMI_CC_TASK_MGT) { in csmi_ioctl_callback()
411 struct atto_csmi_task_mgmt *tm = &ci->data.tsk_mgt; in csmi_ioctl_callback()
413 path = tm->path_id; in csmi_ioctl_callback()
414 tid = tm->target_id; in csmi_ioctl_callback()
415 lun = tm->lun; in csmi_ioctl_callback()
419 rq->func_rsp.ioctl_rsp.csmi.csmi_status = cpu_to_le32( in csmi_ioctl_callback()
424 rq->target_id = tid; in csmi_ioctl_callback()
425 rq->vrq->scsi.flags |= cpu_to_le32(lun); in csmi_ioctl_callback()
427 switch (ci->control_code) { in csmi_ioctl_callback()
430 struct atto_csmi_get_driver_info *gdi = &ioctl_csmi->drvr_info; in csmi_ioctl_callback()
432 strcpy(gdi->description, esas2r_get_model_name(a)); in csmi_ioctl_callback()
433 gdi->csmi_major_rev = CSMI_MAJOR_REV; in csmi_ioctl_callback()
434 gdi->csmi_minor_rev = CSMI_MINOR_REV; in csmi_ioctl_callback()
440 struct atto_csmi_get_cntlr_cfg *gcc = &ioctl_csmi->cntlr_cfg; in csmi_ioctl_callback()
442 gcc->base_io_addr = 0; in csmi_ioctl_callback()
443 pci_read_config_dword(a->pcid, PCI_BASE_ADDRESS_2, in csmi_ioctl_callback()
444 &gcc->base_memaddr_lo); in csmi_ioctl_callback()
445 pci_read_config_dword(a->pcid, PCI_BASE_ADDRESS_3, in csmi_ioctl_callback()
446 &gcc->base_memaddr_hi); in csmi_ioctl_callback()
447 gcc->board_id = MAKEDWORD(a->pcid->subsystem_device, in csmi_ioctl_callback()
448 a->pcid->subsystem_vendor); in csmi_ioctl_callback()
449 gcc->slot_num = CSMI_SLOT_NUM_UNKNOWN; in csmi_ioctl_callback()
450 gcc->cntlr_class = CSMI_CNTLR_CLASS_HBA; in csmi_ioctl_callback()
451 gcc->io_bus_type = CSMI_BUS_TYPE_PCI; in csmi_ioctl_callback()
452 gcc->pci_addr.bus_num = a->pcid->bus->number; in csmi_ioctl_callback()
453 gcc->pci_addr.device_num = PCI_SLOT(a->pcid->devfn); in csmi_ioctl_callback()
454 gcc->pci_addr.function_num = PCI_FUNC(a->pcid->devfn); in csmi_ioctl_callback()
456 memset(gcc->serial_num, 0, sizeof(gcc->serial_num)); in csmi_ioctl_callback()
458 gcc->major_rev = LOBYTE(LOWORD(a->fw_version)); in csmi_ioctl_callback()
459 gcc->minor_rev = HIBYTE(LOWORD(a->fw_version)); in csmi_ioctl_callback()
460 gcc->build_rev = LOBYTE(HIWORD(a->fw_version)); in csmi_ioctl_callback()
461 gcc->release_rev = HIBYTE(HIWORD(a->fw_version)); in csmi_ioctl_callback()
462 gcc->bios_major_rev = HIBYTE(HIWORD(a->flash_ver)); in csmi_ioctl_callback()
463 gcc->bios_minor_rev = LOBYTE(HIWORD(a->flash_ver)); in csmi_ioctl_callback()
464 gcc->bios_build_rev = LOWORD(a->flash_ver); in csmi_ioctl_callback()
466 if (test_bit(AF2_THUNDERLINK, &a->flags2)) in csmi_ioctl_callback()
467 gcc->cntlr_flags = CSMI_CNTLRF_SAS_HBA in csmi_ioctl_callback()
470 gcc->cntlr_flags = CSMI_CNTLRF_SAS_RAID in csmi_ioctl_callback()
473 gcc->rrom_major_rev = 0; in csmi_ioctl_callback()
474 gcc->rrom_minor_rev = 0; in csmi_ioctl_callback()
475 gcc->rrom_build_rev = 0; in csmi_ioctl_callback()
476 gcc->rrom_release_rev = 0; in csmi_ioctl_callback()
477 gcc->rrom_biosmajor_rev = 0; in csmi_ioctl_callback()
478 gcc->rrom_biosminor_rev = 0; in csmi_ioctl_callback()
479 gcc->rrom_biosbuild_rev = 0; in csmi_ioctl_callback()
480 gcc->rrom_biosrelease_rev = 0; in csmi_ioctl_callback()
486 struct atto_csmi_get_cntlr_sts *gcs = &ioctl_csmi->cntlr_sts; in csmi_ioctl_callback()
488 if (test_bit(AF_DEGRADED_MODE, &a->flags)) in csmi_ioctl_callback()
489 gcs->status = CSMI_CNTLR_STS_FAILED; in csmi_ioctl_callback()
491 gcs->status = CSMI_CNTLR_STS_GOOD; in csmi_ioctl_callback()
493 gcs->offline_reason = CSMI_OFFLINE_NO_REASON; in csmi_ioctl_callback()
515 ci->control_code, in csmi_ioctl_callback()
525 struct atto_csmi_get_scsi_addr *gsa = &ioctl_csmi->scsi_addr; in csmi_ioctl_callback()
529 memcpy(&lun, gsa->sas_lun, sizeof(struct scsi_lun)); in csmi_ioctl_callback()
537 spin_lock_irqsave(&a->mem_lock, flags); in csmi_ioctl_callback()
538 t = esas2r_targ_db_find_by_sas_addr(a, (u64 *)gsa->sas_addr); in csmi_ioctl_callback()
539 spin_unlock_irqrestore(&a->mem_lock, flags); in csmi_ioctl_callback()
546 gsa->host_index = 0xFF; in csmi_ioctl_callback()
547 gsa->lun = gsa->sas_lun[1]; in csmi_ioctl_callback()
548 rq->target_id = esas2r_targ_get_id(t, a); in csmi_ioctl_callback()
554 struct atto_csmi_get_dev_addr *gda = &ioctl_csmi->dev_addr; in csmi_ioctl_callback()
557 t = a->targetdb + rq->target_id; in csmi_ioctl_callback()
559 if (t >= a->targetdb_end in csmi_ioctl_callback()
560 || t->target_state != TS_PRESENT in csmi_ioctl_callback()
561 || t->sas_addr == 0) { in csmi_ioctl_callback()
567 *(u64 *)gda->sas_addr = t->sas_addr; in csmi_ioctl_callback()
568 memset(gda->sas_lun, 0, sizeof(gda->sas_lun)); in csmi_ioctl_callback()
569 gda->sas_lun[1] = (u8)le32_to_cpu(rq->vrq->scsi.flags); in csmi_ioctl_callback()
576 t = a->targetdb + rq->target_id; in csmi_ioctl_callback()
578 if (t >= a->targetdb_end in csmi_ioctl_callback()
579 || t->target_state != TS_PRESENT in csmi_ioctl_callback()
580 || !(t->flags & TF_PASS_THRU)) { in csmi_ioctl_callback()
586 ci->control_code, in csmi_ioctl_callback()
587 t->phys_targ_id)) { in csmi_ioctl_callback()
600 rq->func_rsp.ioctl_rsp.csmi.csmi_status = cpu_to_le32(sts); in csmi_ioctl_callback()
613 switch (ci->control_code) { in csmi_ioctl_done_callback()
617 &ioctl_csmi->drvr_info; in csmi_ioctl_done_callback()
619 strcpy(gdi->name, ESAS2R_VERSION_STR); in csmi_ioctl_done_callback()
621 gdi->major_rev = ESAS2R_MAJOR_REV; in csmi_ioctl_done_callback()
622 gdi->minor_rev = ESAS2R_MINOR_REV; in csmi_ioctl_done_callback()
623 gdi->build_rev = 0; in csmi_ioctl_done_callback()
624 gdi->release_rev = 0; in csmi_ioctl_done_callback()
630 struct atto_csmi_get_scsi_addr *gsa = &ioctl_csmi->scsi_addr; in csmi_ioctl_done_callback()
632 if (le32_to_cpu(rq->func_rsp.ioctl_rsp.csmi.csmi_status) == in csmi_ioctl_done_callback()
634 gsa->target_id = rq->target_id; in csmi_ioctl_done_callback()
635 gsa->path_id = 0; in csmi_ioctl_done_callback()
642 ci->status = le32_to_cpu(rq->func_rsp.ioctl_rsp.csmi.csmi_status); in csmi_ioctl_done_callback()
653 bi.ioctl = &ci->data; in handle_csmi_ioctl()
668 struct atto_ioctl *hi, in hba_ioctl_tunnel() argument
672 esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge); in hba_ioctl_tunnel()
674 esas2r_build_ioctl_req(a, rq, sgc->length, VDA_IOCTL_HBA); in hba_ioctl_tunnel()
677 hi->status = ATTO_STS_OUT_OF_RSRC; in hba_ioctl_tunnel()
690 struct atto_ioctl *hi = (struct atto_ioctl *)rq->aux_req_cx; in scsi_passthru_comp_cb() local
691 struct atto_hba_scsi_pass_thru *spt = &hi->data.scsi_pass_thru; in scsi_passthru_comp_cb()
694 spt->scsi_status = rq->func_rsp.scsi_rsp.scsi_stat; in scsi_passthru_comp_cb()
695 spt->sense_length = rq->sense_len; in scsi_passthru_comp_cb()
696 spt->residual_length = in scsi_passthru_comp_cb()
697 le32_to_cpu(rq->func_rsp.scsi_rsp.residual_length); in scsi_passthru_comp_cb()
699 switch (rq->req_stat) { in scsi_passthru_comp_cb()
734 spt->req_status = sts; in scsi_passthru_comp_cb()
737 spt->target_id = in scsi_passthru_comp_cb()
738 esas2r_targ_db_find_next_present(a, (u16)spt->target_id); in scsi_passthru_comp_cb()
741 (*rq->aux_req_cb)(a, rq); in scsi_passthru_comp_cb()
749 struct atto_ioctl *hi = (struct atto_ioctl *)esas2r_buffered_ioctl; in hba_ioctl_callback() local
751 hi->status = ATTO_STS_SUCCESS; in hba_ioctl_callback()
753 switch (hi->function) { in hba_ioctl_callback()
756 u8 *class_code = (u8 *)&a->pcid->class; in hba_ioctl_callback()
759 &hi->data.get_adap_info; in hba_ioctl_callback()
761 if (hi->flags & HBAF_TUNNEL) { in hba_ioctl_callback()
762 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
766 if (hi->version > ATTO_VER_GET_ADAP_INFO0) { in hba_ioctl_callback()
767 hi->status = ATTO_STS_INV_VERSION; in hba_ioctl_callback()
768 hi->version = ATTO_VER_GET_ADAP_INFO0; in hba_ioctl_callback()
774 gai->pci.vendor_id = a->pcid->vendor; in hba_ioctl_callback()
775 gai->pci.device_id = a->pcid->device; in hba_ioctl_callback()
776 gai->pci.ss_vendor_id = a->pcid->subsystem_vendor; in hba_ioctl_callback()
777 gai->pci.ss_device_id = a->pcid->subsystem_device; in hba_ioctl_callback()
778 gai->pci.class_code[0] = class_code[0]; in hba_ioctl_callback()
779 gai->pci.class_code[1] = class_code[1]; in hba_ioctl_callback()
780 gai->pci.class_code[2] = class_code[2]; in hba_ioctl_callback()
781 gai->pci.rev_id = a->pcid->revision; in hba_ioctl_callback()
782 gai->pci.bus_num = a->pcid->bus->number; in hba_ioctl_callback()
783 gai->pci.dev_num = PCI_SLOT(a->pcid->devfn); in hba_ioctl_callback()
784 gai->pci.func_num = PCI_FUNC(a->pcid->devfn); in hba_ioctl_callback()
786 if (pci_is_pcie(a->pcid)) { in hba_ioctl_callback()
790 pcie_capability_read_word(a->pcid, PCI_EXP_LNKSTA, in hba_ioctl_callback()
792 pcie_capability_read_dword(a->pcid, PCI_EXP_LNKCAP, in hba_ioctl_callback()
795 gai->pci.link_speed_curr = in hba_ioctl_callback()
797 gai->pci.link_speed_max = in hba_ioctl_callback()
799 gai->pci.link_width_curr = in hba_ioctl_callback()
802 gai->pci.link_width_max = in hba_ioctl_callback()
807 gai->pci.msi_vector_cnt = 1; in hba_ioctl_callback()
809 if (a->pcid->msix_enabled) in hba_ioctl_callback()
810 gai->pci.interrupt_mode = ATTO_GAI_PCIIM_MSIX; in hba_ioctl_callback()
811 else if (a->pcid->msi_enabled) in hba_ioctl_callback()
812 gai->pci.interrupt_mode = ATTO_GAI_PCIIM_MSI; in hba_ioctl_callback()
814 gai->pci.interrupt_mode = ATTO_GAI_PCIIM_LEGACY; in hba_ioctl_callback()
816 gai->adap_type = ATTO_GAI_AT_ESASRAID2; in hba_ioctl_callback()
818 if (test_bit(AF2_THUNDERLINK, &a->flags2)) in hba_ioctl_callback()
819 gai->adap_type = ATTO_GAI_AT_TLSASHBA; in hba_ioctl_callback()
821 if (test_bit(AF_DEGRADED_MODE, &a->flags)) in hba_ioctl_callback()
822 gai->adap_flags |= ATTO_GAI_AF_DEGRADED; in hba_ioctl_callback()
824 gai->adap_flags |= ATTO_GAI_AF_SPT_SUPP | in hba_ioctl_callback()
827 if (a->pcid->subsystem_device == ATTO_ESAS_R60F in hba_ioctl_callback()
828 || a->pcid->subsystem_device == ATTO_ESAS_R608 in hba_ioctl_callback()
829 || a->pcid->subsystem_device == ATTO_ESAS_R644 in hba_ioctl_callback()
830 || a->pcid->subsystem_device == ATTO_TSSC_3808E) in hba_ioctl_callback()
831 gai->adap_flags |= ATTO_GAI_AF_VIRT_SES; in hba_ioctl_callback()
833 gai->num_ports = ESAS2R_NUM_PHYS; in hba_ioctl_callback()
834 gai->num_phys = ESAS2R_NUM_PHYS; in hba_ioctl_callback()
836 strcpy(gai->firmware_rev, a->fw_rev); in hba_ioctl_callback()
837 strcpy(gai->flash_rev, a->flash_rev); in hba_ioctl_callback()
838 strcpy(gai->model_name_short, esas2r_get_model_name_short(a)); in hba_ioctl_callback()
839 strcpy(gai->model_name, esas2r_get_model_name(a)); in hba_ioctl_callback()
841 gai->num_targets = ESAS2R_MAX_TARGETS; in hba_ioctl_callback()
843 gai->num_busses = 1; in hba_ioctl_callback()
844 gai->num_targsper_bus = gai->num_targets; in hba_ioctl_callback()
845 gai->num_lunsper_targ = 256; in hba_ioctl_callback()
847 if (a->pcid->subsystem_device == ATTO_ESAS_R6F0 in hba_ioctl_callback()
848 || a->pcid->subsystem_device == ATTO_ESAS_R60F) in hba_ioctl_callback()
849 gai->num_connectors = 4; in hba_ioctl_callback()
851 gai->num_connectors = 2; in hba_ioctl_callback()
853 gai->adap_flags2 |= ATTO_GAI_AF2_ADAP_CTRL_SUPP; in hba_ioctl_callback()
855 gai->num_targets_backend = a->num_targets_backend; in hba_ioctl_callback()
857 gai->tunnel_flags = a->ioctl_tunnel in hba_ioctl_callback()
871 &hi->data.get_adap_addr; in hba_ioctl_callback()
873 if (hi->flags & HBAF_TUNNEL) { in hba_ioctl_callback()
874 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
878 if (hi->version > ATTO_VER_GET_ADAP_ADDR0) { in hba_ioctl_callback()
879 hi->status = ATTO_STS_INV_VERSION; in hba_ioctl_callback()
880 hi->version = ATTO_VER_GET_ADAP_ADDR0; in hba_ioctl_callback()
881 } else if (gaa->addr_type == ATTO_GAA_AT_PORT in hba_ioctl_callback()
882 || gaa->addr_type == ATTO_GAA_AT_NODE) { in hba_ioctl_callback()
883 if (gaa->addr_type == ATTO_GAA_AT_PORT in hba_ioctl_callback()
884 && gaa->port_id >= ESAS2R_NUM_PHYS) { in hba_ioctl_callback()
885 hi->status = ATTO_STS_NOT_APPL; in hba_ioctl_callback()
887 memcpy((u64 *)gaa->address, in hba_ioctl_callback()
888 &a->nvram->sas_addr[0], sizeof(u64)); in hba_ioctl_callback()
889 gaa->addr_len = sizeof(u64); in hba_ioctl_callback()
892 hi->status = ATTO_STS_INV_PARAM; in hba_ioctl_callback()
900 if (hi->flags & HBAF_TUNNEL) { in hba_ioctl_callback()
901 if (hba_ioctl_tunnel(a, hi, rq, sgc)) in hba_ioctl_callback()
907 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
914 struct atto_hba_trace *trc = &hi->data.trace; in hba_ioctl_callback()
916 if (hi->flags & HBAF_TUNNEL) { in hba_ioctl_callback()
917 if (hba_ioctl_tunnel(a, hi, rq, sgc)) in hba_ioctl_callback()
923 if (hi->version > ATTO_VER_TRACE1) { in hba_ioctl_callback()
924 hi->status = ATTO_STS_INV_VERSION; in hba_ioctl_callback()
925 hi->version = ATTO_VER_TRACE1; in hba_ioctl_callback()
929 if (trc->trace_type == ATTO_TRC_TT_FWCOREDUMP in hba_ioctl_callback()
930 && hi->version >= ATTO_VER_TRACE1) { in hba_ioctl_callback()
931 if (trc->trace_func == ATTO_TRC_TF_UPLOAD) { in hba_ioctl_callback()
932 u32 len = hi->data_length; in hba_ioctl_callback()
933 u32 offset = trc->current_offset; in hba_ioctl_callback()
937 if (!test_bit(AF2_COREDUMP_SAVED, &a->flags2)) in hba_ioctl_callback()
946 hi->status = ATTO_STS_INV_PARAM; in hba_ioctl_callback()
951 a->fw_coredump_buff + offset, in hba_ioctl_callback()
954 hi->data_length = len; in hba_ioctl_callback()
955 } else if (trc->trace_func == ATTO_TRC_TF_RESET) { in hba_ioctl_callback()
956 memset(a->fw_coredump_buff, 0, in hba_ioctl_callback()
959 clear_bit(AF2_COREDUMP_SAVED, &a->flags2); in hba_ioctl_callback()
960 } else if (trc->trace_func != ATTO_TRC_TF_GET_INFO) { in hba_ioctl_callback()
961 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
966 trc->trace_mask = 0; in hba_ioctl_callback()
967 trc->current_offset = 0; in hba_ioctl_callback()
968 trc->total_length = ESAS2R_FWCOREDUMP_SZ; in hba_ioctl_callback()
971 if (!test_bit(AF2_COREDUMP_SAVED, &a->flags2)) in hba_ioctl_callback()
972 trc->total_length = 0; in hba_ioctl_callback()
974 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
982 struct atto_hba_scsi_pass_thru *spt = &hi->data.scsi_pass_thru; in hba_ioctl_callback()
985 memcpy(&lun, spt->lun, sizeof(struct scsi_lun)); in hba_ioctl_callback()
987 if (hi->flags & HBAF_TUNNEL) { in hba_ioctl_callback()
988 if (hba_ioctl_tunnel(a, hi, rq, sgc)) in hba_ioctl_callback()
994 if (hi->version > ATTO_VER_SCSI_PASS_THRU0) { in hba_ioctl_callback()
995 hi->status = ATTO_STS_INV_VERSION; in hba_ioctl_callback()
996 hi->version = ATTO_VER_SCSI_PASS_THRU0; in hba_ioctl_callback()
1000 if (spt->target_id >= ESAS2R_MAX_TARGETS || !check_lun(lun)) { in hba_ioctl_callback()
1001 hi->status = ATTO_STS_INV_PARAM; in hba_ioctl_callback()
1007 sgc->length = hi->data_length; in hba_ioctl_callback()
1008 sgc->cur_offset += offsetof(struct atto_ioctl, data.byte) in hba_ioctl_callback()
1012 rq->target_id = (u16)spt->target_id; in hba_ioctl_callback()
1013 rq->vrq->scsi.flags |= cpu_to_le32(spt->lun[1]); in hba_ioctl_callback()
1014 memcpy(rq->vrq->scsi.cdb, spt->cdb, 16); in hba_ioctl_callback()
1015 rq->vrq->scsi.length = cpu_to_le32(hi->data_length); in hba_ioctl_callback()
1016 rq->sense_len = spt->sense_length; in hba_ioctl_callback()
1017 rq->sense_buf = (u8 *)spt->sense_data; in hba_ioctl_callback()
1018 /* NOTE: we ignore spt->timeout */ in hba_ioctl_callback()
1025 rq->aux_req_cx = hi; in hba_ioctl_callback()
1026 rq->aux_req_cb = rq->comp_cb; in hba_ioctl_callback()
1027 rq->comp_cb = scsi_passthru_comp_cb; in hba_ioctl_callback()
1029 if (spt->flags & ATTO_SPTF_DATA_IN) { in hba_ioctl_callback()
1030 rq->vrq->scsi.flags |= cpu_to_le32(FCP_CMND_RDD); in hba_ioctl_callback()
1031 } else if (spt->flags & ATTO_SPTF_DATA_OUT) { in hba_ioctl_callback()
1032 rq->vrq->scsi.flags |= cpu_to_le32(FCP_CMND_WRD); in hba_ioctl_callback()
1034 if (sgc->length) { in hba_ioctl_callback()
1035 hi->status = ATTO_STS_INV_PARAM; in hba_ioctl_callback()
1040 if (spt->flags & ATTO_SPTF_ORDERED_Q) in hba_ioctl_callback()
1041 rq->vrq->scsi.flags |= in hba_ioctl_callback()
1043 else if (spt->flags & ATTO_SPTF_HEAD_OF_Q) in hba_ioctl_callback()
1044 rq->vrq->scsi.flags |= cpu_to_le32(FCP_CMND_TA_HEAD_Q); in hba_ioctl_callback()
1048 hi->status = ATTO_STS_OUT_OF_RSRC; in hba_ioctl_callback()
1060 &hi->data.get_dev_addr; in hba_ioctl_callback()
1063 if (hi->flags & HBAF_TUNNEL) { in hba_ioctl_callback()
1064 if (hba_ioctl_tunnel(a, hi, rq, sgc)) in hba_ioctl_callback()
1070 if (hi->version > ATTO_VER_GET_DEV_ADDR0) { in hba_ioctl_callback()
1071 hi->status = ATTO_STS_INV_VERSION; in hba_ioctl_callback()
1072 hi->version = ATTO_VER_GET_DEV_ADDR0; in hba_ioctl_callback()
1076 if (gda->target_id >= ESAS2R_MAX_TARGETS) { in hba_ioctl_callback()
1077 hi->status = ATTO_STS_INV_PARAM; in hba_ioctl_callback()
1081 t = a->targetdb + (u16)gda->target_id; in hba_ioctl_callback()
1083 if (t->target_state != TS_PRESENT) { in hba_ioctl_callback()
1084 hi->status = ATTO_STS_FAILED; in hba_ioctl_callback()
1085 } else if (gda->addr_type == ATTO_GDA_AT_PORT) { in hba_ioctl_callback()
1086 if (t->sas_addr == 0) { in hba_ioctl_callback()
1087 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
1089 *(u64 *)gda->address = t->sas_addr; in hba_ioctl_callback()
1091 gda->addr_len = sizeof(u64); in hba_ioctl_callback()
1093 } else if (gda->addr_type == ATTO_GDA_AT_NODE) { in hba_ioctl_callback()
1094 hi->status = ATTO_STS_NOT_APPL; in hba_ioctl_callback()
1096 hi->status = ATTO_STS_INV_PARAM; in hba_ioctl_callback()
1101 gda->target_id = in hba_ioctl_callback()
1103 (u16)gda->target_id); in hba_ioctl_callback()
1110 if (hba_ioctl_tunnel(a, hi, rq, sgc)) in hba_ioctl_callback()
1118 struct atto_hba_adap_ctrl *ac = &hi->data.adap_ctrl; in hba_ioctl_callback()
1120 if (hi->flags & HBAF_TUNNEL) { in hba_ioctl_callback()
1121 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
1125 if (hi->version > ATTO_VER_ADAP_CTRL0) { in hba_ioctl_callback()
1126 hi->status = ATTO_STS_INV_VERSION; in hba_ioctl_callback()
1127 hi->version = ATTO_VER_ADAP_CTRL0; in hba_ioctl_callback()
1131 if (ac->adap_func == ATTO_AC_AF_HARD_RST) { in hba_ioctl_callback()
1133 } else if (ac->adap_func != ATTO_AC_AF_GET_STATE) { in hba_ioctl_callback()
1134 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
1138 if (test_bit(AF_CHPRST_NEEDED, &a->flags)) in hba_ioctl_callback()
1139 ac->adap_state = ATTO_AC_AS_RST_SCHED; in hba_ioctl_callback()
1140 else if (test_bit(AF_CHPRST_PENDING, &a->flags)) in hba_ioctl_callback()
1141 ac->adap_state = ATTO_AC_AS_RST_IN_PROG; in hba_ioctl_callback()
1142 else if (test_bit(AF_DISC_PENDING, &a->flags)) in hba_ioctl_callback()
1143 ac->adap_state = ATTO_AC_AS_RST_DISC; in hba_ioctl_callback()
1144 else if (test_bit(AF_DISABLED, &a->flags)) in hba_ioctl_callback()
1145 ac->adap_state = ATTO_AC_AS_DISABLED; in hba_ioctl_callback()
1146 else if (test_bit(AF_DEGRADED_MODE, &a->flags)) in hba_ioctl_callback()
1147 ac->adap_state = ATTO_AC_AS_DEGRADED; in hba_ioctl_callback()
1149 ac->adap_state = ATTO_AC_AS_OK; in hba_ioctl_callback()
1156 struct atto_hba_get_device_info *gdi = &hi->data.get_dev_info; in hba_ioctl_callback()
1159 if (hi->flags & HBAF_TUNNEL) { in hba_ioctl_callback()
1160 if (hba_ioctl_tunnel(a, hi, rq, sgc)) in hba_ioctl_callback()
1166 if (hi->version > ATTO_VER_GET_DEV_INFO0) { in hba_ioctl_callback()
1167 hi->status = ATTO_STS_INV_VERSION; in hba_ioctl_callback()
1168 hi->version = ATTO_VER_GET_DEV_INFO0; in hba_ioctl_callback()
1172 if (gdi->target_id >= ESAS2R_MAX_TARGETS) { in hba_ioctl_callback()
1173 hi->status = ATTO_STS_INV_PARAM; in hba_ioctl_callback()
1177 t = a->targetdb + (u16)gdi->target_id; in hba_ioctl_callback()
1181 gdi->target_id = in hba_ioctl_callback()
1183 (u16)gdi->target_id); in hba_ioctl_callback()
1185 if (t->target_state != TS_PRESENT) { in hba_ioctl_callback()
1186 hi->status = ATTO_STS_FAILED; in hba_ioctl_callback()
1190 hi->status = ATTO_STS_UNSUPPORTED; in hba_ioctl_callback()
1196 hi->status = ATTO_STS_INV_FUNC; in hba_ioctl_callback()
1209 esas2r_debug("hba_ioctl_done_callback %d", a->index); in hba_ioctl_done_callback()
1211 if (ioctl_hba->function == ATTO_FUNC_GET_ADAP_INFO) { in hba_ioctl_done_callback()
1213 &ioctl_hba->data.get_adap_info; in hba_ioctl_done_callback()
1217 gai->drvr_rev_major = ESAS2R_MAJOR_REV; in hba_ioctl_done_callback()
1218 gai->drvr_rev_minor = ESAS2R_MINOR_REV; in hba_ioctl_done_callback()
1220 strcpy(gai->drvr_rev_ascii, ESAS2R_VERSION_STR); in hba_ioctl_done_callback()
1221 strcpy(gai->drvr_name, ESAS2R_DRVR_NAME); in hba_ioctl_done_callback()
1223 gai->num_busses = 1; in hba_ioctl_done_callback()
1224 gai->num_targsper_bus = ESAS2R_MAX_ID + 1; in hba_ioctl_done_callback()
1225 gai->num_lunsper_targ = 1; in hba_ioctl_done_callback()
1238 bi.length = sizeof(struct atto_ioctl) + ioctl_hba->data_length; in handle_hba_ioctl()
1254 a->nvram_command_done = 0; in esas2r_write_params()
1255 rq->comp_cb = complete_nvr_req; in esas2r_write_params()
1259 while (!a->nvram_command_done) in esas2r_write_params()
1260 wait_event_interruptible(a->nvram_waiter, in esas2r_write_params()
1261 a->nvram_command_done); in esas2r_write_params()
1265 if (rq->req_stat == RS_SUCCESS) in esas2r_write_params()
1272 /* This function only cares about ATTO-specific ioctls (atto_express_ioctl) */
1286 return -ENOTSUPP; in esas2r_ioctl_handler()
1298 if (memcmp(ioctl->header.signature, in esas2r_ioctl_handler()
1304 return -ENOTSUPP; in esas2r_ioctl_handler()
1309 ioctl->header.return_code = IOCTL_SUCCESS; in esas2r_ioctl_handler()
1320 ioctl->data.chanlist.num_channels = 0; in esas2r_ioctl_handler()
1324 ioctl->data.chanlist.num_channels++; in esas2r_ioctl_handler()
1325 ioctl->data.chanlist.channel[k] = i; in esas2r_ioctl_handler()
1336 if (ioctl->header.channel == 0xFF) { in esas2r_ioctl_handler()
1339 if (ioctl->header.channel >= MAX_ADAPTERS || in esas2r_ioctl_handler()
1340 esas2r_adapters[ioctl->header.channel] == NULL) { in esas2r_ioctl_handler()
1341 ioctl->header.return_code = IOCTL_BAD_CHANNEL; in esas2r_ioctl_handler()
1345 return -ENOTSUPP; in esas2r_ioctl_handler()
1347 a = esas2r_adapters[ioctl->header.channel]; in esas2r_ioctl_handler()
1353 if (ioctl->data.fwrw.img_type == FW_IMG_FM_API) { in esas2r_ioctl_handler()
1355 (char *)ioctl->data.fwrw.image, in esas2r_ioctl_handler()
1362 (char *)ioctl->data.fwrw. in esas2r_ioctl_handler()
1368 } else if (ioctl->data.fwrw.img_type == FW_IMG_FS_API) { in esas2r_ioctl_handler()
1370 (char *)ioctl->data.fwrw.image, in esas2r_ioctl_handler()
1377 (char *)ioctl->data.fwrw. in esas2r_ioctl_handler()
1384 ioctl->header.return_code = IOCTL_BAD_FLASH_IMGTYPE; in esas2r_ioctl_handler()
1391 memcpy(ioctl->data.prw.data_buffer, a->nvram, in esas2r_ioctl_handler()
1393 ioctl->data.prw.code = 1; in esas2r_ioctl_handler()
1403 return -ENOMEM; in esas2r_ioctl_handler()
1407 (struct esas2r_sas_nvram *)ioctl->data.prw.data_buffer); in esas2r_ioctl_handler()
1408 ioctl->data.prw.code = code; in esas2r_ioctl_handler()
1417 (struct esas2r_sas_nvram *)ioctl->data.prw.data_buffer); in esas2r_ioctl_handler()
1418 ioctl->data.prw.code = 1; in esas2r_ioctl_handler()
1423 ioctl->data.chaninfo.major_rev = ESAS2R_MAJOR_REV; in esas2r_ioctl_handler()
1424 ioctl->data.chaninfo.minor_rev = ESAS2R_MINOR_REV; in esas2r_ioctl_handler()
1425 ioctl->data.chaninfo.IRQ = a->pcid->irq; in esas2r_ioctl_handler()
1426 ioctl->data.chaninfo.device_id = a->pcid->device; in esas2r_ioctl_handler()
1427 ioctl->data.chaninfo.vendor_id = a->pcid->vendor; in esas2r_ioctl_handler()
1428 ioctl->data.chaninfo.ven_dev_id = a->pcid->subsystem_device; in esas2r_ioctl_handler()
1429 ioctl->data.chaninfo.revision_id = a->pcid->revision; in esas2r_ioctl_handler()
1430 ioctl->data.chaninfo.pci_bus = a->pcid->bus->number; in esas2r_ioctl_handler()
1431 ioctl->data.chaninfo.pci_dev_func = a->pcid->devfn; in esas2r_ioctl_handler()
1432 ioctl->data.chaninfo.core_rev = 0; in esas2r_ioctl_handler()
1433 ioctl->data.chaninfo.host_no = a->host->host_no; in esas2r_ioctl_handler()
1434 ioctl->data.chaninfo.hbaapi_rev = 0; in esas2r_ioctl_handler()
1438 ioctl->header.return_code = handle_smp_ioctl(a, in esas2r_ioctl_handler()
1439 &ioctl->data. in esas2r_ioctl_handler()
1444 ioctl->header.return_code = in esas2r_ioctl_handler()
1445 handle_csmi_ioctl(a, &ioctl->data.csmi); in esas2r_ioctl_handler()
1449 ioctl->header.return_code = handle_hba_ioctl(a, in esas2r_ioctl_handler()
1450 &ioctl->data. in esas2r_ioctl_handler()
1456 (char *)&ioctl->data.ioctl_vda, in esas2r_ioctl_handler()
1459 ioctl->data.ioctl_vda.data_length); in esas2r_ioctl_handler()
1463 (char *)&ioctl->data.ioctl_vda, in esas2r_ioctl_handler()
1466 ioctl->data.ioctl_vda.data_length); in esas2r_ioctl_handler()
1476 ioctl->data.modinfo.adapter = a; in esas2r_ioctl_handler()
1477 ioctl->data.modinfo.pci_dev = a->pcid; in esas2r_ioctl_handler()
1478 ioctl->data.modinfo.scsi_host = a->host; in esas2r_ioctl_handler()
1479 ioctl->data.modinfo.host_no = a->host->host_no; in esas2r_ioctl_handler()
1485 ioctl->header.return_code = IOCTL_ERR_INVCMD; in esas2r_ioctl_handler()
1495 case -ENOMEM: in esas2r_ioctl_handler()
1496 case -EBUSY: in esas2r_ioctl_handler()
1497 ioctl->header.return_code = IOCTL_OUT_OF_RESOURCES; in esas2r_ioctl_handler()
1500 case -ENOSYS: in esas2r_ioctl_handler()
1501 case -EINVAL: in esas2r_ioctl_handler()
1502 ioctl->header.return_code = IOCTL_INVALID_PARAM; in esas2r_ioctl_handler()
1506 ioctl->header.return_code = IOCTL_GENERAL_ERROR; in esas2r_ioctl_handler()
1520 return -EFAULT; in esas2r_ioctl_handler()
1530 return esas2r_ioctl_handler(sd->host->hostdata, cmd, arg); in esas2r_ioctl()
1535 if (a->firmware.data) { in free_fw_buffers()
1536 dma_free_coherent(&a->pcid->dev, in free_fw_buffers()
1537 (size_t)a->firmware.orig_len, in free_fw_buffers()
1538 a->firmware.data, in free_fw_buffers()
1539 (dma_addr_t)a->firmware.phys); in free_fw_buffers()
1541 a->firmware.data = NULL; in free_fw_buffers()
1549 a->firmware.orig_len = length; in allocate_fw_buffers()
1551 a->firmware.data = dma_alloc_coherent(&a->pcid->dev, in allocate_fw_buffers()
1553 (dma_addr_t *)&a->firmware.phys, in allocate_fw_buffers()
1556 if (!a->firmware.data) { in allocate_fw_buffers()
1569 if (a->firmware.state == FW_STATUS_ST) { in esas2r_read_fw()
1570 int size = min_t(int, count, sizeof(a->firmware.header)); in esas2r_read_fw()
1572 memcpy(buf, &a->firmware.header, size); in esas2r_read_fw()
1582 if (a->firmware.state == FW_COMMAND_ST) { in esas2r_read_fw()
1583 u32 length = a->firmware.header.length; in esas2r_read_fw()
1591 if (a->firmware.header.action == FI_ACT_UP) { in esas2r_read_fw()
1593 return -ENOMEM; in esas2r_read_fw()
1598 memcpy(a->firmware.data, in esas2r_read_fw()
1599 &a->firmware.header, in esas2r_read_fw()
1600 sizeof(a->firmware.header)); in esas2r_read_fw()
1603 (struct esas2r_flash_img *)a->firmware.data); in esas2r_read_fw()
1604 } else if (a->firmware.header.action == FI_ACT_UPSZ) { in esas2r_read_fw()
1607 (int)sizeof(a->firmware.header)); in esas2r_read_fw()
1608 do_fm_api(a, &a->firmware.header); in esas2r_read_fw()
1609 memcpy(buf, &a->firmware.header, size); in esas2r_read_fw()
1614 a->firmware.header.action); in esas2r_read_fw()
1615 return -ENOSYS; in esas2r_read_fw()
1620 count = length - off; in esas2r_read_fw()
1625 if (!a->firmware.data) { in esas2r_read_fw()
1628 return -ENOMEM; in esas2r_read_fw()
1635 memcpy(buf, &a->firmware.data[off], count); in esas2r_read_fw()
1650 a->firmware.state); in esas2r_read_fw()
1652 return -EINVAL; in esas2r_read_fw()
1669 a->firmware.state = FW_INVALID_ST; in esas2r_write_fw()
1674 || header->fi_version > FI_VERSION_1) { in esas2r_write_fw()
1677 return -EINVAL; in esas2r_write_fw()
1682 if (header->fi_version == FI_VERSION_1) in esas2r_write_fw()
1688 return -EINVAL; in esas2r_write_fw()
1692 length = header->length; in esas2r_write_fw()
1697 length, header->fi_version); in esas2r_write_fw()
1698 return -EINVAL; in esas2r_write_fw()
1707 if (header->action == FI_ACT_DOWN) { in esas2r_write_fw()
1709 return -ENOMEM; in esas2r_write_fw()
1715 memcpy(&a->firmware.header, in esas2r_write_fw()
1718 } else if (header->action == FI_ACT_UP in esas2r_write_fw()
1719 || header->action == FI_ACT_UPSZ) { in esas2r_write_fw()
1721 memcpy(&a->firmware.header, in esas2r_write_fw()
1725 a->firmware.state = FW_COMMAND_ST; in esas2r_write_fw()
1729 count, header->action); in esas2r_write_fw()
1739 a->firmware.header.action); in esas2r_write_fw()
1740 return -ENOSYS; in esas2r_write_fw()
1743 length = a->firmware.header.length; in esas2r_write_fw()
1753 count = length - off; in esas2r_write_fw()
1765 if (a->firmware.header.action == FI_ACT_UP) in esas2r_write_fw()
1768 if (!a->firmware.data) { in esas2r_write_fw()
1771 return -ENOMEM; in esas2r_write_fw()
1774 memcpy(&a->firmware.data[off], buf, count); in esas2r_write_fw()
1778 (struct esas2r_flash_img *)a->firmware.data); in esas2r_write_fw()
1784 memcpy(&a->firmware.header, in esas2r_write_fw()
1785 a->firmware.data, in esas2r_write_fw()
1786 sizeof(a->firmware.header)); in esas2r_write_fw()
1788 a->firmware.state = FW_STATUS_ST; in esas2r_write_fw()
1809 a->vda_command_done = 1; in vda_complete_req()
1810 wake_up_interruptible(&a->vda_waiter); in vda_complete_req()
1816 struct esas2r_adapter *a = (struct esas2r_adapter *)sgc->adapter; in get_physaddr_vda()
1817 int offset = (u8 *)sgc->cur_offset - (u8 *)a->vda_buffer; in get_physaddr_vda()
1819 (*addr) = a->ppvda_buffer + offset; in get_physaddr_vda()
1820 return VDA_MAX_BUFFER_SIZE - offset; in get_physaddr_vda()
1826 if (!a->vda_buffer) in esas2r_read_vda()
1827 return -ENOMEM; in esas2r_read_vda()
1832 (struct atto_ioctl_vda *)a->vda_buffer; in esas2r_read_vda()
1846 return -EBUSY; in esas2r_read_vda()
1849 rq->comp_cb = vda_complete_req; in esas2r_read_vda()
1853 sgc.cur_offset = a->vda_buffer + VDA_BUFFER_HEADER_SZ; in esas2r_read_vda()
1856 a->vda_command_done = 0; in esas2r_read_vda()
1864 while (!a->vda_command_done) in esas2r_read_vda()
1865 wait_event_interruptible(a->vda_waiter, in esas2r_read_vda()
1866 a->vda_command_done); in esas2r_read_vda()
1876 count = VDA_MAX_BUFFER_SIZE - off; in esas2r_read_vda()
1881 memcpy(buf, a->vda_buffer + off, count); in esas2r_read_vda()
1895 if (!a->vda_buffer) { in esas2r_write_vda()
1897 a->vda_buffer = dma_alloc_coherent(&a->pcid->dev, in esas2r_write_vda()
1903 a->ppvda_buffer = dma_addr; in esas2r_write_vda()
1906 if (!a->vda_buffer) in esas2r_write_vda()
1907 return -ENOMEM; in esas2r_write_vda()
1913 count = VDA_MAX_BUFFER_SIZE - off; in esas2r_write_vda()
1918 memcpy(a->vda_buffer + off, buf, count); in esas2r_write_vda()
1927 a->fs_api_command_done = 1; in fs_api_complete_req()
1929 wake_up_interruptible(&a->fs_api_waiter); in fs_api_complete_req()
1935 struct esas2r_adapter *a = (struct esas2r_adapter *)sgc->adapter; in get_physaddr_fs_api()
1937 (struct esas2r_ioctl_fs *)a->fs_api_buffer; in get_physaddr_fs_api()
1938 u32 offset = (u8 *)sgc->cur_offset - (u8 *)fs; in get_physaddr_fs_api()
1940 (*addr) = a->ppfs_api_buffer + offset; in get_physaddr_fs_api()
1942 return a->fs_api_buffer_size - offset; in get_physaddr_fs_api()
1948 if (!a->fs_api_buffer) in esas2r_read_fs()
1949 return -ENOMEM; in esas2r_read_fs()
1955 (struct esas2r_ioctl_fs *)a->fs_api_buffer; in esas2r_read_fs()
1958 if (mutex_lock_interruptible(&a->fs_api_mutex)) { in esas2r_read_fs()
1960 fs->status = ATTO_STS_OUT_OF_RSRC; in esas2r_read_fs()
1961 return -EBUSY; in esas2r_read_fs()
1974 mutex_unlock(&a->fs_api_mutex); in esas2r_read_fs()
1978 rq->comp_cb = fs_api_complete_req; in esas2r_read_fs()
1982 sgc.cur_offset = fs->data; in esas2r_read_fs()
1985 a->fs_api_command_done = 0; in esas2r_read_fs()
1988 if (fs->status == ATTO_STS_OUT_OF_RSRC) in esas2r_read_fs()
1989 count = -EBUSY; in esas2r_read_fs()
1996 while (!a->fs_api_command_done) in esas2r_read_fs()
1997 wait_event_interruptible(a->fs_api_waiter, in esas2r_read_fs()
1998 a->fs_api_command_done); in esas2r_read_fs()
2002 mutex_unlock(&a->fs_api_mutex); in esas2r_read_fs()
2010 if (off > a->fs_api_buffer_size) in esas2r_read_fs()
2013 if (count + off > a->fs_api_buffer_size) in esas2r_read_fs()
2014 count = a->fs_api_buffer_size - off; in esas2r_read_fs()
2019 memcpy(buf, a->fs_api_buffer + off, count); in esas2r_read_fs()
2030 u32 length = fs->command.length + offsetof( in esas2r_write_fs()
2039 if (fs->command.command == ESAS2R_FS_CMD_BEGINW) in esas2r_write_fs()
2049 return -EINVAL; in esas2r_write_fs()
2052 if (a->fs_api_buffer) { in esas2r_write_fs()
2053 if (a->fs_api_buffer_size < length) { in esas2r_write_fs()
2054 /* Free too-small buffer and get a new one */ in esas2r_write_fs()
2055 dma_free_coherent(&a->pcid->dev, in esas2r_write_fs()
2056 (size_t)a->fs_api_buffer_size, in esas2r_write_fs()
2057 a->fs_api_buffer, in esas2r_write_fs()
2058 (dma_addr_t)a->ppfs_api_buffer); in esas2r_write_fs()
2064 a->fs_api_buffer_size = length; in esas2r_write_fs()
2066 a->fs_api_buffer = dma_alloc_coherent(&a->pcid->dev, in esas2r_write_fs()
2067 (size_t)a->fs_api_buffer_size, in esas2r_write_fs()
2068 (dma_addr_t *)&a->ppfs_api_buffer, in esas2r_write_fs()
2073 if (!a->fs_api_buffer) in esas2r_write_fs()
2074 return -ENOMEM; in esas2r_write_fs()
2076 if (off > a->fs_api_buffer_size) in esas2r_write_fs()
2079 if (count + off > a->fs_api_buffer_size) in esas2r_write_fs()
2080 count = a->fs_api_buffer_size - off; in esas2r_write_fs()
2085 memcpy(a->fs_api_buffer + off, buf, count); in esas2r_write_fs()