Lines Matching refs:ccb

71 static void mvs_softreset(device_t dev, union ccb *ccb);
76 static void mvs_tfd_read(device_t dev, union ccb *ccb);
77 static void mvs_tfd_write(device_t dev, union ccb *ccb);
80 static void mvs_begin_transaction(device_t dev, union ccb *ccb);
90 static void mvs_process_read_log(device_t dev, union ccb *ccb);
91 static void mvs_process_request_sense(device_t dev, union ccb *ccb);
93 static void mvsaction(struct cam_sim *sim, union ccb *ccb);
438 slot->ccb = NULL; in mvs_slotsalloc()
599 union ccb *ccb; in mvs_phy_check_events() local
610 if ((ccb = xpt_alloc_ccb_nowait()) == NULL) in mvs_phy_check_events()
612 if (xpt_create_path(&ccb->ccb_h.path, NULL, in mvs_phy_check_events()
615 xpt_free_ccb(ccb); in mvs_phy_check_events()
618 xpt_rescan(ccb); in mvs_phy_check_events()
776 ch->slot[i].ccb->ccb_h.target_id != port) in mvs_ch_intr()
840 union ccb *ccb = slot->ccb; in mvs_legacy_intr() local
866 if (ccb->ccb_h.func_code == XPT_ATA_IO) { /* ATA PIO */ in mvs_legacy_intr()
867 ccb->ataio.res.status = status; in mvs_legacy_intr()
869 if ((ccb->ccb_h.flags & CAM_DIR_MASK) != CAM_DIR_NONE) { in mvs_legacy_intr()
871 if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) { in mvs_legacy_intr()
880 (uint16_t *)(ccb->ataio.data_ptr + ch->donecount), in mvs_legacy_intr()
886 if (ccb->ataio.dxfer_len > ch->donecount) { in mvs_legacy_intr()
888 ch->transfersize = min(ccb->ataio.dxfer_len - ch->donecount, in mvs_legacy_intr()
891 if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_OUT) { in mvs_legacy_intr()
901 (uint16_t *)(ccb->ataio.data_ptr + ch->donecount), in mvs_legacy_intr()
906 if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) in mvs_legacy_intr()
931 if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) { in mvs_legacy_intr()
938 (uint16_t *)(ccb->csio.data_ptr + ch->donecount), in mvs_legacy_intr()
945 ch->transfersize = min(ccb->csio.dxfer_len - ch->donecount, in mvs_legacy_intr()
946 ch->curr[ccb->ccb_h.target_id].bytecount); in mvs_legacy_intr()
951 if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_OUT) { in mvs_legacy_intr()
958 (uint16_t *)(ccb->csio.data_ptr + ch->donecount), in mvs_legacy_intr()
963 ((uint8_t *)ccb->csio.data_ptr + ch->donecount + in mvs_legacy_intr()
971 ch->transfersize = min(ccb->csio.dxfer_len - ch->donecount, in mvs_legacy_intr()
972 ch->curr[ccb->ccb_h.target_id].bytecount); in mvs_legacy_intr()
979 if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) { in mvs_legacy_intr()
981 (uint16_t *)(ccb->csio.data_ptr + ch->donecount), in mvs_legacy_intr()
985 else if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_OUT) { in mvs_legacy_intr()
987 (uint16_t *)(ccb->csio.data_ptr + ch->donecount), in mvs_legacy_intr()
1018 union ccb *ccb; in mvs_crbq_intr() local
1056 ccb = ch->slot[slot].ccb; in mvs_crbq_intr()
1057 ccb->ataio.res.status = in mvs_crbq_intr()
1084 mvs_check_collision(device_t dev, union ccb *ccb) in mvs_check_collision() argument
1088 if (ccb->ccb_h.func_code == XPT_ATA_IO) { in mvs_check_collision()
1090 if (ccb->ataio.cmd.flags & CAM_ATAIO_FPDMA) { in mvs_check_collision()
1101 ch->taggedtarget != ccb->ccb_h.target_id) in mvs_check_collision()
1105 } else if (ccb->ataio.cmd.flags & CAM_ATAIO_DMA) { in mvs_check_collision()
1118 if (ccb->ataio.cmd.flags & (CAM_ATAIO_CONTROL | CAM_ATAIO_NEEDRESULT)) { in mvs_check_collision()
1135 mvs_tfd_read(device_t dev, union ccb *ccb) in mvs_tfd_read() argument
1138 struct ata_res *res = &ccb->ataio.res; in mvs_tfd_read()
1156 mvs_tfd_write(device_t dev, union ccb *ccb) in mvs_tfd_write() argument
1159 struct ata_cmd *cmd = &ccb->ataio.cmd; in mvs_tfd_write()
1178 mvs_begin_transaction(device_t dev, union ccb *ccb) in mvs_begin_transaction() argument
1187 if (ccb->ccb_h.func_code == XPT_ATA_IO && in mvs_begin_transaction()
1188 (ccb->ataio.cmd.flags & CAM_ATAIO_CONTROL)) { in mvs_begin_transaction()
1189 mvs_softreset(dev, ccb); in mvs_begin_transaction()
1194 if ((ccb->ccb_h.func_code == XPT_ATA_IO) && in mvs_begin_transaction()
1195 (ccb->ataio.cmd.flags & CAM_ATAIO_FPDMA)) { in mvs_begin_transaction()
1197 tag = ffs(~ch->otagspd[ccb->ccb_h.target_id]) - 1; in mvs_begin_transaction()
1204 slot->ccb = ccb; in mvs_begin_transaction()
1212 ch->numrslotspd[ccb->ccb_h.target_id]++; in mvs_begin_transaction()
1213 if (ccb->ccb_h.func_code == XPT_ATA_IO) { in mvs_begin_transaction()
1214 if (ccb->ataio.cmd.flags & CAM_ATAIO_FPDMA) { in mvs_begin_transaction()
1215 ch->otagspd[ccb->ccb_h.target_id] |= (1 << slot->tag); in mvs_begin_transaction()
1217 ch->numtslotspd[ccb->ccb_h.target_id]++; in mvs_begin_transaction()
1218 ch->taggedtarget = ccb->ccb_h.target_id; in mvs_begin_transaction()
1220 } else if (ccb->ataio.cmd.flags & CAM_ATAIO_DMA) { in mvs_begin_transaction()
1227 if (ccb->ataio.cmd.flags & in mvs_begin_transaction()
1232 uint8_t *cdb = (ccb->ccb_h.flags & CAM_CDB_POINTER) ? in mvs_begin_transaction()
1233 ccb->csio.cdb_io.cdb_ptr : ccb->csio.cdb_io.cdb_bytes; in mvs_begin_transaction()
1236 if ((ccb->ccb_h.flags & CAM_DIR_MASK) != CAM_DIR_NONE && in mvs_begin_transaction()
1237 ch->curr[ccb->ccb_h.target_id].mode >= ATA_DMA && in mvs_begin_transaction()
1255 ccb, mvs_dmasetprd, slot, 0); in mvs_begin_transaction()
1292 ((slot->ccb->ccb_h.flags & CAM_DIR_IN) ? in mvs_dmasetprd()
1306 union ccb *ccb = slot->ccb; in mvs_legacy_execute_transaction() local
1307 int port = ccb->ccb_h.target_id & 0x0f; in mvs_legacy_execute_transaction()
1313 if (ccb->ccb_h.func_code == XPT_ATA_IO) { in mvs_legacy_execute_transaction()
1314 mvs_tfd_write(dev, ccb); in mvs_legacy_execute_transaction()
1316 if (ccb->ataio.cmd.command == ATA_DEVICE_RESET) { in mvs_legacy_execute_transaction()
1320 ccb->ataio.res.status = ATA_INB(ch->r_mem, ATA_STATUS); in mvs_legacy_execute_transaction()
1321 } while (ccb->ataio.res.status & ATA_S_BUSY && timeout--); in mvs_legacy_execute_transaction()
1326 if (ccb->ataio.cmd.command == ATA_READ_MUL || in mvs_legacy_execute_transaction()
1327 ccb->ataio.cmd.command == ATA_READ_MUL48 || in mvs_legacy_execute_transaction()
1328 ccb->ataio.cmd.command == ATA_WRITE_MUL || in mvs_legacy_execute_transaction()
1329 ccb->ataio.cmd.command == ATA_WRITE_MUL48) { in mvs_legacy_execute_transaction()
1330 ch->transfersize = min(ccb->ataio.dxfer_len, in mvs_legacy_execute_transaction()
1333 ch->transfersize = min(ccb->ataio.dxfer_len, 512); in mvs_legacy_execute_transaction()
1334 if ((ccb->ccb_h.flags & CAM_DIR_MASK) != CAM_DIR_NONE) in mvs_legacy_execute_transaction()
1337 if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_OUT) { in mvs_legacy_execute_transaction()
1347 (uint16_t *)(ccb->ataio.data_ptr + ch->donecount), in mvs_legacy_execute_transaction()
1352 ch->transfersize = min(ccb->csio.dxfer_len, in mvs_legacy_execute_transaction()
1394 (uint16_t *)((ccb->ccb_h.flags & CAM_CDB_POINTER) ? in mvs_legacy_execute_transaction()
1395 ccb->csio.cdb_io.cdb_ptr : ccb->csio.cdb_io.cdb_bytes), in mvs_legacy_execute_transaction()
1404 (((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) ? in mvs_legacy_execute_transaction()
1409 callout_reset_sbt(&slot->timeout, SBT_1MS * ccb->ccb_h.timeout, 0, in mvs_legacy_execute_transaction()
1422 union ccb *ccb = slot->ccb; in mvs_execute_transaction() local
1423 int port = ccb->ccb_h.target_id & 0x0f; in mvs_execute_transaction()
1433 ((ccb->ccb_h.flags & CAM_DIR_IN) ? MVS_CRQB2E_READ : 0) | in mvs_execute_transaction()
1447 crqb2e->cmd[2] = ccb->ataio.cmd.command; in mvs_execute_transaction()
1448 crqb2e->cmd[3] = ccb->ataio.cmd.features; in mvs_execute_transaction()
1449 crqb2e->cmd[4] = ccb->ataio.cmd.lba_low; in mvs_execute_transaction()
1450 crqb2e->cmd[5] = ccb->ataio.cmd.lba_mid; in mvs_execute_transaction()
1451 crqb2e->cmd[6] = ccb->ataio.cmd.lba_high; in mvs_execute_transaction()
1452 crqb2e->cmd[7] = ccb->ataio.cmd.device; in mvs_execute_transaction()
1453 crqb2e->cmd[8] = ccb->ataio.cmd.lba_low_exp; in mvs_execute_transaction()
1454 crqb2e->cmd[9] = ccb->ataio.cmd.lba_mid_exp; in mvs_execute_transaction()
1455 crqb2e->cmd[10] = ccb->ataio.cmd.lba_high_exp; in mvs_execute_transaction()
1456 crqb2e->cmd[11] = ccb->ataio.cmd.features_exp; in mvs_execute_transaction()
1457 if (ccb->ataio.cmd.flags & CAM_ATAIO_FPDMA) { in mvs_execute_transaction()
1461 crqb2e->cmd[12] = ccb->ataio.cmd.sector_count; in mvs_execute_transaction()
1462 crqb2e->cmd[13] = ccb->ataio.cmd.sector_count_exp; in mvs_execute_transaction()
1472 ((ccb->ccb_h.flags & CAM_DIR_IN) ? MVS_CRQB_READ : 0) | in mvs_execute_transaction()
1480 if (ccb->ataio.cmd.flags & CAM_ATAIO_FPDMA) { in mvs_execute_transaction()
1481 crqb->cmd[i++] = ccb->ataio.cmd.features_exp; in mvs_execute_transaction()
1484 crqb->cmd[i++] = ccb->ataio.cmd.features; in mvs_execute_transaction()
1486 if (ccb->ataio.cmd.flags & CAM_ATAIO_FPDMA) { in mvs_execute_transaction()
1488 (ccb->ataio.cmd.sector_count & 0x07); in mvs_execute_transaction()
1491 crqb->cmd[i++] = ccb->ataio.cmd.sector_count_exp; in mvs_execute_transaction()
1493 crqb->cmd[i++] = ccb->ataio.cmd.sector_count; in mvs_execute_transaction()
1496 crqb->cmd[i++] = ccb->ataio.cmd.lba_low_exp; in mvs_execute_transaction()
1498 crqb->cmd[i++] = ccb->ataio.cmd.lba_low; in mvs_execute_transaction()
1500 crqb->cmd[i++] = ccb->ataio.cmd.lba_mid_exp; in mvs_execute_transaction()
1502 crqb->cmd[i++] = ccb->ataio.cmd.lba_mid; in mvs_execute_transaction()
1504 crqb->cmd[i++] = ccb->ataio.cmd.lba_high_exp; in mvs_execute_transaction()
1506 crqb->cmd[i++] = ccb->ataio.cmd.lba_high; in mvs_execute_transaction()
1508 crqb->cmd[i++] = ccb->ataio.cmd.device; in mvs_execute_transaction()
1510 crqb->cmd[i++] = ccb->ataio.cmd.command; in mvs_execute_transaction()
1524 callout_reset_sbt(&slot->timeout, SBT_1MS * ccb->ccb_h.timeout, 0, in mvs_execute_transaction()
1563 SBT_1MS * slot->ccb->ccb_h.timeout / 2, 0, in mvs_rearm_timeout()
1606 union ccb *ccb = slot->ccb; in mvs_end_transaction() local
1615 if (ccb->ccb_h.func_code == XPT_ATA_IO) { in mvs_end_transaction()
1616 struct ata_res *res = &ccb->ataio.res; in mvs_end_transaction()
1619 (ccb->ataio.cmd.flags & CAM_ATAIO_NEEDRESULT)) { in mvs_end_transaction()
1620 mvs_tfd_read(dev, ccb); in mvs_end_transaction()
1624 if ((ccb->ccb_h.flags & CAM_DIR_MASK) != CAM_DIR_NONE && in mvs_end_transaction()
1626 ccb->csio.resid = ccb->csio.dxfer_len - ch->donecount; in mvs_end_transaction()
1629 if ((ccb->ccb_h.flags & CAM_DIR_MASK) != CAM_DIR_NONE) { in mvs_end_transaction()
1631 (ccb->ccb_h.flags & CAM_DIR_IN) ? in mvs_end_transaction()
1640 !(ccb->ccb_h.status & CAM_DEV_QFRZN)) { in mvs_end_transaction()
1641 xpt_freeze_devq(ccb->ccb_h.path, 1); in mvs_end_transaction()
1642 ccb->ccb_h.status |= CAM_DEV_QFRZN; in mvs_end_transaction()
1645 ccb->ccb_h.status &= ~CAM_STATUS_MASK; in mvs_end_transaction()
1648 ccb->ccb_h.status |= CAM_REQ_CMP; in mvs_end_transaction()
1649 if (ccb->ccb_h.func_code == XPT_SCSI_IO) in mvs_end_transaction()
1650 ccb->csio.scsi_status = SCSI_STATUS_OK; in mvs_end_transaction()
1654 ccb->ccb_h.status |= CAM_REQ_INVALID; in mvs_end_transaction()
1657 ccb->ccb_h.status |= CAM_REQUEUE_REQ; in mvs_end_transaction()
1661 if (ccb->ccb_h.func_code == XPT_SCSI_IO) { in mvs_end_transaction()
1662 ccb->ccb_h.status |= CAM_SCSI_STATUS_ERROR; in mvs_end_transaction()
1663 ccb->csio.scsi_status = SCSI_STATUS_CHECK_COND; in mvs_end_transaction()
1665 ccb->ccb_h.status |= CAM_ATA_STATUS_ERROR; in mvs_end_transaction()
1672 ccb->ccb_h.status &= ~CAM_STATUS_MASK; in mvs_end_transaction()
1673 ccb->ccb_h.status |= CAM_RELEASE_SIMQ; in mvs_end_transaction()
1675 ccb->ccb_h.status |= CAM_UNCOR_PARITY; in mvs_end_transaction()
1680 ccb->ccb_h.status &= ~CAM_STATUS_MASK; in mvs_end_transaction()
1681 ccb->ccb_h.status |= CAM_RELEASE_SIMQ; in mvs_end_transaction()
1683 ccb->ccb_h.status |= CAM_CMD_TIMEOUT; in mvs_end_transaction()
1687 ccb->ccb_h.status |= CAM_REQ_CMP_ERR; in mvs_end_transaction()
1694 slot->ccb = NULL; in mvs_end_transaction()
1697 ch->numrslotspd[ccb->ccb_h.target_id]--; in mvs_end_transaction()
1698 if (ccb->ccb_h.func_code == XPT_ATA_IO) { in mvs_end_transaction()
1699 if (ccb->ataio.cmd.flags & CAM_ATAIO_FPDMA) { in mvs_end_transaction()
1700 ch->otagspd[ccb->ccb_h.target_id] &= ~(1 << slot->tag); in mvs_end_transaction()
1702 ch->numtslotspd[ccb->ccb_h.target_id]--; in mvs_end_transaction()
1703 } else if (ccb->ataio.cmd.flags & CAM_ATAIO_DMA) { in mvs_end_transaction()
1720 if (ccb->ccb_h.recovery_type == RECOVERY_READ_LOG) { in mvs_end_transaction()
1721 mvs_process_read_log(dev, ccb); in mvs_end_transaction()
1723 } else if (ccb->ccb_h.recovery_type == RECOVERY_REQUEST_SENSE) { in mvs_end_transaction()
1724 mvs_process_request_sense(dev, ccb); in mvs_end_transaction()
1727 ((ccb->ccb_h.status & CAM_STATUS_MASK) == CAM_SCSI_STATUS_ERROR && in mvs_end_transaction()
1728 (ccb->ccb_h.flags & CAM_DIS_AUTOSENSE) == 0)) { in mvs_end_transaction()
1729 ch->hold[slot->slot] = ccb; in mvs_end_transaction()
1733 xpt_done(ccb); in mvs_end_transaction()
1755 union ccb *fccb = ch->frozen; in mvs_end_transaction()
1772 union ccb *ccb; in mvs_issue_recovery() local
1782 ccb = xpt_alloc_ccb_nowait(); in mvs_issue_recovery()
1783 if (ccb == NULL) { in mvs_issue_recovery()
1799 xpt_setup_ccb(&ccb->ccb_h, ch->hold[i]->ccb_h.path, in mvs_issue_recovery()
1803 ccb->ccb_h.recovery_type = RECOVERY_READ_LOG; in mvs_issue_recovery()
1804 ccb->ccb_h.func_code = XPT_ATA_IO; in mvs_issue_recovery()
1805 ccb->ccb_h.flags = CAM_DIR_IN; in mvs_issue_recovery()
1806 ccb->ccb_h.timeout = 1000; /* 1s should be enough. */ in mvs_issue_recovery()
1807 ataio = &ccb->ataio; in mvs_issue_recovery()
1810 xpt_free_ccb(ccb); in mvs_issue_recovery()
1826 ccb->ccb_h.recovery_type = RECOVERY_REQUEST_SENSE; in mvs_issue_recovery()
1827 ccb->ccb_h.recovery_slot = i; in mvs_issue_recovery()
1828 ccb->ccb_h.func_code = XPT_SCSI_IO; in mvs_issue_recovery()
1829 ccb->ccb_h.flags = CAM_DIR_IN; in mvs_issue_recovery()
1830 ccb->ccb_h.status = 0; in mvs_issue_recovery()
1831 ccb->ccb_h.timeout = 1000; /* 1s should be enough. */ in mvs_issue_recovery()
1832 csio = &ccb->csio; in mvs_issue_recovery()
1843 mvs_begin_transaction(dev, ccb); in mvs_issue_recovery()
1847 mvs_process_read_log(device_t dev, union ccb *ccb) in mvs_process_read_log() argument
1856 data = ccb->ataio.data_ptr; in mvs_process_read_log()
1857 if ((ccb->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP && in mvs_process_read_log()
1862 if (ch->hold[i]->ccb_h.target_id != ccb->ccb_h.target_id) in mvs_process_read_log()
1886 if ((ccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_CMP) in mvs_process_read_log()
1895 if (ch->hold[i]->ccb_h.target_id != ccb->ccb_h.target_id) in mvs_process_read_log()
1902 free(ccb->ataio.data_ptr, M_MVS); in mvs_process_read_log()
1903 xpt_free_ccb(ccb); in mvs_process_read_log()
1908 mvs_process_request_sense(device_t dev, union ccb *ccb) in mvs_process_request_sense() argument
1915 i = ccb->ccb_h.recovery_slot; in mvs_process_request_sense()
1916 if ((ccb->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP) { in mvs_process_request_sense()
1925 xpt_free_ccb(ccb); in mvs_process_request_sense()
1951 union ccb *fccb = ch->frozen; in mvs_requeue_frozen()
2107 mvs_softreset(device_t dev, union ccb *ccb) in mvs_softreset() argument
2110 int port = ccb->ccb_h.target_id & 0x0f; in mvs_softreset()
2119 ccb->ccb_h.status &= ~CAM_STATUS_MASK; in mvs_softreset()
2121 if ((i = mvs_wait(dev, 0, ATA_S_BUSY, ccb->ccb_h.timeout)) < 0) { in mvs_softreset()
2122 ccb->ccb_h.status |= CAM_CMD_TIMEOUT; in mvs_softreset()
2127 ccb->ccb_h.status |= CAM_ATA_STATUS_ERROR; in mvs_softreset()
2129 ccb->ccb_h.status |= CAM_REQ_CMP; in mvs_softreset()
2135 mvs_tfd_read(dev, ccb); in mvs_softreset()
2147 mvs_wait(dev, 0, ATA_S_BUSY | ATA_S_DRQ, ccb->ccb_h.timeout); in mvs_softreset()
2150 xpt_done(ccb); in mvs_softreset()
2228 mvs_check_ids(device_t dev, union ccb *ccb) in mvs_check_ids() argument
2232 if (ccb->ccb_h.target_id > ((ch->quirks & MVS_Q_GENI) ? 0 : 15)) { in mvs_check_ids()
2233 ccb->ccb_h.status = CAM_TID_INVALID; in mvs_check_ids()
2234 xpt_done(ccb); in mvs_check_ids()
2237 if (ccb->ccb_h.target_lun != 0) { in mvs_check_ids()
2238 ccb->ccb_h.status = CAM_LUN_INVALID; in mvs_check_ids()
2239 xpt_done(ccb); in mvs_check_ids()
2245 KASSERT(ccb->ccb_h.func_code != XPT_ATA_IO || in mvs_check_ids()
2246 ((ccb->ataio.ata_flags & ATA_FLAG_AUX) == 0), in mvs_check_ids()
2252 mvsaction(struct cam_sim *sim, union ccb *ccb) in mvsaction() argument
2257 CAM_DEBUG(ccb->ccb_h.path, CAM_DEBUG_TRACE, ("mvsaction func_code=%x\n", in mvsaction()
2258 ccb->ccb_h.func_code)); in mvsaction()
2262 switch (ccb->ccb_h.func_code) { in mvsaction()
2266 if (mvs_check_ids(dev, ccb)) in mvsaction()
2270 ccb->ccb_h.target_id > 0 && ccb->ccb_h.target_id < 15)) { in mvsaction()
2271 ccb->ccb_h.status = CAM_SEL_TIMEOUT; in mvsaction()
2274 ccb->ccb_h.recovery_type = RECOVERY_NONE; in mvsaction()
2276 if (mvs_check_collision(dev, ccb)) { in mvsaction()
2278 ch->frozen = ccb; in mvsaction()
2283 mvs_begin_transaction(dev, ccb); in mvsaction()
2287 ccb->ccb_h.status = CAM_REQ_INVALID; in mvsaction()
2291 struct ccb_trans_settings *cts = &ccb->cts; in mvsaction()
2294 if (mvs_check_ids(dev, ccb)) in mvsaction()
2297 d = &ch->curr[ccb->ccb_h.target_id]; in mvsaction()
2299 d = &ch->user[ccb->ccb_h.target_id]; in mvsaction()
2316 ccb->ccb_h.status = CAM_REQ_CMP; in mvsaction()
2322 struct ccb_trans_settings *cts = &ccb->cts; in mvsaction()
2326 if (mvs_check_ids(dev, ccb)) in mvsaction()
2329 d = &ch->curr[ccb->ccb_h.target_id]; in mvsaction()
2331 d = &ch->user[ccb->ccb_h.target_id]; in mvsaction()
2339 (ccb->ccb_h.target_id == 15 || in mvsaction()
2340 (ccb->ccb_h.target_id == 0 && !ch->pm_present))) { in mvsaction()
2353 ch->user[ccb->ccb_h.target_id].caps; in mvsaction()
2374 ccb->ccb_h.status = CAM_REQ_CMP; in mvsaction()
2380 ccb->ccb_h.status = CAM_REQ_CMP; in mvsaction()
2384 ccb->ccb_h.status = CAM_REQ_INVALID; in mvsaction()
2388 struct ccb_pathinq *cpi = &ccb->cpi; in mvsaction()
2429 ccb->ccb_h.status = CAM_REQ_INVALID; in mvsaction()
2432 xpt_done(ccb); in mvsaction()