Lines Matching full:wu

42   * because we assert that only one ccb per WU will ever be active.
115 -sr_crypto_wu_get(struct sr_workunit *wu, int encrypt)
116 +sr_crypto_prepare(struct sr_workunit *wu, int encrypt)
118 struct scsi_xfer *xs = wu->swu_xs;
119 struct sr_discipline *sd = wu->swu_dis;
127 - DNPRINTF(SR_D_DIS, "%s: sr_crypto_wu_get wu: %p encrypt: %d\n",
128 + DNPRINTF(SR_D_DIS, "%s: sr_crypto_prepare wu %p encrypt %d\n",
129 DEVNAME(sd->sd_sc), wu, encrypt);
138 + crwu = (struct sr_crypto_wu *)wu;
155 + blkno = wu->swu_blk_start;
222 - crwu->cr_wu = wu;
238 - struct sr_workunit *wu = crwu->cr_wu;
239 - struct sr_discipline *sd = wu->swu_dis;
242 - DEVNAME(wu->swu_dis->sd_sc), crwu);
528 + struct sr_workunit *wu;
556 - * For each wu allocate the uio, iovec and crypto structures.
576 + TAILQ_FOREACH(wu, &sd->sd_wu, swu_next) {
577 + crwu = (struct sr_crypto_wu *)wu;
632 + struct sr_workunit *wu;
662 + TAILQ_FOREACH(wu, &sd->sd_wu, swu_next) {
663 + crwu = (struct sr_crypto_wu *)wu;
679 sr_crypto_rw(struct sr_workunit *wu)
686 - DNPRINTF(SR_D_DIS, "%s: sr_crypto_rw wu: %p\n",
687 + DNPRINTF(SR_D_DIS, "%s: sr_crypto_rw wu %p\n",
688 DEVNAME(wu->swu_dis->sd_sc), wu);
690 - if (wu->swu_xs->flags & SCSI_DATA_OUT) {
691 - crwu = sr_crypto_wu_get(wu, 1);
694 + if (sr_validate_io(wu, &blkno, "sr_crypto_rw"))
697 + if (wu->swu_xs->flags & SCSI_DATA_OUT) {
698 + crwu = sr_crypto_prepare(wu, 1);
709 - rv = sr_crypto_rw2(wu, NULL);
710 + rv = sr_crypto_dev_rw(wu, NULL);
720 - struct sr_workunit *wu = crwu->cr_wu;
721 + struct sr_workunit *wu = &crwu->cr_wu;
724 - DNPRINTF(SR_D_INTR, "%s: sr_crypto_write: wu %x xs: %x\n",
725 + DNPRINTF(SR_D_INTR, "%s: sr_crypto_write: wu %p xs: %p\n",
726 DEVNAME(wu->swu_dis->sd_sc), wu, wu->swu_xs);
730 wu->swu_xs->error = XS_DRIVER_STUFFUP;
732 - sr_crypto_finish_io(wu);
733 + sr_scsi_done(wu->swu_dis, wu->swu_xs);
737 - return (sr_crypto_rw2(wu, crwu));
738 + sr_crypto_dev_rw(wu, crwu);
742 -sr_crypto_rw2(struct sr_workunit *wu, struct sr_crypto_wu *crwu)
743 +sr_crypto_dev_rw(struct sr_workunit *wu, struct sr_crypto_wu *crwu)
745 struct sr_discipline *sd = wu->swu_dis;
746 struct scsi_xfer *xs = wu->swu_xs;
753 - if (sr_validate_io(wu, &blk, "sr_crypto_rw2"))
755 + blkno = wu->swu_blk_start;
767 sr_wu_enqueue_ccb(wu, ccb);
768 + sr_schedule_wu(wu);
772 - if (sr_check_io_collision(wu))
775 - sr_raid_startwu(wu);
782 /* wu is unwound by sr_wu_put */
786 struct scsi_xfer *xs = wu->swu_xs;
794 - crwu = sr_crypto_wu_get(wu, 0);
796 - panic("sr_crypto_intr: no wu");
797 + crwu = sr_crypto_prepare(wu, 0);
799 - ccb = TAILQ_FIRST(&wu->swu_ccb);
805 DEVNAME(wu->swu_dis->sd_sc), crwu->cr_crp);
814 - sr_crypto_finish_io(wu);
815 + sr_scsi_done(wu->swu_dis, wu->swu_xs);
820 -sr_crypto_finish_io(struct sr_workunit *wu)
822 - struct sr_discipline *sd = wu->swu_dis;
823 - struct scsi_xfer *xs = wu->swu_xs;
831 - DNPRINTF(SR_D_INTR, "%s: sr_crypto_finish_io: wu %x xs: %x\n",
832 - DEVNAME(sc), wu, xs);
834 - if (wu->swu_cb_active == 1)
836 - TAILQ_FOREACH(ccb, &wu->swu_ccb, ccb_link) {
849 - struct sr_workunit *wu = crwu->cr_wu;
850 + struct sr_workunit *wu = &crwu->cr_wu;
853 - DNPRINTF(SR_D_INTR, "%s: sr_crypto_read: wu %x xs: %x\n",
854 + DNPRINTF(SR_D_INTR, "%s: sr_crypto_read: wu %p xs: %p\n",
855 DEVNAME(wu->swu_dis->sd_sc), wu, wu->swu_xs);
858 wu->swu_xs->error = XS_DRIVER_STUFFUP;
861 - sr_crypto_finish_io(wu);
862 + sr_scsi_done(wu->swu_dis, wu->swu_xs);