Lines Matching refs:private

37 	struct ccw_dev_id *id = &cdev->private->dev_id;
38 int mpath = cdev->private->flags.mpath;
39 int pgroup = cdev->private->flags.pgroup;
60 struct ccw_request *req = &cdev->private->req;
61 struct ccw1 *cp = cdev->private->dma_area->iccws;
76 struct ccw_request *req = &cdev->private->req;
79 ~cdev->private->path_noirq_mask);
108 struct ccw_request *req = &cdev->private->req;
115 cdev->private->path_noirq_mask |= req->lpm;
118 cdev->private->path_notoper_mask |= req->lpm;
137 struct ccw_request *req = &cdev->private->req;
138 struct ccw1 *cp = cdev->private->dma_area->iccws;
140 struct pgid *pgid = &cdev->private->dma_area->pgid[i];
161 cdev->private->flags.pgid_unknown = 0;
171 struct ccw_dev_id *id = &cdev->private->dev_id;
172 struct ccw_request *req = &cdev->private->req;
176 id->ssid, id->devno, cdev->private->pgid_valid_mask,
177 cdev->private->path_noirq_mask);
186 if (cdev->private->flags.mpath)
198 struct ccw_request *req = &cdev->private->req;
202 req->lpm = lpm_adjust(req->lpm, cdev->private->pgid_todo_mask);
210 if (cdev->private->flags.mpath)
217 if (cdev->private->flags.pgid_unknown) {
231 struct ccw_request *req = &cdev->private->req;
238 cdev->private->flags.pgid_unknown = 1;
239 cdev->private->path_noirq_mask |= req->lpm;
242 cdev->private->path_notoper_mask |= req->lpm;
245 if (cdev->private->flags.mpath) {
247 cdev->private->flags.mpath = 0;
251 cdev->private->flags.pgroup = 0;
269 struct ccw_request *req = &cdev->private->req;
304 struct pgid *pgid = &cdev->private->dma_area->pgid[0];
313 if ((cdev->private->pgid_valid_mask & lpm) == 0)
344 if ((cdev->private->pgid_valid_mask & lpm) == 0)
346 pgid = &cdev->private->dma_area->pgid[i];
354 if (cdev->private->flags.mpath) {
372 memcpy(&cdev->private->dma_area->pgid[i], pgid,
381 struct ccw_dev_id *id = &cdev->private->dev_id;
392 if (reserved == cdev->private->pgid_valid_mask)
399 cdev->private->pgid_reset_mask |= reset;
400 cdev->private->pgid_todo_mask &=
401 ~(donepm | cdev->private->path_noirq_mask);
407 id->devno, rc, cdev->private->pgid_valid_mask, sch->vpm,
408 cdev->private->pgid_todo_mask, mismatch, reserved, reset);
411 if (cdev->private->flags.pgid_unknown) {
416 if (cdev->private->pgid_todo_mask == 0) {
425 cdev->private->flags.pgroup = 0;
426 cdev->private->flags.mpath = 0;
439 struct ccw_request *req = &cdev->private->req;
440 struct ccw1 *cp = cdev->private->dma_area->iccws;
445 cp->cda = virt_to_dma32(&cdev->private->dma_area->pgid[i]);
457 struct ccw_request *req = &cdev->private->req;
461 ~cdev->private->path_noirq_mask);
469 if (cdev->private->pgid_valid_mask)
471 else if (cdev->private->path_noirq_mask)
483 struct ccw_request *req = &cdev->private->req;
487 cdev->private->pgid_valid_mask |= req->lpm;
490 cdev->private->flags.pgid_unknown = 1;
491 cdev->private->path_noirq_mask |= req->lpm;
494 cdev->private->path_notoper_mask |= req->lpm;
514 struct ccw_request *req = &cdev->private->req;
515 struct ccw_dev_id *devid = &cdev->private->dev_id;
521 memset(cdev->private->dma_area->pgid, 0,
522 sizeof(cdev->private->dma_area->pgid));
523 cdev->private->pgid_valid_mask = 0;
524 cdev->private->pgid_todo_mask = sch->schib.pmcw.pam;
525 cdev->private->path_notoper_mask = 0;
533 if (cdev->private->flags.pgroup) {
560 CIO_HEX_EVENT(4, &cdev->private->dev_id, sizeof(cdev->private->dev_id));
565 cdev->private->flags.pgroup = cdev->private->options.pgroup;
566 cdev->private->flags.mpath = cdev->private->options.mpath;
567 cdev->private->flags.doverify = 0;
568 cdev->private->path_noirq_mask = 0;
578 struct ccw_dev_id *id = &cdev->private->dev_id;
583 cdev->private->flags.mpath = 0;
605 struct ccw_request *req = &cdev->private->req;
609 CIO_HEX_EVENT(4, &cdev->private->dev_id, sizeof(cdev->private->dev_id));
618 if (cdev->private->flags.mpath)
631 struct ccw_request *req = &cdev->private->req;
632 struct ccw1 *cp = cdev->private->dma_area->iccws;
666 struct ccw_request *req = &cdev->private->req;
669 CIO_HEX_EVENT(4, &cdev->private->dev_id, sizeof(cdev->private->dev_id));
693 if (!cdev->private->options.force)
706 cdev->private->state = DEV_STATE_STEAL_LOCK;
721 cdev->private->state = DEV_STATE_BOXED;