Lines Matching refs:private

37 	struct ccw_dev_id *id = &cdev->private->dev_id;  in verify_done()
38 int mpath = cdev->private->flags.mpath; in verify_done()
39 int pgroup = cdev->private->flags.pgroup; in verify_done()
60 struct ccw_request *req = &cdev->private->req; in nop_build_cp()
61 struct ccw1 *cp = cdev->private->dma_area->iccws; in nop_build_cp()
76 struct ccw_request *req = &cdev->private->req; in nop_do()
79 ~cdev->private->path_noirq_mask); in nop_do()
108 struct ccw_request *req = &cdev->private->req; in nop_callback()
115 cdev->private->path_noirq_mask |= req->lpm; in nop_callback()
118 cdev->private->path_notoper_mask |= req->lpm; in nop_callback()
137 struct ccw_request *req = &cdev->private->req; in spid_build_cp()
138 struct ccw1 *cp = cdev->private->dma_area->iccws; in spid_build_cp()
140 struct pgid *pgid = &cdev->private->dma_area->pgid[i]; in spid_build_cp()
161 cdev->private->flags.pgid_unknown = 0; in pgid_wipeout_callback()
171 struct ccw_dev_id *id = &cdev->private->dev_id; in pgid_wipeout_start()
172 struct ccw_request *req = &cdev->private->req; in pgid_wipeout_start()
176 id->ssid, id->devno, cdev->private->pgid_valid_mask, in pgid_wipeout_start()
177 cdev->private->path_noirq_mask); in pgid_wipeout_start()
186 if (cdev->private->flags.mpath) in pgid_wipeout_start()
198 struct ccw_request *req = &cdev->private->req; in spid_do()
202 req->lpm = lpm_adjust(req->lpm, cdev->private->pgid_todo_mask); in spid_do()
210 if (cdev->private->flags.mpath) in spid_do()
217 if (cdev->private->flags.pgid_unknown) { in spid_do()
231 struct ccw_request *req = &cdev->private->req; in spid_callback()
238 cdev->private->flags.pgid_unknown = 1; in spid_callback()
239 cdev->private->path_noirq_mask |= req->lpm; in spid_callback()
242 cdev->private->path_notoper_mask |= req->lpm; in spid_callback()
245 if (cdev->private->flags.mpath) { in spid_callback()
247 cdev->private->flags.mpath = 0; in spid_callback()
251 cdev->private->flags.pgroup = 0; in spid_callback()
269 struct ccw_request *req = &cdev->private->req; in spid_start()
304 struct pgid *pgid = &cdev->private->dma_area->pgid[0]; in pgid_analyze()
313 if ((cdev->private->pgid_valid_mask & lpm) == 0) in pgid_analyze()
344 if ((cdev->private->pgid_valid_mask & lpm) == 0) in pgid_to_donepm()
346 pgid = &cdev->private->dma_area->pgid[i]; in pgid_to_donepm()
354 if (cdev->private->flags.mpath) { in pgid_to_donepm()
372 memcpy(&cdev->private->dma_area->pgid[i], pgid, in pgid_fill()
381 struct ccw_dev_id *id = &cdev->private->dev_id; in snid_done()
392 if (reserved == cdev->private->pgid_valid_mask) in snid_done()
399 cdev->private->pgid_reset_mask |= reset; in snid_done()
400 cdev->private->pgid_todo_mask &= in snid_done()
401 ~(donepm | cdev->private->path_noirq_mask); in snid_done()
407 id->devno, rc, cdev->private->pgid_valid_mask, sch->vpm, in snid_done()
408 cdev->private->pgid_todo_mask, mismatch, reserved, reset); in snid_done()
411 if (cdev->private->flags.pgid_unknown) { in snid_done()
416 if (cdev->private->pgid_todo_mask == 0) { in snid_done()
425 cdev->private->flags.pgroup = 0; in snid_done()
426 cdev->private->flags.mpath = 0; in snid_done()
439 struct ccw_request *req = &cdev->private->req; in snid_build_cp()
440 struct ccw1 *cp = cdev->private->dma_area->iccws; in snid_build_cp()
445 cp->cda = (u32)virt_to_phys(&cdev->private->dma_area->pgid[i]); in snid_build_cp()
457 struct ccw_request *req = &cdev->private->req; in snid_do()
461 ~cdev->private->path_noirq_mask); in snid_do()
469 if (cdev->private->pgid_valid_mask) in snid_do()
471 else if (cdev->private->path_noirq_mask) in snid_do()
483 struct ccw_request *req = &cdev->private->req; in snid_callback()
487 cdev->private->pgid_valid_mask |= req->lpm; in snid_callback()
490 cdev->private->flags.pgid_unknown = 1; in snid_callback()
491 cdev->private->path_noirq_mask |= req->lpm; in snid_callback()
494 cdev->private->path_notoper_mask |= req->lpm; in snid_callback()
514 struct ccw_request *req = &cdev->private->req; in verify_start()
515 struct ccw_dev_id *devid = &cdev->private->dev_id; in verify_start()
521 memset(cdev->private->dma_area->pgid, 0, in verify_start()
522 sizeof(cdev->private->dma_area->pgid)); in verify_start()
523 cdev->private->pgid_valid_mask = 0; in verify_start()
524 cdev->private->pgid_todo_mask = sch->schib.pmcw.pam; in verify_start()
525 cdev->private->path_notoper_mask = 0; in verify_start()
533 if (cdev->private->flags.pgroup) { in verify_start()
560 CIO_HEX_EVENT(4, &cdev->private->dev_id, sizeof(cdev->private->dev_id)); in ccw_device_verify_start()
565 cdev->private->flags.pgroup = cdev->private->options.pgroup; in ccw_device_verify_start()
566 cdev->private->flags.mpath = cdev->private->options.mpath; in ccw_device_verify_start()
567 cdev->private->flags.doverify = 0; in ccw_device_verify_start()
568 cdev->private->path_noirq_mask = 0; in ccw_device_verify_start()
578 struct ccw_dev_id *id = &cdev->private->dev_id; in disband_callback()
583 cdev->private->flags.mpath = 0; in disband_callback()
605 struct ccw_request *req = &cdev->private->req; in ccw_device_disband_start()
609 CIO_HEX_EVENT(4, &cdev->private->dev_id, sizeof(cdev->private->dev_id)); in ccw_device_disband_start()
618 if (cdev->private->flags.mpath) in ccw_device_disband_start()
631 struct ccw_request *req = &cdev->private->req; in stlck_build_cp()
632 struct ccw1 *cp = cdev->private->dma_area->iccws; in stlck_build_cp()
666 struct ccw_request *req = &cdev->private->req; in ccw_device_stlck_start()
669 CIO_HEX_EVENT(4, &cdev->private->dev_id, sizeof(cdev->private->dev_id)); in ccw_device_stlck_start()
693 if (!cdev->private->options.force) in ccw_device_stlck()
706 cdev->private->state = DEV_STATE_STEAL_LOCK; in ccw_device_stlck()
721 cdev->private->state = DEV_STATE_BOXED; in ccw_device_stlck()