Lines Matching refs:cdev

63 	struct ccw_device *cdev = to_ccwdev(dev);  in ccw_bus_match()  local
70 found = ccw_device_id_match(ids, &cdev->id); in ccw_bus_match()
74 cdev->id.driver_info = found->driver_info; in ccw_bus_match()
106 const struct ccw_device *cdev = to_ccwdev(dev); in ccw_uevent() local
107 const struct ccw_device_id *id = &(cdev->id); in ccw_uevent()
200 struct ccw_device *cdev = to_ccwdev(dev); in devtype_show() local
201 struct ccw_device_id *id = &(cdev->id); in devtype_show()
213 struct ccw_device *cdev = to_ccwdev(dev); in cutype_show() local
214 struct ccw_device_id *id = &(cdev->id); in cutype_show()
223 struct ccw_device *cdev = to_ccwdev(dev); in modalias_show() local
224 struct ccw_device_id *id = &(cdev->id); in modalias_show()
235 struct ccw_device *cdev = to_ccwdev(dev); in online_show() local
237 return sprintf(buf, cdev->online ? "1\n" : "0\n"); in online_show()
240 int ccw_device_is_orphan(struct ccw_device *cdev) in ccw_device_is_orphan() argument
242 return sch_is_pseudo_sch(to_subchannel(cdev->dev.parent)); in ccw_device_is_orphan()
245 static void ccw_device_unregister(struct ccw_device *cdev) in ccw_device_unregister() argument
247 mutex_lock(&cdev->reg_mutex); in ccw_device_unregister()
248 if (device_is_registered(&cdev->dev)) { in ccw_device_unregister()
250 device_del(&cdev->dev); in ccw_device_unregister()
252 mutex_unlock(&cdev->reg_mutex); in ccw_device_unregister()
254 if (cdev->private->flags.initialized) { in ccw_device_unregister()
255 cdev->private->flags.initialized = 0; in ccw_device_unregister()
257 put_device(&cdev->dev); in ccw_device_unregister()
274 int ccw_device_set_offline(struct ccw_device *cdev) in ccw_device_set_offline() argument
279 if (!cdev) in ccw_device_set_offline()
281 if (!cdev->online || !cdev->drv) in ccw_device_set_offline()
284 if (cdev->drv->set_offline) { in ccw_device_set_offline()
285 ret = cdev->drv->set_offline(cdev); in ccw_device_set_offline()
289 spin_lock_irq(cdev->ccwlock); in ccw_device_set_offline()
290 sch = to_subchannel(cdev->dev.parent); in ccw_device_set_offline()
291 cdev->online = 0; in ccw_device_set_offline()
293 while (!dev_fsm_final_state(cdev) && in ccw_device_set_offline()
294 cdev->private->state != DEV_STATE_DISCONNECTED) { in ccw_device_set_offline()
295 spin_unlock_irq(cdev->ccwlock); in ccw_device_set_offline()
296 wait_event(cdev->private->wait_q, (dev_fsm_final_state(cdev) || in ccw_device_set_offline()
297 cdev->private->state == DEV_STATE_DISCONNECTED)); in ccw_device_set_offline()
298 spin_lock_irq(cdev->ccwlock); in ccw_device_set_offline()
301 ret = ccw_device_offline(cdev); in ccw_device_set_offline()
305 "0.%x.%04x\n", ret, cdev->private->dev_id.ssid, in ccw_device_set_offline()
306 cdev->private->dev_id.devno); in ccw_device_set_offline()
309 state = cdev->private->state; in ccw_device_set_offline()
310 spin_unlock_irq(cdev->ccwlock); in ccw_device_set_offline()
312 spin_lock_irq(cdev->ccwlock); in ccw_device_set_offline()
313 cdev->private->state = state; in ccw_device_set_offline()
315 spin_unlock_irq(cdev->ccwlock); in ccw_device_set_offline()
316 wait_event(cdev->private->wait_q, (dev_fsm_final_state(cdev) || in ccw_device_set_offline()
317 cdev->private->state == DEV_STATE_DISCONNECTED)); in ccw_device_set_offline()
319 if (cdev->private->state == DEV_STATE_BOXED) { in ccw_device_set_offline()
321 dev_name(&cdev->dev)); in ccw_device_set_offline()
322 } else if (cdev->private->state == DEV_STATE_NOT_OPER) { in ccw_device_set_offline()
324 dev_name(&cdev->dev)); in ccw_device_set_offline()
327 put_device(&cdev->dev); in ccw_device_set_offline()
331 cdev->private->state = DEV_STATE_OFFLINE; in ccw_device_set_offline()
332 dev_fsm_event(cdev, DEV_EVENT_NOTOPER); in ccw_device_set_offline()
333 spin_unlock_irq(cdev->ccwlock); in ccw_device_set_offline()
335 put_device(&cdev->dev); in ccw_device_set_offline()
351 int ccw_device_set_online(struct ccw_device *cdev) in ccw_device_set_online() argument
356 if (!cdev) in ccw_device_set_online()
358 if (cdev->online || !cdev->drv) in ccw_device_set_online()
361 if (!get_device(&cdev->dev)) in ccw_device_set_online()
364 spin_lock_irq(cdev->ccwlock); in ccw_device_set_online()
365 ret = ccw_device_online(cdev); in ccw_device_set_online()
366 spin_unlock_irq(cdev->ccwlock); in ccw_device_set_online()
368 wait_event(cdev->private->wait_q, dev_fsm_final_state(cdev)); in ccw_device_set_online()
372 ret, cdev->private->dev_id.ssid, in ccw_device_set_online()
373 cdev->private->dev_id.devno); in ccw_device_set_online()
375 put_device(&cdev->dev); in ccw_device_set_online()
378 spin_lock_irq(cdev->ccwlock); in ccw_device_set_online()
380 if ((cdev->private->state != DEV_STATE_ONLINE) && in ccw_device_set_online()
381 (cdev->private->state != DEV_STATE_W4SENSE)) { in ccw_device_set_online()
382 spin_unlock_irq(cdev->ccwlock); in ccw_device_set_online()
384 if (cdev->private->state == DEV_STATE_BOXED) { in ccw_device_set_online()
386 dev_name(&cdev->dev)); in ccw_device_set_online()
387 } else if (cdev->private->state == DEV_STATE_NOT_OPER) { in ccw_device_set_online()
389 dev_name(&cdev->dev)); in ccw_device_set_online()
392 put_device(&cdev->dev); in ccw_device_set_online()
395 spin_unlock_irq(cdev->ccwlock); in ccw_device_set_online()
396 if (cdev->drv->set_online) in ccw_device_set_online()
397 ret = cdev->drv->set_online(cdev); in ccw_device_set_online()
401 spin_lock_irq(cdev->ccwlock); in ccw_device_set_online()
402 cdev->online = 1; in ccw_device_set_online()
403 spin_unlock_irq(cdev->ccwlock); in ccw_device_set_online()
407 spin_lock_irq(cdev->ccwlock); in ccw_device_set_online()
409 while (!dev_fsm_final_state(cdev) && in ccw_device_set_online()
410 cdev->private->state != DEV_STATE_DISCONNECTED) { in ccw_device_set_online()
411 spin_unlock_irq(cdev->ccwlock); in ccw_device_set_online()
412 wait_event(cdev->private->wait_q, (dev_fsm_final_state(cdev) || in ccw_device_set_online()
413 cdev->private->state == DEV_STATE_DISCONNECTED)); in ccw_device_set_online()
414 spin_lock_irq(cdev->ccwlock); in ccw_device_set_online()
416 ret2 = ccw_device_offline(cdev); in ccw_device_set_online()
419 spin_unlock_irq(cdev->ccwlock); in ccw_device_set_online()
420 wait_event(cdev->private->wait_q, (dev_fsm_final_state(cdev) || in ccw_device_set_online()
421 cdev->private->state == DEV_STATE_DISCONNECTED)); in ccw_device_set_online()
423 put_device(&cdev->dev); in ccw_device_set_online()
429 ret2, cdev->private->dev_id.ssid, in ccw_device_set_online()
430 cdev->private->dev_id.devno); in ccw_device_set_online()
431 cdev->private->state = DEV_STATE_OFFLINE; in ccw_device_set_online()
432 spin_unlock_irq(cdev->ccwlock); in ccw_device_set_online()
434 put_device(&cdev->dev); in ccw_device_set_online()
438 static int online_store_handle_offline(struct ccw_device *cdev) in online_store_handle_offline() argument
440 if (cdev->private->state == DEV_STATE_DISCONNECTED) { in online_store_handle_offline()
441 spin_lock_irq(cdev->ccwlock); in online_store_handle_offline()
442 ccw_device_sched_todo(cdev, CDEV_TODO_UNREG_EVAL); in online_store_handle_offline()
443 spin_unlock_irq(cdev->ccwlock); in online_store_handle_offline()
446 if (cdev->drv && cdev->drv->set_offline) in online_store_handle_offline()
447 return ccw_device_set_offline(cdev); in online_store_handle_offline()
451 static int online_store_recog_and_online(struct ccw_device *cdev) in online_store_recog_and_online() argument
454 if (cdev->private->state == DEV_STATE_BOXED) { in online_store_recog_and_online()
455 spin_lock_irq(cdev->ccwlock); in online_store_recog_and_online()
456 ccw_device_recognition(cdev); in online_store_recog_and_online()
457 spin_unlock_irq(cdev->ccwlock); in online_store_recog_and_online()
458 wait_event(cdev->private->wait_q, in online_store_recog_and_online()
459 cdev->private->flags.recog_done); in online_store_recog_and_online()
460 if (cdev->private->state != DEV_STATE_OFFLINE) in online_store_recog_and_online()
464 if (cdev->drv && cdev->drv->set_online) in online_store_recog_and_online()
465 return ccw_device_set_online(cdev); in online_store_recog_and_online()
469 static int online_store_handle_online(struct ccw_device *cdev, int force) in online_store_handle_online() argument
473 ret = online_store_recog_and_online(cdev); in online_store_handle_online()
476 if (force && cdev->private->state == DEV_STATE_BOXED) { in online_store_handle_online()
477 ret = ccw_device_stlck(cdev); in online_store_handle_online()
480 if (cdev->id.cu_type == 0) in online_store_handle_online()
481 cdev->private->state = DEV_STATE_NOT_OPER; in online_store_handle_online()
482 ret = online_store_recog_and_online(cdev); in online_store_handle_online()
492 struct ccw_device *cdev = to_ccwdev(dev); in online_store() local
497 if (atomic_cmpxchg(&cdev->private->onoff, 0, 1) != 0) in online_store()
500 if (!dev_fsm_final_state(cdev) && in online_store()
501 cdev->private->state != DEV_STATE_DISCONNECTED) { in online_store()
506 if (work_pending(&cdev->private->todo_work)) { in online_store()
524 ret = online_store_handle_offline(cdev); in online_store()
527 ret = online_store_handle_online(cdev, force); in online_store()
535 atomic_set(&cdev->private->onoff, 0); in online_store()
542 struct ccw_device *cdev = to_ccwdev(dev); in available_show() local
545 if (ccw_device_is_orphan(cdev)) in available_show()
547 switch (cdev->private->state) { in available_show()
629 struct ccw_device *cdev = to_ccwdev(dev); in match_dev_id() local
632 return ccw_dev_id_is_equal(&cdev->private->dev_id, dev_id); in match_dev_id()
655 static void ccw_device_do_unbind_bind(struct ccw_device *cdev) in ccw_device_do_unbind_bind() argument
659 mutex_lock(&cdev->reg_mutex); in ccw_device_do_unbind_bind()
660 if (device_is_registered(&cdev->dev)) { in ccw_device_do_unbind_bind()
661 device_release_driver(&cdev->dev); in ccw_device_do_unbind_bind()
662 ret = device_attach(&cdev->dev); in ccw_device_do_unbind_bind()
665 mutex_unlock(&cdev->reg_mutex); in ccw_device_do_unbind_bind()
671 struct ccw_device *cdev; in ccw_device_release() local
673 cdev = to_ccwdev(dev); in ccw_device_release()
674 cio_gp_dma_free(cdev->private->dma_pool, cdev->private->dma_area, in ccw_device_release()
675 sizeof(*cdev->private->dma_area)); in ccw_device_release()
676 cio_gp_dma_destroy(cdev->private->dma_pool, &cdev->dev); in ccw_device_release()
678 put_device(cdev->dev.parent); in ccw_device_release()
679 kfree(cdev->private); in ccw_device_release()
680 kfree(cdev); in ccw_device_release()
685 struct ccw_device *cdev; in io_subchannel_allocate_dev() local
689 cdev = kzalloc(sizeof(*cdev), GFP_KERNEL); in io_subchannel_allocate_dev()
690 if (!cdev) { in io_subchannel_allocate_dev()
694 cdev->private = kzalloc(sizeof(struct ccw_device_private), in io_subchannel_allocate_dev()
696 if (!cdev->private) { in io_subchannel_allocate_dev()
701 cdev->dev.dma_mask = sch->dev.dma_mask; in io_subchannel_allocate_dev()
702 ret = dma_set_coherent_mask(&cdev->dev, sch->dev.coherent_dma_mask); in io_subchannel_allocate_dev()
706 dma_pool = cio_gp_dma_create(&cdev->dev, 1); in io_subchannel_allocate_dev()
711 cdev->private->dma_pool = dma_pool; in io_subchannel_allocate_dev()
712 cdev->private->dma_area = cio_gp_dma_zalloc(dma_pool, &cdev->dev, in io_subchannel_allocate_dev()
713 sizeof(*cdev->private->dma_area)); in io_subchannel_allocate_dev()
714 if (!cdev->private->dma_area) { in io_subchannel_allocate_dev()
718 return cdev; in io_subchannel_allocate_dev()
720 cio_gp_dma_destroy(dma_pool, &cdev->dev); in io_subchannel_allocate_dev()
723 kfree(cdev->private); in io_subchannel_allocate_dev()
725 kfree(cdev); in io_subchannel_allocate_dev()
733 struct ccw_device *cdev) in io_subchannel_initialize_dev() argument
735 struct ccw_device_private *priv = cdev->private; in io_subchannel_initialize_dev()
738 priv->cdev = cdev; in io_subchannel_initialize_dev()
748 mutex_init(&cdev->reg_mutex); in io_subchannel_initialize_dev()
751 cdev->ccwlock = sch->lock; in io_subchannel_initialize_dev()
752 cdev->dev.parent = &sch->dev; in io_subchannel_initialize_dev()
753 cdev->dev.release = ccw_device_release; in io_subchannel_initialize_dev()
754 cdev->dev.bus = &ccw_bus_type; in io_subchannel_initialize_dev()
755 cdev->dev.groups = ccwdev_attr_groups; in io_subchannel_initialize_dev()
757 device_initialize(&cdev->dev); in io_subchannel_initialize_dev()
758 ret = dev_set_name(&cdev->dev, "0.%x.%04x", cdev->private->dev_id.ssid, in io_subchannel_initialize_dev()
759 cdev->private->dev_id.devno); in io_subchannel_initialize_dev()
768 sch_set_cdev(sch, cdev); in io_subchannel_initialize_dev()
774 put_device(&cdev->dev); in io_subchannel_initialize_dev()
780 struct ccw_device *cdev; in io_subchannel_create_ccwdev() local
783 cdev = io_subchannel_allocate_dev(sch); in io_subchannel_create_ccwdev()
784 if (!IS_ERR(cdev)) { in io_subchannel_create_ccwdev()
785 ret = io_subchannel_initialize_dev(sch, cdev); in io_subchannel_create_ccwdev()
787 cdev = ERR_PTR(ret); in io_subchannel_create_ccwdev()
789 return cdev; in io_subchannel_create_ccwdev()
796 struct ccw_device *cdev; in sch_create_and_recog_new_device() local
799 cdev = io_subchannel_create_ccwdev(sch); in sch_create_and_recog_new_device()
800 if (IS_ERR(cdev)) { in sch_create_and_recog_new_device()
806 io_subchannel_recog(cdev, sch); in sch_create_and_recog_new_device()
812 static void io_subchannel_register(struct ccw_device *cdev) in io_subchannel_register() argument
818 sch = to_subchannel(cdev->dev.parent); in io_subchannel_register()
834 mutex_lock(&cdev->reg_mutex); in io_subchannel_register()
835 if (device_is_registered(&cdev->dev)) { in io_subchannel_register()
836 if (!cdev->drv) { in io_subchannel_register()
837 ret = device_reprobe(&cdev->dev); in io_subchannel_register()
842 cdev->private->dev_id.ssid, in io_subchannel_register()
843 cdev->private->dev_id.devno); in io_subchannel_register()
849 ret = device_add(&cdev->dev); in io_subchannel_register()
852 cdev->private->dev_id.ssid, in io_subchannel_register()
853 cdev->private->dev_id.devno, ret); in io_subchannel_register()
857 mutex_unlock(&cdev->reg_mutex); in io_subchannel_register()
859 put_device(&cdev->dev); in io_subchannel_register()
863 cdev->private->flags.recog_done = 1; in io_subchannel_register()
864 mutex_unlock(&cdev->reg_mutex); in io_subchannel_register()
865 wake_up(&cdev->private->wait_q); in io_subchannel_register()
875 io_subchannel_recog_done(struct ccw_device *cdev) in io_subchannel_recog_done() argument
878 cdev->private->flags.recog_done = 1; in io_subchannel_recog_done()
881 switch (cdev->private->state) { in io_subchannel_recog_done()
885 cdev->private->flags.recog_done = 1; in io_subchannel_recog_done()
887 ccw_device_sched_todo(cdev, CDEV_TODO_UNREG); in io_subchannel_recog_done()
896 ccw_device_sched_todo(cdev, CDEV_TODO_REGISTER); in io_subchannel_recog_done()
901 static void io_subchannel_recog(struct ccw_device *cdev, struct subchannel *sch) in io_subchannel_recog() argument
908 ccw_device_recognition(cdev); in io_subchannel_recog()
912 static int ccw_device_move_to_sch(struct ccw_device *cdev, in ccw_device_move_to_sch() argument
918 old_sch = to_subchannel(cdev->dev.parent); in ccw_device_move_to_sch()
938 rc = device_move(&cdev->dev, &sch->dev, DPM_ORDER_PARENT_BEFORE_DEV); in ccw_device_move_to_sch()
942 cdev->private->dev_id.ssid, in ccw_device_move_to_sch()
943 cdev->private->dev_id.devno, sch->schid.ssid, in ccw_device_move_to_sch()
966 cdev->ccwlock = sch->lock; in ccw_device_move_to_sch()
968 sch_set_cdev(sch, cdev); in ccw_device_move_to_sch()
975 static int ccw_device_move_to_orph(struct ccw_device *cdev) in ccw_device_move_to_orph() argument
977 struct subchannel *sch = to_subchannel(cdev->dev.parent); in ccw_device_move_to_orph()
980 return ccw_device_move_to_sch(cdev, css->pseudo_subchannel); in ccw_device_move_to_orph()
985 struct ccw_device *cdev; in io_subchannel_irq() local
987 cdev = sch_get_cdev(sch); in io_subchannel_irq()
991 if (cdev) in io_subchannel_irq()
992 dev_fsm_event(cdev, DEV_EVENT_INTERRUPT); in io_subchannel_irq()
1028 struct ccw_device *cdev; in io_subchannel_probe() local
1043 cdev = sch_get_cdev(sch); in io_subchannel_probe()
1044 rc = device_add(&cdev->dev); in io_subchannel_probe()
1047 put_device(&cdev->dev); in io_subchannel_probe()
1089 struct ccw_device *cdev; in io_subchannel_remove() local
1091 cdev = sch_get_cdev(sch); in io_subchannel_remove()
1092 if (!cdev) in io_subchannel_remove()
1095 ccw_device_unregister(cdev); in io_subchannel_remove()
1109 struct ccw_device *cdev; in io_subchannel_verify() local
1111 cdev = sch_get_cdev(sch); in io_subchannel_verify()
1112 if (cdev) in io_subchannel_verify()
1113 dev_fsm_event(cdev, DEV_EVENT_VERIFY); in io_subchannel_verify()
1120 struct ccw_device *cdev; in io_subchannel_terminate_path() local
1122 cdev = sch_get_cdev(sch); in io_subchannel_terminate_path()
1123 if (!cdev) in io_subchannel_terminate_path()
1130 if (cdev->private->state == DEV_STATE_ONLINE) { in io_subchannel_terminate_path()
1131 ccw_device_kill_io(cdev); in io_subchannel_terminate_path()
1138 dev_fsm_event(cdev, DEV_EVENT_VERIFY); in io_subchannel_terminate_path()
1142 dev_fsm_event(cdev, DEV_EVENT_NOTOPER); in io_subchannel_terminate_path()
1148 struct ccw_device *cdev = sch_get_cdev(sch); in io_subchannel_chp_event() local
1159 if (cdev) in io_subchannel_chp_event()
1160 cdev->private->path_gone_mask |= mask; in io_subchannel_chp_event()
1166 if (cdev) in io_subchannel_chp_event()
1167 cdev->private->path_new_mask |= mask; in io_subchannel_chp_event()
1173 if (cdev) in io_subchannel_chp_event()
1174 cdev->private->path_gone_mask |= mask; in io_subchannel_chp_event()
1181 if (cdev) in io_subchannel_chp_event()
1182 cdev->private->path_new_mask |= mask; in io_subchannel_chp_event()
1194 if (cdev && cdev->drv && cdev->drv->path_event) in io_subchannel_chp_event()
1195 cdev->drv->path_event(cdev, path_event); in io_subchannel_chp_event()
1203 struct ccw_device *cdev; in io_subchannel_quiesce() local
1207 cdev = sch_get_cdev(sch); in io_subchannel_quiesce()
1215 if (cdev->handler) in io_subchannel_quiesce()
1216 cdev->handler(cdev, cdev->private->intparm, ERR_PTR(-EIO)); in io_subchannel_quiesce()
1218 cdev->private->state = DEV_STATE_QUIESCE; in io_subchannel_quiesce()
1219 cdev->private->iretry = 255; in io_subchannel_quiesce()
1220 ret = ccw_device_cancel_halt_clear(cdev); in io_subchannel_quiesce()
1222 ccw_device_set_timeout(cdev, HZ/10); in io_subchannel_quiesce()
1224 wait_event(cdev->private->wait_q, in io_subchannel_quiesce()
1225 cdev->private->state != DEV_STATE_QUIESCE); in io_subchannel_quiesce()
1239 static int device_is_disconnected(struct ccw_device *cdev) in device_is_disconnected() argument
1241 if (!cdev) in device_is_disconnected()
1243 return (cdev->private->state == DEV_STATE_DISCONNECTED || in device_is_disconnected()
1244 cdev->private->state == DEV_STATE_DISCONNECTED_SENSE_ID); in device_is_disconnected()
1249 struct ccw_device *cdev = to_ccwdev(dev); in recovery_check() local
1253 spin_lock_irq(cdev->ccwlock); in recovery_check()
1254 switch (cdev->private->state) { in recovery_check()
1256 sch = to_subchannel(cdev->dev.parent); in recovery_check()
1262 cdev->private->dev_id.ssid, in recovery_check()
1263 cdev->private->dev_id.devno); in recovery_check()
1264 dev_fsm_event(cdev, DEV_EVENT_VERIFY); in recovery_check()
1271 spin_unlock_irq(cdev->ccwlock); in recovery_check()
1320 struct ccw_device *cdev = to_ccwdev(dev); in purge_fn() local
1321 struct ccw_dev_id *id = &cdev->private->dev_id; in purge_fn()
1322 struct subchannel *sch = to_subchannel(cdev->dev.parent); in purge_fn()
1324 spin_lock_irq(cdev->ccwlock); in purge_fn()
1326 (cdev->private->state == DEV_STATE_OFFLINE) && in purge_fn()
1327 (atomic_cmpxchg(&cdev->private->onoff, 0, 1) == 0)) { in purge_fn()
1330 ccw_device_sched_todo(cdev, CDEV_TODO_UNREG); in purge_fn()
1332 atomic_set(&cdev->private->onoff, 0); in purge_fn()
1334 spin_unlock_irq(cdev->ccwlock); in purge_fn()
1354 void ccw_device_set_disconnected(struct ccw_device *cdev) in ccw_device_set_disconnected() argument
1356 if (!cdev) in ccw_device_set_disconnected()
1358 ccw_device_set_timeout(cdev, 0); in ccw_device_set_disconnected()
1359 cdev->private->flags.fake_irb = 0; in ccw_device_set_disconnected()
1360 cdev->private->state = DEV_STATE_DISCONNECTED; in ccw_device_set_disconnected()
1361 if (cdev->online) in ccw_device_set_disconnected()
1365 void ccw_device_set_notoper(struct ccw_device *cdev) in ccw_device_set_notoper() argument
1367 struct subchannel *sch = to_subchannel(cdev->dev.parent); in ccw_device_set_notoper()
1371 ccw_device_set_timeout(cdev, 0); in ccw_device_set_notoper()
1373 cdev->private->state = DEV_STATE_NOT_OPER; in ccw_device_set_notoper()
1391 struct ccw_device *cdev; in sch_get_action() local
1393 cdev = sch_get_cdev(sch); in sch_get_action()
1396 if (!cdev) in sch_get_action()
1398 if (ccw_device_notify(cdev, CIO_GONE) != NOTIFY_OK) in sch_get_action()
1403 if (!cdev) in sch_get_action()
1405 if (sch->schib.pmcw.dev != cdev->private->dev_id.devno) { in sch_get_action()
1406 if (ccw_device_notify(cdev, CIO_GONE) != NOTIFY_OK) in sch_get_action()
1411 if (ccw_device_notify(cdev, CIO_NO_PATH) != NOTIFY_OK) in sch_get_action()
1415 if (device_is_disconnected(cdev)) in sch_get_action()
1417 if (cdev->online) in sch_get_action()
1419 if (cdev->private->state == DEV_STATE_NOT_OPER) in sch_get_action()
1437 struct ccw_device *cdev; in io_subchannel_sch_event() local
1447 cdev = sch_get_cdev(sch); in io_subchannel_sch_event()
1448 if (cdev && work_pending(&cdev->private->todo_work)) in io_subchannel_sch_event()
1458 ccw_device_trigger_reprobe(cdev); in io_subchannel_sch_event()
1467 ccw_device_set_disconnected(cdev); in io_subchannel_sch_event()
1472 ccw_device_set_disconnected(cdev); in io_subchannel_sch_event()
1477 if (!cdev) in io_subchannel_sch_event()
1479 if (cdev->private->state == DEV_STATE_SENSE_ID) { in io_subchannel_sch_event()
1485 dev_fsm_event(cdev, DEV_EVENT_NOTOPER); in io_subchannel_sch_event()
1487 ccw_device_set_notoper(cdev); in io_subchannel_sch_event()
1504 rc = ccw_device_move_to_orph(cdev); in io_subchannel_sch_event()
1514 ccw_device_unregister(cdev); in io_subchannel_sch_event()
1530 cdev = get_ccwdev_by_dev_id(&dev_id); in io_subchannel_sch_event()
1531 if (!cdev) { in io_subchannel_sch_event()
1535 rc = ccw_device_move_to_sch(cdev, sch); in io_subchannel_sch_event()
1538 put_device(&cdev->dev); in io_subchannel_sch_event()
1542 ccw_device_trigger_reprobe(cdev); in io_subchannel_sch_event()
1545 put_device(&cdev->dev); in io_subchannel_sch_event()
1558 static void ccw_device_set_int_class(struct ccw_device *cdev) in ccw_device_set_int_class() argument
1560 struct ccw_driver *cdrv = cdev->drv; in ccw_device_set_int_class()
1565 cdev->private->int_class = cdrv->int_class; in ccw_device_set_int_class()
1567 cdev->private->int_class = IRQIO_CIO; in ccw_device_set_int_class()
1571 int __init ccw_device_enable_console(struct ccw_device *cdev) in ccw_device_enable_console() argument
1573 struct subchannel *sch = to_subchannel(cdev->dev.parent); in ccw_device_enable_console()
1576 if (!cdev->drv || !cdev->handler) in ccw_device_enable_console()
1584 io_subchannel_recog(cdev, sch); in ccw_device_enable_console()
1586 spin_lock_irq(cdev->ccwlock); in ccw_device_enable_console()
1587 while (!dev_fsm_final_state(cdev)) in ccw_device_enable_console()
1588 ccw_device_wait_idle(cdev); in ccw_device_enable_console()
1591 get_device(&cdev->dev); in ccw_device_enable_console()
1592 rc = ccw_device_online(cdev); in ccw_device_enable_console()
1596 while (!dev_fsm_final_state(cdev)) in ccw_device_enable_console()
1597 ccw_device_wait_idle(cdev); in ccw_device_enable_console()
1599 if (cdev->private->state == DEV_STATE_ONLINE) in ccw_device_enable_console()
1600 cdev->online = 1; in ccw_device_enable_console()
1604 spin_unlock_irq(cdev->ccwlock); in ccw_device_enable_console()
1606 put_device(&cdev->dev); in ccw_device_enable_console()
1613 struct ccw_device *cdev; in ccw_device_create_console() local
1629 cdev = io_subchannel_create_ccwdev(sch); in ccw_device_create_console()
1630 if (IS_ERR(cdev)) { in ccw_device_create_console()
1636 return cdev; in ccw_device_create_console()
1638 cdev->drv = drv; in ccw_device_create_console()
1639 ccw_device_set_int_class(cdev); in ccw_device_create_console()
1640 return cdev; in ccw_device_create_console()
1649 void __init ccw_device_destroy_console(struct ccw_device *cdev) in ccw_device_destroy_console() argument
1651 struct subchannel *sch = to_subchannel(cdev->dev.parent); in ccw_device_destroy_console()
1658 put_device(&cdev->dev); in ccw_device_destroy_console()
1670 void ccw_device_wait_idle(struct ccw_device *cdev) in ccw_device_wait_idle() argument
1672 struct subchannel *sch = to_subchannel(cdev->dev.parent); in ccw_device_wait_idle()
1717 struct ccw_device *cdev = to_ccwdev(dev); in ccw_device_probe() local
1721 cdev->drv = cdrv; /* to let the driver call _set_online */ in ccw_device_probe()
1722 ccw_device_set_int_class(cdev); in ccw_device_probe()
1723 ret = cdrv->probe ? cdrv->probe(cdev) : -ENODEV; in ccw_device_probe()
1725 cdev->drv = NULL; in ccw_device_probe()
1726 cdev->private->int_class = IRQIO_CIO; in ccw_device_probe()
1735 struct ccw_device *cdev = to_ccwdev(dev); in ccw_device_remove() local
1736 struct ccw_driver *cdrv = cdev->drv; in ccw_device_remove()
1741 cdrv->remove(cdev); in ccw_device_remove()
1743 spin_lock_irq(cdev->ccwlock); in ccw_device_remove()
1744 if (cdev->online) { in ccw_device_remove()
1745 cdev->online = 0; in ccw_device_remove()
1746 ret = ccw_device_offline(cdev); in ccw_device_remove()
1747 spin_unlock_irq(cdev->ccwlock); in ccw_device_remove()
1749 wait_event(cdev->private->wait_q, in ccw_device_remove()
1750 dev_fsm_final_state(cdev)); in ccw_device_remove()
1754 ret, cdev->private->dev_id.ssid, in ccw_device_remove()
1755 cdev->private->dev_id.devno); in ccw_device_remove()
1757 put_device(&cdev->dev); in ccw_device_remove()
1758 spin_lock_irq(cdev->ccwlock); in ccw_device_remove()
1760 ccw_device_set_timeout(cdev, 0); in ccw_device_remove()
1761 cdev->drv = NULL; in ccw_device_remove()
1762 cdev->private->int_class = IRQIO_CIO; in ccw_device_remove()
1763 sch = to_subchannel(cdev->dev.parent); in ccw_device_remove()
1764 spin_unlock_irq(cdev->ccwlock); in ccw_device_remove()
1766 __disable_cmf(cdev); in ccw_device_remove()
1771 struct ccw_device *cdev; in ccw_device_shutdown() local
1773 cdev = to_ccwdev(dev); in ccw_device_shutdown()
1774 if (cdev->drv && cdev->drv->shutdown) in ccw_device_shutdown()
1775 cdev->drv->shutdown(cdev); in ccw_device_shutdown()
1776 __disable_cmf(cdev); in ccw_device_shutdown()
1819 struct ccw_device *cdev; in ccw_device_todo() local
1824 cdev = priv->cdev; in ccw_device_todo()
1825 sch = to_subchannel(cdev->dev.parent); in ccw_device_todo()
1827 spin_lock_irq(cdev->ccwlock); in ccw_device_todo()
1832 spin_unlock_irq(cdev->ccwlock); in ccw_device_todo()
1836 cmf_reenable(cdev); in ccw_device_todo()
1839 ccw_device_do_unbind_bind(cdev); in ccw_device_todo()
1842 io_subchannel_register(cdev); in ccw_device_todo()
1852 ccw_device_unregister(cdev); in ccw_device_todo()
1858 put_device(&cdev->dev); in ccw_device_todo()
1870 void ccw_device_sched_todo(struct ccw_device *cdev, enum cdev_todo todo) in ccw_device_sched_todo() argument
1873 cdev->private->dev_id.ssid, cdev->private->dev_id.devno, in ccw_device_sched_todo()
1875 if (cdev->private->todo >= todo) in ccw_device_sched_todo()
1877 cdev->private->todo = todo; in ccw_device_sched_todo()
1879 if (!get_device(&cdev->dev)) in ccw_device_sched_todo()
1881 if (!queue_work(cio_work_q, &cdev->private->todo_work)) { in ccw_device_sched_todo()
1883 put_device(&cdev->dev); in ccw_device_sched_todo()
1894 int ccw_device_siosl(struct ccw_device *cdev) in ccw_device_siosl() argument
1896 struct subchannel *sch = to_subchannel(cdev->dev.parent); in ccw_device_siosl()