Lines Matching refs:shost

66 void scsi_eh_wakeup(struct Scsi_Host *shost)  in scsi_eh_wakeup()  argument
68 lockdep_assert_held(shost->host_lock); in scsi_eh_wakeup()
70 if (scsi_host_busy(shost) == shost->host_failed) { in scsi_eh_wakeup()
71 trace_scsi_eh_wakeup(shost); in scsi_eh_wakeup()
72 wake_up_process(shost->ehandler); in scsi_eh_wakeup()
73 SCSI_LOG_ERROR_RECOVERY(5, shost_printk(KERN_INFO, shost, in scsi_eh_wakeup()
84 void scsi_schedule_eh(struct Scsi_Host *shost) in scsi_schedule_eh() argument
88 spin_lock_irqsave(shost->host_lock, flags); in scsi_schedule_eh()
90 if (scsi_host_set_state(shost, SHOST_RECOVERY) == 0 || in scsi_schedule_eh()
91 scsi_host_set_state(shost, SHOST_CANCEL_RECOVERY) == 0) { in scsi_schedule_eh()
92 shost->host_eh_scheduled++; in scsi_schedule_eh()
93 scsi_eh_wakeup(shost); in scsi_schedule_eh()
96 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_schedule_eh()
100 static int scsi_host_eh_past_deadline(struct Scsi_Host *shost) in scsi_host_eh_past_deadline() argument
102 if (!shost->last_reset || shost->eh_deadline == -1) in scsi_host_eh_past_deadline()
112 if (time_before(jiffies, shost->last_reset + shost->eh_deadline) && in scsi_host_eh_past_deadline()
113 shost->eh_deadline > -1) in scsi_host_eh_past_deadline()
189 struct Scsi_Host *shost = sdev->host; in scsi_abort_command() local
203 spin_lock_irqsave(shost->host_lock, flags); in scsi_abort_command()
204 if (shost->eh_deadline != -1 && !shost->last_reset) in scsi_abort_command()
205 shost->last_reset = jiffies; in scsi_abort_command()
206 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_abort_command()
211 queue_delayed_work(shost->tmf_work_q, &scmd->abort_work, HZ / 100); in scsi_abort_command()
235 struct Scsi_Host *shost = scmd->device->host; in scsi_eh_inc_host_failed() local
238 spin_lock_irqsave(shost->host_lock, flags); in scsi_eh_inc_host_failed()
239 shost->host_failed++; in scsi_eh_inc_host_failed()
240 scsi_eh_wakeup(shost); in scsi_eh_inc_host_failed()
241 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_eh_inc_host_failed()
250 struct Scsi_Host *shost = scmd->device->host; in scsi_eh_scmd_add() local
254 WARN_ON_ONCE(!shost->ehandler); in scsi_eh_scmd_add()
256 spin_lock_irqsave(shost->host_lock, flags); in scsi_eh_scmd_add()
257 if (scsi_host_set_state(shost, SHOST_RECOVERY)) { in scsi_eh_scmd_add()
258 ret = scsi_host_set_state(shost, SHOST_CANCEL_RECOVERY); in scsi_eh_scmd_add()
261 if (shost->eh_deadline != -1 && !shost->last_reset) in scsi_eh_scmd_add()
262 shost->last_reset = jiffies; in scsi_eh_scmd_add()
265 list_add_tail(&scmd->eh_entry, &shost->eh_cmd_q); in scsi_eh_scmd_add()
266 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_eh_scmd_add()
352 static inline void scsi_eh_prt_fail_stats(struct Scsi_Host *shost, in scsi_eh_prt_fail_stats() argument
362 shost_for_each_device(sdev, shost) { in scsi_eh_prt_fail_stats()
375 shost_printk(KERN_INFO, shost, in scsi_eh_prt_fail_stats()
385 SCSI_LOG_ERROR_RECOVERY(2, shost_printk(KERN_INFO, shost, in scsi_eh_prt_fail_stats()
1059 struct Scsi_Host *shost = sdev->host; in scsi_send_eh_cmnd() local
1068 shost->eh_action = &done; in scsi_send_eh_cmnd()
1090 rtn = shost->hostt->queuecommand(shost, scmd); in scsi_send_eh_cmnd()
1110 shost->eh_action = NULL; in scsi_send_eh_cmnd()
1220 struct Scsi_Host *shost; in scsi_eh_get_sense() local
1232 shost = scmd->device->host; in scsi_eh_get_sense()
1233 if (scsi_host_eh_past_deadline(shost)) { in scsi_eh_get_sense()
1402 static int scsi_eh_stu(struct Scsi_Host *shost, in scsi_eh_stu() argument
1409 shost_for_each_device(sdev, shost) { in scsi_eh_stu()
1410 if (scsi_host_eh_past_deadline(shost)) { in scsi_eh_stu()
1467 static int scsi_eh_bus_device_reset(struct Scsi_Host *shost, in scsi_eh_bus_device_reset() argument
1475 shost_for_each_device(sdev, shost) { in scsi_eh_bus_device_reset()
1476 if (scsi_host_eh_past_deadline(shost)) { in scsi_eh_bus_device_reset()
1528 static int scsi_eh_target_reset(struct Scsi_Host *shost, in scsi_eh_target_reset() argument
1542 if (scsi_host_eh_past_deadline(shost)) { in scsi_eh_target_reset()
1547 shost_printk(KERN_INFO, shost, in scsi_eh_target_reset()
1557 shost_printk(KERN_INFO, shost, in scsi_eh_target_reset()
1563 shost_printk(KERN_INFO, shost, in scsi_eh_target_reset()
1590 static int scsi_eh_bus_reset(struct Scsi_Host *shost, in scsi_eh_bus_reset() argument
1606 for (channel = 0; channel <= shost->max_channel; channel++) { in scsi_eh_bus_reset()
1607 if (scsi_host_eh_past_deadline(shost)) { in scsi_eh_bus_reset()
1610 shost_printk(KERN_INFO, shost, in scsi_eh_bus_reset()
1631 shost_printk(KERN_INFO, shost, in scsi_eh_bus_reset()
1648 shost_printk(KERN_INFO, shost, in scsi_eh_bus_reset()
1662 static int scsi_eh_host_reset(struct Scsi_Host *shost, in scsi_eh_host_reset() argument
1675 shost_printk(KERN_INFO, shost, in scsi_eh_host_reset()
1688 shost_printk(KERN_INFO, shost, in scsi_eh_host_reset()
2005 static void scsi_restart_operations(struct Scsi_Host *shost) in scsi_restart_operations() argument
2015 shost_for_each_device(sdev, shost) { in scsi_restart_operations()
2028 shost_printk(KERN_INFO, shost, "waking up host to restart\n")); in scsi_restart_operations()
2030 spin_lock_irqsave(shost->host_lock, flags); in scsi_restart_operations()
2031 if (scsi_host_set_state(shost, SHOST_RUNNING)) in scsi_restart_operations()
2032 if (scsi_host_set_state(shost, SHOST_CANCEL)) in scsi_restart_operations()
2033 BUG_ON(scsi_host_set_state(shost, SHOST_DEL)); in scsi_restart_operations()
2034 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_restart_operations()
2036 wake_up(&shost->host_wait); in scsi_restart_operations()
2044 scsi_run_host_queues(shost); in scsi_restart_operations()
2054 spin_lock_irqsave(shost->host_lock, flags); in scsi_restart_operations()
2055 if (shost->host_eh_scheduled) in scsi_restart_operations()
2056 if (scsi_host_set_state(shost, SHOST_RECOVERY)) in scsi_restart_operations()
2057 WARN_ON(scsi_host_set_state(shost, SHOST_CANCEL_RECOVERY)); in scsi_restart_operations()
2058 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_restart_operations()
2067 void scsi_eh_ready_devs(struct Scsi_Host *shost, in scsi_eh_ready_devs() argument
2071 if (!scsi_eh_stu(shost, work_q, done_q)) in scsi_eh_ready_devs()
2072 if (!scsi_eh_bus_device_reset(shost, work_q, done_q)) in scsi_eh_ready_devs()
2073 if (!scsi_eh_target_reset(shost, work_q, done_q)) in scsi_eh_ready_devs()
2074 if (!scsi_eh_bus_reset(shost, work_q, done_q)) in scsi_eh_ready_devs()
2075 if (!scsi_eh_host_reset(shost, work_q, done_q)) in scsi_eh_ready_devs()
2140 static void scsi_unjam_host(struct Scsi_Host *shost) in scsi_unjam_host() argument
2146 spin_lock_irqsave(shost->host_lock, flags); in scsi_unjam_host()
2147 list_splice_init(&shost->eh_cmd_q, &eh_work_q); in scsi_unjam_host()
2148 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_unjam_host()
2150 SCSI_LOG_ERROR_RECOVERY(1, scsi_eh_prt_fail_stats(shost, &eh_work_q)); in scsi_unjam_host()
2153 scsi_eh_ready_devs(shost, &eh_work_q, &eh_done_q); in scsi_unjam_host()
2155 spin_lock_irqsave(shost->host_lock, flags); in scsi_unjam_host()
2156 if (shost->eh_deadline != -1) in scsi_unjam_host()
2157 shost->last_reset = 0; in scsi_unjam_host()
2158 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_unjam_host()
2172 struct Scsi_Host *shost = data; in scsi_error_handler() local
2191 if ((shost->host_failed == 0 && shost->host_eh_scheduled == 0) || in scsi_error_handler()
2192 shost->host_failed != scsi_host_busy(shost)) { in scsi_error_handler()
2194 shost_printk(KERN_INFO, shost, in scsi_error_handler()
2196 shost->host_no)); in scsi_error_handler()
2203 shost_printk(KERN_INFO, shost, in scsi_error_handler()
2205 shost->host_no, shost->host_eh_scheduled, in scsi_error_handler()
2206 shost->host_failed, in scsi_error_handler()
2207 scsi_host_busy(shost))); in scsi_error_handler()
2214 if (!shost->eh_noresume && scsi_autopm_get_host(shost) != 0) { in scsi_error_handler()
2216 shost_printk(KERN_ERR, shost, in scsi_error_handler()
2218 shost->host_no)); in scsi_error_handler()
2222 if (shost->transportt->eh_strategy_handler) in scsi_error_handler()
2223 shost->transportt->eh_strategy_handler(shost); in scsi_error_handler()
2225 scsi_unjam_host(shost); in scsi_error_handler()
2228 shost->host_failed = 0; in scsi_error_handler()
2237 scsi_restart_operations(shost); in scsi_error_handler()
2238 if (!shost->eh_noresume) in scsi_error_handler()
2239 scsi_autopm_put_host(shost); in scsi_error_handler()
2244 shost_printk(KERN_INFO, shost, in scsi_error_handler()
2246 shost->host_no)); in scsi_error_handler()
2247 shost->ehandler = NULL; in scsi_error_handler()
2272 void scsi_report_bus_reset(struct Scsi_Host *shost, int channel) in scsi_report_bus_reset() argument
2276 __shost_for_each_device(sdev, shost) { in scsi_report_bus_reset()
2305 void scsi_report_device_reset(struct Scsi_Host *shost, int channel, int target) in scsi_report_device_reset() argument
2309 __shost_for_each_device(sdev, shost) { in scsi_report_device_reset()
2331 struct Scsi_Host *shost = dev->host; in scsi_ioctl_reset() local
2343 if (scsi_autopm_get_host(shost) < 0) in scsi_ioctl_reset()
2348 shost->hostt->cmd_size, GFP_KERNEL); in scsi_ioctl_reset()
2365 spin_lock_irqsave(shost->host_lock, flags); in scsi_ioctl_reset()
2366 shost->tmf_in_progress = 1; in scsi_ioctl_reset()
2367 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_ioctl_reset()
2400 spin_lock_irqsave(shost->host_lock, flags); in scsi_ioctl_reset()
2401 shost->tmf_in_progress = 0; in scsi_ioctl_reset()
2402 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_ioctl_reset()
2409 shost_printk(KERN_INFO, shost, in scsi_ioctl_reset()
2412 wake_up(&shost->host_wait); in scsi_ioctl_reset()
2413 scsi_run_host_queues(shost); in scsi_ioctl_reset()
2419 scsi_autopm_put_host(shost); in scsi_ioctl_reset()