Lines Matching refs:bfad

46 	struct bfad_s         *bfad = drv;  in bfa_cb_ioim_done()  local
53 bfa_trc(bfad, scsi_status); in bfa_cb_ioim_done()
57 bfa_trc(bfad, sns_len); in bfa_cb_ioim_done()
64 bfa_trc(bfad, residue); in bfa_cb_ioim_done()
69 bfa_trc(bfad, 0); in bfa_cb_ioim_done()
92 bfa_trc(bfad, cmnd->result); in bfa_cb_ioim_done()
142 struct bfad_s *bfad = drv; in bfa_cb_ioim_abort() local
150 bfa_trc(bfad, cmnd->result); in bfa_cb_ioim_abort()
155 bfa_cb_tskim_done(void *bfad, struct bfad_tskim_s *dtsk, in bfa_cb_tskim_done() argument
182 struct bfad_s *bfad = im_port->bfad; in bfad_im_info() local
187 bfad->pci_name, BFAD_DRIVER_VERSION); in bfad_im_info()
203 struct bfad_s *bfad = im_port->bfad; in bfad_im_abort_handler() local
209 spin_lock_irqsave(&bfad->bfad_lock, flags); in bfad_im_abort_handler()
221 bfa_trc(bfad, hal_io->iotag); in bfad_im_abort_handler()
222 BFA_LOG(KERN_INFO, bfad, bfa_log_level, in bfad_im_abort_handler()
226 spin_unlock_irqrestore(&bfad->bfad_lock, flags); in bfad_im_abort_handler()
238 bfa_trc(bfad, hal_io->iotag); in bfad_im_abort_handler()
239 BFA_LOG(KERN_INFO, bfad, bfa_log_level, in bfad_im_abort_handler()
244 spin_unlock_irqrestore(&bfad->bfad_lock, flags); in bfad_im_abort_handler()
249 bfad_im_target_reset_send(struct bfad_s *bfad, struct scsi_cmnd *cmnd, in bfad_im_target_reset_send() argument
257 tskim = bfa_tskim_alloc(&bfad->bfa, (struct bfad_tskim_s *) cmnd); in bfad_im_target_reset_send()
259 BFA_LOG(KERN_ERR, bfad, bfa_log_level, in bfad_im_target_reset_send()
279 BFA_LOG(KERN_ERR, bfad, bfa_log_level, in bfad_im_target_reset_send()
305 struct bfad_s *bfad = im_port->bfad; in bfad_im_reset_lun_handler() local
315 spin_lock_irqsave(&bfad->bfad_lock, flags); in bfad_im_reset_lun_handler()
318 spin_unlock_irqrestore(&bfad->bfad_lock, flags); in bfad_im_reset_lun_handler()
323 tskim = bfa_tskim_alloc(&bfad->bfa, (struct bfad_tskim_s *) cmnd); in bfad_im_reset_lun_handler()
325 BFA_LOG(KERN_ERR, bfad, bfa_log_level, in bfad_im_reset_lun_handler()
327 spin_unlock_irqrestore(&bfad->bfad_lock, flags); in bfad_im_reset_lun_handler()
347 BFA_LOG(KERN_ERR, bfad, bfa_log_level, in bfad_im_reset_lun_handler()
349 spin_unlock_irqrestore(&bfad->bfad_lock, flags); in bfad_im_reset_lun_handler()
356 spin_unlock_irqrestore(&bfad->bfad_lock, flags); in bfad_im_reset_lun_handler()
363 BFA_LOG(KERN_ERR, bfad, bfa_log_level, in bfad_im_reset_lun_handler()
382 struct bfad_s *bfad = im_port->bfad; in bfad_im_reset_target_handler() local
389 spin_lock_irqsave(&bfad->bfad_lock, flags); in bfad_im_reset_target_handler()
393 rc = bfad_im_target_reset_send(bfad, cmnd, itnim); in bfad_im_reset_target_handler()
396 spin_unlock_irqrestore(&bfad->bfad_lock, flags); in bfad_im_reset_target_handler()
399 spin_lock_irqsave(&bfad->bfad_lock, flags); in bfad_im_reset_target_handler()
403 BFA_LOG(KERN_ERR, bfad, bfa_log_level, in bfad_im_reset_target_handler()
410 spin_unlock_irqrestore(&bfad->bfad_lock, flags); in bfad_im_reset_target_handler()
434 bfa_fcb_itnim_alloc(struct bfad_s *bfad, struct bfa_fcs_itnim_s **itnim, in bfa_fcb_itnim_alloc() argument
441 (*itnim_drv)->im = bfad->im; in bfa_fcb_itnim_alloc()
449 bfad->bfad_flags |= BFAD_RPORT_ONLINE; in bfa_fcb_itnim_alloc()
458 bfa_fcb_itnim_free(struct bfad_s *bfad, struct bfad_itnim_s *itnim_drv) in bfa_fcb_itnim_free() argument
481 BFA_LOG(KERN_INFO, bfad, bfa_log_level, in bfa_fcb_itnim_free()
520 struct bfad_s *bfad; in bfa_fcb_itnim_offline() local
524 bfad = port->bfad; in bfa_fcb_itnim_offline()
525 if ((bfad->pport.flags & BFAD_PORT_DELETE) || in bfa_fcb_itnim_offline()
543 bfad_im_scsi_host_alloc(struct bfad_s *bfad, struct bfad_im_port_s *im_port, in bfad_im_scsi_host_alloc() argument
559 im_port->shost = bfad_scsi_host_alloc(im_port, bfad); in bfad_im_scsi_host_alloc()
572 im_port->shost->can_queue = bfad->cfg_data.ioc_queue_depth; in bfad_im_scsi_host_alloc()
579 error = scsi_add_host_with_dma(im_port->shost, dev, &bfad->pcidev->dev); in bfad_im_scsi_host_alloc()
599 bfad_im_scsi_host_free(struct bfad_s *bfad, struct bfad_im_port_s *im_port) in bfad_im_scsi_host_free() argument
601 bfa_trc(bfad, bfad->inst_no); in bfad_im_scsi_host_free()
602 BFA_LOG(KERN_INFO, bfad, bfa_log_level, "Free scsi%d\n", in bfad_im_scsi_host_free()
628 bfad_im_port_new(struct bfad_s *bfad, struct bfad_port_s *port) in bfad_im_port_new() argument
640 im_port->bfad = bfad; in bfad_im_port_new()
651 bfad_im_port_delete(struct bfad_s *bfad, struct bfad_port_s *port) in bfad_im_port_delete() argument
655 queue_work(bfad->im->drv_workq, in bfad_im_port_delete()
664 struct bfad_s *bfad = im_port->bfad; in bfad_im_port_clean() local
666 spin_lock_irqsave(&bfad->bfad_lock, flags); in bfad_im_port_clean()
676 spin_unlock_irqrestore(&bfad->bfad_lock, flags); in bfad_im_port_clean()
684 struct bfad_s *bfad = im->bfad; in bfad_aen_im_notify_handler() local
685 struct Scsi_Host *shost = bfad->pport.im_port->shost; in bfad_aen_im_notify_handler()
689 while (!list_empty(&bfad->active_aen_q)) { in bfad_aen_im_notify_handler()
690 spin_lock_irqsave(&bfad->bfad_aen_spinlock, flags); in bfad_aen_im_notify_handler()
691 bfa_q_deq(&bfad->active_aen_q, &aen_entry); in bfad_aen_im_notify_handler()
692 spin_unlock_irqrestore(&bfad->bfad_aen_spinlock, flags); in bfad_aen_im_notify_handler()
698 spin_lock_irqsave(&bfad->bfad_aen_spinlock, flags); in bfad_aen_im_notify_handler()
699 list_add_tail(&aen_entry->qe, &bfad->free_aen_q); in bfad_aen_im_notify_handler()
700 spin_unlock_irqrestore(&bfad->bfad_aen_spinlock, flags); in bfad_aen_im_notify_handler()
705 bfad_im_probe(struct bfad_s *bfad) in bfad_im_probe() argument
713 bfad->im = im; in bfad_im_probe()
714 im->bfad = bfad; in bfad_im_probe()
716 if (bfad_thread_workq(bfad) != BFA_STATUS_OK) { in bfad_im_probe()
726 bfad_im_probe_undo(struct bfad_s *bfad) in bfad_im_probe_undo() argument
728 if (bfad->im) { in bfad_im_probe_undo()
729 bfad_destroy_workq(bfad->im); in bfad_im_probe_undo()
730 kfree(bfad->im); in bfad_im_probe_undo()
731 bfad->im = NULL; in bfad_im_probe_undo()
736 bfad_scsi_host_alloc(struct bfad_im_port_s *im_port, struct bfad_s *bfad) in bfad_scsi_host_alloc() argument
748 sht->sg_tablesize = bfad->cfg_data.io_max_sge; in bfad_scsi_host_alloc()
754 bfad_scsi_host_free(struct bfad_s *bfad, struct bfad_im_port_s *im_port) in bfad_scsi_host_free() argument
757 flush_workqueue(bfad->im->drv_workq); in bfad_scsi_host_free()
758 bfad_im_scsi_host_free(im_port->bfad, im_port); in bfad_scsi_host_free()
774 bfad_thread_workq(struct bfad_s *bfad) in bfad_thread_workq() argument
776 struct bfad_im_s *im = bfad->im; in bfad_thread_workq()
778 bfa_trc(bfad, 0); in bfad_thread_workq()
780 bfad->inst_no); in bfad_thread_workq()
1038 struct bfad_s *bfad = im_port->bfad; in bfad_fc_host_init() local
1041 struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(&bfad->bfa); in bfad_fc_host_init()
1047 fc_host_max_npiv_vports(host) = bfa_lps_get_max_vport(&bfad->bfa); in bfad_fc_host_init()
1059 strlcpy(symname, bfad->bfa_fcs.fabric.bport.port_cfg.sym_name.symname, in bfad_fc_host_init()
1063 fc_host_supported_speeds(host) = bfad_im_supported_speeds(&bfad->bfa); in bfad_fc_host_init()
1119 struct bfad_s *bfad = im->bfad; in bfad_im_itnim_work_handler() local
1127 spin_lock_irqsave(&bfad->bfad_lock, flags); in bfad_im_itnim_work_handler()
1129 bfa_trc(bfad, itnim->state); in bfad_im_itnim_work_handler()
1133 spin_unlock_irqrestore(&bfad->bfad_lock, flags); in bfad_im_itnim_work_handler()
1135 spin_lock_irqsave(&bfad->bfad_lock, flags); in bfad_im_itnim_work_handler()
1142 BFA_LOG(KERN_INFO, bfad, bfa_log_level, in bfad_im_itnim_work_handler()
1164 spin_unlock_irqrestore(&bfad->bfad_lock, flags); in bfad_im_itnim_work_handler()
1166 bfa_fcpim_path_tov_get(&bfad->bfa) + 1; in bfad_im_itnim_work_handler()
1168 spin_lock_irqsave(&bfad->bfad_lock, flags); in bfad_im_itnim_work_handler()
1175 BFA_LOG(KERN_INFO, bfad, bfa_log_level, in bfad_im_itnim_work_handler()
1190 spin_unlock_irqrestore(&bfad->bfad_lock, flags); in bfad_im_itnim_work_handler()
1192 bfa_fcpim_path_tov_get(&bfad->bfa) + 1; in bfad_im_itnim_work_handler()
1194 spin_lock_irqsave(&bfad->bfad_lock, flags); in bfad_im_itnim_work_handler()
1206 spin_unlock_irqrestore(&bfad->bfad_lock, flags); in bfad_im_itnim_work_handler()
1217 struct bfad_s *bfad = im_port->bfad; in bfad_im_queuecommand_lck() local
1233 if (bfad->bfad_flags & BFAD_EEH_BUSY) { in bfad_im_queuecommand_lck()
1234 if (bfad->bfad_flags & BFAD_EEH_PCI_CHANNEL_IO_PERM_FAILURE) in bfad_im_queuecommand_lck()
1248 spin_lock_irqsave(&bfad->bfad_lock, flags); in bfad_im_queuecommand_lck()
1249 if (!(bfad->bfad_flags & BFAD_HAL_START_DONE)) { in bfad_im_queuecommand_lck()
1252 bfad->inst_no, cmnd, cmnd->cmnd[0]); in bfad_im_queuecommand_lck()
1264 hal_io = bfa_ioim_alloc(&bfad->bfa, (struct bfad_ioim_s *) cmnd, in bfad_im_queuecommand_lck()
1268 spin_unlock_irqrestore(&bfad->bfad_lock, flags); in bfad_im_queuecommand_lck()
1275 spin_unlock_irqrestore(&bfad->bfad_lock, flags); in bfad_im_queuecommand_lck()
1280 spin_unlock_irqrestore(&bfad->bfad_lock, flags); in bfad_im_queuecommand_lck()
1291 bfad_rport_online_wait(struct bfad_s *bfad) in DEF_SCSI_QCMD()
1296 for (i = 0; !(bfad->bfad_flags & BFAD_PORT_ONLINE) in DEF_SCSI_QCMD()
1302 if (bfad->bfad_flags & BFAD_PORT_ONLINE) { in DEF_SCSI_QCMD()
1305 for (i = 0; !(bfad->bfad_flags & BFAD_RPORT_ONLINE) in DEF_SCSI_QCMD()
1311 if (rport_delay > 0 && (bfad->bfad_flags & BFAD_RPORT_ONLINE)) { in DEF_SCSI_QCMD()
1319 bfad_get_linkup_delay(struct bfad_s *bfad) in bfad_get_linkup_delay() argument
1332 bfa_iocfc_get_bootwwns(&bfad->bfa, &nwwns, wwns); in bfad_get_linkup_delay()