Lines Matching refs:phba

221 lpfc_nvmet_defer_release(struct lpfc_hba *phba, struct lpfc_nvmet_rcv_ctx *ctxp)  in lpfc_nvmet_defer_release()  argument
225 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS, in lpfc_nvmet_defer_release()
229 spin_lock_irqsave(&phba->sli4_hba.abts_nvme_buf_list_lock, iflag); in lpfc_nvmet_defer_release()
231 spin_unlock_irqrestore(&phba->sli4_hba.abts_nvme_buf_list_lock, in lpfc_nvmet_defer_release()
236 list_add_tail(&ctxp->list, &phba->sli4_hba.lpfc_abts_nvmet_ctx_list); in lpfc_nvmet_defer_release()
237 spin_unlock_irqrestore(&phba->sli4_hba.abts_nvme_buf_list_lock, iflag); in lpfc_nvmet_defer_release()
251 lpfc_nvmet_xmt_ls_rsp_cmp(struct lpfc_hba *phba, struct lpfc_iocbq *cmdwqe, in lpfc_nvmet_xmt_ls_rsp_cmp() argument
264 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_xmt_ls_rsp_cmp()
270 if (!phba->targetport) in lpfc_nvmet_xmt_ls_rsp_cmp()
273 tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private; in lpfc_nvmet_xmt_ls_rsp_cmp()
290 lpfc_nvmeio_data(phba, "NVMET LS CMPL: xri x%x stat x%x result x%x\n", in lpfc_nvmet_xmt_ls_rsp_cmp()
293 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_DISC, in lpfc_nvmet_xmt_ls_rsp_cmp()
300 lpfc_sli_release_iocbq(phba, cmdwqe); in lpfc_nvmet_xmt_ls_rsp_cmp()
319 lpfc_nvmet_ctxbuf_post(struct lpfc_hba *phba, struct lpfc_nvmet_ctxbuf *ctx_buf) in lpfc_nvmet_ctxbuf_post() argument
333 dma_pool_free(phba->txrdy_payload_pool, ctxp->txrdy, in lpfc_nvmet_ctxbuf_post()
340 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_ctxbuf_post()
346 spin_lock_irqsave(&phba->sli4_hba.nvmet_io_wait_lock, iflag); in lpfc_nvmet_ctxbuf_post()
347 if (phba->sli4_hba.nvmet_io_wait_cnt) { in lpfc_nvmet_ctxbuf_post()
348 list_remove_head(&phba->sli4_hba.lpfc_nvmet_io_wait_list, in lpfc_nvmet_ctxbuf_post()
351 phba->sli4_hba.nvmet_io_wait_cnt--; in lpfc_nvmet_ctxbuf_post()
352 spin_unlock_irqrestore(&phba->sli4_hba.nvmet_io_wait_lock, in lpfc_nvmet_ctxbuf_post()
357 tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private; in lpfc_nvmet_ctxbuf_post()
366 ctxp->phba = phba; in lpfc_nvmet_ctxbuf_post()
400 rc = nvmet_fc_rcv_fcp_req(phba->targetport, &ctxp->ctx.fcp_req, in lpfc_nvmet_ctxbuf_post()
407 nvmebuf->hrq->rqbp->rqb_free_buffer(phba, nvmebuf); in lpfc_nvmet_ctxbuf_post()
413 lpfc_nvmeio_data(phba, in lpfc_nvmet_ctxbuf_post()
421 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_ctxbuf_post()
428 lpfc_nvmet_defer_release(phba, ctxp); in lpfc_nvmet_ctxbuf_post()
429 lpfc_nvmet_unsol_fcp_issue_abort(phba, ctxp, sid, oxid); in lpfc_nvmet_ctxbuf_post()
430 nvmebuf->hrq->rqbp->rqb_free_buffer(phba, nvmebuf); in lpfc_nvmet_ctxbuf_post()
433 spin_unlock_irqrestore(&phba->sli4_hba.nvmet_io_wait_lock, iflag); in lpfc_nvmet_ctxbuf_post()
440 infop = lpfc_get_ctx_list(phba, cpu, ctxp->idx); in lpfc_nvmet_ctxbuf_post()
450 lpfc_nvmet_ktime(struct lpfc_hba *phba, in lpfc_nvmet_ktime() argument
577 phba->ktime_seg1_total += seg1; in lpfc_nvmet_ktime()
578 if (seg1 < phba->ktime_seg1_min) in lpfc_nvmet_ktime()
579 phba->ktime_seg1_min = seg1; in lpfc_nvmet_ktime()
580 else if (seg1 > phba->ktime_seg1_max) in lpfc_nvmet_ktime()
581 phba->ktime_seg1_max = seg1; in lpfc_nvmet_ktime()
583 phba->ktime_seg2_total += seg2; in lpfc_nvmet_ktime()
584 if (seg2 < phba->ktime_seg2_min) in lpfc_nvmet_ktime()
585 phba->ktime_seg2_min = seg2; in lpfc_nvmet_ktime()
586 else if (seg2 > phba->ktime_seg2_max) in lpfc_nvmet_ktime()
587 phba->ktime_seg2_max = seg2; in lpfc_nvmet_ktime()
589 phba->ktime_seg3_total += seg3; in lpfc_nvmet_ktime()
590 if (seg3 < phba->ktime_seg3_min) in lpfc_nvmet_ktime()
591 phba->ktime_seg3_min = seg3; in lpfc_nvmet_ktime()
592 else if (seg3 > phba->ktime_seg3_max) in lpfc_nvmet_ktime()
593 phba->ktime_seg3_max = seg3; in lpfc_nvmet_ktime()
595 phba->ktime_seg4_total += seg4; in lpfc_nvmet_ktime()
596 if (seg4 < phba->ktime_seg4_min) in lpfc_nvmet_ktime()
597 phba->ktime_seg4_min = seg4; in lpfc_nvmet_ktime()
598 else if (seg4 > phba->ktime_seg4_max) in lpfc_nvmet_ktime()
599 phba->ktime_seg4_max = seg4; in lpfc_nvmet_ktime()
601 phba->ktime_seg5_total += seg5; in lpfc_nvmet_ktime()
602 if (seg5 < phba->ktime_seg5_min) in lpfc_nvmet_ktime()
603 phba->ktime_seg5_min = seg5; in lpfc_nvmet_ktime()
604 else if (seg5 > phba->ktime_seg5_max) in lpfc_nvmet_ktime()
605 phba->ktime_seg5_max = seg5; in lpfc_nvmet_ktime()
607 phba->ktime_data_samples++; in lpfc_nvmet_ktime()
611 phba->ktime_seg6_total += seg6; in lpfc_nvmet_ktime()
612 if (seg6 < phba->ktime_seg6_min) in lpfc_nvmet_ktime()
613 phba->ktime_seg6_min = seg6; in lpfc_nvmet_ktime()
614 else if (seg6 > phba->ktime_seg6_max) in lpfc_nvmet_ktime()
615 phba->ktime_seg6_max = seg6; in lpfc_nvmet_ktime()
617 phba->ktime_seg7_total += seg7; in lpfc_nvmet_ktime()
618 if (seg7 < phba->ktime_seg7_min) in lpfc_nvmet_ktime()
619 phba->ktime_seg7_min = seg7; in lpfc_nvmet_ktime()
620 else if (seg7 > phba->ktime_seg7_max) in lpfc_nvmet_ktime()
621 phba->ktime_seg7_max = seg7; in lpfc_nvmet_ktime()
623 phba->ktime_seg8_total += seg8; in lpfc_nvmet_ktime()
624 if (seg8 < phba->ktime_seg8_min) in lpfc_nvmet_ktime()
625 phba->ktime_seg8_min = seg8; in lpfc_nvmet_ktime()
626 else if (seg8 > phba->ktime_seg8_max) in lpfc_nvmet_ktime()
627 phba->ktime_seg8_max = seg8; in lpfc_nvmet_ktime()
629 phba->ktime_seg9_total += seg9; in lpfc_nvmet_ktime()
630 if (seg9 < phba->ktime_seg9_min) in lpfc_nvmet_ktime()
631 phba->ktime_seg9_min = seg9; in lpfc_nvmet_ktime()
632 else if (seg9 > phba->ktime_seg9_max) in lpfc_nvmet_ktime()
633 phba->ktime_seg9_max = seg9; in lpfc_nvmet_ktime()
635 phba->ktime_seg10_total += seg10; in lpfc_nvmet_ktime()
636 if (seg10 < phba->ktime_seg10_min) in lpfc_nvmet_ktime()
637 phba->ktime_seg10_min = seg10; in lpfc_nvmet_ktime()
638 else if (seg10 > phba->ktime_seg10_max) in lpfc_nvmet_ktime()
639 phba->ktime_seg10_max = seg10; in lpfc_nvmet_ktime()
640 phba->ktime_status_samples++; in lpfc_nvmet_ktime()
655 lpfc_nvmet_xmt_fcp_op_cmp(struct lpfc_hba *phba, struct lpfc_iocbq *cmdwqe, in lpfc_nvmet_xmt_fcp_op_cmp() argument
675 if (phba->targetport) in lpfc_nvmet_xmt_fcp_op_cmp()
676 tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private; in lpfc_nvmet_xmt_fcp_op_cmp()
680 lpfc_nvmeio_data(phba, "NVMET FCP CMPL: xri x%x op x%x status x%x\n", in lpfc_nvmet_xmt_fcp_op_cmp()
705 lpfc_printf_log(phba, KERN_INFO, logerr, in lpfc_nvmet_xmt_fcp_op_cmp()
747 if (phba->cpucheck_on & LPFC_CHECK_NVMET_IO) { in lpfc_nvmet_xmt_fcp_op_cmp()
750 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_xmt_fcp_op_cmp()
755 phba->cpucheck_cmpl_io[id]++; in lpfc_nvmet_xmt_fcp_op_cmp()
761 lpfc_nvmet_ktime(phba, ctxp); in lpfc_nvmet_xmt_fcp_op_cmp()
774 if (phba->cpucheck_on & LPFC_CHECK_NVMET_IO) { in lpfc_nvmet_xmt_fcp_op_cmp()
777 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_xmt_fcp_op_cmp()
782 phba->cpucheck_ccmpl_io[id]++; in lpfc_nvmet_xmt_fcp_op_cmp()
795 struct lpfc_hba *phba = ctxp->phba; in lpfc_nvmet_xmt_ls_rsp() local
804 if (phba->pport->load_flag & FC_UNLOADING) in lpfc_nvmet_xmt_ls_rsp()
807 if (phba->pport->load_flag & FC_UNLOADING) in lpfc_nvmet_xmt_ls_rsp()
810 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_DISC, in lpfc_nvmet_xmt_ls_rsp()
815 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_xmt_ls_rsp()
823 nvmewqeq = lpfc_nvmet_prep_ls_wqe(phba, ctxp, rsp->rspdma, in lpfc_nvmet_xmt_ls_rsp()
827 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_xmt_ls_rsp()
830 lpfc_in_buf_free(phba, &nvmebuf->dbuf); in lpfc_nvmet_xmt_ls_rsp()
832 lpfc_nvmet_unsol_ls_issue_abort(phba, ctxp, in lpfc_nvmet_xmt_ls_rsp()
852 lpfc_nvmeio_data(phba, "NVMET LS RESP: xri x%x wqidx x%x len x%x\n", in lpfc_nvmet_xmt_ls_rsp()
855 rc = lpfc_sli4_issue_wqe(phba, LPFC_ELS_RING, nvmewqeq); in lpfc_nvmet_xmt_ls_rsp()
861 lpfc_in_buf_free(phba, &nvmebuf->dbuf); in lpfc_nvmet_xmt_ls_rsp()
868 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_xmt_ls_rsp()
874 lpfc_in_buf_free(phba, &nvmebuf->dbuf); in lpfc_nvmet_xmt_ls_rsp()
876 lpfc_nvmet_unsol_ls_issue_abort(phba, ctxp, ctxp->sid, ctxp->oxid); in lpfc_nvmet_xmt_ls_rsp()
887 struct lpfc_hba *phba = ctxp->phba; in lpfc_nvmet_xmt_fcp_op() local
894 if (phba->pport->load_flag & FC_UNLOADING) { in lpfc_nvmet_xmt_fcp_op()
899 if (phba->pport->load_flag & FC_UNLOADING) { in lpfc_nvmet_xmt_fcp_op()
911 if (phba->cpucheck_on & LPFC_CHECK_NVMET_IO) { in lpfc_nvmet_xmt_fcp_op()
915 phba->cpucheck_xmt_io[id]++; in lpfc_nvmet_xmt_fcp_op()
917 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_xmt_fcp_op()
930 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_xmt_fcp_op()
937 nvmewqeq = lpfc_nvmet_prep_fcp_wqe(phba, ctxp); in lpfc_nvmet_xmt_fcp_op()
940 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_xmt_fcp_op()
953 lpfc_nvmeio_data(phba, "NVMET FCP CMND: xri x%x op x%x len x%x\n", in lpfc_nvmet_xmt_fcp_op()
957 rc = lpfc_sli4_issue_wqe(phba, LPFC_FCP_RING, nvmewqeq); in lpfc_nvmet_xmt_fcp_op()
976 wq = phba->sli4_hba.nvme_wq[rsp->hwqid]; in lpfc_nvmet_xmt_fcp_op()
988 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_xmt_fcp_op()
1016 struct lpfc_hba *phba = ctxp->phba; in lpfc_nvmet_xmt_fcp_abort() local
1020 if (phba->pport->load_flag & FC_UNLOADING) in lpfc_nvmet_xmt_fcp_abort()
1023 if (phba->pport->load_flag & FC_UNLOADING) in lpfc_nvmet_xmt_fcp_abort()
1026 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS, in lpfc_nvmet_xmt_fcp_abort()
1030 lpfc_nvmeio_data(phba, "NVMET FCP ABRT: xri x%x flg x%x ste x%x\n", in lpfc_nvmet_xmt_fcp_abort()
1048 lpfc_nvmet_unsol_fcp_issue_abort(phba, ctxp, ctxp->sid, in lpfc_nvmet_xmt_fcp_abort()
1050 wq = phba->sli4_hba.nvme_wq[ctxp->wqeq->hba_wqidx]; in lpfc_nvmet_xmt_fcp_abort()
1052 lpfc_nvmet_wqfull_flush(phba, wq, ctxp); in lpfc_nvmet_xmt_fcp_abort()
1061 lpfc_nvmet_unsol_fcp_issue_abort(phba, ctxp, ctxp->sid, in lpfc_nvmet_xmt_fcp_abort()
1064 lpfc_nvmet_sol_fcp_issue_abort(phba, ctxp, ctxp->sid, in lpfc_nvmet_xmt_fcp_abort()
1076 struct lpfc_hba *phba = ctxp->phba; in lpfc_nvmet_xmt_fcp_release() local
1082 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_xmt_fcp_release()
1092 lpfc_nvmet_defer_release(phba, ctxp); in lpfc_nvmet_xmt_fcp_release()
1096 lpfc_nvmeio_data(phba, "NVMET FCP FREE: xri x%x ste %d abt %d\n", ctxp->oxid, in lpfc_nvmet_xmt_fcp_release()
1104 lpfc_nvmet_ctxbuf_post(phba, ctxp->ctxbuf); in lpfc_nvmet_xmt_fcp_release()
1115 struct lpfc_hba *phba = ctxp->phba; in lpfc_nvmet_defer_rcv() local
1117 lpfc_nvmeio_data(phba, "NVMET DEFERRCV: xri x%x sz %d CPU %02x\n", in lpfc_nvmet_defer_rcv()
1121 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_IOERR, in lpfc_nvmet_defer_rcv()
1128 tgtp = phba->targetport->private; in lpfc_nvmet_defer_rcv()
1133 nvmebuf->hrq->rqbp->rqb_free_buffer(phba, nvmebuf); in lpfc_nvmet_defer_rcv()
1156 __lpfc_nvmet_clean_io_for_cpu(struct lpfc_hba *phba, in __lpfc_nvmet_clean_io_for_cpu() argument
1165 spin_lock(&phba->sli4_hba.abts_nvme_buf_list_lock); in __lpfc_nvmet_clean_io_for_cpu()
1167 spin_unlock(&phba->sli4_hba.abts_nvme_buf_list_lock); in __lpfc_nvmet_clean_io_for_cpu()
1169 __lpfc_clear_active_sglq(phba, ctx_buf->sglq->sli4_lxritag); in __lpfc_nvmet_clean_io_for_cpu()
1173 spin_lock(&phba->sli4_hba.sgl_list_lock); in __lpfc_nvmet_clean_io_for_cpu()
1175 &phba->sli4_hba.lpfc_nvmet_sgl_list); in __lpfc_nvmet_clean_io_for_cpu()
1176 spin_unlock(&phba->sli4_hba.sgl_list_lock); in __lpfc_nvmet_clean_io_for_cpu()
1178 lpfc_sli_release_iocbq(phba, ctx_buf->iocbq); in __lpfc_nvmet_clean_io_for_cpu()
1185 lpfc_nvmet_cleanup_io_context(struct lpfc_hba *phba) in lpfc_nvmet_cleanup_io_context() argument
1191 infop = phba->sli4_hba.nvmet_ctx_info; in lpfc_nvmet_cleanup_io_context()
1196 for (i = 0; i < phba->cfg_nvmet_mrq; i++) { in lpfc_nvmet_cleanup_io_context()
1197 for (j = 0; j < phba->sli4_hba.num_present_cpu; j++) { in lpfc_nvmet_cleanup_io_context()
1198 __lpfc_nvmet_clean_io_for_cpu(phba, infop); in lpfc_nvmet_cleanup_io_context()
1202 kfree(phba->sli4_hba.nvmet_ctx_info); in lpfc_nvmet_cleanup_io_context()
1203 phba->sli4_hba.nvmet_ctx_info = NULL; in lpfc_nvmet_cleanup_io_context()
1207 lpfc_nvmet_setup_io_context(struct lpfc_hba *phba) in lpfc_nvmet_setup_io_context() argument
1216 lpfc_printf_log(phba, KERN_INFO, LOG_NVME, in lpfc_nvmet_setup_io_context()
1218 phba->sli4_hba.nvmet_xri_cnt); in lpfc_nvmet_setup_io_context()
1220 phba->sli4_hba.nvmet_ctx_info = kcalloc( in lpfc_nvmet_setup_io_context()
1221 phba->sli4_hba.num_present_cpu * phba->cfg_nvmet_mrq, in lpfc_nvmet_setup_io_context()
1223 if (!phba->sli4_hba.nvmet_ctx_info) { in lpfc_nvmet_setup_io_context()
1224 lpfc_printf_log(phba, KERN_ERR, LOG_INIT, in lpfc_nvmet_setup_io_context()
1249 infop = phba->sli4_hba.nvmet_ctx_info; in lpfc_nvmet_setup_io_context()
1250 for (i = 0; i < phba->sli4_hba.num_present_cpu; i++) { in lpfc_nvmet_setup_io_context()
1251 for (j = 0; j < phba->cfg_nvmet_mrq; j++) { in lpfc_nvmet_setup_io_context()
1264 for (j = 0; j < phba->cfg_nvmet_mrq; j++) { in lpfc_nvmet_setup_io_context()
1265 last_infop = lpfc_get_ctx_list(phba, 0, j); in lpfc_nvmet_setup_io_context()
1266 for (i = phba->sli4_hba.num_present_cpu - 1; i >= 0; i--) { in lpfc_nvmet_setup_io_context()
1267 infop = lpfc_get_ctx_list(phba, i, j); in lpfc_nvmet_setup_io_context()
1277 for (i = 0; i < phba->sli4_hba.nvmet_xri_cnt; i++) { in lpfc_nvmet_setup_io_context()
1280 lpfc_printf_log(phba, KERN_ERR, LOG_NVME, in lpfc_nvmet_setup_io_context()
1289 lpfc_printf_log(phba, KERN_ERR, LOG_NVME, in lpfc_nvmet_setup_io_context()
1297 ctx_buf->iocbq = lpfc_sli_get_iocbq(phba); in lpfc_nvmet_setup_io_context()
1301 lpfc_printf_log(phba, KERN_ERR, LOG_NVME, in lpfc_nvmet_setup_io_context()
1313 spin_lock(&phba->sli4_hba.sgl_list_lock); in lpfc_nvmet_setup_io_context()
1314 ctx_buf->sglq = __lpfc_sli_get_nvmet_sglq(phba, ctx_buf->iocbq); in lpfc_nvmet_setup_io_context()
1315 spin_unlock(&phba->sli4_hba.sgl_list_lock); in lpfc_nvmet_setup_io_context()
1317 lpfc_sli_release_iocbq(phba, ctx_buf->iocbq); in lpfc_nvmet_setup_io_context()
1320 lpfc_printf_log(phba, KERN_ERR, LOG_NVME, in lpfc_nvmet_setup_io_context()
1330 infop = lpfc_get_ctx_list(phba, idx, idx); in lpfc_nvmet_setup_io_context()
1338 if (idx >= phba->cfg_nvmet_mrq) in lpfc_nvmet_setup_io_context()
1342 infop = phba->sli4_hba.nvmet_ctx_info; in lpfc_nvmet_setup_io_context()
1343 for (j = 0; j < phba->cfg_nvmet_mrq; j++) { in lpfc_nvmet_setup_io_context()
1344 for (i = 0; i < phba->sli4_hba.num_present_cpu; i++) { in lpfc_nvmet_setup_io_context()
1345 lpfc_printf_log(phba, KERN_INFO, LOG_NVME | LOG_INIT, in lpfc_nvmet_setup_io_context()
1357 lpfc_nvmet_create_targetport(struct lpfc_hba *phba) in lpfc_nvmet_create_targetport() argument
1359 struct lpfc_vport *vport = phba->pport; in lpfc_nvmet_create_targetport()
1364 if (phba->targetport) in lpfc_nvmet_create_targetport()
1367 error = lpfc_nvmet_setup_io_context(phba); in lpfc_nvmet_create_targetport()
1379 if (phba->cfg_sg_seg_cnt > LPFC_MAX_NVME_SEG_CNT) { in lpfc_nvmet_create_targetport()
1380 lpfc_printf_log(phba, KERN_INFO, LOG_NVME | LOG_INIT, in lpfc_nvmet_create_targetport()
1383 phba->cfg_nvme_seg_cnt = LPFC_MAX_NVME_SEG_CNT; in lpfc_nvmet_create_targetport()
1385 phba->cfg_nvme_seg_cnt = phba->cfg_sg_seg_cnt; in lpfc_nvmet_create_targetport()
1387 lpfc_tgttemplate.max_sgl_segments = phba->cfg_nvme_seg_cnt + 1; in lpfc_nvmet_create_targetport()
1388 lpfc_tgttemplate.max_hw_queues = phba->cfg_nvme_io_channel; in lpfc_nvmet_create_targetport()
1393 &phba->pcidev->dev, in lpfc_nvmet_create_targetport()
1394 &phba->targetport); in lpfc_nvmet_create_targetport()
1399 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_DISC, in lpfc_nvmet_create_targetport()
1406 phba->targetport = NULL; in lpfc_nvmet_create_targetport()
1407 phba->nvmet_support = 0; in lpfc_nvmet_create_targetport()
1409 lpfc_nvmet_cleanup_io_context(phba); in lpfc_nvmet_create_targetport()
1413 phba->targetport->private; in lpfc_nvmet_create_targetport()
1414 tgtp->phba = phba; in lpfc_nvmet_create_targetport()
1416 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_DISC, in lpfc_nvmet_create_targetport()
1420 phba->targetport, tgtp, in lpfc_nvmet_create_targetport()
1465 lpfc_nvmet_update_targetport(struct lpfc_hba *phba) in lpfc_nvmet_update_targetport() argument
1467 struct lpfc_vport *vport = phba->pport; in lpfc_nvmet_update_targetport()
1469 if (!phba->targetport) in lpfc_nvmet_update_targetport()
1474 phba->targetport, vport->fc_myDID); in lpfc_nvmet_update_targetport()
1476 phba->targetport->port_id = vport->fc_myDID; in lpfc_nvmet_update_targetport()
1489 lpfc_sli4_nvmet_xri_aborted(struct lpfc_hba *phba, in lpfc_sli4_nvmet_xri_aborted() argument
1501 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS, in lpfc_sli4_nvmet_xri_aborted()
1504 if (!(phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME)) in lpfc_sli4_nvmet_xri_aborted()
1507 if (phba->targetport) { in lpfc_sli4_nvmet_xri_aborted()
1508 tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private; in lpfc_sli4_nvmet_xri_aborted()
1512 spin_lock_irqsave(&phba->hbalock, iflag); in lpfc_sli4_nvmet_xri_aborted()
1513 spin_lock(&phba->sli4_hba.abts_nvme_buf_list_lock); in lpfc_sli4_nvmet_xri_aborted()
1515 &phba->sli4_hba.lpfc_abts_nvmet_ctx_list, in lpfc_sli4_nvmet_xri_aborted()
1529 spin_unlock(&phba->sli4_hba.abts_nvme_buf_list_lock); in lpfc_sli4_nvmet_xri_aborted()
1531 rrq_empty = list_empty(&phba->active_rrq_list); in lpfc_sli4_nvmet_xri_aborted()
1532 spin_unlock_irqrestore(&phba->hbalock, iflag); in lpfc_sli4_nvmet_xri_aborted()
1533 ndlp = lpfc_findnode_did(phba->pport, ctxp->sid); in lpfc_sli4_nvmet_xri_aborted()
1537 lpfc_set_rrq_active(phba, ndlp, in lpfc_sli4_nvmet_xri_aborted()
1540 lpfc_sli4_abts_err_handler(phba, ndlp, axri); in lpfc_sli4_nvmet_xri_aborted()
1543 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS, in lpfc_sli4_nvmet_xri_aborted()
1547 lpfc_nvmet_ctxbuf_post(phba, ctxp->ctxbuf); in lpfc_sli4_nvmet_xri_aborted()
1550 lpfc_worker_wake_up(phba); in lpfc_sli4_nvmet_xri_aborted()
1553 spin_unlock(&phba->sli4_hba.abts_nvme_buf_list_lock); in lpfc_sli4_nvmet_xri_aborted()
1554 spin_unlock_irqrestore(&phba->hbalock, iflag); in lpfc_sli4_nvmet_xri_aborted()
1563 struct lpfc_hba *phba = vport->phba; in lpfc_nvmet_rcv_unsol_abort() local
1571 spin_lock_irqsave(&phba->hbalock, iflag); in lpfc_nvmet_rcv_unsol_abort()
1572 spin_lock(&phba->sli4_hba.abts_nvme_buf_list_lock); in lpfc_nvmet_rcv_unsol_abort()
1574 &phba->sli4_hba.lpfc_abts_nvmet_ctx_list, in lpfc_nvmet_rcv_unsol_abort()
1579 spin_unlock(&phba->sli4_hba.abts_nvme_buf_list_lock); in lpfc_nvmet_rcv_unsol_abort()
1580 spin_unlock_irqrestore(&phba->hbalock, iflag); in lpfc_nvmet_rcv_unsol_abort()
1586 lpfc_nvmeio_data(phba, in lpfc_nvmet_rcv_unsol_abort()
1590 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS, in lpfc_nvmet_rcv_unsol_abort()
1594 nvmet_fc_rcv_fcp_abort(phba->targetport, rsp); in lpfc_nvmet_rcv_unsol_abort()
1600 spin_unlock(&phba->sli4_hba.abts_nvme_buf_list_lock); in lpfc_nvmet_rcv_unsol_abort()
1601 spin_unlock_irqrestore(&phba->hbalock, iflag); in lpfc_nvmet_rcv_unsol_abort()
1603 lpfc_nvmeio_data(phba, "NVMET ABTS RCV: xri x%x CPU %02x rjt %d\n", in lpfc_nvmet_rcv_unsol_abort()
1606 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS, in lpfc_nvmet_rcv_unsol_abort()
1616 lpfc_nvmet_wqfull_flush(struct lpfc_hba *phba, struct lpfc_queue *wq, in lpfc_nvmet_wqfull_flush() argument
1643 lpfc_nvmet_xmt_fcp_op_cmp(phba, nvmewqeq, in lpfc_nvmet_wqfull_flush()
1652 lpfc_nvmet_xmt_fcp_op_cmp(phba, nvmewqeq, wcqep); in lpfc_nvmet_wqfull_flush()
1662 lpfc_nvmet_wqfull_process(struct lpfc_hba *phba, in lpfc_nvmet_wqfull_process() argument
1681 rc = lpfc_sli4_issue_wqe(phba, LPFC_FCP_RING, nvmewqeq); in lpfc_nvmet_wqfull_process()
1697 lpfc_nvmet_destroy_targetport(struct lpfc_hba *phba) in lpfc_nvmet_destroy_targetport() argument
1704 if (phba->nvmet_support == 0) in lpfc_nvmet_destroy_targetport()
1706 if (phba->targetport) { in lpfc_nvmet_destroy_targetport()
1707 tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private; in lpfc_nvmet_destroy_targetport()
1708 for (qidx = 0; qidx < phba->cfg_nvme_io_channel; qidx++) { in lpfc_nvmet_destroy_targetport()
1709 wq = phba->sli4_hba.nvme_wq[qidx]; in lpfc_nvmet_destroy_targetport()
1710 lpfc_nvmet_wqfull_flush(phba, wq, NULL); in lpfc_nvmet_destroy_targetport()
1713 nvmet_fc_unregister_targetport(phba->targetport); in lpfc_nvmet_destroy_targetport()
1715 lpfc_nvmet_cleanup_io_context(phba); in lpfc_nvmet_destroy_targetport()
1717 phba->targetport = NULL; in lpfc_nvmet_destroy_targetport()
1735 lpfc_nvmet_unsol_ls_buffer(struct lpfc_hba *phba, struct lpfc_sli_ring *pring, in lpfc_nvmet_unsol_ls_buffer() argument
1748 if (!phba->targetport) { in lpfc_nvmet_unsol_ls_buffer()
1749 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_unsol_ls_buffer()
1758 tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private; in lpfc_nvmet_unsol_ls_buffer()
1766 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_unsol_ls_buffer()
1770 lpfc_nvmeio_data(phba, "NVMET LS DROP: " in lpfc_nvmet_unsol_ls_buffer()
1773 lpfc_in_buf_free(phba, &nvmebuf->dbuf); in lpfc_nvmet_unsol_ls_buffer()
1776 ctxp->phba = phba; in lpfc_nvmet_unsol_ls_buffer()
1785 lpfc_nvmeio_data(phba, "NVMET LS RCV: xri x%x sz %d from %06x\n", in lpfc_nvmet_unsol_ls_buffer()
1793 rc = nvmet_fc_rcv_ls_req(phba->targetport, &ctxp->ctx.ls_req, in lpfc_nvmet_unsol_ls_buffer()
1796 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_DISC, in lpfc_nvmet_unsol_ls_buffer()
1807 lpfc_nvmeio_data(phba, "NVMET LS DROP: xri x%x sz %d from %06x\n", in lpfc_nvmet_unsol_ls_buffer()
1811 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_unsol_ls_buffer()
1816 lpfc_in_buf_free(phba, &nvmebuf->dbuf); in lpfc_nvmet_unsol_ls_buffer()
1819 lpfc_nvmet_unsol_ls_issue_abort(phba, ctxp, sid, oxid); in lpfc_nvmet_unsol_ls_buffer()
1824 lpfc_nvmet_replenish_context(struct lpfc_hba *phba, in lpfc_nvmet_replenish_context() argument
1847 for (i = 0; i < phba->sli4_hba.num_present_cpu; i++) { in lpfc_nvmet_replenish_context()
1894 lpfc_nvmet_unsol_fcp_buffer(struct lpfc_hba *phba, in lpfc_nvmet_unsol_fcp_buffer() argument
1916 if (!nvmebuf || !phba->targetport) { in lpfc_nvmet_unsol_fcp_buffer()
1917 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_unsol_fcp_buffer()
1934 current_infop = lpfc_get_ctx_list(phba, current_cpu, idx); in lpfc_nvmet_unsol_fcp_buffer()
1941 ctx_buf = lpfc_nvmet_replenish_context(phba, current_infop); in lpfc_nvmet_unsol_fcp_buffer()
1950 if (phba->cpucheck_on & LPFC_CHECK_NVMET_RCV) { in lpfc_nvmet_unsol_fcp_buffer()
1953 phba->cpucheck_rcv_io[id]++; in lpfc_nvmet_unsol_fcp_buffer()
1957 lpfc_nvmeio_data(phba, "NVMET FCP RCV: xri x%x sz %d CPU %02x\n", in lpfc_nvmet_unsol_fcp_buffer()
1960 tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private; in lpfc_nvmet_unsol_fcp_buffer()
1964 spin_lock_irqsave(&phba->sli4_hba.nvmet_io_wait_lock, iflag); in lpfc_nvmet_unsol_fcp_buffer()
1966 &phba->sli4_hba.lpfc_nvmet_io_wait_list); in lpfc_nvmet_unsol_fcp_buffer()
1967 phba->sli4_hba.nvmet_io_wait_cnt++; in lpfc_nvmet_unsol_fcp_buffer()
1968 phba->sli4_hba.nvmet_io_wait_total++; in lpfc_nvmet_unsol_fcp_buffer()
1969 spin_unlock_irqrestore(&phba->sli4_hba.nvmet_io_wait_lock, in lpfc_nvmet_unsol_fcp_buffer()
1975 phba, phba->sli4_hba.nvmet_mrq_hdr[qno], in lpfc_nvmet_unsol_fcp_buffer()
1976 phba->sli4_hba.nvmet_mrq_data[qno], 1, qno); in lpfc_nvmet_unsol_fcp_buffer()
1987 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_unsol_fcp_buffer()
1994 ctxp->phba = phba; in lpfc_nvmet_unsol_fcp_buffer()
2032 rc = nvmet_fc_rcv_fcp_req(phba->targetport, &ctxp->ctx.fcp_req, in lpfc_nvmet_unsol_fcp_buffer()
2039 lpfc_rq_buf_free(phba, &nvmebuf->hbuf); /* repost */ in lpfc_nvmet_unsol_fcp_buffer()
2051 phba, phba->sli4_hba.nvmet_mrq_hdr[qno], in lpfc_nvmet_unsol_fcp_buffer()
2052 phba->sli4_hba.nvmet_mrq_data[qno], 1, qno); in lpfc_nvmet_unsol_fcp_buffer()
2054 lpfc_nvmeio_data(phba, in lpfc_nvmet_unsol_fcp_buffer()
2064 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_unsol_fcp_buffer()
2071 lpfc_nvmeio_data(phba, "NVMET FCP DROP: xri x%x sz %d from %06x\n", in lpfc_nvmet_unsol_fcp_buffer()
2074 lpfc_nvmet_defer_release(phba, ctxp); in lpfc_nvmet_unsol_fcp_buffer()
2075 lpfc_nvmet_unsol_fcp_issue_abort(phba, ctxp, sid, oxid); in lpfc_nvmet_unsol_fcp_buffer()
2076 lpfc_rq_buf_free(phba, &nvmebuf->hbuf); /* repost */ in lpfc_nvmet_unsol_fcp_buffer()
2081 lpfc_nvmet_ctxbuf_post(phba, ctx_buf); in lpfc_nvmet_unsol_fcp_buffer()
2084 lpfc_rq_buf_free(phba, &nvmebuf->hbuf); /* repost */ in lpfc_nvmet_unsol_fcp_buffer()
2100 lpfc_nvmet_unsol_ls_event(struct lpfc_hba *phba, struct lpfc_sli_ring *pring, in lpfc_nvmet_unsol_ls_event() argument
2109 if (phba->nvmet_support == 0) { in lpfc_nvmet_unsol_ls_event()
2110 lpfc_in_buf_free(phba, &nvmebuf->dbuf); in lpfc_nvmet_unsol_ls_event()
2113 lpfc_nvmet_unsol_ls_buffer(phba, pring, nvmebuf); in lpfc_nvmet_unsol_ls_event()
2129 lpfc_nvmet_unsol_fcp_event(struct lpfc_hba *phba, in lpfc_nvmet_unsol_fcp_event() argument
2134 if (phba->nvmet_support == 0) { in lpfc_nvmet_unsol_fcp_event()
2135 lpfc_rq_buf_free(phba, &nvmebuf->hbuf); in lpfc_nvmet_unsol_fcp_event()
2138 lpfc_nvmet_unsol_fcp_buffer(phba, idx, nvmebuf, in lpfc_nvmet_unsol_fcp_event()
2168 lpfc_nvmet_prep_ls_wqe(struct lpfc_hba *phba, in lpfc_nvmet_prep_ls_wqe() argument
2176 if (!lpfc_is_link_up(phba)) { in lpfc_nvmet_prep_ls_wqe()
2177 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_DISC, in lpfc_nvmet_prep_ls_wqe()
2185 nvmewqe = lpfc_sli_get_iocbq(phba); in lpfc_nvmet_prep_ls_wqe()
2187 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_DISC, in lpfc_nvmet_prep_ls_wqe()
2194 ndlp = lpfc_findnode_did(phba->pport, ctxp->sid); in lpfc_nvmet_prep_ls_wqe()
2198 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_DISC, in lpfc_nvmet_prep_ls_wqe()
2234 phba->sli4_hba.rpi_ids[ndlp->nlp_rpi]); in lpfc_nvmet_prep_ls_wqe()
2269 nvmewqe->vport = phba->pport; in lpfc_nvmet_prep_ls_wqe()
2270 nvmewqe->drvrTimeout = (phba->fc_ratov * 3) + LPFC_DRVR_TIMEOUT; in lpfc_nvmet_prep_ls_wqe()
2274 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_DISC, in lpfc_nvmet_prep_ls_wqe()
2284 lpfc_sli_release_iocbq(phba, nvmewqe); in lpfc_nvmet_prep_ls_wqe()
2290 lpfc_nvmet_prep_fcp_wqe(struct lpfc_hba *phba, in lpfc_nvmet_prep_fcp_wqe() argument
2307 if (!lpfc_is_link_up(phba)) { in lpfc_nvmet_prep_fcp_wqe()
2308 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_prep_fcp_wqe()
2315 ndlp = lpfc_findnode_did(phba->pport, ctxp->sid); in lpfc_nvmet_prep_fcp_wqe()
2319 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_prep_fcp_wqe()
2327 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_prep_fcp_wqe()
2331 phba->cfg_nvme_seg_cnt); in lpfc_nvmet_prep_fcp_wqe()
2335 tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private; in lpfc_nvmet_prep_fcp_wqe()
2341 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_prep_fcp_wqe()
2359 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_prep_fcp_wqe()
2394 phba->sli4_hba.rpi_ids[ndlp->nlp_rpi]); in lpfc_nvmet_prep_fcp_wqe()
2464 txrdy = dma_pool_alloc(phba->txrdy_payload_pool, in lpfc_nvmet_prep_fcp_wqe()
2467 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_prep_fcp_wqe()
2486 phba->sli4_hba.rpi_ids[ndlp->nlp_rpi]); in lpfc_nvmet_prep_fcp_wqe()
2504 if (phba->cfg_enable_pbde) { in lpfc_nvmet_prep_fcp_wqe()
2556 phba->sli4_hba.rpi_ids[ndlp->nlp_rpi]); in lpfc_nvmet_prep_fcp_wqe()
2595 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_IOERR, in lpfc_nvmet_prep_fcp_wqe()
2602 nvmewqe->vport = phba->pport; in lpfc_nvmet_prep_fcp_wqe()
2603 nvmewqe->drvrTimeout = (phba->fc_ratov * 3) + LPFC_DRVR_TIMEOUT; in lpfc_nvmet_prep_fcp_wqe()
2652 lpfc_nvmet_sol_fcp_abort_cmp(struct lpfc_hba *phba, struct lpfc_iocbq *cmdwqe, in lpfc_nvmet_sol_fcp_abort_cmp() argument
2665 tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private; in lpfc_nvmet_sol_fcp_abort_cmp()
2684 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS, in lpfc_nvmet_sol_fcp_abort_cmp()
2698 lpfc_nvmet_ctxbuf_post(phba, ctxp->ctxbuf); in lpfc_nvmet_sol_fcp_abort_cmp()
2701 lpfc_sli_release_iocbq(phba, cmdwqe); in lpfc_nvmet_sol_fcp_abort_cmp()
2720 lpfc_nvmet_unsol_fcp_abort_cmp(struct lpfc_hba *phba, struct lpfc_iocbq *cmdwqe, in lpfc_nvmet_unsol_fcp_abort_cmp() argument
2735 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS, in lpfc_nvmet_unsol_fcp_abort_cmp()
2742 tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private; in lpfc_nvmet_unsol_fcp_abort_cmp()
2748 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_ABTS, in lpfc_nvmet_unsol_fcp_abort_cmp()
2767 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS, in lpfc_nvmet_unsol_fcp_abort_cmp()
2781 lpfc_nvmet_ctxbuf_post(phba, ctxp->ctxbuf); in lpfc_nvmet_unsol_fcp_abort_cmp()
2800 lpfc_nvmet_xmt_ls_abort_cmp(struct lpfc_hba *phba, struct lpfc_iocbq *cmdwqe, in lpfc_nvmet_xmt_ls_abort_cmp() argument
2811 tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private; in lpfc_nvmet_xmt_ls_abort_cmp()
2814 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS, in lpfc_nvmet_xmt_ls_abort_cmp()
2820 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_ABTS, in lpfc_nvmet_xmt_ls_abort_cmp()
2826 lpfc_sli_release_iocbq(phba, cmdwqe); in lpfc_nvmet_xmt_ls_abort_cmp()
2831 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_xmt_ls_abort_cmp()
2839 lpfc_sli_release_iocbq(phba, cmdwqe); in lpfc_nvmet_xmt_ls_abort_cmp()
2844 lpfc_nvmet_unsol_issue_abort(struct lpfc_hba *phba, in lpfc_nvmet_unsol_issue_abort() argument
2853 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS, in lpfc_nvmet_unsol_issue_abort()
2857 tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private; in lpfc_nvmet_unsol_issue_abort()
2859 ndlp = lpfc_findnode_did(phba->pport, sid); in lpfc_nvmet_unsol_issue_abort()
2864 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_ABTS, in lpfc_nvmet_unsol_issue_abort()
2890 phba->sli4_hba.rpi_ids[ndlp->nlp_rpi]); in lpfc_nvmet_unsol_issue_abort()
2923 abts_wqeq->vport = phba->pport; in lpfc_nvmet_unsol_issue_abort()
2932 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS, in lpfc_nvmet_unsol_issue_abort()
2939 lpfc_nvmet_sol_fcp_issue_abort(struct lpfc_hba *phba, in lpfc_nvmet_sol_fcp_issue_abort() argument
2950 tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private; in lpfc_nvmet_sol_fcp_issue_abort()
2956 ndlp = lpfc_findnode_did(phba->pport, sid); in lpfc_nvmet_sol_fcp_issue_abort()
2961 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_ABTS, in lpfc_nvmet_sol_fcp_issue_abort()
2971 ctxp->abort_wqeq = lpfc_sli_get_iocbq(phba); in lpfc_nvmet_sol_fcp_issue_abort()
2974 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_ABTS, in lpfc_nvmet_sol_fcp_issue_abort()
2986 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS, in lpfc_nvmet_sol_fcp_issue_abort()
2994 spin_lock_irqsave(&phba->hbalock, flags); in lpfc_nvmet_sol_fcp_issue_abort()
2996 if (phba->hba_flag & HBA_NVME_IOQ_FLUSH) { in lpfc_nvmet_sol_fcp_issue_abort()
2997 spin_unlock_irqrestore(&phba->hbalock, flags); in lpfc_nvmet_sol_fcp_issue_abort()
2999 lpfc_printf_log(phba, KERN_ERR, LOG_NVME, in lpfc_nvmet_sol_fcp_issue_abort()
3002 phba->hba_flag, ctxp->oxid); in lpfc_nvmet_sol_fcp_issue_abort()
3003 lpfc_sli_release_iocbq(phba, abts_wqeq); in lpfc_nvmet_sol_fcp_issue_abort()
3010 spin_unlock_irqrestore(&phba->hbalock, flags); in lpfc_nvmet_sol_fcp_issue_abort()
3012 lpfc_printf_log(phba, KERN_ERR, LOG_NVME, in lpfc_nvmet_sol_fcp_issue_abort()
3016 lpfc_sli_release_iocbq(phba, abts_wqeq); in lpfc_nvmet_sol_fcp_issue_abort()
3060 abts_wqeq->vport = phba->pport; in lpfc_nvmet_sol_fcp_issue_abort()
3061 rc = lpfc_sli4_issue_wqe(phba, LPFC_FCP_RING, abts_wqeq); in lpfc_nvmet_sol_fcp_issue_abort()
3062 spin_unlock_irqrestore(&phba->hbalock, flags); in lpfc_nvmet_sol_fcp_issue_abort()
3070 lpfc_sli_release_iocbq(phba, abts_wqeq); in lpfc_nvmet_sol_fcp_issue_abort()
3071 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_ABTS, in lpfc_nvmet_sol_fcp_issue_abort()
3080 lpfc_nvmet_unsol_fcp_issue_abort(struct lpfc_hba *phba, in lpfc_nvmet_unsol_fcp_issue_abort() argument
3089 tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private; in lpfc_nvmet_unsol_fcp_issue_abort()
3096 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_unsol_fcp_issue_abort()
3104 rc = lpfc_nvmet_unsol_issue_abort(phba, ctxp, sid, xri); in lpfc_nvmet_unsol_fcp_issue_abort()
3108 spin_lock_irqsave(&phba->hbalock, flags); in lpfc_nvmet_unsol_fcp_issue_abort()
3113 rc = lpfc_sli4_issue_wqe(phba, LPFC_FCP_RING, abts_wqeq); in lpfc_nvmet_unsol_fcp_issue_abort()
3114 spin_unlock_irqrestore(&phba->hbalock, flags); in lpfc_nvmet_unsol_fcp_issue_abort()
3127 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_ABTS, in lpfc_nvmet_unsol_fcp_issue_abort()
3130 lpfc_nvmet_ctxbuf_post(phba, ctxp->ctxbuf); in lpfc_nvmet_unsol_fcp_issue_abort()
3135 lpfc_nvmet_unsol_ls_issue_abort(struct lpfc_hba *phba, in lpfc_nvmet_unsol_ls_issue_abort() argument
3150 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR, in lpfc_nvmet_unsol_ls_issue_abort()
3157 tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private; in lpfc_nvmet_unsol_ls_issue_abort()
3160 ctxp->wqeq = lpfc_sli_get_iocbq(phba); in lpfc_nvmet_unsol_ls_issue_abort()
3162 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_ABTS, in lpfc_nvmet_unsol_ls_issue_abort()
3173 if (lpfc_nvmet_unsol_issue_abort(phba, ctxp, sid, xri) == 0) { in lpfc_nvmet_unsol_ls_issue_abort()
3178 spin_lock_irqsave(&phba->hbalock, flags); in lpfc_nvmet_unsol_ls_issue_abort()
3182 rc = lpfc_sli4_issue_wqe(phba, LPFC_ELS_RING, abts_wqeq); in lpfc_nvmet_unsol_ls_issue_abort()
3183 spin_unlock_irqrestore(&phba->hbalock, flags); in lpfc_nvmet_unsol_ls_issue_abort()
3192 lpfc_sli_release_iocbq(phba, abts_wqeq); in lpfc_nvmet_unsol_ls_issue_abort()
3194 lpfc_printf_log(phba, KERN_ERR, LOG_NVME_ABTS, in lpfc_nvmet_unsol_ls_issue_abort()