Lines Matching refs:fod

139 	struct nvmet_fc_fcp_iod		*fod;		/* array of fcp_iods */  member
167 return (fodptr - fodptr->queue->fod); in nvmet_fc_fodnum()
237 struct nvmet_fc_fcp_iod *fod);
429 struct nvmet_fc_fcp_iod *fod = queue->fod; in nvmet_fc_prep_fcp_iodlist() local
432 for (i = 0; i < queue->sqsize; fod++, i++) { in nvmet_fc_prep_fcp_iodlist()
433 INIT_WORK(&fod->work, nvmet_fc_handle_fcp_rqst_work); in nvmet_fc_prep_fcp_iodlist()
434 INIT_WORK(&fod->done_work, nvmet_fc_fcp_rqst_op_done_work); in nvmet_fc_prep_fcp_iodlist()
435 INIT_WORK(&fod->defer_work, nvmet_fc_fcp_rqst_op_defer_work); in nvmet_fc_prep_fcp_iodlist()
436 fod->tgtport = tgtport; in nvmet_fc_prep_fcp_iodlist()
437 fod->queue = queue; in nvmet_fc_prep_fcp_iodlist()
438 fod->active = false; in nvmet_fc_prep_fcp_iodlist()
439 fod->abort = false; in nvmet_fc_prep_fcp_iodlist()
440 fod->aborted = false; in nvmet_fc_prep_fcp_iodlist()
441 fod->fcpreq = NULL; in nvmet_fc_prep_fcp_iodlist()
442 list_add_tail(&fod->fcp_list, &queue->fod_list); in nvmet_fc_prep_fcp_iodlist()
443 spin_lock_init(&fod->flock); in nvmet_fc_prep_fcp_iodlist()
445 fod->rspdma = fc_dma_map_single(tgtport->dev, &fod->rspiubuf, in nvmet_fc_prep_fcp_iodlist()
446 sizeof(fod->rspiubuf), DMA_TO_DEVICE); in nvmet_fc_prep_fcp_iodlist()
447 if (fc_dma_mapping_error(tgtport->dev, fod->rspdma)) { in nvmet_fc_prep_fcp_iodlist()
448 list_del(&fod->fcp_list); in nvmet_fc_prep_fcp_iodlist()
449 for (fod--, i--; i >= 0; fod--, i--) { in nvmet_fc_prep_fcp_iodlist()
450 fc_dma_unmap_single(tgtport->dev, fod->rspdma, in nvmet_fc_prep_fcp_iodlist()
451 sizeof(fod->rspiubuf), in nvmet_fc_prep_fcp_iodlist()
453 fod->rspdma = 0L; in nvmet_fc_prep_fcp_iodlist()
454 list_del(&fod->fcp_list); in nvmet_fc_prep_fcp_iodlist()
466 struct nvmet_fc_fcp_iod *fod = queue->fod; in nvmet_fc_destroy_fcp_iodlist() local
469 for (i = 0; i < queue->sqsize; fod++, i++) { in nvmet_fc_destroy_fcp_iodlist()
470 if (fod->rspdma) in nvmet_fc_destroy_fcp_iodlist()
471 fc_dma_unmap_single(tgtport->dev, fod->rspdma, in nvmet_fc_destroy_fcp_iodlist()
472 sizeof(fod->rspiubuf), DMA_TO_DEVICE); in nvmet_fc_destroy_fcp_iodlist()
479 struct nvmet_fc_fcp_iod *fod; in nvmet_fc_alloc_fcp_iod() local
483 fod = list_first_entry_or_null(&queue->fod_list, in nvmet_fc_alloc_fcp_iod()
485 if (fod) { in nvmet_fc_alloc_fcp_iod()
486 list_del(&fod->fcp_list); in nvmet_fc_alloc_fcp_iod()
487 fod->active = true; in nvmet_fc_alloc_fcp_iod()
494 return fod; in nvmet_fc_alloc_fcp_iod()
503 struct nvmet_fc_fcp_iod *fod = fcpreq->nvmet_fc_private; in nvmet_fc_queue_fcp_req() local
513 queue_work_on(queue->cpu, queue->work_q, &fod->work); in nvmet_fc_queue_fcp_req()
515 nvmet_fc_handle_fcp_rqst(tgtport, fod); in nvmet_fc_queue_fcp_req()
521 struct nvmet_fc_fcp_iod *fod = in nvmet_fc_fcp_rqst_op_defer_work() local
525 nvmet_fc_queue_fcp_req(fod->tgtport, fod->queue, fod->fcpreq); in nvmet_fc_fcp_rqst_op_defer_work()
531 struct nvmet_fc_fcp_iod *fod) in nvmet_fc_free_fcp_iod() argument
533 struct nvmefc_tgt_fcp_req *fcpreq = fod->fcpreq; in nvmet_fc_free_fcp_iod()
534 struct nvmet_fc_tgtport *tgtport = fod->tgtport; in nvmet_fc_free_fcp_iod()
538 fc_dma_sync_single_for_cpu(tgtport->dev, fod->rspdma, in nvmet_fc_free_fcp_iod()
539 sizeof(fod->rspiubuf), DMA_TO_DEVICE); in nvmet_fc_free_fcp_iod()
543 fod->active = false; in nvmet_fc_free_fcp_iod()
544 fod->abort = false; in nvmet_fc_free_fcp_iod()
545 fod->aborted = false; in nvmet_fc_free_fcp_iod()
546 fod->writedataactive = false; in nvmet_fc_free_fcp_iod()
547 fod->fcpreq = NULL; in nvmet_fc_free_fcp_iod()
558 list_add_tail(&fod->fcp_list, &fod->queue->fod_list); in nvmet_fc_free_fcp_iod()
574 memcpy(&fod->cmdiubuf, fcpreq->rspaddr, fcpreq->rsplen); in nvmet_fc_free_fcp_iod()
579 fcpreq->nvmet_fc_private = fod; in nvmet_fc_free_fcp_iod()
580 fod->fcpreq = fcpreq; in nvmet_fc_free_fcp_iod()
581 fod->active = true; in nvmet_fc_free_fcp_iod()
591 queue_work(queue->work_q, &fod->defer_work); in nvmet_fc_free_fcp_iod()
644 queue->fod = (struct nvmet_fc_fcp_iod *)&queue[1]; in nvmet_fc_alloc_target_queue()
721 struct nvmet_fc_fcp_iod *fod = queue->fod; in nvmet_fc_delete_target_queue() local
731 for (i = 0; i < queue->sqsize; fod++, i++) { in nvmet_fc_delete_target_queue()
732 if (fod->active) { in nvmet_fc_delete_target_queue()
733 spin_lock(&fod->flock); in nvmet_fc_delete_target_queue()
734 fod->abort = true; in nvmet_fc_delete_target_queue()
735 writedataactive = fod->writedataactive; in nvmet_fc_delete_target_queue()
736 spin_unlock(&fod->flock); in nvmet_fc_delete_target_queue()
743 spin_lock(&fod->flock); in nvmet_fc_delete_target_queue()
744 fod->aborted = true; in nvmet_fc_delete_target_queue()
745 spin_unlock(&fod->flock); in nvmet_fc_delete_target_queue()
747 &tgtport->fc_target_port, fod->fcpreq); in nvmet_fc_delete_target_queue()
1710 nvmet_fc_alloc_tgt_pgs(struct nvmet_fc_fcp_iod *fod) in nvmet_fc_alloc_tgt_pgs() argument
1715 sg = sgl_alloc(fod->req.transfer_len, GFP_KERNEL, &nent); in nvmet_fc_alloc_tgt_pgs()
1719 fod->data_sg = sg; in nvmet_fc_alloc_tgt_pgs()
1720 fod->data_sg_cnt = nent; in nvmet_fc_alloc_tgt_pgs()
1721 fod->data_sg_cnt = fc_dma_map_sg(fod->tgtport->dev, sg, nent, in nvmet_fc_alloc_tgt_pgs()
1722 ((fod->io_dir == NVMET_FCP_WRITE) ? in nvmet_fc_alloc_tgt_pgs()
1725 fod->next_sg = fod->data_sg; in nvmet_fc_alloc_tgt_pgs()
1734 nvmet_fc_free_tgt_pgs(struct nvmet_fc_fcp_iod *fod) in nvmet_fc_free_tgt_pgs() argument
1736 if (!fod->data_sg || !fod->data_sg_cnt) in nvmet_fc_free_tgt_pgs()
1739 fc_dma_unmap_sg(fod->tgtport->dev, fod->data_sg, fod->data_sg_cnt, in nvmet_fc_free_tgt_pgs()
1740 ((fod->io_dir == NVMET_FCP_WRITE) ? in nvmet_fc_free_tgt_pgs()
1742 sgl_free(fod->data_sg); in nvmet_fc_free_tgt_pgs()
1743 fod->data_sg = NULL; in nvmet_fc_free_tgt_pgs()
1744 fod->data_sg_cnt = 0; in nvmet_fc_free_tgt_pgs()
1766 struct nvmet_fc_fcp_iod *fod) in nvmet_fc_prep_fcp_rsp() argument
1768 struct nvme_fc_ersp_iu *ersp = &fod->rspiubuf; in nvmet_fc_prep_fcp_rsp()
1769 struct nvme_common_command *sqe = &fod->cmdiubuf.sqe.common; in nvmet_fc_prep_fcp_rsp()
1775 if (fod->fcpreq->op == NVMET_FCOP_READDATA_RSP) in nvmet_fc_prep_fcp_rsp()
1776 xfr_length = fod->req.transfer_len; in nvmet_fc_prep_fcp_rsp()
1778 xfr_length = fod->offset; in nvmet_fc_prep_fcp_rsp()
1799 rspcnt = atomic_inc_return(&fod->queue->zrspcnt); in nvmet_fc_prep_fcp_rsp()
1800 if (!(rspcnt % fod->queue->ersp_ratio) || in nvmet_fc_prep_fcp_rsp()
1802 xfr_length != fod->req.transfer_len || in nvmet_fc_prep_fcp_rsp()
1805 queue_90percent_full(fod->queue, le16_to_cpu(cqe->sq_head))) in nvmet_fc_prep_fcp_rsp()
1809 fod->fcpreq->rspaddr = ersp; in nvmet_fc_prep_fcp_rsp()
1810 fod->fcpreq->rspdma = fod->rspdma; in nvmet_fc_prep_fcp_rsp()
1814 fod->fcpreq->rsplen = NVME_FC_SIZEOF_ZEROS_RSP; in nvmet_fc_prep_fcp_rsp()
1817 rsn = atomic_inc_return(&fod->queue->rsn); in nvmet_fc_prep_fcp_rsp()
1820 fod->fcpreq->rsplen = sizeof(*ersp); in nvmet_fc_prep_fcp_rsp()
1823 fc_dma_sync_single_for_device(tgtport->dev, fod->rspdma, in nvmet_fc_prep_fcp_rsp()
1824 sizeof(fod->rspiubuf), DMA_TO_DEVICE); in nvmet_fc_prep_fcp_rsp()
1831 struct nvmet_fc_fcp_iod *fod) in nvmet_fc_abort_op() argument
1833 struct nvmefc_tgt_fcp_req *fcpreq = fod->fcpreq; in nvmet_fc_abort_op()
1836 nvmet_fc_free_tgt_pgs(fod); in nvmet_fc_abort_op()
1843 if (!fod->aborted) in nvmet_fc_abort_op()
1846 nvmet_fc_free_fcp_iod(fod->queue, fod); in nvmet_fc_abort_op()
1851 struct nvmet_fc_fcp_iod *fod) in nvmet_fc_xmt_fcp_rsp() argument
1855 fod->fcpreq->op = NVMET_FCOP_RSP; in nvmet_fc_xmt_fcp_rsp()
1856 fod->fcpreq->timeout = 0; in nvmet_fc_xmt_fcp_rsp()
1858 nvmet_fc_prep_fcp_rsp(tgtport, fod); in nvmet_fc_xmt_fcp_rsp()
1860 ret = tgtport->ops->fcp_op(&tgtport->fc_target_port, fod->fcpreq); in nvmet_fc_xmt_fcp_rsp()
1862 nvmet_fc_abort_op(tgtport, fod); in nvmet_fc_xmt_fcp_rsp()
1867 struct nvmet_fc_fcp_iod *fod, u8 op) in nvmet_fc_transfer_fcp_data() argument
1869 struct nvmefc_tgt_fcp_req *fcpreq = fod->fcpreq; in nvmet_fc_transfer_fcp_data()
1870 struct scatterlist *sg = fod->next_sg; in nvmet_fc_transfer_fcp_data()
1872 u32 remaininglen = fod->req.transfer_len - fod->offset; in nvmet_fc_transfer_fcp_data()
1877 fcpreq->offset = fod->offset; in nvmet_fc_transfer_fcp_data()
1904 fod->next_sg = sg; in nvmet_fc_transfer_fcp_data()
1906 fod->next_sg = NULL; in nvmet_fc_transfer_fcp_data()
1918 ((fod->offset + fcpreq->transfer_length) == fod->req.transfer_len) && in nvmet_fc_transfer_fcp_data()
1921 nvmet_fc_prep_fcp_rsp(tgtport, fod); in nvmet_fc_transfer_fcp_data()
1924 ret = tgtport->ops->fcp_op(&tgtport->fc_target_port, fod->fcpreq); in nvmet_fc_transfer_fcp_data()
1931 fod->abort = true; in nvmet_fc_transfer_fcp_data()
1934 spin_lock_irqsave(&fod->flock, flags); in nvmet_fc_transfer_fcp_data()
1935 fod->writedataactive = false; in nvmet_fc_transfer_fcp_data()
1936 spin_unlock_irqrestore(&fod->flock, flags); in nvmet_fc_transfer_fcp_data()
1937 nvmet_req_complete(&fod->req, NVME_SC_INTERNAL); in nvmet_fc_transfer_fcp_data()
1941 nvmet_fc_xmt_fcp_op_done(fod->fcpreq); in nvmet_fc_transfer_fcp_data()
1947 __nvmet_fc_fod_op_abort(struct nvmet_fc_fcp_iod *fod, bool abort) in __nvmet_fc_fod_op_abort() argument
1949 struct nvmefc_tgt_fcp_req *fcpreq = fod->fcpreq; in __nvmet_fc_fod_op_abort()
1950 struct nvmet_fc_tgtport *tgtport = fod->tgtport; in __nvmet_fc_fod_op_abort()
1955 nvmet_req_complete(&fod->req, NVME_SC_INTERNAL); in __nvmet_fc_fod_op_abort()
1959 nvmet_fc_abort_op(tgtport, fod); in __nvmet_fc_fod_op_abort()
1970 nvmet_fc_fod_op_done(struct nvmet_fc_fcp_iod *fod) in nvmet_fc_fod_op_done() argument
1972 struct nvmefc_tgt_fcp_req *fcpreq = fod->fcpreq; in nvmet_fc_fod_op_done()
1973 struct nvmet_fc_tgtport *tgtport = fod->tgtport; in nvmet_fc_fod_op_done()
1977 spin_lock_irqsave(&fod->flock, flags); in nvmet_fc_fod_op_done()
1978 abort = fod->abort; in nvmet_fc_fod_op_done()
1979 fod->writedataactive = false; in nvmet_fc_fod_op_done()
1980 spin_unlock_irqrestore(&fod->flock, flags); in nvmet_fc_fod_op_done()
1985 if (__nvmet_fc_fod_op_abort(fod, abort)) in nvmet_fc_fod_op_done()
1989 spin_lock(&fod->flock); in nvmet_fc_fod_op_done()
1990 fod->abort = true; in nvmet_fc_fod_op_done()
1991 spin_unlock(&fod->flock); in nvmet_fc_fod_op_done()
1993 nvmet_req_complete(&fod->req, NVME_SC_INTERNAL); in nvmet_fc_fod_op_done()
1997 fod->offset += fcpreq->transferred_length; in nvmet_fc_fod_op_done()
1998 if (fod->offset != fod->req.transfer_len) { in nvmet_fc_fod_op_done()
1999 spin_lock_irqsave(&fod->flock, flags); in nvmet_fc_fod_op_done()
2000 fod->writedataactive = true; in nvmet_fc_fod_op_done()
2001 spin_unlock_irqrestore(&fod->flock, flags); in nvmet_fc_fod_op_done()
2004 nvmet_fc_transfer_fcp_data(tgtport, fod, in nvmet_fc_fod_op_done()
2010 nvmet_req_execute(&fod->req); in nvmet_fc_fod_op_done()
2015 if (__nvmet_fc_fod_op_abort(fod, abort)) in nvmet_fc_fod_op_done()
2019 nvmet_fc_abort_op(tgtport, fod); in nvmet_fc_fod_op_done()
2027 nvmet_fc_free_tgt_pgs(fod); in nvmet_fc_fod_op_done()
2028 nvmet_fc_free_fcp_iod(fod->queue, fod); in nvmet_fc_fod_op_done()
2032 fod->offset += fcpreq->transferred_length; in nvmet_fc_fod_op_done()
2033 if (fod->offset != fod->req.transfer_len) { in nvmet_fc_fod_op_done()
2035 nvmet_fc_transfer_fcp_data(tgtport, fod, in nvmet_fc_fod_op_done()
2043 nvmet_fc_free_tgt_pgs(fod); in nvmet_fc_fod_op_done()
2045 nvmet_fc_xmt_fcp_rsp(tgtport, fod); in nvmet_fc_fod_op_done()
2050 if (__nvmet_fc_fod_op_abort(fod, abort)) in nvmet_fc_fod_op_done()
2052 nvmet_fc_free_fcp_iod(fod->queue, fod); in nvmet_fc_fod_op_done()
2063 struct nvmet_fc_fcp_iod *fod = in nvmet_fc_fcp_rqst_op_done_work() local
2066 nvmet_fc_fod_op_done(fod); in nvmet_fc_fcp_rqst_op_done_work()
2072 struct nvmet_fc_fcp_iod *fod = fcpreq->nvmet_fc_private; in nvmet_fc_xmt_fcp_op_done() local
2073 struct nvmet_fc_tgt_queue *queue = fod->queue; in nvmet_fc_xmt_fcp_op_done()
2075 if (fod->tgtport->ops->target_features & NVMET_FCTGTFEAT_OPDONE_IN_ISR) in nvmet_fc_xmt_fcp_op_done()
2077 queue_work_on(queue->cpu, queue->work_q, &fod->done_work); in nvmet_fc_xmt_fcp_op_done()
2079 nvmet_fc_fod_op_done(fod); in nvmet_fc_xmt_fcp_op_done()
2087 struct nvmet_fc_fcp_iod *fod, int status) in __nvmet_fc_fcp_nvme_cmd_done() argument
2089 struct nvme_common_command *sqe = &fod->cmdiubuf.sqe.common; in __nvmet_fc_fcp_nvme_cmd_done()
2090 struct nvme_completion *cqe = &fod->rspiubuf.cqe; in __nvmet_fc_fcp_nvme_cmd_done()
2094 spin_lock_irqsave(&fod->flock, flags); in __nvmet_fc_fcp_nvme_cmd_done()
2095 abort = fod->abort; in __nvmet_fc_fcp_nvme_cmd_done()
2096 spin_unlock_irqrestore(&fod->flock, flags); in __nvmet_fc_fcp_nvme_cmd_done()
2100 fod->queue->sqhd = cqe->sq_head; in __nvmet_fc_fcp_nvme_cmd_done()
2103 nvmet_fc_abort_op(tgtport, fod); in __nvmet_fc_fcp_nvme_cmd_done()
2111 cqe->sq_head = fod->queue->sqhd; /* echo last cqe sqhd */ in __nvmet_fc_fcp_nvme_cmd_done()
2112 cqe->sq_id = cpu_to_le16(fod->queue->qid); in __nvmet_fc_fcp_nvme_cmd_done()
2122 if ((fod->io_dir == NVMET_FCP_READ) && (fod->data_sg_cnt)) { in __nvmet_fc_fcp_nvme_cmd_done()
2124 nvmet_fc_transfer_fcp_data(tgtport, fod, in __nvmet_fc_fcp_nvme_cmd_done()
2133 nvmet_fc_free_tgt_pgs(fod); in __nvmet_fc_fcp_nvme_cmd_done()
2135 nvmet_fc_xmt_fcp_rsp(tgtport, fod); in __nvmet_fc_fcp_nvme_cmd_done()
2142 struct nvmet_fc_fcp_iod *fod = nvmet_req_to_fod(nvme_req); in nvmet_fc_fcp_nvme_cmd_done() local
2143 struct nvmet_fc_tgtport *tgtport = fod->tgtport; in nvmet_fc_fcp_nvme_cmd_done()
2145 __nvmet_fc_fcp_nvme_cmd_done(tgtport, fod, 0); in nvmet_fc_fcp_nvme_cmd_done()
2154 struct nvmet_fc_fcp_iod *fod) in nvmet_fc_handle_fcp_rqst() argument
2156 struct nvme_fc_cmd_iu *cmdiu = &fod->cmdiubuf; in nvmet_fc_handle_fcp_rqst()
2169 fod->fcpreq->done = nvmet_fc_xmt_fcp_op_done; in nvmet_fc_handle_fcp_rqst()
2172 fod->io_dir = NVMET_FCP_WRITE; in nvmet_fc_handle_fcp_rqst()
2176 fod->io_dir = NVMET_FCP_READ; in nvmet_fc_handle_fcp_rqst()
2180 fod->io_dir = NVMET_FCP_NODATA; in nvmet_fc_handle_fcp_rqst()
2185 fod->req.cmd = &fod->cmdiubuf.sqe; in nvmet_fc_handle_fcp_rqst()
2186 fod->req.rsp = &fod->rspiubuf.cqe; in nvmet_fc_handle_fcp_rqst()
2187 fod->req.port = fod->queue->port; in nvmet_fc_handle_fcp_rqst()
2190 memset(&fod->rspiubuf, 0, sizeof(fod->rspiubuf)); in nvmet_fc_handle_fcp_rqst()
2192 fod->data_sg = NULL; in nvmet_fc_handle_fcp_rqst()
2193 fod->data_sg_cnt = 0; in nvmet_fc_handle_fcp_rqst()
2195 ret = nvmet_req_init(&fod->req, in nvmet_fc_handle_fcp_rqst()
2196 &fod->queue->nvme_cq, in nvmet_fc_handle_fcp_rqst()
2197 &fod->queue->nvme_sq, in nvmet_fc_handle_fcp_rqst()
2205 fod->req.transfer_len = xfrlen; in nvmet_fc_handle_fcp_rqst()
2208 atomic_inc(&fod->queue->sqtail); in nvmet_fc_handle_fcp_rqst()
2210 if (fod->req.transfer_len) { in nvmet_fc_handle_fcp_rqst()
2211 ret = nvmet_fc_alloc_tgt_pgs(fod); in nvmet_fc_handle_fcp_rqst()
2213 nvmet_req_complete(&fod->req, ret); in nvmet_fc_handle_fcp_rqst()
2217 fod->req.sg = fod->data_sg; in nvmet_fc_handle_fcp_rqst()
2218 fod->req.sg_cnt = fod->data_sg_cnt; in nvmet_fc_handle_fcp_rqst()
2219 fod->offset = 0; in nvmet_fc_handle_fcp_rqst()
2221 if (fod->io_dir == NVMET_FCP_WRITE) { in nvmet_fc_handle_fcp_rqst()
2223 nvmet_fc_transfer_fcp_data(tgtport, fod, NVMET_FCOP_WRITEDATA); in nvmet_fc_handle_fcp_rqst()
2233 nvmet_req_execute(&fod->req); in nvmet_fc_handle_fcp_rqst()
2237 nvmet_fc_abort_op(tgtport, fod); in nvmet_fc_handle_fcp_rqst()
2246 struct nvmet_fc_fcp_iod *fod = in nvmet_fc_handle_fcp_rqst_work() local
2248 struct nvmet_fc_tgtport *tgtport = fod->tgtport; in nvmet_fc_handle_fcp_rqst_work()
2250 nvmet_fc_handle_fcp_rqst(tgtport, fod); in nvmet_fc_handle_fcp_rqst_work()
2308 struct nvmet_fc_fcp_iod *fod; in nvmet_fc_rcv_fcp_req() local
2333 fod = nvmet_fc_alloc_fcp_iod(queue); in nvmet_fc_rcv_fcp_req()
2334 if (fod) { in nvmet_fc_rcv_fcp_req()
2337 fcpreq->nvmet_fc_private = fod; in nvmet_fc_rcv_fcp_req()
2338 fod->fcpreq = fcpreq; in nvmet_fc_rcv_fcp_req()
2340 memcpy(&fod->cmdiubuf, cmdiubuf, cmdiubuf_len); in nvmet_fc_rcv_fcp_req()
2415 struct nvmet_fc_fcp_iod *fod = fcpreq->nvmet_fc_private; in nvmet_fc_rcv_fcp_abort() local
2419 if (!fod || fod->fcpreq != fcpreq) in nvmet_fc_rcv_fcp_abort()
2423 queue = fod->queue; in nvmet_fc_rcv_fcp_abort()
2426 if (fod->active) { in nvmet_fc_rcv_fcp_abort()
2432 spin_lock(&fod->flock); in nvmet_fc_rcv_fcp_abort()
2433 fod->abort = true; in nvmet_fc_rcv_fcp_abort()
2434 fod->aborted = true; in nvmet_fc_rcv_fcp_abort()
2435 spin_unlock(&fod->flock); in nvmet_fc_rcv_fcp_abort()