Lines Matching refs:n_sg
173 u32 n_sg; member
515 skreq->n_sg = 0; in skd_mq_queue_rq()
530 skreq->n_sg * in skd_mq_queue_rq()
624 int n_sg; in skd_preop_sg_list() local
632 n_sg = blk_rq_map_sg(skdev->queue, req, sgl); in skd_preop_sg_list()
633 if (n_sg <= 0) in skd_preop_sg_list()
640 n_sg = dma_map_sg(&skdev->pdev->dev, sgl, n_sg, skreq->data_dir); in skd_preop_sg_list()
641 if (n_sg <= 0) in skd_preop_sg_list()
644 SKD_ASSERT(n_sg <= skdev->sgs_per_request); in skd_preop_sg_list()
646 skreq->n_sg = n_sg; in skd_preop_sg_list()
648 for_each_sg(sgl, sg, n_sg, i) { in skd_preop_sg_list()
660 skreq->sksg_list[n_sg - 1].next_desc_ptr = 0LL; in skd_preop_sg_list()
661 skreq->sksg_list[n_sg - 1].control = FIT_SGD_CONTROL_LAST; in skd_preop_sg_list()
667 for (i = 0; i < n_sg; i++) { in skd_preop_sg_list()
687 skreq->sksg_list[skreq->n_sg - 1].next_desc_ptr = in skd_postop_sg_list()
689 ((skreq->n_sg) * sizeof(struct fit_sg_descriptor)); in skd_postop_sg_list()
690 dma_unmap_sg(&skdev->pdev->dev, &skreq->sg[0], skreq->n_sg, in skd_postop_sg_list()
900 skspcl->req.n_sg = 1; in skd_format_internal_skspcl()
1245 WARN_ON_ONCE(skspcl->req.n_sg != 1); in skd_send_special_fitmsg()
1262 for (i = 0; i < skspcl->req.n_sg; i++) { in skd_send_special_fitmsg()
1553 if (skreq->n_sg > 0) in skd_isr_completion_posted()
1906 if (skreq->n_sg > 0) in skd_recover_request()
2704 u32 n_sg, in skd_cons_sg_list() argument
2716 for (i = 0; i < n_sg - 1; i++) { in skd_cons_sg_list()
3589 skreq->data_dir, skreq->n_sg); in skd_log_skreq()