Lines Matching refs:rqd
599 static void nvm_rq_tgt_to_dev(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd) in nvm_rq_tgt_to_dev() argument
601 if (rqd->nr_ppas == 1) { in nvm_rq_tgt_to_dev()
602 nvm_ppa_tgt_to_dev(tgt_dev, &rqd->ppa_addr, 1); in nvm_rq_tgt_to_dev()
606 nvm_ppa_tgt_to_dev(tgt_dev, rqd->ppa_list, rqd->nr_ppas); in nvm_rq_tgt_to_dev()
609 static void nvm_rq_dev_to_tgt(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd) in nvm_rq_dev_to_tgt() argument
611 if (rqd->nr_ppas == 1) { in nvm_rq_dev_to_tgt()
612 nvm_ppa_dev_to_tgt(tgt_dev, &rqd->ppa_addr, 1); in nvm_rq_dev_to_tgt()
616 nvm_ppa_dev_to_tgt(tgt_dev, rqd->ppa_list, rqd->nr_ppas); in nvm_rq_dev_to_tgt()
670 static int nvm_set_rqd_ppalist(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd, in nvm_set_rqd_ppalist() argument
679 rqd->nr_ppas = nr_ppas; in nvm_set_rqd_ppalist()
680 rqd->ppa_addr = ppas[0]; in nvm_set_rqd_ppalist()
685 rqd->nr_ppas = nr_ppas; in nvm_set_rqd_ppalist()
686 rqd->ppa_list = nvm_dev_dma_alloc(dev, GFP_KERNEL, &rqd->dma_ppa_list); in nvm_set_rqd_ppalist()
687 if (!rqd->ppa_list) { in nvm_set_rqd_ppalist()
693 rqd->nr_ppas *= plane_cnt; in nvm_set_rqd_ppalist()
699 rqd->ppa_list[(pl_idx * nr_ppas) + i] = ppa; in nvm_set_rqd_ppalist()
707 struct nvm_rq *rqd) in nvm_free_rqd_ppalist() argument
709 if (!rqd->ppa_list) in nvm_free_rqd_ppalist()
712 nvm_dev_dma_free(tgt_dev->parent, rqd->ppa_list, rqd->dma_ppa_list); in nvm_free_rqd_ppalist()
731 struct nvm_rq rqd; in nvm_set_tgt_bb_tbl() local
739 memset(&rqd, 0, sizeof(struct nvm_rq)); in nvm_set_tgt_bb_tbl()
741 nvm_set_rqd_ppalist(tgt_dev, &rqd, ppas, nr_ppas); in nvm_set_tgt_bb_tbl()
742 nvm_rq_tgt_to_dev(tgt_dev, &rqd); in nvm_set_tgt_bb_tbl()
744 ret = dev->ops->set_bb_tbl(dev, &rqd.ppa_addr, rqd.nr_ppas, type); in nvm_set_tgt_bb_tbl()
745 nvm_free_rqd_ppalist(tgt_dev, &rqd); in nvm_set_tgt_bb_tbl()
755 int nvm_submit_io(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd) in nvm_submit_io() argument
763 nvm_rq_tgt_to_dev(tgt_dev, rqd); in nvm_submit_io()
765 rqd->dev = tgt_dev; in nvm_submit_io()
768 ret = dev->ops->submit_io(dev, rqd); in nvm_submit_io()
770 nvm_rq_dev_to_tgt(tgt_dev, rqd); in nvm_submit_io()
775 int nvm_submit_io_sync(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd) in nvm_submit_io_sync() argument
783 nvm_rq_tgt_to_dev(tgt_dev, rqd); in nvm_submit_io_sync()
785 rqd->dev = tgt_dev; in nvm_submit_io_sync()
788 ret = dev->ops->submit_io_sync(dev, rqd); in nvm_submit_io_sync()
789 nvm_rq_dev_to_tgt(tgt_dev, rqd); in nvm_submit_io_sync()
795 void nvm_end_io(struct nvm_rq *rqd) in nvm_end_io() argument
797 struct nvm_tgt_dev *tgt_dev = rqd->dev; in nvm_end_io()
801 nvm_rq_dev_to_tgt(tgt_dev, rqd); in nvm_end_io()
803 if (rqd->end_io) in nvm_end_io()
804 rqd->end_io(rqd); in nvm_end_io()