/Linux-v4.19/drivers/misc/genwqe/ |
D | card_utils.c | 298 int genwqe_alloc_sync_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl, in genwqe_alloc_sync_sgl() argument 304 sgl->fpage_offs = offset_in_page((unsigned long)user_addr); in genwqe_alloc_sync_sgl() 305 sgl->fpage_size = min_t(size_t, PAGE_SIZE-sgl->fpage_offs, user_size); in genwqe_alloc_sync_sgl() 306 sgl->nr_pages = DIV_ROUND_UP(sgl->fpage_offs + user_size, PAGE_SIZE); in genwqe_alloc_sync_sgl() 307 sgl->lpage_size = (user_size - sgl->fpage_size) % PAGE_SIZE; in genwqe_alloc_sync_sgl() 310 __func__, user_addr, user_size, sgl->nr_pages, in genwqe_alloc_sync_sgl() 311 sgl->fpage_offs, sgl->fpage_size, sgl->lpage_size); in genwqe_alloc_sync_sgl() 313 sgl->user_addr = user_addr; in genwqe_alloc_sync_sgl() 314 sgl->user_size = user_size; in genwqe_alloc_sync_sgl() 315 sgl->write = write; in genwqe_alloc_sync_sgl() [all …]
|
/Linux-v4.19/lib/ |
D | scatterlist.c | 103 struct scatterlist *sg_last(struct scatterlist *sgl, unsigned int nents) in sg_last() argument 108 for_each_sg(sgl, sg, nents, i) in sg_last() 126 void sg_init_table(struct scatterlist *sgl, unsigned int nents) in sg_init_table() argument 128 memset(sgl, 0, sizeof(*sgl) * nents); in sg_init_table() 129 sg_init_marker(sgl, nents); in sg_init_table() 196 struct scatterlist *sgl, *next; in __sg_free_table() local 198 if (unlikely(!table->sgl)) in __sg_free_table() 201 sgl = table->sgl; in __sg_free_table() 213 next = sg_chain_ptr(&sgl[max_ents - 1]); in __sg_free_table() 225 free_fn(sgl, alloc_size); in __sg_free_table() [all …]
|
D | kfifo.c | 308 static int setup_sgl_buf(struct scatterlist *sgl, void *buf, in setup_sgl_buf() argument 334 sg_set_page(sgl, page, l - off, off); in setup_sgl_buf() 335 sgl = sg_next(sgl); in setup_sgl_buf() 336 if (++n == nents || sgl == NULL) in setup_sgl_buf() 343 sg_set_page(sgl, page, len, off); in setup_sgl_buf() 347 static unsigned int setup_sgl(struct __kfifo *fifo, struct scatterlist *sgl, in setup_sgl() argument 363 n = setup_sgl_buf(sgl, fifo->data + off, nents, l); in setup_sgl() 364 n += setup_sgl_buf(sgl + n, fifo->data, nents - n, len - l); in setup_sgl() 370 struct scatterlist *sgl, int nents, unsigned int len) in __kfifo_dma_in_prepare() argument 378 return setup_sgl(fifo, sgl, nents, len, fifo->in); in __kfifo_dma_in_prepare() [all …]
|
/Linux-v4.19/crypto/ |
D | af_alg.c | 397 int af_alg_make_sg(struct af_alg_sgl *sgl, struct iov_iter *iter, int len) in af_alg_make_sg() argument 403 n = iov_iter_get_pages(iter, sgl->pages, len, ALG_MAX_PAGES, &off); in af_alg_make_sg() 411 sg_init_table(sgl->sg, npages + 1); in af_alg_make_sg() 416 sg_set_page(sgl->sg + i, sgl->pages[i], plen, off); in af_alg_make_sg() 421 sg_mark_end(sgl->sg + npages - 1); in af_alg_make_sg() 422 sgl->npages = npages; in af_alg_make_sg() 435 void af_alg_free_sg(struct af_alg_sgl *sgl) in af_alg_free_sg() argument 439 for (i = 0; i < sgl->npages; i++) in af_alg_free_sg() 440 put_page(sgl->pages[i]); in af_alg_free_sg() 495 struct af_alg_tsgl *sgl; in af_alg_alloc_tsgl() local [all …]
|
/Linux-v4.19/include/linux/ |
D | scatterlist.h | 43 struct scatterlist *sgl; /* the list */ member 165 struct scatterlist *sgl) in sg_chain() argument 177 prv[prv_nents - 1].page_link = ((unsigned long) sgl | SG_CHAIN) in sg_chain() 248 static inline void sg_init_marker(struct scatterlist *sgl, in sg_init_marker() argument 251 sg_mark_end(&sgl[nents - 1]); in sg_init_marker() 288 void sgl_free_n_order(struct scatterlist *sgl, int nents, int order); 289 void sgl_free_order(struct scatterlist *sgl, int order); 290 void sgl_free(struct scatterlist *sgl); 293 size_t sg_copy_buffer(struct scatterlist *sgl, unsigned int nents, void *buf, 296 size_t sg_copy_from_buffer(struct scatterlist *sgl, unsigned int nents, [all …]
|
/Linux-v4.19/kernel/dma/ |
D | noncoherent.c | 21 struct scatterlist *sgl, int nents, enum dma_data_direction dir) in dma_noncoherent_sync_sg_for_device() argument 26 for_each_sg(sgl, sg, nents, i) in dma_noncoherent_sync_sg_for_device() 43 static int dma_noncoherent_map_sg(struct device *dev, struct scatterlist *sgl, in dma_noncoherent_map_sg() argument 46 nents = dma_direct_map_sg(dev, sgl, nents, dir, attrs); in dma_noncoherent_map_sg() 48 dma_noncoherent_sync_sg_for_device(dev, sgl, nents, dir); in dma_noncoherent_map_sg() 62 struct scatterlist *sgl, int nents, enum dma_data_direction dir) in dma_noncoherent_sync_sg_for_cpu() argument 67 for_each_sg(sgl, sg, nents, i) in dma_noncoherent_sync_sg_for_cpu() 79 static void dma_noncoherent_unmap_sg(struct device *dev, struct scatterlist *sgl, in dma_noncoherent_unmap_sg() argument 83 dma_noncoherent_sync_sg_for_cpu(dev, sgl, nents, dir); in dma_noncoherent_unmap_sg()
|
/Linux-v4.19/drivers/target/iscsi/cxgbit/ |
D | cxgbit_ddp.c | 136 struct scatterlist *sg = ttinfo->sgl; in cxgbit_ddp_set_map() 179 struct scatterlist *sgl = ttinfo->sgl; in cxgbit_ddp_reserve() local 181 unsigned int sg_offset = sgl->offset; in cxgbit_ddp_reserve() 191 if (cxgbit_ddp_sgl_check(sgl, sgcnt) < 0) in cxgbit_ddp_reserve() 194 ttinfo->nr_pages = (xferlen + sgl->offset + in cxgbit_ddp_reserve() 206 sgl->offset = 0; in cxgbit_ddp_reserve() 207 ret = dma_map_sg(&ppm->pdev->dev, sgl, sgcnt, DMA_FROM_DEVICE); in cxgbit_ddp_reserve() 208 sgl->offset = sg_offset; in cxgbit_ddp_reserve() 215 cxgbi_ppm_make_ppod_hdr(ppm, ttinfo->tag, csk->tid, sgl->offset, in cxgbit_ddp_reserve() 221 dma_unmap_sg(&ppm->pdev->dev, sgl, sgcnt, DMA_FROM_DEVICE); in cxgbit_ddp_reserve() [all …]
|
/Linux-v4.19/tools/virtio/linux/ |
D | scatterlist.h | 91 struct scatterlist *sgl) in sg_chain() argument 103 prv[prv_nents - 1].page_link = ((unsigned long) sgl | 0x01) & ~0x02; in sg_chain() 149 static inline void sg_init_table(struct scatterlist *sgl, unsigned int nents) in sg_init_table() argument 151 memset(sgl, 0, sizeof(*sgl) * nents); in sg_init_table() 152 sg_mark_end(&sgl[nents - 1]); in sg_init_table()
|
/Linux-v4.19/drivers/gpu/drm/v3d/ |
D | v3d_mmu.c | 90 struct scatterlist *sgl; in v3d_mmu_insert_ptes() local 92 for_each_sg(bo->sgt->sgl, sgl, bo->sgt->nents, count) { in v3d_mmu_insert_ptes() 93 u32 page_address = sg_dma_address(sgl) >> V3D_MMU_PAGE_SHIFT; in v3d_mmu_insert_ptes() 97 BUG_ON(page_address + (sg_dma_len(sgl) >> V3D_MMU_PAGE_SHIFT) >= in v3d_mmu_insert_ptes() 100 for (i = 0; i < sg_dma_len(sgl) >> V3D_MMU_PAGE_SHIFT; i++) in v3d_mmu_insert_ptes()
|
/Linux-v4.19/arch/arm64/mm/ |
D | dma-mapping.c | 183 static int __swiotlb_map_sg_attrs(struct device *dev, struct scatterlist *sgl, in __swiotlb_map_sg_attrs() argument 190 ret = swiotlb_map_sg_attrs(dev, sgl, nelems, dir, attrs); in __swiotlb_map_sg_attrs() 193 for_each_sg(sgl, sg, ret, i) in __swiotlb_map_sg_attrs() 201 struct scatterlist *sgl, int nelems, in __swiotlb_unmap_sg_attrs() argument 210 for_each_sg(sgl, sg, nelems, i) in __swiotlb_unmap_sg_attrs() 213 swiotlb_unmap_sg_attrs(dev, sgl, nelems, dir, attrs); in __swiotlb_unmap_sg_attrs() 235 struct scatterlist *sgl, int nelems, in __swiotlb_sync_sg_for_cpu() argument 242 for_each_sg(sgl, sg, nelems, i) in __swiotlb_sync_sg_for_cpu() 245 swiotlb_sync_sg_for_cpu(dev, sgl, nelems, dir); in __swiotlb_sync_sg_for_cpu() 249 struct scatterlist *sgl, int nelems, in __swiotlb_sync_sg_for_device() argument [all …]
|
/Linux-v4.19/arch/arm/mm/ |
D | dma-mapping-nommu.c | 137 static int arm_nommu_dma_map_sg(struct device *dev, struct scatterlist *sgl, in arm_nommu_dma_map_sg() argument 144 for_each_sg(sgl, sg, nents, i) { in arm_nommu_dma_map_sg() 153 static void arm_nommu_dma_unmap_sg(struct device *dev, struct scatterlist *sgl, in arm_nommu_dma_unmap_sg() argument 160 for_each_sg(sgl, sg, nents, i) in arm_nommu_dma_unmap_sg() 176 static void arm_nommu_dma_sync_sg_for_device(struct device *dev, struct scatterlist *sgl, in arm_nommu_dma_sync_sg_for_device() argument 182 for_each_sg(sgl, sg, nents, i) in arm_nommu_dma_sync_sg_for_device() 186 static void arm_nommu_dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sgl, in arm_nommu_dma_sync_sg_for_cpu() argument 192 for_each_sg(sgl, sg, nents, i) in arm_nommu_dma_sync_sg_for_cpu()
|
/Linux-v4.19/drivers/media/platform/rockchip/rga/ |
D | rga-buf.c | 126 struct scatterlist *sgl; in rga_buf_map() local 139 for_each_sg(sgt->sgl, sgl, sgt->nents, i) { in rga_buf_map() 140 len = sg_dma_len(sgl) >> PAGE_SHIFT; in rga_buf_map() 141 address = sg_phys(sgl); in rga_buf_map()
|
/Linux-v4.19/drivers/scsi/lpfc/ |
D | lpfc_scsi.c | 185 struct sli4_sge *sgl = (struct sli4_sge *)lpfc_cmd->fcp_bpl; in lpfc_sli4_set_rsp_sgl_last() local 186 if (sgl) { in lpfc_sli4_set_rsp_sgl_last() 187 sgl += 1; in lpfc_sli4_set_rsp_sgl_last() 188 sgl->word2 = le32_to_cpu(sgl->word2); in lpfc_sli4_set_rsp_sgl_last() 189 bf_set(lpfc_sli4_sge_last, sgl, 1); in lpfc_sli4_set_rsp_sgl_last() 190 sgl->word2 = cpu_to_le32(sgl->word2); in lpfc_sli4_set_rsp_sgl_last() 799 struct sli4_sge *sgl; in lpfc_new_scsi_buf_s4() local 883 sgl = (struct sli4_sge *)psb->fcp_bpl; in lpfc_new_scsi_buf_s4() 893 sgl->addr_hi = cpu_to_le32(putPaddrHigh(pdma_phys_fcp_cmd)); in lpfc_new_scsi_buf_s4() 894 sgl->addr_lo = cpu_to_le32(putPaddrLow(pdma_phys_fcp_cmd)); in lpfc_new_scsi_buf_s4() [all …]
|
D | lpfc_nvme.c | 767 struct sli4_sge *sgl; in lpfc_nvme_adj_fcp_sgls() local 786 sgl = lpfc_ncmd->nvme_sgl; in lpfc_nvme_adj_fcp_sgls() 787 sgl->sge_len = cpu_to_le32(nCmd->cmdlen); in lpfc_nvme_adj_fcp_sgls() 789 sgl->addr_hi = 0; in lpfc_nvme_adj_fcp_sgls() 790 sgl->addr_lo = 0; in lpfc_nvme_adj_fcp_sgls() 829 sgl->addr_hi = cpu_to_le32(putPaddrHigh(nCmd->cmddma)); in lpfc_nvme_adj_fcp_sgls() 830 sgl->addr_lo = cpu_to_le32(putPaddrLow(nCmd->cmddma)); in lpfc_nvme_adj_fcp_sgls() 835 wqe->generic.bde.addrHigh = sgl->addr_hi; in lpfc_nvme_adj_fcp_sgls() 836 wqe->generic.bde.addrLow = sgl->addr_lo; in lpfc_nvme_adj_fcp_sgls() 843 sgl++; in lpfc_nvme_adj_fcp_sgls() [all …]
|
/Linux-v4.19/drivers/media/platform/ |
D | m2m-deinterlace.c | 243 ctx->xt->sgl[0].size = s_width; in deinterlace_issue_dma() 244 ctx->xt->sgl[0].icg = s_width; in deinterlace_issue_dma() 250 ctx->xt->sgl[0].size = s_width; in deinterlace_issue_dma() 251 ctx->xt->sgl[0].icg = s_width; in deinterlace_issue_dma() 257 ctx->xt->sgl[0].size = s_width / 2; in deinterlace_issue_dma() 258 ctx->xt->sgl[0].icg = s_width / 2; in deinterlace_issue_dma() 264 ctx->xt->sgl[0].size = s_width / 2; in deinterlace_issue_dma() 265 ctx->xt->sgl[0].icg = s_width / 2; in deinterlace_issue_dma() 271 ctx->xt->sgl[0].size = s_width / 2; in deinterlace_issue_dma() 272 ctx->xt->sgl[0].icg = s_width / 2; in deinterlace_issue_dma() [all …]
|
/Linux-v4.19/drivers/media/common/videobuf2/ |
D | videobuf2-dma-contig.c | 52 dma_addr_t expected = sg_dma_address(sgt->sgl); in vb2_dc_get_contiguous_size() 56 for_each_sg(sgt->sgl, s, sgt->nents, i) { in vb2_dc_get_contiguous_size() 102 dma_sync_sg_for_device(buf->dev, sgt->sgl, sgt->orig_nents, in vb2_dc_prepare() 115 dma_sync_sg_for_cpu(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir); in vb2_dc_finish() 249 rd = buf->sgt_base->sgl; in vb2_dc_dmabuf_ops_attach() 250 wr = sgt->sgl; in vb2_dc_dmabuf_ops_attach() 276 dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, in vb2_dc_dmabuf_ops_detach() 302 dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, in vb2_dc_dmabuf_ops_map() 308 sgt->nents = dma_map_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, in vb2_dc_dmabuf_ops_map() 431 dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, in vb2_dc_put_userptr() [all …]
|
D | videobuf2-dma-sg.c | 145 sgt->nents = dma_map_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, in vb2_dma_sg_alloc() 183 dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, in vb2_dma_sg_put() 205 dma_sync_sg_for_device(buf->dev, sgt->sgl, sgt->orig_nents, in vb2_dma_sg_prepare() 218 dma_sync_sg_for_cpu(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir); in vb2_dma_sg_finish() 262 sgt->nents = dma_map_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, in vb2_dma_sg_get_userptr() 290 dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir, in vb2_dma_sg_put_userptr() 398 rd = buf->dma_sgt->sgl; in vb2_dma_sg_dmabuf_ops_attach() 399 wr = sgt->sgl; in vb2_dma_sg_dmabuf_ops_attach() 425 dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, in vb2_dma_sg_dmabuf_ops_detach() 451 dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, in vb2_dma_sg_dmabuf_ops_map() [all …]
|
/Linux-v4.19/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_fcoe.c | 28 ddp->sgl = NULL; in ixgbe_fcoe_clear_ddp() 109 if (ddp->sgl) in ixgbe_fcoe_ddp_put() 110 dma_unmap_sg(&adapter->pdev->dev, ddp->sgl, ddp->sgc, in ixgbe_fcoe_ddp_put() 133 struct scatterlist *sgl, unsigned int sgc, in ixgbe_fcoe_ddp_setup() argument 152 if (!netdev || !sgl) in ixgbe_fcoe_ddp_setup() 168 if (ddp->sgl) { in ixgbe_fcoe_ddp_setup() 170 xid, ddp->sgl, ddp->sgc); in ixgbe_fcoe_ddp_setup() 188 dmacount = dma_map_sg(&adapter->pdev->dev, sgl, sgc, DMA_FROM_DEVICE); in ixgbe_fcoe_ddp_setup() 201 ddp->sgl = sgl; in ixgbe_fcoe_ddp_setup() 205 for_each_sg(sgl, sg, dmacount, i) { in ixgbe_fcoe_ddp_setup() [all …]
|
/Linux-v4.19/drivers/gpu/drm/armada/ |
D | armada_gem.c | 396 for_each_sg(sgt->sgl, sg, count, i) { in armada_gem_prime_map_dma_buf() 408 if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0) { in armada_gem_prime_map_dma_buf() 417 sg_set_page(sgt->sgl, dobj->page, dobj->obj.size, 0); in armada_gem_prime_map_dma_buf() 419 if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0) in armada_gem_prime_map_dma_buf() 425 sg_dma_address(sgt->sgl) = dobj->dev_addr; in armada_gem_prime_map_dma_buf() 426 sg_dma_len(sgt->sgl) = dobj->obj.size; in armada_gem_prime_map_dma_buf() 433 for_each_sg(sgt->sgl, sg, num, i) in armada_gem_prime_map_dma_buf() 450 dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, dir); in armada_gem_prime_unmap_dma_buf() 454 for_each_sg(sgt->sgl, sg, sgt->nents, i) in armada_gem_prime_unmap_dma_buf() 556 if (sg_dma_len(dobj->sgt->sgl) < dobj->obj.size) { in armada_gem_map_import() [all …]
|
/Linux-v4.19/drivers/scsi/esas2r/ |
D | esas2r_io.c | 224 struct esas2r_mem_desc *sgl; in esas2r_build_sg_list_sge() local 231 sgl = esas2r_alloc_sgl(a); in esas2r_build_sg_list_sge() 233 if (unlikely(sgl == NULL)) in esas2r_build_sg_list_sge() 244 memcpy(sgl->virt_addr, sgc->sge.a64.last, sgelen); in esas2r_build_sg_list_sge() 248 (struct atto_vda_sge *)((u8 *)sgl->virt_addr + in esas2r_build_sg_list_sge() 253 (struct atto_vda_sge *)((u8 *)sgl->virt_addr in esas2r_build_sg_list_sge() 260 cpu_to_le64(sgl->phys_addr); in esas2r_build_sg_list_sge() 302 list_add(&sgl->next_desc, &rq->sg_table_head); in esas2r_build_sg_list_sge() 376 struct esas2r_mem_desc *sgl; in esas2r_build_prd_iblk() local 449 sgl = esas2r_alloc_sgl(a); in esas2r_build_prd_iblk() [all …]
|
/Linux-v4.19/drivers/scsi/qedf/ |
D | drv_scsi_fw_funcs.c | 37 val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_addr.lo); in init_scsi_sgl_context() 39 val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_addr.hi); in init_scsi_sgl_context() 41 val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_len); in init_scsi_sgl_context()
|
/Linux-v4.19/drivers/gpu/drm/i915/selftests/ |
D | huge_gem_object.c | 33 for (sg = pages->sgl; sg && nreal--; sg = __sg_next(sg)) in huge_free_pages() 58 sg = pages->sgl; in huge_get_pages() 72 for (end = sg, src = pages->sgl; sg; sg = __sg_next(sg)) { in huge_get_pages() 76 src = pages->sgl; in huge_get_pages()
|
/Linux-v4.19/drivers/dma/ |
D | dma-axi-dmac.c | 384 struct dma_chan *c, struct scatterlist *sgl, in axi_dmac_prep_slave_sg() argument 400 for_each_sg(sgl, sg, sg_len, i) { in axi_dmac_prep_slave_sg() 484 dst_icg = dmaengine_get_dst_icg(xt, &xt->sgl[0]); in axi_dmac_prep_interleaved() 485 src_icg = dmaengine_get_src_icg(xt, &xt->sgl[0]); in axi_dmac_prep_interleaved() 488 if (!axi_dmac_check_len(chan, xt->sgl[0].size) || in axi_dmac_prep_interleaved() 491 if (xt->sgl[0].size + dst_icg > chan->max_length || in axi_dmac_prep_interleaved() 492 xt->sgl[0].size + src_icg > chan->max_length) in axi_dmac_prep_interleaved() 497 if (chan->max_length / xt->sgl[0].size < xt->numf) in axi_dmac_prep_interleaved() 499 if (!axi_dmac_check_len(chan, xt->sgl[0].size * xt->numf)) in axi_dmac_prep_interleaved() 509 desc->sg[0].src_stride = xt->sgl[0].size + src_icg; in axi_dmac_prep_interleaved() [all …]
|
/Linux-v4.19/drivers/infiniband/core/ |
D | umem.c | 54 ib_dma_unmap_sg(dev, umem->sg_head.sgl, in __ib_umem_release() 58 for_each_sg(umem->sg_head.sgl, sg, umem->npages, i) { in __ib_umem_release() 173 sg_list_start = umem->sg_head.sgl; in ib_umem_get() 203 umem->sg_head.sgl, in ib_umem_get() 310 for_each_sg(umem->sg_head.sgl, sg, umem->nmap, i) in ib_umem_page_count() 339 ret = sg_pcopy_to_buffer(umem->sg_head.sgl, umem->npages, dst, length, in ib_umem_copy_from()
|
/Linux-v4.19/drivers/gpu/drm/tegra/ |
D | gem.c | 138 bo->size = iommu_map_sg(tegra->domain, bo->paddr, bo->sgt->sgl, in tegra_bo_iommu_map() 206 dma_unmap_sg(drm->dev, bo->sgt->sgl, bo->sgt->nents, in tegra_bo_free() 232 err = dma_map_sg(drm->dev, bo->sgt->sgl, bo->sgt->nents, in tegra_bo_get_pages() 368 bo->paddr = sg_dma_address(bo->sgt->sgl); in tegra_bo_import() 517 for_each_sg(sgt->sgl, sg, bo->num_pages, i) in tegra_gem_prime_map_dma_buf() 520 if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0) in tegra_gem_prime_map_dma_buf() 526 sg_dma_address(sgt->sgl) = bo->paddr; in tegra_gem_prime_map_dma_buf() 527 sg_dma_len(sgt->sgl) = gem->size; in tegra_gem_prime_map_dma_buf() 546 dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, dir); in tegra_gem_prime_unmap_dma_buf() 565 dma_sync_sg_for_cpu(drm->dev, bo->sgt->sgl, bo->sgt->nents, in tegra_gem_prime_begin_cpu_access() [all …]
|