Home
last modified time | relevance | path

Searched refs:sgl (Results 1 – 25 of 381) sorted by relevance

12345678910>>...16

/Linux-v4.19/drivers/misc/genwqe/
Dcard_utils.c298 int genwqe_alloc_sync_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl, in genwqe_alloc_sync_sgl() argument
304 sgl->fpage_offs = offset_in_page((unsigned long)user_addr); in genwqe_alloc_sync_sgl()
305 sgl->fpage_size = min_t(size_t, PAGE_SIZE-sgl->fpage_offs, user_size); in genwqe_alloc_sync_sgl()
306 sgl->nr_pages = DIV_ROUND_UP(sgl->fpage_offs + user_size, PAGE_SIZE); in genwqe_alloc_sync_sgl()
307 sgl->lpage_size = (user_size - sgl->fpage_size) % PAGE_SIZE; in genwqe_alloc_sync_sgl()
310 __func__, user_addr, user_size, sgl->nr_pages, in genwqe_alloc_sync_sgl()
311 sgl->fpage_offs, sgl->fpage_size, sgl->lpage_size); in genwqe_alloc_sync_sgl()
313 sgl->user_addr = user_addr; in genwqe_alloc_sync_sgl()
314 sgl->user_size = user_size; in genwqe_alloc_sync_sgl()
315 sgl->write = write; in genwqe_alloc_sync_sgl()
[all …]
/Linux-v4.19/lib/
Dscatterlist.c103 struct scatterlist *sg_last(struct scatterlist *sgl, unsigned int nents) in sg_last() argument
108 for_each_sg(sgl, sg, nents, i) in sg_last()
126 void sg_init_table(struct scatterlist *sgl, unsigned int nents) in sg_init_table() argument
128 memset(sgl, 0, sizeof(*sgl) * nents); in sg_init_table()
129 sg_init_marker(sgl, nents); in sg_init_table()
196 struct scatterlist *sgl, *next; in __sg_free_table() local
198 if (unlikely(!table->sgl)) in __sg_free_table()
201 sgl = table->sgl; in __sg_free_table()
213 next = sg_chain_ptr(&sgl[max_ents - 1]); in __sg_free_table()
225 free_fn(sgl, alloc_size); in __sg_free_table()
[all …]
Dkfifo.c308 static int setup_sgl_buf(struct scatterlist *sgl, void *buf, in setup_sgl_buf() argument
334 sg_set_page(sgl, page, l - off, off); in setup_sgl_buf()
335 sgl = sg_next(sgl); in setup_sgl_buf()
336 if (++n == nents || sgl == NULL) in setup_sgl_buf()
343 sg_set_page(sgl, page, len, off); in setup_sgl_buf()
347 static unsigned int setup_sgl(struct __kfifo *fifo, struct scatterlist *sgl, in setup_sgl() argument
363 n = setup_sgl_buf(sgl, fifo->data + off, nents, l); in setup_sgl()
364 n += setup_sgl_buf(sgl + n, fifo->data, nents - n, len - l); in setup_sgl()
370 struct scatterlist *sgl, int nents, unsigned int len) in __kfifo_dma_in_prepare() argument
378 return setup_sgl(fifo, sgl, nents, len, fifo->in); in __kfifo_dma_in_prepare()
[all …]
/Linux-v4.19/crypto/
Daf_alg.c397 int af_alg_make_sg(struct af_alg_sgl *sgl, struct iov_iter *iter, int len) in af_alg_make_sg() argument
403 n = iov_iter_get_pages(iter, sgl->pages, len, ALG_MAX_PAGES, &off); in af_alg_make_sg()
411 sg_init_table(sgl->sg, npages + 1); in af_alg_make_sg()
416 sg_set_page(sgl->sg + i, sgl->pages[i], plen, off); in af_alg_make_sg()
421 sg_mark_end(sgl->sg + npages - 1); in af_alg_make_sg()
422 sgl->npages = npages; in af_alg_make_sg()
435 void af_alg_free_sg(struct af_alg_sgl *sgl) in af_alg_free_sg() argument
439 for (i = 0; i < sgl->npages; i++) in af_alg_free_sg()
440 put_page(sgl->pages[i]); in af_alg_free_sg()
495 struct af_alg_tsgl *sgl; in af_alg_alloc_tsgl() local
[all …]
/Linux-v4.19/include/linux/
Dscatterlist.h43 struct scatterlist *sgl; /* the list */ member
165 struct scatterlist *sgl) in sg_chain() argument
177 prv[prv_nents - 1].page_link = ((unsigned long) sgl | SG_CHAIN) in sg_chain()
248 static inline void sg_init_marker(struct scatterlist *sgl, in sg_init_marker() argument
251 sg_mark_end(&sgl[nents - 1]); in sg_init_marker()
288 void sgl_free_n_order(struct scatterlist *sgl, int nents, int order);
289 void sgl_free_order(struct scatterlist *sgl, int order);
290 void sgl_free(struct scatterlist *sgl);
293 size_t sg_copy_buffer(struct scatterlist *sgl, unsigned int nents, void *buf,
296 size_t sg_copy_from_buffer(struct scatterlist *sgl, unsigned int nents,
[all …]
/Linux-v4.19/kernel/dma/
Dnoncoherent.c21 struct scatterlist *sgl, int nents, enum dma_data_direction dir) in dma_noncoherent_sync_sg_for_device() argument
26 for_each_sg(sgl, sg, nents, i) in dma_noncoherent_sync_sg_for_device()
43 static int dma_noncoherent_map_sg(struct device *dev, struct scatterlist *sgl, in dma_noncoherent_map_sg() argument
46 nents = dma_direct_map_sg(dev, sgl, nents, dir, attrs); in dma_noncoherent_map_sg()
48 dma_noncoherent_sync_sg_for_device(dev, sgl, nents, dir); in dma_noncoherent_map_sg()
62 struct scatterlist *sgl, int nents, enum dma_data_direction dir) in dma_noncoherent_sync_sg_for_cpu() argument
67 for_each_sg(sgl, sg, nents, i) in dma_noncoherent_sync_sg_for_cpu()
79 static void dma_noncoherent_unmap_sg(struct device *dev, struct scatterlist *sgl, in dma_noncoherent_unmap_sg() argument
83 dma_noncoherent_sync_sg_for_cpu(dev, sgl, nents, dir); in dma_noncoherent_unmap_sg()
/Linux-v4.19/drivers/target/iscsi/cxgbit/
Dcxgbit_ddp.c136 struct scatterlist *sg = ttinfo->sgl; in cxgbit_ddp_set_map()
179 struct scatterlist *sgl = ttinfo->sgl; in cxgbit_ddp_reserve() local
181 unsigned int sg_offset = sgl->offset; in cxgbit_ddp_reserve()
191 if (cxgbit_ddp_sgl_check(sgl, sgcnt) < 0) in cxgbit_ddp_reserve()
194 ttinfo->nr_pages = (xferlen + sgl->offset + in cxgbit_ddp_reserve()
206 sgl->offset = 0; in cxgbit_ddp_reserve()
207 ret = dma_map_sg(&ppm->pdev->dev, sgl, sgcnt, DMA_FROM_DEVICE); in cxgbit_ddp_reserve()
208 sgl->offset = sg_offset; in cxgbit_ddp_reserve()
215 cxgbi_ppm_make_ppod_hdr(ppm, ttinfo->tag, csk->tid, sgl->offset, in cxgbit_ddp_reserve()
221 dma_unmap_sg(&ppm->pdev->dev, sgl, sgcnt, DMA_FROM_DEVICE); in cxgbit_ddp_reserve()
[all …]
/Linux-v4.19/tools/virtio/linux/
Dscatterlist.h91 struct scatterlist *sgl) in sg_chain() argument
103 prv[prv_nents - 1].page_link = ((unsigned long) sgl | 0x01) & ~0x02; in sg_chain()
149 static inline void sg_init_table(struct scatterlist *sgl, unsigned int nents) in sg_init_table() argument
151 memset(sgl, 0, sizeof(*sgl) * nents); in sg_init_table()
152 sg_mark_end(&sgl[nents - 1]); in sg_init_table()
/Linux-v4.19/drivers/gpu/drm/v3d/
Dv3d_mmu.c90 struct scatterlist *sgl; in v3d_mmu_insert_ptes() local
92 for_each_sg(bo->sgt->sgl, sgl, bo->sgt->nents, count) { in v3d_mmu_insert_ptes()
93 u32 page_address = sg_dma_address(sgl) >> V3D_MMU_PAGE_SHIFT; in v3d_mmu_insert_ptes()
97 BUG_ON(page_address + (sg_dma_len(sgl) >> V3D_MMU_PAGE_SHIFT) >= in v3d_mmu_insert_ptes()
100 for (i = 0; i < sg_dma_len(sgl) >> V3D_MMU_PAGE_SHIFT; i++) in v3d_mmu_insert_ptes()
/Linux-v4.19/arch/arm64/mm/
Ddma-mapping.c183 static int __swiotlb_map_sg_attrs(struct device *dev, struct scatterlist *sgl, in __swiotlb_map_sg_attrs() argument
190 ret = swiotlb_map_sg_attrs(dev, sgl, nelems, dir, attrs); in __swiotlb_map_sg_attrs()
193 for_each_sg(sgl, sg, ret, i) in __swiotlb_map_sg_attrs()
201 struct scatterlist *sgl, int nelems, in __swiotlb_unmap_sg_attrs() argument
210 for_each_sg(sgl, sg, nelems, i) in __swiotlb_unmap_sg_attrs()
213 swiotlb_unmap_sg_attrs(dev, sgl, nelems, dir, attrs); in __swiotlb_unmap_sg_attrs()
235 struct scatterlist *sgl, int nelems, in __swiotlb_sync_sg_for_cpu() argument
242 for_each_sg(sgl, sg, nelems, i) in __swiotlb_sync_sg_for_cpu()
245 swiotlb_sync_sg_for_cpu(dev, sgl, nelems, dir); in __swiotlb_sync_sg_for_cpu()
249 struct scatterlist *sgl, int nelems, in __swiotlb_sync_sg_for_device() argument
[all …]
/Linux-v4.19/arch/arm/mm/
Ddma-mapping-nommu.c137 static int arm_nommu_dma_map_sg(struct device *dev, struct scatterlist *sgl, in arm_nommu_dma_map_sg() argument
144 for_each_sg(sgl, sg, nents, i) { in arm_nommu_dma_map_sg()
153 static void arm_nommu_dma_unmap_sg(struct device *dev, struct scatterlist *sgl, in arm_nommu_dma_unmap_sg() argument
160 for_each_sg(sgl, sg, nents, i) in arm_nommu_dma_unmap_sg()
176 static void arm_nommu_dma_sync_sg_for_device(struct device *dev, struct scatterlist *sgl, in arm_nommu_dma_sync_sg_for_device() argument
182 for_each_sg(sgl, sg, nents, i) in arm_nommu_dma_sync_sg_for_device()
186 static void arm_nommu_dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sgl, in arm_nommu_dma_sync_sg_for_cpu() argument
192 for_each_sg(sgl, sg, nents, i) in arm_nommu_dma_sync_sg_for_cpu()
/Linux-v4.19/drivers/media/platform/rockchip/rga/
Drga-buf.c126 struct scatterlist *sgl; in rga_buf_map() local
139 for_each_sg(sgt->sgl, sgl, sgt->nents, i) { in rga_buf_map()
140 len = sg_dma_len(sgl) >> PAGE_SHIFT; in rga_buf_map()
141 address = sg_phys(sgl); in rga_buf_map()
/Linux-v4.19/drivers/scsi/lpfc/
Dlpfc_scsi.c185 struct sli4_sge *sgl = (struct sli4_sge *)lpfc_cmd->fcp_bpl; in lpfc_sli4_set_rsp_sgl_last() local
186 if (sgl) { in lpfc_sli4_set_rsp_sgl_last()
187 sgl += 1; in lpfc_sli4_set_rsp_sgl_last()
188 sgl->word2 = le32_to_cpu(sgl->word2); in lpfc_sli4_set_rsp_sgl_last()
189 bf_set(lpfc_sli4_sge_last, sgl, 1); in lpfc_sli4_set_rsp_sgl_last()
190 sgl->word2 = cpu_to_le32(sgl->word2); in lpfc_sli4_set_rsp_sgl_last()
799 struct sli4_sge *sgl; in lpfc_new_scsi_buf_s4() local
883 sgl = (struct sli4_sge *)psb->fcp_bpl; in lpfc_new_scsi_buf_s4()
893 sgl->addr_hi = cpu_to_le32(putPaddrHigh(pdma_phys_fcp_cmd)); in lpfc_new_scsi_buf_s4()
894 sgl->addr_lo = cpu_to_le32(putPaddrLow(pdma_phys_fcp_cmd)); in lpfc_new_scsi_buf_s4()
[all …]
Dlpfc_nvme.c767 struct sli4_sge *sgl; in lpfc_nvme_adj_fcp_sgls() local
786 sgl = lpfc_ncmd->nvme_sgl; in lpfc_nvme_adj_fcp_sgls()
787 sgl->sge_len = cpu_to_le32(nCmd->cmdlen); in lpfc_nvme_adj_fcp_sgls()
789 sgl->addr_hi = 0; in lpfc_nvme_adj_fcp_sgls()
790 sgl->addr_lo = 0; in lpfc_nvme_adj_fcp_sgls()
829 sgl->addr_hi = cpu_to_le32(putPaddrHigh(nCmd->cmddma)); in lpfc_nvme_adj_fcp_sgls()
830 sgl->addr_lo = cpu_to_le32(putPaddrLow(nCmd->cmddma)); in lpfc_nvme_adj_fcp_sgls()
835 wqe->generic.bde.addrHigh = sgl->addr_hi; in lpfc_nvme_adj_fcp_sgls()
836 wqe->generic.bde.addrLow = sgl->addr_lo; in lpfc_nvme_adj_fcp_sgls()
843 sgl++; in lpfc_nvme_adj_fcp_sgls()
[all …]
/Linux-v4.19/drivers/media/platform/
Dm2m-deinterlace.c243 ctx->xt->sgl[0].size = s_width; in deinterlace_issue_dma()
244 ctx->xt->sgl[0].icg = s_width; in deinterlace_issue_dma()
250 ctx->xt->sgl[0].size = s_width; in deinterlace_issue_dma()
251 ctx->xt->sgl[0].icg = s_width; in deinterlace_issue_dma()
257 ctx->xt->sgl[0].size = s_width / 2; in deinterlace_issue_dma()
258 ctx->xt->sgl[0].icg = s_width / 2; in deinterlace_issue_dma()
264 ctx->xt->sgl[0].size = s_width / 2; in deinterlace_issue_dma()
265 ctx->xt->sgl[0].icg = s_width / 2; in deinterlace_issue_dma()
271 ctx->xt->sgl[0].size = s_width / 2; in deinterlace_issue_dma()
272 ctx->xt->sgl[0].icg = s_width / 2; in deinterlace_issue_dma()
[all …]
/Linux-v4.19/drivers/media/common/videobuf2/
Dvideobuf2-dma-contig.c52 dma_addr_t expected = sg_dma_address(sgt->sgl); in vb2_dc_get_contiguous_size()
56 for_each_sg(sgt->sgl, s, sgt->nents, i) { in vb2_dc_get_contiguous_size()
102 dma_sync_sg_for_device(buf->dev, sgt->sgl, sgt->orig_nents, in vb2_dc_prepare()
115 dma_sync_sg_for_cpu(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir); in vb2_dc_finish()
249 rd = buf->sgt_base->sgl; in vb2_dc_dmabuf_ops_attach()
250 wr = sgt->sgl; in vb2_dc_dmabuf_ops_attach()
276 dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, in vb2_dc_dmabuf_ops_detach()
302 dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, in vb2_dc_dmabuf_ops_map()
308 sgt->nents = dma_map_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, in vb2_dc_dmabuf_ops_map()
431 dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, in vb2_dc_put_userptr()
[all …]
Dvideobuf2-dma-sg.c145 sgt->nents = dma_map_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, in vb2_dma_sg_alloc()
183 dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, in vb2_dma_sg_put()
205 dma_sync_sg_for_device(buf->dev, sgt->sgl, sgt->orig_nents, in vb2_dma_sg_prepare()
218 dma_sync_sg_for_cpu(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir); in vb2_dma_sg_finish()
262 sgt->nents = dma_map_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, in vb2_dma_sg_get_userptr()
290 dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir, in vb2_dma_sg_put_userptr()
398 rd = buf->dma_sgt->sgl; in vb2_dma_sg_dmabuf_ops_attach()
399 wr = sgt->sgl; in vb2_dma_sg_dmabuf_ops_attach()
425 dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, in vb2_dma_sg_dmabuf_ops_detach()
451 dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, in vb2_dma_sg_dmabuf_ops_map()
[all …]
/Linux-v4.19/drivers/net/ethernet/intel/ixgbe/
Dixgbe_fcoe.c28 ddp->sgl = NULL; in ixgbe_fcoe_clear_ddp()
109 if (ddp->sgl) in ixgbe_fcoe_ddp_put()
110 dma_unmap_sg(&adapter->pdev->dev, ddp->sgl, ddp->sgc, in ixgbe_fcoe_ddp_put()
133 struct scatterlist *sgl, unsigned int sgc, in ixgbe_fcoe_ddp_setup() argument
152 if (!netdev || !sgl) in ixgbe_fcoe_ddp_setup()
168 if (ddp->sgl) { in ixgbe_fcoe_ddp_setup()
170 xid, ddp->sgl, ddp->sgc); in ixgbe_fcoe_ddp_setup()
188 dmacount = dma_map_sg(&adapter->pdev->dev, sgl, sgc, DMA_FROM_DEVICE); in ixgbe_fcoe_ddp_setup()
201 ddp->sgl = sgl; in ixgbe_fcoe_ddp_setup()
205 for_each_sg(sgl, sg, dmacount, i) { in ixgbe_fcoe_ddp_setup()
[all …]
/Linux-v4.19/drivers/gpu/drm/armada/
Darmada_gem.c396 for_each_sg(sgt->sgl, sg, count, i) { in armada_gem_prime_map_dma_buf()
408 if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0) { in armada_gem_prime_map_dma_buf()
417 sg_set_page(sgt->sgl, dobj->page, dobj->obj.size, 0); in armada_gem_prime_map_dma_buf()
419 if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0) in armada_gem_prime_map_dma_buf()
425 sg_dma_address(sgt->sgl) = dobj->dev_addr; in armada_gem_prime_map_dma_buf()
426 sg_dma_len(sgt->sgl) = dobj->obj.size; in armada_gem_prime_map_dma_buf()
433 for_each_sg(sgt->sgl, sg, num, i) in armada_gem_prime_map_dma_buf()
450 dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, dir); in armada_gem_prime_unmap_dma_buf()
454 for_each_sg(sgt->sgl, sg, sgt->nents, i) in armada_gem_prime_unmap_dma_buf()
556 if (sg_dma_len(dobj->sgt->sgl) < dobj->obj.size) { in armada_gem_map_import()
[all …]
/Linux-v4.19/drivers/scsi/esas2r/
Desas2r_io.c224 struct esas2r_mem_desc *sgl; in esas2r_build_sg_list_sge() local
231 sgl = esas2r_alloc_sgl(a); in esas2r_build_sg_list_sge()
233 if (unlikely(sgl == NULL)) in esas2r_build_sg_list_sge()
244 memcpy(sgl->virt_addr, sgc->sge.a64.last, sgelen); in esas2r_build_sg_list_sge()
248 (struct atto_vda_sge *)((u8 *)sgl->virt_addr + in esas2r_build_sg_list_sge()
253 (struct atto_vda_sge *)((u8 *)sgl->virt_addr in esas2r_build_sg_list_sge()
260 cpu_to_le64(sgl->phys_addr); in esas2r_build_sg_list_sge()
302 list_add(&sgl->next_desc, &rq->sg_table_head); in esas2r_build_sg_list_sge()
376 struct esas2r_mem_desc *sgl; in esas2r_build_prd_iblk() local
449 sgl = esas2r_alloc_sgl(a); in esas2r_build_prd_iblk()
[all …]
/Linux-v4.19/drivers/scsi/qedf/
Ddrv_scsi_fw_funcs.c37 val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_addr.lo); in init_scsi_sgl_context()
39 val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_addr.hi); in init_scsi_sgl_context()
41 val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_len); in init_scsi_sgl_context()
/Linux-v4.19/drivers/gpu/drm/i915/selftests/
Dhuge_gem_object.c33 for (sg = pages->sgl; sg && nreal--; sg = __sg_next(sg)) in huge_free_pages()
58 sg = pages->sgl; in huge_get_pages()
72 for (end = sg, src = pages->sgl; sg; sg = __sg_next(sg)) { in huge_get_pages()
76 src = pages->sgl; in huge_get_pages()
/Linux-v4.19/drivers/dma/
Ddma-axi-dmac.c384 struct dma_chan *c, struct scatterlist *sgl, in axi_dmac_prep_slave_sg() argument
400 for_each_sg(sgl, sg, sg_len, i) { in axi_dmac_prep_slave_sg()
484 dst_icg = dmaengine_get_dst_icg(xt, &xt->sgl[0]); in axi_dmac_prep_interleaved()
485 src_icg = dmaengine_get_src_icg(xt, &xt->sgl[0]); in axi_dmac_prep_interleaved()
488 if (!axi_dmac_check_len(chan, xt->sgl[0].size) || in axi_dmac_prep_interleaved()
491 if (xt->sgl[0].size + dst_icg > chan->max_length || in axi_dmac_prep_interleaved()
492 xt->sgl[0].size + src_icg > chan->max_length) in axi_dmac_prep_interleaved()
497 if (chan->max_length / xt->sgl[0].size < xt->numf) in axi_dmac_prep_interleaved()
499 if (!axi_dmac_check_len(chan, xt->sgl[0].size * xt->numf)) in axi_dmac_prep_interleaved()
509 desc->sg[0].src_stride = xt->sgl[0].size + src_icg; in axi_dmac_prep_interleaved()
[all …]
/Linux-v4.19/drivers/infiniband/core/
Dumem.c54 ib_dma_unmap_sg(dev, umem->sg_head.sgl, in __ib_umem_release()
58 for_each_sg(umem->sg_head.sgl, sg, umem->npages, i) { in __ib_umem_release()
173 sg_list_start = umem->sg_head.sgl; in ib_umem_get()
203 umem->sg_head.sgl, in ib_umem_get()
310 for_each_sg(umem->sg_head.sgl, sg, umem->nmap, i) in ib_umem_page_count()
339 ret = sg_pcopy_to_buffer(umem->sg_head.sgl, umem->npages, dst, length, in ib_umem_copy_from()
/Linux-v4.19/drivers/gpu/drm/tegra/
Dgem.c138 bo->size = iommu_map_sg(tegra->domain, bo->paddr, bo->sgt->sgl, in tegra_bo_iommu_map()
206 dma_unmap_sg(drm->dev, bo->sgt->sgl, bo->sgt->nents, in tegra_bo_free()
232 err = dma_map_sg(drm->dev, bo->sgt->sgl, bo->sgt->nents, in tegra_bo_get_pages()
368 bo->paddr = sg_dma_address(bo->sgt->sgl); in tegra_bo_import()
517 for_each_sg(sgt->sgl, sg, bo->num_pages, i) in tegra_gem_prime_map_dma_buf()
520 if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0) in tegra_gem_prime_map_dma_buf()
526 sg_dma_address(sgt->sgl) = bo->paddr; in tegra_gem_prime_map_dma_buf()
527 sg_dma_len(sgt->sgl) = gem->size; in tegra_gem_prime_map_dma_buf()
546 dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, dir); in tegra_gem_prime_unmap_dma_buf()
565 dma_sync_sg_for_cpu(drm->dev, bo->sgt->sgl, bo->sgt->nents, in tegra_gem_prime_begin_cpu_access()
[all …]

12345678910>>...16