/Linux-v4.19/drivers/parisc/ |
D | iommu-helpers.h | 47 BUG_ON(pdirp && (dma_len != sg_dma_len(dma_sg))); in iommu_fill_pdir() 51 dma_len = sg_dma_len(startsg); in iommu_fill_pdir() 52 sg_dma_len(startsg) = 0; in iommu_fill_pdir() 71 sg_dma_len(dma_sg) += startsg->length; in iommu_fill_pdir() 125 sg_dma_len(startsg) = 0; in iommu_coalesce_chunks() 142 sg_dma_len(startsg) = 0; in iommu_coalesce_chunks() 171 sg_dma_len(contig_sg) = dma_len; in iommu_coalesce_chunks()
|
D | ccio-rm-dma.c | 122 sg_dma_len(sglist) = sglist->length; in ccio_map_sg() 135 ccio_unmap_single(dev, sg_dma_address(sglist), sg_dma_len(sglist), direction); in ccio_unmap_sg()
|
/Linux-v4.19/drivers/media/pci/tw68/ |
D | tw68-risc.c | 66 while (offset && offset >= sg_dma_len(sg)) { in tw68_risc_field() 67 offset -= sg_dma_len(sg); in tw68_risc_field() 70 if (bpl <= sg_dma_len(sg) - offset) { in tw68_risc_field() 85 done = (sg_dma_len(sg) - offset); in tw68_risc_field() 93 while (todo > sg_dma_len(sg)) { in tw68_risc_field() 96 sg_dma_len(sg)); in tw68_risc_field() 98 todo -= sg_dma_len(sg); in tw68_risc_field() 100 done += sg_dma_len(sg); in tw68_risc_field()
|
/Linux-v4.19/net/rds/ |
D | ib_fmr.c | 97 int page_cnt, sg_dma_len; in rds_ib_map_fmr() local 101 sg_dma_len = ib_dma_map_sg(dev, sg, nents, DMA_BIDIRECTIONAL); in rds_ib_map_fmr() 102 if (unlikely(!sg_dma_len)) { in rds_ib_map_fmr() 110 for (i = 0; i < sg_dma_len; ++i) { in rds_ib_map_fmr() 124 if (i < sg_dma_len - 1) { in rds_ib_map_fmr() 150 for (i = 0; i < sg_dma_len; ++i) { in rds_ib_map_fmr() 172 ibmr->sg_dma_len = sg_dma_len; in rds_ib_map_fmr()
|
D | ib_frmr.c | 168 ibmr->sg_dma_len = 0; in rds_ib_map_frmr() 170 WARN_ON(ibmr->sg_dma_len); in rds_ib_map_frmr() 171 ibmr->sg_dma_len = ib_dma_map_sg(dev, ibmr->sg, ibmr->sg_len, in rds_ib_map_frmr() 173 if (unlikely(!ibmr->sg_dma_len)) { in rds_ib_map_frmr() 183 for (i = 0; i < ibmr->sg_dma_len; ++i) { in rds_ib_map_frmr() 196 if (i < ibmr->sg_dma_len - 1) in rds_ib_map_frmr() 225 ibmr->sg_dma_len = 0; in rds_ib_map_frmr() 304 if (ibmr->sg_dma_len) in rds_ib_unreg_frmr()
|
D | ib.h | 327 unsigned int sg_dma_len, in rds_ib_dma_sync_sg_for_cpu() argument 333 for_each_sg(sglist, sg, sg_dma_len, i) { in rds_ib_dma_sync_sg_for_cpu() 344 unsigned int sg_dma_len, in rds_ib_dma_sync_sg_for_device() argument 350 for_each_sg(sglist, sg, sg_dma_len, i) { in rds_ib_dma_sync_sg_for_device()
|
D | ib_rdma.c | 237 ibmr->sg_dma_len, DMA_BIDIRECTIONAL); in rds_ib_sync_mr() 241 ibmr->sg_dma_len, DMA_BIDIRECTIONAL); in rds_ib_sync_mr() 250 if (ibmr->sg_dma_len) { in __rds_ib_teardown_mr() 254 ibmr->sg_dma_len = 0; in __rds_ib_teardown_mr()
|
/Linux-v4.19/lib/ |
D | sg_split.c | 38 sglen = mapped ? sg_dma_len(sg) : sg->length; in sg_calculate_split() 97 sg_dma_len(out_sg) = 0; in sg_split_phys() 116 sg_dma_len(out_sg) = sg_dma_len(in_sg); in sg_split_mapped() 119 sg_dma_len(out_sg) -= split->skip_sg0; in sg_split_mapped() 123 sg_dma_len(--out_sg) = split->length_last_sg; in sg_split_mapped()
|
/Linux-v4.19/drivers/media/pci/bt8xx/ |
D | bttv-risc.c | 85 while (offset && offset >= sg_dma_len(sg)) { in bttv_risc_packed() 86 offset -= sg_dma_len(sg); in bttv_risc_packed() 89 if (bpl <= sg_dma_len(sg)-offset) { in bttv_risc_packed() 99 (sg_dma_len(sg)-offset)); in bttv_risc_packed() 101 todo -= (sg_dma_len(sg)-offset); in bttv_risc_packed() 104 while (todo > sg_dma_len(sg)) { in bttv_risc_packed() 106 sg_dma_len(sg)); in bttv_risc_packed() 108 todo -= sg_dma_len(sg); in bttv_risc_packed() 188 while (yoffset && yoffset >= sg_dma_len(ysg)) { in bttv_risc_planar() 189 yoffset -= sg_dma_len(ysg); in bttv_risc_planar() [all …]
|
/Linux-v4.19/arch/arm/mm/ |
D | dma-mapping-nommu.c | 146 sg_dma_len(sg) = sg->length; in arm_nommu_dma_map_sg() 147 __dma_page_cpu_to_dev(sg_dma_address(sg), sg_dma_len(sg), dir); in arm_nommu_dma_map_sg() 161 __dma_page_dev_to_cpu(sg_dma_address(sg), sg_dma_len(sg), dir); in arm_nommu_dma_unmap_sg() 183 __dma_page_cpu_to_dev(sg_dma_address(sg), sg_dma_len(sg), dir); in arm_nommu_dma_sync_sg_for_device() 193 __dma_page_dev_to_cpu(sg_dma_address(sg), sg_dma_len(sg), dir); in arm_nommu_dma_sync_sg_for_cpu()
|
/Linux-v4.19/drivers/media/pci/cx25821/ |
D | cx25821-core.c | 1024 while (offset && offset >= sg_dma_len(sg)) { in cx25821_risc_field() 1025 offset -= sg_dma_len(sg); in cx25821_risc_field() 1028 if (bpl <= sg_dma_len(sg) - offset) { in cx25821_risc_field() 1039 (sg_dma_len(sg) - offset)); in cx25821_risc_field() 1042 todo -= (sg_dma_len(sg) - offset); in cx25821_risc_field() 1045 while (todo > sg_dma_len(sg)) { in cx25821_risc_field() 1047 sg_dma_len(sg)); in cx25821_risc_field() 1050 todo -= sg_dma_len(sg); in cx25821_risc_field() 1129 while (offset && offset >= sg_dma_len(sg)) { in cx25821_risc_field_audio() 1130 offset -= sg_dma_len(sg); in cx25821_risc_field_audio() [all …]
|
/Linux-v4.19/drivers/iommu/ |
D | dma-iommu.c | 662 unsigned int s_length = sg_dma_len(s); in __finalise_sg() 668 sg_dma_len(s) = 0; in __finalise_sg() 691 sg_dma_len(cur) = cur_len; in __finalise_sg() 712 if (sg_dma_len(s)) in __invalidate_sg() 713 s->length = sg_dma_len(s); in __invalidate_sg() 715 sg_dma_len(s) = 0; in __invalidate_sg() 750 sg_dma_len(s) = s_length; in iommu_dma_map_sg() 809 if (sg_dma_len(tmp) == 0) in iommu_dma_unmap_sg() 813 end = sg_dma_address(sg) + sg_dma_len(sg); in iommu_dma_unmap_sg()
|
/Linux-v4.19/drivers/gpu/drm/v3d/ |
D | v3d_mmu.c | 97 BUG_ON(page_address + (sg_dma_len(sgl) >> V3D_MMU_PAGE_SHIFT) >= in v3d_mmu_insert_ptes() 100 for (i = 0; i < sg_dma_len(sgl) >> V3D_MMU_PAGE_SHIFT; i++) in v3d_mmu_insert_ptes()
|
/Linux-v4.19/drivers/gpu/drm/etnaviv/ |
D | etnaviv_mmu.c | 82 size_t bytes = sg_dma_len(sg) + sg->offset; in etnaviv_iommu_map() 99 size_t bytes = sg_dma_len(sg) + sg->offset; in etnaviv_iommu_map() 116 size_t bytes = sg_dma_len(sg) + sg->offset; in etnaviv_iommu_unmap() 240 if (iova < 0x80000000 - sg_dma_len(sgt->sgl)) { in etnaviv_iommu_map_gem()
|
/Linux-v4.19/drivers/media/pci/ivtv/ |
D | ivtv-udma.c | 80 dma->SGarray[i].size = cpu_to_le32(sg_dma_len(sg)); in ivtv_udma_fill_sg_array() 83 buffer_offset += sg_dma_len(sg); in ivtv_udma_fill_sg_array() 85 split -= sg_dma_len(sg); in ivtv_udma_fill_sg_array()
|
/Linux-v4.19/drivers/media/pci/cx88/ |
D | cx88-core.c | 99 while (offset && offset >= sg_dma_len(sg)) { in cx88_risc_field() 100 offset -= sg_dma_len(sg); in cx88_risc_field() 107 if (bpl <= sg_dma_len(sg) - offset) { in cx88_risc_field() 117 (sg_dma_len(sg) - offset)); in cx88_risc_field() 119 todo -= (sg_dma_len(sg) - offset); in cx88_risc_field() 122 while (todo > sg_dma_len(sg)) { in cx88_risc_field() 124 sg_dma_len(sg)); in cx88_risc_field() 126 todo -= sg_dma_len(sg); in cx88_risc_field()
|
/Linux-v4.19/drivers/crypto/ccp/ |
D | ccp-dmaengine.c | 382 src_len = sg_dma_len(src_sg); in ccp_create_desc() 385 dst_len = sg_dma_len(dst_sg); in ccp_create_desc() 398 src_len = sg_dma_len(src_sg); in ccp_create_desc() 412 dst_len = sg_dma_len(dst_sg); in ccp_create_desc() 491 sg_dma_len(&dst_sg) = len; in ccp_prep_dma_memcpy() 495 sg_dma_len(&src_sg) = len; in ccp_prep_dma_memcpy()
|
/Linux-v4.19/drivers/dma/ |
D | ste_dma40_ll.c | 286 unsigned int len = sg_dma_len(current_sg); in d40_phy_sg_to_lli() 289 total_size += sg_dma_len(current_sg); in d40_phy_sg_to_lli() 435 unsigned int len = sg_dma_len(current_sg); in d40_log_sg_to_lli() 438 total_size += sg_dma_len(current_sg); in d40_log_sg_to_lli()
|
D | altera-msgdma.c | 406 desc_cnt += DIV_ROUND_UP(sg_dma_len(sg), MSGDMA_MAX_TRANS_LEN); in msgdma_prep_slave_sg() 417 avail = sg_dma_len(sgl); in msgdma_prep_slave_sg() 428 dma_src = sg_dma_address(sgl) + sg_dma_len(sgl) - avail; in msgdma_prep_slave_sg() 433 dma_dst = sg_dma_address(sgl) + sg_dma_len(sgl) - avail; in msgdma_prep_slave_sg() 452 avail = sg_dma_len(sgl); in msgdma_prep_slave_sg()
|
D | timb_dma.c | 156 if (sg_dma_len(sg) > USHRT_MAX) { in td_fill_desc() 162 if (sg_dma_len(sg) % sizeof(u32)) { in td_fill_desc() 164 sg_dma_len(sg)); in td_fill_desc() 176 dma_desc[3] = (sg_dma_len(sg) >> 8) & 0xff; in td_fill_desc() 177 dma_desc[2] = (sg_dma_len(sg) >> 0) & 0xff; in td_fill_desc()
|
/Linux-v4.19/drivers/media/pci/cobalt/ |
D | cobalt-omnitek.c | 193 bytes = min(sg_dma_len(scatter_list) - offset, in descriptor_list_create() 236 bytes = min(sg_dma_len(scatter_list) - offset, in descriptor_list_create() 241 if (sg_dma_len(scatter_list) == offset) { in descriptor_list_create()
|
/Linux-v4.19/drivers/infiniband/hw/mlx5/ |
D | mem.c | 77 len = sg_dma_len(sg) >> page_shift; in mlx5_ib_cont_pages() 172 len = sg_dma_len(sg) >> umem_page_shift; in __mlx5_ib_populate_pas()
|
/Linux-v4.19/drivers/crypto/caam/ |
D | sg_sw_qm2.h | 61 sg_dma_len(sg), offset); in sg_to_qm_sg()
|
/Linux-v4.19/drivers/scsi/aic94xx/ |
D | aic94xx_task.c | 101 sg->size = cpu_to_le32((u32)sg_dma_len(sc)); in asd_map_scatterlist() 109 sg_arr[i].size = cpu_to_le32((u32)sg_dma_len(sc)); in asd_map_scatterlist() 121 sg_arr[i].size = cpu_to_le32((u32)sg_dma_len(sc)); in asd_map_scatterlist() 453 cpu_to_le32((u32)sg_dma_len(&task->smp_task.smp_req)-4); in asd_build_smp_ascb() 458 cpu_to_le32((u32)sg_dma_len(&task->smp_task.smp_resp)-4); in asd_build_smp_ascb()
|
/Linux-v4.19/net/smc/ |
D | smc_ib.c | 402 if (!sg_dma_len(sg)) in smc_ib_sync_sg_for_cpu() 406 sg_dma_len(sg), in smc_ib_sync_sg_for_cpu() 422 if (!sg_dma_len(sg)) in smc_ib_sync_sg_for_device() 426 sg_dma_len(sg), in smc_ib_sync_sg_for_device()
|