/Linux-v5.4/drivers/parisc/ |
D | iommu-helpers.h | 22 unsigned long dma_offset = 0, dma_len = 0; in iommu_fill_pdir() local 47 BUG_ON(pdirp && (dma_len != sg_dma_len(dma_sg))); in iommu_fill_pdir() 51 dma_len = sg_dma_len(startsg); in iommu_fill_pdir() 106 unsigned long dma_offset, dma_len; /* start/len of DMA stream */ in iommu_coalesce_chunks() local 120 dma_len = startsg->length; in iommu_coalesce_chunks() 149 if (unlikely(ALIGN(dma_len + dma_offset + startsg->length, IOVP_SIZE) > in iommu_coalesce_chunks() 163 dma_len += startsg->length; in iommu_coalesce_chunks() 171 sg_dma_len(contig_sg) = dma_len; in iommu_coalesce_chunks() 172 dma_len = ALIGN(dma_len + dma_offset, IOVP_SIZE); in iommu_coalesce_chunks() 175 | (iommu_alloc_range(ioc, dev, dma_len) << IOVP_SHIFT) in iommu_coalesce_chunks()
|
/Linux-v5.4/drivers/net/wireless/broadcom/brcm80211/brcmsmac/ |
D | brcms_trace_brcmsmac_tx.h | 76 u16 dma_len), 78 dma_len), 85 __field(u16, dma_len) 93 __entry->dma_len = dma_len; 98 __entry->ampdu_frames, __entry->dma_len)
|
D | ampdu.h | 37 u16 dma_len; member
|
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | xdp.h | 169 u32 dma_len = xdptxd->len; in mlx5e_xdp_mpwqe_add_dseg() local 173 if (session->inline_on && dma_len <= MLX5E_XDP_INLINE_WQE_SZ_THRSD) { in mlx5e_xdp_mpwqe_add_dseg() 176 u16 ds_len = sizeof(*inline_dseg) + dma_len; in mlx5e_xdp_mpwqe_add_dseg() 179 inline_dseg->byte_count = cpu_to_be32(dma_len | MLX5_INLINE_SEG); in mlx5e_xdp_mpwqe_add_dseg() 180 memcpy(inline_dseg->data, xdptxd->data, dma_len); in mlx5e_xdp_mpwqe_add_dseg() 188 dseg->byte_count = cpu_to_be32(dma_len); in mlx5e_xdp_mpwqe_add_dseg()
|
D | xdp.c | 313 u32 dma_len = xdptxd->len; in mlx5e_xmit_xdp_frame() local 319 if (unlikely(dma_len < MLX5E_XDP_MIN_INLINE || sq->hw_mtu < dma_len)) { in mlx5e_xmit_xdp_frame() 335 dma_len -= MLX5E_XDP_MIN_INLINE; in mlx5e_xmit_xdp_frame() 342 dseg->byte_count = cpu_to_be32(dma_len); in mlx5e_xmit_xdp_frame()
|
/Linux-v5.4/drivers/i2c/busses/ |
D | i2c-stm32.c | 84 dma->dma_len = 0; in stm32_i2c_dma_free() 114 dma->dma_len = len; in stm32_i2c_prep_dma_xfer() 117 dma->dma_buf = dma_map_single(chan_dev, buf, dma->dma_len, in stm32_i2c_prep_dma_xfer() 125 dma->dma_len, in stm32_i2c_prep_dma_xfer() 149 dma_unmap_single(chan_dev, dma->dma_buf, dma->dma_len, in stm32_i2c_prep_dma_xfer()
|
D | i2c-stm32.h | 41 unsigned int dma_len; member
|
D | i2c-imx.c | 181 unsigned int dma_len; member 354 dma->dma_len, dma->dma_data_dir); in i2c_imx_dma_callback() 367 dma->dma_len, dma->dma_data_dir); in i2c_imx_dma_xfer() 374 dma->dma_len, dma->dma_transfer_dir, in i2c_imx_dma_xfer() 396 dma->dma_len, dma->dma_data_dir); in i2c_imx_dma_xfer() 406 dma->dma_len = 0; in i2c_imx_dma_free() 623 dma->dma_len = msgs->len - 1; in i2c_imx_dma_write() 687 dma->dma_len = msgs->len - 2; in i2c_imx_dma_read()
|
/Linux-v5.4/net/rds/ |
D | ib_fmr.c | 122 unsigned int dma_len = sg_dma_len(&scat[i]); in rds_ib_map_fmr() local 134 if ((dma_addr + dma_len) & ~PAGE_MASK) { in rds_ib_map_fmr() 144 len += dma_len; in rds_ib_map_fmr() 162 unsigned int dma_len = sg_dma_len(&scat[i]); in rds_ib_map_fmr() local 165 for (j = 0; j < dma_len; j += PAGE_SIZE) in rds_ib_map_fmr()
|
D | ib_frmr.c | 221 unsigned int dma_len = sg_dma_len(&ibmr->sg[i]); in rds_ib_map_frmr() local 224 frmr->sg_byte_len += dma_len; in rds_ib_map_frmr() 232 if ((dma_addr + dma_len) & ~PAGE_MASK) { in rds_ib_map_frmr() 239 len += dma_len; in rds_ib_map_frmr()
|
/Linux-v5.4/drivers/net/ethernet/sfc/ |
D | tx_tso.c | 107 unsigned int dma_len; in efx_tx_queue_insert() local 121 dma_len = tx_queue->efx->type->tx_limit_len(tx_queue, in efx_tx_queue_insert() 125 if (dma_len >= len) in efx_tx_queue_insert() 128 buffer->len = dma_len; in efx_tx_queue_insert() 130 dma_addr += dma_len; in efx_tx_queue_insert() 131 len -= dma_len; in efx_tx_queue_insert()
|
/Linux-v5.4/arch/arm/mach-rpc/ |
D | dma.c | 33 unsigned int dma_len; member 62 end = offset + idma->dma_len; in iomd_get_next_sg() 72 idma->dma_len -= end - offset; in iomd_get_next_sg() 75 if (idma->dma_len == 0) { in iomd_get_next_sg() 79 idma->dma_len = idma->dma.sg->length; in iomd_get_next_sg() 179 idma->dma_len = idma->dma.sg->length; in iomd_enable_dma()
|
/Linux-v5.4/drivers/scsi/ |
D | am53c974.c | 243 static u32 pci_esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len) in pci_esp_dma_length_limit() argument 255 if (dma_len > (1U << dma_limit)) in pci_esp_dma_length_limit() 256 dma_len = (1U << dma_limit); in pci_esp_dma_length_limit() 262 end = base + dma_len; in pci_esp_dma_length_limit() 265 dma_len = end - base; in pci_esp_dma_length_limit() 267 return dma_len; in pci_esp_dma_length_limit()
|
D | mac53c94.c | 368 u32 dma_len; in set_dma_cmds() local 382 dma_len = sg_dma_len(scl); in set_dma_cmds() 383 if (dma_len > 0xffff) in set_dma_cmds() 385 total += dma_len; in set_dma_cmds() 386 dcmds->req_count = cpu_to_le16(dma_len); in set_dma_cmds()
|
D | mac_esp.c | 247 static u32 mac_esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len) in mac_esp_dma_length_limit() argument 249 return dma_len > 0xFFFF ? 0xFFFF : dma_len; in mac_esp_dma_length_limit()
|
D | esp_scsi.c | 519 static u32 esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len) in esp_dma_length_limit() argument 523 if (dma_len > (1U << 24)) in esp_dma_length_limit() 524 dma_len = (1U << 24); in esp_dma_length_limit() 534 if (dma_len > (1U << 16)) in esp_dma_length_limit() 535 dma_len = (1U << 16); in esp_dma_length_limit() 541 end = base + dma_len; in esp_dma_length_limit() 544 dma_len = end - base; in esp_dma_length_limit() 546 return dma_len; in esp_dma_length_limit() 1747 unsigned int dma_len = esp_cur_dma_len(ent, cmd); in esp_process_event() local 1758 dma_len = esp->ops->dma_length_limit(esp, dma_addr, in esp_process_event() [all …]
|
/Linux-v5.4/drivers/staging/wusbcore/host/whci/ |
D | qset.c | 452 size_t dma_len; in qset_add_urb_sg() local 480 dma_len = dma_remaining; in qset_add_urb_sg() 488 if (std->len + dma_len > QTD_MAX_XFER_SIZE) { in qset_add_urb_sg() 489 dma_len = (QTD_MAX_XFER_SIZE / qset->max_packet) in qset_add_urb_sg() 493 std->len += dma_len; in qset_add_urb_sg() 497 ep = dma_addr + dma_len; in qset_add_urb_sg() 517 dma_remaining -= dma_len; in qset_add_urb_sg() 518 remaining -= dma_len; in qset_add_urb_sg()
|
/Linux-v5.4/drivers/net/ethernet/chelsio/cxgb/ |
D | sge.c | 165 DEFINE_DMA_UNMAP_LEN(dma_len); 171 DEFINE_DMA_UNMAP_LEN(dma_len); 513 dma_unmap_len(ce, dma_len), in free_freelQ_buffers() 626 if (likely(dma_unmap_len(ce, dma_len))) { in free_cmdQ_buffers() 628 dma_unmap_len(ce, dma_len), in free_cmdQ_buffers() 829 unsigned int dma_len = q->rx_buffer_size - q->dma_offset; in refill_free_list() local 840 mapping = pci_map_single(pdev, skb->data, dma_len, in refill_free_list() 846 dma_unmap_len_set(ce, dma_len, dma_len); in refill_free_list() 849 e->len_gen = V_CMD_LEN(dma_len) | V_CMD_GEN1(q->genbit); in refill_free_list() 1054 dma_unmap_len(ce, dma_len), in get_packet() [all …]
|
/Linux-v5.4/drivers/net/wireless/mediatek/mt7601u/ |
D | dma.c | 116 u16 dma_len = get_unaligned_le16(data); in mt7601u_rx_next_seg_len() local 119 WARN_ON_ONCE(!dma_len) || in mt7601u_rx_next_seg_len() 120 WARN_ON_ONCE(dma_len + MT_DMA_HDRS > data_len) || in mt7601u_rx_next_seg_len() 121 WARN_ON_ONCE(dma_len & 0x3)) in mt7601u_rx_next_seg_len() 124 return MT_DMA_HDRS + dma_len; in mt7601u_rx_next_seg_len()
|
/Linux-v5.4/drivers/misc/mei/ |
D | client.c | 1557 u32 dma_len; in mei_cl_irq_write() local 1601 } else if (dr_slots && hbuf_len >= hdr_len + sizeof(dma_len)) { in mei_cl_irq_write() 1608 mei_hdr.length = sizeof(dma_len); in mei_cl_irq_write() 1609 dma_len = len; in mei_cl_irq_write() 1610 data = &dma_len; in mei_cl_irq_write() 1665 u32 dma_len; in mei_cl_write() local 1728 } else if (dr_slots && hbuf_len >= hdr_len + sizeof(dma_len)) { in mei_cl_write() 1735 mei_hdr.length = sizeof(dma_len); in mei_cl_write() 1736 dma_len = len; in mei_cl_write() 1737 data = &dma_len; in mei_cl_write()
|
/Linux-v5.4/drivers/net/ethernet/sfc/falcon/ |
D | tx.c | 186 unsigned int dma_len; in ef4_tx_map_chunk() local 191 dma_len = nic_type->tx_limit_len(tx_queue, dma_addr, len); in ef4_tx_map_chunk() 193 buffer->len = dma_len; in ef4_tx_map_chunk() 196 len -= dma_len; in ef4_tx_map_chunk() 197 dma_addr += dma_len; in ef4_tx_map_chunk()
|
/Linux-v5.4/drivers/ata/ |
D | pata_arasan_cf.c | 434 u32 xfer_cnt, sglen, dma_len, xfer_ctr; in sg_xfer() local 474 dma_len = min(xfer_cnt, FIFO_SIZE); in sg_xfer() 475 ret = dma_xfer(acdev, src, dest, dma_len); in sg_xfer() 482 src += dma_len; in sg_xfer() 484 dest += dma_len; in sg_xfer() 486 sglen -= dma_len; in sg_xfer() 487 xfer_cnt -= dma_len; in sg_xfer()
|
/Linux-v5.4/drivers/crypto/ |
D | talitos.c | 335 return ((struct talitos_desc *)(edesc->buf + edesc->dma_len))->hdr1; in get_request_hdr() 495 (edesc->buf + edesc->dma_len))->hdr; 1011 if (edesc->dma_len) in ipsec_esp_unmap() 1012 dma_unmap_single(dev, edesc->dma_link_tbl, edesc->dma_len, in ipsec_esp_unmap() 1062 oicv = edesc->buf + edesc->dma_len; in ipsec_esp_decrypt_swauth_done() 1217 dma_addr_t dma_icv = edesc->dma_link_tbl + edesc->dma_len - authsize; in ipsec_esp() 1307 edesc->dma_len, in ipsec_esp() 1334 int src_nents, dst_nents, alloc_len, dma_len, src_len, dst_len; in talitos_edesc_alloc() local 1382 dma_len = (src_nents ? src_len : 0) + in talitos_edesc_alloc() 1385 dma_len = (src_nents + dst_nents + 2) * in talitos_edesc_alloc() [all …]
|
/Linux-v5.4/drivers/block/rsxx/ |
D | dma.c | 633 unsigned int dma_len, in rsxx_queue_dma() argument 649 dma->sub_page.cnt = (dma_len >> 9); in rsxx_queue_dma() 680 unsigned int dma_len; in rsxx_dma_queue_bio() local 720 dma_len = min(bv_len, in rsxx_dma_queue_bio() 725 dma_off, dma_len, in rsxx_dma_queue_bio() 733 addr8 += dma_len; in rsxx_dma_queue_bio() 734 bv_off += dma_len; in rsxx_dma_queue_bio() 735 bv_len -= dma_len; in rsxx_dma_queue_bio()
|
/Linux-v5.4/arch/ia64/hp/common/ |
D | sba_iommu.c | 1302 unsigned long dma_offset, dma_len; /* start/len of DMA stream */ in sba_coalesce_chunks() local 1314 dma_len = vcontig_len = vcontig_end = startsg->length; in sba_coalesce_chunks() 1341 if (((dma_len + dma_offset + startsg->length + ~iovp_mask) & iovp_mask) in sba_coalesce_chunks() 1345 if (dma_len + startsg->length > max_seg_size) in sba_coalesce_chunks() 1358 dma_len += startsg->length; in sba_coalesce_chunks() 1389 dma_len += vcontig_len; in sba_coalesce_chunks() 1402 dma_len = (dma_len + dma_offset + ~iovp_mask) & iovp_mask; in sba_coalesce_chunks() 1403 ASSERT(dma_len <= DMA_CHUNK_SIZE); in sba_coalesce_chunks() 1404 idx = sba_alloc_range(ioc, dev, dma_len); in sba_coalesce_chunks()
|