/Linux-v5.4/drivers/dma/sh/ |
D | usb-dmac.c | 55 unsigned int sg_len; member 263 static int usb_dmac_desc_alloc(struct usb_dmac_chan *chan, unsigned int sg_len, in usb_dmac_desc_alloc() argument 269 desc = kzalloc(struct_size(desc, sg, sg_len), gfp); in usb_dmac_desc_alloc() 273 desc->sg_allocated_len = sg_len; in usb_dmac_desc_alloc() 299 unsigned int sg_len, gfp_t gfp) in usb_dmac_desc_get() argument 307 if (sg_len <= desc->sg_allocated_len) { in usb_dmac_desc_get() 316 if (!usb_dmac_desc_alloc(chan, sg_len, gfp)) { in usb_dmac_desc_get() 417 unsigned int sg_len, enum dma_transfer_direction dir, in usb_dmac_prep_slave_sg() argument 425 if (!sg_len) { in usb_dmac_prep_slave_sg() 427 "%s: bad parameter: len=%d\n", __func__, sg_len); in usb_dmac_prep_slave_sg() [all …]
|
D | shdma-base.c | 561 struct scatterlist *sgl, unsigned int sg_len, dma_addr_t *addr, in shdma_prep_sg() argument 571 for_each_sg(sgl, sg, sg_len, i) in shdma_prep_sg() 588 for_each_sg(sgl, sg, sg_len, i) { in shdma_prep_sg() 662 struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, in shdma_prep_slave_sg() argument 677 if (slave_id < 0 || !sg_len) { in shdma_prep_slave_sg() 679 __func__, sg_len, slave_id); in shdma_prep_slave_sg() 685 return shdma_prep_sg(schan, sgl, sg_len, &slave_addr, in shdma_prep_slave_sg() 700 unsigned int sg_len = buf_len / period_len; in shdma_prep_dma_cyclic() local 711 if (sg_len > SHDMA_MAX_SG_LEN) { in shdma_prep_dma_cyclic() 713 sg_len, SHDMA_MAX_SG_LEN); in shdma_prep_dma_cyclic() [all …]
|
D | rcar-dmac.c | 895 unsigned int sg_len, dma_addr_t dev_addr, in rcar_dmac_chan_prep_sg() argument 930 for_each_sg(sgl, sg, sg_len, i) { in rcar_dmac_chan_prep_sg() 1167 unsigned int sg_len, enum dma_transfer_direction dir, in rcar_dmac_prep_slave_sg() argument 1173 if (rchan->mid_rid < 0 || !sg_len || !sg_dma_len(sgl)) { in rcar_dmac_prep_slave_sg() 1176 __func__, sg_len, rchan->mid_rid); in rcar_dmac_prep_slave_sg() 1183 return rcar_dmac_chan_prep_sg(rchan, sgl, sg_len, rchan->map.addr, in rcar_dmac_prep_slave_sg() 1197 unsigned int sg_len; in rcar_dmac_prep_dma_cyclic() local 1211 sg_len = buf_len / period_len; in rcar_dmac_prep_dma_cyclic() 1212 if (sg_len > RCAR_DMAC_MAX_SG_LEN) { in rcar_dmac_prep_dma_cyclic() 1215 rchan->index, sg_len, RCAR_DMAC_MAX_SG_LEN); in rcar_dmac_prep_dma_cyclic() [all …]
|
/Linux-v5.4/net/rds/ |
D | ib_frmr.c | 114 atomic_add(ibmr->sg_len, &pool->free_pinned); in rds_ib_free_frmr() 134 ret = ib_map_mr_sg_zbva(frmr->mr, ibmr->sg, ibmr->sg_len, in rds_ib_post_reg_frmr() 136 if (unlikely(ret != ibmr->sg_len)) in rds_ib_post_reg_frmr() 190 struct scatterlist *sg, unsigned int sg_len) in rds_ib_map_frmr() argument 204 ibmr->sg_len = sg_len; in rds_ib_map_frmr() 208 ibmr->sg_dma_len = ib_dma_map_sg(dev, ibmr->sg, ibmr->sg_len, in rds_ib_map_frmr() 260 ib_dma_unmap_sg(rds_ibdev->dev, ibmr->sg, ibmr->sg_len, in rds_ib_map_frmr() 379 *unpinned += ibmr->sg_len; in rds_ib_unreg_frmr()
|
D | ib_rdma.c | 234 ibmr->sg, ibmr->sg_len, in __rds_ib_teardown_mr() 240 if (ibmr->sg_len) { in __rds_ib_teardown_mr() 243 for (i = 0; i < ibmr->sg_len; ++i) { in __rds_ib_teardown_mr() 255 ibmr->sg_len = 0; in __rds_ib_teardown_mr() 261 unsigned int pinned = ibmr->sg_len; in rds_ib_teardown_mr() 483 rdsdebug("RDS/IB: free_mr nents %u\n", ibmr->sg_len); in rds_ib_free_mr() 491 atomic_add(ibmr->sg_len, &pool->free_pinned); in rds_ib_free_mr()
|
/Linux-v5.4/drivers/mmc/host/ |
D | uniphier-sd.c | 101 dma_unmap_sg(mmc_dev(host->mmc), host->sg_ptr, host->sg_len, in uniphier_sd_external_dma_callback() 130 int sg_len; in uniphier_sd_external_dma_start() local 143 sg_len = dma_map_sg(mmc_dev(host->mmc), host->sg_ptr, host->sg_len, in uniphier_sd_external_dma_start() 145 if (sg_len == 0) in uniphier_sd_external_dma_start() 148 desc = dmaengine_prep_slave_sg(priv->chan, host->sg_ptr, sg_len, in uniphier_sd_external_dma_start() 165 dma_unmap_sg(mmc_dev(host->mmc), host->sg_ptr, host->sg_len, in uniphier_sd_external_dma_start() 253 int sg_len; in uniphier_sd_internal_dma_start() local 258 if (WARN_ON(host->sg_len != 1)) in uniphier_sd_internal_dma_start() 272 sg_len = dma_map_sg(mmc_dev(host->mmc), sg, 1, priv->dma_dir); in uniphier_sd_internal_dma_start() 273 if (sg_len == 0) in uniphier_sd_internal_dma_start()
|
D | mmci_stm32_sdmmc.c | 38 for_each_sg(data->sg, sg, data->sg_len - 1, i) { in sdmmc_idma_validate_data() 65 data->sg_len, in _sdmmc_idma_prep_data() 89 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, in sdmmc_idma_unprep_data() 132 if (!host->variant->dma_lli || data->sg_len == 1) { in sdmmc_idma_start() 140 for_each_sg(data->sg, sg, data->sg_len, i) { in sdmmc_idma_start() 149 desc[data->sg_len - 1].idmalar &= ~MMCI_STM32_ULA; in sdmmc_idma_start()
|
D | android-goldfish.c | 121 unsigned int sg_len; member 221 host->sg_len, dma_data_dir); in goldfish_mmc_xfer_done() 225 host->sg_len = 0; in goldfish_mmc_xfer_done() 359 unsigned sg_len; in goldfish_mmc_prepare_data() local 379 sg_len = (data->blocks == 1) ? 1 : data->sg_len; in goldfish_mmc_prepare_data() 383 host->sg_len = dma_map_sg(mmc_dev(mmc_from_priv(host)), data->sg, in goldfish_mmc_prepare_data() 384 sg_len, dma_data_dir); in goldfish_mmc_prepare_data()
|
D | renesas_sdhi_sys_dmac.c | 135 host->sg_ptr, host->sg_len, in renesas_sdhi_sys_dmac_dma_callback() 139 host->sg_ptr, host->sg_len, in renesas_sdhi_sys_dmac_dma_callback() 163 for_each_sg(sg, sg_tmp, host->sg_len, i) { in renesas_sdhi_sys_dmac_start_dma_rx() 172 if ((!aligned && (host->sg_len > 1 || sg->length > PAGE_SIZE || in renesas_sdhi_sys_dmac_start_dma_rx() 188 ret = dma_map_sg(chan->device->dev, sg, host->sg_len, DMA_FROM_DEVICE); in renesas_sdhi_sys_dmac_start_dma_rx() 235 for_each_sg(sg, sg_tmp, host->sg_len, i) { in renesas_sdhi_sys_dmac_start_dma_tx() 244 if ((!aligned && (host->sg_len > 1 || sg->length > PAGE_SIZE || in renesas_sdhi_sys_dmac_start_dma_tx() 265 ret = dma_map_sg(chan->device->dev, sg, host->sg_len, DMA_TO_DEVICE); in renesas_sdhi_sys_dmac_start_dma_tx()
|
D | mxs-mmc.c | 148 data->sg_len, ssp->dma_dir); in mxs_mmc_request_done() 221 unsigned int sg_len; in mxs_mmc_prep_dma() local 226 data->sg_len, ssp->dma_dir); in mxs_mmc_prep_dma() 228 sg_len = data->sg_len; in mxs_mmc_prep_dma() 232 sg_len = SSP_PIO_NUM; in mxs_mmc_prep_dma() 236 sgl, sg_len, ssp->slave_dirn, flags); in mxs_mmc_prep_dma() 243 data->sg_len, ssp->dma_dir); in mxs_mmc_prep_dma() 350 unsigned int sg_len = data->sg_len; in mxs_mmc_adtc() local 395 for_each_sg(sgl, sg, sg_len, i) in mxs_mmc_adtc()
|
D | au1xmmc.c | 352 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, host->dma.dir); in au1xmmc_data_complete() 395 int sg_len, max, count; in au1xmmc_send_pio() local 410 sg_len = data->sg[host->pio.index].length - host->pio.offset; in au1xmmc_send_pio() 413 max = (sg_len > host->pio.len) ? host->pio.len : sg_len; in au1xmmc_send_pio() 433 if (count == sg_len) { in au1xmmc_send_pio() 451 int max, count, sg_len = 0; in au1xmmc_receive_pio() local 468 sg_len = sg_dma_len(&data->sg[host->pio.index]) - host->pio.offset; in au1xmmc_receive_pio() 471 if (sg_len < max) in au1xmmc_receive_pio() 472 max = sg_len; in au1xmmc_receive_pio() 512 if (sg_len && count == sg_len) { in au1xmmc_receive_pio() [all …]
|
D | tifm_sd.c | 102 int sg_len; member 171 if (host->sg_pos == host->sg_len) in tifm_sd_transfer_data() 178 if (host->sg_pos == host->sg_len) { in tifm_sd_transfer_data() 234 if (host->sg_pos == host->sg_len) in tifm_sd_bounce_block() 267 if (host->sg_pos == host->sg_len) in tifm_sd_set_dma_data() 275 if (host->sg_pos == host->sg_len) in tifm_sd_set_dma_data() 283 if (host->sg_pos == host->sg_len) in tifm_sd_set_dma_data() 663 host->sg_len = r_data->sg_len; in tifm_sd_request() 677 host->sg_len = tifm_map_sg(sock, r_data->sg, in tifm_sd_request() 678 r_data->sg_len, in tifm_sd_request() [all …]
|
D | omap.c | 146 unsigned int sg_len; member 424 dma_unmap_sg(dev, data->sg, host->sg_len, dma_data_dir); in mmc_omap_release_dma() 448 host->sg_len = 0; in mmc_omap_xfer_done() 505 host->sg_len = 0; in mmc_omap_abort_xfer() 670 BUG_ON(host->sg_idx == host->sg_len); in mmc_omap_xfer_data() 949 unsigned sg_len; in mmc_omap_prepare_data() local 970 sg_len = (data->blocks == 1) ? 1 : data->sg_len; in mmc_omap_prepare_data() 973 for_each_sg(data->sg, sg, sg_len, i) { in mmc_omap_prepare_data() 1034 host->sg_len = dma_map_sg(c->device->dev, data->sg, sg_len, in mmc_omap_prepare_data() 1036 if (host->sg_len == 0) in mmc_omap_prepare_data() [all …]
|
D | cavium.c | 377 dma_unmap_sg(host->dev, data->sg, data->sg_len, get_dma_dir(data)); in finish_dma_single() 397 dma_unmap_sg(host->dev, data->sg, data->sg_len, get_dma_dir(data)); in finish_dma_sg() 403 if (host->use_sg && data->sg_len > 1) in finish_dma() 522 count = dma_map_sg(host->dev, data->sg, data->sg_len, in prepare_dma_single() 559 count = dma_map_sg(host->dev, data->sg, data->sg_len, in prepare_dma_sg() 612 dma_unmap_sg(host->dev, data->sg, data->sg_len, get_dma_dir(data)); in prepare_dma_sg() 620 if (host->use_sg && data->sg_len > 1) in prepare_dma() 658 if (!mrq->data || !mrq->data->sg || !mrq->data->sg_len || in cvm_mmc_dma_request() 711 sg_miter_start(&host->smi, mrq->data->sg, mrq->data->sg_len, in do_read_request() 724 sg_miter_start(smi, mrq->data->sg, mrq->data->sg_len, SG_MITER_FROM_SG); in do_write_request()
|
D | dw_mmc.c | 438 data->sg_len, in dw_mci_dma_cleanup() 481 data->sg_len, in dw_mci_dmac_complete_dma() 576 unsigned int sg_len) in dw_mci_prepare_desc64() argument 585 for (i = 0; i < sg_len; i++) { in dw_mci_prepare_desc64() 648 unsigned int sg_len) in dw_mci_prepare_desc32() argument 657 for (i = 0; i < sg_len; i++) { in dw_mci_prepare_desc32() 719 static int dw_mci_idmac_start_dma(struct dw_mci *host, unsigned int sg_len) in dw_mci_idmac_start_dma() argument 725 ret = dw_mci_prepare_desc64(host, host->data, sg_len); in dw_mci_idmac_start_dma() 727 ret = dw_mci_prepare_desc32(host, host->data, sg_len); in dw_mci_idmac_start_dma() 773 unsigned int sg_len) in dw_mci_edmac_start_dma() argument [all …]
|
D | renesas_sdhi_internal_dmac.c | 185 if (!dma_map_sg(&host->pdev->dev, sg, host->sg_len, in renesas_sdhi_internal_dmac_start_dma() 215 dma_unmap_sg(&host->pdev->dev, sg, host->sg_len, mmc_get_dma_dir(data)); in renesas_sdhi_internal_dmac_start_dma() 248 dma_unmap_sg(&host->pdev->dev, host->sg_ptr, host->sg_len, dir); in renesas_sdhi_internal_dmac_complete_tasklet_fn()
|
/Linux-v5.4/drivers/mmc/core/ |
D | mmc_test.c | 82 unsigned int sg_len; member 161 unsigned int *sg_len; member 209 struct mmc_request *mrq, struct scatterlist *sg, unsigned sg_len, in mmc_test_prepare_mrq() argument 241 mrq->data->sg_len = sg_len; in mmc_test_prepare_mrq() 410 unsigned int *sg_len, int min_sg_len) in mmc_test_map_sg() argument 420 *sg_len = 0; in mmc_test_map_sg() 439 *sg_len += 1; in mmc_test_map_sg() 463 unsigned int *sg_len) in mmc_test_map_sg_max_scatter() argument 472 *sg_len = 0; in mmc_test_map_sg_max_scatter() 494 *sg_len += 1; in mmc_test_map_sg_max_scatter() [all …]
|
D | sdio_ops.c | 157 data.sg_len = nents; in mmc_io_rw_extended() 159 for_each_sg(data.sg, sg_ptr, data.sg_len, i) { in mmc_io_rw_extended() 166 data.sg_len = 1; in mmc_io_rw_extended()
|
/Linux-v5.4/drivers/dma/ |
D | fsl-edma-common.c | 415 int sg_len) in fsl_edma_alloc_desc() argument 420 fsl_desc = kzalloc(struct_size(fsl_desc, tcd, sg_len), GFP_NOWAIT); in fsl_edma_alloc_desc() 425 fsl_desc->n_tcds = sg_len; in fsl_edma_alloc_desc() 426 for (i = 0; i < sg_len; i++) { in fsl_edma_alloc_desc() 450 int sg_len, i; in fsl_edma_prep_dma_cyclic() local 460 sg_len = buf_len / period_len; in fsl_edma_prep_dma_cyclic() 461 fsl_desc = fsl_edma_alloc_desc(fsl_chan, sg_len); in fsl_edma_prep_dma_cyclic() 482 for (i = 0; i < sg_len; i++) { in fsl_edma_prep_dma_cyclic() 487 last_sg = fsl_desc->tcd[(i + 1) % sg_len].ptcd; in fsl_edma_prep_dma_cyclic() 513 unsigned int sg_len, enum dma_transfer_direction direction, in fsl_edma_prep_slave_sg() argument [all …]
|
D | st_fdma.c | 237 int sg_len) in st_fdma_alloc_desc() argument 242 fdesc = kzalloc(struct_size(fdesc, node, sg_len), GFP_NOWAIT); in st_fdma_alloc_desc() 247 fdesc->n_nodes = sg_len; in st_fdma_alloc_desc() 248 for (i = 0; i < sg_len; i++) { in st_fdma_alloc_desc() 453 int sg_len, i; in st_fdma_prep_dma_cyclic() local 473 sg_len = len / period_len; in st_fdma_prep_dma_cyclic() 474 fdesc = st_fdma_alloc_desc(fchan, sg_len); in st_fdma_prep_dma_cyclic() 482 for (i = 0; i < sg_len; i++) { in st_fdma_prep_dma_cyclic() 485 hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc; in st_fdma_prep_dma_cyclic() 507 unsigned int sg_len, enum dma_transfer_direction direction, in st_fdma_prep_slave_sg() argument [all …]
|
D | uniphier-mdmac.c | 54 unsigned int sg_len; member 209 if (md->sg_cur >= md->sg_len) { in uniphier_mdmac_interrupt() 231 unsigned int sg_len, in uniphier_mdmac_prep_slave_sg() argument 246 md->sg_len = sg_len; in uniphier_mdmac_prep_slave_sg() 319 for (i = md->sg_cur; i < md->sg_len; i++) in uniphier_mdmac_tx_status()
|
D | ste_dma40_ll.c | 266 int sg_len, in d40_phy_sg_to_lli() argument 284 for_each_sg(sg, current_sg, sg_len, i) { in d40_phy_sg_to_lli() 291 if (i == sg_len - 1) in d40_phy_sg_to_lli() 418 int sg_len, in d40_log_sg_to_lli() argument 433 for_each_sg(sg, current_sg, sg_len, i) { in d40_log_sg_to_lli()
|
D | mxs-dma.c | 496 unsigned int sg_len, enum dma_transfer_direction direction, in mxs_dma_prep_slave_sg() argument 510 if (sg_len + idx > NUM_CCW) { in mxs_dma_prep_slave_sg() 513 sg_len, NUM_CCW); in mxs_dma_prep_slave_sg() 539 for (j = 0; j < sg_len;) in mxs_dma_prep_slave_sg() 549 ccw->bits |= BF_CCW(sg_len, PIO_NUM); in mxs_dma_prep_slave_sg() 554 for_each_sg(sgl, sg, sg_len, i) { in mxs_dma_prep_slave_sg() 575 if (i + 1 == sg_len) { in mxs_dma_prep_slave_sg()
|
/Linux-v5.4/drivers/crypto/nx/ |
D | nx.c | 83 unsigned int sg_len = 0; in nx_build_sg_list() local 108 for (sg = sg_head; sg_len < *len; sg++) { in nx_build_sg_list() 117 sg_len += sg->len; in nx_build_sg_list() 120 is_vmalloc_addr(start_addr + sg_len)) { in nx_build_sg_list() 122 start_addr + sg_len)); in nx_build_sg_list() 123 end_addr = sg_addr + *len - sg_len; in nx_build_sg_list() 133 *len = sg_len; in nx_build_sg_list()
|
/Linux-v5.4/drivers/crypto/chelsio/ |
D | chcr_algo.c | 1510 param->sg_len) <= SGE_MAX_WR_LEN; in create_hash_wr() 1511 nents = sg_nents_xlen(req_ctx->hctx_wr.srcsg, param->sg_len, in create_hash_wr() 1515 param->sg_len, 16) : (sgl_len(nents) * 8); in create_hash_wr() 1525 chcr_req->sec_cpl.pldlen = htonl(param->bfr_len + param->sg_len); in create_hash_wr() 1571 (param->sg_len + param->bfr_len) : 0); in create_hash_wr() 1630 params.sg_len = chcr_hash_ent_in_wr(req->src, !!req_ctx->reqlen, in chcr_ahash_update() 1632 if (params.sg_len > req->nbytes) in chcr_ahash_update() 1633 params.sg_len = req->nbytes; in chcr_ahash_update() 1634 params.sg_len = rounddown(params.sg_len + req_ctx->reqlen, bs) - in chcr_ahash_update() 1644 req_ctx->data_len += params.sg_len + params.bfr_len; in chcr_ahash_update() [all …]
|