Home
last modified time | relevance | path

Searched refs:dma_tx (Results 1 – 25 of 40) sorted by relevance

12

/Linux-v5.4/drivers/spi/
Dspi-pxa2xx-dma.c102 chan = drv_data->controller->dma_tx; in pxa2xx_spi_dma_prepare_one()
131 dmaengine_terminate_async(drv_data->controller->dma_tx); in pxa2xx_spi_dma_transfer()
171 dmaengine_terminate_async(drv_data->controller->dma_tx); in pxa2xx_spi_dma_prepare()
179 dma_async_issue_pending(drv_data->controller->dma_tx); in pxa2xx_spi_dma_start()
188 dmaengine_terminate_sync(drv_data->controller->dma_tx); in pxa2xx_spi_dma_stop()
201 controller->dma_tx = dma_request_slave_channel_compat(mask, in pxa2xx_spi_dma_setup()
203 if (!controller->dma_tx) in pxa2xx_spi_dma_setup()
209 dma_release_channel(controller->dma_tx); in pxa2xx_spi_dma_setup()
210 controller->dma_tx = NULL; in pxa2xx_spi_dma_setup()
226 if (controller->dma_tx) { in pxa2xx_spi_dma_release()
[all …]
Dspi-at91-usart.c135 ctlr->dma_tx = dma_request_slave_channel_reason(dev, "tx"); in at91_usart_spi_configure_dma()
136 if (IS_ERR_OR_NULL(ctlr->dma_tx)) { in at91_usart_spi_configure_dma()
137 if (IS_ERR(ctlr->dma_tx)) { in at91_usart_spi_configure_dma()
138 err = PTR_ERR(ctlr->dma_tx); in at91_usart_spi_configure_dma()
178 if (dmaengine_slave_config(ctlr->dma_tx, &slave_config)) { in at91_usart_spi_configure_dma()
189 if (!IS_ERR_OR_NULL(ctlr->dma_tx)) in at91_usart_spi_configure_dma()
190 dma_release_channel(ctlr->dma_tx); in at91_usart_spi_configure_dma()
193 ctlr->dma_tx = NULL; in at91_usart_spi_configure_dma()
204 if (ctlr->dma_tx) in at91_usart_spi_release_dma()
205 dma_release_channel(ctlr->dma_tx); in at91_usart_spi_release_dma()
[all …]
Dspi-fsl-lpspi.c380 ret = dmaengine_slave_config(controller->dma_tx, &tx); in fsl_lpspi_dma_configure()
591 desc_tx = dmaengine_prep_slave_sg(controller->dma_tx, in fsl_lpspi_dma_transfer()
595 dmaengine_terminate_all(controller->dma_tx); in fsl_lpspi_dma_transfer()
603 dma_async_issue_pending(controller->dma_tx); in fsl_lpspi_dma_transfer()
616 dmaengine_terminate_all(controller->dma_tx); in fsl_lpspi_dma_transfer()
626 dmaengine_terminate_all(controller->dma_tx); in fsl_lpspi_dma_transfer()
636 dmaengine_terminate_all(controller->dma_tx); in fsl_lpspi_dma_transfer()
646 dmaengine_terminate_all(controller->dma_tx); in fsl_lpspi_dma_transfer()
665 if (controller->dma_tx) { in fsl_lpspi_dma_exit()
666 dma_release_channel(controller->dma_tx); in fsl_lpspi_dma_exit()
[all …]
Dspi-rockchip.c261 dmaengine_terminate_async(master->dma_tx); in rockchip_spi_handle_err()
423 dmaengine_slave_config(master->dma_tx, &txconf); in rockchip_spi_prepare_dma()
426 master->dma_tx, in rockchip_spi_prepare_dma()
451 dma_async_issue_pending(master->dma_tx); in rockchip_spi_prepare_dma()
691 master->dma_tx = dma_request_chan(rs->dev, "tx"); in rockchip_spi_probe()
692 if (IS_ERR(master->dma_tx)) { in rockchip_spi_probe()
694 if (PTR_ERR(master->dma_tx) == -EPROBE_DEFER) { in rockchip_spi_probe()
699 master->dma_tx = NULL; in rockchip_spi_probe()
712 if (master->dma_tx && master->dma_rx) { in rockchip_spi_probe()
730 if (master->dma_tx) in rockchip_spi_probe()
[all …]
Dspi-bcm2835.c604 dmaengine_terminate_async(ctlr->dma_tx); in bcm2835_spi_dma_rx_done()
676 chan = ctlr->dma_tx; in bcm2835_spi_prepare_sg()
797 dma_async_issue_pending(ctlr->dma_tx); in bcm2835_spi_transfer_one_dma()
811 dmaengine_terminate_sync(ctlr->dma_tx); in bcm2835_spi_transfer_one_dma()
857 if (ctlr->dma_tx) { in bcm2835_dma_release()
858 dmaengine_terminate_sync(ctlr->dma_tx); in bcm2835_dma_release()
864 dma_unmap_page_attrs(ctlr->dma_tx->device->dev, in bcm2835_dma_release()
869 dma_release_channel(ctlr->dma_tx); in bcm2835_dma_release()
870 ctlr->dma_tx = NULL; in bcm2835_dma_release()
908 ctlr->dma_tx = dma_request_slave_channel(dev, "tx"); in bcm2835_dma_init()
[all …]
Dspi-pic32.c309 if (!master->dma_rx || !master->dma_tx) in pic32_spi_dma_transfer()
322 desc_tx = dmaengine_prep_slave_sg(master->dma_tx, in pic32_spi_dma_transfer()
347 dma_async_issue_pending(master->dma_tx); in pic32_spi_dma_transfer()
374 ret = dmaengine_slave_config(master->dma_tx, &cfg); in pic32_spi_dma_config()
554 dmaengine_terminate_all(master->dma_tx); in pic32_spi_one_transfer()
624 master->dma_tx = dma_request_slave_channel_compat(mask, NULL, NULL, in pic32_spi_dma_prep()
626 if (!master->dma_tx) { in pic32_spi_dma_prep()
643 if (master->dma_tx) in pic32_spi_dma_prep()
644 dma_release_channel(master->dma_tx); in pic32_spi_dma_prep()
656 if (pic32s->master->dma_tx) in pic32_spi_dma_unprep()
[all …]
Dspi-omap2-mcspi.c94 struct dma_chan *dma_tx; member
404 if (mcspi_dma->dma_tx) { in omap2_mcspi_tx_dma()
407 dmaengine_slave_config(mcspi_dma->dma_tx, &cfg); in omap2_mcspi_tx_dma()
409 tx = dmaengine_prep_slave_sg(mcspi_dma->dma_tx, xfer->tx_sg.sgl, in omap2_mcspi_tx_dma()
421 dma_async_issue_pending(mcspi_dma->dma_tx); in omap2_mcspi_tx_dma()
644 dmaengine_terminate_sync(mcspi_dma->dma_tx); in omap2_mcspi_txrx_dma()
1009 mcspi_dma->dma_tx = dma_request_chan(&master->dev, in omap2_mcspi_request_dma()
1011 if (IS_ERR(mcspi_dma->dma_tx)) { in omap2_mcspi_request_dma()
1012 ret = PTR_ERR(mcspi_dma->dma_tx); in omap2_mcspi_request_dma()
1013 mcspi_dma->dma_tx = NULL; in omap2_mcspi_request_dma()
[all …]
Dspi-davinci.c121 struct dma_chan *dma_tx; member
401 if (dspi->dma_rx && dspi->dma_tx) in davinci_spi_of_setup()
628 dmaengine_slave_config(dspi->dma_tx, &dma_tx_conf); in davinci_spi_bufs()
646 txdesc = dmaengine_prep_slave_sg(dspi->dma_tx, in davinci_spi_bufs()
664 dma_async_issue_pending(dspi->dma_tx); in davinci_spi_bufs()
758 dspi->dma_tx = dma_request_chan(sdev, "tx"); in davinci_spi_request_dma()
759 if (IS_ERR(dspi->dma_tx)) { in davinci_spi_request_dma()
761 return PTR_ERR(dspi->dma_tx); in davinci_spi_request_dma()
975 dspi->dma_tx = NULL; in davinci_spi_probe()
1013 dma_release_channel(dspi->dma_tx); in davinci_spi_probe()
[all …]
Dspi-ep93xx.c93 struct dma_chan *dma_tx; member
299 chan = espi->dma_tx; in ep93xx_spi_dma_prepare()
383 chan = espi->dma_tx; in ep93xx_spi_dma_finish()
427 dma_async_issue_pending(espi->dma_tx); in ep93xx_spi_dma_transfer()
610 espi->dma_tx = dma_request_channel(mask, ep93xx_spi_dma_filter, in ep93xx_spi_setup_dma()
612 if (!espi->dma_tx) { in ep93xx_spi_setup_dma()
634 if (espi->dma_tx) { in ep93xx_spi_release_dma()
635 dma_release_channel(espi->dma_tx); in ep93xx_spi_release_dma()
Dspi-stm32.c314 struct dma_chan *dma_tx; member
677 if (spi->cur_usedma && spi->dma_tx) in stm32f4_spi_disable()
678 dmaengine_terminate_all(spi->dma_tx); in stm32f4_spi_disable()
741 if (spi->cur_usedma && spi->dma_tx) in stm32h7_spi_disable()
742 dmaengine_terminate_all(spi->dma_tx); in stm32h7_spi_disable()
1300 if (spi->tx_buf && spi->dma_tx) { in stm32_spi_transfer_one_dma()
1302 dmaengine_slave_config(spi->dma_tx, &tx_dma_conf); in stm32_spi_transfer_one_dma()
1305 spi->dma_tx, xfer->tx_sg.sgl, in stm32_spi_transfer_one_dma()
1311 if ((spi->tx_buf && spi->dma_tx && !tx_dma_desc) || in stm32_spi_transfer_one_dma()
1342 dma_async_issue_pending(spi->dma_tx); in stm32_spi_transfer_one_dma()
[all …]
Dspi-sh-msiof.c769 dma_sync_single_for_device(p->ctlr->dma_tx->device->dev, in sh_msiof_dma_once()
771 desc_tx = dmaengine_prep_slave_single(p->ctlr->dma_tx, in sh_msiof_dma_once()
805 dma_async_issue_pending(p->ctlr->dma_tx); in sh_msiof_dma_once()
855 dmaengine_terminate_all(p->ctlr->dma_tx); in sh_msiof_dma_once()
931 while (ctlr->dma_tx && len > 15) { in sh_msiof_transfer_one()
1234 ctlr->dma_tx = sh_msiof_request_dma_chan(dev, DMA_MEM_TO_DEV, in sh_msiof_request_dma()
1236 if (!ctlr->dma_tx) in sh_msiof_request_dma()
1252 tx_dev = ctlr->dma_tx->device->dev; in sh_msiof_request_dma()
1276 dma_release_channel(ctlr->dma_tx); in sh_msiof_request_dma()
1277 ctlr->dma_tx = NULL; in sh_msiof_request_dma()
[all …]
Dspi-atmel.c495 if (dmaengine_slave_config(master->dma_tx, slave_config)) { in atmel_spi_dma_slave_config()
530 master->dma_tx = dma_request_slave_channel_reason(dev, "tx"); in atmel_spi_configure_dma()
531 if (IS_ERR(master->dma_tx)) { in atmel_spi_configure_dma()
532 err = PTR_ERR(master->dma_tx); in atmel_spi_configure_dma()
562 dma_chan_name(master->dma_tx), in atmel_spi_configure_dma()
569 if (!IS_ERR(master->dma_tx)) in atmel_spi_configure_dma()
570 dma_release_channel(master->dma_tx); in atmel_spi_configure_dma()
572 master->dma_tx = master->dma_rx = NULL; in atmel_spi_configure_dma()
580 if (master->dma_tx) in atmel_spi_stop_dma()
581 dmaengine_terminate_all(master->dma_tx); in atmel_spi_stop_dma()
[all …]
Dspi-qup.c400 chan = master->dma_tx; in spi_qup_prep_sg()
420 dmaengine_terminate_all(master->dma_tx); in spi_qup_dma_terminate()
498 dma_async_issue_pending(master->dma_tx); in spi_qup_do_dma()
905 IS_ERR_OR_NULL(master->dma_tx)) in spi_qup_can_dma()
922 if (!IS_ERR_OR_NULL(master->dma_tx)) in spi_qup_release_dma()
923 dma_release_channel(master->dma_tx); in spi_qup_release_dma()
939 master->dma_tx = dma_request_slave_channel_reason(dev, "tx"); in spi_qup_init_dma()
940 if (IS_ERR(master->dma_tx)) { in spi_qup_init_dma()
941 ret = PTR_ERR(master->dma_tx); in spi_qup_init_dma()
962 ret = dmaengine_slave_config(master->dma_tx, tx_conf); in spi_qup_init_dma()
[all …]
Dspi-dw-mid.c41 struct dw_dma_slave *tx = dws->dma_tx; in mid_spi_dma_init()
68 dws->master->dma_tx = dws->txchan; in mid_spi_dma_init()
317 dws->dma_tx = &mid_dma_tx; in dw_spi_mid_init()
Dspi-imx.c1175 ret = dmaengine_slave_config(master->dma_tx, &tx); in spi_imx_dma_configure()
1257 if (master->dma_tx) { in spi_imx_sdma_exit()
1258 dma_release_channel(master->dma_tx); in spi_imx_sdma_exit()
1259 master->dma_tx = NULL; in spi_imx_sdma_exit()
1275 master->dma_tx = dma_request_slave_channel_reason(dev, "tx"); in spi_imx_sdma_init()
1276 if (IS_ERR(master->dma_tx)) { in spi_imx_sdma_init()
1277 ret = PTR_ERR(master->dma_tx); in spi_imx_sdma_init()
1279 master->dma_tx = NULL; in spi_imx_sdma_init()
1383 desc_tx = dmaengine_prep_slave_sg(master->dma_tx, in spi_imx_dma_transfer()
1387 dmaengine_terminate_all(master->dma_tx); in spi_imx_dma_transfer()
[all …]
Dspi-rspi.c552 desc_tx = dmaengine_prep_slave_sg(rspi->ctlr->dma_tx, tx->sgl, in rspi_dma_transfer()
592 dma_async_issue_pending(rspi->ctlr->dma_tx); in rspi_dma_transfer()
604 dmaengine_terminate_all(rspi->ctlr->dma_tx); in rspi_dma_transfer()
1084 ctlr->dma_tx = rspi_request_dma_chan(dev, DMA_MEM_TO_DEV, dma_tx_id, in rspi_request_dma()
1086 if (!ctlr->dma_tx) in rspi_request_dma()
1092 dma_release_channel(ctlr->dma_tx); in rspi_request_dma()
1093 ctlr->dma_tx = NULL; in rspi_request_dma()
1104 if (ctlr->dma_tx) in rspi_release_dma()
1105 dma_release_channel(ctlr->dma_tx); in rspi_release_dma()
/Linux-v5.4/arch/powerpc/platforms/512x/
Dmpc512x_lpbfifo.c169 struct dma_async_tx_descriptor *dma_tx = NULL; in mpc512x_lpbfifo_kick() local
259 dma_tx = dmaengine_prep_slave_sg(lpbfifo.chan, &sg, in mpc512x_lpbfifo_kick()
261 if (!dma_tx) { in mpc512x_lpbfifo_kick()
265 dma_tx->callback = mpc512x_lpbfifo_callback; in mpc512x_lpbfifo_kick()
266 dma_tx->callback_param = NULL; in mpc512x_lpbfifo_kick()
315 cookie = dma_tx->tx_submit(dma_tx); in mpc512x_lpbfifo_kick()
/Linux-v5.4/drivers/i2c/busses/
Di2c-sh_mobile.c134 struct dma_chan *dma_tx; member
450 ? pd->dma_rx : pd->dma_tx; in sh_mobile_i2c_dma_unmap()
465 dmaengine_terminate_all(pd->dma_tx); in sh_mobile_i2c_cleanup_dma()
521 struct dma_chan *chan = read ? pd->dma_rx : pd->dma_tx; in sh_mobile_i2c_xfer_dma()
531 chan = pd->dma_tx = sh_mobile_i2c_request_dma_chan(pd->dev, DMA_MEM_TO_DEV, in sh_mobile_i2c_xfer_dma()
788 if (!IS_ERR(pd->dma_tx)) { in sh_mobile_i2c_release_dma()
789 dma_release_channel(pd->dma_tx); in sh_mobile_i2c_release_dma()
790 pd->dma_tx = ERR_PTR(-EPROBE_DEFER); in sh_mobile_i2c_release_dma()
879 pd->dma_rx = pd->dma_tx = ERR_PTR(-EPROBE_DEFER); in sh_mobile_i2c_probe()
Di2c-rcar.c136 struct dma_chan *dma_tx; member
366 ? priv->dma_rx : priv->dma_tx; in rcar_i2c_dma_unmap()
389 dmaengine_terminate_all(priv->dma_tx); in rcar_i2c_cleanup_dma()
409 struct dma_chan *chan = read ? priv->dma_rx : priv->dma_tx; in rcar_i2c_dma()
729 chan = read ? priv->dma_rx : priv->dma_tx; in rcar_i2c_request_dma()
739 priv->dma_tx = chan; in rcar_i2c_request_dma()
744 if (!IS_ERR(priv->dma_tx)) { in rcar_i2c_release_dma()
745 dma_release_channel(priv->dma_tx); in rcar_i2c_release_dma()
746 priv->dma_tx = ERR_PTR(-EPROBE_DEFER); in rcar_i2c_release_dma()
965 priv->dma_rx = priv->dma_tx = ERR_PTR(-EPROBE_DEFER); in rcar_i2c_probe()
/Linux-v5.4/drivers/net/ethernet/samsung/sxgbe/
Dsxgbe_dma.c44 int fix_burst, int pbl, dma_addr_t dma_tx, in sxgbe_dma_channel_init() argument
66 writel(upper_32_bits(dma_tx), in sxgbe_dma_channel_init()
68 writel(lower_32_bits(dma_tx), in sxgbe_dma_channel_init()
80 dma_addr = dma_tx + ((t_rsize - 1) * SXGBE_DESC_SIZE_BYTES); in sxgbe_dma_channel_init()
/Linux-v5.4/drivers/net/ethernet/micrel/
Dks8842.c147 #define KS8842_USE_DMA(adapter) (((adapter)->dma_tx.channel != -1) && \
159 struct ks8842_tx_dma_ctl dma_tx; member
422 struct ks8842_tx_dma_ctl *ctl = &adapter->dma_tx; in ks8842_tx_frame_dma()
845 struct ks8842_tx_dma_ctl *ctl = &adapter->dma_tx; in ks8842_dma_tx_cb()
861 struct ks8842_tx_dma_ctl *tx_ctl = &adapter->dma_tx; in ks8842_stop_dma()
883 struct ks8842_tx_dma_ctl *tx_ctl = &adapter->dma_tx; in ks8842_dealloc_dma_bufs()
915 struct ks8842_tx_dma_ctl *tx_ctl = &adapter->dma_tx; in ks8842_alloc_dma_bufs()
989 adapter->dma_tx.channel = -1; in ks8842_open()
1043 if (adapter->dma_tx.adesc) in ks8842_xmit_frame()
1170 adapter->dma_tx.channel = pdata->tx_dma_channel; in ks8842_probe()
[all …]
/Linux-v5.4/drivers/mmc/host/
Domap.c129 struct dma_chan *dma_tx; member
411 c = host->dma_tx; in mmc_omap_release_dma()
1001 c = host->dma_tx; in mmc_omap_prepare_data()
1077 host->dma_tx : host->dma_rx; in mmc_omap_start_request()
1389 host->dma_tx = dma_request_chan(&pdev->dev, "tx"); in mmc_omap_probe()
1390 if (IS_ERR(host->dma_tx)) { in mmc_omap_probe()
1391 ret = PTR_ERR(host->dma_tx); in mmc_omap_probe()
1397 host->dma_tx = NULL; in mmc_omap_probe()
1405 if (host->dma_tx) in mmc_omap_probe()
1406 dma_release_channel(host->dma_tx); in mmc_omap_probe()
[all …]
Ddavinci_mmc.c192 struct dma_chan *dma_tx; member
395 sync_dev = host->dma_tx; in davinci_abort_dma()
415 chan = host->dma_tx; in mmc_davinci_send_dma_request()
416 dmaengine_slave_config(host->dma_tx, &dma_tx_conf); in mmc_davinci_send_dma_request()
418 desc = dmaengine_prep_slave_sg(host->dma_tx, in mmc_davinci_send_dma_request()
491 dma_release_channel(host->dma_tx); in davinci_release_dma_channels()
497 host->dma_tx = dma_request_chan(mmc_dev(host->mmc), "tx"); in davinci_acquire_dma_channels()
498 if (IS_ERR(host->dma_tx)) { in davinci_acquire_dma_channels()
500 return PTR_ERR(host->dma_tx); in davinci_acquire_dma_channels()
506 dma_release_channel(host->dma_tx); in davinci_acquire_dma_channels()
/Linux-v5.4/drivers/net/ethernet/calxeda/
Dxgmac.c360 struct xgmac_dma_desc *dma_tx; member
750 priv->dma_tx = dma_alloc_coherent(priv->device, in xgmac_dma_desc_rings_init()
755 if (!priv->dma_tx) in xgmac_dma_desc_rings_init()
760 priv->dma_rx, priv->dma_tx, in xgmac_dma_desc_rings_init()
771 desc_init_tx_desc(priv->dma_tx, DMA_TX_RING_SZ); in xgmac_dma_desc_rings_init()
822 p = priv->dma_tx + i; in xgmac_free_tx_skbufs()
843 if (priv->dma_tx) { in xgmac_free_dma_desc_rings()
846 priv->dma_tx, priv->dma_tx_phy); in xgmac_free_dma_desc_rings()
847 priv->dma_tx = NULL; in xgmac_free_dma_desc_rings()
871 struct xgmac_dma_desc *p = priv->dma_tx + entry; in xgmac_tx_complete()
[all …]
/Linux-v5.4/drivers/net/ethernet/stmicro/stmmac/
Dchain_mode.c27 desc = tx_q->dma_tx + entry; in jumbo_frm()
50 desc = tx_q->dma_tx + entry; in jumbo_frm()

12