Home
last modified time | relevance | path

Searched refs:dma_rx (Results 1 – 25 of 42) sorted by relevance

12

/Linux-v5.4/drivers/spi/
Dspi-pxa2xx-dma.c109 chan = drv_data->controller->dma_rx; in pxa2xx_spi_dma_prepare_one()
130 dmaengine_terminate_async(drv_data->controller->dma_rx); in pxa2xx_spi_dma_transfer()
178 dma_async_issue_pending(drv_data->controller->dma_rx); in pxa2xx_spi_dma_start()
187 dmaengine_terminate_sync(drv_data->controller->dma_rx); in pxa2xx_spi_dma_stop()
206 controller->dma_rx = dma_request_slave_channel_compat(mask, in pxa2xx_spi_dma_setup()
208 if (!controller->dma_rx) { in pxa2xx_spi_dma_setup()
221 if (controller->dma_rx) { in pxa2xx_spi_dma_release()
222 dmaengine_terminate_sync(controller->dma_rx); in pxa2xx_spi_dma_release()
223 dma_release_channel(controller->dma_rx); in pxa2xx_spi_dma_release()
224 controller->dma_rx = NULL; in pxa2xx_spi_dma_release()
Dspi-at91-usart.c148 ctlr->dma_rx = dma_request_slave_channel_reason(dev, "rx"); in at91_usart_spi_configure_dma()
149 if (IS_ERR_OR_NULL(ctlr->dma_rx)) { in at91_usart_spi_configure_dma()
150 if (IS_ERR(ctlr->dma_rx)) { in at91_usart_spi_configure_dma()
151 err = PTR_ERR(ctlr->dma_rx); in at91_usart_spi_configure_dma()
170 if (dmaengine_slave_config(ctlr->dma_rx, &slave_config)) { in at91_usart_spi_configure_dma()
191 if (!IS_ERR_OR_NULL(ctlr->dma_rx)) in at91_usart_spi_configure_dma()
192 dma_release_channel(ctlr->dma_rx); in at91_usart_spi_configure_dma()
194 ctlr->dma_rx = NULL; in at91_usart_spi_configure_dma()
202 if (ctlr->dma_rx) in at91_usart_spi_release_dma()
203 dma_release_channel(ctlr->dma_rx); in at91_usart_spi_release_dma()
[all …]
Dspi-bcm2835.c517 dma_sync_single_for_device(ctlr->dma_rx->device->dev, in bcm2835_spi_transfer_prologue()
641 dmaengine_terminate_async(ctlr->dma_rx); in bcm2835_spi_dma_tx_done()
682 chan = ctlr->dma_rx; in bcm2835_spi_prepare_sg()
817 dma_async_issue_pending(ctlr->dma_rx); in bcm2835_spi_transfer_one_dma()
827 dmaengine_terminate_async(ctlr->dma_rx); in bcm2835_spi_transfer_one_dma()
873 if (ctlr->dma_rx) { in bcm2835_dma_release()
874 dmaengine_terminate_sync(ctlr->dma_rx); in bcm2835_dma_release()
881 dma_unmap_single(ctlr->dma_rx->device->dev, in bcm2835_dma_release()
886 dma_release_channel(ctlr->dma_rx); in bcm2835_dma_release()
887 ctlr->dma_rx = NULL; in bcm2835_dma_release()
[all …]
Dspi-rockchip.c264 dmaengine_terminate_async(master->dma_rx); in rockchip_spi_handle_err()
401 dmaengine_slave_config(master->dma_rx, &rxconf); in rockchip_spi_prepare_dma()
404 master->dma_rx, in rockchip_spi_prepare_dma()
431 dmaengine_terminate_sync(master->dma_rx); in rockchip_spi_prepare_dma()
443 dma_async_issue_pending(master->dma_rx); in rockchip_spi_prepare_dma()
702 master->dma_rx = dma_request_chan(rs->dev, "rx"); in rockchip_spi_probe()
703 if (IS_ERR(master->dma_rx)) { in rockchip_spi_probe()
704 if (PTR_ERR(master->dma_rx) == -EPROBE_DEFER) { in rockchip_spi_probe()
709 master->dma_rx = NULL; in rockchip_spi_probe()
712 if (master->dma_tx && master->dma_rx) { in rockchip_spi_probe()
[all …]
Dspi-fsl-lpspi.c186 if (!controller->dma_rx) in fsl_lpspi_can_dma()
391 ret = dmaengine_slave_config(controller->dma_rx, &rx); in fsl_lpspi_dma_configure()
579 desc_rx = dmaengine_prep_slave_sg(controller->dma_rx, in fsl_lpspi_dma_transfer()
589 dma_async_issue_pending(controller->dma_rx); in fsl_lpspi_dma_transfer()
617 dmaengine_terminate_all(controller->dma_rx); in fsl_lpspi_dma_transfer()
627 dmaengine_terminate_all(controller->dma_rx); in fsl_lpspi_dma_transfer()
637 dmaengine_terminate_all(controller->dma_rx); in fsl_lpspi_dma_transfer()
647 dmaengine_terminate_all(controller->dma_rx); in fsl_lpspi_dma_transfer()
660 if (controller->dma_rx) { in fsl_lpspi_dma_exit()
661 dma_release_channel(controller->dma_rx); in fsl_lpspi_dma_exit()
[all …]
Dspi-ep93xx.c92 struct dma_chan *dma_rx; member
292 chan = espi->dma_rx; in ep93xx_spi_dma_prepare()
380 chan = espi->dma_rx; in ep93xx_spi_dma_finish()
426 dma_async_issue_pending(espi->dma_rx); in ep93xx_spi_dma_transfer()
501 if (espi->dma_rx && xfer->len > SPI_FIFO_SIZE) in ep93xx_spi_transfer_one()
599 espi->dma_rx = dma_request_channel(mask, ep93xx_spi_dma_filter, in ep93xx_spi_setup_dma()
601 if (!espi->dma_rx) { in ep93xx_spi_setup_dma()
620 dma_release_channel(espi->dma_rx); in ep93xx_spi_setup_dma()
621 espi->dma_rx = NULL; in ep93xx_spi_setup_dma()
630 if (espi->dma_rx) { in ep93xx_spi_release_dma()
[all …]
Dspi-davinci.c120 struct dma_chan *dma_rx; member
401 if (dspi->dma_rx && dspi->dma_tx) in davinci_spi_of_setup()
627 dmaengine_slave_config(dspi->dma_rx, &dma_rx_conf); in davinci_spi_bufs()
630 rxdesc = dmaengine_prep_slave_sg(dspi->dma_rx, in davinci_spi_bufs()
663 dma_async_issue_pending(dspi->dma_rx); in davinci_spi_bufs()
754 dspi->dma_rx = dma_request_chan(sdev, "rx"); in davinci_spi_request_dma()
755 if (IS_ERR(dspi->dma_rx)) in davinci_spi_request_dma()
756 return PTR_ERR(dspi->dma_rx); in davinci_spi_request_dma()
760 dma_release_channel(dspi->dma_rx); in davinci_spi_request_dma()
974 dspi->dma_rx = NULL; in davinci_spi_probe()
[all …]
Dspi-pic32.c309 if (!master->dma_rx || !master->dma_tx) in pic32_spi_dma_transfer()
312 desc_rx = dmaengine_prep_slave_sg(master->dma_rx, in pic32_spi_dma_transfer()
346 dma_async_issue_pending(master->dma_rx); in pic32_spi_dma_transfer()
352 dmaengine_terminate_all(master->dma_rx); in pic32_spi_dma_transfer()
382 ret = dmaengine_slave_config(master->dma_rx, &cfg); in pic32_spi_dma_config()
553 dmaengine_terminate_all(master->dma_rx); in pic32_spi_one_transfer()
617 master->dma_rx = dma_request_slave_channel_compat(mask, NULL, NULL, in pic32_spi_dma_prep()
619 if (!master->dma_rx) { in pic32_spi_dma_prep()
640 if (master->dma_rx) in pic32_spi_dma_prep()
641 dma_release_channel(master->dma_rx); in pic32_spi_dma_prep()
[all …]
Dspi-omap2-mcspi.c95 struct dma_chan *dma_rx; member
465 if (mcspi_dma->dma_rx) { in omap2_mcspi_rx_dma()
468 dmaengine_slave_config(mcspi_dma->dma_rx, &cfg); in omap2_mcspi_rx_dma()
502 tx = dmaengine_prep_slave_sg(mcspi_dma->dma_rx, in omap2_mcspi_rx_dma()
516 dma_async_issue_pending(mcspi_dma->dma_rx); in omap2_mcspi_rx_dma()
521 dmaengine_terminate_sync(mcspi_dma->dma_rx); in omap2_mcspi_rx_dma()
1001 mcspi_dma->dma_rx = dma_request_chan(&master->dev, in omap2_mcspi_request_dma()
1003 if (IS_ERR(mcspi_dma->dma_rx)) { in omap2_mcspi_request_dma()
1004 ret = PTR_ERR(mcspi_dma->dma_rx); in omap2_mcspi_request_dma()
1005 mcspi_dma->dma_rx = NULL; in omap2_mcspi_request_dma()
[all …]
Dspi-stm32.c315 struct dma_chan *dma_rx; member
679 if (spi->cur_usedma && spi->dma_rx) in stm32f4_spi_disable()
680 dmaengine_terminate_all(spi->dma_rx); in stm32f4_spi_disable()
743 if (spi->cur_usedma && spi->dma_rx) in stm32h7_spi_disable()
744 dmaengine_terminate_all(spi->dma_rx); in stm32h7_spi_disable()
1284 if (spi->rx_buf && spi->dma_rx) { in stm32_spi_transfer_one_dma()
1286 dmaengine_slave_config(spi->dma_rx, &rx_dma_conf); in stm32_spi_transfer_one_dma()
1293 spi->dma_rx, xfer->rx_sg.sgl, in stm32_spi_transfer_one_dma()
1312 (spi->rx_buf && spi->dma_rx && !rx_dma_desc)) in stm32_spi_transfer_one_dma()
1327 dma_async_issue_pending(spi->dma_rx); in stm32_spi_transfer_one_dma()
[all …]
Dspi-qup.c402 chan = master->dma_rx; in spi_qup_prep_sg()
422 dmaengine_terminate_all(master->dma_rx); in spi_qup_dma_terminate()
489 dma_async_issue_pending(master->dma_rx); in spi_qup_do_dma()
897 IS_ERR_OR_NULL(master->dma_rx)) in spi_qup_can_dma()
920 if (!IS_ERR_OR_NULL(master->dma_rx)) in spi_qup_release_dma()
921 dma_release_channel(master->dma_rx); in spi_qup_release_dma()
935 master->dma_rx = dma_request_slave_channel_reason(dev, "rx"); in spi_qup_init_dma()
936 if (IS_ERR(master->dma_rx)) in spi_qup_init_dma()
937 return PTR_ERR(master->dma_rx); in spi_qup_init_dma()
956 ret = dmaengine_slave_config(master->dma_rx, rx_conf); in spi_qup_init_dma()
[all …]
Dspi-atmel.c510 if (dmaengine_slave_config(master->dma_rx, slave_config)) { in atmel_spi_dma_slave_config()
547 master->dma_rx = dma_request_slave_channel(dev, "rx"); in atmel_spi_configure_dma()
549 if (!master->dma_rx) { in atmel_spi_configure_dma()
563 dma_chan_name(master->dma_rx)); in atmel_spi_configure_dma()
567 if (master->dma_rx) in atmel_spi_configure_dma()
568 dma_release_channel(master->dma_rx); in atmel_spi_configure_dma()
572 master->dma_tx = master->dma_rx = NULL; in atmel_spi_configure_dma()
578 if (master->dma_rx) in atmel_spi_stop_dma()
579 dmaengine_terminate_all(master->dma_rx); in atmel_spi_stop_dma()
586 if (master->dma_rx) { in atmel_spi_release_dma()
[all …]
Dspi-imx.c229 if (!master->dma_rx) in spi_imx_can_dma()
1185 ret = dmaengine_slave_config(master->dma_rx, &rx); in spi_imx_dma_configure()
1252 if (master->dma_rx) { in spi_imx_sdma_exit()
1253 dma_release_channel(master->dma_rx); in spi_imx_sdma_exit()
1254 master->dma_rx = NULL; in spi_imx_sdma_exit()
1284 master->dma_rx = dma_request_slave_channel_reason(dev, "rx"); in spi_imx_sdma_init()
1285 if (IS_ERR(master->dma_rx)) { in spi_imx_sdma_init()
1286 ret = PTR_ERR(master->dma_rx); in spi_imx_sdma_init()
1288 master->dma_rx = NULL; in spi_imx_sdma_init()
1371 desc_rx = dmaengine_prep_slave_sg(master->dma_rx, in spi_imx_dma_transfer()
[all …]
Dspi-dw-mid.c42 struct dw_dma_slave *rx = dws->dma_rx; in mid_spi_dma_init()
61 dws->master->dma_rx = dws->rxchan; in mid_spi_dma_init()
318 dws->dma_rx = &mid_dma_rx; in dw_spi_mid_init()
Dspi-sh-msiof.c754 desc_rx = dmaengine_prep_slave_single(p->ctlr->dma_rx, in sh_msiof_dma_once()
803 dma_async_issue_pending(p->ctlr->dma_rx); in sh_msiof_dma_once()
845 dma_sync_single_for_cpu(p->ctlr->dma_rx->device->dev, in sh_msiof_dma_once()
858 dmaengine_terminate_all(p->ctlr->dma_rx); in sh_msiof_dma_once()
1239 ctlr->dma_rx = sh_msiof_request_dma_chan(dev, DMA_DEV_TO_MEM, in sh_msiof_request_dma()
1241 if (!ctlr->dma_rx) in sh_msiof_request_dma()
1258 rx_dev = ctlr->dma_rx->device->dev; in sh_msiof_request_dma()
1274 dma_release_channel(ctlr->dma_rx); in sh_msiof_request_dma()
1288 dma_unmap_single(ctlr->dma_rx->device->dev, p->rx_dma_addr, PAGE_SIZE, in sh_msiof_release_dma()
1294 dma_release_channel(ctlr->dma_rx); in sh_msiof_release_dma()
Dspi-rspi.c532 desc_rx = dmaengine_prep_slave_sg(rspi->ctlr->dma_rx, rx->sgl, in rspi_dma_transfer()
590 dma_async_issue_pending(rspi->ctlr->dma_rx); in rspi_dma_transfer()
606 dmaengine_terminate_all(rspi->ctlr->dma_rx); in rspi_dma_transfer()
620 dmaengine_terminate_all(rspi->ctlr->dma_rx); in rspi_dma_transfer()
1089 ctlr->dma_rx = rspi_request_dma_chan(dev, DMA_DEV_TO_MEM, dma_rx_id, in rspi_request_dma()
1091 if (!ctlr->dma_rx) { in rspi_request_dma()
1106 if (ctlr->dma_rx) in rspi_release_dma()
1107 dma_release_channel(ctlr->dma_rx); in rspi_release_dma()
Dspi-mem.c45 else if (op->data.dir == SPI_MEM_DATA_IN && ctlr->dma_rx) in spi_controller_dma_map_mem_op_data()
46 dmadev = ctlr->dma_rx->device->dev; in spi_controller_dma_map_mem_op_data()
91 else if (op->data.dir == SPI_MEM_DATA_IN && ctlr->dma_rx) in spi_controller_dma_unmap_mem_op_data()
92 dmadev = ctlr->dma_rx->device->dev; in spi_controller_dma_unmap_mem_op_data()
/Linux-v5.4/drivers/i2c/busses/
Di2c-sh_mobile.c135 struct dma_chan *dma_rx; member
450 ? pd->dma_rx : pd->dma_tx; in sh_mobile_i2c_dma_unmap()
463 dmaengine_terminate_all(pd->dma_rx); in sh_mobile_i2c_cleanup_dma()
521 struct dma_chan *chan = read ? pd->dma_rx : pd->dma_tx; in sh_mobile_i2c_xfer_dma()
528 chan = pd->dma_rx = sh_mobile_i2c_request_dma_chan(pd->dev, DMA_DEV_TO_MEM, in sh_mobile_i2c_xfer_dma()
793 if (!IS_ERR(pd->dma_rx)) { in sh_mobile_i2c_release_dma()
794 dma_release_channel(pd->dma_rx); in sh_mobile_i2c_release_dma()
795 pd->dma_rx = ERR_PTR(-EPROBE_DEFER); in sh_mobile_i2c_release_dma()
879 pd->dma_rx = pd->dma_tx = ERR_PTR(-EPROBE_DEFER); in sh_mobile_i2c_probe()
Di2c-rcar.c137 struct dma_chan *dma_rx; member
366 ? priv->dma_rx : priv->dma_tx; in rcar_i2c_dma_unmap()
387 dmaengine_terminate_all(priv->dma_rx); in rcar_i2c_cleanup_dma()
409 struct dma_chan *chan = read ? priv->dma_rx : priv->dma_tx; in rcar_i2c_dma()
729 chan = read ? priv->dma_rx : priv->dma_tx; in rcar_i2c_request_dma()
737 priv->dma_rx = chan; in rcar_i2c_request_dma()
749 if (!IS_ERR(priv->dma_rx)) { in rcar_i2c_release_dma()
750 dma_release_channel(priv->dma_rx); in rcar_i2c_release_dma()
751 priv->dma_rx = ERR_PTR(-EPROBE_DEFER); in rcar_i2c_release_dma()
965 priv->dma_rx = priv->dma_tx = ERR_PTR(-EPROBE_DEFER); in rcar_i2c_probe()
/Linux-v5.4/drivers/net/ethernet/micrel/
Dks8842.c148 ((adapter)->dma_rx.channel != -1))
160 struct ks8842_rx_dma_ctl dma_rx; member
543 struct ks8842_rx_dma_ctl *ctl = &adapter->dma_rx; in __ks8842_start_new_rx_dma()
594 struct ks8842_rx_dma_ctl *ctl = &adapter->dma_rx; in ks8842_rx_frame_dma_tasklet()
837 if (adapter->dma_rx.adesc) in ks8842_dma_rx_cb()
838 tasklet_schedule(&adapter->dma_rx.tasklet); in ks8842_dma_rx_cb()
862 struct ks8842_rx_dma_ctl *rx_ctl = &adapter->dma_rx; in ks8842_stop_dma()
884 struct ks8842_rx_dma_ctl *rx_ctl = &adapter->dma_rx; in ks8842_dealloc_dma_bufs()
916 struct ks8842_rx_dma_ctl *rx_ctl = &adapter->dma_rx; in ks8842_alloc_dma_bufs()
988 adapter->dma_rx.channel = -1; in ks8842_open()
[all …]
/Linux-v5.4/drivers/net/ethernet/samsung/sxgbe/
Dsxgbe_dma.c45 dma_addr_t dma_rx, int t_rsize, int r_rsize) in sxgbe_dma_channel_init() argument
71 writel(upper_32_bits(dma_rx), in sxgbe_dma_channel_init()
73 writel(lower_32_bits(dma_rx), in sxgbe_dma_channel_init()
84 dma_addr = dma_rx + ((r_rsize - 1) * SXGBE_DESC_SIZE_BYTES); in sxgbe_dma_channel_init()
Dsxgbe_dma.h24 int pbl, dma_addr_t dma_tx, dma_addr_t dma_rx,
/Linux-v5.4/drivers/mmc/host/
Domap.c127 struct dma_chan *dma_rx; member
414 c = host->dma_rx; in mmc_omap_release_dma()
1006 c = host->dma_rx; in mmc_omap_prepare_data()
1077 host->dma_tx : host->dma_rx; in mmc_omap_start_request()
1401 host->dma_rx = dma_request_chan(&pdev->dev, "rx"); in mmc_omap_probe()
1402 if (IS_ERR(host->dma_rx)) { in mmc_omap_probe()
1403 ret = PTR_ERR(host->dma_rx); in mmc_omap_probe()
1411 host->dma_rx = NULL; in mmc_omap_probe()
1456 if (host->dma_rx) in mmc_omap_probe()
1457 dma_release_channel(host->dma_rx); in mmc_omap_probe()
[all …]
Ddavinci_mmc.c193 struct dma_chan *dma_rx; member
393 sync_dev = host->dma_rx; in davinci_abort_dma()
437 chan = host->dma_rx; in mmc_davinci_send_dma_request()
438 dmaengine_slave_config(host->dma_rx, &dma_rx_conf); in mmc_davinci_send_dma_request()
440 desc = dmaengine_prep_slave_sg(host->dma_rx, in mmc_davinci_send_dma_request()
492 dma_release_channel(host->dma_rx); in davinci_release_dma_channels()
503 host->dma_rx = dma_request_chan(mmc_dev(host->mmc), "rx"); in davinci_acquire_dma_channels()
504 if (IS_ERR(host->dma_rx)) { in davinci_acquire_dma_channels()
507 return PTR_ERR(host->dma_rx); in davinci_acquire_dma_channels()
/Linux-v5.4/drivers/net/ethernet/calxeda/
Dxgmac.c355 struct xgmac_dma_desc *dma_rx; member
687 p = priv->dma_rx + entry; in xgmac_rx_refill()
737 priv->dma_rx = dma_alloc_coherent(priv->device, in xgmac_dma_desc_rings_init()
742 if (!priv->dma_rx) in xgmac_dma_desc_rings_init()
760 priv->dma_rx, priv->dma_tx, in xgmac_dma_desc_rings_init()
766 desc_init_rx_desc(priv->dma_rx, DMA_RX_RING_SZ, priv->dma_buf_sz); in xgmac_dma_desc_rings_init()
783 priv->dma_rx, priv->dma_rx_phy); in xgmac_dma_desc_rings_init()
802 p = priv->dma_rx + i; in xgmac_free_rx_skbufs()
849 if (priv->dma_rx) { in xgmac_free_dma_desc_rings()
852 priv->dma_rx, priv->dma_rx_phy); in xgmac_free_dma_desc_rings()
[all …]

12