Home
last modified time | relevance | path

Searched refs:dma_dev (Results 1 – 25 of 50) sorted by relevance

12

/Zephyr-latest/drivers/spi/spi_nxp_lpspi/
Dspi_nxp_lpspi_dma.c21 const struct device *dma_dev; member
70 return dma_config(stream->dma_dev, stream->channel, &stream->dma_cfg); in spi_mcux_dma_tx_load()
92 return dma_config(stream->dma_dev, stream->channel, &stream->dma_cfg); in spi_mcux_dma_rx_load()
115 ret = dma_start(rx->dma_dev, rx->channel); in spi_mcux_dma_rxtx_load()
120 return dma_start(tx->dma_dev, tx->channel); in spi_mcux_dma_rxtx_load()
240 static int lpspi_dma_dev_ready(const struct device *dma_dev) in lpspi_dma_dev_ready() argument
242 if (!device_is_ready(dma_dev)) { in lpspi_dma_dev_ready()
243 LOG_ERR("%s device is not ready", dma_dev->name); in lpspi_dma_dev_ready()
256 if (!lpspi_dma_dev_ready(dma_data->dma_tx.dma_dev) || in spi_mcux_dma_init()
257 !lpspi_dma_dev_ready(dma_data->dma_rx.dma_dev)) { in spi_mcux_dma_init()
[all …]
/Zephyr-latest/drivers/i2s/
Di2s_esp32.c40 const struct device *dma_dev; member
129 static void i2s_esp32_rx_callback(const struct device *dma_dev, void *arg, uint32_t channel,
131 static void i2s_esp32_tx_callback(const struct device *dma_dev, void *arg, uint32_t channel,
153 err = dma_reload(stream->dma_dev, stream->dma_channel, (uint32_t)src, (uint32_t)dst, in i2s_esp32_restart_dma()
164 err = dma_start(stream->dma_dev, stream->dma_channel); in i2s_esp32_restart_dma()
210 err = dma_config(stream->dma_dev, stream->dma_channel, &dma_cfg); in i2s_esp32_start_dma()
221 err = dma_start(stream->dma_dev, stream->dma_channel); in i2s_esp32_start_dma()
299 dma_stop(stream->dma_dev, stream->dma_channel); in i2s_esp32_rx_stop_transfer()
313 dma_stop(stream->dma_dev, stream->dma_channel); in i2s_esp32_tx_stop_transfer()
322 static void i2s_esp32_rx_callback(const struct device *dma_dev, void *arg, uint32_t channel, in i2s_esp32_rx_callback() argument
[all …]
/Zephyr-latest/drivers/video/
Dvideo_mcux_smartdma.c22 const struct device *dma_dev; member
61 dma_stop(data->config->dma_dev, 0); in nxp_video_sdma_callback()
104 return dma_stop(config->dma_dev, 0); in nxp_video_sdma_set_stream()
132 ret = dma_config(config->dma_dev, 0, &sdma_config); in nxp_video_sdma_set_stream()
141 ret = dma_start(config->dma_dev, 0); in nxp_video_sdma_set_stream()
206 dma_stop(config->dma_dev, 0); in nxp_video_sdma_flush()
324 if (!device_is_ready(config->dma_dev)) { in nxp_video_sdma_init()
346 dma_smartdma_install_fw(config->dma_dev, (uint8_t *)s_smartdmaCameraFirmware, in nxp_video_sdma_init()
369 .dma_dev = DEVICE_DT_GET(DT_INST_PARENT(inst)), \
Dvideo_esp32_dvp.c49 const struct device *dma_dev; member
86 ret = dma_reload(cfg->dma_dev, cfg->rx_dma_channel, 0, (uint32_t)data->active_vbuf->buffer, in video_esp32_reload_dma()
93 ret = dma_start(cfg->dma_dev, cfg->rx_dma_channel); in video_esp32_reload_dma()
154 error = dma_stop(cfg->dma_dev, cfg->rx_dma_channel); in video_esp32_set_stream()
171 error = dma_get_status(cfg->dma_dev, cfg->rx_dma_channel, &dma_status); in video_esp32_set_stream()
219 error = dma_config(cfg->dma_dev, cfg->rx_dma_channel, &dma_cfg); in video_esp32_set_stream()
225 error = dma_start(cfg->dma_dev, cfg->rx_dma_channel); in video_esp32_set_stream()
417 if (!device_is_ready(cfg->dma_dev)) { in video_esp32_init()
447 .dma_dev = ESP32_DT_INST_DMA_CTLR(0, rx),
Dvideo_stm32_dcmi.c33 const struct device *dma_dev; member
121 if (!device_is_ready(config->dma.dma_dev)) { in stm32_dma_init()
122 LOG_ERR("%s DMA device not ready", config->dma.dma_dev->name); in stm32_dma_init()
142 ret = dma_config(config->dma.dma_dev, in stm32_dma_init()
396 .dma_dev = DEVICE_DT_GET(DT_INST_DMAS_CTLR_BY_IDX(index, 0)), \
/Zephyr-latest/drivers/spi/
Dspi_silabs_eusart.c43 const struct device *dma_dev; member
76 __ASSERT_NO_MSG(!!data->dma_chan_tx.dma_dev == !!data->dma_chan_rx.dma_dev); in spi_silabs_eusart_is_dma_enabled_instance()
78 return data->dma_chan_rx.dma_dev != NULL; in spi_silabs_eusart_is_dma_enabled_instance()
182 if (!device_is_ready(data->dma_chan_tx.dma_dev)) { in spi_silabs_eusart_configure()
191 dma_request_channel(data->dma_chan_rx.dma_dev, NULL); in spi_silabs_eusart_configure()
201 dma_request_channel(data->dma_chan_tx.dma_dev, NULL); in spi_silabs_eusart_configure()
205 dma_release_channel(data->dma_chan_rx.dma_dev, data->dma_chan_rx.chan_nb); in spi_silabs_eusart_configure()
229 dma_release_channel(data->dma_chan_rx.dma_dev, data->dma_chan_rx.chan_nb); in spi_silabs_eusart_configure()
230 dma_release_channel(data->dma_chan_tx.dma_dev, data->dma_chan_tx.chan_nb); in spi_silabs_eusart_configure()
302 dma_stop(data->dma_chan_tx.dma_dev, data->dma_chan_tx.chan_nb); in spi_silabs_dma_rx_callback()
[all …]
Dspi_mcux_dspi.c31 const struct device *dma_dev; member
113 ret = dma_start(data->tx_dma_config.dma_dev, in spi_mcux_transfer_next_packet()
127 ret = dma_start(data->rx_dma_config.dma_dev, in spi_mcux_transfer_next_packet()
143 dma_start(data->tx_dma_config.dma_dev, data->tx_dma_config.dma_channel); in spi_mcux_transfer_next_packet()
218 dma_start(data->rx_dma_config.dma_dev, data->rx_dma_config.dma_channel); in spi_mcux_isr()
382 dma_config(data->tx_dma_config.dma_dev, data->tx_dma_config.dma_channel, in update_tx_dma()
457 dma_config(data->rx_dma_config.dma_dev, data->rx_dma_config.dma_channel, in update_rx_dma()
476 static void dma_callback(const struct device *dma_dev, void *callback_arg, in dma_callback() argument
770 const struct device *dma_dev; in spi_mcux_init() local
772 dma_dev = data->rx_dma_config.dma_dev; in spi_mcux_init()
[all …]
Dspi_sam0.c40 const struct device *dma_dev; member
404 static void spi_sam0_dma_rx_done(const struct device *dma_dev, void *arg,
440 retval = dma_config(cfg->dma_dev, cfg->rx_dma_channel, in spi_sam0_dma_rx_load()
446 return dma_start(cfg->dma_dev, cfg->rx_dma_channel); in spi_sam0_dma_rx_load()
479 retval = dma_config(cfg->dma_dev, cfg->tx_dma_channel, in spi_sam0_dma_tx_load()
486 return dma_start(cfg->dma_dev, cfg->tx_dma_channel); in spi_sam0_dma_tx_load()
550 static void spi_sam0_dma_rx_done(const struct device *dma_dev, void *arg, in spi_sam0_dma_rx_done() argument
573 dma_stop(cfg->dma_dev, cfg->tx_dma_channel); in spi_sam0_dma_rx_done()
574 dma_stop(cfg->dma_dev, cfg->rx_dma_channel); in spi_sam0_dma_rx_done()
621 dma_stop(cfg->dma_dev, cfg->tx_dma_channel); in spi_sam0_transceive_async()
[all …]
Dspi_andes_atcspi200.c24 const struct device *dma_dev; member
261 dma_stop(data->dma_rx.dma_dev, data->dma_rx.channel); in dma_rx_callback()
269 error = dma_start(data->dma_rx.dma_dev, data->dma_rx.channel); in dma_rx_callback()
282 dma_stop(data->dma_tx.dma_dev, data->dma_tx.channel); in dma_tx_callback()
290 error = dma_start(data->dma_tx.dma_dev, data->dma_tx.channel); in dma_tx_callback()
407 ret = dma_config(data->dma_tx.dma_dev, data->dma_tx.channel, in spi_dma_tx_load()
520 ret = dma_config(data->dma_rx.dma_dev, data->dma_rx.channel, in spi_dma_rx_load()
582 error = dma_start(data->dma_rx.dma_dev, data->dma_rx.channel); in spi_transfer_dma()
589 error = dma_start(data->dma_tx.dma_dev, data->dma_tx.channel); in spi_transfer_dma()
639 if ((data->dma_tx.dma_dev != NULL) && (data->dma_rx.dma_dev != NULL)) { in transceive()
[all …]
Dspi_sam.c42 const struct device *dma_dev; member
292 static void dma_callback(const struct device *dma_dev, void *user_data, in dma_callback() argument
295 ARG_UNUSED(dma_dev); in dma_callback()
398 res = dma_config(drv_cfg->dma_dev, drv_cfg->dma_rx_channel, &rx_dma_cfg); in spi_sam_dma_txrx()
404 res = dma_config(drv_cfg->dma_dev, drv_cfg->dma_tx_channel, &tx_dma_cfg); in spi_sam_dma_txrx()
411 res = dma_start(drv_cfg->dma_dev, drv_cfg->dma_rx_channel); in spi_sam_dma_txrx()
417 res = dma_start(drv_cfg->dma_dev, drv_cfg->dma_tx_channel); in spi_sam_dma_txrx()
420 dma_stop(drv_cfg->dma_dev, drv_cfg->dma_rx_channel); in spi_sam_dma_txrx()
448 if ((rx_buf_len < SAM_SPI_DMA_THRESHOLD || cfg->dma_dev == NULL) && in spi_sam_rx()
476 if ((tx_buf_len < SAM_SPI_DMA_THRESHOLD || cfg->dma_dev == NULL) && in spi_sam_tx()
[all …]
Dspi_mcux_flexcomm.c53 const struct device *dma_dev; member
481 ret = dma_config(data->dma_tx.dma_dev, data->dma_tx.channel, in spi_mcux_dma_tx_load()
508 return dma_start(data->dma_tx.dma_dev, data->dma_tx.channel); in spi_mcux_dma_tx_load()
546 ret = dma_config(data->dma_rx.dma_dev, data->dma_rx.channel, in spi_mcux_dma_rx_load()
554 return dma_start(data->dma_rx.dma_dev, data->dma_rx.channel); in spi_mcux_dma_rx_load()
805 if (!device_is_ready(data->dma_tx.dma_dev)) { in spi_mcux_init()
806 LOG_ERR("%s device is not ready", data->dma_tx.dma_dev->name); in spi_mcux_init()
810 if (!device_is_ready(data->dma_rx.dma_dev)) { in spi_mcux_init()
811 LOG_ERR("%s device is not ready", data->dma_rx.dma_dev->name); in spi_mcux_init()
857 .dma_dev = DEVICE_DT_GET(DT_INST_DMAS_CTLR_BY_NAME(id, tx)), \
[all …]
Dspi_ll_stm32.c126 static void dma_callback(const struct device *dma_dev, void *arg, in dma_callback() argument
129 ARG_UNUSED(dma_dev); in dma_callback()
207 ret = dma_config(data->dma_tx.dma_dev, data->dma_tx.channel, in spi_stm32_dma_tx_load()
215 return dma_start(data->dma_tx.dma_dev, data->dma_tx.channel); in spi_stm32_dma_tx_load()
266 ret = dma_config(data->dma_rx.dma_dev, data->dma_rx.channel, in spi_stm32_dma_rx_load()
274 return dma_start(data->dma_rx.dma_dev, data->dma_rx.channel); in spi_stm32_dma_rx_load()
1331 err = dma_stop(data->dma_rx.dma_dev, data->dma_rx.channel); in transceive_dma()
1335 err = dma_stop(data->dma_tx.dma_dev, data->dma_tx.channel); in transceive_dma()
1363 if ((data->dma_tx.dma_dev != NULL) in spi_stm32_transceive()
1364 && (data->dma_rx.dma_dev != NULL)) { in spi_stm32_transceive()
[all …]
/Zephyr-latest/subsys/mgmt/ec_host_cmd/backends/
Dec_host_cmd_backend_spi_stm32.c149 const struct device *dma_dev; member
188 .dma_dev = DEVICE_DT_GET(DT_DMAS_CTLR_BY_NAME(id, dir)), \
365 if ((hc_spi->dma_rx->dma_dev != NULL) && !device_is_ready(hc_spi->dma_rx->dma_dev)) { in spi_init()
366 LOG_ERR("%s device not ready", hc_spi->dma_rx->dma_dev->name); in spi_init()
370 if ((hc_spi->dma_tx->dma_dev != NULL) && !device_is_ready(hc_spi->dma_tx->dma_dev)) { in spi_init()
371 LOG_ERR("%s device not ready", hc_spi->dma_tx->dma_dev->name); in spi_init()
424 ret = dma_reload(hc_spi->dma_tx->dma_dev, hc_spi->dma_tx->channel, (uint32_t)hc_spi->tx_buf, in reload_dma_tx()
431 ret = dma_start(hc_spi->dma_tx->dma_dev, hc_spi->dma_tx->channel); in reload_dma_tx()
469 ret = dma_config(hc_spi->dma_tx->dma_dev, hc_spi->dma_tx->channel, &stream->dma_cfg); in spi_config_dma_tx()
485 ret = dma_reload(hc_spi->dma_rx->dma_dev, hc_spi->dma_rx->channel, dma_source_addr(spi), in reload_dma_rx()
[all …]
/Zephyr-latest/drivers/serial/
Duart_xmc4xxx.c41 const struct device *dma_dev; member
443 if (dma_get_status(data->dma_rx.dma_dev, data->dma_rx.dma_channel, &stat) == 0) { in async_evt_rx_stopped()
525 if (dma_get_status(data->dma_rx.dma_dev, data->dma_rx.dma_channel, &stat) == 0) { in uart_xmc4xxx_async_rx_timeout()
552 if (!dma_get_status(data->dma_tx.dma_dev, data->dma_tx.dma_channel, &stat)) { in uart_xmc4xxx_async_tx_abort()
556 dma_stop(data->dma_tx.dma_dev, data->dma_tx.dma_channel); in uart_xmc4xxx_async_tx_abort()
582 if (data->dma_rx.dma_dev != NULL) { in uart_xmc4xxx_async_init()
583 if (!device_is_ready(data->dma_rx.dma_dev)) { in uart_xmc4xxx_async_init()
600 if (data->dma_tx.dma_dev != NULL) { in uart_xmc4xxx_async_init()
601 if (!device_is_ready(data->dma_tx.dma_dev)) { in uart_xmc4xxx_async_init()
647 if (data->dma_tx.dma_dev == NULL) { in uart_xmc4xxx_async_tx()
[all …]
Duart_silabs_usart.c32 const struct device *dma_dev; member
420 if (!dma_get_status(data->dma_rx.dma_dev, data->dma_rx.dma_channel, &stat)) { in uart_silabs_dma_rx_flush()
429 void uart_silabs_dma_rx_cb(const struct device *dma_dev, void *user_data, uint32_t channel, in uart_silabs_dma_rx_cb() argument
449 dma_stop(data->dma_rx.dma_dev, data->dma_rx.dma_channel); in uart_silabs_dma_rx_cb()
456 void uart_silabs_dma_tx_cb(const struct device *dma_dev, void *user_data, uint32_t channel, in uart_silabs_dma_tx_cb() argument
462 dma_stop(data->dma_tx.dma_dev, data->dma_tx.dma_channel); in uart_silabs_dma_tx_cb()
473 if (!data->dma_tx.dma_dev) { in uart_silabs_async_tx()
504 ret = dma_config(data->dma_tx.dma_dev, data->dma_tx.dma_channel, &data->dma_tx.dma_cfg); in uart_silabs_async_tx()
510 ret = dma_start(data->dma_tx.dma_dev, data->dma_tx.dma_channel); in uart_silabs_async_tx()
537 if (!dma_get_status(data->dma_tx.dma_dev, data->dma_tx.dma_channel, &stat)) { in uart_silabs_async_tx_abort()
[all …]
Duart_sam0.c50 const struct device *dma_dev; member
131 static void uart_sam0_dma_tx_done(const struct device *dma_dev, void *arg, in uart_sam0_dma_tx_done() argument
134 ARG_UNUSED(dma_dev); in uart_sam0_dma_tx_done()
165 dma_stop(cfg->dma_dev, cfg->tx_dma_channel); in uart_sam0_tx_halt()
169 if (dma_get_status(cfg->dma_dev, cfg->tx_dma_channel, &st) == 0) { in uart_sam0_tx_halt()
220 static void uart_sam0_dma_rx_done(const struct device *dma_dev, void *arg, in uart_sam0_dma_rx_done() argument
223 ARG_UNUSED(dma_dev); in uart_sam0_dma_rx_done()
275 dma_reload(cfg->dma_dev, cfg->rx_dma_channel, in uart_sam0_dma_rx_done()
293 dma_start(cfg->dma_dev, cfg->rx_dma_channel); in uart_sam0_dma_rx_done()
326 dma_stop(cfg->dma_dev, cfg->rx_dma_channel); in uart_sam0_rx_timeout()
[all …]
Duart_esp32.c88 const struct device *dma_dev; member
560 static void IRAM_ATTR uart_esp32_dma_rx_done(const struct device *dma_dev, void *user_data, in uart_esp32_dma_rx_done() argument
573 dma_reload(config->dma_dev, config->rx_dma_channel, 0, in uart_esp32_dma_rx_done()
576 dma_start(config->dma_dev, config->rx_dma_channel); in uart_esp32_dma_rx_done()
620 dma_reload(config->dma_dev, config->rx_dma_channel, 0, (uint32_t)data->async.rx_buf, in uart_esp32_dma_rx_done()
622 dma_start(config->dma_dev, config->rx_dma_channel); in uart_esp32_dma_rx_done()
629 static void IRAM_ATTR uart_esp32_dma_tx_done(const struct device *dma_dev, void *user_data, in uart_esp32_dma_tx_done() argument
662 err = dma_stop(config->dma_dev, config->tx_dma_channel); in uart_esp32_async_tx_abort()
751 err = dma_get_status(config->dma_dev, config->tx_dma_channel, &dma_status); in uart_esp32_async_tx()
775 err = dma_config(config->dma_dev, config->tx_dma_channel, &dma_cfg); in uart_esp32_async_tx()
[all …]
Duart_ns16550.c292 const struct device *dma_dev; member
305 const struct device *dma_dev; member
831 if (data->async.tx_dma_params.dma_dev != NULL) {
849 if (data->async.tx_dma_params.dma_dev != NULL) {
869 dma_intel_lpss_set_base(data->async.tx_dma_params.dma_dev, base);
870 dma_intel_lpss_setup(data->async.tx_dma_params.dma_dev);
1297 if (dev_data->async.tx_dma_params.dma_dev != NULL) {
1310 dma_intel_lpss_isr(dev_data->async.rx_dma_params.dma_dev);
1499 dma_get_status(dma_params->dma_dev,
1518 if (!device_is_ready(dma_params->dma_dev)) {
[all …]
Duart_mcux_lpuart.c49 const struct device *dma_dev; member
514 const int get_status_result = dma_get_status(config->rx_dma_config.dma_dev, in mcux_lpuart_async_rx_flush()
561 const int ret = dma_stop(config->rx_dma_config.dma_dev, in mcux_lpuart_rx_disable()
600 int ret = dma_config(config->rx_dma_config.dma_dev, in configure_and_start_rx_dma()
608 ret = dma_start(config->rx_dma_config.dma_dev, config->rx_dma_config.dma_channel); in configure_and_start_rx_dma()
630 dma_reload(config->rx_dma_config.dma_dev, config->rx_dma_config.dma_channel, in uart_mcux_lpuart_dma_replace_rx_buffer()
641 static void dma_callback(const struct device *dma_dev, void *callback_arg, uint32_t channel, in dma_callback() argument
651 const int get_status_result = dma_get_status(dma_dev, channel, &status); in dma_callback()
739 const int get_status_result = dma_get_status(config->tx_dma_config.dma_dev, in mcux_lpuart_tx()
761 ret = dma_config(config->tx_dma_config.dma_dev, in mcux_lpuart_tx()
[all …]
Duart_stm32.c1234 if (dma_get_status(data->dma_rx.dma_dev, data->dma_rx.dma_channel, &stat) == 0) { in uart_stm32_dma_rx_flush()
1462 dma_stop(data->dma_rx.dma_dev, data->dma_rx.dma_channel); in uart_stm32_async_rx_disable()
1485 void uart_stm32_dma_tx_cb(const struct device *dma_dev, void *user_data, in uart_stm32_dma_tx_cb() argument
1498 if (!dma_get_status(data->dma_tx.dma_dev, in uart_stm32_dma_tx_cb()
1528 dma_reload(data->dma_rx.dma_dev, data->dma_rx.dma_channel, in uart_stm32_dma_replace_buffer()
1533 dma_start(data->dma_rx.dma_dev, data->dma_rx.dma_channel); in uart_stm32_dma_replace_buffer()
1541 void uart_stm32_dma_rx_cb(const struct device *dma_dev, void *user_data, in uart_stm32_dma_rx_cb() argument
1591 if (data->dma_tx.dma_dev == NULL) { in uart_stm32_async_tx()
1622 ret = dma_config(data->dma_tx.dma_dev, data->dma_tx.dma_channel, in uart_stm32_async_tx()
1630 if (dma_start(data->dma_tx.dma_dev, data->dma_tx.dma_channel)) { in uart_stm32_async_tx()
[all …]
/Zephyr-latest/drivers/flash/
Dsoc_flash_silabs_series2.c27 const struct device *dma_dev; member
72 static void flash_silabs_read_callback(const struct device *dma_dev, void *user_data, in flash_silabs_read_callback() argument
108 err = dma_config(dev_data->dma_dev, dev_data->dma_channel, &dev_data->dma_cfg); in flash_silabs_read_dma()
113 err = dma_start(dev_data->dma_dev, dev_data->dma_channel); in flash_silabs_read_dma()
282 dev_data->dma_channel = dma_request_channel(dev_data->dma_dev, NULL); in flash_silabs_init()
305 .dma_dev = DEVICE_DT_GET(DT_INST_DMAS_CTLR(0)),
/Zephyr-latest/drivers/i2c/
Di2c_sam0.c47 const struct device *dma_dev; member
107 dma_stop(cfg->dma_dev, cfg->dma_channel); in i2c_sam0_terminate_on_error()
222 static void i2c_sam0_dma_write_done(const struct device *dma_dev, void *arg, in i2c_sam0_dma_write_done() argument
230 ARG_UNUSED(dma_dev); in i2c_sam0_dma_write_done()
296 retval = dma_config(cfg->dma_dev, cfg->dma_channel, &dma_cfg); in i2c_sam0_dma_write_start()
303 retval = dma_start(cfg->dma_dev, cfg->dma_channel); in i2c_sam0_dma_write_start()
313 static void i2c_sam0_dma_read_done(const struct device *dma_dev, void *arg, in i2c_sam0_dma_read_done() argument
321 ARG_UNUSED(dma_dev); in i2c_sam0_dma_read_done()
389 retval = dma_config(cfg->dma_dev, cfg->dma_channel, &dma_cfg); in i2c_sam0_dma_read_start()
396 retval = dma_start(cfg->dma_dev, cfg->dma_channel); in i2c_sam0_dma_read_start()
[all …]
Di2c_dw.c79 dma_stop(rom->dma_dev, channel); in cb_i2c_idma_transfer()
112 if (!device_is_ready(rom->dma_dev)) { in i2c_dw_idma_rx_transfer()
135 if (dma_config(rom->dma_dev, DMA_INTEL_LPSS_RX_CHAN, &dma_cfg)) { in i2c_dw_idma_rx_transfer()
140 if (dma_start(rom->dma_dev, DMA_INTEL_LPSS_RX_CHAN)) { in i2c_dw_idma_rx_transfer()
159 if (!device_is_ready(rom->dma_dev)) { in i2c_dw_idma_tx_transfer()
182 if (dma_config(rom->dma_dev, DMA_INTEL_LPSS_TX_CHAN, &dma_cfg)) { in i2c_dw_idma_tx_transfer()
187 if (dma_start(rom->dma_dev, DMA_INTEL_LPSS_TX_CHAN)) { in i2c_dw_idma_tx_transfer()
395 dma_intel_lpss_isr(rom->dma_dev); in i2c_dw_isr()
1073 dma_intel_lpss_set_base(rom->dma_dev, base); in i2c_dw_initialize()
1074 dma_intel_lpss_setup(rom->dma_dev); in i2c_dw_initialize()
[all …]
/Zephyr-latest/drivers/mipi_dbi/
Dmipi_dbi_nxp_flexio_lcdif.c23 const struct device *dma_dev; member
277 dma_config(lcdif_data->dma_tx.dma_dev, lcdif_data->dma_tx.channel, &stream->dma_cfg); in mipi_dbi_flexio_ldcif_write_display()
297 dma_start(lcdif_data->dma_tx.dma_dev, lcdif_data->dma_tx.channel); in mipi_dbi_flexio_ldcif_write_display()
379 if (!device_is_ready(lcdif_data->dma_tx.dma_dev)) { in flexio_lcdif_init()
380 LOG_ERR("%s device is not ready", lcdif_data->dma_tx.dma_dev->name); in flexio_lcdif_init()
465 .dma_dev = DEVICE_DT_GET(DT_INST_DMAS_CTLR_BY_NAME(n, tx)), \
/Zephyr-latest/drivers/adc/
Dadc_esp32.c57 const struct device *dma_dev; member
152 static void IRAM_ATTR adc_esp32_dma_conv_done(const struct device *dma_dev, void *user_data, in adc_esp32_dma_conv_done() argument
155 ARG_UNUSED(dma_dev); in adc_esp32_dma_conv_done()
173 err = dma_get_status(conf->dma_dev, conf->dma_channel, &dma_status); in adc_esp32_dma_start()
196 err = dma_config(conf->dma_dev, conf->dma_channel, &dma_cfg); in adc_esp32_dma_start()
202 err = dma_start(conf->dma_dev, conf->dma_channel); in adc_esp32_dma_start()
219 err = dma_stop(conf->dma_dev, conf->dma_channel); in adc_esp32_dma_stop()
742 .dma_dev = COND_CODE_1(DT_INST_NODE_HAS_PROP(n, dmas), \

12