Home
last modified time | relevance | path

Searched refs:dma_tx (Results 1 – 15 of 15) sorted by relevance

/Zephyr-latest/drivers/spi/
Dspi_andes_atcspi200.c45 struct stream dma_tx; member
234 data->dma_tx.dma_blk_cfg.next_block = NULL; in spi_dma_move_buffers()
282 dma_stop(data->dma_tx.dma_dev, data->dma_tx.channel); in dma_tx_callback()
290 error = dma_start(data->dma_tx.dma_dev, data->dma_tx.channel); in dma_tx_callback()
309 memset(&data->dma_tx.dma_blk_cfg, 0, sizeof(struct dma_block_config)); in spi_dma_tx_load()
312 data->dma_tx.dma_blk_cfg.block_size = data->chunk_len / in spi_dma_tx_load()
313 data->dma_tx.dma_cfg.dest_data_size; in spi_dma_tx_load()
315 data->dma_tx.dma_blk_cfg.block_size = ctx->current_tx->len / in spi_dma_tx_load()
316 data->dma_tx.dma_cfg.dest_data_size; in spi_dma_tx_load()
323 data->dma_tx.dma_blk_cfg.source_address = (uintptr_t)&dummy_rx_tx_buffer; in spi_dma_tx_load()
[all …]
Dspi_xmc4xxx.c62 struct spi_xmc4xxx_dma_stream dma_tx; member
81 if (dev_dma == data->dma_tx.dev_dma && dma_channel == data->dma_tx.dma_channel) { in spi_xmc4xxx_dma_callback()
355 struct spi_xmc4xxx_dma_stream *dma_tx = &data->dma_tx; in spi_xmc4xxx_transceive_dma() local
430 dma_tx->blk_cfg.source_address = (uint32_t)ctx->tx_buf; in spi_xmc4xxx_transceive_dma()
431 dma_tx->blk_cfg.source_addr_adj = DMA_ADDR_ADJ_INCREMENT; in spi_xmc4xxx_transceive_dma()
433 dma_tx->blk_cfg.source_address = (uint32_t)&tx_dummy_data; in spi_xmc4xxx_transceive_dma()
434 dma_tx->blk_cfg.source_addr_adj = DMA_ADDR_ADJ_NO_CHANGE; in spi_xmc4xxx_transceive_dma()
437 dma_tx->blk_cfg.block_size = dma_len; in spi_xmc4xxx_transceive_dma()
439 ret = dma_config(dma_tx->dev_dma, dma_tx->dma_channel, &dma_tx->dma_cfg); in spi_xmc4xxx_transceive_dma()
450 ret = dma_start(dma_tx->dev_dma, dma_tx->dma_channel); in spi_xmc4xxx_transceive_dma()
[all …]
Dspi_mcux_flexcomm.c69 struct stream dma_tx; member
314 if (channel == data->dma_tx.channel) { in spi_mcux_dma_callback()
315 if (status != data->dma_tx.wait_for_dma_status) { in spi_mcux_dma_callback()
398 struct stream *stream = &data->dma_tx; in spi_mcux_dma_tx_load()
431 data->dma_tx.wait_for_dma_status = DMA_STATUS_COMPLETE; in spi_mcux_dma_tx_load()
438 data->dma_tx.wait_for_dma_status = DMA_STATUS_BLOCK; in spi_mcux_dma_tx_load()
464 data->dma_tx.wait_for_dma_status = DMA_STATUS_COMPLETE; in spi_mcux_dma_tx_load()
469 data->dma_tx.wait_for_dma_status = DMA_STATUS_BLOCK; in spi_mcux_dma_tx_load()
481 ret = dma_config(data->dma_tx.dma_dev, data->dma_tx.channel, in spi_mcux_dma_tx_load()
508 return dma_start(data->dma_tx.dma_dev, data->dma_tx.channel); in spi_mcux_dma_tx_load()
[all …]
Dspi_ll_stm32.c141 if (channel == spi_dma_data->dma_tx.channel) { in dma_callback()
165 struct stream *stream = &data->dma_tx; in spi_stm32_dma_tx_load()
184 if (data->dma_tx.src_addr_increment) { in spi_stm32_dma_tx_load()
193 if (data->dma_tx.dst_addr_increment) { in spi_stm32_dma_tx_load()
200 blk_cfg->fifo_mode_control = data->dma_tx.fifo_threshold; in spi_stm32_dma_tx_load()
207 ret = dma_config(data->dma_tx.dma_dev, data->dma_tx.channel, in spi_stm32_dma_tx_load()
215 return dma_start(data->dma_tx.dma_dev, data->dma_tx.channel); in spi_stm32_dma_tx_load()
290 dma_segment_len = len * data->dma_tx.dma_cfg.source_data_size; in spi_dma_move_buffers()
1100 err = dma_stop(data->dma_tx.dma_dev, data->dma_tx.channel); in transceive_dma()
1128 if ((data->dma_tx.dma_dev != NULL) in spi_stm32_transceive()
[all …]
Dspi_mcux_lpspi.c87 struct spi_dma_stream dma_tx; member
230 return (data->dma_tx.dma_dev && data->dma_rx.dma_dev); in lpspi_inst_has_dma()
246 if (channel == data->dma_tx.channel) { in spi_mcux_dma_callback()
317 struct spi_dma_stream *stream = &data->dma_tx; in spi_mcux_dma_tx_load()
395 ret = dma_start(data->dma_tx.dma_dev, data->dma_tx.channel); in spi_mcux_dma_rxtx_load()
752 return lpspi_dma_dev_ready(data->dma_tx.dma_dev) | in lpspi_dma_devs_ready()
809 (.dma_tx = {.dma_dev = DEVICE_DT_GET(DT_INST_DMAS_CTLR_BY_NAME(n, tx)), \
Dspi_ll_stm32.h69 struct stream dma_tx; member
/Zephyr-latest/drivers/serial/
Duart_xmc4xxx.c69 struct uart_dma_stream dma_tx; member
484 .data.tx.buf = data->dma_tx.buffer, in async_evt_tx_done()
485 .data.tx.len = data->dma_tx.counter}; in async_evt_tx_done()
487 data->dma_tx.buffer = NULL; in async_evt_tx_done()
488 data->dma_tx.buffer_len = 0; in async_evt_tx_done()
489 data->dma_tx.counter = 0; in async_evt_tx_done()
499 .data.tx.buf = data->dma_tx.buffer, in async_evt_tx_abort()
500 .data.tx.len = data->dma_tx.counter}; in async_evt_tx_abort()
502 data->dma_tx.buffer = NULL; in async_evt_tx_abort()
503 data->dma_tx.buffer_len = 0; in async_evt_tx_abort()
[all …]
Duart_stm32.c1146 LOG_DBG("tx done: %d", data->dma_tx.counter); in async_evt_tx_done()
1150 .data.tx.buf = data->dma_tx.buffer, in async_evt_tx_done()
1151 .data.tx.len = data->dma_tx.counter in async_evt_tx_done()
1155 data->dma_tx.buffer_length = 0; in async_evt_tx_done()
1156 data->dma_tx.counter = 0; in async_evt_tx_done()
1163 LOG_DBG("tx abort: %d", data->dma_tx.counter); in async_evt_tx_abort()
1167 .data.tx.buf = data->dma_tx.buffer, in async_evt_tx_abort()
1168 .data.tx.len = data->dma_tx.counter in async_evt_tx_abort()
1172 data->dma_tx.buffer_length = 0; in async_evt_tx_abort()
1173 data->dma_tx.counter = 0; in async_evt_tx_abort()
[all …]
Duart_stm32.h103 struct uart_dma_stream dma_tx; member
/Zephyr-latest/subsys/mgmt/ec_host_cmd/backends/
Dec_host_cmd_backend_spi_stm32.c171 struct dma_stream *dma_tx; member
220 static struct dma_stream dma_tx = {SPI_DMA_CHANNEL_INIT(id, tx, TX, MEMORY, PERIPHERAL)}
227 .dma_tx = &dma_tx, \
311 if (channel == hc_spi->dma_tx->channel) { in dma_callback()
370 if ((hc_spi->dma_tx->dma_dev != NULL) && !device_is_ready(hc_spi->dma_tx->dma_dev)) { in spi_init()
371 LOG_ERR("%s device not ready", hc_spi->dma_tx->dma_dev->name); in spi_init()
424 ret = dma_reload(hc_spi->dma_tx->dma_dev, hc_spi->dma_tx->channel, (uint32_t)hc_spi->tx_buf, in reload_dma_tx()
431 ret = dma_start(hc_spi->dma_tx->dma_dev, hc_spi->dma_tx->channel); in reload_dma_tx()
447 struct dma_stream *stream = hc_spi->dma_tx; in spi_config_dma_tx()
463 blk_cfg->fifo_mode_control = hc_spi->dma_tx->fifo_threshold; in spi_config_dma_tx()
[all …]
/Zephyr-latest/drivers/mipi_dbi/
Dmipi_dbi_nxp_flexio_lcdif.c42 struct stream dma_tx; member
239 struct stream *stream = &lcdif_data->dma_tx; in mipi_dbi_flexio_ldcif_write_display()
277 dma_config(lcdif_data->dma_tx.dma_dev, lcdif_data->dma_tx.channel, &stream->dma_cfg); in mipi_dbi_flexio_ldcif_write_display()
283 EDMA_SetModulo(DMA0, lcdif_data->dma_tx.channel, kEDMA_ModuloDisable, in mipi_dbi_flexio_ldcif_write_display()
297 dma_start(lcdif_data->dma_tx.dma_dev, lcdif_data->dma_tx.channel); in mipi_dbi_flexio_ldcif_write_display()
379 if (!device_is_ready(lcdif_data->dma_tx.dma_dev)) { in flexio_lcdif_init()
380 LOG_ERR("%s device is not ready", lcdif_data->dma_tx.dma_dev->name); in flexio_lcdif_init()
464 .dma_tx = { \
/Zephyr-latest/drivers/disk/
Dsdmmc_stm32.c91 struct sdmmc_dma_stream dma_tx; member
239 err = stm32_sdmmc_configure_dma(&priv->dma_tx_handle, &priv->dma_tx); in stm32_sdmmc_dma_init()
261 struct sdmmc_dma_stream *dma_tx = &priv->dma_tx; in stm32_sdmmc_dma_deinit() local
264 ret = dma_stop(dma_tx->dev, dma_tx->channel); in stm32_sdmmc_dma_deinit()
/Zephyr-latest/drivers/flash/
Dflash_stm32_xspi.h108 struct stream dma_tx; member
Dflash_stm32_xspi.c2179 if (flash_stm32_xspi_dma_init(&hdma_tx, &dev_data->dma_tx) != 0) { in flash_stm32_xspi_init()
/Zephyr-latest/drivers/i3c/
Di3c_stm32.c141 struct i3c_stm32_dma_stream dma_tx; /* TX DMA channel config */ member
1143 dma_stream = &(data->dma_tx); in i3c_stm32_dma_msg_config()
1450 dev, &data->dma_tx, 0, in i3c_stm32_init_dma()