Home
last modified time | relevance | path

Searched refs:dma_ctx (Results 1 – 15 of 15) sorted by relevance

/Zephyr-latest/include/zephyr/drivers/
Ddma.h577 struct dma_context *dma_ctx = (struct dma_context *)dev->data; in z_impl_dma_request_channel() local
579 if (dma_ctx->magic != DMA_MAGIC) { in z_impl_dma_request_channel()
583 for (i = 0; i < dma_ctx->dma_channels; i++) { in z_impl_dma_request_channel()
584 if (!atomic_test_and_set_bit(dma_ctx->atomic, i)) { in z_impl_dma_request_channel()
587 atomic_clear_bit(dma_ctx->atomic, i); in z_impl_dma_request_channel()
619 struct dma_context *dma_ctx = (struct dma_context *)dev->data; in z_impl_dma_release_channel() local
621 if (dma_ctx->magic != DMA_MAGIC) { in z_impl_dma_release_channel()
625 if ((int)channel < dma_ctx->dma_channels) { in z_impl_dma_release_channel()
630 atomic_clear_bit(dma_ctx->atomic, channel); in z_impl_dma_release_channel()
/Zephyr-latest/drivers/dma/
Ddma_silabs_ldma.c40 struct dma_context dma_ctx; member
275 for (chnum = 0; chnum < data->dma_ctx.dma_channels; chnum++) { in dma_silabs_irq_handler()
310 if (channel > data->dma_ctx.dma_channels) { in dma_silabs_configure()
403 atomic_set_bit(data->dma_ctx.atomic, channel); in dma_silabs_configure()
414 if (channel > data->dma_ctx.dma_channels) { in dma_silabs_start()
430 if (channel > data->dma_ctx.dma_channels) { in dma_silabs_stop()
448 if (channel > data->dma_ctx.dma_channels) { in dma_silabs_get_status()
452 if (!atomic_test_bit(data->dma_ctx.atomic, channel)) { in dma_silabs_get_status()
516 .dma_ctx.magic = DMA_MAGIC, \
517 .dma_ctx.dma_channels = DT_INST_PROP(inst, dma_channels), \
[all …]
Ddma_dw_axi.c231 struct dma_context dma_ctx; member
303 if (channel > (dw_dev_data->dma_ctx.dma_channels - 1)) { in dma_dw_axi_isr()
457 if (channel > (dw_dev_data->dma_ctx.dma_channels - 1)) { in dma_dw_axi_config()
517 if (cfg->channel_priority < dw_dev_data->dma_ctx.dma_channels) { in dma_dw_axi_config()
646 if (channel > (dw_dev_data->dma_ctx.dma_channels - 1)) { in dma_dw_axi_start()
708 if (channel > (dw_dev_data->dma_ctx.dma_channels - 1)) { in dma_dw_axi_stop()
759 if (channel > (dw_dev_data->dma_ctx.dma_channels - 1)) { in dma_dw_axi_resume()
790 if (channel > (dw_dev_data->dma_ctx.dma_channels - 1)) { in dma_dw_axi_suspend()
842 for (i = 0; i < dw_dev_data->dma_ctx.dma_channels; i++) { in dma_dw_axi_init()
883 .dma_ctx = { \
Ddma_smartbond.c156 struct dma_context dma_ctx; member
961 for (int i = 0; i < data->dma_ctx.dma_channels; i++) { in dma_smartbond_is_sleep_allowed()
962 if (atomic_test_bit(data->dma_ctx.atomic, i)) { in dma_smartbond_is_sleep_allowed()
1018 data->dma_ctx.magic = DMA_MAGIC; in dma_smartbond_init()
1019 data->dma_ctx.dma_channels = DMA_CHANNELS_COUNT; in dma_smartbond_init()
1020 data->dma_ctx.atomic = data->channels_atomic; in dma_smartbond_init()
Ddma_intel_adsp_hda.c443 struct dma_context *dma_ctx; in intel_adsp_hda_dma_isr() local
467 dma_ctx = (struct dma_context *)host_dev[i]->data; in intel_adsp_hda_dma_isr()
469 enabled_chs = atomic_get(dma_ctx->atomic); in intel_adsp_hda_dma_isr()
470 for (j = 0; enabled_chs && j < dma_ctx->dma_channels; j++) { in intel_adsp_hda_dma_isr()
Ddma_nxp_sdma.c47 struct dma_context dma_ctx; member
446 data->dma_ctx.magic = DMA_MAGIC; in dma_nxp_sdma_init()
447 data->dma_ctx.dma_channels = FSL_FEATURE_SDMA_MODULE_CHANNEL; in dma_nxp_sdma_init()
448 data->dma_ctx.atomic = data->channels_atomic; in dma_nxp_sdma_init()
Ddma_emul.c62 struct dma_context dma_ctx; member
544 data->dma_ctx.magic = DMA_MAGIC; in dma_emul_init()
545 data->dma_ctx.dma_channels = config->num_channels; in dma_emul_init()
546 data->dma_ctx.atomic = data->channels_atomic; in dma_emul_init()
Ddma_stm32_bdma.h35 struct dma_context dma_ctx; member
Ddma_intel_adsp_gpdma.c471 dev_data->dma_ctx.magic = DMA_MAGIC; in intel_adsp_gpdma_init()
472 dev_data->dma_ctx.dma_channels = DW_MAX_CHAN; in intel_adsp_gpdma_init()
473 dev_data->dma_ctx.atomic = dev_data->channels_atomic; in intel_adsp_gpdma_init()
Ddma_stm32.h34 struct dma_context dma_ctx; member
Ddma_stm32.c661 ((struct dma_stm32_data *)dev->data)->dma_ctx.magic = 0; in dma_stm32_init()
662 ((struct dma_stm32_data *)dev->data)->dma_ctx.dma_channels = 0; in dma_stm32_init()
663 ((struct dma_stm32_data *)dev->data)->dma_ctx.atomic = 0; in dma_stm32_init()
Ddma_stm32_bdma.c810 ((struct bdma_stm32_data *)dev->data)->dma_ctx.magic = 0; in bdma_stm32_init()
811 ((struct bdma_stm32_data *)dev->data)->dma_ctx.dma_channels = 0; in bdma_stm32_init()
812 ((struct bdma_stm32_data *)dev->data)->dma_ctx.atomic = 0; in bdma_stm32_init()
Ddma_stm32u5.c682 ((struct dma_stm32_data *)dev->data)->dma_ctx.magic = 0; in dma_stm32_init()
683 ((struct dma_stm32_data *)dev->data)->dma_ctx.dma_channels = 0; in dma_stm32_init()
684 ((struct dma_stm32_data *)dev->data)->dma_ctx.atomic = 0; in dma_stm32_init()
Ddma_mcux_edma.c111 struct dma_context dma_ctx; member
842 data->dma_ctx.magic = DMA_MAGIC; in dma_mcux_edma_init()
843 data->dma_ctx.dma_channels = config->dma_channels; in dma_mcux_edma_init()
844 data->dma_ctx.atomic = data->channels_atomic; in dma_mcux_edma_init()
Ddma_dw_common.h247 struct dma_context dma_ctx; member