Lines Matching full:async

382 	struct uart_ns16550_async_data async;  member
827 if (data->async.tx_dma_params.dma_dev != NULL) {
845 if (data->async.tx_dma_params.dma_dev != NULL) {
846 data->async.next_rx_buffer = NULL;
847 data->async.next_rx_buffer_len = 0;
848 data->async.uart_dev = dev;
849 k_work_init_delayable(&data->async.rx_dma_params.timeout_work,
851 k_work_init_delayable(&data->async.tx_dma_params.timeout_work,
853 data->async.rx_dma_params.dma_cfg.head_block =
854 &data->async.rx_dma_params.active_dma_block;
855 data->async.tx_dma_params.dma_cfg.head_block =
856 &data->async.tx_dma_params.active_dma_block;
865 dma_intel_lpss_set_base(data->async.tx_dma_params.dma_dev, base);
866 dma_intel_lpss_setup(data->async.tx_dma_params.dma_dev);
1293 if (dev_data->async.tx_dma_params.dma_dev != NULL) {
1299 if (dma_status & BIT(dev_data->async.rx_dma_params.dma_channel)) {
1300 async_timer_start(&dev_data->async.rx_dma_params.timeout_work,
1301 dev_data->async.rx_dma_params.timeout_us);
1303 BIT(dev_data->async.rx_dma_params.dma_channel));
1306 dma_intel_lpss_isr(dev_data->async.rx_dma_params.dma_dev);
1309 async_timer_start(&dev_data->async.rx_dma_params.timeout_work,
1310 dev_data->async.rx_dma_params.timeout_us);
1421 if (data->async.user_callback) {
1422 data->async.user_callback(dev, evt, data->async.user_data);
1430 struct uart_ns16550_tx_dma_params *tx_params = &data->async.tx_dma_params;
1432 (void)k_work_cancel_delayable(&data->async.tx_dma_params.timeout_work);
1450 struct uart_ns16550_rx_dma_params *dma_params = &data->async.rx_dma_params;
1471 .data.rx_buf.buf = data->async.rx_dma_params.buf
1475 data->async.rx_dma_params.buf = NULL;
1476 data->async.rx_dma_params.buf_len = 0U;
1477 data->async.rx_dma_params.offset = 0U;
1478 data->async.rx_dma_params.counter = 0U;
1492 struct uart_ns16550_rx_dma_params *dma_params = &data->async.rx_dma_params;
1510 struct uart_ns16550_rx_dma_params *dma_params = &data->async.rx_dma_params;
1519 (void)k_work_cancel_delayable(&data->async.rx_dma_params.timeout_work);
1524 if (data->async.next_rx_buffer != NULL) {
1525 dma_params->buf = data->async.next_rx_buffer;
1526 dma_params->buf_len = data->async.next_rx_buffer_len;
1527 data->async.next_rx_buffer = NULL;
1528 data->async.next_rx_buffer_len = 0;
1549 struct uart_ns16550_rx_dma_params *rx_dma_params = &data->async.rx_dma_params;
1567 struct uart_ns16550_rx_dma_params *rx_params = &data->async.rx_dma_params;
1568 struct uart_ns16550_tx_dma_params *tx_params = &data->async.tx_dma_params;
1579 rx_params->buf = data->async.next_rx_buffer;
1580 rx_params->buf_len = data->async.next_rx_buffer_len;
1581 data->async.next_rx_buffer = NULL;
1582 data->async.next_rx_buffer_len = 0U;
1602 data->async.user_callback = callback;
1603 data->async.user_data = user_data;
1612 struct uart_ns16550_tx_dma_params *tx_params = &data->async.tx_dma_params;
1639 async_timer_start(&data->async.tx_dma_params.timeout_work, timeout_us);
1650 struct uart_ns16550_tx_dma_params *tx_params = &data->async.tx_dma_params;
1662 (void)k_work_cancel_delayable(&data->async.tx_dma_params.timeout_work);
1688 struct uart_ns16550_rx_dma_params *rx_dma_params = &data->async.rx_dma_params;
1723 assert(data->async.next_rx_buffer == NULL);
1724 assert(data->async.next_rx_buffer_len == 0);
1725 data->async.next_rx_buffer = buf;
1726 data->async.next_rx_buffer_len = len;
1871 .async.tx_dma_params = { \
1890 .async.rx_dma_params = { \
1913 .async.tx_dma_params = { \
1916 .async.rx_dma_params = { \