Lines Matching refs:atchan

80 static struct at_desc *atc_first_active(struct at_dma_chan *atchan)  in atc_first_active()  argument
82 return list_first_entry(&atchan->active_list, in atc_first_active()
86 static struct at_desc *atc_first_queued(struct at_dma_chan *atchan) in atc_first_queued() argument
88 return list_first_entry(&atchan->queue, in atc_first_queued()
126 static struct at_desc *atc_desc_get(struct at_dma_chan *atchan) in atc_desc_get() argument
133 spin_lock_irqsave(&atchan->lock, flags); in atc_desc_get()
134 list_for_each_entry_safe(desc, _desc, &atchan->free_list, desc_node) { in atc_desc_get()
141 dev_dbg(chan2dev(&atchan->chan_common), in atc_desc_get()
144 spin_unlock_irqrestore(&atchan->lock, flags); in atc_desc_get()
145 dev_vdbg(chan2dev(&atchan->chan_common), in atc_desc_get()
150 ret = atc_alloc_descriptor(&atchan->chan_common, GFP_NOWAIT); in atc_desc_get()
160 static void atc_desc_put(struct at_dma_chan *atchan, struct at_desc *desc) in atc_desc_put() argument
166 spin_lock_irqsave(&atchan->lock, flags); in atc_desc_put()
168 dev_vdbg(chan2dev(&atchan->chan_common), in atc_desc_put()
171 list_splice_init(&desc->tx_list, &atchan->free_list); in atc_desc_put()
172 dev_vdbg(chan2dev(&atchan->chan_common), in atc_desc_put()
174 list_add(&desc->desc_node, &atchan->free_list); in atc_desc_put()
175 spin_unlock_irqrestore(&atchan->lock, flags); in atc_desc_put()
209 static void atc_dostart(struct at_dma_chan *atchan, struct at_desc *first) in atc_dostart() argument
211 struct at_dma *atdma = to_at_dma(atchan->chan_common.device); in atc_dostart()
214 if (atc_chan_is_enabled(atchan)) { in atc_dostart()
215 dev_err(chan2dev(&atchan->chan_common), in atc_dostart()
217 dev_err(chan2dev(&atchan->chan_common), in atc_dostart()
219 channel_readl(atchan, SADDR), in atc_dostart()
220 channel_readl(atchan, DADDR), in atc_dostart()
221 channel_readl(atchan, CTRLA), in atc_dostart()
222 channel_readl(atchan, CTRLB), in atc_dostart()
223 channel_readl(atchan, DSCR)); in atc_dostart()
229 vdbg_dump_regs(atchan); in atc_dostart()
231 channel_writel(atchan, SADDR, 0); in atc_dostart()
232 channel_writel(atchan, DADDR, 0); in atc_dostart()
233 channel_writel(atchan, CTRLA, 0); in atc_dostart()
234 channel_writel(atchan, CTRLB, 0); in atc_dostart()
235 channel_writel(atchan, DSCR, first->txd.phys); in atc_dostart()
236 channel_writel(atchan, SPIP, ATC_SPIP_HOLE(first->src_hole) | in atc_dostart()
238 channel_writel(atchan, DPIP, ATC_DPIP_HOLE(first->dst_hole) | in atc_dostart()
240 dma_writel(atdma, CHER, atchan->mask); in atc_dostart()
242 vdbg_dump_regs(atchan); in atc_dostart()
250 static struct at_desc *atc_get_desc_by_cookie(struct at_dma_chan *atchan, in atc_get_desc_by_cookie() argument
255 list_for_each_entry_safe(desc, _desc, &atchan->queue, desc_node) { in atc_get_desc_by_cookie()
260 list_for_each_entry_safe(desc, _desc, &atchan->active_list, desc_node) { in atc_get_desc_by_cookie()
296 struct at_dma_chan *atchan = to_at_dma_chan(chan); in atc_get_bytes_left() local
297 struct at_desc *desc_first = atc_first_active(atchan); in atc_get_bytes_left()
307 desc = atc_get_desc_by_cookie(atchan, cookie); in atc_get_bytes_left()
367 dscr = channel_readl(atchan, DSCR); in atc_get_bytes_left()
369 ctrla = channel_readl(atchan, CTRLA); in atc_get_bytes_left()
374 new_dscr = channel_readl(atchan, DSCR); in atc_get_bytes_left()
394 ctrla = channel_readl(atchan, CTRLA); in atc_get_bytes_left()
418 ctrla = channel_readl(atchan, CTRLA); in atc_get_bytes_left()
431 atc_chain_complete(struct at_dma_chan *atchan, struct at_desc *desc) in atc_chain_complete() argument
434 struct at_dma *atdma = to_at_dma(atchan->chan_common.device); in atc_chain_complete()
437 dev_vdbg(chan2dev(&atchan->chan_common), in atc_chain_complete()
440 spin_lock_irqsave(&atchan->lock, flags); in atc_chain_complete()
443 if (!atc_chan_is_cyclic(atchan)) in atc_chain_complete()
454 list_splice_init(&desc->tx_list, &atchan->free_list); in atc_chain_complete()
456 list_move(&desc->desc_node, &atchan->free_list); in atc_chain_complete()
458 spin_unlock_irqrestore(&atchan->lock, flags); in atc_chain_complete()
463 if (!atc_chan_is_cyclic(atchan)) in atc_chain_complete()
478 static void atc_complete_all(struct at_dma_chan *atchan) in atc_complete_all() argument
484 dev_vdbg(chan2dev(&atchan->chan_common), "complete all\n"); in atc_complete_all()
486 spin_lock_irqsave(&atchan->lock, flags); in atc_complete_all()
492 if (!list_empty(&atchan->queue)) in atc_complete_all()
493 atc_dostart(atchan, atc_first_queued(atchan)); in atc_complete_all()
495 list_splice_init(&atchan->active_list, &list); in atc_complete_all()
497 list_splice_init(&atchan->queue, &atchan->active_list); in atc_complete_all()
499 spin_unlock_irqrestore(&atchan->lock, flags); in atc_complete_all()
502 atc_chain_complete(atchan, desc); in atc_complete_all()
509 static void atc_advance_work(struct at_dma_chan *atchan) in atc_advance_work() argument
514 dev_vdbg(chan2dev(&atchan->chan_common), "advance_work\n"); in atc_advance_work()
516 spin_lock_irqsave(&atchan->lock, flags); in atc_advance_work()
517 ret = atc_chan_is_enabled(atchan); in atc_advance_work()
518 spin_unlock_irqrestore(&atchan->lock, flags); in atc_advance_work()
522 if (list_empty(&atchan->active_list) || in atc_advance_work()
523 list_is_singular(&atchan->active_list)) in atc_advance_work()
524 return atc_complete_all(atchan); in atc_advance_work()
526 atc_chain_complete(atchan, atc_first_active(atchan)); in atc_advance_work()
529 spin_lock_irqsave(&atchan->lock, flags); in atc_advance_work()
530 atc_dostart(atchan, atc_first_active(atchan)); in atc_advance_work()
531 spin_unlock_irqrestore(&atchan->lock, flags); in atc_advance_work()
539 static void atc_handle_error(struct at_dma_chan *atchan) in atc_handle_error() argument
545 spin_lock_irqsave(&atchan->lock, flags); in atc_handle_error()
551 bad_desc = atc_first_active(atchan); in atc_handle_error()
556 list_splice_init(&atchan->queue, atchan->active_list.prev); in atc_handle_error()
559 if (!list_empty(&atchan->active_list)) in atc_handle_error()
560 atc_dostart(atchan, atc_first_active(atchan)); in atc_handle_error()
569 dev_crit(chan2dev(&atchan->chan_common), in atc_handle_error()
571 dev_crit(chan2dev(&atchan->chan_common), in atc_handle_error()
573 atc_dump_lli(atchan, &bad_desc->lli); in atc_handle_error()
575 atc_dump_lli(atchan, &child->lli); in atc_handle_error()
577 spin_unlock_irqrestore(&atchan->lock, flags); in atc_handle_error()
580 atc_chain_complete(atchan, bad_desc); in atc_handle_error()
587 static void atc_handle_cyclic(struct at_dma_chan *atchan) in atc_handle_cyclic() argument
589 struct at_desc *first = atc_first_active(atchan); in atc_handle_cyclic()
592 dev_vdbg(chan2dev(&atchan->chan_common), in atc_handle_cyclic()
594 channel_readl(atchan, DSCR)); in atc_handle_cyclic()
603 struct at_dma_chan *atchan = from_tasklet(atchan, t, tasklet); in atc_tasklet() local
605 if (test_and_clear_bit(ATC_IS_ERROR, &atchan->status)) in atc_tasklet()
606 return atc_handle_error(atchan); in atc_tasklet()
608 if (atc_chan_is_cyclic(atchan)) in atc_tasklet()
609 return atc_handle_cyclic(atchan); in atc_tasklet()
611 atc_advance_work(atchan); in atc_tasklet()
617 struct at_dma_chan *atchan; in at_dma_interrupt() local
635 atchan = &atdma->chan[i]; in at_dma_interrupt()
640 AT_DMA_RES(i) | atchan->mask); in at_dma_interrupt()
642 set_bit(ATC_IS_ERROR, &atchan->status); in at_dma_interrupt()
644 tasklet_schedule(&atchan->tasklet); in at_dma_interrupt()
668 struct at_dma_chan *atchan = to_at_dma_chan(tx->chan); in atc_tx_submit() local
672 spin_lock_irqsave(&atchan->lock, flags); in atc_tx_submit()
675 if (list_empty(&atchan->active_list)) { in atc_tx_submit()
678 atc_dostart(atchan, desc); in atc_tx_submit()
679 list_add_tail(&desc->desc_node, &atchan->active_list); in atc_tx_submit()
683 list_add_tail(&desc->desc_node, &atchan->queue); in atc_tx_submit()
686 spin_unlock_irqrestore(&atchan->lock, flags); in atc_tx_submit()
702 struct at_dma_chan *atchan = to_at_dma_chan(chan); in atc_prep_dma_interleaved() local
763 desc = atc_desc_get(atchan); in atc_prep_dma_interleaved()
802 struct at_dma_chan *atchan = to_at_dma_chan(chan); in atc_prep_dma_memcpy() local
839 desc = atc_desc_get(atchan); in atc_prep_dma_memcpy()
866 atc_desc_put(atchan, first); in atc_prep_dma_memcpy()
875 struct at_dma_chan *atchan = to_at_dma_chan(chan); in atc_create_memset_desc() local
892 desc = atc_desc_get(atchan); in atc_create_memset_desc()
981 struct at_dma_chan *atchan = to_at_dma_chan(chan); in atc_prep_dma_memset_sg() local
1048 atc_desc_put(atchan, first); in atc_prep_dma_memset_sg()
1066 struct at_dma_chan *atchan = to_at_dma_chan(chan); in atc_prep_slave_sg() local
1068 struct dma_slave_config *sconfig = &atchan->dma_sconfig; in atc_prep_slave_sg()
1101 | ATC_SIF(atchan->mem_if) | ATC_DIF(atchan->per_if); in atc_prep_slave_sg()
1108 desc = atc_desc_get(atchan); in atc_prep_slave_sg()
1141 | ATC_SIF(atchan->per_if) | ATC_DIF(atchan->mem_if); in atc_prep_slave_sg()
1149 desc = atc_desc_get(atchan); in atc_prep_slave_sg()
1195 atc_desc_put(atchan, first); in atc_prep_slave_sg()
1229 struct at_dma_chan *atchan = to_at_dma_chan(chan); in atc_dma_cyclic_fill_desc() local
1230 struct dma_slave_config *sconfig = &atchan->dma_sconfig; in atc_dma_cyclic_fill_desc()
1248 | ATC_SIF(atchan->mem_if) in atc_dma_cyclic_fill_desc()
1249 | ATC_DIF(atchan->per_if); in atc_dma_cyclic_fill_desc()
1260 | ATC_SIF(atchan->per_if) in atc_dma_cyclic_fill_desc()
1261 | ATC_DIF(atchan->mem_if); in atc_dma_cyclic_fill_desc()
1286 struct at_dma_chan *atchan = to_at_dma_chan(chan); in atc_prep_dma_cyclic() local
1288 struct dma_slave_config *sconfig = &atchan->dma_sconfig; in atc_prep_dma_cyclic()
1306 was_cyclic = test_and_set_bit(ATC_IS_CYCLIC, &atchan->status); in atc_prep_dma_cyclic()
1328 desc = atc_desc_get(atchan); in atc_prep_dma_cyclic()
1350 atc_desc_put(atchan, first); in atc_prep_dma_cyclic()
1352 clear_bit(ATC_IS_CYCLIC, &atchan->status); in atc_prep_dma_cyclic()
1359 struct at_dma_chan *atchan = to_at_dma_chan(chan); in atc_config() local
1367 memcpy(&atchan->dma_sconfig, sconfig, sizeof(*sconfig)); in atc_config()
1369 convert_burst(&atchan->dma_sconfig.src_maxburst); in atc_config()
1370 convert_burst(&atchan->dma_sconfig.dst_maxburst); in atc_config()
1377 struct at_dma_chan *atchan = to_at_dma_chan(chan); in atc_pause() local
1379 int chan_id = atchan->chan_common.chan_id; in atc_pause()
1384 spin_lock_irqsave(&atchan->lock, flags); in atc_pause()
1387 set_bit(ATC_IS_PAUSED, &atchan->status); in atc_pause()
1389 spin_unlock_irqrestore(&atchan->lock, flags); in atc_pause()
1396 struct at_dma_chan *atchan = to_at_dma_chan(chan); in atc_resume() local
1398 int chan_id = atchan->chan_common.chan_id; in atc_resume()
1403 if (!atc_chan_is_paused(atchan)) in atc_resume()
1406 spin_lock_irqsave(&atchan->lock, flags); in atc_resume()
1409 clear_bit(ATC_IS_PAUSED, &atchan->status); in atc_resume()
1411 spin_unlock_irqrestore(&atchan->lock, flags); in atc_resume()
1418 struct at_dma_chan *atchan = to_at_dma_chan(chan); in atc_terminate_all() local
1420 int chan_id = atchan->chan_common.chan_id; in atc_terminate_all()
1434 spin_lock_irqsave(&atchan->lock, flags); in atc_terminate_all()
1437 dma_writel(atdma, CHDR, AT_DMA_RES(chan_id) | atchan->mask); in atc_terminate_all()
1440 while (dma_readl(atdma, CHSR) & atchan->mask) in atc_terminate_all()
1444 list_splice_init(&atchan->queue, &list); in atc_terminate_all()
1445 list_splice_init(&atchan->active_list, &list); in atc_terminate_all()
1447 spin_unlock_irqrestore(&atchan->lock, flags); in atc_terminate_all()
1451 atc_chain_complete(atchan, desc); in atc_terminate_all()
1453 clear_bit(ATC_IS_PAUSED, &atchan->status); in atc_terminate_all()
1455 clear_bit(ATC_IS_CYCLIC, &atchan->status); in atc_terminate_all()
1475 struct at_dma_chan *atchan = to_at_dma_chan(chan); in atc_tx_status() local
1490 spin_lock_irqsave(&atchan->lock, flags); in atc_tx_status()
1495 spin_unlock_irqrestore(&atchan->lock, flags); in atc_tx_status()
1516 struct at_dma_chan *atchan = to_at_dma_chan(chan); in atc_issue_pending() local
1521 if (atc_chan_is_cyclic(atchan)) in atc_issue_pending()
1524 atc_advance_work(atchan); in atc_issue_pending()
1535 struct at_dma_chan *atchan = to_at_dma_chan(chan); in atc_alloc_chan_resources() local
1545 if (atc_chan_is_enabled(atchan)) { in atc_alloc_chan_resources()
1550 if (!list_empty(&atchan->free_list)) { in atc_alloc_chan_resources()
1578 list_add_tail(&desc->desc_node, &atchan->free_list); in atc_alloc_chan_resources()
1584 channel_writel(atchan, CFG, cfg); in atc_alloc_chan_resources()
1598 struct at_dma_chan *atchan = to_at_dma_chan(chan); in atc_free_chan_resources() local
1604 BUG_ON(!list_empty(&atchan->active_list)); in atc_free_chan_resources()
1605 BUG_ON(!list_empty(&atchan->queue)); in atc_free_chan_resources()
1606 BUG_ON(atc_chan_is_enabled(atchan)); in atc_free_chan_resources()
1608 list_for_each_entry_safe(desc, _desc, &atchan->free_list, desc_node) { in atc_free_chan_resources()
1614 list_splice_init(&atchan->free_list, &list); in atc_free_chan_resources()
1615 atchan->status = 0; in atc_free_chan_resources()
1643 struct at_dma_chan *atchan; in at_dma_xlate() local
1698 atchan = to_at_dma_chan(chan); in at_dma_xlate()
1699 atchan->per_if = dma_spec->args[0] & 0xff; in at_dma_xlate()
1700 atchan->mem_if = (dma_spec->args[0] >> 16) & 0xff; in at_dma_xlate()
1878 struct at_dma_chan *atchan = &atdma->chan[i]; in at_dma_probe() local
1880 atchan->mem_if = AT_DMA_MEM_IF; in at_dma_probe()
1881 atchan->per_if = AT_DMA_PER_IF; in at_dma_probe()
1882 atchan->chan_common.device = &atdma->dma_common; in at_dma_probe()
1883 dma_cookie_init(&atchan->chan_common); in at_dma_probe()
1884 list_add_tail(&atchan->chan_common.device_node, in at_dma_probe()
1887 atchan->ch_regs = atdma->regs + ch_regs(i); in at_dma_probe()
1888 spin_lock_init(&atchan->lock); in at_dma_probe()
1889 atchan->mask = 1 << i; in at_dma_probe()
1891 INIT_LIST_HEAD(&atchan->active_list); in at_dma_probe()
1892 INIT_LIST_HEAD(&atchan->queue); in at_dma_probe()
1893 INIT_LIST_HEAD(&atchan->free_list); in at_dma_probe()
1895 tasklet_setup(&atchan->tasklet, atc_tasklet); in at_dma_probe()
1998 struct at_dma_chan *atchan = to_at_dma_chan(chan); in at_dma_remove() local
2003 tasklet_kill(&atchan->tasklet); in at_dma_remove()
2036 struct at_dma_chan *atchan = to_at_dma_chan(chan); in at_dma_prepare() local
2038 if (atc_chan_is_enabled(atchan) && !atc_chan_is_cyclic(atchan)) in at_dma_prepare()
2044 static void atc_suspend_cyclic(struct at_dma_chan *atchan) in atc_suspend_cyclic() argument
2046 struct dma_chan *chan = &atchan->chan_common; in atc_suspend_cyclic()
2050 if (!atc_chan_is_paused(atchan)) { in atc_suspend_cyclic()
2058 atchan->save_dscr = channel_readl(atchan, DSCR); in atc_suspend_cyclic()
2060 vdbg_dump_regs(atchan); in atc_suspend_cyclic()
2071 struct at_dma_chan *atchan = to_at_dma_chan(chan); in at_dma_suspend_noirq() local
2073 if (atc_chan_is_cyclic(atchan)) in at_dma_suspend_noirq()
2074 atc_suspend_cyclic(atchan); in at_dma_suspend_noirq()
2075 atchan->save_cfg = channel_readl(atchan, CFG); in at_dma_suspend_noirq()
2085 static void atc_resume_cyclic(struct at_dma_chan *atchan) in atc_resume_cyclic() argument
2087 struct at_dma *atdma = to_at_dma(atchan->chan_common.device); in atc_resume_cyclic()
2091 channel_writel(atchan, SADDR, 0); in atc_resume_cyclic()
2092 channel_writel(atchan, DADDR, 0); in atc_resume_cyclic()
2093 channel_writel(atchan, CTRLA, 0); in atc_resume_cyclic()
2094 channel_writel(atchan, CTRLB, 0); in atc_resume_cyclic()
2095 channel_writel(atchan, DSCR, atchan->save_dscr); in atc_resume_cyclic()
2096 dma_writel(atdma, CHER, atchan->mask); in atc_resume_cyclic()
2101 vdbg_dump_regs(atchan); in atc_resume_cyclic()
2121 struct at_dma_chan *atchan = to_at_dma_chan(chan); in at_dma_resume_noirq() local
2123 channel_writel(atchan, CFG, atchan->save_cfg); in at_dma_resume_noirq()
2124 if (atc_chan_is_cyclic(atchan)) in at_dma_resume_noirq()
2125 atc_resume_cyclic(atchan); in at_dma_resume_noirq()