Lines Matching refs:pd_chan
172 struct pch_dma_desc *pdc_first_active(struct pch_dma_chan *pd_chan) in pdc_first_active() argument
174 return list_first_entry(&pd_chan->active_list, in pdc_first_active()
179 struct pch_dma_desc *pdc_first_queued(struct pch_dma_chan *pd_chan) in pdc_first_queued() argument
181 return list_first_entry(&pd_chan->queue, in pdc_first_queued()
211 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pdc_set_dir() local
225 if (pd_chan->dir == DMA_MEM_TO_DEV) in pdc_set_dir()
243 if (pd_chan->dir == DMA_MEM_TO_DEV) in pdc_set_dir()
291 static u32 pdc_get_status0(struct pch_dma_chan *pd_chan) in pdc_get_status0() argument
293 struct pch_dma *pd = to_pd(pd_chan->chan.device); in pdc_get_status0()
298 DMA_STATUS_BITS_PER_CH * pd_chan->chan.chan_id)); in pdc_get_status0()
301 static u32 pdc_get_status2(struct pch_dma_chan *pd_chan) in pdc_get_status2() argument
303 struct pch_dma *pd = to_pd(pd_chan->chan.device); in pdc_get_status2()
308 DMA_STATUS_BITS_PER_CH * (pd_chan->chan.chan_id - 8))); in pdc_get_status2()
311 static bool pdc_is_idle(struct pch_dma_chan *pd_chan) in pdc_is_idle() argument
315 if (pd_chan->chan.chan_id < 8) in pdc_is_idle()
316 sts = pdc_get_status0(pd_chan); in pdc_is_idle()
318 sts = pdc_get_status2(pd_chan); in pdc_is_idle()
327 static void pdc_dostart(struct pch_dma_chan *pd_chan, struct pch_dma_desc* desc) in pdc_dostart() argument
329 if (!pdc_is_idle(pd_chan)) { in pdc_dostart()
330 dev_err(chan2dev(&pd_chan->chan), in pdc_dostart()
335 dev_dbg(chan2dev(&pd_chan->chan), "chan %d -> dev_addr: %x\n", in pdc_dostart()
336 pd_chan->chan.chan_id, desc->regs.dev_addr); in pdc_dostart()
337 dev_dbg(chan2dev(&pd_chan->chan), "chan %d -> mem_addr: %x\n", in pdc_dostart()
338 pd_chan->chan.chan_id, desc->regs.mem_addr); in pdc_dostart()
339 dev_dbg(chan2dev(&pd_chan->chan), "chan %d -> size: %x\n", in pdc_dostart()
340 pd_chan->chan.chan_id, desc->regs.size); in pdc_dostart()
341 dev_dbg(chan2dev(&pd_chan->chan), "chan %d -> next: %x\n", in pdc_dostart()
342 pd_chan->chan.chan_id, desc->regs.next); in pdc_dostart()
345 channel_writel(pd_chan, DEV_ADDR, desc->regs.dev_addr); in pdc_dostart()
346 channel_writel(pd_chan, MEM_ADDR, desc->regs.mem_addr); in pdc_dostart()
347 channel_writel(pd_chan, SIZE, desc->regs.size); in pdc_dostart()
348 channel_writel(pd_chan, NEXT, desc->regs.next); in pdc_dostart()
349 pdc_set_mode(&pd_chan->chan, DMA_CTL0_ONESHOT); in pdc_dostart()
351 channel_writel(pd_chan, NEXT, desc->txd.phys); in pdc_dostart()
352 pdc_set_mode(&pd_chan->chan, DMA_CTL0_SG); in pdc_dostart()
356 static void pdc_chain_complete(struct pch_dma_chan *pd_chan, in pdc_chain_complete() argument
363 list_splice_init(&desc->tx_list, &pd_chan->free_list); in pdc_chain_complete()
364 list_move(&desc->desc_node, &pd_chan->free_list); in pdc_chain_complete()
369 static void pdc_complete_all(struct pch_dma_chan *pd_chan) in pdc_complete_all() argument
374 BUG_ON(!pdc_is_idle(pd_chan)); in pdc_complete_all()
376 if (!list_empty(&pd_chan->queue)) in pdc_complete_all()
377 pdc_dostart(pd_chan, pdc_first_queued(pd_chan)); in pdc_complete_all()
379 list_splice_init(&pd_chan->active_list, &list); in pdc_complete_all()
380 list_splice_init(&pd_chan->queue, &pd_chan->active_list); in pdc_complete_all()
383 pdc_chain_complete(pd_chan, desc); in pdc_complete_all()
386 static void pdc_handle_error(struct pch_dma_chan *pd_chan) in pdc_handle_error() argument
390 bad_desc = pdc_first_active(pd_chan); in pdc_handle_error()
393 list_splice_init(&pd_chan->queue, pd_chan->active_list.prev); in pdc_handle_error()
395 if (!list_empty(&pd_chan->active_list)) in pdc_handle_error()
396 pdc_dostart(pd_chan, pdc_first_active(pd_chan)); in pdc_handle_error()
398 dev_crit(chan2dev(&pd_chan->chan), "Bad descriptor submitted\n"); in pdc_handle_error()
399 dev_crit(chan2dev(&pd_chan->chan), "descriptor cookie: %d\n", in pdc_handle_error()
402 pdc_chain_complete(pd_chan, bad_desc); in pdc_handle_error()
405 static void pdc_advance_work(struct pch_dma_chan *pd_chan) in pdc_advance_work() argument
407 if (list_empty(&pd_chan->active_list) || in pdc_advance_work()
408 list_is_singular(&pd_chan->active_list)) { in pdc_advance_work()
409 pdc_complete_all(pd_chan); in pdc_advance_work()
411 pdc_chain_complete(pd_chan, pdc_first_active(pd_chan)); in pdc_advance_work()
412 pdc_dostart(pd_chan, pdc_first_active(pd_chan)); in pdc_advance_work()
419 struct pch_dma_chan *pd_chan = to_pd_chan(txd->chan); in pd_tx_submit() local
421 spin_lock(&pd_chan->lock); in pd_tx_submit()
423 if (list_empty(&pd_chan->active_list)) { in pd_tx_submit()
424 list_add_tail(&desc->desc_node, &pd_chan->active_list); in pd_tx_submit()
425 pdc_dostart(pd_chan, desc); in pd_tx_submit()
427 list_add_tail(&desc->desc_node, &pd_chan->queue); in pd_tx_submit()
430 spin_unlock(&pd_chan->lock); in pd_tx_submit()
452 static struct pch_dma_desc *pdc_desc_get(struct pch_dma_chan *pd_chan) in pdc_desc_get() argument
458 spin_lock(&pd_chan->lock); in pdc_desc_get()
459 list_for_each_entry_safe(desc, _d, &pd_chan->free_list, desc_node) { in pdc_desc_get()
466 dev_dbg(chan2dev(&pd_chan->chan), "desc %p not ACKed\n", desc); in pdc_desc_get()
468 spin_unlock(&pd_chan->lock); in pdc_desc_get()
469 dev_dbg(chan2dev(&pd_chan->chan), "scanned %d descriptors\n", i); in pdc_desc_get()
472 ret = pdc_alloc_desc(&pd_chan->chan, GFP_ATOMIC); in pdc_desc_get()
474 spin_lock(&pd_chan->lock); in pdc_desc_get()
475 pd_chan->descs_allocated++; in pdc_desc_get()
476 spin_unlock(&pd_chan->lock); in pdc_desc_get()
478 dev_err(chan2dev(&pd_chan->chan), in pdc_desc_get()
486 static void pdc_desc_put(struct pch_dma_chan *pd_chan, in pdc_desc_put() argument
490 spin_lock(&pd_chan->lock); in pdc_desc_put()
491 list_splice_init(&desc->tx_list, &pd_chan->free_list); in pdc_desc_put()
492 list_add(&desc->desc_node, &pd_chan->free_list); in pdc_desc_put()
493 spin_unlock(&pd_chan->lock); in pdc_desc_put()
499 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pd_alloc_chan_resources() local
504 if (!pdc_is_idle(pd_chan)) { in pd_alloc_chan_resources()
509 if (!list_empty(&pd_chan->free_list)) in pd_alloc_chan_resources()
510 return pd_chan->descs_allocated; in pd_alloc_chan_resources()
524 spin_lock_irq(&pd_chan->lock); in pd_alloc_chan_resources()
525 list_splice(&tmp_list, &pd_chan->free_list); in pd_alloc_chan_resources()
526 pd_chan->descs_allocated = i; in pd_alloc_chan_resources()
528 spin_unlock_irq(&pd_chan->lock); in pd_alloc_chan_resources()
532 return pd_chan->descs_allocated; in pd_alloc_chan_resources()
537 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pd_free_chan_resources() local
542 BUG_ON(!pdc_is_idle(pd_chan)); in pd_free_chan_resources()
543 BUG_ON(!list_empty(&pd_chan->active_list)); in pd_free_chan_resources()
544 BUG_ON(!list_empty(&pd_chan->queue)); in pd_free_chan_resources()
546 spin_lock_irq(&pd_chan->lock); in pd_free_chan_resources()
547 list_splice_init(&pd_chan->free_list, &tmp_list); in pd_free_chan_resources()
548 pd_chan->descs_allocated = 0; in pd_free_chan_resources()
549 spin_unlock_irq(&pd_chan->lock); in pd_free_chan_resources()
565 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pd_issue_pending() local
567 if (pdc_is_idle(pd_chan)) { in pd_issue_pending()
568 spin_lock(&pd_chan->lock); in pd_issue_pending()
569 pdc_advance_work(pd_chan); in pd_issue_pending()
570 spin_unlock(&pd_chan->lock); in pd_issue_pending()
579 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pd_prep_slave_sg() local
600 pd_chan->dir = direction; in pd_prep_slave_sg()
604 desc = pdc_desc_get(pd_chan); in pd_prep_slave_sg()
656 pdc_desc_put(pd_chan, first); in pd_prep_slave_sg()
662 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pd_device_terminate_all() local
666 spin_lock_irq(&pd_chan->lock); in pd_device_terminate_all()
668 pdc_set_mode(&pd_chan->chan, DMA_CTL0_DISABLE); in pd_device_terminate_all()
670 list_splice_init(&pd_chan->active_list, &list); in pd_device_terminate_all()
671 list_splice_init(&pd_chan->queue, &list); in pd_device_terminate_all()
674 pdc_chain_complete(pd_chan, desc); in pd_device_terminate_all()
676 spin_unlock_irq(&pd_chan->lock); in pd_device_terminate_all()
683 struct pch_dma_chan *pd_chan = (struct pch_dma_chan *)data; in pdc_tasklet() local
686 if (!pdc_is_idle(pd_chan)) { in pdc_tasklet()
687 dev_err(chan2dev(&pd_chan->chan), in pdc_tasklet()
692 spin_lock_irqsave(&pd_chan->lock, flags); in pdc_tasklet()
693 if (test_and_clear_bit(0, &pd_chan->err_status)) in pdc_tasklet()
694 pdc_handle_error(pd_chan); in pdc_tasklet()
696 pdc_advance_work(pd_chan); in pdc_tasklet()
697 spin_unlock_irqrestore(&pd_chan->lock, flags); in pdc_tasklet()
703 struct pch_dma_chan *pd_chan; in pd_irq() local
716 pd_chan = &pd->channels[i]; in pd_irq()
721 set_bit(0, &pd_chan->err_status); in pd_irq()
723 tasklet_schedule(&pd_chan->tasklet); in pd_irq()
729 set_bit(0, &pd_chan->err_status); in pd_irq()
731 tasklet_schedule(&pd_chan->tasklet); in pd_irq()
749 struct pch_dma_chan *pd_chan; in pch_dma_save_regs() local
759 pd_chan = to_pd_chan(chan); in pch_dma_save_regs()
761 pd->ch_regs[i].dev_addr = channel_readl(pd_chan, DEV_ADDR); in pch_dma_save_regs()
762 pd->ch_regs[i].mem_addr = channel_readl(pd_chan, MEM_ADDR); in pch_dma_save_regs()
763 pd->ch_regs[i].size = channel_readl(pd_chan, SIZE); in pch_dma_save_regs()
764 pd->ch_regs[i].next = channel_readl(pd_chan, NEXT); in pch_dma_save_regs()
772 struct pch_dma_chan *pd_chan; in pch_dma_restore_regs() local
782 pd_chan = to_pd_chan(chan); in pch_dma_restore_regs()
784 channel_writel(pd_chan, DEV_ADDR, pd->ch_regs[i].dev_addr); in pch_dma_restore_regs()
785 channel_writel(pd_chan, MEM_ADDR, pd->ch_regs[i].mem_addr); in pch_dma_restore_regs()
786 channel_writel(pd_chan, SIZE, pd->ch_regs[i].size); in pch_dma_restore_regs()
787 channel_writel(pd_chan, NEXT, pd->ch_regs[i].next); in pch_dma_restore_regs()
896 struct pch_dma_chan *pd_chan = &pd->channels[i]; in pch_dma_probe() local
898 pd_chan->chan.device = &pd->dma; in pch_dma_probe()
899 dma_cookie_init(&pd_chan->chan); in pch_dma_probe()
901 pd_chan->membase = ®s->desc[i]; in pch_dma_probe()
903 spin_lock_init(&pd_chan->lock); in pch_dma_probe()
905 INIT_LIST_HEAD(&pd_chan->active_list); in pch_dma_probe()
906 INIT_LIST_HEAD(&pd_chan->queue); in pch_dma_probe()
907 INIT_LIST_HEAD(&pd_chan->free_list); in pch_dma_probe()
909 tasklet_init(&pd_chan->tasklet, pdc_tasklet, in pch_dma_probe()
910 (unsigned long)pd_chan); in pch_dma_probe()
911 list_add_tail(&pd_chan->chan.device_node, &pd->dma.channels); in pch_dma_probe()
951 struct pch_dma_chan *pd_chan; in pch_dma_remove() local
961 pd_chan = to_pd_chan(chan); in pch_dma_remove()
963 tasklet_kill(&pd_chan->tasklet); in pch_dma_remove()