Lines Matching refs:musb

73 struct musb *hcd_to_musb(struct usb_hcd *hcd)  in hcd_to_musb()
75 return *(struct musb **) hcd->hcd_priv; in hcd_to_musb()
79 static void musb_ep_program(struct musb *musb, u8 epnum,
88 struct musb *musb = ep->musb; in musb_h_tx_flush_fifo() local
113 if (dev_WARN_ONCE(musb->controller, retries-- < 1, in musb_h_tx_flush_fifo()
171 if (is_cppi_enabled(ep->musb)) in musb_h_tx_dma_start()
196 musb_start_urb(struct musb *musb, int is_in, struct musb_qh *qh) in musb_start_urb() argument
199 void __iomem *mbase = musb->mregs; in musb_start_urb()
215 musb->ep0_stage = MUSB_EP0_START; in musb_start_urb()
231 trace_musb_urb_start(musb, urb); in musb_start_urb()
235 musb_ep_program(musb, epnum, urb, !is_in, buf, offset, len); in musb_start_urb()
245 musb_dbg(musb, "check whether there's still time for periodic Tx"); in musb_start_urb()
258 musb_dbg(musb, "SOF for %d", epnum); in musb_start_urb()
266 musb_dbg(musb, "Start TX%d %s", epnum, in musb_start_urb()
271 else if (is_cppi_enabled(musb) || tusb_dma_omap(musb)) in musb_start_urb()
277 static void musb_giveback(struct musb *musb, struct urb *urb, int status) in musb_giveback() argument
278 __releases(musb->lock) in musb_giveback()
279 __acquires(musb->lock) in musb_giveback()
281 trace_musb_urb_gb(musb, urb); in musb_giveback()
283 usb_hcd_unlink_urb_from_ep(musb->hcd, urb); in musb_giveback()
284 spin_unlock(&musb->lock); in musb_giveback()
285 usb_hcd_giveback_urb(musb->hcd, urb, status); in musb_giveback()
286 spin_lock(&musb->lock); in musb_giveback()
316 static void musb_advance_schedule(struct musb *musb, struct urb *urb, in musb_advance_schedule() argument
339 musb_giveback(musb, urb, status); in musb_advance_schedule()
347 struct dma_controller *dma = musb->dma_controller; in musb_advance_schedule()
396 musb_dbg(musb, "... next ep%d %cX urb %p", in musb_advance_schedule()
398 musb_start_urb(musb, is_in, qh); in musb_advance_schedule()
425 musb_host_packet_rx(struct musb *musb, struct urb *urb, u8 epnum, u8 iso_err) in musb_host_packet_rx() argument
433 struct musb_hw_ep *hw_ep = musb->endpoints + epnum; in musb_host_packet_rx()
441 musb_dbg(musb, "RX%d count %d, buffer %p len %d/%d", epnum, rx_count, in musb_host_packet_rx()
463 musb_dbg(musb, "OVERFLOW %d into %d", rx_count, length); in musb_host_packet_rx()
481 musb_dbg(musb, "OVERFLOW %d into %d", rx_count, length); in musb_host_packet_rx()
526 musb_rx_reinit(struct musb *musb, struct musb_qh *qh, u8 epnum) in musb_rx_reinit() argument
528 struct musb_hw_ep *ep = musb->endpoints + epnum; in musb_rx_reinit()
564 if (musb->is_multipoint) { in musb_rx_reinit()
565 musb_write_rxfunaddr(musb, epnum, qh->addr_reg); in musb_rx_reinit()
566 musb_write_rxhubaddr(musb, epnum, qh->h_addr_reg); in musb_rx_reinit()
567 musb_write_rxhubport(musb, epnum, qh->h_port_reg); in musb_rx_reinit()
569 musb_writeb(musb->mregs, MUSB_FADDR, qh->addr_reg); in musb_rx_reinit()
612 can_bulk_split(hw_ep->musb, qh->type))) in musb_tx_dma_set_mode_mentor()
650 if (musb_dma_inventra(hw_ep->musb) || musb_dma_ux500(hw_ep->musb)) in musb_tx_dma_program()
653 else if (is_cppi_enabled(hw_ep->musb) || tusb_dma_omap(hw_ep->musb)) in musb_tx_dma_program()
687 static void musb_ep_program(struct musb *musb, u8 epnum, in musb_ep_program() argument
694 void __iomem *mbase = musb->mregs; in musb_ep_program()
695 struct musb_hw_ep *hw_ep = musb->endpoints + epnum; in musb_ep_program()
702 musb_dbg(musb, "%s hw%d urb %p spd%d dev%d ep%d%s " in musb_ep_program()
721 dma_controller = musb->dma_controller; in musb_ep_program()
746 int_txe = musb->intrtxe; in musb_ep_program()
794 if (musb->is_multipoint) { in musb_ep_program()
795 musb_write_txfunaddr(musb, epnum, qh->addr_reg); in musb_ep_program()
796 musb_write_txhubaddr(musb, epnum, qh->h_addr_reg); in musb_ep_program()
797 musb_write_txhubport(musb, epnum, qh->h_port_reg); in musb_ep_program()
805 if (can_bulk_split(musb, qh->type)) { in musb_ep_program()
818 if (musb->is_multipoint) in musb_ep_program()
823 if (can_bulk_split(musb, qh->type)) in musb_ep_program()
841 dev_err(musb->controller, in musb_ep_program()
866 musb_rx_reinit(musb, qh, epnum); in musb_ep_program()
892 if ((is_cppi_enabled(musb) || tusb_dma_omap(musb)) && dma_channel) { in musb_ep_program()
918 musb_dbg(musb, "RXCSR%d := %04x", epnum, csr); in musb_ep_program()
927 static void musb_bulk_nak_timeout(struct musb *musb, struct musb_hw_ep *ep, in musb_bulk_nak_timeout() argument
932 void __iomem *mbase = musb->mregs; in musb_bulk_nak_timeout()
953 cur_qh = first_qh(&musb->in_bulk); in musb_bulk_nak_timeout()
963 cur_qh = first_qh(&musb->out_bulk); in musb_bulk_nak_timeout()
969 musb->dma_controller->channel_abort(dma); in musb_bulk_nak_timeout()
977 list_move_tail(&cur_qh->ring, &musb->in_bulk); in musb_bulk_nak_timeout()
980 next_qh = first_qh(&musb->in_bulk); in musb_bulk_nak_timeout()
986 list_move_tail(&cur_qh->ring, &musb->out_bulk); in musb_bulk_nak_timeout()
989 next_qh = first_qh(&musb->out_bulk); in musb_bulk_nak_timeout()
996 musb_start_urb(musb, is_in, next_qh); in musb_bulk_nak_timeout()
1004 static bool musb_h_ep0_continue(struct musb *musb, u16 len, struct urb *urb) in musb_h_ep0_continue() argument
1009 struct musb_hw_ep *hw_ep = musb->control_ep; in musb_h_ep0_continue()
1013 switch (musb->ep0_stage) { in musb_h_ep0_continue()
1036 musb_dbg(musb, "start no-DATA"); in musb_h_ep0_continue()
1039 musb_dbg(musb, "start IN-DATA"); in musb_h_ep0_continue()
1040 musb->ep0_stage = MUSB_EP0_IN; in musb_h_ep0_continue()
1044 musb_dbg(musb, "start OUT-DATA"); in musb_h_ep0_continue()
1045 musb->ep0_stage = MUSB_EP0_OUT; in musb_h_ep0_continue()
1056 musb_dbg(musb, "Sending %d byte%s to ep0 fifo %p", in musb_h_ep0_continue()
1067 ERR("bogus ep0 stage %d\n", musb->ep0_stage); in musb_h_ep0_continue()
1080 irqreturn_t musb_h_ep0_irq(struct musb *musb) in musb_h_ep0_irq() argument
1085 void __iomem *mbase = musb->mregs; in musb_h_ep0_irq()
1086 struct musb_hw_ep *hw_ep = musb->control_ep; in musb_h_ep0_irq()
1101 musb_dbg(musb, "<== csr0 %04x, qh %p, count %d, urb %p, stage %d", in musb_h_ep0_irq()
1102 csr, qh, len, urb, musb->ep0_stage); in musb_h_ep0_irq()
1105 if (MUSB_EP0_STATUS == musb->ep0_stage) { in musb_h_ep0_irq()
1112 musb_dbg(musb, "STALLING ENDPOINT"); in musb_h_ep0_irq()
1116 musb_dbg(musb, "no response, csr0 %04x", csr); in musb_h_ep0_irq()
1120 musb_dbg(musb, "control NAK timeout"); in musb_h_ep0_irq()
1135 musb_dbg(musb, "aborting"); in musb_h_ep0_irq()
1168 if (musb_h_ep0_continue(musb, len, urb)) { in musb_h_ep0_irq()
1170 csr = (MUSB_EP0_IN == musb->ep0_stage) in musb_h_ep0_irq()
1186 musb->ep0_stage = MUSB_EP0_STATUS; in musb_h_ep0_irq()
1188 musb_dbg(musb, "ep0 STATUS, csr %04x", csr); in musb_h_ep0_irq()
1194 musb->ep0_stage = MUSB_EP0_IDLE; in musb_h_ep0_irq()
1198 musb_advance_schedule(musb, urb, hw_ep, 1); in musb_h_ep0_irq()
1221 void musb_host_tx(struct musb *musb, u8 epnum) in musb_host_tx() argument
1228 struct musb_hw_ep *hw_ep = musb->endpoints + epnum; in musb_host_tx()
1233 void __iomem *mbase = musb->mregs; in musb_host_tx()
1242 musb_dbg(musb, "extra TX%d ready, csr %04x", epnum, tx_csr); in musb_host_tx()
1248 trace_musb_urb_tx(musb, urb); in musb_host_tx()
1249 musb_dbg(musb, "OUT/TX%d end, csr %04x%s", epnum, tx_csr, in musb_host_tx()
1255 musb_dbg(musb, "TX end %d stall", epnum); in musb_host_tx()
1262 musb_dbg(musb, "TX 3strikes on ep=%d", epnum); in musb_host_tx()
1268 && !list_is_singular(&musb->out_bulk)) { in musb_host_tx()
1269 musb_dbg(musb, "NAK timeout on TX%d ep", epnum); in musb_host_tx()
1270 musb_bulk_nak_timeout(musb, hw_ep, 0); in musb_host_tx()
1272 musb_dbg(musb, "TX ep%d device not responding", epnum); in musb_host_tx()
1293 musb->dma_controller->channel_abort(dma); in musb_host_tx()
1318 musb_dbg(musb, "extra TX%d ready, csr %04x", epnum, tx_csr); in musb_host_tx()
1377 musb_dbg(musb, in musb_host_tx()
1435 musb_advance_schedule(musb, urb, hw_ep, USB_DIR_OUT); in musb_host_tx()
1438 if (musb_tx_dma_program(musb->dma_controller, hw_ep, qh, urb, in musb_host_tx()
1440 if (is_cppi_enabled(musb) || tusb_dma_omap(musb)) in musb_host_tx()
1445 musb_dbg(musb, "not complete, but DMA enabled?"); in musb_host_tx()
1459 usb_hcd_unmap_urb_for_dma(musb->hcd, urb); in musb_host_tx()
1471 dev_err(musb->controller, "error: sg list empty\n"); in musb_host_tx()
1600 if (musb_dma_cppi41(hw_ep->musb)) in musb_rx_dma_inventra_cppi41()
1647 struct musb *musb = hw_ep->musb; in musb_rx_dma_in_inventra_cppi41() local
1672 musb_dbg(musb, "** OVERFLOW %d into %d", in musb_rx_dma_in_inventra_cppi41()
1764 void musb_host_rx(struct musb *musb, u8 epnum) in musb_host_rx() argument
1767 struct musb_hw_ep *hw_ep = musb->endpoints + epnum; in musb_host_rx()
1768 struct dma_controller *c = musb->dma_controller; in musb_host_rx()
1772 void __iomem *mbase = musb->mregs; in musb_host_rx()
1795 musb_dbg(musb, "BOGUS RX%d ready, csr %04x, count %d", in musb_host_rx()
1801 trace_musb_urb_rx(musb, urb); in musb_host_rx()
1806 musb_dbg(musb, "RX end %d STALL", epnum); in musb_host_rx()
1812 musb_dbg(musb, "end %d RX proto error", epnum); in musb_host_rx()
1823 musb_dbg(musb, "RX end %d NAK timeout", epnum); in musb_host_rx()
1835 && !list_is_singular(&musb->in_bulk)) { in musb_host_rx()
1836 musb_bulk_nak_timeout(musb, hw_ep, 1); in musb_host_rx()
1846 musb_dbg(musb, "RX end %d ISO data error", epnum); in musb_host_rx()
1851 musb_dbg(musb, "end %d high bandwidth incomplete ISO packet RX", in musb_host_rx()
1861 musb->dma_controller->channel_abort(dma); in musb_host_rx()
1882 if (!musb_dma_inventra(musb) && !musb_dma_ux500(musb) && in musb_host_rx()
1891 musb->dma_controller->channel_abort(dma); in musb_host_rx()
1896 musb_dbg(musb, "RXCSR%d %04x, reqpkt, len %zu%s", epnum, rx_csr, in musb_host_rx()
1914 if (musb_dma_inventra(musb) || musb_dma_ux500(musb) || in musb_host_rx()
1915 musb_dma_cppi41(musb)) { in musb_host_rx()
1917 musb_dbg(hw_ep->musb, in musb_host_rx()
1943 if ((musb_dma_inventra(musb) || musb_dma_ux500(musb) || in musb_host_rx()
1944 musb_dma_cppi41(musb)) && dma) { in musb_host_rx()
1945 musb_dbg(hw_ep->musb, in musb_host_rx()
1957 dev_err(musb->controller, "error: rx_dma failed\n"); in musb_host_rx()
1964 usb_hcd_unmap_urb_for_dma(musb->hcd, urb); in musb_host_rx()
1978 dev_err(musb->controller, "error: sg list empty\n"); in musb_host_rx()
1987 done = musb_host_packet_rx(musb, urb, epnum, in musb_host_rx()
1995 done = musb_host_packet_rx(musb, urb, in musb_host_rx()
1998 musb_dbg(musb, "read %spacket", done ? "last " : ""); in musb_host_rx()
2011 musb_advance_schedule(musb, urb, hw_ep, USB_DIR_IN); in musb_host_rx()
2021 struct musb *musb, in musb_schedule() argument
2036 head = &musb->control; in musb_schedule()
2037 hw_ep = musb->control_ep; in musb_schedule()
2053 for (epnum = 1, hw_ep = musb->endpoints + 1; in musb_schedule()
2054 epnum < musb->nr_endpoints; in musb_schedule()
2061 if (hw_ep == musb->bulk_ep) in musb_schedule()
2084 hw_ep = musb->endpoints + epnum; in musb_schedule()
2098 hw_ep = musb->bulk_ep; in musb_schedule()
2100 head = &musb->in_bulk; in musb_schedule()
2102 head = &musb->out_bulk; in musb_schedule()
2116 dev_err(musb->controller, in musb_schedule()
2125 hw_ep = musb->endpoints + best_end; in musb_schedule()
2126 musb_dbg(musb, "qh %p periodic slot %d", qh, best_end); in musb_schedule()
2136 musb_start_urb(musb, is_in, qh); in musb_schedule()
2146 struct musb *musb = hcd_to_musb(hcd); in musb_urb_enqueue() local
2155 if (!is_host_active(musb) || !musb->is_active) in musb_urb_enqueue()
2158 trace_musb_urb_enq(musb, urb); in musb_urb_enqueue()
2160 spin_lock_irqsave(&musb->lock, flags); in musb_urb_enqueue()
2165 spin_unlock_irqrestore(&musb->lock, flags); in musb_urb_enqueue()
2186 spin_lock_irqsave(&musb->lock, flags); in musb_urb_enqueue()
2188 spin_unlock_irqrestore(&musb->lock, flags); in musb_urb_enqueue()
2209 ok = (usb_pipein(urb->pipe) && musb->hb_iso_rx) in musb_urb_enqueue()
2210 || (usb_pipeout(urb->pipe) && musb->hb_iso_tx); in musb_urb_enqueue()
2212 dev_err(musb->controller, in musb_urb_enqueue()
2277 if (musb->is_multipoint) { in musb_urb_enqueue()
2299 spin_lock_irqsave(&musb->lock, flags); in musb_urb_enqueue()
2308 ret = musb_schedule(musb, qh, in musb_urb_enqueue()
2317 spin_unlock_irqrestore(&musb->lock, flags); in musb_urb_enqueue()
2321 spin_lock_irqsave(&musb->lock, flags); in musb_urb_enqueue()
2323 spin_unlock_irqrestore(&musb->lock, flags); in musb_urb_enqueue()
2338 struct musb *musb = ep->musb; in musb_cleanup_urb() local
2341 void __iomem *regs = ep->musb->mregs; in musb_cleanup_urb()
2352 status = ep->musb->dma_controller->channel_abort(dma); in musb_cleanup_urb()
2353 musb_dbg(musb, "abort %cX%d DMA for urb %p --> %d", in musb_cleanup_urb()
2367 musb_platform_clear_ep_rxintr(musb, ep->epnum); in musb_cleanup_urb()
2386 musb_advance_schedule(ep->musb, urb, ep, is_in); in musb_cleanup_urb()
2392 struct musb *musb = hcd_to_musb(hcd); in musb_urb_dequeue() local
2398 trace_musb_urb_deq(musb, urb); in musb_urb_dequeue()
2400 spin_lock_irqsave(&musb->lock, flags); in musb_urb_dequeue()
2427 musb_giveback(musb, urb, 0); in musb_urb_dequeue()
2441 spin_unlock_irqrestore(&musb->lock, flags); in musb_urb_dequeue()
2451 struct musb *musb = hcd_to_musb(hcd); in musb_h_disable() local
2455 spin_lock_irqsave(&musb->lock, flags); in musb_h_disable()
2481 musb_advance_schedule(musb, urb, qh->hw_ep, is_in); in musb_h_disable()
2489 musb_giveback(musb, next_urb(qh), -ESHUTDOWN); in musb_h_disable()
2496 spin_unlock_irqrestore(&musb->lock, flags); in musb_h_disable()
2501 struct musb *musb = hcd_to_musb(hcd); in musb_h_get_frame_number() local
2503 return musb_readw(musb->mregs, MUSB_FRAME); in musb_h_get_frame_number()
2508 struct musb *musb = hcd_to_musb(hcd); in musb_h_start() local
2514 musb->port1_status = 0; in musb_h_start()
2526 struct musb *musb = hcd_to_musb(hcd); in musb_bus_suspend() local
2530 ret = musb_port_suspend(musb, true); in musb_bus_suspend()
2534 if (!is_host_active(musb)) in musb_bus_suspend()
2537 switch (musb->xceiv->otg->state) { in musb_bus_suspend()
2545 devctl = musb_readb(musb->mregs, MUSB_DEVCTL); in musb_bus_suspend()
2547 musb->xceiv->otg->state = OTG_STATE_A_WAIT_BCON; in musb_bus_suspend()
2553 if (musb->is_active) { in musb_bus_suspend()
2555 usb_otg_state_string(musb->xceiv->otg->state)); in musb_bus_suspend()
2563 struct musb *musb = hcd_to_musb(hcd); in musb_bus_resume() local
2565 if (musb->config && in musb_bus_resume()
2566 musb->config->host_port_deassert_reset_at_resume) in musb_bus_resume()
2567 musb_port_reset(musb, false); in musb_bus_resume()
2651 struct musb *musb = hcd_to_musb(hcd); in musb_map_urb_for_dma() local
2660 if (musb->hwvers < MUSB_HWVERS_1800) in musb_map_urb_for_dma()
2676 struct musb *musb = hcd_to_musb(hcd); in musb_unmap_urb_for_dma() local
2681 if (musb->hwvers < MUSB_HWVERS_1800) in musb_unmap_urb_for_dma()
2691 .hcd_priv_size = sizeof(struct musb *),
2720 int musb_host_alloc(struct musb *musb) in musb_host_alloc() argument
2722 struct device *dev = musb->controller; in musb_host_alloc()
2725 musb->hcd = usb_create_hcd(&musb_hc_driver, dev, dev_name(dev)); in musb_host_alloc()
2726 if (!musb->hcd) in musb_host_alloc()
2729 *musb->hcd->hcd_priv = (unsigned long) musb; in musb_host_alloc()
2730 musb->hcd->self.uses_pio_for_control = 1; in musb_host_alloc()
2731 musb->hcd->uses_new_polling = 1; in musb_host_alloc()
2732 musb->hcd->has_tt = 1; in musb_host_alloc()
2737 void musb_host_cleanup(struct musb *musb) in musb_host_cleanup() argument
2739 if (musb->port_mode == MUSB_PERIPHERAL) in musb_host_cleanup()
2741 usb_remove_hcd(musb->hcd); in musb_host_cleanup()
2744 void musb_host_free(struct musb *musb) in musb_host_free() argument
2746 usb_put_hcd(musb->hcd); in musb_host_free()
2749 int musb_host_setup(struct musb *musb, int power_budget) in musb_host_setup() argument
2752 struct usb_hcd *hcd = musb->hcd; in musb_host_setup()
2754 if (musb->port_mode == MUSB_HOST) { in musb_host_setup()
2755 MUSB_HST_MODE(musb); in musb_host_setup()
2756 musb->xceiv->otg->state = OTG_STATE_A_IDLE; in musb_host_setup()
2758 otg_set_host(musb->xceiv->otg, &hcd->self); in musb_host_setup()
2761 musb->xceiv->otg->host = &hcd->self; in musb_host_setup()
2773 void musb_host_resume_root_hub(struct musb *musb) in musb_host_resume_root_hub() argument
2775 usb_hcd_resume_root_hub(musb->hcd); in musb_host_resume_root_hub()
2778 void musb_host_poke_root_hub(struct musb *musb) in musb_host_poke_root_hub() argument
2780 MUSB_HST_MODE(musb); in musb_host_poke_root_hub()
2781 if (musb->hcd->status_urb) in musb_host_poke_root_hub()
2782 usb_hcd_poll_rh_status(musb->hcd); in musb_host_poke_root_hub()
2784 usb_hcd_resume_root_hub(musb->hcd); in musb_host_poke_root_hub()