Home
last modified time | relevance | path

Searched refs:enq_seg (Results 1 – 12 of 12) sorted by relevance

/Linux-v6.1/drivers/usb/host/
Dxhci-dbgcap.h167 xhci_trb_virt_to_dma((d)->ring_out->enq_seg, (d)->ring_out->enqueue)
169 xhci_trb_virt_to_dma((d)->ring_in->enq_seg, (d)->ring_in->enqueue)
Dxhci-trace.h464 __field(dma_addr_t, enq_seg)
477 __entry->enq_seg = ring->enq_seg->dma;
482 __entry->enq = xhci_trb_virt_to_dma(ring->enq_seg, ring->enqueue);
487 &__entry->enq, &__entry->enq_seg,
Dxhci-ring.c228 if (last_trb_on_seg(ring->enq_seg, ring->enqueue)) { in inc_enq()
266 ring->enq_seg = ring->enq_seg->next; in inc_enq()
267 ring->enqueue = ring->enq_seg->trbs; in inc_enq()
603 new_seg = ep_ring->enq_seg; in xhci_move_dequeue_past_td()
3207 ep_ring->enq_seg = ep_ring->enq_seg->next; in prepare_ring()
3208 ep_ring->enqueue = ep_ring->enq_seg->trbs; in prepare_ring()
3217 if (last_trb_on_seg(ep_ring->enq_seg, ep_ring->enqueue)) { in prepare_ring()
3268 td->start_seg = ep_ring->enq_seg; in prepare_transfer()
3598 ring->enq_seg)) { in xhci_queue_bulk_tx()
3599 send_addr = ring->enq_seg->bounce_dma; in xhci_queue_bulk_tx()
[all …]
Dxhci-mem.c141 next = ring->enq_seg->next; in xhci_link_rings()
142 xhci_link_segments(ring->enq_seg, first, ring->type, chain_links); in xhci_link_rings()
147 if (ring->type != TYPE_EVENT && ring->enq_seg == ring->last_seg) { in xhci_link_rings()
301 ring->enq_seg = ring->first_seg; in xhci_initialize_ring_info()
1054 ep0_ctx->deq = cpu_to_le64(xhci_trb_virt_to_dma(ep_ring->enq_seg, in xhci_copy_ep0_dequeue_into_input_ctx()
Dxhci-dbgcap.c240 ring->enqueue = ring->enq_seg->trbs; in xhci_dbc_queue_trb()
272 req->trb_dma = xhci_trb_virt_to_dma(ring->enq_seg, ring->enqueue); in xhci_dbc_queue_bulk_tx()
Dxhci-debugfs.c169 dma = xhci_trb_virt_to_dma(ring->enq_seg, ring->enqueue); in xhci_ring_enqueue_show()
Dxhci.h1620 struct xhci_segment *enq_seg; member
Dxhci.c903 ring->enq_seg = ring->deq_seg; in xhci_clear_command_ring()
/Linux-v6.1/drivers/usb/cdns3/
Dcdnsp-ring.c224 ring->enq_seg = ring->enq_seg->next; in cdnsp_inc_enq()
225 ring->enqueue = ring->enq_seg->trbs; in cdnsp_inc_enq()
1655 ep_ring->enq_seg = ep_ring->enq_seg->next; in cdnsp_prepare_ring()
1656 ep_ring->enqueue = ep_ring->enq_seg->trbs; in cdnsp_prepare_ring()
1687 preq->td.start_seg = ep_ring->enq_seg; in cdnsp_prepare_transfer()
1940 ring->enq_seg)) { in cdnsp_queue_bulk_tx()
1941 send_addr = ring->enq_seg->bounce_dma; in cdnsp_queue_bulk_tx()
1943 preq->td.bounce_seg = ring->enq_seg; in cdnsp_queue_bulk_tx()
2300 ep_ring->enq_seg = preq->td.start_seg; in cdnsp_queue_isoc_tx()
Dcdnsp-trace.h708 __field(dma_addr_t, enq_seg)
721 __entry->enq_seg = ring->enq_seg->dma;
726 __entry->enq = cdnsp_trb_virt_to_dma(ring->enq_seg,
734 &__entry->enq, &__entry->enq_seg,
Dcdnsp-mem.c147 next = ring->enq_seg->next; in cdnsp_link_rings()
148 cdnsp_link_segments(pdev, ring->enq_seg, first, ring->type); in cdnsp_link_rings()
153 if (ring->type != TYPE_EVENT && ring->enq_seg == ring->last_seg) { in cdnsp_link_rings()
300 ring->enq_seg = ring->first_seg; in cdnsp_initialize_ring_info()
721 dma = cdnsp_trb_virt_to_dma(ep_ring->enq_seg, ep_ring->enqueue); in cdnsp_copy_ep0_dequeue_into_input_ctx()
Dcdnsp-gadget.h1253 struct cdnsp_segment *enq_seg; member