Lines Matching refs:ep_ctx
606 struct xhci_ep_ctx *ep_ctx; in xhci_get_hw_deq() local
616 ep_ctx = xhci_get_ep_ctx(xhci, vdev->out_ctx, ep_index); in xhci_get_hw_deq()
617 return le64_to_cpu(ep_ctx->deq); in xhci_get_hw_deq()
1095 struct xhci_ep_ctx *ep_ctx; in xhci_handle_cmd_stop_ep() local
1113 ep_ctx = xhci_get_ep_ctx(xhci, ep->vdev->out_ctx, ep_index); in xhci_handle_cmd_stop_ep()
1115 trace_xhci_handle_cmd_stop_ep(ep_ctx); in xhci_handle_cmd_stop_ep()
1132 switch (GET_EP_CTX_STATE(ep_ctx)) { in xhci_handle_cmd_stop_ep()
1327 struct xhci_ep_ctx *ep_ctx; in xhci_handle_cmd_set_deq() local
1345 ep_ctx = xhci_get_ep_ctx(xhci, ep->vdev->out_ctx, ep_index); in xhci_handle_cmd_set_deq()
1348 trace_xhci_handle_cmd_set_deq_ep(ep_ctx); in xhci_handle_cmd_set_deq()
1360 ep_state = GET_EP_CTX_STATE(ep_ctx); in xhci_handle_cmd_set_deq()
1390 deq = le64_to_cpu(ep_ctx->deq) & ~EP_CTX_CYCLE_MASK; in xhci_handle_cmd_set_deq()
1433 struct xhci_ep_ctx *ep_ctx; in xhci_handle_cmd_reset_ep() local
1441 ep_ctx = xhci_get_ep_ctx(xhci, ep->vdev->out_ctx, ep_index); in xhci_handle_cmd_reset_ep()
1442 trace_xhci_handle_cmd_reset_ep(ep_ctx); in xhci_handle_cmd_reset_ep()
1494 struct xhci_ep_ctx *ep_ctx; in xhci_handle_cmd_config_ep() local
1517 ep_ctx = xhci_get_ep_ctx(xhci, virt_dev->out_ctx, ep_index); in xhci_handle_cmd_config_ep()
1518 trace_xhci_handle_cmd_config_ep(ep_ctx); in xhci_handle_cmd_config_ep()
2127 struct xhci_ep_ctx *ep_ctx, in xhci_requires_manual_halt_cleanup() argument
2140 if (GET_EP_CTX_STATE(ep_ctx) == EP_STATE_HALTED) in xhci_requires_manual_halt_cleanup()
2164 struct xhci_ep_ctx *ep_ctx; in finish_td() local
2166 ep_ctx = xhci_get_ep_ctx(xhci, ep->vdev->out_ctx, ep->ep_index); in finish_td()
2192 if (GET_EP_CTX_STATE(ep_ctx) != EP_STATE_HALTED) { in finish_td()
2263 struct xhci_ep_ctx *ep_ctx; in process_ctrl_td() local
2269 ep_ctx = xhci_get_ep_ctx(xhci, ep->vdev->out_ctx, ep->ep_index); in process_ctrl_td()
2314 ep_ctx, trb_comp_code)) in process_ctrl_td()
2554 struct xhci_ep_ctx *ep_ctx; in handle_tx_event() local
2571 ep_ctx = xhci_get_ep_ctx(xhci, ep->vdev->out_ctx, ep_index); in handle_tx_event()
2573 if (GET_EP_CTX_STATE(ep_ctx) == EP_STATE_DISABLED) { in handle_tx_event()
2773 xhci_requires_manual_halt_cleanup(xhci, ep_ctx, in handle_tx_event()
2866 xhci_requires_manual_halt_cleanup(xhci, ep_ctx, in handle_tx_event()
3267 struct xhci_ep_ctx *ep_ctx = xhci_get_ep_ctx(xhci, xdev->out_ctx, ep_index); in prepare_transfer() local
3277 ret = prepare_ring(xhci, ep_ring, GET_EP_CTX_STATE(ep_ctx), in prepare_transfer()
3378 struct xhci_ep_ctx *ep_ctx) in check_interval() argument
3383 xhci_interval = EP_INTERVAL_TO_UFRAMES(le32_to_cpu(ep_ctx->ep_info)); in check_interval()
3416 struct xhci_ep_ctx *ep_ctx; in xhci_queue_intr_tx() local
3418 ep_ctx = xhci_get_ep_ctx(xhci, xhci->devs[slot_id]->out_ctx, ep_index); in xhci_queue_intr_tx()
3419 check_interval(xhci, urb, ep_ctx); in xhci_queue_intr_tx()
4196 struct xhci_ep_ctx *ep_ctx; in xhci_queue_isoc_tx_prepare() local
4206 ep_ctx = xhci_get_ep_ctx(xhci, xdev->out_ctx, ep_index); in xhci_queue_isoc_tx_prepare()
4216 ret = prepare_ring(xhci, ep_ring, GET_EP_CTX_STATE(ep_ctx), in xhci_queue_isoc_tx_prepare()
4225 check_interval(xhci, urb, ep_ctx); in xhci_queue_isoc_tx_prepare()
4229 if (GET_EP_CTX_STATE(ep_ctx) == EP_STATE_RUNNING) { in xhci_queue_isoc_tx_prepare()