Lines Matching refs:hwreq
339 static int add_td_to_list(struct ci_hw_ep *hwep, struct ci_hw_req *hwreq, in add_td_to_list() argument
360 u32 mul = hwreq->req.length / hwep->ep.maxpacket; in add_td_to_list()
362 if (hwreq->req.length == 0 in add_td_to_list()
363 || hwreq->req.length % hwep->ep.maxpacket) in add_td_to_list()
368 temp = (u32) (hwreq->req.dma + hwreq->req.actual); in add_td_to_list()
378 hwreq->req.actual += length; in add_td_to_list()
380 if (!list_empty(&hwreq->tds)) { in add_td_to_list()
382 lastnode = list_entry(hwreq->tds.prev, in add_td_to_list()
388 list_add_tail(&node->td, &hwreq->tds); in add_td_to_list()
409 static int _hardware_enqueue(struct ci_hw_ep *hwep, struct ci_hw_req *hwreq) in _hardware_enqueue() argument
413 unsigned rest = hwreq->req.length; in _hardware_enqueue()
418 if (hwreq->req.status == -EALREADY) in _hardware_enqueue()
421 hwreq->req.status = -EALREADY; in _hardware_enqueue()
424 &hwreq->req, hwep->dir); in _hardware_enqueue()
432 if (hwreq->req.dma % PAGE_SIZE) in _hardware_enqueue()
436 ret = add_td_to_list(hwep, hwreq, 0); in _hardware_enqueue()
442 unsigned count = min(hwreq->req.length - hwreq->req.actual, in _hardware_enqueue()
444 ret = add_td_to_list(hwep, hwreq, count); in _hardware_enqueue()
451 if (hwreq->req.zero && hwreq->req.length && hwep->dir == TX in _hardware_enqueue()
452 && (hwreq->req.length % hwep->ep.maxpacket == 0)) { in _hardware_enqueue()
453 ret = add_td_to_list(hwep, hwreq, 0); in _hardware_enqueue()
458 firstnode = list_first_entry(&hwreq->tds, struct td_node, td); in _hardware_enqueue()
460 lastnode = list_entry(hwreq->tds.prev, in _hardware_enqueue()
464 if (!hwreq->req.no_interrupt) in _hardware_enqueue()
468 hwreq->req.actual = 0; in _hardware_enqueue()
500 u32 mul = hwreq->req.length / hwep->ep.maxpacket; in _hardware_enqueue()
502 if (hwreq->req.length == 0 in _hardware_enqueue()
503 || hwreq->req.length % hwep->ep.maxpacket) in _hardware_enqueue()
545 static int _hardware_dequeue(struct ci_hw_ep *hwep, struct ci_hw_req *hwreq) in _hardware_dequeue() argument
550 unsigned actual = hwreq->req.length; in _hardware_dequeue()
553 if (hwreq->req.status != -EALREADY) in _hardware_dequeue()
556 hwreq->req.status = 0; in _hardware_dequeue()
558 list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) { in _hardware_dequeue()
566 hwreq->req.status = -EALREADY; in _hardware_dequeue()
574 hwreq->req.status = tmptoken & TD_STATUS; in _hardware_dequeue()
575 if ((TD_STATUS_HALTED & hwreq->req.status)) { in _hardware_dequeue()
576 hwreq->req.status = -EPIPE; in _hardware_dequeue()
578 } else if ((TD_STATUS_DT_ERR & hwreq->req.status)) { in _hardware_dequeue()
579 hwreq->req.status = -EPROTO; in _hardware_dequeue()
581 } else if ((TD_STATUS_TR_ERR & hwreq->req.status)) { in _hardware_dequeue()
582 hwreq->req.status = -EILSEQ; in _hardware_dequeue()
588 hwreq->req.status = -EPROTO; in _hardware_dequeue()
605 &hwreq->req, hwep->dir); in _hardware_dequeue()
607 hwreq->req.actual += actual; in _hardware_dequeue()
609 if (hwreq->req.status) in _hardware_dequeue()
610 return hwreq->req.status; in _hardware_dequeue()
612 return hwreq->req.actual; in _hardware_dequeue()
635 struct ci_hw_req *hwreq = list_entry(hwep->qh.queue.next, in _ep_nuke() local
638 list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) { in _ep_nuke()
645 list_del_init(&hwreq->queue); in _ep_nuke()
646 hwreq->req.status = -ESHUTDOWN; in _ep_nuke()
648 if (hwreq->req.complete != NULL) { in _ep_nuke()
650 usb_gadget_giveback_request(&hwep->ep, &hwreq->req); in _ep_nuke()
804 struct ci_hw_req *hwreq = container_of(req, struct ci_hw_req, req); in _ep_queue() local
823 hwreq->req.length > hwep->ep.mult * hwep->ep.maxpacket) { in _ep_queue()
829 if (!list_empty(&hwreq->queue)) { in _ep_queue()
835 hwreq->req.status = -EINPROGRESS; in _ep_queue()
836 hwreq->req.actual = 0; in _ep_queue()
838 retval = _hardware_enqueue(hwep, hwreq); in _ep_queue()
843 list_add_tail(&hwreq->queue, &hwep->qh.queue); in _ep_queue()
973 struct ci_hw_req *hwreq, *hwreqtemp; in isr_tr_complete_low() local
977 list_for_each_entry_safe(hwreq, hwreqtemp, &hwep->qh.queue, in isr_tr_complete_low()
979 retval = _hardware_dequeue(hwep, hwreq); in isr_tr_complete_low()
982 list_del_init(&hwreq->queue); in isr_tr_complete_low()
983 if (hwreq->req.complete != NULL) { in isr_tr_complete_low()
986 hwreq->req.length) in isr_tr_complete_low()
988 usb_gadget_giveback_request(&hweptemp->ep, &hwreq->req); in isr_tr_complete_low()
1331 struct ci_hw_req *hwreq = NULL; in ep_alloc_request() local
1336 hwreq = kzalloc(sizeof(struct ci_hw_req), gfp_flags); in ep_alloc_request()
1337 if (hwreq != NULL) { in ep_alloc_request()
1338 INIT_LIST_HEAD(&hwreq->queue); in ep_alloc_request()
1339 INIT_LIST_HEAD(&hwreq->tds); in ep_alloc_request()
1342 return (hwreq == NULL) ? NULL : &hwreq->req; in ep_alloc_request()
1353 struct ci_hw_req *hwreq = container_of(req, struct ci_hw_req, req); in ep_free_request() local
1359 } else if (!list_empty(&hwreq->queue)) { in ep_free_request()
1366 list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) { in ep_free_request()
1373 kfree(hwreq); in ep_free_request()
1407 struct ci_hw_req *hwreq = container_of(req, struct ci_hw_req, req); in ep_dequeue() local
1411 if (ep == NULL || req == NULL || hwreq->req.status != -EALREADY || in ep_dequeue()
1412 hwep->ep.desc == NULL || list_empty(&hwreq->queue) || in ep_dequeue()
1420 list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) { in ep_dequeue()
1427 list_del_init(&hwreq->queue); in ep_dequeue()
1433 if (hwreq->req.complete != NULL) { in ep_dequeue()
1435 usb_gadget_giveback_request(&hwep->ep, &hwreq->req); in ep_dequeue()