Lines Matching refs:unack
59 struct sk_buff_head unack; /* Unack'ed packets queue */ member
316 spin_lock_irqsave_nested(&bcsp->unack.lock, flags, SINGLE_DEPTH_NESTING); in bcsp_dequeue()
318 if (bcsp->unack.qlen < BCSP_TXWINSIZE) { in bcsp_dequeue()
326 __skb_queue_tail(&bcsp->unack, skb); in bcsp_dequeue()
328 spin_unlock_irqrestore(&bcsp->unack.lock, flags); in bcsp_dequeue()
337 spin_unlock_irqrestore(&bcsp->unack.lock, flags); in bcsp_dequeue()
370 spin_lock_irqsave(&bcsp->unack.lock, flags); in bcsp_pkt_cull()
372 pkts_to_be_removed = skb_queue_len(&bcsp->unack); in bcsp_pkt_cull()
386 pkts_to_be_removed, skb_queue_len(&bcsp->unack), in bcsp_pkt_cull()
390 skb_queue_walk_safe(&bcsp->unack, skb, tmp) { in bcsp_pkt_cull()
395 __skb_unlink(skb, &bcsp->unack); in bcsp_pkt_cull()
399 if (skb_queue_empty(&bcsp->unack)) in bcsp_pkt_cull()
402 spin_unlock_irqrestore(&bcsp->unack.lock, flags); in bcsp_pkt_cull()
708 BT_DBG("hu %p retransmitting %u pkts", hu, bcsp->unack.qlen); in bcsp_timed_event()
710 spin_lock_irqsave_nested(&bcsp->unack.lock, flags, SINGLE_DEPTH_NESTING); in bcsp_timed_event()
712 while ((skb = __skb_dequeue_tail(&bcsp->unack)) != NULL) { in bcsp_timed_event()
717 spin_unlock_irqrestore(&bcsp->unack.lock, flags); in bcsp_timed_event()
734 skb_queue_head_init(&bcsp->unack); in bcsp_open()
758 skb_queue_purge(&bcsp->unack); in bcsp_close()