Searched refs:ack_queue (Results 1 – 24 of 24) sorted by relevance
39 skb_queue_purge(&nr->ack_queue); in nr_clear_queues()58 while (skb_peek(&nrom->ack_queue) != NULL && nrom->va != nr) { in nr_frames_acked()59 skb = skb_dequeue(&nrom->ack_queue); in nr_frames_acked()75 while ((skb = skb_dequeue(&nr_sk(sk)->ack_queue)) != NULL) { in nr_requeue_frames()
105 if ((skb = skb_peek(&nr->ack_queue)) == NULL) in nr_send_nak_frame()140 start = (skb_peek(&nr->ack_queue) == NULL) ? nr->va : nr->vs; in nr_kick()176 skb_queue_tail(&nr->ack_queue, skb); in nr_kick()
447 skb_queue_head_init(&nr->ack_queue); in nr_create()495 skb_queue_head_init(&nr->ack_queue); in nr_make_new()
39 skb_queue_purge(&ax25->ack_queue); in ax25_clear_queues()57 while (skb_peek(&ax25->ack_queue) != NULL && ax25->va != nr) { in ax25_frames_acked()58 skb = skb_dequeue(&ax25->ack_queue); in ax25_frames_acked()74 while ((skb = skb_dequeue_tail(&ax25->ack_queue)) != NULL) in ax25_requeue_frames()
73 …if (ax25->state == AX25_STATE_1 || ax25->state == AX25_STATE_2 || skb_peek(&ax25->ack_queue) != NU… in ax25_ds_enquiry_response()99 …if (ax25o->state == AX25_STATE_1 || ax25o->state == AX25_STATE_2 || skb_peek(&ax25o->ack_queue) !=… in ax25_ds_enquiry_response()
259 start = (skb_peek(&ax25->ack_queue) == NULL) ? ax25->va : ax25->vs; in ax25_kick()316 skb_queue_tail(&ax25->ack_queue, skb); in ax25_kick()
516 skb_queue_head_init(&ax25->ack_queue); in ax25_create_cb()
43 skb_queue_purge(&lapb->ack_queue); in lapb_clear_queues()62 while (skb_peek(&lapb->ack_queue) && lapb->va != nr) { in lapb_frames_acked()63 skb = skb_dequeue(&lapb->ack_queue); in lapb_frames_acked()78 while ((skb = skb_dequeue(&lapb->ack_queue)) != NULL) { in lapb_requeue_frames()
77 start = !skb_peek(&lapb->ack_queue) ? lapb->va : lapb->vs; in lapb_kick()108 skb_queue_tail(&lapb->ack_queue, skb); in lapb_kick()
128 skb_queue_head_init(&lapb->ack_queue); in lapb_create_cb()
63 start = (skb_peek(&rose->ack_queue) == NULL) ? rose->va : rose->vs; in rose_kick()96 skb_queue_tail(&rose->ack_queue, skb); in rose_kick()
38 skb_queue_purge(&rose_sk(sk)->ack_queue); in rose_clear_queues()55 while (skb_peek(&rose->ack_queue) != NULL && rose->va != nr) { in rose_frames_acked()56 skb = skb_dequeue(&rose->ack_queue); in rose_frames_acked()72 while ((skb = skb_dequeue(&rose_sk(sk)->ack_queue)) != NULL) { in rose_requeue_frames()
533 skb_queue_head_init(&rose->ack_queue); in rose_create()572 skb_queue_head_init(&rose->ack_queue); in rose_make_new()
44 skb_queue_purge(&x25->ack_queue); in x25_clear_queues()66 while (skb_peek(&x25->ack_queue) && x25->va != nr) { in x25_frames_acked()67 skb = skb_dequeue(&x25->ack_queue); in x25_frames_acked()82 while ((skb = skb_dequeue(&x25_sk(sk)->ack_queue)) != NULL) { in x25_requeue_frames()
169 start = skb_peek(&x25->ack_queue) ? x25->vs : x25->va; in x25_kick()202 skb_queue_tail(&x25->ack_queue, skb); in x25_kick()
530 skb_queue_head_init(&x25->ack_queue); in x25_alloc_socket()
98 struct sk_buff_head ack_queue; member
144 struct sk_buff_head ack_queue; member
81 struct sk_buff_head ack_queue; member
159 struct sk_buff_head ack_queue; member
244 struct sk_buff_head ack_queue; member
234 wake_up(&vgdev->ctrlq.ack_queue); in virtio_gpu_dequeue_ctrl_func()260 wake_up(&vgdev->cursorq.ack_queue); in virtio_gpu_dequeue_cursor_func()296 wait_event(vgdev->ctrlq.ack_queue, vq->num_free >= outcnt + incnt); in virtio_gpu_queue_ctrl_buffer_locked()340 wait_event(vgdev->ctrlq.ack_queue, vq->num_free >= 3); in virtio_gpu_queue_fenced_ctrl_buffer()371 wait_event(vgdev->cursorq.ack_queue, vq->num_free >= outcnt); in virtio_gpu_queue_cursor()
94 init_waitqueue_head(&vgvq->ack_queue); in virtio_gpu_init_vq()
146 wait_queue_head_t ack_queue; member