Lines Matching full:queue

47 /* Number of bytes allowed on the internal guest Rx queue. */
55 void xenvif_skb_zerocopy_prepare(struct xenvif_queue *queue, in xenvif_skb_zerocopy_prepare() argument
59 atomic_inc(&queue->inflight_packets); in xenvif_skb_zerocopy_prepare()
62 void xenvif_skb_zerocopy_complete(struct xenvif_queue *queue) in xenvif_skb_zerocopy_complete() argument
64 atomic_dec(&queue->inflight_packets); in xenvif_skb_zerocopy_complete()
70 wake_up(&queue->dealloc_wq); in xenvif_skb_zerocopy_complete()
80 static bool xenvif_handle_tx_interrupt(struct xenvif_queue *queue) in xenvif_handle_tx_interrupt() argument
84 rc = RING_HAS_UNCONSUMED_REQUESTS(&queue->tx); in xenvif_handle_tx_interrupt()
86 napi_schedule(&queue->napi); in xenvif_handle_tx_interrupt()
92 struct xenvif_queue *queue = dev_id; in xenvif_tx_interrupt() local
95 old = atomic_fetch_or(NETBK_TX_EOI, &queue->eoi_pending); in xenvif_tx_interrupt()
98 if (!xenvif_handle_tx_interrupt(queue)) { in xenvif_tx_interrupt()
99 atomic_andnot(NETBK_TX_EOI, &queue->eoi_pending); in xenvif_tx_interrupt()
108 struct xenvif_queue *queue = in xenvif_poll() local
116 if (unlikely(queue->vif->disabled)) { in xenvif_poll()
121 work_done = xenvif_tx_action(queue, budget); in xenvif_poll()
125 /* If the queue is rate-limited, it shall be in xenvif_poll()
128 if (likely(!queue->rate_limited)) in xenvif_poll()
129 xenvif_napi_schedule_or_enable_events(queue); in xenvif_poll()
135 static bool xenvif_handle_rx_interrupt(struct xenvif_queue *queue) in xenvif_handle_rx_interrupt() argument
139 rc = xenvif_have_rx_work(queue, false); in xenvif_handle_rx_interrupt()
141 xenvif_kick_thread(queue); in xenvif_handle_rx_interrupt()
147 struct xenvif_queue *queue = dev_id; in xenvif_rx_interrupt() local
150 old = atomic_fetch_or(NETBK_RX_EOI, &queue->eoi_pending); in xenvif_rx_interrupt()
153 if (!xenvif_handle_rx_interrupt(queue)) { in xenvif_rx_interrupt()
154 atomic_andnot(NETBK_RX_EOI, &queue->eoi_pending); in xenvif_rx_interrupt()
163 struct xenvif_queue *queue = dev_id; in xenvif_interrupt() local
167 old = atomic_fetch_or(NETBK_COMMON_EOI, &queue->eoi_pending); in xenvif_interrupt()
170 has_tx = xenvif_handle_tx_interrupt(queue); in xenvif_interrupt()
171 has_rx = xenvif_handle_rx_interrupt(queue); in xenvif_interrupt()
174 atomic_andnot(NETBK_COMMON_EOI, &queue->eoi_pending); in xenvif_interrupt()
181 int xenvif_queue_stopped(struct xenvif_queue *queue) in xenvif_queue_stopped() argument
183 struct net_device *dev = queue->vif->dev; in xenvif_queue_stopped()
184 unsigned int id = queue->id; in xenvif_queue_stopped()
188 void xenvif_wake_queue(struct xenvif_queue *queue) in xenvif_wake_queue() argument
190 struct net_device *dev = queue->vif->dev; in xenvif_wake_queue()
191 unsigned int id = queue->id; in xenvif_wake_queue()
225 struct xenvif_queue *queue = NULL; in xenvif_start_xmit() local
240 /* Obtain the queue to be used to transmit this packet */ in xenvif_start_xmit()
243 pr_warn_ratelimited("Invalid queue %hu for packet on interface %s\n", in xenvif_start_xmit()
247 queue = &vif->queues[index]; in xenvif_start_xmit()
249 /* Drop the packet if queue is not ready */ in xenvif_start_xmit()
250 if (queue->task == NULL || in xenvif_start_xmit()
251 queue->dealloc_task == NULL || in xenvif_start_xmit()
272 xenvif_rx_queue_tail(queue, skb); in xenvif_start_xmit()
273 xenvif_kick_thread(queue); in xenvif_start_xmit()
286 struct xenvif_queue *queue = NULL; in xenvif_get_stats() local
297 /* Aggregate tx and rx stats from each queue */ in xenvif_get_stats()
299 queue = &vif->queues[index]; in xenvif_get_stats()
300 rx_bytes += queue->stats.rx_bytes; in xenvif_get_stats()
301 rx_packets += queue->stats.rx_packets; in xenvif_get_stats()
302 tx_bytes += queue->stats.tx_bytes; in xenvif_get_stats()
303 tx_packets += queue->stats.tx_packets; in xenvif_get_stats()
318 struct xenvif_queue *queue = NULL; in xenvif_up() local
323 queue = &vif->queues[queue_index]; in xenvif_up()
324 napi_enable(&queue->napi); in xenvif_up()
325 enable_irq(queue->tx_irq); in xenvif_up()
326 if (queue->tx_irq != queue->rx_irq) in xenvif_up()
327 enable_irq(queue->rx_irq); in xenvif_up()
328 xenvif_napi_schedule_or_enable_events(queue); in xenvif_up()
334 struct xenvif_queue *queue = NULL; in xenvif_down() local
339 queue = &vif->queues[queue_index]; in xenvif_down()
340 disable_irq(queue->tx_irq); in xenvif_down()
341 if (queue->tx_irq != queue->rx_irq) in xenvif_down()
342 disable_irq(queue->rx_irq); in xenvif_down()
343 napi_disable(&queue->napi); in xenvif_down()
344 del_timer_sync(&queue->credit_timeout); in xenvif_down()
573 int xenvif_init_queue(struct xenvif_queue *queue) in xenvif_init_queue() argument
577 queue->credit_bytes = queue->remaining_credit = ~0UL; in xenvif_init_queue()
578 queue->credit_usec = 0UL; in xenvif_init_queue()
579 timer_setup(&queue->credit_timeout, xenvif_tx_credit_callback, 0); in xenvif_init_queue()
580 queue->credit_window_start = get_jiffies_64(); in xenvif_init_queue()
582 queue->rx_queue_max = XENVIF_RX_QUEUE_BYTES; in xenvif_init_queue()
584 skb_queue_head_init(&queue->rx_queue); in xenvif_init_queue()
585 skb_queue_head_init(&queue->tx_queue); in xenvif_init_queue()
587 queue->pending_cons = 0; in xenvif_init_queue()
588 queue->pending_prod = MAX_PENDING_REQS; in xenvif_init_queue()
590 queue->pending_ring[i] = i; in xenvif_init_queue()
592 spin_lock_init(&queue->callback_lock); in xenvif_init_queue()
593 spin_lock_init(&queue->response_lock); in xenvif_init_queue()
600 queue->mmap_pages); in xenvif_init_queue()
602 netdev_err(queue->vif->dev, "Could not reserve mmap_pages\n"); in xenvif_init_queue()
607 queue->pending_tx_info[i].callback_struct = (struct ubuf_info) in xenvif_init_queue()
611 queue->grant_tx_handle[i] = NETBACK_INVALID_HANDLE; in xenvif_init_queue()
683 static void xenvif_disconnect_queue(struct xenvif_queue *queue) in xenvif_disconnect_queue() argument
685 if (queue->task) { in xenvif_disconnect_queue()
686 kthread_stop(queue->task); in xenvif_disconnect_queue()
687 put_task_struct(queue->task); in xenvif_disconnect_queue()
688 queue->task = NULL; in xenvif_disconnect_queue()
691 if (queue->dealloc_task) { in xenvif_disconnect_queue()
692 kthread_stop(queue->dealloc_task); in xenvif_disconnect_queue()
693 queue->dealloc_task = NULL; in xenvif_disconnect_queue()
696 if (queue->napi.poll) { in xenvif_disconnect_queue()
697 netif_napi_del(&queue->napi); in xenvif_disconnect_queue()
698 queue->napi.poll = NULL; in xenvif_disconnect_queue()
701 if (queue->tx_irq) { in xenvif_disconnect_queue()
702 unbind_from_irqhandler(queue->tx_irq, queue); in xenvif_disconnect_queue()
703 if (queue->tx_irq == queue->rx_irq) in xenvif_disconnect_queue()
704 queue->rx_irq = 0; in xenvif_disconnect_queue()
705 queue->tx_irq = 0; in xenvif_disconnect_queue()
708 if (queue->rx_irq) { in xenvif_disconnect_queue()
709 unbind_from_irqhandler(queue->rx_irq, queue); in xenvif_disconnect_queue()
710 queue->rx_irq = 0; in xenvif_disconnect_queue()
713 xenvif_unmap_frontend_data_rings(queue); in xenvif_disconnect_queue()
716 int xenvif_connect_data(struct xenvif_queue *queue, in xenvif_connect_data() argument
722 struct xenbus_device *dev = xenvif_to_xenbus_device(queue->vif); in xenvif_connect_data()
726 BUG_ON(queue->tx_irq); in xenvif_connect_data()
727 BUG_ON(queue->task); in xenvif_connect_data()
728 BUG_ON(queue->dealloc_task); in xenvif_connect_data()
730 err = xenvif_map_frontend_data_rings(queue, tx_ring_ref, in xenvif_connect_data()
735 init_waitqueue_head(&queue->wq); in xenvif_connect_data()
736 init_waitqueue_head(&queue->dealloc_wq); in xenvif_connect_data()
737 atomic_set(&queue->inflight_packets, 0); in xenvif_connect_data()
739 netif_napi_add(queue->vif->dev, &queue->napi, xenvif_poll, in xenvif_connect_data()
742 queue->stalled = true; in xenvif_connect_data()
744 task = kthread_run(xenvif_kthread_guest_rx, queue, in xenvif_connect_data()
745 "%s-guest-rx", queue->name); in xenvif_connect_data()
748 queue->task = task; in xenvif_connect_data()
755 task = kthread_run(xenvif_dealloc_kthread, queue, in xenvif_connect_data()
756 "%s-dealloc", queue->name); in xenvif_connect_data()
759 queue->dealloc_task = task; in xenvif_connect_data()
765 queue->name, queue); in xenvif_connect_data()
768 queue->tx_irq = queue->rx_irq = err; in xenvif_connect_data()
769 disable_irq(queue->tx_irq); in xenvif_connect_data()
772 snprintf(queue->tx_irq_name, sizeof(queue->tx_irq_name), in xenvif_connect_data()
773 "%s-tx", queue->name); in xenvif_connect_data()
776 queue->tx_irq_name, queue); in xenvif_connect_data()
779 queue->tx_irq = err; in xenvif_connect_data()
780 disable_irq(queue->tx_irq); in xenvif_connect_data()
782 snprintf(queue->rx_irq_name, sizeof(queue->rx_irq_name), in xenvif_connect_data()
783 "%s-rx", queue->name); in xenvif_connect_data()
786 queue->rx_irq_name, queue); in xenvif_connect_data()
789 queue->rx_irq = err; in xenvif_connect_data()
790 disable_irq(queue->rx_irq); in xenvif_connect_data()
796 pr_warn("Could not allocate kthread for %s\n", queue->name); in xenvif_connect_data()
799 xenvif_disconnect_queue(queue); in xenvif_connect_data()
818 struct xenvif_queue *queue = NULL; in xenvif_disconnect_data() local
825 queue = &vif->queues[queue_index]; in xenvif_disconnect_data()
827 xenvif_disconnect_queue(queue); in xenvif_disconnect_data()
849 * Used for queue teardown from xenvif_free(), and on the
852 void xenvif_deinit_queue(struct xenvif_queue *queue) in xenvif_deinit_queue() argument
854 gnttab_free_pages(MAX_PENDING_REQS, queue->mmap_pages); in xenvif_deinit_queue()