Lines Matching refs:VHOST_NET_VQ_TX

85 	VHOST_NET_VQ_TX = 1,  enumerator
589 struct vhost_net_virtqueue *nvq = &net->vqs[VHOST_NET_VQ_TX]; in vhost_exceeds_maxpend()
762 struct vhost_net_virtqueue *nvq = &net->vqs[VHOST_NET_VQ_TX]; in handle_tx_copy()
851 struct vhost_net_virtqueue *nvq = &net->vqs[VHOST_NET_VQ_TX]; in handle_tx_zerocopy()
953 struct vhost_net_virtqueue *nvq = &net->vqs[VHOST_NET_VQ_TX]; in handle_tx()
957 mutex_lock_nested(&vq->mutex, VHOST_NET_VQ_TX); in handle_tx()
1002 struct vhost_net_virtqueue *tnvq = &net->vqs[VHOST_NET_VQ_TX]; in vhost_net_rx_peek_head_len()
1265 poll[VHOST_NET_VQ_TX].work); in handle_tx_net()
1310 n->vqs[VHOST_NET_VQ_TX].xdp = xdp; in vhost_net_open()
1313 vqs[VHOST_NET_VQ_TX] = &n->vqs[VHOST_NET_VQ_TX].vq; in vhost_net_open()
1315 n->vqs[VHOST_NET_VQ_TX].vq.handle_kick = handle_tx_kick; in vhost_net_open()
1333 vhost_poll_init(n->poll + VHOST_NET_VQ_TX, handle_tx_net, EPOLLOUT, dev); in vhost_net_open()
1363 *tx_sock = vhost_net_stop_vq(n, &n->vqs[VHOST_NET_VQ_TX].vq); in vhost_net_stop()
1375 vhost_net_flush_vq(n, VHOST_NET_VQ_TX); in vhost_net_flush()
1377 if (n->vqs[VHOST_NET_VQ_TX].ubufs) { in vhost_net_flush()
1378 mutex_lock(&n->vqs[VHOST_NET_VQ_TX].vq.mutex); in vhost_net_flush()
1380 mutex_unlock(&n->vqs[VHOST_NET_VQ_TX].vq.mutex); in vhost_net_flush()
1382 vhost_net_ubuf_put_and_wait(n->vqs[VHOST_NET_VQ_TX].ubufs); in vhost_net_flush()
1383 mutex_lock(&n->vqs[VHOST_NET_VQ_TX].vq.mutex); in vhost_net_flush()
1385 atomic_set(&n->vqs[VHOST_NET_VQ_TX].ubufs->refcount, 1); in vhost_net_flush()
1386 mutex_unlock(&n->vqs[VHOST_NET_VQ_TX].vq.mutex); in vhost_net_flush()
1411 kfree(n->vqs[VHOST_NET_VQ_TX].xdp); in vhost_net_release()
1785 vhost_net_enable_zcopy(VHOST_NET_VQ_TX); in vhost_net_init()