Lines Matching refs:vqueue

57 	struct virtqueue *vqueue;  member
163 virtio_break_device(vioch->vqueue->vdev); in scmi_vio_channel_cleanup_sync()
233 struct device *dev = &vioch->vqueue->vdev->dev; in scmi_vio_feed_vq_rx()
239 rc = virtqueue_add_inbuf(vioch->vqueue, &sg_in, 1, msg, GFP_ATOMIC); in scmi_vio_feed_vq_rx()
243 virtqueue_kick(vioch->vqueue); in scmi_vio_feed_vq_rx()
263 static void scmi_vio_complete_cb(struct virtqueue *vqueue) in scmi_vio_complete_cb() argument
271 if (WARN_ON_ONCE(!vqueue->vdev->priv)) in scmi_vio_complete_cb()
273 vioch = &((struct scmi_vio_channel *)vqueue->vdev->priv)[vqueue->index]; in scmi_vio_complete_cb()
281 virtqueue_disable_cb(vqueue); in scmi_vio_complete_cb()
285 msg = virtqueue_get_buf(vqueue, &length); in scmi_vio_complete_cb()
287 if (virtqueue_enable_cb(vqueue)) { in scmi_vio_complete_cb()
353 scmi_vio_complete_cb(vioch->vqueue); in scmi_vio_deferred_tx_worker()
526 msg->poll_idx = virtqueue_enable_cb_prepare(vioch->vqueue); in virtio_send_message()
534 rc = virtqueue_add_sgs(vioch->vqueue, sgs, 1, 1, msg, GFP_ATOMIC); in virtio_send_message()
539 virtqueue_kick(vioch->vqueue); in virtio_send_message()
710 pending = virtqueue_poll(vioch->vqueue, msg->poll_idx); in virtio_poll_done()
717 virtqueue_disable_cb(vioch->vqueue); in virtio_poll_done()
723 while ((next_msg = virtqueue_get_buf(vioch->vqueue, &length))) { in virtio_poll_done()
775 pending = !virtqueue_enable_cb(vioch->vqueue); in virtio_poll_done()
777 msg->poll_idx = virtqueue_enable_cb_prepare(vioch->vqueue); in virtio_poll_done()
778 pending = virtqueue_poll(vioch->vqueue, msg->poll_idx); in virtio_poll_done()
846 channels[i].vqueue = vqs[i]; in scmi_vio_probe()
848 sz = virtqueue_get_vring_size(channels[i].vqueue); in scmi_vio_probe()