139d6af07SAsias He #include <linux/virtio_ring.h> 23fdf659dSSasha Levin #include <linux/types.h> 339d6af07SAsias He #include <sys/uio.h> 402eca50cSAsias He #include <stdlib.h> 52caa836dSIngo Molnar 62caa836dSIngo Molnar #include "kvm/barrier.h" 739d6af07SAsias He #include "kvm/virtio.h" 802eca50cSAsias He #include "kvm/virtio-pci.h" 9*755752d6SAsias He #include "kvm/virtio-mmio.h" 1002eca50cSAsias He #include "kvm/util.h" 1102eca50cSAsias He #include "kvm/kvm.h" 1202eca50cSAsias He 1339d6af07SAsias He 143fdf659dSSasha Levin struct vring_used_elem *virt_queue__set_used_elem(struct virt_queue *queue, u32 head, u32 len) 1539d6af07SAsias He { 1639d6af07SAsias He struct vring_used_elem *used_elem; 17407475bfSPekka Enberg 1894902782SSasha Levin used_elem = &queue->vring.used->ring[queue->vring.used->idx % queue->vring.num]; 1939d6af07SAsias He used_elem->id = head; 2039d6af07SAsias He used_elem->len = len; 2194902782SSasha Levin 2294902782SSasha Levin /* 2394902782SSasha Levin * Use wmb to assure that used elem was updated with head and len. 2494902782SSasha Levin * We need a wmb here since we can't advance idx unless we're ready 2594902782SSasha Levin * to pass the used element to the guest. 2694902782SSasha Levin */ 2794902782SSasha Levin wmb(); 2894902782SSasha Levin queue->vring.used->idx++; 2994902782SSasha Levin 3094902782SSasha Levin /* 3194902782SSasha Levin * Use wmb to assure used idx has been increased before we signal the guest. 3294902782SSasha Levin * Without a wmb here the guest may ignore the queue since it won't see 3394902782SSasha Levin * an updated idx. 3494902782SSasha Levin */ 3594902782SSasha Levin wmb(); 3694902782SSasha Levin 3739d6af07SAsias He return used_elem; 3839d6af07SAsias He } 3939d6af07SAsias He 40754c8ce3SSasha Levin /* 41754c8ce3SSasha Levin * Each buffer in the virtqueues is actually a chain of descriptors. This 42754c8ce3SSasha Levin * function returns the next descriptor in the chain, or vq->vring.num if we're 43754c8ce3SSasha Levin * at the end. 44754c8ce3SSasha Levin */ 45754c8ce3SSasha Levin static unsigned next_desc(struct vring_desc *desc, 46754c8ce3SSasha Levin unsigned int i, unsigned int max) 47754c8ce3SSasha Levin { 48754c8ce3SSasha Levin unsigned int next; 49754c8ce3SSasha Levin 50754c8ce3SSasha Levin /* If this descriptor says it doesn't chain, we're done. */ 51754c8ce3SSasha Levin if (!(desc[i].flags & VRING_DESC_F_NEXT)) 52754c8ce3SSasha Levin return max; 53754c8ce3SSasha Levin 54754c8ce3SSasha Levin /* Check they're not leading us off end of descriptors. */ 55754c8ce3SSasha Levin next = desc[i].next; 56754c8ce3SSasha Levin /* Make sure compiler knows to grab that: we don't want it changing! */ 57754c8ce3SSasha Levin wmb(); 58754c8ce3SSasha Levin 59754c8ce3SSasha Levin return next; 60754c8ce3SSasha Levin } 61754c8ce3SSasha Levin 622fddfdb5SAsias He u16 virt_queue__get_head_iov(struct virt_queue *vq, struct iovec iov[], u16 *out, u16 *in, u16 head, struct kvm *kvm) 6339d6af07SAsias He { 6439d6af07SAsias He struct vring_desc *desc; 652fddfdb5SAsias He u16 idx; 66754c8ce3SSasha Levin u16 max; 6739d6af07SAsias He 682fddfdb5SAsias He idx = head; 6939d6af07SAsias He *out = *in = 0; 70754c8ce3SSasha Levin max = vq->vring.num; 71754c8ce3SSasha Levin desc = vq->vring.desc; 72754c8ce3SSasha Levin 73754c8ce3SSasha Levin if (desc[idx].flags & VRING_DESC_F_INDIRECT) { 74754c8ce3SSasha Levin max = desc[idx].len / sizeof(struct vring_desc); 75754c8ce3SSasha Levin desc = guest_flat_to_host(kvm, desc[idx].addr); 76754c8ce3SSasha Levin idx = 0; 77754c8ce3SSasha Levin } 7839d6af07SAsias He 7939d6af07SAsias He do { 80754c8ce3SSasha Levin /* Grab the first descriptor, and check it's OK. */ 81754c8ce3SSasha Levin iov[*out + *in].iov_len = desc[idx].len; 82754c8ce3SSasha Levin iov[*out + *in].iov_base = guest_flat_to_host(kvm, desc[idx].addr); 83754c8ce3SSasha Levin /* If this is an input descriptor, increment that count. */ 84754c8ce3SSasha Levin if (desc[idx].flags & VRING_DESC_F_WRITE) 8539d6af07SAsias He (*in)++; 8639d6af07SAsias He else 8739d6af07SAsias He (*out)++; 88754c8ce3SSasha Levin } while ((idx = next_desc(desc, idx, max)) != max); 8939d6af07SAsias He 9039d6af07SAsias He return head; 9139d6af07SAsias He } 927f5ffaf5SAsias He 932fddfdb5SAsias He u16 virt_queue__get_iov(struct virt_queue *vq, struct iovec iov[], u16 *out, u16 *in, struct kvm *kvm) 942fddfdb5SAsias He { 952fddfdb5SAsias He u16 head; 962fddfdb5SAsias He 972fddfdb5SAsias He head = virt_queue__pop(vq); 982fddfdb5SAsias He 992fddfdb5SAsias He return virt_queue__get_head_iov(vq, iov, out, in, head, kvm); 1002fddfdb5SAsias He } 1012fddfdb5SAsias He 10208861bcfSAneesh Kumar K.V /* in and out are relative to guest */ 10308861bcfSAneesh Kumar K.V u16 virt_queue__get_inout_iov(struct kvm *kvm, struct virt_queue *queue, 10408861bcfSAneesh Kumar K.V struct iovec in_iov[], struct iovec out_iov[], 10508861bcfSAneesh Kumar K.V u16 *in, u16 *out) 10608861bcfSAneesh Kumar K.V { 10708861bcfSAneesh Kumar K.V struct vring_desc *desc; 1082fddfdb5SAsias He u16 head, idx; 10908861bcfSAneesh Kumar K.V 11008861bcfSAneesh Kumar K.V idx = head = virt_queue__pop(queue); 11108861bcfSAneesh Kumar K.V *out = *in = 0; 11208861bcfSAneesh Kumar K.V do { 11308861bcfSAneesh Kumar K.V desc = virt_queue__get_desc(queue, idx); 11408861bcfSAneesh Kumar K.V if (desc->flags & VRING_DESC_F_WRITE) { 11508861bcfSAneesh Kumar K.V in_iov[*in].iov_base = guest_flat_to_host(kvm, 11608861bcfSAneesh Kumar K.V desc->addr); 11708861bcfSAneesh Kumar K.V in_iov[*in].iov_len = desc->len; 11808861bcfSAneesh Kumar K.V (*in)++; 11908861bcfSAneesh Kumar K.V } else { 12008861bcfSAneesh Kumar K.V out_iov[*out].iov_base = guest_flat_to_host(kvm, 12108861bcfSAneesh Kumar K.V desc->addr); 12208861bcfSAneesh Kumar K.V out_iov[*out].iov_len = desc->len; 12308861bcfSAneesh Kumar K.V (*out)++; 12408861bcfSAneesh Kumar K.V } 12508861bcfSAneesh Kumar K.V if (desc->flags & VRING_DESC_F_NEXT) 12608861bcfSAneesh Kumar K.V idx = desc->next; 12708861bcfSAneesh Kumar K.V else 12808861bcfSAneesh Kumar K.V break; 12908861bcfSAneesh Kumar K.V } while (1); 1302fddfdb5SAsias He 13108861bcfSAneesh Kumar K.V return head; 13208861bcfSAneesh Kumar K.V } 13308861bcfSAneesh Kumar K.V 1341382aba0SSasha Levin int virtio__get_dev_specific_field(int offset, bool msix, u32 *config_off) 135c3a79fa1SSasha Levin { 136c3a79fa1SSasha Levin if (msix) { 137c3a79fa1SSasha Levin if (offset < 4) 138c3a79fa1SSasha Levin return VIRTIO_PCI_O_MSIX; 139c3a79fa1SSasha Levin else 140c3a79fa1SSasha Levin offset -= 4; 141c3a79fa1SSasha Levin } 142c3a79fa1SSasha Levin 143c3a79fa1SSasha Levin *config_off = offset; 144c3a79fa1SSasha Levin 145c3a79fa1SSasha Levin return VIRTIO_PCI_O_CONFIG; 146c3a79fa1SSasha Levin } 14751b1454fSAsias He 14851b1454fSAsias He bool virtio_queue__should_signal(struct virt_queue *vq) 14951b1454fSAsias He { 15051b1454fSAsias He u16 old_idx, new_idx, event_idx; 15151b1454fSAsias He 15251b1454fSAsias He old_idx = vq->last_used_signalled; 15351b1454fSAsias He new_idx = vq->vring.used->idx; 15451b1454fSAsias He event_idx = vring_used_event(&vq->vring); 15551b1454fSAsias He 15651b1454fSAsias He if (vring_need_event(event_idx, new_idx, old_idx)) { 15751b1454fSAsias He vq->last_used_signalled = new_idx; 15851b1454fSAsias He return true; 15951b1454fSAsias He } 16051b1454fSAsias He 16151b1454fSAsias He return false; 16251b1454fSAsias He } 16302eca50cSAsias He 16402eca50cSAsias He int virtio_init(struct kvm *kvm, void *dev, struct virtio_device *vdev, 16502eca50cSAsias He struct virtio_ops *ops, enum virtio_trans trans, 16602eca50cSAsias He int device_id, int subsys_id, int class) 16702eca50cSAsias He { 16802eca50cSAsias He void *virtio; 16902eca50cSAsias He 17002eca50cSAsias He switch (trans) { 17102eca50cSAsias He case VIRTIO_PCI: 17202eca50cSAsias He virtio = calloc(sizeof(struct virtio_pci), 1); 17302eca50cSAsias He if (!virtio) 17402eca50cSAsias He return -ENOMEM; 17502eca50cSAsias He vdev->virtio = virtio; 17602eca50cSAsias He vdev->ops = ops; 17702eca50cSAsias He vdev->ops->signal_vq = virtio_pci__signal_vq; 17802eca50cSAsias He vdev->ops->signal_config = virtio_pci__signal_config; 17902eca50cSAsias He vdev->ops->init = virtio_pci__init; 18002eca50cSAsias He vdev->ops->exit = virtio_pci__exit; 18102eca50cSAsias He vdev->ops->init(kvm, dev, vdev, device_id, subsys_id, class); 18202eca50cSAsias He break; 183*755752d6SAsias He case VIRTIO_MMIO: 184*755752d6SAsias He virtio = calloc(sizeof(struct virtio_mmio), 1); 185*755752d6SAsias He if (!virtio) 186*755752d6SAsias He return -ENOMEM; 187*755752d6SAsias He vdev->virtio = virtio; 188*755752d6SAsias He vdev->ops = ops; 189*755752d6SAsias He vdev->ops->signal_vq = virtio_mmio_signal_vq; 190*755752d6SAsias He vdev->ops->signal_config = virtio_mmio_signal_config; 191*755752d6SAsias He vdev->ops->init = virtio_mmio_init; 192*755752d6SAsias He vdev->ops->exit = virtio_mmio_exit; 193*755752d6SAsias He vdev->ops->init(kvm, dev, vdev, device_id, subsys_id, class); 194*755752d6SAsias He break; 19502eca50cSAsias He default: 19602eca50cSAsias He return -1; 19702eca50cSAsias He }; 19802eca50cSAsias He 19902eca50cSAsias He return 0; 20002eca50cSAsias He } 201