131638bcaSCyrill Gorcunov #include "kvm/virtio-pci-dev.h" 2b5ee1ea7SAsias He #include "kvm/virtio-net.h" 34f56d42cSAsias He #include "kvm/virtio.h" 44f56d42cSAsias He #include "kvm/ioport.h" 54f56d42cSAsias He #include "kvm/types.h" 64f56d42cSAsias He #include "kvm/mutex.h" 74f56d42cSAsias He #include "kvm/util.h" 84f56d42cSAsias He #include "kvm/kvm.h" 94f56d42cSAsias He #include "kvm/pci.h" 102449f6e3SSasha Levin #include "kvm/irq.h" 11b5ee1ea7SAsias He #include "kvm/uip.h" 1227ab67f5SSasha Levin #include "kvm/ioeventfd.h" 134f56d42cSAsias He 144f56d42cSAsias He #include <linux/virtio_net.h> 154f56d42cSAsias He #include <linux/if_tun.h> 16c229370aSIngo Molnar 17c229370aSIngo Molnar #include <arpa/inet.h> 184f56d42cSAsias He #include <net/if.h> 19c229370aSIngo Molnar 20c229370aSIngo Molnar #include <unistd.h> 214f56d42cSAsias He #include <assert.h> 224f56d42cSAsias He #include <fcntl.h> 23c229370aSIngo Molnar 24cb7202c1SSasha Levin #include <sys/socket.h> 25c229370aSIngo Molnar #include <sys/ioctl.h> 26c229370aSIngo Molnar #include <sys/types.h> 2773b7d038SAmos Kong #include <sys/wait.h> 284f56d42cSAsias He 294f56d42cSAsias He #define VIRTIO_NET_QUEUE_SIZE 128 304f56d42cSAsias He #define VIRTIO_NET_NUM_QUEUES 2 314f56d42cSAsias He #define VIRTIO_NET_RX_QUEUE 0 324f56d42cSAsias He #define VIRTIO_NET_TX_QUEUE 1 334f56d42cSAsias He 34c229370aSIngo Molnar static struct pci_device_header pci_header = { 352449f6e3SSasha Levin .vendor_id = PCI_VENDOR_ID_REDHAT_QUMRANET, 362449f6e3SSasha Levin .device_id = PCI_DEVICE_ID_VIRTIO_NET, 372449f6e3SSasha Levin .header_type = PCI_HEADER_TYPE_NORMAL, 382449f6e3SSasha Levin .revision_id = 0, 392449f6e3SSasha Levin .class = 0x020000, 402449f6e3SSasha Levin .subsys_vendor_id = PCI_SUBSYSTEM_VENDOR_ID_REDHAT_QUMRANET, 410a7ab0c6SSasha Levin .subsys_id = VIRTIO_ID_NET, 422449f6e3SSasha Levin }; 432449f6e3SSasha Levin 44b4fdde6dSAsias He struct net_dev; 45b4fdde6dSAsias He 46b4fdde6dSAsias He struct net_dev_operations { 47b4fdde6dSAsias He int (*rx)(struct iovec *iov, u16 in, struct net_dev *ndev); 48b4fdde6dSAsias He int (*tx)(struct iovec *iov, u16 in, struct net_dev *ndev); 49b4fdde6dSAsias He }; 50b4fdde6dSAsias He 518626798bSAsias He struct net_dev { 524f56d42cSAsias He pthread_mutex_t mutex; 534f56d42cSAsias He 544f56d42cSAsias He struct virt_queue vqs[VIRTIO_NET_NUM_QUEUES]; 55c229370aSIngo Molnar struct virtio_net_config config; 563fdf659dSSasha Levin u32 host_features; 573fdf659dSSasha Levin u32 guest_features; 583fdf659dSSasha Levin u16 config_vector; 593fdf659dSSasha Levin u8 status; 607f5ffaf5SAsias He u8 isr; 613fdf659dSSasha Levin u16 queue_selector; 6207f90696SSasha Levin u16 base_addr; 63*3395f880SSasha Levin u32 vq_vector[VIRTIO_NET_NUM_QUEUES]; 64*3395f880SSasha Levin u32 gsis[VIRTIO_NET_NUM_QUEUES]; 65*3395f880SSasha Levin u32 msix_io_block; 664f56d42cSAsias He 67c4aa7c02SPekka Enberg pthread_t io_rx_thread; 68c229370aSIngo Molnar pthread_mutex_t io_rx_lock; 69c4aa7c02SPekka Enberg pthread_cond_t io_rx_cond; 70c4aa7c02SPekka Enberg 71c4aa7c02SPekka Enberg pthread_t io_tx_thread; 72c229370aSIngo Molnar pthread_mutex_t io_tx_lock; 73c4aa7c02SPekka Enberg pthread_cond_t io_tx_cond; 74c4aa7c02SPekka Enberg 754f56d42cSAsias He int tap_fd; 764f56d42cSAsias He char tap_name[IFNAMSIZ]; 77bb1a32f1SAsias He 78bb1a32f1SAsias He int mode; 79bb1a32f1SAsias He 80b5ee1ea7SAsias He struct uip_info info; 81b4fdde6dSAsias He struct net_dev_operations *ops; 824f56d42cSAsias He }; 834f56d42cSAsias He 848626798bSAsias He static struct net_dev ndev = { 854f56d42cSAsias He .mutex = PTHREAD_MUTEX_INITIALIZER, 864f56d42cSAsias He 87c229370aSIngo Molnar .config = { 884f56d42cSAsias He .status = VIRTIO_NET_S_LINK_UP, 894f56d42cSAsias He }, 90407475bfSPekka Enberg .host_features = 1UL << VIRTIO_NET_F_MAC 91407475bfSPekka Enberg | 1UL << VIRTIO_NET_F_CSUM 92407475bfSPekka Enberg | 1UL << VIRTIO_NET_F_HOST_UFO 93407475bfSPekka Enberg | 1UL << VIRTIO_NET_F_HOST_TSO4 94407475bfSPekka Enberg | 1UL << VIRTIO_NET_F_HOST_TSO6 95407475bfSPekka Enberg | 1UL << VIRTIO_NET_F_GUEST_UFO 96407475bfSPekka Enberg | 1UL << VIRTIO_NET_F_GUEST_TSO4 97407475bfSPekka Enberg | 1UL << VIRTIO_NET_F_GUEST_TSO6, 98b5ee1ea7SAsias He .info = { 99b5ee1ea7SAsias He .buf_nr = 20, 100b5ee1ea7SAsias He } 1014f56d42cSAsias He }; 1024f56d42cSAsias He 103c4aa7c02SPekka Enberg static void *virtio_net_rx_thread(void *p) 1044f56d42cSAsias He { 1054f56d42cSAsias He struct iovec iov[VIRTIO_NET_QUEUE_SIZE]; 1064f56d42cSAsias He struct virt_queue *vq; 10743835ac9SSasha Levin struct kvm *kvm; 1083fdf659dSSasha Levin u16 out, in; 1093fdf659dSSasha Levin u16 head; 1104f56d42cSAsias He int len; 1114f56d42cSAsias He 11243835ac9SSasha Levin kvm = p; 113c229370aSIngo Molnar vq = &ndev.vqs[VIRTIO_NET_RX_QUEUE]; 114c4aa7c02SPekka Enberg 115c4aa7c02SPekka Enberg while (1) { 116b5ee1ea7SAsias He 117c229370aSIngo Molnar mutex_lock(&ndev.io_rx_lock); 118c4aa7c02SPekka Enberg if (!virt_queue__available(vq)) 119c229370aSIngo Molnar pthread_cond_wait(&ndev.io_rx_cond, &ndev.io_rx_lock); 120c229370aSIngo Molnar mutex_unlock(&ndev.io_rx_lock); 1214f56d42cSAsias He 1224f56d42cSAsias He while (virt_queue__available(vq)) { 123b5ee1ea7SAsias He 12443835ac9SSasha Levin head = virt_queue__get_iov(vq, iov, &out, &in, kvm); 125b5ee1ea7SAsias He 126b4fdde6dSAsias He len = ndev.ops->rx(iov, in, &ndev); 127b5ee1ea7SAsias He 128246c8347SAsias He virt_queue__set_used_elem(vq, head, len); 1297f5ffaf5SAsias He 130c4aa7c02SPekka Enberg /* We should interrupt guest right now, otherwise latency is huge. */ 131*3395f880SSasha Levin kvm__irq_trigger(kvm, ndev.gsis[VIRTIO_NET_RX_QUEUE]); 1324f56d42cSAsias He } 1334f56d42cSAsias He 134c4aa7c02SPekka Enberg } 135c4aa7c02SPekka Enberg 136c4aa7c02SPekka Enberg pthread_exit(NULL); 137c4aa7c02SPekka Enberg return NULL; 138c4aa7c02SPekka Enberg 139c4aa7c02SPekka Enberg } 140c4aa7c02SPekka Enberg 141c4aa7c02SPekka Enberg static void *virtio_net_tx_thread(void *p) 1424f56d42cSAsias He { 1434f56d42cSAsias He struct iovec iov[VIRTIO_NET_QUEUE_SIZE]; 1444f56d42cSAsias He struct virt_queue *vq; 14543835ac9SSasha Levin struct kvm *kvm; 1463fdf659dSSasha Levin u16 out, in; 1473fdf659dSSasha Levin u16 head; 1484f56d42cSAsias He int len; 1494f56d42cSAsias He 15043835ac9SSasha Levin kvm = p; 151c229370aSIngo Molnar vq = &ndev.vqs[VIRTIO_NET_TX_QUEUE]; 152c4aa7c02SPekka Enberg 153c4aa7c02SPekka Enberg while (1) { 154c229370aSIngo Molnar mutex_lock(&ndev.io_tx_lock); 155c4aa7c02SPekka Enberg if (!virt_queue__available(vq)) 156c229370aSIngo Molnar pthread_cond_wait(&ndev.io_tx_cond, &ndev.io_tx_lock); 157c229370aSIngo Molnar mutex_unlock(&ndev.io_tx_lock); 1584f56d42cSAsias He 1594f56d42cSAsias He while (virt_queue__available(vq)) { 160b5ee1ea7SAsias He 16143835ac9SSasha Levin head = virt_queue__get_iov(vq, iov, &out, &in, kvm); 162b5ee1ea7SAsias He 163b4fdde6dSAsias He len = ndev.ops->tx(iov, out, &ndev); 164b5ee1ea7SAsias He 1654f56d42cSAsias He virt_queue__set_used_elem(vq, head, len); 1664f56d42cSAsias He } 1674f56d42cSAsias He 168*3395f880SSasha Levin kvm__irq_trigger(kvm, ndev.gsis[VIRTIO_NET_TX_QUEUE]); 1694f56d42cSAsias He } 1704f56d42cSAsias He 171c4aa7c02SPekka Enberg pthread_exit(NULL); 172407475bfSPekka Enberg 173c4aa7c02SPekka Enberg return NULL; 174c4aa7c02SPekka Enberg 175c4aa7c02SPekka Enberg } 176407475bfSPekka Enberg 1773fdf659dSSasha Levin static bool virtio_net_pci_io_device_specific_in(void *data, unsigned long offset, int size, u32 count) 1784f56d42cSAsias He { 179c229370aSIngo Molnar u8 *config_space = (u8 *)&ndev.config; 1804f56d42cSAsias He 1814f56d42cSAsias He if (size != 1 || count != 1) 1824f56d42cSAsias He return false; 1834f56d42cSAsias He 184b8f43678SSasha Levin if ((offset - VIRTIO_MSI_CONFIG_VECTOR) > sizeof(struct virtio_net_config)) 1854542f276SCyrill Gorcunov pr_error("config offset is too big: %li", offset - VIRTIO_MSI_CONFIG_VECTOR); 1864f56d42cSAsias He 187b8f43678SSasha Levin ioport__write8(data, config_space[offset - VIRTIO_MSI_CONFIG_VECTOR]); 1884f56d42cSAsias He 1894f56d42cSAsias He return true; 1904f56d42cSAsias He } 1914f56d42cSAsias He 1923d62dea6SSasha Levin static bool virtio_net_pci_io_in(struct ioport *ioport, struct kvm *kvm, u16 port, void *data, int size, u32 count) 1934f56d42cSAsias He { 19407f90696SSasha Levin unsigned long offset = port - ndev.base_addr; 1954f56d42cSAsias He bool ret = true; 1964f56d42cSAsias He 197c229370aSIngo Molnar mutex_lock(&ndev.mutex); 1984f56d42cSAsias He 1994f56d42cSAsias He switch (offset) { 2004f56d42cSAsias He case VIRTIO_PCI_HOST_FEATURES: 201c229370aSIngo Molnar ioport__write32(data, ndev.host_features); 2024f56d42cSAsias He break; 2034f56d42cSAsias He case VIRTIO_PCI_GUEST_FEATURES: 2044f56d42cSAsias He ret = false; 2054f56d42cSAsias He break; 2064f56d42cSAsias He case VIRTIO_PCI_QUEUE_PFN: 207c229370aSIngo Molnar ioport__write32(data, ndev.vqs[ndev.queue_selector].pfn); 2084f56d42cSAsias He break; 2094f56d42cSAsias He case VIRTIO_PCI_QUEUE_NUM: 2104f56d42cSAsias He ioport__write16(data, VIRTIO_NET_QUEUE_SIZE); 2114f56d42cSAsias He break; 2124f56d42cSAsias He case VIRTIO_PCI_QUEUE_SEL: 2134f56d42cSAsias He case VIRTIO_PCI_QUEUE_NOTIFY: 2144f56d42cSAsias He ret = false; 2154f56d42cSAsias He break; 2164f56d42cSAsias He case VIRTIO_PCI_STATUS: 217c229370aSIngo Molnar ioport__write8(data, ndev.status); 2184f56d42cSAsias He break; 2194f56d42cSAsias He case VIRTIO_PCI_ISR: 220c229370aSIngo Molnar ioport__write8(data, ndev.isr); 221c229370aSIngo Molnar kvm__irq_line(kvm, pci_header.irq_line, VIRTIO_IRQ_LOW); 222c229370aSIngo Molnar ndev.isr = VIRTIO_IRQ_LOW; 2234f56d42cSAsias He break; 224*3395f880SSasha Levin case VIRTIO_MSI_CONFIG_VECTOR: 225*3395f880SSasha Levin ioport__write16(data, ndev.config_vector); 226*3395f880SSasha Levin break; 227*3395f880SSasha Levin case VIRTIO_MSI_QUEUE_VECTOR: 228*3395f880SSasha Levin ioport__write16(data, ndev.vq_vector[ndev.queue_selector]); 229*3395f880SSasha Levin break; 2304f56d42cSAsias He default: 2314f56d42cSAsias He ret = virtio_net_pci_io_device_specific_in(data, offset, size, count); 2324f56d42cSAsias He }; 2334f56d42cSAsias He 234c229370aSIngo Molnar mutex_unlock(&ndev.mutex); 2354f56d42cSAsias He 2364f56d42cSAsias He return ret; 2374f56d42cSAsias He } 2384f56d42cSAsias He 23943835ac9SSasha Levin static void virtio_net_handle_callback(struct kvm *kvm, u16 queue_index) 2404f56d42cSAsias He { 241407475bfSPekka Enberg switch (queue_index) { 242b5ee1ea7SAsias He case VIRTIO_NET_TX_QUEUE: 243c229370aSIngo Molnar mutex_lock(&ndev.io_tx_lock); 244c229370aSIngo Molnar pthread_cond_signal(&ndev.io_tx_cond); 245c229370aSIngo Molnar mutex_unlock(&ndev.io_tx_lock); 246407475bfSPekka Enberg break; 247b5ee1ea7SAsias He case VIRTIO_NET_RX_QUEUE: 248c229370aSIngo Molnar mutex_lock(&ndev.io_rx_lock); 249c229370aSIngo Molnar pthread_cond_signal(&ndev.io_rx_cond); 250c229370aSIngo Molnar mutex_unlock(&ndev.io_rx_lock); 251407475bfSPekka Enberg break; 252407475bfSPekka Enberg default: 2534542f276SCyrill Gorcunov pr_warning("Unknown queue index %u", queue_index); 254c4aa7c02SPekka Enberg } 2554f56d42cSAsias He } 2564f56d42cSAsias He 2573d62dea6SSasha Levin static bool virtio_net_pci_io_out(struct ioport *ioport, struct kvm *kvm, u16 port, void *data, int size, u32 count) 2584f56d42cSAsias He { 25907f90696SSasha Levin unsigned long offset = port - ndev.base_addr; 2604f56d42cSAsias He bool ret = true; 2614f56d42cSAsias He 262c229370aSIngo Molnar mutex_lock(&ndev.mutex); 2634f56d42cSAsias He 2644f56d42cSAsias He switch (offset) { 2654f56d42cSAsias He case VIRTIO_PCI_GUEST_FEATURES: 266c229370aSIngo Molnar ndev.guest_features = ioport__read32(data); 2674f56d42cSAsias He break; 2684f56d42cSAsias He case VIRTIO_PCI_QUEUE_PFN: { 2694f56d42cSAsias He struct virt_queue *queue; 2704f56d42cSAsias He void *p; 2714f56d42cSAsias He 272c229370aSIngo Molnar assert(ndev.queue_selector < VIRTIO_NET_NUM_QUEUES); 2734f56d42cSAsias He 274c229370aSIngo Molnar queue = &ndev.vqs[ndev.queue_selector]; 2754f56d42cSAsias He queue->pfn = ioport__read32(data); 27643835ac9SSasha Levin p = guest_pfn_to_host(kvm, queue->pfn); 2774f56d42cSAsias He 278b8f43678SSasha Levin vring_init(&queue->vring, VIRTIO_NET_QUEUE_SIZE, p, VIRTIO_PCI_VRING_ALIGN); 2794f56d42cSAsias He 2804f56d42cSAsias He break; 2814f56d42cSAsias He } 2824f56d42cSAsias He case VIRTIO_PCI_QUEUE_SEL: 283c229370aSIngo Molnar ndev.queue_selector = ioport__read16(data); 2844f56d42cSAsias He break; 2854f56d42cSAsias He case VIRTIO_PCI_QUEUE_NOTIFY: { 2863fdf659dSSasha Levin u16 queue_index; 287c229370aSIngo Molnar 2884f56d42cSAsias He queue_index = ioport__read16(data); 28943835ac9SSasha Levin virtio_net_handle_callback(kvm, queue_index); 2904f56d42cSAsias He break; 2914f56d42cSAsias He } 2924f56d42cSAsias He case VIRTIO_PCI_STATUS: 293c229370aSIngo Molnar ndev.status = ioport__read8(data); 2944f56d42cSAsias He break; 2954f56d42cSAsias He case VIRTIO_MSI_CONFIG_VECTOR: 296*3395f880SSasha Levin ndev.config_vector = ioport__read16(data); 2974f56d42cSAsias He break; 298*3395f880SSasha Levin case VIRTIO_MSI_QUEUE_VECTOR: { 299*3395f880SSasha Levin u32 gsi; 300*3395f880SSasha Levin u32 vec; 301*3395f880SSasha Levin 302*3395f880SSasha Levin vec = ndev.vq_vector[ndev.queue_selector] = ioport__read16(data); 303*3395f880SSasha Levin 304*3395f880SSasha Levin gsi = irq__add_msix_route(kvm, 305*3395f880SSasha Levin pci_header.msix.table[vec].low, 306*3395f880SSasha Levin pci_header.msix.table[vec].high, 307*3395f880SSasha Levin pci_header.msix.table[vec].data); 308*3395f880SSasha Levin 309*3395f880SSasha Levin ndev.gsis[ndev.queue_selector] = gsi; 3104f56d42cSAsias He break; 311*3395f880SSasha Levin } 3124f56d42cSAsias He default: 3134f56d42cSAsias He ret = false; 3144f56d42cSAsias He }; 3154f56d42cSAsias He 316c229370aSIngo Molnar mutex_unlock(&ndev.mutex); 317407475bfSPekka Enberg 3184f56d42cSAsias He return ret; 3194f56d42cSAsias He } 3204f56d42cSAsias He 32127ab67f5SSasha Levin static void ioevent_callback(struct kvm *kvm, void *param) 32227ab67f5SSasha Levin { 323926e0e2fSIngo Molnar virtio_net_handle_callback(kvm, (u64)(long)param); 32427ab67f5SSasha Levin } 32527ab67f5SSasha Levin 3264f56d42cSAsias He static struct ioport_operations virtio_net_io_ops = { 3274f56d42cSAsias He .io_in = virtio_net_pci_io_in, 3284f56d42cSAsias He .io_out = virtio_net_pci_io_out, 3294f56d42cSAsias He }; 3304f56d42cSAsias He 331*3395f880SSasha Levin static void callback_mmio(u64 addr, u8 *data, u32 len, u8 is_write, void *ptr) 332*3395f880SSasha Levin { 333*3395f880SSasha Levin void *table = pci_header.msix.table; 334*3395f880SSasha Levin if (is_write) 335*3395f880SSasha Levin memcpy(table + addr - ndev.msix_io_block, data, len); 336*3395f880SSasha Levin else 337*3395f880SSasha Levin memcpy(data, table + addr - ndev.msix_io_block, len); 338*3395f880SSasha Levin } 339*3395f880SSasha Levin 3403b02f580SSasha Levin static bool virtio_net__tap_init(const struct virtio_net_parameters *params) 3414f56d42cSAsias He { 342cb7202c1SSasha Levin int sock = socket(AF_INET, SOCK_STREAM, 0); 343f715177dSAsias He int pid, status, offload, hdr_len; 344cb7202c1SSasha Levin struct sockaddr_in sin = {0}; 345246c8347SAsias He struct ifreq ifr; 3464f56d42cSAsias He 347c229370aSIngo Molnar ndev.tap_fd = open("/dev/net/tun", O_RDWR); 348c229370aSIngo Molnar if (ndev.tap_fd < 0) { 3494542f276SCyrill Gorcunov pr_warning("Unable to open /dev/net/tun"); 3503b02f580SSasha Levin goto fail; 3513b02f580SSasha Levin } 3524f56d42cSAsias He 3534f56d42cSAsias He memset(&ifr, 0, sizeof(ifr)); 354246c8347SAsias He ifr.ifr_flags = IFF_TAP | IFF_NO_PI | IFF_VNET_HDR; 355c229370aSIngo Molnar if (ioctl(ndev.tap_fd, TUNSETIFF, &ifr) < 0) { 3564542f276SCyrill Gorcunov pr_warning("Config tap device error. Are you root?"); 3573b02f580SSasha Levin goto fail; 3583b02f580SSasha Levin } 3594f56d42cSAsias He 360c229370aSIngo Molnar strncpy(ndev.tap_name, ifr.ifr_name, sizeof(ndev.tap_name)); 3614f56d42cSAsias He 362c229370aSIngo Molnar if (ioctl(ndev.tap_fd, TUNSETNOCSUM, 1) < 0) { 3634542f276SCyrill Gorcunov pr_warning("Config tap device TUNSETNOCSUM error"); 364246c8347SAsias He goto fail; 365246c8347SAsias He } 366246c8347SAsias He 367246c8347SAsias He hdr_len = sizeof(struct virtio_net_hdr); 368c229370aSIngo Molnar if (ioctl(ndev.tap_fd, TUNSETVNETHDRSZ, &hdr_len) < 0) { 3694542f276SCyrill Gorcunov pr_warning("Config tap device TUNSETVNETHDRSZ error"); 370246c8347SAsias He } 371246c8347SAsias He 372246c8347SAsias He offload = TUN_F_CSUM | TUN_F_TSO4 | TUN_F_TSO6 | TUN_F_UFO; 373c229370aSIngo Molnar if (ioctl(ndev.tap_fd, TUNSETOFFLOAD, offload) < 0) { 3744542f276SCyrill Gorcunov pr_warning("Config tap device TUNSETOFFLOAD error"); 375246c8347SAsias He goto fail; 376246c8347SAsias He } 3774f56d42cSAsias He 37873b7d038SAmos Kong if (strcmp(params->script, "none")) { 37973b7d038SAmos Kong pid = fork(); 38073b7d038SAmos Kong if (pid == 0) { 381c229370aSIngo Molnar execl(params->script, params->script, ndev.tap_name, NULL); 38273b7d038SAmos Kong _exit(1); 38373b7d038SAmos Kong } else { 38473b7d038SAmos Kong waitpid(pid, &status, 0); 38573b7d038SAmos Kong if (WIFEXITED(status) && WEXITSTATUS(status) != 0) { 3864542f276SCyrill Gorcunov pr_warning("Fail to setup tap by %s", params->script); 38773b7d038SAmos Kong goto fail; 38873b7d038SAmos Kong } 38973b7d038SAmos Kong } 39073b7d038SAmos Kong } else { 391cb7202c1SSasha Levin memset(&ifr, 0, sizeof(ifr)); 392c229370aSIngo Molnar strncpy(ifr.ifr_name, ndev.tap_name, sizeof(ndev.tap_name)); 393bdfcfca6SSasha Levin sin.sin_addr.s_addr = inet_addr(params->host_ip); 394cb7202c1SSasha Levin memcpy(&(ifr.ifr_addr), &sin, sizeof(ifr.ifr_addr)); 395cb7202c1SSasha Levin ifr.ifr_addr.sa_family = AF_INET; 3963b02f580SSasha Levin if (ioctl(sock, SIOCSIFADDR, &ifr) < 0) { 3974542f276SCyrill Gorcunov pr_warning("Could not set ip address on tap device"); 3983b02f580SSasha Levin goto fail; 3993b02f580SSasha Levin } 40073b7d038SAmos Kong } 401cb7202c1SSasha Levin 402cb7202c1SSasha Levin memset(&ifr, 0, sizeof(ifr)); 403c229370aSIngo Molnar strncpy(ifr.ifr_name, ndev.tap_name, sizeof(ndev.tap_name)); 404cb7202c1SSasha Levin ioctl(sock, SIOCGIFFLAGS, &ifr); 405cb7202c1SSasha Levin ifr.ifr_flags |= IFF_UP | IFF_RUNNING; 406cb7202c1SSasha Levin if (ioctl(sock, SIOCSIFFLAGS, &ifr) < 0) 4074542f276SCyrill Gorcunov pr_warning("Could not bring tap device up"); 408cb7202c1SSasha Levin 409cb7202c1SSasha Levin close(sock); 4103b02f580SSasha Levin 4113b02f580SSasha Levin return 1; 4123b02f580SSasha Levin 4133b02f580SSasha Levin fail: 4143b02f580SSasha Levin if (sock >= 0) 4153b02f580SSasha Levin close(sock); 416c229370aSIngo Molnar if (ndev.tap_fd >= 0) 417c229370aSIngo Molnar close(ndev.tap_fd); 4183b02f580SSasha Levin 4193b02f580SSasha Levin return 0; 4204f56d42cSAsias He } 4214f56d42cSAsias He 42243835ac9SSasha Levin static void virtio_net__io_thread_init(struct kvm *kvm) 423c4aa7c02SPekka Enberg { 424c229370aSIngo Molnar pthread_mutex_init(&ndev.io_rx_lock, NULL); 425c229370aSIngo Molnar pthread_cond_init(&ndev.io_tx_cond, NULL); 426c4aa7c02SPekka Enberg 427c229370aSIngo Molnar pthread_mutex_init(&ndev.io_rx_lock, NULL); 428c229370aSIngo Molnar pthread_cond_init(&ndev.io_tx_cond, NULL); 429c4aa7c02SPekka Enberg 430c229370aSIngo Molnar pthread_create(&ndev.io_rx_thread, NULL, virtio_net_rx_thread, (void *)kvm); 431c229370aSIngo Molnar pthread_create(&ndev.io_tx_thread, NULL, virtio_net_tx_thread, (void *)kvm); 432c4aa7c02SPekka Enberg } 433c4aa7c02SPekka Enberg 434b4fdde6dSAsias He static inline int tap_ops_tx(struct iovec *iov, u16 out, struct net_dev *ndev) 435b4fdde6dSAsias He { 436b4fdde6dSAsias He return writev(ndev->tap_fd, iov, out); 437b4fdde6dSAsias He } 438b4fdde6dSAsias He 439b4fdde6dSAsias He static inline int tap_ops_rx(struct iovec *iov, u16 in, struct net_dev *ndev) 440b4fdde6dSAsias He { 441b4fdde6dSAsias He return readv(ndev->tap_fd, iov, in); 442b4fdde6dSAsias He } 443b4fdde6dSAsias He 444b4fdde6dSAsias He static inline int uip_ops_tx(struct iovec *iov, u16 out, struct net_dev *ndev) 445b4fdde6dSAsias He { 446b4fdde6dSAsias He return uip_tx(iov, out, &ndev->info); 447b4fdde6dSAsias He } 448b4fdde6dSAsias He 449b4fdde6dSAsias He static inline int uip_ops_rx(struct iovec *iov, u16 in, struct net_dev *ndev) 450b4fdde6dSAsias He { 451b4fdde6dSAsias He return uip_rx(iov, in, &ndev->info); 452b4fdde6dSAsias He } 453b4fdde6dSAsias He 454b4fdde6dSAsias He static struct net_dev_operations tap_ops = { 455b4fdde6dSAsias He .rx = tap_ops_rx, 456b4fdde6dSAsias He .tx = tap_ops_tx, 457b4fdde6dSAsias He }; 458b4fdde6dSAsias He 459b4fdde6dSAsias He static struct net_dev_operations uip_ops = { 460b4fdde6dSAsias He .rx = uip_ops_rx, 461b4fdde6dSAsias He .tx = uip_ops_tx, 462b4fdde6dSAsias He }; 463b4fdde6dSAsias He 464bdfcfca6SSasha Levin void virtio_net__init(const struct virtio_net_parameters *params) 4654f56d42cSAsias He { 466b5ee1ea7SAsias He struct ioevent ioevent; 4672449f6e3SSasha Levin u8 dev, line, pin; 46807f90696SSasha Levin u16 net_base_addr; 469b5ee1ea7SAsias He int i; 4702449f6e3SSasha Levin 4710a7ab0c6SSasha Levin if (irq__register_device(VIRTIO_ID_NET, &dev, &pin, &line) < 0) 4722449f6e3SSasha Levin return; 4732449f6e3SSasha Levin 474c229370aSIngo Molnar pci_header.irq_pin = pin; 475c229370aSIngo Molnar pci_header.irq_line = line; 47607f90696SSasha Levin net_base_addr = ioport__register(IOPORT_EMPTY, &virtio_net_io_ops, IOPORT_SIZE, NULL); 47707f90696SSasha Levin pci_header.bar[0] = net_base_addr | PCI_BASE_ADDRESS_SPACE_IO; 47807f90696SSasha Levin ndev.base_addr = net_base_addr; 479c229370aSIngo Molnar pci__register(&pci_header, dev); 480c4aa7c02SPekka Enberg 4810c54698eSAsias He for (i = 0 ; i < 6 ; i++) { 482f715177dSAsias He ndev.config.mac[i] = params->guest_mac[i]; 4830c54698eSAsias He ndev.info.guest_mac.addr[i] = params->guest_mac[i]; 4840c54698eSAsias He ndev.info.host_mac.addr[i] = params->host_mac[i]; 4850c54698eSAsias He } 486f715177dSAsias He 487b5ee1ea7SAsias He ndev.mode = params->mode; 488b4fdde6dSAsias He if (ndev.mode == NET_MODE_TAP) { 489b5ee1ea7SAsias He virtio_net__tap_init(params); 490b4fdde6dSAsias He ndev.ops = &tap_ops; 491b4fdde6dSAsias He } else { 4920c54698eSAsias He ndev.info.host_ip = ntohl(inet_addr(params->host_ip)); 4930c54698eSAsias He ndev.info.guest_ip = ntohl(inet_addr(params->guest_ip)); 4940c54698eSAsias He ndev.info.guest_netmask = ntohl(inet_addr("255.255.255.0")); 495b5ee1ea7SAsias He uip_init(&ndev.info); 496b4fdde6dSAsias He ndev.ops = &uip_ops; 497b4fdde6dSAsias He } 498b5ee1ea7SAsias He 499*3395f880SSasha Levin ndev.msix_io_block = pci_get_io_space_block(); 500*3395f880SSasha Levin kvm__register_mmio(params->kvm, ndev.msix_io_block, 0x100, callback_mmio, NULL); 501*3395f880SSasha Levin pci_header.bar[1] = ndev.msix_io_block | 502*3395f880SSasha Levin PCI_BASE_ADDRESS_SPACE_MEMORY | 503*3395f880SSasha Levin PCI_BASE_ADDRESS_MEM_TYPE_64; 504*3395f880SSasha Levin /* bar[2] is the continuation of bar[1] for 64bit addressing */ 505*3395f880SSasha Levin pci_header.bar[2] = 0; 506*3395f880SSasha Levin pci_header.status = PCI_STATUS_CAP_LIST; 507*3395f880SSasha Levin pci_header.capabilities = (void *)&pci_header.msix - (void *)&pci_header; 508*3395f880SSasha Levin 509*3395f880SSasha Levin pci_header.msix.cap = PCI_CAP_ID_MSIX; 510*3395f880SSasha Levin pci_header.msix.next = 0; 511*3395f880SSasha Levin pci_header.msix.table_size = (VIRTIO_NET_NUM_QUEUES + 1) | PCI_MSIX_FLAGS_ENABLE; 512*3395f880SSasha Levin pci_header.msix.table_offset = 1; /* Use BAR 1 */ 513*3395f880SSasha Levin 51443835ac9SSasha Levin virtio_net__io_thread_init(params->kvm); 51527ab67f5SSasha Levin 51627ab67f5SSasha Levin for (i = 0; i < VIRTIO_NET_NUM_QUEUES; i++) { 51727ab67f5SSasha Levin ioevent = (struct ioevent) { 51827ab67f5SSasha Levin .io_addr = net_base_addr + VIRTIO_PCI_QUEUE_NOTIFY, 51927ab67f5SSasha Levin .io_len = sizeof(u16), 52027ab67f5SSasha Levin .fn = ioevent_callback, 52127ab67f5SSasha Levin .datamatch = i, 522926e0e2fSIngo Molnar .fn_ptr = (void *)(long)i, 52327ab67f5SSasha Levin .fn_kvm = params->kvm, 52427ab67f5SSasha Levin .fd = eventfd(0, 0), 52527ab67f5SSasha Levin }; 52627ab67f5SSasha Levin 52727ab67f5SSasha Levin ioeventfd__add_event(&ioevent); 52827ab67f5SSasha Levin } 5294f56d42cSAsias He } 530