/qemu/tests/unit/ |
H A D | test-iov.c | 267 unsigned int iov_cnt; in test_discard_front() local 274 iov_random(&iov, &iov_cnt); in test_discard_front() 276 iov_cnt_tmp = iov_cnt; in test_discard_front() 280 g_assert(iov_cnt_tmp == iov_cnt); in test_discard_front() 281 iov_free(iov, iov_cnt); in test_discard_front() 284 iov_random(&iov, &iov_cnt); in test_discard_front() 286 iov_cnt_tmp = iov_cnt; in test_discard_front() 287 size = iov_size(iov, iov_cnt); in test_discard_front() 291 iov_free(iov, iov_cnt); in test_discard_front() 294 iov_random(&iov, &iov_cnt); in test_discard_front() [all …]
|
/qemu/include/qemu/ |
H A D | iov.h | 22 size_t iov_size(const struct iovec *iov, const unsigned int iov_cnt); 36 size_t iov_from_buf_full(const struct iovec *iov, unsigned int iov_cnt, 38 size_t iov_to_buf_full(const struct iovec *iov, const unsigned int iov_cnt, 42 iov_from_buf(const struct iovec *iov, unsigned int iov_cnt, in iov_from_buf() argument 45 if (__builtin_constant_p(bytes) && iov_cnt && in iov_from_buf() 50 return iov_from_buf_full(iov, iov_cnt, offset, buf, bytes); in iov_from_buf() 55 iov_to_buf(const struct iovec *iov, const unsigned int iov_cnt, in iov_to_buf() argument 58 if (__builtin_constant_p(bytes) && iov_cnt && in iov_to_buf() 63 return iov_to_buf_full(iov, iov_cnt, offset, buf, bytes); in iov_to_buf() 77 size_t iov_memset(const struct iovec *iov, const unsigned int iov_cnt, [all …]
|
H A D | coroutine.h | 295 unsigned iov_cnt, size_t offset, 297 #define qemu_co_recvv(sockfd, iov, iov_cnt, offset, bytes) \ argument 298 qemu_co_sendv_recvv(sockfd, iov, iov_cnt, offset, bytes, false) 299 #define qemu_co_sendv(sockfd, iov, iov_cnt, offset, bytes) \ argument 300 qemu_co_sendv_recvv(sockfd, iov, iov_cnt, offset, bytes, true)
|
H A D | crc32c.h | 33 uint32_t iov_crc32c(uint32_t crc, const struct iovec *iov, size_t iov_cnt);
|
H A D | osdep.h | 657 ssize_t readv(int fd, const struct iovec *iov, int iov_cnt); 658 ssize_t writev(int fd, const struct iovec *iov, int iov_cnt);
|
/qemu/util/ |
H A D | iov.c | 25 size_t iov_from_buf_full(const struct iovec *iov, unsigned int iov_cnt, in iov_from_buf_full() argument 30 for (i = 0, done = 0; (offset || done < bytes) && i < iov_cnt; i++) { in iov_from_buf_full() 43 size_t iov_to_buf_full(const struct iovec *iov, const unsigned int iov_cnt, in iov_to_buf_full() argument 48 for (i = 0, done = 0; (offset || done < bytes) && i < iov_cnt; i++) { in iov_to_buf_full() 61 size_t iov_memset(const struct iovec *iov, const unsigned int iov_cnt, in iov_memset() argument 66 for (i = 0, done = 0; (offset || done < bytes) && i < iov_cnt; i++) { in iov_memset() 79 size_t iov_size(const struct iovec *iov, const unsigned int iov_cnt) in iov_size() argument 85 for (i = 0; i < iov_cnt; i++) { in iov_size() 93 do_send_recv(int sockfd, int flags, struct iovec *iov, unsigned iov_cnt, in do_send_recv() argument 101 msg.msg_iovlen = iov_cnt; in do_send_recv() [all …]
|
H A D | crc32c.c | 116 uint32_t iov_crc32c(uint32_t crc, const struct iovec *iov, size_t iov_cnt) in iov_crc32c() argument 118 while (iov_cnt--) { in iov_crc32c()
|
H A D | osdep.c | 555 readv_writev(int fd, const struct iovec *iov, int iov_cnt, bool do_write) in readv_writev() argument 560 while (i < iov_cnt) { in readv_writev() 589 readv(int fd, const struct iovec *iov, int iov_cnt) in readv() argument 591 return readv_writev(fd, iov, iov_cnt, false); in readv() 595 writev(int fd, const struct iovec *iov, int iov_cnt) in writev() argument 597 return readv_writev(fd, iov, iov_cnt, true); in writev()
|
H A D | qemu-coroutine-io.c | 32 qemu_co_sendv_recvv(int sockfd, struct iovec *iov, unsigned iov_cnt, in qemu_co_sendv_recvv() argument 38 ret = iov_send_recv(sockfd, iov, iov_cnt, in qemu_co_sendv_recvv()
|
/qemu/include/hw/hyperv/ |
H A D | vmbus.h | 209 const struct iovec *iov, size_t iov_cnt); 218 unsigned iov_cnt, size_t len, size_t off); 224 unsigned iov_cnt, size_t accessed);
|
/qemu/hw/display/ |
H A D | virtio-gpu-udmabuf.c | 43 sizeof(struct udmabuf_create_item) * res->iov_cnt); in virtio_gpu_create_udmabuf() 45 for (i = 0; i < res->iov_cnt; i++) { in virtio_gpu_create_udmabuf() 60 list->count = res->iov_cnt; in virtio_gpu_create_udmabuf() 136 if (res->iov_cnt == 1 && in virtio_gpu_init_udmabuf()
|
H A D | virtio-gpu.c | 354 &res->iov_cnt); in virtio_gpu_resource_create_blob() 475 iov_to_buf(res->iov, res->iov_cnt, src_offset, in virtio_gpu_transfer_to_host_2d() 482 iov_to_buf(res->iov, res->iov_cnt, src_offset, in virtio_gpu_transfer_to_host_2d() 896 virtio_gpu_cleanup_mapping_iov(g, res->iov, res->iov_cnt); in virtio_gpu_cleanup_mapping() 898 res->iov_cnt = 0; in virtio_gpu_cleanup_mapping() 933 &res->addrs, &res->iov, &res->iov_cnt); in virtio_gpu_resource_attach_backing() 1239 qemu_put_be32(f, res->iov_cnt); in virtio_gpu_save() 1240 for (i = 0; i < res->iov_cnt; i++) { in virtio_gpu_save() 1257 for (i = 0; i < res->iov_cnt; i++) { in virtio_gpu_load_restore_mapping() 1270 res->iov_cnt = i; in virtio_gpu_load_restore_mapping() [all …]
|
H A D | virtio-gpu-rutabaga.c | 468 cmd, NULL, &res->iov, &res->iov_cnt); in rutabaga_cmd_attach_backing() 472 vecs.num_iovecs = res->iov_cnt; in rutabaga_cmd_attach_backing() 618 &res->iov, &res->iov_cnt); in rutabaga_cmd_resource_create_blob() 628 vecs.num_iovecs = res->iov_cnt; in rutabaga_cmd_resource_create_blob()
|
H A D | virtio-gpu-virgl.c | 703 &res->base.iov, &res->base.iov_cnt); in virgl_cmd_resource_create_blob() 717 virgl_args.num_iovs = res->base.iov_cnt; in virgl_cmd_resource_create_blob()
|
/qemu/hw/virtio/ |
H A D | virtio-iommu.c | 958 unsigned int iov_cnt, in virtio_iommu_iov_to_req() argument 961 size_t sz = iov_to_buf(iov, iov_cnt, 0, req, payload_sz); in virtio_iommu_iov_to_req() 972 unsigned int iov_cnt) \ 975 int ret = virtio_iommu_iov_to_req(iov, iov_cnt, &req, \ 988 unsigned int iov_cnt, in virtio_iommu_handle_req() 992 int ret = virtio_iommu_iov_to_req(iov, iov_cnt, &req, sizeof(req)); in virtio_iommu_handle_req() 1003 unsigned int iov_cnt; in virtio_iommu_handle_command() local 1024 iov_cnt = elem->out_num; in virtio_iommu_handle_command() 1026 sz = iov_to_buf(iov, iov_cnt, 0, &head, sizeof(head)); in virtio_iommu_handle_command() 1037 tail.status = virtio_iommu_handle_attach(s, iov, iov_cnt); in virtio_iommu_handle_command() [all …]
|
/qemu/hw/net/ |
H A D | virtio-net.c | 1020 struct iovec *iov, unsigned int iov_cnt) in virtio_net_handle_rx_mode() argument 1026 s = iov_to_buf(iov, iov_cnt, 0, &on, sizeof(on)); in virtio_net_handle_rx_mode() 1053 struct iovec *iov, unsigned int iov_cnt) in virtio_net_handle_offloads() argument 1063 s = iov_to_buf(iov, iov_cnt, 0, &offloads, sizeof(offloads)); in virtio_net_handle_offloads() 1098 struct iovec *iov, unsigned int iov_cnt) in virtio_net_handle_mac() argument 1106 if (iov_size(iov, iov_cnt) != sizeof(n->mac)) { in virtio_net_handle_mac() 1109 s = iov_to_buf(iov, iov_cnt, 0, &n->mac, sizeof(n->mac)); in virtio_net_handle_mac() 1127 s = iov_to_buf(iov, iov_cnt, 0, &mac_data.entries, in virtio_net_handle_mac() 1133 iov_discard_front(&iov, &iov_cnt, s); in virtio_net_handle_mac() 1135 if (mac_data.entries * ETH_ALEN > iov_size(iov, iov_cnt)) { in virtio_net_handle_mac() [all …]
|
H A D | net_tx_pkt.c | 576 void *opaque, const struct iovec *iov, int iov_cnt, in net_tx_pkt_sendv() argument 584 qemu_sendv_packet(nc, iov, iov_cnt); in net_tx_pkt_sendv()
|
/qemu/net/ |
H A D | checksum.c | 186 net_checksum_add_iov(const struct iovec *iov, const unsigned int iov_cnt, in net_checksum_add_iov() argument 194 for (i = 0; i < iov_cnt && size; i++) { in net_checksum_add_iov()
|
/qemu/contrib/vhost-user-blk/ |
H A D | vhost-user-blk.c | 58 const unsigned int iov_cnt) in vub_iov_size() argument 64 for (i = 0; i < iov_cnt; i++) { in vub_iov_size() 71 const unsigned int iov_cnt, void *buf) in vub_iov_to_buf() argument 77 for (i = 0; i < iov_cnt; i++) { in vub_iov_to_buf()
|
/qemu/include/net/ |
H A D | checksum.h | 63 const unsigned int iov_cnt,
|
/qemu/contrib/vhost-user-gpu/ |
H A D | vhost-user-gpu.c | 38 unsigned int iov_cnt; member 553 res->iov_cnt = ab.nr_entries; in vg_resource_attach_backing() 567 vg_cleanup_mapping_iov(g, res->iov, res->iov_cnt); in vg_cleanup_mapping() 569 res->iov_cnt = 0; in vg_cleanup_mapping() 640 iov_to_buf(res->iov, res->iov_cnt, src_offset, in vg_transfer_to_host_2d() 645 iov_to_buf(res->iov, res->iov_cnt, 0, in vg_transfer_to_host_2d()
|
/qemu/hw/misc/ |
H A D | aspeed_hace.c | 108 const unsigned int iov_cnt) in hace_iov_hexdump() argument 114 for (i = 0; i < iov_cnt; i++) { in hace_iov_hexdump() 124 iov_to_buf(iov, iov_cnt, 0, buf, size); in hace_iov_hexdump()
|
/qemu/hw/hyperv/ |
H A D | vmbus.c | 445 const struct iovec *iov, size_t iov_cnt) in vmbus_iov_to_gpadl() argument 455 for (i = 0; i < iov_cnt; i++) { in vmbus_iov_to_gpadl() 467 unsigned iov_cnt, size_t len, size_t off) in vmbus_map_sgl() argument 490 if (ret_cnt == iov_cnt) { in vmbus_map_sgl() 514 unsigned iov_cnt, size_t accessed) in vmbus_unmap_sgl() argument 519 for (i = 0; i < iov_cnt; i++) { in vmbus_unmap_sgl()
|
/qemu/include/hw/virtio/ |
H A D | virtio-gpu.h | 51 unsigned int iov_cnt; member
|