/qemu/hw/scsi/ |
H A D | megasas.c | 73 size_t iov_size; member 285 size_t iov_size = 0; in megasas_map_sgl() local 311 iov_size += (size_t)iov_size_p; in megasas_map_sgl() 313 if (cmd->iov_size > iov_size) { in megasas_map_sgl() 314 trace_megasas_iovec_overflow(cmd->index, iov_size, cmd->iov_size); in megasas_map_sgl() 316 } else if (cmd->iov_size < iov_size) { in megasas_map_sgl() 317 trace_megasas_iovec_underflow(cmd->index, iov_size, cmd->iov_size); in megasas_map_sgl() 599 cmd->iov_size = 0; in megasas_complete_command() 698 dma_addr_t iov_pa, iov_size; in megasas_map_dcmd() local 705 cmd->iov_size = 0; in megasas_map_dcmd() [all …]
|
/qemu/tests/unit/ |
H A D | test-iov.c | 70 sz = iov_size(iov, niov); in test_to_from_buf_1() 168 sz = iov_size(iov, niov); in test_io() 287 size = iov_size(iov, iov_cnt); in test_discard_front() 297 size = iov_size(iov, iov_cnt); in test_discard_front() 368 size = iov_size(iov, iov_cnt); in test_discard_front_undo() 380 size = iov_size(iov, iov_cnt); in test_discard_front_undo() 443 size = iov_size(iov, iov_cnt); in test_discard_back() 452 size = iov_size(iov, iov_cnt); in test_discard_back() 515 size = iov_size(iov, iov_cnt); in test_discard_back_undo() 526 size = iov_size(iov, iov_cnt); in test_discard_back_undo()
|
/qemu/net/ |
H A D | filter-replay.c | 43 return iov_size(iov, iovcnt); in filter_replay_receive_iov() 49 return iov_size(iov, iovcnt); in filter_replay_receive_iov()
|
H A D | eth.c | 76 size_t size = iov_size(l2hdr_iov, iovcnt); in eth_get_l3_proto() 140 size_t input_size = iov_size(iov, iovcnt); in eth_get_protocols() 382 size_t input_size = iov_size(pkt, pkt_frags); in _eth_get_rss_ex_dst_addr() 413 size_t input_size = iov_size(pkt, pkt_frags); in _eth_get_rss_ex_src_addr() 459 size_t input_size = iov_size(pkt, pkt_frags); in eth_parse_ipv6_hdr()
|
H A D | filter-buffer.c | 84 return iov_size(iov, iovcnt); in filter_buffer_receive_iov()
|
H A D | dump.c | 70 size_t size = iov_size(iov, cnt) - offset; in dump_receive_iov()
|
H A D | hub.c | 67 ssize_t len = iov_size(iov, iovcnt); in net_hub_receive_iov()
|
H A D | vhost-vdpa.c | 698 size_t data_size = iov_size(data_sg, data_num), cmd_size; in vhost_vdpa_net_load_cmd() 708 iov_size(out_cursor, 1) < cmd_size) { in vhost_vdpa_net_load_cmd() 1356 if (iov_size(elem->out_sg, elem->out_num) != cursor) { in vhost_vdpa_net_excessive_mac_filter_cvq_add() 1495 iov_size(elem->out_sg, elem->out_num) > out.iov_len)) { in vhost_vdpa_net_handle_ctrl_avail()
|
H A D | filter-rewriter.c | 263 ssize_t size = iov_size(iov, iovcnt); in colo_rewriter_receive_iov()
|
H A D | filter.c | 119 return iov_size(iov, iovcnt); in qemu_netfilter_pass_to_next()
|
H A D | filter-mirror.c | 110 ssize_t size = iov_size(iov, iovcnt); in filter_send()
|
H A D | net.c | 785 offset = iov_size(iov, iovcnt); in nc_sendv_compat() 814 return iov_size(iov, iovcnt); in qemu_deliver_packet_iov() 860 size_t size = iov_size(iov, iovcnt); in qemu_sendv_packet_async()
|
/qemu/block/export/ |
H A D | virtio-blk-handler.c | 66 if (unlikely(iov_size(iov, iovcnt) > sizeof(desc))) { in virtio_blk_discard_write_zeroes() 163 in_len = iov_size(in_iov, in_num); in virtio_blk_process_req() 220 MIN(iov_size(in_iov, in_num), in virtio_blk_process_req()
|
/qemu/replay/ |
H A D | replay-net.c | 58 event->data = g_malloc(iov_size(iov, iovcnt)); in replay_net_packet_event() 59 event->size = iov_size(iov, iovcnt); in replay_net_packet_event()
|
/qemu/hw/9pfs/ |
H A D | virtio-9p-device.c | 62 if (iov_size(elem->in_sg, elem->in_num) < 7) { in handle_9p_output() 161 size_t buf_size = iov_size(elem->in_sg, elem->in_num); in virtio_init_in_iov_from_pdu() 181 size_t buf_size = iov_size(elem->out_sg, elem->out_num); in virtio_init_out_iov_from_pdu()
|
/qemu/crypto/ |
H A D | hash-afalg.c | 173 0, iov_size(iov, niov), true); in qcrypto_afalg_send_to_kernel() 210 0, iov_size(&outv, 1), false); in qcrypto_afalg_recv_from_kernel()
|
/qemu/hw/audio/ |
H A D | virtio-snd.c | 185 if (iov_size(cmd->elem->in_sg, cmd->elem->in_num) < in virtio_snd_handle_pcm_info() 191 iov_size(cmd->elem->in_sg, cmd->elem->in_num), in virtio_snd_handle_pcm_info() 894 size = iov_size(elem->out_sg, elem->out_num) - msg_sz; in virtio_snd_handle_tx_xfer() 974 size = iov_size(elem->in_sg, elem->in_num) - in virtio_snd_handle_rx_xfer() 1255 max_size = iov_size(buffer->elem->in_sg, buffer->elem->in_num); in virtio_snd_pcm_in_cb()
|
/qemu/io/ |
H A D | channel-null.c | 95 return iov_size(iov, niov); in qio_channel_null_writev()
|
H A D | channel.c | 155 0, iov_size(iov, niov)); in qio_channel_readv_full_all_eof() 262 0, iov_size(iov, niov)); in qio_channel_writev_full_all()
|
H A D | channel-websock.c | 623 assert(size <= iov_size(iov, niov)); in qio_channel_websock_encode() 1136 ssize_t want = iov_size(iov, niov); in qio_channel_websock_writev()
|
/qemu/fsdev/ |
H A D | qemu-fsdev-throttle.c | 108 throttle_account(&fst->ts, direction, iov_size(iov, iovcnt)); in fsdev_co_throttle_request()
|
/qemu/include/qemu/ |
H A D | iov.h | 22 size_t iov_size(const struct iovec *iov, const unsigned int iov_cnt);
|
/qemu/hw/net/rocker/ |
H A D | rocker_tlv.h | 155 size_t len = iov_size(iov, iovcnt); in rocker_tlv_put_iov()
|
/qemu/hw/block/ |
H A D | virtio-blk.c | 763 int64_t len = iov_size(out_iov, out_num); in virtio_blk_handle_zone_append() 822 req->in_len = iov_size(in_iov, in_num); in virtio_blk_handle_request() 906 MIN(iov_size(in_iov, in_num), in virtio_blk_handle_request() 930 size_t out_len = iov_size(out_iov, out_num); in virtio_blk_handle_request()
|
/qemu/hw/virtio/ |
H A D | virtio-iommu.c | 1016 if (iov_size(elem->in_sg, elem->in_num) < sizeof(tail) || in virtio_iommu_handle_command() 1017 iov_size(elem->out_sg, elem->out_num) < sizeof(head)) { in virtio_iommu_handle_command() 1120 if (iov_size(elem->in_sg, elem->in_num) < sizeof(fault)) { in virtio_iommu_report_fault()
|