Searched refs:chunk_end (Results 1 – 3 of 3) sorted by relevance
432 size_t chunk_end = tcg_ctx->gen_insn_end_off[insn]; in tb_gen_code() local433 if (chunk_end > chunk_start) { in tb_gen_code()437 chunk_end - chunk_start); in tb_gen_code()438 chunk_start = chunk_end; in tb_gen_code()
1106 uint8_t *chunk_start, uint8_t *chunk_end) in qemu_rdma_register_and_get_keys() argument1129 uint64_t len = chunk_end - chunk_start; in qemu_rdma_register_and_get_keys()1845 uint8_t *chunk_start, *chunk_end; in qemu_rdma_write_one() local1882 chunk_end = ram_chunk_end(block, chunk + chunks); in qemu_rdma_write_one()1970 chunk_start, chunk_end)) { in qemu_rdma_write_one()1989 chunk_start, chunk_end)) { in qemu_rdma_write_one()2001 chunk_start, chunk_end)) { in qemu_rdma_write_one()2104 uint8_t *chunk_end; in qemu_rdma_buffer_mergeable() local2116 chunk_end = ram_chunk_end(block, rdma->current_chunk); in qemu_rdma_buffer_mergeable()2137 if ((host_addr + len) > chunk_end) { in qemu_rdma_buffer_mergeable()[all …]
781 unsigned long chunk_end = QEMU_ALIGN_UP(start + npages, chunk_pages); in migration_clear_memory_region_dirty_bitmap_range() local787 for (i = chunk_start; i < chunk_end; i += chunk_pages) { in migration_clear_memory_region_dirty_bitmap_range()