/qemu/include/hw/ppc/ |
H A D | xive2_regs.h | 111 #define xive2_end_is_valid(end) (be32_to_cpu((end)->w0) & END2_W0_VALID) argument 112 #define xive2_end_is_enqueue(end) (be32_to_cpu((end)->w0) & END2_W0_ENQUEUE) argument 113 #define xive2_end_is_notify(end) \ argument 114 (be32_to_cpu((end)->w0) & END2_W0_UCOND_NOTIFY) 115 #define xive2_end_is_backlog(end) (be32_to_cpu((end)->w0) & END2_W0_BACKLOG) argument 116 #define xive2_end_is_precluded_escalation(end) \ argument 117 (be32_to_cpu((end)->w0) & END2_W0_PRECL_ESC_CTL) 118 #define xive2_end_is_escalate(end) \ argument 119 (be32_to_cpu((end)->w0) & END2_W0_ESCALATE_CTL) 120 #define xive2_end_is_uncond_escalation(end) \ argument [all …]
|
H A D | xive_regs.h | 272 #define xive_end_is_valid(end) (be32_to_cpu((end)->w0) & END_W0_VALID) argument 273 #define xive_end_is_enqueue(end) (be32_to_cpu((end)->w0) & END_W0_ENQUEUE) argument 274 #define xive_end_is_notify(end) (be32_to_cpu((end)->w0) & END_W0_UCOND_NOTIFY) argument 275 #define xive_end_is_backlog(end) (be32_to_cpu((end)->w0) & END_W0_BACKLOG) argument 276 #define xive_end_is_escalate(end) (be32_to_cpu((end)->w0) & END_W0_ESCALATE_CTL) argument 277 #define xive_end_is_uncond_escalation(end) \ argument 278 (be32_to_cpu((end)->w0) & END_W0_UNCOND_ESCALATE) 279 #define xive_end_is_silent_escalation(end) \ argument 280 (be32_to_cpu((end)->w0) & END_W0_SILENT_ESCALATE) 281 #define xive_end_is_firmware(end) \ argument [all …]
|
/qemu/scripts/ |
H A D | vmstate-static-checker.py | 185 print("Section \"" + sec + "\",", end=' ') 186 print("Description " + "\"" + desc + "\":", end=' ') 187 print("expected field \"" + s_item["field"] + "\",", end=' ') 206 print("Section \"" + sec + "\",", end=' ') 207 print("Description \"" + desc + "\":", end=' ') 208 print("unused size mismatch near \"", end=' ') 220 print("Section \"" + sec + "\",", end=' ') 221 print("Description \"" + desc + "\":", end=' ') 222 print("unused size mismatch near \"", end=' ') 273 print("Section \"" + sec + "\",", end=' ') [all …]
|
/qemu/hw/nvram/ |
H A D | chrp_nvram.c | 52 int end; in chrp_nvram_create_system_partition() local 62 end = sizeof(ChrpNvramPartHdr); in chrp_nvram_create_system_partition() 64 end = chrp_nvram_set_var(data, end, prom_envs[i], max_len - end); in chrp_nvram_create_system_partition() 65 if (end == -1) { in chrp_nvram_create_system_partition() 71 data[end++] = '\0'; in chrp_nvram_create_system_partition() 73 end = (end + 15) & ~15; in chrp_nvram_create_system_partition() 76 if (end < min_len) { in chrp_nvram_create_system_partition() 77 end = min_len; in chrp_nvram_create_system_partition() 79 chrp_nvram_finish_partition(part_header, end); in chrp_nvram_create_system_partition() 81 return end; in chrp_nvram_create_system_partition()
|
/qemu/hw/intc/ |
H A D | xive2.c | 191 void xive2_end_queue_pic_print_info(Xive2End *end, uint32_t width, GString *buf) in xive2_end_queue_pic_print_info() argument 193 uint64_t qaddr_base = xive2_end_qaddr(end); in xive2_end_queue_pic_print_info() 194 uint32_t qsize = xive_get_field32(END2_W3_QSIZE, end->w3); in xive2_end_queue_pic_print_info() 195 uint32_t qindex = xive_get_field32(END2_W1_PAGE_OFF, end->w1); in xive2_end_queue_pic_print_info() 221 void xive2_end_pic_print_info(Xive2End *end, uint32_t end_idx, GString *buf) in xive2_end_pic_print_info() argument 223 uint64_t qaddr_base = xive2_end_qaddr(end); in xive2_end_pic_print_info() 224 uint32_t qindex = xive_get_field32(END2_W1_PAGE_OFF, end->w1); in xive2_end_pic_print_info() 225 uint32_t qgen = xive_get_field32(END2_W1_GENERATION, end->w1); in xive2_end_pic_print_info() 226 uint32_t qsize = xive_get_field32(END2_W3_QSIZE, end->w3); in xive2_end_pic_print_info() 229 uint32_t nvx_blk = xive_get_field32(END2_W6_VP_BLOCK, end->w6); in xive2_end_pic_print_info() [all …]
|
H A D | xive.c | 1441 void xive_end_queue_pic_print_info(XiveEND *end, uint32_t width, GString *buf) in xive_end_queue_pic_print_info() argument 1443 uint64_t qaddr_base = xive_end_qaddr(end); in xive_end_queue_pic_print_info() 1444 uint32_t qsize = xive_get_field32(END_W0_QSIZE, end->w0); in xive_end_queue_pic_print_info() 1445 uint32_t qindex = xive_get_field32(END_W1_PAGE_OFF, end->w1); in xive_end_queue_pic_print_info() 1471 void xive_end_pic_print_info(XiveEND *end, uint32_t end_idx, GString *buf) in xive_end_pic_print_info() argument 1473 uint64_t qaddr_base = xive_end_qaddr(end); in xive_end_pic_print_info() 1474 uint32_t qindex = xive_get_field32(END_W1_PAGE_OFF, end->w1); in xive_end_pic_print_info() 1475 uint32_t qgen = xive_get_field32(END_W1_GENERATION, end->w1); in xive_end_pic_print_info() 1476 uint32_t qsize = xive_get_field32(END_W0_QSIZE, end->w0); in xive_end_pic_print_info() 1479 uint32_t nvt_blk = xive_get_field32(END_W6_NVT_BLOCK, end->w6); in xive_end_pic_print_info() [all …]
|
H A D | spapr_xive.c | 132 static void spapr_xive_end_pic_print_info(SpaprXive *xive, XiveEND *end, in spapr_xive_end_pic_print_info() argument 135 uint64_t qaddr_base = xive_end_qaddr(end); in spapr_xive_end_pic_print_info() 136 uint32_t qindex = xive_get_field32(END_W1_PAGE_OFF, end->w1); in spapr_xive_end_pic_print_info() 137 uint32_t qgen = xive_get_field32(END_W1_GENERATION, end->w1); in spapr_xive_end_pic_print_info() 138 uint32_t qsize = xive_get_field32(END_W0_QSIZE, end->w0); in spapr_xive_end_pic_print_info() 140 uint32_t nvt = xive_get_field32(END_W6_NVT_INDEX, end->w6); in spapr_xive_end_pic_print_info() 141 uint8_t priority = xive_get_field32(END_W7_F0_PRIORITY, end->w7); in spapr_xive_end_pic_print_info() 147 xive_end_queue_pic_print_info(end, 6, buf); in spapr_xive_end_pic_print_info() 192 XiveEND *end; in spapr_xive_pic_print_info() local 195 end = &xive->endt[end_idx]; in spapr_xive_pic_print_info() [all …]
|
/qemu/tests/qtest/ |
H A D | pnv-xive2-common.c | 163 Xive2End end; in set_end() local 171 memset(&end, 0, sizeof(end)); in set_end() 172 end.w0 = xive_set_field32(END2_W0_VALID, 0, 1); in set_end() 173 end.w0 = xive_set_field32(END2_W0_ENQUEUE, end.w0, 1); in set_end() 174 end.w0 = xive_set_field32(END2_W0_UCOND_NOTIFY, end.w0, 1); in set_end() 175 end.w0 = xive_set_field32(END2_W0_BACKLOG, end.w0, 1); in set_end() 177 end.w1 = xive_set_field32(END2_W1_GENERATION, 0, 1); in set_end() 179 end.w2 = cpu_to_be32(queue_hi); in set_end() 181 end.w3 = cpu_to_be32(queue_lo); in set_end() 182 end.w3 = xive_set_field32(END2_W3_QSIZE, end.w3, queue_size); in set_end() [all …]
|
H A D | i440fx-test.c | 159 static gboolean verify_area(uint32_t start, uint32_t end, uint8_t value) in verify_area() argument 161 uint32_t size = end - start + 1; in verify_area() 183 static void write_area(uint32_t start, uint32_t end, uint8_t value) in write_area() argument 185 uint32_t size = end - start + 1; in write_area() 203 uint32_t end; in test_i440fx_pam() member 226 if (pam_area[i].start == pam_area[i].end) { in test_i440fx_pam() 231 pam_area[i].start, pam_area[i].end); in test_i440fx_pam() 235 g_assert(verify_area(pam_area[i].start, pam_area[i].end, 0)); in test_i440fx_pam() 240 write_area(pam_area[i].start, pam_area[i].end, 0x42); in test_i440fx_pam() 248 g_assert(!verify_area(pam_area[i].start, pam_area[i].end, 0x42)); in test_i440fx_pam() [all …]
|
/qemu/bsd-user/ |
H A D | mmap.c | 68 abi_ulong end, host_start, host_end, addr; in target_mprotect() local 79 end = start + len; in target_mprotect() 80 if (end < start) in target_mprotect() 88 host_end = HOST_PAGE_ALIGN(end); in target_mprotect() 96 for (addr = end; addr < host_end; addr += TARGET_PAGE_SIZE) { in target_mprotect() 99 end = host_end; in target_mprotect() 107 if (end < host_end) { in target_mprotect() 109 for (addr = end; addr < host_end; addr += TARGET_PAGE_SIZE) { in target_mprotect() 189 abi_ulong start, abi_ulong end, in mmap_frag() argument 202 if (addr < start || addr >= end) in mmap_frag() [all …]
|
/qemu/docs/interop/ |
H A D | vhost-user.rst | 26 The protocol defines 2 sides of the communication, *front-end* and 27 *back-end*. The *front-end* is the application that shares its virtqueues, in 28 our case QEMU. The *back-end* is the consumer of the virtqueues. 30 In the current implementation QEMU is the *front-end*, and the *back-end* 33 or a block device back-end processing read & write to a virtual 34 disk. In order to facilitate interoperability between various back-end 38 The *front-end* and *back-end* can be either a client (i.e. connecting) or 80 - Bit 2 is the reply flag - needs to be sent on each reply from the back-end 124 back-end will process. This is a free-running index that is not 140 back-end will process. This is a free-running index that is not [all …]
|
/qemu/target/i386/ |
H A D | monitor.c | 252 hwaddr end, int prot) in mem_print() argument 261 addr_canonical(env, end), in mem_print() 262 addr_canonical(env, end - *pstart), in mem_print() 268 *pstart = end; in mem_print() 280 hwaddr start, end; in mem_info_32() local 288 end = l1 << 22; in mem_info_32() 292 mem_print(mon, env, &start, &last_prot, end, prot); in mem_info_32() 297 end = (l1 << 22) + (l2 << 12); in mem_info_32() 304 mem_print(mon, env, &start, &last_prot, end, prot); in mem_info_32() 309 mem_print(mon, env, &start, &last_prot, end, prot); in mem_info_32() [all …]
|
/qemu/tests/migration-stress/ |
H A D | stress.c | 57 char *start, *end; in get_command_arg_str() local 85 end = strstr(start, " "); in get_command_arg_str() 86 if (!end) in get_command_arg_str() 87 end = strstr(start, "\n"); in get_command_arg_str() 89 if (end == start) { in get_command_arg_str() 95 if (end) in get_command_arg_str() 96 *val = g_strndup(start, end - start); in get_command_arg_str() 107 char *end; in get_command_arg_ull() local 114 *val = strtoll(valstr, &end, 10); in get_command_arg_ull() 115 if (errno || *end) { in get_command_arg_ull() [all …]
|
/qemu/ui/ |
H A D | vnc-enc-zywrle.h | 316 int8_t *end; in wavelet_level() local 320 end = px0 + (size >> (l + 1)) * s; in wavelet_level() 324 while (px0 < end) { in wavelet_level() 448 int *end; in wavelet() local 452 end = buf + height * width; in wavelet() 454 while (top < end) { in wavelet() 459 end = buf + width; in wavelet() 461 while (top < end) { in wavelet() 576 end = ph + h * w; \ 577 while (ph < end) { \ [all …]
|
/qemu/hw/hyperv/ |
H A D | hv-balloon-page_range_tree.h | 25 uint64_t end = MIN(endr, start); in page_range_part_before() local 28 if (end > out->start) { in page_range_part_before() 29 out->count = end - out->start; in page_range_part_before() 40 uint64_t end = range->start + range->count; in page_range_part_after() local 44 if (end > out->start) { in page_range_part_after() 45 out->count = end - out->start; in page_range_part_after() 57 uint64_t end = MIN(end1, end2); in page_range_intersect() local 60 out->count = out->start < end ? end - out->start : 0; in page_range_intersect()
|
/qemu/tests/multiboot/ |
H A D | mmap.out | 18 mmap end: 0x9090 19 real mmap end: 0x9090 35 mmap end: 0x9078 36 real mmap end: 0x9078 53 mmap end: 0x9090 54 real mmap end: 0x9090 72 mmap end: 0x90a8 73 real mmap end: 0x90a8 91 mmap end: 0x90a8 92 real mmap end: 0x90a8
|
/qemu/scripts/codeconverter/codeconverter/ |
H A D | patching.py | 32 end: int 65 def end(self) -> int: member in FileMatch 66 return self.match.end() 115 return Patch(self.start(), self.end(), replacement) 117 def make_subpatch(self, start: int, end: int, replacement: str) -> 'Patch': 118 return Patch(self.start() + start, self.start() + end, replacement) 126 return Patch(self.end(), self.end(), s) 144 return other.start() >= self.start() and other.end() <= self.end() 148 end = self.file.line_col(self.end() - 1) 152 end.line, end.col, self.group(0)[:100]) [all …]
|
/qemu/system/ |
H A D | memory-internal.h | 42 static inline bool devend_big_endian(enum device_endian end) in devend_big_endian() argument 44 if (end == DEVICE_NATIVE_ENDIAN) { in devend_big_endian() 47 return end == DEVICE_BIG_ENDIAN; in devend_big_endian() 51 static inline MemOp devend_memop(enum device_endian end) in devend_memop() argument 53 return devend_big_endian(end) ? MO_BE : MO_LE; in devend_memop()
|
/qemu/tests/tcg/alpha/system/ |
H A D | boot.S | 104 .end _start 114 .end _exit 131 .end log_exit 162 .end __sys_outc 192 .end __sys_outs 268 .end __divqu 289 .end __remqu 329 .end __divqs 368 .end __remqs 395 .end __divlu [all …]
|
/qemu/qapi/ |
H A D | string-input-visitor.c | 137 int64_t start, end; in try_parse_int64_list_entry() local 143 end = start; in try_parse_int64_list_entry() 154 if (qemu_strtoi64(endptr + 1, &endptr, 0, &end)) { in try_parse_int64_list_entry() 157 if (start > end || end - start >= RANGE_MAX_ELEMENTS) { in try_parse_int64_list_entry() 178 siv->rangeEnd.i64 = end; in try_parse_int64_list_entry() 227 uint64_t start, end; in try_parse_uint64_list_entry() local 233 end = start; in try_parse_uint64_list_entry() 244 if (qemu_strtou64(endptr + 1, &endptr, 0, &end)) { in try_parse_uint64_list_entry() 247 if (start > end || end - start >= RANGE_MAX_ELEMENTS) { in try_parse_uint64_list_entry() 268 siv->rangeEnd.u64 = end; in try_parse_uint64_list_entry()
|
/qemu/tcg/ |
H A D | region.c | 331 void *start, *end; in tcg_region_bounds() local 334 end = start + region.size; in tcg_region_bounds() 341 end = region.start_aligned + region.total_size; in tcg_region_bounds() 345 *pend = end; in tcg_region_bounds() 350 void *start, *end; in tcg_region_assign() local 352 tcg_region_bounds(curr_region, &start, &end); in tcg_region_assign() 356 s->code_gen_buffer_size = end - start; in tcg_region_assign() 357 s->code_gen_highwater = end - TCG_HIGHWATER; in tcg_region_assign() 505 void *buf, *end; in alloc_code_gen_buffer() local 515 end = static_code_gen_buffer + sizeof(static_code_gen_buffer); in alloc_code_gen_buffer() [all …]
|
/qemu/util/ |
H A D | hbitmap.c | 200 uint64_t end; in hbitmap_next_dirty() local 208 end = count > hb->orig_size - start ? hb->orig_size : start + count; in hbitmap_next_dirty() 213 if (first_dirty_off < 0 || first_dirty_off >= end) { in hbitmap_next_dirty() 273 bool hbitmap_next_dirty_area(const HBitmap *hb, int64_t start, int64_t end, in hbitmap_next_dirty_area() argument 279 assert(start >= 0 && end >= 0 && max_dirty_count > 0); in hbitmap_next_dirty_area() 281 end = MIN(end, hb->orig_size); in hbitmap_next_dirty_area() 282 if (start >= end) { in hbitmap_next_dirty_area() 286 start = hbitmap_next_dirty(hb, start, end - start); in hbitmap_next_dirty_area() 291 end = start + MIN(end - start, max_dirty_count); in hbitmap_next_dirty_area() 293 next_zero = hbitmap_next_zero(hb, start, end - start); in hbitmap_next_dirty_area() [all …]
|
/qemu/hw/vfio/ |
H A D | spapr.c | 67 hwaddr end; in vfio_prereg_listener_region_add() local 90 end = section->offset_within_address_space + int128_get64(section->size); in vfio_prereg_listener_region_add() 91 if (gpa >= end) { in vfio_prereg_listener_region_add() 98 reg.size = end - gpa; in vfio_prereg_listener_region_add() 126 hwaddr end; in vfio_prereg_listener_region_del() local 149 end = section->offset_within_address_space + int128_get64(section->size); in vfio_prereg_listener_region_del() 150 if (gpa >= end) { in vfio_prereg_listener_region_del() 155 reg.size = end - gpa; in vfio_prereg_listener_region_del() 206 hwaddr iova, hwaddr end) in vfio_find_hostwin() argument 212 if (hostwin->min_iova <= iova && end <= hostwin->max_iova) { in vfio_find_hostwin() [all …]
|
/qemu/tests/tcg/mips/user/isa/mips64r6/bit-count/ |
H A D | test_mips64r6_clz.c | 39 struct timeval start, end; in main() local 137 gettimeofday(&end, NULL); in main() 139 elapsed_time = (end.tv_sec - start.tv_sec) * 1000.0; in main() 140 elapsed_time += (end.tv_usec - start.tv_usec) / 1000.0; in main()
|
H A D | test_mips64r6_dclo.c | 39 struct timeval start, end; in main() local 137 gettimeofday(&end, NULL); in main() 139 elapsed_time = (end.tv_sec - start.tv_sec) * 1000.0; in main() 140 elapsed_time += (end.tv_usec - start.tv_usec) / 1000.0; in main()
|