1 #include "kvm/kvm-cpu.h" 2 3 #include "kvm/util.h" 4 #include "kvm/kvm.h" 5 6 #include <asm/msr-index.h> 7 8 #include <sys/ioctl.h> 9 #include <sys/mman.h> 10 #include <signal.h> 11 #include <stdlib.h> 12 #include <errno.h> 13 #include <stdio.h> 14 15 static inline bool is_in_protected_mode(struct kvm_cpu *vcpu) 16 { 17 return vcpu->sregs.cr0 & 0x01; 18 } 19 20 static inline u64 ip_to_flat(struct kvm_cpu *vcpu, u64 ip) 21 { 22 u64 cs; 23 24 /* 25 * NOTE! We should take code segment base address into account here. 26 * Luckily it's usually zero because Linux uses flat memory model. 27 */ 28 if (is_in_protected_mode(vcpu)) 29 return ip; 30 31 cs = vcpu->sregs.cs.selector; 32 33 return ip + (cs << 4); 34 } 35 36 static inline u32 selector_to_base(u16 selector) 37 { 38 /* 39 * KVM on Intel requires 'base' to be 'selector * 16' in real mode. 40 */ 41 return (u32)selector * 16; 42 } 43 44 static struct kvm_cpu *kvm_cpu__new(struct kvm *kvm) 45 { 46 struct kvm_cpu *vcpu; 47 48 vcpu = calloc(1, sizeof *vcpu); 49 if (!vcpu) 50 return NULL; 51 52 vcpu->kvm = kvm; 53 54 return vcpu; 55 } 56 57 void kvm_cpu__delete(struct kvm_cpu *vcpu) 58 { 59 if (vcpu->msrs) 60 free(vcpu->msrs); 61 62 free(vcpu); 63 } 64 65 struct kvm_cpu *kvm_cpu__init(struct kvm *kvm, unsigned long cpu_id) 66 { 67 struct kvm_cpu *vcpu; 68 int mmap_size; 69 70 vcpu = kvm_cpu__new(kvm); 71 if (!vcpu) 72 return NULL; 73 74 vcpu->cpu_id = cpu_id; 75 76 vcpu->vcpu_fd = ioctl(vcpu->kvm->vm_fd, KVM_CREATE_VCPU, cpu_id); 77 if (vcpu->vcpu_fd < 0) 78 die_perror("KVM_CREATE_VCPU ioctl"); 79 80 mmap_size = ioctl(vcpu->kvm->sys_fd, KVM_GET_VCPU_MMAP_SIZE, 0); 81 if (mmap_size < 0) 82 die_perror("KVM_GET_VCPU_MMAP_SIZE ioctl"); 83 84 vcpu->kvm_run = mmap(NULL, mmap_size, PROT_READ|PROT_WRITE, MAP_SHARED, vcpu->vcpu_fd, 0); 85 if (vcpu->kvm_run == MAP_FAILED) 86 die("unable to mmap vcpu fd"); 87 88 return vcpu; 89 } 90 91 void kvm_cpu__enable_singlestep(struct kvm_cpu *vcpu) 92 { 93 struct kvm_guest_debug debug = { 94 .control = KVM_GUESTDBG_ENABLE | KVM_GUESTDBG_SINGLESTEP, 95 }; 96 97 if (ioctl(vcpu->vcpu_fd, KVM_SET_GUEST_DEBUG, &debug) < 0) 98 warning("KVM_SET_GUEST_DEBUG failed"); 99 } 100 101 static struct kvm_msrs *kvm_msrs__new(size_t nmsrs) 102 { 103 struct kvm_msrs *vcpu = calloc(1, sizeof(*vcpu) + (sizeof(struct kvm_msr_entry) * nmsrs)); 104 105 if (!vcpu) 106 die("out of memory"); 107 108 return vcpu; 109 } 110 111 #define KVM_MSR_ENTRY(_index, _data) \ 112 (struct kvm_msr_entry) { .index = _index, .data = _data } 113 114 static void kvm_cpu__setup_msrs(struct kvm_cpu *vcpu) 115 { 116 unsigned long ndx = 0; 117 118 vcpu->msrs = kvm_msrs__new(100); 119 120 vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_IA32_SYSENTER_CS, 0x0); 121 vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_IA32_SYSENTER_ESP, 0x0); 122 vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_IA32_SYSENTER_EIP, 0x0); 123 #ifdef CONFIG_X86_64 124 vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_STAR, 0x0); 125 vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_CSTAR, 0x0); 126 vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_KERNEL_GS_BASE, 0x0); 127 vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_SYSCALL_MASK, 0x0); 128 vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_LSTAR, 0x0); 129 #endif 130 vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_IA32_TSC, 0x0); 131 132 vcpu->msrs->nmsrs = ndx; 133 134 if (ioctl(vcpu->vcpu_fd, KVM_SET_MSRS, vcpu->msrs) < 0) 135 die_perror("KVM_SET_MSRS failed"); 136 } 137 138 static void kvm_cpu__setup_fpu(struct kvm_cpu *vcpu) 139 { 140 vcpu->fpu = (struct kvm_fpu) { 141 .fcw = 0x37f, 142 .mxcsr = 0x1f80, 143 }; 144 145 if (ioctl(vcpu->vcpu_fd, KVM_SET_FPU, &vcpu->fpu) < 0) 146 die_perror("KVM_SET_FPU failed"); 147 } 148 149 static void kvm_cpu__setup_regs(struct kvm_cpu *vcpu) 150 { 151 vcpu->regs = (struct kvm_regs) { 152 /* We start the guest in 16-bit real mode */ 153 .rflags = 0x0000000000000002ULL, 154 155 .rip = vcpu->kvm->boot_ip, 156 .rsp = vcpu->kvm->boot_sp, 157 .rbp = vcpu->kvm->boot_sp, 158 }; 159 160 if (vcpu->regs.rip > USHRT_MAX) 161 die("ip 0x%llx is too high for real mode", (u64) vcpu->regs.rip); 162 163 if (ioctl(vcpu->vcpu_fd, KVM_SET_REGS, &vcpu->regs) < 0) 164 die_perror("KVM_SET_REGS failed"); 165 } 166 167 static void kvm_cpu__setup_sregs(struct kvm_cpu *vcpu) 168 { 169 170 if (ioctl(vcpu->vcpu_fd, KVM_GET_SREGS, &vcpu->sregs) < 0) 171 die_perror("KVM_GET_SREGS failed"); 172 173 vcpu->sregs.cs.selector = vcpu->kvm->boot_selector; 174 vcpu->sregs.cs.base = selector_to_base(vcpu->kvm->boot_selector); 175 vcpu->sregs.ss.selector = vcpu->kvm->boot_selector; 176 vcpu->sregs.ss.base = selector_to_base(vcpu->kvm->boot_selector); 177 vcpu->sregs.ds.selector = vcpu->kvm->boot_selector; 178 vcpu->sregs.ds.base = selector_to_base(vcpu->kvm->boot_selector); 179 vcpu->sregs.es.selector = vcpu->kvm->boot_selector; 180 vcpu->sregs.es.base = selector_to_base(vcpu->kvm->boot_selector); 181 vcpu->sregs.fs.selector = vcpu->kvm->boot_selector; 182 vcpu->sregs.fs.base = selector_to_base(vcpu->kvm->boot_selector); 183 vcpu->sregs.gs.selector = vcpu->kvm->boot_selector; 184 vcpu->sregs.gs.base = selector_to_base(vcpu->kvm->boot_selector); 185 186 if (ioctl(vcpu->vcpu_fd, KVM_SET_SREGS, &vcpu->sregs) < 0) 187 die_perror("KVM_SET_SREGS failed"); 188 } 189 190 /** 191 * kvm_cpu__reset_vcpu - reset virtual CPU to a known state 192 */ 193 void kvm_cpu__reset_vcpu(struct kvm_cpu *vcpu) 194 { 195 kvm_cpu__setup_sregs(vcpu); 196 kvm_cpu__setup_regs(vcpu); 197 kvm_cpu__setup_fpu(vcpu); 198 kvm_cpu__setup_msrs(vcpu); 199 } 200 201 static void print_dtable(const char *name, struct kvm_dtable *dtable) 202 { 203 printf(" %s %016llx %08hx\n", 204 name, (u64) dtable->base, (u16) dtable->limit); 205 } 206 207 static void print_segment(const char *name, struct kvm_segment *seg) 208 { 209 printf(" %s %04hx %016llx %08x %02hhx %x %x %x %x %x %x %x\n", 210 name, (u16) seg->selector, (u64) seg->base, (u32) seg->limit, 211 (u8) seg->type, seg->present, seg->dpl, seg->db, seg->s, seg->l, seg->g, seg->avl); 212 } 213 214 void kvm_cpu__show_registers(struct kvm_cpu *vcpu) 215 { 216 unsigned long cr0, cr2, cr3; 217 unsigned long cr4, cr8; 218 unsigned long rax, rbx, rcx; 219 unsigned long rdx, rsi, rdi; 220 unsigned long rbp, r8, r9; 221 unsigned long r10, r11, r12; 222 unsigned long r13, r14, r15; 223 unsigned long rip, rsp; 224 struct kvm_sregs sregs; 225 unsigned long rflags; 226 struct kvm_regs regs; 227 int i; 228 229 if (ioctl(vcpu->vcpu_fd, KVM_GET_REGS, ®s) < 0) 230 die("KVM_GET_REGS failed"); 231 232 rflags = regs.rflags; 233 234 rip = regs.rip; rsp = regs.rsp; 235 rax = regs.rax; rbx = regs.rbx; rcx = regs.rcx; 236 rdx = regs.rdx; rsi = regs.rsi; rdi = regs.rdi; 237 rbp = regs.rbp; r8 = regs.r8; r9 = regs.r9; 238 r10 = regs.r10; r11 = regs.r11; r12 = regs.r12; 239 r13 = regs.r13; r14 = regs.r14; r15 = regs.r15; 240 241 printf("\n Registers:\n"); 242 printf( " ----------\n"); 243 printf(" rip: %016lx rsp: %016lx flags: %016lx\n", rip, rsp, rflags); 244 printf(" rax: %016lx rbx: %016lx rcx: %016lx\n", rax, rbx, rcx); 245 printf(" rdx: %016lx rsi: %016lx rdi: %016lx\n", rdx, rsi, rdi); 246 printf(" rbp: %016lx r8: %016lx r9: %016lx\n", rbp, r8, r9); 247 printf(" r10: %016lx r11: %016lx r12: %016lx\n", r10, r11, r12); 248 printf(" r13: %016lx r14: %016lx r15: %016lx\n", r13, r14, r15); 249 250 if (ioctl(vcpu->vcpu_fd, KVM_GET_SREGS, &sregs) < 0) 251 die("KVM_GET_REGS failed"); 252 253 cr0 = sregs.cr0; cr2 = sregs.cr2; cr3 = sregs.cr3; 254 cr4 = sregs.cr4; cr8 = sregs.cr8; 255 256 printf(" cr0: %016lx cr2: %016lx cr3: %016lx\n", cr0, cr2, cr3); 257 printf(" cr4: %016lx cr8: %016lx\n", cr4, cr8); 258 printf("\n Segment registers:\n"); 259 printf( " ------------------\n"); 260 printf(" register selector base limit type p dpl db s l g avl\n"); 261 print_segment("cs ", &sregs.cs); 262 print_segment("ss ", &sregs.ss); 263 print_segment("ds ", &sregs.ds); 264 print_segment("es ", &sregs.es); 265 print_segment("fs ", &sregs.fs); 266 print_segment("gs ", &sregs.gs); 267 print_segment("tr ", &sregs.tr); 268 print_segment("ldt", &sregs.ldt); 269 print_dtable("gdt", &sregs.gdt); 270 print_dtable("idt", &sregs.idt); 271 272 printf("\n APIC:\n"); 273 printf( " -----\n"); 274 printf(" efer: %016llx apic base: %016llx nmi: %s\n", 275 (u64) sregs.efer, (u64) sregs.apic_base, 276 (vcpu->kvm->nmi_disabled ? "disabled" : "enabled")); 277 278 printf("\n Interrupt bitmap:\n"); 279 printf( " -----------------\n"); 280 for (i = 0; i < (KVM_NR_INTERRUPTS + 63) / 64; i++) 281 printf(" %016llx", (u64) sregs.interrupt_bitmap[i]); 282 printf("\n"); 283 } 284 285 void kvm_cpu__show_code(struct kvm_cpu *vcpu) 286 { 287 unsigned int code_bytes = 64; 288 unsigned int code_prologue = code_bytes * 43 / 64; 289 unsigned int code_len = code_bytes; 290 unsigned char c; 291 unsigned int i; 292 u8 *ip; 293 294 if (ioctl(vcpu->vcpu_fd, KVM_GET_REGS, &vcpu->regs) < 0) 295 die("KVM_GET_REGS failed"); 296 297 if (ioctl(vcpu->vcpu_fd, KVM_GET_SREGS, &vcpu->sregs) < 0) 298 die("KVM_GET_SREGS failed"); 299 300 ip = guest_flat_to_host(vcpu->kvm, ip_to_flat(vcpu, vcpu->regs.rip) - code_prologue); 301 302 printf("\n Code:\n"); 303 printf( " -----\n"); 304 305 for (i = 0; i < code_len; i++, ip++) { 306 if (!host_ptr_in_ram(vcpu->kvm, ip)) 307 break; 308 309 c = *ip; 310 311 if (ip == guest_flat_to_host(vcpu->kvm, ip_to_flat(vcpu, vcpu->regs.rip))) 312 printf(" <%02x>", c); 313 else 314 printf(" %02x", c); 315 } 316 317 printf("\n"); 318 319 printf("\n Stack:\n"); 320 printf( " ------\n"); 321 kvm__dump_mem(vcpu->kvm, vcpu->regs.rsp, 32); 322 } 323 324 void kvm_cpu__show_page_tables(struct kvm_cpu *vcpu) 325 { 326 u64 *pte1; 327 u64 *pte2; 328 u64 *pte3; 329 u64 *pte4; 330 331 if (!is_in_protected_mode(vcpu)) 332 return; 333 334 if (ioctl(vcpu->vcpu_fd, KVM_GET_SREGS, &vcpu->sregs) < 0) 335 die("KVM_GET_SREGS failed"); 336 337 pte4 = guest_flat_to_host(vcpu->kvm, vcpu->sregs.cr3); 338 if (!host_ptr_in_ram(vcpu->kvm, pte4)) 339 return; 340 341 pte3 = guest_flat_to_host(vcpu->kvm, (*pte4 & ~0xfff)); 342 if (!host_ptr_in_ram(vcpu->kvm, pte3)) 343 return; 344 345 pte2 = guest_flat_to_host(vcpu->kvm, (*pte3 & ~0xfff)); 346 if (!host_ptr_in_ram(vcpu->kvm, pte2)) 347 return; 348 349 pte1 = guest_flat_to_host(vcpu->kvm, (*pte2 & ~0xfff)); 350 if (!host_ptr_in_ram(vcpu->kvm, pte1)) 351 return; 352 353 printf("Page Tables:\n"); 354 if (*pte2 & (1 << 7)) 355 printf(" pte4: %016llx pte3: %016llx" 356 " pte2: %016llx\n", 357 *pte4, *pte3, *pte2); 358 else 359 printf(" pte4: %016llx pte3: %016llx pte2: %016" 360 "llx pte1: %016llx\n", 361 *pte4, *pte3, *pte2, *pte1); 362 } 363 364 void kvm_cpu__run(struct kvm_cpu *vcpu) 365 { 366 int err; 367 368 err = ioctl(vcpu->vcpu_fd, KVM_RUN, 0); 369 if (err && (errno != EINTR && errno != EAGAIN)) 370 die_perror("KVM_RUN failed"); 371 } 372 373 int kvm_cpu__start(struct kvm_cpu *cpu) 374 { 375 sigset_t sigset; 376 377 sigemptyset(&sigset); 378 sigaddset(&sigset, SIGALRM); 379 380 pthread_sigmask(SIG_BLOCK, &sigset, NULL); 381 382 kvm_cpu__setup_cpuid(cpu); 383 kvm_cpu__reset_vcpu(cpu); 384 385 for (;;) { 386 kvm_cpu__run(cpu); 387 388 switch (cpu->kvm_run->exit_reason) { 389 case KVM_EXIT_UNKNOWN: 390 break; 391 case KVM_EXIT_DEBUG: 392 kvm_cpu__show_registers(cpu); 393 kvm_cpu__show_code(cpu); 394 break; 395 case KVM_EXIT_IO: { 396 bool ret; 397 398 ret = kvm__emulate_io(cpu->kvm, 399 cpu->kvm_run->io.port, 400 (u8 *)cpu->kvm_run + 401 cpu->kvm_run->io.data_offset, 402 cpu->kvm_run->io.direction, 403 cpu->kvm_run->io.size, 404 cpu->kvm_run->io.count); 405 406 if (!ret) 407 goto panic_kvm; 408 break; 409 } 410 case KVM_EXIT_MMIO: { 411 bool ret; 412 413 ret = kvm__emulate_mmio(cpu->kvm, 414 cpu->kvm_run->mmio.phys_addr, 415 cpu->kvm_run->mmio.data, 416 cpu->kvm_run->mmio.len, 417 cpu->kvm_run->mmio.is_write); 418 419 if (!ret) 420 goto panic_kvm; 421 break; 422 } 423 case KVM_EXIT_INTR: 424 break; 425 case KVM_EXIT_SHUTDOWN: 426 goto exit_kvm; 427 default: 428 goto panic_kvm; 429 } 430 } 431 432 exit_kvm: 433 return 0; 434 435 panic_kvm: 436 return 1; 437 } 438