xref: /kvmtool/kvm-cpu.c (revision 43835ac90fb7c14c3946772d778da4b62a0bfd7c)
1 #include "kvm/kvm-cpu.h"
2 
3 #include "kvm/util.h"
4 #include "kvm/kvm.h"
5 
6 #include <asm/msr-index.h>
7 
8 #include <sys/ioctl.h>
9 #include <sys/mman.h>
10 #include <signal.h>
11 #include <stdlib.h>
12 #include <errno.h>
13 #include <stdio.h>
14 
15 static inline bool is_in_protected_mode(struct kvm_cpu *vcpu)
16 {
17 	return vcpu->sregs.cr0 & 0x01;
18 }
19 
20 static inline u64 ip_to_flat(struct kvm_cpu *vcpu, u64 ip)
21 {
22 	u64 cs;
23 
24 	/*
25 	 * NOTE! We should take code segment base address into account here.
26 	 * Luckily it's usually zero because Linux uses flat memory model.
27 	 */
28 	if (is_in_protected_mode(vcpu))
29 		return ip;
30 
31 	cs = vcpu->sregs.cs.selector;
32 
33 	return ip + (cs << 4);
34 }
35 
36 static inline u32 selector_to_base(u16 selector)
37 {
38 	/*
39 	 * KVM on Intel requires 'base' to be 'selector * 16' in real mode.
40 	 */
41 	return (u32)selector * 16;
42 }
43 
44 static struct kvm_cpu *kvm_cpu__new(struct kvm *kvm)
45 {
46 	struct kvm_cpu *vcpu;
47 
48 	vcpu		= calloc(1, sizeof *vcpu);
49 	if (!vcpu)
50 		return NULL;
51 
52 	vcpu->kvm	= kvm;
53 
54 	return vcpu;
55 }
56 
57 void kvm_cpu__delete(struct kvm_cpu *vcpu)
58 {
59 	if (vcpu->msrs)
60 		free(vcpu->msrs);
61 
62 	free(vcpu);
63 }
64 
65 struct kvm_cpu *kvm_cpu__init(struct kvm *kvm, unsigned long cpu_id)
66 {
67 	struct kvm_cpu *vcpu;
68 	int mmap_size;
69 
70 	vcpu		= kvm_cpu__new(kvm);
71 	if (!vcpu)
72 		return NULL;
73 
74 	vcpu->cpu_id	= cpu_id;
75 
76 	vcpu->vcpu_fd = ioctl(vcpu->kvm->vm_fd, KVM_CREATE_VCPU, cpu_id);
77 	if (vcpu->vcpu_fd < 0)
78 		die_perror("KVM_CREATE_VCPU ioctl");
79 
80 	mmap_size = ioctl(vcpu->kvm->sys_fd, KVM_GET_VCPU_MMAP_SIZE, 0);
81 	if (mmap_size < 0)
82 		die_perror("KVM_GET_VCPU_MMAP_SIZE ioctl");
83 
84 	vcpu->kvm_run = mmap(NULL, mmap_size, PROT_READ|PROT_WRITE, MAP_SHARED, vcpu->vcpu_fd, 0);
85 	if (vcpu->kvm_run == MAP_FAILED)
86 		die("unable to mmap vcpu fd");
87 
88 	return vcpu;
89 }
90 
91 void kvm_cpu__enable_singlestep(struct kvm_cpu *vcpu)
92 {
93 	struct kvm_guest_debug debug = {
94 		.control	= KVM_GUESTDBG_ENABLE | KVM_GUESTDBG_SINGLESTEP,
95 	};
96 
97 	if (ioctl(vcpu->vcpu_fd, KVM_SET_GUEST_DEBUG, &debug) < 0)
98 		warning("KVM_SET_GUEST_DEBUG failed");
99 }
100 
101 static struct kvm_msrs *kvm_msrs__new(size_t nmsrs)
102 {
103 	struct kvm_msrs *vcpu = calloc(1, sizeof(*vcpu) + (sizeof(struct kvm_msr_entry) * nmsrs));
104 
105 	if (!vcpu)
106 		die("out of memory");
107 
108 	return vcpu;
109 }
110 
111 #define KVM_MSR_ENTRY(_index, _data)	\
112 	(struct kvm_msr_entry) { .index = _index, .data = _data }
113 
114 static void kvm_cpu__setup_msrs(struct kvm_cpu *vcpu)
115 {
116 	unsigned long ndx = 0;
117 
118 	vcpu->msrs = kvm_msrs__new(100);
119 
120 	vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_IA32_SYSENTER_CS,	0x0);
121 	vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_IA32_SYSENTER_ESP,	0x0);
122 	vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_IA32_SYSENTER_EIP,	0x0);
123 #ifdef CONFIG_X86_64
124 	vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_STAR,			0x0);
125 	vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_CSTAR,			0x0);
126 	vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_KERNEL_GS_BASE,		0x0);
127 	vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_SYSCALL_MASK,		0x0);
128 	vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_LSTAR,			0x0);
129 #endif
130 	vcpu->msrs->entries[ndx++] = KVM_MSR_ENTRY(MSR_IA32_TSC,		0x0);
131 
132 	vcpu->msrs->nmsrs	= ndx;
133 
134 	if (ioctl(vcpu->vcpu_fd, KVM_SET_MSRS, vcpu->msrs) < 0)
135 		die_perror("KVM_SET_MSRS failed");
136 }
137 
138 static void kvm_cpu__setup_fpu(struct kvm_cpu *vcpu)
139 {
140 	vcpu->fpu = (struct kvm_fpu) {
141 		.fcw		= 0x37f,
142 		.mxcsr		= 0x1f80,
143 	};
144 
145 	if (ioctl(vcpu->vcpu_fd, KVM_SET_FPU, &vcpu->fpu) < 0)
146 		die_perror("KVM_SET_FPU failed");
147 }
148 
149 static void kvm_cpu__setup_regs(struct kvm_cpu *vcpu)
150 {
151 	vcpu->regs = (struct kvm_regs) {
152 		/* We start the guest in 16-bit real mode  */
153 		.rflags		= 0x0000000000000002ULL,
154 
155 		.rip		= vcpu->kvm->boot_ip,
156 		.rsp		= vcpu->kvm->boot_sp,
157 		.rbp		= vcpu->kvm->boot_sp,
158 	};
159 
160 	if (vcpu->regs.rip > USHRT_MAX)
161 		die("ip 0x%llx is too high for real mode", (u64) vcpu->regs.rip);
162 
163 	if (ioctl(vcpu->vcpu_fd, KVM_SET_REGS, &vcpu->regs) < 0)
164 		die_perror("KVM_SET_REGS failed");
165 }
166 
167 static void kvm_cpu__setup_sregs(struct kvm_cpu *vcpu)
168 {
169 
170 	if (ioctl(vcpu->vcpu_fd, KVM_GET_SREGS, &vcpu->sregs) < 0)
171 		die_perror("KVM_GET_SREGS failed");
172 
173 	vcpu->sregs.cs.selector	= vcpu->kvm->boot_selector;
174 	vcpu->sregs.cs.base	= selector_to_base(vcpu->kvm->boot_selector);
175 	vcpu->sregs.ss.selector	= vcpu->kvm->boot_selector;
176 	vcpu->sregs.ss.base	= selector_to_base(vcpu->kvm->boot_selector);
177 	vcpu->sregs.ds.selector	= vcpu->kvm->boot_selector;
178 	vcpu->sregs.ds.base	= selector_to_base(vcpu->kvm->boot_selector);
179 	vcpu->sregs.es.selector	= vcpu->kvm->boot_selector;
180 	vcpu->sregs.es.base	= selector_to_base(vcpu->kvm->boot_selector);
181 	vcpu->sregs.fs.selector	= vcpu->kvm->boot_selector;
182 	vcpu->sregs.fs.base	= selector_to_base(vcpu->kvm->boot_selector);
183 	vcpu->sregs.gs.selector	= vcpu->kvm->boot_selector;
184 	vcpu->sregs.gs.base	= selector_to_base(vcpu->kvm->boot_selector);
185 
186 	if (ioctl(vcpu->vcpu_fd, KVM_SET_SREGS, &vcpu->sregs) < 0)
187 		die_perror("KVM_SET_SREGS failed");
188 }
189 
190 /**
191  * kvm_cpu__reset_vcpu - reset virtual CPU to a known state
192  */
193 void kvm_cpu__reset_vcpu(struct kvm_cpu *vcpu)
194 {
195 	kvm_cpu__setup_sregs(vcpu);
196 	kvm_cpu__setup_regs(vcpu);
197 	kvm_cpu__setup_fpu(vcpu);
198 	kvm_cpu__setup_msrs(vcpu);
199 }
200 
201 static void print_dtable(const char *name, struct kvm_dtable *dtable)
202 {
203 	printf(" %s                 %016llx  %08hx\n",
204 		name, (u64) dtable->base, (u16) dtable->limit);
205 }
206 
207 static void print_segment(const char *name, struct kvm_segment *seg)
208 {
209 	printf(" %s       %04hx      %016llx  %08x  %02hhx    %x %x   %x  %x %x %x %x\n",
210 		name, (u16) seg->selector, (u64) seg->base, (u32) seg->limit,
211 		(u8) seg->type, seg->present, seg->dpl, seg->db, seg->s, seg->l, seg->g, seg->avl);
212 }
213 
214 void kvm_cpu__show_registers(struct kvm_cpu *vcpu)
215 {
216 	unsigned long cr0, cr2, cr3;
217 	unsigned long cr4, cr8;
218 	unsigned long rax, rbx, rcx;
219 	unsigned long rdx, rsi, rdi;
220 	unsigned long rbp,  r8,  r9;
221 	unsigned long r10, r11, r12;
222 	unsigned long r13, r14, r15;
223 	unsigned long rip, rsp;
224 	struct kvm_sregs sregs;
225 	unsigned long rflags;
226 	struct kvm_regs regs;
227 	int i;
228 
229 	if (ioctl(vcpu->vcpu_fd, KVM_GET_REGS, &regs) < 0)
230 		die("KVM_GET_REGS failed");
231 
232 	rflags = regs.rflags;
233 
234 	rip = regs.rip; rsp = regs.rsp;
235 	rax = regs.rax; rbx = regs.rbx; rcx = regs.rcx;
236 	rdx = regs.rdx; rsi = regs.rsi; rdi = regs.rdi;
237 	rbp = regs.rbp; r8  = regs.r8;  r9  = regs.r9;
238 	r10 = regs.r10; r11 = regs.r11; r12 = regs.r12;
239 	r13 = regs.r13; r14 = regs.r14; r15 = regs.r15;
240 
241 	printf("\n Registers:\n");
242 	printf(  " ----------\n");
243 	printf(" rip: %016lx   rsp: %016lx flags: %016lx\n", rip, rsp, rflags);
244 	printf(" rax: %016lx   rbx: %016lx   rcx: %016lx\n", rax, rbx, rcx);
245 	printf(" rdx: %016lx   rsi: %016lx   rdi: %016lx\n", rdx, rsi, rdi);
246 	printf(" rbp: %016lx    r8: %016lx    r9: %016lx\n", rbp, r8,  r9);
247 	printf(" r10: %016lx   r11: %016lx   r12: %016lx\n", r10, r11, r12);
248 	printf(" r13: %016lx   r14: %016lx   r15: %016lx\n", r13, r14, r15);
249 
250 	if (ioctl(vcpu->vcpu_fd, KVM_GET_SREGS, &sregs) < 0)
251 		die("KVM_GET_REGS failed");
252 
253 	cr0 = sregs.cr0; cr2 = sregs.cr2; cr3 = sregs.cr3;
254 	cr4 = sregs.cr4; cr8 = sregs.cr8;
255 
256 	printf(" cr0: %016lx   cr2: %016lx   cr3: %016lx\n", cr0, cr2, cr3);
257 	printf(" cr4: %016lx   cr8: %016lx\n", cr4, cr8);
258 	printf("\n Segment registers:\n");
259 	printf(  " ------------------\n");
260 	printf(" register  selector  base              limit     type  p dpl db s l g avl\n");
261 	print_segment("cs ", &sregs.cs);
262 	print_segment("ss ", &sregs.ss);
263 	print_segment("ds ", &sregs.ds);
264 	print_segment("es ", &sregs.es);
265 	print_segment("fs ", &sregs.fs);
266 	print_segment("gs ", &sregs.gs);
267 	print_segment("tr ", &sregs.tr);
268 	print_segment("ldt", &sregs.ldt);
269 	print_dtable("gdt", &sregs.gdt);
270 	print_dtable("idt", &sregs.idt);
271 
272 	printf("\n APIC:\n");
273 	printf(  " -----\n");
274 	printf(" efer: %016llx  apic base: %016llx  nmi: %s\n",
275 		(u64) sregs.efer, (u64) sregs.apic_base,
276 		(vcpu->kvm->nmi_disabled ? "disabled" : "enabled"));
277 
278 	printf("\n Interrupt bitmap:\n");
279 	printf(  " -----------------\n");
280 	for (i = 0; i < (KVM_NR_INTERRUPTS + 63) / 64; i++)
281 		printf(" %016llx", (u64) sregs.interrupt_bitmap[i]);
282 	printf("\n");
283 }
284 
285 void kvm_cpu__show_code(struct kvm_cpu *vcpu)
286 {
287 	unsigned int code_bytes = 64;
288 	unsigned int code_prologue = code_bytes * 43 / 64;
289 	unsigned int code_len = code_bytes;
290 	unsigned char c;
291 	unsigned int i;
292 	u8 *ip;
293 
294 	if (ioctl(vcpu->vcpu_fd, KVM_GET_REGS, &vcpu->regs) < 0)
295 		die("KVM_GET_REGS failed");
296 
297 	if (ioctl(vcpu->vcpu_fd, KVM_GET_SREGS, &vcpu->sregs) < 0)
298 		die("KVM_GET_SREGS failed");
299 
300 	ip = guest_flat_to_host(vcpu->kvm, ip_to_flat(vcpu, vcpu->regs.rip) - code_prologue);
301 
302 	printf("\n Code:\n");
303 	printf(  " -----\n");
304 
305 	for (i = 0; i < code_len; i++, ip++) {
306 		if (!host_ptr_in_ram(vcpu->kvm, ip))
307 			break;
308 
309 		c = *ip;
310 
311 		if (ip == guest_flat_to_host(vcpu->kvm, ip_to_flat(vcpu, vcpu->regs.rip)))
312 			printf(" <%02x>", c);
313 		else
314 			printf(" %02x", c);
315 	}
316 
317 	printf("\n");
318 
319 	printf("\n Stack:\n");
320 	printf(  " ------\n");
321 	kvm__dump_mem(vcpu->kvm, vcpu->regs.rsp, 32);
322 }
323 
324 void kvm_cpu__show_page_tables(struct kvm_cpu *vcpu)
325 {
326 	u64 *pte1;
327 	u64 *pte2;
328 	u64 *pte3;
329 	u64 *pte4;
330 
331 	if (!is_in_protected_mode(vcpu))
332 		return;
333 
334 	if (ioctl(vcpu->vcpu_fd, KVM_GET_SREGS, &vcpu->sregs) < 0)
335 		die("KVM_GET_SREGS failed");
336 
337 	pte4	= guest_flat_to_host(vcpu->kvm, vcpu->sregs.cr3);
338 	if (!host_ptr_in_ram(vcpu->kvm, pte4))
339 		return;
340 
341 	pte3	= guest_flat_to_host(vcpu->kvm, (*pte4 & ~0xfff));
342 	if (!host_ptr_in_ram(vcpu->kvm, pte3))
343 		return;
344 
345 	pte2	= guest_flat_to_host(vcpu->kvm, (*pte3 & ~0xfff));
346 	if (!host_ptr_in_ram(vcpu->kvm, pte2))
347 		return;
348 
349 	pte1	= guest_flat_to_host(vcpu->kvm, (*pte2 & ~0xfff));
350 	if (!host_ptr_in_ram(vcpu->kvm, pte1))
351 		return;
352 
353 	printf("Page Tables:\n");
354 	if (*pte2 & (1 << 7))
355 		printf(" pte4: %016llx   pte3: %016llx"
356 			"   pte2: %016llx\n",
357 			*pte4, *pte3, *pte2);
358 	else
359 		printf(" pte4: %016llx  pte3: %016llx   pte2: %016"
360 			"llx   pte1: %016llx\n",
361 			*pte4, *pte3, *pte2, *pte1);
362 }
363 
364 void kvm_cpu__run(struct kvm_cpu *vcpu)
365 {
366 	int err;
367 
368 	err = ioctl(vcpu->vcpu_fd, KVM_RUN, 0);
369 	if (err && (errno != EINTR && errno != EAGAIN))
370 		die_perror("KVM_RUN failed");
371 }
372 
373 int kvm_cpu__start(struct kvm_cpu *cpu)
374 {
375 	sigset_t sigset;
376 
377 	sigemptyset(&sigset);
378 	sigaddset(&sigset, SIGALRM);
379 
380 	pthread_sigmask(SIG_BLOCK, &sigset, NULL);
381 
382 	kvm_cpu__setup_cpuid(cpu);
383 	kvm_cpu__reset_vcpu(cpu);
384 
385 	for (;;) {
386 		kvm_cpu__run(cpu);
387 
388 		switch (cpu->kvm_run->exit_reason) {
389 		case KVM_EXIT_UNKNOWN:
390 			break;
391 		case KVM_EXIT_DEBUG:
392 			kvm_cpu__show_registers(cpu);
393 			kvm_cpu__show_code(cpu);
394 			break;
395 		case KVM_EXIT_IO: {
396 			bool ret;
397 
398 			ret = kvm__emulate_io(cpu->kvm,
399 					cpu->kvm_run->io.port,
400 					(u8 *)cpu->kvm_run +
401 					cpu->kvm_run->io.data_offset,
402 					cpu->kvm_run->io.direction,
403 					cpu->kvm_run->io.size,
404 					cpu->kvm_run->io.count);
405 
406 			if (!ret)
407 				goto panic_kvm;
408 			break;
409 		}
410 		case KVM_EXIT_MMIO: {
411 			bool ret;
412 
413 			ret = kvm__emulate_mmio(cpu->kvm,
414 					cpu->kvm_run->mmio.phys_addr,
415 					cpu->kvm_run->mmio.data,
416 					cpu->kvm_run->mmio.len,
417 					cpu->kvm_run->mmio.is_write);
418 
419 			if (!ret)
420 				goto panic_kvm;
421 			break;
422 		}
423 		case KVM_EXIT_INTR:
424 			break;
425 		case KVM_EXIT_SHUTDOWN:
426 			goto exit_kvm;
427 		default:
428 			goto panic_kvm;
429 		}
430 	}
431 
432 exit_kvm:
433 	return 0;
434 
435 panic_kvm:
436 	return 1;
437 }
438