1ae1fae34SPekka Enberg #include "kvm/kvm.h" 272811558SPekka Enberg #include "kvm/read-write.h" 372811558SPekka Enberg #include "kvm/util.h" 424ed52dbSCyrill Gorcunov #include "kvm/strbuf.h" 54298ddadSSasha Levin #include "kvm/mutex.h" 64298ddadSSasha Levin #include "kvm/kvm-cpu.h" 74b1addaeSSasha Levin #include "kvm/kvm-ipc.h" 8eda03319SPekka Enberg 96c7d8514SPekka Enberg #include <linux/kvm.h> 10495fbd4eSSasha Levin #include <linux/err.h> 11f5ab5f67SPekka Enberg 124b1addaeSSasha Levin #include <sys/un.h> 13e2e876c2SMatt Evans #include <sys/stat.h> 144b1addaeSSasha Levin #include <sys/types.h> 154b1addaeSSasha Levin #include <sys/socket.h> 16ae1fae34SPekka Enberg #include <sys/ioctl.h> 171f9cff23SPekka Enberg #include <sys/mman.h> 182da26a59SPekka Enberg #include <stdbool.h> 1906e41eeaSPekka Enberg #include <limits.h> 20ce79f1caSPekka Enberg #include <signal.h> 21f5ab5f67SPekka Enberg #include <stdarg.h> 22b8f6afcdSPekka Enberg #include <stdlib.h> 23f5ab5f67SPekka Enberg #include <string.h> 240d1f17ecSPekka Enberg #include <unistd.h> 251f9cff23SPekka Enberg #include <stdio.h> 26b8f6afcdSPekka Enberg #include <fcntl.h> 27ce79f1caSPekka Enberg #include <time.h> 284298ddadSSasha Levin #include <sys/eventfd.h> 29c7828731SSasha Levin #include <asm/unistd.h> 3063bc8503SSasha Levin #include <dirent.h> 31b8f6afcdSPekka Enberg 32ae1fae34SPekka Enberg #define DEFINE_KVM_EXIT_REASON(reason) [reason] = #reason 330d1f17ecSPekka Enberg 34ae1fae34SPekka Enberg const char *kvm_exit_reasons[] = { 35ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_UNKNOWN), 36ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_EXCEPTION), 37ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_IO), 38ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_HYPERCALL), 39ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_DEBUG), 40ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_HLT), 41ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_MMIO), 42ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_IRQ_WINDOW_OPEN), 43ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_SHUTDOWN), 44ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_FAIL_ENTRY), 45ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_INTR), 46ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_SET_TPR), 47ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_TPR_ACCESS), 48ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_S390_SIEIC), 49ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_S390_RESET), 50ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_DCR), 51ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_NMI), 52ae1fae34SPekka Enberg DEFINE_KVM_EXIT_REASON(KVM_EXIT_INTERNAL_ERROR), 5363e158a0SMatt Evans #ifdef CONFIG_PPC64 5463e158a0SMatt Evans DEFINE_KVM_EXIT_REASON(KVM_EXIT_PAPR_HCALL), 5563e158a0SMatt Evans #endif 569b1fb1c3SPekka Enberg }; 579b1fb1c3SPekka Enberg 584298ddadSSasha Levin extern struct kvm *kvm; 59e2077857SMatt Evans extern struct kvm_cpu **kvm_cpus; 604298ddadSSasha Levin static int pause_event; 614298ddadSSasha Levin static DEFINE_MUTEX(pause_lock); 62af7b0868SMatt Evans extern struct kvm_ext kvm_req_ext[]; 634298ddadSSasha Levin 649667701cSPekka Enberg static char kvm_dir[PATH_MAX]; 659667701cSPekka Enberg 66495fbd4eSSasha Levin static int set_dir(const char *fmt, va_list args) 679667701cSPekka Enberg { 68dd188f9fSPekka Enberg char tmp[PATH_MAX]; 69dd188f9fSPekka Enberg 70dd188f9fSPekka Enberg vsnprintf(tmp, sizeof(tmp), fmt, args); 71dd188f9fSPekka Enberg 722bc995fbSPekka Enberg mkdir(tmp, 0777); 732bc995fbSPekka Enberg 74dd188f9fSPekka Enberg if (!realpath(tmp, kvm_dir)) 75495fbd4eSSasha Levin return -errno; 76f76a3285SPekka Enberg 77f76a3285SPekka Enberg strcat(kvm_dir, "/"); 78495fbd4eSSasha Levin 79495fbd4eSSasha Levin return 0; 809667701cSPekka Enberg } 819667701cSPekka Enberg 829667701cSPekka Enberg void kvm__set_dir(const char *fmt, ...) 839667701cSPekka Enberg { 849667701cSPekka Enberg va_list args; 859667701cSPekka Enberg 869667701cSPekka Enberg va_start(args, fmt); 879667701cSPekka Enberg set_dir(fmt, args); 889667701cSPekka Enberg va_end(args); 899667701cSPekka Enberg } 909667701cSPekka Enberg 919667701cSPekka Enberg const char *kvm__get_dir(void) 929667701cSPekka Enberg { 939667701cSPekka Enberg return kvm_dir; 949667701cSPekka Enberg } 959667701cSPekka Enberg 961d6fb3f2SSasha Levin bool kvm__supports_extension(struct kvm *kvm, unsigned int extension) 97b8f6afcdSPekka Enberg { 9828fa19c0SPekka Enberg int ret; 99b8f6afcdSPekka Enberg 10043835ac9SSasha Levin ret = ioctl(kvm->sys_fd, KVM_CHECK_EXTENSION, extension); 1014076b041SPekka Enberg if (ret < 0) 1024076b041SPekka Enberg return false; 1034076b041SPekka Enberg 1044076b041SPekka Enberg return ret; 1054076b041SPekka Enberg } 1064076b041SPekka Enberg 10743835ac9SSasha Levin static int kvm__check_extensions(struct kvm *kvm) 10855e19624SCyrill Gorcunov { 109495fbd4eSSasha Levin int i; 11055e19624SCyrill Gorcunov 111af7b0868SMatt Evans for (i = 0; ; i++) { 112af7b0868SMatt Evans if (!kvm_req_ext[i].name) 113af7b0868SMatt Evans break; 11443835ac9SSasha Levin if (!kvm__supports_extension(kvm, kvm_req_ext[i].code)) { 115599ed2a8SCyrill Gorcunov pr_err("Unsuppored KVM extension detected: %s", 11655e19624SCyrill Gorcunov kvm_req_ext[i].name); 117495fbd4eSSasha Levin return -i; 11855e19624SCyrill Gorcunov } 11955e19624SCyrill Gorcunov } 12055e19624SCyrill Gorcunov 12155e19624SCyrill Gorcunov return 0; 12255e19624SCyrill Gorcunov } 12355e19624SCyrill Gorcunov 12447621338SSasha Levin struct kvm *kvm__new(void) 1254076b041SPekka Enberg { 126495fbd4eSSasha Levin struct kvm *kvm = calloc(1, sizeof(*kvm)); 12743835ac9SSasha Levin if (!kvm) 128495fbd4eSSasha Levin return ERR_PTR(-ENOMEM); 1294076b041SPekka Enberg 130d648dbf5SCyrill Gorcunov kvm->sys_fd = -1; 131d648dbf5SCyrill Gorcunov kvm->vm_fd = -1; 132d648dbf5SCyrill Gorcunov 13343835ac9SSasha Levin return kvm; 1344076b041SPekka Enberg } 1354076b041SPekka Enberg 136495fbd4eSSasha Levin int kvm__exit(struct kvm *kvm) 1379ef4c68eSPekka Enberg { 138e56e2de7SLai Jiangshan kvm__arch_delete_ram(kvm); 13943835ac9SSasha Levin free(kvm); 140495fbd4eSSasha Levin 141495fbd4eSSasha Levin return 0; 1429ef4c68eSPekka Enberg } 14349a8afd1SSasha Levin core_exit(kvm__exit); 1449ef4c68eSPekka Enberg 14596feb589SPekka Enberg /* 14696feb589SPekka Enberg * Note: KVM_SET_USER_MEMORY_REGION assumes that we don't pass overlapping 14796feb589SPekka Enberg * memory regions to it. Therefore, be careful if you use this function for 14896feb589SPekka Enberg * registering memory regions for emulating hardware. 14996feb589SPekka Enberg */ 150495fbd4eSSasha Levin int kvm__register_mem(struct kvm *kvm, u64 guest_phys, u64 size, void *userspace_addr) 1514076b041SPekka Enberg { 1522b0e3342SPekka Enberg struct kvm_userspace_memory_region mem; 153839051d9SSasha Levin int ret; 154839051d9SSasha Levin 155839051d9SSasha Levin mem = (struct kvm_userspace_memory_region) { 15696feb589SPekka Enberg .slot = kvm->mem_slots++, 157874467f8SSasha Levin .guest_phys_addr = guest_phys, 158874467f8SSasha Levin .memory_size = size, 159c4acb611SIngo Molnar .userspace_addr = (unsigned long)userspace_addr, 160839051d9SSasha Levin }; 161839051d9SSasha Levin 162874467f8SSasha Levin ret = ioctl(kvm->vm_fd, KVM_SET_USER_MEMORY_REGION, &mem); 163839051d9SSasha Levin if (ret < 0) 164495fbd4eSSasha Levin return -errno; 165495fbd4eSSasha Levin 166495fbd4eSSasha Levin return 0; 167839051d9SSasha Levin } 168839051d9SSasha Levin 1698259b8ccSSasha Levin int kvm__recommended_cpus(struct kvm *kvm) 170384922b3SPekka Enberg { 171384922b3SPekka Enberg int ret; 172384922b3SPekka Enberg 17343835ac9SSasha Levin ret = ioctl(kvm->sys_fd, KVM_CHECK_EXTENSION, KVM_CAP_NR_VCPUS); 1748259b8ccSSasha Levin if (ret <= 0) 1753b9b691dSMatt Evans /* 1763b9b691dSMatt Evans * api.txt states that if KVM_CAP_NR_VCPUS does not exist, 1773b9b691dSMatt Evans * assume 4. 1783b9b691dSMatt Evans */ 1793b9b691dSMatt Evans return 4; 180384922b3SPekka Enberg 181384922b3SPekka Enberg return ret; 182384922b3SPekka Enberg } 183384922b3SPekka Enberg 1848259b8ccSSasha Levin /* 1858259b8ccSSasha Levin * The following hack should be removed once 'x86: Raise the hard 1868259b8ccSSasha Levin * VCPU count limit' makes it's way into the mainline. 1878259b8ccSSasha Levin */ 1888259b8ccSSasha Levin #ifndef KVM_CAP_MAX_VCPUS 1898259b8ccSSasha Levin #define KVM_CAP_MAX_VCPUS 66 1908259b8ccSSasha Levin #endif 1918259b8ccSSasha Levin 1928259b8ccSSasha Levin int kvm__max_cpus(struct kvm *kvm) 1938259b8ccSSasha Levin { 1948259b8ccSSasha Levin int ret; 1958259b8ccSSasha Levin 1968259b8ccSSasha Levin ret = ioctl(kvm->sys_fd, KVM_CHECK_EXTENSION, KVM_CAP_MAX_VCPUS); 1978259b8ccSSasha Levin if (ret <= 0) 1988259b8ccSSasha Levin ret = kvm__recommended_cpus(kvm); 1998259b8ccSSasha Levin 2008259b8ccSSasha Levin return ret; 2018259b8ccSSasha Levin } 2028259b8ccSSasha Levin 20347621338SSasha Levin int kvm__init(struct kvm *kvm) 204839051d9SSasha Levin { 2054076b041SPekka Enberg int ret; 2064076b041SPekka Enberg 207495fbd4eSSasha Levin if (!kvm__arch_cpu_supports_vm()) { 208495fbd4eSSasha Levin pr_err("Your CPU does not support hardware virtualization"); 2096fce7105SYang Bai ret = -ENOSYS; 2106fce7105SYang Bai goto err; 211495fbd4eSSasha Levin } 212c78b8713SAsias He 21347621338SSasha Levin kvm->sys_fd = open(kvm->cfg.dev, O_RDWR); 21443835ac9SSasha Levin if (kvm->sys_fd < 0) { 215d648dbf5SCyrill Gorcunov if (errno == ENOENT) 216495fbd4eSSasha Levin pr_err("'%s' not found. Please make sure your kernel has CONFIG_KVM " 21747621338SSasha Levin "enabled and that the KVM modules are loaded.", kvm->cfg.dev); 218d648dbf5SCyrill Gorcunov else if (errno == ENODEV) 219d648dbf5SCyrill Gorcunov pr_err("'%s' KVM driver not available.\n # (If the KVM " 220495fbd4eSSasha Levin "module is loaded then 'dmesg' may offer further clues " 22147621338SSasha Levin "about the failure.)", kvm->cfg.dev); 222d648dbf5SCyrill Gorcunov else 22347621338SSasha Levin pr_err("Could not open %s: ", kvm->cfg.dev); 224d648dbf5SCyrill Gorcunov 225495fbd4eSSasha Levin ret = -errno; 226d648dbf5SCyrill Gorcunov goto err_free; 2276d7c36ceSPekka Enberg } 228b8f6afcdSPekka Enberg 22943835ac9SSasha Levin ret = ioctl(kvm->sys_fd, KVM_GET_API_VERSION, 0); 230495fbd4eSSasha Levin if (ret != KVM_API_VERSION) { 231495fbd4eSSasha Levin pr_err("KVM_API_VERSION ioctl"); 232495fbd4eSSasha Levin ret = -errno; 233d648dbf5SCyrill Gorcunov goto err_sys_fd; 234495fbd4eSSasha Levin } 2356c7d8514SPekka Enberg 23643835ac9SSasha Levin kvm->vm_fd = ioctl(kvm->sys_fd, KVM_CREATE_VM, 0); 237495fbd4eSSasha Levin if (kvm->vm_fd < 0) { 238495fbd4eSSasha Levin ret = kvm->vm_fd; 239d648dbf5SCyrill Gorcunov goto err_sys_fd; 240495fbd4eSSasha Levin } 24128fa19c0SPekka Enberg 242495fbd4eSSasha Levin if (kvm__check_extensions(kvm)) { 243495fbd4eSSasha Levin pr_err("A required KVM extention is not supported by OS"); 244495fbd4eSSasha Levin ret = -ENOSYS; 2456fce7105SYang Bai goto err_vm_fd; 246495fbd4eSSasha Levin } 2479687927dSAsias He 24847621338SSasha Levin kvm__arch_init(kvm, kvm->cfg.hugetlbfs_path, kvm->cfg.ram_size); 2499687927dSAsias He 250abee258bSSasha Levin kvm__init_ram(kvm); 251abee258bSSasha Levin 252084a1356SSasha Levin if (!kvm->cfg.firmware_filename) { 253084a1356SSasha Levin if (!kvm__load_kernel(kvm, kvm->cfg.kernel_filename, 254084a1356SSasha Levin kvm->cfg.initrd_filename, kvm->cfg.real_cmdline, kvm->cfg.vidmode)) 255084a1356SSasha Levin die("unable to load kernel %s", kvm->cfg.kernel_filename); 256084a1356SSasha Levin } 257084a1356SSasha Levin 258084a1356SSasha Levin if (kvm->cfg.firmware_filename) { 259084a1356SSasha Levin if (!kvm__load_firmware(kvm, kvm->cfg.firmware_filename)) 260084a1356SSasha Levin die("unable to load firmware image %s: %s", kvm->cfg.firmware_filename, strerror(errno)); 261084a1356SSasha Levin } else { 262084a1356SSasha Levin ret = kvm__arch_setup_firmware(kvm); 263084a1356SSasha Levin if (ret < 0) 264084a1356SSasha Levin die("kvm__arch_setup_firmware() failed with error %d\n", ret); 265084a1356SSasha Levin } 266084a1356SSasha Levin 26747621338SSasha Levin return 0; 268d648dbf5SCyrill Gorcunov 2696fce7105SYang Bai err_vm_fd: 270495fbd4eSSasha Levin close(kvm->vm_fd); 271d648dbf5SCyrill Gorcunov err_sys_fd: 272495fbd4eSSasha Levin close(kvm->sys_fd); 273d648dbf5SCyrill Gorcunov err_free: 274495fbd4eSSasha Levin free(kvm); 2756fce7105SYang Bai err: 27647621338SSasha Levin return ret; 2774076b041SPekka Enberg } 27849a8afd1SSasha Levin core_init(kvm__init); 2794076b041SPekka Enberg 28072811558SPekka Enberg /* RFC 1952 */ 28172811558SPekka Enberg #define GZIP_ID1 0x1f 28272811558SPekka Enberg #define GZIP_ID2 0x8b 283663ce1dfSMatt Evans #define CPIO_MAGIC "0707" 284663ce1dfSMatt Evans /* initrd may be gzipped, or a plain cpio */ 28572811558SPekka Enberg static bool initrd_check(int fd) 28672811558SPekka Enberg { 287663ce1dfSMatt Evans unsigned char id[4]; 28872811558SPekka Enberg 28972811558SPekka Enberg if (read_in_full(fd, id, ARRAY_SIZE(id)) < 0) 29072811558SPekka Enberg return false; 29172811558SPekka Enberg 29272811558SPekka Enberg if (lseek(fd, 0, SEEK_SET) < 0) 29372811558SPekka Enberg die_perror("lseek"); 29472811558SPekka Enberg 295663ce1dfSMatt Evans return (id[0] == GZIP_ID1 && id[1] == GZIP_ID2) || 296663ce1dfSMatt Evans !memcmp(id, CPIO_MAGIC, 4); 29772811558SPekka Enberg } 29872811558SPekka Enberg 2996d1f350dSCyrill Gorcunov bool kvm__load_kernel(struct kvm *kvm, const char *kernel_filename, 30053861c74SJohn Floren const char *initrd_filename, const char *kernel_cmdline, u16 vidmode) 301ae1fae34SPekka Enberg { 3027fb218bdSPekka Enberg bool ret; 3032065a6f7SCyrill Gorcunov int fd_kernel = -1, fd_initrd = -1; 304ae1fae34SPekka Enberg 3052065a6f7SCyrill Gorcunov fd_kernel = open(kernel_filename, O_RDONLY); 3062065a6f7SCyrill Gorcunov if (fd_kernel < 0) 3070b62d2bbSPekka Enberg die("Unable to open kernel %s", kernel_filename); 308ae1fae34SPekka Enberg 3092065a6f7SCyrill Gorcunov if (initrd_filename) { 3102065a6f7SCyrill Gorcunov fd_initrd = open(initrd_filename, O_RDONLY); 3112065a6f7SCyrill Gorcunov if (fd_initrd < 0) 3120b62d2bbSPekka Enberg die("Unable to open initrd %s", initrd_filename); 31372811558SPekka Enberg 31472811558SPekka Enberg if (!initrd_check(fd_initrd)) 31572811558SPekka Enberg die("%s is not an initrd", initrd_filename); 3162065a6f7SCyrill Gorcunov } 3172065a6f7SCyrill Gorcunov 31853861c74SJohn Floren ret = load_bzimage(kvm, fd_kernel, fd_initrd, kernel_cmdline, vidmode); 31928972750SCyrill Gorcunov 320009b0758SPekka Enberg if (ret) 321009b0758SPekka Enberg goto found_kernel; 322ae1fae34SPekka Enberg 3234542f276SCyrill Gorcunov pr_warning("%s is not a bzImage. Trying to load it as a flat binary...", kernel_filename); 3240b62d2bbSPekka Enberg 325604dbd63SMatt Evans ret = load_flat_binary(kvm, fd_kernel, fd_initrd, kernel_cmdline); 326604dbd63SMatt Evans 327009b0758SPekka Enberg if (ret) 328009b0758SPekka Enberg goto found_kernel; 329009b0758SPekka Enberg 330604dbd63SMatt Evans if (initrd_filename) 331604dbd63SMatt Evans close(fd_initrd); 3325a6ac675SSasha Levin close(fd_kernel); 3335a6ac675SSasha Levin 334009b0758SPekka Enberg die("%s is not a valid bzImage or flat binary", kernel_filename); 335009b0758SPekka Enberg 336009b0758SPekka Enberg found_kernel: 337604dbd63SMatt Evans if (initrd_filename) 338604dbd63SMatt Evans close(fd_initrd); 3395a6ac675SSasha Levin close(fd_kernel); 3405a6ac675SSasha Levin 341ae1fae34SPekka Enberg return ret; 342ae1fae34SPekka Enberg } 343ae1fae34SPekka Enberg 344ce79f1caSPekka Enberg #define TIMER_INTERVAL_NS 1000000 /* 1 msec */ 345ce79f1caSPekka Enberg 346ce79f1caSPekka Enberg /* 347ce79f1caSPekka Enberg * This function sets up a timer that's used to inject interrupts from the 348ce79f1caSPekka Enberg * userspace hypervisor into the guest at periodical intervals. Please note 349ce79f1caSPekka Enberg * that clock interrupt, for example, is not handled here. 350ce79f1caSPekka Enberg */ 351b4532ca9SSasha Levin int kvm_timer__init(struct kvm *kvm) 352ce79f1caSPekka Enberg { 353ce79f1caSPekka Enberg struct itimerspec its; 354ce79f1caSPekka Enberg struct sigevent sev; 355b4532ca9SSasha Levin int r; 356ce79f1caSPekka Enberg 357ce79f1caSPekka Enberg memset(&sev, 0, sizeof(struct sigevent)); 358ce79f1caSPekka Enberg sev.sigev_value.sival_int = 0; 359c7828731SSasha Levin sev.sigev_notify = SIGEV_THREAD_ID; 360ce79f1caSPekka Enberg sev.sigev_signo = SIGALRM; 361*5002444cSSasha Levin sev.sigev_value.sival_ptr = kvm; 362c7828731SSasha Levin sev._sigev_un._tid = syscall(__NR_gettid); 363ce79f1caSPekka Enberg 364b4532ca9SSasha Levin r = timer_create(CLOCK_REALTIME, &sev, &kvm->timerid); 365b4532ca9SSasha Levin if (r < 0) 366b4532ca9SSasha Levin return r; 367ce79f1caSPekka Enberg 368ce79f1caSPekka Enberg its.it_value.tv_sec = TIMER_INTERVAL_NS / 1000000000; 369ce79f1caSPekka Enberg its.it_value.tv_nsec = TIMER_INTERVAL_NS % 1000000000; 370ce79f1caSPekka Enberg its.it_interval.tv_sec = its.it_value.tv_sec; 371ce79f1caSPekka Enberg its.it_interval.tv_nsec = its.it_value.tv_nsec; 372ce79f1caSPekka Enberg 373b4532ca9SSasha Levin r = timer_settime(kvm->timerid, 0, &its, NULL); 374b4532ca9SSasha Levin if (r < 0) { 375b4532ca9SSasha Levin timer_delete(kvm->timerid); 376b4532ca9SSasha Levin return r; 377ce79f1caSPekka Enberg } 378ce79f1caSPekka Enberg 379b4532ca9SSasha Levin return 0; 380b4532ca9SSasha Levin } 38149a8afd1SSasha Levin firmware_init(kvm_timer__init); 382b4532ca9SSasha Levin 383b4532ca9SSasha Levin int kvm_timer__exit(struct kvm *kvm) 384fbfe68b7SSasha Levin { 38543835ac9SSasha Levin if (kvm->timerid) 38643835ac9SSasha Levin if (timer_delete(kvm->timerid) < 0) 387fbfe68b7SSasha Levin die("timer_delete()"); 388fbfe68b7SSasha Levin 38943835ac9SSasha Levin kvm->timerid = 0; 390b4532ca9SSasha Levin 391b4532ca9SSasha Levin return 0; 392fbfe68b7SSasha Levin } 39349a8afd1SSasha Levin firmware_exit(kvm_timer__exit); 394fbfe68b7SSasha Levin 39543835ac9SSasha Levin void kvm__dump_mem(struct kvm *kvm, unsigned long addr, unsigned long size) 396090f898eSCyrill Gorcunov { 397090f898eSCyrill Gorcunov unsigned char *p; 398090f898eSCyrill Gorcunov unsigned long n; 399090f898eSCyrill Gorcunov 400090f898eSCyrill Gorcunov size &= ~7; /* mod 8 */ 401090f898eSCyrill Gorcunov if (!size) 402090f898eSCyrill Gorcunov return; 403090f898eSCyrill Gorcunov 40443835ac9SSasha Levin p = guest_flat_to_host(kvm, addr); 405090f898eSCyrill Gorcunov 40648cf3877SPekka Enberg for (n = 0; n < size; n += 8) { 40743835ac9SSasha Levin if (!host_ptr_in_ram(kvm, p + n)) 40848cf3877SPekka Enberg break; 40948cf3877SPekka Enberg 410090f898eSCyrill Gorcunov printf(" 0x%08lx: %02x %02x %02x %02x %02x %02x %02x %02x\n", 411090f898eSCyrill Gorcunov addr + n, p[n + 0], p[n + 1], p[n + 2], p[n + 3], 412090f898eSCyrill Gorcunov p[n + 4], p[n + 5], p[n + 6], p[n + 7]); 413090f898eSCyrill Gorcunov } 41448cf3877SPekka Enberg } 4154298ddadSSasha Levin 4164298ddadSSasha Levin void kvm__pause(void) 4174298ddadSSasha Levin { 4184298ddadSSasha Levin int i, paused_vcpus = 0; 4194298ddadSSasha Levin 4204298ddadSSasha Levin /* Check if the guest is running */ 421df4239fbSSasha Levin if (!kvm->cpus[0] || kvm->cpus[0]->thread == 0) 4224298ddadSSasha Levin return; 4234298ddadSSasha Levin 4244298ddadSSasha Levin mutex_lock(&pause_lock); 4254298ddadSSasha Levin 4264298ddadSSasha Levin pause_event = eventfd(0, 0); 4274298ddadSSasha Levin if (pause_event < 0) 4284298ddadSSasha Levin die("Failed creating pause notification event"); 4294298ddadSSasha Levin for (i = 0; i < kvm->nrcpus; i++) 430df4239fbSSasha Levin pthread_kill(kvm->cpus[i]->thread, SIGKVMPAUSE); 4314298ddadSSasha Levin 4324298ddadSSasha Levin while (paused_vcpus < kvm->nrcpus) { 4334298ddadSSasha Levin u64 cur_read; 4344298ddadSSasha Levin 4354298ddadSSasha Levin if (read(pause_event, &cur_read, sizeof(cur_read)) < 0) 4364298ddadSSasha Levin die("Failed reading pause event"); 4374298ddadSSasha Levin paused_vcpus += cur_read; 4384298ddadSSasha Levin } 4394298ddadSSasha Levin close(pause_event); 4404298ddadSSasha Levin } 4414298ddadSSasha Levin 4424298ddadSSasha Levin void kvm__continue(void) 4434298ddadSSasha Levin { 4444298ddadSSasha Levin /* Check if the guest is running */ 445df4239fbSSasha Levin if (!kvm->cpus[0] || kvm->cpus[0]->thread == 0) 4464298ddadSSasha Levin return; 4474298ddadSSasha Levin 4484298ddadSSasha Levin mutex_unlock(&pause_lock); 4494298ddadSSasha Levin } 4504298ddadSSasha Levin 4514298ddadSSasha Levin void kvm__notify_paused(void) 4524298ddadSSasha Levin { 4534298ddadSSasha Levin u64 p = 1; 4544298ddadSSasha Levin 4554298ddadSSasha Levin if (write(pause_event, &p, sizeof(p)) < 0) 4564298ddadSSasha Levin die("Failed notifying of paused VCPU."); 4574298ddadSSasha Levin 4584298ddadSSasha Levin mutex_lock(&pause_lock); 4594298ddadSSasha Levin mutex_unlock(&pause_lock); 4604298ddadSSasha Levin } 461